var/home/core/zuul-output/0000755000175000017500000000000015073371503014531 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015073424353015477 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000006745625315073424342017723 0ustar rootrootOct 14 07:01:03 crc systemd[1]: Starting Kubernetes Kubelet... Oct 14 07:01:03 crc restorecon[4807]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:03 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 14 07:01:04 crc restorecon[4807]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 14 07:01:04 crc restorecon[4807]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Oct 14 07:01:04 crc kubenswrapper[4870]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 14 07:01:04 crc kubenswrapper[4870]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Oct 14 07:01:04 crc kubenswrapper[4870]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 14 07:01:04 crc kubenswrapper[4870]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 14 07:01:04 crc kubenswrapper[4870]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Oct 14 07:01:04 crc kubenswrapper[4870]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.804747 4870 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.809240 4870 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.809256 4870 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.809260 4870 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.809264 4870 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.809268 4870 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.809274 4870 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.809278 4870 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.809282 4870 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.809286 4870 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.809290 4870 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.809293 4870 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.809297 4870 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.809300 4870 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.809303 4870 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.809307 4870 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.809310 4870 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.809314 4870 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.809317 4870 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.809321 4870 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.809324 4870 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.809333 4870 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.809337 4870 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.809341 4870 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.809344 4870 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.809350 4870 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.809354 4870 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.809358 4870 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.809363 4870 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.809367 4870 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.809372 4870 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.809376 4870 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.809379 4870 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.809383 4870 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.809387 4870 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.809390 4870 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.809393 4870 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.809397 4870 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.809401 4870 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.809406 4870 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.809410 4870 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.809419 4870 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.809426 4870 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.809431 4870 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.809450 4870 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.809503 4870 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.809509 4870 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.809513 4870 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.809517 4870 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.809521 4870 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.809524 4870 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.809528 4870 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.809532 4870 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.809536 4870 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.809540 4870 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.809543 4870 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.809547 4870 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.809551 4870 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.809556 4870 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.809562 4870 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.809566 4870 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.809570 4870 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.809574 4870 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.809578 4870 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.809581 4870 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.809585 4870 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.809588 4870 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.809592 4870 feature_gate.go:330] unrecognized feature gate: Example Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.809595 4870 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.809598 4870 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.809603 4870 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.809606 4870 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.811413 4870 flags.go:64] FLAG: --address="0.0.0.0" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.811427 4870 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.811448 4870 flags.go:64] FLAG: --anonymous-auth="true" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.811454 4870 flags.go:64] FLAG: --application-metrics-count-limit="100" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.811459 4870 flags.go:64] FLAG: --authentication-token-webhook="false" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.811464 4870 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.811470 4870 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.811475 4870 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.811479 4870 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.811483 4870 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.811488 4870 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.811492 4870 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.811496 4870 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.811500 4870 flags.go:64] FLAG: --cgroup-root="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.811504 4870 flags.go:64] FLAG: --cgroups-per-qos="true" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.811508 4870 flags.go:64] FLAG: --client-ca-file="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.811512 4870 flags.go:64] FLAG: --cloud-config="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.811516 4870 flags.go:64] FLAG: --cloud-provider="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.811520 4870 flags.go:64] FLAG: --cluster-dns="[]" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.811525 4870 flags.go:64] FLAG: --cluster-domain="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.811528 4870 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.811533 4870 flags.go:64] FLAG: --config-dir="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.811537 4870 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.811541 4870 flags.go:64] FLAG: --container-log-max-files="5" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.811547 4870 flags.go:64] FLAG: --container-log-max-size="10Mi" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.811551 4870 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.811556 4870 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.811560 4870 flags.go:64] FLAG: --containerd-namespace="k8s.io" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.811564 4870 flags.go:64] FLAG: --contention-profiling="false" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.811569 4870 flags.go:64] FLAG: --cpu-cfs-quota="true" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.811573 4870 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.811578 4870 flags.go:64] FLAG: --cpu-manager-policy="none" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.811582 4870 flags.go:64] FLAG: --cpu-manager-policy-options="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.811587 4870 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.811592 4870 flags.go:64] FLAG: --enable-controller-attach-detach="true" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.811596 4870 flags.go:64] FLAG: --enable-debugging-handlers="true" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.811600 4870 flags.go:64] FLAG: --enable-load-reader="false" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.811604 4870 flags.go:64] FLAG: --enable-server="true" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.811608 4870 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.811613 4870 flags.go:64] FLAG: --event-burst="100" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.811617 4870 flags.go:64] FLAG: --event-qps="50" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.811621 4870 flags.go:64] FLAG: --event-storage-age-limit="default=0" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.811625 4870 flags.go:64] FLAG: --event-storage-event-limit="default=0" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.811629 4870 flags.go:64] FLAG: --eviction-hard="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.811635 4870 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.811639 4870 flags.go:64] FLAG: --eviction-minimum-reclaim="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.811643 4870 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.811648 4870 flags.go:64] FLAG: --eviction-soft="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.811652 4870 flags.go:64] FLAG: --eviction-soft-grace-period="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.811655 4870 flags.go:64] FLAG: --exit-on-lock-contention="false" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.811659 4870 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.811663 4870 flags.go:64] FLAG: --experimental-mounter-path="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.811668 4870 flags.go:64] FLAG: --fail-cgroupv1="false" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.811671 4870 flags.go:64] FLAG: --fail-swap-on="true" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.811675 4870 flags.go:64] FLAG: --feature-gates="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.811680 4870 flags.go:64] FLAG: --file-check-frequency="20s" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.811684 4870 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.811689 4870 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.811693 4870 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.811698 4870 flags.go:64] FLAG: --healthz-port="10248" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.811704 4870 flags.go:64] FLAG: --help="false" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.811708 4870 flags.go:64] FLAG: --hostname-override="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.811712 4870 flags.go:64] FLAG: --housekeeping-interval="10s" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.811717 4870 flags.go:64] FLAG: --http-check-frequency="20s" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.811721 4870 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.811726 4870 flags.go:64] FLAG: --image-credential-provider-config="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.811731 4870 flags.go:64] FLAG: --image-gc-high-threshold="85" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.811736 4870 flags.go:64] FLAG: --image-gc-low-threshold="80" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.811740 4870 flags.go:64] FLAG: --image-service-endpoint="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.811744 4870 flags.go:64] FLAG: --kernel-memcg-notification="false" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.811750 4870 flags.go:64] FLAG: --kube-api-burst="100" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.811755 4870 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.811759 4870 flags.go:64] FLAG: --kube-api-qps="50" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.811764 4870 flags.go:64] FLAG: --kube-reserved="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.811768 4870 flags.go:64] FLAG: --kube-reserved-cgroup="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.811771 4870 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.811775 4870 flags.go:64] FLAG: --kubelet-cgroups="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.811779 4870 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.811783 4870 flags.go:64] FLAG: --lock-file="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.811787 4870 flags.go:64] FLAG: --log-cadvisor-usage="false" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.811791 4870 flags.go:64] FLAG: --log-flush-frequency="5s" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.811795 4870 flags.go:64] FLAG: --log-json-info-buffer-size="0" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.811801 4870 flags.go:64] FLAG: --log-json-split-stream="false" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.811805 4870 flags.go:64] FLAG: --log-text-info-buffer-size="0" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.811809 4870 flags.go:64] FLAG: --log-text-split-stream="false" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.811812 4870 flags.go:64] FLAG: --logging-format="text" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.811816 4870 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.811821 4870 flags.go:64] FLAG: --make-iptables-util-chains="true" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.811825 4870 flags.go:64] FLAG: --manifest-url="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.811829 4870 flags.go:64] FLAG: --manifest-url-header="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.811834 4870 flags.go:64] FLAG: --max-housekeeping-interval="15s" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.811838 4870 flags.go:64] FLAG: --max-open-files="1000000" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.811844 4870 flags.go:64] FLAG: --max-pods="110" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.811848 4870 flags.go:64] FLAG: --maximum-dead-containers="-1" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.811852 4870 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.811856 4870 flags.go:64] FLAG: --memory-manager-policy="None" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.811860 4870 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.811864 4870 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.811868 4870 flags.go:64] FLAG: --node-ip="192.168.126.11" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.811873 4870 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.811883 4870 flags.go:64] FLAG: --node-status-max-images="50" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.811887 4870 flags.go:64] FLAG: --node-status-update-frequency="10s" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.811891 4870 flags.go:64] FLAG: --oom-score-adj="-999" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.811895 4870 flags.go:64] FLAG: --pod-cidr="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.811899 4870 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.811904 4870 flags.go:64] FLAG: --pod-manifest-path="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.811909 4870 flags.go:64] FLAG: --pod-max-pids="-1" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.811913 4870 flags.go:64] FLAG: --pods-per-core="0" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.811917 4870 flags.go:64] FLAG: --port="10250" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.811921 4870 flags.go:64] FLAG: --protect-kernel-defaults="false" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.811925 4870 flags.go:64] FLAG: --provider-id="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.811929 4870 flags.go:64] FLAG: --qos-reserved="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.811933 4870 flags.go:64] FLAG: --read-only-port="10255" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.811937 4870 flags.go:64] FLAG: --register-node="true" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.811942 4870 flags.go:64] FLAG: --register-schedulable="true" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.811945 4870 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.811952 4870 flags.go:64] FLAG: --registry-burst="10" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.811956 4870 flags.go:64] FLAG: --registry-qps="5" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.811960 4870 flags.go:64] FLAG: --reserved-cpus="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.811964 4870 flags.go:64] FLAG: --reserved-memory="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.811969 4870 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.811973 4870 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.811977 4870 flags.go:64] FLAG: --rotate-certificates="false" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.811981 4870 flags.go:64] FLAG: --rotate-server-certificates="false" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.811984 4870 flags.go:64] FLAG: --runonce="false" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.811989 4870 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.811993 4870 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.811997 4870 flags.go:64] FLAG: --seccomp-default="false" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.812001 4870 flags.go:64] FLAG: --serialize-image-pulls="true" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.812005 4870 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.812010 4870 flags.go:64] FLAG: --storage-driver-db="cadvisor" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.812014 4870 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.812018 4870 flags.go:64] FLAG: --storage-driver-password="root" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.812022 4870 flags.go:64] FLAG: --storage-driver-secure="false" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.812026 4870 flags.go:64] FLAG: --storage-driver-table="stats" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.812030 4870 flags.go:64] FLAG: --storage-driver-user="root" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.812034 4870 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.812038 4870 flags.go:64] FLAG: --sync-frequency="1m0s" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.812042 4870 flags.go:64] FLAG: --system-cgroups="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.812046 4870 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.812053 4870 flags.go:64] FLAG: --system-reserved-cgroup="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.812057 4870 flags.go:64] FLAG: --tls-cert-file="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.812061 4870 flags.go:64] FLAG: --tls-cipher-suites="[]" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.812066 4870 flags.go:64] FLAG: --tls-min-version="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.812070 4870 flags.go:64] FLAG: --tls-private-key-file="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.812074 4870 flags.go:64] FLAG: --topology-manager-policy="none" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.812078 4870 flags.go:64] FLAG: --topology-manager-policy-options="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.812082 4870 flags.go:64] FLAG: --topology-manager-scope="container" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.812086 4870 flags.go:64] FLAG: --v="2" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.812091 4870 flags.go:64] FLAG: --version="false" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.812097 4870 flags.go:64] FLAG: --vmodule="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.812102 4870 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.812106 4870 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.812306 4870 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.812315 4870 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.812319 4870 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.812323 4870 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.812327 4870 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.812333 4870 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.812342 4870 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.812346 4870 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.812351 4870 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.812354 4870 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.812358 4870 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.812362 4870 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.812365 4870 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.812369 4870 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.812373 4870 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.812376 4870 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.812380 4870 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.812384 4870 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.812390 4870 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.812393 4870 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.812397 4870 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.812401 4870 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.812404 4870 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.812408 4870 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.812411 4870 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.812471 4870 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.812477 4870 feature_gate.go:330] unrecognized feature gate: Example Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.812482 4870 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.812486 4870 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.812491 4870 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.812495 4870 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.812504 4870 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.812509 4870 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.812513 4870 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.812522 4870 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.812527 4870 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.812531 4870 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.812535 4870 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.812539 4870 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.812542 4870 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.812546 4870 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.812550 4870 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.812553 4870 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.812557 4870 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.813859 4870 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.813872 4870 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.813908 4870 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.813917 4870 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.814224 4870 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.814236 4870 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.814243 4870 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.814247 4870 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.814254 4870 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.814259 4870 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.814264 4870 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.814268 4870 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.814273 4870 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.814277 4870 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.814282 4870 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.814287 4870 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.814292 4870 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.814297 4870 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.814302 4870 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.814307 4870 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.814312 4870 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.814321 4870 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.814326 4870 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.814332 4870 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.814336 4870 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.814341 4870 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.814346 4870 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.814355 4870 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.822923 4870 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.822955 4870 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.823027 4870 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.823036 4870 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.823041 4870 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.823048 4870 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.823055 4870 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.823061 4870 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.823065 4870 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.823070 4870 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.823075 4870 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.823080 4870 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.823085 4870 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.823090 4870 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.823095 4870 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.823101 4870 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.823107 4870 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.823112 4870 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.823117 4870 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.823121 4870 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.823125 4870 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.823130 4870 feature_gate.go:330] unrecognized feature gate: Example Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.823134 4870 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.823138 4870 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.823143 4870 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.823147 4870 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.823151 4870 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.823155 4870 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.823159 4870 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.823163 4870 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.823168 4870 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.823172 4870 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.823177 4870 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.823182 4870 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.823186 4870 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.823191 4870 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.823196 4870 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.823200 4870 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.823204 4870 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.823209 4870 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.823213 4870 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.823217 4870 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.823222 4870 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.823226 4870 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.823231 4870 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.823235 4870 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.823239 4870 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.823244 4870 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.823249 4870 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.823253 4870 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.823257 4870 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.823262 4870 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.823266 4870 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.823272 4870 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.823278 4870 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.823283 4870 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.823288 4870 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.823294 4870 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.823299 4870 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.823303 4870 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.823308 4870 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.823314 4870 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.823320 4870 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.823325 4870 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.823331 4870 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.823336 4870 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.823341 4870 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.823349 4870 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.823353 4870 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.823358 4870 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.823362 4870 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.823367 4870 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.823372 4870 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.823380 4870 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.823561 4870 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.823573 4870 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.823578 4870 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.823583 4870 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.823588 4870 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.823593 4870 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.823597 4870 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.823602 4870 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.823606 4870 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.823611 4870 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.823615 4870 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.823620 4870 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.823624 4870 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.823629 4870 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.823635 4870 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.823642 4870 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.823647 4870 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.823652 4870 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.823657 4870 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.823662 4870 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.823666 4870 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.823671 4870 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.823676 4870 feature_gate.go:330] unrecognized feature gate: Example Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.823680 4870 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.823686 4870 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.823690 4870 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.823695 4870 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.823700 4870 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.823705 4870 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.823710 4870 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.823716 4870 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.823723 4870 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.823730 4870 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.823735 4870 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.823739 4870 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.823745 4870 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.823750 4870 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.823755 4870 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.823760 4870 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.823765 4870 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.823770 4870 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.823775 4870 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.823780 4870 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.823784 4870 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.823789 4870 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.823794 4870 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.823800 4870 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.823806 4870 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.823830 4870 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.823836 4870 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.823841 4870 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.823847 4870 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.823852 4870 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.823856 4870 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.823862 4870 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.823868 4870 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.823872 4870 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.823878 4870 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.823882 4870 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.823886 4870 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.823891 4870 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.823896 4870 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.823901 4870 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.823906 4870 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.823911 4870 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.823916 4870 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.823920 4870 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.823925 4870 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.823929 4870 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.823934 4870 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.823938 4870 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.823947 4870 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.824919 4870 server.go:940] "Client rotation is on, will bootstrap in background" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.828831 4870 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.828907 4870 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.830453 4870 server.go:997] "Starting client certificate rotation" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.830472 4870 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.830666 4870 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2025-11-21 10:35:53.095065427 +0000 UTC Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.830809 4870 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 915h34m48.264259679s for next certificate rotation Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.859105 4870 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.860813 4870 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.872091 4870 log.go:25] "Validated CRI v1 runtime API" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.906767 4870 log.go:25] "Validated CRI v1 image API" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.908556 4870 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.914771 4870 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-10-14-06-40-15-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.914813 4870 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:29 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.933688 4870 manager.go:217] Machine: {Timestamp:2025-10-14 07:01:04.930954477 +0000 UTC m=+0.628314858 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2800000 MemoryCapacity:33654124544 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:0adb347d-b90d-4e9f-9962-f1bd97980aed BootID:56448e6b-6b3a-4bb9-91f8-2f0a9b44e9aa Filesystems:[{Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827060224 Type:vfs Inodes:4108169 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:29 Capacity:16827064320 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365408768 Type:vfs Inodes:821633 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:4108169 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:8a:96:d8 Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:8a:96:d8 Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:6c:d8:2b Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:c0:ec:f6 Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:22:4f:f8 Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:97:9e:c8 Speed:-1 Mtu:1496} {Name:ens7.23 MacAddress:52:54:00:67:84:59 Speed:-1 Mtu:1496} {Name:ens7.44 MacAddress:52:54:00:4a:bb:3b Speed:-1 Mtu:1496} {Name:eth10 MacAddress:ea:49:ac:a0:45:c5 Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:b6:b7:62:a7:db:7f Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654124544 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.933924 4870 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.934072 4870 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.935667 4870 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.936004 4870 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.936052 4870 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.936381 4870 topology_manager.go:138] "Creating topology manager with none policy" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.936399 4870 container_manager_linux.go:303] "Creating device plugin manager" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.937208 4870 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.937252 4870 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.937612 4870 state_mem.go:36] "Initialized new in-memory state store" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.937769 4870 server.go:1245] "Using root directory" path="/var/lib/kubelet" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.942957 4870 kubelet.go:418] "Attempting to sync node with API server" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.942996 4870 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.943048 4870 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.943069 4870 kubelet.go:324] "Adding apiserver pod source" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.943088 4870 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.947955 4870 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.949073 4870 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.949308 4870 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.23:6443: connect: connection refused Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.949309 4870 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.23:6443: connect: connection refused Oct 14 07:01:04 crc kubenswrapper[4870]: E1014 07:01:04.949401 4870 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.23:6443: connect: connection refused" logger="UnhandledError" Oct 14 07:01:04 crc kubenswrapper[4870]: E1014 07:01:04.949479 4870 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.23:6443: connect: connection refused" logger="UnhandledError" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.952112 4870 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.953709 4870 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.953751 4870 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.953765 4870 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.953779 4870 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.953801 4870 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.953814 4870 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.953828 4870 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.953849 4870 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.953865 4870 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.953889 4870 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.953910 4870 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.953925 4870 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.954835 4870 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.955551 4870 server.go:1280] "Started kubelet" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.957031 4870 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.957039 4870 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Oct 14 07:01:04 crc systemd[1]: Started Kubernetes Kubelet. Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.957412 4870 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.23:6443: connect: connection refused Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.957424 4870 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.963812 4870 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.963862 4870 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.964642 4870 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-11 05:47:48.155811458 +0000 UTC Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.964702 4870 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 1390h46m43.191119934s for next certificate rotation Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.964928 4870 volume_manager.go:287] "The desired_state_of_world populator starts" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.964992 4870 volume_manager.go:289] "Starting Kubelet Volume Manager" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.965282 4870 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Oct 14 07:01:04 crc kubenswrapper[4870]: E1014 07:01:04.965434 4870 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Oct 14 07:01:04 crc kubenswrapper[4870]: E1014 07:01:04.966398 4870 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.23:6443: connect: connection refused" interval="200ms" Oct 14 07:01:04 crc kubenswrapper[4870]: W1014 07:01:04.966970 4870 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.23:6443: connect: connection refused Oct 14 07:01:04 crc kubenswrapper[4870]: E1014 07:01:04.967076 4870 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.23:6443: connect: connection refused" logger="UnhandledError" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.968034 4870 factory.go:55] Registering systemd factory Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.968687 4870 factory.go:221] Registration of the systemd container factory successfully Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.971390 4870 factory.go:153] Registering CRI-O factory Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.971487 4870 factory.go:221] Registration of the crio container factory successfully Oct 14 07:01:04 crc kubenswrapper[4870]: E1014 07:01:04.967066 4870 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.23:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.186e49738b4bbbbe default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-10-14 07:01:04.955505598 +0000 UTC m=+0.652866029,LastTimestamp:2025-10-14 07:01:04.955505598 +0000 UTC m=+0.652866029,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.971763 4870 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.971811 4870 factory.go:103] Registering Raw factory Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.971920 4870 manager.go:1196] Started watching for new ooms in manager Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.972524 4870 server.go:460] "Adding debug handlers to kubelet server" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.974858 4870 manager.go:319] Starting recovery of all containers Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.980777 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.980871 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.980908 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.980937 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.980963 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.980989 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.981015 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.981040 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.981073 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.981102 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.981126 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.981206 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.981232 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.981268 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.981298 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.981324 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.981349 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.981373 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.981399 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.981424 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.981559 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.981595 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.981622 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.981647 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.981674 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.981700 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.981729 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.981755 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.981781 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.981811 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.981835 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.981872 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.981899 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.981923 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.981949 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.981975 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.981998 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.982022 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.982046 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.982070 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.982093 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.982163 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.982190 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.982213 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.982237 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.982263 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.982288 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.982311 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.982334 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.982356 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.982380 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.982403 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.982517 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.982545 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.982569 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.982596 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.982620 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.982644 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.982668 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.982692 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.982716 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.982743 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.982768 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.982793 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.982820 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.982844 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.982871 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.982897 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.982922 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.982949 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.982972 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.982998 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.983022 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.983043 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.983068 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.983090 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.983116 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.983143 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.983164 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.983187 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.983210 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.983233 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.983255 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.983278 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.983304 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.983328 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.983352 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.983374 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.983397 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.983420 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.983474 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.983497 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.983520 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.983559 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.983584 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.983607 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.983635 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.983661 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.983688 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.983713 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.983740 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.983765 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.983790 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.983815 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.983853 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.983882 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.983912 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.983939 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.983966 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.983998 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.984025 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.984056 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.984086 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.984113 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.984141 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.984166 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.984191 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.984216 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.984242 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.984269 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.984292 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.984379 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.984409 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.984467 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.984497 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.984525 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.984553 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.984577 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.984604 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.984628 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.984654 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.984680 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.984704 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.984728 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.984755 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.984780 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.984807 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.984833 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.984857 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.984880 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.984903 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.984929 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.984949 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.984976 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.985000 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.985024 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.985048 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.985076 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.985099 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.985122 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.985146 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.985172 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.985197 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.985221 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.985249 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.985277 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.985302 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.985329 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.985354 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.985380 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.985414 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.985468 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.985495 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.985518 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.985540 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.985564 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.985599 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.985627 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.985650 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.985675 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.985699 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.985725 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.985752 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.985778 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.985803 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.985829 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.985853 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.985878 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.985904 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.985929 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.985954 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.985978 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.986005 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.986030 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.986055 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.986080 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.986128 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.986157 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.986182 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.986208 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.986233 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.988861 4870 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.988958 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.988994 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.989011 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.989032 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.989053 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.989070 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.989092 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.989106 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.989120 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.989137 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.989153 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.989171 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.989188 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.989209 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.989224 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.989240 4870 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.989251 4870 reconstruct.go:97] "Volume reconstruction finished" Oct 14 07:01:04 crc kubenswrapper[4870]: I1014 07:01:04.989261 4870 reconciler.go:26] "Reconciler: start to sync state" Oct 14 07:01:05 crc kubenswrapper[4870]: I1014 07:01:05.000327 4870 manager.go:324] Recovery completed Oct 14 07:01:05 crc kubenswrapper[4870]: I1014 07:01:05.009491 4870 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 07:01:05 crc kubenswrapper[4870]: I1014 07:01:05.012086 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:05 crc kubenswrapper[4870]: I1014 07:01:05.012131 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:05 crc kubenswrapper[4870]: I1014 07:01:05.012147 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:05 crc kubenswrapper[4870]: I1014 07:01:05.014225 4870 cpu_manager.go:225] "Starting CPU manager" policy="none" Oct 14 07:01:05 crc kubenswrapper[4870]: I1014 07:01:05.014749 4870 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Oct 14 07:01:05 crc kubenswrapper[4870]: I1014 07:01:05.014793 4870 state_mem.go:36] "Initialized new in-memory state store" Oct 14 07:01:05 crc kubenswrapper[4870]: I1014 07:01:05.029738 4870 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Oct 14 07:01:05 crc kubenswrapper[4870]: I1014 07:01:05.032534 4870 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Oct 14 07:01:05 crc kubenswrapper[4870]: I1014 07:01:05.032623 4870 status_manager.go:217] "Starting to sync pod status with apiserver" Oct 14 07:01:05 crc kubenswrapper[4870]: I1014 07:01:05.032657 4870 kubelet.go:2335] "Starting kubelet main sync loop" Oct 14 07:01:05 crc kubenswrapper[4870]: W1014 07:01:05.033132 4870 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.23:6443: connect: connection refused Oct 14 07:01:05 crc kubenswrapper[4870]: E1014 07:01:05.033171 4870 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.23:6443: connect: connection refused" logger="UnhandledError" Oct 14 07:01:05 crc kubenswrapper[4870]: E1014 07:01:05.033238 4870 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Oct 14 07:01:05 crc kubenswrapper[4870]: I1014 07:01:05.038990 4870 policy_none.go:49] "None policy: Start" Oct 14 07:01:05 crc kubenswrapper[4870]: I1014 07:01:05.040383 4870 memory_manager.go:170] "Starting memorymanager" policy="None" Oct 14 07:01:05 crc kubenswrapper[4870]: I1014 07:01:05.040449 4870 state_mem.go:35] "Initializing new in-memory state store" Oct 14 07:01:05 crc kubenswrapper[4870]: E1014 07:01:05.066036 4870 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Oct 14 07:01:05 crc kubenswrapper[4870]: I1014 07:01:05.105808 4870 manager.go:334] "Starting Device Plugin manager" Oct 14 07:01:05 crc kubenswrapper[4870]: I1014 07:01:05.105884 4870 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Oct 14 07:01:05 crc kubenswrapper[4870]: I1014 07:01:05.105900 4870 server.go:79] "Starting device plugin registration server" Oct 14 07:01:05 crc kubenswrapper[4870]: I1014 07:01:05.106527 4870 eviction_manager.go:189] "Eviction manager: starting control loop" Oct 14 07:01:05 crc kubenswrapper[4870]: I1014 07:01:05.106545 4870 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Oct 14 07:01:05 crc kubenswrapper[4870]: I1014 07:01:05.107168 4870 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Oct 14 07:01:05 crc kubenswrapper[4870]: I1014 07:01:05.107266 4870 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Oct 14 07:01:05 crc kubenswrapper[4870]: I1014 07:01:05.107276 4870 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Oct 14 07:01:05 crc kubenswrapper[4870]: E1014 07:01:05.113532 4870 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Oct 14 07:01:05 crc kubenswrapper[4870]: I1014 07:01:05.133937 4870 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc","openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Oct 14 07:01:05 crc kubenswrapper[4870]: I1014 07:01:05.134305 4870 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 07:01:05 crc kubenswrapper[4870]: I1014 07:01:05.135552 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:05 crc kubenswrapper[4870]: I1014 07:01:05.135661 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:05 crc kubenswrapper[4870]: I1014 07:01:05.135743 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:05 crc kubenswrapper[4870]: I1014 07:01:05.135965 4870 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 07:01:05 crc kubenswrapper[4870]: I1014 07:01:05.136390 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Oct 14 07:01:05 crc kubenswrapper[4870]: I1014 07:01:05.136526 4870 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 07:01:05 crc kubenswrapper[4870]: I1014 07:01:05.136644 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:05 crc kubenswrapper[4870]: I1014 07:01:05.136676 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:05 crc kubenswrapper[4870]: I1014 07:01:05.136689 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:05 crc kubenswrapper[4870]: I1014 07:01:05.136793 4870 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 07:01:05 crc kubenswrapper[4870]: I1014 07:01:05.137012 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 14 07:01:05 crc kubenswrapper[4870]: I1014 07:01:05.137084 4870 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 07:01:05 crc kubenswrapper[4870]: I1014 07:01:05.138081 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:05 crc kubenswrapper[4870]: I1014 07:01:05.138105 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:05 crc kubenswrapper[4870]: I1014 07:01:05.138117 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:05 crc kubenswrapper[4870]: I1014 07:01:05.138081 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:05 crc kubenswrapper[4870]: I1014 07:01:05.138160 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:05 crc kubenswrapper[4870]: I1014 07:01:05.138199 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:05 crc kubenswrapper[4870]: I1014 07:01:05.138309 4870 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 07:01:05 crc kubenswrapper[4870]: I1014 07:01:05.138561 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:05 crc kubenswrapper[4870]: I1014 07:01:05.138582 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:05 crc kubenswrapper[4870]: I1014 07:01:05.138595 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:05 crc kubenswrapper[4870]: I1014 07:01:05.138757 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 14 07:01:05 crc kubenswrapper[4870]: I1014 07:01:05.138787 4870 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 07:01:05 crc kubenswrapper[4870]: I1014 07:01:05.138929 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:05 crc kubenswrapper[4870]: I1014 07:01:05.139014 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:05 crc kubenswrapper[4870]: I1014 07:01:05.139032 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:05 crc kubenswrapper[4870]: I1014 07:01:05.139121 4870 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 07:01:05 crc kubenswrapper[4870]: I1014 07:01:05.139254 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 14 07:01:05 crc kubenswrapper[4870]: I1014 07:01:05.139287 4870 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 07:01:05 crc kubenswrapper[4870]: I1014 07:01:05.139646 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:05 crc kubenswrapper[4870]: I1014 07:01:05.139668 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:05 crc kubenswrapper[4870]: I1014 07:01:05.139681 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:05 crc kubenswrapper[4870]: I1014 07:01:05.139845 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:05 crc kubenswrapper[4870]: I1014 07:01:05.139867 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:05 crc kubenswrapper[4870]: I1014 07:01:05.139878 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:05 crc kubenswrapper[4870]: I1014 07:01:05.139977 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 14 07:01:05 crc kubenswrapper[4870]: I1014 07:01:05.139999 4870 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 07:01:05 crc kubenswrapper[4870]: I1014 07:01:05.140610 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:05 crc kubenswrapper[4870]: I1014 07:01:05.140626 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:05 crc kubenswrapper[4870]: I1014 07:01:05.140635 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:05 crc kubenswrapper[4870]: I1014 07:01:05.140733 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:05 crc kubenswrapper[4870]: I1014 07:01:05.140743 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:05 crc kubenswrapper[4870]: I1014 07:01:05.140751 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:05 crc kubenswrapper[4870]: E1014 07:01:05.167395 4870 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.23:6443: connect: connection refused" interval="400ms" Oct 14 07:01:05 crc kubenswrapper[4870]: I1014 07:01:05.191873 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 14 07:01:05 crc kubenswrapper[4870]: I1014 07:01:05.191914 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 14 07:01:05 crc kubenswrapper[4870]: I1014 07:01:05.191938 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 14 07:01:05 crc kubenswrapper[4870]: I1014 07:01:05.191956 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 14 07:01:05 crc kubenswrapper[4870]: I1014 07:01:05.192032 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 14 07:01:05 crc kubenswrapper[4870]: I1014 07:01:05.192111 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 14 07:01:05 crc kubenswrapper[4870]: I1014 07:01:05.192135 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 14 07:01:05 crc kubenswrapper[4870]: I1014 07:01:05.192169 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 14 07:01:05 crc kubenswrapper[4870]: I1014 07:01:05.192192 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 14 07:01:05 crc kubenswrapper[4870]: I1014 07:01:05.192230 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 14 07:01:05 crc kubenswrapper[4870]: I1014 07:01:05.192248 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 14 07:01:05 crc kubenswrapper[4870]: I1014 07:01:05.192278 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 14 07:01:05 crc kubenswrapper[4870]: I1014 07:01:05.192338 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 14 07:01:05 crc kubenswrapper[4870]: I1014 07:01:05.192367 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 14 07:01:05 crc kubenswrapper[4870]: I1014 07:01:05.192421 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 14 07:01:05 crc kubenswrapper[4870]: I1014 07:01:05.206805 4870 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 07:01:05 crc kubenswrapper[4870]: I1014 07:01:05.208176 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:05 crc kubenswrapper[4870]: I1014 07:01:05.208224 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:05 crc kubenswrapper[4870]: I1014 07:01:05.208236 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:05 crc kubenswrapper[4870]: I1014 07:01:05.208266 4870 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 14 07:01:05 crc kubenswrapper[4870]: E1014 07:01:05.208879 4870 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.23:6443: connect: connection refused" node="crc" Oct 14 07:01:05 crc kubenswrapper[4870]: I1014 07:01:05.293362 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 14 07:01:05 crc kubenswrapper[4870]: I1014 07:01:05.293408 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 14 07:01:05 crc kubenswrapper[4870]: I1014 07:01:05.293425 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 14 07:01:05 crc kubenswrapper[4870]: I1014 07:01:05.293461 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 14 07:01:05 crc kubenswrapper[4870]: I1014 07:01:05.293486 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 14 07:01:05 crc kubenswrapper[4870]: I1014 07:01:05.293507 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 14 07:01:05 crc kubenswrapper[4870]: I1014 07:01:05.293527 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 14 07:01:05 crc kubenswrapper[4870]: I1014 07:01:05.293533 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 14 07:01:05 crc kubenswrapper[4870]: I1014 07:01:05.293547 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 14 07:01:05 crc kubenswrapper[4870]: I1014 07:01:05.293575 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 14 07:01:05 crc kubenswrapper[4870]: I1014 07:01:05.293579 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 14 07:01:05 crc kubenswrapper[4870]: I1014 07:01:05.293594 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 14 07:01:05 crc kubenswrapper[4870]: I1014 07:01:05.293615 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 14 07:01:05 crc kubenswrapper[4870]: I1014 07:01:05.293622 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 14 07:01:05 crc kubenswrapper[4870]: I1014 07:01:05.293634 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 14 07:01:05 crc kubenswrapper[4870]: I1014 07:01:05.293637 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 14 07:01:05 crc kubenswrapper[4870]: I1014 07:01:05.293664 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 14 07:01:05 crc kubenswrapper[4870]: I1014 07:01:05.293731 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 14 07:01:05 crc kubenswrapper[4870]: I1014 07:01:05.293774 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 14 07:01:05 crc kubenswrapper[4870]: I1014 07:01:05.293750 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 14 07:01:05 crc kubenswrapper[4870]: I1014 07:01:05.293783 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 14 07:01:05 crc kubenswrapper[4870]: I1014 07:01:05.293785 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 14 07:01:05 crc kubenswrapper[4870]: I1014 07:01:05.293810 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 14 07:01:05 crc kubenswrapper[4870]: I1014 07:01:05.293805 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 14 07:01:05 crc kubenswrapper[4870]: I1014 07:01:05.293849 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 14 07:01:05 crc kubenswrapper[4870]: I1014 07:01:05.293878 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 14 07:01:05 crc kubenswrapper[4870]: I1014 07:01:05.293930 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 14 07:01:05 crc kubenswrapper[4870]: I1014 07:01:05.293978 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 14 07:01:05 crc kubenswrapper[4870]: I1014 07:01:05.294056 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 14 07:01:05 crc kubenswrapper[4870]: I1014 07:01:05.293977 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 14 07:01:05 crc kubenswrapper[4870]: I1014 07:01:05.409316 4870 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 07:01:05 crc kubenswrapper[4870]: I1014 07:01:05.410906 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:05 crc kubenswrapper[4870]: I1014 07:01:05.411000 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:05 crc kubenswrapper[4870]: I1014 07:01:05.411019 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:05 crc kubenswrapper[4870]: I1014 07:01:05.411067 4870 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 14 07:01:05 crc kubenswrapper[4870]: E1014 07:01:05.411899 4870 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.23:6443: connect: connection refused" node="crc" Oct 14 07:01:05 crc kubenswrapper[4870]: I1014 07:01:05.489684 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 14 07:01:05 crc kubenswrapper[4870]: I1014 07:01:05.508564 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Oct 14 07:01:05 crc kubenswrapper[4870]: I1014 07:01:05.537839 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 14 07:01:05 crc kubenswrapper[4870]: I1014 07:01:05.545759 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 14 07:01:05 crc kubenswrapper[4870]: W1014 07:01:05.549486 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-470ec9dd2cd8d771cb4f3ba54ff87a444ebecb441cf22323b8282d4908ff2cda WatchSource:0}: Error finding container 470ec9dd2cd8d771cb4f3ba54ff87a444ebecb441cf22323b8282d4908ff2cda: Status 404 returned error can't find the container with id 470ec9dd2cd8d771cb4f3ba54ff87a444ebecb441cf22323b8282d4908ff2cda Oct 14 07:01:05 crc kubenswrapper[4870]: I1014 07:01:05.549766 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 14 07:01:05 crc kubenswrapper[4870]: W1014 07:01:05.556950 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-2bdd6615483310d106552201618f53a870069273d18a934f85f50ee1a64e58c2 WatchSource:0}: Error finding container 2bdd6615483310d106552201618f53a870069273d18a934f85f50ee1a64e58c2: Status 404 returned error can't find the container with id 2bdd6615483310d106552201618f53a870069273d18a934f85f50ee1a64e58c2 Oct 14 07:01:05 crc kubenswrapper[4870]: E1014 07:01:05.568284 4870 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.23:6443: connect: connection refused" interval="800ms" Oct 14 07:01:05 crc kubenswrapper[4870]: W1014 07:01:05.578115 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-5d0222b1256399aaedda88087760bc8b5091b0219792370ddb2940a373ce2cb2 WatchSource:0}: Error finding container 5d0222b1256399aaedda88087760bc8b5091b0219792370ddb2940a373ce2cb2: Status 404 returned error can't find the container with id 5d0222b1256399aaedda88087760bc8b5091b0219792370ddb2940a373ce2cb2 Oct 14 07:01:05 crc kubenswrapper[4870]: W1014 07:01:05.580185 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-0a248bb268a3d0d0ef1ab641b0f36773c1679f9bbd503a82481f8610373bb700 WatchSource:0}: Error finding container 0a248bb268a3d0d0ef1ab641b0f36773c1679f9bbd503a82481f8610373bb700: Status 404 returned error can't find the container with id 0a248bb268a3d0d0ef1ab641b0f36773c1679f9bbd503a82481f8610373bb700 Oct 14 07:01:05 crc kubenswrapper[4870]: W1014 07:01:05.589743 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-286ecce3f76ae37e5a851cc985fc8d04e07488f0e4b3471647043579e2dd8cb3 WatchSource:0}: Error finding container 286ecce3f76ae37e5a851cc985fc8d04e07488f0e4b3471647043579e2dd8cb3: Status 404 returned error can't find the container with id 286ecce3f76ae37e5a851cc985fc8d04e07488f0e4b3471647043579e2dd8cb3 Oct 14 07:01:05 crc kubenswrapper[4870]: I1014 07:01:05.812750 4870 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 07:01:05 crc kubenswrapper[4870]: I1014 07:01:05.814560 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:05 crc kubenswrapper[4870]: I1014 07:01:05.814608 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:05 crc kubenswrapper[4870]: I1014 07:01:05.814620 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:05 crc kubenswrapper[4870]: I1014 07:01:05.814656 4870 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 14 07:01:05 crc kubenswrapper[4870]: E1014 07:01:05.815138 4870 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.23:6443: connect: connection refused" node="crc" Oct 14 07:01:05 crc kubenswrapper[4870]: I1014 07:01:05.958206 4870 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.23:6443: connect: connection refused Oct 14 07:01:06 crc kubenswrapper[4870]: W1014 07:01:06.038266 4870 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.23:6443: connect: connection refused Oct 14 07:01:06 crc kubenswrapper[4870]: E1014 07:01:06.038350 4870 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.23:6443: connect: connection refused" logger="UnhandledError" Oct 14 07:01:06 crc kubenswrapper[4870]: I1014 07:01:06.039157 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"5d0222b1256399aaedda88087760bc8b5091b0219792370ddb2940a373ce2cb2"} Oct 14 07:01:06 crc kubenswrapper[4870]: I1014 07:01:06.040377 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"470ec9dd2cd8d771cb4f3ba54ff87a444ebecb441cf22323b8282d4908ff2cda"} Oct 14 07:01:06 crc kubenswrapper[4870]: I1014 07:01:06.041640 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"2bdd6615483310d106552201618f53a870069273d18a934f85f50ee1a64e58c2"} Oct 14 07:01:06 crc kubenswrapper[4870]: I1014 07:01:06.042806 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"286ecce3f76ae37e5a851cc985fc8d04e07488f0e4b3471647043579e2dd8cb3"} Oct 14 07:01:06 crc kubenswrapper[4870]: I1014 07:01:06.043675 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"0a248bb268a3d0d0ef1ab641b0f36773c1679f9bbd503a82481f8610373bb700"} Oct 14 07:01:06 crc kubenswrapper[4870]: W1014 07:01:06.078051 4870 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.23:6443: connect: connection refused Oct 14 07:01:06 crc kubenswrapper[4870]: E1014 07:01:06.078117 4870 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.23:6443: connect: connection refused" logger="UnhandledError" Oct 14 07:01:06 crc kubenswrapper[4870]: W1014 07:01:06.097036 4870 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.23:6443: connect: connection refused Oct 14 07:01:06 crc kubenswrapper[4870]: E1014 07:01:06.097105 4870 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.23:6443: connect: connection refused" logger="UnhandledError" Oct 14 07:01:06 crc kubenswrapper[4870]: W1014 07:01:06.311666 4870 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.23:6443: connect: connection refused Oct 14 07:01:06 crc kubenswrapper[4870]: E1014 07:01:06.311982 4870 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.23:6443: connect: connection refused" logger="UnhandledError" Oct 14 07:01:06 crc kubenswrapper[4870]: E1014 07:01:06.369407 4870 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.23:6443: connect: connection refused" interval="1.6s" Oct 14 07:01:06 crc kubenswrapper[4870]: I1014 07:01:06.615570 4870 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 07:01:06 crc kubenswrapper[4870]: I1014 07:01:06.617615 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:06 crc kubenswrapper[4870]: I1014 07:01:06.617666 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:06 crc kubenswrapper[4870]: I1014 07:01:06.617680 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:06 crc kubenswrapper[4870]: I1014 07:01:06.617707 4870 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 14 07:01:06 crc kubenswrapper[4870]: E1014 07:01:06.618139 4870 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.23:6443: connect: connection refused" node="crc" Oct 14 07:01:06 crc kubenswrapper[4870]: I1014 07:01:06.958334 4870 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.23:6443: connect: connection refused Oct 14 07:01:07 crc kubenswrapper[4870]: I1014 07:01:07.049226 4870 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="8b192bd3e0eb8540d7492abc207025471bc9e198e7e78b50215c8d2f5bc57cf4" exitCode=0 Oct 14 07:01:07 crc kubenswrapper[4870]: I1014 07:01:07.049324 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"8b192bd3e0eb8540d7492abc207025471bc9e198e7e78b50215c8d2f5bc57cf4"} Oct 14 07:01:07 crc kubenswrapper[4870]: I1014 07:01:07.049429 4870 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 07:01:07 crc kubenswrapper[4870]: I1014 07:01:07.051070 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:07 crc kubenswrapper[4870]: I1014 07:01:07.051132 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:07 crc kubenswrapper[4870]: I1014 07:01:07.051154 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:07 crc kubenswrapper[4870]: I1014 07:01:07.051419 4870 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="07543849cd70415fe71b7dd0fd5f3806fdb422afa7a2dcce18a5c1b9fc974a76" exitCode=0 Oct 14 07:01:07 crc kubenswrapper[4870]: I1014 07:01:07.051466 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"07543849cd70415fe71b7dd0fd5f3806fdb422afa7a2dcce18a5c1b9fc974a76"} Oct 14 07:01:07 crc kubenswrapper[4870]: I1014 07:01:07.051631 4870 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 07:01:07 crc kubenswrapper[4870]: I1014 07:01:07.053366 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:07 crc kubenswrapper[4870]: I1014 07:01:07.053400 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:07 crc kubenswrapper[4870]: I1014 07:01:07.053411 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:07 crc kubenswrapper[4870]: I1014 07:01:07.055345 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"77a3247a7f65709543d7fd374b2ab69d7d793537d4132d7e22261188d6e58544"} Oct 14 07:01:07 crc kubenswrapper[4870]: I1014 07:01:07.055373 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"d57d938d9d9c92198ae5b3857536ece36012fbfbe8cc8cfd245d1386d39b0be5"} Oct 14 07:01:07 crc kubenswrapper[4870]: I1014 07:01:07.055389 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"25ec4808b09adc44488530b3c4e29fd7fec46b3fcfa31d4449ec4a6dbc880ec8"} Oct 14 07:01:07 crc kubenswrapper[4870]: I1014 07:01:07.055400 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"3f3f5de86fb11231a19034a18e9d7f31476c162e6567d0c136ead8bbbd900d6b"} Oct 14 07:01:07 crc kubenswrapper[4870]: I1014 07:01:07.055420 4870 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 07:01:07 crc kubenswrapper[4870]: I1014 07:01:07.056569 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:07 crc kubenswrapper[4870]: I1014 07:01:07.056613 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:07 crc kubenswrapper[4870]: I1014 07:01:07.056627 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:07 crc kubenswrapper[4870]: I1014 07:01:07.058235 4870 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="48ca3a5017ba650df99b2fc7201c3fe5abfa89d81089c5647cde1b1549f534fc" exitCode=0 Oct 14 07:01:07 crc kubenswrapper[4870]: I1014 07:01:07.058308 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"48ca3a5017ba650df99b2fc7201c3fe5abfa89d81089c5647cde1b1549f534fc"} Oct 14 07:01:07 crc kubenswrapper[4870]: I1014 07:01:07.058498 4870 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 07:01:07 crc kubenswrapper[4870]: I1014 07:01:07.059686 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:07 crc kubenswrapper[4870]: I1014 07:01:07.059711 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:07 crc kubenswrapper[4870]: I1014 07:01:07.059719 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:07 crc kubenswrapper[4870]: I1014 07:01:07.060649 4870 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="82d3c2214e0be679d2827594b159ace362151318689078bc49d3158f42899636" exitCode=0 Oct 14 07:01:07 crc kubenswrapper[4870]: I1014 07:01:07.060659 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"82d3c2214e0be679d2827594b159ace362151318689078bc49d3158f42899636"} Oct 14 07:01:07 crc kubenswrapper[4870]: I1014 07:01:07.060745 4870 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 07:01:07 crc kubenswrapper[4870]: I1014 07:01:07.062008 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:07 crc kubenswrapper[4870]: I1014 07:01:07.062039 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:07 crc kubenswrapper[4870]: I1014 07:01:07.062052 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:07 crc kubenswrapper[4870]: I1014 07:01:07.064829 4870 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 07:01:07 crc kubenswrapper[4870]: I1014 07:01:07.065815 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:07 crc kubenswrapper[4870]: I1014 07:01:07.065838 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:07 crc kubenswrapper[4870]: I1014 07:01:07.065852 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:07 crc kubenswrapper[4870]: W1014 07:01:07.836159 4870 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.23:6443: connect: connection refused Oct 14 07:01:07 crc kubenswrapper[4870]: E1014 07:01:07.836314 4870 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.23:6443: connect: connection refused" logger="UnhandledError" Oct 14 07:01:07 crc kubenswrapper[4870]: I1014 07:01:07.958288 4870 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.23:6443: connect: connection refused Oct 14 07:01:07 crc kubenswrapper[4870]: E1014 07:01:07.970565 4870 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.23:6443: connect: connection refused" interval="3.2s" Oct 14 07:01:08 crc kubenswrapper[4870]: I1014 07:01:08.076565 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"f75b3dc1557917e9c0c35c2810338a076ed92ba89963edafb9d537ad19c2bd8d"} Oct 14 07:01:08 crc kubenswrapper[4870]: I1014 07:01:08.076626 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"7e8a7a5c1f1706ebbd2e8165d8961002f5a9340e2749e377b0058db107ad7dbf"} Oct 14 07:01:08 crc kubenswrapper[4870]: I1014 07:01:08.076640 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"0c859ebc0d20b632d1d5a716caee343c7e6b9b752ed5f257422cd35ad12c170f"} Oct 14 07:01:08 crc kubenswrapper[4870]: I1014 07:01:08.076585 4870 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 07:01:08 crc kubenswrapper[4870]: I1014 07:01:08.077550 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:08 crc kubenswrapper[4870]: I1014 07:01:08.077589 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:08 crc kubenswrapper[4870]: I1014 07:01:08.077601 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:08 crc kubenswrapper[4870]: I1014 07:01:08.079856 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"f8c8ee375ef1ec48a55be4a2d3e9e8860886b88ccb4080918d3e306ab34d2b95"} Oct 14 07:01:08 crc kubenswrapper[4870]: I1014 07:01:08.079946 4870 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 07:01:08 crc kubenswrapper[4870]: I1014 07:01:08.080988 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:08 crc kubenswrapper[4870]: I1014 07:01:08.081016 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:08 crc kubenswrapper[4870]: I1014 07:01:08.081028 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:08 crc kubenswrapper[4870]: I1014 07:01:08.081943 4870 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="19682eca67f76bf2e06e1b80f12e21141fcbce0acb8195f0defacd1b9a10b5e7" exitCode=0 Oct 14 07:01:08 crc kubenswrapper[4870]: I1014 07:01:08.081973 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"19682eca67f76bf2e06e1b80f12e21141fcbce0acb8195f0defacd1b9a10b5e7"} Oct 14 07:01:08 crc kubenswrapper[4870]: I1014 07:01:08.082065 4870 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 07:01:08 crc kubenswrapper[4870]: I1014 07:01:08.082846 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:08 crc kubenswrapper[4870]: I1014 07:01:08.082872 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:08 crc kubenswrapper[4870]: I1014 07:01:08.082881 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:08 crc kubenswrapper[4870]: I1014 07:01:08.084890 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"2bd0044a247bdc5ca4cab7294103af3270b76cde61f3109b85e83ed050ca53b6"} Oct 14 07:01:08 crc kubenswrapper[4870]: I1014 07:01:08.084933 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"2b329b570622155a55f70501b2260e278106a68ebef497914231b4ec860e0125"} Oct 14 07:01:08 crc kubenswrapper[4870]: I1014 07:01:08.084946 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"ba942a0e6fe5ab93a187e0b7b4bd7c440c2e7827d9b5323fc50f90b562075392"} Oct 14 07:01:08 crc kubenswrapper[4870]: I1014 07:01:08.084956 4870 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 07:01:08 crc kubenswrapper[4870]: I1014 07:01:08.084955 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"7815af9a584f6db6660073da289f2e8c987a04ef9ddfac155bdd9a15b014fcfa"} Oct 14 07:01:08 crc kubenswrapper[4870]: I1014 07:01:08.085544 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:08 crc kubenswrapper[4870]: I1014 07:01:08.085569 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:08 crc kubenswrapper[4870]: I1014 07:01:08.085577 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:08 crc kubenswrapper[4870]: E1014 07:01:08.155393 4870 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.23:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.186e49738b4bbbbe default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-10-14 07:01:04.955505598 +0000 UTC m=+0.652866029,LastTimestamp:2025-10-14 07:01:04.955505598 +0000 UTC m=+0.652866029,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Oct 14 07:01:08 crc kubenswrapper[4870]: I1014 07:01:08.218298 4870 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 07:01:08 crc kubenswrapper[4870]: I1014 07:01:08.224821 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:08 crc kubenswrapper[4870]: I1014 07:01:08.224901 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:08 crc kubenswrapper[4870]: I1014 07:01:08.224915 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:08 crc kubenswrapper[4870]: I1014 07:01:08.224957 4870 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 14 07:01:08 crc kubenswrapper[4870]: E1014 07:01:08.226603 4870 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.23:6443: connect: connection refused" node="crc" Oct 14 07:01:08 crc kubenswrapper[4870]: W1014 07:01:08.637226 4870 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.23:6443: connect: connection refused Oct 14 07:01:08 crc kubenswrapper[4870]: E1014 07:01:08.637356 4870 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.23:6443: connect: connection refused" logger="UnhandledError" Oct 14 07:01:08 crc kubenswrapper[4870]: I1014 07:01:08.959071 4870 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.23:6443: connect: connection refused Oct 14 07:01:09 crc kubenswrapper[4870]: W1014 07:01:09.042607 4870 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.23:6443: connect: connection refused Oct 14 07:01:09 crc kubenswrapper[4870]: E1014 07:01:09.042754 4870 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.23:6443: connect: connection refused" logger="UnhandledError" Oct 14 07:01:09 crc kubenswrapper[4870]: I1014 07:01:09.090921 4870 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="042aad6d5f6bee8eef77f9e6a2c265de775a69f3da72aedc68bdf312198f00fa" exitCode=0 Oct 14 07:01:09 crc kubenswrapper[4870]: I1014 07:01:09.090988 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"042aad6d5f6bee8eef77f9e6a2c265de775a69f3da72aedc68bdf312198f00fa"} Oct 14 07:01:09 crc kubenswrapper[4870]: I1014 07:01:09.091124 4870 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 07:01:09 crc kubenswrapper[4870]: I1014 07:01:09.092829 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:09 crc kubenswrapper[4870]: I1014 07:01:09.092882 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:09 crc kubenswrapper[4870]: I1014 07:01:09.092901 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:09 crc kubenswrapper[4870]: I1014 07:01:09.095958 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"32056f51de4479330f592c7d38413b0cfc159d3cec4c4b7158dcec62daf0e8c8"} Oct 14 07:01:09 crc kubenswrapper[4870]: I1014 07:01:09.095998 4870 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 14 07:01:09 crc kubenswrapper[4870]: I1014 07:01:09.096067 4870 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 07:01:09 crc kubenswrapper[4870]: I1014 07:01:09.096160 4870 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 07:01:09 crc kubenswrapper[4870]: I1014 07:01:09.096165 4870 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 07:01:09 crc kubenswrapper[4870]: I1014 07:01:09.097805 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:09 crc kubenswrapper[4870]: I1014 07:01:09.097827 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:09 crc kubenswrapper[4870]: I1014 07:01:09.097836 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:09 crc kubenswrapper[4870]: I1014 07:01:09.097840 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:09 crc kubenswrapper[4870]: I1014 07:01:09.097870 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:09 crc kubenswrapper[4870]: I1014 07:01:09.097886 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:09 crc kubenswrapper[4870]: I1014 07:01:09.097965 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:09 crc kubenswrapper[4870]: I1014 07:01:09.097982 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:09 crc kubenswrapper[4870]: I1014 07:01:09.097996 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:09 crc kubenswrapper[4870]: W1014 07:01:09.141726 4870 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.23:6443: connect: connection refused Oct 14 07:01:09 crc kubenswrapper[4870]: E1014 07:01:09.141862 4870 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.23:6443: connect: connection refused" logger="UnhandledError" Oct 14 07:01:09 crc kubenswrapper[4870]: I1014 07:01:09.814903 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 14 07:01:09 crc kubenswrapper[4870]: I1014 07:01:09.815173 4870 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 07:01:09 crc kubenswrapper[4870]: I1014 07:01:09.818070 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:09 crc kubenswrapper[4870]: I1014 07:01:09.818113 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:09 crc kubenswrapper[4870]: I1014 07:01:09.818125 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:09 crc kubenswrapper[4870]: I1014 07:01:09.824593 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 14 07:01:10 crc kubenswrapper[4870]: I1014 07:01:10.104843 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"cb398668197dedc1c2a14b4ae7d6a29714f3527309ccb9c1622af841100e1d60"} Oct 14 07:01:10 crc kubenswrapper[4870]: I1014 07:01:10.104901 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"3570643ef1b7d5cc3a542f320413b819338e0c8b7ab7f0f09cc909997a39bab8"} Oct 14 07:01:10 crc kubenswrapper[4870]: I1014 07:01:10.104919 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"5a8135a3d01076e4437ae8a784b459bf8d16de57fa16ec7ce4d40c5129c9d4a9"} Oct 14 07:01:10 crc kubenswrapper[4870]: I1014 07:01:10.107042 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Oct 14 07:01:10 crc kubenswrapper[4870]: I1014 07:01:10.109344 4870 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="32056f51de4479330f592c7d38413b0cfc159d3cec4c4b7158dcec62daf0e8c8" exitCode=255 Oct 14 07:01:10 crc kubenswrapper[4870]: I1014 07:01:10.109521 4870 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 07:01:10 crc kubenswrapper[4870]: I1014 07:01:10.110231 4870 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 07:01:10 crc kubenswrapper[4870]: I1014 07:01:10.110687 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"32056f51de4479330f592c7d38413b0cfc159d3cec4c4b7158dcec62daf0e8c8"} Oct 14 07:01:10 crc kubenswrapper[4870]: I1014 07:01:10.110737 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 14 07:01:10 crc kubenswrapper[4870]: I1014 07:01:10.111057 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:10 crc kubenswrapper[4870]: I1014 07:01:10.111091 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:10 crc kubenswrapper[4870]: I1014 07:01:10.111105 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:10 crc kubenswrapper[4870]: I1014 07:01:10.111752 4870 scope.go:117] "RemoveContainer" containerID="32056f51de4479330f592c7d38413b0cfc159d3cec4c4b7158dcec62daf0e8c8" Oct 14 07:01:10 crc kubenswrapper[4870]: I1014 07:01:10.112596 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:10 crc kubenswrapper[4870]: I1014 07:01:10.112618 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:10 crc kubenswrapper[4870]: I1014 07:01:10.112630 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:10 crc kubenswrapper[4870]: I1014 07:01:10.257243 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 14 07:01:10 crc kubenswrapper[4870]: I1014 07:01:10.257461 4870 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 07:01:10 crc kubenswrapper[4870]: I1014 07:01:10.258620 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:10 crc kubenswrapper[4870]: I1014 07:01:10.258679 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:10 crc kubenswrapper[4870]: I1014 07:01:10.258694 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:10 crc kubenswrapper[4870]: I1014 07:01:10.565141 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 14 07:01:11 crc kubenswrapper[4870]: I1014 07:01:11.121043 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"783ff243e395333f3fc80cfae21358c532f0c9c5cb5515586be56b90ed10d912"} Oct 14 07:01:11 crc kubenswrapper[4870]: I1014 07:01:11.121143 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"c1d2b6d068d63bd38cdfb83824525f6d12054f6cabab5d684e03ea5f2152076c"} Oct 14 07:01:11 crc kubenswrapper[4870]: I1014 07:01:11.121177 4870 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 07:01:11 crc kubenswrapper[4870]: I1014 07:01:11.122963 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:11 crc kubenswrapper[4870]: I1014 07:01:11.123033 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:11 crc kubenswrapper[4870]: I1014 07:01:11.123064 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:11 crc kubenswrapper[4870]: I1014 07:01:11.125710 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Oct 14 07:01:11 crc kubenswrapper[4870]: I1014 07:01:11.128784 4870 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 07:01:11 crc kubenswrapper[4870]: I1014 07:01:11.129741 4870 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 14 07:01:11 crc kubenswrapper[4870]: I1014 07:01:11.129820 4870 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 07:01:11 crc kubenswrapper[4870]: I1014 07:01:11.130580 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"978e8db184ba9e430232c6bb2ee530beb92567fdc71ba3c35b0bedca3492d845"} Oct 14 07:01:11 crc kubenswrapper[4870]: I1014 07:01:11.131286 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:11 crc kubenswrapper[4870]: I1014 07:01:11.131364 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:11 crc kubenswrapper[4870]: I1014 07:01:11.131388 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:11 crc kubenswrapper[4870]: I1014 07:01:11.131384 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:11 crc kubenswrapper[4870]: I1014 07:01:11.131652 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:11 crc kubenswrapper[4870]: I1014 07:01:11.131681 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:11 crc kubenswrapper[4870]: I1014 07:01:11.427537 4870 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 07:01:11 crc kubenswrapper[4870]: I1014 07:01:11.429777 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:11 crc kubenswrapper[4870]: I1014 07:01:11.429828 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:11 crc kubenswrapper[4870]: I1014 07:01:11.429842 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:11 crc kubenswrapper[4870]: I1014 07:01:11.429872 4870 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 14 07:01:12 crc kubenswrapper[4870]: I1014 07:01:12.132523 4870 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 14 07:01:12 crc kubenswrapper[4870]: I1014 07:01:12.132624 4870 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 07:01:12 crc kubenswrapper[4870]: I1014 07:01:12.132670 4870 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 07:01:12 crc kubenswrapper[4870]: I1014 07:01:12.132630 4870 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 07:01:12 crc kubenswrapper[4870]: I1014 07:01:12.135064 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:12 crc kubenswrapper[4870]: I1014 07:01:12.135110 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:12 crc kubenswrapper[4870]: I1014 07:01:12.135122 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:12 crc kubenswrapper[4870]: I1014 07:01:12.135259 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:12 crc kubenswrapper[4870]: I1014 07:01:12.135276 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:12 crc kubenswrapper[4870]: I1014 07:01:12.135299 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:12 crc kubenswrapper[4870]: I1014 07:01:12.135691 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:12 crc kubenswrapper[4870]: I1014 07:01:12.135717 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:12 crc kubenswrapper[4870]: I1014 07:01:12.135731 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:12 crc kubenswrapper[4870]: I1014 07:01:12.235586 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 14 07:01:12 crc kubenswrapper[4870]: I1014 07:01:12.870555 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 14 07:01:13 crc kubenswrapper[4870]: I1014 07:01:13.134760 4870 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 14 07:01:13 crc kubenswrapper[4870]: I1014 07:01:13.134822 4870 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 07:01:13 crc kubenswrapper[4870]: I1014 07:01:13.134838 4870 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 07:01:13 crc kubenswrapper[4870]: I1014 07:01:13.136431 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:13 crc kubenswrapper[4870]: I1014 07:01:13.136486 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:13 crc kubenswrapper[4870]: I1014 07:01:13.136502 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:13 crc kubenswrapper[4870]: I1014 07:01:13.136538 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:13 crc kubenswrapper[4870]: I1014 07:01:13.136576 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:13 crc kubenswrapper[4870]: I1014 07:01:13.136589 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:13 crc kubenswrapper[4870]: I1014 07:01:13.487117 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 14 07:01:14 crc kubenswrapper[4870]: I1014 07:01:14.138248 4870 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 14 07:01:14 crc kubenswrapper[4870]: I1014 07:01:14.138334 4870 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 07:01:14 crc kubenswrapper[4870]: I1014 07:01:14.139895 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:14 crc kubenswrapper[4870]: I1014 07:01:14.139964 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:14 crc kubenswrapper[4870]: I1014 07:01:14.139986 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:14 crc kubenswrapper[4870]: I1014 07:01:14.500355 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Oct 14 07:01:14 crc kubenswrapper[4870]: I1014 07:01:14.500590 4870 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 07:01:14 crc kubenswrapper[4870]: I1014 07:01:14.501950 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:14 crc kubenswrapper[4870]: I1014 07:01:14.502007 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:14 crc kubenswrapper[4870]: I1014 07:01:14.502026 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:14 crc kubenswrapper[4870]: I1014 07:01:14.512216 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 14 07:01:15 crc kubenswrapper[4870]: E1014 07:01:15.113820 4870 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Oct 14 07:01:15 crc kubenswrapper[4870]: I1014 07:01:15.140102 4870 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 07:01:15 crc kubenswrapper[4870]: I1014 07:01:15.141866 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:15 crc kubenswrapper[4870]: I1014 07:01:15.141922 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:15 crc kubenswrapper[4870]: I1014 07:01:15.141936 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:15 crc kubenswrapper[4870]: I1014 07:01:15.235913 4870 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Oct 14 07:01:15 crc kubenswrapper[4870]: I1014 07:01:15.236028 4870 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 14 07:01:16 crc kubenswrapper[4870]: I1014 07:01:16.111866 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Oct 14 07:01:16 crc kubenswrapper[4870]: I1014 07:01:16.112174 4870 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 07:01:16 crc kubenswrapper[4870]: I1014 07:01:16.113766 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:16 crc kubenswrapper[4870]: I1014 07:01:16.113837 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:16 crc kubenswrapper[4870]: I1014 07:01:16.113856 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:19 crc kubenswrapper[4870]: I1014 07:01:19.959713 4870 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Oct 14 07:01:20 crc kubenswrapper[4870]: I1014 07:01:20.855265 4870 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Oct 14 07:01:20 crc kubenswrapper[4870]: I1014 07:01:20.855718 4870 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Oct 14 07:01:20 crc kubenswrapper[4870]: I1014 07:01:20.860429 4870 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Oct 14 07:01:20 crc kubenswrapper[4870]: I1014 07:01:20.860524 4870 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Oct 14 07:01:21 crc kubenswrapper[4870]: I1014 07:01:21.154246 4870 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Liveness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Oct 14 07:01:21 crc kubenswrapper[4870]: I1014 07:01:21.154314 4870 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Oct 14 07:01:22 crc kubenswrapper[4870]: I1014 07:01:22.874737 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 14 07:01:22 crc kubenswrapper[4870]: I1014 07:01:22.874951 4870 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 07:01:22 crc kubenswrapper[4870]: I1014 07:01:22.875616 4870 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Oct 14 07:01:22 crc kubenswrapper[4870]: I1014 07:01:22.875727 4870 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Oct 14 07:01:22 crc kubenswrapper[4870]: I1014 07:01:22.876723 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:22 crc kubenswrapper[4870]: I1014 07:01:22.876787 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:22 crc kubenswrapper[4870]: I1014 07:01:22.876804 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:22 crc kubenswrapper[4870]: I1014 07:01:22.880512 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 14 07:01:23 crc kubenswrapper[4870]: I1014 07:01:23.161254 4870 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 07:01:23 crc kubenswrapper[4870]: I1014 07:01:23.161843 4870 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Oct 14 07:01:23 crc kubenswrapper[4870]: I1014 07:01:23.161948 4870 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Oct 14 07:01:23 crc kubenswrapper[4870]: I1014 07:01:23.162879 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:23 crc kubenswrapper[4870]: I1014 07:01:23.162945 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:23 crc kubenswrapper[4870]: I1014 07:01:23.162961 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:24 crc kubenswrapper[4870]: I1014 07:01:24.513054 4870 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Oct 14 07:01:24 crc kubenswrapper[4870]: I1014 07:01:24.513118 4870 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Oct 14 07:01:25 crc kubenswrapper[4870]: E1014 07:01:25.113939 4870 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Oct 14 07:01:25 crc kubenswrapper[4870]: I1014 07:01:25.236510 4870 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Oct 14 07:01:25 crc kubenswrapper[4870]: I1014 07:01:25.236823 4870 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Oct 14 07:01:25 crc kubenswrapper[4870]: I1014 07:01:25.765584 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 14 07:01:25 crc kubenswrapper[4870]: I1014 07:01:25.765759 4870 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 07:01:25 crc kubenswrapper[4870]: I1014 07:01:25.767028 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:25 crc kubenswrapper[4870]: I1014 07:01:25.767071 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:25 crc kubenswrapper[4870]: I1014 07:01:25.767082 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:25 crc kubenswrapper[4870]: E1014 07:01:25.850920 4870 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="6.4s" Oct 14 07:01:25 crc kubenswrapper[4870]: I1014 07:01:25.853742 4870 trace.go:236] Trace[1334288525]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (14-Oct-2025 07:01:14.937) (total time: 10915ms): Oct 14 07:01:25 crc kubenswrapper[4870]: Trace[1334288525]: ---"Objects listed" error: 10915ms (07:01:25.853) Oct 14 07:01:25 crc kubenswrapper[4870]: Trace[1334288525]: [10.915956012s] [10.915956012s] END Oct 14 07:01:25 crc kubenswrapper[4870]: I1014 07:01:25.853772 4870 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Oct 14 07:01:25 crc kubenswrapper[4870]: I1014 07:01:25.854186 4870 trace.go:236] Trace[1351289660]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (14-Oct-2025 07:01:12.955) (total time: 12898ms): Oct 14 07:01:25 crc kubenswrapper[4870]: Trace[1351289660]: ---"Objects listed" error: 12897ms (07:01:25.853) Oct 14 07:01:25 crc kubenswrapper[4870]: Trace[1351289660]: [12.898304306s] [12.898304306s] END Oct 14 07:01:25 crc kubenswrapper[4870]: I1014 07:01:25.854334 4870 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Oct 14 07:01:25 crc kubenswrapper[4870]: I1014 07:01:25.858247 4870 trace.go:236] Trace[1944457370]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (14-Oct-2025 07:01:13.117) (total time: 12740ms): Oct 14 07:01:25 crc kubenswrapper[4870]: Trace[1944457370]: ---"Objects listed" error: 12740ms (07:01:25.858) Oct 14 07:01:25 crc kubenswrapper[4870]: Trace[1944457370]: [12.740893689s] [12.740893689s] END Oct 14 07:01:25 crc kubenswrapper[4870]: I1014 07:01:25.858384 4870 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Oct 14 07:01:25 crc kubenswrapper[4870]: E1014 07:01:25.859420 4870 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Oct 14 07:01:25 crc kubenswrapper[4870]: I1014 07:01:25.860830 4870 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Oct 14 07:01:25 crc kubenswrapper[4870]: I1014 07:01:25.861913 4870 trace.go:236] Trace[499279301]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (14-Oct-2025 07:01:15.002) (total time: 10859ms): Oct 14 07:01:25 crc kubenswrapper[4870]: Trace[499279301]: ---"Objects listed" error: 10859ms (07:01:25.861) Oct 14 07:01:25 crc kubenswrapper[4870]: Trace[499279301]: [10.859359999s] [10.859359999s] END Oct 14 07:01:25 crc kubenswrapper[4870]: I1014 07:01:25.862036 4870 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Oct 14 07:01:25 crc kubenswrapper[4870]: I1014 07:01:25.957143 4870 apiserver.go:52] "Watching apiserver" Oct 14 07:01:25 crc kubenswrapper[4870]: I1014 07:01:25.962248 4870 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Oct 14 07:01:25 crc kubenswrapper[4870]: I1014 07:01:25.962567 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf"] Oct 14 07:01:25 crc kubenswrapper[4870]: I1014 07:01:25.963124 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 14 07:01:25 crc kubenswrapper[4870]: I1014 07:01:25.963185 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 07:01:25 crc kubenswrapper[4870]: I1014 07:01:25.963422 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 07:01:25 crc kubenswrapper[4870]: I1014 07:01:25.963354 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 14 07:01:25 crc kubenswrapper[4870]: E1014 07:01:25.963762 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 07:01:25 crc kubenswrapper[4870]: I1014 07:01:25.963241 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 07:01:25 crc kubenswrapper[4870]: I1014 07:01:25.964038 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 14 07:01:25 crc kubenswrapper[4870]: E1014 07:01:25.964157 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 07:01:25 crc kubenswrapper[4870]: I1014 07:01:25.964791 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Oct 14 07:01:25 crc kubenswrapper[4870]: I1014 07:01:25.965485 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Oct 14 07:01:25 crc kubenswrapper[4870]: I1014 07:01:25.965792 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Oct 14 07:01:25 crc kubenswrapper[4870]: I1014 07:01:25.965960 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Oct 14 07:01:25 crc kubenswrapper[4870]: I1014 07:01:25.966244 4870 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Oct 14 07:01:25 crc kubenswrapper[4870]: I1014 07:01:25.966400 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Oct 14 07:01:25 crc kubenswrapper[4870]: I1014 07:01:25.966980 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Oct 14 07:01:25 crc kubenswrapper[4870]: I1014 07:01:25.967067 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Oct 14 07:01:25 crc kubenswrapper[4870]: I1014 07:01:25.966991 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Oct 14 07:01:25 crc kubenswrapper[4870]: E1014 07:01:25.965674 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 07:01:25 crc kubenswrapper[4870]: I1014 07:01:25.973967 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Oct 14 07:01:25 crc kubenswrapper[4870]: I1014 07:01:25.995430 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.015806 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.026393 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.037497 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.057274 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.061546 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.061598 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.061631 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.061656 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.061816 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.061853 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.061883 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.061914 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.061947 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.061992 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.062020 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.062045 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.062091 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.062116 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.062146 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.062170 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.062191 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.062214 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.062254 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.062283 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.062305 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.062329 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.062355 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.062390 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.062418 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.062459 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.062485 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.062508 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.062537 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.062594 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.062625 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.062650 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.062712 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.062738 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.062798 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.062825 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.062851 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.062874 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.062899 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.062924 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.062947 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.062954 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.062971 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.063053 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.063080 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.063100 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.063124 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.063145 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.063165 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.063184 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.063202 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.063225 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.063242 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.063263 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.063287 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.063307 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.063329 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.063348 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.063370 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.063392 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.063414 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.063434 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.063472 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.063492 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.063494 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.063510 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.063533 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.063554 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.063573 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.063592 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.063612 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.063636 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.063663 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.063682 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.063685 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.063754 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.063788 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.063818 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.063881 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.063885 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.063905 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.063937 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.063967 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.063993 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.064017 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.064043 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.064072 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.064098 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.064123 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.064151 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.064178 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.064204 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.064232 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.064261 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.064289 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.064320 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.064345 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.064371 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.064395 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.064422 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.064472 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.064499 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.064524 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.064552 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.064577 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.064603 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.064634 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.064659 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.064683 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.064707 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.064729 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.064752 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.064775 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.064804 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.064835 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.064859 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.064882 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.064907 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.064930 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.064959 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.064987 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.065014 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.065041 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.065066 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.065093 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.065108 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.065119 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.065126 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.065147 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.065259 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.065287 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.065310 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.065349 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.065352 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.065370 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.065393 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.065426 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.065466 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.065486 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.065507 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.065548 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.065568 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.065588 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.065607 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.065622 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.065642 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.065666 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.065887 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.065908 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.065951 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.066174 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.066223 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.067538 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.067571 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.067643 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.067871 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.067914 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.068293 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.068303 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.068507 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.068716 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.068870 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.068919 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.074214 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.074793 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.075228 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.075622 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.075963 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.076045 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.076279 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.076576 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.076619 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.076693 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.077715 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.078055 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.078094 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.078140 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.078177 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.078212 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.078227 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.078248 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.078265 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.078281 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.078361 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.078798 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.078921 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.078407 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.080629 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.080660 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.080708 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.080744 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.080779 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.080808 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.080833 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.080866 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.080898 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.078791 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.088522 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.079509 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.079835 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.080397 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.081104 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.081234 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.081382 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.083294 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.084336 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.084616 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: E1014 07:01:26.084767 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 07:01:26.584717051 +0000 UTC m=+22.282077412 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.088696 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.085057 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.085330 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.085611 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.088839 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.088920 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.088951 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.088972 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.082777 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.089003 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.089027 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.089047 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.089070 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.089096 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.089123 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.089148 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.089170 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.089192 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.089213 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.089237 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.089258 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.089278 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.095372 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.095489 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.095531 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.095561 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.095588 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.095618 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.095652 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.095679 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.095709 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.095737 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.095764 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.095794 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.095820 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.095849 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.095877 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.095905 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.095935 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.095961 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.095990 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.096018 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.096041 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.096066 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.096105 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.096130 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.096201 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.096241 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.096269 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.096461 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.096506 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.096535 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.096563 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.096588 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.096613 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.096644 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.096678 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.096703 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.096728 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.096757 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.096857 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.096880 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.096898 4870 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.096912 4870 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.096926 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.096947 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.096963 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.096977 4870 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.096990 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.097010 4870 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.097024 4870 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.097038 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.097057 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.097072 4870 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.097088 4870 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.097102 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.097120 4870 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.097132 4870 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.097145 4870 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.097157 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.097174 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.097187 4870 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.097199 4870 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.097211 4870 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.097228 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.097241 4870 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.097254 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.097271 4870 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.097287 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.097303 4870 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.097316 4870 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.097333 4870 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.097346 4870 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.097360 4870 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.097374 4870 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.097396 4870 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.097409 4870 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.097423 4870 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.097458 4870 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.097472 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.097484 4870 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.097497 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.097515 4870 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.097528 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.097539 4870 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.097552 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.097570 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.097585 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.097599 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.097611 4870 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.097625 4870 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.097638 4870 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.097650 4870 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.097665 4870 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.097678 4870 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.097691 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.097703 4870 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.097721 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.084661 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.089901 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.085902 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.086018 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.087929 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.088214 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.088252 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.088279 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.088515 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.089954 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.090096 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.090135 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.090296 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.090379 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.093171 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.093223 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.093505 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.093517 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.093684 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.093695 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.093752 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.094101 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.094143 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.094378 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.085682 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.094777 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.094814 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.094808 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.094936 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.094974 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.095074 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.095170 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.084783 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.095590 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.095824 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.095859 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.096335 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.096929 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.096960 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.098706 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.090770 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.100785 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.101487 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.101669 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.101840 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.102194 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.102566 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.102922 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.103740 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.103804 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.103882 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.103979 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.104513 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.104579 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.105202 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.105338 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.105763 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.105848 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.105938 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.105983 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.106009 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.114579 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.107582 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.120620 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 14 07:01:26 crc kubenswrapper[4870]: E1014 07:01:26.127187 4870 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 14 07:01:26 crc kubenswrapper[4870]: E1014 07:01:26.127235 4870 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 14 07:01:26 crc kubenswrapper[4870]: E1014 07:01:26.127252 4870 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 14 07:01:26 crc kubenswrapper[4870]: E1014 07:01:26.127330 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-14 07:01:26.627309362 +0000 UTC m=+22.324669733 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.106146 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.106553 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.106583 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.106691 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: E1014 07:01:26.106707 4870 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.127828 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.148562 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.108790 4870 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.152331 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.152635 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: E1014 07:01:26.152702 4870 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 14 07:01:26 crc kubenswrapper[4870]: E1014 07:01:26.127759 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-14 07:01:26.627729292 +0000 UTC m=+22.325089663 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 14 07:01:26 crc kubenswrapper[4870]: E1014 07:01:26.152735 4870 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 14 07:01:26 crc kubenswrapper[4870]: E1014 07:01:26.108370 4870 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 14 07:01:26 crc kubenswrapper[4870]: E1014 07:01:26.152756 4870 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.108475 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.108690 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.107661 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: E1014 07:01:26.152815 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-14 07:01:26.652788185 +0000 UTC m=+22.350148556 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 14 07:01:26 crc kubenswrapper[4870]: E1014 07:01:26.152862 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-14 07:01:26.652841116 +0000 UTC m=+22.350201487 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.153661 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.154165 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.154180 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.154523 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.107228 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.154726 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.154908 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.155016 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.155169 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.155187 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.155323 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.155563 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.156046 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.156334 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.156533 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.156652 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.156856 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.157453 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.157525 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.157781 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.157811 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.157875 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.157953 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.159877 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.160277 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.161948 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.162425 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.162716 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.127486 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.162949 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.163683 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.163809 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.165105 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.164103 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.162769 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.164210 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.165519 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.178191 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.178738 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.179044 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.179840 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.180046 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.180407 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.180925 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.181022 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.181036 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.181689 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.181882 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.182175 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.182256 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.189781 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.191558 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.191785 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.198165 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.198284 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.198463 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.198646 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.198686 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.198777 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 14 07:01:26 crc kubenswrapper[4870]: W1014 07:01:26.198833 4870 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.198878 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.198847 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.198977 4870 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.198888 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.198994 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.199009 4870 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.199052 4870 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.199074 4870 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.199087 4870 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.199102 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.199119 4870 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.199113 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.199134 4870 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.199151 4870 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.199168 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.199182 4870 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.199195 4870 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.199207 4870 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.199221 4870 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.199235 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.199248 4870 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.199260 4870 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.199273 4870 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.199313 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.199325 4870 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.199336 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.199347 4870 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.199359 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.199371 4870 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.199384 4870 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.199396 4870 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.199408 4870 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.199419 4870 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.199431 4870 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.199460 4870 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.199474 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.199489 4870 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.199502 4870 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.199515 4870 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.199527 4870 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.199539 4870 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.199554 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.199567 4870 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.199580 4870 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.199591 4870 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.199603 4870 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.199614 4870 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.199626 4870 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.199632 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.199637 4870 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.199685 4870 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.199699 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.199711 4870 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.199723 4870 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.199734 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.199745 4870 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.199755 4870 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.199767 4870 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.199779 4870 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.199791 4870 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.199801 4870 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.199815 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.199826 4870 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.199838 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.199850 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.199862 4870 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.199879 4870 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.199893 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.199904 4870 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.199914 4870 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.199924 4870 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.199934 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.199943 4870 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.199955 4870 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.199965 4870 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.199976 4870 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.199988 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.199999 4870 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.200009 4870 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.200017 4870 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.200026 4870 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.200035 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.200044 4870 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.200053 4870 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.200063 4870 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.200072 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.200081 4870 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.200090 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.200100 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.200111 4870 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.200122 4870 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.200130 4870 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.200140 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.200149 4870 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.200159 4870 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.200167 4870 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.200177 4870 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.200186 4870 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.200195 4870 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.200205 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.200213 4870 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.200222 4870 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.200231 4870 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.200238 4870 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.200250 4870 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.200259 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.200268 4870 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.200276 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.200285 4870 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.200296 4870 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.200309 4870 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.200323 4870 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.200335 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.200346 4870 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.200355 4870 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.200367 4870 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.200410 4870 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.200423 4870 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.200453 4870 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.200466 4870 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.200476 4870 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.200488 4870 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.200499 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.200510 4870 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.200521 4870 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.200533 4870 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.200544 4870 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.201085 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: W1014 07:01:26.203069 4870 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.203137 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.206858 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.207597 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.208104 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.209024 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.209834 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.210671 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.211717 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.211817 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.211953 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.211994 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.212251 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.216254 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.217132 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.224368 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.224611 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.226398 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.233094 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.236387 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.243649 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.258155 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.276931 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.288928 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.291207 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.299228 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.299289 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.306073 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.306966 4870 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.306989 4870 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.307005 4870 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.307017 4870 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.307029 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.307039 4870 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.307050 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.307061 4870 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.307072 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.307084 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.307096 4870 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.307106 4870 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.307117 4870 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.307128 4870 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Oct 14 07:01:26 crc kubenswrapper[4870]: W1014 07:01:26.308417 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod37a5e44f_9a88_4405_be8a_b645485e7312.slice/crio-01a890b3732b139575b96579629a58eb8b7e7f05a1043958e8e994a603632302 WatchSource:0}: Error finding container 01a890b3732b139575b96579629a58eb8b7e7f05a1043958e8e994a603632302: Status 404 returned error can't find the container with id 01a890b3732b139575b96579629a58eb8b7e7f05a1043958e8e994a603632302 Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.318345 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 14 07:01:26 crc kubenswrapper[4870]: W1014 07:01:26.324847 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd75a4c96_2883_4a0b_bab2_0fab2b6c0b49.slice/crio-5f1f856410372deedd5fc11fc0c4ff0e9f085d7db1291a62298d7af922b53d43 WatchSource:0}: Error finding container 5f1f856410372deedd5fc11fc0c4ff0e9f085d7db1291a62298d7af922b53d43: Status 404 returned error can't find the container with id 5f1f856410372deedd5fc11fc0c4ff0e9f085d7db1291a62298d7af922b53d43 Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.332366 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.349912 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.357912 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.367519 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.376886 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-4v2v2"] Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.377211 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-4v2v2" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.379878 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.380103 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.382871 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.384984 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.402672 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33a9493e-9660-48db-b8d1-310e29ebe313\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3570643ef1b7d5cc3a542f320413b819338e0c8b7ab7f0f09cc909997a39bab8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb398668197dedc1c2a14b4ae7d6a29714f3527309ccb9c1622af841100e1d60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1d2b6d068d63bd38cdfb83824525f6d12054f6cabab5d684e03ea5f2152076c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://783ff243e395333f3fc80cfae21358c532f0c9c5cb5515586be56b90ed10d912\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a8135a3d01076e4437ae8a784b459bf8d16de57fa16ec7ce4d40c5129c9d4a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48ca3a5017ba650df99b2fc7201c3fe5abfa89d81089c5647cde1b1549f534fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48ca3a5017ba650df99b2fc7201c3fe5abfa89d81089c5647cde1b1549f534fc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://19682eca67f76bf2e06e1b80f12e21141fcbce0acb8195f0defacd1b9a10b5e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://19682eca67f76bf2e06e1b80f12e21141fcbce0acb8195f0defacd1b9a10b5e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://042aad6d5f6bee8eef77f9e6a2c265de775a69f3da72aedc68bdf312198f00fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://042aad6d5f6bee8eef77f9e6a2c265de775a69f3da72aedc68bdf312198f00fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:05Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.407459 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/d4796ba6-c23b-4764-be96-2e3a3bf5e962-hosts-file\") pod \"node-resolver-4v2v2\" (UID: \"d4796ba6-c23b-4764-be96-2e3a3bf5e962\") " pod="openshift-dns/node-resolver-4v2v2" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.407493 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vnp8b\" (UniqueName: \"kubernetes.io/projected/d4796ba6-c23b-4764-be96-2e3a3bf5e962-kube-api-access-vnp8b\") pod \"node-resolver-4v2v2\" (UID: \"d4796ba6-c23b-4764-be96-2e3a3bf5e962\") " pod="openshift-dns/node-resolver-4v2v2" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.413609 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.428068 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.437082 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-4v2v2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4796ba6-c23b-4764-be96-2e3a3bf5e962\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnp8b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-4v2v2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.449798 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.463918 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.475970 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.489582 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.508198 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/d4796ba6-c23b-4764-be96-2e3a3bf5e962-hosts-file\") pod \"node-resolver-4v2v2\" (UID: \"d4796ba6-c23b-4764-be96-2e3a3bf5e962\") " pod="openshift-dns/node-resolver-4v2v2" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.508238 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vnp8b\" (UniqueName: \"kubernetes.io/projected/d4796ba6-c23b-4764-be96-2e3a3bf5e962-kube-api-access-vnp8b\") pod \"node-resolver-4v2v2\" (UID: \"d4796ba6-c23b-4764-be96-2e3a3bf5e962\") " pod="openshift-dns/node-resolver-4v2v2" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.508407 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/d4796ba6-c23b-4764-be96-2e3a3bf5e962-hosts-file\") pod \"node-resolver-4v2v2\" (UID: \"d4796ba6-c23b-4764-be96-2e3a3bf5e962\") " pod="openshift-dns/node-resolver-4v2v2" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.525118 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vnp8b\" (UniqueName: \"kubernetes.io/projected/d4796ba6-c23b-4764-be96-2e3a3bf5e962-kube-api-access-vnp8b\") pod \"node-resolver-4v2v2\" (UID: \"d4796ba6-c23b-4764-be96-2e3a3bf5e962\") " pod="openshift-dns/node-resolver-4v2v2" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.608498 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 07:01:26 crc kubenswrapper[4870]: E1014 07:01:26.608666 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 07:01:27.608643617 +0000 UTC m=+23.306003988 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.701751 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-4v2v2" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.709860 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.709898 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.709917 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.709938 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 07:01:26 crc kubenswrapper[4870]: E1014 07:01:26.709989 4870 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 14 07:01:26 crc kubenswrapper[4870]: E1014 07:01:26.710018 4870 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 14 07:01:26 crc kubenswrapper[4870]: E1014 07:01:26.710035 4870 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 14 07:01:26 crc kubenswrapper[4870]: E1014 07:01:26.710046 4870 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 14 07:01:26 crc kubenswrapper[4870]: E1014 07:01:26.710035 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-14 07:01:27.710022225 +0000 UTC m=+23.407382596 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 14 07:01:26 crc kubenswrapper[4870]: E1014 07:01:26.710171 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-14 07:01:27.710121297 +0000 UTC m=+23.407481668 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 14 07:01:26 crc kubenswrapper[4870]: E1014 07:01:26.710175 4870 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 14 07:01:26 crc kubenswrapper[4870]: E1014 07:01:26.710237 4870 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 14 07:01:26 crc kubenswrapper[4870]: E1014 07:01:26.710300 4870 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 14 07:01:26 crc kubenswrapper[4870]: E1014 07:01:26.710319 4870 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 14 07:01:26 crc kubenswrapper[4870]: E1014 07:01:26.710260 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-14 07:01:27.71024161 +0000 UTC m=+23.407601981 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 14 07:01:26 crc kubenswrapper[4870]: E1014 07:01:26.710421 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-14 07:01:27.710377884 +0000 UTC m=+23.407738265 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 14 07:01:26 crc kubenswrapper[4870]: W1014 07:01:26.714761 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd4796ba6_c23b_4764_be96_2e3a3bf5e962.slice/crio-8ac7f59d6179818248b9f8a64e5668e80532498586333babacbd555138ebf5ea WatchSource:0}: Error finding container 8ac7f59d6179818248b9f8a64e5668e80532498586333babacbd555138ebf5ea: Status 404 returned error can't find the container with id 8ac7f59d6179818248b9f8a64e5668e80532498586333babacbd555138ebf5ea Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.742207 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-cdjmj"] Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.742779 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-7tvc7"] Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.743222 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-fdxjh"] Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.742803 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-cdjmj" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.743395 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-v7br2"] Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.743802 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.743903 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-fdxjh" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.744772 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-v7br2" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.746089 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.746161 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.746280 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.746762 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.747022 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.747118 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.747400 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.747414 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.747655 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.750742 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.751039 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.752915 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.752870 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.753024 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.753245 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.753831 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.754039 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.754646 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.754704 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.777769 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.798631 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.811040 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/8a3a430a-b540-4a8c-adad-b2893dbf9898-host-run-multus-certs\") pod \"multus-fdxjh\" (UID: \"8a3a430a-b540-4a8c-adad-b2893dbf9898\") " pod="openshift-multus/multus-fdxjh" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.811082 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/8dcaef0e-42e3-424b-bbe5-4a023efe912d-run-ovn\") pod \"ovnkube-node-v7br2\" (UID: \"8dcaef0e-42e3-424b-bbe5-4a023efe912d\") " pod="openshift-ovn-kubernetes/ovnkube-node-v7br2" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.811100 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/8dcaef0e-42e3-424b-bbe5-4a023efe912d-run-systemd\") pod \"ovnkube-node-v7br2\" (UID: \"8dcaef0e-42e3-424b-bbe5-4a023efe912d\") " pod="openshift-ovn-kubernetes/ovnkube-node-v7br2" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.811165 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/8dcaef0e-42e3-424b-bbe5-4a023efe912d-host-run-ovn-kubernetes\") pod \"ovnkube-node-v7br2\" (UID: \"8dcaef0e-42e3-424b-bbe5-4a023efe912d\") " pod="openshift-ovn-kubernetes/ovnkube-node-v7br2" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.811247 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/8a3a430a-b540-4a8c-adad-b2893dbf9898-host-run-netns\") pod \"multus-fdxjh\" (UID: \"8a3a430a-b540-4a8c-adad-b2893dbf9898\") " pod="openshift-multus/multus-fdxjh" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.811271 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/8dcaef0e-42e3-424b-bbe5-4a023efe912d-host-run-netns\") pod \"ovnkube-node-v7br2\" (UID: \"8dcaef0e-42e3-424b-bbe5-4a023efe912d\") " pod="openshift-ovn-kubernetes/ovnkube-node-v7br2" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.811310 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/8dcaef0e-42e3-424b-bbe5-4a023efe912d-systemd-units\") pod \"ovnkube-node-v7br2\" (UID: \"8dcaef0e-42e3-424b-bbe5-4a023efe912d\") " pod="openshift-ovn-kubernetes/ovnkube-node-v7br2" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.811331 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/8dcaef0e-42e3-424b-bbe5-4a023efe912d-node-log\") pod \"ovnkube-node-v7br2\" (UID: \"8dcaef0e-42e3-424b-bbe5-4a023efe912d\") " pod="openshift-ovn-kubernetes/ovnkube-node-v7br2" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.811352 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/8dcaef0e-42e3-424b-bbe5-4a023efe912d-log-socket\") pod \"ovnkube-node-v7br2\" (UID: \"8dcaef0e-42e3-424b-bbe5-4a023efe912d\") " pod="openshift-ovn-kubernetes/ovnkube-node-v7br2" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.811383 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/8a3a430a-b540-4a8c-adad-b2893dbf9898-system-cni-dir\") pod \"multus-fdxjh\" (UID: \"8a3a430a-b540-4a8c-adad-b2893dbf9898\") " pod="openshift-multus/multus-fdxjh" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.811405 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/8a3a430a-b540-4a8c-adad-b2893dbf9898-host-var-lib-kubelet\") pod \"multus-fdxjh\" (UID: \"8a3a430a-b540-4a8c-adad-b2893dbf9898\") " pod="openshift-multus/multus-fdxjh" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.811427 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/8dcaef0e-42e3-424b-bbe5-4a023efe912d-var-lib-openvswitch\") pod \"ovnkube-node-v7br2\" (UID: \"8dcaef0e-42e3-424b-bbe5-4a023efe912d\") " pod="openshift-ovn-kubernetes/ovnkube-node-v7br2" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.811540 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/8dcaef0e-42e3-424b-bbe5-4a023efe912d-host-slash\") pod \"ovnkube-node-v7br2\" (UID: \"8dcaef0e-42e3-424b-bbe5-4a023efe912d\") " pod="openshift-ovn-kubernetes/ovnkube-node-v7br2" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.811572 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/8dcaef0e-42e3-424b-bbe5-4a023efe912d-host-cni-bin\") pod \"ovnkube-node-v7br2\" (UID: \"8dcaef0e-42e3-424b-bbe5-4a023efe912d\") " pod="openshift-ovn-kubernetes/ovnkube-node-v7br2" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.811596 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fcfq9\" (UniqueName: \"kubernetes.io/projected/612e7ab9-bda4-45d8-bc50-ff1afb68aace-kube-api-access-fcfq9\") pod \"multus-additional-cni-plugins-cdjmj\" (UID: \"612e7ab9-bda4-45d8-bc50-ff1afb68aace\") " pod="openshift-multus/multus-additional-cni-plugins-cdjmj" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.811621 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/8a3a430a-b540-4a8c-adad-b2893dbf9898-host-var-lib-cni-bin\") pod \"multus-fdxjh\" (UID: \"8a3a430a-b540-4a8c-adad-b2893dbf9898\") " pod="openshift-multus/multus-fdxjh" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.811640 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/8a3a430a-b540-4a8c-adad-b2893dbf9898-etc-kubernetes\") pod \"multus-fdxjh\" (UID: \"8a3a430a-b540-4a8c-adad-b2893dbf9898\") " pod="openshift-multus/multus-fdxjh" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.811658 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/8dcaef0e-42e3-424b-bbe5-4a023efe912d-run-openvswitch\") pod \"ovnkube-node-v7br2\" (UID: \"8dcaef0e-42e3-424b-bbe5-4a023efe912d\") " pod="openshift-ovn-kubernetes/ovnkube-node-v7br2" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.811681 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-glrll\" (UniqueName: \"kubernetes.io/projected/8dcaef0e-42e3-424b-bbe5-4a023efe912d-kube-api-access-glrll\") pod \"ovnkube-node-v7br2\" (UID: \"8dcaef0e-42e3-424b-bbe5-4a023efe912d\") " pod="openshift-ovn-kubernetes/ovnkube-node-v7br2" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.811705 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/612e7ab9-bda4-45d8-bc50-ff1afb68aace-tuning-conf-dir\") pod \"multus-additional-cni-plugins-cdjmj\" (UID: \"612e7ab9-bda4-45d8-bc50-ff1afb68aace\") " pod="openshift-multus/multus-additional-cni-plugins-cdjmj" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.811723 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/612e7ab9-bda4-45d8-bc50-ff1afb68aace-cnibin\") pod \"multus-additional-cni-plugins-cdjmj\" (UID: \"612e7ab9-bda4-45d8-bc50-ff1afb68aace\") " pod="openshift-multus/multus-additional-cni-plugins-cdjmj" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.811770 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/8a3a430a-b540-4a8c-adad-b2893dbf9898-multus-conf-dir\") pod \"multus-fdxjh\" (UID: \"8a3a430a-b540-4a8c-adad-b2893dbf9898\") " pod="openshift-multus/multus-fdxjh" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.811794 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/8dcaef0e-42e3-424b-bbe5-4a023efe912d-host-kubelet\") pod \"ovnkube-node-v7br2\" (UID: \"8dcaef0e-42e3-424b-bbe5-4a023efe912d\") " pod="openshift-ovn-kubernetes/ovnkube-node-v7br2" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.811845 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/874a5e45-dffd-4d17-b609-b7d3ed2eab07-rootfs\") pod \"machine-config-daemon-7tvc7\" (UID: \"874a5e45-dffd-4d17-b609-b7d3ed2eab07\") " pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.811861 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wj4ks\" (UniqueName: \"kubernetes.io/projected/874a5e45-dffd-4d17-b609-b7d3ed2eab07-kube-api-access-wj4ks\") pod \"machine-config-daemon-7tvc7\" (UID: \"874a5e45-dffd-4d17-b609-b7d3ed2eab07\") " pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.811876 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/8a3a430a-b540-4a8c-adad-b2893dbf9898-hostroot\") pod \"multus-fdxjh\" (UID: \"8a3a430a-b540-4a8c-adad-b2893dbf9898\") " pod="openshift-multus/multus-fdxjh" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.811926 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/8a3a430a-b540-4a8c-adad-b2893dbf9898-multus-daemon-config\") pod \"multus-fdxjh\" (UID: \"8a3a430a-b540-4a8c-adad-b2893dbf9898\") " pod="openshift-multus/multus-fdxjh" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.811967 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/8dcaef0e-42e3-424b-bbe5-4a023efe912d-ovn-node-metrics-cert\") pod \"ovnkube-node-v7br2\" (UID: \"8dcaef0e-42e3-424b-bbe5-4a023efe912d\") " pod="openshift-ovn-kubernetes/ovnkube-node-v7br2" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.811991 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/8a3a430a-b540-4a8c-adad-b2893dbf9898-host-var-lib-cni-multus\") pod \"multus-fdxjh\" (UID: \"8a3a430a-b540-4a8c-adad-b2893dbf9898\") " pod="openshift-multus/multus-fdxjh" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.812013 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/8dcaef0e-42e3-424b-bbe5-4a023efe912d-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-v7br2\" (UID: \"8dcaef0e-42e3-424b-bbe5-4a023efe912d\") " pod="openshift-ovn-kubernetes/ovnkube-node-v7br2" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.812035 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/8dcaef0e-42e3-424b-bbe5-4a023efe912d-ovnkube-script-lib\") pod \"ovnkube-node-v7br2\" (UID: \"8dcaef0e-42e3-424b-bbe5-4a023efe912d\") " pod="openshift-ovn-kubernetes/ovnkube-node-v7br2" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.812055 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/612e7ab9-bda4-45d8-bc50-ff1afb68aace-system-cni-dir\") pod \"multus-additional-cni-plugins-cdjmj\" (UID: \"612e7ab9-bda4-45d8-bc50-ff1afb68aace\") " pod="openshift-multus/multus-additional-cni-plugins-cdjmj" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.812081 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/8a3a430a-b540-4a8c-adad-b2893dbf9898-os-release\") pod \"multus-fdxjh\" (UID: \"8a3a430a-b540-4a8c-adad-b2893dbf9898\") " pod="openshift-multus/multus-fdxjh" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.812101 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/8a3a430a-b540-4a8c-adad-b2893dbf9898-host-run-k8s-cni-cncf-io\") pod \"multus-fdxjh\" (UID: \"8a3a430a-b540-4a8c-adad-b2893dbf9898\") " pod="openshift-multus/multus-fdxjh" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.812156 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/874a5e45-dffd-4d17-b609-b7d3ed2eab07-mcd-auth-proxy-config\") pod \"machine-config-daemon-7tvc7\" (UID: \"874a5e45-dffd-4d17-b609-b7d3ed2eab07\") " pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.812179 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/8a3a430a-b540-4a8c-adad-b2893dbf9898-cnibin\") pod \"multus-fdxjh\" (UID: \"8a3a430a-b540-4a8c-adad-b2893dbf9898\") " pod="openshift-multus/multus-fdxjh" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.812234 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/8a3a430a-b540-4a8c-adad-b2893dbf9898-multus-cni-dir\") pod \"multus-fdxjh\" (UID: \"8a3a430a-b540-4a8c-adad-b2893dbf9898\") " pod="openshift-multus/multus-fdxjh" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.812254 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/612e7ab9-bda4-45d8-bc50-ff1afb68aace-cni-binary-copy\") pod \"multus-additional-cni-plugins-cdjmj\" (UID: \"612e7ab9-bda4-45d8-bc50-ff1afb68aace\") " pod="openshift-multus/multus-additional-cni-plugins-cdjmj" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.812307 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/8dcaef0e-42e3-424b-bbe5-4a023efe912d-host-cni-netd\") pod \"ovnkube-node-v7br2\" (UID: \"8dcaef0e-42e3-424b-bbe5-4a023efe912d\") " pod="openshift-ovn-kubernetes/ovnkube-node-v7br2" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.812326 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/8dcaef0e-42e3-424b-bbe5-4a023efe912d-ovnkube-config\") pod \"ovnkube-node-v7br2\" (UID: \"8dcaef0e-42e3-424b-bbe5-4a023efe912d\") " pod="openshift-ovn-kubernetes/ovnkube-node-v7br2" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.812390 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/874a5e45-dffd-4d17-b609-b7d3ed2eab07-proxy-tls\") pod \"machine-config-daemon-7tvc7\" (UID: \"874a5e45-dffd-4d17-b609-b7d3ed2eab07\") " pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.812456 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/8a3a430a-b540-4a8c-adad-b2893dbf9898-multus-socket-dir-parent\") pod \"multus-fdxjh\" (UID: \"8a3a430a-b540-4a8c-adad-b2893dbf9898\") " pod="openshift-multus/multus-fdxjh" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.812483 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/612e7ab9-bda4-45d8-bc50-ff1afb68aace-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-cdjmj\" (UID: \"612e7ab9-bda4-45d8-bc50-ff1afb68aace\") " pod="openshift-multus/multus-additional-cni-plugins-cdjmj" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.812508 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/8a3a430a-b540-4a8c-adad-b2893dbf9898-cni-binary-copy\") pod \"multus-fdxjh\" (UID: \"8a3a430a-b540-4a8c-adad-b2893dbf9898\") " pod="openshift-multus/multus-fdxjh" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.812530 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/8dcaef0e-42e3-424b-bbe5-4a023efe912d-env-overrides\") pod \"ovnkube-node-v7br2\" (UID: \"8dcaef0e-42e3-424b-bbe5-4a023efe912d\") " pod="openshift-ovn-kubernetes/ovnkube-node-v7br2" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.812551 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/612e7ab9-bda4-45d8-bc50-ff1afb68aace-os-release\") pod \"multus-additional-cni-plugins-cdjmj\" (UID: \"612e7ab9-bda4-45d8-bc50-ff1afb68aace\") " pod="openshift-multus/multus-additional-cni-plugins-cdjmj" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.812569 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/8dcaef0e-42e3-424b-bbe5-4a023efe912d-etc-openvswitch\") pod \"ovnkube-node-v7br2\" (UID: \"8dcaef0e-42e3-424b-bbe5-4a023efe912d\") " pod="openshift-ovn-kubernetes/ovnkube-node-v7br2" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.812600 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k787j\" (UniqueName: \"kubernetes.io/projected/8a3a430a-b540-4a8c-adad-b2893dbf9898-kube-api-access-k787j\") pod \"multus-fdxjh\" (UID: \"8a3a430a-b540-4a8c-adad-b2893dbf9898\") " pod="openshift-multus/multus-fdxjh" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.814306 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.835736 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.852799 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.871619 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.889242 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-4v2v2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4796ba6-c23b-4764-be96-2e3a3bf5e962\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnp8b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-4v2v2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.906502 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-cdjmj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"612e7ab9-bda4-45d8-bc50-ff1afb68aace\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-cdjmj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.913789 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/8a3a430a-b540-4a8c-adad-b2893dbf9898-os-release\") pod \"multus-fdxjh\" (UID: \"8a3a430a-b540-4a8c-adad-b2893dbf9898\") " pod="openshift-multus/multus-fdxjh" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.913839 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/8a3a430a-b540-4a8c-adad-b2893dbf9898-host-run-k8s-cni-cncf-io\") pod \"multus-fdxjh\" (UID: \"8a3a430a-b540-4a8c-adad-b2893dbf9898\") " pod="openshift-multus/multus-fdxjh" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.913869 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/874a5e45-dffd-4d17-b609-b7d3ed2eab07-mcd-auth-proxy-config\") pod \"machine-config-daemon-7tvc7\" (UID: \"874a5e45-dffd-4d17-b609-b7d3ed2eab07\") " pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.913885 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/8a3a430a-b540-4a8c-adad-b2893dbf9898-cnibin\") pod \"multus-fdxjh\" (UID: \"8a3a430a-b540-4a8c-adad-b2893dbf9898\") " pod="openshift-multus/multus-fdxjh" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.913901 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/8a3a430a-b540-4a8c-adad-b2893dbf9898-multus-cni-dir\") pod \"multus-fdxjh\" (UID: \"8a3a430a-b540-4a8c-adad-b2893dbf9898\") " pod="openshift-multus/multus-fdxjh" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.913919 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/612e7ab9-bda4-45d8-bc50-ff1afb68aace-cni-binary-copy\") pod \"multus-additional-cni-plugins-cdjmj\" (UID: \"612e7ab9-bda4-45d8-bc50-ff1afb68aace\") " pod="openshift-multus/multus-additional-cni-plugins-cdjmj" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.913938 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/874a5e45-dffd-4d17-b609-b7d3ed2eab07-proxy-tls\") pod \"machine-config-daemon-7tvc7\" (UID: \"874a5e45-dffd-4d17-b609-b7d3ed2eab07\") " pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.913953 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/8a3a430a-b540-4a8c-adad-b2893dbf9898-multus-socket-dir-parent\") pod \"multus-fdxjh\" (UID: \"8a3a430a-b540-4a8c-adad-b2893dbf9898\") " pod="openshift-multus/multus-fdxjh" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.913968 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/8dcaef0e-42e3-424b-bbe5-4a023efe912d-host-cni-netd\") pod \"ovnkube-node-v7br2\" (UID: \"8dcaef0e-42e3-424b-bbe5-4a023efe912d\") " pod="openshift-ovn-kubernetes/ovnkube-node-v7br2" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.913985 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/8dcaef0e-42e3-424b-bbe5-4a023efe912d-ovnkube-config\") pod \"ovnkube-node-v7br2\" (UID: \"8dcaef0e-42e3-424b-bbe5-4a023efe912d\") " pod="openshift-ovn-kubernetes/ovnkube-node-v7br2" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.914001 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/8a3a430a-b540-4a8c-adad-b2893dbf9898-cni-binary-copy\") pod \"multus-fdxjh\" (UID: \"8a3a430a-b540-4a8c-adad-b2893dbf9898\") " pod="openshift-multus/multus-fdxjh" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.914017 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/8dcaef0e-42e3-424b-bbe5-4a023efe912d-env-overrides\") pod \"ovnkube-node-v7br2\" (UID: \"8dcaef0e-42e3-424b-bbe5-4a023efe912d\") " pod="openshift-ovn-kubernetes/ovnkube-node-v7br2" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.914032 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/612e7ab9-bda4-45d8-bc50-ff1afb68aace-os-release\") pod \"multus-additional-cni-plugins-cdjmj\" (UID: \"612e7ab9-bda4-45d8-bc50-ff1afb68aace\") " pod="openshift-multus/multus-additional-cni-plugins-cdjmj" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.914047 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/612e7ab9-bda4-45d8-bc50-ff1afb68aace-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-cdjmj\" (UID: \"612e7ab9-bda4-45d8-bc50-ff1afb68aace\") " pod="openshift-multus/multus-additional-cni-plugins-cdjmj" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.914065 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k787j\" (UniqueName: \"kubernetes.io/projected/8a3a430a-b540-4a8c-adad-b2893dbf9898-kube-api-access-k787j\") pod \"multus-fdxjh\" (UID: \"8a3a430a-b540-4a8c-adad-b2893dbf9898\") " pod="openshift-multus/multus-fdxjh" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.914080 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/8dcaef0e-42e3-424b-bbe5-4a023efe912d-etc-openvswitch\") pod \"ovnkube-node-v7br2\" (UID: \"8dcaef0e-42e3-424b-bbe5-4a023efe912d\") " pod="openshift-ovn-kubernetes/ovnkube-node-v7br2" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.914098 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/8a3a430a-b540-4a8c-adad-b2893dbf9898-host-run-multus-certs\") pod \"multus-fdxjh\" (UID: \"8a3a430a-b540-4a8c-adad-b2893dbf9898\") " pod="openshift-multus/multus-fdxjh" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.914124 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/8dcaef0e-42e3-424b-bbe5-4a023efe912d-run-ovn\") pod \"ovnkube-node-v7br2\" (UID: \"8dcaef0e-42e3-424b-bbe5-4a023efe912d\") " pod="openshift-ovn-kubernetes/ovnkube-node-v7br2" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.914153 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/8a3a430a-b540-4a8c-adad-b2893dbf9898-host-run-netns\") pod \"multus-fdxjh\" (UID: \"8a3a430a-b540-4a8c-adad-b2893dbf9898\") " pod="openshift-multus/multus-fdxjh" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.914177 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/8dcaef0e-42e3-424b-bbe5-4a023efe912d-host-run-netns\") pod \"ovnkube-node-v7br2\" (UID: \"8dcaef0e-42e3-424b-bbe5-4a023efe912d\") " pod="openshift-ovn-kubernetes/ovnkube-node-v7br2" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.914200 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/8dcaef0e-42e3-424b-bbe5-4a023efe912d-run-systemd\") pod \"ovnkube-node-v7br2\" (UID: \"8dcaef0e-42e3-424b-bbe5-4a023efe912d\") " pod="openshift-ovn-kubernetes/ovnkube-node-v7br2" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.914223 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/8dcaef0e-42e3-424b-bbe5-4a023efe912d-host-run-ovn-kubernetes\") pod \"ovnkube-node-v7br2\" (UID: \"8dcaef0e-42e3-424b-bbe5-4a023efe912d\") " pod="openshift-ovn-kubernetes/ovnkube-node-v7br2" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.914262 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/8dcaef0e-42e3-424b-bbe5-4a023efe912d-systemd-units\") pod \"ovnkube-node-v7br2\" (UID: \"8dcaef0e-42e3-424b-bbe5-4a023efe912d\") " pod="openshift-ovn-kubernetes/ovnkube-node-v7br2" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.914287 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/8a3a430a-b540-4a8c-adad-b2893dbf9898-system-cni-dir\") pod \"multus-fdxjh\" (UID: \"8a3a430a-b540-4a8c-adad-b2893dbf9898\") " pod="openshift-multus/multus-fdxjh" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.914310 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/8a3a430a-b540-4a8c-adad-b2893dbf9898-host-var-lib-kubelet\") pod \"multus-fdxjh\" (UID: \"8a3a430a-b540-4a8c-adad-b2893dbf9898\") " pod="openshift-multus/multus-fdxjh" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.914368 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/8dcaef0e-42e3-424b-bbe5-4a023efe912d-var-lib-openvswitch\") pod \"ovnkube-node-v7br2\" (UID: \"8dcaef0e-42e3-424b-bbe5-4a023efe912d\") " pod="openshift-ovn-kubernetes/ovnkube-node-v7br2" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.914393 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/8dcaef0e-42e3-424b-bbe5-4a023efe912d-node-log\") pod \"ovnkube-node-v7br2\" (UID: \"8dcaef0e-42e3-424b-bbe5-4a023efe912d\") " pod="openshift-ovn-kubernetes/ovnkube-node-v7br2" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.914419 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/8dcaef0e-42e3-424b-bbe5-4a023efe912d-log-socket\") pod \"ovnkube-node-v7br2\" (UID: \"8dcaef0e-42e3-424b-bbe5-4a023efe912d\") " pod="openshift-ovn-kubernetes/ovnkube-node-v7br2" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.914417 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/8dcaef0e-42e3-424b-bbe5-4a023efe912d-host-run-netns\") pod \"ovnkube-node-v7br2\" (UID: \"8dcaef0e-42e3-424b-bbe5-4a023efe912d\") " pod="openshift-ovn-kubernetes/ovnkube-node-v7br2" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.914474 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/8dcaef0e-42e3-424b-bbe5-4a023efe912d-host-slash\") pod \"ovnkube-node-v7br2\" (UID: \"8dcaef0e-42e3-424b-bbe5-4a023efe912d\") " pod="openshift-ovn-kubernetes/ovnkube-node-v7br2" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.914527 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/8dcaef0e-42e3-424b-bbe5-4a023efe912d-host-cni-bin\") pod \"ovnkube-node-v7br2\" (UID: \"8dcaef0e-42e3-424b-bbe5-4a023efe912d\") " pod="openshift-ovn-kubernetes/ovnkube-node-v7br2" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.914553 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/8dcaef0e-42e3-424b-bbe5-4a023efe912d-run-systemd\") pod \"ovnkube-node-v7br2\" (UID: \"8dcaef0e-42e3-424b-bbe5-4a023efe912d\") " pod="openshift-ovn-kubernetes/ovnkube-node-v7br2" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.914554 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fcfq9\" (UniqueName: \"kubernetes.io/projected/612e7ab9-bda4-45d8-bc50-ff1afb68aace-kube-api-access-fcfq9\") pod \"multus-additional-cni-plugins-cdjmj\" (UID: \"612e7ab9-bda4-45d8-bc50-ff1afb68aace\") " pod="openshift-multus/multus-additional-cni-plugins-cdjmj" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.914606 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/8a3a430a-b540-4a8c-adad-b2893dbf9898-etc-kubernetes\") pod \"multus-fdxjh\" (UID: \"8a3a430a-b540-4a8c-adad-b2893dbf9898\") " pod="openshift-multus/multus-fdxjh" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.914625 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/8dcaef0e-42e3-424b-bbe5-4a023efe912d-run-openvswitch\") pod \"ovnkube-node-v7br2\" (UID: \"8dcaef0e-42e3-424b-bbe5-4a023efe912d\") " pod="openshift-ovn-kubernetes/ovnkube-node-v7br2" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.914648 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/8a3a430a-b540-4a8c-adad-b2893dbf9898-multus-socket-dir-parent\") pod \"multus-fdxjh\" (UID: \"8a3a430a-b540-4a8c-adad-b2893dbf9898\") " pod="openshift-multus/multus-fdxjh" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.914701 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/8a3a430a-b540-4a8c-adad-b2893dbf9898-multus-cni-dir\") pod \"multus-fdxjh\" (UID: \"8a3a430a-b540-4a8c-adad-b2893dbf9898\") " pod="openshift-multus/multus-fdxjh" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.914665 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-glrll\" (UniqueName: \"kubernetes.io/projected/8dcaef0e-42e3-424b-bbe5-4a023efe912d-kube-api-access-glrll\") pod \"ovnkube-node-v7br2\" (UID: \"8dcaef0e-42e3-424b-bbe5-4a023efe912d\") " pod="openshift-ovn-kubernetes/ovnkube-node-v7br2" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.914775 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/612e7ab9-bda4-45d8-bc50-ff1afb68aace-tuning-conf-dir\") pod \"multus-additional-cni-plugins-cdjmj\" (UID: \"612e7ab9-bda4-45d8-bc50-ff1afb68aace\") " pod="openshift-multus/multus-additional-cni-plugins-cdjmj" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.914815 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/8a3a430a-b540-4a8c-adad-b2893dbf9898-host-var-lib-cni-bin\") pod \"multus-fdxjh\" (UID: \"8a3a430a-b540-4a8c-adad-b2893dbf9898\") " pod="openshift-multus/multus-fdxjh" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.914844 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/8dcaef0e-42e3-424b-bbe5-4a023efe912d-run-ovn\") pod \"ovnkube-node-v7br2\" (UID: \"8dcaef0e-42e3-424b-bbe5-4a023efe912d\") " pod="openshift-ovn-kubernetes/ovnkube-node-v7br2" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.914849 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/8a3a430a-b540-4a8c-adad-b2893dbf9898-multus-conf-dir\") pod \"multus-fdxjh\" (UID: \"8a3a430a-b540-4a8c-adad-b2893dbf9898\") " pod="openshift-multus/multus-fdxjh" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.914883 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/8a3a430a-b540-4a8c-adad-b2893dbf9898-multus-conf-dir\") pod \"multus-fdxjh\" (UID: \"8a3a430a-b540-4a8c-adad-b2893dbf9898\") " pod="openshift-multus/multus-fdxjh" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.914886 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/8dcaef0e-42e3-424b-bbe5-4a023efe912d-host-kubelet\") pod \"ovnkube-node-v7br2\" (UID: \"8dcaef0e-42e3-424b-bbe5-4a023efe912d\") " pod="openshift-ovn-kubernetes/ovnkube-node-v7br2" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.914901 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/8dcaef0e-42e3-424b-bbe5-4a023efe912d-host-kubelet\") pod \"ovnkube-node-v7br2\" (UID: \"8dcaef0e-42e3-424b-bbe5-4a023efe912d\") " pod="openshift-ovn-kubernetes/ovnkube-node-v7br2" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.914921 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/612e7ab9-bda4-45d8-bc50-ff1afb68aace-cnibin\") pod \"multus-additional-cni-plugins-cdjmj\" (UID: \"612e7ab9-bda4-45d8-bc50-ff1afb68aace\") " pod="openshift-multus/multus-additional-cni-plugins-cdjmj" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.914944 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/8dcaef0e-42e3-424b-bbe5-4a023efe912d-host-cni-bin\") pod \"ovnkube-node-v7br2\" (UID: \"8dcaef0e-42e3-424b-bbe5-4a023efe912d\") " pod="openshift-ovn-kubernetes/ovnkube-node-v7br2" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.914531 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/8dcaef0e-42e3-424b-bbe5-4a023efe912d-host-slash\") pod \"ovnkube-node-v7br2\" (UID: \"8dcaef0e-42e3-424b-bbe5-4a023efe912d\") " pod="openshift-ovn-kubernetes/ovnkube-node-v7br2" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.914981 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/8dcaef0e-42e3-424b-bbe5-4a023efe912d-systemd-units\") pod \"ovnkube-node-v7br2\" (UID: \"8dcaef0e-42e3-424b-bbe5-4a023efe912d\") " pod="openshift-ovn-kubernetes/ovnkube-node-v7br2" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.914951 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wj4ks\" (UniqueName: \"kubernetes.io/projected/874a5e45-dffd-4d17-b609-b7d3ed2eab07-kube-api-access-wj4ks\") pod \"machine-config-daemon-7tvc7\" (UID: \"874a5e45-dffd-4d17-b609-b7d3ed2eab07\") " pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.914999 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/8a3a430a-b540-4a8c-adad-b2893dbf9898-system-cni-dir\") pod \"multus-fdxjh\" (UID: \"8a3a430a-b540-4a8c-adad-b2893dbf9898\") " pod="openshift-multus/multus-fdxjh" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.915014 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/8a3a430a-b540-4a8c-adad-b2893dbf9898-hostroot\") pod \"multus-fdxjh\" (UID: \"8a3a430a-b540-4a8c-adad-b2893dbf9898\") " pod="openshift-multus/multus-fdxjh" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.915035 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/8a3a430a-b540-4a8c-adad-b2893dbf9898-multus-daemon-config\") pod \"multus-fdxjh\" (UID: \"8a3a430a-b540-4a8c-adad-b2893dbf9898\") " pod="openshift-multus/multus-fdxjh" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.915054 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/874a5e45-dffd-4d17-b609-b7d3ed2eab07-rootfs\") pod \"machine-config-daemon-7tvc7\" (UID: \"874a5e45-dffd-4d17-b609-b7d3ed2eab07\") " pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.915059 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/8dcaef0e-42e3-424b-bbe5-4a023efe912d-var-lib-openvswitch\") pod \"ovnkube-node-v7br2\" (UID: \"8dcaef0e-42e3-424b-bbe5-4a023efe912d\") " pod="openshift-ovn-kubernetes/ovnkube-node-v7br2" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.915064 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/8a3a430a-b540-4a8c-adad-b2893dbf9898-cni-binary-copy\") pod \"multus-fdxjh\" (UID: \"8a3a430a-b540-4a8c-adad-b2893dbf9898\") " pod="openshift-multus/multus-fdxjh" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.915100 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/8dcaef0e-42e3-424b-bbe5-4a023efe912d-ovn-node-metrics-cert\") pod \"ovnkube-node-v7br2\" (UID: \"8dcaef0e-42e3-424b-bbe5-4a023efe912d\") " pod="openshift-ovn-kubernetes/ovnkube-node-v7br2" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.914950 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/8dcaef0e-42e3-424b-bbe5-4a023efe912d-host-run-ovn-kubernetes\") pod \"ovnkube-node-v7br2\" (UID: \"8dcaef0e-42e3-424b-bbe5-4a023efe912d\") " pod="openshift-ovn-kubernetes/ovnkube-node-v7br2" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.915102 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/8a3a430a-b540-4a8c-adad-b2893dbf9898-hostroot\") pod \"multus-fdxjh\" (UID: \"8a3a430a-b540-4a8c-adad-b2893dbf9898\") " pod="openshift-multus/multus-fdxjh" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.915143 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/8dcaef0e-42e3-424b-bbe5-4a023efe912d-node-log\") pod \"ovnkube-node-v7br2\" (UID: \"8dcaef0e-42e3-424b-bbe5-4a023efe912d\") " pod="openshift-ovn-kubernetes/ovnkube-node-v7br2" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.915173 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/8dcaef0e-42e3-424b-bbe5-4a023efe912d-host-cni-netd\") pod \"ovnkube-node-v7br2\" (UID: \"8dcaef0e-42e3-424b-bbe5-4a023efe912d\") " pod="openshift-ovn-kubernetes/ovnkube-node-v7br2" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.915037 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/8a3a430a-b540-4a8c-adad-b2893dbf9898-host-var-lib-kubelet\") pod \"multus-fdxjh\" (UID: \"8a3a430a-b540-4a8c-adad-b2893dbf9898\") " pod="openshift-multus/multus-fdxjh" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.915431 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/8a3a430a-b540-4a8c-adad-b2893dbf9898-etc-kubernetes\") pod \"multus-fdxjh\" (UID: \"8a3a430a-b540-4a8c-adad-b2893dbf9898\") " pod="openshift-multus/multus-fdxjh" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.915086 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/874a5e45-dffd-4d17-b609-b7d3ed2eab07-rootfs\") pod \"machine-config-daemon-7tvc7\" (UID: \"874a5e45-dffd-4d17-b609-b7d3ed2eab07\") " pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.915501 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/8dcaef0e-42e3-424b-bbe5-4a023efe912d-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-v7br2\" (UID: \"8dcaef0e-42e3-424b-bbe5-4a023efe912d\") " pod="openshift-ovn-kubernetes/ovnkube-node-v7br2" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.915534 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/8dcaef0e-42e3-424b-bbe5-4a023efe912d-ovnkube-script-lib\") pod \"ovnkube-node-v7br2\" (UID: \"8dcaef0e-42e3-424b-bbe5-4a023efe912d\") " pod="openshift-ovn-kubernetes/ovnkube-node-v7br2" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.915561 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/612e7ab9-bda4-45d8-bc50-ff1afb68aace-system-cni-dir\") pod \"multus-additional-cni-plugins-cdjmj\" (UID: \"612e7ab9-bda4-45d8-bc50-ff1afb68aace\") " pod="openshift-multus/multus-additional-cni-plugins-cdjmj" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.915634 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/8a3a430a-b540-4a8c-adad-b2893dbf9898-host-var-lib-cni-multus\") pod \"multus-fdxjh\" (UID: \"8a3a430a-b540-4a8c-adad-b2893dbf9898\") " pod="openshift-multus/multus-fdxjh" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.915698 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/8a3a430a-b540-4a8c-adad-b2893dbf9898-host-var-lib-cni-multus\") pod \"multus-fdxjh\" (UID: \"8a3a430a-b540-4a8c-adad-b2893dbf9898\") " pod="openshift-multus/multus-fdxjh" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.915726 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/612e7ab9-bda4-45d8-bc50-ff1afb68aace-system-cni-dir\") pod \"multus-additional-cni-plugins-cdjmj\" (UID: \"612e7ab9-bda4-45d8-bc50-ff1afb68aace\") " pod="openshift-multus/multus-additional-cni-plugins-cdjmj" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.915747 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/8dcaef0e-42e3-424b-bbe5-4a023efe912d-log-socket\") pod \"ovnkube-node-v7br2\" (UID: \"8dcaef0e-42e3-424b-bbe5-4a023efe912d\") " pod="openshift-ovn-kubernetes/ovnkube-node-v7br2" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.915771 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/8dcaef0e-42e3-424b-bbe5-4a023efe912d-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-v7br2\" (UID: \"8dcaef0e-42e3-424b-bbe5-4a023efe912d\") " pod="openshift-ovn-kubernetes/ovnkube-node-v7br2" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.915789 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/8a3a430a-b540-4a8c-adad-b2893dbf9898-multus-daemon-config\") pod \"multus-fdxjh\" (UID: \"8a3a430a-b540-4a8c-adad-b2893dbf9898\") " pod="openshift-multus/multus-fdxjh" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.915834 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/8a3a430a-b540-4a8c-adad-b2893dbf9898-host-var-lib-cni-bin\") pod \"multus-fdxjh\" (UID: \"8a3a430a-b540-4a8c-adad-b2893dbf9898\") " pod="openshift-multus/multus-fdxjh" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.915867 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/8a3a430a-b540-4a8c-adad-b2893dbf9898-cnibin\") pod \"multus-fdxjh\" (UID: \"8a3a430a-b540-4a8c-adad-b2893dbf9898\") " pod="openshift-multus/multus-fdxjh" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.915989 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/8a3a430a-b540-4a8c-adad-b2893dbf9898-host-run-multus-certs\") pod \"multus-fdxjh\" (UID: \"8a3a430a-b540-4a8c-adad-b2893dbf9898\") " pod="openshift-multus/multus-fdxjh" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.916008 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/874a5e45-dffd-4d17-b609-b7d3ed2eab07-mcd-auth-proxy-config\") pod \"machine-config-daemon-7tvc7\" (UID: \"874a5e45-dffd-4d17-b609-b7d3ed2eab07\") " pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.916052 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/612e7ab9-bda4-45d8-bc50-ff1afb68aace-cnibin\") pod \"multus-additional-cni-plugins-cdjmj\" (UID: \"612e7ab9-bda4-45d8-bc50-ff1afb68aace\") " pod="openshift-multus/multus-additional-cni-plugins-cdjmj" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.914925 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/8a3a430a-b540-4a8c-adad-b2893dbf9898-host-run-netns\") pod \"multus-fdxjh\" (UID: \"8a3a430a-b540-4a8c-adad-b2893dbf9898\") " pod="openshift-multus/multus-fdxjh" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.916092 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/8dcaef0e-42e3-424b-bbe5-4a023efe912d-ovnkube-config\") pod \"ovnkube-node-v7br2\" (UID: \"8dcaef0e-42e3-424b-bbe5-4a023efe912d\") " pod="openshift-ovn-kubernetes/ovnkube-node-v7br2" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.916156 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/8a3a430a-b540-4a8c-adad-b2893dbf9898-os-release\") pod \"multus-fdxjh\" (UID: \"8a3a430a-b540-4a8c-adad-b2893dbf9898\") " pod="openshift-multus/multus-fdxjh" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.916189 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/8dcaef0e-42e3-424b-bbe5-4a023efe912d-run-openvswitch\") pod \"ovnkube-node-v7br2\" (UID: \"8dcaef0e-42e3-424b-bbe5-4a023efe912d\") " pod="openshift-ovn-kubernetes/ovnkube-node-v7br2" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.916215 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/8a3a430a-b540-4a8c-adad-b2893dbf9898-host-run-k8s-cni-cncf-io\") pod \"multus-fdxjh\" (UID: \"8a3a430a-b540-4a8c-adad-b2893dbf9898\") " pod="openshift-multus/multus-fdxjh" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.916486 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/612e7ab9-bda4-45d8-bc50-ff1afb68aace-os-release\") pod \"multus-additional-cni-plugins-cdjmj\" (UID: \"612e7ab9-bda4-45d8-bc50-ff1afb68aace\") " pod="openshift-multus/multus-additional-cni-plugins-cdjmj" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.916515 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/8dcaef0e-42e3-424b-bbe5-4a023efe912d-etc-openvswitch\") pod \"ovnkube-node-v7br2\" (UID: \"8dcaef0e-42e3-424b-bbe5-4a023efe912d\") " pod="openshift-ovn-kubernetes/ovnkube-node-v7br2" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.916639 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/612e7ab9-bda4-45d8-bc50-ff1afb68aace-cni-binary-copy\") pod \"multus-additional-cni-plugins-cdjmj\" (UID: \"612e7ab9-bda4-45d8-bc50-ff1afb68aace\") " pod="openshift-multus/multus-additional-cni-plugins-cdjmj" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.916658 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/612e7ab9-bda4-45d8-bc50-ff1afb68aace-tuning-conf-dir\") pod \"multus-additional-cni-plugins-cdjmj\" (UID: \"612e7ab9-bda4-45d8-bc50-ff1afb68aace\") " pod="openshift-multus/multus-additional-cni-plugins-cdjmj" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.916758 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/8dcaef0e-42e3-424b-bbe5-4a023efe912d-env-overrides\") pod \"ovnkube-node-v7br2\" (UID: \"8dcaef0e-42e3-424b-bbe5-4a023efe912d\") " pod="openshift-ovn-kubernetes/ovnkube-node-v7br2" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.916950 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/612e7ab9-bda4-45d8-bc50-ff1afb68aace-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-cdjmj\" (UID: \"612e7ab9-bda4-45d8-bc50-ff1afb68aace\") " pod="openshift-multus/multus-additional-cni-plugins-cdjmj" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.917267 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/8dcaef0e-42e3-424b-bbe5-4a023efe912d-ovnkube-script-lib\") pod \"ovnkube-node-v7br2\" (UID: \"8dcaef0e-42e3-424b-bbe5-4a023efe912d\") " pod="openshift-ovn-kubernetes/ovnkube-node-v7br2" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.918911 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/874a5e45-dffd-4d17-b609-b7d3ed2eab07-proxy-tls\") pod \"machine-config-daemon-7tvc7\" (UID: \"874a5e45-dffd-4d17-b609-b7d3ed2eab07\") " pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.920009 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/8dcaef0e-42e3-424b-bbe5-4a023efe912d-ovn-node-metrics-cert\") pod \"ovnkube-node-v7br2\" (UID: \"8dcaef0e-42e3-424b-bbe5-4a023efe912d\") " pod="openshift-ovn-kubernetes/ovnkube-node-v7br2" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.930915 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-glrll\" (UniqueName: \"kubernetes.io/projected/8dcaef0e-42e3-424b-bbe5-4a023efe912d-kube-api-access-glrll\") pod \"ovnkube-node-v7br2\" (UID: \"8dcaef0e-42e3-424b-bbe5-4a023efe912d\") " pod="openshift-ovn-kubernetes/ovnkube-node-v7br2" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.931893 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k787j\" (UniqueName: \"kubernetes.io/projected/8a3a430a-b540-4a8c-adad-b2893dbf9898-kube-api-access-k787j\") pod \"multus-fdxjh\" (UID: \"8a3a430a-b540-4a8c-adad-b2893dbf9898\") " pod="openshift-multus/multus-fdxjh" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.933477 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wj4ks\" (UniqueName: \"kubernetes.io/projected/874a5e45-dffd-4d17-b609-b7d3ed2eab07-kube-api-access-wj4ks\") pod \"machine-config-daemon-7tvc7\" (UID: \"874a5e45-dffd-4d17-b609-b7d3ed2eab07\") " pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.934597 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fcfq9\" (UniqueName: \"kubernetes.io/projected/612e7ab9-bda4-45d8-bc50-ff1afb68aace-kube-api-access-fcfq9\") pod \"multus-additional-cni-plugins-cdjmj\" (UID: \"612e7ab9-bda4-45d8-bc50-ff1afb68aace\") " pod="openshift-multus/multus-additional-cni-plugins-cdjmj" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.937419 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33a9493e-9660-48db-b8d1-310e29ebe313\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3570643ef1b7d5cc3a542f320413b819338e0c8b7ab7f0f09cc909997a39bab8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb398668197dedc1c2a14b4ae7d6a29714f3527309ccb9c1622af841100e1d60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1d2b6d068d63bd38cdfb83824525f6d12054f6cabab5d684e03ea5f2152076c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://783ff243e395333f3fc80cfae21358c532f0c9c5cb5515586be56b90ed10d912\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a8135a3d01076e4437ae8a784b459bf8d16de57fa16ec7ce4d40c5129c9d4a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48ca3a5017ba650df99b2fc7201c3fe5abfa89d81089c5647cde1b1549f534fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48ca3a5017ba650df99b2fc7201c3fe5abfa89d81089c5647cde1b1549f534fc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://19682eca67f76bf2e06e1b80f12e21141fcbce0acb8195f0defacd1b9a10b5e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://19682eca67f76bf2e06e1b80f12e21141fcbce0acb8195f0defacd1b9a10b5e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://042aad6d5f6bee8eef77f9e6a2c265de775a69f3da72aedc68bdf312198f00fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://042aad6d5f6bee8eef77f9e6a2c265de775a69f3da72aedc68bdf312198f00fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:05Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.947433 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.957934 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.967001 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"874a5e45-dffd-4d17-b609-b7d3ed2eab07\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wj4ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wj4ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7tvc7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.979597 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 14 07:01:26 crc kubenswrapper[4870]: I1014 07:01:26.992871 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 14 07:01:27 crc kubenswrapper[4870]: I1014 07:01:27.009557 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-v7br2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dcaef0e-42e3-424b-bbe5-4a023efe912d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-v7br2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 14 07:01:27 crc kubenswrapper[4870]: I1014 07:01:27.023808 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-cdjmj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"612e7ab9-bda4-45d8-bc50-ff1afb68aace\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-cdjmj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 14 07:01:27 crc kubenswrapper[4870]: I1014 07:01:27.033807 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fdxjh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a3a430a-b540-4a8c-adad-b2893dbf9898\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k787j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fdxjh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 14 07:01:27 crc kubenswrapper[4870]: I1014 07:01:27.040111 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Oct 14 07:01:27 crc kubenswrapper[4870]: I1014 07:01:27.041022 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Oct 14 07:01:27 crc kubenswrapper[4870]: I1014 07:01:27.042201 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Oct 14 07:01:27 crc kubenswrapper[4870]: I1014 07:01:27.042927 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Oct 14 07:01:27 crc kubenswrapper[4870]: I1014 07:01:27.043971 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Oct 14 07:01:27 crc kubenswrapper[4870]: I1014 07:01:27.044589 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Oct 14 07:01:27 crc kubenswrapper[4870]: I1014 07:01:27.045279 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Oct 14 07:01:27 crc kubenswrapper[4870]: I1014 07:01:27.046307 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Oct 14 07:01:27 crc kubenswrapper[4870]: I1014 07:01:27.046959 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Oct 14 07:01:27 crc kubenswrapper[4870]: I1014 07:01:27.051850 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Oct 14 07:01:27 crc kubenswrapper[4870]: I1014 07:01:27.052409 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Oct 14 07:01:27 crc kubenswrapper[4870]: I1014 07:01:27.053557 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Oct 14 07:01:27 crc kubenswrapper[4870]: I1014 07:01:27.054130 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Oct 14 07:01:27 crc kubenswrapper[4870]: I1014 07:01:27.054770 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Oct 14 07:01:27 crc kubenswrapper[4870]: I1014 07:01:27.055730 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Oct 14 07:01:27 crc kubenswrapper[4870]: I1014 07:01:27.056375 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Oct 14 07:01:27 crc kubenswrapper[4870]: I1014 07:01:27.057553 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Oct 14 07:01:27 crc kubenswrapper[4870]: I1014 07:01:27.058002 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Oct 14 07:01:27 crc kubenswrapper[4870]: I1014 07:01:27.058687 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Oct 14 07:01:27 crc kubenswrapper[4870]: I1014 07:01:27.059722 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Oct 14 07:01:27 crc kubenswrapper[4870]: I1014 07:01:27.060213 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Oct 14 07:01:27 crc kubenswrapper[4870]: I1014 07:01:27.061292 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Oct 14 07:01:27 crc kubenswrapper[4870]: I1014 07:01:27.061795 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Oct 14 07:01:27 crc kubenswrapper[4870]: I1014 07:01:27.062981 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Oct 14 07:01:27 crc kubenswrapper[4870]: I1014 07:01:27.063502 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Oct 14 07:01:27 crc kubenswrapper[4870]: I1014 07:01:27.064243 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Oct 14 07:01:27 crc kubenswrapper[4870]: I1014 07:01:27.065071 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-cdjmj" Oct 14 07:01:27 crc kubenswrapper[4870]: I1014 07:01:27.065630 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Oct 14 07:01:27 crc kubenswrapper[4870]: I1014 07:01:27.066521 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Oct 14 07:01:27 crc kubenswrapper[4870]: I1014 07:01:27.067378 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33a9493e-9660-48db-b8d1-310e29ebe313\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3570643ef1b7d5cc3a542f320413b819338e0c8b7ab7f0f09cc909997a39bab8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb398668197dedc1c2a14b4ae7d6a29714f3527309ccb9c1622af841100e1d60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1d2b6d068d63bd38cdfb83824525f6d12054f6cabab5d684e03ea5f2152076c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://783ff243e395333f3fc80cfae21358c532f0c9c5cb5515586be56b90ed10d912\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a8135a3d01076e4437ae8a784b459bf8d16de57fa16ec7ce4d40c5129c9d4a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48ca3a5017ba650df99b2fc7201c3fe5abfa89d81089c5647cde1b1549f534fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48ca3a5017ba650df99b2fc7201c3fe5abfa89d81089c5647cde1b1549f534fc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://19682eca67f76bf2e06e1b80f12e21141fcbce0acb8195f0defacd1b9a10b5e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://19682eca67f76bf2e06e1b80f12e21141fcbce0acb8195f0defacd1b9a10b5e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://042aad6d5f6bee8eef77f9e6a2c265de775a69f3da72aedc68bdf312198f00fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://042aad6d5f6bee8eef77f9e6a2c265de775a69f3da72aedc68bdf312198f00fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:05Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 14 07:01:27 crc kubenswrapper[4870]: I1014 07:01:27.068169 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Oct 14 07:01:27 crc kubenswrapper[4870]: I1014 07:01:27.068915 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Oct 14 07:01:27 crc kubenswrapper[4870]: I1014 07:01:27.069483 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" Oct 14 07:01:27 crc kubenswrapper[4870]: I1014 07:01:27.070360 4870 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Oct 14 07:01:27 crc kubenswrapper[4870]: I1014 07:01:27.070590 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Oct 14 07:01:27 crc kubenswrapper[4870]: I1014 07:01:27.072496 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Oct 14 07:01:27 crc kubenswrapper[4870]: I1014 07:01:27.073495 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Oct 14 07:01:27 crc kubenswrapper[4870]: I1014 07:01:27.074062 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Oct 14 07:01:27 crc kubenswrapper[4870]: I1014 07:01:27.076285 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Oct 14 07:01:27 crc kubenswrapper[4870]: I1014 07:01:27.076571 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-4v2v2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4796ba6-c23b-4764-be96-2e3a3bf5e962\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnp8b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-4v2v2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 14 07:01:27 crc kubenswrapper[4870]: I1014 07:01:27.077618 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Oct 14 07:01:27 crc kubenswrapper[4870]: I1014 07:01:27.077968 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-fdxjh" Oct 14 07:01:27 crc kubenswrapper[4870]: I1014 07:01:27.079051 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Oct 14 07:01:27 crc kubenswrapper[4870]: I1014 07:01:27.080211 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Oct 14 07:01:27 crc kubenswrapper[4870]: I1014 07:01:27.081568 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Oct 14 07:01:27 crc kubenswrapper[4870]: I1014 07:01:27.082954 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Oct 14 07:01:27 crc kubenswrapper[4870]: W1014 07:01:27.084030 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod874a5e45_dffd_4d17_b609_b7d3ed2eab07.slice/crio-f3dca5b767ee64960d6e676cbcf55fc953221e14d702ab6bd171ea6963cf37ae WatchSource:0}: Error finding container f3dca5b767ee64960d6e676cbcf55fc953221e14d702ab6bd171ea6963cf37ae: Status 404 returned error can't find the container with id f3dca5b767ee64960d6e676cbcf55fc953221e14d702ab6bd171ea6963cf37ae Oct 14 07:01:27 crc kubenswrapper[4870]: I1014 07:01:27.084144 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Oct 14 07:01:27 crc kubenswrapper[4870]: I1014 07:01:27.085761 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Oct 14 07:01:27 crc kubenswrapper[4870]: I1014 07:01:27.086808 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Oct 14 07:01:27 crc kubenswrapper[4870]: I1014 07:01:27.087356 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-v7br2" Oct 14 07:01:27 crc kubenswrapper[4870]: I1014 07:01:27.087547 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 14 07:01:27 crc kubenswrapper[4870]: I1014 07:01:27.088355 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Oct 14 07:01:27 crc kubenswrapper[4870]: I1014 07:01:27.089838 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Oct 14 07:01:27 crc kubenswrapper[4870]: I1014 07:01:27.091338 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Oct 14 07:01:27 crc kubenswrapper[4870]: I1014 07:01:27.092285 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Oct 14 07:01:27 crc kubenswrapper[4870]: I1014 07:01:27.093255 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Oct 14 07:01:27 crc kubenswrapper[4870]: I1014 07:01:27.093792 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Oct 14 07:01:27 crc kubenswrapper[4870]: I1014 07:01:27.094319 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Oct 14 07:01:27 crc kubenswrapper[4870]: I1014 07:01:27.095285 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Oct 14 07:01:27 crc kubenswrapper[4870]: W1014 07:01:27.095689 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8a3a430a_b540_4a8c_adad_b2893dbf9898.slice/crio-94fa86151c685e964de7c2c4acb9bd6c931ecf077abb2b6d637dcdf97096504e WatchSource:0}: Error finding container 94fa86151c685e964de7c2c4acb9bd6c931ecf077abb2b6d637dcdf97096504e: Status 404 returned error can't find the container with id 94fa86151c685e964de7c2c4acb9bd6c931ecf077abb2b6d637dcdf97096504e Oct 14 07:01:27 crc kubenswrapper[4870]: I1014 07:01:27.095893 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Oct 14 07:01:27 crc kubenswrapper[4870]: I1014 07:01:27.096896 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Oct 14 07:01:27 crc kubenswrapper[4870]: I1014 07:01:27.101682 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 14 07:01:27 crc kubenswrapper[4870]: W1014 07:01:27.140220 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8dcaef0e_42e3_424b_bbe5_4a023efe912d.slice/crio-919facd0f2f00aaf6e6831e4778b6f30f1b1428e0530e02fb7d55e87d66fb222 WatchSource:0}: Error finding container 919facd0f2f00aaf6e6831e4778b6f30f1b1428e0530e02fb7d55e87d66fb222: Status 404 returned error can't find the container with id 919facd0f2f00aaf6e6831e4778b6f30f1b1428e0530e02fb7d55e87d66fb222 Oct 14 07:01:27 crc kubenswrapper[4870]: I1014 07:01:27.213248 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"e26fe49ec2045d20a0cb2dd17e96e79b6eb43c641eb184c688cc472ca5db39d4"} Oct 14 07:01:27 crc kubenswrapper[4870]: I1014 07:01:27.213296 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"aea115f7ee71c700b218fe15c928af7b55e6babe70e65d67f125d87be1fd2e2c"} Oct 14 07:01:27 crc kubenswrapper[4870]: I1014 07:01:27.213308 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"7130d126cda18a7f47bdae9a7bd2b7671774c3419d63365bc6559778ffa25d73"} Oct 14 07:01:27 crc kubenswrapper[4870]: I1014 07:01:27.226695 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-v7br2" event={"ID":"8dcaef0e-42e3-424b-bbe5-4a023efe912d","Type":"ContainerStarted","Data":"919facd0f2f00aaf6e6831e4778b6f30f1b1428e0530e02fb7d55e87d66fb222"} Oct 14 07:01:27 crc kubenswrapper[4870]: I1014 07:01:27.234736 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 14 07:01:27 crc kubenswrapper[4870]: I1014 07:01:27.234760 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-fdxjh" event={"ID":"8a3a430a-b540-4a8c-adad-b2893dbf9898","Type":"ContainerStarted","Data":"94fa86151c685e964de7c2c4acb9bd6c931ecf077abb2b6d637dcdf97096504e"} Oct 14 07:01:27 crc kubenswrapper[4870]: I1014 07:01:27.240931 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"e539fac3d0a2ade314a4b4c1050bb27c37bcff5ba5272fa2d71b97ab5c0d7191"} Oct 14 07:01:27 crc kubenswrapper[4870]: I1014 07:01:27.241161 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"01a890b3732b139575b96579629a58eb8b7e7f05a1043958e8e994a603632302"} Oct 14 07:01:27 crc kubenswrapper[4870]: I1014 07:01:27.244584 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-4v2v2" event={"ID":"d4796ba6-c23b-4764-be96-2e3a3bf5e962","Type":"ContainerStarted","Data":"bf5afb0b99a7101f6d5573fb7eb925c5a1d8becc11b2e3c6f3351d80b6f0e823"} Oct 14 07:01:27 crc kubenswrapper[4870]: I1014 07:01:27.244833 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-4v2v2" event={"ID":"d4796ba6-c23b-4764-be96-2e3a3bf5e962","Type":"ContainerStarted","Data":"8ac7f59d6179818248b9f8a64e5668e80532498586333babacbd555138ebf5ea"} Oct 14 07:01:27 crc kubenswrapper[4870]: I1014 07:01:27.246596 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"5f1f856410372deedd5fc11fc0c4ff0e9f085d7db1291a62298d7af922b53d43"} Oct 14 07:01:27 crc kubenswrapper[4870]: I1014 07:01:27.253014 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" event={"ID":"874a5e45-dffd-4d17-b609-b7d3ed2eab07","Type":"ContainerStarted","Data":"f3dca5b767ee64960d6e676cbcf55fc953221e14d702ab6bd171ea6963cf37ae"} Oct 14 07:01:27 crc kubenswrapper[4870]: I1014 07:01:27.260347 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-cdjmj" event={"ID":"612e7ab9-bda4-45d8-bc50-ff1afb68aace","Type":"ContainerStarted","Data":"61a079e6e4a34faad7f9a9d784c7b840ae34076668bb4bbbde58b7a62d529a54"} Oct 14 07:01:27 crc kubenswrapper[4870]: I1014 07:01:27.262562 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:27Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:27 crc kubenswrapper[4870]: I1014 07:01:27.273555 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Oct 14 07:01:27 crc kubenswrapper[4870]: I1014 07:01:27.274348 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Oct 14 07:01:27 crc kubenswrapper[4870]: I1014 07:01:27.279539 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"874a5e45-dffd-4d17-b609-b7d3ed2eab07\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wj4ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wj4ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7tvc7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:27Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:27 crc kubenswrapper[4870]: I1014 07:01:27.282055 4870 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="978e8db184ba9e430232c6bb2ee530beb92567fdc71ba3c35b0bedca3492d845" exitCode=255 Oct 14 07:01:27 crc kubenswrapper[4870]: I1014 07:01:27.283106 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"978e8db184ba9e430232c6bb2ee530beb92567fdc71ba3c35b0bedca3492d845"} Oct 14 07:01:27 crc kubenswrapper[4870]: I1014 07:01:27.283153 4870 scope.go:117] "RemoveContainer" containerID="32056f51de4479330f592c7d38413b0cfc159d3cec4c4b7158dcec62daf0e8c8" Oct 14 07:01:27 crc kubenswrapper[4870]: I1014 07:01:27.296958 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Oct 14 07:01:27 crc kubenswrapper[4870]: I1014 07:01:27.297582 4870 scope.go:117] "RemoveContainer" containerID="978e8db184ba9e430232c6bb2ee530beb92567fdc71ba3c35b0bedca3492d845" Oct 14 07:01:27 crc kubenswrapper[4870]: E1014 07:01:27.297797 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Oct 14 07:01:27 crc kubenswrapper[4870]: I1014 07:01:27.306777 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-v7br2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dcaef0e-42e3-424b-bbe5-4a023efe912d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-v7br2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:27Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:27 crc kubenswrapper[4870]: I1014 07:01:27.321766 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:27Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:27 crc kubenswrapper[4870]: I1014 07:01:27.338365 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e26fe49ec2045d20a0cb2dd17e96e79b6eb43c641eb184c688cc472ca5db39d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aea115f7ee71c700b218fe15c928af7b55e6babe70e65d67f125d87be1fd2e2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:27Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:27 crc kubenswrapper[4870]: I1014 07:01:27.351818 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-4v2v2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4796ba6-c23b-4764-be96-2e3a3bf5e962\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnp8b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-4v2v2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:27Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:27 crc kubenswrapper[4870]: I1014 07:01:27.373140 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-cdjmj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"612e7ab9-bda4-45d8-bc50-ff1afb68aace\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-cdjmj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:27Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:27 crc kubenswrapper[4870]: I1014 07:01:27.388375 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fdxjh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a3a430a-b540-4a8c-adad-b2893dbf9898\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k787j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fdxjh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:27Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:27 crc kubenswrapper[4870]: I1014 07:01:27.412683 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33a9493e-9660-48db-b8d1-310e29ebe313\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3570643ef1b7d5cc3a542f320413b819338e0c8b7ab7f0f09cc909997a39bab8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb398668197dedc1c2a14b4ae7d6a29714f3527309ccb9c1622af841100e1d60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1d2b6d068d63bd38cdfb83824525f6d12054f6cabab5d684e03ea5f2152076c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://783ff243e395333f3fc80cfae21358c532f0c9c5cb5515586be56b90ed10d912\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a8135a3d01076e4437ae8a784b459bf8d16de57fa16ec7ce4d40c5129c9d4a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48ca3a5017ba650df99b2fc7201c3fe5abfa89d81089c5647cde1b1549f534fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48ca3a5017ba650df99b2fc7201c3fe5abfa89d81089c5647cde1b1549f534fc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://19682eca67f76bf2e06e1b80f12e21141fcbce0acb8195f0defacd1b9a10b5e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://19682eca67f76bf2e06e1b80f12e21141fcbce0acb8195f0defacd1b9a10b5e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://042aad6d5f6bee8eef77f9e6a2c265de775a69f3da72aedc68bdf312198f00fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://042aad6d5f6bee8eef77f9e6a2c265de775a69f3da72aedc68bdf312198f00fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:05Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:27Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:27 crc kubenswrapper[4870]: I1014 07:01:27.429904 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:27Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:27 crc kubenswrapper[4870]: I1014 07:01:27.446151 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:27Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:27 crc kubenswrapper[4870]: I1014 07:01:27.461540 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fdxjh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a3a430a-b540-4a8c-adad-b2893dbf9898\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k787j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fdxjh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:27Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:27 crc kubenswrapper[4870]: I1014 07:01:27.484559 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33a9493e-9660-48db-b8d1-310e29ebe313\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3570643ef1b7d5cc3a542f320413b819338e0c8b7ab7f0f09cc909997a39bab8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb398668197dedc1c2a14b4ae7d6a29714f3527309ccb9c1622af841100e1d60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1d2b6d068d63bd38cdfb83824525f6d12054f6cabab5d684e03ea5f2152076c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://783ff243e395333f3fc80cfae21358c532f0c9c5cb5515586be56b90ed10d912\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a8135a3d01076e4437ae8a784b459bf8d16de57fa16ec7ce4d40c5129c9d4a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48ca3a5017ba650df99b2fc7201c3fe5abfa89d81089c5647cde1b1549f534fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48ca3a5017ba650df99b2fc7201c3fe5abfa89d81089c5647cde1b1549f534fc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://19682eca67f76bf2e06e1b80f12e21141fcbce0acb8195f0defacd1b9a10b5e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://19682eca67f76bf2e06e1b80f12e21141fcbce0acb8195f0defacd1b9a10b5e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://042aad6d5f6bee8eef77f9e6a2c265de775a69f3da72aedc68bdf312198f00fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://042aad6d5f6bee8eef77f9e6a2c265de775a69f3da72aedc68bdf312198f00fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:05Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:27Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:27 crc kubenswrapper[4870]: I1014 07:01:27.497383 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-4v2v2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4796ba6-c23b-4764-be96-2e3a3bf5e962\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf5afb0b99a7101f6d5573fb7eb925c5a1d8becc11b2e3c6f3351d80b6f0e823\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnp8b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-4v2v2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:27Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:27 crc kubenswrapper[4870]: I1014 07:01:27.512836 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-cdjmj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"612e7ab9-bda4-45d8-bc50-ff1afb68aace\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-cdjmj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:27Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:27 crc kubenswrapper[4870]: I1014 07:01:27.543543 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e539fac3d0a2ade314a4b4c1050bb27c37bcff5ba5272fa2d71b97ab5c0d7191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:27Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:27 crc kubenswrapper[4870]: I1014 07:01:27.586482 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:27Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:27 crc kubenswrapper[4870]: I1014 07:01:27.624996 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:27Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:27 crc kubenswrapper[4870]: I1014 07:01:27.628300 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 07:01:27 crc kubenswrapper[4870]: E1014 07:01:27.628551 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 07:01:29.628509295 +0000 UTC m=+25.325869686 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:01:27 crc kubenswrapper[4870]: I1014 07:01:27.663572 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:27Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:27 crc kubenswrapper[4870]: I1014 07:01:27.701141 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"874a5e45-dffd-4d17-b609-b7d3ed2eab07\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wj4ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wj4ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7tvc7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:27Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:27 crc kubenswrapper[4870]: I1014 07:01:27.730128 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 07:01:27 crc kubenswrapper[4870]: I1014 07:01:27.730174 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 07:01:27 crc kubenswrapper[4870]: I1014 07:01:27.730197 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 07:01:27 crc kubenswrapper[4870]: I1014 07:01:27.730220 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 07:01:27 crc kubenswrapper[4870]: E1014 07:01:27.730304 4870 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 14 07:01:27 crc kubenswrapper[4870]: E1014 07:01:27.730324 4870 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 14 07:01:27 crc kubenswrapper[4870]: E1014 07:01:27.730351 4870 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 14 07:01:27 crc kubenswrapper[4870]: E1014 07:01:27.730366 4870 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 14 07:01:27 crc kubenswrapper[4870]: E1014 07:01:27.730354 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-14 07:01:29.730339074 +0000 UTC m=+25.427699445 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 14 07:01:27 crc kubenswrapper[4870]: E1014 07:01:27.730412 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-14 07:01:29.730401426 +0000 UTC m=+25.427761797 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 14 07:01:27 crc kubenswrapper[4870]: E1014 07:01:27.730410 4870 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 14 07:01:27 crc kubenswrapper[4870]: E1014 07:01:27.730470 4870 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 14 07:01:27 crc kubenswrapper[4870]: E1014 07:01:27.730474 4870 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 14 07:01:27 crc kubenswrapper[4870]: E1014 07:01:27.730493 4870 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 14 07:01:27 crc kubenswrapper[4870]: E1014 07:01:27.730517 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-14 07:01:29.730505728 +0000 UTC m=+25.427866109 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 14 07:01:27 crc kubenswrapper[4870]: E1014 07:01:27.730555 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-14 07:01:29.730535099 +0000 UTC m=+25.427895530 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 14 07:01:27 crc kubenswrapper[4870]: I1014 07:01:27.742411 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77a0d93a-5391-4112-b502-c474f45de209\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7815af9a584f6db6660073da289f2e8c987a04ef9ddfac155bdd9a15b014fcfa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b329b570622155a55f70501b2260e278106a68ebef497914231b4ec860e0125\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba942a0e6fe5ab93a187e0b7b4bd7c440c2e7827d9b5323fc50f90b562075392\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://978e8db184ba9e430232c6bb2ee530beb92567fdc71ba3c35b0bedca3492d845\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://32056f51de4479330f592c7d38413b0cfc159d3cec4c4b7158dcec62daf0e8c8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T07:01:09Z\\\",\\\"message\\\":\\\"W1014 07:01:08.265859 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1014 07:01:08.266350 1 crypto.go:601] Generating new CA for check-endpoints-signer@1760425268 cert, and key in /tmp/serving-cert-4254447587/serving-signer.crt, /tmp/serving-cert-4254447587/serving-signer.key\\\\nI1014 07:01:08.616317 1 observer_polling.go:159] Starting file observer\\\\nW1014 07:01:08.619480 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1014 07:01:08.619647 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1014 07:01:08.620554 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4254447587/tls.crt::/tmp/serving-cert-4254447587/tls.key\\\\\\\"\\\\nF1014 07:01:09.093392 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:08Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://978e8db184ba9e430232c6bb2ee530beb92567fdc71ba3c35b0bedca3492d845\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"file observer\\\\nW1014 07:01:25.870672 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1014 07:01:25.870795 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1014 07:01:25.871521 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-756714914/tls.crt::/tmp/serving-cert-756714914/tls.key\\\\\\\"\\\\nI1014 07:01:26.460037 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1014 07:01:26.464876 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1014 07:01:26.464902 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1014 07:01:26.464925 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1014 07:01:26.464931 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1014 07:01:26.481374 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1014 07:01:26.481400 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:01:26.481405 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:01:26.481410 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1014 07:01:26.481414 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1014 07:01:26.481417 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1014 07:01:26.481420 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1014 07:01:26.481610 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1014 07:01:26.482872 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bd0044a247bdc5ca4cab7294103af3270b76cde61f3109b85e83ed050ca53b6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82d3c2214e0be679d2827594b159ace362151318689078bc49d3158f42899636\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82d3c2214e0be679d2827594b159ace362151318689078bc49d3158f42899636\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:05Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:27Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:27 crc kubenswrapper[4870]: I1014 07:01:27.783943 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:27Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:27 crc kubenswrapper[4870]: I1014 07:01:27.824969 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e26fe49ec2045d20a0cb2dd17e96e79b6eb43c641eb184c688cc472ca5db39d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aea115f7ee71c700b218fe15c928af7b55e6babe70e65d67f125d87be1fd2e2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:27Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:27 crc kubenswrapper[4870]: I1014 07:01:27.868741 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-v7br2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dcaef0e-42e3-424b-bbe5-4a023efe912d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-v7br2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:27Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:28 crc kubenswrapper[4870]: I1014 07:01:28.032926 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 07:01:28 crc kubenswrapper[4870]: I1014 07:01:28.032948 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 07:01:28 crc kubenswrapper[4870]: I1014 07:01:28.033016 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 07:01:28 crc kubenswrapper[4870]: E1014 07:01:28.033085 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 07:01:28 crc kubenswrapper[4870]: E1014 07:01:28.033277 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 07:01:28 crc kubenswrapper[4870]: E1014 07:01:28.033413 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 07:01:28 crc kubenswrapper[4870]: I1014 07:01:28.287085 4870 generic.go:334] "Generic (PLEG): container finished" podID="8dcaef0e-42e3-424b-bbe5-4a023efe912d" containerID="1ba24c2e067b9f3a14d9d018b66300b54d866530d30174fbdb3c899299575d28" exitCode=0 Oct 14 07:01:28 crc kubenswrapper[4870]: I1014 07:01:28.287178 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-v7br2" event={"ID":"8dcaef0e-42e3-424b-bbe5-4a023efe912d","Type":"ContainerDied","Data":"1ba24c2e067b9f3a14d9d018b66300b54d866530d30174fbdb3c899299575d28"} Oct 14 07:01:28 crc kubenswrapper[4870]: I1014 07:01:28.289086 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-fdxjh" event={"ID":"8a3a430a-b540-4a8c-adad-b2893dbf9898","Type":"ContainerStarted","Data":"93e41d5e927eed7ec62eb9c3607cc804cd8999556663363fd123b2a784b00ff1"} Oct 14 07:01:28 crc kubenswrapper[4870]: I1014 07:01:28.292383 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" event={"ID":"874a5e45-dffd-4d17-b609-b7d3ed2eab07","Type":"ContainerStarted","Data":"d98d044836ab44a405196915ef7bf572b75113f2c81ec7e64effa63b47ad3543"} Oct 14 07:01:28 crc kubenswrapper[4870]: I1014 07:01:28.292450 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" event={"ID":"874a5e45-dffd-4d17-b609-b7d3ed2eab07","Type":"ContainerStarted","Data":"8004752575df4a1c44efbcafe64799f09126d0f97285d92388c76a0b90e022f8"} Oct 14 07:01:28 crc kubenswrapper[4870]: I1014 07:01:28.294249 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Oct 14 07:01:28 crc kubenswrapper[4870]: I1014 07:01:28.296688 4870 scope.go:117] "RemoveContainer" containerID="978e8db184ba9e430232c6bb2ee530beb92567fdc71ba3c35b0bedca3492d845" Oct 14 07:01:28 crc kubenswrapper[4870]: E1014 07:01:28.296889 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Oct 14 07:01:28 crc kubenswrapper[4870]: I1014 07:01:28.298736 4870 generic.go:334] "Generic (PLEG): container finished" podID="612e7ab9-bda4-45d8-bc50-ff1afb68aace" containerID="834444b5350ee12df4e4652c3009880d8d69e1edf8ba0c0156ab96fb2bbe28e1" exitCode=0 Oct 14 07:01:28 crc kubenswrapper[4870]: I1014 07:01:28.299666 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-cdjmj" event={"ID":"612e7ab9-bda4-45d8-bc50-ff1afb68aace","Type":"ContainerDied","Data":"834444b5350ee12df4e4652c3009880d8d69e1edf8ba0c0156ab96fb2bbe28e1"} Oct 14 07:01:28 crc kubenswrapper[4870]: I1014 07:01:28.320414 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:28Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:28 crc kubenswrapper[4870]: I1014 07:01:28.339999 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:28Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:28 crc kubenswrapper[4870]: I1014 07:01:28.364839 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"874a5e45-dffd-4d17-b609-b7d3ed2eab07\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wj4ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wj4ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7tvc7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:28Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:28 crc kubenswrapper[4870]: I1014 07:01:28.381238 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77a0d93a-5391-4112-b502-c474f45de209\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7815af9a584f6db6660073da289f2e8c987a04ef9ddfac155bdd9a15b014fcfa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b329b570622155a55f70501b2260e278106a68ebef497914231b4ec860e0125\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba942a0e6fe5ab93a187e0b7b4bd7c440c2e7827d9b5323fc50f90b562075392\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://978e8db184ba9e430232c6bb2ee530beb92567fdc71ba3c35b0bedca3492d845\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://32056f51de4479330f592c7d38413b0cfc159d3cec4c4b7158dcec62daf0e8c8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T07:01:09Z\\\",\\\"message\\\":\\\"W1014 07:01:08.265859 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1014 07:01:08.266350 1 crypto.go:601] Generating new CA for check-endpoints-signer@1760425268 cert, and key in /tmp/serving-cert-4254447587/serving-signer.crt, /tmp/serving-cert-4254447587/serving-signer.key\\\\nI1014 07:01:08.616317 1 observer_polling.go:159] Starting file observer\\\\nW1014 07:01:08.619480 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1014 07:01:08.619647 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1014 07:01:08.620554 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4254447587/tls.crt::/tmp/serving-cert-4254447587/tls.key\\\\\\\"\\\\nF1014 07:01:09.093392 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:08Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://978e8db184ba9e430232c6bb2ee530beb92567fdc71ba3c35b0bedca3492d845\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"file observer\\\\nW1014 07:01:25.870672 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1014 07:01:25.870795 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1014 07:01:25.871521 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-756714914/tls.crt::/tmp/serving-cert-756714914/tls.key\\\\\\\"\\\\nI1014 07:01:26.460037 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1014 07:01:26.464876 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1014 07:01:26.464902 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1014 07:01:26.464925 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1014 07:01:26.464931 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1014 07:01:26.481374 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1014 07:01:26.481400 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:01:26.481405 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:01:26.481410 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1014 07:01:26.481414 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1014 07:01:26.481417 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1014 07:01:26.481420 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1014 07:01:26.481610 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1014 07:01:26.482872 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bd0044a247bdc5ca4cab7294103af3270b76cde61f3109b85e83ed050ca53b6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82d3c2214e0be679d2827594b159ace362151318689078bc49d3158f42899636\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82d3c2214e0be679d2827594b159ace362151318689078bc49d3158f42899636\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:05Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:28Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:28 crc kubenswrapper[4870]: I1014 07:01:28.398672 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:28Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:28 crc kubenswrapper[4870]: I1014 07:01:28.417782 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e26fe49ec2045d20a0cb2dd17e96e79b6eb43c641eb184c688cc472ca5db39d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aea115f7ee71c700b218fe15c928af7b55e6babe70e65d67f125d87be1fd2e2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:28Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:28 crc kubenswrapper[4870]: I1014 07:01:28.439645 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-v7br2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dcaef0e-42e3-424b-bbe5-4a023efe912d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ba24c2e067b9f3a14d9d018b66300b54d866530d30174fbdb3c899299575d28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ba24c2e067b9f3a14d9d018b66300b54d866530d30174fbdb3c899299575d28\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-v7br2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:28Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:28 crc kubenswrapper[4870]: I1014 07:01:28.460008 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33a9493e-9660-48db-b8d1-310e29ebe313\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3570643ef1b7d5cc3a542f320413b819338e0c8b7ab7f0f09cc909997a39bab8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb398668197dedc1c2a14b4ae7d6a29714f3527309ccb9c1622af841100e1d60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1d2b6d068d63bd38cdfb83824525f6d12054f6cabab5d684e03ea5f2152076c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://783ff243e395333f3fc80cfae21358c532f0c9c5cb5515586be56b90ed10d912\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a8135a3d01076e4437ae8a784b459bf8d16de57fa16ec7ce4d40c5129c9d4a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48ca3a5017ba650df99b2fc7201c3fe5abfa89d81089c5647cde1b1549f534fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48ca3a5017ba650df99b2fc7201c3fe5abfa89d81089c5647cde1b1549f534fc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://19682eca67f76bf2e06e1b80f12e21141fcbce0acb8195f0defacd1b9a10b5e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://19682eca67f76bf2e06e1b80f12e21141fcbce0acb8195f0defacd1b9a10b5e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://042aad6d5f6bee8eef77f9e6a2c265de775a69f3da72aedc68bdf312198f00fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://042aad6d5f6bee8eef77f9e6a2c265de775a69f3da72aedc68bdf312198f00fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:05Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:28Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:28 crc kubenswrapper[4870]: I1014 07:01:28.471132 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-4v2v2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4796ba6-c23b-4764-be96-2e3a3bf5e962\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf5afb0b99a7101f6d5573fb7eb925c5a1d8becc11b2e3c6f3351d80b6f0e823\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnp8b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-4v2v2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:28Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:28 crc kubenswrapper[4870]: I1014 07:01:28.485003 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-cdjmj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"612e7ab9-bda4-45d8-bc50-ff1afb68aace\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-cdjmj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:28Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:28 crc kubenswrapper[4870]: I1014 07:01:28.499761 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fdxjh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a3a430a-b540-4a8c-adad-b2893dbf9898\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k787j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fdxjh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:28Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:28 crc kubenswrapper[4870]: I1014 07:01:28.524012 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e539fac3d0a2ade314a4b4c1050bb27c37bcff5ba5272fa2d71b97ab5c0d7191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:28Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:28 crc kubenswrapper[4870]: I1014 07:01:28.539200 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:28Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:28 crc kubenswrapper[4870]: I1014 07:01:28.559764 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e539fac3d0a2ade314a4b4c1050bb27c37bcff5ba5272fa2d71b97ab5c0d7191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:28Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:28 crc kubenswrapper[4870]: I1014 07:01:28.572672 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:28Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:28 crc kubenswrapper[4870]: I1014 07:01:28.586025 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:28Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:28 crc kubenswrapper[4870]: I1014 07:01:28.597878 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:28Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:28 crc kubenswrapper[4870]: I1014 07:01:28.613929 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"874a5e45-dffd-4d17-b609-b7d3ed2eab07\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d98d044836ab44a405196915ef7bf572b75113f2c81ec7e64effa63b47ad3543\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wj4ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8004752575df4a1c44efbcafe64799f09126d0f97285d92388c76a0b90e022f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wj4ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7tvc7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:28Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:28 crc kubenswrapper[4870]: I1014 07:01:28.629314 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77a0d93a-5391-4112-b502-c474f45de209\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7815af9a584f6db6660073da289f2e8c987a04ef9ddfac155bdd9a15b014fcfa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b329b570622155a55f70501b2260e278106a68ebef497914231b4ec860e0125\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba942a0e6fe5ab93a187e0b7b4bd7c440c2e7827d9b5323fc50f90b562075392\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://978e8db184ba9e430232c6bb2ee530beb92567fdc71ba3c35b0bedca3492d845\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://978e8db184ba9e430232c6bb2ee530beb92567fdc71ba3c35b0bedca3492d845\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"file observer\\\\nW1014 07:01:25.870672 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1014 07:01:25.870795 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1014 07:01:25.871521 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-756714914/tls.crt::/tmp/serving-cert-756714914/tls.key\\\\\\\"\\\\nI1014 07:01:26.460037 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1014 07:01:26.464876 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1014 07:01:26.464902 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1014 07:01:26.464925 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1014 07:01:26.464931 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1014 07:01:26.481374 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1014 07:01:26.481400 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:01:26.481405 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:01:26.481410 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1014 07:01:26.481414 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1014 07:01:26.481417 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1014 07:01:26.481420 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1014 07:01:26.481610 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1014 07:01:26.482872 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bd0044a247bdc5ca4cab7294103af3270b76cde61f3109b85e83ed050ca53b6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82d3c2214e0be679d2827594b159ace362151318689078bc49d3158f42899636\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82d3c2214e0be679d2827594b159ace362151318689078bc49d3158f42899636\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:05Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:28Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:28 crc kubenswrapper[4870]: I1014 07:01:28.662209 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:28Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:28 crc kubenswrapper[4870]: I1014 07:01:28.702937 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e26fe49ec2045d20a0cb2dd17e96e79b6eb43c641eb184c688cc472ca5db39d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aea115f7ee71c700b218fe15c928af7b55e6babe70e65d67f125d87be1fd2e2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:28Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:28 crc kubenswrapper[4870]: I1014 07:01:28.747774 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-v7br2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dcaef0e-42e3-424b-bbe5-4a023efe912d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ba24c2e067b9f3a14d9d018b66300b54d866530d30174fbdb3c899299575d28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ba24c2e067b9f3a14d9d018b66300b54d866530d30174fbdb3c899299575d28\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-v7br2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:28Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:28 crc kubenswrapper[4870]: I1014 07:01:28.783305 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fdxjh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a3a430a-b540-4a8c-adad-b2893dbf9898\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93e41d5e927eed7ec62eb9c3607cc804cd8999556663363fd123b2a784b00ff1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k787j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fdxjh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:28Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:28 crc kubenswrapper[4870]: I1014 07:01:28.826177 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33a9493e-9660-48db-b8d1-310e29ebe313\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3570643ef1b7d5cc3a542f320413b819338e0c8b7ab7f0f09cc909997a39bab8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb398668197dedc1c2a14b4ae7d6a29714f3527309ccb9c1622af841100e1d60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1d2b6d068d63bd38cdfb83824525f6d12054f6cabab5d684e03ea5f2152076c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://783ff243e395333f3fc80cfae21358c532f0c9c5cb5515586be56b90ed10d912\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a8135a3d01076e4437ae8a784b459bf8d16de57fa16ec7ce4d40c5129c9d4a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48ca3a5017ba650df99b2fc7201c3fe5abfa89d81089c5647cde1b1549f534fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48ca3a5017ba650df99b2fc7201c3fe5abfa89d81089c5647cde1b1549f534fc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://19682eca67f76bf2e06e1b80f12e21141fcbce0acb8195f0defacd1b9a10b5e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://19682eca67f76bf2e06e1b80f12e21141fcbce0acb8195f0defacd1b9a10b5e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://042aad6d5f6bee8eef77f9e6a2c265de775a69f3da72aedc68bdf312198f00fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://042aad6d5f6bee8eef77f9e6a2c265de775a69f3da72aedc68bdf312198f00fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:05Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:28Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:28 crc kubenswrapper[4870]: I1014 07:01:28.866939 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-4v2v2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4796ba6-c23b-4764-be96-2e3a3bf5e962\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf5afb0b99a7101f6d5573fb7eb925c5a1d8becc11b2e3c6f3351d80b6f0e823\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnp8b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-4v2v2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:28Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:28 crc kubenswrapper[4870]: I1014 07:01:28.907726 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-cdjmj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"612e7ab9-bda4-45d8-bc50-ff1afb68aace\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://834444b5350ee12df4e4652c3009880d8d69e1edf8ba0c0156ab96fb2bbe28e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://834444b5350ee12df4e4652c3009880d8d69e1edf8ba0c0156ab96fb2bbe28e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-cdjmj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:28Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:28 crc kubenswrapper[4870]: I1014 07:01:28.963017 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-kz77f"] Oct 14 07:01:28 crc kubenswrapper[4870]: I1014 07:01:28.963806 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-kz77f" Oct 14 07:01:28 crc kubenswrapper[4870]: I1014 07:01:28.966045 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Oct 14 07:01:28 crc kubenswrapper[4870]: I1014 07:01:28.966178 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Oct 14 07:01:28 crc kubenswrapper[4870]: I1014 07:01:28.966496 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Oct 14 07:01:28 crc kubenswrapper[4870]: I1014 07:01:28.971959 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Oct 14 07:01:29 crc kubenswrapper[4870]: I1014 07:01:29.025631 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e539fac3d0a2ade314a4b4c1050bb27c37bcff5ba5272fa2d71b97ab5c0d7191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:29Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:29 crc kubenswrapper[4870]: I1014 07:01:29.043125 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/d3864a91-9bf6-4540-a908-cc41aa39656c-serviceca\") pod \"node-ca-kz77f\" (UID: \"d3864a91-9bf6-4540-a908-cc41aa39656c\") " pod="openshift-image-registry/node-ca-kz77f" Oct 14 07:01:29 crc kubenswrapper[4870]: I1014 07:01:29.043180 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v4h8v\" (UniqueName: \"kubernetes.io/projected/d3864a91-9bf6-4540-a908-cc41aa39656c-kube-api-access-v4h8v\") pod \"node-ca-kz77f\" (UID: \"d3864a91-9bf6-4540-a908-cc41aa39656c\") " pod="openshift-image-registry/node-ca-kz77f" Oct 14 07:01:29 crc kubenswrapper[4870]: I1014 07:01:29.043210 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d3864a91-9bf6-4540-a908-cc41aa39656c-host\") pod \"node-ca-kz77f\" (UID: \"d3864a91-9bf6-4540-a908-cc41aa39656c\") " pod="openshift-image-registry/node-ca-kz77f" Oct 14 07:01:29 crc kubenswrapper[4870]: I1014 07:01:29.062942 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:29Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:29 crc kubenswrapper[4870]: I1014 07:01:29.102586 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-kz77f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d3864a91-9bf6-4540-a908-cc41aa39656c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4h8v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:28Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-kz77f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:29Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:29 crc kubenswrapper[4870]: I1014 07:01:29.140958 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:29Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:29 crc kubenswrapper[4870]: I1014 07:01:29.144935 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d3864a91-9bf6-4540-a908-cc41aa39656c-host\") pod \"node-ca-kz77f\" (UID: \"d3864a91-9bf6-4540-a908-cc41aa39656c\") " pod="openshift-image-registry/node-ca-kz77f" Oct 14 07:01:29 crc kubenswrapper[4870]: I1014 07:01:29.144641 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d3864a91-9bf6-4540-a908-cc41aa39656c-host\") pod \"node-ca-kz77f\" (UID: \"d3864a91-9bf6-4540-a908-cc41aa39656c\") " pod="openshift-image-registry/node-ca-kz77f" Oct 14 07:01:29 crc kubenswrapper[4870]: I1014 07:01:29.145091 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/d3864a91-9bf6-4540-a908-cc41aa39656c-serviceca\") pod \"node-ca-kz77f\" (UID: \"d3864a91-9bf6-4540-a908-cc41aa39656c\") " pod="openshift-image-registry/node-ca-kz77f" Oct 14 07:01:29 crc kubenswrapper[4870]: I1014 07:01:29.146245 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v4h8v\" (UniqueName: \"kubernetes.io/projected/d3864a91-9bf6-4540-a908-cc41aa39656c-kube-api-access-v4h8v\") pod \"node-ca-kz77f\" (UID: \"d3864a91-9bf6-4540-a908-cc41aa39656c\") " pod="openshift-image-registry/node-ca-kz77f" Oct 14 07:01:29 crc kubenswrapper[4870]: I1014 07:01:29.146274 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/d3864a91-9bf6-4540-a908-cc41aa39656c-serviceca\") pod \"node-ca-kz77f\" (UID: \"d3864a91-9bf6-4540-a908-cc41aa39656c\") " pod="openshift-image-registry/node-ca-kz77f" Oct 14 07:01:29 crc kubenswrapper[4870]: I1014 07:01:29.193970 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v4h8v\" (UniqueName: \"kubernetes.io/projected/d3864a91-9bf6-4540-a908-cc41aa39656c-kube-api-access-v4h8v\") pod \"node-ca-kz77f\" (UID: \"d3864a91-9bf6-4540-a908-cc41aa39656c\") " pod="openshift-image-registry/node-ca-kz77f" Oct 14 07:01:29 crc kubenswrapper[4870]: I1014 07:01:29.202975 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"874a5e45-dffd-4d17-b609-b7d3ed2eab07\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d98d044836ab44a405196915ef7bf572b75113f2c81ec7e64effa63b47ad3543\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wj4ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8004752575df4a1c44efbcafe64799f09126d0f97285d92388c76a0b90e022f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wj4ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7tvc7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:29Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:29 crc kubenswrapper[4870]: I1014 07:01:29.243172 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:29Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:29 crc kubenswrapper[4870]: I1014 07:01:29.287191 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-kz77f" Oct 14 07:01:29 crc kubenswrapper[4870]: I1014 07:01:29.305805 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:29Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:29 crc kubenswrapper[4870]: I1014 07:01:29.333170 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e26fe49ec2045d20a0cb2dd17e96e79b6eb43c641eb184c688cc472ca5db39d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aea115f7ee71c700b218fe15c928af7b55e6babe70e65d67f125d87be1fd2e2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:29Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:29 crc kubenswrapper[4870]: I1014 07:01:29.334321 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-v7br2" event={"ID":"8dcaef0e-42e3-424b-bbe5-4a023efe912d","Type":"ContainerStarted","Data":"d16ed1c65925549dcb69ab02c3013e25e3f123a83179f1af3a8e08bf91a59896"} Oct 14 07:01:29 crc kubenswrapper[4870]: I1014 07:01:29.334360 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-v7br2" event={"ID":"8dcaef0e-42e3-424b-bbe5-4a023efe912d","Type":"ContainerStarted","Data":"e72cff0723e2ce520340244b4e597d8016c58e1b071c4b087f2f6cd718741594"} Oct 14 07:01:29 crc kubenswrapper[4870]: I1014 07:01:29.334372 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-v7br2" event={"ID":"8dcaef0e-42e3-424b-bbe5-4a023efe912d","Type":"ContainerStarted","Data":"71dec55e859b11a081dbb90d7d6cf6b4846cf591382303d8329003ff50c71f2a"} Oct 14 07:01:29 crc kubenswrapper[4870]: I1014 07:01:29.334382 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-v7br2" event={"ID":"8dcaef0e-42e3-424b-bbe5-4a023efe912d","Type":"ContainerStarted","Data":"ed59243b59c39ce82bf9bc8f33c0a8b84854846afff6a89f5d37ffc97a096b7c"} Oct 14 07:01:29 crc kubenswrapper[4870]: I1014 07:01:29.334391 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-v7br2" event={"ID":"8dcaef0e-42e3-424b-bbe5-4a023efe912d","Type":"ContainerStarted","Data":"f3b6d20cf459b1f6d3aff5b9290f9391909baf27ee40c49e2b0cd2d5cfe889ad"} Oct 14 07:01:29 crc kubenswrapper[4870]: I1014 07:01:29.339238 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-cdjmj" event={"ID":"612e7ab9-bda4-45d8-bc50-ff1afb68aace","Type":"ContainerStarted","Data":"9c1972e7b2f75790db22a07a078eee97072bffbdb7f8e594bf2be5f800632a7b"} Oct 14 07:01:29 crc kubenswrapper[4870]: I1014 07:01:29.344310 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"3eb0b397c05ffd6650607f1be4a75683e0342d3333fc096bf1cd63c7688834f9"} Oct 14 07:01:29 crc kubenswrapper[4870]: I1014 07:01:29.369831 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-v7br2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dcaef0e-42e3-424b-bbe5-4a023efe912d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ba24c2e067b9f3a14d9d018b66300b54d866530d30174fbdb3c899299575d28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ba24c2e067b9f3a14d9d018b66300b54d866530d30174fbdb3c899299575d28\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-v7br2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:29Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:29 crc kubenswrapper[4870]: I1014 07:01:29.407107 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77a0d93a-5391-4112-b502-c474f45de209\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7815af9a584f6db6660073da289f2e8c987a04ef9ddfac155bdd9a15b014fcfa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b329b570622155a55f70501b2260e278106a68ebef497914231b4ec860e0125\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba942a0e6fe5ab93a187e0b7b4bd7c440c2e7827d9b5323fc50f90b562075392\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://978e8db184ba9e430232c6bb2ee530beb92567fdc71ba3c35b0bedca3492d845\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://978e8db184ba9e430232c6bb2ee530beb92567fdc71ba3c35b0bedca3492d845\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"file observer\\\\nW1014 07:01:25.870672 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1014 07:01:25.870795 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1014 07:01:25.871521 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-756714914/tls.crt::/tmp/serving-cert-756714914/tls.key\\\\\\\"\\\\nI1014 07:01:26.460037 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1014 07:01:26.464876 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1014 07:01:26.464902 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1014 07:01:26.464925 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1014 07:01:26.464931 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1014 07:01:26.481374 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1014 07:01:26.481400 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:01:26.481405 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:01:26.481410 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1014 07:01:26.481414 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1014 07:01:26.481417 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1014 07:01:26.481420 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1014 07:01:26.481610 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1014 07:01:26.482872 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bd0044a247bdc5ca4cab7294103af3270b76cde61f3109b85e83ed050ca53b6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82d3c2214e0be679d2827594b159ace362151318689078bc49d3158f42899636\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82d3c2214e0be679d2827594b159ace362151318689078bc49d3158f42899636\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:05Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:29Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:29 crc kubenswrapper[4870]: I1014 07:01:29.444085 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-4v2v2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4796ba6-c23b-4764-be96-2e3a3bf5e962\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf5afb0b99a7101f6d5573fb7eb925c5a1d8becc11b2e3c6f3351d80b6f0e823\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnp8b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-4v2v2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:29Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:29 crc kubenswrapper[4870]: I1014 07:01:29.485486 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-cdjmj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"612e7ab9-bda4-45d8-bc50-ff1afb68aace\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://834444b5350ee12df4e4652c3009880d8d69e1edf8ba0c0156ab96fb2bbe28e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://834444b5350ee12df4e4652c3009880d8d69e1edf8ba0c0156ab96fb2bbe28e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-cdjmj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:29Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:29 crc kubenswrapper[4870]: I1014 07:01:29.523493 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fdxjh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a3a430a-b540-4a8c-adad-b2893dbf9898\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93e41d5e927eed7ec62eb9c3607cc804cd8999556663363fd123b2a784b00ff1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k787j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fdxjh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:29Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:29 crc kubenswrapper[4870]: I1014 07:01:29.569545 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33a9493e-9660-48db-b8d1-310e29ebe313\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3570643ef1b7d5cc3a542f320413b819338e0c8b7ab7f0f09cc909997a39bab8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb398668197dedc1c2a14b4ae7d6a29714f3527309ccb9c1622af841100e1d60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1d2b6d068d63bd38cdfb83824525f6d12054f6cabab5d684e03ea5f2152076c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://783ff243e395333f3fc80cfae21358c532f0c9c5cb5515586be56b90ed10d912\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a8135a3d01076e4437ae8a784b459bf8d16de57fa16ec7ce4d40c5129c9d4a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48ca3a5017ba650df99b2fc7201c3fe5abfa89d81089c5647cde1b1549f534fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48ca3a5017ba650df99b2fc7201c3fe5abfa89d81089c5647cde1b1549f534fc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://19682eca67f76bf2e06e1b80f12e21141fcbce0acb8195f0defacd1b9a10b5e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://19682eca67f76bf2e06e1b80f12e21141fcbce0acb8195f0defacd1b9a10b5e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://042aad6d5f6bee8eef77f9e6a2c265de775a69f3da72aedc68bdf312198f00fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://042aad6d5f6bee8eef77f9e6a2c265de775a69f3da72aedc68bdf312198f00fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:05Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:29Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:29 crc kubenswrapper[4870]: I1014 07:01:29.602252 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"874a5e45-dffd-4d17-b609-b7d3ed2eab07\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d98d044836ab44a405196915ef7bf572b75113f2c81ec7e64effa63b47ad3543\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wj4ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8004752575df4a1c44efbcafe64799f09126d0f97285d92388c76a0b90e022f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wj4ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7tvc7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:29Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:29 crc kubenswrapper[4870]: I1014 07:01:29.645954 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:29Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:29 crc kubenswrapper[4870]: I1014 07:01:29.651318 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 07:01:29 crc kubenswrapper[4870]: E1014 07:01:29.651543 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 07:01:33.651503662 +0000 UTC m=+29.348864173 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:01:29 crc kubenswrapper[4870]: I1014 07:01:29.684066 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3eb0b397c05ffd6650607f1be4a75683e0342d3333fc096bf1cd63c7688834f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:29Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:29 crc kubenswrapper[4870]: I1014 07:01:29.724898 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e26fe49ec2045d20a0cb2dd17e96e79b6eb43c641eb184c688cc472ca5db39d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aea115f7ee71c700b218fe15c928af7b55e6babe70e65d67f125d87be1fd2e2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:29Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:29 crc kubenswrapper[4870]: I1014 07:01:29.753099 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 07:01:29 crc kubenswrapper[4870]: I1014 07:01:29.753215 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 07:01:29 crc kubenswrapper[4870]: I1014 07:01:29.753293 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 07:01:29 crc kubenswrapper[4870]: E1014 07:01:29.753341 4870 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 14 07:01:29 crc kubenswrapper[4870]: E1014 07:01:29.753411 4870 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 14 07:01:29 crc kubenswrapper[4870]: I1014 07:01:29.753356 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 07:01:29 crc kubenswrapper[4870]: E1014 07:01:29.753500 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-14 07:01:33.753466084 +0000 UTC m=+29.450826455 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 14 07:01:29 crc kubenswrapper[4870]: E1014 07:01:29.753597 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-14 07:01:33.753577587 +0000 UTC m=+29.450937958 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 14 07:01:29 crc kubenswrapper[4870]: E1014 07:01:29.753596 4870 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 14 07:01:29 crc kubenswrapper[4870]: E1014 07:01:29.753631 4870 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 14 07:01:29 crc kubenswrapper[4870]: E1014 07:01:29.753653 4870 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 14 07:01:29 crc kubenswrapper[4870]: E1014 07:01:29.753737 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-14 07:01:33.75371083 +0000 UTC m=+29.451071231 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 14 07:01:29 crc kubenswrapper[4870]: E1014 07:01:29.753547 4870 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 14 07:01:29 crc kubenswrapper[4870]: E1014 07:01:29.753784 4870 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 14 07:01:29 crc kubenswrapper[4870]: E1014 07:01:29.753799 4870 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 14 07:01:29 crc kubenswrapper[4870]: E1014 07:01:29.753844 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-14 07:01:33.753831583 +0000 UTC m=+29.451191994 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 14 07:01:29 crc kubenswrapper[4870]: I1014 07:01:29.775933 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-v7br2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dcaef0e-42e3-424b-bbe5-4a023efe912d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ba24c2e067b9f3a14d9d018b66300b54d866530d30174fbdb3c899299575d28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ba24c2e067b9f3a14d9d018b66300b54d866530d30174fbdb3c899299575d28\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-v7br2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:29Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:29 crc kubenswrapper[4870]: I1014 07:01:29.806841 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77a0d93a-5391-4112-b502-c474f45de209\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7815af9a584f6db6660073da289f2e8c987a04ef9ddfac155bdd9a15b014fcfa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b329b570622155a55f70501b2260e278106a68ebef497914231b4ec860e0125\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba942a0e6fe5ab93a187e0b7b4bd7c440c2e7827d9b5323fc50f90b562075392\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://978e8db184ba9e430232c6bb2ee530beb92567fdc71ba3c35b0bedca3492d845\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://978e8db184ba9e430232c6bb2ee530beb92567fdc71ba3c35b0bedca3492d845\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"file observer\\\\nW1014 07:01:25.870672 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1014 07:01:25.870795 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1014 07:01:25.871521 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-756714914/tls.crt::/tmp/serving-cert-756714914/tls.key\\\\\\\"\\\\nI1014 07:01:26.460037 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1014 07:01:26.464876 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1014 07:01:26.464902 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1014 07:01:26.464925 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1014 07:01:26.464931 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1014 07:01:26.481374 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1014 07:01:26.481400 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:01:26.481405 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:01:26.481410 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1014 07:01:26.481414 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1014 07:01:26.481417 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1014 07:01:26.481420 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1014 07:01:26.481610 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1014 07:01:26.482872 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bd0044a247bdc5ca4cab7294103af3270b76cde61f3109b85e83ed050ca53b6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82d3c2214e0be679d2827594b159ace362151318689078bc49d3158f42899636\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82d3c2214e0be679d2827594b159ace362151318689078bc49d3158f42899636\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:05Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:29Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:29 crc kubenswrapper[4870]: I1014 07:01:29.844216 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:29Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:29 crc kubenswrapper[4870]: I1014 07:01:29.884195 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-4v2v2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4796ba6-c23b-4764-be96-2e3a3bf5e962\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf5afb0b99a7101f6d5573fb7eb925c5a1d8becc11b2e3c6f3351d80b6f0e823\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnp8b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-4v2v2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:29Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:29 crc kubenswrapper[4870]: I1014 07:01:29.929413 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-cdjmj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"612e7ab9-bda4-45d8-bc50-ff1afb68aace\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://834444b5350ee12df4e4652c3009880d8d69e1edf8ba0c0156ab96fb2bbe28e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://834444b5350ee12df4e4652c3009880d8d69e1edf8ba0c0156ab96fb2bbe28e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c1972e7b2f75790db22a07a078eee97072bffbdb7f8e594bf2be5f800632a7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-cdjmj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:29Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:29 crc kubenswrapper[4870]: I1014 07:01:29.968679 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fdxjh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a3a430a-b540-4a8c-adad-b2893dbf9898\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93e41d5e927eed7ec62eb9c3607cc804cd8999556663363fd123b2a784b00ff1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k787j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fdxjh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:29Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:30 crc kubenswrapper[4870]: I1014 07:01:30.015673 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33a9493e-9660-48db-b8d1-310e29ebe313\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3570643ef1b7d5cc3a542f320413b819338e0c8b7ab7f0f09cc909997a39bab8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb398668197dedc1c2a14b4ae7d6a29714f3527309ccb9c1622af841100e1d60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1d2b6d068d63bd38cdfb83824525f6d12054f6cabab5d684e03ea5f2152076c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://783ff243e395333f3fc80cfae21358c532f0c9c5cb5515586be56b90ed10d912\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a8135a3d01076e4437ae8a784b459bf8d16de57fa16ec7ce4d40c5129c9d4a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48ca3a5017ba650df99b2fc7201c3fe5abfa89d81089c5647cde1b1549f534fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48ca3a5017ba650df99b2fc7201c3fe5abfa89d81089c5647cde1b1549f534fc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://19682eca67f76bf2e06e1b80f12e21141fcbce0acb8195f0defacd1b9a10b5e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://19682eca67f76bf2e06e1b80f12e21141fcbce0acb8195f0defacd1b9a10b5e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://042aad6d5f6bee8eef77f9e6a2c265de775a69f3da72aedc68bdf312198f00fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://042aad6d5f6bee8eef77f9e6a2c265de775a69f3da72aedc68bdf312198f00fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:05Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:30Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:30 crc kubenswrapper[4870]: I1014 07:01:30.033499 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 07:01:30 crc kubenswrapper[4870]: I1014 07:01:30.033544 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 07:01:30 crc kubenswrapper[4870]: E1014 07:01:30.033598 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 07:01:30 crc kubenswrapper[4870]: E1014 07:01:30.033687 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 07:01:30 crc kubenswrapper[4870]: I1014 07:01:30.033775 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 07:01:30 crc kubenswrapper[4870]: E1014 07:01:30.033821 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 07:01:30 crc kubenswrapper[4870]: I1014 07:01:30.045619 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:30Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:30 crc kubenswrapper[4870]: I1014 07:01:30.082735 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-kz77f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d3864a91-9bf6-4540-a908-cc41aa39656c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4h8v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:28Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-kz77f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:30Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:30 crc kubenswrapper[4870]: I1014 07:01:30.125555 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e539fac3d0a2ade314a4b4c1050bb27c37bcff5ba5272fa2d71b97ab5c0d7191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:30Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:30 crc kubenswrapper[4870]: I1014 07:01:30.352915 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-v7br2" event={"ID":"8dcaef0e-42e3-424b-bbe5-4a023efe912d","Type":"ContainerStarted","Data":"fc9d4db944f837289d1923176d60cfc4a566a282c2d2e3a703e220df884703c7"} Oct 14 07:01:30 crc kubenswrapper[4870]: I1014 07:01:30.355641 4870 generic.go:334] "Generic (PLEG): container finished" podID="612e7ab9-bda4-45d8-bc50-ff1afb68aace" containerID="9c1972e7b2f75790db22a07a078eee97072bffbdb7f8e594bf2be5f800632a7b" exitCode=0 Oct 14 07:01:30 crc kubenswrapper[4870]: I1014 07:01:30.355729 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-cdjmj" event={"ID":"612e7ab9-bda4-45d8-bc50-ff1afb68aace","Type":"ContainerDied","Data":"9c1972e7b2f75790db22a07a078eee97072bffbdb7f8e594bf2be5f800632a7b"} Oct 14 07:01:30 crc kubenswrapper[4870]: I1014 07:01:30.358401 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-kz77f" event={"ID":"d3864a91-9bf6-4540-a908-cc41aa39656c","Type":"ContainerStarted","Data":"1d4ded3d86030f5ea548c8b1cc3caddbe14a45c9a3ae4af836e892880584f57d"} Oct 14 07:01:30 crc kubenswrapper[4870]: I1014 07:01:30.358486 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-kz77f" event={"ID":"d3864a91-9bf6-4540-a908-cc41aa39656c","Type":"ContainerStarted","Data":"43ebd4e2452c44b3ca4cb20a05e4bf88194278714e1ac587f243a06e557865ff"} Oct 14 07:01:30 crc kubenswrapper[4870]: I1014 07:01:30.380473 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e539fac3d0a2ade314a4b4c1050bb27c37bcff5ba5272fa2d71b97ab5c0d7191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:30Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:30 crc kubenswrapper[4870]: I1014 07:01:30.401311 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:30Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:30 crc kubenswrapper[4870]: I1014 07:01:30.414022 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-kz77f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d3864a91-9bf6-4540-a908-cc41aa39656c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4h8v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:28Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-kz77f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:30Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:30 crc kubenswrapper[4870]: I1014 07:01:30.429534 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:30Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:30 crc kubenswrapper[4870]: I1014 07:01:30.444289 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3eb0b397c05ffd6650607f1be4a75683e0342d3333fc096bf1cd63c7688834f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:30Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:30 crc kubenswrapper[4870]: I1014 07:01:30.456720 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"874a5e45-dffd-4d17-b609-b7d3ed2eab07\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d98d044836ab44a405196915ef7bf572b75113f2c81ec7e64effa63b47ad3543\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wj4ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8004752575df4a1c44efbcafe64799f09126d0f97285d92388c76a0b90e022f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wj4ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7tvc7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:30Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:30 crc kubenswrapper[4870]: I1014 07:01:30.482599 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77a0d93a-5391-4112-b502-c474f45de209\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7815af9a584f6db6660073da289f2e8c987a04ef9ddfac155bdd9a15b014fcfa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b329b570622155a55f70501b2260e278106a68ebef497914231b4ec860e0125\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba942a0e6fe5ab93a187e0b7b4bd7c440c2e7827d9b5323fc50f90b562075392\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://978e8db184ba9e430232c6bb2ee530beb92567fdc71ba3c35b0bedca3492d845\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://978e8db184ba9e430232c6bb2ee530beb92567fdc71ba3c35b0bedca3492d845\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"file observer\\\\nW1014 07:01:25.870672 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1014 07:01:25.870795 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1014 07:01:25.871521 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-756714914/tls.crt::/tmp/serving-cert-756714914/tls.key\\\\\\\"\\\\nI1014 07:01:26.460037 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1014 07:01:26.464876 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1014 07:01:26.464902 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1014 07:01:26.464925 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1014 07:01:26.464931 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1014 07:01:26.481374 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1014 07:01:26.481400 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:01:26.481405 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:01:26.481410 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1014 07:01:26.481414 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1014 07:01:26.481417 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1014 07:01:26.481420 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1014 07:01:26.481610 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1014 07:01:26.482872 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bd0044a247bdc5ca4cab7294103af3270b76cde61f3109b85e83ed050ca53b6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82d3c2214e0be679d2827594b159ace362151318689078bc49d3158f42899636\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82d3c2214e0be679d2827594b159ace362151318689078bc49d3158f42899636\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:05Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:30Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:30 crc kubenswrapper[4870]: I1014 07:01:30.503404 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:30Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:30 crc kubenswrapper[4870]: I1014 07:01:30.528561 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e26fe49ec2045d20a0cb2dd17e96e79b6eb43c641eb184c688cc472ca5db39d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aea115f7ee71c700b218fe15c928af7b55e6babe70e65d67f125d87be1fd2e2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:30Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:30 crc kubenswrapper[4870]: I1014 07:01:30.550296 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-v7br2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dcaef0e-42e3-424b-bbe5-4a023efe912d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ba24c2e067b9f3a14d9d018b66300b54d866530d30174fbdb3c899299575d28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ba24c2e067b9f3a14d9d018b66300b54d866530d30174fbdb3c899299575d28\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-v7br2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:30Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:30 crc kubenswrapper[4870]: I1014 07:01:30.576880 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33a9493e-9660-48db-b8d1-310e29ebe313\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3570643ef1b7d5cc3a542f320413b819338e0c8b7ab7f0f09cc909997a39bab8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb398668197dedc1c2a14b4ae7d6a29714f3527309ccb9c1622af841100e1d60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1d2b6d068d63bd38cdfb83824525f6d12054f6cabab5d684e03ea5f2152076c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://783ff243e395333f3fc80cfae21358c532f0c9c5cb5515586be56b90ed10d912\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a8135a3d01076e4437ae8a784b459bf8d16de57fa16ec7ce4d40c5129c9d4a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48ca3a5017ba650df99b2fc7201c3fe5abfa89d81089c5647cde1b1549f534fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48ca3a5017ba650df99b2fc7201c3fe5abfa89d81089c5647cde1b1549f534fc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://19682eca67f76bf2e06e1b80f12e21141fcbce0acb8195f0defacd1b9a10b5e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://19682eca67f76bf2e06e1b80f12e21141fcbce0acb8195f0defacd1b9a10b5e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://042aad6d5f6bee8eef77f9e6a2c265de775a69f3da72aedc68bdf312198f00fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://042aad6d5f6bee8eef77f9e6a2c265de775a69f3da72aedc68bdf312198f00fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:05Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:30Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:30 crc kubenswrapper[4870]: I1014 07:01:30.601518 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-4v2v2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4796ba6-c23b-4764-be96-2e3a3bf5e962\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf5afb0b99a7101f6d5573fb7eb925c5a1d8becc11b2e3c6f3351d80b6f0e823\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnp8b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-4v2v2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:30Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:30 crc kubenswrapper[4870]: I1014 07:01:30.646866 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-cdjmj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"612e7ab9-bda4-45d8-bc50-ff1afb68aace\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://834444b5350ee12df4e4652c3009880d8d69e1edf8ba0c0156ab96fb2bbe28e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://834444b5350ee12df4e4652c3009880d8d69e1edf8ba0c0156ab96fb2bbe28e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c1972e7b2f75790db22a07a078eee97072bffbdb7f8e594bf2be5f800632a7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c1972e7b2f75790db22a07a078eee97072bffbdb7f8e594bf2be5f800632a7b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-cdjmj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:30Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:30 crc kubenswrapper[4870]: I1014 07:01:30.684286 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fdxjh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a3a430a-b540-4a8c-adad-b2893dbf9898\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93e41d5e927eed7ec62eb9c3607cc804cd8999556663363fd123b2a784b00ff1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k787j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fdxjh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:30Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:30 crc kubenswrapper[4870]: I1014 07:01:30.734866 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:30Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:30 crc kubenswrapper[4870]: I1014 07:01:30.763999 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3eb0b397c05ffd6650607f1be4a75683e0342d3333fc096bf1cd63c7688834f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:30Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:30 crc kubenswrapper[4870]: I1014 07:01:30.803536 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"874a5e45-dffd-4d17-b609-b7d3ed2eab07\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d98d044836ab44a405196915ef7bf572b75113f2c81ec7e64effa63b47ad3543\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wj4ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8004752575df4a1c44efbcafe64799f09126d0f97285d92388c76a0b90e022f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wj4ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7tvc7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:30Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:30 crc kubenswrapper[4870]: I1014 07:01:30.846428 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77a0d93a-5391-4112-b502-c474f45de209\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7815af9a584f6db6660073da289f2e8c987a04ef9ddfac155bdd9a15b014fcfa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b329b570622155a55f70501b2260e278106a68ebef497914231b4ec860e0125\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba942a0e6fe5ab93a187e0b7b4bd7c440c2e7827d9b5323fc50f90b562075392\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://978e8db184ba9e430232c6bb2ee530beb92567fdc71ba3c35b0bedca3492d845\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://978e8db184ba9e430232c6bb2ee530beb92567fdc71ba3c35b0bedca3492d845\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"file observer\\\\nW1014 07:01:25.870672 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1014 07:01:25.870795 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1014 07:01:25.871521 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-756714914/tls.crt::/tmp/serving-cert-756714914/tls.key\\\\\\\"\\\\nI1014 07:01:26.460037 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1014 07:01:26.464876 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1014 07:01:26.464902 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1014 07:01:26.464925 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1014 07:01:26.464931 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1014 07:01:26.481374 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1014 07:01:26.481400 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:01:26.481405 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:01:26.481410 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1014 07:01:26.481414 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1014 07:01:26.481417 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1014 07:01:26.481420 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1014 07:01:26.481610 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1014 07:01:26.482872 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bd0044a247bdc5ca4cab7294103af3270b76cde61f3109b85e83ed050ca53b6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82d3c2214e0be679d2827594b159ace362151318689078bc49d3158f42899636\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82d3c2214e0be679d2827594b159ace362151318689078bc49d3158f42899636\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:05Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:30Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:30 crc kubenswrapper[4870]: I1014 07:01:30.888898 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:30Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:30 crc kubenswrapper[4870]: I1014 07:01:30.925555 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e26fe49ec2045d20a0cb2dd17e96e79b6eb43c641eb184c688cc472ca5db39d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aea115f7ee71c700b218fe15c928af7b55e6babe70e65d67f125d87be1fd2e2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:30Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:30 crc kubenswrapper[4870]: I1014 07:01:30.971805 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-v7br2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dcaef0e-42e3-424b-bbe5-4a023efe912d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ba24c2e067b9f3a14d9d018b66300b54d866530d30174fbdb3c899299575d28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ba24c2e067b9f3a14d9d018b66300b54d866530d30174fbdb3c899299575d28\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-v7br2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:30Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:31 crc kubenswrapper[4870]: I1014 07:01:31.013578 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-cdjmj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"612e7ab9-bda4-45d8-bc50-ff1afb68aace\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://834444b5350ee12df4e4652c3009880d8d69e1edf8ba0c0156ab96fb2bbe28e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://834444b5350ee12df4e4652c3009880d8d69e1edf8ba0c0156ab96fb2bbe28e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c1972e7b2f75790db22a07a078eee97072bffbdb7f8e594bf2be5f800632a7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c1972e7b2f75790db22a07a078eee97072bffbdb7f8e594bf2be5f800632a7b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-cdjmj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:31Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:31 crc kubenswrapper[4870]: I1014 07:01:31.044785 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fdxjh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a3a430a-b540-4a8c-adad-b2893dbf9898\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93e41d5e927eed7ec62eb9c3607cc804cd8999556663363fd123b2a784b00ff1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k787j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fdxjh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:31Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:31 crc kubenswrapper[4870]: I1014 07:01:31.093158 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33a9493e-9660-48db-b8d1-310e29ebe313\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3570643ef1b7d5cc3a542f320413b819338e0c8b7ab7f0f09cc909997a39bab8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb398668197dedc1c2a14b4ae7d6a29714f3527309ccb9c1622af841100e1d60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1d2b6d068d63bd38cdfb83824525f6d12054f6cabab5d684e03ea5f2152076c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://783ff243e395333f3fc80cfae21358c532f0c9c5cb5515586be56b90ed10d912\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a8135a3d01076e4437ae8a784b459bf8d16de57fa16ec7ce4d40c5129c9d4a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48ca3a5017ba650df99b2fc7201c3fe5abfa89d81089c5647cde1b1549f534fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48ca3a5017ba650df99b2fc7201c3fe5abfa89d81089c5647cde1b1549f534fc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://19682eca67f76bf2e06e1b80f12e21141fcbce0acb8195f0defacd1b9a10b5e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://19682eca67f76bf2e06e1b80f12e21141fcbce0acb8195f0defacd1b9a10b5e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://042aad6d5f6bee8eef77f9e6a2c265de775a69f3da72aedc68bdf312198f00fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://042aad6d5f6bee8eef77f9e6a2c265de775a69f3da72aedc68bdf312198f00fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:05Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:31Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:31 crc kubenswrapper[4870]: I1014 07:01:31.125271 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-4v2v2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4796ba6-c23b-4764-be96-2e3a3bf5e962\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf5afb0b99a7101f6d5573fb7eb925c5a1d8becc11b2e3c6f3351d80b6f0e823\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnp8b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-4v2v2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:31Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:31 crc kubenswrapper[4870]: I1014 07:01:31.154528 4870 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 14 07:01:31 crc kubenswrapper[4870]: I1014 07:01:31.155467 4870 scope.go:117] "RemoveContainer" containerID="978e8db184ba9e430232c6bb2ee530beb92567fdc71ba3c35b0bedca3492d845" Oct 14 07:01:31 crc kubenswrapper[4870]: E1014 07:01:31.155675 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Oct 14 07:01:31 crc kubenswrapper[4870]: I1014 07:01:31.164940 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e539fac3d0a2ade314a4b4c1050bb27c37bcff5ba5272fa2d71b97ab5c0d7191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:31Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:31 crc kubenswrapper[4870]: I1014 07:01:31.207415 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:31Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:31 crc kubenswrapper[4870]: I1014 07:01:31.242195 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-kz77f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d3864a91-9bf6-4540-a908-cc41aa39656c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d4ded3d86030f5ea548c8b1cc3caddbe14a45c9a3ae4af836e892880584f57d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4h8v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:28Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-kz77f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:31Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:31 crc kubenswrapper[4870]: I1014 07:01:31.364487 4870 generic.go:334] "Generic (PLEG): container finished" podID="612e7ab9-bda4-45d8-bc50-ff1afb68aace" containerID="5b124b42a44c6a23da1d92b9a5326987eb74430f80489a71e8913829475654ab" exitCode=0 Oct 14 07:01:31 crc kubenswrapper[4870]: I1014 07:01:31.364576 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-cdjmj" event={"ID":"612e7ab9-bda4-45d8-bc50-ff1afb68aace","Type":"ContainerDied","Data":"5b124b42a44c6a23da1d92b9a5326987eb74430f80489a71e8913829475654ab"} Oct 14 07:01:31 crc kubenswrapper[4870]: I1014 07:01:31.385833 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-4v2v2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4796ba6-c23b-4764-be96-2e3a3bf5e962\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf5afb0b99a7101f6d5573fb7eb925c5a1d8becc11b2e3c6f3351d80b6f0e823\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnp8b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-4v2v2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:31Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:31 crc kubenswrapper[4870]: I1014 07:01:31.404376 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-cdjmj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"612e7ab9-bda4-45d8-bc50-ff1afb68aace\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://834444b5350ee12df4e4652c3009880d8d69e1edf8ba0c0156ab96fb2bbe28e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://834444b5350ee12df4e4652c3009880d8d69e1edf8ba0c0156ab96fb2bbe28e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c1972e7b2f75790db22a07a078eee97072bffbdb7f8e594bf2be5f800632a7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c1972e7b2f75790db22a07a078eee97072bffbdb7f8e594bf2be5f800632a7b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b124b42a44c6a23da1d92b9a5326987eb74430f80489a71e8913829475654ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b124b42a44c6a23da1d92b9a5326987eb74430f80489a71e8913829475654ab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-cdjmj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:31Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:31 crc kubenswrapper[4870]: I1014 07:01:31.429190 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fdxjh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a3a430a-b540-4a8c-adad-b2893dbf9898\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93e41d5e927eed7ec62eb9c3607cc804cd8999556663363fd123b2a784b00ff1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k787j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fdxjh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:31Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:31 crc kubenswrapper[4870]: I1014 07:01:31.459523 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33a9493e-9660-48db-b8d1-310e29ebe313\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3570643ef1b7d5cc3a542f320413b819338e0c8b7ab7f0f09cc909997a39bab8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb398668197dedc1c2a14b4ae7d6a29714f3527309ccb9c1622af841100e1d60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1d2b6d068d63bd38cdfb83824525f6d12054f6cabab5d684e03ea5f2152076c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://783ff243e395333f3fc80cfae21358c532f0c9c5cb5515586be56b90ed10d912\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a8135a3d01076e4437ae8a784b459bf8d16de57fa16ec7ce4d40c5129c9d4a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48ca3a5017ba650df99b2fc7201c3fe5abfa89d81089c5647cde1b1549f534fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48ca3a5017ba650df99b2fc7201c3fe5abfa89d81089c5647cde1b1549f534fc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://19682eca67f76bf2e06e1b80f12e21141fcbce0acb8195f0defacd1b9a10b5e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://19682eca67f76bf2e06e1b80f12e21141fcbce0acb8195f0defacd1b9a10b5e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://042aad6d5f6bee8eef77f9e6a2c265de775a69f3da72aedc68bdf312198f00fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://042aad6d5f6bee8eef77f9e6a2c265de775a69f3da72aedc68bdf312198f00fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:05Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:31Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:31 crc kubenswrapper[4870]: I1014 07:01:31.477092 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-kz77f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d3864a91-9bf6-4540-a908-cc41aa39656c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d4ded3d86030f5ea548c8b1cc3caddbe14a45c9a3ae4af836e892880584f57d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4h8v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:28Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-kz77f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:31Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:31 crc kubenswrapper[4870]: I1014 07:01:31.495639 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e539fac3d0a2ade314a4b4c1050bb27c37bcff5ba5272fa2d71b97ab5c0d7191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:31Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:31 crc kubenswrapper[4870]: I1014 07:01:31.525017 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:31Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:31 crc kubenswrapper[4870]: I1014 07:01:31.564716 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:31Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:31 crc kubenswrapper[4870]: I1014 07:01:31.602628 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3eb0b397c05ffd6650607f1be4a75683e0342d3333fc096bf1cd63c7688834f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:31Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:31 crc kubenswrapper[4870]: I1014 07:01:31.643839 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"874a5e45-dffd-4d17-b609-b7d3ed2eab07\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d98d044836ab44a405196915ef7bf572b75113f2c81ec7e64effa63b47ad3543\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wj4ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8004752575df4a1c44efbcafe64799f09126d0f97285d92388c76a0b90e022f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wj4ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7tvc7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:31Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:31 crc kubenswrapper[4870]: I1014 07:01:31.688422 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-v7br2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dcaef0e-42e3-424b-bbe5-4a023efe912d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ba24c2e067b9f3a14d9d018b66300b54d866530d30174fbdb3c899299575d28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ba24c2e067b9f3a14d9d018b66300b54d866530d30174fbdb3c899299575d28\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-v7br2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:31Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:31 crc kubenswrapper[4870]: I1014 07:01:31.726078 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77a0d93a-5391-4112-b502-c474f45de209\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7815af9a584f6db6660073da289f2e8c987a04ef9ddfac155bdd9a15b014fcfa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b329b570622155a55f70501b2260e278106a68ebef497914231b4ec860e0125\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba942a0e6fe5ab93a187e0b7b4bd7c440c2e7827d9b5323fc50f90b562075392\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://978e8db184ba9e430232c6bb2ee530beb92567fdc71ba3c35b0bedca3492d845\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://978e8db184ba9e430232c6bb2ee530beb92567fdc71ba3c35b0bedca3492d845\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"file observer\\\\nW1014 07:01:25.870672 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1014 07:01:25.870795 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1014 07:01:25.871521 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-756714914/tls.crt::/tmp/serving-cert-756714914/tls.key\\\\\\\"\\\\nI1014 07:01:26.460037 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1014 07:01:26.464876 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1014 07:01:26.464902 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1014 07:01:26.464925 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1014 07:01:26.464931 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1014 07:01:26.481374 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1014 07:01:26.481400 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:01:26.481405 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:01:26.481410 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1014 07:01:26.481414 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1014 07:01:26.481417 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1014 07:01:26.481420 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1014 07:01:26.481610 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1014 07:01:26.482872 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bd0044a247bdc5ca4cab7294103af3270b76cde61f3109b85e83ed050ca53b6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82d3c2214e0be679d2827594b159ace362151318689078bc49d3158f42899636\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82d3c2214e0be679d2827594b159ace362151318689078bc49d3158f42899636\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:05Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:31Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:31 crc kubenswrapper[4870]: I1014 07:01:31.767758 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:31Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:31 crc kubenswrapper[4870]: I1014 07:01:31.803231 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e26fe49ec2045d20a0cb2dd17e96e79b6eb43c641eb184c688cc472ca5db39d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aea115f7ee71c700b218fe15c928af7b55e6babe70e65d67f125d87be1fd2e2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:31Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:32 crc kubenswrapper[4870]: I1014 07:01:32.033355 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 07:01:32 crc kubenswrapper[4870]: E1014 07:01:32.033536 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 07:01:32 crc kubenswrapper[4870]: I1014 07:01:32.033536 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 07:01:32 crc kubenswrapper[4870]: I1014 07:01:32.033564 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 07:01:32 crc kubenswrapper[4870]: E1014 07:01:32.033639 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 07:01:32 crc kubenswrapper[4870]: E1014 07:01:32.033787 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 07:01:32 crc kubenswrapper[4870]: I1014 07:01:32.241816 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 14 07:01:32 crc kubenswrapper[4870]: I1014 07:01:32.248232 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 14 07:01:32 crc kubenswrapper[4870]: I1014 07:01:32.256413 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Oct 14 07:01:32 crc kubenswrapper[4870]: I1014 07:01:32.260096 4870 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 14 07:01:32 crc kubenswrapper[4870]: I1014 07:01:32.263111 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:32 crc kubenswrapper[4870]: I1014 07:01:32.263171 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:32 crc kubenswrapper[4870]: I1014 07:01:32.263187 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:32 crc kubenswrapper[4870]: I1014 07:01:32.263354 4870 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 14 07:01:32 crc kubenswrapper[4870]: I1014 07:01:32.270415 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33a9493e-9660-48db-b8d1-310e29ebe313\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3570643ef1b7d5cc3a542f320413b819338e0c8b7ab7f0f09cc909997a39bab8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb398668197dedc1c2a14b4ae7d6a29714f3527309ccb9c1622af841100e1d60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1d2b6d068d63bd38cdfb83824525f6d12054f6cabab5d684e03ea5f2152076c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://783ff243e395333f3fc80cfae21358c532f0c9c5cb5515586be56b90ed10d912\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a8135a3d01076e4437ae8a784b459bf8d16de57fa16ec7ce4d40c5129c9d4a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48ca3a5017ba650df99b2fc7201c3fe5abfa89d81089c5647cde1b1549f534fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48ca3a5017ba650df99b2fc7201c3fe5abfa89d81089c5647cde1b1549f534fc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://19682eca67f76bf2e06e1b80f12e21141fcbce0acb8195f0defacd1b9a10b5e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://19682eca67f76bf2e06e1b80f12e21141fcbce0acb8195f0defacd1b9a10b5e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://042aad6d5f6bee8eef77f9e6a2c265de775a69f3da72aedc68bdf312198f00fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://042aad6d5f6bee8eef77f9e6a2c265de775a69f3da72aedc68bdf312198f00fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:05Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:32Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:32 crc kubenswrapper[4870]: I1014 07:01:32.273405 4870 kubelet_node_status.go:115] "Node was previously registered" node="crc" Oct 14 07:01:32 crc kubenswrapper[4870]: I1014 07:01:32.273779 4870 kubelet_node_status.go:79] "Successfully registered node" node="crc" Oct 14 07:01:32 crc kubenswrapper[4870]: I1014 07:01:32.275594 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:32 crc kubenswrapper[4870]: I1014 07:01:32.275656 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:32 crc kubenswrapper[4870]: I1014 07:01:32.275669 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:32 crc kubenswrapper[4870]: I1014 07:01:32.275696 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:32 crc kubenswrapper[4870]: I1014 07:01:32.275712 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:32Z","lastTransitionTime":"2025-10-14T07:01:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:32 crc kubenswrapper[4870]: I1014 07:01:32.289185 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-4v2v2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4796ba6-c23b-4764-be96-2e3a3bf5e962\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf5afb0b99a7101f6d5573fb7eb925c5a1d8becc11b2e3c6f3351d80b6f0e823\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnp8b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-4v2v2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:32Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:32 crc kubenswrapper[4870]: E1014 07:01:32.291862 4870 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:01:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:01:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:32Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:01:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:01:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:32Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"56448e6b-6b3a-4bb9-91f8-2f0a9b44e9aa\\\",\\\"systemUUID\\\":\\\"0adb347d-b90d-4e9f-9962-f1bd97980aed\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:32Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:32 crc kubenswrapper[4870]: I1014 07:01:32.296869 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:32 crc kubenswrapper[4870]: I1014 07:01:32.296910 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:32 crc kubenswrapper[4870]: I1014 07:01:32.296922 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:32 crc kubenswrapper[4870]: I1014 07:01:32.296940 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:32 crc kubenswrapper[4870]: I1014 07:01:32.296954 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:32Z","lastTransitionTime":"2025-10-14T07:01:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:32 crc kubenswrapper[4870]: I1014 07:01:32.311963 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-cdjmj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"612e7ab9-bda4-45d8-bc50-ff1afb68aace\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://834444b5350ee12df4e4652c3009880d8d69e1edf8ba0c0156ab96fb2bbe28e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://834444b5350ee12df4e4652c3009880d8d69e1edf8ba0c0156ab96fb2bbe28e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c1972e7b2f75790db22a07a078eee97072bffbdb7f8e594bf2be5f800632a7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c1972e7b2f75790db22a07a078eee97072bffbdb7f8e594bf2be5f800632a7b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b124b42a44c6a23da1d92b9a5326987eb74430f80489a71e8913829475654ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b124b42a44c6a23da1d92b9a5326987eb74430f80489a71e8913829475654ab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-cdjmj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:32Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:32 crc kubenswrapper[4870]: E1014 07:01:32.314013 4870 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:01:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:01:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:32Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:01:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:01:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:32Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"56448e6b-6b3a-4bb9-91f8-2f0a9b44e9aa\\\",\\\"systemUUID\\\":\\\"0adb347d-b90d-4e9f-9962-f1bd97980aed\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:32Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:32 crc kubenswrapper[4870]: I1014 07:01:32.323342 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:32 crc kubenswrapper[4870]: I1014 07:01:32.323400 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:32 crc kubenswrapper[4870]: I1014 07:01:32.323414 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:32 crc kubenswrapper[4870]: I1014 07:01:32.323454 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:32 crc kubenswrapper[4870]: I1014 07:01:32.323470 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:32Z","lastTransitionTime":"2025-10-14T07:01:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:32 crc kubenswrapper[4870]: I1014 07:01:32.334985 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fdxjh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a3a430a-b540-4a8c-adad-b2893dbf9898\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93e41d5e927eed7ec62eb9c3607cc804cd8999556663363fd123b2a784b00ff1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k787j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fdxjh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:32Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:32 crc kubenswrapper[4870]: E1014 07:01:32.341356 4870 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:01:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:01:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:32Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:01:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:01:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:32Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"56448e6b-6b3a-4bb9-91f8-2f0a9b44e9aa\\\",\\\"systemUUID\\\":\\\"0adb347d-b90d-4e9f-9962-f1bd97980aed\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:32Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:32 crc kubenswrapper[4870]: I1014 07:01:32.346584 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:32 crc kubenswrapper[4870]: I1014 07:01:32.346651 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:32 crc kubenswrapper[4870]: I1014 07:01:32.346667 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:32 crc kubenswrapper[4870]: I1014 07:01:32.346687 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:32 crc kubenswrapper[4870]: I1014 07:01:32.346700 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:32Z","lastTransitionTime":"2025-10-14T07:01:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:32 crc kubenswrapper[4870]: I1014 07:01:32.355516 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e539fac3d0a2ade314a4b4c1050bb27c37bcff5ba5272fa2d71b97ab5c0d7191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:32Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:32 crc kubenswrapper[4870]: E1014 07:01:32.363261 4870 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:01:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:01:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:32Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:01:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:01:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:32Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"56448e6b-6b3a-4bb9-91f8-2f0a9b44e9aa\\\",\\\"systemUUID\\\":\\\"0adb347d-b90d-4e9f-9962-f1bd97980aed\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:32Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:32 crc kubenswrapper[4870]: I1014 07:01:32.368285 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:32 crc kubenswrapper[4870]: I1014 07:01:32.368376 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:32 crc kubenswrapper[4870]: I1014 07:01:32.368466 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:32 crc kubenswrapper[4870]: I1014 07:01:32.368492 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:32 crc kubenswrapper[4870]: I1014 07:01:32.368508 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:32Z","lastTransitionTime":"2025-10-14T07:01:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:32 crc kubenswrapper[4870]: I1014 07:01:32.371053 4870 generic.go:334] "Generic (PLEG): container finished" podID="612e7ab9-bda4-45d8-bc50-ff1afb68aace" containerID="62d0deb6b20ea65fbcbaf94c2fc14d69adb22e7a46cc043ccd4def2668ddd4b7" exitCode=0 Oct 14 07:01:32 crc kubenswrapper[4870]: I1014 07:01:32.371110 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-cdjmj" event={"ID":"612e7ab9-bda4-45d8-bc50-ff1afb68aace","Type":"ContainerDied","Data":"62d0deb6b20ea65fbcbaf94c2fc14d69adb22e7a46cc043ccd4def2668ddd4b7"} Oct 14 07:01:32 crc kubenswrapper[4870]: I1014 07:01:32.376770 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:32Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:32 crc kubenswrapper[4870]: I1014 07:01:32.379315 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-v7br2" event={"ID":"8dcaef0e-42e3-424b-bbe5-4a023efe912d","Type":"ContainerStarted","Data":"9fe3468cdd612fc4378e1f6f55a2a9d00274f7ddb65cb57b2a368dfc065cfb4e"} Oct 14 07:01:32 crc kubenswrapper[4870]: E1014 07:01:32.385114 4870 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:01:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:01:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:32Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:01:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:01:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:32Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"56448e6b-6b3a-4bb9-91f8-2f0a9b44e9aa\\\",\\\"systemUUID\\\":\\\"0adb347d-b90d-4e9f-9962-f1bd97980aed\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:32Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:32 crc kubenswrapper[4870]: E1014 07:01:32.385267 4870 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 14 07:01:32 crc kubenswrapper[4870]: I1014 07:01:32.387361 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:32 crc kubenswrapper[4870]: I1014 07:01:32.387389 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:32 crc kubenswrapper[4870]: I1014 07:01:32.387399 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:32 crc kubenswrapper[4870]: I1014 07:01:32.387415 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:32 crc kubenswrapper[4870]: I1014 07:01:32.387462 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:32Z","lastTransitionTime":"2025-10-14T07:01:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:32 crc kubenswrapper[4870]: I1014 07:01:32.393807 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-kz77f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d3864a91-9bf6-4540-a908-cc41aa39656c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d4ded3d86030f5ea548c8b1cc3caddbe14a45c9a3ae4af836e892880584f57d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4h8v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:28Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-kz77f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:32Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:32 crc kubenswrapper[4870]: I1014 07:01:32.413482 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:32Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:32 crc kubenswrapper[4870]: I1014 07:01:32.425294 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3eb0b397c05ffd6650607f1be4a75683e0342d3333fc096bf1cd63c7688834f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:32Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:32 crc kubenswrapper[4870]: I1014 07:01:32.437920 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"874a5e45-dffd-4d17-b609-b7d3ed2eab07\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d98d044836ab44a405196915ef7bf572b75113f2c81ec7e64effa63b47ad3543\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wj4ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8004752575df4a1c44efbcafe64799f09126d0f97285d92388c76a0b90e022f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wj4ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7tvc7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:32Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:32 crc kubenswrapper[4870]: I1014 07:01:32.453422 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77a0d93a-5391-4112-b502-c474f45de209\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7815af9a584f6db6660073da289f2e8c987a04ef9ddfac155bdd9a15b014fcfa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b329b570622155a55f70501b2260e278106a68ebef497914231b4ec860e0125\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba942a0e6fe5ab93a187e0b7b4bd7c440c2e7827d9b5323fc50f90b562075392\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://978e8db184ba9e430232c6bb2ee530beb92567fdc71ba3c35b0bedca3492d845\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://978e8db184ba9e430232c6bb2ee530beb92567fdc71ba3c35b0bedca3492d845\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"file observer\\\\nW1014 07:01:25.870672 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1014 07:01:25.870795 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1014 07:01:25.871521 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-756714914/tls.crt::/tmp/serving-cert-756714914/tls.key\\\\\\\"\\\\nI1014 07:01:26.460037 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1014 07:01:26.464876 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1014 07:01:26.464902 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1014 07:01:26.464925 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1014 07:01:26.464931 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1014 07:01:26.481374 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1014 07:01:26.481400 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:01:26.481405 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:01:26.481410 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1014 07:01:26.481414 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1014 07:01:26.481417 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1014 07:01:26.481420 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1014 07:01:26.481610 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1014 07:01:26.482872 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bd0044a247bdc5ca4cab7294103af3270b76cde61f3109b85e83ed050ca53b6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82d3c2214e0be679d2827594b159ace362151318689078bc49d3158f42899636\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82d3c2214e0be679d2827594b159ace362151318689078bc49d3158f42899636\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:05Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:32Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:32 crc kubenswrapper[4870]: I1014 07:01:32.469119 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:32Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:32 crc kubenswrapper[4870]: I1014 07:01:32.483974 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e26fe49ec2045d20a0cb2dd17e96e79b6eb43c641eb184c688cc472ca5db39d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aea115f7ee71c700b218fe15c928af7b55e6babe70e65d67f125d87be1fd2e2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:32Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:32 crc kubenswrapper[4870]: I1014 07:01:32.493094 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:32 crc kubenswrapper[4870]: I1014 07:01:32.493184 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:32 crc kubenswrapper[4870]: I1014 07:01:32.493201 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:32 crc kubenswrapper[4870]: I1014 07:01:32.493231 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:32 crc kubenswrapper[4870]: I1014 07:01:32.493247 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:32Z","lastTransitionTime":"2025-10-14T07:01:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:32 crc kubenswrapper[4870]: I1014 07:01:32.512511 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-v7br2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dcaef0e-42e3-424b-bbe5-4a023efe912d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ba24c2e067b9f3a14d9d018b66300b54d866530d30174fbdb3c899299575d28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ba24c2e067b9f3a14d9d018b66300b54d866530d30174fbdb3c899299575d28\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-v7br2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:32Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:32 crc kubenswrapper[4870]: I1014 07:01:32.533044 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:32Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:32 crc kubenswrapper[4870]: I1014 07:01:32.550178 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3eb0b397c05ffd6650607f1be4a75683e0342d3333fc096bf1cd63c7688834f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:32Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:32 crc kubenswrapper[4870]: I1014 07:01:32.563950 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"874a5e45-dffd-4d17-b609-b7d3ed2eab07\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d98d044836ab44a405196915ef7bf572b75113f2c81ec7e64effa63b47ad3543\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wj4ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8004752575df4a1c44efbcafe64799f09126d0f97285d92388c76a0b90e022f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wj4ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7tvc7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:32Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:32 crc kubenswrapper[4870]: I1014 07:01:32.583989 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77a0d93a-5391-4112-b502-c474f45de209\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7815af9a584f6db6660073da289f2e8c987a04ef9ddfac155bdd9a15b014fcfa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b329b570622155a55f70501b2260e278106a68ebef497914231b4ec860e0125\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba942a0e6fe5ab93a187e0b7b4bd7c440c2e7827d9b5323fc50f90b562075392\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://978e8db184ba9e430232c6bb2ee530beb92567fdc71ba3c35b0bedca3492d845\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://978e8db184ba9e430232c6bb2ee530beb92567fdc71ba3c35b0bedca3492d845\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"file observer\\\\nW1014 07:01:25.870672 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1014 07:01:25.870795 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1014 07:01:25.871521 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-756714914/tls.crt::/tmp/serving-cert-756714914/tls.key\\\\\\\"\\\\nI1014 07:01:26.460037 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1014 07:01:26.464876 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1014 07:01:26.464902 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1014 07:01:26.464925 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1014 07:01:26.464931 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1014 07:01:26.481374 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1014 07:01:26.481400 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:01:26.481405 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:01:26.481410 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1014 07:01:26.481414 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1014 07:01:26.481417 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1014 07:01:26.481420 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1014 07:01:26.481610 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1014 07:01:26.482872 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bd0044a247bdc5ca4cab7294103af3270b76cde61f3109b85e83ed050ca53b6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82d3c2214e0be679d2827594b159ace362151318689078bc49d3158f42899636\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82d3c2214e0be679d2827594b159ace362151318689078bc49d3158f42899636\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:05Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:32Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:32 crc kubenswrapper[4870]: I1014 07:01:32.596304 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:32 crc kubenswrapper[4870]: I1014 07:01:32.596341 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:32 crc kubenswrapper[4870]: I1014 07:01:32.596354 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:32 crc kubenswrapper[4870]: I1014 07:01:32.596372 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:32 crc kubenswrapper[4870]: I1014 07:01:32.596383 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:32Z","lastTransitionTime":"2025-10-14T07:01:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:32 crc kubenswrapper[4870]: I1014 07:01:32.623816 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:32Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:32 crc kubenswrapper[4870]: I1014 07:01:32.666993 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e26fe49ec2045d20a0cb2dd17e96e79b6eb43c641eb184c688cc472ca5db39d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aea115f7ee71c700b218fe15c928af7b55e6babe70e65d67f125d87be1fd2e2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:32Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:32 crc kubenswrapper[4870]: I1014 07:01:32.699556 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:32 crc kubenswrapper[4870]: I1014 07:01:32.699612 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:32 crc kubenswrapper[4870]: I1014 07:01:32.699626 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:32 crc kubenswrapper[4870]: I1014 07:01:32.699649 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:32 crc kubenswrapper[4870]: I1014 07:01:32.699665 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:32Z","lastTransitionTime":"2025-10-14T07:01:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:32 crc kubenswrapper[4870]: I1014 07:01:32.764554 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-v7br2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dcaef0e-42e3-424b-bbe5-4a023efe912d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ba24c2e067b9f3a14d9d018b66300b54d866530d30174fbdb3c899299575d28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ba24c2e067b9f3a14d9d018b66300b54d866530d30174fbdb3c899299575d28\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-v7br2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:32Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:32 crc kubenswrapper[4870]: I1014 07:01:32.781589 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fdxjh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a3a430a-b540-4a8c-adad-b2893dbf9898\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93e41d5e927eed7ec62eb9c3607cc804cd8999556663363fd123b2a784b00ff1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k787j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fdxjh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:32Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:32 crc kubenswrapper[4870]: I1014 07:01:32.803314 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:32 crc kubenswrapper[4870]: I1014 07:01:32.803362 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:32 crc kubenswrapper[4870]: I1014 07:01:32.803386 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:32 crc kubenswrapper[4870]: I1014 07:01:32.803402 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:32 crc kubenswrapper[4870]: I1014 07:01:32.803415 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:32Z","lastTransitionTime":"2025-10-14T07:01:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:32 crc kubenswrapper[4870]: I1014 07:01:32.807068 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33a9493e-9660-48db-b8d1-310e29ebe313\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3570643ef1b7d5cc3a542f320413b819338e0c8b7ab7f0f09cc909997a39bab8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb398668197dedc1c2a14b4ae7d6a29714f3527309ccb9c1622af841100e1d60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1d2b6d068d63bd38cdfb83824525f6d12054f6cabab5d684e03ea5f2152076c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://783ff243e395333f3fc80cfae21358c532f0c9c5cb5515586be56b90ed10d912\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a8135a3d01076e4437ae8a784b459bf8d16de57fa16ec7ce4d40c5129c9d4a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48ca3a5017ba650df99b2fc7201c3fe5abfa89d81089c5647cde1b1549f534fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48ca3a5017ba650df99b2fc7201c3fe5abfa89d81089c5647cde1b1549f534fc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://19682eca67f76bf2e06e1b80f12e21141fcbce0acb8195f0defacd1b9a10b5e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://19682eca67f76bf2e06e1b80f12e21141fcbce0acb8195f0defacd1b9a10b5e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://042aad6d5f6bee8eef77f9e6a2c265de775a69f3da72aedc68bdf312198f00fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://042aad6d5f6bee8eef77f9e6a2c265de775a69f3da72aedc68bdf312198f00fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:05Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:32Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:32 crc kubenswrapper[4870]: I1014 07:01:32.828871 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-4v2v2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4796ba6-c23b-4764-be96-2e3a3bf5e962\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf5afb0b99a7101f6d5573fb7eb925c5a1d8becc11b2e3c6f3351d80b6f0e823\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnp8b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-4v2v2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:32Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:32 crc kubenswrapper[4870]: I1014 07:01:32.868639 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-cdjmj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"612e7ab9-bda4-45d8-bc50-ff1afb68aace\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://834444b5350ee12df4e4652c3009880d8d69e1edf8ba0c0156ab96fb2bbe28e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://834444b5350ee12df4e4652c3009880d8d69e1edf8ba0c0156ab96fb2bbe28e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c1972e7b2f75790db22a07a078eee97072bffbdb7f8e594bf2be5f800632a7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c1972e7b2f75790db22a07a078eee97072bffbdb7f8e594bf2be5f800632a7b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b124b42a44c6a23da1d92b9a5326987eb74430f80489a71e8913829475654ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b124b42a44c6a23da1d92b9a5326987eb74430f80489a71e8913829475654ab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62d0deb6b20ea65fbcbaf94c2fc14d69adb22e7a46cc043ccd4def2668ddd4b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62d0deb6b20ea65fbcbaf94c2fc14d69adb22e7a46cc043ccd4def2668ddd4b7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-cdjmj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:32Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:32 crc kubenswrapper[4870]: I1014 07:01:32.901901 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bcbc6dd6-aab0-4dbc-a62e-c58f06ec1337\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25ec4808b09adc44488530b3c4e29fd7fec46b3fcfa31d4449ec4a6dbc880ec8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f3f5de86fb11231a19034a18e9d7f31476c162e6567d0c136ead8bbbd900d6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d57d938d9d9c92198ae5b3857536ece36012fbfbe8cc8cfd245d1386d39b0be5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77a3247a7f65709543d7fd374b2ab69d7d793537d4132d7e22261188d6e58544\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:05Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:32Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:32 crc kubenswrapper[4870]: I1014 07:01:32.905785 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:32 crc kubenswrapper[4870]: I1014 07:01:32.906024 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:32 crc kubenswrapper[4870]: I1014 07:01:32.906034 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:32 crc kubenswrapper[4870]: I1014 07:01:32.906049 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:32 crc kubenswrapper[4870]: I1014 07:01:32.906063 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:32Z","lastTransitionTime":"2025-10-14T07:01:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:32 crc kubenswrapper[4870]: I1014 07:01:32.943940 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e539fac3d0a2ade314a4b4c1050bb27c37bcff5ba5272fa2d71b97ab5c0d7191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:32Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:32 crc kubenswrapper[4870]: I1014 07:01:32.983291 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:32Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:33 crc kubenswrapper[4870]: I1014 07:01:33.008846 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:33 crc kubenswrapper[4870]: I1014 07:01:33.008887 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:33 crc kubenswrapper[4870]: I1014 07:01:33.008898 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:33 crc kubenswrapper[4870]: I1014 07:01:33.008916 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:33 crc kubenswrapper[4870]: I1014 07:01:33.008926 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:33Z","lastTransitionTime":"2025-10-14T07:01:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:33 crc kubenswrapper[4870]: I1014 07:01:33.020673 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-kz77f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d3864a91-9bf6-4540-a908-cc41aa39656c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d4ded3d86030f5ea548c8b1cc3caddbe14a45c9a3ae4af836e892880584f57d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4h8v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:28Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-kz77f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:33Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:33 crc kubenswrapper[4870]: I1014 07:01:33.113158 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:33 crc kubenswrapper[4870]: I1014 07:01:33.113224 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:33 crc kubenswrapper[4870]: I1014 07:01:33.113236 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:33 crc kubenswrapper[4870]: I1014 07:01:33.113261 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:33 crc kubenswrapper[4870]: I1014 07:01:33.113273 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:33Z","lastTransitionTime":"2025-10-14T07:01:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:33 crc kubenswrapper[4870]: I1014 07:01:33.217153 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:33 crc kubenswrapper[4870]: I1014 07:01:33.217228 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:33 crc kubenswrapper[4870]: I1014 07:01:33.217245 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:33 crc kubenswrapper[4870]: I1014 07:01:33.217277 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:33 crc kubenswrapper[4870]: I1014 07:01:33.217297 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:33Z","lastTransitionTime":"2025-10-14T07:01:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:33 crc kubenswrapper[4870]: I1014 07:01:33.320539 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:33 crc kubenswrapper[4870]: I1014 07:01:33.320595 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:33 crc kubenswrapper[4870]: I1014 07:01:33.320612 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:33 crc kubenswrapper[4870]: I1014 07:01:33.320638 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:33 crc kubenswrapper[4870]: I1014 07:01:33.320656 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:33Z","lastTransitionTime":"2025-10-14T07:01:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:33 crc kubenswrapper[4870]: I1014 07:01:33.388194 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-cdjmj" event={"ID":"612e7ab9-bda4-45d8-bc50-ff1afb68aace","Type":"ContainerStarted","Data":"f7f0d7cf29ca737d0effe58731a8b890379527c97f78c144d4eebb910260fadb"} Oct 14 07:01:33 crc kubenswrapper[4870]: I1014 07:01:33.412173 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-v7br2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dcaef0e-42e3-424b-bbe5-4a023efe912d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ba24c2e067b9f3a14d9d018b66300b54d866530d30174fbdb3c899299575d28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ba24c2e067b9f3a14d9d018b66300b54d866530d30174fbdb3c899299575d28\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-v7br2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:33Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:33 crc kubenswrapper[4870]: I1014 07:01:33.423873 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:33 crc kubenswrapper[4870]: I1014 07:01:33.423938 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:33 crc kubenswrapper[4870]: I1014 07:01:33.423953 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:33 crc kubenswrapper[4870]: I1014 07:01:33.423981 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:33 crc kubenswrapper[4870]: I1014 07:01:33.423998 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:33Z","lastTransitionTime":"2025-10-14T07:01:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:33 crc kubenswrapper[4870]: I1014 07:01:33.429939 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77a0d93a-5391-4112-b502-c474f45de209\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7815af9a584f6db6660073da289f2e8c987a04ef9ddfac155bdd9a15b014fcfa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b329b570622155a55f70501b2260e278106a68ebef497914231b4ec860e0125\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba942a0e6fe5ab93a187e0b7b4bd7c440c2e7827d9b5323fc50f90b562075392\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://978e8db184ba9e430232c6bb2ee530beb92567fdc71ba3c35b0bedca3492d845\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://978e8db184ba9e430232c6bb2ee530beb92567fdc71ba3c35b0bedca3492d845\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"file observer\\\\nW1014 07:01:25.870672 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1014 07:01:25.870795 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1014 07:01:25.871521 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-756714914/tls.crt::/tmp/serving-cert-756714914/tls.key\\\\\\\"\\\\nI1014 07:01:26.460037 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1014 07:01:26.464876 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1014 07:01:26.464902 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1014 07:01:26.464925 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1014 07:01:26.464931 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1014 07:01:26.481374 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1014 07:01:26.481400 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:01:26.481405 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:01:26.481410 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1014 07:01:26.481414 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1014 07:01:26.481417 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1014 07:01:26.481420 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1014 07:01:26.481610 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1014 07:01:26.482872 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bd0044a247bdc5ca4cab7294103af3270b76cde61f3109b85e83ed050ca53b6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82d3c2214e0be679d2827594b159ace362151318689078bc49d3158f42899636\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82d3c2214e0be679d2827594b159ace362151318689078bc49d3158f42899636\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:05Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:33Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:33 crc kubenswrapper[4870]: I1014 07:01:33.449493 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:33Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:33 crc kubenswrapper[4870]: I1014 07:01:33.466653 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e26fe49ec2045d20a0cb2dd17e96e79b6eb43c641eb184c688cc472ca5db39d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aea115f7ee71c700b218fe15c928af7b55e6babe70e65d67f125d87be1fd2e2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:33Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:33 crc kubenswrapper[4870]: I1014 07:01:33.479582 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-4v2v2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4796ba6-c23b-4764-be96-2e3a3bf5e962\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf5afb0b99a7101f6d5573fb7eb925c5a1d8becc11b2e3c6f3351d80b6f0e823\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnp8b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-4v2v2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:33Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:33 crc kubenswrapper[4870]: I1014 07:01:33.498969 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-cdjmj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"612e7ab9-bda4-45d8-bc50-ff1afb68aace\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://834444b5350ee12df4e4652c3009880d8d69e1edf8ba0c0156ab96fb2bbe28e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://834444b5350ee12df4e4652c3009880d8d69e1edf8ba0c0156ab96fb2bbe28e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c1972e7b2f75790db22a07a078eee97072bffbdb7f8e594bf2be5f800632a7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c1972e7b2f75790db22a07a078eee97072bffbdb7f8e594bf2be5f800632a7b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b124b42a44c6a23da1d92b9a5326987eb74430f80489a71e8913829475654ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b124b42a44c6a23da1d92b9a5326987eb74430f80489a71e8913829475654ab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62d0deb6b20ea65fbcbaf94c2fc14d69adb22e7a46cc043ccd4def2668ddd4b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62d0deb6b20ea65fbcbaf94c2fc14d69adb22e7a46cc043ccd4def2668ddd4b7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7f0d7cf29ca737d0effe58731a8b890379527c97f78c144d4eebb910260fadb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-cdjmj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:33Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:33 crc kubenswrapper[4870]: I1014 07:01:33.518121 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fdxjh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a3a430a-b540-4a8c-adad-b2893dbf9898\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93e41d5e927eed7ec62eb9c3607cc804cd8999556663363fd123b2a784b00ff1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k787j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fdxjh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:33Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:33 crc kubenswrapper[4870]: I1014 07:01:33.526688 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:33 crc kubenswrapper[4870]: I1014 07:01:33.526725 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:33 crc kubenswrapper[4870]: I1014 07:01:33.526734 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:33 crc kubenswrapper[4870]: I1014 07:01:33.526752 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:33 crc kubenswrapper[4870]: I1014 07:01:33.526762 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:33Z","lastTransitionTime":"2025-10-14T07:01:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:33 crc kubenswrapper[4870]: I1014 07:01:33.544482 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33a9493e-9660-48db-b8d1-310e29ebe313\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3570643ef1b7d5cc3a542f320413b819338e0c8b7ab7f0f09cc909997a39bab8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb398668197dedc1c2a14b4ae7d6a29714f3527309ccb9c1622af841100e1d60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1d2b6d068d63bd38cdfb83824525f6d12054f6cabab5d684e03ea5f2152076c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://783ff243e395333f3fc80cfae21358c532f0c9c5cb5515586be56b90ed10d912\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a8135a3d01076e4437ae8a784b459bf8d16de57fa16ec7ce4d40c5129c9d4a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48ca3a5017ba650df99b2fc7201c3fe5abfa89d81089c5647cde1b1549f534fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48ca3a5017ba650df99b2fc7201c3fe5abfa89d81089c5647cde1b1549f534fc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://19682eca67f76bf2e06e1b80f12e21141fcbce0acb8195f0defacd1b9a10b5e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://19682eca67f76bf2e06e1b80f12e21141fcbce0acb8195f0defacd1b9a10b5e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://042aad6d5f6bee8eef77f9e6a2c265de775a69f3da72aedc68bdf312198f00fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://042aad6d5f6bee8eef77f9e6a2c265de775a69f3da72aedc68bdf312198f00fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:05Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:33Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:33 crc kubenswrapper[4870]: I1014 07:01:33.557878 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-kz77f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d3864a91-9bf6-4540-a908-cc41aa39656c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d4ded3d86030f5ea548c8b1cc3caddbe14a45c9a3ae4af836e892880584f57d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4h8v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:28Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-kz77f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:33Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:33 crc kubenswrapper[4870]: I1014 07:01:33.575728 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bcbc6dd6-aab0-4dbc-a62e-c58f06ec1337\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25ec4808b09adc44488530b3c4e29fd7fec46b3fcfa31d4449ec4a6dbc880ec8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f3f5de86fb11231a19034a18e9d7f31476c162e6567d0c136ead8bbbd900d6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d57d938d9d9c92198ae5b3857536ece36012fbfbe8cc8cfd245d1386d39b0be5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77a3247a7f65709543d7fd374b2ab69d7d793537d4132d7e22261188d6e58544\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:05Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:33Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:33 crc kubenswrapper[4870]: I1014 07:01:33.596867 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e539fac3d0a2ade314a4b4c1050bb27c37bcff5ba5272fa2d71b97ab5c0d7191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:33Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:33 crc kubenswrapper[4870]: I1014 07:01:33.616591 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:33Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:33 crc kubenswrapper[4870]: I1014 07:01:33.630050 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:33 crc kubenswrapper[4870]: I1014 07:01:33.630097 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:33 crc kubenswrapper[4870]: I1014 07:01:33.630127 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:33 crc kubenswrapper[4870]: I1014 07:01:33.630146 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:33 crc kubenswrapper[4870]: I1014 07:01:33.630159 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:33Z","lastTransitionTime":"2025-10-14T07:01:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:33 crc kubenswrapper[4870]: I1014 07:01:33.639220 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:33Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:33 crc kubenswrapper[4870]: I1014 07:01:33.655614 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3eb0b397c05ffd6650607f1be4a75683e0342d3333fc096bf1cd63c7688834f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:33Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:33 crc kubenswrapper[4870]: I1014 07:01:33.668420 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"874a5e45-dffd-4d17-b609-b7d3ed2eab07\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d98d044836ab44a405196915ef7bf572b75113f2c81ec7e64effa63b47ad3543\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wj4ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8004752575df4a1c44efbcafe64799f09126d0f97285d92388c76a0b90e022f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wj4ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7tvc7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:33Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:33 crc kubenswrapper[4870]: I1014 07:01:33.697887 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 07:01:33 crc kubenswrapper[4870]: E1014 07:01:33.698063 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 07:01:41.69803456 +0000 UTC m=+37.395394951 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:01:33 crc kubenswrapper[4870]: I1014 07:01:33.734870 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:33 crc kubenswrapper[4870]: I1014 07:01:33.735200 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:33 crc kubenswrapper[4870]: I1014 07:01:33.735215 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:33 crc kubenswrapper[4870]: I1014 07:01:33.735238 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:33 crc kubenswrapper[4870]: I1014 07:01:33.735254 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:33Z","lastTransitionTime":"2025-10-14T07:01:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:33 crc kubenswrapper[4870]: I1014 07:01:33.800147 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 07:01:33 crc kubenswrapper[4870]: I1014 07:01:33.800217 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 07:01:33 crc kubenswrapper[4870]: I1014 07:01:33.800272 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 07:01:33 crc kubenswrapper[4870]: I1014 07:01:33.800315 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 07:01:33 crc kubenswrapper[4870]: E1014 07:01:33.800474 4870 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 14 07:01:33 crc kubenswrapper[4870]: E1014 07:01:33.800531 4870 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 14 07:01:33 crc kubenswrapper[4870]: E1014 07:01:33.800477 4870 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 14 07:01:33 crc kubenswrapper[4870]: E1014 07:01:33.800549 4870 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 14 07:01:33 crc kubenswrapper[4870]: E1014 07:01:33.800540 4870 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 14 07:01:33 crc kubenswrapper[4870]: E1014 07:01:33.800573 4870 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 14 07:01:33 crc kubenswrapper[4870]: E1014 07:01:33.800612 4870 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 14 07:01:33 crc kubenswrapper[4870]: E1014 07:01:33.800629 4870 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 14 07:01:33 crc kubenswrapper[4870]: E1014 07:01:33.800639 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-14 07:01:41.800607717 +0000 UTC m=+37.497968108 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 14 07:01:33 crc kubenswrapper[4870]: E1014 07:01:33.800666 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-14 07:01:41.800655778 +0000 UTC m=+37.498016169 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 14 07:01:33 crc kubenswrapper[4870]: E1014 07:01:33.800688 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-14 07:01:41.800677689 +0000 UTC m=+37.498038080 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 14 07:01:33 crc kubenswrapper[4870]: E1014 07:01:33.800709 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-14 07:01:41.800699529 +0000 UTC m=+37.498059910 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 14 07:01:33 crc kubenswrapper[4870]: I1014 07:01:33.838005 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:33 crc kubenswrapper[4870]: I1014 07:01:33.838055 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:33 crc kubenswrapper[4870]: I1014 07:01:33.838071 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:33 crc kubenswrapper[4870]: I1014 07:01:33.838094 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:33 crc kubenswrapper[4870]: I1014 07:01:33.838185 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:33Z","lastTransitionTime":"2025-10-14T07:01:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:33 crc kubenswrapper[4870]: I1014 07:01:33.944910 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:33 crc kubenswrapper[4870]: I1014 07:01:33.945107 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:33 crc kubenswrapper[4870]: I1014 07:01:33.945132 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:33 crc kubenswrapper[4870]: I1014 07:01:33.945165 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:33 crc kubenswrapper[4870]: I1014 07:01:33.945193 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:33Z","lastTransitionTime":"2025-10-14T07:01:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:34 crc kubenswrapper[4870]: I1014 07:01:34.033737 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 07:01:34 crc kubenswrapper[4870]: I1014 07:01:34.033801 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 07:01:34 crc kubenswrapper[4870]: I1014 07:01:34.033864 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 07:01:34 crc kubenswrapper[4870]: E1014 07:01:34.033974 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 07:01:34 crc kubenswrapper[4870]: E1014 07:01:34.034085 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 07:01:34 crc kubenswrapper[4870]: E1014 07:01:34.034210 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 07:01:34 crc kubenswrapper[4870]: I1014 07:01:34.049720 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:34 crc kubenswrapper[4870]: I1014 07:01:34.049810 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:34 crc kubenswrapper[4870]: I1014 07:01:34.049834 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:34 crc kubenswrapper[4870]: I1014 07:01:34.050067 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:34 crc kubenswrapper[4870]: I1014 07:01:34.050092 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:34Z","lastTransitionTime":"2025-10-14T07:01:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:34 crc kubenswrapper[4870]: I1014 07:01:34.153150 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:34 crc kubenswrapper[4870]: I1014 07:01:34.153200 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:34 crc kubenswrapper[4870]: I1014 07:01:34.153212 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:34 crc kubenswrapper[4870]: I1014 07:01:34.153231 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:34 crc kubenswrapper[4870]: I1014 07:01:34.153244 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:34Z","lastTransitionTime":"2025-10-14T07:01:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:34 crc kubenswrapper[4870]: I1014 07:01:34.255221 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:34 crc kubenswrapper[4870]: I1014 07:01:34.255267 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:34 crc kubenswrapper[4870]: I1014 07:01:34.255279 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:34 crc kubenswrapper[4870]: I1014 07:01:34.255296 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:34 crc kubenswrapper[4870]: I1014 07:01:34.255308 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:34Z","lastTransitionTime":"2025-10-14T07:01:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:34 crc kubenswrapper[4870]: I1014 07:01:34.358104 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:34 crc kubenswrapper[4870]: I1014 07:01:34.358145 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:34 crc kubenswrapper[4870]: I1014 07:01:34.358156 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:34 crc kubenswrapper[4870]: I1014 07:01:34.358174 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:34 crc kubenswrapper[4870]: I1014 07:01:34.358186 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:34Z","lastTransitionTime":"2025-10-14T07:01:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:34 crc kubenswrapper[4870]: I1014 07:01:34.394184 4870 generic.go:334] "Generic (PLEG): container finished" podID="612e7ab9-bda4-45d8-bc50-ff1afb68aace" containerID="f7f0d7cf29ca737d0effe58731a8b890379527c97f78c144d4eebb910260fadb" exitCode=0 Oct 14 07:01:34 crc kubenswrapper[4870]: I1014 07:01:34.394263 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-cdjmj" event={"ID":"612e7ab9-bda4-45d8-bc50-ff1afb68aace","Type":"ContainerDied","Data":"f7f0d7cf29ca737d0effe58731a8b890379527c97f78c144d4eebb910260fadb"} Oct 14 07:01:34 crc kubenswrapper[4870]: I1014 07:01:34.400788 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-v7br2" event={"ID":"8dcaef0e-42e3-424b-bbe5-4a023efe912d","Type":"ContainerStarted","Data":"7c43b07bd7ecc0a98ea65b6f59752ad8651568236dc78ee216e0ca42dabcbe4f"} Oct 14 07:01:34 crc kubenswrapper[4870]: I1014 07:01:34.401362 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-v7br2" Oct 14 07:01:34 crc kubenswrapper[4870]: I1014 07:01:34.417256 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:34Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:34 crc kubenswrapper[4870]: I1014 07:01:34.434532 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3eb0b397c05ffd6650607f1be4a75683e0342d3333fc096bf1cd63c7688834f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:34Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:34 crc kubenswrapper[4870]: I1014 07:01:34.451164 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"874a5e45-dffd-4d17-b609-b7d3ed2eab07\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d98d044836ab44a405196915ef7bf572b75113f2c81ec7e64effa63b47ad3543\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wj4ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8004752575df4a1c44efbcafe64799f09126d0f97285d92388c76a0b90e022f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wj4ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7tvc7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:34Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:34 crc kubenswrapper[4870]: I1014 07:01:34.467267 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:34 crc kubenswrapper[4870]: I1014 07:01:34.467333 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:34 crc kubenswrapper[4870]: I1014 07:01:34.467350 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:34 crc kubenswrapper[4870]: I1014 07:01:34.467376 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:34 crc kubenswrapper[4870]: I1014 07:01:34.467396 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:34Z","lastTransitionTime":"2025-10-14T07:01:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:34 crc kubenswrapper[4870]: I1014 07:01:34.469525 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-v7br2" Oct 14 07:01:34 crc kubenswrapper[4870]: I1014 07:01:34.472373 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77a0d93a-5391-4112-b502-c474f45de209\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7815af9a584f6db6660073da289f2e8c987a04ef9ddfac155bdd9a15b014fcfa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b329b570622155a55f70501b2260e278106a68ebef497914231b4ec860e0125\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba942a0e6fe5ab93a187e0b7b4bd7c440c2e7827d9b5323fc50f90b562075392\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://978e8db184ba9e430232c6bb2ee530beb92567fdc71ba3c35b0bedca3492d845\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://978e8db184ba9e430232c6bb2ee530beb92567fdc71ba3c35b0bedca3492d845\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"file observer\\\\nW1014 07:01:25.870672 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1014 07:01:25.870795 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1014 07:01:25.871521 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-756714914/tls.crt::/tmp/serving-cert-756714914/tls.key\\\\\\\"\\\\nI1014 07:01:26.460037 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1014 07:01:26.464876 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1014 07:01:26.464902 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1014 07:01:26.464925 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1014 07:01:26.464931 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1014 07:01:26.481374 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1014 07:01:26.481400 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:01:26.481405 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:01:26.481410 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1014 07:01:26.481414 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1014 07:01:26.481417 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1014 07:01:26.481420 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1014 07:01:26.481610 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1014 07:01:26.482872 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bd0044a247bdc5ca4cab7294103af3270b76cde61f3109b85e83ed050ca53b6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82d3c2214e0be679d2827594b159ace362151318689078bc49d3158f42899636\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82d3c2214e0be679d2827594b159ace362151318689078bc49d3158f42899636\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:05Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:34Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:34 crc kubenswrapper[4870]: I1014 07:01:34.489382 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:34Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:34 crc kubenswrapper[4870]: I1014 07:01:34.506553 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e26fe49ec2045d20a0cb2dd17e96e79b6eb43c641eb184c688cc472ca5db39d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aea115f7ee71c700b218fe15c928af7b55e6babe70e65d67f125d87be1fd2e2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:34Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:34 crc kubenswrapper[4870]: I1014 07:01:34.530582 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-v7br2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dcaef0e-42e3-424b-bbe5-4a023efe912d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ba24c2e067b9f3a14d9d018b66300b54d866530d30174fbdb3c899299575d28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ba24c2e067b9f3a14d9d018b66300b54d866530d30174fbdb3c899299575d28\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-v7br2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:34Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:34 crc kubenswrapper[4870]: I1014 07:01:34.547535 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fdxjh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a3a430a-b540-4a8c-adad-b2893dbf9898\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93e41d5e927eed7ec62eb9c3607cc804cd8999556663363fd123b2a784b00ff1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k787j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fdxjh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:34Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:34 crc kubenswrapper[4870]: I1014 07:01:34.571007 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:34 crc kubenswrapper[4870]: I1014 07:01:34.571065 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:34 crc kubenswrapper[4870]: I1014 07:01:34.571080 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:34 crc kubenswrapper[4870]: I1014 07:01:34.571104 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:34 crc kubenswrapper[4870]: I1014 07:01:34.571117 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:34Z","lastTransitionTime":"2025-10-14T07:01:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:34 crc kubenswrapper[4870]: I1014 07:01:34.575059 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33a9493e-9660-48db-b8d1-310e29ebe313\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3570643ef1b7d5cc3a542f320413b819338e0c8b7ab7f0f09cc909997a39bab8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb398668197dedc1c2a14b4ae7d6a29714f3527309ccb9c1622af841100e1d60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1d2b6d068d63bd38cdfb83824525f6d12054f6cabab5d684e03ea5f2152076c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://783ff243e395333f3fc80cfae21358c532f0c9c5cb5515586be56b90ed10d912\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a8135a3d01076e4437ae8a784b459bf8d16de57fa16ec7ce4d40c5129c9d4a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48ca3a5017ba650df99b2fc7201c3fe5abfa89d81089c5647cde1b1549f534fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48ca3a5017ba650df99b2fc7201c3fe5abfa89d81089c5647cde1b1549f534fc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://19682eca67f76bf2e06e1b80f12e21141fcbce0acb8195f0defacd1b9a10b5e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://19682eca67f76bf2e06e1b80f12e21141fcbce0acb8195f0defacd1b9a10b5e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://042aad6d5f6bee8eef77f9e6a2c265de775a69f3da72aedc68bdf312198f00fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://042aad6d5f6bee8eef77f9e6a2c265de775a69f3da72aedc68bdf312198f00fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:05Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:34Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:34 crc kubenswrapper[4870]: I1014 07:01:34.588228 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-4v2v2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4796ba6-c23b-4764-be96-2e3a3bf5e962\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf5afb0b99a7101f6d5573fb7eb925c5a1d8becc11b2e3c6f3351d80b6f0e823\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnp8b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-4v2v2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:34Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:34 crc kubenswrapper[4870]: I1014 07:01:34.604271 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-cdjmj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"612e7ab9-bda4-45d8-bc50-ff1afb68aace\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://834444b5350ee12df4e4652c3009880d8d69e1edf8ba0c0156ab96fb2bbe28e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://834444b5350ee12df4e4652c3009880d8d69e1edf8ba0c0156ab96fb2bbe28e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c1972e7b2f75790db22a07a078eee97072bffbdb7f8e594bf2be5f800632a7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c1972e7b2f75790db22a07a078eee97072bffbdb7f8e594bf2be5f800632a7b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b124b42a44c6a23da1d92b9a5326987eb74430f80489a71e8913829475654ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b124b42a44c6a23da1d92b9a5326987eb74430f80489a71e8913829475654ab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62d0deb6b20ea65fbcbaf94c2fc14d69adb22e7a46cc043ccd4def2668ddd4b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62d0deb6b20ea65fbcbaf94c2fc14d69adb22e7a46cc043ccd4def2668ddd4b7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7f0d7cf29ca737d0effe58731a8b890379527c97f78c144d4eebb910260fadb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7f0d7cf29ca737d0effe58731a8b890379527c97f78c144d4eebb910260fadb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-cdjmj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:34Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:34 crc kubenswrapper[4870]: I1014 07:01:34.620693 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bcbc6dd6-aab0-4dbc-a62e-c58f06ec1337\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25ec4808b09adc44488530b3c4e29fd7fec46b3fcfa31d4449ec4a6dbc880ec8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f3f5de86fb11231a19034a18e9d7f31476c162e6567d0c136ead8bbbd900d6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d57d938d9d9c92198ae5b3857536ece36012fbfbe8cc8cfd245d1386d39b0be5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77a3247a7f65709543d7fd374b2ab69d7d793537d4132d7e22261188d6e58544\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:05Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:34Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:34 crc kubenswrapper[4870]: I1014 07:01:34.637720 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e539fac3d0a2ade314a4b4c1050bb27c37bcff5ba5272fa2d71b97ab5c0d7191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:34Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:34 crc kubenswrapper[4870]: I1014 07:01:34.651941 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:34Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:34 crc kubenswrapper[4870]: I1014 07:01:34.662694 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-kz77f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d3864a91-9bf6-4540-a908-cc41aa39656c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d4ded3d86030f5ea548c8b1cc3caddbe14a45c9a3ae4af836e892880584f57d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4h8v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:28Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-kz77f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:34Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:34 crc kubenswrapper[4870]: I1014 07:01:34.674252 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:34 crc kubenswrapper[4870]: I1014 07:01:34.674287 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:34 crc kubenswrapper[4870]: I1014 07:01:34.674301 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:34 crc kubenswrapper[4870]: I1014 07:01:34.674323 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:34 crc kubenswrapper[4870]: I1014 07:01:34.674338 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:34Z","lastTransitionTime":"2025-10-14T07:01:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:34 crc kubenswrapper[4870]: I1014 07:01:34.677028 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:34Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:34 crc kubenswrapper[4870]: I1014 07:01:34.687631 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3eb0b397c05ffd6650607f1be4a75683e0342d3333fc096bf1cd63c7688834f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:34Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:34 crc kubenswrapper[4870]: I1014 07:01:34.700206 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"874a5e45-dffd-4d17-b609-b7d3ed2eab07\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d98d044836ab44a405196915ef7bf572b75113f2c81ec7e64effa63b47ad3543\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wj4ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8004752575df4a1c44efbcafe64799f09126d0f97285d92388c76a0b90e022f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wj4ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7tvc7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:34Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:34 crc kubenswrapper[4870]: I1014 07:01:34.715606 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77a0d93a-5391-4112-b502-c474f45de209\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7815af9a584f6db6660073da289f2e8c987a04ef9ddfac155bdd9a15b014fcfa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b329b570622155a55f70501b2260e278106a68ebef497914231b4ec860e0125\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba942a0e6fe5ab93a187e0b7b4bd7c440c2e7827d9b5323fc50f90b562075392\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://978e8db184ba9e430232c6bb2ee530beb92567fdc71ba3c35b0bedca3492d845\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://978e8db184ba9e430232c6bb2ee530beb92567fdc71ba3c35b0bedca3492d845\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"file observer\\\\nW1014 07:01:25.870672 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1014 07:01:25.870795 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1014 07:01:25.871521 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-756714914/tls.crt::/tmp/serving-cert-756714914/tls.key\\\\\\\"\\\\nI1014 07:01:26.460037 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1014 07:01:26.464876 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1014 07:01:26.464902 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1014 07:01:26.464925 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1014 07:01:26.464931 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1014 07:01:26.481374 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1014 07:01:26.481400 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:01:26.481405 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:01:26.481410 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1014 07:01:26.481414 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1014 07:01:26.481417 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1014 07:01:26.481420 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1014 07:01:26.481610 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1014 07:01:26.482872 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bd0044a247bdc5ca4cab7294103af3270b76cde61f3109b85e83ed050ca53b6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82d3c2214e0be679d2827594b159ace362151318689078bc49d3158f42899636\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82d3c2214e0be679d2827594b159ace362151318689078bc49d3158f42899636\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:05Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:34Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:34 crc kubenswrapper[4870]: I1014 07:01:34.731215 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:34Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:34 crc kubenswrapper[4870]: I1014 07:01:34.746315 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e26fe49ec2045d20a0cb2dd17e96e79b6eb43c641eb184c688cc472ca5db39d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aea115f7ee71c700b218fe15c928af7b55e6babe70e65d67f125d87be1fd2e2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:34Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:34 crc kubenswrapper[4870]: I1014 07:01:34.766926 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-v7br2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dcaef0e-42e3-424b-bbe5-4a023efe912d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71dec55e859b11a081dbb90d7d6cf6b4846cf591382303d8329003ff50c71f2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e72cff0723e2ce520340244b4e597d8016c58e1b071c4b087f2f6cd718741594\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc9d4db944f837289d1923176d60cfc4a566a282c2d2e3a703e220df884703c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d16ed1c65925549dcb69ab02c3013e25e3f123a83179f1af3a8e08bf91a59896\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed59243b59c39ce82bf9bc8f33c0a8b84854846afff6a89f5d37ffc97a096b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3b6d20cf459b1f6d3aff5b9290f9391909baf27ee40c49e2b0cd2d5cfe889ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c43b07bd7ecc0a98ea65b6f59752ad8651568236dc78ee216e0ca42dabcbe4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fe3468cdd612fc4378e1f6f55a2a9d00274f7ddb65cb57b2a368dfc065cfb4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ba24c2e067b9f3a14d9d018b66300b54d866530d30174fbdb3c899299575d28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ba24c2e067b9f3a14d9d018b66300b54d866530d30174fbdb3c899299575d28\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-v7br2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:34Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:34 crc kubenswrapper[4870]: I1014 07:01:34.776765 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:34 crc kubenswrapper[4870]: I1014 07:01:34.776806 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:34 crc kubenswrapper[4870]: I1014 07:01:34.776819 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:34 crc kubenswrapper[4870]: I1014 07:01:34.776836 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:34 crc kubenswrapper[4870]: I1014 07:01:34.776847 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:34Z","lastTransitionTime":"2025-10-14T07:01:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:34 crc kubenswrapper[4870]: I1014 07:01:34.783197 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fdxjh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a3a430a-b540-4a8c-adad-b2893dbf9898\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93e41d5e927eed7ec62eb9c3607cc804cd8999556663363fd123b2a784b00ff1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k787j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fdxjh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:34Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:34 crc kubenswrapper[4870]: I1014 07:01:34.806767 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33a9493e-9660-48db-b8d1-310e29ebe313\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3570643ef1b7d5cc3a542f320413b819338e0c8b7ab7f0f09cc909997a39bab8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb398668197dedc1c2a14b4ae7d6a29714f3527309ccb9c1622af841100e1d60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1d2b6d068d63bd38cdfb83824525f6d12054f6cabab5d684e03ea5f2152076c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://783ff243e395333f3fc80cfae21358c532f0c9c5cb5515586be56b90ed10d912\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a8135a3d01076e4437ae8a784b459bf8d16de57fa16ec7ce4d40c5129c9d4a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48ca3a5017ba650df99b2fc7201c3fe5abfa89d81089c5647cde1b1549f534fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48ca3a5017ba650df99b2fc7201c3fe5abfa89d81089c5647cde1b1549f534fc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://19682eca67f76bf2e06e1b80f12e21141fcbce0acb8195f0defacd1b9a10b5e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://19682eca67f76bf2e06e1b80f12e21141fcbce0acb8195f0defacd1b9a10b5e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://042aad6d5f6bee8eef77f9e6a2c265de775a69f3da72aedc68bdf312198f00fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://042aad6d5f6bee8eef77f9e6a2c265de775a69f3da72aedc68bdf312198f00fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:05Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:34Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:34 crc kubenswrapper[4870]: I1014 07:01:34.821463 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-4v2v2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4796ba6-c23b-4764-be96-2e3a3bf5e962\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf5afb0b99a7101f6d5573fb7eb925c5a1d8becc11b2e3c6f3351d80b6f0e823\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnp8b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-4v2v2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:34Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:34 crc kubenswrapper[4870]: I1014 07:01:34.839649 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-cdjmj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"612e7ab9-bda4-45d8-bc50-ff1afb68aace\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://834444b5350ee12df4e4652c3009880d8d69e1edf8ba0c0156ab96fb2bbe28e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://834444b5350ee12df4e4652c3009880d8d69e1edf8ba0c0156ab96fb2bbe28e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c1972e7b2f75790db22a07a078eee97072bffbdb7f8e594bf2be5f800632a7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c1972e7b2f75790db22a07a078eee97072bffbdb7f8e594bf2be5f800632a7b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b124b42a44c6a23da1d92b9a5326987eb74430f80489a71e8913829475654ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b124b42a44c6a23da1d92b9a5326987eb74430f80489a71e8913829475654ab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62d0deb6b20ea65fbcbaf94c2fc14d69adb22e7a46cc043ccd4def2668ddd4b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62d0deb6b20ea65fbcbaf94c2fc14d69adb22e7a46cc043ccd4def2668ddd4b7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7f0d7cf29ca737d0effe58731a8b890379527c97f78c144d4eebb910260fadb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7f0d7cf29ca737d0effe58731a8b890379527c97f78c144d4eebb910260fadb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-cdjmj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:34Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:34 crc kubenswrapper[4870]: I1014 07:01:34.858029 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bcbc6dd6-aab0-4dbc-a62e-c58f06ec1337\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25ec4808b09adc44488530b3c4e29fd7fec46b3fcfa31d4449ec4a6dbc880ec8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f3f5de86fb11231a19034a18e9d7f31476c162e6567d0c136ead8bbbd900d6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d57d938d9d9c92198ae5b3857536ece36012fbfbe8cc8cfd245d1386d39b0be5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77a3247a7f65709543d7fd374b2ab69d7d793537d4132d7e22261188d6e58544\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:05Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:34Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:34 crc kubenswrapper[4870]: I1014 07:01:34.876157 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e539fac3d0a2ade314a4b4c1050bb27c37bcff5ba5272fa2d71b97ab5c0d7191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:34Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:34 crc kubenswrapper[4870]: I1014 07:01:34.878921 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:34 crc kubenswrapper[4870]: I1014 07:01:34.878975 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:34 crc kubenswrapper[4870]: I1014 07:01:34.878992 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:34 crc kubenswrapper[4870]: I1014 07:01:34.879013 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:34 crc kubenswrapper[4870]: I1014 07:01:34.879026 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:34Z","lastTransitionTime":"2025-10-14T07:01:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:34 crc kubenswrapper[4870]: I1014 07:01:34.893218 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:34Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:34 crc kubenswrapper[4870]: I1014 07:01:34.909525 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-kz77f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d3864a91-9bf6-4540-a908-cc41aa39656c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d4ded3d86030f5ea548c8b1cc3caddbe14a45c9a3ae4af836e892880584f57d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4h8v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:28Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-kz77f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:34Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:34 crc kubenswrapper[4870]: I1014 07:01:34.981294 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:34 crc kubenswrapper[4870]: I1014 07:01:34.981330 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:34 crc kubenswrapper[4870]: I1014 07:01:34.981340 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:34 crc kubenswrapper[4870]: I1014 07:01:34.981355 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:34 crc kubenswrapper[4870]: I1014 07:01:34.981364 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:34Z","lastTransitionTime":"2025-10-14T07:01:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:35 crc kubenswrapper[4870]: I1014 07:01:35.048606 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:35Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:35 crc kubenswrapper[4870]: I1014 07:01:35.060279 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3eb0b397c05ffd6650607f1be4a75683e0342d3333fc096bf1cd63c7688834f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:35Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:35 crc kubenswrapper[4870]: I1014 07:01:35.073647 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"874a5e45-dffd-4d17-b609-b7d3ed2eab07\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d98d044836ab44a405196915ef7bf572b75113f2c81ec7e64effa63b47ad3543\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wj4ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8004752575df4a1c44efbcafe64799f09126d0f97285d92388c76a0b90e022f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wj4ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7tvc7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:35Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:35 crc kubenswrapper[4870]: I1014 07:01:35.084059 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:35 crc kubenswrapper[4870]: I1014 07:01:35.084111 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:35 crc kubenswrapper[4870]: I1014 07:01:35.084125 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:35 crc kubenswrapper[4870]: I1014 07:01:35.084145 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:35 crc kubenswrapper[4870]: I1014 07:01:35.084157 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:35Z","lastTransitionTime":"2025-10-14T07:01:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:35 crc kubenswrapper[4870]: I1014 07:01:35.096275 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-v7br2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dcaef0e-42e3-424b-bbe5-4a023efe912d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71dec55e859b11a081dbb90d7d6cf6b4846cf591382303d8329003ff50c71f2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e72cff0723e2ce520340244b4e597d8016c58e1b071c4b087f2f6cd718741594\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc9d4db944f837289d1923176d60cfc4a566a282c2d2e3a703e220df884703c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d16ed1c65925549dcb69ab02c3013e25e3f123a83179f1af3a8e08bf91a59896\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed59243b59c39ce82bf9bc8f33c0a8b84854846afff6a89f5d37ffc97a096b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3b6d20cf459b1f6d3aff5b9290f9391909baf27ee40c49e2b0cd2d5cfe889ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c43b07bd7ecc0a98ea65b6f59752ad8651568236dc78ee216e0ca42dabcbe4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fe3468cdd612fc4378e1f6f55a2a9d00274f7ddb65cb57b2a368dfc065cfb4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ba24c2e067b9f3a14d9d018b66300b54d866530d30174fbdb3c899299575d28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ba24c2e067b9f3a14d9d018b66300b54d866530d30174fbdb3c899299575d28\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-v7br2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:35Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:35 crc kubenswrapper[4870]: I1014 07:01:35.112505 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77a0d93a-5391-4112-b502-c474f45de209\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7815af9a584f6db6660073da289f2e8c987a04ef9ddfac155bdd9a15b014fcfa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b329b570622155a55f70501b2260e278106a68ebef497914231b4ec860e0125\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba942a0e6fe5ab93a187e0b7b4bd7c440c2e7827d9b5323fc50f90b562075392\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://978e8db184ba9e430232c6bb2ee530beb92567fdc71ba3c35b0bedca3492d845\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://978e8db184ba9e430232c6bb2ee530beb92567fdc71ba3c35b0bedca3492d845\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"file observer\\\\nW1014 07:01:25.870672 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1014 07:01:25.870795 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1014 07:01:25.871521 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-756714914/tls.crt::/tmp/serving-cert-756714914/tls.key\\\\\\\"\\\\nI1014 07:01:26.460037 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1014 07:01:26.464876 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1014 07:01:26.464902 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1014 07:01:26.464925 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1014 07:01:26.464931 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1014 07:01:26.481374 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1014 07:01:26.481400 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:01:26.481405 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:01:26.481410 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1014 07:01:26.481414 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1014 07:01:26.481417 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1014 07:01:26.481420 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1014 07:01:26.481610 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1014 07:01:26.482872 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bd0044a247bdc5ca4cab7294103af3270b76cde61f3109b85e83ed050ca53b6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82d3c2214e0be679d2827594b159ace362151318689078bc49d3158f42899636\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82d3c2214e0be679d2827594b159ace362151318689078bc49d3158f42899636\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:05Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:35Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:35 crc kubenswrapper[4870]: I1014 07:01:35.132472 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:35Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:35 crc kubenswrapper[4870]: I1014 07:01:35.147257 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e26fe49ec2045d20a0cb2dd17e96e79b6eb43c641eb184c688cc472ca5db39d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aea115f7ee71c700b218fe15c928af7b55e6babe70e65d67f125d87be1fd2e2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:35Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:35 crc kubenswrapper[4870]: I1014 07:01:35.159834 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-4v2v2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4796ba6-c23b-4764-be96-2e3a3bf5e962\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf5afb0b99a7101f6d5573fb7eb925c5a1d8becc11b2e3c6f3351d80b6f0e823\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnp8b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-4v2v2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:35Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:35 crc kubenswrapper[4870]: I1014 07:01:35.186584 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:35 crc kubenswrapper[4870]: I1014 07:01:35.186620 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:35 crc kubenswrapper[4870]: I1014 07:01:35.186629 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:35 crc kubenswrapper[4870]: I1014 07:01:35.186646 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:35 crc kubenswrapper[4870]: I1014 07:01:35.186657 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:35Z","lastTransitionTime":"2025-10-14T07:01:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:35 crc kubenswrapper[4870]: I1014 07:01:35.187100 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-cdjmj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"612e7ab9-bda4-45d8-bc50-ff1afb68aace\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://834444b5350ee12df4e4652c3009880d8d69e1edf8ba0c0156ab96fb2bbe28e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://834444b5350ee12df4e4652c3009880d8d69e1edf8ba0c0156ab96fb2bbe28e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c1972e7b2f75790db22a07a078eee97072bffbdb7f8e594bf2be5f800632a7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c1972e7b2f75790db22a07a078eee97072bffbdb7f8e594bf2be5f800632a7b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b124b42a44c6a23da1d92b9a5326987eb74430f80489a71e8913829475654ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b124b42a44c6a23da1d92b9a5326987eb74430f80489a71e8913829475654ab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62d0deb6b20ea65fbcbaf94c2fc14d69adb22e7a46cc043ccd4def2668ddd4b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62d0deb6b20ea65fbcbaf94c2fc14d69adb22e7a46cc043ccd4def2668ddd4b7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7f0d7cf29ca737d0effe58731a8b890379527c97f78c144d4eebb910260fadb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7f0d7cf29ca737d0effe58731a8b890379527c97f78c144d4eebb910260fadb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-cdjmj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:35Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:35 crc kubenswrapper[4870]: I1014 07:01:35.222501 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fdxjh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a3a430a-b540-4a8c-adad-b2893dbf9898\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93e41d5e927eed7ec62eb9c3607cc804cd8999556663363fd123b2a784b00ff1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k787j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fdxjh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:35Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:35 crc kubenswrapper[4870]: I1014 07:01:35.267648 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33a9493e-9660-48db-b8d1-310e29ebe313\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3570643ef1b7d5cc3a542f320413b819338e0c8b7ab7f0f09cc909997a39bab8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb398668197dedc1c2a14b4ae7d6a29714f3527309ccb9c1622af841100e1d60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1d2b6d068d63bd38cdfb83824525f6d12054f6cabab5d684e03ea5f2152076c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://783ff243e395333f3fc80cfae21358c532f0c9c5cb5515586be56b90ed10d912\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a8135a3d01076e4437ae8a784b459bf8d16de57fa16ec7ce4d40c5129c9d4a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48ca3a5017ba650df99b2fc7201c3fe5abfa89d81089c5647cde1b1549f534fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48ca3a5017ba650df99b2fc7201c3fe5abfa89d81089c5647cde1b1549f534fc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://19682eca67f76bf2e06e1b80f12e21141fcbce0acb8195f0defacd1b9a10b5e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://19682eca67f76bf2e06e1b80f12e21141fcbce0acb8195f0defacd1b9a10b5e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://042aad6d5f6bee8eef77f9e6a2c265de775a69f3da72aedc68bdf312198f00fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://042aad6d5f6bee8eef77f9e6a2c265de775a69f3da72aedc68bdf312198f00fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:05Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:35Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:35 crc kubenswrapper[4870]: I1014 07:01:35.289795 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:35 crc kubenswrapper[4870]: I1014 07:01:35.289836 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:35 crc kubenswrapper[4870]: I1014 07:01:35.289844 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:35 crc kubenswrapper[4870]: I1014 07:01:35.289871 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:35 crc kubenswrapper[4870]: I1014 07:01:35.289880 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:35Z","lastTransitionTime":"2025-10-14T07:01:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:35 crc kubenswrapper[4870]: I1014 07:01:35.302773 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-kz77f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d3864a91-9bf6-4540-a908-cc41aa39656c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d4ded3d86030f5ea548c8b1cc3caddbe14a45c9a3ae4af836e892880584f57d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4h8v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:28Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-kz77f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:35Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:35 crc kubenswrapper[4870]: I1014 07:01:35.347077 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bcbc6dd6-aab0-4dbc-a62e-c58f06ec1337\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25ec4808b09adc44488530b3c4e29fd7fec46b3fcfa31d4449ec4a6dbc880ec8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f3f5de86fb11231a19034a18e9d7f31476c162e6567d0c136ead8bbbd900d6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d57d938d9d9c92198ae5b3857536ece36012fbfbe8cc8cfd245d1386d39b0be5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77a3247a7f65709543d7fd374b2ab69d7d793537d4132d7e22261188d6e58544\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:05Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:35Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:35 crc kubenswrapper[4870]: I1014 07:01:35.389084 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e539fac3d0a2ade314a4b4c1050bb27c37bcff5ba5272fa2d71b97ab5c0d7191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:35Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:35 crc kubenswrapper[4870]: I1014 07:01:35.392406 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:35 crc kubenswrapper[4870]: I1014 07:01:35.392477 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:35 crc kubenswrapper[4870]: I1014 07:01:35.392494 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:35 crc kubenswrapper[4870]: I1014 07:01:35.392529 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:35 crc kubenswrapper[4870]: I1014 07:01:35.392551 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:35Z","lastTransitionTime":"2025-10-14T07:01:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:35 crc kubenswrapper[4870]: I1014 07:01:35.410000 4870 generic.go:334] "Generic (PLEG): container finished" podID="612e7ab9-bda4-45d8-bc50-ff1afb68aace" containerID="b5959abb7c8e7ab32bed3aca709ae0334e8400f991ba724246c9fda56cbed822" exitCode=0 Oct 14 07:01:35 crc kubenswrapper[4870]: I1014 07:01:35.410224 4870 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 14 07:01:35 crc kubenswrapper[4870]: I1014 07:01:35.410264 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-cdjmj" event={"ID":"612e7ab9-bda4-45d8-bc50-ff1afb68aace","Type":"ContainerDied","Data":"b5959abb7c8e7ab32bed3aca709ae0334e8400f991ba724246c9fda56cbed822"} Oct 14 07:01:35 crc kubenswrapper[4870]: I1014 07:01:35.410688 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-v7br2" Oct 14 07:01:35 crc kubenswrapper[4870]: I1014 07:01:35.423808 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:35Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:35 crc kubenswrapper[4870]: I1014 07:01:35.447350 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-v7br2" Oct 14 07:01:35 crc kubenswrapper[4870]: I1014 07:01:35.462505 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77a0d93a-5391-4112-b502-c474f45de209\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7815af9a584f6db6660073da289f2e8c987a04ef9ddfac155bdd9a15b014fcfa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b329b570622155a55f70501b2260e278106a68ebef497914231b4ec860e0125\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba942a0e6fe5ab93a187e0b7b4bd7c440c2e7827d9b5323fc50f90b562075392\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://978e8db184ba9e430232c6bb2ee530beb92567fdc71ba3c35b0bedca3492d845\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://978e8db184ba9e430232c6bb2ee530beb92567fdc71ba3c35b0bedca3492d845\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"file observer\\\\nW1014 07:01:25.870672 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1014 07:01:25.870795 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1014 07:01:25.871521 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-756714914/tls.crt::/tmp/serving-cert-756714914/tls.key\\\\\\\"\\\\nI1014 07:01:26.460037 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1014 07:01:26.464876 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1014 07:01:26.464902 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1014 07:01:26.464925 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1014 07:01:26.464931 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1014 07:01:26.481374 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1014 07:01:26.481400 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:01:26.481405 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:01:26.481410 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1014 07:01:26.481414 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1014 07:01:26.481417 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1014 07:01:26.481420 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1014 07:01:26.481610 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1014 07:01:26.482872 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bd0044a247bdc5ca4cab7294103af3270b76cde61f3109b85e83ed050ca53b6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82d3c2214e0be679d2827594b159ace362151318689078bc49d3158f42899636\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82d3c2214e0be679d2827594b159ace362151318689078bc49d3158f42899636\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:05Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:35Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:35 crc kubenswrapper[4870]: I1014 07:01:35.499724 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:35 crc kubenswrapper[4870]: I1014 07:01:35.499787 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:35 crc kubenswrapper[4870]: I1014 07:01:35.499803 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:35 crc kubenswrapper[4870]: I1014 07:01:35.499824 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:35 crc kubenswrapper[4870]: I1014 07:01:35.499839 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:35Z","lastTransitionTime":"2025-10-14T07:01:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:35 crc kubenswrapper[4870]: I1014 07:01:35.504362 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:35Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:35 crc kubenswrapper[4870]: I1014 07:01:35.545697 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e26fe49ec2045d20a0cb2dd17e96e79b6eb43c641eb184c688cc472ca5db39d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aea115f7ee71c700b218fe15c928af7b55e6babe70e65d67f125d87be1fd2e2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:35Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:35 crc kubenswrapper[4870]: I1014 07:01:35.587594 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-v7br2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dcaef0e-42e3-424b-bbe5-4a023efe912d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71dec55e859b11a081dbb90d7d6cf6b4846cf591382303d8329003ff50c71f2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e72cff0723e2ce520340244b4e597d8016c58e1b071c4b087f2f6cd718741594\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc9d4db944f837289d1923176d60cfc4a566a282c2d2e3a703e220df884703c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d16ed1c65925549dcb69ab02c3013e25e3f123a83179f1af3a8e08bf91a59896\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed59243b59c39ce82bf9bc8f33c0a8b84854846afff6a89f5d37ffc97a096b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3b6d20cf459b1f6d3aff5b9290f9391909baf27ee40c49e2b0cd2d5cfe889ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c43b07bd7ecc0a98ea65b6f59752ad8651568236dc78ee216e0ca42dabcbe4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fe3468cdd612fc4378e1f6f55a2a9d00274f7ddb65cb57b2a368dfc065cfb4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ba24c2e067b9f3a14d9d018b66300b54d866530d30174fbdb3c899299575d28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ba24c2e067b9f3a14d9d018b66300b54d866530d30174fbdb3c899299575d28\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-v7br2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:35Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:35 crc kubenswrapper[4870]: I1014 07:01:35.602929 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:35 crc kubenswrapper[4870]: I1014 07:01:35.602968 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:35 crc kubenswrapper[4870]: I1014 07:01:35.602979 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:35 crc kubenswrapper[4870]: I1014 07:01:35.602999 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:35 crc kubenswrapper[4870]: I1014 07:01:35.603010 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:35Z","lastTransitionTime":"2025-10-14T07:01:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:35 crc kubenswrapper[4870]: I1014 07:01:35.624129 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-cdjmj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"612e7ab9-bda4-45d8-bc50-ff1afb68aace\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://834444b5350ee12df4e4652c3009880d8d69e1edf8ba0c0156ab96fb2bbe28e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://834444b5350ee12df4e4652c3009880d8d69e1edf8ba0c0156ab96fb2bbe28e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c1972e7b2f75790db22a07a078eee97072bffbdb7f8e594bf2be5f800632a7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c1972e7b2f75790db22a07a078eee97072bffbdb7f8e594bf2be5f800632a7b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b124b42a44c6a23da1d92b9a5326987eb74430f80489a71e8913829475654ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b124b42a44c6a23da1d92b9a5326987eb74430f80489a71e8913829475654ab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62d0deb6b20ea65fbcbaf94c2fc14d69adb22e7a46cc043ccd4def2668ddd4b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62d0deb6b20ea65fbcbaf94c2fc14d69adb22e7a46cc043ccd4def2668ddd4b7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7f0d7cf29ca737d0effe58731a8b890379527c97f78c144d4eebb910260fadb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7f0d7cf29ca737d0effe58731a8b890379527c97f78c144d4eebb910260fadb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5959abb7c8e7ab32bed3aca709ae0334e8400f991ba724246c9fda56cbed822\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5959abb7c8e7ab32bed3aca709ae0334e8400f991ba724246c9fda56cbed822\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-cdjmj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:35Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:35 crc kubenswrapper[4870]: I1014 07:01:35.669000 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fdxjh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a3a430a-b540-4a8c-adad-b2893dbf9898\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93e41d5e927eed7ec62eb9c3607cc804cd8999556663363fd123b2a784b00ff1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k787j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fdxjh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:35Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:35 crc kubenswrapper[4870]: I1014 07:01:35.707759 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:35 crc kubenswrapper[4870]: I1014 07:01:35.707813 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:35 crc kubenswrapper[4870]: I1014 07:01:35.707828 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:35 crc kubenswrapper[4870]: I1014 07:01:35.707851 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:35 crc kubenswrapper[4870]: I1014 07:01:35.707865 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:35Z","lastTransitionTime":"2025-10-14T07:01:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:35 crc kubenswrapper[4870]: I1014 07:01:35.709879 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33a9493e-9660-48db-b8d1-310e29ebe313\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3570643ef1b7d5cc3a542f320413b819338e0c8b7ab7f0f09cc909997a39bab8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb398668197dedc1c2a14b4ae7d6a29714f3527309ccb9c1622af841100e1d60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1d2b6d068d63bd38cdfb83824525f6d12054f6cabab5d684e03ea5f2152076c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://783ff243e395333f3fc80cfae21358c532f0c9c5cb5515586be56b90ed10d912\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a8135a3d01076e4437ae8a784b459bf8d16de57fa16ec7ce4d40c5129c9d4a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48ca3a5017ba650df99b2fc7201c3fe5abfa89d81089c5647cde1b1549f534fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48ca3a5017ba650df99b2fc7201c3fe5abfa89d81089c5647cde1b1549f534fc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://19682eca67f76bf2e06e1b80f12e21141fcbce0acb8195f0defacd1b9a10b5e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://19682eca67f76bf2e06e1b80f12e21141fcbce0acb8195f0defacd1b9a10b5e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://042aad6d5f6bee8eef77f9e6a2c265de775a69f3da72aedc68bdf312198f00fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://042aad6d5f6bee8eef77f9e6a2c265de775a69f3da72aedc68bdf312198f00fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:05Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:35Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:35 crc kubenswrapper[4870]: I1014 07:01:35.741727 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-4v2v2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4796ba6-c23b-4764-be96-2e3a3bf5e962\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf5afb0b99a7101f6d5573fb7eb925c5a1d8becc11b2e3c6f3351d80b6f0e823\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnp8b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-4v2v2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:35Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:35 crc kubenswrapper[4870]: I1014 07:01:35.787531 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bcbc6dd6-aab0-4dbc-a62e-c58f06ec1337\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25ec4808b09adc44488530b3c4e29fd7fec46b3fcfa31d4449ec4a6dbc880ec8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f3f5de86fb11231a19034a18e9d7f31476c162e6567d0c136ead8bbbd900d6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d57d938d9d9c92198ae5b3857536ece36012fbfbe8cc8cfd245d1386d39b0be5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77a3247a7f65709543d7fd374b2ab69d7d793537d4132d7e22261188d6e58544\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:05Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:35Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:35 crc kubenswrapper[4870]: I1014 07:01:35.811903 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:35 crc kubenswrapper[4870]: I1014 07:01:35.811959 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:35 crc kubenswrapper[4870]: I1014 07:01:35.811972 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:35 crc kubenswrapper[4870]: I1014 07:01:35.811994 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:35 crc kubenswrapper[4870]: I1014 07:01:35.812010 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:35Z","lastTransitionTime":"2025-10-14T07:01:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:35 crc kubenswrapper[4870]: I1014 07:01:35.823106 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e539fac3d0a2ade314a4b4c1050bb27c37bcff5ba5272fa2d71b97ab5c0d7191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:35Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:35 crc kubenswrapper[4870]: I1014 07:01:35.865005 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:35Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:35 crc kubenswrapper[4870]: I1014 07:01:35.902529 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-kz77f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d3864a91-9bf6-4540-a908-cc41aa39656c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d4ded3d86030f5ea548c8b1cc3caddbe14a45c9a3ae4af836e892880584f57d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4h8v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:28Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-kz77f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:35Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:35 crc kubenswrapper[4870]: I1014 07:01:35.915544 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:35 crc kubenswrapper[4870]: I1014 07:01:35.915590 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:35 crc kubenswrapper[4870]: I1014 07:01:35.915604 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:35 crc kubenswrapper[4870]: I1014 07:01:35.915624 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:35 crc kubenswrapper[4870]: I1014 07:01:35.915639 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:35Z","lastTransitionTime":"2025-10-14T07:01:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:35 crc kubenswrapper[4870]: I1014 07:01:35.949314 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:35Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:35 crc kubenswrapper[4870]: I1014 07:01:35.983625 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3eb0b397c05ffd6650607f1be4a75683e0342d3333fc096bf1cd63c7688834f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:35Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:36 crc kubenswrapper[4870]: I1014 07:01:36.018395 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:36 crc kubenswrapper[4870]: I1014 07:01:36.018428 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:36 crc kubenswrapper[4870]: I1014 07:01:36.018457 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:36 crc kubenswrapper[4870]: I1014 07:01:36.018474 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:36 crc kubenswrapper[4870]: I1014 07:01:36.018487 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:36Z","lastTransitionTime":"2025-10-14T07:01:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:36 crc kubenswrapper[4870]: I1014 07:01:36.021795 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"874a5e45-dffd-4d17-b609-b7d3ed2eab07\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d98d044836ab44a405196915ef7bf572b75113f2c81ec7e64effa63b47ad3543\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wj4ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8004752575df4a1c44efbcafe64799f09126d0f97285d92388c76a0b90e022f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wj4ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7tvc7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:36Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:36 crc kubenswrapper[4870]: I1014 07:01:36.033697 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 07:01:36 crc kubenswrapper[4870]: E1014 07:01:36.033804 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 07:01:36 crc kubenswrapper[4870]: I1014 07:01:36.033697 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 07:01:36 crc kubenswrapper[4870]: I1014 07:01:36.033713 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 07:01:36 crc kubenswrapper[4870]: E1014 07:01:36.033935 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 07:01:36 crc kubenswrapper[4870]: E1014 07:01:36.033968 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 07:01:36 crc kubenswrapper[4870]: I1014 07:01:36.060518 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:36Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:36 crc kubenswrapper[4870]: I1014 07:01:36.100641 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3eb0b397c05ffd6650607f1be4a75683e0342d3333fc096bf1cd63c7688834f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:36Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:36 crc kubenswrapper[4870]: I1014 07:01:36.120715 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:36 crc kubenswrapper[4870]: I1014 07:01:36.120782 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:36 crc kubenswrapper[4870]: I1014 07:01:36.120808 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:36 crc kubenswrapper[4870]: I1014 07:01:36.120838 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:36 crc kubenswrapper[4870]: I1014 07:01:36.120861 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:36Z","lastTransitionTime":"2025-10-14T07:01:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:36 crc kubenswrapper[4870]: I1014 07:01:36.139178 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"874a5e45-dffd-4d17-b609-b7d3ed2eab07\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d98d044836ab44a405196915ef7bf572b75113f2c81ec7e64effa63b47ad3543\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wj4ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8004752575df4a1c44efbcafe64799f09126d0f97285d92388c76a0b90e022f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wj4ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7tvc7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:36Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:36 crc kubenswrapper[4870]: I1014 07:01:36.181883 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77a0d93a-5391-4112-b502-c474f45de209\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7815af9a584f6db6660073da289f2e8c987a04ef9ddfac155bdd9a15b014fcfa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b329b570622155a55f70501b2260e278106a68ebef497914231b4ec860e0125\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba942a0e6fe5ab93a187e0b7b4bd7c440c2e7827d9b5323fc50f90b562075392\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://978e8db184ba9e430232c6bb2ee530beb92567fdc71ba3c35b0bedca3492d845\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://978e8db184ba9e430232c6bb2ee530beb92567fdc71ba3c35b0bedca3492d845\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"file observer\\\\nW1014 07:01:25.870672 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1014 07:01:25.870795 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1014 07:01:25.871521 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-756714914/tls.crt::/tmp/serving-cert-756714914/tls.key\\\\\\\"\\\\nI1014 07:01:26.460037 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1014 07:01:26.464876 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1014 07:01:26.464902 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1014 07:01:26.464925 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1014 07:01:26.464931 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1014 07:01:26.481374 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1014 07:01:26.481400 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:01:26.481405 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:01:26.481410 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1014 07:01:26.481414 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1014 07:01:26.481417 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1014 07:01:26.481420 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1014 07:01:26.481610 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1014 07:01:26.482872 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bd0044a247bdc5ca4cab7294103af3270b76cde61f3109b85e83ed050ca53b6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82d3c2214e0be679d2827594b159ace362151318689078bc49d3158f42899636\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82d3c2214e0be679d2827594b159ace362151318689078bc49d3158f42899636\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:05Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:36Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:36 crc kubenswrapper[4870]: I1014 07:01:36.220595 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:36Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:36 crc kubenswrapper[4870]: I1014 07:01:36.223053 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:36 crc kubenswrapper[4870]: I1014 07:01:36.223072 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:36 crc kubenswrapper[4870]: I1014 07:01:36.223080 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:36 crc kubenswrapper[4870]: I1014 07:01:36.223092 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:36 crc kubenswrapper[4870]: I1014 07:01:36.223103 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:36Z","lastTransitionTime":"2025-10-14T07:01:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:36 crc kubenswrapper[4870]: I1014 07:01:36.264118 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e26fe49ec2045d20a0cb2dd17e96e79b6eb43c641eb184c688cc472ca5db39d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aea115f7ee71c700b218fe15c928af7b55e6babe70e65d67f125d87be1fd2e2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:36Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:36 crc kubenswrapper[4870]: I1014 07:01:36.308635 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-v7br2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dcaef0e-42e3-424b-bbe5-4a023efe912d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71dec55e859b11a081dbb90d7d6cf6b4846cf591382303d8329003ff50c71f2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e72cff0723e2ce520340244b4e597d8016c58e1b071c4b087f2f6cd718741594\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc9d4db944f837289d1923176d60cfc4a566a282c2d2e3a703e220df884703c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d16ed1c65925549dcb69ab02c3013e25e3f123a83179f1af3a8e08bf91a59896\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed59243b59c39ce82bf9bc8f33c0a8b84854846afff6a89f5d37ffc97a096b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3b6d20cf459b1f6d3aff5b9290f9391909baf27ee40c49e2b0cd2d5cfe889ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c43b07bd7ecc0a98ea65b6f59752ad8651568236dc78ee216e0ca42dabcbe4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fe3468cdd612fc4378e1f6f55a2a9d00274f7ddb65cb57b2a368dfc065cfb4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ba24c2e067b9f3a14d9d018b66300b54d866530d30174fbdb3c899299575d28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ba24c2e067b9f3a14d9d018b66300b54d866530d30174fbdb3c899299575d28\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-v7br2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:36Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:36 crc kubenswrapper[4870]: I1014 07:01:36.325571 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:36 crc kubenswrapper[4870]: I1014 07:01:36.325619 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:36 crc kubenswrapper[4870]: I1014 07:01:36.325632 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:36 crc kubenswrapper[4870]: I1014 07:01:36.325652 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:36 crc kubenswrapper[4870]: I1014 07:01:36.325666 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:36Z","lastTransitionTime":"2025-10-14T07:01:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:36 crc kubenswrapper[4870]: I1014 07:01:36.344033 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-cdjmj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"612e7ab9-bda4-45d8-bc50-ff1afb68aace\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://834444b5350ee12df4e4652c3009880d8d69e1edf8ba0c0156ab96fb2bbe28e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://834444b5350ee12df4e4652c3009880d8d69e1edf8ba0c0156ab96fb2bbe28e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c1972e7b2f75790db22a07a078eee97072bffbdb7f8e594bf2be5f800632a7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c1972e7b2f75790db22a07a078eee97072bffbdb7f8e594bf2be5f800632a7b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b124b42a44c6a23da1d92b9a5326987eb74430f80489a71e8913829475654ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b124b42a44c6a23da1d92b9a5326987eb74430f80489a71e8913829475654ab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62d0deb6b20ea65fbcbaf94c2fc14d69adb22e7a46cc043ccd4def2668ddd4b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62d0deb6b20ea65fbcbaf94c2fc14d69adb22e7a46cc043ccd4def2668ddd4b7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7f0d7cf29ca737d0effe58731a8b890379527c97f78c144d4eebb910260fadb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7f0d7cf29ca737d0effe58731a8b890379527c97f78c144d4eebb910260fadb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5959abb7c8e7ab32bed3aca709ae0334e8400f991ba724246c9fda56cbed822\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5959abb7c8e7ab32bed3aca709ae0334e8400f991ba724246c9fda56cbed822\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-cdjmj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:36Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:36 crc kubenswrapper[4870]: I1014 07:01:36.386147 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fdxjh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a3a430a-b540-4a8c-adad-b2893dbf9898\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93e41d5e927eed7ec62eb9c3607cc804cd8999556663363fd123b2a784b00ff1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k787j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fdxjh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:36Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:36 crc kubenswrapper[4870]: I1014 07:01:36.418266 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-cdjmj" event={"ID":"612e7ab9-bda4-45d8-bc50-ff1afb68aace","Type":"ContainerStarted","Data":"7776352d62b814fc63d01e4c16ad13f878b0597eb37a370bf9da0a88abab6cfa"} Oct 14 07:01:36 crc kubenswrapper[4870]: I1014 07:01:36.418338 4870 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 14 07:01:36 crc kubenswrapper[4870]: I1014 07:01:36.428916 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:36 crc kubenswrapper[4870]: I1014 07:01:36.429010 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:36 crc kubenswrapper[4870]: I1014 07:01:36.429032 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:36 crc kubenswrapper[4870]: I1014 07:01:36.429058 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:36 crc kubenswrapper[4870]: I1014 07:01:36.429079 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:36Z","lastTransitionTime":"2025-10-14T07:01:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:36 crc kubenswrapper[4870]: I1014 07:01:36.433389 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33a9493e-9660-48db-b8d1-310e29ebe313\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3570643ef1b7d5cc3a542f320413b819338e0c8b7ab7f0f09cc909997a39bab8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb398668197dedc1c2a14b4ae7d6a29714f3527309ccb9c1622af841100e1d60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1d2b6d068d63bd38cdfb83824525f6d12054f6cabab5d684e03ea5f2152076c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://783ff243e395333f3fc80cfae21358c532f0c9c5cb5515586be56b90ed10d912\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a8135a3d01076e4437ae8a784b459bf8d16de57fa16ec7ce4d40c5129c9d4a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48ca3a5017ba650df99b2fc7201c3fe5abfa89d81089c5647cde1b1549f534fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48ca3a5017ba650df99b2fc7201c3fe5abfa89d81089c5647cde1b1549f534fc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://19682eca67f76bf2e06e1b80f12e21141fcbce0acb8195f0defacd1b9a10b5e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://19682eca67f76bf2e06e1b80f12e21141fcbce0acb8195f0defacd1b9a10b5e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://042aad6d5f6bee8eef77f9e6a2c265de775a69f3da72aedc68bdf312198f00fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://042aad6d5f6bee8eef77f9e6a2c265de775a69f3da72aedc68bdf312198f00fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:05Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:36Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:36 crc kubenswrapper[4870]: I1014 07:01:36.465045 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-4v2v2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4796ba6-c23b-4764-be96-2e3a3bf5e962\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf5afb0b99a7101f6d5573fb7eb925c5a1d8becc11b2e3c6f3351d80b6f0e823\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnp8b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-4v2v2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:36Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:36 crc kubenswrapper[4870]: I1014 07:01:36.509212 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bcbc6dd6-aab0-4dbc-a62e-c58f06ec1337\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25ec4808b09adc44488530b3c4e29fd7fec46b3fcfa31d4449ec4a6dbc880ec8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f3f5de86fb11231a19034a18e9d7f31476c162e6567d0c136ead8bbbd900d6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d57d938d9d9c92198ae5b3857536ece36012fbfbe8cc8cfd245d1386d39b0be5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77a3247a7f65709543d7fd374b2ab69d7d793537d4132d7e22261188d6e58544\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:05Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:36Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:36 crc kubenswrapper[4870]: I1014 07:01:36.531341 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:36 crc kubenswrapper[4870]: I1014 07:01:36.531391 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:36 crc kubenswrapper[4870]: I1014 07:01:36.531402 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:36 crc kubenswrapper[4870]: I1014 07:01:36.531420 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:36 crc kubenswrapper[4870]: I1014 07:01:36.531432 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:36Z","lastTransitionTime":"2025-10-14T07:01:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:36 crc kubenswrapper[4870]: I1014 07:01:36.548207 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e539fac3d0a2ade314a4b4c1050bb27c37bcff5ba5272fa2d71b97ab5c0d7191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:36Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:36 crc kubenswrapper[4870]: I1014 07:01:36.585124 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:36Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:36 crc kubenswrapper[4870]: I1014 07:01:36.626924 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-kz77f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d3864a91-9bf6-4540-a908-cc41aa39656c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d4ded3d86030f5ea548c8b1cc3caddbe14a45c9a3ae4af836e892880584f57d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4h8v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:28Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-kz77f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:36Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:36 crc kubenswrapper[4870]: I1014 07:01:36.634599 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:36 crc kubenswrapper[4870]: I1014 07:01:36.634667 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:36 crc kubenswrapper[4870]: I1014 07:01:36.634718 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:36 crc kubenswrapper[4870]: I1014 07:01:36.634800 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:36 crc kubenswrapper[4870]: I1014 07:01:36.634821 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:36Z","lastTransitionTime":"2025-10-14T07:01:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:36 crc kubenswrapper[4870]: I1014 07:01:36.678514 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33a9493e-9660-48db-b8d1-310e29ebe313\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3570643ef1b7d5cc3a542f320413b819338e0c8b7ab7f0f09cc909997a39bab8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb398668197dedc1c2a14b4ae7d6a29714f3527309ccb9c1622af841100e1d60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1d2b6d068d63bd38cdfb83824525f6d12054f6cabab5d684e03ea5f2152076c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://783ff243e395333f3fc80cfae21358c532f0c9c5cb5515586be56b90ed10d912\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a8135a3d01076e4437ae8a784b459bf8d16de57fa16ec7ce4d40c5129c9d4a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48ca3a5017ba650df99b2fc7201c3fe5abfa89d81089c5647cde1b1549f534fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48ca3a5017ba650df99b2fc7201c3fe5abfa89d81089c5647cde1b1549f534fc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://19682eca67f76bf2e06e1b80f12e21141fcbce0acb8195f0defacd1b9a10b5e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://19682eca67f76bf2e06e1b80f12e21141fcbce0acb8195f0defacd1b9a10b5e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://042aad6d5f6bee8eef77f9e6a2c265de775a69f3da72aedc68bdf312198f00fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://042aad6d5f6bee8eef77f9e6a2c265de775a69f3da72aedc68bdf312198f00fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:05Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:36Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:36 crc kubenswrapper[4870]: I1014 07:01:36.701647 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-4v2v2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4796ba6-c23b-4764-be96-2e3a3bf5e962\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf5afb0b99a7101f6d5573fb7eb925c5a1d8becc11b2e3c6f3351d80b6f0e823\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnp8b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-4v2v2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:36Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:36 crc kubenswrapper[4870]: I1014 07:01:36.737662 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:36 crc kubenswrapper[4870]: I1014 07:01:36.737738 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:36 crc kubenswrapper[4870]: I1014 07:01:36.737758 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:36 crc kubenswrapper[4870]: I1014 07:01:36.737788 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:36 crc kubenswrapper[4870]: I1014 07:01:36.737806 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:36Z","lastTransitionTime":"2025-10-14T07:01:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:36 crc kubenswrapper[4870]: I1014 07:01:36.746643 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-cdjmj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"612e7ab9-bda4-45d8-bc50-ff1afb68aace\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7776352d62b814fc63d01e4c16ad13f878b0597eb37a370bf9da0a88abab6cfa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://834444b5350ee12df4e4652c3009880d8d69e1edf8ba0c0156ab96fb2bbe28e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://834444b5350ee12df4e4652c3009880d8d69e1edf8ba0c0156ab96fb2bbe28e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c1972e7b2f75790db22a07a078eee97072bffbdb7f8e594bf2be5f800632a7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c1972e7b2f75790db22a07a078eee97072bffbdb7f8e594bf2be5f800632a7b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b124b42a44c6a23da1d92b9a5326987eb74430f80489a71e8913829475654ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b124b42a44c6a23da1d92b9a5326987eb74430f80489a71e8913829475654ab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62d0deb6b20ea65fbcbaf94c2fc14d69adb22e7a46cc043ccd4def2668ddd4b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62d0deb6b20ea65fbcbaf94c2fc14d69adb22e7a46cc043ccd4def2668ddd4b7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7f0d7cf29ca737d0effe58731a8b890379527c97f78c144d4eebb910260fadb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7f0d7cf29ca737d0effe58731a8b890379527c97f78c144d4eebb910260fadb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5959abb7c8e7ab32bed3aca709ae0334e8400f991ba724246c9fda56cbed822\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5959abb7c8e7ab32bed3aca709ae0334e8400f991ba724246c9fda56cbed822\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-cdjmj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:36Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:36 crc kubenswrapper[4870]: I1014 07:01:36.795059 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fdxjh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a3a430a-b540-4a8c-adad-b2893dbf9898\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93e41d5e927eed7ec62eb9c3607cc804cd8999556663363fd123b2a784b00ff1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k787j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fdxjh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:36Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:36 crc kubenswrapper[4870]: I1014 07:01:36.829532 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bcbc6dd6-aab0-4dbc-a62e-c58f06ec1337\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25ec4808b09adc44488530b3c4e29fd7fec46b3fcfa31d4449ec4a6dbc880ec8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f3f5de86fb11231a19034a18e9d7f31476c162e6567d0c136ead8bbbd900d6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d57d938d9d9c92198ae5b3857536ece36012fbfbe8cc8cfd245d1386d39b0be5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77a3247a7f65709543d7fd374b2ab69d7d793537d4132d7e22261188d6e58544\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:05Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:36Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:36 crc kubenswrapper[4870]: I1014 07:01:36.841467 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:36 crc kubenswrapper[4870]: I1014 07:01:36.841998 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:36 crc kubenswrapper[4870]: I1014 07:01:36.842192 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:36 crc kubenswrapper[4870]: I1014 07:01:36.842466 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:36 crc kubenswrapper[4870]: I1014 07:01:36.842655 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:36Z","lastTransitionTime":"2025-10-14T07:01:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:36 crc kubenswrapper[4870]: I1014 07:01:36.867218 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e539fac3d0a2ade314a4b4c1050bb27c37bcff5ba5272fa2d71b97ab5c0d7191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:36Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:36 crc kubenswrapper[4870]: I1014 07:01:36.900867 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:36Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:36 crc kubenswrapper[4870]: I1014 07:01:36.938693 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-kz77f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d3864a91-9bf6-4540-a908-cc41aa39656c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d4ded3d86030f5ea548c8b1cc3caddbe14a45c9a3ae4af836e892880584f57d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4h8v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:28Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-kz77f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:36Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:36 crc kubenswrapper[4870]: I1014 07:01:36.945280 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:36 crc kubenswrapper[4870]: I1014 07:01:36.945305 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:36 crc kubenswrapper[4870]: I1014 07:01:36.945314 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:36 crc kubenswrapper[4870]: I1014 07:01:36.945328 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:36 crc kubenswrapper[4870]: I1014 07:01:36.945339 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:36Z","lastTransitionTime":"2025-10-14T07:01:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:36 crc kubenswrapper[4870]: I1014 07:01:36.980246 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:36Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:37 crc kubenswrapper[4870]: I1014 07:01:37.022532 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3eb0b397c05ffd6650607f1be4a75683e0342d3333fc096bf1cd63c7688834f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:37Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:37 crc kubenswrapper[4870]: I1014 07:01:37.048110 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:37 crc kubenswrapper[4870]: I1014 07:01:37.048168 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:37 crc kubenswrapper[4870]: I1014 07:01:37.048179 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:37 crc kubenswrapper[4870]: I1014 07:01:37.048202 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:37 crc kubenswrapper[4870]: I1014 07:01:37.048219 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:37Z","lastTransitionTime":"2025-10-14T07:01:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:37 crc kubenswrapper[4870]: I1014 07:01:37.061588 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"874a5e45-dffd-4d17-b609-b7d3ed2eab07\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d98d044836ab44a405196915ef7bf572b75113f2c81ec7e64effa63b47ad3543\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wj4ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8004752575df4a1c44efbcafe64799f09126d0f97285d92388c76a0b90e022f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wj4ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7tvc7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:37Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:37 crc kubenswrapper[4870]: I1014 07:01:37.103618 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77a0d93a-5391-4112-b502-c474f45de209\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7815af9a584f6db6660073da289f2e8c987a04ef9ddfac155bdd9a15b014fcfa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b329b570622155a55f70501b2260e278106a68ebef497914231b4ec860e0125\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba942a0e6fe5ab93a187e0b7b4bd7c440c2e7827d9b5323fc50f90b562075392\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://978e8db184ba9e430232c6bb2ee530beb92567fdc71ba3c35b0bedca3492d845\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://978e8db184ba9e430232c6bb2ee530beb92567fdc71ba3c35b0bedca3492d845\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"file observer\\\\nW1014 07:01:25.870672 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1014 07:01:25.870795 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1014 07:01:25.871521 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-756714914/tls.crt::/tmp/serving-cert-756714914/tls.key\\\\\\\"\\\\nI1014 07:01:26.460037 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1014 07:01:26.464876 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1014 07:01:26.464902 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1014 07:01:26.464925 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1014 07:01:26.464931 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1014 07:01:26.481374 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1014 07:01:26.481400 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:01:26.481405 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:01:26.481410 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1014 07:01:26.481414 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1014 07:01:26.481417 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1014 07:01:26.481420 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1014 07:01:26.481610 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1014 07:01:26.482872 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bd0044a247bdc5ca4cab7294103af3270b76cde61f3109b85e83ed050ca53b6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82d3c2214e0be679d2827594b159ace362151318689078bc49d3158f42899636\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82d3c2214e0be679d2827594b159ace362151318689078bc49d3158f42899636\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:05Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:37Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:37 crc kubenswrapper[4870]: I1014 07:01:37.141579 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:37Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:37 crc kubenswrapper[4870]: I1014 07:01:37.150485 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:37 crc kubenswrapper[4870]: I1014 07:01:37.150524 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:37 crc kubenswrapper[4870]: I1014 07:01:37.150542 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:37 crc kubenswrapper[4870]: I1014 07:01:37.150567 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:37 crc kubenswrapper[4870]: I1014 07:01:37.150583 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:37Z","lastTransitionTime":"2025-10-14T07:01:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:37 crc kubenswrapper[4870]: I1014 07:01:37.181634 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e26fe49ec2045d20a0cb2dd17e96e79b6eb43c641eb184c688cc472ca5db39d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aea115f7ee71c700b218fe15c928af7b55e6babe70e65d67f125d87be1fd2e2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:37Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:37 crc kubenswrapper[4870]: I1014 07:01:37.225799 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-v7br2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dcaef0e-42e3-424b-bbe5-4a023efe912d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71dec55e859b11a081dbb90d7d6cf6b4846cf591382303d8329003ff50c71f2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e72cff0723e2ce520340244b4e597d8016c58e1b071c4b087f2f6cd718741594\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc9d4db944f837289d1923176d60cfc4a566a282c2d2e3a703e220df884703c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d16ed1c65925549dcb69ab02c3013e25e3f123a83179f1af3a8e08bf91a59896\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed59243b59c39ce82bf9bc8f33c0a8b84854846afff6a89f5d37ffc97a096b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3b6d20cf459b1f6d3aff5b9290f9391909baf27ee40c49e2b0cd2d5cfe889ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c43b07bd7ecc0a98ea65b6f59752ad8651568236dc78ee216e0ca42dabcbe4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fe3468cdd612fc4378e1f6f55a2a9d00274f7ddb65cb57b2a368dfc065cfb4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ba24c2e067b9f3a14d9d018b66300b54d866530d30174fbdb3c899299575d28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ba24c2e067b9f3a14d9d018b66300b54d866530d30174fbdb3c899299575d28\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-v7br2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:37Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:37 crc kubenswrapper[4870]: I1014 07:01:37.253494 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:37 crc kubenswrapper[4870]: I1014 07:01:37.253554 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:37 crc kubenswrapper[4870]: I1014 07:01:37.253572 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:37 crc kubenswrapper[4870]: I1014 07:01:37.253597 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:37 crc kubenswrapper[4870]: I1014 07:01:37.253617 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:37Z","lastTransitionTime":"2025-10-14T07:01:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:37 crc kubenswrapper[4870]: I1014 07:01:37.356864 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:37 crc kubenswrapper[4870]: I1014 07:01:37.356917 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:37 crc kubenswrapper[4870]: I1014 07:01:37.356929 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:37 crc kubenswrapper[4870]: I1014 07:01:37.356947 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:37 crc kubenswrapper[4870]: I1014 07:01:37.356960 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:37Z","lastTransitionTime":"2025-10-14T07:01:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:37 crc kubenswrapper[4870]: I1014 07:01:37.421058 4870 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 14 07:01:37 crc kubenswrapper[4870]: I1014 07:01:37.459417 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:37 crc kubenswrapper[4870]: I1014 07:01:37.459488 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:37 crc kubenswrapper[4870]: I1014 07:01:37.459500 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:37 crc kubenswrapper[4870]: I1014 07:01:37.459518 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:37 crc kubenswrapper[4870]: I1014 07:01:37.459530 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:37Z","lastTransitionTime":"2025-10-14T07:01:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:37 crc kubenswrapper[4870]: I1014 07:01:37.484730 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-v7br2" Oct 14 07:01:37 crc kubenswrapper[4870]: I1014 07:01:37.562355 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:37 crc kubenswrapper[4870]: I1014 07:01:37.562418 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:37 crc kubenswrapper[4870]: I1014 07:01:37.562462 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:37 crc kubenswrapper[4870]: I1014 07:01:37.562486 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:37 crc kubenswrapper[4870]: I1014 07:01:37.562500 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:37Z","lastTransitionTime":"2025-10-14T07:01:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:37 crc kubenswrapper[4870]: I1014 07:01:37.665677 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:37 crc kubenswrapper[4870]: I1014 07:01:37.665787 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:37 crc kubenswrapper[4870]: I1014 07:01:37.665809 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:37 crc kubenswrapper[4870]: I1014 07:01:37.665840 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:37 crc kubenswrapper[4870]: I1014 07:01:37.665859 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:37Z","lastTransitionTime":"2025-10-14T07:01:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:37 crc kubenswrapper[4870]: I1014 07:01:37.769886 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:37 crc kubenswrapper[4870]: I1014 07:01:37.769988 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:37 crc kubenswrapper[4870]: I1014 07:01:37.770016 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:37 crc kubenswrapper[4870]: I1014 07:01:37.770047 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:37 crc kubenswrapper[4870]: I1014 07:01:37.770068 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:37Z","lastTransitionTime":"2025-10-14T07:01:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:37 crc kubenswrapper[4870]: I1014 07:01:37.873545 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:37 crc kubenswrapper[4870]: I1014 07:01:37.873631 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:37 crc kubenswrapper[4870]: I1014 07:01:37.873652 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:37 crc kubenswrapper[4870]: I1014 07:01:37.873691 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:37 crc kubenswrapper[4870]: I1014 07:01:37.873716 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:37Z","lastTransitionTime":"2025-10-14T07:01:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:37 crc kubenswrapper[4870]: I1014 07:01:37.976583 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:37 crc kubenswrapper[4870]: I1014 07:01:37.976638 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:37 crc kubenswrapper[4870]: I1014 07:01:37.976650 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:37 crc kubenswrapper[4870]: I1014 07:01:37.976666 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:37 crc kubenswrapper[4870]: I1014 07:01:37.976675 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:37Z","lastTransitionTime":"2025-10-14T07:01:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:38 crc kubenswrapper[4870]: I1014 07:01:38.033645 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 07:01:38 crc kubenswrapper[4870]: E1014 07:01:38.033827 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 07:01:38 crc kubenswrapper[4870]: I1014 07:01:38.033932 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 07:01:38 crc kubenswrapper[4870]: I1014 07:01:38.033997 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 07:01:38 crc kubenswrapper[4870]: E1014 07:01:38.034169 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 07:01:38 crc kubenswrapper[4870]: E1014 07:01:38.034264 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 07:01:38 crc kubenswrapper[4870]: I1014 07:01:38.079702 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:38 crc kubenswrapper[4870]: I1014 07:01:38.079760 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:38 crc kubenswrapper[4870]: I1014 07:01:38.079770 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:38 crc kubenswrapper[4870]: I1014 07:01:38.079790 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:38 crc kubenswrapper[4870]: I1014 07:01:38.079802 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:38Z","lastTransitionTime":"2025-10-14T07:01:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:38 crc kubenswrapper[4870]: I1014 07:01:38.182104 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:38 crc kubenswrapper[4870]: I1014 07:01:38.182153 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:38 crc kubenswrapper[4870]: I1014 07:01:38.182171 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:38 crc kubenswrapper[4870]: I1014 07:01:38.182188 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:38 crc kubenswrapper[4870]: I1014 07:01:38.182197 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:38Z","lastTransitionTime":"2025-10-14T07:01:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:38 crc kubenswrapper[4870]: I1014 07:01:38.284787 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:38 crc kubenswrapper[4870]: I1014 07:01:38.284837 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:38 crc kubenswrapper[4870]: I1014 07:01:38.284847 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:38 crc kubenswrapper[4870]: I1014 07:01:38.284867 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:38 crc kubenswrapper[4870]: I1014 07:01:38.284878 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:38Z","lastTransitionTime":"2025-10-14T07:01:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:38 crc kubenswrapper[4870]: I1014 07:01:38.387845 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:38 crc kubenswrapper[4870]: I1014 07:01:38.387881 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:38 crc kubenswrapper[4870]: I1014 07:01:38.387890 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:38 crc kubenswrapper[4870]: I1014 07:01:38.387905 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:38 crc kubenswrapper[4870]: I1014 07:01:38.387916 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:38Z","lastTransitionTime":"2025-10-14T07:01:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:38 crc kubenswrapper[4870]: I1014 07:01:38.443635 4870 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-ovn-kubernetes/ovnkube-node-v7br2" podUID="8dcaef0e-42e3-424b-bbe5-4a023efe912d" containerName="ovnkube-controller" probeResult="failure" output="" Oct 14 07:01:38 crc kubenswrapper[4870]: I1014 07:01:38.503752 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:38 crc kubenswrapper[4870]: I1014 07:01:38.503810 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:38 crc kubenswrapper[4870]: I1014 07:01:38.503825 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:38 crc kubenswrapper[4870]: I1014 07:01:38.503850 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:38 crc kubenswrapper[4870]: I1014 07:01:38.503864 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:38Z","lastTransitionTime":"2025-10-14T07:01:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:38 crc kubenswrapper[4870]: I1014 07:01:38.606235 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:38 crc kubenswrapper[4870]: I1014 07:01:38.606265 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:38 crc kubenswrapper[4870]: I1014 07:01:38.606274 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:38 crc kubenswrapper[4870]: I1014 07:01:38.606289 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:38 crc kubenswrapper[4870]: I1014 07:01:38.606299 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:38Z","lastTransitionTime":"2025-10-14T07:01:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:38 crc kubenswrapper[4870]: I1014 07:01:38.708760 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:38 crc kubenswrapper[4870]: I1014 07:01:38.708807 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:38 crc kubenswrapper[4870]: I1014 07:01:38.708821 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:38 crc kubenswrapper[4870]: I1014 07:01:38.708837 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:38 crc kubenswrapper[4870]: I1014 07:01:38.708849 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:38Z","lastTransitionTime":"2025-10-14T07:01:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:38 crc kubenswrapper[4870]: I1014 07:01:38.839329 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:38 crc kubenswrapper[4870]: I1014 07:01:38.839366 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:38 crc kubenswrapper[4870]: I1014 07:01:38.839376 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:38 crc kubenswrapper[4870]: I1014 07:01:38.839392 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:38 crc kubenswrapper[4870]: I1014 07:01:38.839401 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:38Z","lastTransitionTime":"2025-10-14T07:01:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:38 crc kubenswrapper[4870]: I1014 07:01:38.942526 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:38 crc kubenswrapper[4870]: I1014 07:01:38.942688 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:38 crc kubenswrapper[4870]: I1014 07:01:38.942708 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:38 crc kubenswrapper[4870]: I1014 07:01:38.942726 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:38 crc kubenswrapper[4870]: I1014 07:01:38.942743 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:38Z","lastTransitionTime":"2025-10-14T07:01:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:39 crc kubenswrapper[4870]: I1014 07:01:39.045315 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:39 crc kubenswrapper[4870]: I1014 07:01:39.045371 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:39 crc kubenswrapper[4870]: I1014 07:01:39.045383 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:39 crc kubenswrapper[4870]: I1014 07:01:39.045424 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:39 crc kubenswrapper[4870]: I1014 07:01:39.045513 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:39Z","lastTransitionTime":"2025-10-14T07:01:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:39 crc kubenswrapper[4870]: I1014 07:01:39.148029 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:39 crc kubenswrapper[4870]: I1014 07:01:39.148061 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:39 crc kubenswrapper[4870]: I1014 07:01:39.148071 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:39 crc kubenswrapper[4870]: I1014 07:01:39.148086 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:39 crc kubenswrapper[4870]: I1014 07:01:39.148095 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:39Z","lastTransitionTime":"2025-10-14T07:01:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:39 crc kubenswrapper[4870]: I1014 07:01:39.251326 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:39 crc kubenswrapper[4870]: I1014 07:01:39.251391 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:39 crc kubenswrapper[4870]: I1014 07:01:39.251412 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:39 crc kubenswrapper[4870]: I1014 07:01:39.251472 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:39 crc kubenswrapper[4870]: I1014 07:01:39.251494 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:39Z","lastTransitionTime":"2025-10-14T07:01:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:39 crc kubenswrapper[4870]: I1014 07:01:39.355162 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:39 crc kubenswrapper[4870]: I1014 07:01:39.355225 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:39 crc kubenswrapper[4870]: I1014 07:01:39.355244 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:39 crc kubenswrapper[4870]: I1014 07:01:39.355273 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:39 crc kubenswrapper[4870]: I1014 07:01:39.355293 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:39Z","lastTransitionTime":"2025-10-14T07:01:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:39 crc kubenswrapper[4870]: I1014 07:01:39.458016 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:39 crc kubenswrapper[4870]: I1014 07:01:39.458129 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:39 crc kubenswrapper[4870]: I1014 07:01:39.458144 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:39 crc kubenswrapper[4870]: I1014 07:01:39.458169 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:39 crc kubenswrapper[4870]: I1014 07:01:39.458184 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:39Z","lastTransitionTime":"2025-10-14T07:01:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:39 crc kubenswrapper[4870]: I1014 07:01:39.502502 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-khtq8"] Oct 14 07:01:39 crc kubenswrapper[4870]: I1014 07:01:39.503252 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-khtq8" Oct 14 07:01:39 crc kubenswrapper[4870]: I1014 07:01:39.506320 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Oct 14 07:01:39 crc kubenswrapper[4870]: I1014 07:01:39.510149 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Oct 14 07:01:39 crc kubenswrapper[4870]: I1014 07:01:39.527841 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77a0d93a-5391-4112-b502-c474f45de209\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7815af9a584f6db6660073da289f2e8c987a04ef9ddfac155bdd9a15b014fcfa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b329b570622155a55f70501b2260e278106a68ebef497914231b4ec860e0125\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba942a0e6fe5ab93a187e0b7b4bd7c440c2e7827d9b5323fc50f90b562075392\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://978e8db184ba9e430232c6bb2ee530beb92567fdc71ba3c35b0bedca3492d845\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://978e8db184ba9e430232c6bb2ee530beb92567fdc71ba3c35b0bedca3492d845\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"file observer\\\\nW1014 07:01:25.870672 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1014 07:01:25.870795 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1014 07:01:25.871521 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-756714914/tls.crt::/tmp/serving-cert-756714914/tls.key\\\\\\\"\\\\nI1014 07:01:26.460037 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1014 07:01:26.464876 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1014 07:01:26.464902 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1014 07:01:26.464925 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1014 07:01:26.464931 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1014 07:01:26.481374 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1014 07:01:26.481400 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:01:26.481405 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:01:26.481410 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1014 07:01:26.481414 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1014 07:01:26.481417 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1014 07:01:26.481420 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1014 07:01:26.481610 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1014 07:01:26.482872 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bd0044a247bdc5ca4cab7294103af3270b76cde61f3109b85e83ed050ca53b6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82d3c2214e0be679d2827594b159ace362151318689078bc49d3158f42899636\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82d3c2214e0be679d2827594b159ace362151318689078bc49d3158f42899636\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:05Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:39Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:39 crc kubenswrapper[4870]: I1014 07:01:39.544922 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:39Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:39 crc kubenswrapper[4870]: I1014 07:01:39.560386 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/28deeea6-a1ef-4a4a-a377-d954ccd87de9-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-khtq8\" (UID: \"28deeea6-a1ef-4a4a-a377-d954ccd87de9\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-khtq8" Oct 14 07:01:39 crc kubenswrapper[4870]: I1014 07:01:39.560522 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/28deeea6-a1ef-4a4a-a377-d954ccd87de9-env-overrides\") pod \"ovnkube-control-plane-749d76644c-khtq8\" (UID: \"28deeea6-a1ef-4a4a-a377-d954ccd87de9\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-khtq8" Oct 14 07:01:39 crc kubenswrapper[4870]: I1014 07:01:39.560582 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rb79p\" (UniqueName: \"kubernetes.io/projected/28deeea6-a1ef-4a4a-a377-d954ccd87de9-kube-api-access-rb79p\") pod \"ovnkube-control-plane-749d76644c-khtq8\" (UID: \"28deeea6-a1ef-4a4a-a377-d954ccd87de9\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-khtq8" Oct 14 07:01:39 crc kubenswrapper[4870]: I1014 07:01:39.560636 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/28deeea6-a1ef-4a4a-a377-d954ccd87de9-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-khtq8\" (UID: \"28deeea6-a1ef-4a4a-a377-d954ccd87de9\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-khtq8" Oct 14 07:01:39 crc kubenswrapper[4870]: I1014 07:01:39.561663 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:39 crc kubenswrapper[4870]: I1014 07:01:39.561693 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:39 crc kubenswrapper[4870]: I1014 07:01:39.561703 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:39 crc kubenswrapper[4870]: I1014 07:01:39.561720 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:39 crc kubenswrapper[4870]: I1014 07:01:39.561731 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:39Z","lastTransitionTime":"2025-10-14T07:01:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:39 crc kubenswrapper[4870]: I1014 07:01:39.567294 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e26fe49ec2045d20a0cb2dd17e96e79b6eb43c641eb184c688cc472ca5db39d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aea115f7ee71c700b218fe15c928af7b55e6babe70e65d67f125d87be1fd2e2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:39Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:39 crc kubenswrapper[4870]: I1014 07:01:39.596172 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-v7br2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dcaef0e-42e3-424b-bbe5-4a023efe912d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71dec55e859b11a081dbb90d7d6cf6b4846cf591382303d8329003ff50c71f2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e72cff0723e2ce520340244b4e597d8016c58e1b071c4b087f2f6cd718741594\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc9d4db944f837289d1923176d60cfc4a566a282c2d2e3a703e220df884703c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d16ed1c65925549dcb69ab02c3013e25e3f123a83179f1af3a8e08bf91a59896\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed59243b59c39ce82bf9bc8f33c0a8b84854846afff6a89f5d37ffc97a096b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3b6d20cf459b1f6d3aff5b9290f9391909baf27ee40c49e2b0cd2d5cfe889ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c43b07bd7ecc0a98ea65b6f59752ad8651568236dc78ee216e0ca42dabcbe4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fe3468cdd612fc4378e1f6f55a2a9d00274f7ddb65cb57b2a368dfc065cfb4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ba24c2e067b9f3a14d9d018b66300b54d866530d30174fbdb3c899299575d28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ba24c2e067b9f3a14d9d018b66300b54d866530d30174fbdb3c899299575d28\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-v7br2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:39Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:39 crc kubenswrapper[4870]: I1014 07:01:39.627253 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33a9493e-9660-48db-b8d1-310e29ebe313\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3570643ef1b7d5cc3a542f320413b819338e0c8b7ab7f0f09cc909997a39bab8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb398668197dedc1c2a14b4ae7d6a29714f3527309ccb9c1622af841100e1d60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1d2b6d068d63bd38cdfb83824525f6d12054f6cabab5d684e03ea5f2152076c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://783ff243e395333f3fc80cfae21358c532f0c9c5cb5515586be56b90ed10d912\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a8135a3d01076e4437ae8a784b459bf8d16de57fa16ec7ce4d40c5129c9d4a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48ca3a5017ba650df99b2fc7201c3fe5abfa89d81089c5647cde1b1549f534fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48ca3a5017ba650df99b2fc7201c3fe5abfa89d81089c5647cde1b1549f534fc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://19682eca67f76bf2e06e1b80f12e21141fcbce0acb8195f0defacd1b9a10b5e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://19682eca67f76bf2e06e1b80f12e21141fcbce0acb8195f0defacd1b9a10b5e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://042aad6d5f6bee8eef77f9e6a2c265de775a69f3da72aedc68bdf312198f00fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://042aad6d5f6bee8eef77f9e6a2c265de775a69f3da72aedc68bdf312198f00fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:05Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:39Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:39 crc kubenswrapper[4870]: I1014 07:01:39.642269 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-4v2v2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4796ba6-c23b-4764-be96-2e3a3bf5e962\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf5afb0b99a7101f6d5573fb7eb925c5a1d8becc11b2e3c6f3351d80b6f0e823\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnp8b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-4v2v2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:39Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:39 crc kubenswrapper[4870]: I1014 07:01:39.661268 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/28deeea6-a1ef-4a4a-a377-d954ccd87de9-env-overrides\") pod \"ovnkube-control-plane-749d76644c-khtq8\" (UID: \"28deeea6-a1ef-4a4a-a377-d954ccd87de9\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-khtq8" Oct 14 07:01:39 crc kubenswrapper[4870]: I1014 07:01:39.661335 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rb79p\" (UniqueName: \"kubernetes.io/projected/28deeea6-a1ef-4a4a-a377-d954ccd87de9-kube-api-access-rb79p\") pod \"ovnkube-control-plane-749d76644c-khtq8\" (UID: \"28deeea6-a1ef-4a4a-a377-d954ccd87de9\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-khtq8" Oct 14 07:01:39 crc kubenswrapper[4870]: I1014 07:01:39.661393 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/28deeea6-a1ef-4a4a-a377-d954ccd87de9-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-khtq8\" (UID: \"28deeea6-a1ef-4a4a-a377-d954ccd87de9\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-khtq8" Oct 14 07:01:39 crc kubenswrapper[4870]: I1014 07:01:39.661468 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/28deeea6-a1ef-4a4a-a377-d954ccd87de9-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-khtq8\" (UID: \"28deeea6-a1ef-4a4a-a377-d954ccd87de9\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-khtq8" Oct 14 07:01:39 crc kubenswrapper[4870]: I1014 07:01:39.662352 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/28deeea6-a1ef-4a4a-a377-d954ccd87de9-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-khtq8\" (UID: \"28deeea6-a1ef-4a4a-a377-d954ccd87de9\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-khtq8" Oct 14 07:01:39 crc kubenswrapper[4870]: I1014 07:01:39.662977 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/28deeea6-a1ef-4a4a-a377-d954ccd87de9-env-overrides\") pod \"ovnkube-control-plane-749d76644c-khtq8\" (UID: \"28deeea6-a1ef-4a4a-a377-d954ccd87de9\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-khtq8" Oct 14 07:01:39 crc kubenswrapper[4870]: I1014 07:01:39.665708 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:39 crc kubenswrapper[4870]: I1014 07:01:39.666231 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:39 crc kubenswrapper[4870]: I1014 07:01:39.666253 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:39 crc kubenswrapper[4870]: I1014 07:01:39.666277 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:39 crc kubenswrapper[4870]: I1014 07:01:39.666294 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:39Z","lastTransitionTime":"2025-10-14T07:01:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:39 crc kubenswrapper[4870]: I1014 07:01:39.668530 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-cdjmj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"612e7ab9-bda4-45d8-bc50-ff1afb68aace\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7776352d62b814fc63d01e4c16ad13f878b0597eb37a370bf9da0a88abab6cfa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://834444b5350ee12df4e4652c3009880d8d69e1edf8ba0c0156ab96fb2bbe28e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://834444b5350ee12df4e4652c3009880d8d69e1edf8ba0c0156ab96fb2bbe28e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c1972e7b2f75790db22a07a078eee97072bffbdb7f8e594bf2be5f800632a7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c1972e7b2f75790db22a07a078eee97072bffbdb7f8e594bf2be5f800632a7b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b124b42a44c6a23da1d92b9a5326987eb74430f80489a71e8913829475654ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b124b42a44c6a23da1d92b9a5326987eb74430f80489a71e8913829475654ab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62d0deb6b20ea65fbcbaf94c2fc14d69adb22e7a46cc043ccd4def2668ddd4b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62d0deb6b20ea65fbcbaf94c2fc14d69adb22e7a46cc043ccd4def2668ddd4b7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7f0d7cf29ca737d0effe58731a8b890379527c97f78c144d4eebb910260fadb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7f0d7cf29ca737d0effe58731a8b890379527c97f78c144d4eebb910260fadb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5959abb7c8e7ab32bed3aca709ae0334e8400f991ba724246c9fda56cbed822\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5959abb7c8e7ab32bed3aca709ae0334e8400f991ba724246c9fda56cbed822\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-cdjmj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:39Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:39 crc kubenswrapper[4870]: I1014 07:01:39.673327 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/28deeea6-a1ef-4a4a-a377-d954ccd87de9-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-khtq8\" (UID: \"28deeea6-a1ef-4a4a-a377-d954ccd87de9\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-khtq8" Oct 14 07:01:39 crc kubenswrapper[4870]: I1014 07:01:39.690426 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rb79p\" (UniqueName: \"kubernetes.io/projected/28deeea6-a1ef-4a4a-a377-d954ccd87de9-kube-api-access-rb79p\") pod \"ovnkube-control-plane-749d76644c-khtq8\" (UID: \"28deeea6-a1ef-4a4a-a377-d954ccd87de9\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-khtq8" Oct 14 07:01:39 crc kubenswrapper[4870]: I1014 07:01:39.691413 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fdxjh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a3a430a-b540-4a8c-adad-b2893dbf9898\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93e41d5e927eed7ec62eb9c3607cc804cd8999556663363fd123b2a784b00ff1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k787j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fdxjh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:39Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:39 crc kubenswrapper[4870]: I1014 07:01:39.704916 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-khtq8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28deeea6-a1ef-4a4a-a377-d954ccd87de9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rb79p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rb79p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-khtq8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:39Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:39 crc kubenswrapper[4870]: I1014 07:01:39.722555 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bcbc6dd6-aab0-4dbc-a62e-c58f06ec1337\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25ec4808b09adc44488530b3c4e29fd7fec46b3fcfa31d4449ec4a6dbc880ec8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f3f5de86fb11231a19034a18e9d7f31476c162e6567d0c136ead8bbbd900d6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d57d938d9d9c92198ae5b3857536ece36012fbfbe8cc8cfd245d1386d39b0be5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77a3247a7f65709543d7fd374b2ab69d7d793537d4132d7e22261188d6e58544\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:05Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:39Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:39 crc kubenswrapper[4870]: I1014 07:01:39.740457 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e539fac3d0a2ade314a4b4c1050bb27c37bcff5ba5272fa2d71b97ab5c0d7191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:39Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:39 crc kubenswrapper[4870]: I1014 07:01:39.753230 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:39Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:39 crc kubenswrapper[4870]: I1014 07:01:39.765391 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-kz77f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d3864a91-9bf6-4540-a908-cc41aa39656c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d4ded3d86030f5ea548c8b1cc3caddbe14a45c9a3ae4af836e892880584f57d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4h8v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:28Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-kz77f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:39Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:39 crc kubenswrapper[4870]: I1014 07:01:39.769908 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:39 crc kubenswrapper[4870]: I1014 07:01:39.769968 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:39 crc kubenswrapper[4870]: I1014 07:01:39.769985 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:39 crc kubenswrapper[4870]: I1014 07:01:39.770014 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:39 crc kubenswrapper[4870]: I1014 07:01:39.770029 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:39Z","lastTransitionTime":"2025-10-14T07:01:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:39 crc kubenswrapper[4870]: I1014 07:01:39.780747 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:39Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:39 crc kubenswrapper[4870]: I1014 07:01:39.797185 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3eb0b397c05ffd6650607f1be4a75683e0342d3333fc096bf1cd63c7688834f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:39Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:39 crc kubenswrapper[4870]: I1014 07:01:39.809487 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"874a5e45-dffd-4d17-b609-b7d3ed2eab07\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d98d044836ab44a405196915ef7bf572b75113f2c81ec7e64effa63b47ad3543\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wj4ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8004752575df4a1c44efbcafe64799f09126d0f97285d92388c76a0b90e022f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wj4ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7tvc7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:39Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:39 crc kubenswrapper[4870]: I1014 07:01:39.826046 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-khtq8" Oct 14 07:01:39 crc kubenswrapper[4870]: W1014 07:01:39.836855 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod28deeea6_a1ef_4a4a_a377_d954ccd87de9.slice/crio-519feb50a3108a33c9f93811d1d44293ad01682655a3d20f11ed2e4eb8eb9ab1 WatchSource:0}: Error finding container 519feb50a3108a33c9f93811d1d44293ad01682655a3d20f11ed2e4eb8eb9ab1: Status 404 returned error can't find the container with id 519feb50a3108a33c9f93811d1d44293ad01682655a3d20f11ed2e4eb8eb9ab1 Oct 14 07:01:39 crc kubenswrapper[4870]: I1014 07:01:39.872949 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:39 crc kubenswrapper[4870]: I1014 07:01:39.872981 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:39 crc kubenswrapper[4870]: I1014 07:01:39.872988 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:39 crc kubenswrapper[4870]: I1014 07:01:39.873003 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:39 crc kubenswrapper[4870]: I1014 07:01:39.873012 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:39Z","lastTransitionTime":"2025-10-14T07:01:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:39 crc kubenswrapper[4870]: I1014 07:01:39.975272 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:39 crc kubenswrapper[4870]: I1014 07:01:39.975344 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:39 crc kubenswrapper[4870]: I1014 07:01:39.975363 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:39 crc kubenswrapper[4870]: I1014 07:01:39.975394 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:39 crc kubenswrapper[4870]: I1014 07:01:39.975413 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:39Z","lastTransitionTime":"2025-10-14T07:01:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:40 crc kubenswrapper[4870]: I1014 07:01:40.033031 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 07:01:40 crc kubenswrapper[4870]: I1014 07:01:40.033121 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 07:01:40 crc kubenswrapper[4870]: E1014 07:01:40.033187 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 07:01:40 crc kubenswrapper[4870]: E1014 07:01:40.033322 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 07:01:40 crc kubenswrapper[4870]: I1014 07:01:40.033474 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 07:01:40 crc kubenswrapper[4870]: E1014 07:01:40.033565 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 07:01:40 crc kubenswrapper[4870]: I1014 07:01:40.079187 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:40 crc kubenswrapper[4870]: I1014 07:01:40.079225 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:40 crc kubenswrapper[4870]: I1014 07:01:40.079235 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:40 crc kubenswrapper[4870]: I1014 07:01:40.079252 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:40 crc kubenswrapper[4870]: I1014 07:01:40.079263 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:40Z","lastTransitionTime":"2025-10-14T07:01:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:40 crc kubenswrapper[4870]: I1014 07:01:40.182909 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:40 crc kubenswrapper[4870]: I1014 07:01:40.182962 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:40 crc kubenswrapper[4870]: I1014 07:01:40.182979 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:40 crc kubenswrapper[4870]: I1014 07:01:40.183004 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:40 crc kubenswrapper[4870]: I1014 07:01:40.183020 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:40Z","lastTransitionTime":"2025-10-14T07:01:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:40 crc kubenswrapper[4870]: I1014 07:01:40.288269 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:40 crc kubenswrapper[4870]: I1014 07:01:40.288346 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:40 crc kubenswrapper[4870]: I1014 07:01:40.288360 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:40 crc kubenswrapper[4870]: I1014 07:01:40.288383 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:40 crc kubenswrapper[4870]: I1014 07:01:40.288396 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:40Z","lastTransitionTime":"2025-10-14T07:01:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:40 crc kubenswrapper[4870]: I1014 07:01:40.391706 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:40 crc kubenswrapper[4870]: I1014 07:01:40.391783 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:40 crc kubenswrapper[4870]: I1014 07:01:40.391818 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:40 crc kubenswrapper[4870]: I1014 07:01:40.391847 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:40 crc kubenswrapper[4870]: I1014 07:01:40.391863 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:40Z","lastTransitionTime":"2025-10-14T07:01:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:40 crc kubenswrapper[4870]: I1014 07:01:40.431726 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-v7br2_8dcaef0e-42e3-424b-bbe5-4a023efe912d/ovnkube-controller/0.log" Oct 14 07:01:40 crc kubenswrapper[4870]: I1014 07:01:40.434928 4870 generic.go:334] "Generic (PLEG): container finished" podID="8dcaef0e-42e3-424b-bbe5-4a023efe912d" containerID="7c43b07bd7ecc0a98ea65b6f59752ad8651568236dc78ee216e0ca42dabcbe4f" exitCode=1 Oct 14 07:01:40 crc kubenswrapper[4870]: I1014 07:01:40.435024 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-v7br2" event={"ID":"8dcaef0e-42e3-424b-bbe5-4a023efe912d","Type":"ContainerDied","Data":"7c43b07bd7ecc0a98ea65b6f59752ad8651568236dc78ee216e0ca42dabcbe4f"} Oct 14 07:01:40 crc kubenswrapper[4870]: I1014 07:01:40.436118 4870 scope.go:117] "RemoveContainer" containerID="7c43b07bd7ecc0a98ea65b6f59752ad8651568236dc78ee216e0ca42dabcbe4f" Oct 14 07:01:40 crc kubenswrapper[4870]: I1014 07:01:40.437310 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-khtq8" event={"ID":"28deeea6-a1ef-4a4a-a377-d954ccd87de9","Type":"ContainerStarted","Data":"f2b7bc4c93307e80b778f2282ced993898f1c4cfc350a7d74828b8f9dea9906d"} Oct 14 07:01:40 crc kubenswrapper[4870]: I1014 07:01:40.437335 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-khtq8" event={"ID":"28deeea6-a1ef-4a4a-a377-d954ccd87de9","Type":"ContainerStarted","Data":"296bb048a1d99649beb57d609573833bf547cd4d628ebc823c6af83fe5736c9c"} Oct 14 07:01:40 crc kubenswrapper[4870]: I1014 07:01:40.437347 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-khtq8" event={"ID":"28deeea6-a1ef-4a4a-a377-d954ccd87de9","Type":"ContainerStarted","Data":"519feb50a3108a33c9f93811d1d44293ad01682655a3d20f11ed2e4eb8eb9ab1"} Oct 14 07:01:40 crc kubenswrapper[4870]: I1014 07:01:40.450541 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-4v2v2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4796ba6-c23b-4764-be96-2e3a3bf5e962\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf5afb0b99a7101f6d5573fb7eb925c5a1d8becc11b2e3c6f3351d80b6f0e823\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnp8b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-4v2v2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:40Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:40 crc kubenswrapper[4870]: I1014 07:01:40.469047 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-cdjmj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"612e7ab9-bda4-45d8-bc50-ff1afb68aace\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7776352d62b814fc63d01e4c16ad13f878b0597eb37a370bf9da0a88abab6cfa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://834444b5350ee12df4e4652c3009880d8d69e1edf8ba0c0156ab96fb2bbe28e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://834444b5350ee12df4e4652c3009880d8d69e1edf8ba0c0156ab96fb2bbe28e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c1972e7b2f75790db22a07a078eee97072bffbdb7f8e594bf2be5f800632a7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c1972e7b2f75790db22a07a078eee97072bffbdb7f8e594bf2be5f800632a7b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b124b42a44c6a23da1d92b9a5326987eb74430f80489a71e8913829475654ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b124b42a44c6a23da1d92b9a5326987eb74430f80489a71e8913829475654ab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62d0deb6b20ea65fbcbaf94c2fc14d69adb22e7a46cc043ccd4def2668ddd4b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62d0deb6b20ea65fbcbaf94c2fc14d69adb22e7a46cc043ccd4def2668ddd4b7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7f0d7cf29ca737d0effe58731a8b890379527c97f78c144d4eebb910260fadb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7f0d7cf29ca737d0effe58731a8b890379527c97f78c144d4eebb910260fadb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5959abb7c8e7ab32bed3aca709ae0334e8400f991ba724246c9fda56cbed822\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5959abb7c8e7ab32bed3aca709ae0334e8400f991ba724246c9fda56cbed822\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-cdjmj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:40Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:40 crc kubenswrapper[4870]: I1014 07:01:40.485521 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fdxjh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a3a430a-b540-4a8c-adad-b2893dbf9898\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93e41d5e927eed7ec62eb9c3607cc804cd8999556663363fd123b2a784b00ff1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k787j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fdxjh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:40Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:40 crc kubenswrapper[4870]: I1014 07:01:40.494647 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:40 crc kubenswrapper[4870]: I1014 07:01:40.494686 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:40 crc kubenswrapper[4870]: I1014 07:01:40.494696 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:40 crc kubenswrapper[4870]: I1014 07:01:40.494714 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:40 crc kubenswrapper[4870]: I1014 07:01:40.494726 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:40Z","lastTransitionTime":"2025-10-14T07:01:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:40 crc kubenswrapper[4870]: I1014 07:01:40.498656 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-khtq8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28deeea6-a1ef-4a4a-a377-d954ccd87de9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rb79p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rb79p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-khtq8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:40Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:40 crc kubenswrapper[4870]: I1014 07:01:40.519526 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33a9493e-9660-48db-b8d1-310e29ebe313\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3570643ef1b7d5cc3a542f320413b819338e0c8b7ab7f0f09cc909997a39bab8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb398668197dedc1c2a14b4ae7d6a29714f3527309ccb9c1622af841100e1d60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1d2b6d068d63bd38cdfb83824525f6d12054f6cabab5d684e03ea5f2152076c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://783ff243e395333f3fc80cfae21358c532f0c9c5cb5515586be56b90ed10d912\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a8135a3d01076e4437ae8a784b459bf8d16de57fa16ec7ce4d40c5129c9d4a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48ca3a5017ba650df99b2fc7201c3fe5abfa89d81089c5647cde1b1549f534fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48ca3a5017ba650df99b2fc7201c3fe5abfa89d81089c5647cde1b1549f534fc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://19682eca67f76bf2e06e1b80f12e21141fcbce0acb8195f0defacd1b9a10b5e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://19682eca67f76bf2e06e1b80f12e21141fcbce0acb8195f0defacd1b9a10b5e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://042aad6d5f6bee8eef77f9e6a2c265de775a69f3da72aedc68bdf312198f00fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://042aad6d5f6bee8eef77f9e6a2c265de775a69f3da72aedc68bdf312198f00fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:05Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:40Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:40 crc kubenswrapper[4870]: I1014 07:01:40.534414 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e539fac3d0a2ade314a4b4c1050bb27c37bcff5ba5272fa2d71b97ab5c0d7191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:40Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:40 crc kubenswrapper[4870]: I1014 07:01:40.548671 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:40Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:40 crc kubenswrapper[4870]: I1014 07:01:40.559609 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-kz77f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d3864a91-9bf6-4540-a908-cc41aa39656c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d4ded3d86030f5ea548c8b1cc3caddbe14a45c9a3ae4af836e892880584f57d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4h8v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:28Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-kz77f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:40Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:40 crc kubenswrapper[4870]: I1014 07:01:40.575377 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bcbc6dd6-aab0-4dbc-a62e-c58f06ec1337\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25ec4808b09adc44488530b3c4e29fd7fec46b3fcfa31d4449ec4a6dbc880ec8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f3f5de86fb11231a19034a18e9d7f31476c162e6567d0c136ead8bbbd900d6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d57d938d9d9c92198ae5b3857536ece36012fbfbe8cc8cfd245d1386d39b0be5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77a3247a7f65709543d7fd374b2ab69d7d793537d4132d7e22261188d6e58544\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:05Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:40Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:40 crc kubenswrapper[4870]: I1014 07:01:40.590935 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3eb0b397c05ffd6650607f1be4a75683e0342d3333fc096bf1cd63c7688834f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:40Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:40 crc kubenswrapper[4870]: I1014 07:01:40.597125 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:40 crc kubenswrapper[4870]: I1014 07:01:40.597192 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:40 crc kubenswrapper[4870]: I1014 07:01:40.597209 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:40 crc kubenswrapper[4870]: I1014 07:01:40.597237 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:40 crc kubenswrapper[4870]: I1014 07:01:40.597255 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:40Z","lastTransitionTime":"2025-10-14T07:01:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:40 crc kubenswrapper[4870]: I1014 07:01:40.608184 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"874a5e45-dffd-4d17-b609-b7d3ed2eab07\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d98d044836ab44a405196915ef7bf572b75113f2c81ec7e64effa63b47ad3543\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wj4ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8004752575df4a1c44efbcafe64799f09126d0f97285d92388c76a0b90e022f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wj4ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7tvc7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:40Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:40 crc kubenswrapper[4870]: I1014 07:01:40.626132 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:40Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:40 crc kubenswrapper[4870]: I1014 07:01:40.639615 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:40Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:40 crc kubenswrapper[4870]: I1014 07:01:40.660209 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-zc5j6"] Oct 14 07:01:40 crc kubenswrapper[4870]: I1014 07:01:40.661225 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zc5j6" Oct 14 07:01:40 crc kubenswrapper[4870]: E1014 07:01:40.661355 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zc5j6" podUID="7d7a4f41-388a-4320-b896-43c17ff13da6" Oct 14 07:01:40 crc kubenswrapper[4870]: I1014 07:01:40.663541 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e26fe49ec2045d20a0cb2dd17e96e79b6eb43c641eb184c688cc472ca5db39d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aea115f7ee71c700b218fe15c928af7b55e6babe70e65d67f125d87be1fd2e2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:40Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:40 crc kubenswrapper[4870]: I1014 07:01:40.688294 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-v7br2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dcaef0e-42e3-424b-bbe5-4a023efe912d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71dec55e859b11a081dbb90d7d6cf6b4846cf591382303d8329003ff50c71f2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e72cff0723e2ce520340244b4e597d8016c58e1b071c4b087f2f6cd718741594\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc9d4db944f837289d1923176d60cfc4a566a282c2d2e3a703e220df884703c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d16ed1c65925549dcb69ab02c3013e25e3f123a83179f1af3a8e08bf91a59896\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed59243b59c39ce82bf9bc8f33c0a8b84854846afff6a89f5d37ffc97a096b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3b6d20cf459b1f6d3aff5b9290f9391909baf27ee40c49e2b0cd2d5cfe889ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c43b07bd7ecc0a98ea65b6f59752ad8651568236dc78ee216e0ca42dabcbe4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c43b07bd7ecc0a98ea65b6f59752ad8651568236dc78ee216e0ca42dabcbe4f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-14T07:01:39Z\\\",\\\"message\\\":\\\"ormers/externalversions/factory.go:141\\\\nI1014 07:01:39.675747 6159 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1014 07:01:39.675746 6159 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1014 07:01:39.675779 6159 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1014 07:01:39.675818 6159 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1014 07:01:39.675880 6159 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1014 07:01:39.676734 6159 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1014 07:01:39.676776 6159 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1014 07:01:39.676790 6159 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1014 07:01:39.676817 6159 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1014 07:01:39.676835 6159 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1014 07:01:39.676870 6159 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1014 07:01:39.676971 6159 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1014 07:01:39.677038 6159 factory.go:656] Stopping watch factory\\\\nI1014 07:01:39.677058 6159 ovnkube.go:599] Stopped ovnkube\\\\nI1014 0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fe3468cdd612fc4378e1f6f55a2a9d00274f7ddb65cb57b2a368dfc065cfb4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ba24c2e067b9f3a14d9d018b66300b54d866530d30174fbdb3c899299575d28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ba24c2e067b9f3a14d9d018b66300b54d866530d30174fbdb3c899299575d28\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-v7br2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:40Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:40 crc kubenswrapper[4870]: I1014 07:01:40.704555 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:40 crc kubenswrapper[4870]: I1014 07:01:40.704601 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:40 crc kubenswrapper[4870]: I1014 07:01:40.704615 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:40 crc kubenswrapper[4870]: I1014 07:01:40.704650 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:40 crc kubenswrapper[4870]: I1014 07:01:40.704674 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:40Z","lastTransitionTime":"2025-10-14T07:01:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:40 crc kubenswrapper[4870]: I1014 07:01:40.717821 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77a0d93a-5391-4112-b502-c474f45de209\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7815af9a584f6db6660073da289f2e8c987a04ef9ddfac155bdd9a15b014fcfa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b329b570622155a55f70501b2260e278106a68ebef497914231b4ec860e0125\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba942a0e6fe5ab93a187e0b7b4bd7c440c2e7827d9b5323fc50f90b562075392\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://978e8db184ba9e430232c6bb2ee530beb92567fdc71ba3c35b0bedca3492d845\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://978e8db184ba9e430232c6bb2ee530beb92567fdc71ba3c35b0bedca3492d845\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"file observer\\\\nW1014 07:01:25.870672 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1014 07:01:25.870795 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1014 07:01:25.871521 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-756714914/tls.crt::/tmp/serving-cert-756714914/tls.key\\\\\\\"\\\\nI1014 07:01:26.460037 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1014 07:01:26.464876 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1014 07:01:26.464902 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1014 07:01:26.464925 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1014 07:01:26.464931 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1014 07:01:26.481374 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1014 07:01:26.481400 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:01:26.481405 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:01:26.481410 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1014 07:01:26.481414 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1014 07:01:26.481417 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1014 07:01:26.481420 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1014 07:01:26.481610 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1014 07:01:26.482872 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bd0044a247bdc5ca4cab7294103af3270b76cde61f3109b85e83ed050ca53b6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82d3c2214e0be679d2827594b159ace362151318689078bc49d3158f42899636\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82d3c2214e0be679d2827594b159ace362151318689078bc49d3158f42899636\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:05Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:40Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:40 crc kubenswrapper[4870]: I1014 07:01:40.751861 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-khtq8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28deeea6-a1ef-4a4a-a377-d954ccd87de9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://296bb048a1d99649beb57d609573833bf547cd4d628ebc823c6af83fe5736c9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rb79p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2b7bc4c93307e80b778f2282ced993898f1c4cfc350a7d74828b8f9dea9906d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rb79p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-khtq8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:40Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:40 crc kubenswrapper[4870]: I1014 07:01:40.773248 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7d7a4f41-388a-4320-b896-43c17ff13da6-metrics-certs\") pod \"network-metrics-daemon-zc5j6\" (UID: \"7d7a4f41-388a-4320-b896-43c17ff13da6\") " pod="openshift-multus/network-metrics-daemon-zc5j6" Oct 14 07:01:40 crc kubenswrapper[4870]: I1014 07:01:40.773331 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6p2wh\" (UniqueName: \"kubernetes.io/projected/7d7a4f41-388a-4320-b896-43c17ff13da6-kube-api-access-6p2wh\") pod \"network-metrics-daemon-zc5j6\" (UID: \"7d7a4f41-388a-4320-b896-43c17ff13da6\") " pod="openshift-multus/network-metrics-daemon-zc5j6" Oct 14 07:01:40 crc kubenswrapper[4870]: I1014 07:01:40.784685 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33a9493e-9660-48db-b8d1-310e29ebe313\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3570643ef1b7d5cc3a542f320413b819338e0c8b7ab7f0f09cc909997a39bab8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb398668197dedc1c2a14b4ae7d6a29714f3527309ccb9c1622af841100e1d60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1d2b6d068d63bd38cdfb83824525f6d12054f6cabab5d684e03ea5f2152076c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://783ff243e395333f3fc80cfae21358c532f0c9c5cb5515586be56b90ed10d912\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a8135a3d01076e4437ae8a784b459bf8d16de57fa16ec7ce4d40c5129c9d4a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48ca3a5017ba650df99b2fc7201c3fe5abfa89d81089c5647cde1b1549f534fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48ca3a5017ba650df99b2fc7201c3fe5abfa89d81089c5647cde1b1549f534fc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://19682eca67f76bf2e06e1b80f12e21141fcbce0acb8195f0defacd1b9a10b5e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://19682eca67f76bf2e06e1b80f12e21141fcbce0acb8195f0defacd1b9a10b5e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://042aad6d5f6bee8eef77f9e6a2c265de775a69f3da72aedc68bdf312198f00fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://042aad6d5f6bee8eef77f9e6a2c265de775a69f3da72aedc68bdf312198f00fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:05Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:40Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:40 crc kubenswrapper[4870]: I1014 07:01:40.795474 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-4v2v2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4796ba6-c23b-4764-be96-2e3a3bf5e962\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf5afb0b99a7101f6d5573fb7eb925c5a1d8becc11b2e3c6f3351d80b6f0e823\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnp8b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-4v2v2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:40Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:40 crc kubenswrapper[4870]: I1014 07:01:40.807284 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:40 crc kubenswrapper[4870]: I1014 07:01:40.807339 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:40 crc kubenswrapper[4870]: I1014 07:01:40.807356 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:40 crc kubenswrapper[4870]: I1014 07:01:40.807381 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:40 crc kubenswrapper[4870]: I1014 07:01:40.807402 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:40Z","lastTransitionTime":"2025-10-14T07:01:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:40 crc kubenswrapper[4870]: I1014 07:01:40.809260 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-cdjmj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"612e7ab9-bda4-45d8-bc50-ff1afb68aace\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7776352d62b814fc63d01e4c16ad13f878b0597eb37a370bf9da0a88abab6cfa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://834444b5350ee12df4e4652c3009880d8d69e1edf8ba0c0156ab96fb2bbe28e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://834444b5350ee12df4e4652c3009880d8d69e1edf8ba0c0156ab96fb2bbe28e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c1972e7b2f75790db22a07a078eee97072bffbdb7f8e594bf2be5f800632a7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c1972e7b2f75790db22a07a078eee97072bffbdb7f8e594bf2be5f800632a7b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b124b42a44c6a23da1d92b9a5326987eb74430f80489a71e8913829475654ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b124b42a44c6a23da1d92b9a5326987eb74430f80489a71e8913829475654ab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62d0deb6b20ea65fbcbaf94c2fc14d69adb22e7a46cc043ccd4def2668ddd4b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62d0deb6b20ea65fbcbaf94c2fc14d69adb22e7a46cc043ccd4def2668ddd4b7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7f0d7cf29ca737d0effe58731a8b890379527c97f78c144d4eebb910260fadb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7f0d7cf29ca737d0effe58731a8b890379527c97f78c144d4eebb910260fadb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5959abb7c8e7ab32bed3aca709ae0334e8400f991ba724246c9fda56cbed822\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5959abb7c8e7ab32bed3aca709ae0334e8400f991ba724246c9fda56cbed822\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-cdjmj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:40Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:40 crc kubenswrapper[4870]: I1014 07:01:40.820463 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fdxjh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a3a430a-b540-4a8c-adad-b2893dbf9898\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93e41d5e927eed7ec62eb9c3607cc804cd8999556663363fd123b2a784b00ff1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k787j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fdxjh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:40Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:40 crc kubenswrapper[4870]: I1014 07:01:40.839297 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bcbc6dd6-aab0-4dbc-a62e-c58f06ec1337\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25ec4808b09adc44488530b3c4e29fd7fec46b3fcfa31d4449ec4a6dbc880ec8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f3f5de86fb11231a19034a18e9d7f31476c162e6567d0c136ead8bbbd900d6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d57d938d9d9c92198ae5b3857536ece36012fbfbe8cc8cfd245d1386d39b0be5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77a3247a7f65709543d7fd374b2ab69d7d793537d4132d7e22261188d6e58544\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:05Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:40Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:40 crc kubenswrapper[4870]: I1014 07:01:40.861637 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e539fac3d0a2ade314a4b4c1050bb27c37bcff5ba5272fa2d71b97ab5c0d7191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:40Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:40 crc kubenswrapper[4870]: I1014 07:01:40.874725 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6p2wh\" (UniqueName: \"kubernetes.io/projected/7d7a4f41-388a-4320-b896-43c17ff13da6-kube-api-access-6p2wh\") pod \"network-metrics-daemon-zc5j6\" (UID: \"7d7a4f41-388a-4320-b896-43c17ff13da6\") " pod="openshift-multus/network-metrics-daemon-zc5j6" Oct 14 07:01:40 crc kubenswrapper[4870]: I1014 07:01:40.874833 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7d7a4f41-388a-4320-b896-43c17ff13da6-metrics-certs\") pod \"network-metrics-daemon-zc5j6\" (UID: \"7d7a4f41-388a-4320-b896-43c17ff13da6\") " pod="openshift-multus/network-metrics-daemon-zc5j6" Oct 14 07:01:40 crc kubenswrapper[4870]: E1014 07:01:40.875126 4870 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 14 07:01:40 crc kubenswrapper[4870]: E1014 07:01:40.875287 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7d7a4f41-388a-4320-b896-43c17ff13da6-metrics-certs podName:7d7a4f41-388a-4320-b896-43c17ff13da6 nodeName:}" failed. No retries permitted until 2025-10-14 07:01:41.375255129 +0000 UTC m=+37.072615520 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/7d7a4f41-388a-4320-b896-43c17ff13da6-metrics-certs") pod "network-metrics-daemon-zc5j6" (UID: "7d7a4f41-388a-4320-b896-43c17ff13da6") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 14 07:01:40 crc kubenswrapper[4870]: I1014 07:01:40.875560 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:40Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:40 crc kubenswrapper[4870]: I1014 07:01:40.888984 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-kz77f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d3864a91-9bf6-4540-a908-cc41aa39656c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d4ded3d86030f5ea548c8b1cc3caddbe14a45c9a3ae4af836e892880584f57d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4h8v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:28Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-kz77f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:40Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:40 crc kubenswrapper[4870]: I1014 07:01:40.895459 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6p2wh\" (UniqueName: \"kubernetes.io/projected/7d7a4f41-388a-4320-b896-43c17ff13da6-kube-api-access-6p2wh\") pod \"network-metrics-daemon-zc5j6\" (UID: \"7d7a4f41-388a-4320-b896-43c17ff13da6\") " pod="openshift-multus/network-metrics-daemon-zc5j6" Oct 14 07:01:40 crc kubenswrapper[4870]: I1014 07:01:40.905534 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:40Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:40 crc kubenswrapper[4870]: I1014 07:01:40.910009 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:40 crc kubenswrapper[4870]: I1014 07:01:40.910048 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:40 crc kubenswrapper[4870]: I1014 07:01:40.910059 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:40 crc kubenswrapper[4870]: I1014 07:01:40.910076 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:40 crc kubenswrapper[4870]: I1014 07:01:40.910086 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:40Z","lastTransitionTime":"2025-10-14T07:01:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:40 crc kubenswrapper[4870]: I1014 07:01:40.918910 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3eb0b397c05ffd6650607f1be4a75683e0342d3333fc096bf1cd63c7688834f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:40Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:40 crc kubenswrapper[4870]: I1014 07:01:40.935981 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"874a5e45-dffd-4d17-b609-b7d3ed2eab07\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d98d044836ab44a405196915ef7bf572b75113f2c81ec7e64effa63b47ad3543\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wj4ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8004752575df4a1c44efbcafe64799f09126d0f97285d92388c76a0b90e022f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wj4ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7tvc7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:40Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:40 crc kubenswrapper[4870]: I1014 07:01:40.949198 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zc5j6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d7a4f41-388a-4320-b896-43c17ff13da6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6p2wh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6p2wh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:40Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zc5j6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:40Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:40 crc kubenswrapper[4870]: I1014 07:01:40.968884 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77a0d93a-5391-4112-b502-c474f45de209\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7815af9a584f6db6660073da289f2e8c987a04ef9ddfac155bdd9a15b014fcfa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b329b570622155a55f70501b2260e278106a68ebef497914231b4ec860e0125\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba942a0e6fe5ab93a187e0b7b4bd7c440c2e7827d9b5323fc50f90b562075392\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://978e8db184ba9e430232c6bb2ee530beb92567fdc71ba3c35b0bedca3492d845\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://978e8db184ba9e430232c6bb2ee530beb92567fdc71ba3c35b0bedca3492d845\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"file observer\\\\nW1014 07:01:25.870672 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1014 07:01:25.870795 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1014 07:01:25.871521 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-756714914/tls.crt::/tmp/serving-cert-756714914/tls.key\\\\\\\"\\\\nI1014 07:01:26.460037 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1014 07:01:26.464876 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1014 07:01:26.464902 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1014 07:01:26.464925 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1014 07:01:26.464931 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1014 07:01:26.481374 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1014 07:01:26.481400 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:01:26.481405 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:01:26.481410 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1014 07:01:26.481414 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1014 07:01:26.481417 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1014 07:01:26.481420 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1014 07:01:26.481610 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1014 07:01:26.482872 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bd0044a247bdc5ca4cab7294103af3270b76cde61f3109b85e83ed050ca53b6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82d3c2214e0be679d2827594b159ace362151318689078bc49d3158f42899636\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82d3c2214e0be679d2827594b159ace362151318689078bc49d3158f42899636\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:05Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:40Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:40 crc kubenswrapper[4870]: I1014 07:01:40.982352 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:40Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:41 crc kubenswrapper[4870]: I1014 07:01:41.000820 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e26fe49ec2045d20a0cb2dd17e96e79b6eb43c641eb184c688cc472ca5db39d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aea115f7ee71c700b218fe15c928af7b55e6babe70e65d67f125d87be1fd2e2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:40Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:41 crc kubenswrapper[4870]: I1014 07:01:41.012776 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:41 crc kubenswrapper[4870]: I1014 07:01:41.013223 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:41 crc kubenswrapper[4870]: I1014 07:01:41.013232 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:41 crc kubenswrapper[4870]: I1014 07:01:41.013248 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:41 crc kubenswrapper[4870]: I1014 07:01:41.013259 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:41Z","lastTransitionTime":"2025-10-14T07:01:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:41 crc kubenswrapper[4870]: I1014 07:01:41.030799 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-v7br2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dcaef0e-42e3-424b-bbe5-4a023efe912d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71dec55e859b11a081dbb90d7d6cf6b4846cf591382303d8329003ff50c71f2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e72cff0723e2ce520340244b4e597d8016c58e1b071c4b087f2f6cd718741594\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc9d4db944f837289d1923176d60cfc4a566a282c2d2e3a703e220df884703c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d16ed1c65925549dcb69ab02c3013e25e3f123a83179f1af3a8e08bf91a59896\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed59243b59c39ce82bf9bc8f33c0a8b84854846afff6a89f5d37ffc97a096b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3b6d20cf459b1f6d3aff5b9290f9391909baf27ee40c49e2b0cd2d5cfe889ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c43b07bd7ecc0a98ea65b6f59752ad8651568236dc78ee216e0ca42dabcbe4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c43b07bd7ecc0a98ea65b6f59752ad8651568236dc78ee216e0ca42dabcbe4f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-14T07:01:39Z\\\",\\\"message\\\":\\\"ormers/externalversions/factory.go:141\\\\nI1014 07:01:39.675747 6159 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1014 07:01:39.675746 6159 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1014 07:01:39.675779 6159 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1014 07:01:39.675818 6159 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1014 07:01:39.675880 6159 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1014 07:01:39.676734 6159 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1014 07:01:39.676776 6159 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1014 07:01:39.676790 6159 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1014 07:01:39.676817 6159 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1014 07:01:39.676835 6159 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1014 07:01:39.676870 6159 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1014 07:01:39.676971 6159 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1014 07:01:39.677038 6159 factory.go:656] Stopping watch factory\\\\nI1014 07:01:39.677058 6159 ovnkube.go:599] Stopped ovnkube\\\\nI1014 0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fe3468cdd612fc4378e1f6f55a2a9d00274f7ddb65cb57b2a368dfc065cfb4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ba24c2e067b9f3a14d9d018b66300b54d866530d30174fbdb3c899299575d28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ba24c2e067b9f3a14d9d018b66300b54d866530d30174fbdb3c899299575d28\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-v7br2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:41Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:41 crc kubenswrapper[4870]: I1014 07:01:41.116290 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:41 crc kubenswrapper[4870]: I1014 07:01:41.116345 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:41 crc kubenswrapper[4870]: I1014 07:01:41.116363 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:41 crc kubenswrapper[4870]: I1014 07:01:41.116389 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:41 crc kubenswrapper[4870]: I1014 07:01:41.116403 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:41Z","lastTransitionTime":"2025-10-14T07:01:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:41 crc kubenswrapper[4870]: I1014 07:01:41.219718 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:41 crc kubenswrapper[4870]: I1014 07:01:41.219789 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:41 crc kubenswrapper[4870]: I1014 07:01:41.219806 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:41 crc kubenswrapper[4870]: I1014 07:01:41.219837 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:41 crc kubenswrapper[4870]: I1014 07:01:41.219856 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:41Z","lastTransitionTime":"2025-10-14T07:01:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:41 crc kubenswrapper[4870]: I1014 07:01:41.322563 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:41 crc kubenswrapper[4870]: I1014 07:01:41.322635 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:41 crc kubenswrapper[4870]: I1014 07:01:41.322655 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:41 crc kubenswrapper[4870]: I1014 07:01:41.322688 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:41 crc kubenswrapper[4870]: I1014 07:01:41.322717 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:41Z","lastTransitionTime":"2025-10-14T07:01:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:41 crc kubenswrapper[4870]: I1014 07:01:41.380582 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7d7a4f41-388a-4320-b896-43c17ff13da6-metrics-certs\") pod \"network-metrics-daemon-zc5j6\" (UID: \"7d7a4f41-388a-4320-b896-43c17ff13da6\") " pod="openshift-multus/network-metrics-daemon-zc5j6" Oct 14 07:01:41 crc kubenswrapper[4870]: E1014 07:01:41.380748 4870 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 14 07:01:41 crc kubenswrapper[4870]: E1014 07:01:41.380802 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7d7a4f41-388a-4320-b896-43c17ff13da6-metrics-certs podName:7d7a4f41-388a-4320-b896-43c17ff13da6 nodeName:}" failed. No retries permitted until 2025-10-14 07:01:42.380789456 +0000 UTC m=+38.078149827 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/7d7a4f41-388a-4320-b896-43c17ff13da6-metrics-certs") pod "network-metrics-daemon-zc5j6" (UID: "7d7a4f41-388a-4320-b896-43c17ff13da6") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 14 07:01:41 crc kubenswrapper[4870]: I1014 07:01:41.432942 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:41 crc kubenswrapper[4870]: I1014 07:01:41.432977 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:41 crc kubenswrapper[4870]: I1014 07:01:41.432986 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:41 crc kubenswrapper[4870]: I1014 07:01:41.433002 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:41 crc kubenswrapper[4870]: I1014 07:01:41.433013 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:41Z","lastTransitionTime":"2025-10-14T07:01:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:41 crc kubenswrapper[4870]: I1014 07:01:41.443198 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-v7br2_8dcaef0e-42e3-424b-bbe5-4a023efe912d/ovnkube-controller/0.log" Oct 14 07:01:41 crc kubenswrapper[4870]: I1014 07:01:41.446106 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-v7br2" event={"ID":"8dcaef0e-42e3-424b-bbe5-4a023efe912d","Type":"ContainerStarted","Data":"a05f64fdac0425bbf7ea39d9d9975279c4a68dc9a93e27f58f2d7017a328f428"} Oct 14 07:01:41 crc kubenswrapper[4870]: I1014 07:01:41.446524 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-v7br2" Oct 14 07:01:41 crc kubenswrapper[4870]: I1014 07:01:41.461288 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:41Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:41 crc kubenswrapper[4870]: I1014 07:01:41.473394 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-kz77f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d3864a91-9bf6-4540-a908-cc41aa39656c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d4ded3d86030f5ea548c8b1cc3caddbe14a45c9a3ae4af836e892880584f57d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4h8v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:28Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-kz77f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:41Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:41 crc kubenswrapper[4870]: I1014 07:01:41.488418 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bcbc6dd6-aab0-4dbc-a62e-c58f06ec1337\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25ec4808b09adc44488530b3c4e29fd7fec46b3fcfa31d4449ec4a6dbc880ec8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f3f5de86fb11231a19034a18e9d7f31476c162e6567d0c136ead8bbbd900d6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d57d938d9d9c92198ae5b3857536ece36012fbfbe8cc8cfd245d1386d39b0be5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77a3247a7f65709543d7fd374b2ab69d7d793537d4132d7e22261188d6e58544\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:05Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:41Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:41 crc kubenswrapper[4870]: I1014 07:01:41.501616 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e539fac3d0a2ade314a4b4c1050bb27c37bcff5ba5272fa2d71b97ab5c0d7191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:41Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:41 crc kubenswrapper[4870]: I1014 07:01:41.511692 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"874a5e45-dffd-4d17-b609-b7d3ed2eab07\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d98d044836ab44a405196915ef7bf572b75113f2c81ec7e64effa63b47ad3543\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wj4ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8004752575df4a1c44efbcafe64799f09126d0f97285d92388c76a0b90e022f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wj4ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7tvc7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:41Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:41 crc kubenswrapper[4870]: I1014 07:01:41.523309 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zc5j6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d7a4f41-388a-4320-b896-43c17ff13da6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6p2wh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6p2wh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:40Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zc5j6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:41Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:41 crc kubenswrapper[4870]: I1014 07:01:41.535980 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:41 crc kubenswrapper[4870]: I1014 07:01:41.536037 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:41 crc kubenswrapper[4870]: I1014 07:01:41.536053 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:41 crc kubenswrapper[4870]: I1014 07:01:41.536078 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:41 crc kubenswrapper[4870]: I1014 07:01:41.536093 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:41Z","lastTransitionTime":"2025-10-14T07:01:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:41 crc kubenswrapper[4870]: I1014 07:01:41.538180 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:41Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:41 crc kubenswrapper[4870]: I1014 07:01:41.551894 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3eb0b397c05ffd6650607f1be4a75683e0342d3333fc096bf1cd63c7688834f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:41Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:41 crc kubenswrapper[4870]: I1014 07:01:41.564958 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e26fe49ec2045d20a0cb2dd17e96e79b6eb43c641eb184c688cc472ca5db39d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aea115f7ee71c700b218fe15c928af7b55e6babe70e65d67f125d87be1fd2e2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:41Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:41 crc kubenswrapper[4870]: I1014 07:01:41.583404 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-v7br2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dcaef0e-42e3-424b-bbe5-4a023efe912d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71dec55e859b11a081dbb90d7d6cf6b4846cf591382303d8329003ff50c71f2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e72cff0723e2ce520340244b4e597d8016c58e1b071c4b087f2f6cd718741594\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc9d4db944f837289d1923176d60cfc4a566a282c2d2e3a703e220df884703c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d16ed1c65925549dcb69ab02c3013e25e3f123a83179f1af3a8e08bf91a59896\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed59243b59c39ce82bf9bc8f33c0a8b84854846afff6a89f5d37ffc97a096b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3b6d20cf459b1f6d3aff5b9290f9391909baf27ee40c49e2b0cd2d5cfe889ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a05f64fdac0425bbf7ea39d9d9975279c4a68dc9a93e27f58f2d7017a328f428\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c43b07bd7ecc0a98ea65b6f59752ad8651568236dc78ee216e0ca42dabcbe4f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-14T07:01:39Z\\\",\\\"message\\\":\\\"ormers/externalversions/factory.go:141\\\\nI1014 07:01:39.675747 6159 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1014 07:01:39.675746 6159 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1014 07:01:39.675779 6159 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1014 07:01:39.675818 6159 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1014 07:01:39.675880 6159 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1014 07:01:39.676734 6159 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1014 07:01:39.676776 6159 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1014 07:01:39.676790 6159 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1014 07:01:39.676817 6159 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1014 07:01:39.676835 6159 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1014 07:01:39.676870 6159 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1014 07:01:39.676971 6159 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1014 07:01:39.677038 6159 factory.go:656] Stopping watch factory\\\\nI1014 07:01:39.677058 6159 ovnkube.go:599] Stopped ovnkube\\\\nI1014 0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:33Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fe3468cdd612fc4378e1f6f55a2a9d00274f7ddb65cb57b2a368dfc065cfb4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ba24c2e067b9f3a14d9d018b66300b54d866530d30174fbdb3c899299575d28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ba24c2e067b9f3a14d9d018b66300b54d866530d30174fbdb3c899299575d28\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-v7br2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:41Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:41 crc kubenswrapper[4870]: I1014 07:01:41.600388 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77a0d93a-5391-4112-b502-c474f45de209\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7815af9a584f6db6660073da289f2e8c987a04ef9ddfac155bdd9a15b014fcfa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b329b570622155a55f70501b2260e278106a68ebef497914231b4ec860e0125\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba942a0e6fe5ab93a187e0b7b4bd7c440c2e7827d9b5323fc50f90b562075392\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://978e8db184ba9e430232c6bb2ee530beb92567fdc71ba3c35b0bedca3492d845\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://978e8db184ba9e430232c6bb2ee530beb92567fdc71ba3c35b0bedca3492d845\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"file observer\\\\nW1014 07:01:25.870672 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1014 07:01:25.870795 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1014 07:01:25.871521 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-756714914/tls.crt::/tmp/serving-cert-756714914/tls.key\\\\\\\"\\\\nI1014 07:01:26.460037 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1014 07:01:26.464876 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1014 07:01:26.464902 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1014 07:01:26.464925 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1014 07:01:26.464931 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1014 07:01:26.481374 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1014 07:01:26.481400 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:01:26.481405 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:01:26.481410 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1014 07:01:26.481414 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1014 07:01:26.481417 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1014 07:01:26.481420 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1014 07:01:26.481610 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1014 07:01:26.482872 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bd0044a247bdc5ca4cab7294103af3270b76cde61f3109b85e83ed050ca53b6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82d3c2214e0be679d2827594b159ace362151318689078bc49d3158f42899636\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82d3c2214e0be679d2827594b159ace362151318689078bc49d3158f42899636\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:05Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:41Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:41 crc kubenswrapper[4870]: I1014 07:01:41.616059 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:41Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:41 crc kubenswrapper[4870]: I1014 07:01:41.630060 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-4v2v2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4796ba6-c23b-4764-be96-2e3a3bf5e962\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf5afb0b99a7101f6d5573fb7eb925c5a1d8becc11b2e3c6f3351d80b6f0e823\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnp8b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-4v2v2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:41Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:41 crc kubenswrapper[4870]: I1014 07:01:41.638461 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:41 crc kubenswrapper[4870]: I1014 07:01:41.638489 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:41 crc kubenswrapper[4870]: I1014 07:01:41.638501 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:41 crc kubenswrapper[4870]: I1014 07:01:41.638517 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:41 crc kubenswrapper[4870]: I1014 07:01:41.638526 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:41Z","lastTransitionTime":"2025-10-14T07:01:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:41 crc kubenswrapper[4870]: I1014 07:01:41.650815 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-cdjmj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"612e7ab9-bda4-45d8-bc50-ff1afb68aace\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7776352d62b814fc63d01e4c16ad13f878b0597eb37a370bf9da0a88abab6cfa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://834444b5350ee12df4e4652c3009880d8d69e1edf8ba0c0156ab96fb2bbe28e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://834444b5350ee12df4e4652c3009880d8d69e1edf8ba0c0156ab96fb2bbe28e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c1972e7b2f75790db22a07a078eee97072bffbdb7f8e594bf2be5f800632a7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c1972e7b2f75790db22a07a078eee97072bffbdb7f8e594bf2be5f800632a7b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b124b42a44c6a23da1d92b9a5326987eb74430f80489a71e8913829475654ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b124b42a44c6a23da1d92b9a5326987eb74430f80489a71e8913829475654ab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62d0deb6b20ea65fbcbaf94c2fc14d69adb22e7a46cc043ccd4def2668ddd4b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62d0deb6b20ea65fbcbaf94c2fc14d69adb22e7a46cc043ccd4def2668ddd4b7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7f0d7cf29ca737d0effe58731a8b890379527c97f78c144d4eebb910260fadb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7f0d7cf29ca737d0effe58731a8b890379527c97f78c144d4eebb910260fadb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5959abb7c8e7ab32bed3aca709ae0334e8400f991ba724246c9fda56cbed822\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5959abb7c8e7ab32bed3aca709ae0334e8400f991ba724246c9fda56cbed822\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-cdjmj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:41Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:41 crc kubenswrapper[4870]: I1014 07:01:41.665610 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fdxjh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a3a430a-b540-4a8c-adad-b2893dbf9898\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93e41d5e927eed7ec62eb9c3607cc804cd8999556663363fd123b2a784b00ff1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k787j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fdxjh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:41Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:41 crc kubenswrapper[4870]: I1014 07:01:41.687235 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-khtq8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28deeea6-a1ef-4a4a-a377-d954ccd87de9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://296bb048a1d99649beb57d609573833bf547cd4d628ebc823c6af83fe5736c9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rb79p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2b7bc4c93307e80b778f2282ced993898f1c4cfc350a7d74828b8f9dea9906d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rb79p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-khtq8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:41Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:41 crc kubenswrapper[4870]: I1014 07:01:41.709232 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33a9493e-9660-48db-b8d1-310e29ebe313\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3570643ef1b7d5cc3a542f320413b819338e0c8b7ab7f0f09cc909997a39bab8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb398668197dedc1c2a14b4ae7d6a29714f3527309ccb9c1622af841100e1d60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1d2b6d068d63bd38cdfb83824525f6d12054f6cabab5d684e03ea5f2152076c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://783ff243e395333f3fc80cfae21358c532f0c9c5cb5515586be56b90ed10d912\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a8135a3d01076e4437ae8a784b459bf8d16de57fa16ec7ce4d40c5129c9d4a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48ca3a5017ba650df99b2fc7201c3fe5abfa89d81089c5647cde1b1549f534fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48ca3a5017ba650df99b2fc7201c3fe5abfa89d81089c5647cde1b1549f534fc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://19682eca67f76bf2e06e1b80f12e21141fcbce0acb8195f0defacd1b9a10b5e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://19682eca67f76bf2e06e1b80f12e21141fcbce0acb8195f0defacd1b9a10b5e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://042aad6d5f6bee8eef77f9e6a2c265de775a69f3da72aedc68bdf312198f00fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://042aad6d5f6bee8eef77f9e6a2c265de775a69f3da72aedc68bdf312198f00fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:05Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:41Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:41 crc kubenswrapper[4870]: I1014 07:01:41.744135 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:41 crc kubenswrapper[4870]: I1014 07:01:41.744186 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:41 crc kubenswrapper[4870]: I1014 07:01:41.744205 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:41 crc kubenswrapper[4870]: I1014 07:01:41.744224 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:41 crc kubenswrapper[4870]: I1014 07:01:41.744236 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:41Z","lastTransitionTime":"2025-10-14T07:01:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:41 crc kubenswrapper[4870]: I1014 07:01:41.785564 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 07:01:41 crc kubenswrapper[4870]: E1014 07:01:41.785966 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 07:01:57.785911008 +0000 UTC m=+53.483271519 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:01:41 crc kubenswrapper[4870]: I1014 07:01:41.847929 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:41 crc kubenswrapper[4870]: I1014 07:01:41.848025 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:41 crc kubenswrapper[4870]: I1014 07:01:41.848037 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:41 crc kubenswrapper[4870]: I1014 07:01:41.848053 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:41 crc kubenswrapper[4870]: I1014 07:01:41.848062 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:41Z","lastTransitionTime":"2025-10-14T07:01:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:41 crc kubenswrapper[4870]: I1014 07:01:41.887882 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 07:01:41 crc kubenswrapper[4870]: E1014 07:01:41.887964 4870 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 14 07:01:41 crc kubenswrapper[4870]: I1014 07:01:41.888044 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 07:01:41 crc kubenswrapper[4870]: E1014 07:01:41.888110 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-14 07:01:57.888076565 +0000 UTC m=+53.585437176 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 14 07:01:41 crc kubenswrapper[4870]: I1014 07:01:41.888206 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 07:01:41 crc kubenswrapper[4870]: I1014 07:01:41.888250 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 07:01:41 crc kubenswrapper[4870]: E1014 07:01:41.888252 4870 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 14 07:01:41 crc kubenswrapper[4870]: E1014 07:01:41.888352 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-14 07:01:57.888326081 +0000 UTC m=+53.585686492 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 14 07:01:41 crc kubenswrapper[4870]: E1014 07:01:41.888464 4870 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 14 07:01:41 crc kubenswrapper[4870]: E1014 07:01:41.888502 4870 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 14 07:01:41 crc kubenswrapper[4870]: E1014 07:01:41.888523 4870 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 14 07:01:41 crc kubenswrapper[4870]: E1014 07:01:41.888613 4870 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 14 07:01:41 crc kubenswrapper[4870]: E1014 07:01:41.888651 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-14 07:01:57.888638349 +0000 UTC m=+53.585998950 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 14 07:01:41 crc kubenswrapper[4870]: E1014 07:01:41.888662 4870 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 14 07:01:41 crc kubenswrapper[4870]: E1014 07:01:41.888699 4870 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 14 07:01:41 crc kubenswrapper[4870]: E1014 07:01:41.888771 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-14 07:01:57.888753842 +0000 UTC m=+53.586114253 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 14 07:01:41 crc kubenswrapper[4870]: I1014 07:01:41.950460 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:41 crc kubenswrapper[4870]: I1014 07:01:41.950510 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:41 crc kubenswrapper[4870]: I1014 07:01:41.950524 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:41 crc kubenswrapper[4870]: I1014 07:01:41.950542 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:41 crc kubenswrapper[4870]: I1014 07:01:41.950553 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:41Z","lastTransitionTime":"2025-10-14T07:01:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:42 crc kubenswrapper[4870]: I1014 07:01:42.033503 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 07:01:42 crc kubenswrapper[4870]: I1014 07:01:42.033552 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 07:01:42 crc kubenswrapper[4870]: I1014 07:01:42.033690 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 07:01:42 crc kubenswrapper[4870]: E1014 07:01:42.033798 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 07:01:42 crc kubenswrapper[4870]: E1014 07:01:42.033939 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 07:01:42 crc kubenswrapper[4870]: I1014 07:01:42.034025 4870 scope.go:117] "RemoveContainer" containerID="978e8db184ba9e430232c6bb2ee530beb92567fdc71ba3c35b0bedca3492d845" Oct 14 07:01:42 crc kubenswrapper[4870]: E1014 07:01:42.034162 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 07:01:42 crc kubenswrapper[4870]: I1014 07:01:42.054950 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:42 crc kubenswrapper[4870]: I1014 07:01:42.054998 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:42 crc kubenswrapper[4870]: I1014 07:01:42.055009 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:42 crc kubenswrapper[4870]: I1014 07:01:42.055024 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:42 crc kubenswrapper[4870]: I1014 07:01:42.055036 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:42Z","lastTransitionTime":"2025-10-14T07:01:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:42 crc kubenswrapper[4870]: I1014 07:01:42.157697 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:42 crc kubenswrapper[4870]: I1014 07:01:42.157727 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:42 crc kubenswrapper[4870]: I1014 07:01:42.157738 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:42 crc kubenswrapper[4870]: I1014 07:01:42.157753 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:42 crc kubenswrapper[4870]: I1014 07:01:42.157762 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:42Z","lastTransitionTime":"2025-10-14T07:01:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:42 crc kubenswrapper[4870]: I1014 07:01:42.260347 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:42 crc kubenswrapper[4870]: I1014 07:01:42.260402 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:42 crc kubenswrapper[4870]: I1014 07:01:42.260418 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:42 crc kubenswrapper[4870]: I1014 07:01:42.260471 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:42 crc kubenswrapper[4870]: I1014 07:01:42.260491 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:42Z","lastTransitionTime":"2025-10-14T07:01:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:42 crc kubenswrapper[4870]: I1014 07:01:42.362764 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:42 crc kubenswrapper[4870]: I1014 07:01:42.362799 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:42 crc kubenswrapper[4870]: I1014 07:01:42.362808 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:42 crc kubenswrapper[4870]: I1014 07:01:42.362821 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:42 crc kubenswrapper[4870]: I1014 07:01:42.362831 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:42Z","lastTransitionTime":"2025-10-14T07:01:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:42 crc kubenswrapper[4870]: I1014 07:01:42.393691 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7d7a4f41-388a-4320-b896-43c17ff13da6-metrics-certs\") pod \"network-metrics-daemon-zc5j6\" (UID: \"7d7a4f41-388a-4320-b896-43c17ff13da6\") " pod="openshift-multus/network-metrics-daemon-zc5j6" Oct 14 07:01:42 crc kubenswrapper[4870]: E1014 07:01:42.393810 4870 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 14 07:01:42 crc kubenswrapper[4870]: E1014 07:01:42.393857 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7d7a4f41-388a-4320-b896-43c17ff13da6-metrics-certs podName:7d7a4f41-388a-4320-b896-43c17ff13da6 nodeName:}" failed. No retries permitted until 2025-10-14 07:01:44.393844728 +0000 UTC m=+40.091205089 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/7d7a4f41-388a-4320-b896-43c17ff13da6-metrics-certs") pod "network-metrics-daemon-zc5j6" (UID: "7d7a4f41-388a-4320-b896-43c17ff13da6") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 14 07:01:42 crc kubenswrapper[4870]: I1014 07:01:42.454793 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Oct 14 07:01:42 crc kubenswrapper[4870]: I1014 07:01:42.456758 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"5b9018b078c5c8276243c1696860f0aa73e2d8dd917913ae9d03b30dc9b5113a"} Oct 14 07:01:42 crc kubenswrapper[4870]: I1014 07:01:42.457084 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 14 07:01:42 crc kubenswrapper[4870]: I1014 07:01:42.458518 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-v7br2_8dcaef0e-42e3-424b-bbe5-4a023efe912d/ovnkube-controller/1.log" Oct 14 07:01:42 crc kubenswrapper[4870]: I1014 07:01:42.459135 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-v7br2_8dcaef0e-42e3-424b-bbe5-4a023efe912d/ovnkube-controller/0.log" Oct 14 07:01:42 crc kubenswrapper[4870]: I1014 07:01:42.462019 4870 generic.go:334] "Generic (PLEG): container finished" podID="8dcaef0e-42e3-424b-bbe5-4a023efe912d" containerID="a05f64fdac0425bbf7ea39d9d9975279c4a68dc9a93e27f58f2d7017a328f428" exitCode=1 Oct 14 07:01:42 crc kubenswrapper[4870]: I1014 07:01:42.462050 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-v7br2" event={"ID":"8dcaef0e-42e3-424b-bbe5-4a023efe912d","Type":"ContainerDied","Data":"a05f64fdac0425bbf7ea39d9d9975279c4a68dc9a93e27f58f2d7017a328f428"} Oct 14 07:01:42 crc kubenswrapper[4870]: I1014 07:01:42.462077 4870 scope.go:117] "RemoveContainer" containerID="7c43b07bd7ecc0a98ea65b6f59752ad8651568236dc78ee216e0ca42dabcbe4f" Oct 14 07:01:42 crc kubenswrapper[4870]: I1014 07:01:42.462778 4870 scope.go:117] "RemoveContainer" containerID="a05f64fdac0425bbf7ea39d9d9975279c4a68dc9a93e27f58f2d7017a328f428" Oct 14 07:01:42 crc kubenswrapper[4870]: E1014 07:01:42.462953 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-v7br2_openshift-ovn-kubernetes(8dcaef0e-42e3-424b-bbe5-4a023efe912d)\"" pod="openshift-ovn-kubernetes/ovnkube-node-v7br2" podUID="8dcaef0e-42e3-424b-bbe5-4a023efe912d" Oct 14 07:01:42 crc kubenswrapper[4870]: I1014 07:01:42.470023 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:42 crc kubenswrapper[4870]: I1014 07:01:42.470051 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:42 crc kubenswrapper[4870]: I1014 07:01:42.470061 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:42 crc kubenswrapper[4870]: I1014 07:01:42.470075 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:42 crc kubenswrapper[4870]: I1014 07:01:42.470085 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:42Z","lastTransitionTime":"2025-10-14T07:01:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:42 crc kubenswrapper[4870]: I1014 07:01:42.475751 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bcbc6dd6-aab0-4dbc-a62e-c58f06ec1337\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25ec4808b09adc44488530b3c4e29fd7fec46b3fcfa31d4449ec4a6dbc880ec8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f3f5de86fb11231a19034a18e9d7f31476c162e6567d0c136ead8bbbd900d6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d57d938d9d9c92198ae5b3857536ece36012fbfbe8cc8cfd245d1386d39b0be5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77a3247a7f65709543d7fd374b2ab69d7d793537d4132d7e22261188d6e58544\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:05Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:42Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:42 crc kubenswrapper[4870]: I1014 07:01:42.476397 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:42 crc kubenswrapper[4870]: I1014 07:01:42.476420 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:42 crc kubenswrapper[4870]: I1014 07:01:42.476429 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:42 crc kubenswrapper[4870]: I1014 07:01:42.476451 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:42 crc kubenswrapper[4870]: I1014 07:01:42.476462 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:42Z","lastTransitionTime":"2025-10-14T07:01:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:42 crc kubenswrapper[4870]: I1014 07:01:42.487485 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e539fac3d0a2ade314a4b4c1050bb27c37bcff5ba5272fa2d71b97ab5c0d7191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:42Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:42 crc kubenswrapper[4870]: E1014 07:01:42.489072 4870 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:01:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:01:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:42Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:01:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:01:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:42Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"56448e6b-6b3a-4bb9-91f8-2f0a9b44e9aa\\\",\\\"systemUUID\\\":\\\"0adb347d-b90d-4e9f-9962-f1bd97980aed\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:42Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:42 crc kubenswrapper[4870]: I1014 07:01:42.493276 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:42 crc kubenswrapper[4870]: I1014 07:01:42.493318 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:42 crc kubenswrapper[4870]: I1014 07:01:42.493329 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:42 crc kubenswrapper[4870]: I1014 07:01:42.493346 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:42 crc kubenswrapper[4870]: I1014 07:01:42.493358 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:42Z","lastTransitionTime":"2025-10-14T07:01:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:42 crc kubenswrapper[4870]: I1014 07:01:42.503732 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:42Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:42 crc kubenswrapper[4870]: E1014 07:01:42.511835 4870 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:01:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:01:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:42Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:01:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:01:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:42Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"56448e6b-6b3a-4bb9-91f8-2f0a9b44e9aa\\\",\\\"systemUUID\\\":\\\"0adb347d-b90d-4e9f-9962-f1bd97980aed\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:42Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:42 crc kubenswrapper[4870]: I1014 07:01:42.515098 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:42 crc kubenswrapper[4870]: I1014 07:01:42.515137 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:42 crc kubenswrapper[4870]: I1014 07:01:42.515150 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:42 crc kubenswrapper[4870]: I1014 07:01:42.515169 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:42 crc kubenswrapper[4870]: I1014 07:01:42.515180 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:42Z","lastTransitionTime":"2025-10-14T07:01:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:42 crc kubenswrapper[4870]: I1014 07:01:42.515899 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-kz77f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d3864a91-9bf6-4540-a908-cc41aa39656c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d4ded3d86030f5ea548c8b1cc3caddbe14a45c9a3ae4af836e892880584f57d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4h8v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:28Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-kz77f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:42Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:42 crc kubenswrapper[4870]: E1014 07:01:42.527185 4870 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:01:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:01:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:42Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:01:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:01:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:42Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"56448e6b-6b3a-4bb9-91f8-2f0a9b44e9aa\\\",\\\"systemUUID\\\":\\\"0adb347d-b90d-4e9f-9962-f1bd97980aed\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:42Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:42 crc kubenswrapper[4870]: I1014 07:01:42.531010 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:42 crc kubenswrapper[4870]: I1014 07:01:42.531038 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:42 crc kubenswrapper[4870]: I1014 07:01:42.531049 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:42 crc kubenswrapper[4870]: I1014 07:01:42.531064 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:42 crc kubenswrapper[4870]: I1014 07:01:42.531074 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:42Z","lastTransitionTime":"2025-10-14T07:01:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:42 crc kubenswrapper[4870]: I1014 07:01:42.532846 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:42Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:42 crc kubenswrapper[4870]: I1014 07:01:42.544070 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3eb0b397c05ffd6650607f1be4a75683e0342d3333fc096bf1cd63c7688834f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:42Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:42 crc kubenswrapper[4870]: E1014 07:01:42.546901 4870 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:01:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:01:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:42Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:01:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:01:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:42Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"56448e6b-6b3a-4bb9-91f8-2f0a9b44e9aa\\\",\\\"systemUUID\\\":\\\"0adb347d-b90d-4e9f-9962-f1bd97980aed\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:42Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:42 crc kubenswrapper[4870]: I1014 07:01:42.550814 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:42 crc kubenswrapper[4870]: I1014 07:01:42.550859 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:42 crc kubenswrapper[4870]: I1014 07:01:42.550872 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:42 crc kubenswrapper[4870]: I1014 07:01:42.550896 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:42 crc kubenswrapper[4870]: I1014 07:01:42.550908 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:42Z","lastTransitionTime":"2025-10-14T07:01:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:42 crc kubenswrapper[4870]: I1014 07:01:42.559847 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"874a5e45-dffd-4d17-b609-b7d3ed2eab07\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d98d044836ab44a405196915ef7bf572b75113f2c81ec7e64effa63b47ad3543\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wj4ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8004752575df4a1c44efbcafe64799f09126d0f97285d92388c76a0b90e022f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wj4ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7tvc7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:42Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:42 crc kubenswrapper[4870]: E1014 07:01:42.572987 4870 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:01:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:01:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:42Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:01:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:01:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:42Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"56448e6b-6b3a-4bb9-91f8-2f0a9b44e9aa\\\",\\\"systemUUID\\\":\\\"0adb347d-b90d-4e9f-9962-f1bd97980aed\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:42Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:42 crc kubenswrapper[4870]: E1014 07:01:42.573134 4870 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 14 07:01:42 crc kubenswrapper[4870]: I1014 07:01:42.575085 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zc5j6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d7a4f41-388a-4320-b896-43c17ff13da6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6p2wh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6p2wh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:40Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zc5j6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:42Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:42 crc kubenswrapper[4870]: I1014 07:01:42.575222 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:42 crc kubenswrapper[4870]: I1014 07:01:42.575244 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:42 crc kubenswrapper[4870]: I1014 07:01:42.575253 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:42 crc kubenswrapper[4870]: I1014 07:01:42.575269 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:42 crc kubenswrapper[4870]: I1014 07:01:42.575279 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:42Z","lastTransitionTime":"2025-10-14T07:01:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:42 crc kubenswrapper[4870]: I1014 07:01:42.589411 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77a0d93a-5391-4112-b502-c474f45de209\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7815af9a584f6db6660073da289f2e8c987a04ef9ddfac155bdd9a15b014fcfa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b329b570622155a55f70501b2260e278106a68ebef497914231b4ec860e0125\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba942a0e6fe5ab93a187e0b7b4bd7c440c2e7827d9b5323fc50f90b562075392\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b9018b078c5c8276243c1696860f0aa73e2d8dd917913ae9d03b30dc9b5113a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://978e8db184ba9e430232c6bb2ee530beb92567fdc71ba3c35b0bedca3492d845\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"file observer\\\\nW1014 07:01:25.870672 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1014 07:01:25.870795 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1014 07:01:25.871521 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-756714914/tls.crt::/tmp/serving-cert-756714914/tls.key\\\\\\\"\\\\nI1014 07:01:26.460037 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1014 07:01:26.464876 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1014 07:01:26.464902 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1014 07:01:26.464925 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1014 07:01:26.464931 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1014 07:01:26.481374 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1014 07:01:26.481400 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:01:26.481405 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:01:26.481410 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1014 07:01:26.481414 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1014 07:01:26.481417 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1014 07:01:26.481420 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1014 07:01:26.481610 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1014 07:01:26.482872 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bd0044a247bdc5ca4cab7294103af3270b76cde61f3109b85e83ed050ca53b6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82d3c2214e0be679d2827594b159ace362151318689078bc49d3158f42899636\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82d3c2214e0be679d2827594b159ace362151318689078bc49d3158f42899636\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:05Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:42Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:42 crc kubenswrapper[4870]: I1014 07:01:42.602637 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:42Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:42 crc kubenswrapper[4870]: I1014 07:01:42.615125 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e26fe49ec2045d20a0cb2dd17e96e79b6eb43c641eb184c688cc472ca5db39d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aea115f7ee71c700b218fe15c928af7b55e6babe70e65d67f125d87be1fd2e2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:42Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:42 crc kubenswrapper[4870]: I1014 07:01:42.639113 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-v7br2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dcaef0e-42e3-424b-bbe5-4a023efe912d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71dec55e859b11a081dbb90d7d6cf6b4846cf591382303d8329003ff50c71f2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e72cff0723e2ce520340244b4e597d8016c58e1b071c4b087f2f6cd718741594\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc9d4db944f837289d1923176d60cfc4a566a282c2d2e3a703e220df884703c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d16ed1c65925549dcb69ab02c3013e25e3f123a83179f1af3a8e08bf91a59896\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed59243b59c39ce82bf9bc8f33c0a8b84854846afff6a89f5d37ffc97a096b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3b6d20cf459b1f6d3aff5b9290f9391909baf27ee40c49e2b0cd2d5cfe889ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a05f64fdac0425bbf7ea39d9d9975279c4a68dc9a93e27f58f2d7017a328f428\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c43b07bd7ecc0a98ea65b6f59752ad8651568236dc78ee216e0ca42dabcbe4f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-14T07:01:39Z\\\",\\\"message\\\":\\\"ormers/externalversions/factory.go:141\\\\nI1014 07:01:39.675747 6159 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1014 07:01:39.675746 6159 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1014 07:01:39.675779 6159 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1014 07:01:39.675818 6159 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1014 07:01:39.675880 6159 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1014 07:01:39.676734 6159 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1014 07:01:39.676776 6159 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1014 07:01:39.676790 6159 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1014 07:01:39.676817 6159 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1014 07:01:39.676835 6159 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1014 07:01:39.676870 6159 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1014 07:01:39.676971 6159 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1014 07:01:39.677038 6159 factory.go:656] Stopping watch factory\\\\nI1014 07:01:39.677058 6159 ovnkube.go:599] Stopped ovnkube\\\\nI1014 0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:33Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fe3468cdd612fc4378e1f6f55a2a9d00274f7ddb65cb57b2a368dfc065cfb4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ba24c2e067b9f3a14d9d018b66300b54d866530d30174fbdb3c899299575d28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ba24c2e067b9f3a14d9d018b66300b54d866530d30174fbdb3c899299575d28\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-v7br2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:42Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:42 crc kubenswrapper[4870]: I1014 07:01:42.659249 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-khtq8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28deeea6-a1ef-4a4a-a377-d954ccd87de9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://296bb048a1d99649beb57d609573833bf547cd4d628ebc823c6af83fe5736c9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rb79p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2b7bc4c93307e80b778f2282ced993898f1c4cfc350a7d74828b8f9dea9906d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rb79p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-khtq8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:42Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:42 crc kubenswrapper[4870]: I1014 07:01:42.678591 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:42 crc kubenswrapper[4870]: I1014 07:01:42.678632 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:42 crc kubenswrapper[4870]: I1014 07:01:42.678643 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:42 crc kubenswrapper[4870]: I1014 07:01:42.678663 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:42 crc kubenswrapper[4870]: I1014 07:01:42.678675 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:42Z","lastTransitionTime":"2025-10-14T07:01:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:42 crc kubenswrapper[4870]: I1014 07:01:42.699165 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33a9493e-9660-48db-b8d1-310e29ebe313\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3570643ef1b7d5cc3a542f320413b819338e0c8b7ab7f0f09cc909997a39bab8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb398668197dedc1c2a14b4ae7d6a29714f3527309ccb9c1622af841100e1d60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1d2b6d068d63bd38cdfb83824525f6d12054f6cabab5d684e03ea5f2152076c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://783ff243e395333f3fc80cfae21358c532f0c9c5cb5515586be56b90ed10d912\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a8135a3d01076e4437ae8a784b459bf8d16de57fa16ec7ce4d40c5129c9d4a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48ca3a5017ba650df99b2fc7201c3fe5abfa89d81089c5647cde1b1549f534fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48ca3a5017ba650df99b2fc7201c3fe5abfa89d81089c5647cde1b1549f534fc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://19682eca67f76bf2e06e1b80f12e21141fcbce0acb8195f0defacd1b9a10b5e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://19682eca67f76bf2e06e1b80f12e21141fcbce0acb8195f0defacd1b9a10b5e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://042aad6d5f6bee8eef77f9e6a2c265de775a69f3da72aedc68bdf312198f00fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://042aad6d5f6bee8eef77f9e6a2c265de775a69f3da72aedc68bdf312198f00fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:05Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:42Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:42 crc kubenswrapper[4870]: I1014 07:01:42.714682 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-4v2v2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4796ba6-c23b-4764-be96-2e3a3bf5e962\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf5afb0b99a7101f6d5573fb7eb925c5a1d8becc11b2e3c6f3351d80b6f0e823\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnp8b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-4v2v2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:42Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:42 crc kubenswrapper[4870]: I1014 07:01:42.737326 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-cdjmj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"612e7ab9-bda4-45d8-bc50-ff1afb68aace\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7776352d62b814fc63d01e4c16ad13f878b0597eb37a370bf9da0a88abab6cfa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://834444b5350ee12df4e4652c3009880d8d69e1edf8ba0c0156ab96fb2bbe28e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://834444b5350ee12df4e4652c3009880d8d69e1edf8ba0c0156ab96fb2bbe28e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c1972e7b2f75790db22a07a078eee97072bffbdb7f8e594bf2be5f800632a7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c1972e7b2f75790db22a07a078eee97072bffbdb7f8e594bf2be5f800632a7b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b124b42a44c6a23da1d92b9a5326987eb74430f80489a71e8913829475654ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b124b42a44c6a23da1d92b9a5326987eb74430f80489a71e8913829475654ab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62d0deb6b20ea65fbcbaf94c2fc14d69adb22e7a46cc043ccd4def2668ddd4b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62d0deb6b20ea65fbcbaf94c2fc14d69adb22e7a46cc043ccd4def2668ddd4b7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7f0d7cf29ca737d0effe58731a8b890379527c97f78c144d4eebb910260fadb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7f0d7cf29ca737d0effe58731a8b890379527c97f78c144d4eebb910260fadb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5959abb7c8e7ab32bed3aca709ae0334e8400f991ba724246c9fda56cbed822\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5959abb7c8e7ab32bed3aca709ae0334e8400f991ba724246c9fda56cbed822\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-cdjmj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:42Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:42 crc kubenswrapper[4870]: I1014 07:01:42.755798 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fdxjh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a3a430a-b540-4a8c-adad-b2893dbf9898\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93e41d5e927eed7ec62eb9c3607cc804cd8999556663363fd123b2a784b00ff1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k787j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fdxjh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:42Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:42 crc kubenswrapper[4870]: I1014 07:01:42.773110 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:42Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:42 crc kubenswrapper[4870]: I1014 07:01:42.781911 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:42 crc kubenswrapper[4870]: I1014 07:01:42.781965 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:42 crc kubenswrapper[4870]: I1014 07:01:42.781981 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:42 crc kubenswrapper[4870]: I1014 07:01:42.782005 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:42 crc kubenswrapper[4870]: I1014 07:01:42.782022 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:42Z","lastTransitionTime":"2025-10-14T07:01:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:42 crc kubenswrapper[4870]: I1014 07:01:42.787219 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3eb0b397c05ffd6650607f1be4a75683e0342d3333fc096bf1cd63c7688834f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:42Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:42 crc kubenswrapper[4870]: I1014 07:01:42.798955 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"874a5e45-dffd-4d17-b609-b7d3ed2eab07\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d98d044836ab44a405196915ef7bf572b75113f2c81ec7e64effa63b47ad3543\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wj4ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8004752575df4a1c44efbcafe64799f09126d0f97285d92388c76a0b90e022f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wj4ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7tvc7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:42Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:42 crc kubenswrapper[4870]: I1014 07:01:42.813403 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zc5j6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d7a4f41-388a-4320-b896-43c17ff13da6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6p2wh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6p2wh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:40Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zc5j6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:42Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:42 crc kubenswrapper[4870]: I1014 07:01:42.832357 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77a0d93a-5391-4112-b502-c474f45de209\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7815af9a584f6db6660073da289f2e8c987a04ef9ddfac155bdd9a15b014fcfa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b329b570622155a55f70501b2260e278106a68ebef497914231b4ec860e0125\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba942a0e6fe5ab93a187e0b7b4bd7c440c2e7827d9b5323fc50f90b562075392\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b9018b078c5c8276243c1696860f0aa73e2d8dd917913ae9d03b30dc9b5113a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://978e8db184ba9e430232c6bb2ee530beb92567fdc71ba3c35b0bedca3492d845\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"file observer\\\\nW1014 07:01:25.870672 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1014 07:01:25.870795 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1014 07:01:25.871521 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-756714914/tls.crt::/tmp/serving-cert-756714914/tls.key\\\\\\\"\\\\nI1014 07:01:26.460037 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1014 07:01:26.464876 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1014 07:01:26.464902 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1014 07:01:26.464925 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1014 07:01:26.464931 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1014 07:01:26.481374 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1014 07:01:26.481400 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:01:26.481405 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:01:26.481410 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1014 07:01:26.481414 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1014 07:01:26.481417 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1014 07:01:26.481420 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1014 07:01:26.481610 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1014 07:01:26.482872 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bd0044a247bdc5ca4cab7294103af3270b76cde61f3109b85e83ed050ca53b6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82d3c2214e0be679d2827594b159ace362151318689078bc49d3158f42899636\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82d3c2214e0be679d2827594b159ace362151318689078bc49d3158f42899636\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:05Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:42Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:42 crc kubenswrapper[4870]: I1014 07:01:42.847428 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:42Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:42 crc kubenswrapper[4870]: I1014 07:01:42.866721 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e26fe49ec2045d20a0cb2dd17e96e79b6eb43c641eb184c688cc472ca5db39d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aea115f7ee71c700b218fe15c928af7b55e6babe70e65d67f125d87be1fd2e2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:42Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:42 crc kubenswrapper[4870]: I1014 07:01:42.884413 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:42 crc kubenswrapper[4870]: I1014 07:01:42.884537 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:42 crc kubenswrapper[4870]: I1014 07:01:42.884567 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:42 crc kubenswrapper[4870]: I1014 07:01:42.884603 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:42 crc kubenswrapper[4870]: I1014 07:01:42.884634 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:42Z","lastTransitionTime":"2025-10-14T07:01:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:42 crc kubenswrapper[4870]: I1014 07:01:42.892469 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-v7br2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dcaef0e-42e3-424b-bbe5-4a023efe912d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71dec55e859b11a081dbb90d7d6cf6b4846cf591382303d8329003ff50c71f2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e72cff0723e2ce520340244b4e597d8016c58e1b071c4b087f2f6cd718741594\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc9d4db944f837289d1923176d60cfc4a566a282c2d2e3a703e220df884703c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d16ed1c65925549dcb69ab02c3013e25e3f123a83179f1af3a8e08bf91a59896\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed59243b59c39ce82bf9bc8f33c0a8b84854846afff6a89f5d37ffc97a096b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3b6d20cf459b1f6d3aff5b9290f9391909baf27ee40c49e2b0cd2d5cfe889ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a05f64fdac0425bbf7ea39d9d9975279c4a68dc9a93e27f58f2d7017a328f428\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c43b07bd7ecc0a98ea65b6f59752ad8651568236dc78ee216e0ca42dabcbe4f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-14T07:01:39Z\\\",\\\"message\\\":\\\"ormers/externalversions/factory.go:141\\\\nI1014 07:01:39.675747 6159 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1014 07:01:39.675746 6159 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1014 07:01:39.675779 6159 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1014 07:01:39.675818 6159 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1014 07:01:39.675880 6159 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1014 07:01:39.676734 6159 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1014 07:01:39.676776 6159 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1014 07:01:39.676790 6159 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1014 07:01:39.676817 6159 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1014 07:01:39.676835 6159 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1014 07:01:39.676870 6159 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1014 07:01:39.676971 6159 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1014 07:01:39.677038 6159 factory.go:656] Stopping watch factory\\\\nI1014 07:01:39.677058 6159 ovnkube.go:599] Stopped ovnkube\\\\nI1014 0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:33Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a05f64fdac0425bbf7ea39d9d9975279c4a68dc9a93e27f58f2d7017a328f428\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-14T07:01:41Z\\\",\\\"message\\\":\\\"rvices do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-console/downloads_TCP_cluster\\\\\\\", UUID:\\\\\\\"2ead45b3-c313-4fbc-a7bc-2b3c4ffd610c\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-console/downloads\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-console/downloads_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-console/downloads\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.213\\\\\\\", Port:80, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nF1014 07:01:41.689119 6388 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fe3468cdd612fc4378e1f6f55a2a9d00274f7ddb65cb57b2a368dfc065cfb4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ba24c2e067b9f3a14d9d018b66300b54d866530d30174fbdb3c899299575d28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ba24c2e067b9f3a14d9d018b66300b54d866530d30174fbdb3c899299575d28\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-v7br2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:42Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:42 crc kubenswrapper[4870]: I1014 07:01:42.908973 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-khtq8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28deeea6-a1ef-4a4a-a377-d954ccd87de9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://296bb048a1d99649beb57d609573833bf547cd4d628ebc823c6af83fe5736c9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rb79p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2b7bc4c93307e80b778f2282ced993898f1c4cfc350a7d74828b8f9dea9906d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rb79p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-khtq8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:42Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:42 crc kubenswrapper[4870]: I1014 07:01:42.933036 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33a9493e-9660-48db-b8d1-310e29ebe313\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3570643ef1b7d5cc3a542f320413b819338e0c8b7ab7f0f09cc909997a39bab8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb398668197dedc1c2a14b4ae7d6a29714f3527309ccb9c1622af841100e1d60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1d2b6d068d63bd38cdfb83824525f6d12054f6cabab5d684e03ea5f2152076c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://783ff243e395333f3fc80cfae21358c532f0c9c5cb5515586be56b90ed10d912\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a8135a3d01076e4437ae8a784b459bf8d16de57fa16ec7ce4d40c5129c9d4a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48ca3a5017ba650df99b2fc7201c3fe5abfa89d81089c5647cde1b1549f534fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48ca3a5017ba650df99b2fc7201c3fe5abfa89d81089c5647cde1b1549f534fc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://19682eca67f76bf2e06e1b80f12e21141fcbce0acb8195f0defacd1b9a10b5e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://19682eca67f76bf2e06e1b80f12e21141fcbce0acb8195f0defacd1b9a10b5e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://042aad6d5f6bee8eef77f9e6a2c265de775a69f3da72aedc68bdf312198f00fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://042aad6d5f6bee8eef77f9e6a2c265de775a69f3da72aedc68bdf312198f00fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:05Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:42Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:42 crc kubenswrapper[4870]: I1014 07:01:42.946188 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-4v2v2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4796ba6-c23b-4764-be96-2e3a3bf5e962\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf5afb0b99a7101f6d5573fb7eb925c5a1d8becc11b2e3c6f3351d80b6f0e823\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnp8b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-4v2v2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:42Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:42 crc kubenswrapper[4870]: I1014 07:01:42.968257 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-cdjmj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"612e7ab9-bda4-45d8-bc50-ff1afb68aace\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7776352d62b814fc63d01e4c16ad13f878b0597eb37a370bf9da0a88abab6cfa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://834444b5350ee12df4e4652c3009880d8d69e1edf8ba0c0156ab96fb2bbe28e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://834444b5350ee12df4e4652c3009880d8d69e1edf8ba0c0156ab96fb2bbe28e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c1972e7b2f75790db22a07a078eee97072bffbdb7f8e594bf2be5f800632a7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c1972e7b2f75790db22a07a078eee97072bffbdb7f8e594bf2be5f800632a7b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b124b42a44c6a23da1d92b9a5326987eb74430f80489a71e8913829475654ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b124b42a44c6a23da1d92b9a5326987eb74430f80489a71e8913829475654ab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62d0deb6b20ea65fbcbaf94c2fc14d69adb22e7a46cc043ccd4def2668ddd4b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62d0deb6b20ea65fbcbaf94c2fc14d69adb22e7a46cc043ccd4def2668ddd4b7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7f0d7cf29ca737d0effe58731a8b890379527c97f78c144d4eebb910260fadb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7f0d7cf29ca737d0effe58731a8b890379527c97f78c144d4eebb910260fadb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5959abb7c8e7ab32bed3aca709ae0334e8400f991ba724246c9fda56cbed822\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5959abb7c8e7ab32bed3aca709ae0334e8400f991ba724246c9fda56cbed822\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-cdjmj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:42Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:42 crc kubenswrapper[4870]: I1014 07:01:42.986402 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fdxjh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a3a430a-b540-4a8c-adad-b2893dbf9898\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93e41d5e927eed7ec62eb9c3607cc804cd8999556663363fd123b2a784b00ff1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k787j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fdxjh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:42Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:42 crc kubenswrapper[4870]: I1014 07:01:42.987806 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:42 crc kubenswrapper[4870]: I1014 07:01:42.987849 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:42 crc kubenswrapper[4870]: I1014 07:01:42.987865 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:42 crc kubenswrapper[4870]: I1014 07:01:42.987891 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:42 crc kubenswrapper[4870]: I1014 07:01:42.987905 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:42Z","lastTransitionTime":"2025-10-14T07:01:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:43 crc kubenswrapper[4870]: I1014 07:01:43.002364 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bcbc6dd6-aab0-4dbc-a62e-c58f06ec1337\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25ec4808b09adc44488530b3c4e29fd7fec46b3fcfa31d4449ec4a6dbc880ec8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f3f5de86fb11231a19034a18e9d7f31476c162e6567d0c136ead8bbbd900d6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d57d938d9d9c92198ae5b3857536ece36012fbfbe8cc8cfd245d1386d39b0be5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77a3247a7f65709543d7fd374b2ab69d7d793537d4132d7e22261188d6e58544\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:05Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:42Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:43 crc kubenswrapper[4870]: I1014 07:01:43.018529 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e539fac3d0a2ade314a4b4c1050bb27c37bcff5ba5272fa2d71b97ab5c0d7191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:43Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:43 crc kubenswrapper[4870]: I1014 07:01:43.033840 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zc5j6" Oct 14 07:01:43 crc kubenswrapper[4870]: E1014 07:01:43.034029 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zc5j6" podUID="7d7a4f41-388a-4320-b896-43c17ff13da6" Oct 14 07:01:43 crc kubenswrapper[4870]: I1014 07:01:43.034044 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:43Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:43 crc kubenswrapper[4870]: I1014 07:01:43.048149 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-kz77f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d3864a91-9bf6-4540-a908-cc41aa39656c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d4ded3d86030f5ea548c8b1cc3caddbe14a45c9a3ae4af836e892880584f57d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4h8v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:28Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-kz77f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:43Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:43 crc kubenswrapper[4870]: I1014 07:01:43.090643 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:43 crc kubenswrapper[4870]: I1014 07:01:43.090727 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:43 crc kubenswrapper[4870]: I1014 07:01:43.090748 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:43 crc kubenswrapper[4870]: I1014 07:01:43.090780 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:43 crc kubenswrapper[4870]: I1014 07:01:43.090800 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:43Z","lastTransitionTime":"2025-10-14T07:01:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:43 crc kubenswrapper[4870]: I1014 07:01:43.194217 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:43 crc kubenswrapper[4870]: I1014 07:01:43.194278 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:43 crc kubenswrapper[4870]: I1014 07:01:43.194294 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:43 crc kubenswrapper[4870]: I1014 07:01:43.194318 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:43 crc kubenswrapper[4870]: I1014 07:01:43.194332 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:43Z","lastTransitionTime":"2025-10-14T07:01:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:43 crc kubenswrapper[4870]: I1014 07:01:43.303102 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:43 crc kubenswrapper[4870]: I1014 07:01:43.303164 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:43 crc kubenswrapper[4870]: I1014 07:01:43.303179 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:43 crc kubenswrapper[4870]: I1014 07:01:43.303204 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:43 crc kubenswrapper[4870]: I1014 07:01:43.303222 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:43Z","lastTransitionTime":"2025-10-14T07:01:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:43 crc kubenswrapper[4870]: I1014 07:01:43.406115 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:43 crc kubenswrapper[4870]: I1014 07:01:43.406157 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:43 crc kubenswrapper[4870]: I1014 07:01:43.406174 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:43 crc kubenswrapper[4870]: I1014 07:01:43.406195 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:43 crc kubenswrapper[4870]: I1014 07:01:43.406207 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:43Z","lastTransitionTime":"2025-10-14T07:01:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:43 crc kubenswrapper[4870]: I1014 07:01:43.469500 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-v7br2_8dcaef0e-42e3-424b-bbe5-4a023efe912d/ovnkube-controller/1.log" Oct 14 07:01:43 crc kubenswrapper[4870]: I1014 07:01:43.475651 4870 scope.go:117] "RemoveContainer" containerID="a05f64fdac0425bbf7ea39d9d9975279c4a68dc9a93e27f58f2d7017a328f428" Oct 14 07:01:43 crc kubenswrapper[4870]: E1014 07:01:43.475831 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-v7br2_openshift-ovn-kubernetes(8dcaef0e-42e3-424b-bbe5-4a023efe912d)\"" pod="openshift-ovn-kubernetes/ovnkube-node-v7br2" podUID="8dcaef0e-42e3-424b-bbe5-4a023efe912d" Oct 14 07:01:43 crc kubenswrapper[4870]: I1014 07:01:43.498495 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:43Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:43 crc kubenswrapper[4870]: I1014 07:01:43.509954 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:43 crc kubenswrapper[4870]: I1014 07:01:43.510007 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:43 crc kubenswrapper[4870]: I1014 07:01:43.510020 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:43 crc kubenswrapper[4870]: I1014 07:01:43.510041 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:43 crc kubenswrapper[4870]: I1014 07:01:43.510054 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:43Z","lastTransitionTime":"2025-10-14T07:01:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:43 crc kubenswrapper[4870]: I1014 07:01:43.517223 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3eb0b397c05ffd6650607f1be4a75683e0342d3333fc096bf1cd63c7688834f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:43Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:43 crc kubenswrapper[4870]: I1014 07:01:43.535238 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"874a5e45-dffd-4d17-b609-b7d3ed2eab07\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d98d044836ab44a405196915ef7bf572b75113f2c81ec7e64effa63b47ad3543\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wj4ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8004752575df4a1c44efbcafe64799f09126d0f97285d92388c76a0b90e022f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wj4ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7tvc7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:43Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:43 crc kubenswrapper[4870]: I1014 07:01:43.547969 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zc5j6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d7a4f41-388a-4320-b896-43c17ff13da6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6p2wh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6p2wh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:40Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zc5j6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:43Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:43 crc kubenswrapper[4870]: I1014 07:01:43.567480 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77a0d93a-5391-4112-b502-c474f45de209\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7815af9a584f6db6660073da289f2e8c987a04ef9ddfac155bdd9a15b014fcfa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b329b570622155a55f70501b2260e278106a68ebef497914231b4ec860e0125\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba942a0e6fe5ab93a187e0b7b4bd7c440c2e7827d9b5323fc50f90b562075392\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b9018b078c5c8276243c1696860f0aa73e2d8dd917913ae9d03b30dc9b5113a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://978e8db184ba9e430232c6bb2ee530beb92567fdc71ba3c35b0bedca3492d845\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"file observer\\\\nW1014 07:01:25.870672 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1014 07:01:25.870795 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1014 07:01:25.871521 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-756714914/tls.crt::/tmp/serving-cert-756714914/tls.key\\\\\\\"\\\\nI1014 07:01:26.460037 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1014 07:01:26.464876 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1014 07:01:26.464902 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1014 07:01:26.464925 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1014 07:01:26.464931 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1014 07:01:26.481374 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1014 07:01:26.481400 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:01:26.481405 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:01:26.481410 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1014 07:01:26.481414 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1014 07:01:26.481417 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1014 07:01:26.481420 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1014 07:01:26.481610 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1014 07:01:26.482872 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bd0044a247bdc5ca4cab7294103af3270b76cde61f3109b85e83ed050ca53b6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82d3c2214e0be679d2827594b159ace362151318689078bc49d3158f42899636\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82d3c2214e0be679d2827594b159ace362151318689078bc49d3158f42899636\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:05Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:43Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:43 crc kubenswrapper[4870]: I1014 07:01:43.584924 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:43Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:43 crc kubenswrapper[4870]: I1014 07:01:43.603262 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e26fe49ec2045d20a0cb2dd17e96e79b6eb43c641eb184c688cc472ca5db39d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aea115f7ee71c700b218fe15c928af7b55e6babe70e65d67f125d87be1fd2e2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:43Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:43 crc kubenswrapper[4870]: I1014 07:01:43.613838 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:43 crc kubenswrapper[4870]: I1014 07:01:43.613886 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:43 crc kubenswrapper[4870]: I1014 07:01:43.613897 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:43 crc kubenswrapper[4870]: I1014 07:01:43.613919 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:43 crc kubenswrapper[4870]: I1014 07:01:43.613930 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:43Z","lastTransitionTime":"2025-10-14T07:01:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:43 crc kubenswrapper[4870]: I1014 07:01:43.638028 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-v7br2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dcaef0e-42e3-424b-bbe5-4a023efe912d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71dec55e859b11a081dbb90d7d6cf6b4846cf591382303d8329003ff50c71f2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e72cff0723e2ce520340244b4e597d8016c58e1b071c4b087f2f6cd718741594\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc9d4db944f837289d1923176d60cfc4a566a282c2d2e3a703e220df884703c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d16ed1c65925549dcb69ab02c3013e25e3f123a83179f1af3a8e08bf91a59896\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed59243b59c39ce82bf9bc8f33c0a8b84854846afff6a89f5d37ffc97a096b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3b6d20cf459b1f6d3aff5b9290f9391909baf27ee40c49e2b0cd2d5cfe889ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a05f64fdac0425bbf7ea39d9d9975279c4a68dc9a93e27f58f2d7017a328f428\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a05f64fdac0425bbf7ea39d9d9975279c4a68dc9a93e27f58f2d7017a328f428\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-14T07:01:41Z\\\",\\\"message\\\":\\\"rvices do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-console/downloads_TCP_cluster\\\\\\\", UUID:\\\\\\\"2ead45b3-c313-4fbc-a7bc-2b3c4ffd610c\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-console/downloads\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-console/downloads_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-console/downloads\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.213\\\\\\\", Port:80, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nF1014 07:01:41.689119 6388 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:40Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-v7br2_openshift-ovn-kubernetes(8dcaef0e-42e3-424b-bbe5-4a023efe912d)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fe3468cdd612fc4378e1f6f55a2a9d00274f7ddb65cb57b2a368dfc065cfb4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ba24c2e067b9f3a14d9d018b66300b54d866530d30174fbdb3c899299575d28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ba24c2e067b9f3a14d9d018b66300b54d866530d30174fbdb3c899299575d28\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-v7br2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:43Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:43 crc kubenswrapper[4870]: I1014 07:01:43.656358 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-cdjmj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"612e7ab9-bda4-45d8-bc50-ff1afb68aace\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7776352d62b814fc63d01e4c16ad13f878b0597eb37a370bf9da0a88abab6cfa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://834444b5350ee12df4e4652c3009880d8d69e1edf8ba0c0156ab96fb2bbe28e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://834444b5350ee12df4e4652c3009880d8d69e1edf8ba0c0156ab96fb2bbe28e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c1972e7b2f75790db22a07a078eee97072bffbdb7f8e594bf2be5f800632a7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c1972e7b2f75790db22a07a078eee97072bffbdb7f8e594bf2be5f800632a7b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b124b42a44c6a23da1d92b9a5326987eb74430f80489a71e8913829475654ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b124b42a44c6a23da1d92b9a5326987eb74430f80489a71e8913829475654ab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62d0deb6b20ea65fbcbaf94c2fc14d69adb22e7a46cc043ccd4def2668ddd4b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62d0deb6b20ea65fbcbaf94c2fc14d69adb22e7a46cc043ccd4def2668ddd4b7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7f0d7cf29ca737d0effe58731a8b890379527c97f78c144d4eebb910260fadb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7f0d7cf29ca737d0effe58731a8b890379527c97f78c144d4eebb910260fadb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5959abb7c8e7ab32bed3aca709ae0334e8400f991ba724246c9fda56cbed822\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5959abb7c8e7ab32bed3aca709ae0334e8400f991ba724246c9fda56cbed822\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-cdjmj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:43Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:43 crc kubenswrapper[4870]: I1014 07:01:43.676778 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fdxjh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a3a430a-b540-4a8c-adad-b2893dbf9898\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93e41d5e927eed7ec62eb9c3607cc804cd8999556663363fd123b2a784b00ff1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k787j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fdxjh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:43Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:43 crc kubenswrapper[4870]: I1014 07:01:43.695021 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-khtq8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28deeea6-a1ef-4a4a-a377-d954ccd87de9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://296bb048a1d99649beb57d609573833bf547cd4d628ebc823c6af83fe5736c9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rb79p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2b7bc4c93307e80b778f2282ced993898f1c4cfc350a7d74828b8f9dea9906d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rb79p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-khtq8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:43Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:43 crc kubenswrapper[4870]: I1014 07:01:43.717668 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:43 crc kubenswrapper[4870]: I1014 07:01:43.717731 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:43 crc kubenswrapper[4870]: I1014 07:01:43.717744 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:43 crc kubenswrapper[4870]: I1014 07:01:43.717760 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:43 crc kubenswrapper[4870]: I1014 07:01:43.717771 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:43Z","lastTransitionTime":"2025-10-14T07:01:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:43 crc kubenswrapper[4870]: I1014 07:01:43.719754 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33a9493e-9660-48db-b8d1-310e29ebe313\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3570643ef1b7d5cc3a542f320413b819338e0c8b7ab7f0f09cc909997a39bab8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb398668197dedc1c2a14b4ae7d6a29714f3527309ccb9c1622af841100e1d60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1d2b6d068d63bd38cdfb83824525f6d12054f6cabab5d684e03ea5f2152076c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://783ff243e395333f3fc80cfae21358c532f0c9c5cb5515586be56b90ed10d912\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a8135a3d01076e4437ae8a784b459bf8d16de57fa16ec7ce4d40c5129c9d4a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48ca3a5017ba650df99b2fc7201c3fe5abfa89d81089c5647cde1b1549f534fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48ca3a5017ba650df99b2fc7201c3fe5abfa89d81089c5647cde1b1549f534fc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://19682eca67f76bf2e06e1b80f12e21141fcbce0acb8195f0defacd1b9a10b5e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://19682eca67f76bf2e06e1b80f12e21141fcbce0acb8195f0defacd1b9a10b5e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://042aad6d5f6bee8eef77f9e6a2c265de775a69f3da72aedc68bdf312198f00fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://042aad6d5f6bee8eef77f9e6a2c265de775a69f3da72aedc68bdf312198f00fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:05Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:43Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:43 crc kubenswrapper[4870]: I1014 07:01:43.736687 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-4v2v2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4796ba6-c23b-4764-be96-2e3a3bf5e962\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf5afb0b99a7101f6d5573fb7eb925c5a1d8becc11b2e3c6f3351d80b6f0e823\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnp8b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-4v2v2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:43Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:43 crc kubenswrapper[4870]: I1014 07:01:43.751373 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bcbc6dd6-aab0-4dbc-a62e-c58f06ec1337\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25ec4808b09adc44488530b3c4e29fd7fec46b3fcfa31d4449ec4a6dbc880ec8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f3f5de86fb11231a19034a18e9d7f31476c162e6567d0c136ead8bbbd900d6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d57d938d9d9c92198ae5b3857536ece36012fbfbe8cc8cfd245d1386d39b0be5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77a3247a7f65709543d7fd374b2ab69d7d793537d4132d7e22261188d6e58544\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:05Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:43Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:43 crc kubenswrapper[4870]: I1014 07:01:43.771121 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e539fac3d0a2ade314a4b4c1050bb27c37bcff5ba5272fa2d71b97ab5c0d7191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:43Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:43 crc kubenswrapper[4870]: I1014 07:01:43.786098 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:43Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:43 crc kubenswrapper[4870]: I1014 07:01:43.802236 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-kz77f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d3864a91-9bf6-4540-a908-cc41aa39656c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d4ded3d86030f5ea548c8b1cc3caddbe14a45c9a3ae4af836e892880584f57d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4h8v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:28Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-kz77f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:43Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:43 crc kubenswrapper[4870]: I1014 07:01:43.821639 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:43 crc kubenswrapper[4870]: I1014 07:01:43.821704 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:43 crc kubenswrapper[4870]: I1014 07:01:43.821726 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:43 crc kubenswrapper[4870]: I1014 07:01:43.821758 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:43 crc kubenswrapper[4870]: I1014 07:01:43.821780 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:43Z","lastTransitionTime":"2025-10-14T07:01:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:43 crc kubenswrapper[4870]: I1014 07:01:43.925433 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:43 crc kubenswrapper[4870]: I1014 07:01:43.925516 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:43 crc kubenswrapper[4870]: I1014 07:01:43.925534 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:43 crc kubenswrapper[4870]: I1014 07:01:43.925621 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:43 crc kubenswrapper[4870]: I1014 07:01:43.925662 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:43Z","lastTransitionTime":"2025-10-14T07:01:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:44 crc kubenswrapper[4870]: I1014 07:01:44.029762 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:44 crc kubenswrapper[4870]: I1014 07:01:44.029830 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:44 crc kubenswrapper[4870]: I1014 07:01:44.029855 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:44 crc kubenswrapper[4870]: I1014 07:01:44.029885 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:44 crc kubenswrapper[4870]: I1014 07:01:44.029903 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:44Z","lastTransitionTime":"2025-10-14T07:01:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:44 crc kubenswrapper[4870]: I1014 07:01:44.033084 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 07:01:44 crc kubenswrapper[4870]: I1014 07:01:44.033127 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 07:01:44 crc kubenswrapper[4870]: I1014 07:01:44.033106 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 07:01:44 crc kubenswrapper[4870]: E1014 07:01:44.033334 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 07:01:44 crc kubenswrapper[4870]: E1014 07:01:44.033520 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 07:01:44 crc kubenswrapper[4870]: E1014 07:01:44.033761 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 07:01:44 crc kubenswrapper[4870]: I1014 07:01:44.132824 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:44 crc kubenswrapper[4870]: I1014 07:01:44.132864 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:44 crc kubenswrapper[4870]: I1014 07:01:44.132905 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:44 crc kubenswrapper[4870]: I1014 07:01:44.132920 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:44 crc kubenswrapper[4870]: I1014 07:01:44.132929 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:44Z","lastTransitionTime":"2025-10-14T07:01:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:44 crc kubenswrapper[4870]: I1014 07:01:44.236594 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:44 crc kubenswrapper[4870]: I1014 07:01:44.236675 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:44 crc kubenswrapper[4870]: I1014 07:01:44.236696 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:44 crc kubenswrapper[4870]: I1014 07:01:44.236731 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:44 crc kubenswrapper[4870]: I1014 07:01:44.236761 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:44Z","lastTransitionTime":"2025-10-14T07:01:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:44 crc kubenswrapper[4870]: I1014 07:01:44.341330 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:44 crc kubenswrapper[4870]: I1014 07:01:44.341379 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:44 crc kubenswrapper[4870]: I1014 07:01:44.341397 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:44 crc kubenswrapper[4870]: I1014 07:01:44.341419 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:44 crc kubenswrapper[4870]: I1014 07:01:44.341460 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:44Z","lastTransitionTime":"2025-10-14T07:01:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:44 crc kubenswrapper[4870]: I1014 07:01:44.419076 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7d7a4f41-388a-4320-b896-43c17ff13da6-metrics-certs\") pod \"network-metrics-daemon-zc5j6\" (UID: \"7d7a4f41-388a-4320-b896-43c17ff13da6\") " pod="openshift-multus/network-metrics-daemon-zc5j6" Oct 14 07:01:44 crc kubenswrapper[4870]: E1014 07:01:44.419338 4870 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 14 07:01:44 crc kubenswrapper[4870]: E1014 07:01:44.419516 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7d7a4f41-388a-4320-b896-43c17ff13da6-metrics-certs podName:7d7a4f41-388a-4320-b896-43c17ff13da6 nodeName:}" failed. No retries permitted until 2025-10-14 07:01:48.419484549 +0000 UTC m=+44.116844960 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/7d7a4f41-388a-4320-b896-43c17ff13da6-metrics-certs") pod "network-metrics-daemon-zc5j6" (UID: "7d7a4f41-388a-4320-b896-43c17ff13da6") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 14 07:01:44 crc kubenswrapper[4870]: I1014 07:01:44.449431 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:44 crc kubenswrapper[4870]: I1014 07:01:44.449552 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:44 crc kubenswrapper[4870]: I1014 07:01:44.449574 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:44 crc kubenswrapper[4870]: I1014 07:01:44.449607 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:44 crc kubenswrapper[4870]: I1014 07:01:44.449628 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:44Z","lastTransitionTime":"2025-10-14T07:01:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:44 crc kubenswrapper[4870]: I1014 07:01:44.553362 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:44 crc kubenswrapper[4870]: I1014 07:01:44.553479 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:44 crc kubenswrapper[4870]: I1014 07:01:44.553505 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:44 crc kubenswrapper[4870]: I1014 07:01:44.553538 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:44 crc kubenswrapper[4870]: I1014 07:01:44.553565 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:44Z","lastTransitionTime":"2025-10-14T07:01:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:44 crc kubenswrapper[4870]: I1014 07:01:44.657498 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:44 crc kubenswrapper[4870]: I1014 07:01:44.657655 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:44 crc kubenswrapper[4870]: I1014 07:01:44.657693 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:44 crc kubenswrapper[4870]: I1014 07:01:44.657724 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:44 crc kubenswrapper[4870]: I1014 07:01:44.657765 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:44Z","lastTransitionTime":"2025-10-14T07:01:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:44 crc kubenswrapper[4870]: I1014 07:01:44.761109 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:44 crc kubenswrapper[4870]: I1014 07:01:44.761187 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:44 crc kubenswrapper[4870]: I1014 07:01:44.761199 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:44 crc kubenswrapper[4870]: I1014 07:01:44.761218 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:44 crc kubenswrapper[4870]: I1014 07:01:44.761230 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:44Z","lastTransitionTime":"2025-10-14T07:01:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:44 crc kubenswrapper[4870]: I1014 07:01:44.864296 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:44 crc kubenswrapper[4870]: I1014 07:01:44.864487 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:44 crc kubenswrapper[4870]: I1014 07:01:44.864520 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:44 crc kubenswrapper[4870]: I1014 07:01:44.864551 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:44 crc kubenswrapper[4870]: I1014 07:01:44.864575 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:44Z","lastTransitionTime":"2025-10-14T07:01:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:44 crc kubenswrapper[4870]: I1014 07:01:44.985978 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:44 crc kubenswrapper[4870]: I1014 07:01:44.986047 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:44 crc kubenswrapper[4870]: I1014 07:01:44.986065 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:44 crc kubenswrapper[4870]: I1014 07:01:44.986089 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:44 crc kubenswrapper[4870]: I1014 07:01:44.986105 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:44Z","lastTransitionTime":"2025-10-14T07:01:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:45 crc kubenswrapper[4870]: I1014 07:01:45.033274 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zc5j6" Oct 14 07:01:45 crc kubenswrapper[4870]: E1014 07:01:45.033498 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zc5j6" podUID="7d7a4f41-388a-4320-b896-43c17ff13da6" Oct 14 07:01:45 crc kubenswrapper[4870]: I1014 07:01:45.052482 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:45Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:45 crc kubenswrapper[4870]: I1014 07:01:45.071283 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3eb0b397c05ffd6650607f1be4a75683e0342d3333fc096bf1cd63c7688834f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:45Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:45 crc kubenswrapper[4870]: I1014 07:01:45.088799 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:45 crc kubenswrapper[4870]: I1014 07:01:45.088838 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:45 crc kubenswrapper[4870]: I1014 07:01:45.088847 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:45 crc kubenswrapper[4870]: I1014 07:01:45.088863 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:45 crc kubenswrapper[4870]: I1014 07:01:45.088875 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:45Z","lastTransitionTime":"2025-10-14T07:01:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:45 crc kubenswrapper[4870]: I1014 07:01:45.088806 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"874a5e45-dffd-4d17-b609-b7d3ed2eab07\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d98d044836ab44a405196915ef7bf572b75113f2c81ec7e64effa63b47ad3543\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wj4ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8004752575df4a1c44efbcafe64799f09126d0f97285d92388c76a0b90e022f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wj4ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7tvc7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:45Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:45 crc kubenswrapper[4870]: I1014 07:01:45.100772 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zc5j6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d7a4f41-388a-4320-b896-43c17ff13da6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6p2wh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6p2wh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:40Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zc5j6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:45Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:45 crc kubenswrapper[4870]: I1014 07:01:45.118501 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77a0d93a-5391-4112-b502-c474f45de209\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7815af9a584f6db6660073da289f2e8c987a04ef9ddfac155bdd9a15b014fcfa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b329b570622155a55f70501b2260e278106a68ebef497914231b4ec860e0125\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba942a0e6fe5ab93a187e0b7b4bd7c440c2e7827d9b5323fc50f90b562075392\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b9018b078c5c8276243c1696860f0aa73e2d8dd917913ae9d03b30dc9b5113a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://978e8db184ba9e430232c6bb2ee530beb92567fdc71ba3c35b0bedca3492d845\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"file observer\\\\nW1014 07:01:25.870672 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1014 07:01:25.870795 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1014 07:01:25.871521 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-756714914/tls.crt::/tmp/serving-cert-756714914/tls.key\\\\\\\"\\\\nI1014 07:01:26.460037 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1014 07:01:26.464876 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1014 07:01:26.464902 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1014 07:01:26.464925 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1014 07:01:26.464931 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1014 07:01:26.481374 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1014 07:01:26.481400 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:01:26.481405 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:01:26.481410 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1014 07:01:26.481414 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1014 07:01:26.481417 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1014 07:01:26.481420 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1014 07:01:26.481610 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1014 07:01:26.482872 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bd0044a247bdc5ca4cab7294103af3270b76cde61f3109b85e83ed050ca53b6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82d3c2214e0be679d2827594b159ace362151318689078bc49d3158f42899636\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82d3c2214e0be679d2827594b159ace362151318689078bc49d3158f42899636\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:05Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:45Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:45 crc kubenswrapper[4870]: I1014 07:01:45.135847 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:45Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:45 crc kubenswrapper[4870]: I1014 07:01:45.159241 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e26fe49ec2045d20a0cb2dd17e96e79b6eb43c641eb184c688cc472ca5db39d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aea115f7ee71c700b218fe15c928af7b55e6babe70e65d67f125d87be1fd2e2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:45Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:45 crc kubenswrapper[4870]: I1014 07:01:45.191515 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:45 crc kubenswrapper[4870]: I1014 07:01:45.191554 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:45 crc kubenswrapper[4870]: I1014 07:01:45.191567 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:45 crc kubenswrapper[4870]: I1014 07:01:45.191584 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:45 crc kubenswrapper[4870]: I1014 07:01:45.191596 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:45Z","lastTransitionTime":"2025-10-14T07:01:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:45 crc kubenswrapper[4870]: I1014 07:01:45.198685 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-v7br2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dcaef0e-42e3-424b-bbe5-4a023efe912d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71dec55e859b11a081dbb90d7d6cf6b4846cf591382303d8329003ff50c71f2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e72cff0723e2ce520340244b4e597d8016c58e1b071c4b087f2f6cd718741594\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc9d4db944f837289d1923176d60cfc4a566a282c2d2e3a703e220df884703c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d16ed1c65925549dcb69ab02c3013e25e3f123a83179f1af3a8e08bf91a59896\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed59243b59c39ce82bf9bc8f33c0a8b84854846afff6a89f5d37ffc97a096b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3b6d20cf459b1f6d3aff5b9290f9391909baf27ee40c49e2b0cd2d5cfe889ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a05f64fdac0425bbf7ea39d9d9975279c4a68dc9a93e27f58f2d7017a328f428\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a05f64fdac0425bbf7ea39d9d9975279c4a68dc9a93e27f58f2d7017a328f428\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-14T07:01:41Z\\\",\\\"message\\\":\\\"rvices do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-console/downloads_TCP_cluster\\\\\\\", UUID:\\\\\\\"2ead45b3-c313-4fbc-a7bc-2b3c4ffd610c\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-console/downloads\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-console/downloads_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-console/downloads\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.213\\\\\\\", Port:80, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nF1014 07:01:41.689119 6388 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:40Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-v7br2_openshift-ovn-kubernetes(8dcaef0e-42e3-424b-bbe5-4a023efe912d)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fe3468cdd612fc4378e1f6f55a2a9d00274f7ddb65cb57b2a368dfc065cfb4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ba24c2e067b9f3a14d9d018b66300b54d866530d30174fbdb3c899299575d28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ba24c2e067b9f3a14d9d018b66300b54d866530d30174fbdb3c899299575d28\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-v7br2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:45Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:45 crc kubenswrapper[4870]: I1014 07:01:45.221202 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-cdjmj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"612e7ab9-bda4-45d8-bc50-ff1afb68aace\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7776352d62b814fc63d01e4c16ad13f878b0597eb37a370bf9da0a88abab6cfa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://834444b5350ee12df4e4652c3009880d8d69e1edf8ba0c0156ab96fb2bbe28e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://834444b5350ee12df4e4652c3009880d8d69e1edf8ba0c0156ab96fb2bbe28e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c1972e7b2f75790db22a07a078eee97072bffbdb7f8e594bf2be5f800632a7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c1972e7b2f75790db22a07a078eee97072bffbdb7f8e594bf2be5f800632a7b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b124b42a44c6a23da1d92b9a5326987eb74430f80489a71e8913829475654ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b124b42a44c6a23da1d92b9a5326987eb74430f80489a71e8913829475654ab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62d0deb6b20ea65fbcbaf94c2fc14d69adb22e7a46cc043ccd4def2668ddd4b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62d0deb6b20ea65fbcbaf94c2fc14d69adb22e7a46cc043ccd4def2668ddd4b7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7f0d7cf29ca737d0effe58731a8b890379527c97f78c144d4eebb910260fadb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7f0d7cf29ca737d0effe58731a8b890379527c97f78c144d4eebb910260fadb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5959abb7c8e7ab32bed3aca709ae0334e8400f991ba724246c9fda56cbed822\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5959abb7c8e7ab32bed3aca709ae0334e8400f991ba724246c9fda56cbed822\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-cdjmj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:45Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:45 crc kubenswrapper[4870]: I1014 07:01:45.246317 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fdxjh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a3a430a-b540-4a8c-adad-b2893dbf9898\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93e41d5e927eed7ec62eb9c3607cc804cd8999556663363fd123b2a784b00ff1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k787j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fdxjh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:45Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:45 crc kubenswrapper[4870]: I1014 07:01:45.258799 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-khtq8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28deeea6-a1ef-4a4a-a377-d954ccd87de9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://296bb048a1d99649beb57d609573833bf547cd4d628ebc823c6af83fe5736c9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rb79p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2b7bc4c93307e80b778f2282ced993898f1c4cfc350a7d74828b8f9dea9906d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rb79p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-khtq8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:45Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:45 crc kubenswrapper[4870]: I1014 07:01:45.280768 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33a9493e-9660-48db-b8d1-310e29ebe313\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3570643ef1b7d5cc3a542f320413b819338e0c8b7ab7f0f09cc909997a39bab8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb398668197dedc1c2a14b4ae7d6a29714f3527309ccb9c1622af841100e1d60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1d2b6d068d63bd38cdfb83824525f6d12054f6cabab5d684e03ea5f2152076c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://783ff243e395333f3fc80cfae21358c532f0c9c5cb5515586be56b90ed10d912\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a8135a3d01076e4437ae8a784b459bf8d16de57fa16ec7ce4d40c5129c9d4a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48ca3a5017ba650df99b2fc7201c3fe5abfa89d81089c5647cde1b1549f534fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48ca3a5017ba650df99b2fc7201c3fe5abfa89d81089c5647cde1b1549f534fc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://19682eca67f76bf2e06e1b80f12e21141fcbce0acb8195f0defacd1b9a10b5e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://19682eca67f76bf2e06e1b80f12e21141fcbce0acb8195f0defacd1b9a10b5e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://042aad6d5f6bee8eef77f9e6a2c265de775a69f3da72aedc68bdf312198f00fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://042aad6d5f6bee8eef77f9e6a2c265de775a69f3da72aedc68bdf312198f00fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:05Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:45Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:45 crc kubenswrapper[4870]: I1014 07:01:45.293663 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-4v2v2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4796ba6-c23b-4764-be96-2e3a3bf5e962\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf5afb0b99a7101f6d5573fb7eb925c5a1d8becc11b2e3c6f3351d80b6f0e823\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnp8b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-4v2v2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:45Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:45 crc kubenswrapper[4870]: I1014 07:01:45.295554 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:45 crc kubenswrapper[4870]: I1014 07:01:45.295598 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:45 crc kubenswrapper[4870]: I1014 07:01:45.295610 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:45 crc kubenswrapper[4870]: I1014 07:01:45.295631 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:45 crc kubenswrapper[4870]: I1014 07:01:45.295644 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:45Z","lastTransitionTime":"2025-10-14T07:01:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:45 crc kubenswrapper[4870]: I1014 07:01:45.306774 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bcbc6dd6-aab0-4dbc-a62e-c58f06ec1337\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25ec4808b09adc44488530b3c4e29fd7fec46b3fcfa31d4449ec4a6dbc880ec8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f3f5de86fb11231a19034a18e9d7f31476c162e6567d0c136ead8bbbd900d6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d57d938d9d9c92198ae5b3857536ece36012fbfbe8cc8cfd245d1386d39b0be5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77a3247a7f65709543d7fd374b2ab69d7d793537d4132d7e22261188d6e58544\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:05Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:45Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:45 crc kubenswrapper[4870]: I1014 07:01:45.319948 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e539fac3d0a2ade314a4b4c1050bb27c37bcff5ba5272fa2d71b97ab5c0d7191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:45Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:45 crc kubenswrapper[4870]: I1014 07:01:45.331827 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:45Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:45 crc kubenswrapper[4870]: I1014 07:01:45.341980 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-kz77f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d3864a91-9bf6-4540-a908-cc41aa39656c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d4ded3d86030f5ea548c8b1cc3caddbe14a45c9a3ae4af836e892880584f57d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4h8v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:28Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-kz77f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:45Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:45 crc kubenswrapper[4870]: I1014 07:01:45.398760 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:45 crc kubenswrapper[4870]: I1014 07:01:45.398795 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:45 crc kubenswrapper[4870]: I1014 07:01:45.398809 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:45 crc kubenswrapper[4870]: I1014 07:01:45.398831 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:45 crc kubenswrapper[4870]: I1014 07:01:45.398846 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:45Z","lastTransitionTime":"2025-10-14T07:01:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:45 crc kubenswrapper[4870]: I1014 07:01:45.501849 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:45 crc kubenswrapper[4870]: I1014 07:01:45.501895 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:45 crc kubenswrapper[4870]: I1014 07:01:45.501909 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:45 crc kubenswrapper[4870]: I1014 07:01:45.501927 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:45 crc kubenswrapper[4870]: I1014 07:01:45.501938 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:45Z","lastTransitionTime":"2025-10-14T07:01:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:45 crc kubenswrapper[4870]: I1014 07:01:45.605301 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:45 crc kubenswrapper[4870]: I1014 07:01:45.605378 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:45 crc kubenswrapper[4870]: I1014 07:01:45.605402 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:45 crc kubenswrapper[4870]: I1014 07:01:45.605473 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:45 crc kubenswrapper[4870]: I1014 07:01:45.605496 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:45Z","lastTransitionTime":"2025-10-14T07:01:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:45 crc kubenswrapper[4870]: I1014 07:01:45.709004 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:45 crc kubenswrapper[4870]: I1014 07:01:45.709062 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:45 crc kubenswrapper[4870]: I1014 07:01:45.709080 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:45 crc kubenswrapper[4870]: I1014 07:01:45.709107 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:45 crc kubenswrapper[4870]: I1014 07:01:45.709128 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:45Z","lastTransitionTime":"2025-10-14T07:01:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:45 crc kubenswrapper[4870]: I1014 07:01:45.812647 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:45 crc kubenswrapper[4870]: I1014 07:01:45.812741 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:45 crc kubenswrapper[4870]: I1014 07:01:45.812773 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:45 crc kubenswrapper[4870]: I1014 07:01:45.812806 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:45 crc kubenswrapper[4870]: I1014 07:01:45.812825 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:45Z","lastTransitionTime":"2025-10-14T07:01:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:45 crc kubenswrapper[4870]: I1014 07:01:45.915576 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:45 crc kubenswrapper[4870]: I1014 07:01:45.915633 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:45 crc kubenswrapper[4870]: I1014 07:01:45.915651 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:45 crc kubenswrapper[4870]: I1014 07:01:45.915676 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:45 crc kubenswrapper[4870]: I1014 07:01:45.915696 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:45Z","lastTransitionTime":"2025-10-14T07:01:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:46 crc kubenswrapper[4870]: I1014 07:01:46.020112 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:46 crc kubenswrapper[4870]: I1014 07:01:46.020203 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:46 crc kubenswrapper[4870]: I1014 07:01:46.020226 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:46 crc kubenswrapper[4870]: I1014 07:01:46.020259 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:46 crc kubenswrapper[4870]: I1014 07:01:46.020283 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:46Z","lastTransitionTime":"2025-10-14T07:01:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:46 crc kubenswrapper[4870]: I1014 07:01:46.033717 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 07:01:46 crc kubenswrapper[4870]: I1014 07:01:46.033787 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 07:01:46 crc kubenswrapper[4870]: I1014 07:01:46.033739 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 07:01:46 crc kubenswrapper[4870]: E1014 07:01:46.033919 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 07:01:46 crc kubenswrapper[4870]: E1014 07:01:46.034068 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 07:01:46 crc kubenswrapper[4870]: E1014 07:01:46.034179 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 07:01:46 crc kubenswrapper[4870]: I1014 07:01:46.124051 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:46 crc kubenswrapper[4870]: I1014 07:01:46.124145 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:46 crc kubenswrapper[4870]: I1014 07:01:46.124176 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:46 crc kubenswrapper[4870]: I1014 07:01:46.124218 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:46 crc kubenswrapper[4870]: I1014 07:01:46.124245 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:46Z","lastTransitionTime":"2025-10-14T07:01:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:46 crc kubenswrapper[4870]: I1014 07:01:46.228282 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:46 crc kubenswrapper[4870]: I1014 07:01:46.228343 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:46 crc kubenswrapper[4870]: I1014 07:01:46.228362 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:46 crc kubenswrapper[4870]: I1014 07:01:46.228388 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:46 crc kubenswrapper[4870]: I1014 07:01:46.228405 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:46Z","lastTransitionTime":"2025-10-14T07:01:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:46 crc kubenswrapper[4870]: I1014 07:01:46.331962 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:46 crc kubenswrapper[4870]: I1014 07:01:46.332030 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:46 crc kubenswrapper[4870]: I1014 07:01:46.332049 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:46 crc kubenswrapper[4870]: I1014 07:01:46.332076 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:46 crc kubenswrapper[4870]: I1014 07:01:46.332092 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:46Z","lastTransitionTime":"2025-10-14T07:01:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:46 crc kubenswrapper[4870]: I1014 07:01:46.436127 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:46 crc kubenswrapper[4870]: I1014 07:01:46.436193 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:46 crc kubenswrapper[4870]: I1014 07:01:46.436216 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:46 crc kubenswrapper[4870]: I1014 07:01:46.436247 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:46 crc kubenswrapper[4870]: I1014 07:01:46.436265 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:46Z","lastTransitionTime":"2025-10-14T07:01:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:46 crc kubenswrapper[4870]: I1014 07:01:46.539877 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:46 crc kubenswrapper[4870]: I1014 07:01:46.539924 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:46 crc kubenswrapper[4870]: I1014 07:01:46.539940 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:46 crc kubenswrapper[4870]: I1014 07:01:46.539960 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:46 crc kubenswrapper[4870]: I1014 07:01:46.539972 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:46Z","lastTransitionTime":"2025-10-14T07:01:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:46 crc kubenswrapper[4870]: I1014 07:01:46.643817 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:46 crc kubenswrapper[4870]: I1014 07:01:46.643888 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:46 crc kubenswrapper[4870]: I1014 07:01:46.643908 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:46 crc kubenswrapper[4870]: I1014 07:01:46.643968 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:46 crc kubenswrapper[4870]: I1014 07:01:46.643987 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:46Z","lastTransitionTime":"2025-10-14T07:01:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:46 crc kubenswrapper[4870]: I1014 07:01:46.746968 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:46 crc kubenswrapper[4870]: I1014 07:01:46.747023 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:46 crc kubenswrapper[4870]: I1014 07:01:46.747037 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:46 crc kubenswrapper[4870]: I1014 07:01:46.747058 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:46 crc kubenswrapper[4870]: I1014 07:01:46.747072 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:46Z","lastTransitionTime":"2025-10-14T07:01:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:46 crc kubenswrapper[4870]: I1014 07:01:46.850586 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:46 crc kubenswrapper[4870]: I1014 07:01:46.850613 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:46 crc kubenswrapper[4870]: I1014 07:01:46.850624 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:46 crc kubenswrapper[4870]: I1014 07:01:46.850655 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:46 crc kubenswrapper[4870]: I1014 07:01:46.850666 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:46Z","lastTransitionTime":"2025-10-14T07:01:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:46 crc kubenswrapper[4870]: I1014 07:01:46.953728 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:46 crc kubenswrapper[4870]: I1014 07:01:46.953797 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:46 crc kubenswrapper[4870]: I1014 07:01:46.953818 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:46 crc kubenswrapper[4870]: I1014 07:01:46.953844 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:46 crc kubenswrapper[4870]: I1014 07:01:46.953867 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:46Z","lastTransitionTime":"2025-10-14T07:01:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:47 crc kubenswrapper[4870]: I1014 07:01:47.033142 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zc5j6" Oct 14 07:01:47 crc kubenswrapper[4870]: E1014 07:01:47.033354 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zc5j6" podUID="7d7a4f41-388a-4320-b896-43c17ff13da6" Oct 14 07:01:47 crc kubenswrapper[4870]: I1014 07:01:47.060297 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:47 crc kubenswrapper[4870]: I1014 07:01:47.060391 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:47 crc kubenswrapper[4870]: I1014 07:01:47.060412 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:47 crc kubenswrapper[4870]: I1014 07:01:47.060504 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:47 crc kubenswrapper[4870]: I1014 07:01:47.060543 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:47Z","lastTransitionTime":"2025-10-14T07:01:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:47 crc kubenswrapper[4870]: I1014 07:01:47.164522 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:47 crc kubenswrapper[4870]: I1014 07:01:47.165100 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:47 crc kubenswrapper[4870]: I1014 07:01:47.165119 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:47 crc kubenswrapper[4870]: I1014 07:01:47.165150 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:47 crc kubenswrapper[4870]: I1014 07:01:47.165174 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:47Z","lastTransitionTime":"2025-10-14T07:01:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:47 crc kubenswrapper[4870]: I1014 07:01:47.267968 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:47 crc kubenswrapper[4870]: I1014 07:01:47.268027 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:47 crc kubenswrapper[4870]: I1014 07:01:47.268043 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:47 crc kubenswrapper[4870]: I1014 07:01:47.268067 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:47 crc kubenswrapper[4870]: I1014 07:01:47.268084 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:47Z","lastTransitionTime":"2025-10-14T07:01:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:47 crc kubenswrapper[4870]: I1014 07:01:47.372033 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:47 crc kubenswrapper[4870]: I1014 07:01:47.372089 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:47 crc kubenswrapper[4870]: I1014 07:01:47.372106 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:47 crc kubenswrapper[4870]: I1014 07:01:47.372132 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:47 crc kubenswrapper[4870]: I1014 07:01:47.372150 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:47Z","lastTransitionTime":"2025-10-14T07:01:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:47 crc kubenswrapper[4870]: I1014 07:01:47.474994 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:47 crc kubenswrapper[4870]: I1014 07:01:47.475094 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:47 crc kubenswrapper[4870]: I1014 07:01:47.475113 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:47 crc kubenswrapper[4870]: I1014 07:01:47.475146 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:47 crc kubenswrapper[4870]: I1014 07:01:47.475167 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:47Z","lastTransitionTime":"2025-10-14T07:01:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:47 crc kubenswrapper[4870]: I1014 07:01:47.578829 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:47 crc kubenswrapper[4870]: I1014 07:01:47.578903 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:47 crc kubenswrapper[4870]: I1014 07:01:47.578924 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:47 crc kubenswrapper[4870]: I1014 07:01:47.578952 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:47 crc kubenswrapper[4870]: I1014 07:01:47.578974 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:47Z","lastTransitionTime":"2025-10-14T07:01:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:47 crc kubenswrapper[4870]: I1014 07:01:47.683001 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:47 crc kubenswrapper[4870]: I1014 07:01:47.683100 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:47 crc kubenswrapper[4870]: I1014 07:01:47.683129 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:47 crc kubenswrapper[4870]: I1014 07:01:47.683172 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:47 crc kubenswrapper[4870]: I1014 07:01:47.683196 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:47Z","lastTransitionTime":"2025-10-14T07:01:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:47 crc kubenswrapper[4870]: I1014 07:01:47.787131 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:47 crc kubenswrapper[4870]: I1014 07:01:47.787187 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:47 crc kubenswrapper[4870]: I1014 07:01:47.787204 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:47 crc kubenswrapper[4870]: I1014 07:01:47.787227 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:47 crc kubenswrapper[4870]: I1014 07:01:47.787242 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:47Z","lastTransitionTime":"2025-10-14T07:01:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:47 crc kubenswrapper[4870]: I1014 07:01:47.890062 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:47 crc kubenswrapper[4870]: I1014 07:01:47.890424 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:47 crc kubenswrapper[4870]: I1014 07:01:47.890628 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:47 crc kubenswrapper[4870]: I1014 07:01:47.890737 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:47 crc kubenswrapper[4870]: I1014 07:01:47.890822 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:47Z","lastTransitionTime":"2025-10-14T07:01:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:47 crc kubenswrapper[4870]: I1014 07:01:47.994072 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:47 crc kubenswrapper[4870]: I1014 07:01:47.994133 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:47 crc kubenswrapper[4870]: I1014 07:01:47.994153 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:47 crc kubenswrapper[4870]: I1014 07:01:47.994186 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:47 crc kubenswrapper[4870]: I1014 07:01:47.994209 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:47Z","lastTransitionTime":"2025-10-14T07:01:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:48 crc kubenswrapper[4870]: I1014 07:01:48.033398 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 07:01:48 crc kubenswrapper[4870]: I1014 07:01:48.033433 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 07:01:48 crc kubenswrapper[4870]: E1014 07:01:48.033592 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 07:01:48 crc kubenswrapper[4870]: I1014 07:01:48.033398 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 07:01:48 crc kubenswrapper[4870]: E1014 07:01:48.033750 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 07:01:48 crc kubenswrapper[4870]: E1014 07:01:48.033935 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 07:01:48 crc kubenswrapper[4870]: I1014 07:01:48.097408 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:48 crc kubenswrapper[4870]: I1014 07:01:48.097572 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:48 crc kubenswrapper[4870]: I1014 07:01:48.097596 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:48 crc kubenswrapper[4870]: I1014 07:01:48.097625 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:48 crc kubenswrapper[4870]: I1014 07:01:48.097645 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:48Z","lastTransitionTime":"2025-10-14T07:01:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:48 crc kubenswrapper[4870]: I1014 07:01:48.201717 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:48 crc kubenswrapper[4870]: I1014 07:01:48.201807 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:48 crc kubenswrapper[4870]: I1014 07:01:48.201827 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:48 crc kubenswrapper[4870]: I1014 07:01:48.201861 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:48 crc kubenswrapper[4870]: I1014 07:01:48.201883 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:48Z","lastTransitionTime":"2025-10-14T07:01:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:48 crc kubenswrapper[4870]: I1014 07:01:48.305783 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:48 crc kubenswrapper[4870]: I1014 07:01:48.305850 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:48 crc kubenswrapper[4870]: I1014 07:01:48.305873 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:48 crc kubenswrapper[4870]: I1014 07:01:48.305902 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:48 crc kubenswrapper[4870]: I1014 07:01:48.305924 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:48Z","lastTransitionTime":"2025-10-14T07:01:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:48 crc kubenswrapper[4870]: I1014 07:01:48.408952 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:48 crc kubenswrapper[4870]: I1014 07:01:48.409016 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:48 crc kubenswrapper[4870]: I1014 07:01:48.409042 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:48 crc kubenswrapper[4870]: I1014 07:01:48.409066 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:48 crc kubenswrapper[4870]: I1014 07:01:48.409080 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:48Z","lastTransitionTime":"2025-10-14T07:01:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:48 crc kubenswrapper[4870]: I1014 07:01:48.470477 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7d7a4f41-388a-4320-b896-43c17ff13da6-metrics-certs\") pod \"network-metrics-daemon-zc5j6\" (UID: \"7d7a4f41-388a-4320-b896-43c17ff13da6\") " pod="openshift-multus/network-metrics-daemon-zc5j6" Oct 14 07:01:48 crc kubenswrapper[4870]: E1014 07:01:48.470665 4870 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 14 07:01:48 crc kubenswrapper[4870]: E1014 07:01:48.470767 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7d7a4f41-388a-4320-b896-43c17ff13da6-metrics-certs podName:7d7a4f41-388a-4320-b896-43c17ff13da6 nodeName:}" failed. No retries permitted until 2025-10-14 07:01:56.470742452 +0000 UTC m=+52.168102863 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/7d7a4f41-388a-4320-b896-43c17ff13da6-metrics-certs") pod "network-metrics-daemon-zc5j6" (UID: "7d7a4f41-388a-4320-b896-43c17ff13da6") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 14 07:01:48 crc kubenswrapper[4870]: I1014 07:01:48.512376 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:48 crc kubenswrapper[4870]: I1014 07:01:48.512471 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:48 crc kubenswrapper[4870]: I1014 07:01:48.512492 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:48 crc kubenswrapper[4870]: I1014 07:01:48.512519 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:48 crc kubenswrapper[4870]: I1014 07:01:48.512536 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:48Z","lastTransitionTime":"2025-10-14T07:01:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:48 crc kubenswrapper[4870]: I1014 07:01:48.614643 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:48 crc kubenswrapper[4870]: I1014 07:01:48.614686 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:48 crc kubenswrapper[4870]: I1014 07:01:48.614698 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:48 crc kubenswrapper[4870]: I1014 07:01:48.614715 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:48 crc kubenswrapper[4870]: I1014 07:01:48.614727 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:48Z","lastTransitionTime":"2025-10-14T07:01:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:48 crc kubenswrapper[4870]: I1014 07:01:48.717376 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:48 crc kubenswrapper[4870]: I1014 07:01:48.717478 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:48 crc kubenswrapper[4870]: I1014 07:01:48.717499 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:48 crc kubenswrapper[4870]: I1014 07:01:48.717525 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:48 crc kubenswrapper[4870]: I1014 07:01:48.717552 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:48Z","lastTransitionTime":"2025-10-14T07:01:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:48 crc kubenswrapper[4870]: I1014 07:01:48.820610 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:48 crc kubenswrapper[4870]: I1014 07:01:48.820667 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:48 crc kubenswrapper[4870]: I1014 07:01:48.820683 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:48 crc kubenswrapper[4870]: I1014 07:01:48.820705 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:48 crc kubenswrapper[4870]: I1014 07:01:48.820717 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:48Z","lastTransitionTime":"2025-10-14T07:01:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:48 crc kubenswrapper[4870]: I1014 07:01:48.924490 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:48 crc kubenswrapper[4870]: I1014 07:01:48.924654 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:48 crc kubenswrapper[4870]: I1014 07:01:48.924689 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:48 crc kubenswrapper[4870]: I1014 07:01:48.924722 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:48 crc kubenswrapper[4870]: I1014 07:01:48.924741 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:48Z","lastTransitionTime":"2025-10-14T07:01:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:49 crc kubenswrapper[4870]: I1014 07:01:49.028403 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:49 crc kubenswrapper[4870]: I1014 07:01:49.028480 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:49 crc kubenswrapper[4870]: I1014 07:01:49.028494 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:49 crc kubenswrapper[4870]: I1014 07:01:49.028517 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:49 crc kubenswrapper[4870]: I1014 07:01:49.028529 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:49Z","lastTransitionTime":"2025-10-14T07:01:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:49 crc kubenswrapper[4870]: I1014 07:01:49.033234 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zc5j6" Oct 14 07:01:49 crc kubenswrapper[4870]: E1014 07:01:49.033481 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zc5j6" podUID="7d7a4f41-388a-4320-b896-43c17ff13da6" Oct 14 07:01:49 crc kubenswrapper[4870]: I1014 07:01:49.134815 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:49 crc kubenswrapper[4870]: I1014 07:01:49.134890 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:49 crc kubenswrapper[4870]: I1014 07:01:49.134910 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:49 crc kubenswrapper[4870]: I1014 07:01:49.134933 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:49 crc kubenswrapper[4870]: I1014 07:01:49.134955 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:49Z","lastTransitionTime":"2025-10-14T07:01:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:49 crc kubenswrapper[4870]: I1014 07:01:49.238117 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:49 crc kubenswrapper[4870]: I1014 07:01:49.238170 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:49 crc kubenswrapper[4870]: I1014 07:01:49.238184 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:49 crc kubenswrapper[4870]: I1014 07:01:49.238207 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:49 crc kubenswrapper[4870]: I1014 07:01:49.238219 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:49Z","lastTransitionTime":"2025-10-14T07:01:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:49 crc kubenswrapper[4870]: I1014 07:01:49.341320 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:49 crc kubenswrapper[4870]: I1014 07:01:49.341362 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:49 crc kubenswrapper[4870]: I1014 07:01:49.341371 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:49 crc kubenswrapper[4870]: I1014 07:01:49.341389 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:49 crc kubenswrapper[4870]: I1014 07:01:49.341399 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:49Z","lastTransitionTime":"2025-10-14T07:01:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:49 crc kubenswrapper[4870]: I1014 07:01:49.445752 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:49 crc kubenswrapper[4870]: I1014 07:01:49.445840 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:49 crc kubenswrapper[4870]: I1014 07:01:49.445858 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:49 crc kubenswrapper[4870]: I1014 07:01:49.445886 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:49 crc kubenswrapper[4870]: I1014 07:01:49.445902 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:49Z","lastTransitionTime":"2025-10-14T07:01:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:49 crc kubenswrapper[4870]: I1014 07:01:49.550617 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:49 crc kubenswrapper[4870]: I1014 07:01:49.550691 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:49 crc kubenswrapper[4870]: I1014 07:01:49.550704 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:49 crc kubenswrapper[4870]: I1014 07:01:49.550728 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:49 crc kubenswrapper[4870]: I1014 07:01:49.550745 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:49Z","lastTransitionTime":"2025-10-14T07:01:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:49 crc kubenswrapper[4870]: I1014 07:01:49.653648 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:49 crc kubenswrapper[4870]: I1014 07:01:49.653684 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:49 crc kubenswrapper[4870]: I1014 07:01:49.653693 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:49 crc kubenswrapper[4870]: I1014 07:01:49.653709 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:49 crc kubenswrapper[4870]: I1014 07:01:49.653717 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:49Z","lastTransitionTime":"2025-10-14T07:01:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:49 crc kubenswrapper[4870]: I1014 07:01:49.757082 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:49 crc kubenswrapper[4870]: I1014 07:01:49.757188 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:49 crc kubenswrapper[4870]: I1014 07:01:49.757215 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:49 crc kubenswrapper[4870]: I1014 07:01:49.757248 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:49 crc kubenswrapper[4870]: I1014 07:01:49.757270 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:49Z","lastTransitionTime":"2025-10-14T07:01:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:49 crc kubenswrapper[4870]: I1014 07:01:49.859921 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:49 crc kubenswrapper[4870]: I1014 07:01:49.859961 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:49 crc kubenswrapper[4870]: I1014 07:01:49.859974 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:49 crc kubenswrapper[4870]: I1014 07:01:49.859992 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:49 crc kubenswrapper[4870]: I1014 07:01:49.860004 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:49Z","lastTransitionTime":"2025-10-14T07:01:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:49 crc kubenswrapper[4870]: I1014 07:01:49.962584 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:49 crc kubenswrapper[4870]: I1014 07:01:49.962626 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:49 crc kubenswrapper[4870]: I1014 07:01:49.962636 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:49 crc kubenswrapper[4870]: I1014 07:01:49.962653 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:49 crc kubenswrapper[4870]: I1014 07:01:49.962664 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:49Z","lastTransitionTime":"2025-10-14T07:01:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:50 crc kubenswrapper[4870]: I1014 07:01:50.033459 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 07:01:50 crc kubenswrapper[4870]: I1014 07:01:50.033472 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 07:01:50 crc kubenswrapper[4870]: I1014 07:01:50.033472 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 07:01:50 crc kubenswrapper[4870]: E1014 07:01:50.033736 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 07:01:50 crc kubenswrapper[4870]: E1014 07:01:50.033775 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 07:01:50 crc kubenswrapper[4870]: E1014 07:01:50.033592 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 07:01:50 crc kubenswrapper[4870]: I1014 07:01:50.065027 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:50 crc kubenswrapper[4870]: I1014 07:01:50.065095 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:50 crc kubenswrapper[4870]: I1014 07:01:50.065117 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:50 crc kubenswrapper[4870]: I1014 07:01:50.065147 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:50 crc kubenswrapper[4870]: I1014 07:01:50.065169 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:50Z","lastTransitionTime":"2025-10-14T07:01:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:50 crc kubenswrapper[4870]: I1014 07:01:50.167701 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:50 crc kubenswrapper[4870]: I1014 07:01:50.167748 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:50 crc kubenswrapper[4870]: I1014 07:01:50.167759 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:50 crc kubenswrapper[4870]: I1014 07:01:50.167780 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:50 crc kubenswrapper[4870]: I1014 07:01:50.167792 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:50Z","lastTransitionTime":"2025-10-14T07:01:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:50 crc kubenswrapper[4870]: I1014 07:01:50.270850 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:50 crc kubenswrapper[4870]: I1014 07:01:50.270895 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:50 crc kubenswrapper[4870]: I1014 07:01:50.270910 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:50 crc kubenswrapper[4870]: I1014 07:01:50.270934 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:50 crc kubenswrapper[4870]: I1014 07:01:50.270949 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:50Z","lastTransitionTime":"2025-10-14T07:01:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:50 crc kubenswrapper[4870]: I1014 07:01:50.374333 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:50 crc kubenswrapper[4870]: I1014 07:01:50.374434 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:50 crc kubenswrapper[4870]: I1014 07:01:50.374487 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:50 crc kubenswrapper[4870]: I1014 07:01:50.374517 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:50 crc kubenswrapper[4870]: I1014 07:01:50.374537 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:50Z","lastTransitionTime":"2025-10-14T07:01:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:50 crc kubenswrapper[4870]: I1014 07:01:50.477400 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:50 crc kubenswrapper[4870]: I1014 07:01:50.477471 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:50 crc kubenswrapper[4870]: I1014 07:01:50.477483 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:50 crc kubenswrapper[4870]: I1014 07:01:50.477501 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:50 crc kubenswrapper[4870]: I1014 07:01:50.477516 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:50Z","lastTransitionTime":"2025-10-14T07:01:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:50 crc kubenswrapper[4870]: I1014 07:01:50.580649 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:50 crc kubenswrapper[4870]: I1014 07:01:50.580704 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:50 crc kubenswrapper[4870]: I1014 07:01:50.580715 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:50 crc kubenswrapper[4870]: I1014 07:01:50.580733 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:50 crc kubenswrapper[4870]: I1014 07:01:50.580744 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:50Z","lastTransitionTime":"2025-10-14T07:01:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:50 crc kubenswrapper[4870]: I1014 07:01:50.684262 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:50 crc kubenswrapper[4870]: I1014 07:01:50.684325 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:50 crc kubenswrapper[4870]: I1014 07:01:50.684342 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:50 crc kubenswrapper[4870]: I1014 07:01:50.684367 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:50 crc kubenswrapper[4870]: I1014 07:01:50.684386 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:50Z","lastTransitionTime":"2025-10-14T07:01:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:50 crc kubenswrapper[4870]: I1014 07:01:50.787611 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:50 crc kubenswrapper[4870]: I1014 07:01:50.787653 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:50 crc kubenswrapper[4870]: I1014 07:01:50.787666 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:50 crc kubenswrapper[4870]: I1014 07:01:50.787683 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:50 crc kubenswrapper[4870]: I1014 07:01:50.787695 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:50Z","lastTransitionTime":"2025-10-14T07:01:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:50 crc kubenswrapper[4870]: I1014 07:01:50.893529 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:50 crc kubenswrapper[4870]: I1014 07:01:50.893622 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:50 crc kubenswrapper[4870]: I1014 07:01:50.893639 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:50 crc kubenswrapper[4870]: I1014 07:01:50.893665 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:50 crc kubenswrapper[4870]: I1014 07:01:50.893688 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:50Z","lastTransitionTime":"2025-10-14T07:01:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:50 crc kubenswrapper[4870]: I1014 07:01:50.996882 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:50 crc kubenswrapper[4870]: I1014 07:01:50.996926 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:50 crc kubenswrapper[4870]: I1014 07:01:50.996939 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:50 crc kubenswrapper[4870]: I1014 07:01:50.996958 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:50 crc kubenswrapper[4870]: I1014 07:01:50.996970 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:50Z","lastTransitionTime":"2025-10-14T07:01:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:51 crc kubenswrapper[4870]: I1014 07:01:51.033690 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zc5j6" Oct 14 07:01:51 crc kubenswrapper[4870]: E1014 07:01:51.033879 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zc5j6" podUID="7d7a4f41-388a-4320-b896-43c17ff13da6" Oct 14 07:01:51 crc kubenswrapper[4870]: I1014 07:01:51.099353 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:51 crc kubenswrapper[4870]: I1014 07:01:51.099390 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:51 crc kubenswrapper[4870]: I1014 07:01:51.099398 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:51 crc kubenswrapper[4870]: I1014 07:01:51.099413 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:51 crc kubenswrapper[4870]: I1014 07:01:51.099421 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:51Z","lastTransitionTime":"2025-10-14T07:01:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:51 crc kubenswrapper[4870]: I1014 07:01:51.202902 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:51 crc kubenswrapper[4870]: I1014 07:01:51.202977 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:51 crc kubenswrapper[4870]: I1014 07:01:51.202996 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:51 crc kubenswrapper[4870]: I1014 07:01:51.203019 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:51 crc kubenswrapper[4870]: I1014 07:01:51.203034 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:51Z","lastTransitionTime":"2025-10-14T07:01:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:51 crc kubenswrapper[4870]: I1014 07:01:51.306898 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:51 crc kubenswrapper[4870]: I1014 07:01:51.306956 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:51 crc kubenswrapper[4870]: I1014 07:01:51.306973 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:51 crc kubenswrapper[4870]: I1014 07:01:51.306996 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:51 crc kubenswrapper[4870]: I1014 07:01:51.307011 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:51Z","lastTransitionTime":"2025-10-14T07:01:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:51 crc kubenswrapper[4870]: I1014 07:01:51.409929 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:51 crc kubenswrapper[4870]: I1014 07:01:51.409960 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:51 crc kubenswrapper[4870]: I1014 07:01:51.409970 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:51 crc kubenswrapper[4870]: I1014 07:01:51.409982 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:51 crc kubenswrapper[4870]: I1014 07:01:51.409991 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:51Z","lastTransitionTime":"2025-10-14T07:01:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:51 crc kubenswrapper[4870]: I1014 07:01:51.511570 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:51 crc kubenswrapper[4870]: I1014 07:01:51.511627 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:51 crc kubenswrapper[4870]: I1014 07:01:51.511639 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:51 crc kubenswrapper[4870]: I1014 07:01:51.511660 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:51 crc kubenswrapper[4870]: I1014 07:01:51.511676 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:51Z","lastTransitionTime":"2025-10-14T07:01:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:51 crc kubenswrapper[4870]: I1014 07:01:51.614638 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:51 crc kubenswrapper[4870]: I1014 07:01:51.614694 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:51 crc kubenswrapper[4870]: I1014 07:01:51.614708 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:51 crc kubenswrapper[4870]: I1014 07:01:51.614771 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:51 crc kubenswrapper[4870]: I1014 07:01:51.614790 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:51Z","lastTransitionTime":"2025-10-14T07:01:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:51 crc kubenswrapper[4870]: I1014 07:01:51.717533 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:51 crc kubenswrapper[4870]: I1014 07:01:51.717568 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:51 crc kubenswrapper[4870]: I1014 07:01:51.717577 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:51 crc kubenswrapper[4870]: I1014 07:01:51.717592 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:51 crc kubenswrapper[4870]: I1014 07:01:51.717605 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:51Z","lastTransitionTime":"2025-10-14T07:01:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:51 crc kubenswrapper[4870]: I1014 07:01:51.823796 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:51 crc kubenswrapper[4870]: I1014 07:01:51.823878 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:51 crc kubenswrapper[4870]: I1014 07:01:51.823905 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:51 crc kubenswrapper[4870]: I1014 07:01:51.823944 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:51 crc kubenswrapper[4870]: I1014 07:01:51.823982 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:51Z","lastTransitionTime":"2025-10-14T07:01:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:51 crc kubenswrapper[4870]: I1014 07:01:51.927974 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:51 crc kubenswrapper[4870]: I1014 07:01:51.928029 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:51 crc kubenswrapper[4870]: I1014 07:01:51.928091 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:51 crc kubenswrapper[4870]: I1014 07:01:51.928111 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:51 crc kubenswrapper[4870]: I1014 07:01:51.928121 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:51Z","lastTransitionTime":"2025-10-14T07:01:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:52 crc kubenswrapper[4870]: I1014 07:01:52.030824 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:52 crc kubenswrapper[4870]: I1014 07:01:52.030890 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:52 crc kubenswrapper[4870]: I1014 07:01:52.030909 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:52 crc kubenswrapper[4870]: I1014 07:01:52.030937 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:52 crc kubenswrapper[4870]: I1014 07:01:52.030956 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:52Z","lastTransitionTime":"2025-10-14T07:01:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:52 crc kubenswrapper[4870]: I1014 07:01:52.033314 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 07:01:52 crc kubenswrapper[4870]: I1014 07:01:52.033355 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 07:01:52 crc kubenswrapper[4870]: E1014 07:01:52.033544 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 07:01:52 crc kubenswrapper[4870]: E1014 07:01:52.033632 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 07:01:52 crc kubenswrapper[4870]: I1014 07:01:52.033887 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 07:01:52 crc kubenswrapper[4870]: E1014 07:01:52.034069 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 07:01:52 crc kubenswrapper[4870]: I1014 07:01:52.135383 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:52 crc kubenswrapper[4870]: I1014 07:01:52.135467 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:52 crc kubenswrapper[4870]: I1014 07:01:52.135484 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:52 crc kubenswrapper[4870]: I1014 07:01:52.135507 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:52 crc kubenswrapper[4870]: I1014 07:01:52.135526 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:52Z","lastTransitionTime":"2025-10-14T07:01:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:52 crc kubenswrapper[4870]: I1014 07:01:52.239515 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:52 crc kubenswrapper[4870]: I1014 07:01:52.239570 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:52 crc kubenswrapper[4870]: I1014 07:01:52.239586 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:52 crc kubenswrapper[4870]: I1014 07:01:52.239603 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:52 crc kubenswrapper[4870]: I1014 07:01:52.239614 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:52Z","lastTransitionTime":"2025-10-14T07:01:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:52 crc kubenswrapper[4870]: I1014 07:01:52.343212 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:52 crc kubenswrapper[4870]: I1014 07:01:52.343263 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:52 crc kubenswrapper[4870]: I1014 07:01:52.343277 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:52 crc kubenswrapper[4870]: I1014 07:01:52.343299 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:52 crc kubenswrapper[4870]: I1014 07:01:52.343319 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:52Z","lastTransitionTime":"2025-10-14T07:01:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:52 crc kubenswrapper[4870]: I1014 07:01:52.447187 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:52 crc kubenswrapper[4870]: I1014 07:01:52.447272 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:52 crc kubenswrapper[4870]: I1014 07:01:52.447291 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:52 crc kubenswrapper[4870]: I1014 07:01:52.447327 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:52 crc kubenswrapper[4870]: I1014 07:01:52.447347 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:52Z","lastTransitionTime":"2025-10-14T07:01:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:52 crc kubenswrapper[4870]: I1014 07:01:52.550916 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:52 crc kubenswrapper[4870]: I1014 07:01:52.551011 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:52 crc kubenswrapper[4870]: I1014 07:01:52.551042 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:52 crc kubenswrapper[4870]: I1014 07:01:52.551080 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:52 crc kubenswrapper[4870]: I1014 07:01:52.551105 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:52Z","lastTransitionTime":"2025-10-14T07:01:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:52 crc kubenswrapper[4870]: I1014 07:01:52.654537 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:52 crc kubenswrapper[4870]: I1014 07:01:52.654622 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:52 crc kubenswrapper[4870]: I1014 07:01:52.654639 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:52 crc kubenswrapper[4870]: I1014 07:01:52.654665 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:52 crc kubenswrapper[4870]: I1014 07:01:52.654684 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:52Z","lastTransitionTime":"2025-10-14T07:01:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:52 crc kubenswrapper[4870]: I1014 07:01:52.734940 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:52 crc kubenswrapper[4870]: I1014 07:01:52.734986 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:52 crc kubenswrapper[4870]: I1014 07:01:52.735005 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:52 crc kubenswrapper[4870]: I1014 07:01:52.735030 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:52 crc kubenswrapper[4870]: I1014 07:01:52.735048 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:52Z","lastTransitionTime":"2025-10-14T07:01:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:52 crc kubenswrapper[4870]: E1014 07:01:52.758274 4870 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:01:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:01:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:52Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:01:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:01:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:52Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"56448e6b-6b3a-4bb9-91f8-2f0a9b44e9aa\\\",\\\"systemUUID\\\":\\\"0adb347d-b90d-4e9f-9962-f1bd97980aed\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:52Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:52 crc kubenswrapper[4870]: I1014 07:01:52.764650 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:52 crc kubenswrapper[4870]: I1014 07:01:52.764714 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:52 crc kubenswrapper[4870]: I1014 07:01:52.764732 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:52 crc kubenswrapper[4870]: I1014 07:01:52.764753 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:52 crc kubenswrapper[4870]: I1014 07:01:52.764769 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:52Z","lastTransitionTime":"2025-10-14T07:01:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:52 crc kubenswrapper[4870]: E1014 07:01:52.790779 4870 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:01:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:01:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:52Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:01:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:01:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:52Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"56448e6b-6b3a-4bb9-91f8-2f0a9b44e9aa\\\",\\\"systemUUID\\\":\\\"0adb347d-b90d-4e9f-9962-f1bd97980aed\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:52Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:52 crc kubenswrapper[4870]: I1014 07:01:52.797302 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:52 crc kubenswrapper[4870]: I1014 07:01:52.797342 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:52 crc kubenswrapper[4870]: I1014 07:01:52.797352 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:52 crc kubenswrapper[4870]: I1014 07:01:52.797367 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:52 crc kubenswrapper[4870]: I1014 07:01:52.797378 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:52Z","lastTransitionTime":"2025-10-14T07:01:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:52 crc kubenswrapper[4870]: E1014 07:01:52.818020 4870 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:01:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:01:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:52Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:01:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:01:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:52Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"56448e6b-6b3a-4bb9-91f8-2f0a9b44e9aa\\\",\\\"systemUUID\\\":\\\"0adb347d-b90d-4e9f-9962-f1bd97980aed\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:52Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:52 crc kubenswrapper[4870]: I1014 07:01:52.824088 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:52 crc kubenswrapper[4870]: I1014 07:01:52.824157 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:52 crc kubenswrapper[4870]: I1014 07:01:52.824178 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:52 crc kubenswrapper[4870]: I1014 07:01:52.824204 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:52 crc kubenswrapper[4870]: I1014 07:01:52.824222 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:52Z","lastTransitionTime":"2025-10-14T07:01:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:52 crc kubenswrapper[4870]: E1014 07:01:52.846907 4870 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:01:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:01:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:52Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:01:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:01:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:52Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"56448e6b-6b3a-4bb9-91f8-2f0a9b44e9aa\\\",\\\"systemUUID\\\":\\\"0adb347d-b90d-4e9f-9962-f1bd97980aed\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:52Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:52 crc kubenswrapper[4870]: I1014 07:01:52.851873 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:52 crc kubenswrapper[4870]: I1014 07:01:52.851913 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:52 crc kubenswrapper[4870]: I1014 07:01:52.851924 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:52 crc kubenswrapper[4870]: I1014 07:01:52.851941 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:52 crc kubenswrapper[4870]: I1014 07:01:52.851954 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:52Z","lastTransitionTime":"2025-10-14T07:01:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:52 crc kubenswrapper[4870]: E1014 07:01:52.871959 4870 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:01:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:01:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:52Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:01:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:01:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:52Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"56448e6b-6b3a-4bb9-91f8-2f0a9b44e9aa\\\",\\\"systemUUID\\\":\\\"0adb347d-b90d-4e9f-9962-f1bd97980aed\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:52Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:52 crc kubenswrapper[4870]: E1014 07:01:52.872076 4870 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 14 07:01:52 crc kubenswrapper[4870]: I1014 07:01:52.874054 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:52 crc kubenswrapper[4870]: I1014 07:01:52.874093 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:52 crc kubenswrapper[4870]: I1014 07:01:52.874103 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:52 crc kubenswrapper[4870]: I1014 07:01:52.874116 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:52 crc kubenswrapper[4870]: I1014 07:01:52.874127 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:52Z","lastTransitionTime":"2025-10-14T07:01:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:52 crc kubenswrapper[4870]: I1014 07:01:52.976411 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:52 crc kubenswrapper[4870]: I1014 07:01:52.976512 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:52 crc kubenswrapper[4870]: I1014 07:01:52.976603 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:52 crc kubenswrapper[4870]: I1014 07:01:52.976632 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:52 crc kubenswrapper[4870]: I1014 07:01:52.976649 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:52Z","lastTransitionTime":"2025-10-14T07:01:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:53 crc kubenswrapper[4870]: I1014 07:01:53.033842 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zc5j6" Oct 14 07:01:53 crc kubenswrapper[4870]: E1014 07:01:53.034099 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zc5j6" podUID="7d7a4f41-388a-4320-b896-43c17ff13da6" Oct 14 07:01:53 crc kubenswrapper[4870]: I1014 07:01:53.079648 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:53 crc kubenswrapper[4870]: I1014 07:01:53.079953 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:53 crc kubenswrapper[4870]: I1014 07:01:53.080062 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:53 crc kubenswrapper[4870]: I1014 07:01:53.080170 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:53 crc kubenswrapper[4870]: I1014 07:01:53.080280 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:53Z","lastTransitionTime":"2025-10-14T07:01:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:53 crc kubenswrapper[4870]: I1014 07:01:53.182814 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:53 crc kubenswrapper[4870]: I1014 07:01:53.182847 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:53 crc kubenswrapper[4870]: I1014 07:01:53.182858 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:53 crc kubenswrapper[4870]: I1014 07:01:53.182874 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:53 crc kubenswrapper[4870]: I1014 07:01:53.182886 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:53Z","lastTransitionTime":"2025-10-14T07:01:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:53 crc kubenswrapper[4870]: I1014 07:01:53.286130 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:53 crc kubenswrapper[4870]: I1014 07:01:53.286190 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:53 crc kubenswrapper[4870]: I1014 07:01:53.286206 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:53 crc kubenswrapper[4870]: I1014 07:01:53.286228 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:53 crc kubenswrapper[4870]: I1014 07:01:53.286244 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:53Z","lastTransitionTime":"2025-10-14T07:01:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:53 crc kubenswrapper[4870]: I1014 07:01:53.389287 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:53 crc kubenswrapper[4870]: I1014 07:01:53.389648 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:53 crc kubenswrapper[4870]: I1014 07:01:53.389753 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:53 crc kubenswrapper[4870]: I1014 07:01:53.389850 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:53 crc kubenswrapper[4870]: I1014 07:01:53.389951 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:53Z","lastTransitionTime":"2025-10-14T07:01:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:53 crc kubenswrapper[4870]: I1014 07:01:53.492861 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:53 crc kubenswrapper[4870]: I1014 07:01:53.492930 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:53 crc kubenswrapper[4870]: I1014 07:01:53.492951 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:53 crc kubenswrapper[4870]: I1014 07:01:53.492980 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:53 crc kubenswrapper[4870]: I1014 07:01:53.492997 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:53Z","lastTransitionTime":"2025-10-14T07:01:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:53 crc kubenswrapper[4870]: I1014 07:01:53.596369 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:53 crc kubenswrapper[4870]: I1014 07:01:53.596425 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:53 crc kubenswrapper[4870]: I1014 07:01:53.596483 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:53 crc kubenswrapper[4870]: I1014 07:01:53.596513 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:53 crc kubenswrapper[4870]: I1014 07:01:53.596531 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:53Z","lastTransitionTime":"2025-10-14T07:01:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:53 crc kubenswrapper[4870]: I1014 07:01:53.699537 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:53 crc kubenswrapper[4870]: I1014 07:01:53.699569 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:53 crc kubenswrapper[4870]: I1014 07:01:53.699578 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:53 crc kubenswrapper[4870]: I1014 07:01:53.699592 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:53 crc kubenswrapper[4870]: I1014 07:01:53.699600 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:53Z","lastTransitionTime":"2025-10-14T07:01:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:53 crc kubenswrapper[4870]: I1014 07:01:53.803051 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:53 crc kubenswrapper[4870]: I1014 07:01:53.803095 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:53 crc kubenswrapper[4870]: I1014 07:01:53.803106 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:53 crc kubenswrapper[4870]: I1014 07:01:53.803124 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:53 crc kubenswrapper[4870]: I1014 07:01:53.803135 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:53Z","lastTransitionTime":"2025-10-14T07:01:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:53 crc kubenswrapper[4870]: I1014 07:01:53.907242 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:53 crc kubenswrapper[4870]: I1014 07:01:53.907958 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:53 crc kubenswrapper[4870]: I1014 07:01:53.908065 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:53 crc kubenswrapper[4870]: I1014 07:01:53.908159 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:53 crc kubenswrapper[4870]: I1014 07:01:53.908241 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:53Z","lastTransitionTime":"2025-10-14T07:01:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:54 crc kubenswrapper[4870]: I1014 07:01:54.011204 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:54 crc kubenswrapper[4870]: I1014 07:01:54.011366 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:54 crc kubenswrapper[4870]: I1014 07:01:54.011387 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:54 crc kubenswrapper[4870]: I1014 07:01:54.011406 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:54 crc kubenswrapper[4870]: I1014 07:01:54.011419 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:54Z","lastTransitionTime":"2025-10-14T07:01:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:54 crc kubenswrapper[4870]: I1014 07:01:54.033739 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 07:01:54 crc kubenswrapper[4870]: I1014 07:01:54.033813 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 07:01:54 crc kubenswrapper[4870]: E1014 07:01:54.033895 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 07:01:54 crc kubenswrapper[4870]: E1014 07:01:54.033957 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 07:01:54 crc kubenswrapper[4870]: I1014 07:01:54.033739 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 07:01:54 crc kubenswrapper[4870]: E1014 07:01:54.034054 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 07:01:54 crc kubenswrapper[4870]: I1014 07:01:54.114762 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:54 crc kubenswrapper[4870]: I1014 07:01:54.115037 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:54 crc kubenswrapper[4870]: I1014 07:01:54.115122 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:54 crc kubenswrapper[4870]: I1014 07:01:54.115205 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:54 crc kubenswrapper[4870]: I1014 07:01:54.115294 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:54Z","lastTransitionTime":"2025-10-14T07:01:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:54 crc kubenswrapper[4870]: I1014 07:01:54.217654 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:54 crc kubenswrapper[4870]: I1014 07:01:54.217692 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:54 crc kubenswrapper[4870]: I1014 07:01:54.217704 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:54 crc kubenswrapper[4870]: I1014 07:01:54.217721 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:54 crc kubenswrapper[4870]: I1014 07:01:54.217732 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:54Z","lastTransitionTime":"2025-10-14T07:01:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:54 crc kubenswrapper[4870]: I1014 07:01:54.320584 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:54 crc kubenswrapper[4870]: I1014 07:01:54.320656 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:54 crc kubenswrapper[4870]: I1014 07:01:54.320682 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:54 crc kubenswrapper[4870]: I1014 07:01:54.320715 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:54 crc kubenswrapper[4870]: I1014 07:01:54.320740 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:54Z","lastTransitionTime":"2025-10-14T07:01:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:54 crc kubenswrapper[4870]: I1014 07:01:54.423700 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:54 crc kubenswrapper[4870]: I1014 07:01:54.423744 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:54 crc kubenswrapper[4870]: I1014 07:01:54.423754 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:54 crc kubenswrapper[4870]: I1014 07:01:54.423771 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:54 crc kubenswrapper[4870]: I1014 07:01:54.423782 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:54Z","lastTransitionTime":"2025-10-14T07:01:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:54 crc kubenswrapper[4870]: I1014 07:01:54.519311 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 14 07:01:54 crc kubenswrapper[4870]: I1014 07:01:54.526953 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:54 crc kubenswrapper[4870]: I1014 07:01:54.527005 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:54 crc kubenswrapper[4870]: I1014 07:01:54.527024 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:54 crc kubenswrapper[4870]: I1014 07:01:54.527048 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:54 crc kubenswrapper[4870]: I1014 07:01:54.527069 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:54Z","lastTransitionTime":"2025-10-14T07:01:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:54 crc kubenswrapper[4870]: I1014 07:01:54.542923 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bcbc6dd6-aab0-4dbc-a62e-c58f06ec1337\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25ec4808b09adc44488530b3c4e29fd7fec46b3fcfa31d4449ec4a6dbc880ec8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f3f5de86fb11231a19034a18e9d7f31476c162e6567d0c136ead8bbbd900d6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d57d938d9d9c92198ae5b3857536ece36012fbfbe8cc8cfd245d1386d39b0be5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77a3247a7f65709543d7fd374b2ab69d7d793537d4132d7e22261188d6e58544\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:05Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:54Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:54 crc kubenswrapper[4870]: I1014 07:01:54.562853 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e539fac3d0a2ade314a4b4c1050bb27c37bcff5ba5272fa2d71b97ab5c0d7191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:54Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:54 crc kubenswrapper[4870]: I1014 07:01:54.577665 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:54Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:54 crc kubenswrapper[4870]: I1014 07:01:54.589246 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-kz77f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d3864a91-9bf6-4540-a908-cc41aa39656c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d4ded3d86030f5ea548c8b1cc3caddbe14a45c9a3ae4af836e892880584f57d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4h8v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:28Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-kz77f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:54Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:54 crc kubenswrapper[4870]: I1014 07:01:54.604422 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:54Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:54 crc kubenswrapper[4870]: I1014 07:01:54.621002 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3eb0b397c05ffd6650607f1be4a75683e0342d3333fc096bf1cd63c7688834f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:54Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:54 crc kubenswrapper[4870]: I1014 07:01:54.629523 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:54 crc kubenswrapper[4870]: I1014 07:01:54.629569 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:54 crc kubenswrapper[4870]: I1014 07:01:54.629580 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:54 crc kubenswrapper[4870]: I1014 07:01:54.629598 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:54 crc kubenswrapper[4870]: I1014 07:01:54.629611 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:54Z","lastTransitionTime":"2025-10-14T07:01:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:54 crc kubenswrapper[4870]: I1014 07:01:54.632056 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"874a5e45-dffd-4d17-b609-b7d3ed2eab07\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d98d044836ab44a405196915ef7bf572b75113f2c81ec7e64effa63b47ad3543\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wj4ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8004752575df4a1c44efbcafe64799f09126d0f97285d92388c76a0b90e022f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wj4ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7tvc7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:54Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:54 crc kubenswrapper[4870]: I1014 07:01:54.644353 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zc5j6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d7a4f41-388a-4320-b896-43c17ff13da6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6p2wh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6p2wh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:40Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zc5j6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:54Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:54 crc kubenswrapper[4870]: I1014 07:01:54.657456 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77a0d93a-5391-4112-b502-c474f45de209\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7815af9a584f6db6660073da289f2e8c987a04ef9ddfac155bdd9a15b014fcfa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b329b570622155a55f70501b2260e278106a68ebef497914231b4ec860e0125\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba942a0e6fe5ab93a187e0b7b4bd7c440c2e7827d9b5323fc50f90b562075392\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b9018b078c5c8276243c1696860f0aa73e2d8dd917913ae9d03b30dc9b5113a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://978e8db184ba9e430232c6bb2ee530beb92567fdc71ba3c35b0bedca3492d845\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"file observer\\\\nW1014 07:01:25.870672 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1014 07:01:25.870795 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1014 07:01:25.871521 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-756714914/tls.crt::/tmp/serving-cert-756714914/tls.key\\\\\\\"\\\\nI1014 07:01:26.460037 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1014 07:01:26.464876 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1014 07:01:26.464902 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1014 07:01:26.464925 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1014 07:01:26.464931 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1014 07:01:26.481374 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1014 07:01:26.481400 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:01:26.481405 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:01:26.481410 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1014 07:01:26.481414 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1014 07:01:26.481417 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1014 07:01:26.481420 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1014 07:01:26.481610 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1014 07:01:26.482872 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bd0044a247bdc5ca4cab7294103af3270b76cde61f3109b85e83ed050ca53b6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82d3c2214e0be679d2827594b159ace362151318689078bc49d3158f42899636\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82d3c2214e0be679d2827594b159ace362151318689078bc49d3158f42899636\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:05Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:54Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:54 crc kubenswrapper[4870]: I1014 07:01:54.670190 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:54Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:54 crc kubenswrapper[4870]: I1014 07:01:54.681756 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e26fe49ec2045d20a0cb2dd17e96e79b6eb43c641eb184c688cc472ca5db39d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aea115f7ee71c700b218fe15c928af7b55e6babe70e65d67f125d87be1fd2e2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:54Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:54 crc kubenswrapper[4870]: I1014 07:01:54.701807 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-v7br2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dcaef0e-42e3-424b-bbe5-4a023efe912d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71dec55e859b11a081dbb90d7d6cf6b4846cf591382303d8329003ff50c71f2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e72cff0723e2ce520340244b4e597d8016c58e1b071c4b087f2f6cd718741594\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc9d4db944f837289d1923176d60cfc4a566a282c2d2e3a703e220df884703c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d16ed1c65925549dcb69ab02c3013e25e3f123a83179f1af3a8e08bf91a59896\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed59243b59c39ce82bf9bc8f33c0a8b84854846afff6a89f5d37ffc97a096b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3b6d20cf459b1f6d3aff5b9290f9391909baf27ee40c49e2b0cd2d5cfe889ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a05f64fdac0425bbf7ea39d9d9975279c4a68dc9a93e27f58f2d7017a328f428\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a05f64fdac0425bbf7ea39d9d9975279c4a68dc9a93e27f58f2d7017a328f428\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-14T07:01:41Z\\\",\\\"message\\\":\\\"rvices do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-console/downloads_TCP_cluster\\\\\\\", UUID:\\\\\\\"2ead45b3-c313-4fbc-a7bc-2b3c4ffd610c\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-console/downloads\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-console/downloads_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-console/downloads\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.213\\\\\\\", Port:80, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nF1014 07:01:41.689119 6388 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:40Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-v7br2_openshift-ovn-kubernetes(8dcaef0e-42e3-424b-bbe5-4a023efe912d)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fe3468cdd612fc4378e1f6f55a2a9d00274f7ddb65cb57b2a368dfc065cfb4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ba24c2e067b9f3a14d9d018b66300b54d866530d30174fbdb3c899299575d28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ba24c2e067b9f3a14d9d018b66300b54d866530d30174fbdb3c899299575d28\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-v7br2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:54Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:54 crc kubenswrapper[4870]: I1014 07:01:54.720333 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-cdjmj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"612e7ab9-bda4-45d8-bc50-ff1afb68aace\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7776352d62b814fc63d01e4c16ad13f878b0597eb37a370bf9da0a88abab6cfa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://834444b5350ee12df4e4652c3009880d8d69e1edf8ba0c0156ab96fb2bbe28e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://834444b5350ee12df4e4652c3009880d8d69e1edf8ba0c0156ab96fb2bbe28e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c1972e7b2f75790db22a07a078eee97072bffbdb7f8e594bf2be5f800632a7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c1972e7b2f75790db22a07a078eee97072bffbdb7f8e594bf2be5f800632a7b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b124b42a44c6a23da1d92b9a5326987eb74430f80489a71e8913829475654ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b124b42a44c6a23da1d92b9a5326987eb74430f80489a71e8913829475654ab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62d0deb6b20ea65fbcbaf94c2fc14d69adb22e7a46cc043ccd4def2668ddd4b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62d0deb6b20ea65fbcbaf94c2fc14d69adb22e7a46cc043ccd4def2668ddd4b7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7f0d7cf29ca737d0effe58731a8b890379527c97f78c144d4eebb910260fadb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7f0d7cf29ca737d0effe58731a8b890379527c97f78c144d4eebb910260fadb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5959abb7c8e7ab32bed3aca709ae0334e8400f991ba724246c9fda56cbed822\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5959abb7c8e7ab32bed3aca709ae0334e8400f991ba724246c9fda56cbed822\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-cdjmj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:54Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:54 crc kubenswrapper[4870]: I1014 07:01:54.733349 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:54 crc kubenswrapper[4870]: I1014 07:01:54.733622 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:54 crc kubenswrapper[4870]: I1014 07:01:54.733830 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:54 crc kubenswrapper[4870]: I1014 07:01:54.734027 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:54 crc kubenswrapper[4870]: I1014 07:01:54.734213 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:54Z","lastTransitionTime":"2025-10-14T07:01:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:54 crc kubenswrapper[4870]: I1014 07:01:54.738915 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fdxjh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a3a430a-b540-4a8c-adad-b2893dbf9898\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93e41d5e927eed7ec62eb9c3607cc804cd8999556663363fd123b2a784b00ff1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k787j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fdxjh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:54Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:54 crc kubenswrapper[4870]: I1014 07:01:54.754025 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-khtq8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28deeea6-a1ef-4a4a-a377-d954ccd87de9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://296bb048a1d99649beb57d609573833bf547cd4d628ebc823c6af83fe5736c9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rb79p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2b7bc4c93307e80b778f2282ced993898f1c4cfc350a7d74828b8f9dea9906d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rb79p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-khtq8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:54Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:54 crc kubenswrapper[4870]: I1014 07:01:54.774417 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33a9493e-9660-48db-b8d1-310e29ebe313\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3570643ef1b7d5cc3a542f320413b819338e0c8b7ab7f0f09cc909997a39bab8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb398668197dedc1c2a14b4ae7d6a29714f3527309ccb9c1622af841100e1d60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1d2b6d068d63bd38cdfb83824525f6d12054f6cabab5d684e03ea5f2152076c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://783ff243e395333f3fc80cfae21358c532f0c9c5cb5515586be56b90ed10d912\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a8135a3d01076e4437ae8a784b459bf8d16de57fa16ec7ce4d40c5129c9d4a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48ca3a5017ba650df99b2fc7201c3fe5abfa89d81089c5647cde1b1549f534fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48ca3a5017ba650df99b2fc7201c3fe5abfa89d81089c5647cde1b1549f534fc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://19682eca67f76bf2e06e1b80f12e21141fcbce0acb8195f0defacd1b9a10b5e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://19682eca67f76bf2e06e1b80f12e21141fcbce0acb8195f0defacd1b9a10b5e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://042aad6d5f6bee8eef77f9e6a2c265de775a69f3da72aedc68bdf312198f00fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://042aad6d5f6bee8eef77f9e6a2c265de775a69f3da72aedc68bdf312198f00fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:05Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:54Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:54 crc kubenswrapper[4870]: I1014 07:01:54.785396 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-4v2v2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4796ba6-c23b-4764-be96-2e3a3bf5e962\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf5afb0b99a7101f6d5573fb7eb925c5a1d8becc11b2e3c6f3351d80b6f0e823\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnp8b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-4v2v2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:54Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:54 crc kubenswrapper[4870]: I1014 07:01:54.837144 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:54 crc kubenswrapper[4870]: I1014 07:01:54.837251 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:54 crc kubenswrapper[4870]: I1014 07:01:54.837279 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:54 crc kubenswrapper[4870]: I1014 07:01:54.837317 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:54 crc kubenswrapper[4870]: I1014 07:01:54.837346 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:54Z","lastTransitionTime":"2025-10-14T07:01:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:54 crc kubenswrapper[4870]: I1014 07:01:54.940456 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:54 crc kubenswrapper[4870]: I1014 07:01:54.940491 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:54 crc kubenswrapper[4870]: I1014 07:01:54.940503 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:54 crc kubenswrapper[4870]: I1014 07:01:54.940518 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:54 crc kubenswrapper[4870]: I1014 07:01:54.940529 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:54Z","lastTransitionTime":"2025-10-14T07:01:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:55 crc kubenswrapper[4870]: I1014 07:01:55.033036 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zc5j6" Oct 14 07:01:55 crc kubenswrapper[4870]: E1014 07:01:55.033197 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zc5j6" podUID="7d7a4f41-388a-4320-b896-43c17ff13da6" Oct 14 07:01:55 crc kubenswrapper[4870]: I1014 07:01:55.042306 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:55 crc kubenswrapper[4870]: I1014 07:01:55.042345 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:55 crc kubenswrapper[4870]: I1014 07:01:55.042358 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:55 crc kubenswrapper[4870]: I1014 07:01:55.042376 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:55 crc kubenswrapper[4870]: I1014 07:01:55.042389 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:55Z","lastTransitionTime":"2025-10-14T07:01:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:55 crc kubenswrapper[4870]: I1014 07:01:55.051106 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:55Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:55 crc kubenswrapper[4870]: I1014 07:01:55.064690 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3eb0b397c05ffd6650607f1be4a75683e0342d3333fc096bf1cd63c7688834f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:55Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:55 crc kubenswrapper[4870]: I1014 07:01:55.080405 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"874a5e45-dffd-4d17-b609-b7d3ed2eab07\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d98d044836ab44a405196915ef7bf572b75113f2c81ec7e64effa63b47ad3543\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wj4ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8004752575df4a1c44efbcafe64799f09126d0f97285d92388c76a0b90e022f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wj4ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7tvc7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:55Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:55 crc kubenswrapper[4870]: I1014 07:01:55.092346 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zc5j6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d7a4f41-388a-4320-b896-43c17ff13da6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6p2wh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6p2wh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:40Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zc5j6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:55Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:55 crc kubenswrapper[4870]: I1014 07:01:55.107220 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77a0d93a-5391-4112-b502-c474f45de209\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7815af9a584f6db6660073da289f2e8c987a04ef9ddfac155bdd9a15b014fcfa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b329b570622155a55f70501b2260e278106a68ebef497914231b4ec860e0125\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba942a0e6fe5ab93a187e0b7b4bd7c440c2e7827d9b5323fc50f90b562075392\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b9018b078c5c8276243c1696860f0aa73e2d8dd917913ae9d03b30dc9b5113a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://978e8db184ba9e430232c6bb2ee530beb92567fdc71ba3c35b0bedca3492d845\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"file observer\\\\nW1014 07:01:25.870672 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1014 07:01:25.870795 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1014 07:01:25.871521 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-756714914/tls.crt::/tmp/serving-cert-756714914/tls.key\\\\\\\"\\\\nI1014 07:01:26.460037 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1014 07:01:26.464876 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1014 07:01:26.464902 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1014 07:01:26.464925 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1014 07:01:26.464931 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1014 07:01:26.481374 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1014 07:01:26.481400 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:01:26.481405 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:01:26.481410 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1014 07:01:26.481414 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1014 07:01:26.481417 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1014 07:01:26.481420 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1014 07:01:26.481610 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1014 07:01:26.482872 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bd0044a247bdc5ca4cab7294103af3270b76cde61f3109b85e83ed050ca53b6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82d3c2214e0be679d2827594b159ace362151318689078bc49d3158f42899636\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82d3c2214e0be679d2827594b159ace362151318689078bc49d3158f42899636\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:05Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:55Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:55 crc kubenswrapper[4870]: I1014 07:01:55.122945 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:55Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:55 crc kubenswrapper[4870]: I1014 07:01:55.139888 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e26fe49ec2045d20a0cb2dd17e96e79b6eb43c641eb184c688cc472ca5db39d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aea115f7ee71c700b218fe15c928af7b55e6babe70e65d67f125d87be1fd2e2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:55Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:55 crc kubenswrapper[4870]: I1014 07:01:55.144007 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:55 crc kubenswrapper[4870]: I1014 07:01:55.144047 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:55 crc kubenswrapper[4870]: I1014 07:01:55.144057 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:55 crc kubenswrapper[4870]: I1014 07:01:55.144076 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:55 crc kubenswrapper[4870]: I1014 07:01:55.144088 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:55Z","lastTransitionTime":"2025-10-14T07:01:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:55 crc kubenswrapper[4870]: I1014 07:01:55.158258 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-v7br2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dcaef0e-42e3-424b-bbe5-4a023efe912d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71dec55e859b11a081dbb90d7d6cf6b4846cf591382303d8329003ff50c71f2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e72cff0723e2ce520340244b4e597d8016c58e1b071c4b087f2f6cd718741594\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc9d4db944f837289d1923176d60cfc4a566a282c2d2e3a703e220df884703c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d16ed1c65925549dcb69ab02c3013e25e3f123a83179f1af3a8e08bf91a59896\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed59243b59c39ce82bf9bc8f33c0a8b84854846afff6a89f5d37ffc97a096b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3b6d20cf459b1f6d3aff5b9290f9391909baf27ee40c49e2b0cd2d5cfe889ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a05f64fdac0425bbf7ea39d9d9975279c4a68dc9a93e27f58f2d7017a328f428\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a05f64fdac0425bbf7ea39d9d9975279c4a68dc9a93e27f58f2d7017a328f428\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-14T07:01:41Z\\\",\\\"message\\\":\\\"rvices do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-console/downloads_TCP_cluster\\\\\\\", UUID:\\\\\\\"2ead45b3-c313-4fbc-a7bc-2b3c4ffd610c\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-console/downloads\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-console/downloads_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-console/downloads\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.213\\\\\\\", Port:80, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nF1014 07:01:41.689119 6388 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:40Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-v7br2_openshift-ovn-kubernetes(8dcaef0e-42e3-424b-bbe5-4a023efe912d)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fe3468cdd612fc4378e1f6f55a2a9d00274f7ddb65cb57b2a368dfc065cfb4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ba24c2e067b9f3a14d9d018b66300b54d866530d30174fbdb3c899299575d28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ba24c2e067b9f3a14d9d018b66300b54d866530d30174fbdb3c899299575d28\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-v7br2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:55Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:55 crc kubenswrapper[4870]: I1014 07:01:55.176090 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33a9493e-9660-48db-b8d1-310e29ebe313\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3570643ef1b7d5cc3a542f320413b819338e0c8b7ab7f0f09cc909997a39bab8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb398668197dedc1c2a14b4ae7d6a29714f3527309ccb9c1622af841100e1d60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1d2b6d068d63bd38cdfb83824525f6d12054f6cabab5d684e03ea5f2152076c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://783ff243e395333f3fc80cfae21358c532f0c9c5cb5515586be56b90ed10d912\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a8135a3d01076e4437ae8a784b459bf8d16de57fa16ec7ce4d40c5129c9d4a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48ca3a5017ba650df99b2fc7201c3fe5abfa89d81089c5647cde1b1549f534fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48ca3a5017ba650df99b2fc7201c3fe5abfa89d81089c5647cde1b1549f534fc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://19682eca67f76bf2e06e1b80f12e21141fcbce0acb8195f0defacd1b9a10b5e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://19682eca67f76bf2e06e1b80f12e21141fcbce0acb8195f0defacd1b9a10b5e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://042aad6d5f6bee8eef77f9e6a2c265de775a69f3da72aedc68bdf312198f00fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://042aad6d5f6bee8eef77f9e6a2c265de775a69f3da72aedc68bdf312198f00fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:05Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:55Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:55 crc kubenswrapper[4870]: I1014 07:01:55.185981 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-4v2v2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4796ba6-c23b-4764-be96-2e3a3bf5e962\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf5afb0b99a7101f6d5573fb7eb925c5a1d8becc11b2e3c6f3351d80b6f0e823\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnp8b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-4v2v2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:55Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:55 crc kubenswrapper[4870]: I1014 07:01:55.200085 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-cdjmj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"612e7ab9-bda4-45d8-bc50-ff1afb68aace\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7776352d62b814fc63d01e4c16ad13f878b0597eb37a370bf9da0a88abab6cfa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://834444b5350ee12df4e4652c3009880d8d69e1edf8ba0c0156ab96fb2bbe28e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://834444b5350ee12df4e4652c3009880d8d69e1edf8ba0c0156ab96fb2bbe28e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c1972e7b2f75790db22a07a078eee97072bffbdb7f8e594bf2be5f800632a7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c1972e7b2f75790db22a07a078eee97072bffbdb7f8e594bf2be5f800632a7b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b124b42a44c6a23da1d92b9a5326987eb74430f80489a71e8913829475654ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b124b42a44c6a23da1d92b9a5326987eb74430f80489a71e8913829475654ab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62d0deb6b20ea65fbcbaf94c2fc14d69adb22e7a46cc043ccd4def2668ddd4b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62d0deb6b20ea65fbcbaf94c2fc14d69adb22e7a46cc043ccd4def2668ddd4b7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7f0d7cf29ca737d0effe58731a8b890379527c97f78c144d4eebb910260fadb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7f0d7cf29ca737d0effe58731a8b890379527c97f78c144d4eebb910260fadb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5959abb7c8e7ab32bed3aca709ae0334e8400f991ba724246c9fda56cbed822\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5959abb7c8e7ab32bed3aca709ae0334e8400f991ba724246c9fda56cbed822\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-cdjmj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:55Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:55 crc kubenswrapper[4870]: I1014 07:01:55.213507 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fdxjh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a3a430a-b540-4a8c-adad-b2893dbf9898\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93e41d5e927eed7ec62eb9c3607cc804cd8999556663363fd123b2a784b00ff1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k787j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fdxjh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:55Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:55 crc kubenswrapper[4870]: I1014 07:01:55.224550 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-khtq8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28deeea6-a1ef-4a4a-a377-d954ccd87de9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://296bb048a1d99649beb57d609573833bf547cd4d628ebc823c6af83fe5736c9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rb79p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2b7bc4c93307e80b778f2282ced993898f1c4cfc350a7d74828b8f9dea9906d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rb79p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-khtq8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:55Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:55 crc kubenswrapper[4870]: I1014 07:01:55.236232 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bcbc6dd6-aab0-4dbc-a62e-c58f06ec1337\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25ec4808b09adc44488530b3c4e29fd7fec46b3fcfa31d4449ec4a6dbc880ec8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f3f5de86fb11231a19034a18e9d7f31476c162e6567d0c136ead8bbbd900d6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d57d938d9d9c92198ae5b3857536ece36012fbfbe8cc8cfd245d1386d39b0be5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77a3247a7f65709543d7fd374b2ab69d7d793537d4132d7e22261188d6e58544\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:05Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:55Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:55 crc kubenswrapper[4870]: I1014 07:01:55.245825 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:55 crc kubenswrapper[4870]: I1014 07:01:55.245867 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:55 crc kubenswrapper[4870]: I1014 07:01:55.245878 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:55 crc kubenswrapper[4870]: I1014 07:01:55.245892 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:55 crc kubenswrapper[4870]: I1014 07:01:55.245900 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:55Z","lastTransitionTime":"2025-10-14T07:01:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:55 crc kubenswrapper[4870]: I1014 07:01:55.251255 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e539fac3d0a2ade314a4b4c1050bb27c37bcff5ba5272fa2d71b97ab5c0d7191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:55Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:55 crc kubenswrapper[4870]: I1014 07:01:55.264345 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:55Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:55 crc kubenswrapper[4870]: I1014 07:01:55.275834 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-kz77f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d3864a91-9bf6-4540-a908-cc41aa39656c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d4ded3d86030f5ea548c8b1cc3caddbe14a45c9a3ae4af836e892880584f57d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4h8v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:28Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-kz77f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:55Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:55 crc kubenswrapper[4870]: I1014 07:01:55.347730 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:55 crc kubenswrapper[4870]: I1014 07:01:55.347775 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:55 crc kubenswrapper[4870]: I1014 07:01:55.347786 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:55 crc kubenswrapper[4870]: I1014 07:01:55.347803 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:55 crc kubenswrapper[4870]: I1014 07:01:55.347815 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:55Z","lastTransitionTime":"2025-10-14T07:01:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:55 crc kubenswrapper[4870]: I1014 07:01:55.450694 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:55 crc kubenswrapper[4870]: I1014 07:01:55.450765 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:55 crc kubenswrapper[4870]: I1014 07:01:55.450788 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:55 crc kubenswrapper[4870]: I1014 07:01:55.450816 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:55 crc kubenswrapper[4870]: I1014 07:01:55.450838 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:55Z","lastTransitionTime":"2025-10-14T07:01:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:55 crc kubenswrapper[4870]: I1014 07:01:55.553167 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:55 crc kubenswrapper[4870]: I1014 07:01:55.553235 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:55 crc kubenswrapper[4870]: I1014 07:01:55.553257 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:55 crc kubenswrapper[4870]: I1014 07:01:55.553285 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:55 crc kubenswrapper[4870]: I1014 07:01:55.553307 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:55Z","lastTransitionTime":"2025-10-14T07:01:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:55 crc kubenswrapper[4870]: I1014 07:01:55.656012 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:55 crc kubenswrapper[4870]: I1014 07:01:55.656095 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:55 crc kubenswrapper[4870]: I1014 07:01:55.656119 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:55 crc kubenswrapper[4870]: I1014 07:01:55.656148 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:55 crc kubenswrapper[4870]: I1014 07:01:55.656171 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:55Z","lastTransitionTime":"2025-10-14T07:01:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:55 crc kubenswrapper[4870]: I1014 07:01:55.759198 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:55 crc kubenswrapper[4870]: I1014 07:01:55.759279 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:55 crc kubenswrapper[4870]: I1014 07:01:55.759294 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:55 crc kubenswrapper[4870]: I1014 07:01:55.759315 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:55 crc kubenswrapper[4870]: I1014 07:01:55.759333 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:55Z","lastTransitionTime":"2025-10-14T07:01:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:55 crc kubenswrapper[4870]: I1014 07:01:55.862321 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:55 crc kubenswrapper[4870]: I1014 07:01:55.862383 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:55 crc kubenswrapper[4870]: I1014 07:01:55.862402 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:55 crc kubenswrapper[4870]: I1014 07:01:55.862431 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:55 crc kubenswrapper[4870]: I1014 07:01:55.862489 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:55Z","lastTransitionTime":"2025-10-14T07:01:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:55 crc kubenswrapper[4870]: I1014 07:01:55.965493 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:55 crc kubenswrapper[4870]: I1014 07:01:55.965554 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:55 crc kubenswrapper[4870]: I1014 07:01:55.965574 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:55 crc kubenswrapper[4870]: I1014 07:01:55.965616 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:55 crc kubenswrapper[4870]: I1014 07:01:55.965644 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:55Z","lastTransitionTime":"2025-10-14T07:01:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:56 crc kubenswrapper[4870]: I1014 07:01:56.033149 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 07:01:56 crc kubenswrapper[4870]: I1014 07:01:56.033202 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 07:01:56 crc kubenswrapper[4870]: E1014 07:01:56.033262 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 07:01:56 crc kubenswrapper[4870]: I1014 07:01:56.033325 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 07:01:56 crc kubenswrapper[4870]: E1014 07:01:56.033509 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 07:01:56 crc kubenswrapper[4870]: E1014 07:01:56.033637 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 07:01:56 crc kubenswrapper[4870]: I1014 07:01:56.034687 4870 scope.go:117] "RemoveContainer" containerID="a05f64fdac0425bbf7ea39d9d9975279c4a68dc9a93e27f58f2d7017a328f428" Oct 14 07:01:56 crc kubenswrapper[4870]: I1014 07:01:56.068399 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:56 crc kubenswrapper[4870]: I1014 07:01:56.068481 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:56 crc kubenswrapper[4870]: I1014 07:01:56.068494 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:56 crc kubenswrapper[4870]: I1014 07:01:56.068515 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:56 crc kubenswrapper[4870]: I1014 07:01:56.068529 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:56Z","lastTransitionTime":"2025-10-14T07:01:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:56 crc kubenswrapper[4870]: I1014 07:01:56.171556 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:56 crc kubenswrapper[4870]: I1014 07:01:56.171585 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:56 crc kubenswrapper[4870]: I1014 07:01:56.171594 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:56 crc kubenswrapper[4870]: I1014 07:01:56.171611 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:56 crc kubenswrapper[4870]: I1014 07:01:56.171620 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:56Z","lastTransitionTime":"2025-10-14T07:01:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:56 crc kubenswrapper[4870]: I1014 07:01:56.275101 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:56 crc kubenswrapper[4870]: I1014 07:01:56.275169 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:56 crc kubenswrapper[4870]: I1014 07:01:56.275187 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:56 crc kubenswrapper[4870]: I1014 07:01:56.275209 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:56 crc kubenswrapper[4870]: I1014 07:01:56.275226 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:56Z","lastTransitionTime":"2025-10-14T07:01:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:56 crc kubenswrapper[4870]: I1014 07:01:56.378059 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:56 crc kubenswrapper[4870]: I1014 07:01:56.378091 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:56 crc kubenswrapper[4870]: I1014 07:01:56.378101 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:56 crc kubenswrapper[4870]: I1014 07:01:56.378114 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:56 crc kubenswrapper[4870]: I1014 07:01:56.378123 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:56Z","lastTransitionTime":"2025-10-14T07:01:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:56 crc kubenswrapper[4870]: I1014 07:01:56.481158 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:56 crc kubenswrapper[4870]: I1014 07:01:56.481192 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:56 crc kubenswrapper[4870]: I1014 07:01:56.481202 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:56 crc kubenswrapper[4870]: I1014 07:01:56.481216 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:56 crc kubenswrapper[4870]: I1014 07:01:56.481225 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:56Z","lastTransitionTime":"2025-10-14T07:01:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:56 crc kubenswrapper[4870]: I1014 07:01:56.532870 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-v7br2_8dcaef0e-42e3-424b-bbe5-4a023efe912d/ovnkube-controller/1.log" Oct 14 07:01:56 crc kubenswrapper[4870]: I1014 07:01:56.535692 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-v7br2" event={"ID":"8dcaef0e-42e3-424b-bbe5-4a023efe912d","Type":"ContainerStarted","Data":"3fc9bfcb356b506d6c439adc22e4caaed4fad1c5a085744d44ebd50cf8898777"} Oct 14 07:01:56 crc kubenswrapper[4870]: I1014 07:01:56.536088 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-v7br2" Oct 14 07:01:56 crc kubenswrapper[4870]: I1014 07:01:56.546773 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"874a5e45-dffd-4d17-b609-b7d3ed2eab07\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d98d044836ab44a405196915ef7bf572b75113f2c81ec7e64effa63b47ad3543\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wj4ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8004752575df4a1c44efbcafe64799f09126d0f97285d92388c76a0b90e022f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wj4ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7tvc7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:56Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:56 crc kubenswrapper[4870]: I1014 07:01:56.558239 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zc5j6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d7a4f41-388a-4320-b896-43c17ff13da6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6p2wh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6p2wh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:40Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zc5j6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:56Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:56 crc kubenswrapper[4870]: I1014 07:01:56.564213 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7d7a4f41-388a-4320-b896-43c17ff13da6-metrics-certs\") pod \"network-metrics-daemon-zc5j6\" (UID: \"7d7a4f41-388a-4320-b896-43c17ff13da6\") " pod="openshift-multus/network-metrics-daemon-zc5j6" Oct 14 07:01:56 crc kubenswrapper[4870]: E1014 07:01:56.564316 4870 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 14 07:01:56 crc kubenswrapper[4870]: E1014 07:01:56.564371 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7d7a4f41-388a-4320-b896-43c17ff13da6-metrics-certs podName:7d7a4f41-388a-4320-b896-43c17ff13da6 nodeName:}" failed. No retries permitted until 2025-10-14 07:02:12.564357731 +0000 UTC m=+68.261718102 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/7d7a4f41-388a-4320-b896-43c17ff13da6-metrics-certs") pod "network-metrics-daemon-zc5j6" (UID: "7d7a4f41-388a-4320-b896-43c17ff13da6") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 14 07:01:56 crc kubenswrapper[4870]: I1014 07:01:56.573932 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:56Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:56 crc kubenswrapper[4870]: I1014 07:01:56.583582 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:56 crc kubenswrapper[4870]: I1014 07:01:56.583642 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:56 crc kubenswrapper[4870]: I1014 07:01:56.583655 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:56 crc kubenswrapper[4870]: I1014 07:01:56.583678 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:56 crc kubenswrapper[4870]: I1014 07:01:56.583696 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:56Z","lastTransitionTime":"2025-10-14T07:01:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:56 crc kubenswrapper[4870]: I1014 07:01:56.587743 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3eb0b397c05ffd6650607f1be4a75683e0342d3333fc096bf1cd63c7688834f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:56Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:56 crc kubenswrapper[4870]: I1014 07:01:56.604133 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e26fe49ec2045d20a0cb2dd17e96e79b6eb43c641eb184c688cc472ca5db39d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aea115f7ee71c700b218fe15c928af7b55e6babe70e65d67f125d87be1fd2e2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:56Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:56 crc kubenswrapper[4870]: I1014 07:01:56.624481 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-v7br2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dcaef0e-42e3-424b-bbe5-4a023efe912d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71dec55e859b11a081dbb90d7d6cf6b4846cf591382303d8329003ff50c71f2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e72cff0723e2ce520340244b4e597d8016c58e1b071c4b087f2f6cd718741594\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc9d4db944f837289d1923176d60cfc4a566a282c2d2e3a703e220df884703c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d16ed1c65925549dcb69ab02c3013e25e3f123a83179f1af3a8e08bf91a59896\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed59243b59c39ce82bf9bc8f33c0a8b84854846afff6a89f5d37ffc97a096b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3b6d20cf459b1f6d3aff5b9290f9391909baf27ee40c49e2b0cd2d5cfe889ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fc9bfcb356b506d6c439adc22e4caaed4fad1c5a085744d44ebd50cf8898777\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a05f64fdac0425bbf7ea39d9d9975279c4a68dc9a93e27f58f2d7017a328f428\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-14T07:01:41Z\\\",\\\"message\\\":\\\"rvices do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-console/downloads_TCP_cluster\\\\\\\", UUID:\\\\\\\"2ead45b3-c313-4fbc-a7bc-2b3c4ffd610c\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-console/downloads\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-console/downloads_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-console/downloads\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.213\\\\\\\", Port:80, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nF1014 07:01:41.689119 6388 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:40Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fe3468cdd612fc4378e1f6f55a2a9d00274f7ddb65cb57b2a368dfc065cfb4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ba24c2e067b9f3a14d9d018b66300b54d866530d30174fbdb3c899299575d28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ba24c2e067b9f3a14d9d018b66300b54d866530d30174fbdb3c899299575d28\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-v7br2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:56Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:56 crc kubenswrapper[4870]: I1014 07:01:56.647370 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77a0d93a-5391-4112-b502-c474f45de209\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7815af9a584f6db6660073da289f2e8c987a04ef9ddfac155bdd9a15b014fcfa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b329b570622155a55f70501b2260e278106a68ebef497914231b4ec860e0125\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba942a0e6fe5ab93a187e0b7b4bd7c440c2e7827d9b5323fc50f90b562075392\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b9018b078c5c8276243c1696860f0aa73e2d8dd917913ae9d03b30dc9b5113a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://978e8db184ba9e430232c6bb2ee530beb92567fdc71ba3c35b0bedca3492d845\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"file observer\\\\nW1014 07:01:25.870672 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1014 07:01:25.870795 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1014 07:01:25.871521 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-756714914/tls.crt::/tmp/serving-cert-756714914/tls.key\\\\\\\"\\\\nI1014 07:01:26.460037 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1014 07:01:26.464876 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1014 07:01:26.464902 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1014 07:01:26.464925 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1014 07:01:26.464931 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1014 07:01:26.481374 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1014 07:01:26.481400 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:01:26.481405 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:01:26.481410 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1014 07:01:26.481414 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1014 07:01:26.481417 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1014 07:01:26.481420 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1014 07:01:26.481610 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1014 07:01:26.482872 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bd0044a247bdc5ca4cab7294103af3270b76cde61f3109b85e83ed050ca53b6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82d3c2214e0be679d2827594b159ace362151318689078bc49d3158f42899636\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82d3c2214e0be679d2827594b159ace362151318689078bc49d3158f42899636\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:05Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:56Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:56 crc kubenswrapper[4870]: I1014 07:01:56.667193 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:56Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:56 crc kubenswrapper[4870]: I1014 07:01:56.679630 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-4v2v2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4796ba6-c23b-4764-be96-2e3a3bf5e962\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf5afb0b99a7101f6d5573fb7eb925c5a1d8becc11b2e3c6f3351d80b6f0e823\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnp8b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-4v2v2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:56Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:56 crc kubenswrapper[4870]: I1014 07:01:56.686742 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:56 crc kubenswrapper[4870]: I1014 07:01:56.686784 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:56 crc kubenswrapper[4870]: I1014 07:01:56.686794 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:56 crc kubenswrapper[4870]: I1014 07:01:56.686809 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:56 crc kubenswrapper[4870]: I1014 07:01:56.686820 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:56Z","lastTransitionTime":"2025-10-14T07:01:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:56 crc kubenswrapper[4870]: I1014 07:01:56.695012 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-cdjmj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"612e7ab9-bda4-45d8-bc50-ff1afb68aace\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7776352d62b814fc63d01e4c16ad13f878b0597eb37a370bf9da0a88abab6cfa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://834444b5350ee12df4e4652c3009880d8d69e1edf8ba0c0156ab96fb2bbe28e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://834444b5350ee12df4e4652c3009880d8d69e1edf8ba0c0156ab96fb2bbe28e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c1972e7b2f75790db22a07a078eee97072bffbdb7f8e594bf2be5f800632a7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c1972e7b2f75790db22a07a078eee97072bffbdb7f8e594bf2be5f800632a7b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b124b42a44c6a23da1d92b9a5326987eb74430f80489a71e8913829475654ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b124b42a44c6a23da1d92b9a5326987eb74430f80489a71e8913829475654ab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62d0deb6b20ea65fbcbaf94c2fc14d69adb22e7a46cc043ccd4def2668ddd4b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62d0deb6b20ea65fbcbaf94c2fc14d69adb22e7a46cc043ccd4def2668ddd4b7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7f0d7cf29ca737d0effe58731a8b890379527c97f78c144d4eebb910260fadb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7f0d7cf29ca737d0effe58731a8b890379527c97f78c144d4eebb910260fadb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5959abb7c8e7ab32bed3aca709ae0334e8400f991ba724246c9fda56cbed822\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5959abb7c8e7ab32bed3aca709ae0334e8400f991ba724246c9fda56cbed822\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-cdjmj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:56Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:56 crc kubenswrapper[4870]: I1014 07:01:56.712862 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fdxjh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a3a430a-b540-4a8c-adad-b2893dbf9898\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93e41d5e927eed7ec62eb9c3607cc804cd8999556663363fd123b2a784b00ff1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k787j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fdxjh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:56Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:56 crc kubenswrapper[4870]: I1014 07:01:56.725534 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-khtq8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28deeea6-a1ef-4a4a-a377-d954ccd87de9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://296bb048a1d99649beb57d609573833bf547cd4d628ebc823c6af83fe5736c9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rb79p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2b7bc4c93307e80b778f2282ced993898f1c4cfc350a7d74828b8f9dea9906d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rb79p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-khtq8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:56Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:56 crc kubenswrapper[4870]: I1014 07:01:56.748087 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33a9493e-9660-48db-b8d1-310e29ebe313\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3570643ef1b7d5cc3a542f320413b819338e0c8b7ab7f0f09cc909997a39bab8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb398668197dedc1c2a14b4ae7d6a29714f3527309ccb9c1622af841100e1d60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1d2b6d068d63bd38cdfb83824525f6d12054f6cabab5d684e03ea5f2152076c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://783ff243e395333f3fc80cfae21358c532f0c9c5cb5515586be56b90ed10d912\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a8135a3d01076e4437ae8a784b459bf8d16de57fa16ec7ce4d40c5129c9d4a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48ca3a5017ba650df99b2fc7201c3fe5abfa89d81089c5647cde1b1549f534fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48ca3a5017ba650df99b2fc7201c3fe5abfa89d81089c5647cde1b1549f534fc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://19682eca67f76bf2e06e1b80f12e21141fcbce0acb8195f0defacd1b9a10b5e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://19682eca67f76bf2e06e1b80f12e21141fcbce0acb8195f0defacd1b9a10b5e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://042aad6d5f6bee8eef77f9e6a2c265de775a69f3da72aedc68bdf312198f00fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://042aad6d5f6bee8eef77f9e6a2c265de775a69f3da72aedc68bdf312198f00fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:05Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:56Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:56 crc kubenswrapper[4870]: I1014 07:01:56.761329 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:56Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:56 crc kubenswrapper[4870]: I1014 07:01:56.772793 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-kz77f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d3864a91-9bf6-4540-a908-cc41aa39656c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d4ded3d86030f5ea548c8b1cc3caddbe14a45c9a3ae4af836e892880584f57d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4h8v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:28Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-kz77f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:56Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:56 crc kubenswrapper[4870]: I1014 07:01:56.783980 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bcbc6dd6-aab0-4dbc-a62e-c58f06ec1337\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25ec4808b09adc44488530b3c4e29fd7fec46b3fcfa31d4449ec4a6dbc880ec8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f3f5de86fb11231a19034a18e9d7f31476c162e6567d0c136ead8bbbd900d6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d57d938d9d9c92198ae5b3857536ece36012fbfbe8cc8cfd245d1386d39b0be5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77a3247a7f65709543d7fd374b2ab69d7d793537d4132d7e22261188d6e58544\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:05Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:56Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:56 crc kubenswrapper[4870]: I1014 07:01:56.789643 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:56 crc kubenswrapper[4870]: I1014 07:01:56.789703 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:56 crc kubenswrapper[4870]: I1014 07:01:56.789714 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:56 crc kubenswrapper[4870]: I1014 07:01:56.789732 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:56 crc kubenswrapper[4870]: I1014 07:01:56.789743 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:56Z","lastTransitionTime":"2025-10-14T07:01:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:56 crc kubenswrapper[4870]: I1014 07:01:56.795549 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e539fac3d0a2ade314a4b4c1050bb27c37bcff5ba5272fa2d71b97ab5c0d7191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:56Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:56 crc kubenswrapper[4870]: I1014 07:01:56.892290 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:56 crc kubenswrapper[4870]: I1014 07:01:56.892634 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:56 crc kubenswrapper[4870]: I1014 07:01:56.892648 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:56 crc kubenswrapper[4870]: I1014 07:01:56.892663 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:56 crc kubenswrapper[4870]: I1014 07:01:56.892673 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:56Z","lastTransitionTime":"2025-10-14T07:01:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:56 crc kubenswrapper[4870]: I1014 07:01:56.994984 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:56 crc kubenswrapper[4870]: I1014 07:01:56.995023 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:56 crc kubenswrapper[4870]: I1014 07:01:56.995038 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:56 crc kubenswrapper[4870]: I1014 07:01:56.995062 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:56 crc kubenswrapper[4870]: I1014 07:01:56.995073 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:56Z","lastTransitionTime":"2025-10-14T07:01:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:57 crc kubenswrapper[4870]: I1014 07:01:57.032826 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zc5j6" Oct 14 07:01:57 crc kubenswrapper[4870]: E1014 07:01:57.032959 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zc5j6" podUID="7d7a4f41-388a-4320-b896-43c17ff13da6" Oct 14 07:01:57 crc kubenswrapper[4870]: I1014 07:01:57.098744 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:57 crc kubenswrapper[4870]: I1014 07:01:57.098784 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:57 crc kubenswrapper[4870]: I1014 07:01:57.098797 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:57 crc kubenswrapper[4870]: I1014 07:01:57.098815 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:57 crc kubenswrapper[4870]: I1014 07:01:57.098828 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:57Z","lastTransitionTime":"2025-10-14T07:01:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:57 crc kubenswrapper[4870]: I1014 07:01:57.201367 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:57 crc kubenswrapper[4870]: I1014 07:01:57.201429 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:57 crc kubenswrapper[4870]: I1014 07:01:57.201515 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:57 crc kubenswrapper[4870]: I1014 07:01:57.201553 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:57 crc kubenswrapper[4870]: I1014 07:01:57.201577 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:57Z","lastTransitionTime":"2025-10-14T07:01:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:57 crc kubenswrapper[4870]: I1014 07:01:57.304552 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:57 crc kubenswrapper[4870]: I1014 07:01:57.304625 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:57 crc kubenswrapper[4870]: I1014 07:01:57.304646 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:57 crc kubenswrapper[4870]: I1014 07:01:57.304684 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:57 crc kubenswrapper[4870]: I1014 07:01:57.304704 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:57Z","lastTransitionTime":"2025-10-14T07:01:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:57 crc kubenswrapper[4870]: I1014 07:01:57.408508 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:57 crc kubenswrapper[4870]: I1014 07:01:57.408572 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:57 crc kubenswrapper[4870]: I1014 07:01:57.408582 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:57 crc kubenswrapper[4870]: I1014 07:01:57.408617 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:57 crc kubenswrapper[4870]: I1014 07:01:57.408631 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:57Z","lastTransitionTime":"2025-10-14T07:01:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:57 crc kubenswrapper[4870]: I1014 07:01:57.512110 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:57 crc kubenswrapper[4870]: I1014 07:01:57.512181 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:57 crc kubenswrapper[4870]: I1014 07:01:57.512195 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:57 crc kubenswrapper[4870]: I1014 07:01:57.512224 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:57 crc kubenswrapper[4870]: I1014 07:01:57.512241 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:57Z","lastTransitionTime":"2025-10-14T07:01:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:57 crc kubenswrapper[4870]: I1014 07:01:57.541488 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-v7br2_8dcaef0e-42e3-424b-bbe5-4a023efe912d/ovnkube-controller/2.log" Oct 14 07:01:57 crc kubenswrapper[4870]: I1014 07:01:57.542362 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-v7br2_8dcaef0e-42e3-424b-bbe5-4a023efe912d/ovnkube-controller/1.log" Oct 14 07:01:57 crc kubenswrapper[4870]: I1014 07:01:57.546285 4870 generic.go:334] "Generic (PLEG): container finished" podID="8dcaef0e-42e3-424b-bbe5-4a023efe912d" containerID="3fc9bfcb356b506d6c439adc22e4caaed4fad1c5a085744d44ebd50cf8898777" exitCode=1 Oct 14 07:01:57 crc kubenswrapper[4870]: I1014 07:01:57.546366 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-v7br2" event={"ID":"8dcaef0e-42e3-424b-bbe5-4a023efe912d","Type":"ContainerDied","Data":"3fc9bfcb356b506d6c439adc22e4caaed4fad1c5a085744d44ebd50cf8898777"} Oct 14 07:01:57 crc kubenswrapper[4870]: I1014 07:01:57.546431 4870 scope.go:117] "RemoveContainer" containerID="a05f64fdac0425bbf7ea39d9d9975279c4a68dc9a93e27f58f2d7017a328f428" Oct 14 07:01:57 crc kubenswrapper[4870]: I1014 07:01:57.547467 4870 scope.go:117] "RemoveContainer" containerID="3fc9bfcb356b506d6c439adc22e4caaed4fad1c5a085744d44ebd50cf8898777" Oct 14 07:01:57 crc kubenswrapper[4870]: E1014 07:01:57.547645 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-v7br2_openshift-ovn-kubernetes(8dcaef0e-42e3-424b-bbe5-4a023efe912d)\"" pod="openshift-ovn-kubernetes/ovnkube-node-v7br2" podUID="8dcaef0e-42e3-424b-bbe5-4a023efe912d" Oct 14 07:01:57 crc kubenswrapper[4870]: I1014 07:01:57.568544 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:57Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:57 crc kubenswrapper[4870]: I1014 07:01:57.582831 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3eb0b397c05ffd6650607f1be4a75683e0342d3333fc096bf1cd63c7688834f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:57Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:57 crc kubenswrapper[4870]: I1014 07:01:57.596148 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"874a5e45-dffd-4d17-b609-b7d3ed2eab07\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d98d044836ab44a405196915ef7bf572b75113f2c81ec7e64effa63b47ad3543\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wj4ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8004752575df4a1c44efbcafe64799f09126d0f97285d92388c76a0b90e022f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wj4ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7tvc7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:57Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:57 crc kubenswrapper[4870]: I1014 07:01:57.611386 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zc5j6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d7a4f41-388a-4320-b896-43c17ff13da6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6p2wh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6p2wh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:40Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zc5j6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:57Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:57 crc kubenswrapper[4870]: I1014 07:01:57.615561 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:57 crc kubenswrapper[4870]: I1014 07:01:57.615586 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:57 crc kubenswrapper[4870]: I1014 07:01:57.615596 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:57 crc kubenswrapper[4870]: I1014 07:01:57.615614 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:57 crc kubenswrapper[4870]: I1014 07:01:57.615625 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:57Z","lastTransitionTime":"2025-10-14T07:01:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:57 crc kubenswrapper[4870]: I1014 07:01:57.628687 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77a0d93a-5391-4112-b502-c474f45de209\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7815af9a584f6db6660073da289f2e8c987a04ef9ddfac155bdd9a15b014fcfa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b329b570622155a55f70501b2260e278106a68ebef497914231b4ec860e0125\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba942a0e6fe5ab93a187e0b7b4bd7c440c2e7827d9b5323fc50f90b562075392\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b9018b078c5c8276243c1696860f0aa73e2d8dd917913ae9d03b30dc9b5113a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://978e8db184ba9e430232c6bb2ee530beb92567fdc71ba3c35b0bedca3492d845\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"file observer\\\\nW1014 07:01:25.870672 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1014 07:01:25.870795 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1014 07:01:25.871521 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-756714914/tls.crt::/tmp/serving-cert-756714914/tls.key\\\\\\\"\\\\nI1014 07:01:26.460037 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1014 07:01:26.464876 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1014 07:01:26.464902 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1014 07:01:26.464925 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1014 07:01:26.464931 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1014 07:01:26.481374 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1014 07:01:26.481400 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:01:26.481405 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:01:26.481410 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1014 07:01:26.481414 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1014 07:01:26.481417 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1014 07:01:26.481420 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1014 07:01:26.481610 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1014 07:01:26.482872 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bd0044a247bdc5ca4cab7294103af3270b76cde61f3109b85e83ed050ca53b6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82d3c2214e0be679d2827594b159ace362151318689078bc49d3158f42899636\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82d3c2214e0be679d2827594b159ace362151318689078bc49d3158f42899636\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:05Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:57Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:57 crc kubenswrapper[4870]: I1014 07:01:57.645532 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:57Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:57 crc kubenswrapper[4870]: I1014 07:01:57.663639 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e26fe49ec2045d20a0cb2dd17e96e79b6eb43c641eb184c688cc472ca5db39d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aea115f7ee71c700b218fe15c928af7b55e6babe70e65d67f125d87be1fd2e2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:57Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:57 crc kubenswrapper[4870]: I1014 07:01:57.690549 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-v7br2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dcaef0e-42e3-424b-bbe5-4a023efe912d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71dec55e859b11a081dbb90d7d6cf6b4846cf591382303d8329003ff50c71f2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e72cff0723e2ce520340244b4e597d8016c58e1b071c4b087f2f6cd718741594\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc9d4db944f837289d1923176d60cfc4a566a282c2d2e3a703e220df884703c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d16ed1c65925549dcb69ab02c3013e25e3f123a83179f1af3a8e08bf91a59896\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed59243b59c39ce82bf9bc8f33c0a8b84854846afff6a89f5d37ffc97a096b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3b6d20cf459b1f6d3aff5b9290f9391909baf27ee40c49e2b0cd2d5cfe889ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fc9bfcb356b506d6c439adc22e4caaed4fad1c5a085744d44ebd50cf8898777\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a05f64fdac0425bbf7ea39d9d9975279c4a68dc9a93e27f58f2d7017a328f428\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-14T07:01:41Z\\\",\\\"message\\\":\\\"rvices do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-console/downloads_TCP_cluster\\\\\\\", UUID:\\\\\\\"2ead45b3-c313-4fbc-a7bc-2b3c4ffd610c\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-console/downloads\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-console/downloads_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-console/downloads\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.213\\\\\\\", Port:80, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nF1014 07:01:41.689119 6388 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:40Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3fc9bfcb356b506d6c439adc22e4caaed4fad1c5a085744d44ebd50cf8898777\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-14T07:01:56Z\\\",\\\"message\\\":\\\" 6555 ovn.go:134] Ensuring zone local for Pod openshift-image-registry/node-ca-kz77f in node crc\\\\nI1014 07:01:56.821005 6555 obj_retry.go:386] Retry successful for *v1.Pod openshift-image-registry/node-ca-kz77f after 0 failed attempt(s)\\\\nI1014 07:01:56.820992 6555 model_client.go:382] Update operations generated as: [{Op:update Table:Logical_Switch_Port Row:map[addresses:{GoSet:[0a:58:0a:d9:00:5c 10.217.0.92]} options:{GoMap:map[iface-id-ver:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 requested-chassis:crc]} port_security:{GoSet:[0a:58:0a:d9:00:5c 10.217.0.92]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c94130be-172c-477c-88c4-40cc7eba30fe}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1014 07:01:56.821025 6555 default_network_controller.go:776] Recording success event on pod openshift-image-registry/node-ca-kz77f\\\\nI1014 07:01:56.820931 6555 default_network_controller.go:776] Recording success event on pod openshift-dns/node-resolver-4v2v2\\\\nI1014 07:01:56.820614 6555 services_controller.go:454] Service openshift-marketplace/redhat-marketplace for network=default has 1 cluster-wide, 0 per-node configs, 0 template configs, making 1 (cluster) 0 (per node) and 0 (template) load balancers\\\\nI1014 07:01:56.821037 6555 model_client.go:382] Update operations generated as: [{Op:update Table:N\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fe3468cdd612fc4378e1f6f55a2a9d00274f7ddb65cb57b2a368dfc065cfb4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ba24c2e067b9f3a14d9d018b66300b54d866530d30174fbdb3c899299575d28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ba24c2e067b9f3a14d9d018b66300b54d866530d30174fbdb3c899299575d28\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-v7br2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:57Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:57 crc kubenswrapper[4870]: I1014 07:01:57.710067 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33a9493e-9660-48db-b8d1-310e29ebe313\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3570643ef1b7d5cc3a542f320413b819338e0c8b7ab7f0f09cc909997a39bab8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb398668197dedc1c2a14b4ae7d6a29714f3527309ccb9c1622af841100e1d60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1d2b6d068d63bd38cdfb83824525f6d12054f6cabab5d684e03ea5f2152076c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://783ff243e395333f3fc80cfae21358c532f0c9c5cb5515586be56b90ed10d912\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a8135a3d01076e4437ae8a784b459bf8d16de57fa16ec7ce4d40c5129c9d4a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48ca3a5017ba650df99b2fc7201c3fe5abfa89d81089c5647cde1b1549f534fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48ca3a5017ba650df99b2fc7201c3fe5abfa89d81089c5647cde1b1549f534fc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://19682eca67f76bf2e06e1b80f12e21141fcbce0acb8195f0defacd1b9a10b5e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://19682eca67f76bf2e06e1b80f12e21141fcbce0acb8195f0defacd1b9a10b5e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://042aad6d5f6bee8eef77f9e6a2c265de775a69f3da72aedc68bdf312198f00fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://042aad6d5f6bee8eef77f9e6a2c265de775a69f3da72aedc68bdf312198f00fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:05Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:57Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:57 crc kubenswrapper[4870]: I1014 07:01:57.718686 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:57 crc kubenswrapper[4870]: I1014 07:01:57.718744 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:57 crc kubenswrapper[4870]: I1014 07:01:57.718755 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:57 crc kubenswrapper[4870]: I1014 07:01:57.718779 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:57 crc kubenswrapper[4870]: I1014 07:01:57.718790 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:57Z","lastTransitionTime":"2025-10-14T07:01:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:57 crc kubenswrapper[4870]: I1014 07:01:57.721978 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-4v2v2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4796ba6-c23b-4764-be96-2e3a3bf5e962\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf5afb0b99a7101f6d5573fb7eb925c5a1d8becc11b2e3c6f3351d80b6f0e823\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnp8b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-4v2v2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:57Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:57 crc kubenswrapper[4870]: I1014 07:01:57.736613 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-cdjmj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"612e7ab9-bda4-45d8-bc50-ff1afb68aace\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7776352d62b814fc63d01e4c16ad13f878b0597eb37a370bf9da0a88abab6cfa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://834444b5350ee12df4e4652c3009880d8d69e1edf8ba0c0156ab96fb2bbe28e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://834444b5350ee12df4e4652c3009880d8d69e1edf8ba0c0156ab96fb2bbe28e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c1972e7b2f75790db22a07a078eee97072bffbdb7f8e594bf2be5f800632a7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c1972e7b2f75790db22a07a078eee97072bffbdb7f8e594bf2be5f800632a7b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b124b42a44c6a23da1d92b9a5326987eb74430f80489a71e8913829475654ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b124b42a44c6a23da1d92b9a5326987eb74430f80489a71e8913829475654ab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62d0deb6b20ea65fbcbaf94c2fc14d69adb22e7a46cc043ccd4def2668ddd4b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62d0deb6b20ea65fbcbaf94c2fc14d69adb22e7a46cc043ccd4def2668ddd4b7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7f0d7cf29ca737d0effe58731a8b890379527c97f78c144d4eebb910260fadb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7f0d7cf29ca737d0effe58731a8b890379527c97f78c144d4eebb910260fadb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5959abb7c8e7ab32bed3aca709ae0334e8400f991ba724246c9fda56cbed822\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5959abb7c8e7ab32bed3aca709ae0334e8400f991ba724246c9fda56cbed822\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-cdjmj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:57Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:57 crc kubenswrapper[4870]: I1014 07:01:57.750890 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fdxjh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a3a430a-b540-4a8c-adad-b2893dbf9898\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93e41d5e927eed7ec62eb9c3607cc804cd8999556663363fd123b2a784b00ff1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k787j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fdxjh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:57Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:57 crc kubenswrapper[4870]: I1014 07:01:57.761734 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-khtq8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28deeea6-a1ef-4a4a-a377-d954ccd87de9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://296bb048a1d99649beb57d609573833bf547cd4d628ebc823c6af83fe5736c9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rb79p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2b7bc4c93307e80b778f2282ced993898f1c4cfc350a7d74828b8f9dea9906d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rb79p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-khtq8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:57Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:57 crc kubenswrapper[4870]: I1014 07:01:57.774332 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bcbc6dd6-aab0-4dbc-a62e-c58f06ec1337\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25ec4808b09adc44488530b3c4e29fd7fec46b3fcfa31d4449ec4a6dbc880ec8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f3f5de86fb11231a19034a18e9d7f31476c162e6567d0c136ead8bbbd900d6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d57d938d9d9c92198ae5b3857536ece36012fbfbe8cc8cfd245d1386d39b0be5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77a3247a7f65709543d7fd374b2ab69d7d793537d4132d7e22261188d6e58544\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:05Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:57Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:57 crc kubenswrapper[4870]: I1014 07:01:57.787576 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e539fac3d0a2ade314a4b4c1050bb27c37bcff5ba5272fa2d71b97ab5c0d7191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:57Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:57 crc kubenswrapper[4870]: I1014 07:01:57.799252 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:57Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:57 crc kubenswrapper[4870]: I1014 07:01:57.808064 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-kz77f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d3864a91-9bf6-4540-a908-cc41aa39656c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d4ded3d86030f5ea548c8b1cc3caddbe14a45c9a3ae4af836e892880584f57d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4h8v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:28Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-kz77f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:57Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:57 crc kubenswrapper[4870]: I1014 07:01:57.820659 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:57 crc kubenswrapper[4870]: I1014 07:01:57.820774 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:57 crc kubenswrapper[4870]: I1014 07:01:57.820849 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:57 crc kubenswrapper[4870]: I1014 07:01:57.820955 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:57 crc kubenswrapper[4870]: I1014 07:01:57.821058 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:57Z","lastTransitionTime":"2025-10-14T07:01:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:57 crc kubenswrapper[4870]: I1014 07:01:57.878787 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 07:01:57 crc kubenswrapper[4870]: E1014 07:01:57.879160 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 07:02:29.879135908 +0000 UTC m=+85.576496299 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:01:57 crc kubenswrapper[4870]: I1014 07:01:57.923548 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:57 crc kubenswrapper[4870]: I1014 07:01:57.923585 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:57 crc kubenswrapper[4870]: I1014 07:01:57.923594 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:57 crc kubenswrapper[4870]: I1014 07:01:57.923609 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:57 crc kubenswrapper[4870]: I1014 07:01:57.923617 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:57Z","lastTransitionTime":"2025-10-14T07:01:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:57 crc kubenswrapper[4870]: E1014 07:01:57.980036 4870 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 14 07:01:57 crc kubenswrapper[4870]: E1014 07:01:57.980347 4870 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 14 07:01:57 crc kubenswrapper[4870]: I1014 07:01:57.979838 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 07:01:57 crc kubenswrapper[4870]: I1014 07:01:57.980653 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 07:01:57 crc kubenswrapper[4870]: I1014 07:01:57.980722 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 07:01:57 crc kubenswrapper[4870]: I1014 07:01:57.980780 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 07:01:57 crc kubenswrapper[4870]: E1014 07:01:57.980532 4870 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 14 07:01:57 crc kubenswrapper[4870]: E1014 07:01:57.981064 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-14 07:02:29.981037329 +0000 UTC m=+85.678397740 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 14 07:01:57 crc kubenswrapper[4870]: E1014 07:01:57.980971 4870 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 14 07:01:57 crc kubenswrapper[4870]: E1014 07:01:57.981493 4870 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 14 07:01:57 crc kubenswrapper[4870]: E1014 07:01:57.981514 4870 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 14 07:01:57 crc kubenswrapper[4870]: E1014 07:01:57.981524 4870 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 14 07:01:57 crc kubenswrapper[4870]: E1014 07:01:57.981511 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-14 07:02:29.981477139 +0000 UTC m=+85.678837510 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 14 07:01:57 crc kubenswrapper[4870]: E1014 07:01:57.981600 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-14 07:02:29.981566582 +0000 UTC m=+85.678926983 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 14 07:01:57 crc kubenswrapper[4870]: E1014 07:01:57.982012 4870 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 14 07:01:57 crc kubenswrapper[4870]: E1014 07:01:57.982211 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-14 07:02:29.982185147 +0000 UTC m=+85.679545578 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 14 07:01:58 crc kubenswrapper[4870]: I1014 07:01:58.026166 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:58 crc kubenswrapper[4870]: I1014 07:01:58.026210 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:58 crc kubenswrapper[4870]: I1014 07:01:58.026219 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:58 crc kubenswrapper[4870]: I1014 07:01:58.026233 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:58 crc kubenswrapper[4870]: I1014 07:01:58.026246 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:58Z","lastTransitionTime":"2025-10-14T07:01:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:58 crc kubenswrapper[4870]: I1014 07:01:58.033558 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 07:01:58 crc kubenswrapper[4870]: I1014 07:01:58.033601 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 07:01:58 crc kubenswrapper[4870]: I1014 07:01:58.033555 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 07:01:58 crc kubenswrapper[4870]: E1014 07:01:58.033658 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 07:01:58 crc kubenswrapper[4870]: E1014 07:01:58.033756 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 07:01:58 crc kubenswrapper[4870]: E1014 07:01:58.033822 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 07:01:58 crc kubenswrapper[4870]: I1014 07:01:58.128672 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:58 crc kubenswrapper[4870]: I1014 07:01:58.128903 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:58 crc kubenswrapper[4870]: I1014 07:01:58.128912 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:58 crc kubenswrapper[4870]: I1014 07:01:58.128926 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:58 crc kubenswrapper[4870]: I1014 07:01:58.128936 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:58Z","lastTransitionTime":"2025-10-14T07:01:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:58 crc kubenswrapper[4870]: I1014 07:01:58.231410 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:58 crc kubenswrapper[4870]: I1014 07:01:58.231470 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:58 crc kubenswrapper[4870]: I1014 07:01:58.231483 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:58 crc kubenswrapper[4870]: I1014 07:01:58.231501 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:58 crc kubenswrapper[4870]: I1014 07:01:58.231515 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:58Z","lastTransitionTime":"2025-10-14T07:01:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:58 crc kubenswrapper[4870]: I1014 07:01:58.333925 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:58 crc kubenswrapper[4870]: I1014 07:01:58.333970 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:58 crc kubenswrapper[4870]: I1014 07:01:58.333982 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:58 crc kubenswrapper[4870]: I1014 07:01:58.334008 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:58 crc kubenswrapper[4870]: I1014 07:01:58.334021 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:58Z","lastTransitionTime":"2025-10-14T07:01:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:58 crc kubenswrapper[4870]: I1014 07:01:58.436206 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:58 crc kubenswrapper[4870]: I1014 07:01:58.436261 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:58 crc kubenswrapper[4870]: I1014 07:01:58.436274 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:58 crc kubenswrapper[4870]: I1014 07:01:58.436310 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:58 crc kubenswrapper[4870]: I1014 07:01:58.436321 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:58Z","lastTransitionTime":"2025-10-14T07:01:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:58 crc kubenswrapper[4870]: I1014 07:01:58.539027 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:58 crc kubenswrapper[4870]: I1014 07:01:58.539064 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:58 crc kubenswrapper[4870]: I1014 07:01:58.539073 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:58 crc kubenswrapper[4870]: I1014 07:01:58.539088 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:58 crc kubenswrapper[4870]: I1014 07:01:58.539098 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:58Z","lastTransitionTime":"2025-10-14T07:01:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:58 crc kubenswrapper[4870]: I1014 07:01:58.551008 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-v7br2_8dcaef0e-42e3-424b-bbe5-4a023efe912d/ovnkube-controller/2.log" Oct 14 07:01:58 crc kubenswrapper[4870]: I1014 07:01:58.554716 4870 scope.go:117] "RemoveContainer" containerID="3fc9bfcb356b506d6c439adc22e4caaed4fad1c5a085744d44ebd50cf8898777" Oct 14 07:01:58 crc kubenswrapper[4870]: E1014 07:01:58.555019 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-v7br2_openshift-ovn-kubernetes(8dcaef0e-42e3-424b-bbe5-4a023efe912d)\"" pod="openshift-ovn-kubernetes/ovnkube-node-v7br2" podUID="8dcaef0e-42e3-424b-bbe5-4a023efe912d" Oct 14 07:01:58 crc kubenswrapper[4870]: I1014 07:01:58.565389 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-kz77f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d3864a91-9bf6-4540-a908-cc41aa39656c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d4ded3d86030f5ea548c8b1cc3caddbe14a45c9a3ae4af836e892880584f57d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4h8v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:28Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-kz77f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:58Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:58 crc kubenswrapper[4870]: I1014 07:01:58.577411 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bcbc6dd6-aab0-4dbc-a62e-c58f06ec1337\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25ec4808b09adc44488530b3c4e29fd7fec46b3fcfa31d4449ec4a6dbc880ec8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f3f5de86fb11231a19034a18e9d7f31476c162e6567d0c136ead8bbbd900d6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d57d938d9d9c92198ae5b3857536ece36012fbfbe8cc8cfd245d1386d39b0be5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77a3247a7f65709543d7fd374b2ab69d7d793537d4132d7e22261188d6e58544\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:05Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:58Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:58 crc kubenswrapper[4870]: I1014 07:01:58.588381 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e539fac3d0a2ade314a4b4c1050bb27c37bcff5ba5272fa2d71b97ab5c0d7191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:58Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:58 crc kubenswrapper[4870]: I1014 07:01:58.609702 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:58Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:58 crc kubenswrapper[4870]: I1014 07:01:58.623425 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zc5j6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d7a4f41-388a-4320-b896-43c17ff13da6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6p2wh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6p2wh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:40Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zc5j6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:58Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:58 crc kubenswrapper[4870]: I1014 07:01:58.635258 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:58Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:58 crc kubenswrapper[4870]: I1014 07:01:58.641089 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:58 crc kubenswrapper[4870]: I1014 07:01:58.641142 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:58 crc kubenswrapper[4870]: I1014 07:01:58.641151 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:58 crc kubenswrapper[4870]: I1014 07:01:58.641167 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:58 crc kubenswrapper[4870]: I1014 07:01:58.641178 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:58Z","lastTransitionTime":"2025-10-14T07:01:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:58 crc kubenswrapper[4870]: I1014 07:01:58.647674 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3eb0b397c05ffd6650607f1be4a75683e0342d3333fc096bf1cd63c7688834f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:58Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:58 crc kubenswrapper[4870]: I1014 07:01:58.657732 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"874a5e45-dffd-4d17-b609-b7d3ed2eab07\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d98d044836ab44a405196915ef7bf572b75113f2c81ec7e64effa63b47ad3543\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wj4ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8004752575df4a1c44efbcafe64799f09126d0f97285d92388c76a0b90e022f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wj4ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7tvc7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:58Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:58 crc kubenswrapper[4870]: I1014 07:01:58.673025 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-v7br2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dcaef0e-42e3-424b-bbe5-4a023efe912d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71dec55e859b11a081dbb90d7d6cf6b4846cf591382303d8329003ff50c71f2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e72cff0723e2ce520340244b4e597d8016c58e1b071c4b087f2f6cd718741594\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc9d4db944f837289d1923176d60cfc4a566a282c2d2e3a703e220df884703c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d16ed1c65925549dcb69ab02c3013e25e3f123a83179f1af3a8e08bf91a59896\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed59243b59c39ce82bf9bc8f33c0a8b84854846afff6a89f5d37ffc97a096b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3b6d20cf459b1f6d3aff5b9290f9391909baf27ee40c49e2b0cd2d5cfe889ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fc9bfcb356b506d6c439adc22e4caaed4fad1c5a085744d44ebd50cf8898777\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3fc9bfcb356b506d6c439adc22e4caaed4fad1c5a085744d44ebd50cf8898777\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-14T07:01:56Z\\\",\\\"message\\\":\\\" 6555 ovn.go:134] Ensuring zone local for Pod openshift-image-registry/node-ca-kz77f in node crc\\\\nI1014 07:01:56.821005 6555 obj_retry.go:386] Retry successful for *v1.Pod openshift-image-registry/node-ca-kz77f after 0 failed attempt(s)\\\\nI1014 07:01:56.820992 6555 model_client.go:382] Update operations generated as: [{Op:update Table:Logical_Switch_Port Row:map[addresses:{GoSet:[0a:58:0a:d9:00:5c 10.217.0.92]} options:{GoMap:map[iface-id-ver:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 requested-chassis:crc]} port_security:{GoSet:[0a:58:0a:d9:00:5c 10.217.0.92]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c94130be-172c-477c-88c4-40cc7eba30fe}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1014 07:01:56.821025 6555 default_network_controller.go:776] Recording success event on pod openshift-image-registry/node-ca-kz77f\\\\nI1014 07:01:56.820931 6555 default_network_controller.go:776] Recording success event on pod openshift-dns/node-resolver-4v2v2\\\\nI1014 07:01:56.820614 6555 services_controller.go:454] Service openshift-marketplace/redhat-marketplace for network=default has 1 cluster-wide, 0 per-node configs, 0 template configs, making 1 (cluster) 0 (per node) and 0 (template) load balancers\\\\nI1014 07:01:56.821037 6555 model_client.go:382] Update operations generated as: [{Op:update Table:N\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:56Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-v7br2_openshift-ovn-kubernetes(8dcaef0e-42e3-424b-bbe5-4a023efe912d)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fe3468cdd612fc4378e1f6f55a2a9d00274f7ddb65cb57b2a368dfc065cfb4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ba24c2e067b9f3a14d9d018b66300b54d866530d30174fbdb3c899299575d28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ba24c2e067b9f3a14d9d018b66300b54d866530d30174fbdb3c899299575d28\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-v7br2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:58Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:58 crc kubenswrapper[4870]: I1014 07:01:58.683900 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77a0d93a-5391-4112-b502-c474f45de209\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7815af9a584f6db6660073da289f2e8c987a04ef9ddfac155bdd9a15b014fcfa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b329b570622155a55f70501b2260e278106a68ebef497914231b4ec860e0125\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba942a0e6fe5ab93a187e0b7b4bd7c440c2e7827d9b5323fc50f90b562075392\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b9018b078c5c8276243c1696860f0aa73e2d8dd917913ae9d03b30dc9b5113a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://978e8db184ba9e430232c6bb2ee530beb92567fdc71ba3c35b0bedca3492d845\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"file observer\\\\nW1014 07:01:25.870672 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1014 07:01:25.870795 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1014 07:01:25.871521 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-756714914/tls.crt::/tmp/serving-cert-756714914/tls.key\\\\\\\"\\\\nI1014 07:01:26.460037 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1014 07:01:26.464876 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1014 07:01:26.464902 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1014 07:01:26.464925 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1014 07:01:26.464931 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1014 07:01:26.481374 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1014 07:01:26.481400 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:01:26.481405 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:01:26.481410 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1014 07:01:26.481414 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1014 07:01:26.481417 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1014 07:01:26.481420 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1014 07:01:26.481610 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1014 07:01:26.482872 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bd0044a247bdc5ca4cab7294103af3270b76cde61f3109b85e83ed050ca53b6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82d3c2214e0be679d2827594b159ace362151318689078bc49d3158f42899636\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82d3c2214e0be679d2827594b159ace362151318689078bc49d3158f42899636\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:05Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:58Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:58 crc kubenswrapper[4870]: I1014 07:01:58.693633 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:58Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:58 crc kubenswrapper[4870]: I1014 07:01:58.707691 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e26fe49ec2045d20a0cb2dd17e96e79b6eb43c641eb184c688cc472ca5db39d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aea115f7ee71c700b218fe15c928af7b55e6babe70e65d67f125d87be1fd2e2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:58Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:58 crc kubenswrapper[4870]: I1014 07:01:58.718960 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-4v2v2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4796ba6-c23b-4764-be96-2e3a3bf5e962\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf5afb0b99a7101f6d5573fb7eb925c5a1d8becc11b2e3c6f3351d80b6f0e823\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnp8b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-4v2v2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:58Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:58 crc kubenswrapper[4870]: I1014 07:01:58.732286 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-cdjmj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"612e7ab9-bda4-45d8-bc50-ff1afb68aace\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7776352d62b814fc63d01e4c16ad13f878b0597eb37a370bf9da0a88abab6cfa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://834444b5350ee12df4e4652c3009880d8d69e1edf8ba0c0156ab96fb2bbe28e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://834444b5350ee12df4e4652c3009880d8d69e1edf8ba0c0156ab96fb2bbe28e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c1972e7b2f75790db22a07a078eee97072bffbdb7f8e594bf2be5f800632a7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c1972e7b2f75790db22a07a078eee97072bffbdb7f8e594bf2be5f800632a7b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b124b42a44c6a23da1d92b9a5326987eb74430f80489a71e8913829475654ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b124b42a44c6a23da1d92b9a5326987eb74430f80489a71e8913829475654ab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62d0deb6b20ea65fbcbaf94c2fc14d69adb22e7a46cc043ccd4def2668ddd4b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62d0deb6b20ea65fbcbaf94c2fc14d69adb22e7a46cc043ccd4def2668ddd4b7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7f0d7cf29ca737d0effe58731a8b890379527c97f78c144d4eebb910260fadb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7f0d7cf29ca737d0effe58731a8b890379527c97f78c144d4eebb910260fadb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5959abb7c8e7ab32bed3aca709ae0334e8400f991ba724246c9fda56cbed822\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5959abb7c8e7ab32bed3aca709ae0334e8400f991ba724246c9fda56cbed822\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-cdjmj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:58Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:58 crc kubenswrapper[4870]: I1014 07:01:58.744394 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:58 crc kubenswrapper[4870]: I1014 07:01:58.744431 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:58 crc kubenswrapper[4870]: I1014 07:01:58.744516 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:58 crc kubenswrapper[4870]: I1014 07:01:58.744532 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:58 crc kubenswrapper[4870]: I1014 07:01:58.744542 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:58Z","lastTransitionTime":"2025-10-14T07:01:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:58 crc kubenswrapper[4870]: I1014 07:01:58.745825 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fdxjh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a3a430a-b540-4a8c-adad-b2893dbf9898\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93e41d5e927eed7ec62eb9c3607cc804cd8999556663363fd123b2a784b00ff1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k787j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fdxjh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:58Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:58 crc kubenswrapper[4870]: I1014 07:01:58.756323 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-khtq8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28deeea6-a1ef-4a4a-a377-d954ccd87de9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://296bb048a1d99649beb57d609573833bf547cd4d628ebc823c6af83fe5736c9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rb79p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2b7bc4c93307e80b778f2282ced993898f1c4cfc350a7d74828b8f9dea9906d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rb79p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-khtq8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:58Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:58 crc kubenswrapper[4870]: I1014 07:01:58.771967 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33a9493e-9660-48db-b8d1-310e29ebe313\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3570643ef1b7d5cc3a542f320413b819338e0c8b7ab7f0f09cc909997a39bab8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb398668197dedc1c2a14b4ae7d6a29714f3527309ccb9c1622af841100e1d60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1d2b6d068d63bd38cdfb83824525f6d12054f6cabab5d684e03ea5f2152076c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://783ff243e395333f3fc80cfae21358c532f0c9c5cb5515586be56b90ed10d912\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a8135a3d01076e4437ae8a784b459bf8d16de57fa16ec7ce4d40c5129c9d4a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48ca3a5017ba650df99b2fc7201c3fe5abfa89d81089c5647cde1b1549f534fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48ca3a5017ba650df99b2fc7201c3fe5abfa89d81089c5647cde1b1549f534fc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://19682eca67f76bf2e06e1b80f12e21141fcbce0acb8195f0defacd1b9a10b5e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://19682eca67f76bf2e06e1b80f12e21141fcbce0acb8195f0defacd1b9a10b5e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://042aad6d5f6bee8eef77f9e6a2c265de775a69f3da72aedc68bdf312198f00fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://042aad6d5f6bee8eef77f9e6a2c265de775a69f3da72aedc68bdf312198f00fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:05Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:01:58Z is after 2025-08-24T17:21:41Z" Oct 14 07:01:58 crc kubenswrapper[4870]: I1014 07:01:58.846777 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:58 crc kubenswrapper[4870]: I1014 07:01:58.846820 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:58 crc kubenswrapper[4870]: I1014 07:01:58.846831 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:58 crc kubenswrapper[4870]: I1014 07:01:58.846850 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:58 crc kubenswrapper[4870]: I1014 07:01:58.846861 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:58Z","lastTransitionTime":"2025-10-14T07:01:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:58 crc kubenswrapper[4870]: I1014 07:01:58.950408 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:58 crc kubenswrapper[4870]: I1014 07:01:58.950508 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:58 crc kubenswrapper[4870]: I1014 07:01:58.950564 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:58 crc kubenswrapper[4870]: I1014 07:01:58.950597 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:58 crc kubenswrapper[4870]: I1014 07:01:58.950621 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:58Z","lastTransitionTime":"2025-10-14T07:01:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:59 crc kubenswrapper[4870]: I1014 07:01:59.033153 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zc5j6" Oct 14 07:01:59 crc kubenswrapper[4870]: E1014 07:01:59.033319 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zc5j6" podUID="7d7a4f41-388a-4320-b896-43c17ff13da6" Oct 14 07:01:59 crc kubenswrapper[4870]: I1014 07:01:59.053057 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:59 crc kubenswrapper[4870]: I1014 07:01:59.053127 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:59 crc kubenswrapper[4870]: I1014 07:01:59.053138 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:59 crc kubenswrapper[4870]: I1014 07:01:59.053156 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:59 crc kubenswrapper[4870]: I1014 07:01:59.053168 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:59Z","lastTransitionTime":"2025-10-14T07:01:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:59 crc kubenswrapper[4870]: I1014 07:01:59.156412 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:59 crc kubenswrapper[4870]: I1014 07:01:59.156532 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:59 crc kubenswrapper[4870]: I1014 07:01:59.156556 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:59 crc kubenswrapper[4870]: I1014 07:01:59.156586 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:59 crc kubenswrapper[4870]: I1014 07:01:59.156610 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:59Z","lastTransitionTime":"2025-10-14T07:01:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:59 crc kubenswrapper[4870]: I1014 07:01:59.260099 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:59 crc kubenswrapper[4870]: I1014 07:01:59.260149 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:59 crc kubenswrapper[4870]: I1014 07:01:59.260160 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:59 crc kubenswrapper[4870]: I1014 07:01:59.260177 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:59 crc kubenswrapper[4870]: I1014 07:01:59.260189 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:59Z","lastTransitionTime":"2025-10-14T07:01:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:59 crc kubenswrapper[4870]: I1014 07:01:59.362503 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:59 crc kubenswrapper[4870]: I1014 07:01:59.362571 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:59 crc kubenswrapper[4870]: I1014 07:01:59.362585 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:59 crc kubenswrapper[4870]: I1014 07:01:59.362604 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:59 crc kubenswrapper[4870]: I1014 07:01:59.362616 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:59Z","lastTransitionTime":"2025-10-14T07:01:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:59 crc kubenswrapper[4870]: I1014 07:01:59.465183 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:59 crc kubenswrapper[4870]: I1014 07:01:59.465236 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:59 crc kubenswrapper[4870]: I1014 07:01:59.465253 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:59 crc kubenswrapper[4870]: I1014 07:01:59.465279 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:59 crc kubenswrapper[4870]: I1014 07:01:59.465296 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:59Z","lastTransitionTime":"2025-10-14T07:01:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:59 crc kubenswrapper[4870]: I1014 07:01:59.567206 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:59 crc kubenswrapper[4870]: I1014 07:01:59.567250 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:59 crc kubenswrapper[4870]: I1014 07:01:59.567262 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:59 crc kubenswrapper[4870]: I1014 07:01:59.567279 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:59 crc kubenswrapper[4870]: I1014 07:01:59.567291 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:59Z","lastTransitionTime":"2025-10-14T07:01:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:59 crc kubenswrapper[4870]: I1014 07:01:59.670790 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:59 crc kubenswrapper[4870]: I1014 07:01:59.670868 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:59 crc kubenswrapper[4870]: I1014 07:01:59.670876 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:59 crc kubenswrapper[4870]: I1014 07:01:59.670891 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:59 crc kubenswrapper[4870]: I1014 07:01:59.670900 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:59Z","lastTransitionTime":"2025-10-14T07:01:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:59 crc kubenswrapper[4870]: I1014 07:01:59.773618 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:59 crc kubenswrapper[4870]: I1014 07:01:59.773654 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:59 crc kubenswrapper[4870]: I1014 07:01:59.773664 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:59 crc kubenswrapper[4870]: I1014 07:01:59.773680 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:59 crc kubenswrapper[4870]: I1014 07:01:59.773689 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:59Z","lastTransitionTime":"2025-10-14T07:01:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:59 crc kubenswrapper[4870]: I1014 07:01:59.876412 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:59 crc kubenswrapper[4870]: I1014 07:01:59.876787 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:59 crc kubenswrapper[4870]: I1014 07:01:59.876807 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:59 crc kubenswrapper[4870]: I1014 07:01:59.876964 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:59 crc kubenswrapper[4870]: I1014 07:01:59.877026 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:59Z","lastTransitionTime":"2025-10-14T07:01:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:01:59 crc kubenswrapper[4870]: I1014 07:01:59.980939 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:01:59 crc kubenswrapper[4870]: I1014 07:01:59.980978 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:01:59 crc kubenswrapper[4870]: I1014 07:01:59.980988 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:01:59 crc kubenswrapper[4870]: I1014 07:01:59.981006 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:01:59 crc kubenswrapper[4870]: I1014 07:01:59.981016 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:01:59Z","lastTransitionTime":"2025-10-14T07:01:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:00 crc kubenswrapper[4870]: I1014 07:02:00.033067 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 07:02:00 crc kubenswrapper[4870]: I1014 07:02:00.033118 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 07:02:00 crc kubenswrapper[4870]: E1014 07:02:00.033224 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 07:02:00 crc kubenswrapper[4870]: E1014 07:02:00.033312 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 07:02:00 crc kubenswrapper[4870]: I1014 07:02:00.033402 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 07:02:00 crc kubenswrapper[4870]: E1014 07:02:00.033659 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 07:02:00 crc kubenswrapper[4870]: I1014 07:02:00.084340 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:00 crc kubenswrapper[4870]: I1014 07:02:00.084383 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:00 crc kubenswrapper[4870]: I1014 07:02:00.084398 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:00 crc kubenswrapper[4870]: I1014 07:02:00.084416 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:00 crc kubenswrapper[4870]: I1014 07:02:00.084428 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:00Z","lastTransitionTime":"2025-10-14T07:02:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:00 crc kubenswrapper[4870]: I1014 07:02:00.186897 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:00 crc kubenswrapper[4870]: I1014 07:02:00.186988 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:00 crc kubenswrapper[4870]: I1014 07:02:00.187010 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:00 crc kubenswrapper[4870]: I1014 07:02:00.187035 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:00 crc kubenswrapper[4870]: I1014 07:02:00.187051 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:00Z","lastTransitionTime":"2025-10-14T07:02:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:00 crc kubenswrapper[4870]: I1014 07:02:00.262572 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 14 07:02:00 crc kubenswrapper[4870]: I1014 07:02:00.274667 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Oct 14 07:02:00 crc kubenswrapper[4870]: I1014 07:02:00.283977 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"874a5e45-dffd-4d17-b609-b7d3ed2eab07\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d98d044836ab44a405196915ef7bf572b75113f2c81ec7e64effa63b47ad3543\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wj4ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8004752575df4a1c44efbcafe64799f09126d0f97285d92388c76a0b90e022f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wj4ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7tvc7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:00Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:00 crc kubenswrapper[4870]: I1014 07:02:00.289241 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:00 crc kubenswrapper[4870]: I1014 07:02:00.289266 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:00 crc kubenswrapper[4870]: I1014 07:02:00.289278 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:00 crc kubenswrapper[4870]: I1014 07:02:00.289294 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:00 crc kubenswrapper[4870]: I1014 07:02:00.289305 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:00Z","lastTransitionTime":"2025-10-14T07:02:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:00 crc kubenswrapper[4870]: I1014 07:02:00.299416 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zc5j6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d7a4f41-388a-4320-b896-43c17ff13da6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6p2wh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6p2wh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:40Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zc5j6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:00Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:00 crc kubenswrapper[4870]: I1014 07:02:00.316713 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:00Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:00 crc kubenswrapper[4870]: I1014 07:02:00.331091 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3eb0b397c05ffd6650607f1be4a75683e0342d3333fc096bf1cd63c7688834f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:00Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:00 crc kubenswrapper[4870]: I1014 07:02:00.345653 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e26fe49ec2045d20a0cb2dd17e96e79b6eb43c641eb184c688cc472ca5db39d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aea115f7ee71c700b218fe15c928af7b55e6babe70e65d67f125d87be1fd2e2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:00Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:00 crc kubenswrapper[4870]: I1014 07:02:00.365844 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-v7br2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dcaef0e-42e3-424b-bbe5-4a023efe912d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71dec55e859b11a081dbb90d7d6cf6b4846cf591382303d8329003ff50c71f2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e72cff0723e2ce520340244b4e597d8016c58e1b071c4b087f2f6cd718741594\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc9d4db944f837289d1923176d60cfc4a566a282c2d2e3a703e220df884703c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d16ed1c65925549dcb69ab02c3013e25e3f123a83179f1af3a8e08bf91a59896\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed59243b59c39ce82bf9bc8f33c0a8b84854846afff6a89f5d37ffc97a096b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3b6d20cf459b1f6d3aff5b9290f9391909baf27ee40c49e2b0cd2d5cfe889ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fc9bfcb356b506d6c439adc22e4caaed4fad1c5a085744d44ebd50cf8898777\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3fc9bfcb356b506d6c439adc22e4caaed4fad1c5a085744d44ebd50cf8898777\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-14T07:01:56Z\\\",\\\"message\\\":\\\" 6555 ovn.go:134] Ensuring zone local for Pod openshift-image-registry/node-ca-kz77f in node crc\\\\nI1014 07:01:56.821005 6555 obj_retry.go:386] Retry successful for *v1.Pod openshift-image-registry/node-ca-kz77f after 0 failed attempt(s)\\\\nI1014 07:01:56.820992 6555 model_client.go:382] Update operations generated as: [{Op:update Table:Logical_Switch_Port Row:map[addresses:{GoSet:[0a:58:0a:d9:00:5c 10.217.0.92]} options:{GoMap:map[iface-id-ver:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 requested-chassis:crc]} port_security:{GoSet:[0a:58:0a:d9:00:5c 10.217.0.92]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c94130be-172c-477c-88c4-40cc7eba30fe}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1014 07:01:56.821025 6555 default_network_controller.go:776] Recording success event on pod openshift-image-registry/node-ca-kz77f\\\\nI1014 07:01:56.820931 6555 default_network_controller.go:776] Recording success event on pod openshift-dns/node-resolver-4v2v2\\\\nI1014 07:01:56.820614 6555 services_controller.go:454] Service openshift-marketplace/redhat-marketplace for network=default has 1 cluster-wide, 0 per-node configs, 0 template configs, making 1 (cluster) 0 (per node) and 0 (template) load balancers\\\\nI1014 07:01:56.821037 6555 model_client.go:382] Update operations generated as: [{Op:update Table:N\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:56Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-v7br2_openshift-ovn-kubernetes(8dcaef0e-42e3-424b-bbe5-4a023efe912d)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fe3468cdd612fc4378e1f6f55a2a9d00274f7ddb65cb57b2a368dfc065cfb4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ba24c2e067b9f3a14d9d018b66300b54d866530d30174fbdb3c899299575d28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ba24c2e067b9f3a14d9d018b66300b54d866530d30174fbdb3c899299575d28\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-v7br2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:00Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:00 crc kubenswrapper[4870]: I1014 07:02:00.381567 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77a0d93a-5391-4112-b502-c474f45de209\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7815af9a584f6db6660073da289f2e8c987a04ef9ddfac155bdd9a15b014fcfa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b329b570622155a55f70501b2260e278106a68ebef497914231b4ec860e0125\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba942a0e6fe5ab93a187e0b7b4bd7c440c2e7827d9b5323fc50f90b562075392\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b9018b078c5c8276243c1696860f0aa73e2d8dd917913ae9d03b30dc9b5113a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://978e8db184ba9e430232c6bb2ee530beb92567fdc71ba3c35b0bedca3492d845\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"file observer\\\\nW1014 07:01:25.870672 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1014 07:01:25.870795 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1014 07:01:25.871521 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-756714914/tls.crt::/tmp/serving-cert-756714914/tls.key\\\\\\\"\\\\nI1014 07:01:26.460037 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1014 07:01:26.464876 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1014 07:01:26.464902 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1014 07:01:26.464925 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1014 07:01:26.464931 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1014 07:01:26.481374 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1014 07:01:26.481400 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:01:26.481405 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:01:26.481410 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1014 07:01:26.481414 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1014 07:01:26.481417 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1014 07:01:26.481420 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1014 07:01:26.481610 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1014 07:01:26.482872 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bd0044a247bdc5ca4cab7294103af3270b76cde61f3109b85e83ed050ca53b6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82d3c2214e0be679d2827594b159ace362151318689078bc49d3158f42899636\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82d3c2214e0be679d2827594b159ace362151318689078bc49d3158f42899636\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:05Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:00Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:00 crc kubenswrapper[4870]: I1014 07:02:00.392073 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:00 crc kubenswrapper[4870]: I1014 07:02:00.392114 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:00 crc kubenswrapper[4870]: I1014 07:02:00.392127 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:00 crc kubenswrapper[4870]: I1014 07:02:00.392146 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:00 crc kubenswrapper[4870]: I1014 07:02:00.392157 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:00Z","lastTransitionTime":"2025-10-14T07:02:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:00 crc kubenswrapper[4870]: I1014 07:02:00.394418 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:00Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:00 crc kubenswrapper[4870]: I1014 07:02:00.404655 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-4v2v2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4796ba6-c23b-4764-be96-2e3a3bf5e962\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf5afb0b99a7101f6d5573fb7eb925c5a1d8becc11b2e3c6f3351d80b6f0e823\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnp8b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-4v2v2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:00Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:00 crc kubenswrapper[4870]: I1014 07:02:00.417008 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-cdjmj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"612e7ab9-bda4-45d8-bc50-ff1afb68aace\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7776352d62b814fc63d01e4c16ad13f878b0597eb37a370bf9da0a88abab6cfa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://834444b5350ee12df4e4652c3009880d8d69e1edf8ba0c0156ab96fb2bbe28e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://834444b5350ee12df4e4652c3009880d8d69e1edf8ba0c0156ab96fb2bbe28e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c1972e7b2f75790db22a07a078eee97072bffbdb7f8e594bf2be5f800632a7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c1972e7b2f75790db22a07a078eee97072bffbdb7f8e594bf2be5f800632a7b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b124b42a44c6a23da1d92b9a5326987eb74430f80489a71e8913829475654ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b124b42a44c6a23da1d92b9a5326987eb74430f80489a71e8913829475654ab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62d0deb6b20ea65fbcbaf94c2fc14d69adb22e7a46cc043ccd4def2668ddd4b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62d0deb6b20ea65fbcbaf94c2fc14d69adb22e7a46cc043ccd4def2668ddd4b7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7f0d7cf29ca737d0effe58731a8b890379527c97f78c144d4eebb910260fadb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7f0d7cf29ca737d0effe58731a8b890379527c97f78c144d4eebb910260fadb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5959abb7c8e7ab32bed3aca709ae0334e8400f991ba724246c9fda56cbed822\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5959abb7c8e7ab32bed3aca709ae0334e8400f991ba724246c9fda56cbed822\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-cdjmj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:00Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:00 crc kubenswrapper[4870]: I1014 07:02:00.429041 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fdxjh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a3a430a-b540-4a8c-adad-b2893dbf9898\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93e41d5e927eed7ec62eb9c3607cc804cd8999556663363fd123b2a784b00ff1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k787j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fdxjh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:00Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:00 crc kubenswrapper[4870]: I1014 07:02:00.439810 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-khtq8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28deeea6-a1ef-4a4a-a377-d954ccd87de9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://296bb048a1d99649beb57d609573833bf547cd4d628ebc823c6af83fe5736c9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rb79p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2b7bc4c93307e80b778f2282ced993898f1c4cfc350a7d74828b8f9dea9906d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rb79p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-khtq8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:00Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:00 crc kubenswrapper[4870]: I1014 07:02:00.458768 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33a9493e-9660-48db-b8d1-310e29ebe313\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3570643ef1b7d5cc3a542f320413b819338e0c8b7ab7f0f09cc909997a39bab8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb398668197dedc1c2a14b4ae7d6a29714f3527309ccb9c1622af841100e1d60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1d2b6d068d63bd38cdfb83824525f6d12054f6cabab5d684e03ea5f2152076c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://783ff243e395333f3fc80cfae21358c532f0c9c5cb5515586be56b90ed10d912\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a8135a3d01076e4437ae8a784b459bf8d16de57fa16ec7ce4d40c5129c9d4a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48ca3a5017ba650df99b2fc7201c3fe5abfa89d81089c5647cde1b1549f534fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48ca3a5017ba650df99b2fc7201c3fe5abfa89d81089c5647cde1b1549f534fc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://19682eca67f76bf2e06e1b80f12e21141fcbce0acb8195f0defacd1b9a10b5e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://19682eca67f76bf2e06e1b80f12e21141fcbce0acb8195f0defacd1b9a10b5e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://042aad6d5f6bee8eef77f9e6a2c265de775a69f3da72aedc68bdf312198f00fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://042aad6d5f6bee8eef77f9e6a2c265de775a69f3da72aedc68bdf312198f00fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:05Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:00Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:00 crc kubenswrapper[4870]: I1014 07:02:00.469802 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:00Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:00 crc kubenswrapper[4870]: I1014 07:02:00.479490 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-kz77f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d3864a91-9bf6-4540-a908-cc41aa39656c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d4ded3d86030f5ea548c8b1cc3caddbe14a45c9a3ae4af836e892880584f57d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4h8v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:28Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-kz77f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:00Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:00 crc kubenswrapper[4870]: I1014 07:02:00.492013 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bcbc6dd6-aab0-4dbc-a62e-c58f06ec1337\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25ec4808b09adc44488530b3c4e29fd7fec46b3fcfa31d4449ec4a6dbc880ec8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f3f5de86fb11231a19034a18e9d7f31476c162e6567d0c136ead8bbbd900d6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d57d938d9d9c92198ae5b3857536ece36012fbfbe8cc8cfd245d1386d39b0be5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77a3247a7f65709543d7fd374b2ab69d7d793537d4132d7e22261188d6e58544\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:05Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:00Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:00 crc kubenswrapper[4870]: I1014 07:02:00.493763 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:00 crc kubenswrapper[4870]: I1014 07:02:00.493859 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:00 crc kubenswrapper[4870]: I1014 07:02:00.493941 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:00 crc kubenswrapper[4870]: I1014 07:02:00.494026 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:00 crc kubenswrapper[4870]: I1014 07:02:00.494113 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:00Z","lastTransitionTime":"2025-10-14T07:02:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:00 crc kubenswrapper[4870]: I1014 07:02:00.505509 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e539fac3d0a2ade314a4b4c1050bb27c37bcff5ba5272fa2d71b97ab5c0d7191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:00Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:00 crc kubenswrapper[4870]: I1014 07:02:00.597225 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:00 crc kubenswrapper[4870]: I1014 07:02:00.597275 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:00 crc kubenswrapper[4870]: I1014 07:02:00.597287 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:00 crc kubenswrapper[4870]: I1014 07:02:00.597302 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:00 crc kubenswrapper[4870]: I1014 07:02:00.597313 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:00Z","lastTransitionTime":"2025-10-14T07:02:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:00 crc kubenswrapper[4870]: I1014 07:02:00.701597 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:00 crc kubenswrapper[4870]: I1014 07:02:00.701644 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:00 crc kubenswrapper[4870]: I1014 07:02:00.701654 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:00 crc kubenswrapper[4870]: I1014 07:02:00.701672 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:00 crc kubenswrapper[4870]: I1014 07:02:00.701684 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:00Z","lastTransitionTime":"2025-10-14T07:02:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:00 crc kubenswrapper[4870]: I1014 07:02:00.804271 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:00 crc kubenswrapper[4870]: I1014 07:02:00.804317 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:00 crc kubenswrapper[4870]: I1014 07:02:00.804338 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:00 crc kubenswrapper[4870]: I1014 07:02:00.804359 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:00 crc kubenswrapper[4870]: I1014 07:02:00.804371 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:00Z","lastTransitionTime":"2025-10-14T07:02:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:00 crc kubenswrapper[4870]: I1014 07:02:00.907053 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:00 crc kubenswrapper[4870]: I1014 07:02:00.907414 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:00 crc kubenswrapper[4870]: I1014 07:02:00.907637 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:00 crc kubenswrapper[4870]: I1014 07:02:00.907810 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:00 crc kubenswrapper[4870]: I1014 07:02:00.907958 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:00Z","lastTransitionTime":"2025-10-14T07:02:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:01 crc kubenswrapper[4870]: I1014 07:02:01.011154 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:01 crc kubenswrapper[4870]: I1014 07:02:01.011614 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:01 crc kubenswrapper[4870]: I1014 07:02:01.011661 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:01 crc kubenswrapper[4870]: I1014 07:02:01.011696 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:01 crc kubenswrapper[4870]: I1014 07:02:01.011721 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:01Z","lastTransitionTime":"2025-10-14T07:02:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:01 crc kubenswrapper[4870]: I1014 07:02:01.034097 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zc5j6" Oct 14 07:02:01 crc kubenswrapper[4870]: E1014 07:02:01.034382 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zc5j6" podUID="7d7a4f41-388a-4320-b896-43c17ff13da6" Oct 14 07:02:01 crc kubenswrapper[4870]: I1014 07:02:01.115002 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:01 crc kubenswrapper[4870]: I1014 07:02:01.115050 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:01 crc kubenswrapper[4870]: I1014 07:02:01.115059 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:01 crc kubenswrapper[4870]: I1014 07:02:01.115075 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:01 crc kubenswrapper[4870]: I1014 07:02:01.115086 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:01Z","lastTransitionTime":"2025-10-14T07:02:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:01 crc kubenswrapper[4870]: I1014 07:02:01.218299 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:01 crc kubenswrapper[4870]: I1014 07:02:01.218350 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:01 crc kubenswrapper[4870]: I1014 07:02:01.218367 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:01 crc kubenswrapper[4870]: I1014 07:02:01.218393 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:01 crc kubenswrapper[4870]: I1014 07:02:01.218410 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:01Z","lastTransitionTime":"2025-10-14T07:02:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:01 crc kubenswrapper[4870]: I1014 07:02:01.320853 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:01 crc kubenswrapper[4870]: I1014 07:02:01.320888 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:01 crc kubenswrapper[4870]: I1014 07:02:01.320898 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:01 crc kubenswrapper[4870]: I1014 07:02:01.320912 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:01 crc kubenswrapper[4870]: I1014 07:02:01.320922 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:01Z","lastTransitionTime":"2025-10-14T07:02:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:01 crc kubenswrapper[4870]: I1014 07:02:01.428294 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:01 crc kubenswrapper[4870]: I1014 07:02:01.428335 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:01 crc kubenswrapper[4870]: I1014 07:02:01.428346 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:01 crc kubenswrapper[4870]: I1014 07:02:01.428395 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:01 crc kubenswrapper[4870]: I1014 07:02:01.428406 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:01Z","lastTransitionTime":"2025-10-14T07:02:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:01 crc kubenswrapper[4870]: I1014 07:02:01.531203 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:01 crc kubenswrapper[4870]: I1014 07:02:01.531283 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:01 crc kubenswrapper[4870]: I1014 07:02:01.531326 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:01 crc kubenswrapper[4870]: I1014 07:02:01.531361 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:01 crc kubenswrapper[4870]: I1014 07:02:01.531396 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:01Z","lastTransitionTime":"2025-10-14T07:02:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:01 crc kubenswrapper[4870]: I1014 07:02:01.633644 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:01 crc kubenswrapper[4870]: I1014 07:02:01.633684 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:01 crc kubenswrapper[4870]: I1014 07:02:01.633693 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:01 crc kubenswrapper[4870]: I1014 07:02:01.633706 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:01 crc kubenswrapper[4870]: I1014 07:02:01.633715 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:01Z","lastTransitionTime":"2025-10-14T07:02:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:01 crc kubenswrapper[4870]: I1014 07:02:01.736630 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:01 crc kubenswrapper[4870]: I1014 07:02:01.736672 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:01 crc kubenswrapper[4870]: I1014 07:02:01.736682 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:01 crc kubenswrapper[4870]: I1014 07:02:01.736700 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:01 crc kubenswrapper[4870]: I1014 07:02:01.736712 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:01Z","lastTransitionTime":"2025-10-14T07:02:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:01 crc kubenswrapper[4870]: I1014 07:02:01.840033 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:01 crc kubenswrapper[4870]: I1014 07:02:01.840105 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:01 crc kubenswrapper[4870]: I1014 07:02:01.840115 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:01 crc kubenswrapper[4870]: I1014 07:02:01.840130 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:01 crc kubenswrapper[4870]: I1014 07:02:01.840139 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:01Z","lastTransitionTime":"2025-10-14T07:02:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:01 crc kubenswrapper[4870]: I1014 07:02:01.942305 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:01 crc kubenswrapper[4870]: I1014 07:02:01.942354 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:01 crc kubenswrapper[4870]: I1014 07:02:01.942365 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:01 crc kubenswrapper[4870]: I1014 07:02:01.942383 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:01 crc kubenswrapper[4870]: I1014 07:02:01.942397 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:01Z","lastTransitionTime":"2025-10-14T07:02:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:02 crc kubenswrapper[4870]: I1014 07:02:02.033747 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 07:02:02 crc kubenswrapper[4870]: I1014 07:02:02.033763 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 07:02:02 crc kubenswrapper[4870]: I1014 07:02:02.033903 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 07:02:02 crc kubenswrapper[4870]: E1014 07:02:02.034082 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 07:02:02 crc kubenswrapper[4870]: E1014 07:02:02.034154 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 07:02:02 crc kubenswrapper[4870]: E1014 07:02:02.034319 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 07:02:02 crc kubenswrapper[4870]: I1014 07:02:02.044795 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:02 crc kubenswrapper[4870]: I1014 07:02:02.044854 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:02 crc kubenswrapper[4870]: I1014 07:02:02.044871 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:02 crc kubenswrapper[4870]: I1014 07:02:02.044894 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:02 crc kubenswrapper[4870]: I1014 07:02:02.044911 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:02Z","lastTransitionTime":"2025-10-14T07:02:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:02 crc kubenswrapper[4870]: I1014 07:02:02.147146 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:02 crc kubenswrapper[4870]: I1014 07:02:02.147200 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:02 crc kubenswrapper[4870]: I1014 07:02:02.147214 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:02 crc kubenswrapper[4870]: I1014 07:02:02.147236 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:02 crc kubenswrapper[4870]: I1014 07:02:02.147284 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:02Z","lastTransitionTime":"2025-10-14T07:02:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:02 crc kubenswrapper[4870]: I1014 07:02:02.249505 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:02 crc kubenswrapper[4870]: I1014 07:02:02.249573 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:02 crc kubenswrapper[4870]: I1014 07:02:02.249591 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:02 crc kubenswrapper[4870]: I1014 07:02:02.249617 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:02 crc kubenswrapper[4870]: I1014 07:02:02.249634 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:02Z","lastTransitionTime":"2025-10-14T07:02:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:02 crc kubenswrapper[4870]: I1014 07:02:02.352298 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:02 crc kubenswrapper[4870]: I1014 07:02:02.352340 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:02 crc kubenswrapper[4870]: I1014 07:02:02.352354 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:02 crc kubenswrapper[4870]: I1014 07:02:02.352373 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:02 crc kubenswrapper[4870]: I1014 07:02:02.352387 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:02Z","lastTransitionTime":"2025-10-14T07:02:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:02 crc kubenswrapper[4870]: I1014 07:02:02.455636 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:02 crc kubenswrapper[4870]: I1014 07:02:02.455886 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:02 crc kubenswrapper[4870]: I1014 07:02:02.455993 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:02 crc kubenswrapper[4870]: I1014 07:02:02.456069 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:02 crc kubenswrapper[4870]: I1014 07:02:02.456146 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:02Z","lastTransitionTime":"2025-10-14T07:02:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:02 crc kubenswrapper[4870]: I1014 07:02:02.558345 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:02 crc kubenswrapper[4870]: I1014 07:02:02.558382 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:02 crc kubenswrapper[4870]: I1014 07:02:02.558409 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:02 crc kubenswrapper[4870]: I1014 07:02:02.558426 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:02 crc kubenswrapper[4870]: I1014 07:02:02.558457 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:02Z","lastTransitionTime":"2025-10-14T07:02:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:02 crc kubenswrapper[4870]: I1014 07:02:02.661687 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:02 crc kubenswrapper[4870]: I1014 07:02:02.661753 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:02 crc kubenswrapper[4870]: I1014 07:02:02.661775 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:02 crc kubenswrapper[4870]: I1014 07:02:02.661804 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:02 crc kubenswrapper[4870]: I1014 07:02:02.661826 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:02Z","lastTransitionTime":"2025-10-14T07:02:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:02 crc kubenswrapper[4870]: I1014 07:02:02.764647 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:02 crc kubenswrapper[4870]: I1014 07:02:02.764764 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:02 crc kubenswrapper[4870]: I1014 07:02:02.764789 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:02 crc kubenswrapper[4870]: I1014 07:02:02.764816 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:02 crc kubenswrapper[4870]: I1014 07:02:02.764833 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:02Z","lastTransitionTime":"2025-10-14T07:02:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:02 crc kubenswrapper[4870]: I1014 07:02:02.867871 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:02 crc kubenswrapper[4870]: I1014 07:02:02.868252 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:02 crc kubenswrapper[4870]: I1014 07:02:02.868378 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:02 crc kubenswrapper[4870]: I1014 07:02:02.868521 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:02 crc kubenswrapper[4870]: I1014 07:02:02.868622 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:02Z","lastTransitionTime":"2025-10-14T07:02:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:02 crc kubenswrapper[4870]: I1014 07:02:02.972049 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:02 crc kubenswrapper[4870]: I1014 07:02:02.972580 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:02 crc kubenswrapper[4870]: I1014 07:02:02.972807 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:02 crc kubenswrapper[4870]: I1014 07:02:02.972998 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:02 crc kubenswrapper[4870]: I1014 07:02:02.973125 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:02Z","lastTransitionTime":"2025-10-14T07:02:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:03 crc kubenswrapper[4870]: I1014 07:02:03.033873 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zc5j6" Oct 14 07:02:03 crc kubenswrapper[4870]: E1014 07:02:03.034091 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zc5j6" podUID="7d7a4f41-388a-4320-b896-43c17ff13da6" Oct 14 07:02:03 crc kubenswrapper[4870]: I1014 07:02:03.075875 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:03 crc kubenswrapper[4870]: I1014 07:02:03.075939 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:03 crc kubenswrapper[4870]: I1014 07:02:03.075960 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:03 crc kubenswrapper[4870]: I1014 07:02:03.075986 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:03 crc kubenswrapper[4870]: I1014 07:02:03.076003 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:03Z","lastTransitionTime":"2025-10-14T07:02:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:03 crc kubenswrapper[4870]: I1014 07:02:03.150788 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:03 crc kubenswrapper[4870]: I1014 07:02:03.150834 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:03 crc kubenswrapper[4870]: I1014 07:02:03.150852 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:03 crc kubenswrapper[4870]: I1014 07:02:03.150874 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:03 crc kubenswrapper[4870]: I1014 07:02:03.150886 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:03Z","lastTransitionTime":"2025-10-14T07:02:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:03 crc kubenswrapper[4870]: E1014 07:02:03.162112 4870 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:02:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:02:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:02:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:02:03Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:02:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:02:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:02:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:02:03Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"56448e6b-6b3a-4bb9-91f8-2f0a9b44e9aa\\\",\\\"systemUUID\\\":\\\"0adb347d-b90d-4e9f-9962-f1bd97980aed\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:03Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:03 crc kubenswrapper[4870]: I1014 07:02:03.165821 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:03 crc kubenswrapper[4870]: I1014 07:02:03.165867 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:03 crc kubenswrapper[4870]: I1014 07:02:03.165891 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:03 crc kubenswrapper[4870]: I1014 07:02:03.165914 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:03 crc kubenswrapper[4870]: I1014 07:02:03.165932 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:03Z","lastTransitionTime":"2025-10-14T07:02:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:03 crc kubenswrapper[4870]: E1014 07:02:03.177823 4870 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:02:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:02:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:02:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:02:03Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:02:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:02:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:02:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:02:03Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"56448e6b-6b3a-4bb9-91f8-2f0a9b44e9aa\\\",\\\"systemUUID\\\":\\\"0adb347d-b90d-4e9f-9962-f1bd97980aed\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:03Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:03 crc kubenswrapper[4870]: I1014 07:02:03.181945 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:03 crc kubenswrapper[4870]: I1014 07:02:03.181973 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:03 crc kubenswrapper[4870]: I1014 07:02:03.181983 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:03 crc kubenswrapper[4870]: I1014 07:02:03.181998 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:03 crc kubenswrapper[4870]: I1014 07:02:03.182008 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:03Z","lastTransitionTime":"2025-10-14T07:02:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:03 crc kubenswrapper[4870]: E1014 07:02:03.199176 4870 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:02:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:02:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:02:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:02:03Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:02:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:02:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:02:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:02:03Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"56448e6b-6b3a-4bb9-91f8-2f0a9b44e9aa\\\",\\\"systemUUID\\\":\\\"0adb347d-b90d-4e9f-9962-f1bd97980aed\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:03Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:03 crc kubenswrapper[4870]: I1014 07:02:03.203472 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:03 crc kubenswrapper[4870]: I1014 07:02:03.203510 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:03 crc kubenswrapper[4870]: I1014 07:02:03.203523 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:03 crc kubenswrapper[4870]: I1014 07:02:03.203540 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:03 crc kubenswrapper[4870]: I1014 07:02:03.203553 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:03Z","lastTransitionTime":"2025-10-14T07:02:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:03 crc kubenswrapper[4870]: E1014 07:02:03.217838 4870 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:02:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:02:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:02:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:02:03Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:02:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:02:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:02:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:02:03Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"56448e6b-6b3a-4bb9-91f8-2f0a9b44e9aa\\\",\\\"systemUUID\\\":\\\"0adb347d-b90d-4e9f-9962-f1bd97980aed\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:03Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:03 crc kubenswrapper[4870]: I1014 07:02:03.221608 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:03 crc kubenswrapper[4870]: I1014 07:02:03.221651 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:03 crc kubenswrapper[4870]: I1014 07:02:03.221667 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:03 crc kubenswrapper[4870]: I1014 07:02:03.221688 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:03 crc kubenswrapper[4870]: I1014 07:02:03.221703 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:03Z","lastTransitionTime":"2025-10-14T07:02:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:03 crc kubenswrapper[4870]: E1014 07:02:03.234318 4870 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:02:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:02:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:02:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:02:03Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:02:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:02:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:02:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:02:03Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"56448e6b-6b3a-4bb9-91f8-2f0a9b44e9aa\\\",\\\"systemUUID\\\":\\\"0adb347d-b90d-4e9f-9962-f1bd97980aed\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:03Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:03 crc kubenswrapper[4870]: E1014 07:02:03.234563 4870 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 14 07:02:03 crc kubenswrapper[4870]: I1014 07:02:03.236395 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:03 crc kubenswrapper[4870]: I1014 07:02:03.236422 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:03 crc kubenswrapper[4870]: I1014 07:02:03.236431 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:03 crc kubenswrapper[4870]: I1014 07:02:03.236462 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:03 crc kubenswrapper[4870]: I1014 07:02:03.236472 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:03Z","lastTransitionTime":"2025-10-14T07:02:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:03 crc kubenswrapper[4870]: I1014 07:02:03.338899 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:03 crc kubenswrapper[4870]: I1014 07:02:03.338928 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:03 crc kubenswrapper[4870]: I1014 07:02:03.338936 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:03 crc kubenswrapper[4870]: I1014 07:02:03.338948 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:03 crc kubenswrapper[4870]: I1014 07:02:03.338957 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:03Z","lastTransitionTime":"2025-10-14T07:02:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:03 crc kubenswrapper[4870]: I1014 07:02:03.442388 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:03 crc kubenswrapper[4870]: I1014 07:02:03.442429 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:03 crc kubenswrapper[4870]: I1014 07:02:03.442459 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:03 crc kubenswrapper[4870]: I1014 07:02:03.442476 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:03 crc kubenswrapper[4870]: I1014 07:02:03.442486 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:03Z","lastTransitionTime":"2025-10-14T07:02:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:03 crc kubenswrapper[4870]: I1014 07:02:03.545991 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:03 crc kubenswrapper[4870]: I1014 07:02:03.546066 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:03 crc kubenswrapper[4870]: I1014 07:02:03.546079 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:03 crc kubenswrapper[4870]: I1014 07:02:03.546097 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:03 crc kubenswrapper[4870]: I1014 07:02:03.546110 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:03Z","lastTransitionTime":"2025-10-14T07:02:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:03 crc kubenswrapper[4870]: I1014 07:02:03.649758 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:03 crc kubenswrapper[4870]: I1014 07:02:03.649822 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:03 crc kubenswrapper[4870]: I1014 07:02:03.649839 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:03 crc kubenswrapper[4870]: I1014 07:02:03.649866 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:03 crc kubenswrapper[4870]: I1014 07:02:03.649884 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:03Z","lastTransitionTime":"2025-10-14T07:02:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:03 crc kubenswrapper[4870]: I1014 07:02:03.752727 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:03 crc kubenswrapper[4870]: I1014 07:02:03.752789 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:03 crc kubenswrapper[4870]: I1014 07:02:03.752802 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:03 crc kubenswrapper[4870]: I1014 07:02:03.752818 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:03 crc kubenswrapper[4870]: I1014 07:02:03.752828 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:03Z","lastTransitionTime":"2025-10-14T07:02:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:03 crc kubenswrapper[4870]: I1014 07:02:03.854845 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:03 crc kubenswrapper[4870]: I1014 07:02:03.854911 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:03 crc kubenswrapper[4870]: I1014 07:02:03.854929 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:03 crc kubenswrapper[4870]: I1014 07:02:03.854955 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:03 crc kubenswrapper[4870]: I1014 07:02:03.854973 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:03Z","lastTransitionTime":"2025-10-14T07:02:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:03 crc kubenswrapper[4870]: I1014 07:02:03.957506 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:03 crc kubenswrapper[4870]: I1014 07:02:03.957544 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:03 crc kubenswrapper[4870]: I1014 07:02:03.957555 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:03 crc kubenswrapper[4870]: I1014 07:02:03.957570 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:03 crc kubenswrapper[4870]: I1014 07:02:03.957580 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:03Z","lastTransitionTime":"2025-10-14T07:02:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:04 crc kubenswrapper[4870]: I1014 07:02:04.033345 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 07:02:04 crc kubenswrapper[4870]: E1014 07:02:04.033494 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 07:02:04 crc kubenswrapper[4870]: I1014 07:02:04.033663 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 07:02:04 crc kubenswrapper[4870]: E1014 07:02:04.033751 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 07:02:04 crc kubenswrapper[4870]: I1014 07:02:04.033762 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 07:02:04 crc kubenswrapper[4870]: E1014 07:02:04.033949 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 07:02:04 crc kubenswrapper[4870]: I1014 07:02:04.059882 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:04 crc kubenswrapper[4870]: I1014 07:02:04.059919 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:04 crc kubenswrapper[4870]: I1014 07:02:04.059944 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:04 crc kubenswrapper[4870]: I1014 07:02:04.059961 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:04 crc kubenswrapper[4870]: I1014 07:02:04.059970 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:04Z","lastTransitionTime":"2025-10-14T07:02:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:04 crc kubenswrapper[4870]: I1014 07:02:04.162782 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:04 crc kubenswrapper[4870]: I1014 07:02:04.162861 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:04 crc kubenswrapper[4870]: I1014 07:02:04.162879 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:04 crc kubenswrapper[4870]: I1014 07:02:04.162904 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:04 crc kubenswrapper[4870]: I1014 07:02:04.162923 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:04Z","lastTransitionTime":"2025-10-14T07:02:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:04 crc kubenswrapper[4870]: I1014 07:02:04.265729 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:04 crc kubenswrapper[4870]: I1014 07:02:04.265765 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:04 crc kubenswrapper[4870]: I1014 07:02:04.265776 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:04 crc kubenswrapper[4870]: I1014 07:02:04.265791 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:04 crc kubenswrapper[4870]: I1014 07:02:04.265804 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:04Z","lastTransitionTime":"2025-10-14T07:02:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:04 crc kubenswrapper[4870]: I1014 07:02:04.369370 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:04 crc kubenswrapper[4870]: I1014 07:02:04.369504 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:04 crc kubenswrapper[4870]: I1014 07:02:04.369530 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:04 crc kubenswrapper[4870]: I1014 07:02:04.369561 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:04 crc kubenswrapper[4870]: I1014 07:02:04.369583 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:04Z","lastTransitionTime":"2025-10-14T07:02:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:04 crc kubenswrapper[4870]: I1014 07:02:04.472398 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:04 crc kubenswrapper[4870]: I1014 07:02:04.472458 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:04 crc kubenswrapper[4870]: I1014 07:02:04.472476 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:04 crc kubenswrapper[4870]: I1014 07:02:04.472493 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:04 crc kubenswrapper[4870]: I1014 07:02:04.472504 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:04Z","lastTransitionTime":"2025-10-14T07:02:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:04 crc kubenswrapper[4870]: I1014 07:02:04.574558 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:04 crc kubenswrapper[4870]: I1014 07:02:04.574611 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:04 crc kubenswrapper[4870]: I1014 07:02:04.574625 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:04 crc kubenswrapper[4870]: I1014 07:02:04.574644 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:04 crc kubenswrapper[4870]: I1014 07:02:04.574658 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:04Z","lastTransitionTime":"2025-10-14T07:02:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:04 crc kubenswrapper[4870]: I1014 07:02:04.677233 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:04 crc kubenswrapper[4870]: I1014 07:02:04.677276 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:04 crc kubenswrapper[4870]: I1014 07:02:04.677288 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:04 crc kubenswrapper[4870]: I1014 07:02:04.677304 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:04 crc kubenswrapper[4870]: I1014 07:02:04.677316 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:04Z","lastTransitionTime":"2025-10-14T07:02:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:04 crc kubenswrapper[4870]: I1014 07:02:04.779549 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:04 crc kubenswrapper[4870]: I1014 07:02:04.779582 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:04 crc kubenswrapper[4870]: I1014 07:02:04.779590 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:04 crc kubenswrapper[4870]: I1014 07:02:04.779603 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:04 crc kubenswrapper[4870]: I1014 07:02:04.779613 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:04Z","lastTransitionTime":"2025-10-14T07:02:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:04 crc kubenswrapper[4870]: I1014 07:02:04.883010 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:04 crc kubenswrapper[4870]: I1014 07:02:04.883084 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:04 crc kubenswrapper[4870]: I1014 07:02:04.883107 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:04 crc kubenswrapper[4870]: I1014 07:02:04.883140 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:04 crc kubenswrapper[4870]: I1014 07:02:04.883160 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:04Z","lastTransitionTime":"2025-10-14T07:02:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:04 crc kubenswrapper[4870]: I1014 07:02:04.985149 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:04 crc kubenswrapper[4870]: I1014 07:02:04.985217 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:04 crc kubenswrapper[4870]: I1014 07:02:04.985253 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:04 crc kubenswrapper[4870]: I1014 07:02:04.989582 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:04 crc kubenswrapper[4870]: I1014 07:02:04.989674 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:04Z","lastTransitionTime":"2025-10-14T07:02:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:05 crc kubenswrapper[4870]: I1014 07:02:05.033602 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zc5j6" Oct 14 07:02:05 crc kubenswrapper[4870]: E1014 07:02:05.033779 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zc5j6" podUID="7d7a4f41-388a-4320-b896-43c17ff13da6" Oct 14 07:02:05 crc kubenswrapper[4870]: I1014 07:02:05.048124 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"874a5e45-dffd-4d17-b609-b7d3ed2eab07\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d98d044836ab44a405196915ef7bf572b75113f2c81ec7e64effa63b47ad3543\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wj4ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8004752575df4a1c44efbcafe64799f09126d0f97285d92388c76a0b90e022f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wj4ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7tvc7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:05Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:05 crc kubenswrapper[4870]: I1014 07:02:05.063739 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zc5j6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d7a4f41-388a-4320-b896-43c17ff13da6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6p2wh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6p2wh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:40Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zc5j6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:05Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:05 crc kubenswrapper[4870]: I1014 07:02:05.078355 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:05Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:05 crc kubenswrapper[4870]: I1014 07:02:05.093583 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:05 crc kubenswrapper[4870]: I1014 07:02:05.093635 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:05 crc kubenswrapper[4870]: I1014 07:02:05.093646 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:05 crc kubenswrapper[4870]: I1014 07:02:05.093664 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:05 crc kubenswrapper[4870]: I1014 07:02:05.093674 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:05Z","lastTransitionTime":"2025-10-14T07:02:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:05 crc kubenswrapper[4870]: I1014 07:02:05.098315 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3eb0b397c05ffd6650607f1be4a75683e0342d3333fc096bf1cd63c7688834f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:05Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:05 crc kubenswrapper[4870]: I1014 07:02:05.115382 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e26fe49ec2045d20a0cb2dd17e96e79b6eb43c641eb184c688cc472ca5db39d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aea115f7ee71c700b218fe15c928af7b55e6babe70e65d67f125d87be1fd2e2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:05Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:05 crc kubenswrapper[4870]: I1014 07:02:05.137530 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-v7br2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dcaef0e-42e3-424b-bbe5-4a023efe912d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71dec55e859b11a081dbb90d7d6cf6b4846cf591382303d8329003ff50c71f2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e72cff0723e2ce520340244b4e597d8016c58e1b071c4b087f2f6cd718741594\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc9d4db944f837289d1923176d60cfc4a566a282c2d2e3a703e220df884703c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d16ed1c65925549dcb69ab02c3013e25e3f123a83179f1af3a8e08bf91a59896\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed59243b59c39ce82bf9bc8f33c0a8b84854846afff6a89f5d37ffc97a096b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3b6d20cf459b1f6d3aff5b9290f9391909baf27ee40c49e2b0cd2d5cfe889ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fc9bfcb356b506d6c439adc22e4caaed4fad1c5a085744d44ebd50cf8898777\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3fc9bfcb356b506d6c439adc22e4caaed4fad1c5a085744d44ebd50cf8898777\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-14T07:01:56Z\\\",\\\"message\\\":\\\" 6555 ovn.go:134] Ensuring zone local for Pod openshift-image-registry/node-ca-kz77f in node crc\\\\nI1014 07:01:56.821005 6555 obj_retry.go:386] Retry successful for *v1.Pod openshift-image-registry/node-ca-kz77f after 0 failed attempt(s)\\\\nI1014 07:01:56.820992 6555 model_client.go:382] Update operations generated as: [{Op:update Table:Logical_Switch_Port Row:map[addresses:{GoSet:[0a:58:0a:d9:00:5c 10.217.0.92]} options:{GoMap:map[iface-id-ver:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 requested-chassis:crc]} port_security:{GoSet:[0a:58:0a:d9:00:5c 10.217.0.92]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c94130be-172c-477c-88c4-40cc7eba30fe}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1014 07:01:56.821025 6555 default_network_controller.go:776] Recording success event on pod openshift-image-registry/node-ca-kz77f\\\\nI1014 07:01:56.820931 6555 default_network_controller.go:776] Recording success event on pod openshift-dns/node-resolver-4v2v2\\\\nI1014 07:01:56.820614 6555 services_controller.go:454] Service openshift-marketplace/redhat-marketplace for network=default has 1 cluster-wide, 0 per-node configs, 0 template configs, making 1 (cluster) 0 (per node) and 0 (template) load balancers\\\\nI1014 07:01:56.821037 6555 model_client.go:382] Update operations generated as: [{Op:update Table:N\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:56Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-v7br2_openshift-ovn-kubernetes(8dcaef0e-42e3-424b-bbe5-4a023efe912d)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fe3468cdd612fc4378e1f6f55a2a9d00274f7ddb65cb57b2a368dfc065cfb4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ba24c2e067b9f3a14d9d018b66300b54d866530d30174fbdb3c899299575d28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ba24c2e067b9f3a14d9d018b66300b54d866530d30174fbdb3c899299575d28\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-v7br2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:05Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:05 crc kubenswrapper[4870]: I1014 07:02:05.158470 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77a0d93a-5391-4112-b502-c474f45de209\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7815af9a584f6db6660073da289f2e8c987a04ef9ddfac155bdd9a15b014fcfa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b329b570622155a55f70501b2260e278106a68ebef497914231b4ec860e0125\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba942a0e6fe5ab93a187e0b7b4bd7c440c2e7827d9b5323fc50f90b562075392\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b9018b078c5c8276243c1696860f0aa73e2d8dd917913ae9d03b30dc9b5113a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://978e8db184ba9e430232c6bb2ee530beb92567fdc71ba3c35b0bedca3492d845\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"file observer\\\\nW1014 07:01:25.870672 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1014 07:01:25.870795 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1014 07:01:25.871521 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-756714914/tls.crt::/tmp/serving-cert-756714914/tls.key\\\\\\\"\\\\nI1014 07:01:26.460037 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1014 07:01:26.464876 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1014 07:01:26.464902 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1014 07:01:26.464925 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1014 07:01:26.464931 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1014 07:01:26.481374 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1014 07:01:26.481400 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:01:26.481405 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:01:26.481410 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1014 07:01:26.481414 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1014 07:01:26.481417 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1014 07:01:26.481420 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1014 07:01:26.481610 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1014 07:01:26.482872 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bd0044a247bdc5ca4cab7294103af3270b76cde61f3109b85e83ed050ca53b6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82d3c2214e0be679d2827594b159ace362151318689078bc49d3158f42899636\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82d3c2214e0be679d2827594b159ace362151318689078bc49d3158f42899636\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:05Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:05Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:05 crc kubenswrapper[4870]: I1014 07:02:05.171943 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:05Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:05 crc kubenswrapper[4870]: I1014 07:02:05.181618 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a77729fb-a187-48c6-80db-bee90f7bdac4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:02:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:02:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c859ebc0d20b632d1d5a716caee343c7e6b9b752ed5f257422cd35ad12c170f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e8a7a5c1f1706ebbd2e8165d8961002f5a9340e2749e377b0058db107ad7dbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f75b3dc1557917e9c0c35c2810338a076ed92ba89963edafb9d537ad19c2bd8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b192bd3e0eb8540d7492abc207025471bc9e198e7e78b50215c8d2f5bc57cf4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b192bd3e0eb8540d7492abc207025471bc9e198e7e78b50215c8d2f5bc57cf4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:05Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:05Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:05 crc kubenswrapper[4870]: I1014 07:02:05.189874 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-4v2v2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4796ba6-c23b-4764-be96-2e3a3bf5e962\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf5afb0b99a7101f6d5573fb7eb925c5a1d8becc11b2e3c6f3351d80b6f0e823\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnp8b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-4v2v2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:05Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:05 crc kubenswrapper[4870]: I1014 07:02:05.197054 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:05 crc kubenswrapper[4870]: I1014 07:02:05.197090 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:05 crc kubenswrapper[4870]: I1014 07:02:05.197107 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:05 crc kubenswrapper[4870]: I1014 07:02:05.197123 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:05 crc kubenswrapper[4870]: I1014 07:02:05.197134 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:05Z","lastTransitionTime":"2025-10-14T07:02:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:05 crc kubenswrapper[4870]: I1014 07:02:05.203050 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-cdjmj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"612e7ab9-bda4-45d8-bc50-ff1afb68aace\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7776352d62b814fc63d01e4c16ad13f878b0597eb37a370bf9da0a88abab6cfa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://834444b5350ee12df4e4652c3009880d8d69e1edf8ba0c0156ab96fb2bbe28e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://834444b5350ee12df4e4652c3009880d8d69e1edf8ba0c0156ab96fb2bbe28e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c1972e7b2f75790db22a07a078eee97072bffbdb7f8e594bf2be5f800632a7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c1972e7b2f75790db22a07a078eee97072bffbdb7f8e594bf2be5f800632a7b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b124b42a44c6a23da1d92b9a5326987eb74430f80489a71e8913829475654ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b124b42a44c6a23da1d92b9a5326987eb74430f80489a71e8913829475654ab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62d0deb6b20ea65fbcbaf94c2fc14d69adb22e7a46cc043ccd4def2668ddd4b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62d0deb6b20ea65fbcbaf94c2fc14d69adb22e7a46cc043ccd4def2668ddd4b7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7f0d7cf29ca737d0effe58731a8b890379527c97f78c144d4eebb910260fadb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7f0d7cf29ca737d0effe58731a8b890379527c97f78c144d4eebb910260fadb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5959abb7c8e7ab32bed3aca709ae0334e8400f991ba724246c9fda56cbed822\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5959abb7c8e7ab32bed3aca709ae0334e8400f991ba724246c9fda56cbed822\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-cdjmj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:05Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:05 crc kubenswrapper[4870]: I1014 07:02:05.215477 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fdxjh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a3a430a-b540-4a8c-adad-b2893dbf9898\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93e41d5e927eed7ec62eb9c3607cc804cd8999556663363fd123b2a784b00ff1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k787j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fdxjh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:05Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:05 crc kubenswrapper[4870]: I1014 07:02:05.228210 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-khtq8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28deeea6-a1ef-4a4a-a377-d954ccd87de9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://296bb048a1d99649beb57d609573833bf547cd4d628ebc823c6af83fe5736c9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rb79p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2b7bc4c93307e80b778f2282ced993898f1c4cfc350a7d74828b8f9dea9906d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rb79p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-khtq8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:05Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:05 crc kubenswrapper[4870]: I1014 07:02:05.248325 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33a9493e-9660-48db-b8d1-310e29ebe313\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3570643ef1b7d5cc3a542f320413b819338e0c8b7ab7f0f09cc909997a39bab8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb398668197dedc1c2a14b4ae7d6a29714f3527309ccb9c1622af841100e1d60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1d2b6d068d63bd38cdfb83824525f6d12054f6cabab5d684e03ea5f2152076c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://783ff243e395333f3fc80cfae21358c532f0c9c5cb5515586be56b90ed10d912\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a8135a3d01076e4437ae8a784b459bf8d16de57fa16ec7ce4d40c5129c9d4a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48ca3a5017ba650df99b2fc7201c3fe5abfa89d81089c5647cde1b1549f534fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48ca3a5017ba650df99b2fc7201c3fe5abfa89d81089c5647cde1b1549f534fc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://19682eca67f76bf2e06e1b80f12e21141fcbce0acb8195f0defacd1b9a10b5e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://19682eca67f76bf2e06e1b80f12e21141fcbce0acb8195f0defacd1b9a10b5e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://042aad6d5f6bee8eef77f9e6a2c265de775a69f3da72aedc68bdf312198f00fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://042aad6d5f6bee8eef77f9e6a2c265de775a69f3da72aedc68bdf312198f00fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:05Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:05Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:05 crc kubenswrapper[4870]: I1014 07:02:05.264082 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:05Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:05 crc kubenswrapper[4870]: I1014 07:02:05.276800 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-kz77f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d3864a91-9bf6-4540-a908-cc41aa39656c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d4ded3d86030f5ea548c8b1cc3caddbe14a45c9a3ae4af836e892880584f57d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4h8v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:28Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-kz77f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:05Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:05 crc kubenswrapper[4870]: I1014 07:02:05.294062 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bcbc6dd6-aab0-4dbc-a62e-c58f06ec1337\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25ec4808b09adc44488530b3c4e29fd7fec46b3fcfa31d4449ec4a6dbc880ec8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f3f5de86fb11231a19034a18e9d7f31476c162e6567d0c136ead8bbbd900d6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d57d938d9d9c92198ae5b3857536ece36012fbfbe8cc8cfd245d1386d39b0be5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77a3247a7f65709543d7fd374b2ab69d7d793537d4132d7e22261188d6e58544\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:05Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:05Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:05 crc kubenswrapper[4870]: I1014 07:02:05.299910 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:05 crc kubenswrapper[4870]: I1014 07:02:05.299937 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:05 crc kubenswrapper[4870]: I1014 07:02:05.299946 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:05 crc kubenswrapper[4870]: I1014 07:02:05.299959 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:05 crc kubenswrapper[4870]: I1014 07:02:05.299971 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:05Z","lastTransitionTime":"2025-10-14T07:02:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:05 crc kubenswrapper[4870]: I1014 07:02:05.309778 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e539fac3d0a2ade314a4b4c1050bb27c37bcff5ba5272fa2d71b97ab5c0d7191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:05Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:05 crc kubenswrapper[4870]: I1014 07:02:05.403331 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:05 crc kubenswrapper[4870]: I1014 07:02:05.403379 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:05 crc kubenswrapper[4870]: I1014 07:02:05.403389 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:05 crc kubenswrapper[4870]: I1014 07:02:05.403406 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:05 crc kubenswrapper[4870]: I1014 07:02:05.403417 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:05Z","lastTransitionTime":"2025-10-14T07:02:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:05 crc kubenswrapper[4870]: I1014 07:02:05.506299 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:05 crc kubenswrapper[4870]: I1014 07:02:05.506359 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:05 crc kubenswrapper[4870]: I1014 07:02:05.506372 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:05 crc kubenswrapper[4870]: I1014 07:02:05.506389 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:05 crc kubenswrapper[4870]: I1014 07:02:05.506402 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:05Z","lastTransitionTime":"2025-10-14T07:02:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:05 crc kubenswrapper[4870]: I1014 07:02:05.609947 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:05 crc kubenswrapper[4870]: I1014 07:02:05.609989 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:05 crc kubenswrapper[4870]: I1014 07:02:05.610011 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:05 crc kubenswrapper[4870]: I1014 07:02:05.610027 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:05 crc kubenswrapper[4870]: I1014 07:02:05.610039 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:05Z","lastTransitionTime":"2025-10-14T07:02:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:05 crc kubenswrapper[4870]: I1014 07:02:05.712130 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:05 crc kubenswrapper[4870]: I1014 07:02:05.712167 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:05 crc kubenswrapper[4870]: I1014 07:02:05.712190 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:05 crc kubenswrapper[4870]: I1014 07:02:05.712205 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:05 crc kubenswrapper[4870]: I1014 07:02:05.712214 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:05Z","lastTransitionTime":"2025-10-14T07:02:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:05 crc kubenswrapper[4870]: I1014 07:02:05.816121 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:05 crc kubenswrapper[4870]: I1014 07:02:05.816162 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:05 crc kubenswrapper[4870]: I1014 07:02:05.816172 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:05 crc kubenswrapper[4870]: I1014 07:02:05.816187 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:05 crc kubenswrapper[4870]: I1014 07:02:05.816200 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:05Z","lastTransitionTime":"2025-10-14T07:02:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:05 crc kubenswrapper[4870]: I1014 07:02:05.918563 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:05 crc kubenswrapper[4870]: I1014 07:02:05.918920 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:05 crc kubenswrapper[4870]: I1014 07:02:05.918934 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:05 crc kubenswrapper[4870]: I1014 07:02:05.918953 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:05 crc kubenswrapper[4870]: I1014 07:02:05.918964 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:05Z","lastTransitionTime":"2025-10-14T07:02:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:06 crc kubenswrapper[4870]: I1014 07:02:06.021775 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:06 crc kubenswrapper[4870]: I1014 07:02:06.021813 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:06 crc kubenswrapper[4870]: I1014 07:02:06.021823 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:06 crc kubenswrapper[4870]: I1014 07:02:06.021839 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:06 crc kubenswrapper[4870]: I1014 07:02:06.021848 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:06Z","lastTransitionTime":"2025-10-14T07:02:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:06 crc kubenswrapper[4870]: I1014 07:02:06.033467 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 07:02:06 crc kubenswrapper[4870]: I1014 07:02:06.033516 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 07:02:06 crc kubenswrapper[4870]: E1014 07:02:06.033641 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 07:02:06 crc kubenswrapper[4870]: I1014 07:02:06.033680 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 07:02:06 crc kubenswrapper[4870]: E1014 07:02:06.033792 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 07:02:06 crc kubenswrapper[4870]: E1014 07:02:06.033856 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 07:02:06 crc kubenswrapper[4870]: I1014 07:02:06.124717 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:06 crc kubenswrapper[4870]: I1014 07:02:06.124761 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:06 crc kubenswrapper[4870]: I1014 07:02:06.124773 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:06 crc kubenswrapper[4870]: I1014 07:02:06.124790 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:06 crc kubenswrapper[4870]: I1014 07:02:06.124802 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:06Z","lastTransitionTime":"2025-10-14T07:02:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:06 crc kubenswrapper[4870]: I1014 07:02:06.227884 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:06 crc kubenswrapper[4870]: I1014 07:02:06.227974 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:06 crc kubenswrapper[4870]: I1014 07:02:06.228013 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:06 crc kubenswrapper[4870]: I1014 07:02:06.228038 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:06 crc kubenswrapper[4870]: I1014 07:02:06.228056 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:06Z","lastTransitionTime":"2025-10-14T07:02:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:06 crc kubenswrapper[4870]: I1014 07:02:06.331025 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:06 crc kubenswrapper[4870]: I1014 07:02:06.331085 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:06 crc kubenswrapper[4870]: I1014 07:02:06.331110 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:06 crc kubenswrapper[4870]: I1014 07:02:06.331138 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:06 crc kubenswrapper[4870]: I1014 07:02:06.331158 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:06Z","lastTransitionTime":"2025-10-14T07:02:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:06 crc kubenswrapper[4870]: I1014 07:02:06.434328 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:06 crc kubenswrapper[4870]: I1014 07:02:06.434423 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:06 crc kubenswrapper[4870]: I1014 07:02:06.434462 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:06 crc kubenswrapper[4870]: I1014 07:02:06.434486 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:06 crc kubenswrapper[4870]: I1014 07:02:06.434505 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:06Z","lastTransitionTime":"2025-10-14T07:02:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:06 crc kubenswrapper[4870]: I1014 07:02:06.538292 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:06 crc kubenswrapper[4870]: I1014 07:02:06.538345 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:06 crc kubenswrapper[4870]: I1014 07:02:06.538354 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:06 crc kubenswrapper[4870]: I1014 07:02:06.538369 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:06 crc kubenswrapper[4870]: I1014 07:02:06.538381 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:06Z","lastTransitionTime":"2025-10-14T07:02:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:06 crc kubenswrapper[4870]: I1014 07:02:06.640813 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:06 crc kubenswrapper[4870]: I1014 07:02:06.640852 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:06 crc kubenswrapper[4870]: I1014 07:02:06.640866 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:06 crc kubenswrapper[4870]: I1014 07:02:06.640882 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:06 crc kubenswrapper[4870]: I1014 07:02:06.640892 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:06Z","lastTransitionTime":"2025-10-14T07:02:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:06 crc kubenswrapper[4870]: I1014 07:02:06.743902 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:06 crc kubenswrapper[4870]: I1014 07:02:06.743952 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:06 crc kubenswrapper[4870]: I1014 07:02:06.743963 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:06 crc kubenswrapper[4870]: I1014 07:02:06.743984 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:06 crc kubenswrapper[4870]: I1014 07:02:06.743997 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:06Z","lastTransitionTime":"2025-10-14T07:02:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:06 crc kubenswrapper[4870]: I1014 07:02:06.846421 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:06 crc kubenswrapper[4870]: I1014 07:02:06.846486 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:06 crc kubenswrapper[4870]: I1014 07:02:06.846499 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:06 crc kubenswrapper[4870]: I1014 07:02:06.846516 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:06 crc kubenswrapper[4870]: I1014 07:02:06.846528 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:06Z","lastTransitionTime":"2025-10-14T07:02:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:06 crc kubenswrapper[4870]: I1014 07:02:06.949424 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:06 crc kubenswrapper[4870]: I1014 07:02:06.949486 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:06 crc kubenswrapper[4870]: I1014 07:02:06.949551 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:06 crc kubenswrapper[4870]: I1014 07:02:06.949575 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:06 crc kubenswrapper[4870]: I1014 07:02:06.949587 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:06Z","lastTransitionTime":"2025-10-14T07:02:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:07 crc kubenswrapper[4870]: I1014 07:02:07.033668 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zc5j6" Oct 14 07:02:07 crc kubenswrapper[4870]: E1014 07:02:07.033867 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zc5j6" podUID="7d7a4f41-388a-4320-b896-43c17ff13da6" Oct 14 07:02:07 crc kubenswrapper[4870]: I1014 07:02:07.052311 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:07 crc kubenswrapper[4870]: I1014 07:02:07.052354 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:07 crc kubenswrapper[4870]: I1014 07:02:07.052365 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:07 crc kubenswrapper[4870]: I1014 07:02:07.052382 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:07 crc kubenswrapper[4870]: I1014 07:02:07.052395 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:07Z","lastTransitionTime":"2025-10-14T07:02:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:07 crc kubenswrapper[4870]: I1014 07:02:07.155211 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:07 crc kubenswrapper[4870]: I1014 07:02:07.155272 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:07 crc kubenswrapper[4870]: I1014 07:02:07.155289 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:07 crc kubenswrapper[4870]: I1014 07:02:07.155312 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:07 crc kubenswrapper[4870]: I1014 07:02:07.155332 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:07Z","lastTransitionTime":"2025-10-14T07:02:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:07 crc kubenswrapper[4870]: I1014 07:02:07.257854 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:07 crc kubenswrapper[4870]: I1014 07:02:07.257915 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:07 crc kubenswrapper[4870]: I1014 07:02:07.257933 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:07 crc kubenswrapper[4870]: I1014 07:02:07.257958 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:07 crc kubenswrapper[4870]: I1014 07:02:07.257975 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:07Z","lastTransitionTime":"2025-10-14T07:02:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:07 crc kubenswrapper[4870]: I1014 07:02:07.360227 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:07 crc kubenswrapper[4870]: I1014 07:02:07.360297 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:07 crc kubenswrapper[4870]: I1014 07:02:07.360315 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:07 crc kubenswrapper[4870]: I1014 07:02:07.360340 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:07 crc kubenswrapper[4870]: I1014 07:02:07.360358 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:07Z","lastTransitionTime":"2025-10-14T07:02:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:07 crc kubenswrapper[4870]: I1014 07:02:07.462850 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:07 crc kubenswrapper[4870]: I1014 07:02:07.462907 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:07 crc kubenswrapper[4870]: I1014 07:02:07.462919 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:07 crc kubenswrapper[4870]: I1014 07:02:07.462939 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:07 crc kubenswrapper[4870]: I1014 07:02:07.462952 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:07Z","lastTransitionTime":"2025-10-14T07:02:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:07 crc kubenswrapper[4870]: I1014 07:02:07.565734 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:07 crc kubenswrapper[4870]: I1014 07:02:07.565804 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:07 crc kubenswrapper[4870]: I1014 07:02:07.565832 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:07 crc kubenswrapper[4870]: I1014 07:02:07.565861 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:07 crc kubenswrapper[4870]: I1014 07:02:07.565884 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:07Z","lastTransitionTime":"2025-10-14T07:02:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:07 crc kubenswrapper[4870]: I1014 07:02:07.668834 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:07 crc kubenswrapper[4870]: I1014 07:02:07.668880 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:07 crc kubenswrapper[4870]: I1014 07:02:07.668893 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:07 crc kubenswrapper[4870]: I1014 07:02:07.668912 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:07 crc kubenswrapper[4870]: I1014 07:02:07.668924 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:07Z","lastTransitionTime":"2025-10-14T07:02:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:07 crc kubenswrapper[4870]: I1014 07:02:07.772291 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:07 crc kubenswrapper[4870]: I1014 07:02:07.772378 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:07 crc kubenswrapper[4870]: I1014 07:02:07.772411 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:07 crc kubenswrapper[4870]: I1014 07:02:07.772478 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:07 crc kubenswrapper[4870]: I1014 07:02:07.772517 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:07Z","lastTransitionTime":"2025-10-14T07:02:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:07 crc kubenswrapper[4870]: I1014 07:02:07.875510 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:07 crc kubenswrapper[4870]: I1014 07:02:07.875574 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:07 crc kubenswrapper[4870]: I1014 07:02:07.875592 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:07 crc kubenswrapper[4870]: I1014 07:02:07.875616 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:07 crc kubenswrapper[4870]: I1014 07:02:07.875634 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:07Z","lastTransitionTime":"2025-10-14T07:02:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:07 crc kubenswrapper[4870]: I1014 07:02:07.978709 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:07 crc kubenswrapper[4870]: I1014 07:02:07.978750 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:07 crc kubenswrapper[4870]: I1014 07:02:07.978759 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:07 crc kubenswrapper[4870]: I1014 07:02:07.978773 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:07 crc kubenswrapper[4870]: I1014 07:02:07.978782 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:07Z","lastTransitionTime":"2025-10-14T07:02:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:08 crc kubenswrapper[4870]: I1014 07:02:08.033544 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 07:02:08 crc kubenswrapper[4870]: I1014 07:02:08.033584 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 07:02:08 crc kubenswrapper[4870]: I1014 07:02:08.033610 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 07:02:08 crc kubenswrapper[4870]: E1014 07:02:08.033682 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 07:02:08 crc kubenswrapper[4870]: E1014 07:02:08.033737 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 07:02:08 crc kubenswrapper[4870]: E1014 07:02:08.033799 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 07:02:08 crc kubenswrapper[4870]: I1014 07:02:08.081384 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:08 crc kubenswrapper[4870]: I1014 07:02:08.081418 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:08 crc kubenswrapper[4870]: I1014 07:02:08.081457 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:08 crc kubenswrapper[4870]: I1014 07:02:08.081475 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:08 crc kubenswrapper[4870]: I1014 07:02:08.081488 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:08Z","lastTransitionTime":"2025-10-14T07:02:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:08 crc kubenswrapper[4870]: I1014 07:02:08.184757 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:08 crc kubenswrapper[4870]: I1014 07:02:08.184827 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:08 crc kubenswrapper[4870]: I1014 07:02:08.184841 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:08 crc kubenswrapper[4870]: I1014 07:02:08.184860 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:08 crc kubenswrapper[4870]: I1014 07:02:08.184873 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:08Z","lastTransitionTime":"2025-10-14T07:02:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:08 crc kubenswrapper[4870]: I1014 07:02:08.287531 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:08 crc kubenswrapper[4870]: I1014 07:02:08.287591 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:08 crc kubenswrapper[4870]: I1014 07:02:08.287609 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:08 crc kubenswrapper[4870]: I1014 07:02:08.287633 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:08 crc kubenswrapper[4870]: I1014 07:02:08.287650 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:08Z","lastTransitionTime":"2025-10-14T07:02:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:08 crc kubenswrapper[4870]: I1014 07:02:08.390774 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:08 crc kubenswrapper[4870]: I1014 07:02:08.391207 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:08 crc kubenswrapper[4870]: I1014 07:02:08.391329 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:08 crc kubenswrapper[4870]: I1014 07:02:08.391432 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:08 crc kubenswrapper[4870]: I1014 07:02:08.391543 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:08Z","lastTransitionTime":"2025-10-14T07:02:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:08 crc kubenswrapper[4870]: I1014 07:02:08.493867 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:08 crc kubenswrapper[4870]: I1014 07:02:08.493902 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:08 crc kubenswrapper[4870]: I1014 07:02:08.493912 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:08 crc kubenswrapper[4870]: I1014 07:02:08.493926 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:08 crc kubenswrapper[4870]: I1014 07:02:08.493937 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:08Z","lastTransitionTime":"2025-10-14T07:02:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:08 crc kubenswrapper[4870]: I1014 07:02:08.596817 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:08 crc kubenswrapper[4870]: I1014 07:02:08.597097 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:08 crc kubenswrapper[4870]: I1014 07:02:08.597123 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:08 crc kubenswrapper[4870]: I1014 07:02:08.597156 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:08 crc kubenswrapper[4870]: I1014 07:02:08.597180 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:08Z","lastTransitionTime":"2025-10-14T07:02:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:08 crc kubenswrapper[4870]: I1014 07:02:08.700199 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:08 crc kubenswrapper[4870]: I1014 07:02:08.700260 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:08 crc kubenswrapper[4870]: I1014 07:02:08.700272 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:08 crc kubenswrapper[4870]: I1014 07:02:08.700289 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:08 crc kubenswrapper[4870]: I1014 07:02:08.700300 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:08Z","lastTransitionTime":"2025-10-14T07:02:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:08 crc kubenswrapper[4870]: I1014 07:02:08.802688 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:08 crc kubenswrapper[4870]: I1014 07:02:08.802725 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:08 crc kubenswrapper[4870]: I1014 07:02:08.802737 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:08 crc kubenswrapper[4870]: I1014 07:02:08.802756 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:08 crc kubenswrapper[4870]: I1014 07:02:08.802766 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:08Z","lastTransitionTime":"2025-10-14T07:02:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:08 crc kubenswrapper[4870]: I1014 07:02:08.904863 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:08 crc kubenswrapper[4870]: I1014 07:02:08.904902 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:08 crc kubenswrapper[4870]: I1014 07:02:08.904911 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:08 crc kubenswrapper[4870]: I1014 07:02:08.904924 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:08 crc kubenswrapper[4870]: I1014 07:02:08.904934 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:08Z","lastTransitionTime":"2025-10-14T07:02:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:09 crc kubenswrapper[4870]: I1014 07:02:09.007542 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:09 crc kubenswrapper[4870]: I1014 07:02:09.007589 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:09 crc kubenswrapper[4870]: I1014 07:02:09.007604 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:09 crc kubenswrapper[4870]: I1014 07:02:09.007619 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:09 crc kubenswrapper[4870]: I1014 07:02:09.007631 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:09Z","lastTransitionTime":"2025-10-14T07:02:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:09 crc kubenswrapper[4870]: I1014 07:02:09.033648 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zc5j6" Oct 14 07:02:09 crc kubenswrapper[4870]: E1014 07:02:09.033788 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zc5j6" podUID="7d7a4f41-388a-4320-b896-43c17ff13da6" Oct 14 07:02:09 crc kubenswrapper[4870]: I1014 07:02:09.111742 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:09 crc kubenswrapper[4870]: I1014 07:02:09.111803 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:09 crc kubenswrapper[4870]: I1014 07:02:09.111814 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:09 crc kubenswrapper[4870]: I1014 07:02:09.111837 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:09 crc kubenswrapper[4870]: I1014 07:02:09.111849 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:09Z","lastTransitionTime":"2025-10-14T07:02:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:09 crc kubenswrapper[4870]: I1014 07:02:09.215346 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:09 crc kubenswrapper[4870]: I1014 07:02:09.215396 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:09 crc kubenswrapper[4870]: I1014 07:02:09.215408 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:09 crc kubenswrapper[4870]: I1014 07:02:09.215426 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:09 crc kubenswrapper[4870]: I1014 07:02:09.215473 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:09Z","lastTransitionTime":"2025-10-14T07:02:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:09 crc kubenswrapper[4870]: I1014 07:02:09.317129 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:09 crc kubenswrapper[4870]: I1014 07:02:09.317167 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:09 crc kubenswrapper[4870]: I1014 07:02:09.317187 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:09 crc kubenswrapper[4870]: I1014 07:02:09.317201 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:09 crc kubenswrapper[4870]: I1014 07:02:09.317214 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:09Z","lastTransitionTime":"2025-10-14T07:02:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:09 crc kubenswrapper[4870]: I1014 07:02:09.420235 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:09 crc kubenswrapper[4870]: I1014 07:02:09.420273 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:09 crc kubenswrapper[4870]: I1014 07:02:09.420284 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:09 crc kubenswrapper[4870]: I1014 07:02:09.420300 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:09 crc kubenswrapper[4870]: I1014 07:02:09.420313 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:09Z","lastTransitionTime":"2025-10-14T07:02:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:09 crc kubenswrapper[4870]: I1014 07:02:09.523170 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:09 crc kubenswrapper[4870]: I1014 07:02:09.523221 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:09 crc kubenswrapper[4870]: I1014 07:02:09.523233 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:09 crc kubenswrapper[4870]: I1014 07:02:09.523254 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:09 crc kubenswrapper[4870]: I1014 07:02:09.523268 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:09Z","lastTransitionTime":"2025-10-14T07:02:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:09 crc kubenswrapper[4870]: I1014 07:02:09.626190 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:09 crc kubenswrapper[4870]: I1014 07:02:09.626237 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:09 crc kubenswrapper[4870]: I1014 07:02:09.626251 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:09 crc kubenswrapper[4870]: I1014 07:02:09.626268 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:09 crc kubenswrapper[4870]: I1014 07:02:09.626280 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:09Z","lastTransitionTime":"2025-10-14T07:02:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:09 crc kubenswrapper[4870]: I1014 07:02:09.729123 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:09 crc kubenswrapper[4870]: I1014 07:02:09.729203 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:09 crc kubenswrapper[4870]: I1014 07:02:09.729219 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:09 crc kubenswrapper[4870]: I1014 07:02:09.729241 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:09 crc kubenswrapper[4870]: I1014 07:02:09.729257 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:09Z","lastTransitionTime":"2025-10-14T07:02:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:09 crc kubenswrapper[4870]: I1014 07:02:09.831890 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:09 crc kubenswrapper[4870]: I1014 07:02:09.831932 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:09 crc kubenswrapper[4870]: I1014 07:02:09.831947 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:09 crc kubenswrapper[4870]: I1014 07:02:09.831964 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:09 crc kubenswrapper[4870]: I1014 07:02:09.831976 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:09Z","lastTransitionTime":"2025-10-14T07:02:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:09 crc kubenswrapper[4870]: I1014 07:02:09.934844 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:09 crc kubenswrapper[4870]: I1014 07:02:09.934885 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:09 crc kubenswrapper[4870]: I1014 07:02:09.934897 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:09 crc kubenswrapper[4870]: I1014 07:02:09.934917 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:09 crc kubenswrapper[4870]: I1014 07:02:09.934932 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:09Z","lastTransitionTime":"2025-10-14T07:02:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:10 crc kubenswrapper[4870]: I1014 07:02:10.033130 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 07:02:10 crc kubenswrapper[4870]: I1014 07:02:10.033392 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 07:02:10 crc kubenswrapper[4870]: I1014 07:02:10.033130 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 07:02:10 crc kubenswrapper[4870]: E1014 07:02:10.033568 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 07:02:10 crc kubenswrapper[4870]: E1014 07:02:10.033609 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 07:02:10 crc kubenswrapper[4870]: E1014 07:02:10.033667 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 07:02:10 crc kubenswrapper[4870]: I1014 07:02:10.037722 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:10 crc kubenswrapper[4870]: I1014 07:02:10.037753 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:10 crc kubenswrapper[4870]: I1014 07:02:10.037780 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:10 crc kubenswrapper[4870]: I1014 07:02:10.037801 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:10 crc kubenswrapper[4870]: I1014 07:02:10.037815 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:10Z","lastTransitionTime":"2025-10-14T07:02:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:10 crc kubenswrapper[4870]: I1014 07:02:10.140212 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:10 crc kubenswrapper[4870]: I1014 07:02:10.140294 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:10 crc kubenswrapper[4870]: I1014 07:02:10.140311 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:10 crc kubenswrapper[4870]: I1014 07:02:10.140336 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:10 crc kubenswrapper[4870]: I1014 07:02:10.140350 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:10Z","lastTransitionTime":"2025-10-14T07:02:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:10 crc kubenswrapper[4870]: I1014 07:02:10.242950 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:10 crc kubenswrapper[4870]: I1014 07:02:10.243002 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:10 crc kubenswrapper[4870]: I1014 07:02:10.243011 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:10 crc kubenswrapper[4870]: I1014 07:02:10.243028 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:10 crc kubenswrapper[4870]: I1014 07:02:10.243040 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:10Z","lastTransitionTime":"2025-10-14T07:02:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:10 crc kubenswrapper[4870]: I1014 07:02:10.345458 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:10 crc kubenswrapper[4870]: I1014 07:02:10.345518 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:10 crc kubenswrapper[4870]: I1014 07:02:10.345530 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:10 crc kubenswrapper[4870]: I1014 07:02:10.345546 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:10 crc kubenswrapper[4870]: I1014 07:02:10.345558 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:10Z","lastTransitionTime":"2025-10-14T07:02:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:10 crc kubenswrapper[4870]: I1014 07:02:10.448499 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:10 crc kubenswrapper[4870]: I1014 07:02:10.448592 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:10 crc kubenswrapper[4870]: I1014 07:02:10.448610 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:10 crc kubenswrapper[4870]: I1014 07:02:10.448625 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:10 crc kubenswrapper[4870]: I1014 07:02:10.448641 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:10Z","lastTransitionTime":"2025-10-14T07:02:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:10 crc kubenswrapper[4870]: I1014 07:02:10.552003 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:10 crc kubenswrapper[4870]: I1014 07:02:10.552047 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:10 crc kubenswrapper[4870]: I1014 07:02:10.552055 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:10 crc kubenswrapper[4870]: I1014 07:02:10.552069 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:10 crc kubenswrapper[4870]: I1014 07:02:10.552077 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:10Z","lastTransitionTime":"2025-10-14T07:02:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:10 crc kubenswrapper[4870]: I1014 07:02:10.654501 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:10 crc kubenswrapper[4870]: I1014 07:02:10.654552 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:10 crc kubenswrapper[4870]: I1014 07:02:10.654562 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:10 crc kubenswrapper[4870]: I1014 07:02:10.654575 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:10 crc kubenswrapper[4870]: I1014 07:02:10.654584 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:10Z","lastTransitionTime":"2025-10-14T07:02:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:10 crc kubenswrapper[4870]: I1014 07:02:10.757019 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:10 crc kubenswrapper[4870]: I1014 07:02:10.757078 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:10 crc kubenswrapper[4870]: I1014 07:02:10.757090 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:10 crc kubenswrapper[4870]: I1014 07:02:10.757107 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:10 crc kubenswrapper[4870]: I1014 07:02:10.757117 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:10Z","lastTransitionTime":"2025-10-14T07:02:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:10 crc kubenswrapper[4870]: I1014 07:02:10.859513 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:10 crc kubenswrapper[4870]: I1014 07:02:10.859556 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:10 crc kubenswrapper[4870]: I1014 07:02:10.859568 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:10 crc kubenswrapper[4870]: I1014 07:02:10.859584 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:10 crc kubenswrapper[4870]: I1014 07:02:10.859595 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:10Z","lastTransitionTime":"2025-10-14T07:02:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:10 crc kubenswrapper[4870]: I1014 07:02:10.961953 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:10 crc kubenswrapper[4870]: I1014 07:02:10.962022 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:10 crc kubenswrapper[4870]: I1014 07:02:10.962033 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:10 crc kubenswrapper[4870]: I1014 07:02:10.962053 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:10 crc kubenswrapper[4870]: I1014 07:02:10.962065 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:10Z","lastTransitionTime":"2025-10-14T07:02:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:11 crc kubenswrapper[4870]: I1014 07:02:11.033089 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zc5j6" Oct 14 07:02:11 crc kubenswrapper[4870]: E1014 07:02:11.033203 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zc5j6" podUID="7d7a4f41-388a-4320-b896-43c17ff13da6" Oct 14 07:02:11 crc kubenswrapper[4870]: I1014 07:02:11.064132 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:11 crc kubenswrapper[4870]: I1014 07:02:11.064190 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:11 crc kubenswrapper[4870]: I1014 07:02:11.064200 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:11 crc kubenswrapper[4870]: I1014 07:02:11.064212 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:11 crc kubenswrapper[4870]: I1014 07:02:11.064222 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:11Z","lastTransitionTime":"2025-10-14T07:02:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:11 crc kubenswrapper[4870]: I1014 07:02:11.166551 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:11 crc kubenswrapper[4870]: I1014 07:02:11.166600 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:11 crc kubenswrapper[4870]: I1014 07:02:11.166613 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:11 crc kubenswrapper[4870]: I1014 07:02:11.166631 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:11 crc kubenswrapper[4870]: I1014 07:02:11.166643 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:11Z","lastTransitionTime":"2025-10-14T07:02:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:11 crc kubenswrapper[4870]: I1014 07:02:11.269786 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:11 crc kubenswrapper[4870]: I1014 07:02:11.269839 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:11 crc kubenswrapper[4870]: I1014 07:02:11.269851 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:11 crc kubenswrapper[4870]: I1014 07:02:11.269869 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:11 crc kubenswrapper[4870]: I1014 07:02:11.269879 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:11Z","lastTransitionTime":"2025-10-14T07:02:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:11 crc kubenswrapper[4870]: I1014 07:02:11.371832 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:11 crc kubenswrapper[4870]: I1014 07:02:11.371868 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:11 crc kubenswrapper[4870]: I1014 07:02:11.371906 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:11 crc kubenswrapper[4870]: I1014 07:02:11.371922 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:11 crc kubenswrapper[4870]: I1014 07:02:11.371931 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:11Z","lastTransitionTime":"2025-10-14T07:02:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:11 crc kubenswrapper[4870]: I1014 07:02:11.474540 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:11 crc kubenswrapper[4870]: I1014 07:02:11.474586 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:11 crc kubenswrapper[4870]: I1014 07:02:11.474599 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:11 crc kubenswrapper[4870]: I1014 07:02:11.474618 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:11 crc kubenswrapper[4870]: I1014 07:02:11.474630 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:11Z","lastTransitionTime":"2025-10-14T07:02:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:11 crc kubenswrapper[4870]: I1014 07:02:11.576947 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:11 crc kubenswrapper[4870]: I1014 07:02:11.577021 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:11 crc kubenswrapper[4870]: I1014 07:02:11.577033 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:11 crc kubenswrapper[4870]: I1014 07:02:11.577048 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:11 crc kubenswrapper[4870]: I1014 07:02:11.577060 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:11Z","lastTransitionTime":"2025-10-14T07:02:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:11 crc kubenswrapper[4870]: I1014 07:02:11.680210 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:11 crc kubenswrapper[4870]: I1014 07:02:11.680239 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:11 crc kubenswrapper[4870]: I1014 07:02:11.680248 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:11 crc kubenswrapper[4870]: I1014 07:02:11.680263 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:11 crc kubenswrapper[4870]: I1014 07:02:11.680273 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:11Z","lastTransitionTime":"2025-10-14T07:02:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:11 crc kubenswrapper[4870]: I1014 07:02:11.782875 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:11 crc kubenswrapper[4870]: I1014 07:02:11.782903 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:11 crc kubenswrapper[4870]: I1014 07:02:11.782911 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:11 crc kubenswrapper[4870]: I1014 07:02:11.782925 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:11 crc kubenswrapper[4870]: I1014 07:02:11.782934 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:11Z","lastTransitionTime":"2025-10-14T07:02:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:11 crc kubenswrapper[4870]: I1014 07:02:11.884731 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:11 crc kubenswrapper[4870]: I1014 07:02:11.884768 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:11 crc kubenswrapper[4870]: I1014 07:02:11.884778 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:11 crc kubenswrapper[4870]: I1014 07:02:11.884792 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:11 crc kubenswrapper[4870]: I1014 07:02:11.884801 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:11Z","lastTransitionTime":"2025-10-14T07:02:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:11 crc kubenswrapper[4870]: I1014 07:02:11.986697 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:11 crc kubenswrapper[4870]: I1014 07:02:11.986738 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:11 crc kubenswrapper[4870]: I1014 07:02:11.986750 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:11 crc kubenswrapper[4870]: I1014 07:02:11.986769 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:11 crc kubenswrapper[4870]: I1014 07:02:11.986782 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:11Z","lastTransitionTime":"2025-10-14T07:02:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:12 crc kubenswrapper[4870]: I1014 07:02:12.032979 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 07:02:12 crc kubenswrapper[4870]: I1014 07:02:12.033025 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 07:02:12 crc kubenswrapper[4870]: E1014 07:02:12.033105 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 07:02:12 crc kubenswrapper[4870]: I1014 07:02:12.033029 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 07:02:12 crc kubenswrapper[4870]: E1014 07:02:12.033226 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 07:02:12 crc kubenswrapper[4870]: E1014 07:02:12.033311 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 07:02:12 crc kubenswrapper[4870]: I1014 07:02:12.089035 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:12 crc kubenswrapper[4870]: I1014 07:02:12.089071 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:12 crc kubenswrapper[4870]: I1014 07:02:12.089082 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:12 crc kubenswrapper[4870]: I1014 07:02:12.089098 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:12 crc kubenswrapper[4870]: I1014 07:02:12.089108 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:12Z","lastTransitionTime":"2025-10-14T07:02:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:12 crc kubenswrapper[4870]: I1014 07:02:12.192492 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:12 crc kubenswrapper[4870]: I1014 07:02:12.192532 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:12 crc kubenswrapper[4870]: I1014 07:02:12.192560 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:12 crc kubenswrapper[4870]: I1014 07:02:12.192575 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:12 crc kubenswrapper[4870]: I1014 07:02:12.192585 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:12Z","lastTransitionTime":"2025-10-14T07:02:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:12 crc kubenswrapper[4870]: I1014 07:02:12.294865 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:12 crc kubenswrapper[4870]: I1014 07:02:12.294926 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:12 crc kubenswrapper[4870]: I1014 07:02:12.294937 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:12 crc kubenswrapper[4870]: I1014 07:02:12.294955 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:12 crc kubenswrapper[4870]: I1014 07:02:12.294966 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:12Z","lastTransitionTime":"2025-10-14T07:02:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:12 crc kubenswrapper[4870]: I1014 07:02:12.397350 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:12 crc kubenswrapper[4870]: I1014 07:02:12.397388 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:12 crc kubenswrapper[4870]: I1014 07:02:12.397399 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:12 crc kubenswrapper[4870]: I1014 07:02:12.397414 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:12 crc kubenswrapper[4870]: I1014 07:02:12.397426 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:12Z","lastTransitionTime":"2025-10-14T07:02:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:12 crc kubenswrapper[4870]: I1014 07:02:12.500518 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:12 crc kubenswrapper[4870]: I1014 07:02:12.500572 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:12 crc kubenswrapper[4870]: I1014 07:02:12.500584 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:12 crc kubenswrapper[4870]: I1014 07:02:12.500602 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:12 crc kubenswrapper[4870]: I1014 07:02:12.500620 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:12Z","lastTransitionTime":"2025-10-14T07:02:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:12 crc kubenswrapper[4870]: I1014 07:02:12.602890 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:12 crc kubenswrapper[4870]: I1014 07:02:12.602919 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:12 crc kubenswrapper[4870]: I1014 07:02:12.602929 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:12 crc kubenswrapper[4870]: I1014 07:02:12.602946 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:12 crc kubenswrapper[4870]: I1014 07:02:12.602956 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:12Z","lastTransitionTime":"2025-10-14T07:02:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:12 crc kubenswrapper[4870]: I1014 07:02:12.644175 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7d7a4f41-388a-4320-b896-43c17ff13da6-metrics-certs\") pod \"network-metrics-daemon-zc5j6\" (UID: \"7d7a4f41-388a-4320-b896-43c17ff13da6\") " pod="openshift-multus/network-metrics-daemon-zc5j6" Oct 14 07:02:12 crc kubenswrapper[4870]: E1014 07:02:12.644378 4870 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 14 07:02:12 crc kubenswrapper[4870]: E1014 07:02:12.644490 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7d7a4f41-388a-4320-b896-43c17ff13da6-metrics-certs podName:7d7a4f41-388a-4320-b896-43c17ff13da6 nodeName:}" failed. No retries permitted until 2025-10-14 07:02:44.644467822 +0000 UTC m=+100.341828263 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/7d7a4f41-388a-4320-b896-43c17ff13da6-metrics-certs") pod "network-metrics-daemon-zc5j6" (UID: "7d7a4f41-388a-4320-b896-43c17ff13da6") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 14 07:02:12 crc kubenswrapper[4870]: I1014 07:02:12.705089 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:12 crc kubenswrapper[4870]: I1014 07:02:12.705132 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:12 crc kubenswrapper[4870]: I1014 07:02:12.705143 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:12 crc kubenswrapper[4870]: I1014 07:02:12.705174 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:12 crc kubenswrapper[4870]: I1014 07:02:12.705187 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:12Z","lastTransitionTime":"2025-10-14T07:02:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:12 crc kubenswrapper[4870]: I1014 07:02:12.807820 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:12 crc kubenswrapper[4870]: I1014 07:02:12.807882 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:12 crc kubenswrapper[4870]: I1014 07:02:12.807894 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:12 crc kubenswrapper[4870]: I1014 07:02:12.807910 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:12 crc kubenswrapper[4870]: I1014 07:02:12.807926 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:12Z","lastTransitionTime":"2025-10-14T07:02:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:12 crc kubenswrapper[4870]: I1014 07:02:12.910665 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:12 crc kubenswrapper[4870]: I1014 07:02:12.910709 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:12 crc kubenswrapper[4870]: I1014 07:02:12.910743 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:12 crc kubenswrapper[4870]: I1014 07:02:12.910762 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:12 crc kubenswrapper[4870]: I1014 07:02:12.910775 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:12Z","lastTransitionTime":"2025-10-14T07:02:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:13 crc kubenswrapper[4870]: I1014 07:02:13.019714 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:13 crc kubenswrapper[4870]: I1014 07:02:13.019786 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:13 crc kubenswrapper[4870]: I1014 07:02:13.019795 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:13 crc kubenswrapper[4870]: I1014 07:02:13.019810 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:13 crc kubenswrapper[4870]: I1014 07:02:13.019842 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:13Z","lastTransitionTime":"2025-10-14T07:02:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:13 crc kubenswrapper[4870]: I1014 07:02:13.033145 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zc5j6" Oct 14 07:02:13 crc kubenswrapper[4870]: E1014 07:02:13.033384 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zc5j6" podUID="7d7a4f41-388a-4320-b896-43c17ff13da6" Oct 14 07:02:13 crc kubenswrapper[4870]: I1014 07:02:13.121517 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:13 crc kubenswrapper[4870]: I1014 07:02:13.121556 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:13 crc kubenswrapper[4870]: I1014 07:02:13.121570 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:13 crc kubenswrapper[4870]: I1014 07:02:13.121586 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:13 crc kubenswrapper[4870]: I1014 07:02:13.121597 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:13Z","lastTransitionTime":"2025-10-14T07:02:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:13 crc kubenswrapper[4870]: I1014 07:02:13.223959 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:13 crc kubenswrapper[4870]: I1014 07:02:13.224022 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:13 crc kubenswrapper[4870]: I1014 07:02:13.224035 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:13 crc kubenswrapper[4870]: I1014 07:02:13.224051 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:13 crc kubenswrapper[4870]: I1014 07:02:13.224086 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:13Z","lastTransitionTime":"2025-10-14T07:02:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:13 crc kubenswrapper[4870]: I1014 07:02:13.326230 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:13 crc kubenswrapper[4870]: I1014 07:02:13.326271 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:13 crc kubenswrapper[4870]: I1014 07:02:13.326305 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:13 crc kubenswrapper[4870]: I1014 07:02:13.326323 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:13 crc kubenswrapper[4870]: I1014 07:02:13.326336 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:13Z","lastTransitionTime":"2025-10-14T07:02:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:13 crc kubenswrapper[4870]: I1014 07:02:13.428894 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:13 crc kubenswrapper[4870]: I1014 07:02:13.428937 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:13 crc kubenswrapper[4870]: I1014 07:02:13.428950 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:13 crc kubenswrapper[4870]: I1014 07:02:13.428969 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:13 crc kubenswrapper[4870]: I1014 07:02:13.428981 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:13Z","lastTransitionTime":"2025-10-14T07:02:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:13 crc kubenswrapper[4870]: I1014 07:02:13.457079 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:13 crc kubenswrapper[4870]: I1014 07:02:13.457128 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:13 crc kubenswrapper[4870]: I1014 07:02:13.457140 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:13 crc kubenswrapper[4870]: I1014 07:02:13.457156 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:13 crc kubenswrapper[4870]: I1014 07:02:13.457168 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:13Z","lastTransitionTime":"2025-10-14T07:02:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:13 crc kubenswrapper[4870]: E1014 07:02:13.471833 4870 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:02:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:02:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:02:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:02:13Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:02:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:02:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:02:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:02:13Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"56448e6b-6b3a-4bb9-91f8-2f0a9b44e9aa\\\",\\\"systemUUID\\\":\\\"0adb347d-b90d-4e9f-9962-f1bd97980aed\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:13Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:13 crc kubenswrapper[4870]: I1014 07:02:13.478813 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:13 crc kubenswrapper[4870]: I1014 07:02:13.478862 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:13 crc kubenswrapper[4870]: I1014 07:02:13.478873 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:13 crc kubenswrapper[4870]: I1014 07:02:13.478891 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:13 crc kubenswrapper[4870]: I1014 07:02:13.478903 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:13Z","lastTransitionTime":"2025-10-14T07:02:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:13 crc kubenswrapper[4870]: E1014 07:02:13.494043 4870 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:02:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:02:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:02:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:02:13Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:02:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:02:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:02:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:02:13Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"56448e6b-6b3a-4bb9-91f8-2f0a9b44e9aa\\\",\\\"systemUUID\\\":\\\"0adb347d-b90d-4e9f-9962-f1bd97980aed\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:13Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:13 crc kubenswrapper[4870]: I1014 07:02:13.498428 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:13 crc kubenswrapper[4870]: I1014 07:02:13.498497 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:13 crc kubenswrapper[4870]: I1014 07:02:13.498509 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:13 crc kubenswrapper[4870]: I1014 07:02:13.498528 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:13 crc kubenswrapper[4870]: I1014 07:02:13.498537 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:13Z","lastTransitionTime":"2025-10-14T07:02:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:13 crc kubenswrapper[4870]: E1014 07:02:13.513625 4870 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:02:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:02:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:02:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:02:13Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:02:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:02:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:02:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:02:13Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"56448e6b-6b3a-4bb9-91f8-2f0a9b44e9aa\\\",\\\"systemUUID\\\":\\\"0adb347d-b90d-4e9f-9962-f1bd97980aed\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:13Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:13 crc kubenswrapper[4870]: I1014 07:02:13.518051 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:13 crc kubenswrapper[4870]: I1014 07:02:13.518098 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:13 crc kubenswrapper[4870]: I1014 07:02:13.518108 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:13 crc kubenswrapper[4870]: I1014 07:02:13.518130 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:13 crc kubenswrapper[4870]: I1014 07:02:13.518145 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:13Z","lastTransitionTime":"2025-10-14T07:02:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:13 crc kubenswrapper[4870]: E1014 07:02:13.530703 4870 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:02:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:02:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:02:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:02:13Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:02:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:02:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:02:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:02:13Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"56448e6b-6b3a-4bb9-91f8-2f0a9b44e9aa\\\",\\\"systemUUID\\\":\\\"0adb347d-b90d-4e9f-9962-f1bd97980aed\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:13Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:13 crc kubenswrapper[4870]: I1014 07:02:13.535580 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:13 crc kubenswrapper[4870]: I1014 07:02:13.535643 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:13 crc kubenswrapper[4870]: I1014 07:02:13.535660 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:13 crc kubenswrapper[4870]: I1014 07:02:13.535686 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:13 crc kubenswrapper[4870]: I1014 07:02:13.535707 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:13Z","lastTransitionTime":"2025-10-14T07:02:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:13 crc kubenswrapper[4870]: E1014 07:02:13.551296 4870 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:02:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:02:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:02:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:02:13Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:02:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:02:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:02:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:02:13Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"56448e6b-6b3a-4bb9-91f8-2f0a9b44e9aa\\\",\\\"systemUUID\\\":\\\"0adb347d-b90d-4e9f-9962-f1bd97980aed\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:13Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:13 crc kubenswrapper[4870]: E1014 07:02:13.551555 4870 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 14 07:02:13 crc kubenswrapper[4870]: I1014 07:02:13.553246 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:13 crc kubenswrapper[4870]: I1014 07:02:13.553299 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:13 crc kubenswrapper[4870]: I1014 07:02:13.553313 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:13 crc kubenswrapper[4870]: I1014 07:02:13.553335 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:13 crc kubenswrapper[4870]: I1014 07:02:13.553352 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:13Z","lastTransitionTime":"2025-10-14T07:02:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:13 crc kubenswrapper[4870]: I1014 07:02:13.656034 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:13 crc kubenswrapper[4870]: I1014 07:02:13.656115 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:13 crc kubenswrapper[4870]: I1014 07:02:13.656139 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:13 crc kubenswrapper[4870]: I1014 07:02:13.656172 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:13 crc kubenswrapper[4870]: I1014 07:02:13.656201 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:13Z","lastTransitionTime":"2025-10-14T07:02:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:13 crc kubenswrapper[4870]: I1014 07:02:13.758706 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:13 crc kubenswrapper[4870]: I1014 07:02:13.758766 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:13 crc kubenswrapper[4870]: I1014 07:02:13.758776 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:13 crc kubenswrapper[4870]: I1014 07:02:13.758796 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:13 crc kubenswrapper[4870]: I1014 07:02:13.758826 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:13Z","lastTransitionTime":"2025-10-14T07:02:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:13 crc kubenswrapper[4870]: I1014 07:02:13.861113 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:13 crc kubenswrapper[4870]: I1014 07:02:13.861155 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:13 crc kubenswrapper[4870]: I1014 07:02:13.861166 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:13 crc kubenswrapper[4870]: I1014 07:02:13.861183 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:13 crc kubenswrapper[4870]: I1014 07:02:13.861195 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:13Z","lastTransitionTime":"2025-10-14T07:02:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:13 crc kubenswrapper[4870]: I1014 07:02:13.964755 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:13 crc kubenswrapper[4870]: I1014 07:02:13.964792 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:13 crc kubenswrapper[4870]: I1014 07:02:13.964803 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:13 crc kubenswrapper[4870]: I1014 07:02:13.964818 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:13 crc kubenswrapper[4870]: I1014 07:02:13.964829 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:13Z","lastTransitionTime":"2025-10-14T07:02:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:14 crc kubenswrapper[4870]: I1014 07:02:14.033393 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 07:02:14 crc kubenswrapper[4870]: E1014 07:02:14.033545 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 07:02:14 crc kubenswrapper[4870]: I1014 07:02:14.033615 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 07:02:14 crc kubenswrapper[4870]: I1014 07:02:14.033670 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 07:02:14 crc kubenswrapper[4870]: E1014 07:02:14.033949 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 07:02:14 crc kubenswrapper[4870]: E1014 07:02:14.034105 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 07:02:14 crc kubenswrapper[4870]: I1014 07:02:14.034538 4870 scope.go:117] "RemoveContainer" containerID="3fc9bfcb356b506d6c439adc22e4caaed4fad1c5a085744d44ebd50cf8898777" Oct 14 07:02:14 crc kubenswrapper[4870]: E1014 07:02:14.034902 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-v7br2_openshift-ovn-kubernetes(8dcaef0e-42e3-424b-bbe5-4a023efe912d)\"" pod="openshift-ovn-kubernetes/ovnkube-node-v7br2" podUID="8dcaef0e-42e3-424b-bbe5-4a023efe912d" Oct 14 07:02:14 crc kubenswrapper[4870]: I1014 07:02:14.067141 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:14 crc kubenswrapper[4870]: I1014 07:02:14.067286 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:14 crc kubenswrapper[4870]: I1014 07:02:14.067353 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:14 crc kubenswrapper[4870]: I1014 07:02:14.067417 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:14 crc kubenswrapper[4870]: I1014 07:02:14.067540 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:14Z","lastTransitionTime":"2025-10-14T07:02:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:14 crc kubenswrapper[4870]: I1014 07:02:14.169861 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:14 crc kubenswrapper[4870]: I1014 07:02:14.170142 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:14 crc kubenswrapper[4870]: I1014 07:02:14.170215 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:14 crc kubenswrapper[4870]: I1014 07:02:14.170333 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:14 crc kubenswrapper[4870]: I1014 07:02:14.170427 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:14Z","lastTransitionTime":"2025-10-14T07:02:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:14 crc kubenswrapper[4870]: I1014 07:02:14.273260 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:14 crc kubenswrapper[4870]: I1014 07:02:14.273629 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:14 crc kubenswrapper[4870]: I1014 07:02:14.273702 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:14 crc kubenswrapper[4870]: I1014 07:02:14.273784 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:14 crc kubenswrapper[4870]: I1014 07:02:14.273854 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:14Z","lastTransitionTime":"2025-10-14T07:02:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:14 crc kubenswrapper[4870]: I1014 07:02:14.376403 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:14 crc kubenswrapper[4870]: I1014 07:02:14.376467 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:14 crc kubenswrapper[4870]: I1014 07:02:14.376478 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:14 crc kubenswrapper[4870]: I1014 07:02:14.376495 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:14 crc kubenswrapper[4870]: I1014 07:02:14.376506 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:14Z","lastTransitionTime":"2025-10-14T07:02:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:14 crc kubenswrapper[4870]: I1014 07:02:14.479576 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:14 crc kubenswrapper[4870]: I1014 07:02:14.480129 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:14 crc kubenswrapper[4870]: I1014 07:02:14.480206 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:14 crc kubenswrapper[4870]: I1014 07:02:14.480304 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:14 crc kubenswrapper[4870]: I1014 07:02:14.480372 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:14Z","lastTransitionTime":"2025-10-14T07:02:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:14 crc kubenswrapper[4870]: I1014 07:02:14.583354 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:14 crc kubenswrapper[4870]: I1014 07:02:14.583415 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:14 crc kubenswrapper[4870]: I1014 07:02:14.583427 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:14 crc kubenswrapper[4870]: I1014 07:02:14.583471 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:14 crc kubenswrapper[4870]: I1014 07:02:14.583483 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:14Z","lastTransitionTime":"2025-10-14T07:02:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:14 crc kubenswrapper[4870]: I1014 07:02:14.602028 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-fdxjh_8a3a430a-b540-4a8c-adad-b2893dbf9898/kube-multus/0.log" Oct 14 07:02:14 crc kubenswrapper[4870]: I1014 07:02:14.602105 4870 generic.go:334] "Generic (PLEG): container finished" podID="8a3a430a-b540-4a8c-adad-b2893dbf9898" containerID="93e41d5e927eed7ec62eb9c3607cc804cd8999556663363fd123b2a784b00ff1" exitCode=1 Oct 14 07:02:14 crc kubenswrapper[4870]: I1014 07:02:14.602150 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-fdxjh" event={"ID":"8a3a430a-b540-4a8c-adad-b2893dbf9898","Type":"ContainerDied","Data":"93e41d5e927eed7ec62eb9c3607cc804cd8999556663363fd123b2a784b00ff1"} Oct 14 07:02:14 crc kubenswrapper[4870]: I1014 07:02:14.603251 4870 scope.go:117] "RemoveContainer" containerID="93e41d5e927eed7ec62eb9c3607cc804cd8999556663363fd123b2a784b00ff1" Oct 14 07:02:14 crc kubenswrapper[4870]: I1014 07:02:14.621375 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:14Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:14 crc kubenswrapper[4870]: I1014 07:02:14.637807 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3eb0b397c05ffd6650607f1be4a75683e0342d3333fc096bf1cd63c7688834f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:14Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:14 crc kubenswrapper[4870]: I1014 07:02:14.651387 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"874a5e45-dffd-4d17-b609-b7d3ed2eab07\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d98d044836ab44a405196915ef7bf572b75113f2c81ec7e64effa63b47ad3543\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wj4ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8004752575df4a1c44efbcafe64799f09126d0f97285d92388c76a0b90e022f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wj4ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7tvc7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:14Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:14 crc kubenswrapper[4870]: I1014 07:02:14.668179 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zc5j6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d7a4f41-388a-4320-b896-43c17ff13da6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6p2wh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6p2wh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:40Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zc5j6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:14Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:14 crc kubenswrapper[4870]: I1014 07:02:14.686627 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:14 crc kubenswrapper[4870]: I1014 07:02:14.686688 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:14 crc kubenswrapper[4870]: I1014 07:02:14.686705 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:14 crc kubenswrapper[4870]: I1014 07:02:14.686732 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:14 crc kubenswrapper[4870]: I1014 07:02:14.686752 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:14Z","lastTransitionTime":"2025-10-14T07:02:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:14 crc kubenswrapper[4870]: I1014 07:02:14.687428 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77a0d93a-5391-4112-b502-c474f45de209\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7815af9a584f6db6660073da289f2e8c987a04ef9ddfac155bdd9a15b014fcfa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b329b570622155a55f70501b2260e278106a68ebef497914231b4ec860e0125\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba942a0e6fe5ab93a187e0b7b4bd7c440c2e7827d9b5323fc50f90b562075392\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b9018b078c5c8276243c1696860f0aa73e2d8dd917913ae9d03b30dc9b5113a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://978e8db184ba9e430232c6bb2ee530beb92567fdc71ba3c35b0bedca3492d845\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"file observer\\\\nW1014 07:01:25.870672 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1014 07:01:25.870795 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1014 07:01:25.871521 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-756714914/tls.crt::/tmp/serving-cert-756714914/tls.key\\\\\\\"\\\\nI1014 07:01:26.460037 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1014 07:01:26.464876 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1014 07:01:26.464902 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1014 07:01:26.464925 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1014 07:01:26.464931 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1014 07:01:26.481374 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1014 07:01:26.481400 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:01:26.481405 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:01:26.481410 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1014 07:01:26.481414 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1014 07:01:26.481417 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1014 07:01:26.481420 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1014 07:01:26.481610 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1014 07:01:26.482872 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bd0044a247bdc5ca4cab7294103af3270b76cde61f3109b85e83ed050ca53b6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82d3c2214e0be679d2827594b159ace362151318689078bc49d3158f42899636\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82d3c2214e0be679d2827594b159ace362151318689078bc49d3158f42899636\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:05Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:14Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:14 crc kubenswrapper[4870]: I1014 07:02:14.701150 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:14Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:14 crc kubenswrapper[4870]: I1014 07:02:14.715091 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e26fe49ec2045d20a0cb2dd17e96e79b6eb43c641eb184c688cc472ca5db39d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aea115f7ee71c700b218fe15c928af7b55e6babe70e65d67f125d87be1fd2e2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:14Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:14 crc kubenswrapper[4870]: I1014 07:02:14.736997 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-v7br2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dcaef0e-42e3-424b-bbe5-4a023efe912d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71dec55e859b11a081dbb90d7d6cf6b4846cf591382303d8329003ff50c71f2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e72cff0723e2ce520340244b4e597d8016c58e1b071c4b087f2f6cd718741594\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc9d4db944f837289d1923176d60cfc4a566a282c2d2e3a703e220df884703c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d16ed1c65925549dcb69ab02c3013e25e3f123a83179f1af3a8e08bf91a59896\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed59243b59c39ce82bf9bc8f33c0a8b84854846afff6a89f5d37ffc97a096b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3b6d20cf459b1f6d3aff5b9290f9391909baf27ee40c49e2b0cd2d5cfe889ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fc9bfcb356b506d6c439adc22e4caaed4fad1c5a085744d44ebd50cf8898777\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3fc9bfcb356b506d6c439adc22e4caaed4fad1c5a085744d44ebd50cf8898777\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-14T07:01:56Z\\\",\\\"message\\\":\\\" 6555 ovn.go:134] Ensuring zone local for Pod openshift-image-registry/node-ca-kz77f in node crc\\\\nI1014 07:01:56.821005 6555 obj_retry.go:386] Retry successful for *v1.Pod openshift-image-registry/node-ca-kz77f after 0 failed attempt(s)\\\\nI1014 07:01:56.820992 6555 model_client.go:382] Update operations generated as: [{Op:update Table:Logical_Switch_Port Row:map[addresses:{GoSet:[0a:58:0a:d9:00:5c 10.217.0.92]} options:{GoMap:map[iface-id-ver:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 requested-chassis:crc]} port_security:{GoSet:[0a:58:0a:d9:00:5c 10.217.0.92]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c94130be-172c-477c-88c4-40cc7eba30fe}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1014 07:01:56.821025 6555 default_network_controller.go:776] Recording success event on pod openshift-image-registry/node-ca-kz77f\\\\nI1014 07:01:56.820931 6555 default_network_controller.go:776] Recording success event on pod openshift-dns/node-resolver-4v2v2\\\\nI1014 07:01:56.820614 6555 services_controller.go:454] Service openshift-marketplace/redhat-marketplace for network=default has 1 cluster-wide, 0 per-node configs, 0 template configs, making 1 (cluster) 0 (per node) and 0 (template) load balancers\\\\nI1014 07:01:56.821037 6555 model_client.go:382] Update operations generated as: [{Op:update Table:N\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:56Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-v7br2_openshift-ovn-kubernetes(8dcaef0e-42e3-424b-bbe5-4a023efe912d)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fe3468cdd612fc4378e1f6f55a2a9d00274f7ddb65cb57b2a368dfc065cfb4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ba24c2e067b9f3a14d9d018b66300b54d866530d30174fbdb3c899299575d28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ba24c2e067b9f3a14d9d018b66300b54d866530d30174fbdb3c899299575d28\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-v7br2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:14Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:14 crc kubenswrapper[4870]: I1014 07:02:14.757428 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-cdjmj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"612e7ab9-bda4-45d8-bc50-ff1afb68aace\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7776352d62b814fc63d01e4c16ad13f878b0597eb37a370bf9da0a88abab6cfa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://834444b5350ee12df4e4652c3009880d8d69e1edf8ba0c0156ab96fb2bbe28e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://834444b5350ee12df4e4652c3009880d8d69e1edf8ba0c0156ab96fb2bbe28e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c1972e7b2f75790db22a07a078eee97072bffbdb7f8e594bf2be5f800632a7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c1972e7b2f75790db22a07a078eee97072bffbdb7f8e594bf2be5f800632a7b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b124b42a44c6a23da1d92b9a5326987eb74430f80489a71e8913829475654ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b124b42a44c6a23da1d92b9a5326987eb74430f80489a71e8913829475654ab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62d0deb6b20ea65fbcbaf94c2fc14d69adb22e7a46cc043ccd4def2668ddd4b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62d0deb6b20ea65fbcbaf94c2fc14d69adb22e7a46cc043ccd4def2668ddd4b7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7f0d7cf29ca737d0effe58731a8b890379527c97f78c144d4eebb910260fadb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7f0d7cf29ca737d0effe58731a8b890379527c97f78c144d4eebb910260fadb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5959abb7c8e7ab32bed3aca709ae0334e8400f991ba724246c9fda56cbed822\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5959abb7c8e7ab32bed3aca709ae0334e8400f991ba724246c9fda56cbed822\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-cdjmj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:14Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:14 crc kubenswrapper[4870]: I1014 07:02:14.774470 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fdxjh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a3a430a-b540-4a8c-adad-b2893dbf9898\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:02:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:02:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93e41d5e927eed7ec62eb9c3607cc804cd8999556663363fd123b2a784b00ff1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://93e41d5e927eed7ec62eb9c3607cc804cd8999556663363fd123b2a784b00ff1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-14T07:02:13Z\\\",\\\"message\\\":\\\"2025-10-14T07:01:28+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_847000ef-6815-4615-8665-86c348661ad2\\\\n2025-10-14T07:01:28+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_847000ef-6815-4615-8665-86c348661ad2 to /host/opt/cni/bin/\\\\n2025-10-14T07:01:28Z [verbose] multus-daemon started\\\\n2025-10-14T07:01:28Z [verbose] Readiness Indicator file check\\\\n2025-10-14T07:02:13Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k787j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fdxjh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:14Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:14 crc kubenswrapper[4870]: I1014 07:02:14.789389 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:14 crc kubenswrapper[4870]: I1014 07:02:14.789441 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:14 crc kubenswrapper[4870]: I1014 07:02:14.789454 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:14 crc kubenswrapper[4870]: I1014 07:02:14.789535 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:14 crc kubenswrapper[4870]: I1014 07:02:14.789551 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:14Z","lastTransitionTime":"2025-10-14T07:02:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:14 crc kubenswrapper[4870]: I1014 07:02:14.792265 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-khtq8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28deeea6-a1ef-4a4a-a377-d954ccd87de9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://296bb048a1d99649beb57d609573833bf547cd4d628ebc823c6af83fe5736c9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rb79p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2b7bc4c93307e80b778f2282ced993898f1c4cfc350a7d74828b8f9dea9906d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rb79p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-khtq8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:14Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:14 crc kubenswrapper[4870]: I1014 07:02:14.811303 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33a9493e-9660-48db-b8d1-310e29ebe313\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3570643ef1b7d5cc3a542f320413b819338e0c8b7ab7f0f09cc909997a39bab8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb398668197dedc1c2a14b4ae7d6a29714f3527309ccb9c1622af841100e1d60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1d2b6d068d63bd38cdfb83824525f6d12054f6cabab5d684e03ea5f2152076c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://783ff243e395333f3fc80cfae21358c532f0c9c5cb5515586be56b90ed10d912\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a8135a3d01076e4437ae8a784b459bf8d16de57fa16ec7ce4d40c5129c9d4a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48ca3a5017ba650df99b2fc7201c3fe5abfa89d81089c5647cde1b1549f534fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48ca3a5017ba650df99b2fc7201c3fe5abfa89d81089c5647cde1b1549f534fc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://19682eca67f76bf2e06e1b80f12e21141fcbce0acb8195f0defacd1b9a10b5e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://19682eca67f76bf2e06e1b80f12e21141fcbce0acb8195f0defacd1b9a10b5e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://042aad6d5f6bee8eef77f9e6a2c265de775a69f3da72aedc68bdf312198f00fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://042aad6d5f6bee8eef77f9e6a2c265de775a69f3da72aedc68bdf312198f00fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:05Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:14Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:14 crc kubenswrapper[4870]: I1014 07:02:14.823990 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a77729fb-a187-48c6-80db-bee90f7bdac4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:02:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:02:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c859ebc0d20b632d1d5a716caee343c7e6b9b752ed5f257422cd35ad12c170f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e8a7a5c1f1706ebbd2e8165d8961002f5a9340e2749e377b0058db107ad7dbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f75b3dc1557917e9c0c35c2810338a076ed92ba89963edafb9d537ad19c2bd8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b192bd3e0eb8540d7492abc207025471bc9e198e7e78b50215c8d2f5bc57cf4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b192bd3e0eb8540d7492abc207025471bc9e198e7e78b50215c8d2f5bc57cf4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:05Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:14Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:14 crc kubenswrapper[4870]: I1014 07:02:14.835285 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-4v2v2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4796ba6-c23b-4764-be96-2e3a3bf5e962\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf5afb0b99a7101f6d5573fb7eb925c5a1d8becc11b2e3c6f3351d80b6f0e823\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnp8b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-4v2v2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:14Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:14 crc kubenswrapper[4870]: I1014 07:02:14.847596 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bcbc6dd6-aab0-4dbc-a62e-c58f06ec1337\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25ec4808b09adc44488530b3c4e29fd7fec46b3fcfa31d4449ec4a6dbc880ec8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f3f5de86fb11231a19034a18e9d7f31476c162e6567d0c136ead8bbbd900d6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d57d938d9d9c92198ae5b3857536ece36012fbfbe8cc8cfd245d1386d39b0be5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77a3247a7f65709543d7fd374b2ab69d7d793537d4132d7e22261188d6e58544\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:05Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:14Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:14 crc kubenswrapper[4870]: I1014 07:02:14.860391 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e539fac3d0a2ade314a4b4c1050bb27c37bcff5ba5272fa2d71b97ab5c0d7191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:14Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:14 crc kubenswrapper[4870]: I1014 07:02:14.875188 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:14Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:14 crc kubenswrapper[4870]: I1014 07:02:14.892767 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:14 crc kubenswrapper[4870]: I1014 07:02:14.892812 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:14 crc kubenswrapper[4870]: I1014 07:02:14.892823 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:14 crc kubenswrapper[4870]: I1014 07:02:14.892842 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:14 crc kubenswrapper[4870]: I1014 07:02:14.892852 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:14Z","lastTransitionTime":"2025-10-14T07:02:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:14 crc kubenswrapper[4870]: I1014 07:02:14.901223 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-kz77f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d3864a91-9bf6-4540-a908-cc41aa39656c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d4ded3d86030f5ea548c8b1cc3caddbe14a45c9a3ae4af836e892880584f57d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4h8v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:28Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-kz77f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:14Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:14 crc kubenswrapper[4870]: I1014 07:02:14.996165 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:14 crc kubenswrapper[4870]: I1014 07:02:14.996232 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:14 crc kubenswrapper[4870]: I1014 07:02:14.996437 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:14 crc kubenswrapper[4870]: I1014 07:02:14.996737 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:14 crc kubenswrapper[4870]: I1014 07:02:14.996767 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:14Z","lastTransitionTime":"2025-10-14T07:02:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:15 crc kubenswrapper[4870]: I1014 07:02:15.032973 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zc5j6" Oct 14 07:02:15 crc kubenswrapper[4870]: E1014 07:02:15.033128 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zc5j6" podUID="7d7a4f41-388a-4320-b896-43c17ff13da6" Oct 14 07:02:15 crc kubenswrapper[4870]: I1014 07:02:15.053005 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77a0d93a-5391-4112-b502-c474f45de209\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7815af9a584f6db6660073da289f2e8c987a04ef9ddfac155bdd9a15b014fcfa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b329b570622155a55f70501b2260e278106a68ebef497914231b4ec860e0125\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba942a0e6fe5ab93a187e0b7b4bd7c440c2e7827d9b5323fc50f90b562075392\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b9018b078c5c8276243c1696860f0aa73e2d8dd917913ae9d03b30dc9b5113a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://978e8db184ba9e430232c6bb2ee530beb92567fdc71ba3c35b0bedca3492d845\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"file observer\\\\nW1014 07:01:25.870672 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1014 07:01:25.870795 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1014 07:01:25.871521 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-756714914/tls.crt::/tmp/serving-cert-756714914/tls.key\\\\\\\"\\\\nI1014 07:01:26.460037 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1014 07:01:26.464876 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1014 07:01:26.464902 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1014 07:01:26.464925 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1014 07:01:26.464931 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1014 07:01:26.481374 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1014 07:01:26.481400 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:01:26.481405 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:01:26.481410 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1014 07:01:26.481414 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1014 07:01:26.481417 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1014 07:01:26.481420 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1014 07:01:26.481610 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1014 07:01:26.482872 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bd0044a247bdc5ca4cab7294103af3270b76cde61f3109b85e83ed050ca53b6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82d3c2214e0be679d2827594b159ace362151318689078bc49d3158f42899636\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82d3c2214e0be679d2827594b159ace362151318689078bc49d3158f42899636\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:05Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:15Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:15 crc kubenswrapper[4870]: I1014 07:02:15.067039 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:15Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:15 crc kubenswrapper[4870]: I1014 07:02:15.081691 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e26fe49ec2045d20a0cb2dd17e96e79b6eb43c641eb184c688cc472ca5db39d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aea115f7ee71c700b218fe15c928af7b55e6babe70e65d67f125d87be1fd2e2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:15Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:15 crc kubenswrapper[4870]: I1014 07:02:15.100130 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:15 crc kubenswrapper[4870]: I1014 07:02:15.100189 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:15 crc kubenswrapper[4870]: I1014 07:02:15.100205 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:15 crc kubenswrapper[4870]: I1014 07:02:15.100274 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:15 crc kubenswrapper[4870]: I1014 07:02:15.100291 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:15Z","lastTransitionTime":"2025-10-14T07:02:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:15 crc kubenswrapper[4870]: I1014 07:02:15.106760 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-v7br2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dcaef0e-42e3-424b-bbe5-4a023efe912d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71dec55e859b11a081dbb90d7d6cf6b4846cf591382303d8329003ff50c71f2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e72cff0723e2ce520340244b4e597d8016c58e1b071c4b087f2f6cd718741594\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc9d4db944f837289d1923176d60cfc4a566a282c2d2e3a703e220df884703c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d16ed1c65925549dcb69ab02c3013e25e3f123a83179f1af3a8e08bf91a59896\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed59243b59c39ce82bf9bc8f33c0a8b84854846afff6a89f5d37ffc97a096b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3b6d20cf459b1f6d3aff5b9290f9391909baf27ee40c49e2b0cd2d5cfe889ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fc9bfcb356b506d6c439adc22e4caaed4fad1c5a085744d44ebd50cf8898777\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3fc9bfcb356b506d6c439adc22e4caaed4fad1c5a085744d44ebd50cf8898777\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-14T07:01:56Z\\\",\\\"message\\\":\\\" 6555 ovn.go:134] Ensuring zone local for Pod openshift-image-registry/node-ca-kz77f in node crc\\\\nI1014 07:01:56.821005 6555 obj_retry.go:386] Retry successful for *v1.Pod openshift-image-registry/node-ca-kz77f after 0 failed attempt(s)\\\\nI1014 07:01:56.820992 6555 model_client.go:382] Update operations generated as: [{Op:update Table:Logical_Switch_Port Row:map[addresses:{GoSet:[0a:58:0a:d9:00:5c 10.217.0.92]} options:{GoMap:map[iface-id-ver:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 requested-chassis:crc]} port_security:{GoSet:[0a:58:0a:d9:00:5c 10.217.0.92]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c94130be-172c-477c-88c4-40cc7eba30fe}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1014 07:01:56.821025 6555 default_network_controller.go:776] Recording success event on pod openshift-image-registry/node-ca-kz77f\\\\nI1014 07:01:56.820931 6555 default_network_controller.go:776] Recording success event on pod openshift-dns/node-resolver-4v2v2\\\\nI1014 07:01:56.820614 6555 services_controller.go:454] Service openshift-marketplace/redhat-marketplace for network=default has 1 cluster-wide, 0 per-node configs, 0 template configs, making 1 (cluster) 0 (per node) and 0 (template) load balancers\\\\nI1014 07:01:56.821037 6555 model_client.go:382] Update operations generated as: [{Op:update Table:N\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:56Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-v7br2_openshift-ovn-kubernetes(8dcaef0e-42e3-424b-bbe5-4a023efe912d)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fe3468cdd612fc4378e1f6f55a2a9d00274f7ddb65cb57b2a368dfc065cfb4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ba24c2e067b9f3a14d9d018b66300b54d866530d30174fbdb3c899299575d28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ba24c2e067b9f3a14d9d018b66300b54d866530d30174fbdb3c899299575d28\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-v7br2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:15Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:15 crc kubenswrapper[4870]: I1014 07:02:15.127400 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-cdjmj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"612e7ab9-bda4-45d8-bc50-ff1afb68aace\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7776352d62b814fc63d01e4c16ad13f878b0597eb37a370bf9da0a88abab6cfa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://834444b5350ee12df4e4652c3009880d8d69e1edf8ba0c0156ab96fb2bbe28e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://834444b5350ee12df4e4652c3009880d8d69e1edf8ba0c0156ab96fb2bbe28e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c1972e7b2f75790db22a07a078eee97072bffbdb7f8e594bf2be5f800632a7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c1972e7b2f75790db22a07a078eee97072bffbdb7f8e594bf2be5f800632a7b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b124b42a44c6a23da1d92b9a5326987eb74430f80489a71e8913829475654ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b124b42a44c6a23da1d92b9a5326987eb74430f80489a71e8913829475654ab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62d0deb6b20ea65fbcbaf94c2fc14d69adb22e7a46cc043ccd4def2668ddd4b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62d0deb6b20ea65fbcbaf94c2fc14d69adb22e7a46cc043ccd4def2668ddd4b7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7f0d7cf29ca737d0effe58731a8b890379527c97f78c144d4eebb910260fadb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7f0d7cf29ca737d0effe58731a8b890379527c97f78c144d4eebb910260fadb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5959abb7c8e7ab32bed3aca709ae0334e8400f991ba724246c9fda56cbed822\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5959abb7c8e7ab32bed3aca709ae0334e8400f991ba724246c9fda56cbed822\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-cdjmj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:15Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:15 crc kubenswrapper[4870]: I1014 07:02:15.145070 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fdxjh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a3a430a-b540-4a8c-adad-b2893dbf9898\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:02:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:02:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93e41d5e927eed7ec62eb9c3607cc804cd8999556663363fd123b2a784b00ff1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://93e41d5e927eed7ec62eb9c3607cc804cd8999556663363fd123b2a784b00ff1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-14T07:02:13Z\\\",\\\"message\\\":\\\"2025-10-14T07:01:28+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_847000ef-6815-4615-8665-86c348661ad2\\\\n2025-10-14T07:01:28+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_847000ef-6815-4615-8665-86c348661ad2 to /host/opt/cni/bin/\\\\n2025-10-14T07:01:28Z [verbose] multus-daemon started\\\\n2025-10-14T07:01:28Z [verbose] Readiness Indicator file check\\\\n2025-10-14T07:02:13Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k787j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fdxjh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:15Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:15 crc kubenswrapper[4870]: I1014 07:02:15.203974 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:15 crc kubenswrapper[4870]: I1014 07:02:15.204036 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:15 crc kubenswrapper[4870]: I1014 07:02:15.204053 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:15 crc kubenswrapper[4870]: I1014 07:02:15.204074 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:15 crc kubenswrapper[4870]: I1014 07:02:15.204092 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:15Z","lastTransitionTime":"2025-10-14T07:02:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:15 crc kubenswrapper[4870]: I1014 07:02:15.214891 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-khtq8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28deeea6-a1ef-4a4a-a377-d954ccd87de9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://296bb048a1d99649beb57d609573833bf547cd4d628ebc823c6af83fe5736c9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rb79p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2b7bc4c93307e80b778f2282ced993898f1c4cfc350a7d74828b8f9dea9906d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rb79p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-khtq8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:15Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:15 crc kubenswrapper[4870]: I1014 07:02:15.232880 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33a9493e-9660-48db-b8d1-310e29ebe313\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3570643ef1b7d5cc3a542f320413b819338e0c8b7ab7f0f09cc909997a39bab8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb398668197dedc1c2a14b4ae7d6a29714f3527309ccb9c1622af841100e1d60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1d2b6d068d63bd38cdfb83824525f6d12054f6cabab5d684e03ea5f2152076c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://783ff243e395333f3fc80cfae21358c532f0c9c5cb5515586be56b90ed10d912\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a8135a3d01076e4437ae8a784b459bf8d16de57fa16ec7ce4d40c5129c9d4a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48ca3a5017ba650df99b2fc7201c3fe5abfa89d81089c5647cde1b1549f534fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48ca3a5017ba650df99b2fc7201c3fe5abfa89d81089c5647cde1b1549f534fc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://19682eca67f76bf2e06e1b80f12e21141fcbce0acb8195f0defacd1b9a10b5e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://19682eca67f76bf2e06e1b80f12e21141fcbce0acb8195f0defacd1b9a10b5e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://042aad6d5f6bee8eef77f9e6a2c265de775a69f3da72aedc68bdf312198f00fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://042aad6d5f6bee8eef77f9e6a2c265de775a69f3da72aedc68bdf312198f00fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:05Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:15Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:15 crc kubenswrapper[4870]: I1014 07:02:15.244553 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a77729fb-a187-48c6-80db-bee90f7bdac4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:02:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:02:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c859ebc0d20b632d1d5a716caee343c7e6b9b752ed5f257422cd35ad12c170f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e8a7a5c1f1706ebbd2e8165d8961002f5a9340e2749e377b0058db107ad7dbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f75b3dc1557917e9c0c35c2810338a076ed92ba89963edafb9d537ad19c2bd8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b192bd3e0eb8540d7492abc207025471bc9e198e7e78b50215c8d2f5bc57cf4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b192bd3e0eb8540d7492abc207025471bc9e198e7e78b50215c8d2f5bc57cf4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:05Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:15Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:15 crc kubenswrapper[4870]: I1014 07:02:15.258554 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-4v2v2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4796ba6-c23b-4764-be96-2e3a3bf5e962\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf5afb0b99a7101f6d5573fb7eb925c5a1d8becc11b2e3c6f3351d80b6f0e823\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnp8b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-4v2v2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:15Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:15 crc kubenswrapper[4870]: I1014 07:02:15.278825 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bcbc6dd6-aab0-4dbc-a62e-c58f06ec1337\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25ec4808b09adc44488530b3c4e29fd7fec46b3fcfa31d4449ec4a6dbc880ec8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f3f5de86fb11231a19034a18e9d7f31476c162e6567d0c136ead8bbbd900d6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d57d938d9d9c92198ae5b3857536ece36012fbfbe8cc8cfd245d1386d39b0be5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77a3247a7f65709543d7fd374b2ab69d7d793537d4132d7e22261188d6e58544\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:05Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:15Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:15 crc kubenswrapper[4870]: I1014 07:02:15.294577 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e539fac3d0a2ade314a4b4c1050bb27c37bcff5ba5272fa2d71b97ab5c0d7191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:15Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:15 crc kubenswrapper[4870]: I1014 07:02:15.307077 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:15 crc kubenswrapper[4870]: I1014 07:02:15.307120 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:15 crc kubenswrapper[4870]: I1014 07:02:15.307134 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:15 crc kubenswrapper[4870]: I1014 07:02:15.307156 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:15 crc kubenswrapper[4870]: I1014 07:02:15.307170 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:15Z","lastTransitionTime":"2025-10-14T07:02:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:15 crc kubenswrapper[4870]: I1014 07:02:15.310269 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:15Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:15 crc kubenswrapper[4870]: I1014 07:02:15.324169 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-kz77f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d3864a91-9bf6-4540-a908-cc41aa39656c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d4ded3d86030f5ea548c8b1cc3caddbe14a45c9a3ae4af836e892880584f57d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4h8v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:28Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-kz77f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:15Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:15 crc kubenswrapper[4870]: I1014 07:02:15.342313 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:15Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:15 crc kubenswrapper[4870]: I1014 07:02:15.356383 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3eb0b397c05ffd6650607f1be4a75683e0342d3333fc096bf1cd63c7688834f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:15Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:15 crc kubenswrapper[4870]: I1014 07:02:15.370092 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"874a5e45-dffd-4d17-b609-b7d3ed2eab07\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d98d044836ab44a405196915ef7bf572b75113f2c81ec7e64effa63b47ad3543\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wj4ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8004752575df4a1c44efbcafe64799f09126d0f97285d92388c76a0b90e022f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wj4ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7tvc7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:15Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:15 crc kubenswrapper[4870]: I1014 07:02:15.380347 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zc5j6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d7a4f41-388a-4320-b896-43c17ff13da6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6p2wh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6p2wh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:40Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zc5j6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:15Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:15 crc kubenswrapper[4870]: I1014 07:02:15.411159 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:15 crc kubenswrapper[4870]: I1014 07:02:15.411192 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:15 crc kubenswrapper[4870]: I1014 07:02:15.411202 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:15 crc kubenswrapper[4870]: I1014 07:02:15.411220 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:15 crc kubenswrapper[4870]: I1014 07:02:15.411230 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:15Z","lastTransitionTime":"2025-10-14T07:02:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:15 crc kubenswrapper[4870]: I1014 07:02:15.514749 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:15 crc kubenswrapper[4870]: I1014 07:02:15.514798 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:15 crc kubenswrapper[4870]: I1014 07:02:15.514811 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:15 crc kubenswrapper[4870]: I1014 07:02:15.514833 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:15 crc kubenswrapper[4870]: I1014 07:02:15.514844 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:15Z","lastTransitionTime":"2025-10-14T07:02:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:15 crc kubenswrapper[4870]: I1014 07:02:15.609389 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-fdxjh_8a3a430a-b540-4a8c-adad-b2893dbf9898/kube-multus/0.log" Oct 14 07:02:15 crc kubenswrapper[4870]: I1014 07:02:15.609508 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-fdxjh" event={"ID":"8a3a430a-b540-4a8c-adad-b2893dbf9898","Type":"ContainerStarted","Data":"88614373bfe4779bbe61cc41cc4c35f419e438bdae3cec4964271bf0f74ad3f5"} Oct 14 07:02:15 crc kubenswrapper[4870]: I1014 07:02:15.617293 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:15 crc kubenswrapper[4870]: I1014 07:02:15.617343 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:15 crc kubenswrapper[4870]: I1014 07:02:15.617363 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:15 crc kubenswrapper[4870]: I1014 07:02:15.617385 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:15 crc kubenswrapper[4870]: I1014 07:02:15.617400 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:15Z","lastTransitionTime":"2025-10-14T07:02:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:15 crc kubenswrapper[4870]: I1014 07:02:15.633169 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33a9493e-9660-48db-b8d1-310e29ebe313\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3570643ef1b7d5cc3a542f320413b819338e0c8b7ab7f0f09cc909997a39bab8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb398668197dedc1c2a14b4ae7d6a29714f3527309ccb9c1622af841100e1d60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1d2b6d068d63bd38cdfb83824525f6d12054f6cabab5d684e03ea5f2152076c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://783ff243e395333f3fc80cfae21358c532f0c9c5cb5515586be56b90ed10d912\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a8135a3d01076e4437ae8a784b459bf8d16de57fa16ec7ce4d40c5129c9d4a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48ca3a5017ba650df99b2fc7201c3fe5abfa89d81089c5647cde1b1549f534fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48ca3a5017ba650df99b2fc7201c3fe5abfa89d81089c5647cde1b1549f534fc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://19682eca67f76bf2e06e1b80f12e21141fcbce0acb8195f0defacd1b9a10b5e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://19682eca67f76bf2e06e1b80f12e21141fcbce0acb8195f0defacd1b9a10b5e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://042aad6d5f6bee8eef77f9e6a2c265de775a69f3da72aedc68bdf312198f00fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://042aad6d5f6bee8eef77f9e6a2c265de775a69f3da72aedc68bdf312198f00fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:05Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:15Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:15 crc kubenswrapper[4870]: I1014 07:02:15.650192 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a77729fb-a187-48c6-80db-bee90f7bdac4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:02:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:02:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c859ebc0d20b632d1d5a716caee343c7e6b9b752ed5f257422cd35ad12c170f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e8a7a5c1f1706ebbd2e8165d8961002f5a9340e2749e377b0058db107ad7dbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f75b3dc1557917e9c0c35c2810338a076ed92ba89963edafb9d537ad19c2bd8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b192bd3e0eb8540d7492abc207025471bc9e198e7e78b50215c8d2f5bc57cf4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b192bd3e0eb8540d7492abc207025471bc9e198e7e78b50215c8d2f5bc57cf4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:05Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:15Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:15 crc kubenswrapper[4870]: I1014 07:02:15.668563 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-4v2v2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4796ba6-c23b-4764-be96-2e3a3bf5e962\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf5afb0b99a7101f6d5573fb7eb925c5a1d8becc11b2e3c6f3351d80b6f0e823\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnp8b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-4v2v2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:15Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:15 crc kubenswrapper[4870]: I1014 07:02:15.689413 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-cdjmj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"612e7ab9-bda4-45d8-bc50-ff1afb68aace\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7776352d62b814fc63d01e4c16ad13f878b0597eb37a370bf9da0a88abab6cfa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://834444b5350ee12df4e4652c3009880d8d69e1edf8ba0c0156ab96fb2bbe28e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://834444b5350ee12df4e4652c3009880d8d69e1edf8ba0c0156ab96fb2bbe28e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c1972e7b2f75790db22a07a078eee97072bffbdb7f8e594bf2be5f800632a7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c1972e7b2f75790db22a07a078eee97072bffbdb7f8e594bf2be5f800632a7b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b124b42a44c6a23da1d92b9a5326987eb74430f80489a71e8913829475654ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b124b42a44c6a23da1d92b9a5326987eb74430f80489a71e8913829475654ab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62d0deb6b20ea65fbcbaf94c2fc14d69adb22e7a46cc043ccd4def2668ddd4b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62d0deb6b20ea65fbcbaf94c2fc14d69adb22e7a46cc043ccd4def2668ddd4b7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7f0d7cf29ca737d0effe58731a8b890379527c97f78c144d4eebb910260fadb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7f0d7cf29ca737d0effe58731a8b890379527c97f78c144d4eebb910260fadb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5959abb7c8e7ab32bed3aca709ae0334e8400f991ba724246c9fda56cbed822\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5959abb7c8e7ab32bed3aca709ae0334e8400f991ba724246c9fda56cbed822\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-cdjmj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:15Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:15 crc kubenswrapper[4870]: I1014 07:02:15.706832 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fdxjh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a3a430a-b540-4a8c-adad-b2893dbf9898\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:02:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:02:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88614373bfe4779bbe61cc41cc4c35f419e438bdae3cec4964271bf0f74ad3f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://93e41d5e927eed7ec62eb9c3607cc804cd8999556663363fd123b2a784b00ff1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-14T07:02:13Z\\\",\\\"message\\\":\\\"2025-10-14T07:01:28+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_847000ef-6815-4615-8665-86c348661ad2\\\\n2025-10-14T07:01:28+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_847000ef-6815-4615-8665-86c348661ad2 to /host/opt/cni/bin/\\\\n2025-10-14T07:01:28Z [verbose] multus-daemon started\\\\n2025-10-14T07:01:28Z [verbose] Readiness Indicator file check\\\\n2025-10-14T07:02:13Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:02:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k787j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fdxjh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:15Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:15 crc kubenswrapper[4870]: I1014 07:02:15.720042 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:15 crc kubenswrapper[4870]: I1014 07:02:15.720091 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:15 crc kubenswrapper[4870]: I1014 07:02:15.720104 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:15 crc kubenswrapper[4870]: I1014 07:02:15.720121 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:15 crc kubenswrapper[4870]: I1014 07:02:15.720133 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:15Z","lastTransitionTime":"2025-10-14T07:02:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:15 crc kubenswrapper[4870]: I1014 07:02:15.720239 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-khtq8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28deeea6-a1ef-4a4a-a377-d954ccd87de9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://296bb048a1d99649beb57d609573833bf547cd4d628ebc823c6af83fe5736c9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rb79p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2b7bc4c93307e80b778f2282ced993898f1c4cfc350a7d74828b8f9dea9906d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rb79p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-khtq8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:15Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:15 crc kubenswrapper[4870]: I1014 07:02:15.734068 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bcbc6dd6-aab0-4dbc-a62e-c58f06ec1337\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25ec4808b09adc44488530b3c4e29fd7fec46b3fcfa31d4449ec4a6dbc880ec8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f3f5de86fb11231a19034a18e9d7f31476c162e6567d0c136ead8bbbd900d6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d57d938d9d9c92198ae5b3857536ece36012fbfbe8cc8cfd245d1386d39b0be5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77a3247a7f65709543d7fd374b2ab69d7d793537d4132d7e22261188d6e58544\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:05Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:15Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:15 crc kubenswrapper[4870]: I1014 07:02:15.746423 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e539fac3d0a2ade314a4b4c1050bb27c37bcff5ba5272fa2d71b97ab5c0d7191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:15Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:15 crc kubenswrapper[4870]: I1014 07:02:15.760439 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:15Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:15 crc kubenswrapper[4870]: I1014 07:02:15.772688 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-kz77f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d3864a91-9bf6-4540-a908-cc41aa39656c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d4ded3d86030f5ea548c8b1cc3caddbe14a45c9a3ae4af836e892880584f57d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4h8v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:28Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-kz77f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:15Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:15 crc kubenswrapper[4870]: I1014 07:02:15.786902 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:15Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:15 crc kubenswrapper[4870]: I1014 07:02:15.800023 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3eb0b397c05ffd6650607f1be4a75683e0342d3333fc096bf1cd63c7688834f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:15Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:15 crc kubenswrapper[4870]: I1014 07:02:15.813653 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"874a5e45-dffd-4d17-b609-b7d3ed2eab07\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d98d044836ab44a405196915ef7bf572b75113f2c81ec7e64effa63b47ad3543\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wj4ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8004752575df4a1c44efbcafe64799f09126d0f97285d92388c76a0b90e022f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wj4ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7tvc7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:15Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:15 crc kubenswrapper[4870]: I1014 07:02:15.823672 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:15 crc kubenswrapper[4870]: I1014 07:02:15.823736 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:15 crc kubenswrapper[4870]: I1014 07:02:15.823749 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:15 crc kubenswrapper[4870]: I1014 07:02:15.823772 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:15 crc kubenswrapper[4870]: I1014 07:02:15.823784 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:15Z","lastTransitionTime":"2025-10-14T07:02:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:15 crc kubenswrapper[4870]: I1014 07:02:15.827258 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zc5j6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d7a4f41-388a-4320-b896-43c17ff13da6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6p2wh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6p2wh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:40Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zc5j6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:15Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:15 crc kubenswrapper[4870]: I1014 07:02:15.844760 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77a0d93a-5391-4112-b502-c474f45de209\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7815af9a584f6db6660073da289f2e8c987a04ef9ddfac155bdd9a15b014fcfa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b329b570622155a55f70501b2260e278106a68ebef497914231b4ec860e0125\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba942a0e6fe5ab93a187e0b7b4bd7c440c2e7827d9b5323fc50f90b562075392\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b9018b078c5c8276243c1696860f0aa73e2d8dd917913ae9d03b30dc9b5113a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://978e8db184ba9e430232c6bb2ee530beb92567fdc71ba3c35b0bedca3492d845\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"file observer\\\\nW1014 07:01:25.870672 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1014 07:01:25.870795 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1014 07:01:25.871521 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-756714914/tls.crt::/tmp/serving-cert-756714914/tls.key\\\\\\\"\\\\nI1014 07:01:26.460037 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1014 07:01:26.464876 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1014 07:01:26.464902 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1014 07:01:26.464925 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1014 07:01:26.464931 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1014 07:01:26.481374 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1014 07:01:26.481400 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:01:26.481405 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:01:26.481410 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1014 07:01:26.481414 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1014 07:01:26.481417 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1014 07:01:26.481420 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1014 07:01:26.481610 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1014 07:01:26.482872 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bd0044a247bdc5ca4cab7294103af3270b76cde61f3109b85e83ed050ca53b6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82d3c2214e0be679d2827594b159ace362151318689078bc49d3158f42899636\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82d3c2214e0be679d2827594b159ace362151318689078bc49d3158f42899636\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:05Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:15Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:15 crc kubenswrapper[4870]: I1014 07:02:15.861603 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:15Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:15 crc kubenswrapper[4870]: I1014 07:02:15.877778 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e26fe49ec2045d20a0cb2dd17e96e79b6eb43c641eb184c688cc472ca5db39d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aea115f7ee71c700b218fe15c928af7b55e6babe70e65d67f125d87be1fd2e2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:15Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:15 crc kubenswrapper[4870]: I1014 07:02:15.897359 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-v7br2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dcaef0e-42e3-424b-bbe5-4a023efe912d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71dec55e859b11a081dbb90d7d6cf6b4846cf591382303d8329003ff50c71f2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e72cff0723e2ce520340244b4e597d8016c58e1b071c4b087f2f6cd718741594\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc9d4db944f837289d1923176d60cfc4a566a282c2d2e3a703e220df884703c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d16ed1c65925549dcb69ab02c3013e25e3f123a83179f1af3a8e08bf91a59896\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed59243b59c39ce82bf9bc8f33c0a8b84854846afff6a89f5d37ffc97a096b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3b6d20cf459b1f6d3aff5b9290f9391909baf27ee40c49e2b0cd2d5cfe889ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fc9bfcb356b506d6c439adc22e4caaed4fad1c5a085744d44ebd50cf8898777\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3fc9bfcb356b506d6c439adc22e4caaed4fad1c5a085744d44ebd50cf8898777\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-14T07:01:56Z\\\",\\\"message\\\":\\\" 6555 ovn.go:134] Ensuring zone local for Pod openshift-image-registry/node-ca-kz77f in node crc\\\\nI1014 07:01:56.821005 6555 obj_retry.go:386] Retry successful for *v1.Pod openshift-image-registry/node-ca-kz77f after 0 failed attempt(s)\\\\nI1014 07:01:56.820992 6555 model_client.go:382] Update operations generated as: [{Op:update Table:Logical_Switch_Port Row:map[addresses:{GoSet:[0a:58:0a:d9:00:5c 10.217.0.92]} options:{GoMap:map[iface-id-ver:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 requested-chassis:crc]} port_security:{GoSet:[0a:58:0a:d9:00:5c 10.217.0.92]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c94130be-172c-477c-88c4-40cc7eba30fe}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1014 07:01:56.821025 6555 default_network_controller.go:776] Recording success event on pod openshift-image-registry/node-ca-kz77f\\\\nI1014 07:01:56.820931 6555 default_network_controller.go:776] Recording success event on pod openshift-dns/node-resolver-4v2v2\\\\nI1014 07:01:56.820614 6555 services_controller.go:454] Service openshift-marketplace/redhat-marketplace for network=default has 1 cluster-wide, 0 per-node configs, 0 template configs, making 1 (cluster) 0 (per node) and 0 (template) load balancers\\\\nI1014 07:01:56.821037 6555 model_client.go:382] Update operations generated as: [{Op:update Table:N\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:56Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-v7br2_openshift-ovn-kubernetes(8dcaef0e-42e3-424b-bbe5-4a023efe912d)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fe3468cdd612fc4378e1f6f55a2a9d00274f7ddb65cb57b2a368dfc065cfb4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ba24c2e067b9f3a14d9d018b66300b54d866530d30174fbdb3c899299575d28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ba24c2e067b9f3a14d9d018b66300b54d866530d30174fbdb3c899299575d28\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-v7br2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:15Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:15 crc kubenswrapper[4870]: I1014 07:02:15.926539 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:15 crc kubenswrapper[4870]: I1014 07:02:15.926601 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:15 crc kubenswrapper[4870]: I1014 07:02:15.926617 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:15 crc kubenswrapper[4870]: I1014 07:02:15.926642 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:15 crc kubenswrapper[4870]: I1014 07:02:15.926660 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:15Z","lastTransitionTime":"2025-10-14T07:02:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:16 crc kubenswrapper[4870]: I1014 07:02:16.030096 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:16 crc kubenswrapper[4870]: I1014 07:02:16.030146 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:16 crc kubenswrapper[4870]: I1014 07:02:16.030157 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:16 crc kubenswrapper[4870]: I1014 07:02:16.030175 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:16 crc kubenswrapper[4870]: I1014 07:02:16.030186 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:16Z","lastTransitionTime":"2025-10-14T07:02:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:16 crc kubenswrapper[4870]: I1014 07:02:16.033486 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 07:02:16 crc kubenswrapper[4870]: I1014 07:02:16.033517 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 07:02:16 crc kubenswrapper[4870]: I1014 07:02:16.033489 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 07:02:16 crc kubenswrapper[4870]: E1014 07:02:16.033617 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 07:02:16 crc kubenswrapper[4870]: E1014 07:02:16.033676 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 07:02:16 crc kubenswrapper[4870]: E1014 07:02:16.033760 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 07:02:16 crc kubenswrapper[4870]: I1014 07:02:16.135969 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:16 crc kubenswrapper[4870]: I1014 07:02:16.136050 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:16 crc kubenswrapper[4870]: I1014 07:02:16.136069 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:16 crc kubenswrapper[4870]: I1014 07:02:16.136107 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:16 crc kubenswrapper[4870]: I1014 07:02:16.136125 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:16Z","lastTransitionTime":"2025-10-14T07:02:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:16 crc kubenswrapper[4870]: I1014 07:02:16.240518 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:16 crc kubenswrapper[4870]: I1014 07:02:16.240605 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:16 crc kubenswrapper[4870]: I1014 07:02:16.240623 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:16 crc kubenswrapper[4870]: I1014 07:02:16.240646 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:16 crc kubenswrapper[4870]: I1014 07:02:16.240671 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:16Z","lastTransitionTime":"2025-10-14T07:02:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:16 crc kubenswrapper[4870]: I1014 07:02:16.343893 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:16 crc kubenswrapper[4870]: I1014 07:02:16.343940 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:16 crc kubenswrapper[4870]: I1014 07:02:16.343950 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:16 crc kubenswrapper[4870]: I1014 07:02:16.343967 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:16 crc kubenswrapper[4870]: I1014 07:02:16.343977 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:16Z","lastTransitionTime":"2025-10-14T07:02:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:16 crc kubenswrapper[4870]: I1014 07:02:16.447009 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:16 crc kubenswrapper[4870]: I1014 07:02:16.447038 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:16 crc kubenswrapper[4870]: I1014 07:02:16.447046 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:16 crc kubenswrapper[4870]: I1014 07:02:16.447062 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:16 crc kubenswrapper[4870]: I1014 07:02:16.447071 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:16Z","lastTransitionTime":"2025-10-14T07:02:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:16 crc kubenswrapper[4870]: I1014 07:02:16.550876 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:16 crc kubenswrapper[4870]: I1014 07:02:16.550930 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:16 crc kubenswrapper[4870]: I1014 07:02:16.550943 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:16 crc kubenswrapper[4870]: I1014 07:02:16.550961 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:16 crc kubenswrapper[4870]: I1014 07:02:16.550975 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:16Z","lastTransitionTime":"2025-10-14T07:02:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:16 crc kubenswrapper[4870]: I1014 07:02:16.653959 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:16 crc kubenswrapper[4870]: I1014 07:02:16.654015 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:16 crc kubenswrapper[4870]: I1014 07:02:16.654030 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:16 crc kubenswrapper[4870]: I1014 07:02:16.654050 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:16 crc kubenswrapper[4870]: I1014 07:02:16.654065 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:16Z","lastTransitionTime":"2025-10-14T07:02:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:16 crc kubenswrapper[4870]: I1014 07:02:16.756666 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:16 crc kubenswrapper[4870]: I1014 07:02:16.756705 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:16 crc kubenswrapper[4870]: I1014 07:02:16.756714 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:16 crc kubenswrapper[4870]: I1014 07:02:16.756729 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:16 crc kubenswrapper[4870]: I1014 07:02:16.756740 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:16Z","lastTransitionTime":"2025-10-14T07:02:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:16 crc kubenswrapper[4870]: I1014 07:02:16.859074 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:16 crc kubenswrapper[4870]: I1014 07:02:16.859137 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:16 crc kubenswrapper[4870]: I1014 07:02:16.859157 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:16 crc kubenswrapper[4870]: I1014 07:02:16.859184 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:16 crc kubenswrapper[4870]: I1014 07:02:16.859203 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:16Z","lastTransitionTime":"2025-10-14T07:02:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:16 crc kubenswrapper[4870]: I1014 07:02:16.961354 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:16 crc kubenswrapper[4870]: I1014 07:02:16.961399 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:16 crc kubenswrapper[4870]: I1014 07:02:16.961411 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:16 crc kubenswrapper[4870]: I1014 07:02:16.961427 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:16 crc kubenswrapper[4870]: I1014 07:02:16.961461 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:16Z","lastTransitionTime":"2025-10-14T07:02:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:17 crc kubenswrapper[4870]: I1014 07:02:17.033676 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zc5j6" Oct 14 07:02:17 crc kubenswrapper[4870]: E1014 07:02:17.033872 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zc5j6" podUID="7d7a4f41-388a-4320-b896-43c17ff13da6" Oct 14 07:02:17 crc kubenswrapper[4870]: I1014 07:02:17.064580 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:17 crc kubenswrapper[4870]: I1014 07:02:17.064651 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:17 crc kubenswrapper[4870]: I1014 07:02:17.064667 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:17 crc kubenswrapper[4870]: I1014 07:02:17.064694 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:17 crc kubenswrapper[4870]: I1014 07:02:17.064713 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:17Z","lastTransitionTime":"2025-10-14T07:02:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:17 crc kubenswrapper[4870]: I1014 07:02:17.168186 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:17 crc kubenswrapper[4870]: I1014 07:02:17.168237 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:17 crc kubenswrapper[4870]: I1014 07:02:17.168247 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:17 crc kubenswrapper[4870]: I1014 07:02:17.168269 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:17 crc kubenswrapper[4870]: I1014 07:02:17.168283 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:17Z","lastTransitionTime":"2025-10-14T07:02:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:17 crc kubenswrapper[4870]: I1014 07:02:17.273595 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:17 crc kubenswrapper[4870]: I1014 07:02:17.273652 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:17 crc kubenswrapper[4870]: I1014 07:02:17.274488 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:17 crc kubenswrapper[4870]: I1014 07:02:17.274532 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:17 crc kubenswrapper[4870]: I1014 07:02:17.274552 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:17Z","lastTransitionTime":"2025-10-14T07:02:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:17 crc kubenswrapper[4870]: I1014 07:02:17.377626 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:17 crc kubenswrapper[4870]: I1014 07:02:17.377767 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:17 crc kubenswrapper[4870]: I1014 07:02:17.377785 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:17 crc kubenswrapper[4870]: I1014 07:02:17.377806 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:17 crc kubenswrapper[4870]: I1014 07:02:17.377818 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:17Z","lastTransitionTime":"2025-10-14T07:02:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:17 crc kubenswrapper[4870]: I1014 07:02:17.480387 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:17 crc kubenswrapper[4870]: I1014 07:02:17.480435 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:17 crc kubenswrapper[4870]: I1014 07:02:17.480464 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:17 crc kubenswrapper[4870]: I1014 07:02:17.480481 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:17 crc kubenswrapper[4870]: I1014 07:02:17.480491 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:17Z","lastTransitionTime":"2025-10-14T07:02:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:17 crc kubenswrapper[4870]: I1014 07:02:17.582910 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:17 crc kubenswrapper[4870]: I1014 07:02:17.582969 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:17 crc kubenswrapper[4870]: I1014 07:02:17.582981 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:17 crc kubenswrapper[4870]: I1014 07:02:17.583003 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:17 crc kubenswrapper[4870]: I1014 07:02:17.583016 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:17Z","lastTransitionTime":"2025-10-14T07:02:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:17 crc kubenswrapper[4870]: I1014 07:02:17.686239 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:17 crc kubenswrapper[4870]: I1014 07:02:17.686293 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:17 crc kubenswrapper[4870]: I1014 07:02:17.686310 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:17 crc kubenswrapper[4870]: I1014 07:02:17.686336 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:17 crc kubenswrapper[4870]: I1014 07:02:17.686353 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:17Z","lastTransitionTime":"2025-10-14T07:02:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:17 crc kubenswrapper[4870]: I1014 07:02:17.788524 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:17 crc kubenswrapper[4870]: I1014 07:02:17.788577 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:17 crc kubenswrapper[4870]: I1014 07:02:17.788589 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:17 crc kubenswrapper[4870]: I1014 07:02:17.788607 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:17 crc kubenswrapper[4870]: I1014 07:02:17.788620 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:17Z","lastTransitionTime":"2025-10-14T07:02:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:17 crc kubenswrapper[4870]: I1014 07:02:17.891826 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:17 crc kubenswrapper[4870]: I1014 07:02:17.891887 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:17 crc kubenswrapper[4870]: I1014 07:02:17.891899 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:17 crc kubenswrapper[4870]: I1014 07:02:17.891920 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:17 crc kubenswrapper[4870]: I1014 07:02:17.891933 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:17Z","lastTransitionTime":"2025-10-14T07:02:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:17 crc kubenswrapper[4870]: I1014 07:02:17.995326 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:17 crc kubenswrapper[4870]: I1014 07:02:17.995361 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:17 crc kubenswrapper[4870]: I1014 07:02:17.995370 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:17 crc kubenswrapper[4870]: I1014 07:02:17.995386 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:17 crc kubenswrapper[4870]: I1014 07:02:17.995395 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:17Z","lastTransitionTime":"2025-10-14T07:02:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:18 crc kubenswrapper[4870]: I1014 07:02:18.033340 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 07:02:18 crc kubenswrapper[4870]: I1014 07:02:18.033389 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 07:02:18 crc kubenswrapper[4870]: I1014 07:02:18.033357 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 07:02:18 crc kubenswrapper[4870]: E1014 07:02:18.033562 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 07:02:18 crc kubenswrapper[4870]: E1014 07:02:18.033742 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 07:02:18 crc kubenswrapper[4870]: E1014 07:02:18.033807 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 07:02:18 crc kubenswrapper[4870]: I1014 07:02:18.098369 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:18 crc kubenswrapper[4870]: I1014 07:02:18.098410 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:18 crc kubenswrapper[4870]: I1014 07:02:18.098420 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:18 crc kubenswrapper[4870]: I1014 07:02:18.098437 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:18 crc kubenswrapper[4870]: I1014 07:02:18.098480 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:18Z","lastTransitionTime":"2025-10-14T07:02:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:18 crc kubenswrapper[4870]: I1014 07:02:18.201858 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:18 crc kubenswrapper[4870]: I1014 07:02:18.201924 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:18 crc kubenswrapper[4870]: I1014 07:02:18.201936 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:18 crc kubenswrapper[4870]: I1014 07:02:18.201960 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:18 crc kubenswrapper[4870]: I1014 07:02:18.201974 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:18Z","lastTransitionTime":"2025-10-14T07:02:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:18 crc kubenswrapper[4870]: I1014 07:02:18.305349 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:18 crc kubenswrapper[4870]: I1014 07:02:18.305406 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:18 crc kubenswrapper[4870]: I1014 07:02:18.305420 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:18 crc kubenswrapper[4870]: I1014 07:02:18.305443 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:18 crc kubenswrapper[4870]: I1014 07:02:18.305485 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:18Z","lastTransitionTime":"2025-10-14T07:02:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:18 crc kubenswrapper[4870]: I1014 07:02:18.409525 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:18 crc kubenswrapper[4870]: I1014 07:02:18.409618 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:18 crc kubenswrapper[4870]: I1014 07:02:18.409645 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:18 crc kubenswrapper[4870]: I1014 07:02:18.409685 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:18 crc kubenswrapper[4870]: I1014 07:02:18.409709 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:18Z","lastTransitionTime":"2025-10-14T07:02:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:18 crc kubenswrapper[4870]: I1014 07:02:18.512562 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:18 crc kubenswrapper[4870]: I1014 07:02:18.512614 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:18 crc kubenswrapper[4870]: I1014 07:02:18.512628 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:18 crc kubenswrapper[4870]: I1014 07:02:18.512656 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:18 crc kubenswrapper[4870]: I1014 07:02:18.512673 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:18Z","lastTransitionTime":"2025-10-14T07:02:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:18 crc kubenswrapper[4870]: I1014 07:02:18.616041 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:18 crc kubenswrapper[4870]: I1014 07:02:18.616097 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:18 crc kubenswrapper[4870]: I1014 07:02:18.616108 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:18 crc kubenswrapper[4870]: I1014 07:02:18.616126 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:18 crc kubenswrapper[4870]: I1014 07:02:18.616138 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:18Z","lastTransitionTime":"2025-10-14T07:02:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:18 crc kubenswrapper[4870]: I1014 07:02:18.718885 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:18 crc kubenswrapper[4870]: I1014 07:02:18.718958 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:18 crc kubenswrapper[4870]: I1014 07:02:18.718974 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:18 crc kubenswrapper[4870]: I1014 07:02:18.719000 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:18 crc kubenswrapper[4870]: I1014 07:02:18.719017 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:18Z","lastTransitionTime":"2025-10-14T07:02:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:18 crc kubenswrapper[4870]: I1014 07:02:18.822494 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:18 crc kubenswrapper[4870]: I1014 07:02:18.822522 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:18 crc kubenswrapper[4870]: I1014 07:02:18.822531 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:18 crc kubenswrapper[4870]: I1014 07:02:18.822545 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:18 crc kubenswrapper[4870]: I1014 07:02:18.822554 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:18Z","lastTransitionTime":"2025-10-14T07:02:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:18 crc kubenswrapper[4870]: I1014 07:02:18.925975 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:18 crc kubenswrapper[4870]: I1014 07:02:18.926017 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:18 crc kubenswrapper[4870]: I1014 07:02:18.926028 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:18 crc kubenswrapper[4870]: I1014 07:02:18.926042 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:18 crc kubenswrapper[4870]: I1014 07:02:18.926053 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:18Z","lastTransitionTime":"2025-10-14T07:02:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:19 crc kubenswrapper[4870]: I1014 07:02:19.032120 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:19 crc kubenswrapper[4870]: I1014 07:02:19.032222 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:19 crc kubenswrapper[4870]: I1014 07:02:19.032239 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:19 crc kubenswrapper[4870]: I1014 07:02:19.032268 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:19 crc kubenswrapper[4870]: I1014 07:02:19.032285 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:19Z","lastTransitionTime":"2025-10-14T07:02:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:19 crc kubenswrapper[4870]: I1014 07:02:19.033034 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zc5j6" Oct 14 07:02:19 crc kubenswrapper[4870]: E1014 07:02:19.033213 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zc5j6" podUID="7d7a4f41-388a-4320-b896-43c17ff13da6" Oct 14 07:02:19 crc kubenswrapper[4870]: I1014 07:02:19.134888 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:19 crc kubenswrapper[4870]: I1014 07:02:19.134937 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:19 crc kubenswrapper[4870]: I1014 07:02:19.134950 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:19 crc kubenswrapper[4870]: I1014 07:02:19.134968 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:19 crc kubenswrapper[4870]: I1014 07:02:19.134984 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:19Z","lastTransitionTime":"2025-10-14T07:02:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:19 crc kubenswrapper[4870]: I1014 07:02:19.237881 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:19 crc kubenswrapper[4870]: I1014 07:02:19.237925 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:19 crc kubenswrapper[4870]: I1014 07:02:19.237941 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:19 crc kubenswrapper[4870]: I1014 07:02:19.238030 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:19 crc kubenswrapper[4870]: I1014 07:02:19.238047 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:19Z","lastTransitionTime":"2025-10-14T07:02:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:19 crc kubenswrapper[4870]: I1014 07:02:19.341119 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:19 crc kubenswrapper[4870]: I1014 07:02:19.341166 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:19 crc kubenswrapper[4870]: I1014 07:02:19.341180 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:19 crc kubenswrapper[4870]: I1014 07:02:19.341199 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:19 crc kubenswrapper[4870]: I1014 07:02:19.341212 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:19Z","lastTransitionTime":"2025-10-14T07:02:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:19 crc kubenswrapper[4870]: I1014 07:02:19.443288 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:19 crc kubenswrapper[4870]: I1014 07:02:19.443341 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:19 crc kubenswrapper[4870]: I1014 07:02:19.443353 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:19 crc kubenswrapper[4870]: I1014 07:02:19.443368 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:19 crc kubenswrapper[4870]: I1014 07:02:19.443379 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:19Z","lastTransitionTime":"2025-10-14T07:02:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:19 crc kubenswrapper[4870]: I1014 07:02:19.546500 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:19 crc kubenswrapper[4870]: I1014 07:02:19.546603 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:19 crc kubenswrapper[4870]: I1014 07:02:19.546625 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:19 crc kubenswrapper[4870]: I1014 07:02:19.546656 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:19 crc kubenswrapper[4870]: I1014 07:02:19.546682 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:19Z","lastTransitionTime":"2025-10-14T07:02:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:19 crc kubenswrapper[4870]: I1014 07:02:19.650144 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:19 crc kubenswrapper[4870]: I1014 07:02:19.650190 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:19 crc kubenswrapper[4870]: I1014 07:02:19.650202 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:19 crc kubenswrapper[4870]: I1014 07:02:19.650218 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:19 crc kubenswrapper[4870]: I1014 07:02:19.650229 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:19Z","lastTransitionTime":"2025-10-14T07:02:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:19 crc kubenswrapper[4870]: I1014 07:02:19.753647 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:19 crc kubenswrapper[4870]: I1014 07:02:19.753691 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:19 crc kubenswrapper[4870]: I1014 07:02:19.753704 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:19 crc kubenswrapper[4870]: I1014 07:02:19.753727 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:19 crc kubenswrapper[4870]: I1014 07:02:19.753741 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:19Z","lastTransitionTime":"2025-10-14T07:02:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:19 crc kubenswrapper[4870]: I1014 07:02:19.856050 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:19 crc kubenswrapper[4870]: I1014 07:02:19.856123 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:19 crc kubenswrapper[4870]: I1014 07:02:19.856143 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:19 crc kubenswrapper[4870]: I1014 07:02:19.856170 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:19 crc kubenswrapper[4870]: I1014 07:02:19.856188 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:19Z","lastTransitionTime":"2025-10-14T07:02:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:19 crc kubenswrapper[4870]: I1014 07:02:19.959256 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:19 crc kubenswrapper[4870]: I1014 07:02:19.959323 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:19 crc kubenswrapper[4870]: I1014 07:02:19.959342 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:19 crc kubenswrapper[4870]: I1014 07:02:19.959371 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:19 crc kubenswrapper[4870]: I1014 07:02:19.959391 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:19Z","lastTransitionTime":"2025-10-14T07:02:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:20 crc kubenswrapper[4870]: I1014 07:02:20.033162 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 07:02:20 crc kubenswrapper[4870]: I1014 07:02:20.033197 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 07:02:20 crc kubenswrapper[4870]: I1014 07:02:20.033249 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 07:02:20 crc kubenswrapper[4870]: E1014 07:02:20.033339 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 07:02:20 crc kubenswrapper[4870]: E1014 07:02:20.033394 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 07:02:20 crc kubenswrapper[4870]: E1014 07:02:20.033476 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 07:02:20 crc kubenswrapper[4870]: I1014 07:02:20.061315 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:20 crc kubenswrapper[4870]: I1014 07:02:20.061559 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:20 crc kubenswrapper[4870]: I1014 07:02:20.061571 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:20 crc kubenswrapper[4870]: I1014 07:02:20.061588 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:20 crc kubenswrapper[4870]: I1014 07:02:20.061599 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:20Z","lastTransitionTime":"2025-10-14T07:02:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:20 crc kubenswrapper[4870]: I1014 07:02:20.164847 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:20 crc kubenswrapper[4870]: I1014 07:02:20.164896 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:20 crc kubenswrapper[4870]: I1014 07:02:20.164912 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:20 crc kubenswrapper[4870]: I1014 07:02:20.164933 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:20 crc kubenswrapper[4870]: I1014 07:02:20.164944 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:20Z","lastTransitionTime":"2025-10-14T07:02:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:20 crc kubenswrapper[4870]: I1014 07:02:20.268248 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:20 crc kubenswrapper[4870]: I1014 07:02:20.268293 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:20 crc kubenswrapper[4870]: I1014 07:02:20.268304 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:20 crc kubenswrapper[4870]: I1014 07:02:20.268323 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:20 crc kubenswrapper[4870]: I1014 07:02:20.268334 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:20Z","lastTransitionTime":"2025-10-14T07:02:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:20 crc kubenswrapper[4870]: I1014 07:02:20.370835 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:20 crc kubenswrapper[4870]: I1014 07:02:20.370920 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:20 crc kubenswrapper[4870]: I1014 07:02:20.370930 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:20 crc kubenswrapper[4870]: I1014 07:02:20.370943 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:20 crc kubenswrapper[4870]: I1014 07:02:20.370952 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:20Z","lastTransitionTime":"2025-10-14T07:02:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:20 crc kubenswrapper[4870]: I1014 07:02:20.473529 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:20 crc kubenswrapper[4870]: I1014 07:02:20.473618 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:20 crc kubenswrapper[4870]: I1014 07:02:20.473642 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:20 crc kubenswrapper[4870]: I1014 07:02:20.473676 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:20 crc kubenswrapper[4870]: I1014 07:02:20.473696 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:20Z","lastTransitionTime":"2025-10-14T07:02:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:20 crc kubenswrapper[4870]: I1014 07:02:20.576803 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:20 crc kubenswrapper[4870]: I1014 07:02:20.576835 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:20 crc kubenswrapper[4870]: I1014 07:02:20.576846 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:20 crc kubenswrapper[4870]: I1014 07:02:20.576859 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:20 crc kubenswrapper[4870]: I1014 07:02:20.576867 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:20Z","lastTransitionTime":"2025-10-14T07:02:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:20 crc kubenswrapper[4870]: I1014 07:02:20.678733 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:20 crc kubenswrapper[4870]: I1014 07:02:20.678770 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:20 crc kubenswrapper[4870]: I1014 07:02:20.678778 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:20 crc kubenswrapper[4870]: I1014 07:02:20.678791 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:20 crc kubenswrapper[4870]: I1014 07:02:20.678800 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:20Z","lastTransitionTime":"2025-10-14T07:02:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:20 crc kubenswrapper[4870]: I1014 07:02:20.781009 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:20 crc kubenswrapper[4870]: I1014 07:02:20.781053 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:20 crc kubenswrapper[4870]: I1014 07:02:20.781064 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:20 crc kubenswrapper[4870]: I1014 07:02:20.781079 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:20 crc kubenswrapper[4870]: I1014 07:02:20.781090 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:20Z","lastTransitionTime":"2025-10-14T07:02:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:20 crc kubenswrapper[4870]: I1014 07:02:20.888855 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:20 crc kubenswrapper[4870]: I1014 07:02:20.888952 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:20 crc kubenswrapper[4870]: I1014 07:02:20.888984 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:20 crc kubenswrapper[4870]: I1014 07:02:20.889020 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:20 crc kubenswrapper[4870]: I1014 07:02:20.889040 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:20Z","lastTransitionTime":"2025-10-14T07:02:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:20 crc kubenswrapper[4870]: I1014 07:02:20.993663 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:20 crc kubenswrapper[4870]: I1014 07:02:20.993741 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:20 crc kubenswrapper[4870]: I1014 07:02:20.993758 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:20 crc kubenswrapper[4870]: I1014 07:02:20.993783 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:20 crc kubenswrapper[4870]: I1014 07:02:20.993799 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:20Z","lastTransitionTime":"2025-10-14T07:02:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:21 crc kubenswrapper[4870]: I1014 07:02:21.033385 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zc5j6" Oct 14 07:02:21 crc kubenswrapper[4870]: E1014 07:02:21.033676 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zc5j6" podUID="7d7a4f41-388a-4320-b896-43c17ff13da6" Oct 14 07:02:21 crc kubenswrapper[4870]: I1014 07:02:21.096682 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:21 crc kubenswrapper[4870]: I1014 07:02:21.096718 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:21 crc kubenswrapper[4870]: I1014 07:02:21.096728 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:21 crc kubenswrapper[4870]: I1014 07:02:21.096743 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:21 crc kubenswrapper[4870]: I1014 07:02:21.096753 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:21Z","lastTransitionTime":"2025-10-14T07:02:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:21 crc kubenswrapper[4870]: I1014 07:02:21.201975 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:21 crc kubenswrapper[4870]: I1014 07:02:21.202077 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:21 crc kubenswrapper[4870]: I1014 07:02:21.202106 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:21 crc kubenswrapper[4870]: I1014 07:02:21.202201 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:21 crc kubenswrapper[4870]: I1014 07:02:21.202354 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:21Z","lastTransitionTime":"2025-10-14T07:02:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:21 crc kubenswrapper[4870]: I1014 07:02:21.306922 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:21 crc kubenswrapper[4870]: I1014 07:02:21.306988 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:21 crc kubenswrapper[4870]: I1014 07:02:21.306999 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:21 crc kubenswrapper[4870]: I1014 07:02:21.307022 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:21 crc kubenswrapper[4870]: I1014 07:02:21.307035 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:21Z","lastTransitionTime":"2025-10-14T07:02:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:21 crc kubenswrapper[4870]: I1014 07:02:21.411064 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:21 crc kubenswrapper[4870]: I1014 07:02:21.411139 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:21 crc kubenswrapper[4870]: I1014 07:02:21.411208 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:21 crc kubenswrapper[4870]: I1014 07:02:21.411248 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:21 crc kubenswrapper[4870]: I1014 07:02:21.411281 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:21Z","lastTransitionTime":"2025-10-14T07:02:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:21 crc kubenswrapper[4870]: I1014 07:02:21.514779 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:21 crc kubenswrapper[4870]: I1014 07:02:21.515296 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:21 crc kubenswrapper[4870]: I1014 07:02:21.515609 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:21 crc kubenswrapper[4870]: I1014 07:02:21.515880 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:21 crc kubenswrapper[4870]: I1014 07:02:21.516167 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:21Z","lastTransitionTime":"2025-10-14T07:02:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:21 crc kubenswrapper[4870]: I1014 07:02:21.619202 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:21 crc kubenswrapper[4870]: I1014 07:02:21.619264 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:21 crc kubenswrapper[4870]: I1014 07:02:21.619279 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:21 crc kubenswrapper[4870]: I1014 07:02:21.619305 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:21 crc kubenswrapper[4870]: I1014 07:02:21.619320 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:21Z","lastTransitionTime":"2025-10-14T07:02:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:21 crc kubenswrapper[4870]: I1014 07:02:21.723268 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:21 crc kubenswrapper[4870]: I1014 07:02:21.723332 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:21 crc kubenswrapper[4870]: I1014 07:02:21.723351 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:21 crc kubenswrapper[4870]: I1014 07:02:21.723379 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:21 crc kubenswrapper[4870]: I1014 07:02:21.723398 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:21Z","lastTransitionTime":"2025-10-14T07:02:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:21 crc kubenswrapper[4870]: I1014 07:02:21.826684 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:21 crc kubenswrapper[4870]: I1014 07:02:21.826765 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:21 crc kubenswrapper[4870]: I1014 07:02:21.826782 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:21 crc kubenswrapper[4870]: I1014 07:02:21.826812 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:21 crc kubenswrapper[4870]: I1014 07:02:21.826826 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:21Z","lastTransitionTime":"2025-10-14T07:02:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:21 crc kubenswrapper[4870]: I1014 07:02:21.929722 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:21 crc kubenswrapper[4870]: I1014 07:02:21.929791 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:21 crc kubenswrapper[4870]: I1014 07:02:21.929804 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:21 crc kubenswrapper[4870]: I1014 07:02:21.929824 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:21 crc kubenswrapper[4870]: I1014 07:02:21.929839 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:21Z","lastTransitionTime":"2025-10-14T07:02:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:22 crc kubenswrapper[4870]: I1014 07:02:22.033185 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 07:02:22 crc kubenswrapper[4870]: I1014 07:02:22.033328 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 07:02:22 crc kubenswrapper[4870]: I1014 07:02:22.033427 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 07:02:22 crc kubenswrapper[4870]: E1014 07:02:22.033647 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 07:02:22 crc kubenswrapper[4870]: I1014 07:02:22.033814 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:22 crc kubenswrapper[4870]: E1014 07:02:22.033834 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 07:02:22 crc kubenswrapper[4870]: I1014 07:02:22.033866 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:22 crc kubenswrapper[4870]: I1014 07:02:22.033884 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:22 crc kubenswrapper[4870]: I1014 07:02:22.033938 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:22 crc kubenswrapper[4870]: I1014 07:02:22.033955 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:22Z","lastTransitionTime":"2025-10-14T07:02:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:22 crc kubenswrapper[4870]: E1014 07:02:22.033992 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 07:02:22 crc kubenswrapper[4870]: I1014 07:02:22.137709 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:22 crc kubenswrapper[4870]: I1014 07:02:22.137794 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:22 crc kubenswrapper[4870]: I1014 07:02:22.137817 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:22 crc kubenswrapper[4870]: I1014 07:02:22.137851 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:22 crc kubenswrapper[4870]: I1014 07:02:22.137871 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:22Z","lastTransitionTime":"2025-10-14T07:02:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:22 crc kubenswrapper[4870]: I1014 07:02:22.240890 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:22 crc kubenswrapper[4870]: I1014 07:02:22.240928 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:22 crc kubenswrapper[4870]: I1014 07:02:22.240942 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:22 crc kubenswrapper[4870]: I1014 07:02:22.240958 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:22 crc kubenswrapper[4870]: I1014 07:02:22.240969 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:22Z","lastTransitionTime":"2025-10-14T07:02:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:22 crc kubenswrapper[4870]: I1014 07:02:22.344965 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:22 crc kubenswrapper[4870]: I1014 07:02:22.345064 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:22 crc kubenswrapper[4870]: I1014 07:02:22.345092 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:22 crc kubenswrapper[4870]: I1014 07:02:22.345135 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:22 crc kubenswrapper[4870]: I1014 07:02:22.345162 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:22Z","lastTransitionTime":"2025-10-14T07:02:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:22 crc kubenswrapper[4870]: I1014 07:02:22.448203 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:22 crc kubenswrapper[4870]: I1014 07:02:22.448251 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:22 crc kubenswrapper[4870]: I1014 07:02:22.448264 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:22 crc kubenswrapper[4870]: I1014 07:02:22.448282 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:22 crc kubenswrapper[4870]: I1014 07:02:22.448296 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:22Z","lastTransitionTime":"2025-10-14T07:02:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:22 crc kubenswrapper[4870]: I1014 07:02:22.551095 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:22 crc kubenswrapper[4870]: I1014 07:02:22.551165 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:22 crc kubenswrapper[4870]: I1014 07:02:22.551188 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:22 crc kubenswrapper[4870]: I1014 07:02:22.551217 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:22 crc kubenswrapper[4870]: I1014 07:02:22.551241 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:22Z","lastTransitionTime":"2025-10-14T07:02:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:22 crc kubenswrapper[4870]: I1014 07:02:22.653642 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:22 crc kubenswrapper[4870]: I1014 07:02:22.653676 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:22 crc kubenswrapper[4870]: I1014 07:02:22.653684 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:22 crc kubenswrapper[4870]: I1014 07:02:22.653696 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:22 crc kubenswrapper[4870]: I1014 07:02:22.653707 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:22Z","lastTransitionTime":"2025-10-14T07:02:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:22 crc kubenswrapper[4870]: I1014 07:02:22.756528 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:22 crc kubenswrapper[4870]: I1014 07:02:22.756606 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:22 crc kubenswrapper[4870]: I1014 07:02:22.756623 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:22 crc kubenswrapper[4870]: I1014 07:02:22.756655 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:22 crc kubenswrapper[4870]: I1014 07:02:22.756674 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:22Z","lastTransitionTime":"2025-10-14T07:02:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:22 crc kubenswrapper[4870]: I1014 07:02:22.859362 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:22 crc kubenswrapper[4870]: I1014 07:02:22.859404 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:22 crc kubenswrapper[4870]: I1014 07:02:22.859417 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:22 crc kubenswrapper[4870]: I1014 07:02:22.859432 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:22 crc kubenswrapper[4870]: I1014 07:02:22.859478 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:22Z","lastTransitionTime":"2025-10-14T07:02:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:22 crc kubenswrapper[4870]: I1014 07:02:22.962539 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:22 crc kubenswrapper[4870]: I1014 07:02:22.962610 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:22 crc kubenswrapper[4870]: I1014 07:02:22.962630 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:22 crc kubenswrapper[4870]: I1014 07:02:22.962655 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:22 crc kubenswrapper[4870]: I1014 07:02:22.962673 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:22Z","lastTransitionTime":"2025-10-14T07:02:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:23 crc kubenswrapper[4870]: I1014 07:02:23.033051 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zc5j6" Oct 14 07:02:23 crc kubenswrapper[4870]: E1014 07:02:23.033273 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zc5j6" podUID="7d7a4f41-388a-4320-b896-43c17ff13da6" Oct 14 07:02:23 crc kubenswrapper[4870]: I1014 07:02:23.069133 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:23 crc kubenswrapper[4870]: I1014 07:02:23.069187 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:23 crc kubenswrapper[4870]: I1014 07:02:23.069200 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:23 crc kubenswrapper[4870]: I1014 07:02:23.069217 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:23 crc kubenswrapper[4870]: I1014 07:02:23.069230 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:23Z","lastTransitionTime":"2025-10-14T07:02:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:23 crc kubenswrapper[4870]: I1014 07:02:23.171847 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:23 crc kubenswrapper[4870]: I1014 07:02:23.171893 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:23 crc kubenswrapper[4870]: I1014 07:02:23.171906 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:23 crc kubenswrapper[4870]: I1014 07:02:23.171922 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:23 crc kubenswrapper[4870]: I1014 07:02:23.171934 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:23Z","lastTransitionTime":"2025-10-14T07:02:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:23 crc kubenswrapper[4870]: I1014 07:02:23.274869 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:23 crc kubenswrapper[4870]: I1014 07:02:23.274937 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:23 crc kubenswrapper[4870]: I1014 07:02:23.274960 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:23 crc kubenswrapper[4870]: I1014 07:02:23.274990 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:23 crc kubenswrapper[4870]: I1014 07:02:23.275012 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:23Z","lastTransitionTime":"2025-10-14T07:02:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:23 crc kubenswrapper[4870]: I1014 07:02:23.376792 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:23 crc kubenswrapper[4870]: I1014 07:02:23.376853 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:23 crc kubenswrapper[4870]: I1014 07:02:23.376863 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:23 crc kubenswrapper[4870]: I1014 07:02:23.376877 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:23 crc kubenswrapper[4870]: I1014 07:02:23.376886 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:23Z","lastTransitionTime":"2025-10-14T07:02:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:23 crc kubenswrapper[4870]: I1014 07:02:23.479223 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:23 crc kubenswrapper[4870]: I1014 07:02:23.479293 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:23 crc kubenswrapper[4870]: I1014 07:02:23.479311 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:23 crc kubenswrapper[4870]: I1014 07:02:23.479334 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:23 crc kubenswrapper[4870]: I1014 07:02:23.479352 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:23Z","lastTransitionTime":"2025-10-14T07:02:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:23 crc kubenswrapper[4870]: I1014 07:02:23.583276 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:23 crc kubenswrapper[4870]: I1014 07:02:23.583317 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:23 crc kubenswrapper[4870]: I1014 07:02:23.583327 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:23 crc kubenswrapper[4870]: I1014 07:02:23.583344 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:23 crc kubenswrapper[4870]: I1014 07:02:23.583355 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:23Z","lastTransitionTime":"2025-10-14T07:02:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:23 crc kubenswrapper[4870]: I1014 07:02:23.687777 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:23 crc kubenswrapper[4870]: I1014 07:02:23.687853 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:23 crc kubenswrapper[4870]: I1014 07:02:23.687873 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:23 crc kubenswrapper[4870]: I1014 07:02:23.687901 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:23 crc kubenswrapper[4870]: I1014 07:02:23.687920 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:23Z","lastTransitionTime":"2025-10-14T07:02:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:23 crc kubenswrapper[4870]: I1014 07:02:23.790878 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:23 crc kubenswrapper[4870]: I1014 07:02:23.790954 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:23 crc kubenswrapper[4870]: I1014 07:02:23.790979 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:23 crc kubenswrapper[4870]: I1014 07:02:23.791009 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:23 crc kubenswrapper[4870]: I1014 07:02:23.791030 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:23Z","lastTransitionTime":"2025-10-14T07:02:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:23 crc kubenswrapper[4870]: I1014 07:02:23.869545 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:23 crc kubenswrapper[4870]: I1014 07:02:23.869598 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:23 crc kubenswrapper[4870]: I1014 07:02:23.869610 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:23 crc kubenswrapper[4870]: I1014 07:02:23.869629 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:23 crc kubenswrapper[4870]: I1014 07:02:23.869643 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:23Z","lastTransitionTime":"2025-10-14T07:02:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:23 crc kubenswrapper[4870]: E1014 07:02:23.889589 4870 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:02:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:02:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:02:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:02:23Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:02:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:02:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:02:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:02:23Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"56448e6b-6b3a-4bb9-91f8-2f0a9b44e9aa\\\",\\\"systemUUID\\\":\\\"0adb347d-b90d-4e9f-9962-f1bd97980aed\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:23Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:23 crc kubenswrapper[4870]: I1014 07:02:23.896891 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:23 crc kubenswrapper[4870]: I1014 07:02:23.896952 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:23 crc kubenswrapper[4870]: I1014 07:02:23.896977 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:23 crc kubenswrapper[4870]: I1014 07:02:23.897007 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:23 crc kubenswrapper[4870]: I1014 07:02:23.897029 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:23Z","lastTransitionTime":"2025-10-14T07:02:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:23 crc kubenswrapper[4870]: E1014 07:02:23.916730 4870 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:02:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:02:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:02:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:02:23Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:02:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:02:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:02:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:02:23Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"56448e6b-6b3a-4bb9-91f8-2f0a9b44e9aa\\\",\\\"systemUUID\\\":\\\"0adb347d-b90d-4e9f-9962-f1bd97980aed\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:23Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:23 crc kubenswrapper[4870]: I1014 07:02:23.923863 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:23 crc kubenswrapper[4870]: I1014 07:02:23.923983 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:23 crc kubenswrapper[4870]: I1014 07:02:23.924177 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:23 crc kubenswrapper[4870]: I1014 07:02:23.924579 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:23 crc kubenswrapper[4870]: I1014 07:02:23.924704 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:23Z","lastTransitionTime":"2025-10-14T07:02:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:23 crc kubenswrapper[4870]: E1014 07:02:23.948941 4870 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:02:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:02:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:02:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:02:23Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:02:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:02:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:02:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:02:23Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"56448e6b-6b3a-4bb9-91f8-2f0a9b44e9aa\\\",\\\"systemUUID\\\":\\\"0adb347d-b90d-4e9f-9962-f1bd97980aed\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:23Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:23 crc kubenswrapper[4870]: I1014 07:02:23.954733 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:23 crc kubenswrapper[4870]: I1014 07:02:23.954786 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:23 crc kubenswrapper[4870]: I1014 07:02:23.954797 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:23 crc kubenswrapper[4870]: I1014 07:02:23.954813 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:23 crc kubenswrapper[4870]: I1014 07:02:23.954821 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:23Z","lastTransitionTime":"2025-10-14T07:02:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:23 crc kubenswrapper[4870]: E1014 07:02:23.975695 4870 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:02:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:02:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:02:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:02:23Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:02:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:02:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:02:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:02:23Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"56448e6b-6b3a-4bb9-91f8-2f0a9b44e9aa\\\",\\\"systemUUID\\\":\\\"0adb347d-b90d-4e9f-9962-f1bd97980aed\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:23Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:23 crc kubenswrapper[4870]: I1014 07:02:23.980480 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:23 crc kubenswrapper[4870]: I1014 07:02:23.980545 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:23 crc kubenswrapper[4870]: I1014 07:02:23.980566 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:23 crc kubenswrapper[4870]: I1014 07:02:23.980591 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:23 crc kubenswrapper[4870]: I1014 07:02:23.980609 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:23Z","lastTransitionTime":"2025-10-14T07:02:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:24 crc kubenswrapper[4870]: E1014 07:02:24.000760 4870 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:02:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:02:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:02:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:02:23Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:02:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:02:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:02:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:02:23Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"56448e6b-6b3a-4bb9-91f8-2f0a9b44e9aa\\\",\\\"systemUUID\\\":\\\"0adb347d-b90d-4e9f-9962-f1bd97980aed\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:23Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:24 crc kubenswrapper[4870]: E1014 07:02:24.001144 4870 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 14 07:02:24 crc kubenswrapper[4870]: I1014 07:02:24.003682 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:24 crc kubenswrapper[4870]: I1014 07:02:24.003738 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:24 crc kubenswrapper[4870]: I1014 07:02:24.003764 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:24 crc kubenswrapper[4870]: I1014 07:02:24.003795 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:24 crc kubenswrapper[4870]: I1014 07:02:24.003822 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:24Z","lastTransitionTime":"2025-10-14T07:02:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:24 crc kubenswrapper[4870]: I1014 07:02:24.033314 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 07:02:24 crc kubenswrapper[4870]: I1014 07:02:24.033367 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 07:02:24 crc kubenswrapper[4870]: I1014 07:02:24.033331 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 07:02:24 crc kubenswrapper[4870]: E1014 07:02:24.033650 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 07:02:24 crc kubenswrapper[4870]: E1014 07:02:24.033777 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 07:02:24 crc kubenswrapper[4870]: E1014 07:02:24.033879 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 07:02:24 crc kubenswrapper[4870]: I1014 07:02:24.106788 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:24 crc kubenswrapper[4870]: I1014 07:02:24.107173 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:24 crc kubenswrapper[4870]: I1014 07:02:24.107235 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:24 crc kubenswrapper[4870]: I1014 07:02:24.107302 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:24 crc kubenswrapper[4870]: I1014 07:02:24.107377 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:24Z","lastTransitionTime":"2025-10-14T07:02:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:24 crc kubenswrapper[4870]: I1014 07:02:24.209557 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:24 crc kubenswrapper[4870]: I1014 07:02:24.209896 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:24 crc kubenswrapper[4870]: I1014 07:02:24.209989 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:24 crc kubenswrapper[4870]: I1014 07:02:24.210129 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:24 crc kubenswrapper[4870]: I1014 07:02:24.210267 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:24Z","lastTransitionTime":"2025-10-14T07:02:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:24 crc kubenswrapper[4870]: I1014 07:02:24.313303 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:24 crc kubenswrapper[4870]: I1014 07:02:24.313613 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:24 crc kubenswrapper[4870]: I1014 07:02:24.313754 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:24 crc kubenswrapper[4870]: I1014 07:02:24.313853 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:24 crc kubenswrapper[4870]: I1014 07:02:24.313948 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:24Z","lastTransitionTime":"2025-10-14T07:02:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:24 crc kubenswrapper[4870]: I1014 07:02:24.417102 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:24 crc kubenswrapper[4870]: I1014 07:02:24.417356 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:24 crc kubenswrapper[4870]: I1014 07:02:24.417428 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:24 crc kubenswrapper[4870]: I1014 07:02:24.417564 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:24 crc kubenswrapper[4870]: I1014 07:02:24.417631 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:24Z","lastTransitionTime":"2025-10-14T07:02:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:24 crc kubenswrapper[4870]: I1014 07:02:24.520418 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:24 crc kubenswrapper[4870]: I1014 07:02:24.520490 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:24 crc kubenswrapper[4870]: I1014 07:02:24.520504 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:24 crc kubenswrapper[4870]: I1014 07:02:24.520522 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:24 crc kubenswrapper[4870]: I1014 07:02:24.520536 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:24Z","lastTransitionTime":"2025-10-14T07:02:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:24 crc kubenswrapper[4870]: I1014 07:02:24.623334 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:24 crc kubenswrapper[4870]: I1014 07:02:24.623398 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:24 crc kubenswrapper[4870]: I1014 07:02:24.623416 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:24 crc kubenswrapper[4870]: I1014 07:02:24.623472 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:24 crc kubenswrapper[4870]: I1014 07:02:24.623491 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:24Z","lastTransitionTime":"2025-10-14T07:02:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:24 crc kubenswrapper[4870]: I1014 07:02:24.726748 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:24 crc kubenswrapper[4870]: I1014 07:02:24.726813 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:24 crc kubenswrapper[4870]: I1014 07:02:24.726830 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:24 crc kubenswrapper[4870]: I1014 07:02:24.726847 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:24 crc kubenswrapper[4870]: I1014 07:02:24.726860 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:24Z","lastTransitionTime":"2025-10-14T07:02:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:24 crc kubenswrapper[4870]: I1014 07:02:24.830815 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:24 crc kubenswrapper[4870]: I1014 07:02:24.830889 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:24 crc kubenswrapper[4870]: I1014 07:02:24.830909 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:24 crc kubenswrapper[4870]: I1014 07:02:24.830940 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:24 crc kubenswrapper[4870]: I1014 07:02:24.830959 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:24Z","lastTransitionTime":"2025-10-14T07:02:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:24 crc kubenswrapper[4870]: I1014 07:02:24.934513 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:24 crc kubenswrapper[4870]: I1014 07:02:24.934576 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:24 crc kubenswrapper[4870]: I1014 07:02:24.934590 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:24 crc kubenswrapper[4870]: I1014 07:02:24.934628 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:24 crc kubenswrapper[4870]: I1014 07:02:24.934640 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:24Z","lastTransitionTime":"2025-10-14T07:02:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:25 crc kubenswrapper[4870]: I1014 07:02:25.032942 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zc5j6" Oct 14 07:02:25 crc kubenswrapper[4870]: E1014 07:02:25.033084 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zc5j6" podUID="7d7a4f41-388a-4320-b896-43c17ff13da6" Oct 14 07:02:25 crc kubenswrapper[4870]: I1014 07:02:25.037477 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:25 crc kubenswrapper[4870]: I1014 07:02:25.037661 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:25 crc kubenswrapper[4870]: I1014 07:02:25.037726 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:25 crc kubenswrapper[4870]: I1014 07:02:25.037796 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:25 crc kubenswrapper[4870]: I1014 07:02:25.037865 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:25Z","lastTransitionTime":"2025-10-14T07:02:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:25 crc kubenswrapper[4870]: I1014 07:02:25.053570 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77a0d93a-5391-4112-b502-c474f45de209\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7815af9a584f6db6660073da289f2e8c987a04ef9ddfac155bdd9a15b014fcfa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b329b570622155a55f70501b2260e278106a68ebef497914231b4ec860e0125\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba942a0e6fe5ab93a187e0b7b4bd7c440c2e7827d9b5323fc50f90b562075392\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b9018b078c5c8276243c1696860f0aa73e2d8dd917913ae9d03b30dc9b5113a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://978e8db184ba9e430232c6bb2ee530beb92567fdc71ba3c35b0bedca3492d845\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"file observer\\\\nW1014 07:01:25.870672 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1014 07:01:25.870795 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1014 07:01:25.871521 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-756714914/tls.crt::/tmp/serving-cert-756714914/tls.key\\\\\\\"\\\\nI1014 07:01:26.460037 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1014 07:01:26.464876 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1014 07:01:26.464902 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1014 07:01:26.464925 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1014 07:01:26.464931 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1014 07:01:26.481374 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1014 07:01:26.481400 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:01:26.481405 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:01:26.481410 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1014 07:01:26.481414 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1014 07:01:26.481417 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1014 07:01:26.481420 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1014 07:01:26.481610 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1014 07:01:26.482872 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bd0044a247bdc5ca4cab7294103af3270b76cde61f3109b85e83ed050ca53b6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82d3c2214e0be679d2827594b159ace362151318689078bc49d3158f42899636\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82d3c2214e0be679d2827594b159ace362151318689078bc49d3158f42899636\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:05Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:25Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:25 crc kubenswrapper[4870]: I1014 07:02:25.072107 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:25Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:25 crc kubenswrapper[4870]: I1014 07:02:25.086358 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e26fe49ec2045d20a0cb2dd17e96e79b6eb43c641eb184c688cc472ca5db39d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aea115f7ee71c700b218fe15c928af7b55e6babe70e65d67f125d87be1fd2e2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:25Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:25 crc kubenswrapper[4870]: I1014 07:02:25.106942 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-v7br2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dcaef0e-42e3-424b-bbe5-4a023efe912d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71dec55e859b11a081dbb90d7d6cf6b4846cf591382303d8329003ff50c71f2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e72cff0723e2ce520340244b4e597d8016c58e1b071c4b087f2f6cd718741594\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc9d4db944f837289d1923176d60cfc4a566a282c2d2e3a703e220df884703c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d16ed1c65925549dcb69ab02c3013e25e3f123a83179f1af3a8e08bf91a59896\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed59243b59c39ce82bf9bc8f33c0a8b84854846afff6a89f5d37ffc97a096b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3b6d20cf459b1f6d3aff5b9290f9391909baf27ee40c49e2b0cd2d5cfe889ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fc9bfcb356b506d6c439adc22e4caaed4fad1c5a085744d44ebd50cf8898777\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3fc9bfcb356b506d6c439adc22e4caaed4fad1c5a085744d44ebd50cf8898777\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-14T07:01:56Z\\\",\\\"message\\\":\\\" 6555 ovn.go:134] Ensuring zone local for Pod openshift-image-registry/node-ca-kz77f in node crc\\\\nI1014 07:01:56.821005 6555 obj_retry.go:386] Retry successful for *v1.Pod openshift-image-registry/node-ca-kz77f after 0 failed attempt(s)\\\\nI1014 07:01:56.820992 6555 model_client.go:382] Update operations generated as: [{Op:update Table:Logical_Switch_Port Row:map[addresses:{GoSet:[0a:58:0a:d9:00:5c 10.217.0.92]} options:{GoMap:map[iface-id-ver:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 requested-chassis:crc]} port_security:{GoSet:[0a:58:0a:d9:00:5c 10.217.0.92]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c94130be-172c-477c-88c4-40cc7eba30fe}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1014 07:01:56.821025 6555 default_network_controller.go:776] Recording success event on pod openshift-image-registry/node-ca-kz77f\\\\nI1014 07:01:56.820931 6555 default_network_controller.go:776] Recording success event on pod openshift-dns/node-resolver-4v2v2\\\\nI1014 07:01:56.820614 6555 services_controller.go:454] Service openshift-marketplace/redhat-marketplace for network=default has 1 cluster-wide, 0 per-node configs, 0 template configs, making 1 (cluster) 0 (per node) and 0 (template) load balancers\\\\nI1014 07:01:56.821037 6555 model_client.go:382] Update operations generated as: [{Op:update Table:N\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:56Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-v7br2_openshift-ovn-kubernetes(8dcaef0e-42e3-424b-bbe5-4a023efe912d)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fe3468cdd612fc4378e1f6f55a2a9d00274f7ddb65cb57b2a368dfc065cfb4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ba24c2e067b9f3a14d9d018b66300b54d866530d30174fbdb3c899299575d28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ba24c2e067b9f3a14d9d018b66300b54d866530d30174fbdb3c899299575d28\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-v7br2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:25Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:25 crc kubenswrapper[4870]: I1014 07:02:25.127004 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33a9493e-9660-48db-b8d1-310e29ebe313\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3570643ef1b7d5cc3a542f320413b819338e0c8b7ab7f0f09cc909997a39bab8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb398668197dedc1c2a14b4ae7d6a29714f3527309ccb9c1622af841100e1d60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1d2b6d068d63bd38cdfb83824525f6d12054f6cabab5d684e03ea5f2152076c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://783ff243e395333f3fc80cfae21358c532f0c9c5cb5515586be56b90ed10d912\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a8135a3d01076e4437ae8a784b459bf8d16de57fa16ec7ce4d40c5129c9d4a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48ca3a5017ba650df99b2fc7201c3fe5abfa89d81089c5647cde1b1549f534fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48ca3a5017ba650df99b2fc7201c3fe5abfa89d81089c5647cde1b1549f534fc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://19682eca67f76bf2e06e1b80f12e21141fcbce0acb8195f0defacd1b9a10b5e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://19682eca67f76bf2e06e1b80f12e21141fcbce0acb8195f0defacd1b9a10b5e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://042aad6d5f6bee8eef77f9e6a2c265de775a69f3da72aedc68bdf312198f00fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://042aad6d5f6bee8eef77f9e6a2c265de775a69f3da72aedc68bdf312198f00fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:05Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:25Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:25 crc kubenswrapper[4870]: I1014 07:02:25.139464 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:25 crc kubenswrapper[4870]: I1014 07:02:25.139848 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:25 crc kubenswrapper[4870]: I1014 07:02:25.139921 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:25 crc kubenswrapper[4870]: I1014 07:02:25.139994 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:25 crc kubenswrapper[4870]: I1014 07:02:25.140059 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:25Z","lastTransitionTime":"2025-10-14T07:02:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:25 crc kubenswrapper[4870]: I1014 07:02:25.143213 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a77729fb-a187-48c6-80db-bee90f7bdac4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:02:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:02:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c859ebc0d20b632d1d5a716caee343c7e6b9b752ed5f257422cd35ad12c170f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e8a7a5c1f1706ebbd2e8165d8961002f5a9340e2749e377b0058db107ad7dbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f75b3dc1557917e9c0c35c2810338a076ed92ba89963edafb9d537ad19c2bd8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b192bd3e0eb8540d7492abc207025471bc9e198e7e78b50215c8d2f5bc57cf4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b192bd3e0eb8540d7492abc207025471bc9e198e7e78b50215c8d2f5bc57cf4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:05Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:25Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:25 crc kubenswrapper[4870]: I1014 07:02:25.156688 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-4v2v2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4796ba6-c23b-4764-be96-2e3a3bf5e962\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf5afb0b99a7101f6d5573fb7eb925c5a1d8becc11b2e3c6f3351d80b6f0e823\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnp8b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-4v2v2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:25Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:25 crc kubenswrapper[4870]: I1014 07:02:25.174639 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-cdjmj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"612e7ab9-bda4-45d8-bc50-ff1afb68aace\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7776352d62b814fc63d01e4c16ad13f878b0597eb37a370bf9da0a88abab6cfa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://834444b5350ee12df4e4652c3009880d8d69e1edf8ba0c0156ab96fb2bbe28e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://834444b5350ee12df4e4652c3009880d8d69e1edf8ba0c0156ab96fb2bbe28e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c1972e7b2f75790db22a07a078eee97072bffbdb7f8e594bf2be5f800632a7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c1972e7b2f75790db22a07a078eee97072bffbdb7f8e594bf2be5f800632a7b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b124b42a44c6a23da1d92b9a5326987eb74430f80489a71e8913829475654ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b124b42a44c6a23da1d92b9a5326987eb74430f80489a71e8913829475654ab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62d0deb6b20ea65fbcbaf94c2fc14d69adb22e7a46cc043ccd4def2668ddd4b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62d0deb6b20ea65fbcbaf94c2fc14d69adb22e7a46cc043ccd4def2668ddd4b7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7f0d7cf29ca737d0effe58731a8b890379527c97f78c144d4eebb910260fadb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7f0d7cf29ca737d0effe58731a8b890379527c97f78c144d4eebb910260fadb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5959abb7c8e7ab32bed3aca709ae0334e8400f991ba724246c9fda56cbed822\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5959abb7c8e7ab32bed3aca709ae0334e8400f991ba724246c9fda56cbed822\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-cdjmj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:25Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:25 crc kubenswrapper[4870]: I1014 07:02:25.194507 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fdxjh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a3a430a-b540-4a8c-adad-b2893dbf9898\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:02:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:02:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88614373bfe4779bbe61cc41cc4c35f419e438bdae3cec4964271bf0f74ad3f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://93e41d5e927eed7ec62eb9c3607cc804cd8999556663363fd123b2a784b00ff1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-14T07:02:13Z\\\",\\\"message\\\":\\\"2025-10-14T07:01:28+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_847000ef-6815-4615-8665-86c348661ad2\\\\n2025-10-14T07:01:28+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_847000ef-6815-4615-8665-86c348661ad2 to /host/opt/cni/bin/\\\\n2025-10-14T07:01:28Z [verbose] multus-daemon started\\\\n2025-10-14T07:01:28Z [verbose] Readiness Indicator file check\\\\n2025-10-14T07:02:13Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:02:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k787j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fdxjh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:25Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:25 crc kubenswrapper[4870]: I1014 07:02:25.209853 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-khtq8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28deeea6-a1ef-4a4a-a377-d954ccd87de9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://296bb048a1d99649beb57d609573833bf547cd4d628ebc823c6af83fe5736c9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rb79p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2b7bc4c93307e80b778f2282ced993898f1c4cfc350a7d74828b8f9dea9906d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rb79p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-khtq8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:25Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:25 crc kubenswrapper[4870]: I1014 07:02:25.226848 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bcbc6dd6-aab0-4dbc-a62e-c58f06ec1337\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25ec4808b09adc44488530b3c4e29fd7fec46b3fcfa31d4449ec4a6dbc880ec8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f3f5de86fb11231a19034a18e9d7f31476c162e6567d0c136ead8bbbd900d6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d57d938d9d9c92198ae5b3857536ece36012fbfbe8cc8cfd245d1386d39b0be5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77a3247a7f65709543d7fd374b2ab69d7d793537d4132d7e22261188d6e58544\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:05Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:25Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:25 crc kubenswrapper[4870]: I1014 07:02:25.243694 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:25 crc kubenswrapper[4870]: I1014 07:02:25.243729 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:25 crc kubenswrapper[4870]: I1014 07:02:25.243739 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:25 crc kubenswrapper[4870]: I1014 07:02:25.243754 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:25 crc kubenswrapper[4870]: I1014 07:02:25.243765 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:25Z","lastTransitionTime":"2025-10-14T07:02:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:25 crc kubenswrapper[4870]: I1014 07:02:25.244775 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e539fac3d0a2ade314a4b4c1050bb27c37bcff5ba5272fa2d71b97ab5c0d7191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:25Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:25 crc kubenswrapper[4870]: I1014 07:02:25.261005 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:25Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:25 crc kubenswrapper[4870]: I1014 07:02:25.272885 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-kz77f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d3864a91-9bf6-4540-a908-cc41aa39656c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d4ded3d86030f5ea548c8b1cc3caddbe14a45c9a3ae4af836e892880584f57d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4h8v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:28Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-kz77f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:25Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:25 crc kubenswrapper[4870]: I1014 07:02:25.288025 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:25Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:25 crc kubenswrapper[4870]: I1014 07:02:25.301045 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3eb0b397c05ffd6650607f1be4a75683e0342d3333fc096bf1cd63c7688834f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:25Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:25 crc kubenswrapper[4870]: I1014 07:02:25.314887 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"874a5e45-dffd-4d17-b609-b7d3ed2eab07\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d98d044836ab44a405196915ef7bf572b75113f2c81ec7e64effa63b47ad3543\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wj4ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8004752575df4a1c44efbcafe64799f09126d0f97285d92388c76a0b90e022f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wj4ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7tvc7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:25Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:25 crc kubenswrapper[4870]: I1014 07:02:25.328103 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zc5j6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d7a4f41-388a-4320-b896-43c17ff13da6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6p2wh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6p2wh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:40Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zc5j6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:25Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:25 crc kubenswrapper[4870]: I1014 07:02:25.346630 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:25 crc kubenswrapper[4870]: I1014 07:02:25.346679 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:25 crc kubenswrapper[4870]: I1014 07:02:25.346698 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:25 crc kubenswrapper[4870]: I1014 07:02:25.346720 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:25 crc kubenswrapper[4870]: I1014 07:02:25.346738 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:25Z","lastTransitionTime":"2025-10-14T07:02:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:25 crc kubenswrapper[4870]: I1014 07:02:25.450843 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:25 crc kubenswrapper[4870]: I1014 07:02:25.450906 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:25 crc kubenswrapper[4870]: I1014 07:02:25.450919 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:25 crc kubenswrapper[4870]: I1014 07:02:25.450942 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:25 crc kubenswrapper[4870]: I1014 07:02:25.450955 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:25Z","lastTransitionTime":"2025-10-14T07:02:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:25 crc kubenswrapper[4870]: I1014 07:02:25.554115 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:25 crc kubenswrapper[4870]: I1014 07:02:25.554214 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:25 crc kubenswrapper[4870]: I1014 07:02:25.554236 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:25 crc kubenswrapper[4870]: I1014 07:02:25.554261 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:25 crc kubenswrapper[4870]: I1014 07:02:25.554279 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:25Z","lastTransitionTime":"2025-10-14T07:02:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:25 crc kubenswrapper[4870]: I1014 07:02:25.657338 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:25 crc kubenswrapper[4870]: I1014 07:02:25.657409 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:25 crc kubenswrapper[4870]: I1014 07:02:25.657433 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:25 crc kubenswrapper[4870]: I1014 07:02:25.657502 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:25 crc kubenswrapper[4870]: I1014 07:02:25.657526 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:25Z","lastTransitionTime":"2025-10-14T07:02:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:25 crc kubenswrapper[4870]: I1014 07:02:25.760881 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:25 crc kubenswrapper[4870]: I1014 07:02:25.760934 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:25 crc kubenswrapper[4870]: I1014 07:02:25.760970 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:25 crc kubenswrapper[4870]: I1014 07:02:25.761000 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:25 crc kubenswrapper[4870]: I1014 07:02:25.761026 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:25Z","lastTransitionTime":"2025-10-14T07:02:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:25 crc kubenswrapper[4870]: I1014 07:02:25.864808 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:25 crc kubenswrapper[4870]: I1014 07:02:25.864855 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:25 crc kubenswrapper[4870]: I1014 07:02:25.864870 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:25 crc kubenswrapper[4870]: I1014 07:02:25.864888 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:25 crc kubenswrapper[4870]: I1014 07:02:25.864903 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:25Z","lastTransitionTime":"2025-10-14T07:02:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:25 crc kubenswrapper[4870]: I1014 07:02:25.968226 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:25 crc kubenswrapper[4870]: I1014 07:02:25.968288 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:25 crc kubenswrapper[4870]: I1014 07:02:25.968314 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:25 crc kubenswrapper[4870]: I1014 07:02:25.968342 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:25 crc kubenswrapper[4870]: I1014 07:02:25.968361 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:25Z","lastTransitionTime":"2025-10-14T07:02:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:26 crc kubenswrapper[4870]: I1014 07:02:26.033382 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 07:02:26 crc kubenswrapper[4870]: E1014 07:02:26.033617 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 07:02:26 crc kubenswrapper[4870]: I1014 07:02:26.033664 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 07:02:26 crc kubenswrapper[4870]: I1014 07:02:26.033728 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 07:02:26 crc kubenswrapper[4870]: E1014 07:02:26.033833 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 07:02:26 crc kubenswrapper[4870]: E1014 07:02:26.034923 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 07:02:26 crc kubenswrapper[4870]: I1014 07:02:26.036231 4870 scope.go:117] "RemoveContainer" containerID="3fc9bfcb356b506d6c439adc22e4caaed4fad1c5a085744d44ebd50cf8898777" Oct 14 07:02:26 crc kubenswrapper[4870]: I1014 07:02:26.072380 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:26 crc kubenswrapper[4870]: I1014 07:02:26.072435 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:26 crc kubenswrapper[4870]: I1014 07:02:26.072489 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:26 crc kubenswrapper[4870]: I1014 07:02:26.072522 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:26 crc kubenswrapper[4870]: I1014 07:02:26.072546 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:26Z","lastTransitionTime":"2025-10-14T07:02:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:26 crc kubenswrapper[4870]: I1014 07:02:26.175496 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:26 crc kubenswrapper[4870]: I1014 07:02:26.175590 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:26 crc kubenswrapper[4870]: I1014 07:02:26.175619 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:26 crc kubenswrapper[4870]: I1014 07:02:26.175659 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:26 crc kubenswrapper[4870]: I1014 07:02:26.175684 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:26Z","lastTransitionTime":"2025-10-14T07:02:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:26 crc kubenswrapper[4870]: I1014 07:02:26.279852 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:26 crc kubenswrapper[4870]: I1014 07:02:26.279927 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:26 crc kubenswrapper[4870]: I1014 07:02:26.279955 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:26 crc kubenswrapper[4870]: I1014 07:02:26.279989 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:26 crc kubenswrapper[4870]: I1014 07:02:26.280011 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:26Z","lastTransitionTime":"2025-10-14T07:02:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:26 crc kubenswrapper[4870]: I1014 07:02:26.383284 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:26 crc kubenswrapper[4870]: I1014 07:02:26.383326 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:26 crc kubenswrapper[4870]: I1014 07:02:26.383337 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:26 crc kubenswrapper[4870]: I1014 07:02:26.383357 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:26 crc kubenswrapper[4870]: I1014 07:02:26.383369 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:26Z","lastTransitionTime":"2025-10-14T07:02:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:26 crc kubenswrapper[4870]: I1014 07:02:26.485893 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:26 crc kubenswrapper[4870]: I1014 07:02:26.485949 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:26 crc kubenswrapper[4870]: I1014 07:02:26.485963 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:26 crc kubenswrapper[4870]: I1014 07:02:26.485988 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:26 crc kubenswrapper[4870]: I1014 07:02:26.486003 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:26Z","lastTransitionTime":"2025-10-14T07:02:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:26 crc kubenswrapper[4870]: I1014 07:02:26.590139 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:26 crc kubenswrapper[4870]: I1014 07:02:26.590191 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:26 crc kubenswrapper[4870]: I1014 07:02:26.590205 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:26 crc kubenswrapper[4870]: I1014 07:02:26.590233 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:26 crc kubenswrapper[4870]: I1014 07:02:26.590248 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:26Z","lastTransitionTime":"2025-10-14T07:02:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:26 crc kubenswrapper[4870]: I1014 07:02:26.655331 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-v7br2_8dcaef0e-42e3-424b-bbe5-4a023efe912d/ovnkube-controller/2.log" Oct 14 07:02:26 crc kubenswrapper[4870]: I1014 07:02:26.657680 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-v7br2" event={"ID":"8dcaef0e-42e3-424b-bbe5-4a023efe912d","Type":"ContainerStarted","Data":"ab4295721b6ec6d64040d321899ab478c612e91cbc3b99b33753869578a2d1e8"} Oct 14 07:02:26 crc kubenswrapper[4870]: I1014 07:02:26.658220 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-v7br2" Oct 14 07:02:26 crc kubenswrapper[4870]: I1014 07:02:26.672434 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:26Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:26 crc kubenswrapper[4870]: I1014 07:02:26.689286 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-kz77f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d3864a91-9bf6-4540-a908-cc41aa39656c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d4ded3d86030f5ea548c8b1cc3caddbe14a45c9a3ae4af836e892880584f57d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4h8v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:28Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-kz77f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:26Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:26 crc kubenswrapper[4870]: I1014 07:02:26.693335 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:26 crc kubenswrapper[4870]: I1014 07:02:26.693380 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:26 crc kubenswrapper[4870]: I1014 07:02:26.693393 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:26 crc kubenswrapper[4870]: I1014 07:02:26.693420 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:26 crc kubenswrapper[4870]: I1014 07:02:26.693437 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:26Z","lastTransitionTime":"2025-10-14T07:02:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:26 crc kubenswrapper[4870]: I1014 07:02:26.706704 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bcbc6dd6-aab0-4dbc-a62e-c58f06ec1337\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25ec4808b09adc44488530b3c4e29fd7fec46b3fcfa31d4449ec4a6dbc880ec8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f3f5de86fb11231a19034a18e9d7f31476c162e6567d0c136ead8bbbd900d6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d57d938d9d9c92198ae5b3857536ece36012fbfbe8cc8cfd245d1386d39b0be5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77a3247a7f65709543d7fd374b2ab69d7d793537d4132d7e22261188d6e58544\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:05Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:26Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:26 crc kubenswrapper[4870]: I1014 07:02:26.723773 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e539fac3d0a2ade314a4b4c1050bb27c37bcff5ba5272fa2d71b97ab5c0d7191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:26Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:26 crc kubenswrapper[4870]: I1014 07:02:26.741230 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"874a5e45-dffd-4d17-b609-b7d3ed2eab07\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d98d044836ab44a405196915ef7bf572b75113f2c81ec7e64effa63b47ad3543\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wj4ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8004752575df4a1c44efbcafe64799f09126d0f97285d92388c76a0b90e022f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wj4ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7tvc7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:26Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:26 crc kubenswrapper[4870]: I1014 07:02:26.754423 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zc5j6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d7a4f41-388a-4320-b896-43c17ff13da6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6p2wh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6p2wh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:40Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zc5j6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:26Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:26 crc kubenswrapper[4870]: I1014 07:02:26.769028 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:26Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:26 crc kubenswrapper[4870]: I1014 07:02:26.783556 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3eb0b397c05ffd6650607f1be4a75683e0342d3333fc096bf1cd63c7688834f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:26Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:26 crc kubenswrapper[4870]: I1014 07:02:26.795685 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:26 crc kubenswrapper[4870]: I1014 07:02:26.795721 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:26 crc kubenswrapper[4870]: I1014 07:02:26.795730 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:26 crc kubenswrapper[4870]: I1014 07:02:26.795745 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:26 crc kubenswrapper[4870]: I1014 07:02:26.795753 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:26Z","lastTransitionTime":"2025-10-14T07:02:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:26 crc kubenswrapper[4870]: I1014 07:02:26.795874 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e26fe49ec2045d20a0cb2dd17e96e79b6eb43c641eb184c688cc472ca5db39d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aea115f7ee71c700b218fe15c928af7b55e6babe70e65d67f125d87be1fd2e2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:26Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:26 crc kubenswrapper[4870]: I1014 07:02:26.830503 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-v7br2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dcaef0e-42e3-424b-bbe5-4a023efe912d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71dec55e859b11a081dbb90d7d6cf6b4846cf591382303d8329003ff50c71f2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e72cff0723e2ce520340244b4e597d8016c58e1b071c4b087f2f6cd718741594\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc9d4db944f837289d1923176d60cfc4a566a282c2d2e3a703e220df884703c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d16ed1c65925549dcb69ab02c3013e25e3f123a83179f1af3a8e08bf91a59896\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed59243b59c39ce82bf9bc8f33c0a8b84854846afff6a89f5d37ffc97a096b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3b6d20cf459b1f6d3aff5b9290f9391909baf27ee40c49e2b0cd2d5cfe889ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab4295721b6ec6d64040d321899ab478c612e91cbc3b99b33753869578a2d1e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3fc9bfcb356b506d6c439adc22e4caaed4fad1c5a085744d44ebd50cf8898777\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-14T07:01:56Z\\\",\\\"message\\\":\\\" 6555 ovn.go:134] Ensuring zone local for Pod openshift-image-registry/node-ca-kz77f in node crc\\\\nI1014 07:01:56.821005 6555 obj_retry.go:386] Retry successful for *v1.Pod openshift-image-registry/node-ca-kz77f after 0 failed attempt(s)\\\\nI1014 07:01:56.820992 6555 model_client.go:382] Update operations generated as: [{Op:update Table:Logical_Switch_Port Row:map[addresses:{GoSet:[0a:58:0a:d9:00:5c 10.217.0.92]} options:{GoMap:map[iface-id-ver:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 requested-chassis:crc]} port_security:{GoSet:[0a:58:0a:d9:00:5c 10.217.0.92]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c94130be-172c-477c-88c4-40cc7eba30fe}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1014 07:01:56.821025 6555 default_network_controller.go:776] Recording success event on pod openshift-image-registry/node-ca-kz77f\\\\nI1014 07:01:56.820931 6555 default_network_controller.go:776] Recording success event on pod openshift-dns/node-resolver-4v2v2\\\\nI1014 07:01:56.820614 6555 services_controller.go:454] Service openshift-marketplace/redhat-marketplace for network=default has 1 cluster-wide, 0 per-node configs, 0 template configs, making 1 (cluster) 0 (per node) and 0 (template) load balancers\\\\nI1014 07:01:56.821037 6555 model_client.go:382] Update operations generated as: [{Op:update Table:N\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:56Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:02:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fe3468cdd612fc4378e1f6f55a2a9d00274f7ddb65cb57b2a368dfc065cfb4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ba24c2e067b9f3a14d9d018b66300b54d866530d30174fbdb3c899299575d28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ba24c2e067b9f3a14d9d018b66300b54d866530d30174fbdb3c899299575d28\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-v7br2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:26Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:26 crc kubenswrapper[4870]: I1014 07:02:26.848162 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77a0d93a-5391-4112-b502-c474f45de209\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7815af9a584f6db6660073da289f2e8c987a04ef9ddfac155bdd9a15b014fcfa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b329b570622155a55f70501b2260e278106a68ebef497914231b4ec860e0125\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba942a0e6fe5ab93a187e0b7b4bd7c440c2e7827d9b5323fc50f90b562075392\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b9018b078c5c8276243c1696860f0aa73e2d8dd917913ae9d03b30dc9b5113a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://978e8db184ba9e430232c6bb2ee530beb92567fdc71ba3c35b0bedca3492d845\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"file observer\\\\nW1014 07:01:25.870672 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1014 07:01:25.870795 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1014 07:01:25.871521 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-756714914/tls.crt::/tmp/serving-cert-756714914/tls.key\\\\\\\"\\\\nI1014 07:01:26.460037 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1014 07:01:26.464876 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1014 07:01:26.464902 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1014 07:01:26.464925 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1014 07:01:26.464931 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1014 07:01:26.481374 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1014 07:01:26.481400 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:01:26.481405 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:01:26.481410 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1014 07:01:26.481414 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1014 07:01:26.481417 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1014 07:01:26.481420 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1014 07:01:26.481610 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1014 07:01:26.482872 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bd0044a247bdc5ca4cab7294103af3270b76cde61f3109b85e83ed050ca53b6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82d3c2214e0be679d2827594b159ace362151318689078bc49d3158f42899636\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82d3c2214e0be679d2827594b159ace362151318689078bc49d3158f42899636\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:05Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:26Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:26 crc kubenswrapper[4870]: I1014 07:02:26.862029 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:26Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:26 crc kubenswrapper[4870]: I1014 07:02:26.874412 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a77729fb-a187-48c6-80db-bee90f7bdac4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:02:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:02:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c859ebc0d20b632d1d5a716caee343c7e6b9b752ed5f257422cd35ad12c170f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e8a7a5c1f1706ebbd2e8165d8961002f5a9340e2749e377b0058db107ad7dbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f75b3dc1557917e9c0c35c2810338a076ed92ba89963edafb9d537ad19c2bd8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b192bd3e0eb8540d7492abc207025471bc9e198e7e78b50215c8d2f5bc57cf4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b192bd3e0eb8540d7492abc207025471bc9e198e7e78b50215c8d2f5bc57cf4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:05Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:26Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:26 crc kubenswrapper[4870]: I1014 07:02:26.886007 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-4v2v2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4796ba6-c23b-4764-be96-2e3a3bf5e962\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf5afb0b99a7101f6d5573fb7eb925c5a1d8becc11b2e3c6f3351d80b6f0e823\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnp8b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-4v2v2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:26Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:26 crc kubenswrapper[4870]: I1014 07:02:26.897836 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:26 crc kubenswrapper[4870]: I1014 07:02:26.897909 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:26 crc kubenswrapper[4870]: I1014 07:02:26.897922 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:26 crc kubenswrapper[4870]: I1014 07:02:26.897946 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:26 crc kubenswrapper[4870]: I1014 07:02:26.897959 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:26Z","lastTransitionTime":"2025-10-14T07:02:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:26 crc kubenswrapper[4870]: I1014 07:02:26.902051 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-cdjmj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"612e7ab9-bda4-45d8-bc50-ff1afb68aace\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7776352d62b814fc63d01e4c16ad13f878b0597eb37a370bf9da0a88abab6cfa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://834444b5350ee12df4e4652c3009880d8d69e1edf8ba0c0156ab96fb2bbe28e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://834444b5350ee12df4e4652c3009880d8d69e1edf8ba0c0156ab96fb2bbe28e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c1972e7b2f75790db22a07a078eee97072bffbdb7f8e594bf2be5f800632a7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c1972e7b2f75790db22a07a078eee97072bffbdb7f8e594bf2be5f800632a7b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b124b42a44c6a23da1d92b9a5326987eb74430f80489a71e8913829475654ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b124b42a44c6a23da1d92b9a5326987eb74430f80489a71e8913829475654ab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62d0deb6b20ea65fbcbaf94c2fc14d69adb22e7a46cc043ccd4def2668ddd4b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62d0deb6b20ea65fbcbaf94c2fc14d69adb22e7a46cc043ccd4def2668ddd4b7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7f0d7cf29ca737d0effe58731a8b890379527c97f78c144d4eebb910260fadb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7f0d7cf29ca737d0effe58731a8b890379527c97f78c144d4eebb910260fadb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5959abb7c8e7ab32bed3aca709ae0334e8400f991ba724246c9fda56cbed822\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5959abb7c8e7ab32bed3aca709ae0334e8400f991ba724246c9fda56cbed822\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-cdjmj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:26Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:26 crc kubenswrapper[4870]: I1014 07:02:26.915330 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fdxjh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a3a430a-b540-4a8c-adad-b2893dbf9898\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:02:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:02:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88614373bfe4779bbe61cc41cc4c35f419e438bdae3cec4964271bf0f74ad3f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://93e41d5e927eed7ec62eb9c3607cc804cd8999556663363fd123b2a784b00ff1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-14T07:02:13Z\\\",\\\"message\\\":\\\"2025-10-14T07:01:28+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_847000ef-6815-4615-8665-86c348661ad2\\\\n2025-10-14T07:01:28+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_847000ef-6815-4615-8665-86c348661ad2 to /host/opt/cni/bin/\\\\n2025-10-14T07:01:28Z [verbose] multus-daemon started\\\\n2025-10-14T07:01:28Z [verbose] Readiness Indicator file check\\\\n2025-10-14T07:02:13Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:02:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k787j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fdxjh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:26Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:26 crc kubenswrapper[4870]: I1014 07:02:26.927511 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-khtq8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28deeea6-a1ef-4a4a-a377-d954ccd87de9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://296bb048a1d99649beb57d609573833bf547cd4d628ebc823c6af83fe5736c9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rb79p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2b7bc4c93307e80b778f2282ced993898f1c4cfc350a7d74828b8f9dea9906d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rb79p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-khtq8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:26Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:26 crc kubenswrapper[4870]: I1014 07:02:26.950609 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33a9493e-9660-48db-b8d1-310e29ebe313\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3570643ef1b7d5cc3a542f320413b819338e0c8b7ab7f0f09cc909997a39bab8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb398668197dedc1c2a14b4ae7d6a29714f3527309ccb9c1622af841100e1d60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1d2b6d068d63bd38cdfb83824525f6d12054f6cabab5d684e03ea5f2152076c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://783ff243e395333f3fc80cfae21358c532f0c9c5cb5515586be56b90ed10d912\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a8135a3d01076e4437ae8a784b459bf8d16de57fa16ec7ce4d40c5129c9d4a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48ca3a5017ba650df99b2fc7201c3fe5abfa89d81089c5647cde1b1549f534fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48ca3a5017ba650df99b2fc7201c3fe5abfa89d81089c5647cde1b1549f534fc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://19682eca67f76bf2e06e1b80f12e21141fcbce0acb8195f0defacd1b9a10b5e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://19682eca67f76bf2e06e1b80f12e21141fcbce0acb8195f0defacd1b9a10b5e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://042aad6d5f6bee8eef77f9e6a2c265de775a69f3da72aedc68bdf312198f00fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://042aad6d5f6bee8eef77f9e6a2c265de775a69f3da72aedc68bdf312198f00fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:05Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:26Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:27 crc kubenswrapper[4870]: I1014 07:02:27.001976 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:27 crc kubenswrapper[4870]: I1014 07:02:27.002043 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:27 crc kubenswrapper[4870]: I1014 07:02:27.002055 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:27 crc kubenswrapper[4870]: I1014 07:02:27.002077 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:27 crc kubenswrapper[4870]: I1014 07:02:27.002092 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:27Z","lastTransitionTime":"2025-10-14T07:02:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:27 crc kubenswrapper[4870]: I1014 07:02:27.033542 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zc5j6" Oct 14 07:02:27 crc kubenswrapper[4870]: E1014 07:02:27.033733 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zc5j6" podUID="7d7a4f41-388a-4320-b896-43c17ff13da6" Oct 14 07:02:27 crc kubenswrapper[4870]: I1014 07:02:27.046978 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Oct 14 07:02:27 crc kubenswrapper[4870]: I1014 07:02:27.105563 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:27 crc kubenswrapper[4870]: I1014 07:02:27.105635 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:27 crc kubenswrapper[4870]: I1014 07:02:27.105656 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:27 crc kubenswrapper[4870]: I1014 07:02:27.105686 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:27 crc kubenswrapper[4870]: I1014 07:02:27.105705 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:27Z","lastTransitionTime":"2025-10-14T07:02:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:27 crc kubenswrapper[4870]: I1014 07:02:27.208727 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:27 crc kubenswrapper[4870]: I1014 07:02:27.208791 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:27 crc kubenswrapper[4870]: I1014 07:02:27.208804 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:27 crc kubenswrapper[4870]: I1014 07:02:27.208823 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:27 crc kubenswrapper[4870]: I1014 07:02:27.208837 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:27Z","lastTransitionTime":"2025-10-14T07:02:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:27 crc kubenswrapper[4870]: I1014 07:02:27.311733 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:27 crc kubenswrapper[4870]: I1014 07:02:27.311777 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:27 crc kubenswrapper[4870]: I1014 07:02:27.311789 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:27 crc kubenswrapper[4870]: I1014 07:02:27.311813 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:27 crc kubenswrapper[4870]: I1014 07:02:27.311825 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:27Z","lastTransitionTime":"2025-10-14T07:02:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:27 crc kubenswrapper[4870]: I1014 07:02:27.415283 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:27 crc kubenswrapper[4870]: I1014 07:02:27.415341 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:27 crc kubenswrapper[4870]: I1014 07:02:27.415359 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:27 crc kubenswrapper[4870]: I1014 07:02:27.415386 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:27 crc kubenswrapper[4870]: I1014 07:02:27.415404 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:27Z","lastTransitionTime":"2025-10-14T07:02:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:27 crc kubenswrapper[4870]: I1014 07:02:27.517926 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:27 crc kubenswrapper[4870]: I1014 07:02:27.518025 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:27 crc kubenswrapper[4870]: I1014 07:02:27.518045 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:27 crc kubenswrapper[4870]: I1014 07:02:27.518073 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:27 crc kubenswrapper[4870]: I1014 07:02:27.518091 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:27Z","lastTransitionTime":"2025-10-14T07:02:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:27 crc kubenswrapper[4870]: I1014 07:02:27.620942 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:27 crc kubenswrapper[4870]: I1014 07:02:27.621424 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:27 crc kubenswrapper[4870]: I1014 07:02:27.621471 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:27 crc kubenswrapper[4870]: I1014 07:02:27.621506 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:27 crc kubenswrapper[4870]: I1014 07:02:27.621527 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:27Z","lastTransitionTime":"2025-10-14T07:02:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:27 crc kubenswrapper[4870]: I1014 07:02:27.664175 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-v7br2_8dcaef0e-42e3-424b-bbe5-4a023efe912d/ovnkube-controller/3.log" Oct 14 07:02:27 crc kubenswrapper[4870]: I1014 07:02:27.665030 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-v7br2_8dcaef0e-42e3-424b-bbe5-4a023efe912d/ovnkube-controller/2.log" Oct 14 07:02:27 crc kubenswrapper[4870]: I1014 07:02:27.669305 4870 generic.go:334] "Generic (PLEG): container finished" podID="8dcaef0e-42e3-424b-bbe5-4a023efe912d" containerID="ab4295721b6ec6d64040d321899ab478c612e91cbc3b99b33753869578a2d1e8" exitCode=1 Oct 14 07:02:27 crc kubenswrapper[4870]: I1014 07:02:27.669431 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-v7br2" event={"ID":"8dcaef0e-42e3-424b-bbe5-4a023efe912d","Type":"ContainerDied","Data":"ab4295721b6ec6d64040d321899ab478c612e91cbc3b99b33753869578a2d1e8"} Oct 14 07:02:27 crc kubenswrapper[4870]: I1014 07:02:27.669578 4870 scope.go:117] "RemoveContainer" containerID="3fc9bfcb356b506d6c439adc22e4caaed4fad1c5a085744d44ebd50cf8898777" Oct 14 07:02:27 crc kubenswrapper[4870]: I1014 07:02:27.670346 4870 scope.go:117] "RemoveContainer" containerID="ab4295721b6ec6d64040d321899ab478c612e91cbc3b99b33753869578a2d1e8" Oct 14 07:02:27 crc kubenswrapper[4870]: E1014 07:02:27.670582 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-v7br2_openshift-ovn-kubernetes(8dcaef0e-42e3-424b-bbe5-4a023efe912d)\"" pod="openshift-ovn-kubernetes/ovnkube-node-v7br2" podUID="8dcaef0e-42e3-424b-bbe5-4a023efe912d" Oct 14 07:02:27 crc kubenswrapper[4870]: I1014 07:02:27.689979 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77a0d93a-5391-4112-b502-c474f45de209\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7815af9a584f6db6660073da289f2e8c987a04ef9ddfac155bdd9a15b014fcfa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b329b570622155a55f70501b2260e278106a68ebef497914231b4ec860e0125\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba942a0e6fe5ab93a187e0b7b4bd7c440c2e7827d9b5323fc50f90b562075392\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b9018b078c5c8276243c1696860f0aa73e2d8dd917913ae9d03b30dc9b5113a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://978e8db184ba9e430232c6bb2ee530beb92567fdc71ba3c35b0bedca3492d845\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"file observer\\\\nW1014 07:01:25.870672 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1014 07:01:25.870795 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1014 07:01:25.871521 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-756714914/tls.crt::/tmp/serving-cert-756714914/tls.key\\\\\\\"\\\\nI1014 07:01:26.460037 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1014 07:01:26.464876 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1014 07:01:26.464902 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1014 07:01:26.464925 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1014 07:01:26.464931 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1014 07:01:26.481374 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1014 07:01:26.481400 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:01:26.481405 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:01:26.481410 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1014 07:01:26.481414 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1014 07:01:26.481417 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1014 07:01:26.481420 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1014 07:01:26.481610 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1014 07:01:26.482872 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bd0044a247bdc5ca4cab7294103af3270b76cde61f3109b85e83ed050ca53b6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82d3c2214e0be679d2827594b159ace362151318689078bc49d3158f42899636\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82d3c2214e0be679d2827594b159ace362151318689078bc49d3158f42899636\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:05Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:27Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:27 crc kubenswrapper[4870]: I1014 07:02:27.703684 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:27Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:27 crc kubenswrapper[4870]: I1014 07:02:27.720136 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e26fe49ec2045d20a0cb2dd17e96e79b6eb43c641eb184c688cc472ca5db39d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aea115f7ee71c700b218fe15c928af7b55e6babe70e65d67f125d87be1fd2e2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:27Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:27 crc kubenswrapper[4870]: I1014 07:02:27.725436 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:27 crc kubenswrapper[4870]: I1014 07:02:27.725487 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:27 crc kubenswrapper[4870]: I1014 07:02:27.725497 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:27 crc kubenswrapper[4870]: I1014 07:02:27.725516 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:27 crc kubenswrapper[4870]: I1014 07:02:27.725531 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:27Z","lastTransitionTime":"2025-10-14T07:02:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:27 crc kubenswrapper[4870]: I1014 07:02:27.741267 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-v7br2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dcaef0e-42e3-424b-bbe5-4a023efe912d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71dec55e859b11a081dbb90d7d6cf6b4846cf591382303d8329003ff50c71f2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e72cff0723e2ce520340244b4e597d8016c58e1b071c4b087f2f6cd718741594\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc9d4db944f837289d1923176d60cfc4a566a282c2d2e3a703e220df884703c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d16ed1c65925549dcb69ab02c3013e25e3f123a83179f1af3a8e08bf91a59896\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed59243b59c39ce82bf9bc8f33c0a8b84854846afff6a89f5d37ffc97a096b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3b6d20cf459b1f6d3aff5b9290f9391909baf27ee40c49e2b0cd2d5cfe889ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab4295721b6ec6d64040d321899ab478c612e91cbc3b99b33753869578a2d1e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3fc9bfcb356b506d6c439adc22e4caaed4fad1c5a085744d44ebd50cf8898777\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-14T07:01:56Z\\\",\\\"message\\\":\\\" 6555 ovn.go:134] Ensuring zone local for Pod openshift-image-registry/node-ca-kz77f in node crc\\\\nI1014 07:01:56.821005 6555 obj_retry.go:386] Retry successful for *v1.Pod openshift-image-registry/node-ca-kz77f after 0 failed attempt(s)\\\\nI1014 07:01:56.820992 6555 model_client.go:382] Update operations generated as: [{Op:update Table:Logical_Switch_Port Row:map[addresses:{GoSet:[0a:58:0a:d9:00:5c 10.217.0.92]} options:{GoMap:map[iface-id-ver:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 requested-chassis:crc]} port_security:{GoSet:[0a:58:0a:d9:00:5c 10.217.0.92]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c94130be-172c-477c-88c4-40cc7eba30fe}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1014 07:01:56.821025 6555 default_network_controller.go:776] Recording success event on pod openshift-image-registry/node-ca-kz77f\\\\nI1014 07:01:56.820931 6555 default_network_controller.go:776] Recording success event on pod openshift-dns/node-resolver-4v2v2\\\\nI1014 07:01:56.820614 6555 services_controller.go:454] Service openshift-marketplace/redhat-marketplace for network=default has 1 cluster-wide, 0 per-node configs, 0 template configs, making 1 (cluster) 0 (per node) and 0 (template) load balancers\\\\nI1014 07:01:56.821037 6555 model_client.go:382] Update operations generated as: [{Op:update Table:N\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:56Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ab4295721b6ec6d64040d321899ab478c612e91cbc3b99b33753869578a2d1e8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-14T07:02:27Z\\\",\\\"message\\\":\\\"ckageserver-service]} name:Service_openshift-operator-lifecycle-manager/packageserver-service_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.153:5443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {5e50827b-d271-442b-b8a7-7f33b2cd6b11}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1014 07:02:27.009850 6956 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-operator-lifecycle-manager/packageserver-service]} name:Service_openshift-operator-lifecycle-manager/packageserver-service_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.153:5443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {5e50827b-d271-442b-b8a7-7f33b2cd6b11}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1014 07:02:27.009895 6956 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T07:02:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fe3468cdd612fc4378e1f6f55a2a9d00274f7ddb65cb57b2a368dfc065cfb4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ba24c2e067b9f3a14d9d018b66300b54d866530d30174fbdb3c899299575d28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ba24c2e067b9f3a14d9d018b66300b54d866530d30174fbdb3c899299575d28\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-v7br2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:27Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:27 crc kubenswrapper[4870]: I1014 07:02:27.754840 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-khtq8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28deeea6-a1ef-4a4a-a377-d954ccd87de9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://296bb048a1d99649beb57d609573833bf547cd4d628ebc823c6af83fe5736c9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rb79p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2b7bc4c93307e80b778f2282ced993898f1c4cfc350a7d74828b8f9dea9906d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rb79p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-khtq8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:27Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:27 crc kubenswrapper[4870]: I1014 07:02:27.777121 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33a9493e-9660-48db-b8d1-310e29ebe313\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3570643ef1b7d5cc3a542f320413b819338e0c8b7ab7f0f09cc909997a39bab8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb398668197dedc1c2a14b4ae7d6a29714f3527309ccb9c1622af841100e1d60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1d2b6d068d63bd38cdfb83824525f6d12054f6cabab5d684e03ea5f2152076c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://783ff243e395333f3fc80cfae21358c532f0c9c5cb5515586be56b90ed10d912\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a8135a3d01076e4437ae8a784b459bf8d16de57fa16ec7ce4d40c5129c9d4a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48ca3a5017ba650df99b2fc7201c3fe5abfa89d81089c5647cde1b1549f534fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48ca3a5017ba650df99b2fc7201c3fe5abfa89d81089c5647cde1b1549f534fc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://19682eca67f76bf2e06e1b80f12e21141fcbce0acb8195f0defacd1b9a10b5e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://19682eca67f76bf2e06e1b80f12e21141fcbce0acb8195f0defacd1b9a10b5e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://042aad6d5f6bee8eef77f9e6a2c265de775a69f3da72aedc68bdf312198f00fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://042aad6d5f6bee8eef77f9e6a2c265de775a69f3da72aedc68bdf312198f00fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:05Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:27Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:27 crc kubenswrapper[4870]: I1014 07:02:27.789536 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"221727c7-4b56-4fc4-b745-1b6e71666779\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8c8ee375ef1ec48a55be4a2d3e9e8860886b88ccb4080918d3e306ab34d2b95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07543849cd70415fe71b7dd0fd5f3806fdb422afa7a2dcce18a5c1b9fc974a76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://07543849cd70415fe71b7dd0fd5f3806fdb422afa7a2dcce18a5c1b9fc974a76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:27Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:27 crc kubenswrapper[4870]: I1014 07:02:27.804293 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a77729fb-a187-48c6-80db-bee90f7bdac4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:02:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:02:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c859ebc0d20b632d1d5a716caee343c7e6b9b752ed5f257422cd35ad12c170f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e8a7a5c1f1706ebbd2e8165d8961002f5a9340e2749e377b0058db107ad7dbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f75b3dc1557917e9c0c35c2810338a076ed92ba89963edafb9d537ad19c2bd8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b192bd3e0eb8540d7492abc207025471bc9e198e7e78b50215c8d2f5bc57cf4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b192bd3e0eb8540d7492abc207025471bc9e198e7e78b50215c8d2f5bc57cf4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:05Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:27Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:27 crc kubenswrapper[4870]: I1014 07:02:27.814954 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-4v2v2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4796ba6-c23b-4764-be96-2e3a3bf5e962\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf5afb0b99a7101f6d5573fb7eb925c5a1d8becc11b2e3c6f3351d80b6f0e823\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnp8b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-4v2v2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:27Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:27 crc kubenswrapper[4870]: I1014 07:02:27.828597 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:27 crc kubenswrapper[4870]: I1014 07:02:27.828665 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:27 crc kubenswrapper[4870]: I1014 07:02:27.828698 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:27 crc kubenswrapper[4870]: I1014 07:02:27.828723 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:27 crc kubenswrapper[4870]: I1014 07:02:27.828734 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:27Z","lastTransitionTime":"2025-10-14T07:02:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:27 crc kubenswrapper[4870]: I1014 07:02:27.833685 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-cdjmj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"612e7ab9-bda4-45d8-bc50-ff1afb68aace\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7776352d62b814fc63d01e4c16ad13f878b0597eb37a370bf9da0a88abab6cfa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://834444b5350ee12df4e4652c3009880d8d69e1edf8ba0c0156ab96fb2bbe28e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://834444b5350ee12df4e4652c3009880d8d69e1edf8ba0c0156ab96fb2bbe28e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c1972e7b2f75790db22a07a078eee97072bffbdb7f8e594bf2be5f800632a7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c1972e7b2f75790db22a07a078eee97072bffbdb7f8e594bf2be5f800632a7b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b124b42a44c6a23da1d92b9a5326987eb74430f80489a71e8913829475654ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b124b42a44c6a23da1d92b9a5326987eb74430f80489a71e8913829475654ab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62d0deb6b20ea65fbcbaf94c2fc14d69adb22e7a46cc043ccd4def2668ddd4b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62d0deb6b20ea65fbcbaf94c2fc14d69adb22e7a46cc043ccd4def2668ddd4b7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7f0d7cf29ca737d0effe58731a8b890379527c97f78c144d4eebb910260fadb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7f0d7cf29ca737d0effe58731a8b890379527c97f78c144d4eebb910260fadb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5959abb7c8e7ab32bed3aca709ae0334e8400f991ba724246c9fda56cbed822\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5959abb7c8e7ab32bed3aca709ae0334e8400f991ba724246c9fda56cbed822\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-cdjmj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:27Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:27 crc kubenswrapper[4870]: I1014 07:02:27.847932 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fdxjh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a3a430a-b540-4a8c-adad-b2893dbf9898\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:02:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:02:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88614373bfe4779bbe61cc41cc4c35f419e438bdae3cec4964271bf0f74ad3f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://93e41d5e927eed7ec62eb9c3607cc804cd8999556663363fd123b2a784b00ff1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-14T07:02:13Z\\\",\\\"message\\\":\\\"2025-10-14T07:01:28+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_847000ef-6815-4615-8665-86c348661ad2\\\\n2025-10-14T07:01:28+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_847000ef-6815-4615-8665-86c348661ad2 to /host/opt/cni/bin/\\\\n2025-10-14T07:01:28Z [verbose] multus-daemon started\\\\n2025-10-14T07:01:28Z [verbose] Readiness Indicator file check\\\\n2025-10-14T07:02:13Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:02:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k787j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fdxjh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:27Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:27 crc kubenswrapper[4870]: I1014 07:02:27.863401 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bcbc6dd6-aab0-4dbc-a62e-c58f06ec1337\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25ec4808b09adc44488530b3c4e29fd7fec46b3fcfa31d4449ec4a6dbc880ec8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f3f5de86fb11231a19034a18e9d7f31476c162e6567d0c136ead8bbbd900d6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d57d938d9d9c92198ae5b3857536ece36012fbfbe8cc8cfd245d1386d39b0be5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77a3247a7f65709543d7fd374b2ab69d7d793537d4132d7e22261188d6e58544\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:05Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:27Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:27 crc kubenswrapper[4870]: I1014 07:02:27.881158 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e539fac3d0a2ade314a4b4c1050bb27c37bcff5ba5272fa2d71b97ab5c0d7191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:27Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:27 crc kubenswrapper[4870]: I1014 07:02:27.896002 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:27Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:27 crc kubenswrapper[4870]: I1014 07:02:27.907075 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-kz77f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d3864a91-9bf6-4540-a908-cc41aa39656c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d4ded3d86030f5ea548c8b1cc3caddbe14a45c9a3ae4af836e892880584f57d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4h8v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:28Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-kz77f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:27Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:27 crc kubenswrapper[4870]: I1014 07:02:27.923843 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:27Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:27 crc kubenswrapper[4870]: I1014 07:02:27.933509 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:27 crc kubenswrapper[4870]: I1014 07:02:27.933579 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:27 crc kubenswrapper[4870]: I1014 07:02:27.933610 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:27 crc kubenswrapper[4870]: I1014 07:02:27.933643 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:27 crc kubenswrapper[4870]: I1014 07:02:27.933659 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:27Z","lastTransitionTime":"2025-10-14T07:02:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:27 crc kubenswrapper[4870]: I1014 07:02:27.942394 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3eb0b397c05ffd6650607f1be4a75683e0342d3333fc096bf1cd63c7688834f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:27Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:27 crc kubenswrapper[4870]: I1014 07:02:27.960774 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"874a5e45-dffd-4d17-b609-b7d3ed2eab07\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d98d044836ab44a405196915ef7bf572b75113f2c81ec7e64effa63b47ad3543\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wj4ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8004752575df4a1c44efbcafe64799f09126d0f97285d92388c76a0b90e022f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wj4ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7tvc7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:27Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:27 crc kubenswrapper[4870]: I1014 07:02:27.977290 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zc5j6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d7a4f41-388a-4320-b896-43c17ff13da6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6p2wh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6p2wh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:40Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zc5j6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:27Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:28 crc kubenswrapper[4870]: I1014 07:02:28.033242 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 07:02:28 crc kubenswrapper[4870]: I1014 07:02:28.033242 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 07:02:28 crc kubenswrapper[4870]: E1014 07:02:28.033412 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 07:02:28 crc kubenswrapper[4870]: I1014 07:02:28.033427 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 07:02:28 crc kubenswrapper[4870]: E1014 07:02:28.033508 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 07:02:28 crc kubenswrapper[4870]: E1014 07:02:28.033582 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 07:02:28 crc kubenswrapper[4870]: I1014 07:02:28.035993 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:28 crc kubenswrapper[4870]: I1014 07:02:28.036019 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:28 crc kubenswrapper[4870]: I1014 07:02:28.036030 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:28 crc kubenswrapper[4870]: I1014 07:02:28.036044 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:28 crc kubenswrapper[4870]: I1014 07:02:28.036058 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:28Z","lastTransitionTime":"2025-10-14T07:02:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:28 crc kubenswrapper[4870]: I1014 07:02:28.140336 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:28 crc kubenswrapper[4870]: I1014 07:02:28.140405 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:28 crc kubenswrapper[4870]: I1014 07:02:28.140422 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:28 crc kubenswrapper[4870]: I1014 07:02:28.140472 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:28 crc kubenswrapper[4870]: I1014 07:02:28.140487 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:28Z","lastTransitionTime":"2025-10-14T07:02:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:28 crc kubenswrapper[4870]: I1014 07:02:28.242834 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:28 crc kubenswrapper[4870]: I1014 07:02:28.242873 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:28 crc kubenswrapper[4870]: I1014 07:02:28.242883 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:28 crc kubenswrapper[4870]: I1014 07:02:28.242897 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:28 crc kubenswrapper[4870]: I1014 07:02:28.242907 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:28Z","lastTransitionTime":"2025-10-14T07:02:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:28 crc kubenswrapper[4870]: I1014 07:02:28.348201 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:28 crc kubenswrapper[4870]: I1014 07:02:28.348250 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:28 crc kubenswrapper[4870]: I1014 07:02:28.348263 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:28 crc kubenswrapper[4870]: I1014 07:02:28.348282 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:28 crc kubenswrapper[4870]: I1014 07:02:28.348292 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:28Z","lastTransitionTime":"2025-10-14T07:02:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:28 crc kubenswrapper[4870]: I1014 07:02:28.450718 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:28 crc kubenswrapper[4870]: I1014 07:02:28.450774 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:28 crc kubenswrapper[4870]: I1014 07:02:28.450791 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:28 crc kubenswrapper[4870]: I1014 07:02:28.450819 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:28 crc kubenswrapper[4870]: I1014 07:02:28.450837 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:28Z","lastTransitionTime":"2025-10-14T07:02:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:28 crc kubenswrapper[4870]: I1014 07:02:28.554022 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:28 crc kubenswrapper[4870]: I1014 07:02:28.554069 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:28 crc kubenswrapper[4870]: I1014 07:02:28.554082 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:28 crc kubenswrapper[4870]: I1014 07:02:28.554103 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:28 crc kubenswrapper[4870]: I1014 07:02:28.554115 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:28Z","lastTransitionTime":"2025-10-14T07:02:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:28 crc kubenswrapper[4870]: I1014 07:02:28.657907 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:28 crc kubenswrapper[4870]: I1014 07:02:28.658003 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:28 crc kubenswrapper[4870]: I1014 07:02:28.658040 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:28 crc kubenswrapper[4870]: I1014 07:02:28.658075 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:28 crc kubenswrapper[4870]: I1014 07:02:28.658104 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:28Z","lastTransitionTime":"2025-10-14T07:02:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:28 crc kubenswrapper[4870]: I1014 07:02:28.676632 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-v7br2_8dcaef0e-42e3-424b-bbe5-4a023efe912d/ovnkube-controller/3.log" Oct 14 07:02:28 crc kubenswrapper[4870]: I1014 07:02:28.681345 4870 scope.go:117] "RemoveContainer" containerID="ab4295721b6ec6d64040d321899ab478c612e91cbc3b99b33753869578a2d1e8" Oct 14 07:02:28 crc kubenswrapper[4870]: E1014 07:02:28.681526 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-v7br2_openshift-ovn-kubernetes(8dcaef0e-42e3-424b-bbe5-4a023efe912d)\"" pod="openshift-ovn-kubernetes/ovnkube-node-v7br2" podUID="8dcaef0e-42e3-424b-bbe5-4a023efe912d" Oct 14 07:02:28 crc kubenswrapper[4870]: I1014 07:02:28.705581 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77a0d93a-5391-4112-b502-c474f45de209\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7815af9a584f6db6660073da289f2e8c987a04ef9ddfac155bdd9a15b014fcfa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b329b570622155a55f70501b2260e278106a68ebef497914231b4ec860e0125\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba942a0e6fe5ab93a187e0b7b4bd7c440c2e7827d9b5323fc50f90b562075392\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b9018b078c5c8276243c1696860f0aa73e2d8dd917913ae9d03b30dc9b5113a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://978e8db184ba9e430232c6bb2ee530beb92567fdc71ba3c35b0bedca3492d845\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"file observer\\\\nW1014 07:01:25.870672 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1014 07:01:25.870795 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1014 07:01:25.871521 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-756714914/tls.crt::/tmp/serving-cert-756714914/tls.key\\\\\\\"\\\\nI1014 07:01:26.460037 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1014 07:01:26.464876 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1014 07:01:26.464902 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1014 07:01:26.464925 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1014 07:01:26.464931 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1014 07:01:26.481374 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1014 07:01:26.481400 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:01:26.481405 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:01:26.481410 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1014 07:01:26.481414 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1014 07:01:26.481417 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1014 07:01:26.481420 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1014 07:01:26.481610 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1014 07:01:26.482872 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bd0044a247bdc5ca4cab7294103af3270b76cde61f3109b85e83ed050ca53b6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82d3c2214e0be679d2827594b159ace362151318689078bc49d3158f42899636\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82d3c2214e0be679d2827594b159ace362151318689078bc49d3158f42899636\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:05Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:28Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:28 crc kubenswrapper[4870]: I1014 07:02:28.728220 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:28Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:28 crc kubenswrapper[4870]: I1014 07:02:28.751256 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e26fe49ec2045d20a0cb2dd17e96e79b6eb43c641eb184c688cc472ca5db39d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aea115f7ee71c700b218fe15c928af7b55e6babe70e65d67f125d87be1fd2e2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:28Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:28 crc kubenswrapper[4870]: I1014 07:02:28.762094 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:28 crc kubenswrapper[4870]: I1014 07:02:28.762148 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:28 crc kubenswrapper[4870]: I1014 07:02:28.762163 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:28 crc kubenswrapper[4870]: I1014 07:02:28.762184 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:28 crc kubenswrapper[4870]: I1014 07:02:28.762198 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:28Z","lastTransitionTime":"2025-10-14T07:02:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:28 crc kubenswrapper[4870]: I1014 07:02:28.780106 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-v7br2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dcaef0e-42e3-424b-bbe5-4a023efe912d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71dec55e859b11a081dbb90d7d6cf6b4846cf591382303d8329003ff50c71f2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e72cff0723e2ce520340244b4e597d8016c58e1b071c4b087f2f6cd718741594\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc9d4db944f837289d1923176d60cfc4a566a282c2d2e3a703e220df884703c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d16ed1c65925549dcb69ab02c3013e25e3f123a83179f1af3a8e08bf91a59896\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed59243b59c39ce82bf9bc8f33c0a8b84854846afff6a89f5d37ffc97a096b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3b6d20cf459b1f6d3aff5b9290f9391909baf27ee40c49e2b0cd2d5cfe889ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab4295721b6ec6d64040d321899ab478c612e91cbc3b99b33753869578a2d1e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ab4295721b6ec6d64040d321899ab478c612e91cbc3b99b33753869578a2d1e8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-14T07:02:27Z\\\",\\\"message\\\":\\\"ckageserver-service]} name:Service_openshift-operator-lifecycle-manager/packageserver-service_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.153:5443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {5e50827b-d271-442b-b8a7-7f33b2cd6b11}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1014 07:02:27.009850 6956 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-operator-lifecycle-manager/packageserver-service]} name:Service_openshift-operator-lifecycle-manager/packageserver-service_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.153:5443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {5e50827b-d271-442b-b8a7-7f33b2cd6b11}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1014 07:02:27.009895 6956 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T07:02:26Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-v7br2_openshift-ovn-kubernetes(8dcaef0e-42e3-424b-bbe5-4a023efe912d)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fe3468cdd612fc4378e1f6f55a2a9d00274f7ddb65cb57b2a368dfc065cfb4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ba24c2e067b9f3a14d9d018b66300b54d866530d30174fbdb3c899299575d28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ba24c2e067b9f3a14d9d018b66300b54d866530d30174fbdb3c899299575d28\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-v7br2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:28Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:28 crc kubenswrapper[4870]: I1014 07:02:28.800045 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-khtq8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28deeea6-a1ef-4a4a-a377-d954ccd87de9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://296bb048a1d99649beb57d609573833bf547cd4d628ebc823c6af83fe5736c9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rb79p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2b7bc4c93307e80b778f2282ced993898f1c4cfc350a7d74828b8f9dea9906d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rb79p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-khtq8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:28Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:28 crc kubenswrapper[4870]: I1014 07:02:28.827770 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33a9493e-9660-48db-b8d1-310e29ebe313\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3570643ef1b7d5cc3a542f320413b819338e0c8b7ab7f0f09cc909997a39bab8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb398668197dedc1c2a14b4ae7d6a29714f3527309ccb9c1622af841100e1d60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1d2b6d068d63bd38cdfb83824525f6d12054f6cabab5d684e03ea5f2152076c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://783ff243e395333f3fc80cfae21358c532f0c9c5cb5515586be56b90ed10d912\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a8135a3d01076e4437ae8a784b459bf8d16de57fa16ec7ce4d40c5129c9d4a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48ca3a5017ba650df99b2fc7201c3fe5abfa89d81089c5647cde1b1549f534fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48ca3a5017ba650df99b2fc7201c3fe5abfa89d81089c5647cde1b1549f534fc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://19682eca67f76bf2e06e1b80f12e21141fcbce0acb8195f0defacd1b9a10b5e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://19682eca67f76bf2e06e1b80f12e21141fcbce0acb8195f0defacd1b9a10b5e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://042aad6d5f6bee8eef77f9e6a2c265de775a69f3da72aedc68bdf312198f00fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://042aad6d5f6bee8eef77f9e6a2c265de775a69f3da72aedc68bdf312198f00fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:05Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:28Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:28 crc kubenswrapper[4870]: I1014 07:02:28.842056 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"221727c7-4b56-4fc4-b745-1b6e71666779\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8c8ee375ef1ec48a55be4a2d3e9e8860886b88ccb4080918d3e306ab34d2b95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07543849cd70415fe71b7dd0fd5f3806fdb422afa7a2dcce18a5c1b9fc974a76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://07543849cd70415fe71b7dd0fd5f3806fdb422afa7a2dcce18a5c1b9fc974a76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:28Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:28 crc kubenswrapper[4870]: I1014 07:02:28.858881 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a77729fb-a187-48c6-80db-bee90f7bdac4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:02:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:02:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c859ebc0d20b632d1d5a716caee343c7e6b9b752ed5f257422cd35ad12c170f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e8a7a5c1f1706ebbd2e8165d8961002f5a9340e2749e377b0058db107ad7dbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f75b3dc1557917e9c0c35c2810338a076ed92ba89963edafb9d537ad19c2bd8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b192bd3e0eb8540d7492abc207025471bc9e198e7e78b50215c8d2f5bc57cf4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b192bd3e0eb8540d7492abc207025471bc9e198e7e78b50215c8d2f5bc57cf4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:05Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:28Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:28 crc kubenswrapper[4870]: I1014 07:02:28.865340 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:28 crc kubenswrapper[4870]: I1014 07:02:28.865434 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:28 crc kubenswrapper[4870]: I1014 07:02:28.865508 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:28 crc kubenswrapper[4870]: I1014 07:02:28.865612 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:28 crc kubenswrapper[4870]: I1014 07:02:28.865637 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:28Z","lastTransitionTime":"2025-10-14T07:02:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:28 crc kubenswrapper[4870]: I1014 07:02:28.880382 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-4v2v2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4796ba6-c23b-4764-be96-2e3a3bf5e962\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf5afb0b99a7101f6d5573fb7eb925c5a1d8becc11b2e3c6f3351d80b6f0e823\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnp8b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-4v2v2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:28Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:28 crc kubenswrapper[4870]: I1014 07:02:28.906343 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-cdjmj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"612e7ab9-bda4-45d8-bc50-ff1afb68aace\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7776352d62b814fc63d01e4c16ad13f878b0597eb37a370bf9da0a88abab6cfa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://834444b5350ee12df4e4652c3009880d8d69e1edf8ba0c0156ab96fb2bbe28e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://834444b5350ee12df4e4652c3009880d8d69e1edf8ba0c0156ab96fb2bbe28e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c1972e7b2f75790db22a07a078eee97072bffbdb7f8e594bf2be5f800632a7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c1972e7b2f75790db22a07a078eee97072bffbdb7f8e594bf2be5f800632a7b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b124b42a44c6a23da1d92b9a5326987eb74430f80489a71e8913829475654ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b124b42a44c6a23da1d92b9a5326987eb74430f80489a71e8913829475654ab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62d0deb6b20ea65fbcbaf94c2fc14d69adb22e7a46cc043ccd4def2668ddd4b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62d0deb6b20ea65fbcbaf94c2fc14d69adb22e7a46cc043ccd4def2668ddd4b7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7f0d7cf29ca737d0effe58731a8b890379527c97f78c144d4eebb910260fadb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7f0d7cf29ca737d0effe58731a8b890379527c97f78c144d4eebb910260fadb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5959abb7c8e7ab32bed3aca709ae0334e8400f991ba724246c9fda56cbed822\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5959abb7c8e7ab32bed3aca709ae0334e8400f991ba724246c9fda56cbed822\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-cdjmj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:28Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:28 crc kubenswrapper[4870]: I1014 07:02:28.924348 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fdxjh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a3a430a-b540-4a8c-adad-b2893dbf9898\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:02:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:02:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88614373bfe4779bbe61cc41cc4c35f419e438bdae3cec4964271bf0f74ad3f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://93e41d5e927eed7ec62eb9c3607cc804cd8999556663363fd123b2a784b00ff1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-14T07:02:13Z\\\",\\\"message\\\":\\\"2025-10-14T07:01:28+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_847000ef-6815-4615-8665-86c348661ad2\\\\n2025-10-14T07:01:28+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_847000ef-6815-4615-8665-86c348661ad2 to /host/opt/cni/bin/\\\\n2025-10-14T07:01:28Z [verbose] multus-daemon started\\\\n2025-10-14T07:01:28Z [verbose] Readiness Indicator file check\\\\n2025-10-14T07:02:13Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:02:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k787j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fdxjh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:28Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:28 crc kubenswrapper[4870]: I1014 07:02:28.941733 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bcbc6dd6-aab0-4dbc-a62e-c58f06ec1337\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25ec4808b09adc44488530b3c4e29fd7fec46b3fcfa31d4449ec4a6dbc880ec8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f3f5de86fb11231a19034a18e9d7f31476c162e6567d0c136ead8bbbd900d6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d57d938d9d9c92198ae5b3857536ece36012fbfbe8cc8cfd245d1386d39b0be5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77a3247a7f65709543d7fd374b2ab69d7d793537d4132d7e22261188d6e58544\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:05Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:28Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:28 crc kubenswrapper[4870]: I1014 07:02:28.960088 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e539fac3d0a2ade314a4b4c1050bb27c37bcff5ba5272fa2d71b97ab5c0d7191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:28Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:28 crc kubenswrapper[4870]: I1014 07:02:28.969006 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:28 crc kubenswrapper[4870]: I1014 07:02:28.969061 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:28 crc kubenswrapper[4870]: I1014 07:02:28.969078 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:28 crc kubenswrapper[4870]: I1014 07:02:28.969104 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:28 crc kubenswrapper[4870]: I1014 07:02:28.969124 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:28Z","lastTransitionTime":"2025-10-14T07:02:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:28 crc kubenswrapper[4870]: I1014 07:02:28.976588 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:28Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:28 crc kubenswrapper[4870]: I1014 07:02:28.990702 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-kz77f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d3864a91-9bf6-4540-a908-cc41aa39656c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d4ded3d86030f5ea548c8b1cc3caddbe14a45c9a3ae4af836e892880584f57d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4h8v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:28Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-kz77f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:28Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:29 crc kubenswrapper[4870]: I1014 07:02:29.010690 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:29Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:29 crc kubenswrapper[4870]: I1014 07:02:29.026734 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3eb0b397c05ffd6650607f1be4a75683e0342d3333fc096bf1cd63c7688834f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:29Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:29 crc kubenswrapper[4870]: I1014 07:02:29.033151 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zc5j6" Oct 14 07:02:29 crc kubenswrapper[4870]: E1014 07:02:29.033357 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zc5j6" podUID="7d7a4f41-388a-4320-b896-43c17ff13da6" Oct 14 07:02:29 crc kubenswrapper[4870]: I1014 07:02:29.040997 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"874a5e45-dffd-4d17-b609-b7d3ed2eab07\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d98d044836ab44a405196915ef7bf572b75113f2c81ec7e64effa63b47ad3543\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wj4ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8004752575df4a1c44efbcafe64799f09126d0f97285d92388c76a0b90e022f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wj4ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7tvc7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:29Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:29 crc kubenswrapper[4870]: I1014 07:02:29.053506 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zc5j6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d7a4f41-388a-4320-b896-43c17ff13da6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6p2wh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6p2wh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:40Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zc5j6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:29Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:29 crc kubenswrapper[4870]: I1014 07:02:29.071785 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:29 crc kubenswrapper[4870]: I1014 07:02:29.071836 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:29 crc kubenswrapper[4870]: I1014 07:02:29.071847 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:29 crc kubenswrapper[4870]: I1014 07:02:29.071868 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:29 crc kubenswrapper[4870]: I1014 07:02:29.071882 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:29Z","lastTransitionTime":"2025-10-14T07:02:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:29 crc kubenswrapper[4870]: I1014 07:02:29.174566 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:29 crc kubenswrapper[4870]: I1014 07:02:29.174613 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:29 crc kubenswrapper[4870]: I1014 07:02:29.174624 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:29 crc kubenswrapper[4870]: I1014 07:02:29.174638 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:29 crc kubenswrapper[4870]: I1014 07:02:29.174649 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:29Z","lastTransitionTime":"2025-10-14T07:02:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:29 crc kubenswrapper[4870]: I1014 07:02:29.278047 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:29 crc kubenswrapper[4870]: I1014 07:02:29.278136 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:29 crc kubenswrapper[4870]: I1014 07:02:29.278158 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:29 crc kubenswrapper[4870]: I1014 07:02:29.278194 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:29 crc kubenswrapper[4870]: I1014 07:02:29.278229 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:29Z","lastTransitionTime":"2025-10-14T07:02:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:29 crc kubenswrapper[4870]: I1014 07:02:29.380387 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:29 crc kubenswrapper[4870]: I1014 07:02:29.380416 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:29 crc kubenswrapper[4870]: I1014 07:02:29.380423 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:29 crc kubenswrapper[4870]: I1014 07:02:29.380456 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:29 crc kubenswrapper[4870]: I1014 07:02:29.380465 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:29Z","lastTransitionTime":"2025-10-14T07:02:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:29 crc kubenswrapper[4870]: I1014 07:02:29.483854 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:29 crc kubenswrapper[4870]: I1014 07:02:29.483908 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:29 crc kubenswrapper[4870]: I1014 07:02:29.483920 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:29 crc kubenswrapper[4870]: I1014 07:02:29.483938 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:29 crc kubenswrapper[4870]: I1014 07:02:29.483950 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:29Z","lastTransitionTime":"2025-10-14T07:02:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:29 crc kubenswrapper[4870]: I1014 07:02:29.586730 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:29 crc kubenswrapper[4870]: I1014 07:02:29.586809 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:29 crc kubenswrapper[4870]: I1014 07:02:29.586829 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:29 crc kubenswrapper[4870]: I1014 07:02:29.586854 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:29 crc kubenswrapper[4870]: I1014 07:02:29.586871 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:29Z","lastTransitionTime":"2025-10-14T07:02:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:29 crc kubenswrapper[4870]: I1014 07:02:29.690920 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:29 crc kubenswrapper[4870]: I1014 07:02:29.690981 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:29 crc kubenswrapper[4870]: I1014 07:02:29.690999 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:29 crc kubenswrapper[4870]: I1014 07:02:29.691023 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:29 crc kubenswrapper[4870]: I1014 07:02:29.691042 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:29Z","lastTransitionTime":"2025-10-14T07:02:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:29 crc kubenswrapper[4870]: I1014 07:02:29.793194 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:29 crc kubenswrapper[4870]: I1014 07:02:29.793228 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:29 crc kubenswrapper[4870]: I1014 07:02:29.793239 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:29 crc kubenswrapper[4870]: I1014 07:02:29.793255 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:29 crc kubenswrapper[4870]: I1014 07:02:29.793266 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:29Z","lastTransitionTime":"2025-10-14T07:02:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:29 crc kubenswrapper[4870]: I1014 07:02:29.896010 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:29 crc kubenswrapper[4870]: I1014 07:02:29.896076 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:29 crc kubenswrapper[4870]: I1014 07:02:29.896093 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:29 crc kubenswrapper[4870]: I1014 07:02:29.896117 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:29 crc kubenswrapper[4870]: I1014 07:02:29.896136 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:29Z","lastTransitionTime":"2025-10-14T07:02:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:29 crc kubenswrapper[4870]: I1014 07:02:29.948757 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 07:02:29 crc kubenswrapper[4870]: E1014 07:02:29.949010 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 07:03:33.948976668 +0000 UTC m=+149.646337079 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:02:30 crc kubenswrapper[4870]: I1014 07:02:30.000123 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:30 crc kubenswrapper[4870]: I1014 07:02:30.000176 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:30 crc kubenswrapper[4870]: I1014 07:02:30.000194 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:30 crc kubenswrapper[4870]: I1014 07:02:30.000217 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:30 crc kubenswrapper[4870]: I1014 07:02:30.000236 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:30Z","lastTransitionTime":"2025-10-14T07:02:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:30 crc kubenswrapper[4870]: I1014 07:02:30.033610 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 07:02:30 crc kubenswrapper[4870]: I1014 07:02:30.033641 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 07:02:30 crc kubenswrapper[4870]: I1014 07:02:30.033610 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 07:02:30 crc kubenswrapper[4870]: E1014 07:02:30.033794 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 07:02:30 crc kubenswrapper[4870]: E1014 07:02:30.034098 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 07:02:30 crc kubenswrapper[4870]: E1014 07:02:30.034283 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 07:02:30 crc kubenswrapper[4870]: I1014 07:02:30.050777 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 07:02:30 crc kubenswrapper[4870]: I1014 07:02:30.050880 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 07:02:30 crc kubenswrapper[4870]: E1014 07:02:30.051025 4870 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 14 07:02:30 crc kubenswrapper[4870]: E1014 07:02:30.051125 4870 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 14 07:02:30 crc kubenswrapper[4870]: I1014 07:02:30.051030 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 07:02:30 crc kubenswrapper[4870]: E1014 07:02:30.051140 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-14 07:03:34.051104305 +0000 UTC m=+149.748464736 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 14 07:02:30 crc kubenswrapper[4870]: E1014 07:02:30.051275 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-14 07:03:34.051247278 +0000 UTC m=+149.748607699 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 14 07:02:30 crc kubenswrapper[4870]: I1014 07:02:30.051336 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 07:02:30 crc kubenswrapper[4870]: E1014 07:02:30.051378 4870 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 14 07:02:30 crc kubenswrapper[4870]: E1014 07:02:30.051573 4870 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 14 07:02:30 crc kubenswrapper[4870]: E1014 07:02:30.051607 4870 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 14 07:02:30 crc kubenswrapper[4870]: E1014 07:02:30.051632 4870 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 14 07:02:30 crc kubenswrapper[4870]: E1014 07:02:30.051735 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-14 07:03:34.05171159 +0000 UTC m=+149.749072001 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 14 07:02:30 crc kubenswrapper[4870]: E1014 07:02:30.051850 4870 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 14 07:02:30 crc kubenswrapper[4870]: E1014 07:02:30.051937 4870 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 14 07:02:30 crc kubenswrapper[4870]: E1014 07:02:30.052021 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-14 07:03:34.051995397 +0000 UTC m=+149.749355818 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 14 07:02:30 crc kubenswrapper[4870]: I1014 07:02:30.103491 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:30 crc kubenswrapper[4870]: I1014 07:02:30.103564 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:30 crc kubenswrapper[4870]: I1014 07:02:30.103583 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:30 crc kubenswrapper[4870]: I1014 07:02:30.103610 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:30 crc kubenswrapper[4870]: I1014 07:02:30.103628 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:30Z","lastTransitionTime":"2025-10-14T07:02:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:30 crc kubenswrapper[4870]: I1014 07:02:30.206178 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:30 crc kubenswrapper[4870]: I1014 07:02:30.206230 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:30 crc kubenswrapper[4870]: I1014 07:02:30.206252 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:30 crc kubenswrapper[4870]: I1014 07:02:30.206289 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:30 crc kubenswrapper[4870]: I1014 07:02:30.206325 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:30Z","lastTransitionTime":"2025-10-14T07:02:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:30 crc kubenswrapper[4870]: I1014 07:02:30.310018 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:30 crc kubenswrapper[4870]: I1014 07:02:30.310075 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:30 crc kubenswrapper[4870]: I1014 07:02:30.310097 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:30 crc kubenswrapper[4870]: I1014 07:02:30.310127 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:30 crc kubenswrapper[4870]: I1014 07:02:30.310149 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:30Z","lastTransitionTime":"2025-10-14T07:02:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:30 crc kubenswrapper[4870]: I1014 07:02:30.412593 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:30 crc kubenswrapper[4870]: I1014 07:02:30.412625 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:30 crc kubenswrapper[4870]: I1014 07:02:30.412634 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:30 crc kubenswrapper[4870]: I1014 07:02:30.412646 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:30 crc kubenswrapper[4870]: I1014 07:02:30.412655 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:30Z","lastTransitionTime":"2025-10-14T07:02:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:30 crc kubenswrapper[4870]: I1014 07:02:30.514574 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:30 crc kubenswrapper[4870]: I1014 07:02:30.514626 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:30 crc kubenswrapper[4870]: I1014 07:02:30.514644 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:30 crc kubenswrapper[4870]: I1014 07:02:30.514668 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:30 crc kubenswrapper[4870]: I1014 07:02:30.514686 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:30Z","lastTransitionTime":"2025-10-14T07:02:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:30 crc kubenswrapper[4870]: I1014 07:02:30.617204 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:30 crc kubenswrapper[4870]: I1014 07:02:30.617261 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:30 crc kubenswrapper[4870]: I1014 07:02:30.617280 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:30 crc kubenswrapper[4870]: I1014 07:02:30.617301 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:30 crc kubenswrapper[4870]: I1014 07:02:30.617321 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:30Z","lastTransitionTime":"2025-10-14T07:02:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:30 crc kubenswrapper[4870]: I1014 07:02:30.719845 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:30 crc kubenswrapper[4870]: I1014 07:02:30.719905 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:30 crc kubenswrapper[4870]: I1014 07:02:30.719923 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:30 crc kubenswrapper[4870]: I1014 07:02:30.719951 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:30 crc kubenswrapper[4870]: I1014 07:02:30.719969 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:30Z","lastTransitionTime":"2025-10-14T07:02:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:30 crc kubenswrapper[4870]: I1014 07:02:30.823605 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:30 crc kubenswrapper[4870]: I1014 07:02:30.823683 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:30 crc kubenswrapper[4870]: I1014 07:02:30.823706 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:30 crc kubenswrapper[4870]: I1014 07:02:30.823734 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:30 crc kubenswrapper[4870]: I1014 07:02:30.823757 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:30Z","lastTransitionTime":"2025-10-14T07:02:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:30 crc kubenswrapper[4870]: I1014 07:02:30.926924 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:30 crc kubenswrapper[4870]: I1014 07:02:30.926988 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:30 crc kubenswrapper[4870]: I1014 07:02:30.927014 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:30 crc kubenswrapper[4870]: I1014 07:02:30.927045 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:30 crc kubenswrapper[4870]: I1014 07:02:30.927067 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:30Z","lastTransitionTime":"2025-10-14T07:02:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:31 crc kubenswrapper[4870]: I1014 07:02:31.030962 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:31 crc kubenswrapper[4870]: I1014 07:02:31.031025 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:31 crc kubenswrapper[4870]: I1014 07:02:31.031045 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:31 crc kubenswrapper[4870]: I1014 07:02:31.031069 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:31 crc kubenswrapper[4870]: I1014 07:02:31.031088 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:31Z","lastTransitionTime":"2025-10-14T07:02:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:31 crc kubenswrapper[4870]: I1014 07:02:31.033243 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zc5j6" Oct 14 07:02:31 crc kubenswrapper[4870]: E1014 07:02:31.033418 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zc5j6" podUID="7d7a4f41-388a-4320-b896-43c17ff13da6" Oct 14 07:02:31 crc kubenswrapper[4870]: I1014 07:02:31.134632 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:31 crc kubenswrapper[4870]: I1014 07:02:31.134733 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:31 crc kubenswrapper[4870]: I1014 07:02:31.134782 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:31 crc kubenswrapper[4870]: I1014 07:02:31.134807 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:31 crc kubenswrapper[4870]: I1014 07:02:31.134826 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:31Z","lastTransitionTime":"2025-10-14T07:02:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:31 crc kubenswrapper[4870]: I1014 07:02:31.238140 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:31 crc kubenswrapper[4870]: I1014 07:02:31.238215 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:31 crc kubenswrapper[4870]: I1014 07:02:31.238239 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:31 crc kubenswrapper[4870]: I1014 07:02:31.238268 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:31 crc kubenswrapper[4870]: I1014 07:02:31.238295 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:31Z","lastTransitionTime":"2025-10-14T07:02:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:31 crc kubenswrapper[4870]: I1014 07:02:31.341289 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:31 crc kubenswrapper[4870]: I1014 07:02:31.341363 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:31 crc kubenswrapper[4870]: I1014 07:02:31.341382 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:31 crc kubenswrapper[4870]: I1014 07:02:31.341406 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:31 crc kubenswrapper[4870]: I1014 07:02:31.341424 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:31Z","lastTransitionTime":"2025-10-14T07:02:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:31 crc kubenswrapper[4870]: I1014 07:02:31.445680 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:31 crc kubenswrapper[4870]: I1014 07:02:31.445762 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:31 crc kubenswrapper[4870]: I1014 07:02:31.445789 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:31 crc kubenswrapper[4870]: I1014 07:02:31.445819 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:31 crc kubenswrapper[4870]: I1014 07:02:31.445842 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:31Z","lastTransitionTime":"2025-10-14T07:02:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:31 crc kubenswrapper[4870]: I1014 07:02:31.548102 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:31 crc kubenswrapper[4870]: I1014 07:02:31.548168 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:31 crc kubenswrapper[4870]: I1014 07:02:31.548181 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:31 crc kubenswrapper[4870]: I1014 07:02:31.548199 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:31 crc kubenswrapper[4870]: I1014 07:02:31.548210 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:31Z","lastTransitionTime":"2025-10-14T07:02:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:31 crc kubenswrapper[4870]: I1014 07:02:31.651302 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:31 crc kubenswrapper[4870]: I1014 07:02:31.651401 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:31 crc kubenswrapper[4870]: I1014 07:02:31.651434 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:31 crc kubenswrapper[4870]: I1014 07:02:31.651505 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:31 crc kubenswrapper[4870]: I1014 07:02:31.651528 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:31Z","lastTransitionTime":"2025-10-14T07:02:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:31 crc kubenswrapper[4870]: I1014 07:02:31.754369 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:31 crc kubenswrapper[4870]: I1014 07:02:31.754414 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:31 crc kubenswrapper[4870]: I1014 07:02:31.754427 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:31 crc kubenswrapper[4870]: I1014 07:02:31.754473 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:31 crc kubenswrapper[4870]: I1014 07:02:31.754487 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:31Z","lastTransitionTime":"2025-10-14T07:02:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:31 crc kubenswrapper[4870]: I1014 07:02:31.857838 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:31 crc kubenswrapper[4870]: I1014 07:02:31.857885 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:31 crc kubenswrapper[4870]: I1014 07:02:31.857904 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:31 crc kubenswrapper[4870]: I1014 07:02:31.857929 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:31 crc kubenswrapper[4870]: I1014 07:02:31.857947 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:31Z","lastTransitionTime":"2025-10-14T07:02:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:31 crc kubenswrapper[4870]: I1014 07:02:31.961986 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:31 crc kubenswrapper[4870]: I1014 07:02:31.962066 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:31 crc kubenswrapper[4870]: I1014 07:02:31.962084 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:31 crc kubenswrapper[4870]: I1014 07:02:31.962114 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:31 crc kubenswrapper[4870]: I1014 07:02:31.962138 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:31Z","lastTransitionTime":"2025-10-14T07:02:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:32 crc kubenswrapper[4870]: I1014 07:02:32.033113 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 07:02:32 crc kubenswrapper[4870]: E1014 07:02:32.033348 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 07:02:32 crc kubenswrapper[4870]: I1014 07:02:32.033422 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 07:02:32 crc kubenswrapper[4870]: I1014 07:02:32.033432 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 07:02:32 crc kubenswrapper[4870]: E1014 07:02:32.033847 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 07:02:32 crc kubenswrapper[4870]: E1014 07:02:32.033976 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 07:02:32 crc kubenswrapper[4870]: I1014 07:02:32.065795 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:32 crc kubenswrapper[4870]: I1014 07:02:32.065837 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:32 crc kubenswrapper[4870]: I1014 07:02:32.065849 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:32 crc kubenswrapper[4870]: I1014 07:02:32.065868 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:32 crc kubenswrapper[4870]: I1014 07:02:32.065882 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:32Z","lastTransitionTime":"2025-10-14T07:02:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:32 crc kubenswrapper[4870]: I1014 07:02:32.169485 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:32 crc kubenswrapper[4870]: I1014 07:02:32.169532 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:32 crc kubenswrapper[4870]: I1014 07:02:32.169551 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:32 crc kubenswrapper[4870]: I1014 07:02:32.169571 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:32 crc kubenswrapper[4870]: I1014 07:02:32.169583 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:32Z","lastTransitionTime":"2025-10-14T07:02:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:32 crc kubenswrapper[4870]: I1014 07:02:32.272384 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:32 crc kubenswrapper[4870]: I1014 07:02:32.272429 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:32 crc kubenswrapper[4870]: I1014 07:02:32.272478 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:32 crc kubenswrapper[4870]: I1014 07:02:32.272498 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:32 crc kubenswrapper[4870]: I1014 07:02:32.272510 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:32Z","lastTransitionTime":"2025-10-14T07:02:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:32 crc kubenswrapper[4870]: I1014 07:02:32.375847 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:32 crc kubenswrapper[4870]: I1014 07:02:32.375917 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:32 crc kubenswrapper[4870]: I1014 07:02:32.375929 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:32 crc kubenswrapper[4870]: I1014 07:02:32.375951 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:32 crc kubenswrapper[4870]: I1014 07:02:32.375965 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:32Z","lastTransitionTime":"2025-10-14T07:02:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:32 crc kubenswrapper[4870]: I1014 07:02:32.479740 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:32 crc kubenswrapper[4870]: I1014 07:02:32.479805 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:32 crc kubenswrapper[4870]: I1014 07:02:32.479823 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:32 crc kubenswrapper[4870]: I1014 07:02:32.479848 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:32 crc kubenswrapper[4870]: I1014 07:02:32.479866 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:32Z","lastTransitionTime":"2025-10-14T07:02:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:32 crc kubenswrapper[4870]: I1014 07:02:32.583606 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:32 crc kubenswrapper[4870]: I1014 07:02:32.583675 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:32 crc kubenswrapper[4870]: I1014 07:02:32.583696 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:32 crc kubenswrapper[4870]: I1014 07:02:32.583728 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:32 crc kubenswrapper[4870]: I1014 07:02:32.583749 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:32Z","lastTransitionTime":"2025-10-14T07:02:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:32 crc kubenswrapper[4870]: I1014 07:02:32.687686 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:32 crc kubenswrapper[4870]: I1014 07:02:32.687751 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:32 crc kubenswrapper[4870]: I1014 07:02:32.687771 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:32 crc kubenswrapper[4870]: I1014 07:02:32.687797 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:32 crc kubenswrapper[4870]: I1014 07:02:32.687815 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:32Z","lastTransitionTime":"2025-10-14T07:02:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:32 crc kubenswrapper[4870]: I1014 07:02:32.792618 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:32 crc kubenswrapper[4870]: I1014 07:02:32.792686 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:32 crc kubenswrapper[4870]: I1014 07:02:32.792711 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:32 crc kubenswrapper[4870]: I1014 07:02:32.792746 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:32 crc kubenswrapper[4870]: I1014 07:02:32.792770 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:32Z","lastTransitionTime":"2025-10-14T07:02:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:32 crc kubenswrapper[4870]: I1014 07:02:32.896345 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:32 crc kubenswrapper[4870]: I1014 07:02:32.896398 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:32 crc kubenswrapper[4870]: I1014 07:02:32.896415 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:32 crc kubenswrapper[4870]: I1014 07:02:32.896480 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:32 crc kubenswrapper[4870]: I1014 07:02:32.896497 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:32Z","lastTransitionTime":"2025-10-14T07:02:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:33 crc kubenswrapper[4870]: I1014 07:02:33.001409 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:33 crc kubenswrapper[4870]: I1014 07:02:33.001464 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:33 crc kubenswrapper[4870]: I1014 07:02:33.001475 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:33 crc kubenswrapper[4870]: I1014 07:02:33.001491 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:33 crc kubenswrapper[4870]: I1014 07:02:33.001502 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:33Z","lastTransitionTime":"2025-10-14T07:02:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:33 crc kubenswrapper[4870]: I1014 07:02:33.033632 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zc5j6" Oct 14 07:02:33 crc kubenswrapper[4870]: E1014 07:02:33.033874 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zc5j6" podUID="7d7a4f41-388a-4320-b896-43c17ff13da6" Oct 14 07:02:33 crc kubenswrapper[4870]: I1014 07:02:33.105013 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:33 crc kubenswrapper[4870]: I1014 07:02:33.105073 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:33 crc kubenswrapper[4870]: I1014 07:02:33.105098 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:33 crc kubenswrapper[4870]: I1014 07:02:33.105128 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:33 crc kubenswrapper[4870]: I1014 07:02:33.105149 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:33Z","lastTransitionTime":"2025-10-14T07:02:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:33 crc kubenswrapper[4870]: I1014 07:02:33.208434 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:33 crc kubenswrapper[4870]: I1014 07:02:33.208536 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:33 crc kubenswrapper[4870]: I1014 07:02:33.208593 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:33 crc kubenswrapper[4870]: I1014 07:02:33.208619 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:33 crc kubenswrapper[4870]: I1014 07:02:33.208636 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:33Z","lastTransitionTime":"2025-10-14T07:02:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:33 crc kubenswrapper[4870]: I1014 07:02:33.311787 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:33 crc kubenswrapper[4870]: I1014 07:02:33.311890 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:33 crc kubenswrapper[4870]: I1014 07:02:33.311909 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:33 crc kubenswrapper[4870]: I1014 07:02:33.311966 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:33 crc kubenswrapper[4870]: I1014 07:02:33.311992 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:33Z","lastTransitionTime":"2025-10-14T07:02:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:33 crc kubenswrapper[4870]: I1014 07:02:33.415630 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:33 crc kubenswrapper[4870]: I1014 07:02:33.415719 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:33 crc kubenswrapper[4870]: I1014 07:02:33.415766 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:33 crc kubenswrapper[4870]: I1014 07:02:33.415790 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:33 crc kubenswrapper[4870]: I1014 07:02:33.415807 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:33Z","lastTransitionTime":"2025-10-14T07:02:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:33 crc kubenswrapper[4870]: I1014 07:02:33.518848 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:33 crc kubenswrapper[4870]: I1014 07:02:33.518938 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:33 crc kubenswrapper[4870]: I1014 07:02:33.518986 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:33 crc kubenswrapper[4870]: I1014 07:02:33.519012 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:33 crc kubenswrapper[4870]: I1014 07:02:33.519029 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:33Z","lastTransitionTime":"2025-10-14T07:02:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:33 crc kubenswrapper[4870]: I1014 07:02:33.622512 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:33 crc kubenswrapper[4870]: I1014 07:02:33.622581 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:33 crc kubenswrapper[4870]: I1014 07:02:33.622606 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:33 crc kubenswrapper[4870]: I1014 07:02:33.622636 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:33 crc kubenswrapper[4870]: I1014 07:02:33.622660 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:33Z","lastTransitionTime":"2025-10-14T07:02:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:33 crc kubenswrapper[4870]: I1014 07:02:33.725874 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:33 crc kubenswrapper[4870]: I1014 07:02:33.725937 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:33 crc kubenswrapper[4870]: I1014 07:02:33.725971 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:33 crc kubenswrapper[4870]: I1014 07:02:33.726001 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:33 crc kubenswrapper[4870]: I1014 07:02:33.726023 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:33Z","lastTransitionTime":"2025-10-14T07:02:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:33 crc kubenswrapper[4870]: I1014 07:02:33.829176 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:33 crc kubenswrapper[4870]: I1014 07:02:33.829236 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:33 crc kubenswrapper[4870]: I1014 07:02:33.829255 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:33 crc kubenswrapper[4870]: I1014 07:02:33.829281 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:33 crc kubenswrapper[4870]: I1014 07:02:33.829298 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:33Z","lastTransitionTime":"2025-10-14T07:02:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:33 crc kubenswrapper[4870]: I1014 07:02:33.932001 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:33 crc kubenswrapper[4870]: I1014 07:02:33.932051 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:33 crc kubenswrapper[4870]: I1014 07:02:33.932068 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:33 crc kubenswrapper[4870]: I1014 07:02:33.932092 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:33 crc kubenswrapper[4870]: I1014 07:02:33.932108 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:33Z","lastTransitionTime":"2025-10-14T07:02:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:34 crc kubenswrapper[4870]: I1014 07:02:34.033171 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 07:02:34 crc kubenswrapper[4870]: E1014 07:02:34.033957 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 07:02:34 crc kubenswrapper[4870]: I1014 07:02:34.033208 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 07:02:34 crc kubenswrapper[4870]: I1014 07:02:34.033180 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 07:02:34 crc kubenswrapper[4870]: E1014 07:02:34.035206 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 07:02:34 crc kubenswrapper[4870]: E1014 07:02:34.035044 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 07:02:34 crc kubenswrapper[4870]: I1014 07:02:34.035478 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:34 crc kubenswrapper[4870]: I1014 07:02:34.035582 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:34 crc kubenswrapper[4870]: I1014 07:02:34.035603 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:34 crc kubenswrapper[4870]: I1014 07:02:34.035667 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:34 crc kubenswrapper[4870]: I1014 07:02:34.035684 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:34Z","lastTransitionTime":"2025-10-14T07:02:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:34 crc kubenswrapper[4870]: I1014 07:02:34.139610 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:34 crc kubenswrapper[4870]: I1014 07:02:34.140031 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:34 crc kubenswrapper[4870]: I1014 07:02:34.140242 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:34 crc kubenswrapper[4870]: I1014 07:02:34.140488 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:34 crc kubenswrapper[4870]: I1014 07:02:34.140728 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:34Z","lastTransitionTime":"2025-10-14T07:02:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:34 crc kubenswrapper[4870]: I1014 07:02:34.238505 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:34 crc kubenswrapper[4870]: I1014 07:02:34.239028 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:34 crc kubenswrapper[4870]: I1014 07:02:34.239179 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:34 crc kubenswrapper[4870]: I1014 07:02:34.239483 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:34 crc kubenswrapper[4870]: I1014 07:02:34.239729 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:34Z","lastTransitionTime":"2025-10-14T07:02:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:34 crc kubenswrapper[4870]: E1014 07:02:34.260374 4870 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:02:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:02:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:02:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:02:34Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:02:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:02:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:02:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:02:34Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"56448e6b-6b3a-4bb9-91f8-2f0a9b44e9aa\\\",\\\"systemUUID\\\":\\\"0adb347d-b90d-4e9f-9962-f1bd97980aed\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:34Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:34 crc kubenswrapper[4870]: I1014 07:02:34.265262 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:34 crc kubenswrapper[4870]: I1014 07:02:34.265685 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:34 crc kubenswrapper[4870]: I1014 07:02:34.265863 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:34 crc kubenswrapper[4870]: I1014 07:02:34.265995 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:34 crc kubenswrapper[4870]: I1014 07:02:34.266145 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:34Z","lastTransitionTime":"2025-10-14T07:02:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:34 crc kubenswrapper[4870]: E1014 07:02:34.287076 4870 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:02:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:02:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:02:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:02:34Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:02:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:02:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:02:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:02:34Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"56448e6b-6b3a-4bb9-91f8-2f0a9b44e9aa\\\",\\\"systemUUID\\\":\\\"0adb347d-b90d-4e9f-9962-f1bd97980aed\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:34Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:34 crc kubenswrapper[4870]: I1014 07:02:34.291679 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:34 crc kubenswrapper[4870]: I1014 07:02:34.291731 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:34 crc kubenswrapper[4870]: I1014 07:02:34.291745 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:34 crc kubenswrapper[4870]: I1014 07:02:34.291766 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:34 crc kubenswrapper[4870]: I1014 07:02:34.291782 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:34Z","lastTransitionTime":"2025-10-14T07:02:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:34 crc kubenswrapper[4870]: E1014 07:02:34.307137 4870 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:02:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:02:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:02:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:02:34Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:02:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:02:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:02:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:02:34Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"56448e6b-6b3a-4bb9-91f8-2f0a9b44e9aa\\\",\\\"systemUUID\\\":\\\"0adb347d-b90d-4e9f-9962-f1bd97980aed\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:34Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:34 crc kubenswrapper[4870]: I1014 07:02:34.314193 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:34 crc kubenswrapper[4870]: I1014 07:02:34.314421 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:34 crc kubenswrapper[4870]: I1014 07:02:34.314464 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:34 crc kubenswrapper[4870]: I1014 07:02:34.314491 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:34 crc kubenswrapper[4870]: I1014 07:02:34.314506 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:34Z","lastTransitionTime":"2025-10-14T07:02:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:34 crc kubenswrapper[4870]: E1014 07:02:34.338012 4870 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:02:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:02:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:02:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:02:34Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:02:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:02:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:02:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:02:34Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"56448e6b-6b3a-4bb9-91f8-2f0a9b44e9aa\\\",\\\"systemUUID\\\":\\\"0adb347d-b90d-4e9f-9962-f1bd97980aed\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:34Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:34 crc kubenswrapper[4870]: I1014 07:02:34.342335 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:34 crc kubenswrapper[4870]: I1014 07:02:34.342365 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:34 crc kubenswrapper[4870]: I1014 07:02:34.342378 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:34 crc kubenswrapper[4870]: I1014 07:02:34.342399 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:34 crc kubenswrapper[4870]: I1014 07:02:34.342414 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:34Z","lastTransitionTime":"2025-10-14T07:02:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:34 crc kubenswrapper[4870]: E1014 07:02:34.361723 4870 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:02:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:02:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:02:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:02:34Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:02:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:02:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-14T07:02:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-14T07:02:34Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"56448e6b-6b3a-4bb9-91f8-2f0a9b44e9aa\\\",\\\"systemUUID\\\":\\\"0adb347d-b90d-4e9f-9962-f1bd97980aed\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:34Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:34 crc kubenswrapper[4870]: E1014 07:02:34.361873 4870 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 14 07:02:34 crc kubenswrapper[4870]: I1014 07:02:34.363669 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:34 crc kubenswrapper[4870]: I1014 07:02:34.363700 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:34 crc kubenswrapper[4870]: I1014 07:02:34.363711 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:34 crc kubenswrapper[4870]: I1014 07:02:34.363726 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:34 crc kubenswrapper[4870]: I1014 07:02:34.363738 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:34Z","lastTransitionTime":"2025-10-14T07:02:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:34 crc kubenswrapper[4870]: I1014 07:02:34.466488 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:34 crc kubenswrapper[4870]: I1014 07:02:34.466537 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:34 crc kubenswrapper[4870]: I1014 07:02:34.466573 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:34 crc kubenswrapper[4870]: I1014 07:02:34.466591 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:34 crc kubenswrapper[4870]: I1014 07:02:34.466603 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:34Z","lastTransitionTime":"2025-10-14T07:02:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:34 crc kubenswrapper[4870]: I1014 07:02:34.569495 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:34 crc kubenswrapper[4870]: I1014 07:02:34.569543 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:34 crc kubenswrapper[4870]: I1014 07:02:34.569556 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:34 crc kubenswrapper[4870]: I1014 07:02:34.569573 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:34 crc kubenswrapper[4870]: I1014 07:02:34.569585 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:34Z","lastTransitionTime":"2025-10-14T07:02:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:34 crc kubenswrapper[4870]: I1014 07:02:34.672485 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:34 crc kubenswrapper[4870]: I1014 07:02:34.672561 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:34 crc kubenswrapper[4870]: I1014 07:02:34.672593 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:34 crc kubenswrapper[4870]: I1014 07:02:34.672625 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:34 crc kubenswrapper[4870]: I1014 07:02:34.672656 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:34Z","lastTransitionTime":"2025-10-14T07:02:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:34 crc kubenswrapper[4870]: I1014 07:02:34.776751 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:34 crc kubenswrapper[4870]: I1014 07:02:34.776874 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:34 crc kubenswrapper[4870]: I1014 07:02:34.776893 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:34 crc kubenswrapper[4870]: I1014 07:02:34.776925 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:34 crc kubenswrapper[4870]: I1014 07:02:34.776946 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:34Z","lastTransitionTime":"2025-10-14T07:02:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:34 crc kubenswrapper[4870]: I1014 07:02:34.880388 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:34 crc kubenswrapper[4870]: I1014 07:02:34.880498 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:34 crc kubenswrapper[4870]: I1014 07:02:34.880519 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:34 crc kubenswrapper[4870]: I1014 07:02:34.880554 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:34 crc kubenswrapper[4870]: I1014 07:02:34.880577 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:34Z","lastTransitionTime":"2025-10-14T07:02:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:34 crc kubenswrapper[4870]: I1014 07:02:34.983582 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:34 crc kubenswrapper[4870]: I1014 07:02:34.983632 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:34 crc kubenswrapper[4870]: I1014 07:02:34.983643 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:34 crc kubenswrapper[4870]: I1014 07:02:34.983661 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:34 crc kubenswrapper[4870]: I1014 07:02:34.983678 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:34Z","lastTransitionTime":"2025-10-14T07:02:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:35 crc kubenswrapper[4870]: I1014 07:02:35.033261 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zc5j6" Oct 14 07:02:35 crc kubenswrapper[4870]: E1014 07:02:35.033576 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zc5j6" podUID="7d7a4f41-388a-4320-b896-43c17ff13da6" Oct 14 07:02:35 crc kubenswrapper[4870]: I1014 07:02:35.050324 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-zc5j6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d7a4f41-388a-4320-b896-43c17ff13da6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:40Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6p2wh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6p2wh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:40Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-zc5j6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:35Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:35 crc kubenswrapper[4870]: I1014 07:02:35.072910 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:35Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:35 crc kubenswrapper[4870]: I1014 07:02:35.086155 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:35 crc kubenswrapper[4870]: I1014 07:02:35.086226 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:35 crc kubenswrapper[4870]: I1014 07:02:35.086249 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:35 crc kubenswrapper[4870]: I1014 07:02:35.086279 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:35 crc kubenswrapper[4870]: I1014 07:02:35.086305 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:35Z","lastTransitionTime":"2025-10-14T07:02:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:35 crc kubenswrapper[4870]: I1014 07:02:35.092860 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:29Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:29Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3eb0b397c05ffd6650607f1be4a75683e0342d3333fc096bf1cd63c7688834f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:35Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:35 crc kubenswrapper[4870]: I1014 07:02:35.110639 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"874a5e45-dffd-4d17-b609-b7d3ed2eab07\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d98d044836ab44a405196915ef7bf572b75113f2c81ec7e64effa63b47ad3543\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wj4ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8004752575df4a1c44efbcafe64799f09126d0f97285d92388c76a0b90e022f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wj4ks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7tvc7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:35Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:35 crc kubenswrapper[4870]: I1014 07:02:35.141170 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-v7br2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dcaef0e-42e3-424b-bbe5-4a023efe912d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71dec55e859b11a081dbb90d7d6cf6b4846cf591382303d8329003ff50c71f2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e72cff0723e2ce520340244b4e597d8016c58e1b071c4b087f2f6cd718741594\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc9d4db944f837289d1923176d60cfc4a566a282c2d2e3a703e220df884703c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d16ed1c65925549dcb69ab02c3013e25e3f123a83179f1af3a8e08bf91a59896\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed59243b59c39ce82bf9bc8f33c0a8b84854846afff6a89f5d37ffc97a096b7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3b6d20cf459b1f6d3aff5b9290f9391909baf27ee40c49e2b0cd2d5cfe889ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab4295721b6ec6d64040d321899ab478c612e91cbc3b99b33753869578a2d1e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ab4295721b6ec6d64040d321899ab478c612e91cbc3b99b33753869578a2d1e8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-14T07:02:27Z\\\",\\\"message\\\":\\\"ckageserver-service]} name:Service_openshift-operator-lifecycle-manager/packageserver-service_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.153:5443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {5e50827b-d271-442b-b8a7-7f33b2cd6b11}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1014 07:02:27.009850 6956 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-operator-lifecycle-manager/packageserver-service]} name:Service_openshift-operator-lifecycle-manager/packageserver-service_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.153:5443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {5e50827b-d271-442b-b8a7-7f33b2cd6b11}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1014 07:02:27.009895 6956 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T07:02:26Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-v7br2_openshift-ovn-kubernetes(8dcaef0e-42e3-424b-bbe5-4a023efe912d)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9fe3468cdd612fc4378e1f6f55a2a9d00274f7ddb65cb57b2a368dfc065cfb4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ba24c2e067b9f3a14d9d018b66300b54d866530d30174fbdb3c899299575d28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ba24c2e067b9f3a14d9d018b66300b54d866530d30174fbdb3c899299575d28\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-glrll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-v7br2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:35Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:35 crc kubenswrapper[4870]: I1014 07:02:35.160085 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"77a0d93a-5391-4112-b502-c474f45de209\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7815af9a584f6db6660073da289f2e8c987a04ef9ddfac155bdd9a15b014fcfa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b329b570622155a55f70501b2260e278106a68ebef497914231b4ec860e0125\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ba942a0e6fe5ab93a187e0b7b4bd7c440c2e7827d9b5323fc50f90b562075392\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b9018b078c5c8276243c1696860f0aa73e2d8dd917913ae9d03b30dc9b5113a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://978e8db184ba9e430232c6bb2ee530beb92567fdc71ba3c35b0bedca3492d845\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"message\\\":\\\"file observer\\\\nW1014 07:01:25.870672 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1014 07:01:25.870795 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1014 07:01:25.871521 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-756714914/tls.crt::/tmp/serving-cert-756714914/tls.key\\\\\\\"\\\\nI1014 07:01:26.460037 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1014 07:01:26.464876 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1014 07:01:26.464902 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1014 07:01:26.464925 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1014 07:01:26.464931 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1014 07:01:26.481374 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1014 07:01:26.481400 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:01:26.481405 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1014 07:01:26.481410 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1014 07:01:26.481414 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1014 07:01:26.481417 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1014 07:01:26.481420 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1014 07:01:26.481610 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1014 07:01:26.482872 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:10Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bd0044a247bdc5ca4cab7294103af3270b76cde61f3109b85e83ed050ca53b6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82d3c2214e0be679d2827594b159ace362151318689078bc49d3158f42899636\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82d3c2214e0be679d2827594b159ace362151318689078bc49d3158f42899636\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:05Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:35Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:35 crc kubenswrapper[4870]: I1014 07:02:35.174597 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:35Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:35 crc kubenswrapper[4870]: I1014 07:02:35.187594 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e26fe49ec2045d20a0cb2dd17e96e79b6eb43c641eb184c688cc472ca5db39d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aea115f7ee71c700b218fe15c928af7b55e6babe70e65d67f125d87be1fd2e2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:35Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:35 crc kubenswrapper[4870]: I1014 07:02:35.188936 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:35 crc kubenswrapper[4870]: I1014 07:02:35.189008 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:35 crc kubenswrapper[4870]: I1014 07:02:35.189023 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:35 crc kubenswrapper[4870]: I1014 07:02:35.189043 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:35 crc kubenswrapper[4870]: I1014 07:02:35.189057 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:35Z","lastTransitionTime":"2025-10-14T07:02:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:35 crc kubenswrapper[4870]: I1014 07:02:35.199172 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-4v2v2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4796ba6-c23b-4764-be96-2e3a3bf5e962\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf5afb0b99a7101f6d5573fb7eb925c5a1d8becc11b2e3c6f3351d80b6f0e823\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnp8b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-4v2v2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:35Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:35 crc kubenswrapper[4870]: I1014 07:02:35.216422 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-cdjmj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"612e7ab9-bda4-45d8-bc50-ff1afb68aace\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7776352d62b814fc63d01e4c16ad13f878b0597eb37a370bf9da0a88abab6cfa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://834444b5350ee12df4e4652c3009880d8d69e1edf8ba0c0156ab96fb2bbe28e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://834444b5350ee12df4e4652c3009880d8d69e1edf8ba0c0156ab96fb2bbe28e1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c1972e7b2f75790db22a07a078eee97072bffbdb7f8e594bf2be5f800632a7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c1972e7b2f75790db22a07a078eee97072bffbdb7f8e594bf2be5f800632a7b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b124b42a44c6a23da1d92b9a5326987eb74430f80489a71e8913829475654ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b124b42a44c6a23da1d92b9a5326987eb74430f80489a71e8913829475654ab\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62d0deb6b20ea65fbcbaf94c2fc14d69adb22e7a46cc043ccd4def2668ddd4b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62d0deb6b20ea65fbcbaf94c2fc14d69adb22e7a46cc043ccd4def2668ddd4b7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7f0d7cf29ca737d0effe58731a8b890379527c97f78c144d4eebb910260fadb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7f0d7cf29ca737d0effe58731a8b890379527c97f78c144d4eebb910260fadb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5959abb7c8e7ab32bed3aca709ae0334e8400f991ba724246c9fda56cbed822\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5959abb7c8e7ab32bed3aca709ae0334e8400f991ba724246c9fda56cbed822\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcfq9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-cdjmj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:35Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:35 crc kubenswrapper[4870]: I1014 07:02:35.237622 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fdxjh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a3a430a-b540-4a8c-adad-b2893dbf9898\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:02:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:02:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://88614373bfe4779bbe61cc41cc4c35f419e438bdae3cec4964271bf0f74ad3f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://93e41d5e927eed7ec62eb9c3607cc804cd8999556663363fd123b2a784b00ff1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-14T07:02:13Z\\\",\\\"message\\\":\\\"2025-10-14T07:01:28+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_847000ef-6815-4615-8665-86c348661ad2\\\\n2025-10-14T07:01:28+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_847000ef-6815-4615-8665-86c348661ad2 to /host/opt/cni/bin/\\\\n2025-10-14T07:01:28Z [verbose] multus-daemon started\\\\n2025-10-14T07:01:28Z [verbose] Readiness Indicator file check\\\\n2025-10-14T07:02:13Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:27Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:02:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k787j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fdxjh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:35Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:35 crc kubenswrapper[4870]: I1014 07:02:35.251614 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-khtq8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28deeea6-a1ef-4a4a-a377-d954ccd87de9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://296bb048a1d99649beb57d609573833bf547cd4d628ebc823c6af83fe5736c9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rb79p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2b7bc4c93307e80b778f2282ced993898f1c4cfc350a7d74828b8f9dea9906d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rb79p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-khtq8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:35Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:35 crc kubenswrapper[4870]: I1014 07:02:35.281487 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33a9493e-9660-48db-b8d1-310e29ebe313\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3570643ef1b7d5cc3a542f320413b819338e0c8b7ab7f0f09cc909997a39bab8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb398668197dedc1c2a14b4ae7d6a29714f3527309ccb9c1622af841100e1d60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1d2b6d068d63bd38cdfb83824525f6d12054f6cabab5d684e03ea5f2152076c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://783ff243e395333f3fc80cfae21358c532f0c9c5cb5515586be56b90ed10d912\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a8135a3d01076e4437ae8a784b459bf8d16de57fa16ec7ce4d40c5129c9d4a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48ca3a5017ba650df99b2fc7201c3fe5abfa89d81089c5647cde1b1549f534fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48ca3a5017ba650df99b2fc7201c3fe5abfa89d81089c5647cde1b1549f534fc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://19682eca67f76bf2e06e1b80f12e21141fcbce0acb8195f0defacd1b9a10b5e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://19682eca67f76bf2e06e1b80f12e21141fcbce0acb8195f0defacd1b9a10b5e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://042aad6d5f6bee8eef77f9e6a2c265de775a69f3da72aedc68bdf312198f00fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://042aad6d5f6bee8eef77f9e6a2c265de775a69f3da72aedc68bdf312198f00fa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:05Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:35Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:35 crc kubenswrapper[4870]: I1014 07:02:35.291223 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:35 crc kubenswrapper[4870]: I1014 07:02:35.291508 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:35 crc kubenswrapper[4870]: I1014 07:02:35.291584 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:35 crc kubenswrapper[4870]: I1014 07:02:35.291686 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:35 crc kubenswrapper[4870]: I1014 07:02:35.291792 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:35Z","lastTransitionTime":"2025-10-14T07:02:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:35 crc kubenswrapper[4870]: I1014 07:02:35.295226 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"221727c7-4b56-4fc4-b745-1b6e71666779\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8c8ee375ef1ec48a55be4a2d3e9e8860886b88ccb4080918d3e306ab34d2b95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07543849cd70415fe71b7dd0fd5f3806fdb422afa7a2dcce18a5c1b9fc974a76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://07543849cd70415fe71b7dd0fd5f3806fdb422afa7a2dcce18a5c1b9fc974a76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:35Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:35 crc kubenswrapper[4870]: I1014 07:02:35.308420 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a77729fb-a187-48c6-80db-bee90f7bdac4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:02:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:02:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c859ebc0d20b632d1d5a716caee343c7e6b9b752ed5f257422cd35ad12c170f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e8a7a5c1f1706ebbd2e8165d8961002f5a9340e2749e377b0058db107ad7dbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f75b3dc1557917e9c0c35c2810338a076ed92ba89963edafb9d537ad19c2bd8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b192bd3e0eb8540d7492abc207025471bc9e198e7e78b50215c8d2f5bc57cf4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8b192bd3e0eb8540d7492abc207025471bc9e198e7e78b50215c8d2f5bc57cf4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-14T07:01:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:05Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:35Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:35 crc kubenswrapper[4870]: I1014 07:02:35.318422 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-kz77f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d3864a91-9bf6-4540-a908-cc41aa39656c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d4ded3d86030f5ea548c8b1cc3caddbe14a45c9a3ae4af836e892880584f57d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v4h8v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:28Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-kz77f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:35Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:35 crc kubenswrapper[4870]: I1014 07:02:35.328861 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bcbc6dd6-aab0-4dbc-a62e-c58f06ec1337\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://25ec4808b09adc44488530b3c4e29fd7fec46b3fcfa31d4449ec4a6dbc880ec8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f3f5de86fb11231a19034a18e9d7f31476c162e6567d0c136ead8bbbd900d6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d57d938d9d9c92198ae5b3857536ece36012fbfbe8cc8cfd245d1386d39b0be5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77a3247a7f65709543d7fd374b2ab69d7d793537d4132d7e22261188d6e58544\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-14T07:01:05Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:35Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:35 crc kubenswrapper[4870]: I1014 07:02:35.341035 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e539fac3d0a2ade314a4b4c1050bb27c37bcff5ba5272fa2d71b97ab5c0d7191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-14T07:01:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:35Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:35 crc kubenswrapper[4870]: I1014 07:02:35.356940 4870 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-14T07:01:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-14T07:02:35Z is after 2025-08-24T17:21:41Z" Oct 14 07:02:35 crc kubenswrapper[4870]: I1014 07:02:35.394154 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:35 crc kubenswrapper[4870]: I1014 07:02:35.394641 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:35 crc kubenswrapper[4870]: I1014 07:02:35.394797 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:35 crc kubenswrapper[4870]: I1014 07:02:35.395170 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:35 crc kubenswrapper[4870]: I1014 07:02:35.395425 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:35Z","lastTransitionTime":"2025-10-14T07:02:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:35 crc kubenswrapper[4870]: I1014 07:02:35.498705 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:35 crc kubenswrapper[4870]: I1014 07:02:35.499056 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:35 crc kubenswrapper[4870]: I1014 07:02:35.499066 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:35 crc kubenswrapper[4870]: I1014 07:02:35.499082 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:35 crc kubenswrapper[4870]: I1014 07:02:35.499093 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:35Z","lastTransitionTime":"2025-10-14T07:02:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:35 crc kubenswrapper[4870]: I1014 07:02:35.602114 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:35 crc kubenswrapper[4870]: I1014 07:02:35.602195 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:35 crc kubenswrapper[4870]: I1014 07:02:35.602220 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:35 crc kubenswrapper[4870]: I1014 07:02:35.602249 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:35 crc kubenswrapper[4870]: I1014 07:02:35.602265 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:35Z","lastTransitionTime":"2025-10-14T07:02:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:35 crc kubenswrapper[4870]: I1014 07:02:35.705712 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:35 crc kubenswrapper[4870]: I1014 07:02:35.705780 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:35 crc kubenswrapper[4870]: I1014 07:02:35.705801 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:35 crc kubenswrapper[4870]: I1014 07:02:35.705832 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:35 crc kubenswrapper[4870]: I1014 07:02:35.705853 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:35Z","lastTransitionTime":"2025-10-14T07:02:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:35 crc kubenswrapper[4870]: I1014 07:02:35.809365 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:35 crc kubenswrapper[4870]: I1014 07:02:35.809424 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:35 crc kubenswrapper[4870]: I1014 07:02:35.809461 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:35 crc kubenswrapper[4870]: I1014 07:02:35.809486 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:35 crc kubenswrapper[4870]: I1014 07:02:35.809496 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:35Z","lastTransitionTime":"2025-10-14T07:02:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:35 crc kubenswrapper[4870]: I1014 07:02:35.913023 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:35 crc kubenswrapper[4870]: I1014 07:02:35.913101 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:35 crc kubenswrapper[4870]: I1014 07:02:35.913119 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:35 crc kubenswrapper[4870]: I1014 07:02:35.913148 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:35 crc kubenswrapper[4870]: I1014 07:02:35.913166 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:35Z","lastTransitionTime":"2025-10-14T07:02:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:36 crc kubenswrapper[4870]: I1014 07:02:36.016275 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:36 crc kubenswrapper[4870]: I1014 07:02:36.016750 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:36 crc kubenswrapper[4870]: I1014 07:02:36.016935 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:36 crc kubenswrapper[4870]: I1014 07:02:36.017067 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:36 crc kubenswrapper[4870]: I1014 07:02:36.017185 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:36Z","lastTransitionTime":"2025-10-14T07:02:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:36 crc kubenswrapper[4870]: I1014 07:02:36.033113 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 07:02:36 crc kubenswrapper[4870]: I1014 07:02:36.033323 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 07:02:36 crc kubenswrapper[4870]: I1014 07:02:36.033400 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 07:02:36 crc kubenswrapper[4870]: E1014 07:02:36.033512 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 07:02:36 crc kubenswrapper[4870]: E1014 07:02:36.033644 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 07:02:36 crc kubenswrapper[4870]: E1014 07:02:36.033849 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 07:02:36 crc kubenswrapper[4870]: I1014 07:02:36.120154 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:36 crc kubenswrapper[4870]: I1014 07:02:36.120192 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:36 crc kubenswrapper[4870]: I1014 07:02:36.120200 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:36 crc kubenswrapper[4870]: I1014 07:02:36.120218 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:36 crc kubenswrapper[4870]: I1014 07:02:36.120228 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:36Z","lastTransitionTime":"2025-10-14T07:02:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:36 crc kubenswrapper[4870]: I1014 07:02:36.223384 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:36 crc kubenswrapper[4870]: I1014 07:02:36.223434 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:36 crc kubenswrapper[4870]: I1014 07:02:36.223467 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:36 crc kubenswrapper[4870]: I1014 07:02:36.223489 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:36 crc kubenswrapper[4870]: I1014 07:02:36.223501 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:36Z","lastTransitionTime":"2025-10-14T07:02:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:36 crc kubenswrapper[4870]: I1014 07:02:36.327388 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:36 crc kubenswrapper[4870]: I1014 07:02:36.327503 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:36 crc kubenswrapper[4870]: I1014 07:02:36.327531 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:36 crc kubenswrapper[4870]: I1014 07:02:36.327567 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:36 crc kubenswrapper[4870]: I1014 07:02:36.327585 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:36Z","lastTransitionTime":"2025-10-14T07:02:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:36 crc kubenswrapper[4870]: I1014 07:02:36.430899 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:36 crc kubenswrapper[4870]: I1014 07:02:36.430959 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:36 crc kubenswrapper[4870]: I1014 07:02:36.430981 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:36 crc kubenswrapper[4870]: I1014 07:02:36.431010 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:36 crc kubenswrapper[4870]: I1014 07:02:36.431036 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:36Z","lastTransitionTime":"2025-10-14T07:02:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:36 crc kubenswrapper[4870]: I1014 07:02:36.534197 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:36 crc kubenswrapper[4870]: I1014 07:02:36.534237 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:36 crc kubenswrapper[4870]: I1014 07:02:36.534245 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:36 crc kubenswrapper[4870]: I1014 07:02:36.534260 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:36 crc kubenswrapper[4870]: I1014 07:02:36.534269 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:36Z","lastTransitionTime":"2025-10-14T07:02:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:36 crc kubenswrapper[4870]: I1014 07:02:36.637621 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:36 crc kubenswrapper[4870]: I1014 07:02:36.637661 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:36 crc kubenswrapper[4870]: I1014 07:02:36.637675 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:36 crc kubenswrapper[4870]: I1014 07:02:36.637693 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:36 crc kubenswrapper[4870]: I1014 07:02:36.637706 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:36Z","lastTransitionTime":"2025-10-14T07:02:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:36 crc kubenswrapper[4870]: I1014 07:02:36.743256 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:36 crc kubenswrapper[4870]: I1014 07:02:36.743335 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:36 crc kubenswrapper[4870]: I1014 07:02:36.743362 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:36 crc kubenswrapper[4870]: I1014 07:02:36.743398 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:36 crc kubenswrapper[4870]: I1014 07:02:36.743424 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:36Z","lastTransitionTime":"2025-10-14T07:02:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:36 crc kubenswrapper[4870]: I1014 07:02:36.846572 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:36 crc kubenswrapper[4870]: I1014 07:02:36.846650 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:36 crc kubenswrapper[4870]: I1014 07:02:36.846671 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:36 crc kubenswrapper[4870]: I1014 07:02:36.846694 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:36 crc kubenswrapper[4870]: I1014 07:02:36.846710 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:36Z","lastTransitionTime":"2025-10-14T07:02:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:36 crc kubenswrapper[4870]: I1014 07:02:36.949569 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:36 crc kubenswrapper[4870]: I1014 07:02:36.949630 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:36 crc kubenswrapper[4870]: I1014 07:02:36.949644 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:36 crc kubenswrapper[4870]: I1014 07:02:36.949667 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:36 crc kubenswrapper[4870]: I1014 07:02:36.950024 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:36Z","lastTransitionTime":"2025-10-14T07:02:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:37 crc kubenswrapper[4870]: I1014 07:02:37.032934 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zc5j6" Oct 14 07:02:37 crc kubenswrapper[4870]: E1014 07:02:37.033092 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zc5j6" podUID="7d7a4f41-388a-4320-b896-43c17ff13da6" Oct 14 07:02:37 crc kubenswrapper[4870]: I1014 07:02:37.053191 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:37 crc kubenswrapper[4870]: I1014 07:02:37.053273 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:37 crc kubenswrapper[4870]: I1014 07:02:37.053312 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:37 crc kubenswrapper[4870]: I1014 07:02:37.053350 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:37 crc kubenswrapper[4870]: I1014 07:02:37.053377 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:37Z","lastTransitionTime":"2025-10-14T07:02:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:37 crc kubenswrapper[4870]: I1014 07:02:37.158253 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:37 crc kubenswrapper[4870]: I1014 07:02:37.158319 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:37 crc kubenswrapper[4870]: I1014 07:02:37.158338 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:37 crc kubenswrapper[4870]: I1014 07:02:37.158367 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:37 crc kubenswrapper[4870]: I1014 07:02:37.158391 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:37Z","lastTransitionTime":"2025-10-14T07:02:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:37 crc kubenswrapper[4870]: I1014 07:02:37.262687 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:37 crc kubenswrapper[4870]: I1014 07:02:37.262787 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:37 crc kubenswrapper[4870]: I1014 07:02:37.262817 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:37 crc kubenswrapper[4870]: I1014 07:02:37.262856 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:37 crc kubenswrapper[4870]: I1014 07:02:37.262885 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:37Z","lastTransitionTime":"2025-10-14T07:02:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:37 crc kubenswrapper[4870]: I1014 07:02:37.365751 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:37 crc kubenswrapper[4870]: I1014 07:02:37.365837 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:37 crc kubenswrapper[4870]: I1014 07:02:37.365867 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:37 crc kubenswrapper[4870]: I1014 07:02:37.365902 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:37 crc kubenswrapper[4870]: I1014 07:02:37.365930 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:37Z","lastTransitionTime":"2025-10-14T07:02:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:37 crc kubenswrapper[4870]: I1014 07:02:37.469202 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:37 crc kubenswrapper[4870]: I1014 07:02:37.469260 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:37 crc kubenswrapper[4870]: I1014 07:02:37.469271 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:37 crc kubenswrapper[4870]: I1014 07:02:37.469288 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:37 crc kubenswrapper[4870]: I1014 07:02:37.469317 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:37Z","lastTransitionTime":"2025-10-14T07:02:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:37 crc kubenswrapper[4870]: I1014 07:02:37.573294 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:37 crc kubenswrapper[4870]: I1014 07:02:37.573377 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:37 crc kubenswrapper[4870]: I1014 07:02:37.573401 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:37 crc kubenswrapper[4870]: I1014 07:02:37.573434 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:37 crc kubenswrapper[4870]: I1014 07:02:37.573504 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:37Z","lastTransitionTime":"2025-10-14T07:02:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:37 crc kubenswrapper[4870]: I1014 07:02:37.677623 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:37 crc kubenswrapper[4870]: I1014 07:02:37.677729 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:37 crc kubenswrapper[4870]: I1014 07:02:37.677753 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:37 crc kubenswrapper[4870]: I1014 07:02:37.677788 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:37 crc kubenswrapper[4870]: I1014 07:02:37.677881 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:37Z","lastTransitionTime":"2025-10-14T07:02:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:37 crc kubenswrapper[4870]: I1014 07:02:37.781866 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:37 crc kubenswrapper[4870]: I1014 07:02:37.781944 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:37 crc kubenswrapper[4870]: I1014 07:02:37.781963 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:37 crc kubenswrapper[4870]: I1014 07:02:37.781993 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:37 crc kubenswrapper[4870]: I1014 07:02:37.782013 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:37Z","lastTransitionTime":"2025-10-14T07:02:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:37 crc kubenswrapper[4870]: I1014 07:02:37.886414 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:37 crc kubenswrapper[4870]: I1014 07:02:37.886474 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:37 crc kubenswrapper[4870]: I1014 07:02:37.886487 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:37 crc kubenswrapper[4870]: I1014 07:02:37.886504 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:37 crc kubenswrapper[4870]: I1014 07:02:37.886514 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:37Z","lastTransitionTime":"2025-10-14T07:02:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:37 crc kubenswrapper[4870]: I1014 07:02:37.989081 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:37 crc kubenswrapper[4870]: I1014 07:02:37.989122 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:37 crc kubenswrapper[4870]: I1014 07:02:37.989131 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:37 crc kubenswrapper[4870]: I1014 07:02:37.989146 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:37 crc kubenswrapper[4870]: I1014 07:02:37.989155 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:37Z","lastTransitionTime":"2025-10-14T07:02:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:38 crc kubenswrapper[4870]: I1014 07:02:38.033848 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 07:02:38 crc kubenswrapper[4870]: I1014 07:02:38.033946 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 07:02:38 crc kubenswrapper[4870]: E1014 07:02:38.034059 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 07:02:38 crc kubenswrapper[4870]: I1014 07:02:38.034074 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 07:02:38 crc kubenswrapper[4870]: E1014 07:02:38.034202 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 07:02:38 crc kubenswrapper[4870]: E1014 07:02:38.034310 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 07:02:38 crc kubenswrapper[4870]: I1014 07:02:38.092642 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:38 crc kubenswrapper[4870]: I1014 07:02:38.092717 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:38 crc kubenswrapper[4870]: I1014 07:02:38.092731 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:38 crc kubenswrapper[4870]: I1014 07:02:38.092758 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:38 crc kubenswrapper[4870]: I1014 07:02:38.092774 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:38Z","lastTransitionTime":"2025-10-14T07:02:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:38 crc kubenswrapper[4870]: I1014 07:02:38.195854 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:38 crc kubenswrapper[4870]: I1014 07:02:38.195891 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:38 crc kubenswrapper[4870]: I1014 07:02:38.195903 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:38 crc kubenswrapper[4870]: I1014 07:02:38.195918 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:38 crc kubenswrapper[4870]: I1014 07:02:38.195927 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:38Z","lastTransitionTime":"2025-10-14T07:02:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:38 crc kubenswrapper[4870]: I1014 07:02:38.299842 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:38 crc kubenswrapper[4870]: I1014 07:02:38.299927 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:38 crc kubenswrapper[4870]: I1014 07:02:38.299945 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:38 crc kubenswrapper[4870]: I1014 07:02:38.299971 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:38 crc kubenswrapper[4870]: I1014 07:02:38.299995 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:38Z","lastTransitionTime":"2025-10-14T07:02:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:38 crc kubenswrapper[4870]: I1014 07:02:38.404040 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:38 crc kubenswrapper[4870]: I1014 07:02:38.404096 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:38 crc kubenswrapper[4870]: I1014 07:02:38.404109 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:38 crc kubenswrapper[4870]: I1014 07:02:38.404135 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:38 crc kubenswrapper[4870]: I1014 07:02:38.404149 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:38Z","lastTransitionTime":"2025-10-14T07:02:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:38 crc kubenswrapper[4870]: I1014 07:02:38.507885 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:38 crc kubenswrapper[4870]: I1014 07:02:38.507964 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:38 crc kubenswrapper[4870]: I1014 07:02:38.507980 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:38 crc kubenswrapper[4870]: I1014 07:02:38.508050 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:38 crc kubenswrapper[4870]: I1014 07:02:38.508070 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:38Z","lastTransitionTime":"2025-10-14T07:02:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:38 crc kubenswrapper[4870]: I1014 07:02:38.612283 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:38 crc kubenswrapper[4870]: I1014 07:02:38.612352 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:38 crc kubenswrapper[4870]: I1014 07:02:38.612370 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:38 crc kubenswrapper[4870]: I1014 07:02:38.612403 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:38 crc kubenswrapper[4870]: I1014 07:02:38.612426 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:38Z","lastTransitionTime":"2025-10-14T07:02:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:38 crc kubenswrapper[4870]: I1014 07:02:38.715326 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:38 crc kubenswrapper[4870]: I1014 07:02:38.715384 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:38 crc kubenswrapper[4870]: I1014 07:02:38.715398 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:38 crc kubenswrapper[4870]: I1014 07:02:38.715421 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:38 crc kubenswrapper[4870]: I1014 07:02:38.715454 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:38Z","lastTransitionTime":"2025-10-14T07:02:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:38 crc kubenswrapper[4870]: I1014 07:02:38.819247 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:38 crc kubenswrapper[4870]: I1014 07:02:38.819314 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:38 crc kubenswrapper[4870]: I1014 07:02:38.819339 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:38 crc kubenswrapper[4870]: I1014 07:02:38.819365 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:38 crc kubenswrapper[4870]: I1014 07:02:38.819381 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:38Z","lastTransitionTime":"2025-10-14T07:02:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:38 crc kubenswrapper[4870]: I1014 07:02:38.921941 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:38 crc kubenswrapper[4870]: I1014 07:02:38.921975 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:38 crc kubenswrapper[4870]: I1014 07:02:38.921986 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:38 crc kubenswrapper[4870]: I1014 07:02:38.922001 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:38 crc kubenswrapper[4870]: I1014 07:02:38.922012 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:38Z","lastTransitionTime":"2025-10-14T07:02:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:39 crc kubenswrapper[4870]: I1014 07:02:39.025510 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:39 crc kubenswrapper[4870]: I1014 07:02:39.025603 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:39 crc kubenswrapper[4870]: I1014 07:02:39.025627 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:39 crc kubenswrapper[4870]: I1014 07:02:39.025659 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:39 crc kubenswrapper[4870]: I1014 07:02:39.025680 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:39Z","lastTransitionTime":"2025-10-14T07:02:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:39 crc kubenswrapper[4870]: I1014 07:02:39.033880 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zc5j6" Oct 14 07:02:39 crc kubenswrapper[4870]: E1014 07:02:39.034145 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zc5j6" podUID="7d7a4f41-388a-4320-b896-43c17ff13da6" Oct 14 07:02:39 crc kubenswrapper[4870]: I1014 07:02:39.129531 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:39 crc kubenswrapper[4870]: I1014 07:02:39.129612 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:39 crc kubenswrapper[4870]: I1014 07:02:39.129639 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:39 crc kubenswrapper[4870]: I1014 07:02:39.129680 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:39 crc kubenswrapper[4870]: I1014 07:02:39.129706 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:39Z","lastTransitionTime":"2025-10-14T07:02:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:39 crc kubenswrapper[4870]: I1014 07:02:39.233216 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:39 crc kubenswrapper[4870]: I1014 07:02:39.233296 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:39 crc kubenswrapper[4870]: I1014 07:02:39.233316 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:39 crc kubenswrapper[4870]: I1014 07:02:39.233346 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:39 crc kubenswrapper[4870]: I1014 07:02:39.233369 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:39Z","lastTransitionTime":"2025-10-14T07:02:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:39 crc kubenswrapper[4870]: I1014 07:02:39.337224 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:39 crc kubenswrapper[4870]: I1014 07:02:39.337324 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:39 crc kubenswrapper[4870]: I1014 07:02:39.337351 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:39 crc kubenswrapper[4870]: I1014 07:02:39.337386 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:39 crc kubenswrapper[4870]: I1014 07:02:39.337409 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:39Z","lastTransitionTime":"2025-10-14T07:02:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:39 crc kubenswrapper[4870]: I1014 07:02:39.439810 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:39 crc kubenswrapper[4870]: I1014 07:02:39.439857 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:39 crc kubenswrapper[4870]: I1014 07:02:39.439868 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:39 crc kubenswrapper[4870]: I1014 07:02:39.439885 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:39 crc kubenswrapper[4870]: I1014 07:02:39.439896 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:39Z","lastTransitionTime":"2025-10-14T07:02:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:39 crc kubenswrapper[4870]: I1014 07:02:39.543824 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:39 crc kubenswrapper[4870]: I1014 07:02:39.543885 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:39 crc kubenswrapper[4870]: I1014 07:02:39.543903 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:39 crc kubenswrapper[4870]: I1014 07:02:39.543928 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:39 crc kubenswrapper[4870]: I1014 07:02:39.543948 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:39Z","lastTransitionTime":"2025-10-14T07:02:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:39 crc kubenswrapper[4870]: I1014 07:02:39.646859 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:39 crc kubenswrapper[4870]: I1014 07:02:39.646932 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:39 crc kubenswrapper[4870]: I1014 07:02:39.646948 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:39 crc kubenswrapper[4870]: I1014 07:02:39.646976 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:39 crc kubenswrapper[4870]: I1014 07:02:39.646994 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:39Z","lastTransitionTime":"2025-10-14T07:02:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:39 crc kubenswrapper[4870]: I1014 07:02:39.750033 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:39 crc kubenswrapper[4870]: I1014 07:02:39.750109 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:39 crc kubenswrapper[4870]: I1014 07:02:39.750127 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:39 crc kubenswrapper[4870]: I1014 07:02:39.750165 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:39 crc kubenswrapper[4870]: I1014 07:02:39.750190 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:39Z","lastTransitionTime":"2025-10-14T07:02:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:39 crc kubenswrapper[4870]: I1014 07:02:39.854396 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:39 crc kubenswrapper[4870]: I1014 07:02:39.854516 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:39 crc kubenswrapper[4870]: I1014 07:02:39.854544 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:39 crc kubenswrapper[4870]: I1014 07:02:39.854576 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:39 crc kubenswrapper[4870]: I1014 07:02:39.854598 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:39Z","lastTransitionTime":"2025-10-14T07:02:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:39 crc kubenswrapper[4870]: I1014 07:02:39.957561 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:39 crc kubenswrapper[4870]: I1014 07:02:39.957605 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:39 crc kubenswrapper[4870]: I1014 07:02:39.957621 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:39 crc kubenswrapper[4870]: I1014 07:02:39.957646 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:39 crc kubenswrapper[4870]: I1014 07:02:39.957664 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:39Z","lastTransitionTime":"2025-10-14T07:02:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:40 crc kubenswrapper[4870]: I1014 07:02:40.033219 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 07:02:40 crc kubenswrapper[4870]: I1014 07:02:40.033289 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 07:02:40 crc kubenswrapper[4870]: I1014 07:02:40.033219 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 07:02:40 crc kubenswrapper[4870]: E1014 07:02:40.033406 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 07:02:40 crc kubenswrapper[4870]: E1014 07:02:40.033582 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 07:02:40 crc kubenswrapper[4870]: E1014 07:02:40.033739 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 07:02:40 crc kubenswrapper[4870]: I1014 07:02:40.061503 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:40 crc kubenswrapper[4870]: I1014 07:02:40.061566 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:40 crc kubenswrapper[4870]: I1014 07:02:40.061583 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:40 crc kubenswrapper[4870]: I1014 07:02:40.061604 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:40 crc kubenswrapper[4870]: I1014 07:02:40.061625 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:40Z","lastTransitionTime":"2025-10-14T07:02:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:40 crc kubenswrapper[4870]: I1014 07:02:40.164090 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:40 crc kubenswrapper[4870]: I1014 07:02:40.164161 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:40 crc kubenswrapper[4870]: I1014 07:02:40.164181 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:40 crc kubenswrapper[4870]: I1014 07:02:40.164205 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:40 crc kubenswrapper[4870]: I1014 07:02:40.164223 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:40Z","lastTransitionTime":"2025-10-14T07:02:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:40 crc kubenswrapper[4870]: I1014 07:02:40.266873 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:40 crc kubenswrapper[4870]: I1014 07:02:40.266935 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:40 crc kubenswrapper[4870]: I1014 07:02:40.266955 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:40 crc kubenswrapper[4870]: I1014 07:02:40.266980 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:40 crc kubenswrapper[4870]: I1014 07:02:40.266999 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:40Z","lastTransitionTime":"2025-10-14T07:02:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:40 crc kubenswrapper[4870]: I1014 07:02:40.370733 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:40 crc kubenswrapper[4870]: I1014 07:02:40.370826 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:40 crc kubenswrapper[4870]: I1014 07:02:40.370857 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:40 crc kubenswrapper[4870]: I1014 07:02:40.370898 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:40 crc kubenswrapper[4870]: I1014 07:02:40.370924 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:40Z","lastTransitionTime":"2025-10-14T07:02:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:40 crc kubenswrapper[4870]: I1014 07:02:40.473969 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:40 crc kubenswrapper[4870]: I1014 07:02:40.474023 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:40 crc kubenswrapper[4870]: I1014 07:02:40.474035 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:40 crc kubenswrapper[4870]: I1014 07:02:40.474059 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:40 crc kubenswrapper[4870]: I1014 07:02:40.474076 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:40Z","lastTransitionTime":"2025-10-14T07:02:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:40 crc kubenswrapper[4870]: I1014 07:02:40.579425 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:40 crc kubenswrapper[4870]: I1014 07:02:40.579568 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:40 crc kubenswrapper[4870]: I1014 07:02:40.579594 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:40 crc kubenswrapper[4870]: I1014 07:02:40.579632 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:40 crc kubenswrapper[4870]: I1014 07:02:40.579665 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:40Z","lastTransitionTime":"2025-10-14T07:02:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:40 crc kubenswrapper[4870]: I1014 07:02:40.683291 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:40 crc kubenswrapper[4870]: I1014 07:02:40.683374 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:40 crc kubenswrapper[4870]: I1014 07:02:40.683400 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:40 crc kubenswrapper[4870]: I1014 07:02:40.683435 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:40 crc kubenswrapper[4870]: I1014 07:02:40.683517 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:40Z","lastTransitionTime":"2025-10-14T07:02:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:40 crc kubenswrapper[4870]: I1014 07:02:40.786902 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:40 crc kubenswrapper[4870]: I1014 07:02:40.786960 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:40 crc kubenswrapper[4870]: I1014 07:02:40.786973 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:40 crc kubenswrapper[4870]: I1014 07:02:40.786990 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:40 crc kubenswrapper[4870]: I1014 07:02:40.787006 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:40Z","lastTransitionTime":"2025-10-14T07:02:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:40 crc kubenswrapper[4870]: I1014 07:02:40.890255 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:40 crc kubenswrapper[4870]: I1014 07:02:40.890315 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:40 crc kubenswrapper[4870]: I1014 07:02:40.890326 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:40 crc kubenswrapper[4870]: I1014 07:02:40.890348 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:40 crc kubenswrapper[4870]: I1014 07:02:40.890362 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:40Z","lastTransitionTime":"2025-10-14T07:02:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:40 crc kubenswrapper[4870]: I1014 07:02:40.993695 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:40 crc kubenswrapper[4870]: I1014 07:02:40.993756 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:40 crc kubenswrapper[4870]: I1014 07:02:40.993766 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:40 crc kubenswrapper[4870]: I1014 07:02:40.993790 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:40 crc kubenswrapper[4870]: I1014 07:02:40.993802 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:40Z","lastTransitionTime":"2025-10-14T07:02:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:41 crc kubenswrapper[4870]: I1014 07:02:41.033166 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zc5j6" Oct 14 07:02:41 crc kubenswrapper[4870]: E1014 07:02:41.033410 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zc5j6" podUID="7d7a4f41-388a-4320-b896-43c17ff13da6" Oct 14 07:02:41 crc kubenswrapper[4870]: I1014 07:02:41.096940 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:41 crc kubenswrapper[4870]: I1014 07:02:41.096984 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:41 crc kubenswrapper[4870]: I1014 07:02:41.096994 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:41 crc kubenswrapper[4870]: I1014 07:02:41.097012 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:41 crc kubenswrapper[4870]: I1014 07:02:41.097024 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:41Z","lastTransitionTime":"2025-10-14T07:02:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:41 crc kubenswrapper[4870]: I1014 07:02:41.201971 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:41 crc kubenswrapper[4870]: I1014 07:02:41.202054 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:41 crc kubenswrapper[4870]: I1014 07:02:41.202075 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:41 crc kubenswrapper[4870]: I1014 07:02:41.202107 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:41 crc kubenswrapper[4870]: I1014 07:02:41.202127 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:41Z","lastTransitionTime":"2025-10-14T07:02:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:41 crc kubenswrapper[4870]: I1014 07:02:41.305404 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:41 crc kubenswrapper[4870]: I1014 07:02:41.305505 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:41 crc kubenswrapper[4870]: I1014 07:02:41.305526 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:41 crc kubenswrapper[4870]: I1014 07:02:41.305550 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:41 crc kubenswrapper[4870]: I1014 07:02:41.305567 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:41Z","lastTransitionTime":"2025-10-14T07:02:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:41 crc kubenswrapper[4870]: I1014 07:02:41.409762 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:41 crc kubenswrapper[4870]: I1014 07:02:41.409841 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:41 crc kubenswrapper[4870]: I1014 07:02:41.409859 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:41 crc kubenswrapper[4870]: I1014 07:02:41.409925 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:41 crc kubenswrapper[4870]: I1014 07:02:41.409945 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:41Z","lastTransitionTime":"2025-10-14T07:02:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:41 crc kubenswrapper[4870]: I1014 07:02:41.512603 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:41 crc kubenswrapper[4870]: I1014 07:02:41.512674 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:41 crc kubenswrapper[4870]: I1014 07:02:41.512692 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:41 crc kubenswrapper[4870]: I1014 07:02:41.512722 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:41 crc kubenswrapper[4870]: I1014 07:02:41.512747 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:41Z","lastTransitionTime":"2025-10-14T07:02:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:41 crc kubenswrapper[4870]: I1014 07:02:41.616323 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:41 crc kubenswrapper[4870]: I1014 07:02:41.616372 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:41 crc kubenswrapper[4870]: I1014 07:02:41.616383 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:41 crc kubenswrapper[4870]: I1014 07:02:41.616401 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:41 crc kubenswrapper[4870]: I1014 07:02:41.616412 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:41Z","lastTransitionTime":"2025-10-14T07:02:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:41 crc kubenswrapper[4870]: I1014 07:02:41.720474 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:41 crc kubenswrapper[4870]: I1014 07:02:41.720575 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:41 crc kubenswrapper[4870]: I1014 07:02:41.720595 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:41 crc kubenswrapper[4870]: I1014 07:02:41.720629 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:41 crc kubenswrapper[4870]: I1014 07:02:41.720656 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:41Z","lastTransitionTime":"2025-10-14T07:02:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:41 crc kubenswrapper[4870]: I1014 07:02:41.824686 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:41 crc kubenswrapper[4870]: I1014 07:02:41.824751 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:41 crc kubenswrapper[4870]: I1014 07:02:41.824776 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:41 crc kubenswrapper[4870]: I1014 07:02:41.824807 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:41 crc kubenswrapper[4870]: I1014 07:02:41.824829 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:41Z","lastTransitionTime":"2025-10-14T07:02:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:41 crc kubenswrapper[4870]: I1014 07:02:41.928016 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:41 crc kubenswrapper[4870]: I1014 07:02:41.928112 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:41 crc kubenswrapper[4870]: I1014 07:02:41.928884 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:41 crc kubenswrapper[4870]: I1014 07:02:41.929157 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:41 crc kubenswrapper[4870]: I1014 07:02:41.929177 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:41Z","lastTransitionTime":"2025-10-14T07:02:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:42 crc kubenswrapper[4870]: I1014 07:02:42.032933 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 07:02:42 crc kubenswrapper[4870]: I1014 07:02:42.032933 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 07:02:42 crc kubenswrapper[4870]: I1014 07:02:42.032953 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 07:02:42 crc kubenswrapper[4870]: I1014 07:02:42.033255 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:42 crc kubenswrapper[4870]: I1014 07:02:42.033306 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:42 crc kubenswrapper[4870]: I1014 07:02:42.033324 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:42 crc kubenswrapper[4870]: I1014 07:02:42.033355 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:42 crc kubenswrapper[4870]: I1014 07:02:42.033374 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:42Z","lastTransitionTime":"2025-10-14T07:02:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:42 crc kubenswrapper[4870]: E1014 07:02:42.033463 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 07:02:42 crc kubenswrapper[4870]: E1014 07:02:42.033644 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 07:02:42 crc kubenswrapper[4870]: E1014 07:02:42.033768 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 07:02:42 crc kubenswrapper[4870]: I1014 07:02:42.033802 4870 scope.go:117] "RemoveContainer" containerID="ab4295721b6ec6d64040d321899ab478c612e91cbc3b99b33753869578a2d1e8" Oct 14 07:02:42 crc kubenswrapper[4870]: E1014 07:02:42.034034 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-v7br2_openshift-ovn-kubernetes(8dcaef0e-42e3-424b-bbe5-4a023efe912d)\"" pod="openshift-ovn-kubernetes/ovnkube-node-v7br2" podUID="8dcaef0e-42e3-424b-bbe5-4a023efe912d" Oct 14 07:02:42 crc kubenswrapper[4870]: I1014 07:02:42.136780 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:42 crc kubenswrapper[4870]: I1014 07:02:42.136847 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:42 crc kubenswrapper[4870]: I1014 07:02:42.136867 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:42 crc kubenswrapper[4870]: I1014 07:02:42.136897 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:42 crc kubenswrapper[4870]: I1014 07:02:42.136917 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:42Z","lastTransitionTime":"2025-10-14T07:02:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:42 crc kubenswrapper[4870]: I1014 07:02:42.241505 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:42 crc kubenswrapper[4870]: I1014 07:02:42.241560 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:42 crc kubenswrapper[4870]: I1014 07:02:42.241580 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:42 crc kubenswrapper[4870]: I1014 07:02:42.241612 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:42 crc kubenswrapper[4870]: I1014 07:02:42.241634 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:42Z","lastTransitionTime":"2025-10-14T07:02:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:42 crc kubenswrapper[4870]: I1014 07:02:42.345165 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:42 crc kubenswrapper[4870]: I1014 07:02:42.345232 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:42 crc kubenswrapper[4870]: I1014 07:02:42.345249 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:42 crc kubenswrapper[4870]: I1014 07:02:42.345273 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:42 crc kubenswrapper[4870]: I1014 07:02:42.345290 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:42Z","lastTransitionTime":"2025-10-14T07:02:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:42 crc kubenswrapper[4870]: I1014 07:02:42.448566 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:42 crc kubenswrapper[4870]: I1014 07:02:42.448658 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:42 crc kubenswrapper[4870]: I1014 07:02:42.448677 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:42 crc kubenswrapper[4870]: I1014 07:02:42.448714 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:42 crc kubenswrapper[4870]: I1014 07:02:42.448734 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:42Z","lastTransitionTime":"2025-10-14T07:02:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:42 crc kubenswrapper[4870]: I1014 07:02:42.551993 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:42 crc kubenswrapper[4870]: I1014 07:02:42.552099 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:42 crc kubenswrapper[4870]: I1014 07:02:42.552119 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:42 crc kubenswrapper[4870]: I1014 07:02:42.552151 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:42 crc kubenswrapper[4870]: I1014 07:02:42.552170 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:42Z","lastTransitionTime":"2025-10-14T07:02:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:42 crc kubenswrapper[4870]: I1014 07:02:42.656129 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:42 crc kubenswrapper[4870]: I1014 07:02:42.656211 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:42 crc kubenswrapper[4870]: I1014 07:02:42.656238 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:42 crc kubenswrapper[4870]: I1014 07:02:42.656274 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:42 crc kubenswrapper[4870]: I1014 07:02:42.656295 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:42Z","lastTransitionTime":"2025-10-14T07:02:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:42 crc kubenswrapper[4870]: I1014 07:02:42.759386 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:42 crc kubenswrapper[4870]: I1014 07:02:42.759510 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:42 crc kubenswrapper[4870]: I1014 07:02:42.759530 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:42 crc kubenswrapper[4870]: I1014 07:02:42.759559 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:42 crc kubenswrapper[4870]: I1014 07:02:42.759580 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:42Z","lastTransitionTime":"2025-10-14T07:02:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:42 crc kubenswrapper[4870]: I1014 07:02:42.863117 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:42 crc kubenswrapper[4870]: I1014 07:02:42.863237 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:42 crc kubenswrapper[4870]: I1014 07:02:42.863261 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:42 crc kubenswrapper[4870]: I1014 07:02:42.863293 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:42 crc kubenswrapper[4870]: I1014 07:02:42.863314 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:42Z","lastTransitionTime":"2025-10-14T07:02:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:42 crc kubenswrapper[4870]: I1014 07:02:42.980821 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:42 crc kubenswrapper[4870]: I1014 07:02:42.980894 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:42 crc kubenswrapper[4870]: I1014 07:02:42.980914 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:42 crc kubenswrapper[4870]: I1014 07:02:42.980947 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:42 crc kubenswrapper[4870]: I1014 07:02:42.980972 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:42Z","lastTransitionTime":"2025-10-14T07:02:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:43 crc kubenswrapper[4870]: I1014 07:02:43.034079 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zc5j6" Oct 14 07:02:43 crc kubenswrapper[4870]: E1014 07:02:43.034308 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zc5j6" podUID="7d7a4f41-388a-4320-b896-43c17ff13da6" Oct 14 07:02:43 crc kubenswrapper[4870]: I1014 07:02:43.083916 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:43 crc kubenswrapper[4870]: I1014 07:02:43.083985 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:43 crc kubenswrapper[4870]: I1014 07:02:43.084005 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:43 crc kubenswrapper[4870]: I1014 07:02:43.084037 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:43 crc kubenswrapper[4870]: I1014 07:02:43.084058 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:43Z","lastTransitionTime":"2025-10-14T07:02:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:43 crc kubenswrapper[4870]: I1014 07:02:43.189868 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:43 crc kubenswrapper[4870]: I1014 07:02:43.189927 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:43 crc kubenswrapper[4870]: I1014 07:02:43.189941 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:43 crc kubenswrapper[4870]: I1014 07:02:43.189962 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:43 crc kubenswrapper[4870]: I1014 07:02:43.189973 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:43Z","lastTransitionTime":"2025-10-14T07:02:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:43 crc kubenswrapper[4870]: I1014 07:02:43.294933 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:43 crc kubenswrapper[4870]: I1014 07:02:43.295358 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:43 crc kubenswrapper[4870]: I1014 07:02:43.295594 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:43 crc kubenswrapper[4870]: I1014 07:02:43.295741 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:43 crc kubenswrapper[4870]: I1014 07:02:43.295879 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:43Z","lastTransitionTime":"2025-10-14T07:02:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:43 crc kubenswrapper[4870]: I1014 07:02:43.400245 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:43 crc kubenswrapper[4870]: I1014 07:02:43.400340 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:43 crc kubenswrapper[4870]: I1014 07:02:43.400362 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:43 crc kubenswrapper[4870]: I1014 07:02:43.400414 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:43 crc kubenswrapper[4870]: I1014 07:02:43.400430 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:43Z","lastTransitionTime":"2025-10-14T07:02:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:43 crc kubenswrapper[4870]: I1014 07:02:43.503723 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:43 crc kubenswrapper[4870]: I1014 07:02:43.503815 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:43 crc kubenswrapper[4870]: I1014 07:02:43.503836 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:43 crc kubenswrapper[4870]: I1014 07:02:43.503871 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:43 crc kubenswrapper[4870]: I1014 07:02:43.503893 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:43Z","lastTransitionTime":"2025-10-14T07:02:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:43 crc kubenswrapper[4870]: I1014 07:02:43.607074 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:43 crc kubenswrapper[4870]: I1014 07:02:43.607137 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:43 crc kubenswrapper[4870]: I1014 07:02:43.607156 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:43 crc kubenswrapper[4870]: I1014 07:02:43.607188 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:43 crc kubenswrapper[4870]: I1014 07:02:43.607208 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:43Z","lastTransitionTime":"2025-10-14T07:02:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:43 crc kubenswrapper[4870]: I1014 07:02:43.710873 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:43 crc kubenswrapper[4870]: I1014 07:02:43.710942 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:43 crc kubenswrapper[4870]: I1014 07:02:43.710962 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:43 crc kubenswrapper[4870]: I1014 07:02:43.710992 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:43 crc kubenswrapper[4870]: I1014 07:02:43.711018 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:43Z","lastTransitionTime":"2025-10-14T07:02:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:43 crc kubenswrapper[4870]: I1014 07:02:43.815554 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:43 crc kubenswrapper[4870]: I1014 07:02:43.815669 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:43 crc kubenswrapper[4870]: I1014 07:02:43.815724 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:43 crc kubenswrapper[4870]: I1014 07:02:43.815759 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:43 crc kubenswrapper[4870]: I1014 07:02:43.815777 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:43Z","lastTransitionTime":"2025-10-14T07:02:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:43 crc kubenswrapper[4870]: I1014 07:02:43.919815 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:43 crc kubenswrapper[4870]: I1014 07:02:43.919889 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:43 crc kubenswrapper[4870]: I1014 07:02:43.919914 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:43 crc kubenswrapper[4870]: I1014 07:02:43.919946 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:43 crc kubenswrapper[4870]: I1014 07:02:43.919968 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:43Z","lastTransitionTime":"2025-10-14T07:02:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:44 crc kubenswrapper[4870]: I1014 07:02:44.025051 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:44 crc kubenswrapper[4870]: I1014 07:02:44.025120 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:44 crc kubenswrapper[4870]: I1014 07:02:44.025138 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:44 crc kubenswrapper[4870]: I1014 07:02:44.025169 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:44 crc kubenswrapper[4870]: I1014 07:02:44.025189 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:44Z","lastTransitionTime":"2025-10-14T07:02:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:44 crc kubenswrapper[4870]: I1014 07:02:44.033682 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 07:02:44 crc kubenswrapper[4870]: I1014 07:02:44.033765 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 07:02:44 crc kubenswrapper[4870]: I1014 07:02:44.034087 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 07:02:44 crc kubenswrapper[4870]: E1014 07:02:44.034257 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 07:02:44 crc kubenswrapper[4870]: E1014 07:02:44.034470 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 07:02:44 crc kubenswrapper[4870]: E1014 07:02:44.034841 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 07:02:44 crc kubenswrapper[4870]: I1014 07:02:44.128944 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:44 crc kubenswrapper[4870]: I1014 07:02:44.129020 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:44 crc kubenswrapper[4870]: I1014 07:02:44.129041 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:44 crc kubenswrapper[4870]: I1014 07:02:44.129075 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:44 crc kubenswrapper[4870]: I1014 07:02:44.129096 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:44Z","lastTransitionTime":"2025-10-14T07:02:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:44 crc kubenswrapper[4870]: I1014 07:02:44.232692 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:44 crc kubenswrapper[4870]: I1014 07:02:44.232781 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:44 crc kubenswrapper[4870]: I1014 07:02:44.232801 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:44 crc kubenswrapper[4870]: I1014 07:02:44.232834 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:44 crc kubenswrapper[4870]: I1014 07:02:44.232855 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:44Z","lastTransitionTime":"2025-10-14T07:02:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:44 crc kubenswrapper[4870]: I1014 07:02:44.335950 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:44 crc kubenswrapper[4870]: I1014 07:02:44.336018 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:44 crc kubenswrapper[4870]: I1014 07:02:44.336038 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:44 crc kubenswrapper[4870]: I1014 07:02:44.336067 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:44 crc kubenswrapper[4870]: I1014 07:02:44.336087 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:44Z","lastTransitionTime":"2025-10-14T07:02:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:44 crc kubenswrapper[4870]: I1014 07:02:44.439660 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:44 crc kubenswrapper[4870]: I1014 07:02:44.439739 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:44 crc kubenswrapper[4870]: I1014 07:02:44.439758 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:44 crc kubenswrapper[4870]: I1014 07:02:44.439791 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:44 crc kubenswrapper[4870]: I1014 07:02:44.439818 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:44Z","lastTransitionTime":"2025-10-14T07:02:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:44 crc kubenswrapper[4870]: I1014 07:02:44.545476 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:44 crc kubenswrapper[4870]: I1014 07:02:44.545542 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:44 crc kubenswrapper[4870]: I1014 07:02:44.545557 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:44 crc kubenswrapper[4870]: I1014 07:02:44.545584 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:44 crc kubenswrapper[4870]: I1014 07:02:44.545598 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:44Z","lastTransitionTime":"2025-10-14T07:02:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:44 crc kubenswrapper[4870]: I1014 07:02:44.648277 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:44 crc kubenswrapper[4870]: I1014 07:02:44.648370 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:44 crc kubenswrapper[4870]: I1014 07:02:44.648392 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:44 crc kubenswrapper[4870]: I1014 07:02:44.648429 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:44 crc kubenswrapper[4870]: I1014 07:02:44.648496 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:44Z","lastTransitionTime":"2025-10-14T07:02:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:44 crc kubenswrapper[4870]: I1014 07:02:44.651356 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 14 07:02:44 crc kubenswrapper[4870]: I1014 07:02:44.651410 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 14 07:02:44 crc kubenswrapper[4870]: I1014 07:02:44.651421 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 14 07:02:44 crc kubenswrapper[4870]: I1014 07:02:44.651456 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 14 07:02:44 crc kubenswrapper[4870]: I1014 07:02:44.651469 4870 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-14T07:02:44Z","lastTransitionTime":"2025-10-14T07:02:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 14 07:02:44 crc kubenswrapper[4870]: I1014 07:02:44.719227 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-4r29b"] Oct 14 07:02:44 crc kubenswrapper[4870]: I1014 07:02:44.720314 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-4r29b" Oct 14 07:02:44 crc kubenswrapper[4870]: I1014 07:02:44.726686 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Oct 14 07:02:44 crc kubenswrapper[4870]: I1014 07:02:44.726808 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Oct 14 07:02:44 crc kubenswrapper[4870]: I1014 07:02:44.726910 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Oct 14 07:02:44 crc kubenswrapper[4870]: I1014 07:02:44.728690 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Oct 14 07:02:44 crc kubenswrapper[4870]: I1014 07:02:44.729297 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7d7a4f41-388a-4320-b896-43c17ff13da6-metrics-certs\") pod \"network-metrics-daemon-zc5j6\" (UID: \"7d7a4f41-388a-4320-b896-43c17ff13da6\") " pod="openshift-multus/network-metrics-daemon-zc5j6" Oct 14 07:02:44 crc kubenswrapper[4870]: I1014 07:02:44.729381 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/da3eb795-7ffa-47d8-8fe7-ae4e1c84e0dd-service-ca\") pod \"cluster-version-operator-5c965bbfc6-4r29b\" (UID: \"da3eb795-7ffa-47d8-8fe7-ae4e1c84e0dd\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-4r29b" Oct 14 07:02:44 crc kubenswrapper[4870]: I1014 07:02:44.729422 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/da3eb795-7ffa-47d8-8fe7-ae4e1c84e0dd-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-4r29b\" (UID: \"da3eb795-7ffa-47d8-8fe7-ae4e1c84e0dd\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-4r29b" Oct 14 07:02:44 crc kubenswrapper[4870]: I1014 07:02:44.729475 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/da3eb795-7ffa-47d8-8fe7-ae4e1c84e0dd-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-4r29b\" (UID: \"da3eb795-7ffa-47d8-8fe7-ae4e1c84e0dd\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-4r29b" Oct 14 07:02:44 crc kubenswrapper[4870]: I1014 07:02:44.729502 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/da3eb795-7ffa-47d8-8fe7-ae4e1c84e0dd-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-4r29b\" (UID: \"da3eb795-7ffa-47d8-8fe7-ae4e1c84e0dd\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-4r29b" Oct 14 07:02:44 crc kubenswrapper[4870]: E1014 07:02:44.729515 4870 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 14 07:02:44 crc kubenswrapper[4870]: I1014 07:02:44.729528 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/da3eb795-7ffa-47d8-8fe7-ae4e1c84e0dd-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-4r29b\" (UID: \"da3eb795-7ffa-47d8-8fe7-ae4e1c84e0dd\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-4r29b" Oct 14 07:02:44 crc kubenswrapper[4870]: E1014 07:02:44.729566 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7d7a4f41-388a-4320-b896-43c17ff13da6-metrics-certs podName:7d7a4f41-388a-4320-b896-43c17ff13da6 nodeName:}" failed. No retries permitted until 2025-10-14 07:03:48.729547429 +0000 UTC m=+164.426907800 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/7d7a4f41-388a-4320-b896-43c17ff13da6-metrics-certs") pod "network-metrics-daemon-zc5j6" (UID: "7d7a4f41-388a-4320-b896-43c17ff13da6") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 14 07:02:44 crc kubenswrapper[4870]: I1014 07:02:44.774774 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=77.774742196 podStartE2EDuration="1m17.774742196s" podCreationTimestamp="2025-10-14 07:01:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:02:44.75595875 +0000 UTC m=+100.453319121" watchObservedRunningTime="2025-10-14 07:02:44.774742196 +0000 UTC m=+100.472102567" Oct 14 07:02:44 crc kubenswrapper[4870]: I1014 07:02:44.830725 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/da3eb795-7ffa-47d8-8fe7-ae4e1c84e0dd-service-ca\") pod \"cluster-version-operator-5c965bbfc6-4r29b\" (UID: \"da3eb795-7ffa-47d8-8fe7-ae4e1c84e0dd\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-4r29b" Oct 14 07:02:44 crc kubenswrapper[4870]: I1014 07:02:44.830819 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/da3eb795-7ffa-47d8-8fe7-ae4e1c84e0dd-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-4r29b\" (UID: \"da3eb795-7ffa-47d8-8fe7-ae4e1c84e0dd\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-4r29b" Oct 14 07:02:44 crc kubenswrapper[4870]: I1014 07:02:44.830877 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/da3eb795-7ffa-47d8-8fe7-ae4e1c84e0dd-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-4r29b\" (UID: \"da3eb795-7ffa-47d8-8fe7-ae4e1c84e0dd\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-4r29b" Oct 14 07:02:44 crc kubenswrapper[4870]: I1014 07:02:44.830902 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/da3eb795-7ffa-47d8-8fe7-ae4e1c84e0dd-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-4r29b\" (UID: \"da3eb795-7ffa-47d8-8fe7-ae4e1c84e0dd\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-4r29b" Oct 14 07:02:44 crc kubenswrapper[4870]: I1014 07:02:44.830933 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/da3eb795-7ffa-47d8-8fe7-ae4e1c84e0dd-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-4r29b\" (UID: \"da3eb795-7ffa-47d8-8fe7-ae4e1c84e0dd\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-4r29b" Oct 14 07:02:44 crc kubenswrapper[4870]: I1014 07:02:44.831017 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/da3eb795-7ffa-47d8-8fe7-ae4e1c84e0dd-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-4r29b\" (UID: \"da3eb795-7ffa-47d8-8fe7-ae4e1c84e0dd\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-4r29b" Oct 14 07:02:44 crc kubenswrapper[4870]: I1014 07:02:44.831127 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/da3eb795-7ffa-47d8-8fe7-ae4e1c84e0dd-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-4r29b\" (UID: \"da3eb795-7ffa-47d8-8fe7-ae4e1c84e0dd\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-4r29b" Oct 14 07:02:44 crc kubenswrapper[4870]: I1014 07:02:44.832697 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/da3eb795-7ffa-47d8-8fe7-ae4e1c84e0dd-service-ca\") pod \"cluster-version-operator-5c965bbfc6-4r29b\" (UID: \"da3eb795-7ffa-47d8-8fe7-ae4e1c84e0dd\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-4r29b" Oct 14 07:02:44 crc kubenswrapper[4870]: I1014 07:02:44.839045 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/da3eb795-7ffa-47d8-8fe7-ae4e1c84e0dd-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-4r29b\" (UID: \"da3eb795-7ffa-47d8-8fe7-ae4e1c84e0dd\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-4r29b" Oct 14 07:02:44 crc kubenswrapper[4870]: I1014 07:02:44.849608 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/da3eb795-7ffa-47d8-8fe7-ae4e1c84e0dd-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-4r29b\" (UID: \"da3eb795-7ffa-47d8-8fe7-ae4e1c84e0dd\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-4r29b" Oct 14 07:02:44 crc kubenswrapper[4870]: I1014 07:02:44.854183 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-cdjmj" podStartSLOduration=78.854153822 podStartE2EDuration="1m18.854153822s" podCreationTimestamp="2025-10-14 07:01:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:02:44.854120531 +0000 UTC m=+100.551480902" watchObservedRunningTime="2025-10-14 07:02:44.854153822 +0000 UTC m=+100.551514193" Oct 14 07:02:44 crc kubenswrapper[4870]: I1014 07:02:44.890036 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-khtq8" podStartSLOduration=78.890007781 podStartE2EDuration="1m18.890007781s" podCreationTimestamp="2025-10-14 07:01:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:02:44.88953378 +0000 UTC m=+100.586894181" watchObservedRunningTime="2025-10-14 07:02:44.890007781 +0000 UTC m=+100.587368182" Oct 14 07:02:44 crc kubenswrapper[4870]: I1014 07:02:44.897203 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-fdxjh" podStartSLOduration=78.897187915 podStartE2EDuration="1m18.897187915s" podCreationTimestamp="2025-10-14 07:01:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:02:44.87470927 +0000 UTC m=+100.572069661" watchObservedRunningTime="2025-10-14 07:02:44.897187915 +0000 UTC m=+100.594548296" Oct 14 07:02:44 crc kubenswrapper[4870]: I1014 07:02:44.952389 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=78.952364974 podStartE2EDuration="1m18.952364974s" podCreationTimestamp="2025-10-14 07:01:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:02:44.940424214 +0000 UTC m=+100.637784605" watchObservedRunningTime="2025-10-14 07:02:44.952364974 +0000 UTC m=+100.649725345" Oct 14 07:02:44 crc kubenswrapper[4870]: I1014 07:02:44.953100 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=17.953094881 podStartE2EDuration="17.953094881s" podCreationTimestamp="2025-10-14 07:02:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:02:44.95220347 +0000 UTC m=+100.649563841" watchObservedRunningTime="2025-10-14 07:02:44.953094881 +0000 UTC m=+100.650455252" Oct 14 07:02:44 crc kubenswrapper[4870]: I1014 07:02:44.965723 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=44.965704607 podStartE2EDuration="44.965704607s" podCreationTimestamp="2025-10-14 07:02:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:02:44.964954589 +0000 UTC m=+100.662314970" watchObservedRunningTime="2025-10-14 07:02:44.965704607 +0000 UTC m=+100.663064978" Oct 14 07:02:45 crc kubenswrapper[4870]: I1014 07:02:45.009608 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-4v2v2" podStartSLOduration=80.009588561 podStartE2EDuration="1m20.009588561s" podCreationTimestamp="2025-10-14 07:01:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:02:44.983382766 +0000 UTC m=+100.680743147" watchObservedRunningTime="2025-10-14 07:02:45.009588561 +0000 UTC m=+100.706948932" Oct 14 07:02:45 crc kubenswrapper[4870]: I1014 07:02:45.036863 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zc5j6" Oct 14 07:02:45 crc kubenswrapper[4870]: E1014 07:02:45.037036 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zc5j6" podUID="7d7a4f41-388a-4320-b896-43c17ff13da6" Oct 14 07:02:45 crc kubenswrapper[4870]: I1014 07:02:45.045673 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=73.045658116 podStartE2EDuration="1m13.045658116s" podCreationTimestamp="2025-10-14 07:01:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:02:45.010562545 +0000 UTC m=+100.707922916" watchObservedRunningTime="2025-10-14 07:02:45.045658116 +0000 UTC m=+100.743018497" Oct 14 07:02:45 crc kubenswrapper[4870]: I1014 07:02:45.051885 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-4r29b" Oct 14 07:02:45 crc kubenswrapper[4870]: W1014 07:02:45.067215 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podda3eb795_7ffa_47d8_8fe7_ae4e1c84e0dd.slice/crio-fda48f49f2602c25d6afba195cfccf8346e6ab8bd34385d8aa2a27944e385f7a WatchSource:0}: Error finding container fda48f49f2602c25d6afba195cfccf8346e6ab8bd34385d8aa2a27944e385f7a: Status 404 returned error can't find the container with id fda48f49f2602c25d6afba195cfccf8346e6ab8bd34385d8aa2a27944e385f7a Oct 14 07:02:45 crc kubenswrapper[4870]: I1014 07:02:45.081123 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-kz77f" podStartSLOduration=80.081103016 podStartE2EDuration="1m20.081103016s" podCreationTimestamp="2025-10-14 07:01:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:02:45.079874936 +0000 UTC m=+100.777235347" watchObservedRunningTime="2025-10-14 07:02:45.081103016 +0000 UTC m=+100.778463387" Oct 14 07:02:45 crc kubenswrapper[4870]: I1014 07:02:45.128423 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podStartSLOduration=80.128401593 podStartE2EDuration="1m20.128401593s" podCreationTimestamp="2025-10-14 07:01:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:02:45.128185088 +0000 UTC m=+100.825545459" watchObservedRunningTime="2025-10-14 07:02:45.128401593 +0000 UTC m=+100.825761964" Oct 14 07:02:45 crc kubenswrapper[4870]: I1014 07:02:45.755429 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-4r29b" event={"ID":"da3eb795-7ffa-47d8-8fe7-ae4e1c84e0dd","Type":"ContainerStarted","Data":"9fa13c63a482ccfd3cbc26b4e2c504476fbf772f943223e3b835543a9493ebc7"} Oct 14 07:02:45 crc kubenswrapper[4870]: I1014 07:02:45.756669 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-4r29b" event={"ID":"da3eb795-7ffa-47d8-8fe7-ae4e1c84e0dd","Type":"ContainerStarted","Data":"fda48f49f2602c25d6afba195cfccf8346e6ab8bd34385d8aa2a27944e385f7a"} Oct 14 07:02:46 crc kubenswrapper[4870]: I1014 07:02:46.033406 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 07:02:46 crc kubenswrapper[4870]: I1014 07:02:46.033406 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 07:02:46 crc kubenswrapper[4870]: I1014 07:02:46.033532 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 07:02:46 crc kubenswrapper[4870]: E1014 07:02:46.034165 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 07:02:46 crc kubenswrapper[4870]: E1014 07:02:46.034311 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 07:02:46 crc kubenswrapper[4870]: E1014 07:02:46.034684 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 07:02:47 crc kubenswrapper[4870]: I1014 07:02:47.033614 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zc5j6" Oct 14 07:02:47 crc kubenswrapper[4870]: E1014 07:02:47.035122 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zc5j6" podUID="7d7a4f41-388a-4320-b896-43c17ff13da6" Oct 14 07:02:48 crc kubenswrapper[4870]: I1014 07:02:48.033332 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 07:02:48 crc kubenswrapper[4870]: I1014 07:02:48.033516 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 07:02:48 crc kubenswrapper[4870]: I1014 07:02:48.033614 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 07:02:48 crc kubenswrapper[4870]: E1014 07:02:48.033624 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 07:02:48 crc kubenswrapper[4870]: E1014 07:02:48.033761 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 07:02:48 crc kubenswrapper[4870]: E1014 07:02:48.033933 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 07:02:49 crc kubenswrapper[4870]: I1014 07:02:49.033997 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zc5j6" Oct 14 07:02:49 crc kubenswrapper[4870]: E1014 07:02:49.034226 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zc5j6" podUID="7d7a4f41-388a-4320-b896-43c17ff13da6" Oct 14 07:02:50 crc kubenswrapper[4870]: I1014 07:02:50.033397 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 07:02:50 crc kubenswrapper[4870]: I1014 07:02:50.033412 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 07:02:50 crc kubenswrapper[4870]: E1014 07:02:50.033623 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 07:02:50 crc kubenswrapper[4870]: I1014 07:02:50.033397 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 07:02:50 crc kubenswrapper[4870]: E1014 07:02:50.034152 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 07:02:50 crc kubenswrapper[4870]: E1014 07:02:50.034357 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 07:02:51 crc kubenswrapper[4870]: I1014 07:02:51.033295 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zc5j6" Oct 14 07:02:51 crc kubenswrapper[4870]: E1014 07:02:51.034039 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zc5j6" podUID="7d7a4f41-388a-4320-b896-43c17ff13da6" Oct 14 07:02:52 crc kubenswrapper[4870]: I1014 07:02:52.033202 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 07:02:52 crc kubenswrapper[4870]: I1014 07:02:52.033319 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 07:02:52 crc kubenswrapper[4870]: E1014 07:02:52.033434 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 07:02:52 crc kubenswrapper[4870]: E1014 07:02:52.033551 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 07:02:52 crc kubenswrapper[4870]: I1014 07:02:52.035089 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 07:02:52 crc kubenswrapper[4870]: E1014 07:02:52.035317 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 07:02:53 crc kubenswrapper[4870]: I1014 07:02:53.033498 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zc5j6" Oct 14 07:02:53 crc kubenswrapper[4870]: I1014 07:02:53.034120 4870 scope.go:117] "RemoveContainer" containerID="ab4295721b6ec6d64040d321899ab478c612e91cbc3b99b33753869578a2d1e8" Oct 14 07:02:53 crc kubenswrapper[4870]: E1014 07:02:53.034345 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-v7br2_openshift-ovn-kubernetes(8dcaef0e-42e3-424b-bbe5-4a023efe912d)\"" pod="openshift-ovn-kubernetes/ovnkube-node-v7br2" podUID="8dcaef0e-42e3-424b-bbe5-4a023efe912d" Oct 14 07:02:53 crc kubenswrapper[4870]: E1014 07:02:53.035078 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zc5j6" podUID="7d7a4f41-388a-4320-b896-43c17ff13da6" Oct 14 07:02:54 crc kubenswrapper[4870]: I1014 07:02:54.033665 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 07:02:54 crc kubenswrapper[4870]: I1014 07:02:54.033722 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 07:02:54 crc kubenswrapper[4870]: E1014 07:02:54.033877 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 07:02:54 crc kubenswrapper[4870]: E1014 07:02:54.034064 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 07:02:54 crc kubenswrapper[4870]: I1014 07:02:54.034246 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 07:02:54 crc kubenswrapper[4870]: E1014 07:02:54.034854 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 07:02:55 crc kubenswrapper[4870]: I1014 07:02:55.033275 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zc5j6" Oct 14 07:02:55 crc kubenswrapper[4870]: E1014 07:02:55.035664 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zc5j6" podUID="7d7a4f41-388a-4320-b896-43c17ff13da6" Oct 14 07:02:56 crc kubenswrapper[4870]: I1014 07:02:56.034058 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 07:02:56 crc kubenswrapper[4870]: I1014 07:02:56.034244 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 07:02:56 crc kubenswrapper[4870]: E1014 07:02:56.034349 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 07:02:56 crc kubenswrapper[4870]: I1014 07:02:56.034246 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 07:02:56 crc kubenswrapper[4870]: E1014 07:02:56.034642 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 07:02:56 crc kubenswrapper[4870]: E1014 07:02:56.034781 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 07:02:57 crc kubenswrapper[4870]: I1014 07:02:57.033613 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zc5j6" Oct 14 07:02:57 crc kubenswrapper[4870]: E1014 07:02:57.033892 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zc5j6" podUID="7d7a4f41-388a-4320-b896-43c17ff13da6" Oct 14 07:02:58 crc kubenswrapper[4870]: I1014 07:02:58.033685 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 07:02:58 crc kubenswrapper[4870]: I1014 07:02:58.033863 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 07:02:58 crc kubenswrapper[4870]: E1014 07:02:58.033951 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 07:02:58 crc kubenswrapper[4870]: E1014 07:02:58.034154 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 07:02:58 crc kubenswrapper[4870]: I1014 07:02:58.034682 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 07:02:58 crc kubenswrapper[4870]: E1014 07:02:58.034873 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 07:02:59 crc kubenswrapper[4870]: I1014 07:02:59.033723 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zc5j6" Oct 14 07:02:59 crc kubenswrapper[4870]: E1014 07:02:59.033965 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zc5j6" podUID="7d7a4f41-388a-4320-b896-43c17ff13da6" Oct 14 07:03:00 crc kubenswrapper[4870]: I1014 07:03:00.033481 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 07:03:00 crc kubenswrapper[4870]: E1014 07:03:00.033683 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 07:03:00 crc kubenswrapper[4870]: I1014 07:03:00.033847 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 07:03:00 crc kubenswrapper[4870]: I1014 07:03:00.033938 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 07:03:00 crc kubenswrapper[4870]: E1014 07:03:00.034126 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 07:03:00 crc kubenswrapper[4870]: E1014 07:03:00.034267 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 07:03:00 crc kubenswrapper[4870]: I1014 07:03:00.820155 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-fdxjh_8a3a430a-b540-4a8c-adad-b2893dbf9898/kube-multus/1.log" Oct 14 07:03:00 crc kubenswrapper[4870]: I1014 07:03:00.821048 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-fdxjh_8a3a430a-b540-4a8c-adad-b2893dbf9898/kube-multus/0.log" Oct 14 07:03:00 crc kubenswrapper[4870]: I1014 07:03:00.821142 4870 generic.go:334] "Generic (PLEG): container finished" podID="8a3a430a-b540-4a8c-adad-b2893dbf9898" containerID="88614373bfe4779bbe61cc41cc4c35f419e438bdae3cec4964271bf0f74ad3f5" exitCode=1 Oct 14 07:03:00 crc kubenswrapper[4870]: I1014 07:03:00.821200 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-fdxjh" event={"ID":"8a3a430a-b540-4a8c-adad-b2893dbf9898","Type":"ContainerDied","Data":"88614373bfe4779bbe61cc41cc4c35f419e438bdae3cec4964271bf0f74ad3f5"} Oct 14 07:03:00 crc kubenswrapper[4870]: I1014 07:03:00.821265 4870 scope.go:117] "RemoveContainer" containerID="93e41d5e927eed7ec62eb9c3607cc804cd8999556663363fd123b2a784b00ff1" Oct 14 07:03:00 crc kubenswrapper[4870]: I1014 07:03:00.823750 4870 scope.go:117] "RemoveContainer" containerID="88614373bfe4779bbe61cc41cc4c35f419e438bdae3cec4964271bf0f74ad3f5" Oct 14 07:03:00 crc kubenswrapper[4870]: E1014 07:03:00.824136 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-fdxjh_openshift-multus(8a3a430a-b540-4a8c-adad-b2893dbf9898)\"" pod="openshift-multus/multus-fdxjh" podUID="8a3a430a-b540-4a8c-adad-b2893dbf9898" Oct 14 07:03:00 crc kubenswrapper[4870]: I1014 07:03:00.852961 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-4r29b" podStartSLOduration=95.852922127 podStartE2EDuration="1m35.852922127s" podCreationTimestamp="2025-10-14 07:01:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:02:45.798000964 +0000 UTC m=+101.495361375" watchObservedRunningTime="2025-10-14 07:03:00.852922127 +0000 UTC m=+116.550282528" Oct 14 07:03:01 crc kubenswrapper[4870]: I1014 07:03:01.033583 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zc5j6" Oct 14 07:03:01 crc kubenswrapper[4870]: E1014 07:03:01.033833 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zc5j6" podUID="7d7a4f41-388a-4320-b896-43c17ff13da6" Oct 14 07:03:01 crc kubenswrapper[4870]: I1014 07:03:01.828677 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-fdxjh_8a3a430a-b540-4a8c-adad-b2893dbf9898/kube-multus/1.log" Oct 14 07:03:02 crc kubenswrapper[4870]: I1014 07:03:02.033139 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 07:03:02 crc kubenswrapper[4870]: I1014 07:03:02.033300 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 07:03:02 crc kubenswrapper[4870]: E1014 07:03:02.033384 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 07:03:02 crc kubenswrapper[4870]: E1014 07:03:02.033590 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 07:03:02 crc kubenswrapper[4870]: I1014 07:03:02.033139 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 07:03:02 crc kubenswrapper[4870]: E1014 07:03:02.033831 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 07:03:03 crc kubenswrapper[4870]: I1014 07:03:03.033798 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zc5j6" Oct 14 07:03:03 crc kubenswrapper[4870]: E1014 07:03:03.034108 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zc5j6" podUID="7d7a4f41-388a-4320-b896-43c17ff13da6" Oct 14 07:03:04 crc kubenswrapper[4870]: I1014 07:03:04.033775 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 07:03:04 crc kubenswrapper[4870]: I1014 07:03:04.033775 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 07:03:04 crc kubenswrapper[4870]: E1014 07:03:04.034139 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 07:03:04 crc kubenswrapper[4870]: E1014 07:03:04.034222 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 07:03:04 crc kubenswrapper[4870]: I1014 07:03:04.034607 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 07:03:04 crc kubenswrapper[4870]: E1014 07:03:04.034850 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 07:03:05 crc kubenswrapper[4870]: E1014 07:03:05.033096 4870 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Oct 14 07:03:05 crc kubenswrapper[4870]: I1014 07:03:05.033202 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zc5j6" Oct 14 07:03:05 crc kubenswrapper[4870]: E1014 07:03:05.035984 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zc5j6" podUID="7d7a4f41-388a-4320-b896-43c17ff13da6" Oct 14 07:03:05 crc kubenswrapper[4870]: I1014 07:03:05.037827 4870 scope.go:117] "RemoveContainer" containerID="ab4295721b6ec6d64040d321899ab478c612e91cbc3b99b33753869578a2d1e8" Oct 14 07:03:05 crc kubenswrapper[4870]: E1014 07:03:05.038142 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-v7br2_openshift-ovn-kubernetes(8dcaef0e-42e3-424b-bbe5-4a023efe912d)\"" pod="openshift-ovn-kubernetes/ovnkube-node-v7br2" podUID="8dcaef0e-42e3-424b-bbe5-4a023efe912d" Oct 14 07:03:05 crc kubenswrapper[4870]: E1014 07:03:05.136134 4870 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 14 07:03:06 crc kubenswrapper[4870]: I1014 07:03:06.033904 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 07:03:06 crc kubenswrapper[4870]: I1014 07:03:06.033904 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 07:03:06 crc kubenswrapper[4870]: E1014 07:03:06.034628 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 07:03:06 crc kubenswrapper[4870]: I1014 07:03:06.034036 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 07:03:06 crc kubenswrapper[4870]: E1014 07:03:06.034784 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 07:03:06 crc kubenswrapper[4870]: E1014 07:03:06.034860 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 07:03:07 crc kubenswrapper[4870]: I1014 07:03:07.033323 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zc5j6" Oct 14 07:03:07 crc kubenswrapper[4870]: E1014 07:03:07.033611 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zc5j6" podUID="7d7a4f41-388a-4320-b896-43c17ff13da6" Oct 14 07:03:08 crc kubenswrapper[4870]: I1014 07:03:08.034119 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 07:03:08 crc kubenswrapper[4870]: I1014 07:03:08.034213 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 07:03:08 crc kubenswrapper[4870]: I1014 07:03:08.034283 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 07:03:08 crc kubenswrapper[4870]: E1014 07:03:08.034473 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 07:03:08 crc kubenswrapper[4870]: E1014 07:03:08.034637 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 07:03:08 crc kubenswrapper[4870]: E1014 07:03:08.034838 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 07:03:09 crc kubenswrapper[4870]: I1014 07:03:09.033099 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zc5j6" Oct 14 07:03:09 crc kubenswrapper[4870]: E1014 07:03:09.034054 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zc5j6" podUID="7d7a4f41-388a-4320-b896-43c17ff13da6" Oct 14 07:03:10 crc kubenswrapper[4870]: I1014 07:03:10.033684 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 07:03:10 crc kubenswrapper[4870]: I1014 07:03:10.033785 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 07:03:10 crc kubenswrapper[4870]: I1014 07:03:10.033892 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 07:03:10 crc kubenswrapper[4870]: E1014 07:03:10.034044 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 07:03:10 crc kubenswrapper[4870]: E1014 07:03:10.034336 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 07:03:10 crc kubenswrapper[4870]: E1014 07:03:10.034600 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 07:03:10 crc kubenswrapper[4870]: E1014 07:03:10.138180 4870 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 14 07:03:11 crc kubenswrapper[4870]: I1014 07:03:11.033118 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zc5j6" Oct 14 07:03:11 crc kubenswrapper[4870]: E1014 07:03:11.033723 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zc5j6" podUID="7d7a4f41-388a-4320-b896-43c17ff13da6" Oct 14 07:03:12 crc kubenswrapper[4870]: I1014 07:03:12.033266 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 07:03:12 crc kubenswrapper[4870]: I1014 07:03:12.033363 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 07:03:12 crc kubenswrapper[4870]: I1014 07:03:12.033265 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 07:03:12 crc kubenswrapper[4870]: E1014 07:03:12.033507 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 07:03:12 crc kubenswrapper[4870]: E1014 07:03:12.033617 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 07:03:12 crc kubenswrapper[4870]: E1014 07:03:12.033805 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 07:03:13 crc kubenswrapper[4870]: I1014 07:03:13.032948 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zc5j6" Oct 14 07:03:13 crc kubenswrapper[4870]: E1014 07:03:13.033152 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zc5j6" podUID="7d7a4f41-388a-4320-b896-43c17ff13da6" Oct 14 07:03:14 crc kubenswrapper[4870]: I1014 07:03:14.033663 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 07:03:14 crc kubenswrapper[4870]: I1014 07:03:14.033781 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 07:03:14 crc kubenswrapper[4870]: E1014 07:03:14.034125 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 07:03:14 crc kubenswrapper[4870]: I1014 07:03:14.034171 4870 scope.go:117] "RemoveContainer" containerID="88614373bfe4779bbe61cc41cc4c35f419e438bdae3cec4964271bf0f74ad3f5" Oct 14 07:03:14 crc kubenswrapper[4870]: E1014 07:03:14.034407 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 07:03:14 crc kubenswrapper[4870]: I1014 07:03:14.039010 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 07:03:14 crc kubenswrapper[4870]: E1014 07:03:14.039514 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 07:03:14 crc kubenswrapper[4870]: I1014 07:03:14.885096 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-fdxjh_8a3a430a-b540-4a8c-adad-b2893dbf9898/kube-multus/1.log" Oct 14 07:03:14 crc kubenswrapper[4870]: I1014 07:03:14.885202 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-fdxjh" event={"ID":"8a3a430a-b540-4a8c-adad-b2893dbf9898","Type":"ContainerStarted","Data":"33696bfa1ec686bccf5f7def87cbf3dc2186321f9ba1c180148db5604f27cc90"} Oct 14 07:03:15 crc kubenswrapper[4870]: I1014 07:03:15.035828 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zc5j6" Oct 14 07:03:15 crc kubenswrapper[4870]: E1014 07:03:15.038231 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zc5j6" podUID="7d7a4f41-388a-4320-b896-43c17ff13da6" Oct 14 07:03:15 crc kubenswrapper[4870]: E1014 07:03:15.139513 4870 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 14 07:03:16 crc kubenswrapper[4870]: I1014 07:03:16.033315 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 07:03:16 crc kubenswrapper[4870]: I1014 07:03:16.033639 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 07:03:16 crc kubenswrapper[4870]: I1014 07:03:16.034301 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 07:03:16 crc kubenswrapper[4870]: E1014 07:03:16.034580 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 07:03:16 crc kubenswrapper[4870]: E1014 07:03:16.034779 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 07:03:16 crc kubenswrapper[4870]: E1014 07:03:16.034977 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 07:03:16 crc kubenswrapper[4870]: I1014 07:03:16.035177 4870 scope.go:117] "RemoveContainer" containerID="ab4295721b6ec6d64040d321899ab478c612e91cbc3b99b33753869578a2d1e8" Oct 14 07:03:16 crc kubenswrapper[4870]: I1014 07:03:16.897509 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-v7br2_8dcaef0e-42e3-424b-bbe5-4a023efe912d/ovnkube-controller/3.log" Oct 14 07:03:16 crc kubenswrapper[4870]: I1014 07:03:16.901490 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-v7br2" event={"ID":"8dcaef0e-42e3-424b-bbe5-4a023efe912d","Type":"ContainerStarted","Data":"a23b08d1e6780f4c2ce9ff447bd40050abb1e353c16ebccc8c3ecc7bc9a35e52"} Oct 14 07:03:16 crc kubenswrapper[4870]: I1014 07:03:16.902059 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-v7br2" Oct 14 07:03:16 crc kubenswrapper[4870]: I1014 07:03:16.942177 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-v7br2" podStartSLOduration=110.942149561 podStartE2EDuration="1m50.942149561s" podCreationTimestamp="2025-10-14 07:01:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:03:16.940501491 +0000 UTC m=+132.637861872" watchObservedRunningTime="2025-10-14 07:03:16.942149561 +0000 UTC m=+132.639509952" Oct 14 07:03:17 crc kubenswrapper[4870]: I1014 07:03:17.033871 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zc5j6" Oct 14 07:03:17 crc kubenswrapper[4870]: E1014 07:03:17.034098 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zc5j6" podUID="7d7a4f41-388a-4320-b896-43c17ff13da6" Oct 14 07:03:17 crc kubenswrapper[4870]: I1014 07:03:17.155423 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-zc5j6"] Oct 14 07:03:17 crc kubenswrapper[4870]: I1014 07:03:17.903867 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zc5j6" Oct 14 07:03:17 crc kubenswrapper[4870]: E1014 07:03:17.904244 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zc5j6" podUID="7d7a4f41-388a-4320-b896-43c17ff13da6" Oct 14 07:03:18 crc kubenswrapper[4870]: I1014 07:03:18.033397 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 07:03:18 crc kubenswrapper[4870]: I1014 07:03:18.033424 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 07:03:18 crc kubenswrapper[4870]: E1014 07:03:18.033569 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 07:03:18 crc kubenswrapper[4870]: I1014 07:03:18.033679 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 07:03:18 crc kubenswrapper[4870]: E1014 07:03:18.033905 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 07:03:18 crc kubenswrapper[4870]: E1014 07:03:18.034016 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 07:03:20 crc kubenswrapper[4870]: I1014 07:03:20.032912 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 07:03:20 crc kubenswrapper[4870]: I1014 07:03:20.033086 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 07:03:20 crc kubenswrapper[4870]: I1014 07:03:20.033125 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zc5j6" Oct 14 07:03:20 crc kubenswrapper[4870]: E1014 07:03:20.033819 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 07:03:20 crc kubenswrapper[4870]: I1014 07:03:20.032953 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 07:03:20 crc kubenswrapper[4870]: E1014 07:03:20.034005 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zc5j6" podUID="7d7a4f41-388a-4320-b896-43c17ff13da6" Oct 14 07:03:20 crc kubenswrapper[4870]: E1014 07:03:20.034194 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 07:03:20 crc kubenswrapper[4870]: E1014 07:03:20.034333 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 07:03:20 crc kubenswrapper[4870]: E1014 07:03:20.141021 4870 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 14 07:03:22 crc kubenswrapper[4870]: I1014 07:03:22.033184 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 07:03:22 crc kubenswrapper[4870]: I1014 07:03:22.033313 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 07:03:22 crc kubenswrapper[4870]: I1014 07:03:22.033369 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 07:03:22 crc kubenswrapper[4870]: I1014 07:03:22.033548 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zc5j6" Oct 14 07:03:22 crc kubenswrapper[4870]: E1014 07:03:22.033536 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 07:03:22 crc kubenswrapper[4870]: E1014 07:03:22.033735 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 07:03:22 crc kubenswrapper[4870]: E1014 07:03:22.033751 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zc5j6" podUID="7d7a4f41-388a-4320-b896-43c17ff13da6" Oct 14 07:03:22 crc kubenswrapper[4870]: E1014 07:03:22.033817 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 07:03:24 crc kubenswrapper[4870]: I1014 07:03:24.033184 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 07:03:24 crc kubenswrapper[4870]: I1014 07:03:24.033237 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zc5j6" Oct 14 07:03:24 crc kubenswrapper[4870]: I1014 07:03:24.033365 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 07:03:24 crc kubenswrapper[4870]: E1014 07:03:24.033371 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 14 07:03:24 crc kubenswrapper[4870]: E1014 07:03:24.033481 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 14 07:03:24 crc kubenswrapper[4870]: E1014 07:03:24.033581 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zc5j6" podUID="7d7a4f41-388a-4320-b896-43c17ff13da6" Oct 14 07:03:24 crc kubenswrapper[4870]: I1014 07:03:24.034258 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 07:03:24 crc kubenswrapper[4870]: E1014 07:03:24.034396 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 14 07:03:25 crc kubenswrapper[4870]: I1014 07:03:25.940356 4870 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Oct 14 07:03:25 crc kubenswrapper[4870]: I1014 07:03:25.993875 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-nc79f"] Oct 14 07:03:25 crc kubenswrapper[4870]: I1014 07:03:25.994428 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-nc79f" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.001038 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.002387 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.003023 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.003194 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.006042 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.011460 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-4wddg"] Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.011963 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-4wddg" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.012094 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.013143 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-q9knv"] Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.013612 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-q9knv" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.021765 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-r9t2k"] Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.022540 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-r9t2k" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.022606 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.023130 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.023546 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.023679 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-x8qtb"] Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.023837 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.024088 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.024210 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.024260 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-x8qtb" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.024387 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.024518 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.024813 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.025996 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-2qq5l"] Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.026530 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.026902 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-2qq5l" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.032759 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-9lgs8"] Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.033595 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zc5j6" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.033741 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.034133 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.034242 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.034289 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.034327 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-9lgs8" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.034409 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.034461 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.034478 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.034820 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.034858 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.034882 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.035053 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.035096 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.035104 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.036292 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.036603 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.036768 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.038431 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.038520 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.038803 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.039875 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.040047 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.040170 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.039881 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.040537 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.040664 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.040819 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.040909 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.044517 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-n8z6k"] Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.045394 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-n8z6k" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.046076 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.046573 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.046595 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.048864 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.049195 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.049528 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.049670 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.049820 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.050000 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.050216 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.050471 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.050715 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.050755 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.052996 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tfmnf\" (UniqueName: \"kubernetes.io/projected/172ac2db-a387-4718-9874-fc770b806a17-kube-api-access-tfmnf\") pod \"machine-api-operator-5694c8668f-nc79f\" (UID: \"172ac2db-a387-4718-9874-fc770b806a17\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-nc79f" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.053055 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/172ac2db-a387-4718-9874-fc770b806a17-images\") pod \"machine-api-operator-5694c8668f-nc79f\" (UID: \"172ac2db-a387-4718-9874-fc770b806a17\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-nc79f" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.053073 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/172ac2db-a387-4718-9874-fc770b806a17-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-nc79f\" (UID: \"172ac2db-a387-4718-9874-fc770b806a17\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-nc79f" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.053111 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/172ac2db-a387-4718-9874-fc770b806a17-config\") pod \"machine-api-operator-5694c8668f-nc79f\" (UID: \"172ac2db-a387-4718-9874-fc770b806a17\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-nc79f" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.055628 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.055906 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.056183 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.057564 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.059314 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-4wngz"] Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.060008 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-4wngz" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.072558 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-55gqt"] Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.077302 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.086773 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.091159 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.092361 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.092585 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.093167 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.094351 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.097789 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.099001 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.102058 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.105186 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.105425 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.105677 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.105791 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.108232 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-zpgbg"] Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.108646 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-55gqt" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.108238 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.110575 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-cvh5p"] Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.110808 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-zpgbg" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.110951 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-p769c"] Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.111360 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-d7cdf"] Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.111792 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-d7cdf" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.112210 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-cgwzv"] Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.112888 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-cgwzv" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.112906 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-cvh5p" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.112897 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.112995 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-p769c" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.120420 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.120985 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.123286 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-7jz27"] Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.124120 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-7jz27" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.125295 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-szwj9"] Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.126078 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-6z4z4"] Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.126571 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-szwj9" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.126690 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-6z4z4" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.134738 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-lbq8g"] Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.135357 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-lbq8g" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.139086 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.146862 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-h82f7"] Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.153144 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.174969 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.175128 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.175188 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.175384 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/03d1e16b-e364-4902-9412-0dc138fc76e3-etcd-serving-ca\") pod \"apiserver-76f77b778f-n8z6k\" (UID: \"03d1e16b-e364-4902-9412-0dc138fc76e3\") " pod="openshift-apiserver/apiserver-76f77b778f-n8z6k" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.175448 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/9b9caf4b-0443-4cfe-b9d3-6d7093d07db7-auth-proxy-config\") pod \"machine-approver-56656f9798-9lgs8\" (UID: \"9b9caf4b-0443-4cfe-b9d3-6d7093d07db7\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-9lgs8" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.175473 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v4v5r\" (UniqueName: \"kubernetes.io/projected/6f0fb8e4-f3ee-44ea-a437-d8f3d900c7fe-kube-api-access-v4v5r\") pod \"cluster-image-registry-operator-dc59b4c8b-4wddg\" (UID: \"6f0fb8e4-f3ee-44ea-a437-d8f3d900c7fe\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-4wddg" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.175472 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.175546 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-h82f7" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.175498 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3638c2bd-8853-410f-bac4-0d46d9dfa8a1-serving-cert\") pod \"apiserver-7bbb656c7d-q9knv\" (UID: \"3638c2bd-8853-410f-bac4-0d46d9dfa8a1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-q9knv" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.175694 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/03d1e16b-e364-4902-9412-0dc138fc76e3-serving-cert\") pod \"apiserver-76f77b778f-n8z6k\" (UID: \"03d1e16b-e364-4902-9412-0dc138fc76e3\") " pod="openshift-apiserver/apiserver-76f77b778f-n8z6k" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.175722 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9b9caf4b-0443-4cfe-b9d3-6d7093d07db7-config\") pod \"machine-approver-56656f9798-9lgs8\" (UID: \"9b9caf4b-0443-4cfe-b9d3-6d7093d07db7\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-9lgs8" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.175746 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/50732192-84e4-41fa-9c91-e5f77d85fc5e-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-x8qtb\" (UID: \"50732192-84e4-41fa-9c91-e5f77d85fc5e\") " pod="openshift-authentication/oauth-openshift-558db77b4-x8qtb" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.175767 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/50732192-84e4-41fa-9c91-e5f77d85fc5e-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-x8qtb\" (UID: \"50732192-84e4-41fa-9c91-e5f77d85fc5e\") " pod="openshift-authentication/oauth-openshift-558db77b4-x8qtb" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.175794 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/dce0e109-9f42-4320-b7bb-36dfd8aecb90-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-4wngz\" (UID: \"dce0e109-9f42-4320-b7bb-36dfd8aecb90\") " pod="openshift-controller-manager/controller-manager-879f6c89f-4wngz" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.175823 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/b38b0739-21a6-4050-af6c-6accabefcc45-oauth-serving-cert\") pod \"console-f9d7485db-r9t2k\" (UID: \"b38b0739-21a6-4050-af6c-6accabefcc45\") " pod="openshift-console/console-f9d7485db-r9t2k" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.175864 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/50732192-84e4-41fa-9c91-e5f77d85fc5e-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-x8qtb\" (UID: \"50732192-84e4-41fa-9c91-e5f77d85fc5e\") " pod="openshift-authentication/oauth-openshift-558db77b4-x8qtb" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.175887 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/50732192-84e4-41fa-9c91-e5f77d85fc5e-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-x8qtb\" (UID: \"50732192-84e4-41fa-9c91-e5f77d85fc5e\") " pod="openshift-authentication/oauth-openshift-558db77b4-x8qtb" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.175912 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/50732192-84e4-41fa-9c91-e5f77d85fc5e-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-x8qtb\" (UID: \"50732192-84e4-41fa-9c91-e5f77d85fc5e\") " pod="openshift-authentication/oauth-openshift-558db77b4-x8qtb" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.175941 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d5lhs\" (UniqueName: \"kubernetes.io/projected/dce0e109-9f42-4320-b7bb-36dfd8aecb90-kube-api-access-d5lhs\") pod \"controller-manager-879f6c89f-4wngz\" (UID: \"dce0e109-9f42-4320-b7bb-36dfd8aecb90\") " pod="openshift-controller-manager/controller-manager-879f6c89f-4wngz" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.175960 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t8p86\" (UniqueName: \"kubernetes.io/projected/b38b0739-21a6-4050-af6c-6accabefcc45-kube-api-access-t8p86\") pod \"console-f9d7485db-r9t2k\" (UID: \"b38b0739-21a6-4050-af6c-6accabefcc45\") " pod="openshift-console/console-f9d7485db-r9t2k" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.176001 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dce0e109-9f42-4320-b7bb-36dfd8aecb90-config\") pod \"controller-manager-879f6c89f-4wngz\" (UID: \"dce0e109-9f42-4320-b7bb-36dfd8aecb90\") " pod="openshift-controller-manager/controller-manager-879f6c89f-4wngz" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.176062 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/b38b0739-21a6-4050-af6c-6accabefcc45-console-oauth-config\") pod \"console-f9d7485db-r9t2k\" (UID: \"b38b0739-21a6-4050-af6c-6accabefcc45\") " pod="openshift-console/console-f9d7485db-r9t2k" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.176167 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r4vbf\" (UniqueName: \"kubernetes.io/projected/3638c2bd-8853-410f-bac4-0d46d9dfa8a1-kube-api-access-r4vbf\") pod \"apiserver-7bbb656c7d-q9knv\" (UID: \"3638c2bd-8853-410f-bac4-0d46d9dfa8a1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-q9knv" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.176222 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3638c2bd-8853-410f-bac4-0d46d9dfa8a1-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-q9knv\" (UID: \"3638c2bd-8853-410f-bac4-0d46d9dfa8a1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-q9knv" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.176255 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/03d1e16b-e364-4902-9412-0dc138fc76e3-config\") pod \"apiserver-76f77b778f-n8z6k\" (UID: \"03d1e16b-e364-4902-9412-0dc138fc76e3\") " pod="openshift-apiserver/apiserver-76f77b778f-n8z6k" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.176303 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/03d1e16b-e364-4902-9412-0dc138fc76e3-etcd-client\") pod \"apiserver-76f77b778f-n8z6k\" (UID: \"03d1e16b-e364-4902-9412-0dc138fc76e3\") " pod="openshift-apiserver/apiserver-76f77b778f-n8z6k" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.176330 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8hk97\" (UniqueName: \"kubernetes.io/projected/9b9caf4b-0443-4cfe-b9d3-6d7093d07db7-kube-api-access-8hk97\") pod \"machine-approver-56656f9798-9lgs8\" (UID: \"9b9caf4b-0443-4cfe-b9d3-6d7093d07db7\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-9lgs8" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.176348 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/50732192-84e4-41fa-9c91-e5f77d85fc5e-audit-dir\") pod \"oauth-openshift-558db77b4-x8qtb\" (UID: \"50732192-84e4-41fa-9c91-e5f77d85fc5e\") " pod="openshift-authentication/oauth-openshift-558db77b4-x8qtb" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.176371 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tfmnf\" (UniqueName: \"kubernetes.io/projected/172ac2db-a387-4718-9874-fc770b806a17-kube-api-access-tfmnf\") pod \"machine-api-operator-5694c8668f-nc79f\" (UID: \"172ac2db-a387-4718-9874-fc770b806a17\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-nc79f" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.176406 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/3638c2bd-8853-410f-bac4-0d46d9dfa8a1-audit-dir\") pod \"apiserver-7bbb656c7d-q9knv\" (UID: \"3638c2bd-8853-410f-bac4-0d46d9dfa8a1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-q9knv" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.176450 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/dce0e109-9f42-4320-b7bb-36dfd8aecb90-client-ca\") pod \"controller-manager-879f6c89f-4wngz\" (UID: \"dce0e109-9f42-4320-b7bb-36dfd8aecb90\") " pod="openshift-controller-manager/controller-manager-879f6c89f-4wngz" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.176463 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.176473 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/3638c2bd-8853-410f-bac4-0d46d9dfa8a1-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-q9knv\" (UID: \"3638c2bd-8853-410f-bac4-0d46d9dfa8a1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-q9knv" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.176490 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dce0e109-9f42-4320-b7bb-36dfd8aecb90-serving-cert\") pod \"controller-manager-879f6c89f-4wngz\" (UID: \"dce0e109-9f42-4320-b7bb-36dfd8aecb90\") " pod="openshift-controller-manager/controller-manager-879f6c89f-4wngz" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.176511 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/50732192-84e4-41fa-9c91-e5f77d85fc5e-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-x8qtb\" (UID: \"50732192-84e4-41fa-9c91-e5f77d85fc5e\") " pod="openshift-authentication/oauth-openshift-558db77b4-x8qtb" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.176532 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/03d1e16b-e364-4902-9412-0dc138fc76e3-trusted-ca-bundle\") pod \"apiserver-76f77b778f-n8z6k\" (UID: \"03d1e16b-e364-4902-9412-0dc138fc76e3\") " pod="openshift-apiserver/apiserver-76f77b778f-n8z6k" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.176553 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/03d1e16b-e364-4902-9412-0dc138fc76e3-audit-dir\") pod \"apiserver-76f77b778f-n8z6k\" (UID: \"03d1e16b-e364-4902-9412-0dc138fc76e3\") " pod="openshift-apiserver/apiserver-76f77b778f-n8z6k" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.176575 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.176583 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/50732192-84e4-41fa-9c91-e5f77d85fc5e-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-x8qtb\" (UID: \"50732192-84e4-41fa-9c91-e5f77d85fc5e\") " pod="openshift-authentication/oauth-openshift-558db77b4-x8qtb" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.176603 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/50732192-84e4-41fa-9c91-e5f77d85fc5e-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-x8qtb\" (UID: \"50732192-84e4-41fa-9c91-e5f77d85fc5e\") " pod="openshift-authentication/oauth-openshift-558db77b4-x8qtb" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.176615 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.176619 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9f15578c-a74c-4876-aaf3-09eb81cb0987-config\") pod \"console-operator-58897d9998-2qq5l\" (UID: \"9f15578c-a74c-4876-aaf3-09eb81cb0987\") " pod="openshift-console-operator/console-operator-58897d9998-2qq5l" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.176672 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b38b0739-21a6-4050-af6c-6accabefcc45-trusted-ca-bundle\") pod \"console-f9d7485db-r9t2k\" (UID: \"b38b0739-21a6-4050-af6c-6accabefcc45\") " pod="openshift-console/console-f9d7485db-r9t2k" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.176693 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.176698 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/b38b0739-21a6-4050-af6c-6accabefcc45-console-config\") pod \"console-f9d7485db-r9t2k\" (UID: \"b38b0739-21a6-4050-af6c-6accabefcc45\") " pod="openshift-console/console-f9d7485db-r9t2k" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.176717 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/3638c2bd-8853-410f-bac4-0d46d9dfa8a1-etcd-client\") pod \"apiserver-7bbb656c7d-q9knv\" (UID: \"3638c2bd-8853-410f-bac4-0d46d9dfa8a1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-q9knv" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.176733 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/b38b0739-21a6-4050-af6c-6accabefcc45-console-serving-cert\") pod \"console-f9d7485db-r9t2k\" (UID: \"b38b0739-21a6-4050-af6c-6accabefcc45\") " pod="openshift-console/console-f9d7485db-r9t2k" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.176754 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/6f0fb8e4-f3ee-44ea-a437-d8f3d900c7fe-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-4wddg\" (UID: \"6f0fb8e4-f3ee-44ea-a437-d8f3d900c7fe\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-4wddg" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.176776 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9f15578c-a74c-4876-aaf3-09eb81cb0987-trusted-ca\") pod \"console-operator-58897d9998-2qq5l\" (UID: \"9f15578c-a74c-4876-aaf3-09eb81cb0987\") " pod="openshift-console-operator/console-operator-58897d9998-2qq5l" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.176758 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.176798 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/b38b0739-21a6-4050-af6c-6accabefcc45-service-ca\") pod \"console-f9d7485db-r9t2k\" (UID: \"b38b0739-21a6-4050-af6c-6accabefcc45\") " pod="openshift-console/console-f9d7485db-r9t2k" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.176835 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/50732192-84e4-41fa-9c91-e5f77d85fc5e-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-x8qtb\" (UID: \"50732192-84e4-41fa-9c91-e5f77d85fc5e\") " pod="openshift-authentication/oauth-openshift-558db77b4-x8qtb" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.176842 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.176781 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.176864 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/172ac2db-a387-4718-9874-fc770b806a17-images\") pod \"machine-api-operator-5694c8668f-nc79f\" (UID: \"172ac2db-a387-4718-9874-fc770b806a17\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-nc79f" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.176893 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/3638c2bd-8853-410f-bac4-0d46d9dfa8a1-audit-policies\") pod \"apiserver-7bbb656c7d-q9knv\" (UID: \"3638c2bd-8853-410f-bac4-0d46d9dfa8a1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-q9knv" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.176914 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7rpgv\" (UniqueName: \"kubernetes.io/projected/9f15578c-a74c-4876-aaf3-09eb81cb0987-kube-api-access-7rpgv\") pod \"console-operator-58897d9998-2qq5l\" (UID: \"9f15578c-a74c-4876-aaf3-09eb81cb0987\") " pod="openshift-console-operator/console-operator-58897d9998-2qq5l" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.176929 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.176946 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/172ac2db-a387-4718-9874-fc770b806a17-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-nc79f\" (UID: \"172ac2db-a387-4718-9874-fc770b806a17\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-nc79f" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.176963 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.176964 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/50732192-84e4-41fa-9c91-e5f77d85fc5e-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-x8qtb\" (UID: \"50732192-84e4-41fa-9c91-e5f77d85fc5e\") " pod="openshift-authentication/oauth-openshift-558db77b4-x8qtb" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.176991 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.177098 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.177114 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.177145 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/03d1e16b-e364-4902-9412-0dc138fc76e3-node-pullsecrets\") pod \"apiserver-76f77b778f-n8z6k\" (UID: \"03d1e16b-e364-4902-9412-0dc138fc76e3\") " pod="openshift-apiserver/apiserver-76f77b778f-n8z6k" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.177182 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.177195 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.177192 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9b6cj\" (UniqueName: \"kubernetes.io/projected/03d1e16b-e364-4902-9412-0dc138fc76e3-kube-api-access-9b6cj\") pod \"apiserver-76f77b778f-n8z6k\" (UID: \"03d1e16b-e364-4902-9412-0dc138fc76e3\") " pod="openshift-apiserver/apiserver-76f77b778f-n8z6k" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.177230 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/9b9caf4b-0443-4cfe-b9d3-6d7093d07db7-machine-approver-tls\") pod \"machine-approver-56656f9798-9lgs8\" (UID: \"9b9caf4b-0443-4cfe-b9d3-6d7093d07db7\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-9lgs8" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.177257 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/50732192-84e4-41fa-9c91-e5f77d85fc5e-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-x8qtb\" (UID: \"50732192-84e4-41fa-9c91-e5f77d85fc5e\") " pod="openshift-authentication/oauth-openshift-558db77b4-x8qtb" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.177017 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.177295 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.177286 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/6f0fb8e4-f3ee-44ea-a437-d8f3d900c7fe-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-4wddg\" (UID: \"6f0fb8e4-f3ee-44ea-a437-d8f3d900c7fe\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-4wddg" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.177319 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9plcl\" (UniqueName: \"kubernetes.io/projected/50732192-84e4-41fa-9c91-e5f77d85fc5e-kube-api-access-9plcl\") pod \"oauth-openshift-558db77b4-x8qtb\" (UID: \"50732192-84e4-41fa-9c91-e5f77d85fc5e\") " pod="openshift-authentication/oauth-openshift-558db77b4-x8qtb" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.177332 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.177340 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/172ac2db-a387-4718-9874-fc770b806a17-config\") pod \"machine-api-operator-5694c8668f-nc79f\" (UID: \"172ac2db-a387-4718-9874-fc770b806a17\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-nc79f" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.177378 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/50732192-84e4-41fa-9c91-e5f77d85fc5e-audit-policies\") pod \"oauth-openshift-558db77b4-x8qtb\" (UID: \"50732192-84e4-41fa-9c91-e5f77d85fc5e\") " pod="openshift-authentication/oauth-openshift-558db77b4-x8qtb" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.177396 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/03d1e16b-e364-4902-9412-0dc138fc76e3-audit\") pod \"apiserver-76f77b778f-n8z6k\" (UID: \"03d1e16b-e364-4902-9412-0dc138fc76e3\") " pod="openshift-apiserver/apiserver-76f77b778f-n8z6k" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.177411 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/03d1e16b-e364-4902-9412-0dc138fc76e3-encryption-config\") pod \"apiserver-76f77b778f-n8z6k\" (UID: \"03d1e16b-e364-4902-9412-0dc138fc76e3\") " pod="openshift-apiserver/apiserver-76f77b778f-n8z6k" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.177431 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/6f0fb8e4-f3ee-44ea-a437-d8f3d900c7fe-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-4wddg\" (UID: \"6f0fb8e4-f3ee-44ea-a437-d8f3d900c7fe\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-4wddg" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.177472 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/3638c2bd-8853-410f-bac4-0d46d9dfa8a1-encryption-config\") pod \"apiserver-7bbb656c7d-q9knv\" (UID: \"3638c2bd-8853-410f-bac4-0d46d9dfa8a1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-q9knv" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.177547 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/03d1e16b-e364-4902-9412-0dc138fc76e3-image-import-ca\") pod \"apiserver-76f77b778f-n8z6k\" (UID: \"03d1e16b-e364-4902-9412-0dc138fc76e3\") " pod="openshift-apiserver/apiserver-76f77b778f-n8z6k" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.177575 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9f15578c-a74c-4876-aaf3-09eb81cb0987-serving-cert\") pod \"console-operator-58897d9998-2qq5l\" (UID: \"9f15578c-a74c-4876-aaf3-09eb81cb0987\") " pod="openshift-console-operator/console-operator-58897d9998-2qq5l" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.177777 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/172ac2db-a387-4718-9874-fc770b806a17-images\") pod \"machine-api-operator-5694c8668f-nc79f\" (UID: \"172ac2db-a387-4718-9874-fc770b806a17\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-nc79f" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.177943 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.177953 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/172ac2db-a387-4718-9874-fc770b806a17-config\") pod \"machine-api-operator-5694c8668f-nc79f\" (UID: \"172ac2db-a387-4718-9874-fc770b806a17\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-nc79f" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.178311 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.178378 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.178478 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-g7crn"] Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.178515 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.179016 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-g7crn" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.179378 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ss574"] Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.179756 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ss574" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.180509 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-5dsrq"] Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.181103 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-5dsrq" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.183109 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.187998 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.188678 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/172ac2db-a387-4718-9874-fc770b806a17-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-nc79f\" (UID: \"172ac2db-a387-4718-9874-fc770b806a17\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-nc79f" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.188955 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.192272 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.194252 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-kkxjz"] Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.195147 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-kkxjz" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.195611 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vc8ks"] Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.196147 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vc8ks" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.196204 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-mlqqj"] Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.196897 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-mlqqj" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.198941 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-hwfsm"] Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.199786 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-hwfsm" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.200012 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wjm9z"] Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.200477 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.201547 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wjm9z" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.204873 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-kqxpq"] Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.205789 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-kqxpq" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.206072 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-l24lg"] Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.206739 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-l24lg" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.208212 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-6vdcc"] Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.209360 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-6vdcc" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.211789 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rnf2p"] Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.212381 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rnf2p" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.212925 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-5ptfj"] Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.213712 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-5ptfj" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.214274 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-52qbc"] Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.215120 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-52qbc" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.215990 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29340420-t6cgj"] Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.216478 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29340420-t6cgj" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.218006 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.218063 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-6ztwm"] Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.224349 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-6ztwm" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.224693 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-69t6j"] Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.225863 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-69t6j" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.226369 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-wkc9x"] Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.228700 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-wkc9x" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.229206 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-nc79f"] Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.230495 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-q9knv"] Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.234487 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-r9t2k"] Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.235947 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-4wddg"] Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.237596 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-ptjzm"] Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.238354 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.238997 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-ptjzm" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.240156 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-zpgbg"] Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.252795 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-2qq5l"] Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.255351 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-p769c"] Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.257842 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-d7cdf"] Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.258085 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.259586 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-55gqt"] Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.265621 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-g7crn"] Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.267235 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ss574"] Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.268534 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-cvh5p"] Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.269917 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-x8qtb"] Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.271108 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-hwfsm"] Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.272189 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-4wngz"] Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.273264 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-cgwzv"] Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.275326 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-n8z6k"] Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.278166 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.278354 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-52qbc"] Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.278505 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/03d1e16b-e364-4902-9412-0dc138fc76e3-audit\") pod \"apiserver-76f77b778f-n8z6k\" (UID: \"03d1e16b-e364-4902-9412-0dc138fc76e3\") " pod="openshift-apiserver/apiserver-76f77b778f-n8z6k" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.278540 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/03d1e16b-e364-4902-9412-0dc138fc76e3-encryption-config\") pod \"apiserver-76f77b778f-n8z6k\" (UID: \"03d1e16b-e364-4902-9412-0dc138fc76e3\") " pod="openshift-apiserver/apiserver-76f77b778f-n8z6k" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.278559 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/50732192-84e4-41fa-9c91-e5f77d85fc5e-audit-policies\") pod \"oauth-openshift-558db77b4-x8qtb\" (UID: \"50732192-84e4-41fa-9c91-e5f77d85fc5e\") " pod="openshift-authentication/oauth-openshift-558db77b4-x8qtb" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.278577 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/6f0fb8e4-f3ee-44ea-a437-d8f3d900c7fe-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-4wddg\" (UID: \"6f0fb8e4-f3ee-44ea-a437-d8f3d900c7fe\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-4wddg" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.278597 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/3638c2bd-8853-410f-bac4-0d46d9dfa8a1-encryption-config\") pod \"apiserver-7bbb656c7d-q9knv\" (UID: \"3638c2bd-8853-410f-bac4-0d46d9dfa8a1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-q9knv" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.278621 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/337df1ca-c52e-499c-ab1d-a32aee9c9922-service-ca-bundle\") pod \"authentication-operator-69f744f599-cvh5p\" (UID: \"337df1ca-c52e-499c-ab1d-a32aee9c9922\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-cvh5p" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.278638 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/03d1e16b-e364-4902-9412-0dc138fc76e3-image-import-ca\") pod \"apiserver-76f77b778f-n8z6k\" (UID: \"03d1e16b-e364-4902-9412-0dc138fc76e3\") " pod="openshift-apiserver/apiserver-76f77b778f-n8z6k" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.278657 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9f15578c-a74c-4876-aaf3-09eb81cb0987-serving-cert\") pod \"console-operator-58897d9998-2qq5l\" (UID: \"9f15578c-a74c-4876-aaf3-09eb81cb0987\") " pod="openshift-console-operator/console-operator-58897d9998-2qq5l" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.278698 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b419b072-5f0d-4103-82c6-c296f71eee8a-client-ca\") pod \"route-controller-manager-6576b87f9c-7jz27\" (UID: \"b419b072-5f0d-4103-82c6-c296f71eee8a\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-7jz27" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.278717 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/3f44b737-38ee-4c89-85fb-264a66ce8771-etcd-service-ca\") pod \"etcd-operator-b45778765-5dsrq\" (UID: \"3f44b737-38ee-4c89-85fb-264a66ce8771\") " pod="openshift-etcd-operator/etcd-operator-b45778765-5dsrq" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.278737 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jh6v9\" (UniqueName: \"kubernetes.io/projected/88363344-bc9a-4050-87e2-cf6f07708835-kube-api-access-jh6v9\") pod \"openshift-apiserver-operator-796bbdcf4f-cgwzv\" (UID: \"88363344-bc9a-4050-87e2-cf6f07708835\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-cgwzv" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.278761 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/03d1e16b-e364-4902-9412-0dc138fc76e3-etcd-serving-ca\") pod \"apiserver-76f77b778f-n8z6k\" (UID: \"03d1e16b-e364-4902-9412-0dc138fc76e3\") " pod="openshift-apiserver/apiserver-76f77b778f-n8z6k" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.278782 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/9b9caf4b-0443-4cfe-b9d3-6d7093d07db7-auth-proxy-config\") pod \"machine-approver-56656f9798-9lgs8\" (UID: \"9b9caf4b-0443-4cfe-b9d3-6d7093d07db7\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-9lgs8" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.278803 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v4v5r\" (UniqueName: \"kubernetes.io/projected/6f0fb8e4-f3ee-44ea-a437-d8f3d900c7fe-kube-api-access-v4v5r\") pod \"cluster-image-registry-operator-dc59b4c8b-4wddg\" (UID: \"6f0fb8e4-f3ee-44ea-a437-d8f3d900c7fe\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-4wddg" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.278822 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3638c2bd-8853-410f-bac4-0d46d9dfa8a1-serving-cert\") pod \"apiserver-7bbb656c7d-q9knv\" (UID: \"3638c2bd-8853-410f-bac4-0d46d9dfa8a1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-q9knv" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.278845 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tl684\" (UniqueName: \"kubernetes.io/projected/8070623a-4696-490e-af90-5b2155c0506c-kube-api-access-tl684\") pod \"cluster-samples-operator-665b6dd947-szwj9\" (UID: \"8070623a-4696-490e-af90-5b2155c0506c\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-szwj9" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.278865 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9b9caf4b-0443-4cfe-b9d3-6d7093d07db7-config\") pod \"machine-approver-56656f9798-9lgs8\" (UID: \"9b9caf4b-0443-4cfe-b9d3-6d7093d07db7\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-9lgs8" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.278885 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/50732192-84e4-41fa-9c91-e5f77d85fc5e-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-x8qtb\" (UID: \"50732192-84e4-41fa-9c91-e5f77d85fc5e\") " pod="openshift-authentication/oauth-openshift-558db77b4-x8qtb" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.278904 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/50732192-84e4-41fa-9c91-e5f77d85fc5e-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-x8qtb\" (UID: \"50732192-84e4-41fa-9c91-e5f77d85fc5e\") " pod="openshift-authentication/oauth-openshift-558db77b4-x8qtb" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.278925 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/88363344-bc9a-4050-87e2-cf6f07708835-config\") pod \"openshift-apiserver-operator-796bbdcf4f-cgwzv\" (UID: \"88363344-bc9a-4050-87e2-cf6f07708835\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-cgwzv" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.278947 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/c0cd54e9-09ac-49a6-a774-a9148024b4b9-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-l24lg\" (UID: \"c0cd54e9-09ac-49a6-a774-a9148024b4b9\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-l24lg" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.278969 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/03d1e16b-e364-4902-9412-0dc138fc76e3-serving-cert\") pod \"apiserver-76f77b778f-n8z6k\" (UID: \"03d1e16b-e364-4902-9412-0dc138fc76e3\") " pod="openshift-apiserver/apiserver-76f77b778f-n8z6k" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.278987 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/dce0e109-9f42-4320-b7bb-36dfd8aecb90-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-4wngz\" (UID: \"dce0e109-9f42-4320-b7bb-36dfd8aecb90\") " pod="openshift-controller-manager/controller-manager-879f6c89f-4wngz" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.279008 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/b38b0739-21a6-4050-af6c-6accabefcc45-oauth-serving-cert\") pod \"console-f9d7485db-r9t2k\" (UID: \"b38b0739-21a6-4050-af6c-6accabefcc45\") " pod="openshift-console/console-f9d7485db-r9t2k" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.279025 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/50732192-84e4-41fa-9c91-e5f77d85fc5e-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-x8qtb\" (UID: \"50732192-84e4-41fa-9c91-e5f77d85fc5e\") " pod="openshift-authentication/oauth-openshift-558db77b4-x8qtb" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.279045 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/50732192-84e4-41fa-9c91-e5f77d85fc5e-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-x8qtb\" (UID: \"50732192-84e4-41fa-9c91-e5f77d85fc5e\") " pod="openshift-authentication/oauth-openshift-558db77b4-x8qtb" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.279068 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/50732192-84e4-41fa-9c91-e5f77d85fc5e-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-x8qtb\" (UID: \"50732192-84e4-41fa-9c91-e5f77d85fc5e\") " pod="openshift-authentication/oauth-openshift-558db77b4-x8qtb" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.279086 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/337df1ca-c52e-499c-ab1d-a32aee9c9922-serving-cert\") pod \"authentication-operator-69f744f599-cvh5p\" (UID: \"337df1ca-c52e-499c-ab1d-a32aee9c9922\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-cvh5p" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.279106 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/16295052-1474-4dc9-b1d6-10983e6bfc6f-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-g7crn\" (UID: \"16295052-1474-4dc9-b1d6-10983e6bfc6f\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-g7crn" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.279129 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d9cb408e-a00b-4f76-b8fb-bf144bfee740-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-ss574\" (UID: \"d9cb408e-a00b-4f76-b8fb-bf144bfee740\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ss574" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.279152 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d5lhs\" (UniqueName: \"kubernetes.io/projected/dce0e109-9f42-4320-b7bb-36dfd8aecb90-kube-api-access-d5lhs\") pod \"controller-manager-879f6c89f-4wngz\" (UID: \"dce0e109-9f42-4320-b7bb-36dfd8aecb90\") " pod="openshift-controller-manager/controller-manager-879f6c89f-4wngz" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.279170 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t8p86\" (UniqueName: \"kubernetes.io/projected/b38b0739-21a6-4050-af6c-6accabefcc45-kube-api-access-t8p86\") pod \"console-f9d7485db-r9t2k\" (UID: \"b38b0739-21a6-4050-af6c-6accabefcc45\") " pod="openshift-console/console-f9d7485db-r9t2k" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.279190 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dce0e109-9f42-4320-b7bb-36dfd8aecb90-config\") pod \"controller-manager-879f6c89f-4wngz\" (UID: \"dce0e109-9f42-4320-b7bb-36dfd8aecb90\") " pod="openshift-controller-manager/controller-manager-879f6c89f-4wngz" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.279209 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/b38b0739-21a6-4050-af6c-6accabefcc45-console-oauth-config\") pod \"console-f9d7485db-r9t2k\" (UID: \"b38b0739-21a6-4050-af6c-6accabefcc45\") " pod="openshift-console/console-f9d7485db-r9t2k" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.279311 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2tjr6\" (UniqueName: \"kubernetes.io/projected/146160a7-2b1c-43f4-bc42-7b92d9663dcc-kube-api-access-2tjr6\") pod \"openshift-config-operator-7777fb866f-p769c\" (UID: \"146160a7-2b1c-43f4-bc42-7b92d9663dcc\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-p769c" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.279359 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r4vbf\" (UniqueName: \"kubernetes.io/projected/3638c2bd-8853-410f-bac4-0d46d9dfa8a1-kube-api-access-r4vbf\") pod \"apiserver-7bbb656c7d-q9knv\" (UID: \"3638c2bd-8853-410f-bac4-0d46d9dfa8a1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-q9knv" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.279379 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b419b072-5f0d-4103-82c6-c296f71eee8a-serving-cert\") pod \"route-controller-manager-6576b87f9c-7jz27\" (UID: \"b419b072-5f0d-4103-82c6-c296f71eee8a\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-7jz27" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.279399 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d9cb408e-a00b-4f76-b8fb-bf144bfee740-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-ss574\" (UID: \"d9cb408e-a00b-4f76-b8fb-bf144bfee740\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ss574" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.279419 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hdxrc\" (UniqueName: \"kubernetes.io/projected/337df1ca-c52e-499c-ab1d-a32aee9c9922-kube-api-access-hdxrc\") pod \"authentication-operator-69f744f599-cvh5p\" (UID: \"337df1ca-c52e-499c-ab1d-a32aee9c9922\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-cvh5p" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.279463 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3638c2bd-8853-410f-bac4-0d46d9dfa8a1-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-q9knv\" (UID: \"3638c2bd-8853-410f-bac4-0d46d9dfa8a1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-q9knv" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.279536 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/03d1e16b-e364-4902-9412-0dc138fc76e3-config\") pod \"apiserver-76f77b778f-n8z6k\" (UID: \"03d1e16b-e364-4902-9412-0dc138fc76e3\") " pod="openshift-apiserver/apiserver-76f77b778f-n8z6k" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.279556 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3f44b737-38ee-4c89-85fb-264a66ce8771-config\") pod \"etcd-operator-b45778765-5dsrq\" (UID: \"3f44b737-38ee-4c89-85fb-264a66ce8771\") " pod="openshift-etcd-operator/etcd-operator-b45778765-5dsrq" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.279574 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/03d1e16b-e364-4902-9412-0dc138fc76e3-etcd-client\") pod \"apiserver-76f77b778f-n8z6k\" (UID: \"03d1e16b-e364-4902-9412-0dc138fc76e3\") " pod="openshift-apiserver/apiserver-76f77b778f-n8z6k" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.279595 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8hk97\" (UniqueName: \"kubernetes.io/projected/9b9caf4b-0443-4cfe-b9d3-6d7093d07db7-kube-api-access-8hk97\") pod \"machine-approver-56656f9798-9lgs8\" (UID: \"9b9caf4b-0443-4cfe-b9d3-6d7093d07db7\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-9lgs8" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.279638 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/50732192-84e4-41fa-9c91-e5f77d85fc5e-audit-dir\") pod \"oauth-openshift-558db77b4-x8qtb\" (UID: \"50732192-84e4-41fa-9c91-e5f77d85fc5e\") " pod="openshift-authentication/oauth-openshift-558db77b4-x8qtb" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.279661 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b419b072-5f0d-4103-82c6-c296f71eee8a-config\") pod \"route-controller-manager-6576b87f9c-7jz27\" (UID: \"b419b072-5f0d-4103-82c6-c296f71eee8a\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-7jz27" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.279687 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3f44b737-38ee-4c89-85fb-264a66ce8771-serving-cert\") pod \"etcd-operator-b45778765-5dsrq\" (UID: \"3f44b737-38ee-4c89-85fb-264a66ce8771\") " pod="openshift-etcd-operator/etcd-operator-b45778765-5dsrq" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.279708 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f07c76fa-f754-4175-8bc5-2aae95985b3b-profile-collector-cert\") pod \"olm-operator-6b444d44fb-wjm9z\" (UID: \"f07c76fa-f754-4175-8bc5-2aae95985b3b\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wjm9z" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.279730 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-46lfv\" (UniqueName: \"kubernetes.io/projected/3f44b737-38ee-4c89-85fb-264a66ce8771-kube-api-access-46lfv\") pod \"etcd-operator-b45778765-5dsrq\" (UID: \"3f44b737-38ee-4c89-85fb-264a66ce8771\") " pod="openshift-etcd-operator/etcd-operator-b45778765-5dsrq" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.279754 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/16295052-1474-4dc9-b1d6-10983e6bfc6f-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-g7crn\" (UID: \"16295052-1474-4dc9-b1d6-10983e6bfc6f\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-g7crn" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.279786 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/3638c2bd-8853-410f-bac4-0d46d9dfa8a1-audit-dir\") pod \"apiserver-7bbb656c7d-q9knv\" (UID: \"3638c2bd-8853-410f-bac4-0d46d9dfa8a1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-q9knv" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.279807 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/dce0e109-9f42-4320-b7bb-36dfd8aecb90-client-ca\") pod \"controller-manager-879f6c89f-4wngz\" (UID: \"dce0e109-9f42-4320-b7bb-36dfd8aecb90\") " pod="openshift-controller-manager/controller-manager-879f6c89f-4wngz" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.279829 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/146160a7-2b1c-43f4-bc42-7b92d9663dcc-serving-cert\") pod \"openshift-config-operator-7777fb866f-p769c\" (UID: \"146160a7-2b1c-43f4-bc42-7b92d9663dcc\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-p769c" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.279850 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f07c76fa-f754-4175-8bc5-2aae95985b3b-srv-cert\") pod \"olm-operator-6b444d44fb-wjm9z\" (UID: \"f07c76fa-f754-4175-8bc5-2aae95985b3b\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wjm9z" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.279897 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8nr6p\" (UniqueName: \"kubernetes.io/projected/c0cd54e9-09ac-49a6-a774-a9148024b4b9-kube-api-access-8nr6p\") pod \"multus-admission-controller-857f4d67dd-l24lg\" (UID: \"c0cd54e9-09ac-49a6-a774-a9148024b4b9\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-l24lg" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.279922 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/3638c2bd-8853-410f-bac4-0d46d9dfa8a1-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-q9knv\" (UID: \"3638c2bd-8853-410f-bac4-0d46d9dfa8a1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-q9knv" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.279944 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dce0e109-9f42-4320-b7bb-36dfd8aecb90-serving-cert\") pod \"controller-manager-879f6c89f-4wngz\" (UID: \"dce0e109-9f42-4320-b7bb-36dfd8aecb90\") " pod="openshift-controller-manager/controller-manager-879f6c89f-4wngz" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.279964 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/88363344-bc9a-4050-87e2-cf6f07708835-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-cgwzv\" (UID: \"88363344-bc9a-4050-87e2-cf6f07708835\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-cgwzv" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.279985 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/03d1e16b-e364-4902-9412-0dc138fc76e3-trusted-ca-bundle\") pod \"apiserver-76f77b778f-n8z6k\" (UID: \"03d1e16b-e364-4902-9412-0dc138fc76e3\") " pod="openshift-apiserver/apiserver-76f77b778f-n8z6k" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.280005 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/03d1e16b-e364-4902-9412-0dc138fc76e3-audit-dir\") pod \"apiserver-76f77b778f-n8z6k\" (UID: \"03d1e16b-e364-4902-9412-0dc138fc76e3\") " pod="openshift-apiserver/apiserver-76f77b778f-n8z6k" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.280031 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/50732192-84e4-41fa-9c91-e5f77d85fc5e-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-x8qtb\" (UID: \"50732192-84e4-41fa-9c91-e5f77d85fc5e\") " pod="openshift-authentication/oauth-openshift-558db77b4-x8qtb" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.280068 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tk4zz\" (UniqueName: \"kubernetes.io/projected/49dc86fe-e718-4220-81d7-2ebd7fdd6c2f-kube-api-access-tk4zz\") pod \"openshift-controller-manager-operator-756b6f6bc6-55gqt\" (UID: \"49dc86fe-e718-4220-81d7-2ebd7fdd6c2f\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-55gqt" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.280130 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/50732192-84e4-41fa-9c91-e5f77d85fc5e-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-x8qtb\" (UID: \"50732192-84e4-41fa-9c91-e5f77d85fc5e\") " pod="openshift-authentication/oauth-openshift-558db77b4-x8qtb" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.280148 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/50732192-84e4-41fa-9c91-e5f77d85fc5e-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-x8qtb\" (UID: \"50732192-84e4-41fa-9c91-e5f77d85fc5e\") " pod="openshift-authentication/oauth-openshift-558db77b4-x8qtb" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.280169 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cb6jb\" (UniqueName: \"kubernetes.io/projected/b419b072-5f0d-4103-82c6-c296f71eee8a-kube-api-access-cb6jb\") pod \"route-controller-manager-6576b87f9c-7jz27\" (UID: \"b419b072-5f0d-4103-82c6-c296f71eee8a\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-7jz27" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.280189 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/50732192-84e4-41fa-9c91-e5f77d85fc5e-audit-policies\") pod \"oauth-openshift-558db77b4-x8qtb\" (UID: \"50732192-84e4-41fa-9c91-e5f77d85fc5e\") " pod="openshift-authentication/oauth-openshift-558db77b4-x8qtb" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.280205 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9f15578c-a74c-4876-aaf3-09eb81cb0987-config\") pod \"console-operator-58897d9998-2qq5l\" (UID: \"9f15578c-a74c-4876-aaf3-09eb81cb0987\") " pod="openshift-console-operator/console-operator-58897d9998-2qq5l" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.280219 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/03d1e16b-e364-4902-9412-0dc138fc76e3-audit\") pod \"apiserver-76f77b778f-n8z6k\" (UID: \"03d1e16b-e364-4902-9412-0dc138fc76e3\") " pod="openshift-apiserver/apiserver-76f77b778f-n8z6k" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.280230 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b38b0739-21a6-4050-af6c-6accabefcc45-trusted-ca-bundle\") pod \"console-f9d7485db-r9t2k\" (UID: \"b38b0739-21a6-4050-af6c-6accabefcc45\") " pod="openshift-console/console-f9d7485db-r9t2k" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.280351 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-56bq2\" (UniqueName: \"kubernetes.io/projected/222f1962-d6c1-4e69-a3bc-0b0a12703cff-kube-api-access-56bq2\") pod \"dns-operator-744455d44c-zpgbg\" (UID: \"222f1962-d6c1-4e69-a3bc-0b0a12703cff\") " pod="openshift-dns-operator/dns-operator-744455d44c-zpgbg" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.280395 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/3638c2bd-8853-410f-bac4-0d46d9dfa8a1-etcd-client\") pod \"apiserver-7bbb656c7d-q9knv\" (UID: \"3638c2bd-8853-410f-bac4-0d46d9dfa8a1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-q9knv" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.280431 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/b38b0739-21a6-4050-af6c-6accabefcc45-console-serving-cert\") pod \"console-f9d7485db-r9t2k\" (UID: \"b38b0739-21a6-4050-af6c-6accabefcc45\") " pod="openshift-console/console-f9d7485db-r9t2k" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.280482 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/b38b0739-21a6-4050-af6c-6accabefcc45-console-config\") pod \"console-f9d7485db-r9t2k\" (UID: \"b38b0739-21a6-4050-af6c-6accabefcc45\") " pod="openshift-console/console-f9d7485db-r9t2k" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.280513 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t25cc\" (UniqueName: \"kubernetes.io/projected/7b7a521b-c795-43cc-9576-5a3be3cb7409-kube-api-access-t25cc\") pod \"migrator-59844c95c7-hwfsm\" (UID: \"7b7a521b-c795-43cc-9576-5a3be3cb7409\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-hwfsm" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.280553 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/49dc86fe-e718-4220-81d7-2ebd7fdd6c2f-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-55gqt\" (UID: \"49dc86fe-e718-4220-81d7-2ebd7fdd6c2f\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-55gqt" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.280584 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/49dc86fe-e718-4220-81d7-2ebd7fdd6c2f-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-55gqt\" (UID: \"49dc86fe-e718-4220-81d7-2ebd7fdd6c2f\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-55gqt" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.280616 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/6f0fb8e4-f3ee-44ea-a437-d8f3d900c7fe-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-4wddg\" (UID: \"6f0fb8e4-f3ee-44ea-a437-d8f3d900c7fe\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-4wddg" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.280646 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9f15578c-a74c-4876-aaf3-09eb81cb0987-trusted-ca\") pod \"console-operator-58897d9998-2qq5l\" (UID: \"9f15578c-a74c-4876-aaf3-09eb81cb0987\") " pod="openshift-console-operator/console-operator-58897d9998-2qq5l" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.280675 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d9cb408e-a00b-4f76-b8fb-bf144bfee740-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-ss574\" (UID: \"d9cb408e-a00b-4f76-b8fb-bf144bfee740\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ss574" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.280714 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/3638c2bd-8853-410f-bac4-0d46d9dfa8a1-audit-policies\") pod \"apiserver-7bbb656c7d-q9knv\" (UID: \"3638c2bd-8853-410f-bac4-0d46d9dfa8a1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-q9knv" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.280712 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/03d1e16b-e364-4902-9412-0dc138fc76e3-etcd-serving-ca\") pod \"apiserver-76f77b778f-n8z6k\" (UID: \"03d1e16b-e364-4902-9412-0dc138fc76e3\") " pod="openshift-apiserver/apiserver-76f77b778f-n8z6k" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.280749 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7rpgv\" (UniqueName: \"kubernetes.io/projected/9f15578c-a74c-4876-aaf3-09eb81cb0987-kube-api-access-7rpgv\") pod \"console-operator-58897d9998-2qq5l\" (UID: \"9f15578c-a74c-4876-aaf3-09eb81cb0987\") " pod="openshift-console-operator/console-operator-58897d9998-2qq5l" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.280784 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/b38b0739-21a6-4050-af6c-6accabefcc45-service-ca\") pod \"console-f9d7485db-r9t2k\" (UID: \"b38b0739-21a6-4050-af6c-6accabefcc45\") " pod="openshift-console/console-f9d7485db-r9t2k" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.280815 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/50732192-84e4-41fa-9c91-e5f77d85fc5e-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-x8qtb\" (UID: \"50732192-84e4-41fa-9c91-e5f77d85fc5e\") " pod="openshift-authentication/oauth-openshift-558db77b4-x8qtb" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.280823 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/50732192-84e4-41fa-9c91-e5f77d85fc5e-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-x8qtb\" (UID: \"50732192-84e4-41fa-9c91-e5f77d85fc5e\") " pod="openshift-authentication/oauth-openshift-558db77b4-x8qtb" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.280845 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gb2b5\" (UniqueName: \"kubernetes.io/projected/f07c76fa-f754-4175-8bc5-2aae95985b3b-kube-api-access-gb2b5\") pod \"olm-operator-6b444d44fb-wjm9z\" (UID: \"f07c76fa-f754-4175-8bc5-2aae95985b3b\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wjm9z" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.280933 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/50732192-84e4-41fa-9c91-e5f77d85fc5e-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-x8qtb\" (UID: \"50732192-84e4-41fa-9c91-e5f77d85fc5e\") " pod="openshift-authentication/oauth-openshift-558db77b4-x8qtb" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.280972 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/8070623a-4696-490e-af90-5b2155c0506c-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-szwj9\" (UID: \"8070623a-4696-490e-af90-5b2155c0506c\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-szwj9" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.281000 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/222f1962-d6c1-4e69-a3bc-0b0a12703cff-metrics-tls\") pod \"dns-operator-744455d44c-zpgbg\" (UID: \"222f1962-d6c1-4e69-a3bc-0b0a12703cff\") " pod="openshift-dns-operator/dns-operator-744455d44c-zpgbg" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.281035 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/337df1ca-c52e-499c-ab1d-a32aee9c9922-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-cvh5p\" (UID: \"337df1ca-c52e-499c-ab1d-a32aee9c9922\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-cvh5p" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.281064 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/3f44b737-38ee-4c89-85fb-264a66ce8771-etcd-ca\") pod \"etcd-operator-b45778765-5dsrq\" (UID: \"3f44b737-38ee-4c89-85fb-264a66ce8771\") " pod="openshift-etcd-operator/etcd-operator-b45778765-5dsrq" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.281112 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/03d1e16b-e364-4902-9412-0dc138fc76e3-node-pullsecrets\") pod \"apiserver-76f77b778f-n8z6k\" (UID: \"03d1e16b-e364-4902-9412-0dc138fc76e3\") " pod="openshift-apiserver/apiserver-76f77b778f-n8z6k" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.281147 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9b6cj\" (UniqueName: \"kubernetes.io/projected/03d1e16b-e364-4902-9412-0dc138fc76e3-kube-api-access-9b6cj\") pod \"apiserver-76f77b778f-n8z6k\" (UID: \"03d1e16b-e364-4902-9412-0dc138fc76e3\") " pod="openshift-apiserver/apiserver-76f77b778f-n8z6k" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.281175 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/9b9caf4b-0443-4cfe-b9d3-6d7093d07db7-machine-approver-tls\") pod \"machine-approver-56656f9798-9lgs8\" (UID: \"9b9caf4b-0443-4cfe-b9d3-6d7093d07db7\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-9lgs8" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.281203 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/50732192-84e4-41fa-9c91-e5f77d85fc5e-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-x8qtb\" (UID: \"50732192-84e4-41fa-9c91-e5f77d85fc5e\") " pod="openshift-authentication/oauth-openshift-558db77b4-x8qtb" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.281228 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/3f44b737-38ee-4c89-85fb-264a66ce8771-etcd-client\") pod \"etcd-operator-b45778765-5dsrq\" (UID: \"3f44b737-38ee-4c89-85fb-264a66ce8771\") " pod="openshift-etcd-operator/etcd-operator-b45778765-5dsrq" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.281252 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/146160a7-2b1c-43f4-bc42-7b92d9663dcc-available-featuregates\") pod \"openshift-config-operator-7777fb866f-p769c\" (UID: \"146160a7-2b1c-43f4-bc42-7b92d9663dcc\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-p769c" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.281285 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/6f0fb8e4-f3ee-44ea-a437-d8f3d900c7fe-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-4wddg\" (UID: \"6f0fb8e4-f3ee-44ea-a437-d8f3d900c7fe\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-4wddg" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.281315 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9plcl\" (UniqueName: \"kubernetes.io/projected/50732192-84e4-41fa-9c91-e5f77d85fc5e-kube-api-access-9plcl\") pod \"oauth-openshift-558db77b4-x8qtb\" (UID: \"50732192-84e4-41fa-9c91-e5f77d85fc5e\") " pod="openshift-authentication/oauth-openshift-558db77b4-x8qtb" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.281356 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/9b9caf4b-0443-4cfe-b9d3-6d7093d07db7-auth-proxy-config\") pod \"machine-approver-56656f9798-9lgs8\" (UID: \"9b9caf4b-0443-4cfe-b9d3-6d7093d07db7\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-9lgs8" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.281381 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/16295052-1474-4dc9-b1d6-10983e6bfc6f-config\") pod \"kube-controller-manager-operator-78b949d7b-g7crn\" (UID: \"16295052-1474-4dc9-b1d6-10983e6bfc6f\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-g7crn" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.281433 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/337df1ca-c52e-499c-ab1d-a32aee9c9922-config\") pod \"authentication-operator-69f744f599-cvh5p\" (UID: \"337df1ca-c52e-499c-ab1d-a32aee9c9922\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-cvh5p" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.281626 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b38b0739-21a6-4050-af6c-6accabefcc45-trusted-ca-bundle\") pod \"console-f9d7485db-r9t2k\" (UID: \"b38b0739-21a6-4050-af6c-6accabefcc45\") " pod="openshift-console/console-f9d7485db-r9t2k" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.281869 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-kkxjz"] Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.281911 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-kqxpq"] Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.281921 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-6ztwm"] Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.282602 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/50732192-84e4-41fa-9c91-e5f77d85fc5e-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-x8qtb\" (UID: \"50732192-84e4-41fa-9c91-e5f77d85fc5e\") " pod="openshift-authentication/oauth-openshift-558db77b4-x8qtb" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.282614 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/03d1e16b-e364-4902-9412-0dc138fc76e3-serving-cert\") pod \"apiserver-76f77b778f-n8z6k\" (UID: \"03d1e16b-e364-4902-9412-0dc138fc76e3\") " pod="openshift-apiserver/apiserver-76f77b778f-n8z6k" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.282817 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/03d1e16b-e364-4902-9412-0dc138fc76e3-node-pullsecrets\") pod \"apiserver-76f77b778f-n8z6k\" (UID: \"03d1e16b-e364-4902-9412-0dc138fc76e3\") " pod="openshift-apiserver/apiserver-76f77b778f-n8z6k" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.282881 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/50732192-84e4-41fa-9c91-e5f77d85fc5e-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-x8qtb\" (UID: \"50732192-84e4-41fa-9c91-e5f77d85fc5e\") " pod="openshift-authentication/oauth-openshift-558db77b4-x8qtb" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.282934 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dce0e109-9f42-4320-b7bb-36dfd8aecb90-config\") pod \"controller-manager-879f6c89f-4wngz\" (UID: \"dce0e109-9f42-4320-b7bb-36dfd8aecb90\") " pod="openshift-controller-manager/controller-manager-879f6c89f-4wngz" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.282994 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/dce0e109-9f42-4320-b7bb-36dfd8aecb90-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-4wngz\" (UID: \"dce0e109-9f42-4320-b7bb-36dfd8aecb90\") " pod="openshift-controller-manager/controller-manager-879f6c89f-4wngz" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.283530 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/3638c2bd-8853-410f-bac4-0d46d9dfa8a1-etcd-client\") pod \"apiserver-7bbb656c7d-q9knv\" (UID: \"3638c2bd-8853-410f-bac4-0d46d9dfa8a1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-q9knv" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.283842 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/3638c2bd-8853-410f-bac4-0d46d9dfa8a1-audit-policies\") pod \"apiserver-7bbb656c7d-q9knv\" (UID: \"3638c2bd-8853-410f-bac4-0d46d9dfa8a1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-q9knv" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.284114 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/6f0fb8e4-f3ee-44ea-a437-d8f3d900c7fe-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-4wddg\" (UID: \"6f0fb8e4-f3ee-44ea-a437-d8f3d900c7fe\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-4wddg" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.284116 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/3638c2bd-8853-410f-bac4-0d46d9dfa8a1-encryption-config\") pod \"apiserver-7bbb656c7d-q9knv\" (UID: \"3638c2bd-8853-410f-bac4-0d46d9dfa8a1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-q9knv" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.284390 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/b38b0739-21a6-4050-af6c-6accabefcc45-oauth-serving-cert\") pod \"console-f9d7485db-r9t2k\" (UID: \"b38b0739-21a6-4050-af6c-6accabefcc45\") " pod="openshift-console/console-f9d7485db-r9t2k" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.284463 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/b38b0739-21a6-4050-af6c-6accabefcc45-service-ca\") pod \"console-f9d7485db-r9t2k\" (UID: \"b38b0739-21a6-4050-af6c-6accabefcc45\") " pod="openshift-console/console-f9d7485db-r9t2k" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.284969 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9f15578c-a74c-4876-aaf3-09eb81cb0987-trusted-ca\") pod \"console-operator-58897d9998-2qq5l\" (UID: \"9f15578c-a74c-4876-aaf3-09eb81cb0987\") " pod="openshift-console-operator/console-operator-58897d9998-2qq5l" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.284997 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-q26kj"] Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.285075 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/03d1e16b-e364-4902-9412-0dc138fc76e3-image-import-ca\") pod \"apiserver-76f77b778f-n8z6k\" (UID: \"03d1e16b-e364-4902-9412-0dc138fc76e3\") " pod="openshift-apiserver/apiserver-76f77b778f-n8z6k" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.285262 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/50732192-84e4-41fa-9c91-e5f77d85fc5e-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-x8qtb\" (UID: \"50732192-84e4-41fa-9c91-e5f77d85fc5e\") " pod="openshift-authentication/oauth-openshift-558db77b4-x8qtb" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.285708 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/b38b0739-21a6-4050-af6c-6accabefcc45-console-oauth-config\") pod \"console-f9d7485db-r9t2k\" (UID: \"b38b0739-21a6-4050-af6c-6accabefcc45\") " pod="openshift-console/console-f9d7485db-r9t2k" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.285764 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/3638c2bd-8853-410f-bac4-0d46d9dfa8a1-audit-dir\") pod \"apiserver-7bbb656c7d-q9knv\" (UID: \"3638c2bd-8853-410f-bac4-0d46d9dfa8a1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-q9knv" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.285793 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-d4fhp"] Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.286032 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3638c2bd-8853-410f-bac4-0d46d9dfa8a1-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-q9knv\" (UID: \"3638c2bd-8853-410f-bac4-0d46d9dfa8a1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-q9knv" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.286261 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/b38b0739-21a6-4050-af6c-6accabefcc45-console-config\") pod \"console-f9d7485db-r9t2k\" (UID: \"b38b0739-21a6-4050-af6c-6accabefcc45\") " pod="openshift-console/console-f9d7485db-r9t2k" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.286362 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/dce0e109-9f42-4320-b7bb-36dfd8aecb90-client-ca\") pod \"controller-manager-879f6c89f-4wngz\" (UID: \"dce0e109-9f42-4320-b7bb-36dfd8aecb90\") " pod="openshift-controller-manager/controller-manager-879f6c89f-4wngz" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.286459 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29340420-t6cgj"] Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.286570 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-d4fhp" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.286608 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/03d1e16b-e364-4902-9412-0dc138fc76e3-config\") pod \"apiserver-76f77b778f-n8z6k\" (UID: \"03d1e16b-e364-4902-9412-0dc138fc76e3\") " pod="openshift-apiserver/apiserver-76f77b778f-n8z6k" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.286747 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/50732192-84e4-41fa-9c91-e5f77d85fc5e-audit-dir\") pod \"oauth-openshift-558db77b4-x8qtb\" (UID: \"50732192-84e4-41fa-9c91-e5f77d85fc5e\") " pod="openshift-authentication/oauth-openshift-558db77b4-x8qtb" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.286883 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/3638c2bd-8853-410f-bac4-0d46d9dfa8a1-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-q9knv\" (UID: \"3638c2bd-8853-410f-bac4-0d46d9dfa8a1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-q9knv" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.286932 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-q26kj" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.287489 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/9b9caf4b-0443-4cfe-b9d3-6d7093d07db7-machine-approver-tls\") pod \"machine-approver-56656f9798-9lgs8\" (UID: \"9b9caf4b-0443-4cfe-b9d3-6d7093d07db7\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-9lgs8" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.287754 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9f15578c-a74c-4876-aaf3-09eb81cb0987-serving-cert\") pod \"console-operator-58897d9998-2qq5l\" (UID: \"9f15578c-a74c-4876-aaf3-09eb81cb0987\") " pod="openshift-console-operator/console-operator-58897d9998-2qq5l" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.287802 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-6vdcc"] Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.287828 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9b9caf4b-0443-4cfe-b9d3-6d7093d07db7-config\") pod \"machine-approver-56656f9798-9lgs8\" (UID: \"9b9caf4b-0443-4cfe-b9d3-6d7093d07db7\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-9lgs8" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.288029 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/03d1e16b-e364-4902-9412-0dc138fc76e3-audit-dir\") pod \"apiserver-76f77b778f-n8z6k\" (UID: \"03d1e16b-e364-4902-9412-0dc138fc76e3\") " pod="openshift-apiserver/apiserver-76f77b778f-n8z6k" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.288461 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/50732192-84e4-41fa-9c91-e5f77d85fc5e-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-x8qtb\" (UID: \"50732192-84e4-41fa-9c91-e5f77d85fc5e\") " pod="openshift-authentication/oauth-openshift-558db77b4-x8qtb" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.288523 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/03d1e16b-e364-4902-9412-0dc138fc76e3-encryption-config\") pod \"apiserver-76f77b778f-n8z6k\" (UID: \"03d1e16b-e364-4902-9412-0dc138fc76e3\") " pod="openshift-apiserver/apiserver-76f77b778f-n8z6k" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.289085 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9f15578c-a74c-4876-aaf3-09eb81cb0987-config\") pod \"console-operator-58897d9998-2qq5l\" (UID: \"9f15578c-a74c-4876-aaf3-09eb81cb0987\") " pod="openshift-console-operator/console-operator-58897d9998-2qq5l" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.288631 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/03d1e16b-e364-4902-9412-0dc138fc76e3-trusted-ca-bundle\") pod \"apiserver-76f77b778f-n8z6k\" (UID: \"03d1e16b-e364-4902-9412-0dc138fc76e3\") " pod="openshift-apiserver/apiserver-76f77b778f-n8z6k" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.289162 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-szwj9"] Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.289237 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dce0e109-9f42-4320-b7bb-36dfd8aecb90-serving-cert\") pod \"controller-manager-879f6c89f-4wngz\" (UID: \"dce0e109-9f42-4320-b7bb-36dfd8aecb90\") " pod="openshift-controller-manager/controller-manager-879f6c89f-4wngz" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.288546 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/50732192-84e4-41fa-9c91-e5f77d85fc5e-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-x8qtb\" (UID: \"50732192-84e4-41fa-9c91-e5f77d85fc5e\") " pod="openshift-authentication/oauth-openshift-558db77b4-x8qtb" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.289487 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3638c2bd-8853-410f-bac4-0d46d9dfa8a1-serving-cert\") pod \"apiserver-7bbb656c7d-q9knv\" (UID: \"3638c2bd-8853-410f-bac4-0d46d9dfa8a1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-q9knv" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.289788 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/50732192-84e4-41fa-9c91-e5f77d85fc5e-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-x8qtb\" (UID: \"50732192-84e4-41fa-9c91-e5f77d85fc5e\") " pod="openshift-authentication/oauth-openshift-558db77b4-x8qtb" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.290031 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-5dsrq"] Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.290140 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/50732192-84e4-41fa-9c91-e5f77d85fc5e-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-x8qtb\" (UID: \"50732192-84e4-41fa-9c91-e5f77d85fc5e\") " pod="openshift-authentication/oauth-openshift-558db77b4-x8qtb" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.291068 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-mlqqj"] Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.291078 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/03d1e16b-e364-4902-9412-0dc138fc76e3-etcd-client\") pod \"apiserver-76f77b778f-n8z6k\" (UID: \"03d1e16b-e364-4902-9412-0dc138fc76e3\") " pod="openshift-apiserver/apiserver-76f77b778f-n8z6k" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.291394 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/50732192-84e4-41fa-9c91-e5f77d85fc5e-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-x8qtb\" (UID: \"50732192-84e4-41fa-9c91-e5f77d85fc5e\") " pod="openshift-authentication/oauth-openshift-558db77b4-x8qtb" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.291403 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/50732192-84e4-41fa-9c91-e5f77d85fc5e-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-x8qtb\" (UID: \"50732192-84e4-41fa-9c91-e5f77d85fc5e\") " pod="openshift-authentication/oauth-openshift-558db77b4-x8qtb" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.291580 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/50732192-84e4-41fa-9c91-e5f77d85fc5e-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-x8qtb\" (UID: \"50732192-84e4-41fa-9c91-e5f77d85fc5e\") " pod="openshift-authentication/oauth-openshift-558db77b4-x8qtb" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.293031 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-5ptfj"] Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.293068 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/6f0fb8e4-f3ee-44ea-a437-d8f3d900c7fe-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-4wddg\" (UID: \"6f0fb8e4-f3ee-44ea-a437-d8f3d900c7fe\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-4wddg" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.294752 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vc8ks"] Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.295787 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-l24lg"] Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.296765 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-7jz27"] Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.298026 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-6z4z4"] Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.299631 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rnf2p"] Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.300899 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-lbq8g"] Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.301421 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.303053 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wjm9z"] Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.304018 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-q26kj"] Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.304727 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/b38b0739-21a6-4050-af6c-6accabefcc45-console-serving-cert\") pod \"console-f9d7485db-r9t2k\" (UID: \"b38b0739-21a6-4050-af6c-6accabefcc45\") " pod="openshift-console/console-f9d7485db-r9t2k" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.306969 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-69t6j"] Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.306999 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-ptjzm"] Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.307011 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-d4fhp"] Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.321514 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-wkc9x"] Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.322726 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-klh8n"] Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.323398 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-klh8n" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.326683 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.342057 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.358118 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.381986 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b419b072-5f0d-4103-82c6-c296f71eee8a-config\") pod \"route-controller-manager-6576b87f9c-7jz27\" (UID: \"b419b072-5f0d-4103-82c6-c296f71eee8a\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-7jz27" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.382019 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3f44b737-38ee-4c89-85fb-264a66ce8771-serving-cert\") pod \"etcd-operator-b45778765-5dsrq\" (UID: \"3f44b737-38ee-4c89-85fb-264a66ce8771\") " pod="openshift-etcd-operator/etcd-operator-b45778765-5dsrq" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.382040 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-46lfv\" (UniqueName: \"kubernetes.io/projected/3f44b737-38ee-4c89-85fb-264a66ce8771-kube-api-access-46lfv\") pod \"etcd-operator-b45778765-5dsrq\" (UID: \"3f44b737-38ee-4c89-85fb-264a66ce8771\") " pod="openshift-etcd-operator/etcd-operator-b45778765-5dsrq" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.382057 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/16295052-1474-4dc9-b1d6-10983e6bfc6f-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-g7crn\" (UID: \"16295052-1474-4dc9-b1d6-10983e6bfc6f\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-g7crn" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.382077 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f07c76fa-f754-4175-8bc5-2aae95985b3b-profile-collector-cert\") pod \"olm-operator-6b444d44fb-wjm9z\" (UID: \"f07c76fa-f754-4175-8bc5-2aae95985b3b\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wjm9z" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.382112 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/146160a7-2b1c-43f4-bc42-7b92d9663dcc-serving-cert\") pod \"openshift-config-operator-7777fb866f-p769c\" (UID: \"146160a7-2b1c-43f4-bc42-7b92d9663dcc\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-p769c" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.382127 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f07c76fa-f754-4175-8bc5-2aae95985b3b-srv-cert\") pod \"olm-operator-6b444d44fb-wjm9z\" (UID: \"f07c76fa-f754-4175-8bc5-2aae95985b3b\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wjm9z" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.382151 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8nr6p\" (UniqueName: \"kubernetes.io/projected/c0cd54e9-09ac-49a6-a774-a9148024b4b9-kube-api-access-8nr6p\") pod \"multus-admission-controller-857f4d67dd-l24lg\" (UID: \"c0cd54e9-09ac-49a6-a774-a9148024b4b9\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-l24lg" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.382175 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/88363344-bc9a-4050-87e2-cf6f07708835-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-cgwzv\" (UID: \"88363344-bc9a-4050-87e2-cf6f07708835\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-cgwzv" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.382194 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tk4zz\" (UniqueName: \"kubernetes.io/projected/49dc86fe-e718-4220-81d7-2ebd7fdd6c2f-kube-api-access-tk4zz\") pod \"openshift-controller-manager-operator-756b6f6bc6-55gqt\" (UID: \"49dc86fe-e718-4220-81d7-2ebd7fdd6c2f\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-55gqt" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.382214 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cb6jb\" (UniqueName: \"kubernetes.io/projected/b419b072-5f0d-4103-82c6-c296f71eee8a-kube-api-access-cb6jb\") pod \"route-controller-manager-6576b87f9c-7jz27\" (UID: \"b419b072-5f0d-4103-82c6-c296f71eee8a\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-7jz27" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.382250 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-56bq2\" (UniqueName: \"kubernetes.io/projected/222f1962-d6c1-4e69-a3bc-0b0a12703cff-kube-api-access-56bq2\") pod \"dns-operator-744455d44c-zpgbg\" (UID: \"222f1962-d6c1-4e69-a3bc-0b0a12703cff\") " pod="openshift-dns-operator/dns-operator-744455d44c-zpgbg" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.382271 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t25cc\" (UniqueName: \"kubernetes.io/projected/7b7a521b-c795-43cc-9576-5a3be3cb7409-kube-api-access-t25cc\") pod \"migrator-59844c95c7-hwfsm\" (UID: \"7b7a521b-c795-43cc-9576-5a3be3cb7409\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-hwfsm" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.382289 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/49dc86fe-e718-4220-81d7-2ebd7fdd6c2f-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-55gqt\" (UID: \"49dc86fe-e718-4220-81d7-2ebd7fdd6c2f\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-55gqt" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.382305 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/49dc86fe-e718-4220-81d7-2ebd7fdd6c2f-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-55gqt\" (UID: \"49dc86fe-e718-4220-81d7-2ebd7fdd6c2f\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-55gqt" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.382321 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d9cb408e-a00b-4f76-b8fb-bf144bfee740-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-ss574\" (UID: \"d9cb408e-a00b-4f76-b8fb-bf144bfee740\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ss574" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.382345 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gb2b5\" (UniqueName: \"kubernetes.io/projected/f07c76fa-f754-4175-8bc5-2aae95985b3b-kube-api-access-gb2b5\") pod \"olm-operator-6b444d44fb-wjm9z\" (UID: \"f07c76fa-f754-4175-8bc5-2aae95985b3b\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wjm9z" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.382361 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/222f1962-d6c1-4e69-a3bc-0b0a12703cff-metrics-tls\") pod \"dns-operator-744455d44c-zpgbg\" (UID: \"222f1962-d6c1-4e69-a3bc-0b0a12703cff\") " pod="openshift-dns-operator/dns-operator-744455d44c-zpgbg" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.382378 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/8070623a-4696-490e-af90-5b2155c0506c-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-szwj9\" (UID: \"8070623a-4696-490e-af90-5b2155c0506c\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-szwj9" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.382395 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/337df1ca-c52e-499c-ab1d-a32aee9c9922-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-cvh5p\" (UID: \"337df1ca-c52e-499c-ab1d-a32aee9c9922\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-cvh5p" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.382411 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/3f44b737-38ee-4c89-85fb-264a66ce8771-etcd-ca\") pod \"etcd-operator-b45778765-5dsrq\" (UID: \"3f44b737-38ee-4c89-85fb-264a66ce8771\") " pod="openshift-etcd-operator/etcd-operator-b45778765-5dsrq" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.382457 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/3f44b737-38ee-4c89-85fb-264a66ce8771-etcd-client\") pod \"etcd-operator-b45778765-5dsrq\" (UID: \"3f44b737-38ee-4c89-85fb-264a66ce8771\") " pod="openshift-etcd-operator/etcd-operator-b45778765-5dsrq" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.382477 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/146160a7-2b1c-43f4-bc42-7b92d9663dcc-available-featuregates\") pod \"openshift-config-operator-7777fb866f-p769c\" (UID: \"146160a7-2b1c-43f4-bc42-7b92d9663dcc\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-p769c" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.382501 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/16295052-1474-4dc9-b1d6-10983e6bfc6f-config\") pod \"kube-controller-manager-operator-78b949d7b-g7crn\" (UID: \"16295052-1474-4dc9-b1d6-10983e6bfc6f\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-g7crn" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.382524 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/337df1ca-c52e-499c-ab1d-a32aee9c9922-config\") pod \"authentication-operator-69f744f599-cvh5p\" (UID: \"337df1ca-c52e-499c-ab1d-a32aee9c9922\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-cvh5p" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.382541 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/337df1ca-c52e-499c-ab1d-a32aee9c9922-service-ca-bundle\") pod \"authentication-operator-69f744f599-cvh5p\" (UID: \"337df1ca-c52e-499c-ab1d-a32aee9c9922\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-cvh5p" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.382563 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b419b072-5f0d-4103-82c6-c296f71eee8a-client-ca\") pod \"route-controller-manager-6576b87f9c-7jz27\" (UID: \"b419b072-5f0d-4103-82c6-c296f71eee8a\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-7jz27" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.382578 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/3f44b737-38ee-4c89-85fb-264a66ce8771-etcd-service-ca\") pod \"etcd-operator-b45778765-5dsrq\" (UID: \"3f44b737-38ee-4c89-85fb-264a66ce8771\") " pod="openshift-etcd-operator/etcd-operator-b45778765-5dsrq" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.382593 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jh6v9\" (UniqueName: \"kubernetes.io/projected/88363344-bc9a-4050-87e2-cf6f07708835-kube-api-access-jh6v9\") pod \"openshift-apiserver-operator-796bbdcf4f-cgwzv\" (UID: \"88363344-bc9a-4050-87e2-cf6f07708835\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-cgwzv" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.382610 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tl684\" (UniqueName: \"kubernetes.io/projected/8070623a-4696-490e-af90-5b2155c0506c-kube-api-access-tl684\") pod \"cluster-samples-operator-665b6dd947-szwj9\" (UID: \"8070623a-4696-490e-af90-5b2155c0506c\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-szwj9" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.382630 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/88363344-bc9a-4050-87e2-cf6f07708835-config\") pod \"openshift-apiserver-operator-796bbdcf4f-cgwzv\" (UID: \"88363344-bc9a-4050-87e2-cf6f07708835\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-cgwzv" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.382645 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/c0cd54e9-09ac-49a6-a774-a9148024b4b9-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-l24lg\" (UID: \"c0cd54e9-09ac-49a6-a774-a9148024b4b9\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-l24lg" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.382664 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/337df1ca-c52e-499c-ab1d-a32aee9c9922-serving-cert\") pod \"authentication-operator-69f744f599-cvh5p\" (UID: \"337df1ca-c52e-499c-ab1d-a32aee9c9922\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-cvh5p" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.382678 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/16295052-1474-4dc9-b1d6-10983e6bfc6f-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-g7crn\" (UID: \"16295052-1474-4dc9-b1d6-10983e6bfc6f\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-g7crn" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.382694 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d9cb408e-a00b-4f76-b8fb-bf144bfee740-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-ss574\" (UID: \"d9cb408e-a00b-4f76-b8fb-bf144bfee740\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ss574" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.382815 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2tjr6\" (UniqueName: \"kubernetes.io/projected/146160a7-2b1c-43f4-bc42-7b92d9663dcc-kube-api-access-2tjr6\") pod \"openshift-config-operator-7777fb866f-p769c\" (UID: \"146160a7-2b1c-43f4-bc42-7b92d9663dcc\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-p769c" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.382843 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b419b072-5f0d-4103-82c6-c296f71eee8a-serving-cert\") pod \"route-controller-manager-6576b87f9c-7jz27\" (UID: \"b419b072-5f0d-4103-82c6-c296f71eee8a\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-7jz27" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.382858 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d9cb408e-a00b-4f76-b8fb-bf144bfee740-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-ss574\" (UID: \"d9cb408e-a00b-4f76-b8fb-bf144bfee740\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ss574" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.382873 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hdxrc\" (UniqueName: \"kubernetes.io/projected/337df1ca-c52e-499c-ab1d-a32aee9c9922-kube-api-access-hdxrc\") pod \"authentication-operator-69f744f599-cvh5p\" (UID: \"337df1ca-c52e-499c-ab1d-a32aee9c9922\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-cvh5p" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.382896 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3f44b737-38ee-4c89-85fb-264a66ce8771-config\") pod \"etcd-operator-b45778765-5dsrq\" (UID: \"3f44b737-38ee-4c89-85fb-264a66ce8771\") " pod="openshift-etcd-operator/etcd-operator-b45778765-5dsrq" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.385234 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/146160a7-2b1c-43f4-bc42-7b92d9663dcc-serving-cert\") pod \"openshift-config-operator-7777fb866f-p769c\" (UID: \"146160a7-2b1c-43f4-bc42-7b92d9663dcc\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-p769c" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.385678 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/88363344-bc9a-4050-87e2-cf6f07708835-config\") pod \"openshift-apiserver-operator-796bbdcf4f-cgwzv\" (UID: \"88363344-bc9a-4050-87e2-cf6f07708835\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-cgwzv" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.387112 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/337df1ca-c52e-499c-ab1d-a32aee9c9922-service-ca-bundle\") pod \"authentication-operator-69f744f599-cvh5p\" (UID: \"337df1ca-c52e-499c-ab1d-a32aee9c9922\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-cvh5p" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.387274 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/146160a7-2b1c-43f4-bc42-7b92d9663dcc-available-featuregates\") pod \"openshift-config-operator-7777fb866f-p769c\" (UID: \"146160a7-2b1c-43f4-bc42-7b92d9663dcc\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-p769c" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.387401 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b419b072-5f0d-4103-82c6-c296f71eee8a-config\") pod \"route-controller-manager-6576b87f9c-7jz27\" (UID: \"b419b072-5f0d-4103-82c6-c296f71eee8a\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-7jz27" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.387493 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/337df1ca-c52e-499c-ab1d-a32aee9c9922-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-cvh5p\" (UID: \"337df1ca-c52e-499c-ab1d-a32aee9c9922\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-cvh5p" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.387834 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b419b072-5f0d-4103-82c6-c296f71eee8a-client-ca\") pod \"route-controller-manager-6576b87f9c-7jz27\" (UID: \"b419b072-5f0d-4103-82c6-c296f71eee8a\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-7jz27" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.387979 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/49dc86fe-e718-4220-81d7-2ebd7fdd6c2f-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-55gqt\" (UID: \"49dc86fe-e718-4220-81d7-2ebd7fdd6c2f\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-55gqt" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.388134 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/337df1ca-c52e-499c-ab1d-a32aee9c9922-config\") pod \"authentication-operator-69f744f599-cvh5p\" (UID: \"337df1ca-c52e-499c-ab1d-a32aee9c9922\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-cvh5p" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.388655 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/49dc86fe-e718-4220-81d7-2ebd7fdd6c2f-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-55gqt\" (UID: \"49dc86fe-e718-4220-81d7-2ebd7fdd6c2f\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-55gqt" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.389386 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/337df1ca-c52e-499c-ab1d-a32aee9c9922-serving-cert\") pod \"authentication-operator-69f744f599-cvh5p\" (UID: \"337df1ca-c52e-499c-ab1d-a32aee9c9922\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-cvh5p" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.390334 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.393008 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/222f1962-d6c1-4e69-a3bc-0b0a12703cff-metrics-tls\") pod \"dns-operator-744455d44c-zpgbg\" (UID: \"222f1962-d6c1-4e69-a3bc-0b0a12703cff\") " pod="openshift-dns-operator/dns-operator-744455d44c-zpgbg" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.394902 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/8070623a-4696-490e-af90-5b2155c0506c-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-szwj9\" (UID: \"8070623a-4696-490e-af90-5b2155c0506c\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-szwj9" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.395516 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b419b072-5f0d-4103-82c6-c296f71eee8a-serving-cert\") pod \"route-controller-manager-6576b87f9c-7jz27\" (UID: \"b419b072-5f0d-4103-82c6-c296f71eee8a\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-7jz27" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.398740 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.399148 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/88363344-bc9a-4050-87e2-cf6f07708835-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-cgwzv\" (UID: \"88363344-bc9a-4050-87e2-cf6f07708835\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-cgwzv" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.418187 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.438782 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.478112 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.498617 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.519044 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.538570 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.558612 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.578419 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.598461 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.618705 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.638162 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.658127 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.679549 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.698923 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.735128 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tfmnf\" (UniqueName: \"kubernetes.io/projected/172ac2db-a387-4718-9874-fc770b806a17-kube-api-access-tfmnf\") pod \"machine-api-operator-5694c8668f-nc79f\" (UID: \"172ac2db-a387-4718-9874-fc770b806a17\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-nc79f" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.738350 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.760304 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.780311 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.790093 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/16295052-1474-4dc9-b1d6-10983e6bfc6f-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-g7crn\" (UID: \"16295052-1474-4dc9-b1d6-10983e6bfc6f\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-g7crn" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.798526 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.805179 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/16295052-1474-4dc9-b1d6-10983e6bfc6f-config\") pod \"kube-controller-manager-operator-78b949d7b-g7crn\" (UID: \"16295052-1474-4dc9-b1d6-10983e6bfc6f\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-g7crn" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.819898 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.830771 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d9cb408e-a00b-4f76-b8fb-bf144bfee740-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-ss574\" (UID: \"d9cb408e-a00b-4f76-b8fb-bf144bfee740\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ss574" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.838897 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.850994 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d9cb408e-a00b-4f76-b8fb-bf144bfee740-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-ss574\" (UID: \"d9cb408e-a00b-4f76-b8fb-bf144bfee740\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ss574" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.859287 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.878987 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.899021 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.907243 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/3f44b737-38ee-4c89-85fb-264a66ce8771-etcd-ca\") pod \"etcd-operator-b45778765-5dsrq\" (UID: \"3f44b737-38ee-4c89-85fb-264a66ce8771\") " pod="openshift-etcd-operator/etcd-operator-b45778765-5dsrq" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.919725 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.930891 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-nc79f" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.940433 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.959302 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.974603 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3f44b737-38ee-4c89-85fb-264a66ce8771-serving-cert\") pod \"etcd-operator-b45778765-5dsrq\" (UID: \"3f44b737-38ee-4c89-85fb-264a66ce8771\") " pod="openshift-etcd-operator/etcd-operator-b45778765-5dsrq" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.978154 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.992106 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/3f44b737-38ee-4c89-85fb-264a66ce8771-etcd-client\") pod \"etcd-operator-b45778765-5dsrq\" (UID: \"3f44b737-38ee-4c89-85fb-264a66ce8771\") " pod="openshift-etcd-operator/etcd-operator-b45778765-5dsrq" Oct 14 07:03:26 crc kubenswrapper[4870]: I1014 07:03:26.999584 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Oct 14 07:03:27 crc kubenswrapper[4870]: I1014 07:03:27.019084 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Oct 14 07:03:27 crc kubenswrapper[4870]: I1014 07:03:27.027299 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/3f44b737-38ee-4c89-85fb-264a66ce8771-etcd-service-ca\") pod \"etcd-operator-b45778765-5dsrq\" (UID: \"3f44b737-38ee-4c89-85fb-264a66ce8771\") " pod="openshift-etcd-operator/etcd-operator-b45778765-5dsrq" Oct 14 07:03:27 crc kubenswrapper[4870]: I1014 07:03:27.038173 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Oct 14 07:03:27 crc kubenswrapper[4870]: I1014 07:03:27.044689 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3f44b737-38ee-4c89-85fb-264a66ce8771-config\") pod \"etcd-operator-b45778765-5dsrq\" (UID: \"3f44b737-38ee-4c89-85fb-264a66ce8771\") " pod="openshift-etcd-operator/etcd-operator-b45778765-5dsrq" Oct 14 07:03:27 crc kubenswrapper[4870]: I1014 07:03:27.059470 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Oct 14 07:03:27 crc kubenswrapper[4870]: I1014 07:03:27.118370 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Oct 14 07:03:27 crc kubenswrapper[4870]: I1014 07:03:27.118376 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Oct 14 07:03:27 crc kubenswrapper[4870]: I1014 07:03:27.118704 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Oct 14 07:03:27 crc kubenswrapper[4870]: I1014 07:03:27.143766 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Oct 14 07:03:27 crc kubenswrapper[4870]: I1014 07:03:27.159481 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Oct 14 07:03:27 crc kubenswrapper[4870]: I1014 07:03:27.178564 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Oct 14 07:03:27 crc kubenswrapper[4870]: I1014 07:03:27.197559 4870 request.go:700] Waited for 1.000363722s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-machine-config-operator/secrets?fieldSelector=metadata.name%3Dmcc-proxy-tls&limit=500&resourceVersion=0 Oct 14 07:03:27 crc kubenswrapper[4870]: I1014 07:03:27.200267 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Oct 14 07:03:27 crc kubenswrapper[4870]: I1014 07:03:27.218719 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Oct 14 07:03:27 crc kubenswrapper[4870]: I1014 07:03:27.239059 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Oct 14 07:03:27 crc kubenswrapper[4870]: I1014 07:03:27.241099 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-nc79f"] Oct 14 07:03:27 crc kubenswrapper[4870]: I1014 07:03:27.262928 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Oct 14 07:03:27 crc kubenswrapper[4870]: I1014 07:03:27.279368 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Oct 14 07:03:27 crc kubenswrapper[4870]: I1014 07:03:27.299246 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Oct 14 07:03:27 crc kubenswrapper[4870]: I1014 07:03:27.318388 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Oct 14 07:03:27 crc kubenswrapper[4870]: I1014 07:03:27.337863 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Oct 14 07:03:27 crc kubenswrapper[4870]: I1014 07:03:27.350358 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f07c76fa-f754-4175-8bc5-2aae95985b3b-srv-cert\") pod \"olm-operator-6b444d44fb-wjm9z\" (UID: \"f07c76fa-f754-4175-8bc5-2aae95985b3b\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wjm9z" Oct 14 07:03:27 crc kubenswrapper[4870]: I1014 07:03:27.358778 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Oct 14 07:03:27 crc kubenswrapper[4870]: I1014 07:03:27.370281 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f07c76fa-f754-4175-8bc5-2aae95985b3b-profile-collector-cert\") pod \"olm-operator-6b444d44fb-wjm9z\" (UID: \"f07c76fa-f754-4175-8bc5-2aae95985b3b\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wjm9z" Oct 14 07:03:27 crc kubenswrapper[4870]: I1014 07:03:27.378192 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Oct 14 07:03:27 crc kubenswrapper[4870]: E1014 07:03:27.385940 4870 secret.go:188] Couldn't get secret openshift-multus/multus-admission-controller-secret: failed to sync secret cache: timed out waiting for the condition Oct 14 07:03:27 crc kubenswrapper[4870]: E1014 07:03:27.386011 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c0cd54e9-09ac-49a6-a774-a9148024b4b9-webhook-certs podName:c0cd54e9-09ac-49a6-a774-a9148024b4b9 nodeName:}" failed. No retries permitted until 2025-10-14 07:03:27.885993102 +0000 UTC m=+143.583353473 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/c0cd54e9-09ac-49a6-a774-a9148024b4b9-webhook-certs") pod "multus-admission-controller-857f4d67dd-l24lg" (UID: "c0cd54e9-09ac-49a6-a774-a9148024b4b9") : failed to sync secret cache: timed out waiting for the condition Oct 14 07:03:27 crc kubenswrapper[4870]: I1014 07:03:27.399132 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Oct 14 07:03:27 crc kubenswrapper[4870]: I1014 07:03:27.418622 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Oct 14 07:03:27 crc kubenswrapper[4870]: I1014 07:03:27.439466 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Oct 14 07:03:27 crc kubenswrapper[4870]: I1014 07:03:27.477904 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Oct 14 07:03:27 crc kubenswrapper[4870]: I1014 07:03:27.498389 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Oct 14 07:03:27 crc kubenswrapper[4870]: I1014 07:03:27.518901 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Oct 14 07:03:27 crc kubenswrapper[4870]: I1014 07:03:27.537894 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Oct 14 07:03:27 crc kubenswrapper[4870]: I1014 07:03:27.557705 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Oct 14 07:03:27 crc kubenswrapper[4870]: I1014 07:03:27.578207 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Oct 14 07:03:27 crc kubenswrapper[4870]: I1014 07:03:27.598910 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Oct 14 07:03:27 crc kubenswrapper[4870]: I1014 07:03:27.618300 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Oct 14 07:03:27 crc kubenswrapper[4870]: I1014 07:03:27.639176 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Oct 14 07:03:27 crc kubenswrapper[4870]: I1014 07:03:27.658476 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 14 07:03:27 crc kubenswrapper[4870]: I1014 07:03:27.679112 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 14 07:03:27 crc kubenswrapper[4870]: I1014 07:03:27.697912 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Oct 14 07:03:27 crc kubenswrapper[4870]: I1014 07:03:27.729780 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Oct 14 07:03:27 crc kubenswrapper[4870]: I1014 07:03:27.738123 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Oct 14 07:03:27 crc kubenswrapper[4870]: I1014 07:03:27.758217 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Oct 14 07:03:27 crc kubenswrapper[4870]: I1014 07:03:27.778039 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Oct 14 07:03:27 crc kubenswrapper[4870]: I1014 07:03:27.799272 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Oct 14 07:03:27 crc kubenswrapper[4870]: I1014 07:03:27.819157 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Oct 14 07:03:27 crc kubenswrapper[4870]: I1014 07:03:27.839265 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Oct 14 07:03:27 crc kubenswrapper[4870]: I1014 07:03:27.859730 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Oct 14 07:03:27 crc kubenswrapper[4870]: I1014 07:03:27.879663 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Oct 14 07:03:27 crc kubenswrapper[4870]: I1014 07:03:27.899951 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Oct 14 07:03:27 crc kubenswrapper[4870]: I1014 07:03:27.918423 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Oct 14 07:03:27 crc kubenswrapper[4870]: I1014 07:03:27.926179 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/c0cd54e9-09ac-49a6-a774-a9148024b4b9-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-l24lg\" (UID: \"c0cd54e9-09ac-49a6-a774-a9148024b4b9\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-l24lg" Oct 14 07:03:27 crc kubenswrapper[4870]: I1014 07:03:27.932128 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/c0cd54e9-09ac-49a6-a774-a9148024b4b9-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-l24lg\" (UID: \"c0cd54e9-09ac-49a6-a774-a9148024b4b9\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-l24lg" Oct 14 07:03:27 crc kubenswrapper[4870]: I1014 07:03:27.939059 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Oct 14 07:03:27 crc kubenswrapper[4870]: I1014 07:03:27.947647 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-nc79f" event={"ID":"172ac2db-a387-4718-9874-fc770b806a17","Type":"ContainerStarted","Data":"d7b0d7471773571cac91b78be1211aed75f03a1d38026516a90af2f65175019f"} Oct 14 07:03:27 crc kubenswrapper[4870]: I1014 07:03:27.947734 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-nc79f" event={"ID":"172ac2db-a387-4718-9874-fc770b806a17","Type":"ContainerStarted","Data":"97e054b4566894afb59286993b7fa04ccfc61c5c41f1cf73dafe1da50b4b9fb0"} Oct 14 07:03:27 crc kubenswrapper[4870]: I1014 07:03:27.947754 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-nc79f" event={"ID":"172ac2db-a387-4718-9874-fc770b806a17","Type":"ContainerStarted","Data":"8bce298aed0fe7cc9225ed177714c568c3cc9aee8db7cc5cf013fdecfd2ec687"} Oct 14 07:03:27 crc kubenswrapper[4870]: I1014 07:03:27.960076 4870 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Oct 14 07:03:27 crc kubenswrapper[4870]: I1014 07:03:27.978129 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Oct 14 07:03:27 crc kubenswrapper[4870]: I1014 07:03:27.999534 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.033461 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t8p86\" (UniqueName: \"kubernetes.io/projected/b38b0739-21a6-4050-af6c-6accabefcc45-kube-api-access-t8p86\") pod \"console-f9d7485db-r9t2k\" (UID: \"b38b0739-21a6-4050-af6c-6accabefcc45\") " pod="openshift-console/console-f9d7485db-r9t2k" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.054611 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v4v5r\" (UniqueName: \"kubernetes.io/projected/6f0fb8e4-f3ee-44ea-a437-d8f3d900c7fe-kube-api-access-v4v5r\") pod \"cluster-image-registry-operator-dc59b4c8b-4wddg\" (UID: \"6f0fb8e4-f3ee-44ea-a437-d8f3d900c7fe\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-4wddg" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.073434 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/6f0fb8e4-f3ee-44ea-a437-d8f3d900c7fe-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-4wddg\" (UID: \"6f0fb8e4-f3ee-44ea-a437-d8f3d900c7fe\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-4wddg" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.093459 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d5lhs\" (UniqueName: \"kubernetes.io/projected/dce0e109-9f42-4320-b7bb-36dfd8aecb90-kube-api-access-d5lhs\") pod \"controller-manager-879f6c89f-4wngz\" (UID: \"dce0e109-9f42-4320-b7bb-36dfd8aecb90\") " pod="openshift-controller-manager/controller-manager-879f6c89f-4wngz" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.118050 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9b6cj\" (UniqueName: \"kubernetes.io/projected/03d1e16b-e364-4902-9412-0dc138fc76e3-kube-api-access-9b6cj\") pod \"apiserver-76f77b778f-n8z6k\" (UID: \"03d1e16b-e364-4902-9412-0dc138fc76e3\") " pod="openshift-apiserver/apiserver-76f77b778f-n8z6k" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.132542 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7rpgv\" (UniqueName: \"kubernetes.io/projected/9f15578c-a74c-4876-aaf3-09eb81cb0987-kube-api-access-7rpgv\") pod \"console-operator-58897d9998-2qq5l\" (UID: \"9f15578c-a74c-4876-aaf3-09eb81cb0987\") " pod="openshift-console-operator/console-operator-58897d9998-2qq5l" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.149382 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-4wddg" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.154081 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r4vbf\" (UniqueName: \"kubernetes.io/projected/3638c2bd-8853-410f-bac4-0d46d9dfa8a1-kube-api-access-r4vbf\") pod \"apiserver-7bbb656c7d-q9knv\" (UID: \"3638c2bd-8853-410f-bac4-0d46d9dfa8a1\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-q9knv" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.159937 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-q9knv" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.179009 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.181391 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8hk97\" (UniqueName: \"kubernetes.io/projected/9b9caf4b-0443-4cfe-b9d3-6d7093d07db7-kube-api-access-8hk97\") pod \"machine-approver-56656f9798-9lgs8\" (UID: \"9b9caf4b-0443-4cfe-b9d3-6d7093d07db7\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-9lgs8" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.197782 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-r9t2k" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.207697 4870 request.go:700] Waited for 1.916083181s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-dns/configmaps?fieldSelector=metadata.name%3Ddns-default&limit=500&resourceVersion=0 Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.211551 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.219312 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.237358 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-4wngz" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.240889 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.258986 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.279751 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-2qq5l" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.280806 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.299097 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.338685 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-9lgs8" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.338894 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.344151 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9plcl\" (UniqueName: \"kubernetes.io/projected/50732192-84e4-41fa-9c91-e5f77d85fc5e-kube-api-access-9plcl\") pod \"oauth-openshift-558db77b4-x8qtb\" (UID: \"50732192-84e4-41fa-9c91-e5f77d85fc5e\") " pod="openshift-authentication/oauth-openshift-558db77b4-x8qtb" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.346138 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-n8z6k" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.362823 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.384995 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.393062 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-4wddg"] Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.429876 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8nr6p\" (UniqueName: \"kubernetes.io/projected/c0cd54e9-09ac-49a6-a774-a9148024b4b9-kube-api-access-8nr6p\") pod \"multus-admission-controller-857f4d67dd-l24lg\" (UID: \"c0cd54e9-09ac-49a6-a774-a9148024b4b9\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-l24lg" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.442977 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-46lfv\" (UniqueName: \"kubernetes.io/projected/3f44b737-38ee-4c89-85fb-264a66ce8771-kube-api-access-46lfv\") pod \"etcd-operator-b45778765-5dsrq\" (UID: \"3f44b737-38ee-4c89-85fb-264a66ce8771\") " pod="openshift-etcd-operator/etcd-operator-b45778765-5dsrq" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.444331 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-q9knv"] Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.465111 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tk4zz\" (UniqueName: \"kubernetes.io/projected/49dc86fe-e718-4220-81d7-2ebd7fdd6c2f-kube-api-access-tk4zz\") pod \"openshift-controller-manager-operator-756b6f6bc6-55gqt\" (UID: \"49dc86fe-e718-4220-81d7-2ebd7fdd6c2f\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-55gqt" Oct 14 07:03:28 crc kubenswrapper[4870]: W1014 07:03:28.467582 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3638c2bd_8853_410f_bac4_0d46d9dfa8a1.slice/crio-465acc0f57b2f5317bc9eb4b635206d87cdaf34fc82b2560f7f8145c63616122 WatchSource:0}: Error finding container 465acc0f57b2f5317bc9eb4b635206d87cdaf34fc82b2560f7f8145c63616122: Status 404 returned error can't find the container with id 465acc0f57b2f5317bc9eb4b635206d87cdaf34fc82b2560f7f8145c63616122 Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.471333 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-r9t2k"] Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.475049 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cb6jb\" (UniqueName: \"kubernetes.io/projected/b419b072-5f0d-4103-82c6-c296f71eee8a-kube-api-access-cb6jb\") pod \"route-controller-manager-6576b87f9c-7jz27\" (UID: \"b419b072-5f0d-4103-82c6-c296f71eee8a\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-7jz27" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.494575 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-56bq2\" (UniqueName: \"kubernetes.io/projected/222f1962-d6c1-4e69-a3bc-0b0a12703cff-kube-api-access-56bq2\") pod \"dns-operator-744455d44c-zpgbg\" (UID: \"222f1962-d6c1-4e69-a3bc-0b0a12703cff\") " pod="openshift-dns-operator/dns-operator-744455d44c-zpgbg" Oct 14 07:03:28 crc kubenswrapper[4870]: W1014 07:03:28.497563 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb38b0739_21a6_4050_af6c_6accabefcc45.slice/crio-e1b6a976c48ab09e1c019e72e377ebd7c1e1da64ad1c680caf61df4052a61312 WatchSource:0}: Error finding container e1b6a976c48ab09e1c019e72e377ebd7c1e1da64ad1c680caf61df4052a61312: Status 404 returned error can't find the container with id e1b6a976c48ab09e1c019e72e377ebd7c1e1da64ad1c680caf61df4052a61312 Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.514777 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t25cc\" (UniqueName: \"kubernetes.io/projected/7b7a521b-c795-43cc-9576-5a3be3cb7409-kube-api-access-t25cc\") pod \"migrator-59844c95c7-hwfsm\" (UID: \"7b7a521b-c795-43cc-9576-5a3be3cb7409\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-hwfsm" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.519365 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-5dsrq" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.534769 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jh6v9\" (UniqueName: \"kubernetes.io/projected/88363344-bc9a-4050-87e2-cf6f07708835-kube-api-access-jh6v9\") pod \"openshift-apiserver-operator-796bbdcf4f-cgwzv\" (UID: \"88363344-bc9a-4050-87e2-cf6f07708835\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-cgwzv" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.547363 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-4wngz"] Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.548661 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-hwfsm" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.557152 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-x8qtb" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.559554 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tl684\" (UniqueName: \"kubernetes.io/projected/8070623a-4696-490e-af90-5b2155c0506c-kube-api-access-tl684\") pod \"cluster-samples-operator-665b6dd947-szwj9\" (UID: \"8070623a-4696-490e-af90-5b2155c0506c\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-szwj9" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.572424 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-l24lg" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.576351 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hdxrc\" (UniqueName: \"kubernetes.io/projected/337df1ca-c52e-499c-ab1d-a32aee9c9922-kube-api-access-hdxrc\") pod \"authentication-operator-69f744f599-cvh5p\" (UID: \"337df1ca-c52e-499c-ab1d-a32aee9c9922\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-cvh5p" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.595302 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-2qq5l"] Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.596395 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2tjr6\" (UniqueName: \"kubernetes.io/projected/146160a7-2b1c-43f4-bc42-7b92d9663dcc-kube-api-access-2tjr6\") pod \"openshift-config-operator-7777fb866f-p769c\" (UID: \"146160a7-2b1c-43f4-bc42-7b92d9663dcc\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-p769c" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.614372 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-n8z6k"] Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.619192 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/16295052-1474-4dc9-b1d6-10983e6bfc6f-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-g7crn\" (UID: \"16295052-1474-4dc9-b1d6-10983e6bfc6f\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-g7crn" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.638967 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gb2b5\" (UniqueName: \"kubernetes.io/projected/f07c76fa-f754-4175-8bc5-2aae95985b3b-kube-api-access-gb2b5\") pod \"olm-operator-6b444d44fb-wjm9z\" (UID: \"f07c76fa-f754-4175-8bc5-2aae95985b3b\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wjm9z" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.662033 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d9cb408e-a00b-4f76-b8fb-bf144bfee740-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-ss574\" (UID: \"d9cb408e-a00b-4f76-b8fb-bf144bfee740\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ss574" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.665827 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-cgwzv" Oct 14 07:03:28 crc kubenswrapper[4870]: W1014 07:03:28.666591 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod03d1e16b_e364_4902_9412_0dc138fc76e3.slice/crio-0fcc7060db62ce67d318635d3dcd8d4731c01fbcd332a66740c946b2cf7ed236 WatchSource:0}: Error finding container 0fcc7060db62ce67d318635d3dcd8d4731c01fbcd332a66740c946b2cf7ed236: Status 404 returned error can't find the container with id 0fcc7060db62ce67d318635d3dcd8d4731c01fbcd332a66740c946b2cf7ed236 Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.675675 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-zpgbg" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.687422 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-p769c" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.734320 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-cvh5p" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.737644 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/28a417a5-b0b0-465a-ade0-3bf3ceb5ebdd-stats-auth\") pod \"router-default-5444994796-h82f7\" (UID: \"28a417a5-b0b0-465a-ade0-3bf3ceb5ebdd\") " pod="openshift-ingress/router-default-5444994796-h82f7" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.737689 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/be735c57-60b4-44cb-abf3-888cf00d6c31-trusted-ca\") pod \"ingress-operator-5b745b69d9-6z4z4\" (UID: \"be735c57-60b4-44cb-abf3-888cf00d6c31\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-6z4z4" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.737711 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/be735c57-60b4-44cb-abf3-888cf00d6c31-metrics-tls\") pod \"ingress-operator-5b745b69d9-6z4z4\" (UID: \"be735c57-60b4-44cb-abf3-888cf00d6c31\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-6z4z4" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.737730 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rz2pg\" (UniqueName: \"kubernetes.io/projected/a3280409-7095-4407-aa58-04be41daabf7-kube-api-access-rz2pg\") pod \"kube-storage-version-migrator-operator-b67b599dd-lbq8g\" (UID: \"a3280409-7095-4407-aa58-04be41daabf7\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-lbq8g" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.737752 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3-bound-sa-token\") pod \"image-registry-697d97f7c8-kkxjz\" (UID: \"2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3\") " pod="openshift-image-registry/image-registry-697d97f7c8-kkxjz" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.737772 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/7562fe32-ac93-483f-b017-3eab4d7f94d4-proxy-tls\") pod \"machine-config-controller-84d6567774-mlqqj\" (UID: \"7562fe32-ac93-483f-b017-3eab4d7f94d4\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-mlqqj" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.737800 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h8l2l\" (UniqueName: \"kubernetes.io/projected/7562fe32-ac93-483f-b017-3eab4d7f94d4-kube-api-access-h8l2l\") pod \"machine-config-controller-84d6567774-mlqqj\" (UID: \"7562fe32-ac93-483f-b017-3eab4d7f94d4\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-mlqqj" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.737840 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5848d6ea-90e8-4c07-92e0-8849b5b32c67-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-vc8ks\" (UID: \"5848d6ea-90e8-4c07-92e0-8849b5b32c67\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vc8ks" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.737865 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a3280409-7095-4407-aa58-04be41daabf7-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-lbq8g\" (UID: \"a3280409-7095-4407-aa58-04be41daabf7\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-lbq8g" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.737904 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m77k7\" (UniqueName: \"kubernetes.io/projected/28a417a5-b0b0-465a-ade0-3bf3ceb5ebdd-kube-api-access-m77k7\") pod \"router-default-5444994796-h82f7\" (UID: \"28a417a5-b0b0-465a-ade0-3bf3ceb5ebdd\") " pod="openshift-ingress/router-default-5444994796-h82f7" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.737922 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5848d6ea-90e8-4c07-92e0-8849b5b32c67-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-vc8ks\" (UID: \"5848d6ea-90e8-4c07-92e0-8849b5b32c67\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vc8ks" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.737948 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3-registry-tls\") pod \"image-registry-697d97f7c8-kkxjz\" (UID: \"2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3\") " pod="openshift-image-registry/image-registry-697d97f7c8-kkxjz" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.737992 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-24b47\" (UniqueName: \"kubernetes.io/projected/2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3-kube-api-access-24b47\") pod \"image-registry-697d97f7c8-kkxjz\" (UID: \"2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3\") " pod="openshift-image-registry/image-registry-697d97f7c8-kkxjz" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.738019 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3-trusted-ca\") pod \"image-registry-697d97f7c8-kkxjz\" (UID: \"2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3\") " pod="openshift-image-registry/image-registry-697d97f7c8-kkxjz" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.738034 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5848d6ea-90e8-4c07-92e0-8849b5b32c67-config\") pod \"kube-apiserver-operator-766d6c64bb-vc8ks\" (UID: \"5848d6ea-90e8-4c07-92e0-8849b5b32c67\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vc8ks" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.738056 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s2st6\" (UniqueName: \"kubernetes.io/projected/2a72b0a0-287b-42dc-8c39-1976188955de-kube-api-access-s2st6\") pod \"downloads-7954f5f757-d7cdf\" (UID: \"2a72b0a0-287b-42dc-8c39-1976188955de\") " pod="openshift-console/downloads-7954f5f757-d7cdf" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.738101 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/a0207141-7032-4930-a837-4736469c0974-profile-collector-cert\") pod \"catalog-operator-68c6474976-kqxpq\" (UID: \"a0207141-7032-4930-a837-4736469c0974\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-kqxpq" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.738131 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dpm2h\" (UniqueName: \"kubernetes.io/projected/a0207141-7032-4930-a837-4736469c0974-kube-api-access-dpm2h\") pod \"catalog-operator-68c6474976-kqxpq\" (UID: \"a0207141-7032-4930-a837-4736469c0974\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-kqxpq" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.738155 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a3280409-7095-4407-aa58-04be41daabf7-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-lbq8g\" (UID: \"a3280409-7095-4407-aa58-04be41daabf7\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-lbq8g" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.738187 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kkxjz\" (UID: \"2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3\") " pod="openshift-image-registry/image-registry-697d97f7c8-kkxjz" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.738212 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3-ca-trust-extracted\") pod \"image-registry-697d97f7c8-kkxjz\" (UID: \"2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3\") " pod="openshift-image-registry/image-registry-697d97f7c8-kkxjz" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.738231 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/28a417a5-b0b0-465a-ade0-3bf3ceb5ebdd-default-certificate\") pod \"router-default-5444994796-h82f7\" (UID: \"28a417a5-b0b0-465a-ade0-3bf3ceb5ebdd\") " pod="openshift-ingress/router-default-5444994796-h82f7" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.738269 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/28a417a5-b0b0-465a-ade0-3bf3ceb5ebdd-service-ca-bundle\") pod \"router-default-5444994796-h82f7\" (UID: \"28a417a5-b0b0-465a-ade0-3bf3ceb5ebdd\") " pod="openshift-ingress/router-default-5444994796-h82f7" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.738306 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3-registry-certificates\") pod \"image-registry-697d97f7c8-kkxjz\" (UID: \"2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3\") " pod="openshift-image-registry/image-registry-697d97f7c8-kkxjz" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.738322 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8xg8z\" (UniqueName: \"kubernetes.io/projected/be735c57-60b4-44cb-abf3-888cf00d6c31-kube-api-access-8xg8z\") pod \"ingress-operator-5b745b69d9-6z4z4\" (UID: \"be735c57-60b4-44cb-abf3-888cf00d6c31\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-6z4z4" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.738340 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/7562fe32-ac93-483f-b017-3eab4d7f94d4-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-mlqqj\" (UID: \"7562fe32-ac93-483f-b017-3eab4d7f94d4\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-mlqqj" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.738366 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/be735c57-60b4-44cb-abf3-888cf00d6c31-bound-sa-token\") pod \"ingress-operator-5b745b69d9-6z4z4\" (UID: \"be735c57-60b4-44cb-abf3-888cf00d6c31\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-6z4z4" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.738426 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3-installation-pull-secrets\") pod \"image-registry-697d97f7c8-kkxjz\" (UID: \"2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3\") " pod="openshift-image-registry/image-registry-697d97f7c8-kkxjz" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.738459 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/28a417a5-b0b0-465a-ade0-3bf3ceb5ebdd-metrics-certs\") pod \"router-default-5444994796-h82f7\" (UID: \"28a417a5-b0b0-465a-ade0-3bf3ceb5ebdd\") " pod="openshift-ingress/router-default-5444994796-h82f7" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.738488 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/a0207141-7032-4930-a837-4736469c0974-srv-cert\") pod \"catalog-operator-68c6474976-kqxpq\" (UID: \"a0207141-7032-4930-a837-4736469c0974\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-kqxpq" Oct 14 07:03:28 crc kubenswrapper[4870]: E1014 07:03:28.741985 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-14 07:03:29.241967508 +0000 UTC m=+144.939328069 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kkxjz" (UID: "2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.743428 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-55gqt" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.759519 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-7jz27" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.767224 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-szwj9" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.788749 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-g7crn" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.806790 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ss574" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.839841 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.840032 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/be735c57-60b4-44cb-abf3-888cf00d6c31-metrics-tls\") pod \"ingress-operator-5b745b69d9-6z4z4\" (UID: \"be735c57-60b4-44cb-abf3-888cf00d6c31\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-6z4z4" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.840057 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rz2pg\" (UniqueName: \"kubernetes.io/projected/a3280409-7095-4407-aa58-04be41daabf7-kube-api-access-rz2pg\") pod \"kube-storage-version-migrator-operator-b67b599dd-lbq8g\" (UID: \"a3280409-7095-4407-aa58-04be41daabf7\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-lbq8g" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.840086 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q84wc\" (UniqueName: \"kubernetes.io/projected/9057841b-b3a4-49f2-85a6-812074c635e6-kube-api-access-q84wc\") pod \"dns-default-d4fhp\" (UID: \"9057841b-b3a4-49f2-85a6-812074c635e6\") " pod="openshift-dns/dns-default-d4fhp" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.840135 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3-bound-sa-token\") pod \"image-registry-697d97f7c8-kkxjz\" (UID: \"2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3\") " pod="openshift-image-registry/image-registry-697d97f7c8-kkxjz" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.840158 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/7562fe32-ac93-483f-b017-3eab4d7f94d4-proxy-tls\") pod \"machine-config-controller-84d6567774-mlqqj\" (UID: \"7562fe32-ac93-483f-b017-3eab4d7f94d4\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-mlqqj" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.840177 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/9057841b-b3a4-49f2-85a6-812074c635e6-metrics-tls\") pod \"dns-default-d4fhp\" (UID: \"9057841b-b3a4-49f2-85a6-812074c635e6\") " pod="openshift-dns/dns-default-d4fhp" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.840196 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/1d891345-93bd-4726-905f-8a946a4a885f-apiservice-cert\") pod \"packageserver-d55dfcdfc-rnf2p\" (UID: \"1d891345-93bd-4726-905f-8a946a4a885f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rnf2p" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.840227 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v2khw\" (UniqueName: \"kubernetes.io/projected/1d891345-93bd-4726-905f-8a946a4a885f-kube-api-access-v2khw\") pod \"packageserver-d55dfcdfc-rnf2p\" (UID: \"1d891345-93bd-4726-905f-8a946a4a885f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rnf2p" Oct 14 07:03:28 crc kubenswrapper[4870]: E1014 07:03:28.840502 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 07:03:29.340465897 +0000 UTC m=+145.037826378 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.842899 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h8l2l\" (UniqueName: \"kubernetes.io/projected/7562fe32-ac93-483f-b017-3eab4d7f94d4-kube-api-access-h8l2l\") pod \"machine-config-controller-84d6567774-mlqqj\" (UID: \"7562fe32-ac93-483f-b017-3eab4d7f94d4\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-mlqqj" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.842955 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q92sd\" (UniqueName: \"kubernetes.io/projected/0a83e6b8-c65a-42ef-b4e3-8748503bcd84-kube-api-access-q92sd\") pod \"service-ca-operator-777779d784-wkc9x\" (UID: \"0a83e6b8-c65a-42ef-b4e3-8748503bcd84\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-wkc9x" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.842989 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/edd471fb-addd-407d-b7cb-55dbffd7a1ee-mountpoint-dir\") pod \"csi-hostpathplugin-ptjzm\" (UID: \"edd471fb-addd-407d-b7cb-55dbffd7a1ee\") " pod="hostpath-provisioner/csi-hostpathplugin-ptjzm" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.843016 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5848d6ea-90e8-4c07-92e0-8849b5b32c67-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-vc8ks\" (UID: \"5848d6ea-90e8-4c07-92e0-8849b5b32c67\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vc8ks" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.843063 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hfvbs\" (UniqueName: \"kubernetes.io/projected/63176391-0acc-4316-a0fe-520669eb1744-kube-api-access-hfvbs\") pod \"package-server-manager-789f6589d5-52qbc\" (UID: \"63176391-0acc-4316-a0fe-520669eb1744\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-52qbc" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.843102 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a3280409-7095-4407-aa58-04be41daabf7-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-lbq8g\" (UID: \"a3280409-7095-4407-aa58-04be41daabf7\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-lbq8g" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.843131 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/edd471fb-addd-407d-b7cb-55dbffd7a1ee-registration-dir\") pod \"csi-hostpathplugin-ptjzm\" (UID: \"edd471fb-addd-407d-b7cb-55dbffd7a1ee\") " pod="hostpath-provisioner/csi-hostpathplugin-ptjzm" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.843147 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hrhqh\" (UniqueName: \"kubernetes.io/projected/1df54538-19c2-472b-b93e-b7f947190db0-kube-api-access-hrhqh\") pod \"machine-config-server-klh8n\" (UID: \"1df54538-19c2-472b-b93e-b7f947190db0\") " pod="openshift-machine-config-operator/machine-config-server-klh8n" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.843188 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m77k7\" (UniqueName: \"kubernetes.io/projected/28a417a5-b0b0-465a-ade0-3bf3ceb5ebdd-kube-api-access-m77k7\") pod \"router-default-5444994796-h82f7\" (UID: \"28a417a5-b0b0-465a-ade0-3bf3ceb5ebdd\") " pod="openshift-ingress/router-default-5444994796-h82f7" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.843205 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/1df54538-19c2-472b-b93e-b7f947190db0-node-bootstrap-token\") pod \"machine-config-server-klh8n\" (UID: \"1df54538-19c2-472b-b93e-b7f947190db0\") " pod="openshift-machine-config-operator/machine-config-server-klh8n" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.843231 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/8432ae60-6e63-45cf-b21b-ab1010687654-signing-key\") pod \"service-ca-9c57cc56f-5ptfj\" (UID: \"8432ae60-6e63-45cf-b21b-ab1010687654\") " pod="openshift-service-ca/service-ca-9c57cc56f-5ptfj" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.843268 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5848d6ea-90e8-4c07-92e0-8849b5b32c67-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-vc8ks\" (UID: \"5848d6ea-90e8-4c07-92e0-8849b5b32c67\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vc8ks" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.843298 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/27bd2071-dddb-48c6-905c-22bcb4e84482-auth-proxy-config\") pod \"machine-config-operator-74547568cd-69t6j\" (UID: \"27bd2071-dddb-48c6-905c-22bcb4e84482\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-69t6j" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.843334 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/737f304d-12ac-4d4b-97fb-e99506dd08a8-config-volume\") pod \"collect-profiles-29340420-t6cgj\" (UID: \"737f304d-12ac-4d4b-97fb-e99506dd08a8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340420-t6cgj" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.843377 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/8f610aec-a4a3-4d54-b805-4f01c0de5dce-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-6vdcc\" (UID: \"8f610aec-a4a3-4d54-b805-4f01c0de5dce\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-6vdcc" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.843397 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/737f304d-12ac-4d4b-97fb-e99506dd08a8-secret-volume\") pod \"collect-profiles-29340420-t6cgj\" (UID: \"737f304d-12ac-4d4b-97fb-e99506dd08a8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340420-t6cgj" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.843412 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0a83e6b8-c65a-42ef-b4e3-8748503bcd84-serving-cert\") pod \"service-ca-operator-777779d784-wkc9x\" (UID: \"0a83e6b8-c65a-42ef-b4e3-8748503bcd84\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-wkc9x" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.843430 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/1d891345-93bd-4726-905f-8a946a4a885f-webhook-cert\") pod \"packageserver-d55dfcdfc-rnf2p\" (UID: \"1d891345-93bd-4726-905f-8a946a4a885f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rnf2p" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.843466 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3-registry-tls\") pod \"image-registry-697d97f7c8-kkxjz\" (UID: \"2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3\") " pod="openshift-image-registry/image-registry-697d97f7c8-kkxjz" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.843494 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/1d891345-93bd-4726-905f-8a946a4a885f-tmpfs\") pod \"packageserver-d55dfcdfc-rnf2p\" (UID: \"1d891345-93bd-4726-905f-8a946a4a885f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rnf2p" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.843521 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0a83e6b8-c65a-42ef-b4e3-8748503bcd84-config\") pod \"service-ca-operator-777779d784-wkc9x\" (UID: \"0a83e6b8-c65a-42ef-b4e3-8748503bcd84\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-wkc9x" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.843556 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-24b47\" (UniqueName: \"kubernetes.io/projected/2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3-kube-api-access-24b47\") pod \"image-registry-697d97f7c8-kkxjz\" (UID: \"2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3\") " pod="openshift-image-registry/image-registry-697d97f7c8-kkxjz" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.843572 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/bb92d614-7f4e-470e-a68d-0aff68d3894d-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-6ztwm\" (UID: \"bb92d614-7f4e-470e-a68d-0aff68d3894d\") " pod="openshift-marketplace/marketplace-operator-79b997595-6ztwm" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.843602 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3-trusted-ca\") pod \"image-registry-697d97f7c8-kkxjz\" (UID: \"2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3\") " pod="openshift-image-registry/image-registry-697d97f7c8-kkxjz" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.843618 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5848d6ea-90e8-4c07-92e0-8849b5b32c67-config\") pod \"kube-apiserver-operator-766d6c64bb-vc8ks\" (UID: \"5848d6ea-90e8-4c07-92e0-8849b5b32c67\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vc8ks" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.843633 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2st6\" (UniqueName: \"kubernetes.io/projected/2a72b0a0-287b-42dc-8c39-1976188955de-kube-api-access-s2st6\") pod \"downloads-7954f5f757-d7cdf\" (UID: \"2a72b0a0-287b-42dc-8c39-1976188955de\") " pod="openshift-console/downloads-7954f5f757-d7cdf" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.843652 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/a0207141-7032-4930-a837-4736469c0974-profile-collector-cert\") pod \"catalog-operator-68c6474976-kqxpq\" (UID: \"a0207141-7032-4930-a837-4736469c0974\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-kqxpq" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.843667 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/30ec1e91-85db-4abb-b3dc-19cdcc8d4234-cert\") pod \"ingress-canary-q26kj\" (UID: \"30ec1e91-85db-4abb-b3dc-19cdcc8d4234\") " pod="openshift-ingress-canary/ingress-canary-q26kj" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.843713 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dpm2h\" (UniqueName: \"kubernetes.io/projected/a0207141-7032-4930-a837-4736469c0974-kube-api-access-dpm2h\") pod \"catalog-operator-68c6474976-kqxpq\" (UID: \"a0207141-7032-4930-a837-4736469c0974\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-kqxpq" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.843729 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jhf85\" (UniqueName: \"kubernetes.io/projected/edd471fb-addd-407d-b7cb-55dbffd7a1ee-kube-api-access-jhf85\") pod \"csi-hostpathplugin-ptjzm\" (UID: \"edd471fb-addd-407d-b7cb-55dbffd7a1ee\") " pod="hostpath-provisioner/csi-hostpathplugin-ptjzm" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.843744 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zxf4h\" (UniqueName: \"kubernetes.io/projected/8432ae60-6e63-45cf-b21b-ab1010687654-kube-api-access-zxf4h\") pod \"service-ca-9c57cc56f-5ptfj\" (UID: \"8432ae60-6e63-45cf-b21b-ab1010687654\") " pod="openshift-service-ca/service-ca-9c57cc56f-5ptfj" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.843838 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a3280409-7095-4407-aa58-04be41daabf7-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-lbq8g\" (UID: \"a3280409-7095-4407-aa58-04be41daabf7\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-lbq8g" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.843855 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/63176391-0acc-4316-a0fe-520669eb1744-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-52qbc\" (UID: \"63176391-0acc-4316-a0fe-520669eb1744\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-52qbc" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.843881 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kkxjz\" (UID: \"2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3\") " pod="openshift-image-registry/image-registry-697d97f7c8-kkxjz" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.843898 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3-ca-trust-extracted\") pod \"image-registry-697d97f7c8-kkxjz\" (UID: \"2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3\") " pod="openshift-image-registry/image-registry-697d97f7c8-kkxjz" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.843932 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/28a417a5-b0b0-465a-ade0-3bf3ceb5ebdd-default-certificate\") pod \"router-default-5444994796-h82f7\" (UID: \"28a417a5-b0b0-465a-ade0-3bf3ceb5ebdd\") " pod="openshift-ingress/router-default-5444994796-h82f7" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.843962 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/28a417a5-b0b0-465a-ade0-3bf3ceb5ebdd-service-ca-bundle\") pod \"router-default-5444994796-h82f7\" (UID: \"28a417a5-b0b0-465a-ade0-3bf3ceb5ebdd\") " pod="openshift-ingress/router-default-5444994796-h82f7" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.843980 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j7bm9\" (UniqueName: \"kubernetes.io/projected/8f610aec-a4a3-4d54-b805-4f01c0de5dce-kube-api-access-j7bm9\") pod \"control-plane-machine-set-operator-78cbb6b69f-6vdcc\" (UID: \"8f610aec-a4a3-4d54-b805-4f01c0de5dce\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-6vdcc" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.844011 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bb92d614-7f4e-470e-a68d-0aff68d3894d-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-6ztwm\" (UID: \"bb92d614-7f4e-470e-a68d-0aff68d3894d\") " pod="openshift-marketplace/marketplace-operator-79b997595-6ztwm" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.844028 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/8432ae60-6e63-45cf-b21b-ab1010687654-signing-cabundle\") pod \"service-ca-9c57cc56f-5ptfj\" (UID: \"8432ae60-6e63-45cf-b21b-ab1010687654\") " pod="openshift-service-ca/service-ca-9c57cc56f-5ptfj" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.844048 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3-registry-certificates\") pod \"image-registry-697d97f7c8-kkxjz\" (UID: \"2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3\") " pod="openshift-image-registry/image-registry-697d97f7c8-kkxjz" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.844066 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8xg8z\" (UniqueName: \"kubernetes.io/projected/be735c57-60b4-44cb-abf3-888cf00d6c31-kube-api-access-8xg8z\") pod \"ingress-operator-5b745b69d9-6z4z4\" (UID: \"be735c57-60b4-44cb-abf3-888cf00d6c31\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-6z4z4" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.844115 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/7562fe32-ac93-483f-b017-3eab4d7f94d4-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-mlqqj\" (UID: \"7562fe32-ac93-483f-b017-3eab4d7f94d4\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-mlqqj" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.844136 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/be735c57-60b4-44cb-abf3-888cf00d6c31-bound-sa-token\") pod \"ingress-operator-5b745b69d9-6z4z4\" (UID: \"be735c57-60b4-44cb-abf3-888cf00d6c31\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-6z4z4" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.845073 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9057841b-b3a4-49f2-85a6-812074c635e6-config-volume\") pod \"dns-default-d4fhp\" (UID: \"9057841b-b3a4-49f2-85a6-812074c635e6\") " pod="openshift-dns/dns-default-d4fhp" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.845117 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6m4zp\" (UniqueName: \"kubernetes.io/projected/737f304d-12ac-4d4b-97fb-e99506dd08a8-kube-api-access-6m4zp\") pod \"collect-profiles-29340420-t6cgj\" (UID: \"737f304d-12ac-4d4b-97fb-e99506dd08a8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340420-t6cgj" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.845302 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3-installation-pull-secrets\") pod \"image-registry-697d97f7c8-kkxjz\" (UID: \"2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3\") " pod="openshift-image-registry/image-registry-697d97f7c8-kkxjz" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.845324 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/28a417a5-b0b0-465a-ade0-3bf3ceb5ebdd-metrics-certs\") pod \"router-default-5444994796-h82f7\" (UID: \"28a417a5-b0b0-465a-ade0-3bf3ceb5ebdd\") " pod="openshift-ingress/router-default-5444994796-h82f7" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.845346 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/27bd2071-dddb-48c6-905c-22bcb4e84482-proxy-tls\") pod \"machine-config-operator-74547568cd-69t6j\" (UID: \"27bd2071-dddb-48c6-905c-22bcb4e84482\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-69t6j" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.845403 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/a0207141-7032-4930-a837-4736469c0974-srv-cert\") pod \"catalog-operator-68c6474976-kqxpq\" (UID: \"a0207141-7032-4930-a837-4736469c0974\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-kqxpq" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.845422 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c7cj6\" (UniqueName: \"kubernetes.io/projected/30ec1e91-85db-4abb-b3dc-19cdcc8d4234-kube-api-access-c7cj6\") pod \"ingress-canary-q26kj\" (UID: \"30ec1e91-85db-4abb-b3dc-19cdcc8d4234\") " pod="openshift-ingress-canary/ingress-canary-q26kj" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.845460 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7np8q\" (UniqueName: \"kubernetes.io/projected/bb92d614-7f4e-470e-a68d-0aff68d3894d-kube-api-access-7np8q\") pod \"marketplace-operator-79b997595-6ztwm\" (UID: \"bb92d614-7f4e-470e-a68d-0aff68d3894d\") " pod="openshift-marketplace/marketplace-operator-79b997595-6ztwm" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.845533 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h2ptt\" (UniqueName: \"kubernetes.io/projected/27bd2071-dddb-48c6-905c-22bcb4e84482-kube-api-access-h2ptt\") pod \"machine-config-operator-74547568cd-69t6j\" (UID: \"27bd2071-dddb-48c6-905c-22bcb4e84482\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-69t6j" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.845562 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/27bd2071-dddb-48c6-905c-22bcb4e84482-images\") pod \"machine-config-operator-74547568cd-69t6j\" (UID: \"27bd2071-dddb-48c6-905c-22bcb4e84482\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-69t6j" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.845613 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/edd471fb-addd-407d-b7cb-55dbffd7a1ee-plugins-dir\") pod \"csi-hostpathplugin-ptjzm\" (UID: \"edd471fb-addd-407d-b7cb-55dbffd7a1ee\") " pod="hostpath-provisioner/csi-hostpathplugin-ptjzm" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.845904 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/28a417a5-b0b0-465a-ade0-3bf3ceb5ebdd-stats-auth\") pod \"router-default-5444994796-h82f7\" (UID: \"28a417a5-b0b0-465a-ade0-3bf3ceb5ebdd\") " pod="openshift-ingress/router-default-5444994796-h82f7" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.845946 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/edd471fb-addd-407d-b7cb-55dbffd7a1ee-socket-dir\") pod \"csi-hostpathplugin-ptjzm\" (UID: \"edd471fb-addd-407d-b7cb-55dbffd7a1ee\") " pod="hostpath-provisioner/csi-hostpathplugin-ptjzm" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.845973 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/edd471fb-addd-407d-b7cb-55dbffd7a1ee-csi-data-dir\") pod \"csi-hostpathplugin-ptjzm\" (UID: \"edd471fb-addd-407d-b7cb-55dbffd7a1ee\") " pod="hostpath-provisioner/csi-hostpathplugin-ptjzm" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.846030 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/be735c57-60b4-44cb-abf3-888cf00d6c31-trusted-ca\") pod \"ingress-operator-5b745b69d9-6z4z4\" (UID: \"be735c57-60b4-44cb-abf3-888cf00d6c31\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-6z4z4" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.846053 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/1df54538-19c2-472b-b93e-b7f947190db0-certs\") pod \"machine-config-server-klh8n\" (UID: \"1df54538-19c2-472b-b93e-b7f947190db0\") " pod="openshift-machine-config-operator/machine-config-server-klh8n" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.846092 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/28a417a5-b0b0-465a-ade0-3bf3ceb5ebdd-service-ca-bundle\") pod \"router-default-5444994796-h82f7\" (UID: \"28a417a5-b0b0-465a-ade0-3bf3ceb5ebdd\") " pod="openshift-ingress/router-default-5444994796-h82f7" Oct 14 07:03:28 crc kubenswrapper[4870]: E1014 07:03:28.846661 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-14 07:03:29.346636867 +0000 UTC m=+145.043997438 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kkxjz" (UID: "2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.849178 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/7562fe32-ac93-483f-b017-3eab4d7f94d4-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-mlqqj\" (UID: \"7562fe32-ac93-483f-b017-3eab4d7f94d4\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-mlqqj" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.853566 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3-trusted-ca\") pod \"image-registry-697d97f7c8-kkxjz\" (UID: \"2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3\") " pod="openshift-image-registry/image-registry-697d97f7c8-kkxjz" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.854164 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5848d6ea-90e8-4c07-92e0-8849b5b32c67-config\") pod \"kube-apiserver-operator-766d6c64bb-vc8ks\" (UID: \"5848d6ea-90e8-4c07-92e0-8849b5b32c67\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vc8ks" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.854533 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a3280409-7095-4407-aa58-04be41daabf7-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-lbq8g\" (UID: \"a3280409-7095-4407-aa58-04be41daabf7\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-lbq8g" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.855010 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/28a417a5-b0b0-465a-ade0-3bf3ceb5ebdd-metrics-certs\") pod \"router-default-5444994796-h82f7\" (UID: \"28a417a5-b0b0-465a-ade0-3bf3ceb5ebdd\") " pod="openshift-ingress/router-default-5444994796-h82f7" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.855746 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a3280409-7095-4407-aa58-04be41daabf7-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-lbq8g\" (UID: \"a3280409-7095-4407-aa58-04be41daabf7\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-lbq8g" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.855897 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/28a417a5-b0b0-465a-ade0-3bf3ceb5ebdd-default-certificate\") pod \"router-default-5444994796-h82f7\" (UID: \"28a417a5-b0b0-465a-ade0-3bf3ceb5ebdd\") " pod="openshift-ingress/router-default-5444994796-h82f7" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.857390 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/be735c57-60b4-44cb-abf3-888cf00d6c31-trusted-ca\") pod \"ingress-operator-5b745b69d9-6z4z4\" (UID: \"be735c57-60b4-44cb-abf3-888cf00d6c31\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-6z4z4" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.858116 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3-ca-trust-extracted\") pod \"image-registry-697d97f7c8-kkxjz\" (UID: \"2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3\") " pod="openshift-image-registry/image-registry-697d97f7c8-kkxjz" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.858355 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3-registry-certificates\") pod \"image-registry-697d97f7c8-kkxjz\" (UID: \"2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3\") " pod="openshift-image-registry/image-registry-697d97f7c8-kkxjz" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.858851 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wjm9z" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.861895 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3-registry-tls\") pod \"image-registry-697d97f7c8-kkxjz\" (UID: \"2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3\") " pod="openshift-image-registry/image-registry-697d97f7c8-kkxjz" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.870297 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/28a417a5-b0b0-465a-ade0-3bf3ceb5ebdd-stats-auth\") pod \"router-default-5444994796-h82f7\" (UID: \"28a417a5-b0b0-465a-ade0-3bf3ceb5ebdd\") " pod="openshift-ingress/router-default-5444994796-h82f7" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.871061 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/be735c57-60b4-44cb-abf3-888cf00d6c31-metrics-tls\") pod \"ingress-operator-5b745b69d9-6z4z4\" (UID: \"be735c57-60b4-44cb-abf3-888cf00d6c31\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-6z4z4" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.871363 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/a0207141-7032-4930-a837-4736469c0974-profile-collector-cert\") pod \"catalog-operator-68c6474976-kqxpq\" (UID: \"a0207141-7032-4930-a837-4736469c0974\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-kqxpq" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.871947 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3-installation-pull-secrets\") pod \"image-registry-697d97f7c8-kkxjz\" (UID: \"2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3\") " pod="openshift-image-registry/image-registry-697d97f7c8-kkxjz" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.872180 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5848d6ea-90e8-4c07-92e0-8849b5b32c67-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-vc8ks\" (UID: \"5848d6ea-90e8-4c07-92e0-8849b5b32c67\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vc8ks" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.877235 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/7562fe32-ac93-483f-b017-3eab4d7f94d4-proxy-tls\") pod \"machine-config-controller-84d6567774-mlqqj\" (UID: \"7562fe32-ac93-483f-b017-3eab4d7f94d4\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-mlqqj" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.880084 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rz2pg\" (UniqueName: \"kubernetes.io/projected/a3280409-7095-4407-aa58-04be41daabf7-kube-api-access-rz2pg\") pod \"kube-storage-version-migrator-operator-b67b599dd-lbq8g\" (UID: \"a3280409-7095-4407-aa58-04be41daabf7\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-lbq8g" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.891864 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/a0207141-7032-4930-a837-4736469c0974-srv-cert\") pod \"catalog-operator-68c6474976-kqxpq\" (UID: \"a0207141-7032-4930-a837-4736469c0974\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-kqxpq" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.910196 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-x8qtb"] Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.920829 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3-bound-sa-token\") pod \"image-registry-697d97f7c8-kkxjz\" (UID: \"2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3\") " pod="openshift-image-registry/image-registry-697d97f7c8-kkxjz" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.923108 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5848d6ea-90e8-4c07-92e0-8849b5b32c67-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-vc8ks\" (UID: \"5848d6ea-90e8-4c07-92e0-8849b5b32c67\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vc8ks" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.949939 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.950180 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/27bd2071-dddb-48c6-905c-22bcb4e84482-proxy-tls\") pod \"machine-config-operator-74547568cd-69t6j\" (UID: \"27bd2071-dddb-48c6-905c-22bcb4e84482\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-69t6j" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.950211 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c7cj6\" (UniqueName: \"kubernetes.io/projected/30ec1e91-85db-4abb-b3dc-19cdcc8d4234-kube-api-access-c7cj6\") pod \"ingress-canary-q26kj\" (UID: \"30ec1e91-85db-4abb-b3dc-19cdcc8d4234\") " pod="openshift-ingress-canary/ingress-canary-q26kj" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.950229 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7np8q\" (UniqueName: \"kubernetes.io/projected/bb92d614-7f4e-470e-a68d-0aff68d3894d-kube-api-access-7np8q\") pod \"marketplace-operator-79b997595-6ztwm\" (UID: \"bb92d614-7f4e-470e-a68d-0aff68d3894d\") " pod="openshift-marketplace/marketplace-operator-79b997595-6ztwm" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.950261 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h2ptt\" (UniqueName: \"kubernetes.io/projected/27bd2071-dddb-48c6-905c-22bcb4e84482-kube-api-access-h2ptt\") pod \"machine-config-operator-74547568cd-69t6j\" (UID: \"27bd2071-dddb-48c6-905c-22bcb4e84482\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-69t6j" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.950286 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/27bd2071-dddb-48c6-905c-22bcb4e84482-images\") pod \"machine-config-operator-74547568cd-69t6j\" (UID: \"27bd2071-dddb-48c6-905c-22bcb4e84482\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-69t6j" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.950305 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/edd471fb-addd-407d-b7cb-55dbffd7a1ee-plugins-dir\") pod \"csi-hostpathplugin-ptjzm\" (UID: \"edd471fb-addd-407d-b7cb-55dbffd7a1ee\") " pod="hostpath-provisioner/csi-hostpathplugin-ptjzm" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.950323 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/edd471fb-addd-407d-b7cb-55dbffd7a1ee-socket-dir\") pod \"csi-hostpathplugin-ptjzm\" (UID: \"edd471fb-addd-407d-b7cb-55dbffd7a1ee\") " pod="hostpath-provisioner/csi-hostpathplugin-ptjzm" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.950338 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/edd471fb-addd-407d-b7cb-55dbffd7a1ee-csi-data-dir\") pod \"csi-hostpathplugin-ptjzm\" (UID: \"edd471fb-addd-407d-b7cb-55dbffd7a1ee\") " pod="hostpath-provisioner/csi-hostpathplugin-ptjzm" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.950354 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/1df54538-19c2-472b-b93e-b7f947190db0-certs\") pod \"machine-config-server-klh8n\" (UID: \"1df54538-19c2-472b-b93e-b7f947190db0\") " pod="openshift-machine-config-operator/machine-config-server-klh8n" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.950375 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q84wc\" (UniqueName: \"kubernetes.io/projected/9057841b-b3a4-49f2-85a6-812074c635e6-kube-api-access-q84wc\") pod \"dns-default-d4fhp\" (UID: \"9057841b-b3a4-49f2-85a6-812074c635e6\") " pod="openshift-dns/dns-default-d4fhp" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.950401 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/9057841b-b3a4-49f2-85a6-812074c635e6-metrics-tls\") pod \"dns-default-d4fhp\" (UID: \"9057841b-b3a4-49f2-85a6-812074c635e6\") " pod="openshift-dns/dns-default-d4fhp" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.950421 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/1d891345-93bd-4726-905f-8a946a4a885f-apiservice-cert\") pod \"packageserver-d55dfcdfc-rnf2p\" (UID: \"1d891345-93bd-4726-905f-8a946a4a885f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rnf2p" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.950467 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v2khw\" (UniqueName: \"kubernetes.io/projected/1d891345-93bd-4726-905f-8a946a4a885f-kube-api-access-v2khw\") pod \"packageserver-d55dfcdfc-rnf2p\" (UID: \"1d891345-93bd-4726-905f-8a946a4a885f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rnf2p" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.950496 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q92sd\" (UniqueName: \"kubernetes.io/projected/0a83e6b8-c65a-42ef-b4e3-8748503bcd84-kube-api-access-q92sd\") pod \"service-ca-operator-777779d784-wkc9x\" (UID: \"0a83e6b8-c65a-42ef-b4e3-8748503bcd84\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-wkc9x" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.950514 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/edd471fb-addd-407d-b7cb-55dbffd7a1ee-mountpoint-dir\") pod \"csi-hostpathplugin-ptjzm\" (UID: \"edd471fb-addd-407d-b7cb-55dbffd7a1ee\") " pod="hostpath-provisioner/csi-hostpathplugin-ptjzm" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.950543 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hfvbs\" (UniqueName: \"kubernetes.io/projected/63176391-0acc-4316-a0fe-520669eb1744-kube-api-access-hfvbs\") pod \"package-server-manager-789f6589d5-52qbc\" (UID: \"63176391-0acc-4316-a0fe-520669eb1744\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-52qbc" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.950565 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/edd471fb-addd-407d-b7cb-55dbffd7a1ee-registration-dir\") pod \"csi-hostpathplugin-ptjzm\" (UID: \"edd471fb-addd-407d-b7cb-55dbffd7a1ee\") " pod="hostpath-provisioner/csi-hostpathplugin-ptjzm" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.950580 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hrhqh\" (UniqueName: \"kubernetes.io/projected/1df54538-19c2-472b-b93e-b7f947190db0-kube-api-access-hrhqh\") pod \"machine-config-server-klh8n\" (UID: \"1df54538-19c2-472b-b93e-b7f947190db0\") " pod="openshift-machine-config-operator/machine-config-server-klh8n" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.950604 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/1df54538-19c2-472b-b93e-b7f947190db0-node-bootstrap-token\") pod \"machine-config-server-klh8n\" (UID: \"1df54538-19c2-472b-b93e-b7f947190db0\") " pod="openshift-machine-config-operator/machine-config-server-klh8n" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.950621 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/8432ae60-6e63-45cf-b21b-ab1010687654-signing-key\") pod \"service-ca-9c57cc56f-5ptfj\" (UID: \"8432ae60-6e63-45cf-b21b-ab1010687654\") " pod="openshift-service-ca/service-ca-9c57cc56f-5ptfj" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.950639 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/27bd2071-dddb-48c6-905c-22bcb4e84482-auth-proxy-config\") pod \"machine-config-operator-74547568cd-69t6j\" (UID: \"27bd2071-dddb-48c6-905c-22bcb4e84482\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-69t6j" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.950658 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/737f304d-12ac-4d4b-97fb-e99506dd08a8-config-volume\") pod \"collect-profiles-29340420-t6cgj\" (UID: \"737f304d-12ac-4d4b-97fb-e99506dd08a8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340420-t6cgj" Oct 14 07:03:28 crc kubenswrapper[4870]: E1014 07:03:28.950704 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 07:03:29.45067106 +0000 UTC m=+145.148031431 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.950764 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/8f610aec-a4a3-4d54-b805-4f01c0de5dce-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-6vdcc\" (UID: \"8f610aec-a4a3-4d54-b805-4f01c0de5dce\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-6vdcc" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.950803 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/737f304d-12ac-4d4b-97fb-e99506dd08a8-secret-volume\") pod \"collect-profiles-29340420-t6cgj\" (UID: \"737f304d-12ac-4d4b-97fb-e99506dd08a8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340420-t6cgj" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.950822 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0a83e6b8-c65a-42ef-b4e3-8748503bcd84-serving-cert\") pod \"service-ca-operator-777779d784-wkc9x\" (UID: \"0a83e6b8-c65a-42ef-b4e3-8748503bcd84\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-wkc9x" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.950841 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/1d891345-93bd-4726-905f-8a946a4a885f-webhook-cert\") pod \"packageserver-d55dfcdfc-rnf2p\" (UID: \"1d891345-93bd-4726-905f-8a946a4a885f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rnf2p" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.950863 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/1d891345-93bd-4726-905f-8a946a4a885f-tmpfs\") pod \"packageserver-d55dfcdfc-rnf2p\" (UID: \"1d891345-93bd-4726-905f-8a946a4a885f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rnf2p" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.950887 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0a83e6b8-c65a-42ef-b4e3-8748503bcd84-config\") pod \"service-ca-operator-777779d784-wkc9x\" (UID: \"0a83e6b8-c65a-42ef-b4e3-8748503bcd84\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-wkc9x" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.950924 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/bb92d614-7f4e-470e-a68d-0aff68d3894d-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-6ztwm\" (UID: \"bb92d614-7f4e-470e-a68d-0aff68d3894d\") " pod="openshift-marketplace/marketplace-operator-79b997595-6ztwm" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.950976 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/30ec1e91-85db-4abb-b3dc-19cdcc8d4234-cert\") pod \"ingress-canary-q26kj\" (UID: \"30ec1e91-85db-4abb-b3dc-19cdcc8d4234\") " pod="openshift-ingress-canary/ingress-canary-q26kj" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.951018 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jhf85\" (UniqueName: \"kubernetes.io/projected/edd471fb-addd-407d-b7cb-55dbffd7a1ee-kube-api-access-jhf85\") pod \"csi-hostpathplugin-ptjzm\" (UID: \"edd471fb-addd-407d-b7cb-55dbffd7a1ee\") " pod="hostpath-provisioner/csi-hostpathplugin-ptjzm" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.951035 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zxf4h\" (UniqueName: \"kubernetes.io/projected/8432ae60-6e63-45cf-b21b-ab1010687654-kube-api-access-zxf4h\") pod \"service-ca-9c57cc56f-5ptfj\" (UID: \"8432ae60-6e63-45cf-b21b-ab1010687654\") " pod="openshift-service-ca/service-ca-9c57cc56f-5ptfj" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.951069 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/63176391-0acc-4316-a0fe-520669eb1744-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-52qbc\" (UID: \"63176391-0acc-4316-a0fe-520669eb1744\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-52qbc" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.951095 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kkxjz\" (UID: \"2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3\") " pod="openshift-image-registry/image-registry-697d97f7c8-kkxjz" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.951146 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j7bm9\" (UniqueName: \"kubernetes.io/projected/8f610aec-a4a3-4d54-b805-4f01c0de5dce-kube-api-access-j7bm9\") pod \"control-plane-machine-set-operator-78cbb6b69f-6vdcc\" (UID: \"8f610aec-a4a3-4d54-b805-4f01c0de5dce\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-6vdcc" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.951166 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bb92d614-7f4e-470e-a68d-0aff68d3894d-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-6ztwm\" (UID: \"bb92d614-7f4e-470e-a68d-0aff68d3894d\") " pod="openshift-marketplace/marketplace-operator-79b997595-6ztwm" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.951184 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/8432ae60-6e63-45cf-b21b-ab1010687654-signing-cabundle\") pod \"service-ca-9c57cc56f-5ptfj\" (UID: \"8432ae60-6e63-45cf-b21b-ab1010687654\") " pod="openshift-service-ca/service-ca-9c57cc56f-5ptfj" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.951251 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9057841b-b3a4-49f2-85a6-812074c635e6-config-volume\") pod \"dns-default-d4fhp\" (UID: \"9057841b-b3a4-49f2-85a6-812074c635e6\") " pod="openshift-dns/dns-default-d4fhp" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.951272 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6m4zp\" (UniqueName: \"kubernetes.io/projected/737f304d-12ac-4d4b-97fb-e99506dd08a8-kube-api-access-6m4zp\") pod \"collect-profiles-29340420-t6cgj\" (UID: \"737f304d-12ac-4d4b-97fb-e99506dd08a8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340420-t6cgj" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.951866 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/737f304d-12ac-4d4b-97fb-e99506dd08a8-config-volume\") pod \"collect-profiles-29340420-t6cgj\" (UID: \"737f304d-12ac-4d4b-97fb-e99506dd08a8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340420-t6cgj" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.952136 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/edd471fb-addd-407d-b7cb-55dbffd7a1ee-mountpoint-dir\") pod \"csi-hostpathplugin-ptjzm\" (UID: \"edd471fb-addd-407d-b7cb-55dbffd7a1ee\") " pod="hostpath-provisioner/csi-hostpathplugin-ptjzm" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.952331 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/edd471fb-addd-407d-b7cb-55dbffd7a1ee-registration-dir\") pod \"csi-hostpathplugin-ptjzm\" (UID: \"edd471fb-addd-407d-b7cb-55dbffd7a1ee\") " pod="hostpath-provisioner/csi-hostpathplugin-ptjzm" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.955877 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/27bd2071-dddb-48c6-905c-22bcb4e84482-proxy-tls\") pod \"machine-config-operator-74547568cd-69t6j\" (UID: \"27bd2071-dddb-48c6-905c-22bcb4e84482\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-69t6j" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.956241 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/27bd2071-dddb-48c6-905c-22bcb4e84482-auth-proxy-config\") pod \"machine-config-operator-74547568cd-69t6j\" (UID: \"27bd2071-dddb-48c6-905c-22bcb4e84482\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-69t6j" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.959493 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/1d891345-93bd-4726-905f-8a946a4a885f-apiservice-cert\") pod \"packageserver-d55dfcdfc-rnf2p\" (UID: \"1d891345-93bd-4726-905f-8a946a4a885f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rnf2p" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.960382 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/27bd2071-dddb-48c6-905c-22bcb4e84482-images\") pod \"machine-config-operator-74547568cd-69t6j\" (UID: \"27bd2071-dddb-48c6-905c-22bcb4e84482\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-69t6j" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.961218 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8xg8z\" (UniqueName: \"kubernetes.io/projected/be735c57-60b4-44cb-abf3-888cf00d6c31-kube-api-access-8xg8z\") pod \"ingress-operator-5b745b69d9-6z4z4\" (UID: \"be735c57-60b4-44cb-abf3-888cf00d6c31\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-6z4z4" Oct 14 07:03:28 crc kubenswrapper[4870]: E1014 07:03:28.962019 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-14 07:03:29.461992995 +0000 UTC m=+145.159353556 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kkxjz" (UID: "2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.965169 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/edd471fb-addd-407d-b7cb-55dbffd7a1ee-socket-dir\") pod \"csi-hostpathplugin-ptjzm\" (UID: \"edd471fb-addd-407d-b7cb-55dbffd7a1ee\") " pod="hostpath-provisioner/csi-hostpathplugin-ptjzm" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.965945 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/8432ae60-6e63-45cf-b21b-ab1010687654-signing-cabundle\") pod \"service-ca-9c57cc56f-5ptfj\" (UID: \"8432ae60-6e63-45cf-b21b-ab1010687654\") " pod="openshift-service-ca/service-ca-9c57cc56f-5ptfj" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.968116 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/8432ae60-6e63-45cf-b21b-ab1010687654-signing-key\") pod \"service-ca-9c57cc56f-5ptfj\" (UID: \"8432ae60-6e63-45cf-b21b-ab1010687654\") " pod="openshift-service-ca/service-ca-9c57cc56f-5ptfj" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.968606 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/1df54538-19c2-472b-b93e-b7f947190db0-certs\") pod \"machine-config-server-klh8n\" (UID: \"1df54538-19c2-472b-b93e-b7f947190db0\") " pod="openshift-machine-config-operator/machine-config-server-klh8n" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.969291 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0a83e6b8-c65a-42ef-b4e3-8748503bcd84-config\") pod \"service-ca-operator-777779d784-wkc9x\" (UID: \"0a83e6b8-c65a-42ef-b4e3-8748503bcd84\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-wkc9x" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.969418 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/737f304d-12ac-4d4b-97fb-e99506dd08a8-secret-volume\") pod \"collect-profiles-29340420-t6cgj\" (UID: \"737f304d-12ac-4d4b-97fb-e99506dd08a8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340420-t6cgj" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.970344 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0a83e6b8-c65a-42ef-b4e3-8748503bcd84-serving-cert\") pod \"service-ca-operator-777779d784-wkc9x\" (UID: \"0a83e6b8-c65a-42ef-b4e3-8748503bcd84\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-wkc9x" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.972026 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/1d891345-93bd-4726-905f-8a946a4a885f-tmpfs\") pod \"packageserver-d55dfcdfc-rnf2p\" (UID: \"1d891345-93bd-4726-905f-8a946a4a885f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rnf2p" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.972050 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/edd471fb-addd-407d-b7cb-55dbffd7a1ee-csi-data-dir\") pod \"csi-hostpathplugin-ptjzm\" (UID: \"edd471fb-addd-407d-b7cb-55dbffd7a1ee\") " pod="hostpath-provisioner/csi-hostpathplugin-ptjzm" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.972091 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/63176391-0acc-4316-a0fe-520669eb1744-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-52qbc\" (UID: \"63176391-0acc-4316-a0fe-520669eb1744\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-52qbc" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.972315 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9057841b-b3a4-49f2-85a6-812074c635e6-config-volume\") pod \"dns-default-d4fhp\" (UID: \"9057841b-b3a4-49f2-85a6-812074c635e6\") " pod="openshift-dns/dns-default-d4fhp" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.972393 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/edd471fb-addd-407d-b7cb-55dbffd7a1ee-plugins-dir\") pod \"csi-hostpathplugin-ptjzm\" (UID: \"edd471fb-addd-407d-b7cb-55dbffd7a1ee\") " pod="hostpath-provisioner/csi-hostpathplugin-ptjzm" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.974935 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/bb92d614-7f4e-470e-a68d-0aff68d3894d-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-6ztwm\" (UID: \"bb92d614-7f4e-470e-a68d-0aff68d3894d\") " pod="openshift-marketplace/marketplace-operator-79b997595-6ztwm" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.981068 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/8f610aec-a4a3-4d54-b805-4f01c0de5dce-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-6vdcc\" (UID: \"8f610aec-a4a3-4d54-b805-4f01c0de5dce\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-6vdcc" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.983598 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-hwfsm"] Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.985062 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/1df54538-19c2-472b-b93e-b7f947190db0-node-bootstrap-token\") pod \"machine-config-server-klh8n\" (UID: \"1df54538-19c2-472b-b93e-b7f947190db0\") " pod="openshift-machine-config-operator/machine-config-server-klh8n" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.985850 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/30ec1e91-85db-4abb-b3dc-19cdcc8d4234-cert\") pod \"ingress-canary-q26kj\" (UID: \"30ec1e91-85db-4abb-b3dc-19cdcc8d4234\") " pod="openshift-ingress-canary/ingress-canary-q26kj" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.991792 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/9057841b-b3a4-49f2-85a6-812074c635e6-metrics-tls\") pod \"dns-default-d4fhp\" (UID: \"9057841b-b3a4-49f2-85a6-812074c635e6\") " pod="openshift-dns/dns-default-d4fhp" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.992873 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bb92d614-7f4e-470e-a68d-0aff68d3894d-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-6ztwm\" (UID: \"bb92d614-7f4e-470e-a68d-0aff68d3894d\") " pod="openshift-marketplace/marketplace-operator-79b997595-6ztwm" Oct 14 07:03:28 crc kubenswrapper[4870]: I1014 07:03:28.995222 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h8l2l\" (UniqueName: \"kubernetes.io/projected/7562fe32-ac93-483f-b017-3eab4d7f94d4-kube-api-access-h8l2l\") pod \"machine-config-controller-84d6567774-mlqqj\" (UID: \"7562fe32-ac93-483f-b017-3eab4d7f94d4\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-mlqqj" Oct 14 07:03:29 crc kubenswrapper[4870]: I1014 07:03:29.005321 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/1d891345-93bd-4726-905f-8a946a4a885f-webhook-cert\") pod \"packageserver-d55dfcdfc-rnf2p\" (UID: \"1d891345-93bd-4726-905f-8a946a4a885f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rnf2p" Oct 14 07:03:29 crc kubenswrapper[4870]: I1014 07:03:29.006046 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-r9t2k" event={"ID":"b38b0739-21a6-4050-af6c-6accabefcc45","Type":"ContainerStarted","Data":"326a7f990605a108949aeb6d5053860eea43df99e0119fb3ddc3efb1c5302546"} Oct 14 07:03:29 crc kubenswrapper[4870]: I1014 07:03:29.006088 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-r9t2k" event={"ID":"b38b0739-21a6-4050-af6c-6accabefcc45","Type":"ContainerStarted","Data":"e1b6a976c48ab09e1c019e72e377ebd7c1e1da64ad1c680caf61df4052a61312"} Oct 14 07:03:29 crc kubenswrapper[4870]: I1014 07:03:29.010544 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-cgwzv"] Oct 14 07:03:29 crc kubenswrapper[4870]: I1014 07:03:29.013844 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/be735c57-60b4-44cb-abf3-888cf00d6c31-bound-sa-token\") pod \"ingress-operator-5b745b69d9-6z4z4\" (UID: \"be735c57-60b4-44cb-abf3-888cf00d6c31\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-6z4z4" Oct 14 07:03:29 crc kubenswrapper[4870]: I1014 07:03:29.017485 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-2qq5l" event={"ID":"9f15578c-a74c-4876-aaf3-09eb81cb0987","Type":"ContainerStarted","Data":"95c9ccdbba8f444ba3dc33aa028de95dc6fef11c8d0f5c737c68f3c95b3219ef"} Oct 14 07:03:29 crc kubenswrapper[4870]: I1014 07:03:29.018428 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-2qq5l" event={"ID":"9f15578c-a74c-4876-aaf3-09eb81cb0987","Type":"ContainerStarted","Data":"9e6a11e776c09e527384321a2c6cdc308da3febb93b6ace89ecd0b64cfff75ea"} Oct 14 07:03:29 crc kubenswrapper[4870]: I1014 07:03:29.018463 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-2qq5l" Oct 14 07:03:29 crc kubenswrapper[4870]: I1014 07:03:29.020805 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-x8qtb" event={"ID":"50732192-84e4-41fa-9c91-e5f77d85fc5e","Type":"ContainerStarted","Data":"67de1d7ccbe5e226b042604620c783e69b32f30a5b52159c97f5d396a579cbaf"} Oct 14 07:03:29 crc kubenswrapper[4870]: I1014 07:03:29.025504 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-n8z6k" event={"ID":"03d1e16b-e364-4902-9412-0dc138fc76e3","Type":"ContainerStarted","Data":"0fcc7060db62ce67d318635d3dcd8d4731c01fbcd332a66740c946b2cf7ed236"} Oct 14 07:03:29 crc kubenswrapper[4870]: I1014 07:03:29.029849 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-5dsrq"] Oct 14 07:03:29 crc kubenswrapper[4870]: I1014 07:03:29.032398 4870 patch_prober.go:28] interesting pod/console-operator-58897d9998-2qq5l container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.9:8443/readyz\": dial tcp 10.217.0.9:8443: connect: connection refused" start-of-body= Oct 14 07:03:29 crc kubenswrapper[4870]: I1014 07:03:29.032460 4870 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-2qq5l" podUID="9f15578c-a74c-4876-aaf3-09eb81cb0987" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.9:8443/readyz\": dial tcp 10.217.0.9:8443: connect: connection refused" Oct 14 07:03:29 crc kubenswrapper[4870]: I1014 07:03:29.037848 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-24b47\" (UniqueName: \"kubernetes.io/projected/2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3-kube-api-access-24b47\") pod \"image-registry-697d97f7c8-kkxjz\" (UID: \"2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3\") " pod="openshift-image-registry/image-registry-697d97f7c8-kkxjz" Oct 14 07:03:29 crc kubenswrapper[4870]: I1014 07:03:29.045767 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2st6\" (UniqueName: \"kubernetes.io/projected/2a72b0a0-287b-42dc-8c39-1976188955de-kube-api-access-s2st6\") pod \"downloads-7954f5f757-d7cdf\" (UID: \"2a72b0a0-287b-42dc-8c39-1976188955de\") " pod="openshift-console/downloads-7954f5f757-d7cdf" Oct 14 07:03:29 crc kubenswrapper[4870]: I1014 07:03:29.052311 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-6z4z4" Oct 14 07:03:29 crc kubenswrapper[4870]: I1014 07:03:29.052570 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 07:03:29 crc kubenswrapper[4870]: E1014 07:03:29.053022 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 07:03:29.552998912 +0000 UTC m=+145.250359283 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:03:29 crc kubenswrapper[4870]: I1014 07:03:29.059041 4870 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-4wngz container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.19:8443/healthz\": dial tcp 10.217.0.19:8443: connect: connection refused" start-of-body= Oct 14 07:03:29 crc kubenswrapper[4870]: I1014 07:03:29.059084 4870 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-4wngz" podUID="dce0e109-9f42-4320-b7bb-36dfd8aecb90" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.19:8443/healthz\": dial tcp 10.217.0.19:8443: connect: connection refused" Oct 14 07:03:29 crc kubenswrapper[4870]: I1014 07:03:29.065325 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m77k7\" (UniqueName: \"kubernetes.io/projected/28a417a5-b0b0-465a-ade0-3bf3ceb5ebdd-kube-api-access-m77k7\") pod \"router-default-5444994796-h82f7\" (UID: \"28a417a5-b0b0-465a-ade0-3bf3ceb5ebdd\") " pod="openshift-ingress/router-default-5444994796-h82f7" Oct 14 07:03:29 crc kubenswrapper[4870]: I1014 07:03:29.072479 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-lbq8g" Oct 14 07:03:29 crc kubenswrapper[4870]: I1014 07:03:29.076208 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dpm2h\" (UniqueName: \"kubernetes.io/projected/a0207141-7032-4930-a837-4736469c0974-kube-api-access-dpm2h\") pod \"catalog-operator-68c6474976-kqxpq\" (UID: \"a0207141-7032-4930-a837-4736469c0974\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-kqxpq" Oct 14 07:03:29 crc kubenswrapper[4870]: I1014 07:03:29.083378 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-4wngz" event={"ID":"dce0e109-9f42-4320-b7bb-36dfd8aecb90","Type":"ContainerStarted","Data":"5691df69a48cbc54d505808e821a44380c74c285fce936e31d76f5cf21d595e5"} Oct 14 07:03:29 crc kubenswrapper[4870]: I1014 07:03:29.083427 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-l24lg"] Oct 14 07:03:29 crc kubenswrapper[4870]: I1014 07:03:29.083502 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-4wngz" Oct 14 07:03:29 crc kubenswrapper[4870]: I1014 07:03:29.083520 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-4wngz" event={"ID":"dce0e109-9f42-4320-b7bb-36dfd8aecb90","Type":"ContainerStarted","Data":"8e036a0d8d7b4bbcdf48f51122e9b5e7af15d873287975925dd048f5a3f01332"} Oct 14 07:03:29 crc kubenswrapper[4870]: I1014 07:03:29.083534 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-4wddg" event={"ID":"6f0fb8e4-f3ee-44ea-a437-d8f3d900c7fe","Type":"ContainerStarted","Data":"bf657100582abae93c6d4465e161450b11e40e14824f6588e6afe378e7cfbd93"} Oct 14 07:03:29 crc kubenswrapper[4870]: I1014 07:03:29.083554 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-4wddg" event={"ID":"6f0fb8e4-f3ee-44ea-a437-d8f3d900c7fe","Type":"ContainerStarted","Data":"f120f69a7af858627829ce3e64790cecbb60622c91d8a97f3f1fe9ee27ca15b4"} Oct 14 07:03:29 crc kubenswrapper[4870]: I1014 07:03:29.083732 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-h82f7" Oct 14 07:03:29 crc kubenswrapper[4870]: I1014 07:03:29.089678 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-q9knv" event={"ID":"3638c2bd-8853-410f-bac4-0d46d9dfa8a1","Type":"ContainerStarted","Data":"465acc0f57b2f5317bc9eb4b635206d87cdaf34fc82b2560f7f8145c63616122"} Oct 14 07:03:29 crc kubenswrapper[4870]: I1014 07:03:29.122076 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-9lgs8" event={"ID":"9b9caf4b-0443-4cfe-b9d3-6d7093d07db7","Type":"ContainerStarted","Data":"31720bad635d368a5a59ee1f16f1141d0aa2eb5750e3df7871522851c8d97b20"} Oct 14 07:03:29 crc kubenswrapper[4870]: I1014 07:03:29.122131 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-9lgs8" event={"ID":"9b9caf4b-0443-4cfe-b9d3-6d7093d07db7","Type":"ContainerStarted","Data":"8a0950719d7a468e7e56ad6094f87c1a7e52d3390fb36c66df263d28f518fe43"} Oct 14 07:03:29 crc kubenswrapper[4870]: I1014 07:03:29.130955 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vc8ks" Oct 14 07:03:29 crc kubenswrapper[4870]: I1014 07:03:29.131622 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c7cj6\" (UniqueName: \"kubernetes.io/projected/30ec1e91-85db-4abb-b3dc-19cdcc8d4234-kube-api-access-c7cj6\") pod \"ingress-canary-q26kj\" (UID: \"30ec1e91-85db-4abb-b3dc-19cdcc8d4234\") " pod="openshift-ingress-canary/ingress-canary-q26kj" Oct 14 07:03:29 crc kubenswrapper[4870]: I1014 07:03:29.139289 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-mlqqj" Oct 14 07:03:29 crc kubenswrapper[4870]: I1014 07:03:29.142374 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6m4zp\" (UniqueName: \"kubernetes.io/projected/737f304d-12ac-4d4b-97fb-e99506dd08a8-kube-api-access-6m4zp\") pod \"collect-profiles-29340420-t6cgj\" (UID: \"737f304d-12ac-4d4b-97fb-e99506dd08a8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340420-t6cgj" Oct 14 07:03:29 crc kubenswrapper[4870]: I1014 07:03:29.155975 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kkxjz\" (UID: \"2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3\") " pod="openshift-image-registry/image-registry-697d97f7c8-kkxjz" Oct 14 07:03:29 crc kubenswrapper[4870]: I1014 07:03:29.160934 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-kqxpq" Oct 14 07:03:29 crc kubenswrapper[4870]: I1014 07:03:29.177291 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7np8q\" (UniqueName: \"kubernetes.io/projected/bb92d614-7f4e-470e-a68d-0aff68d3894d-kube-api-access-7np8q\") pod \"marketplace-operator-79b997595-6ztwm\" (UID: \"bb92d614-7f4e-470e-a68d-0aff68d3894d\") " pod="openshift-marketplace/marketplace-operator-79b997595-6ztwm" Oct 14 07:03:29 crc kubenswrapper[4870]: E1014 07:03:29.179965 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-14 07:03:29.679946991 +0000 UTC m=+145.377307352 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kkxjz" (UID: "2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:03:29 crc kubenswrapper[4870]: I1014 07:03:29.196829 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v2khw\" (UniqueName: \"kubernetes.io/projected/1d891345-93bd-4726-905f-8a946a4a885f-kube-api-access-v2khw\") pod \"packageserver-d55dfcdfc-rnf2p\" (UID: \"1d891345-93bd-4726-905f-8a946a4a885f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rnf2p" Oct 14 07:03:29 crc kubenswrapper[4870]: I1014 07:03:29.209433 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29340420-t6cgj" Oct 14 07:03:29 crc kubenswrapper[4870]: I1014 07:03:29.213191 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q92sd\" (UniqueName: \"kubernetes.io/projected/0a83e6b8-c65a-42ef-b4e3-8748503bcd84-kube-api-access-q92sd\") pod \"service-ca-operator-777779d784-wkc9x\" (UID: \"0a83e6b8-c65a-42ef-b4e3-8748503bcd84\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-wkc9x" Oct 14 07:03:29 crc kubenswrapper[4870]: I1014 07:03:29.217732 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-6ztwm" Oct 14 07:03:29 crc kubenswrapper[4870]: I1014 07:03:29.223611 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hfvbs\" (UniqueName: \"kubernetes.io/projected/63176391-0acc-4316-a0fe-520669eb1744-kube-api-access-hfvbs\") pod \"package-server-manager-789f6589d5-52qbc\" (UID: \"63176391-0acc-4316-a0fe-520669eb1744\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-52qbc" Oct 14 07:03:29 crc kubenswrapper[4870]: I1014 07:03:29.236357 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-wkc9x" Oct 14 07:03:29 crc kubenswrapper[4870]: I1014 07:03:29.242087 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hrhqh\" (UniqueName: \"kubernetes.io/projected/1df54538-19c2-472b-b93e-b7f947190db0-kube-api-access-hrhqh\") pod \"machine-config-server-klh8n\" (UID: \"1df54538-19c2-472b-b93e-b7f947190db0\") " pod="openshift-machine-config-operator/machine-config-server-klh8n" Oct 14 07:03:29 crc kubenswrapper[4870]: I1014 07:03:29.257254 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 07:03:29 crc kubenswrapper[4870]: E1014 07:03:29.259000 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 07:03:29.758971547 +0000 UTC m=+145.456331918 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:03:29 crc kubenswrapper[4870]: I1014 07:03:29.260281 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-d7cdf" Oct 14 07:03:29 crc kubenswrapper[4870]: I1014 07:03:29.270017 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jhf85\" (UniqueName: \"kubernetes.io/projected/edd471fb-addd-407d-b7cb-55dbffd7a1ee-kube-api-access-jhf85\") pod \"csi-hostpathplugin-ptjzm\" (UID: \"edd471fb-addd-407d-b7cb-55dbffd7a1ee\") " pod="hostpath-provisioner/csi-hostpathplugin-ptjzm" Oct 14 07:03:29 crc kubenswrapper[4870]: I1014 07:03:29.275068 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-q26kj" Oct 14 07:03:29 crc kubenswrapper[4870]: I1014 07:03:29.282374 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-klh8n" Oct 14 07:03:29 crc kubenswrapper[4870]: I1014 07:03:29.296313 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h2ptt\" (UniqueName: \"kubernetes.io/projected/27bd2071-dddb-48c6-905c-22bcb4e84482-kube-api-access-h2ptt\") pod \"machine-config-operator-74547568cd-69t6j\" (UID: \"27bd2071-dddb-48c6-905c-22bcb4e84482\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-69t6j" Oct 14 07:03:29 crc kubenswrapper[4870]: I1014 07:03:29.297759 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j7bm9\" (UniqueName: \"kubernetes.io/projected/8f610aec-a4a3-4d54-b805-4f01c0de5dce-kube-api-access-j7bm9\") pod \"control-plane-machine-set-operator-78cbb6b69f-6vdcc\" (UID: \"8f610aec-a4a3-4d54-b805-4f01c0de5dce\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-6vdcc" Oct 14 07:03:29 crc kubenswrapper[4870]: I1014 07:03:29.316512 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zxf4h\" (UniqueName: \"kubernetes.io/projected/8432ae60-6e63-45cf-b21b-ab1010687654-kube-api-access-zxf4h\") pod \"service-ca-9c57cc56f-5ptfj\" (UID: \"8432ae60-6e63-45cf-b21b-ab1010687654\") " pod="openshift-service-ca/service-ca-9c57cc56f-5ptfj" Oct 14 07:03:29 crc kubenswrapper[4870]: I1014 07:03:29.356656 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q84wc\" (UniqueName: \"kubernetes.io/projected/9057841b-b3a4-49f2-85a6-812074c635e6-kube-api-access-q84wc\") pod \"dns-default-d4fhp\" (UID: \"9057841b-b3a4-49f2-85a6-812074c635e6\") " pod="openshift-dns/dns-default-d4fhp" Oct 14 07:03:29 crc kubenswrapper[4870]: I1014 07:03:29.368802 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kkxjz\" (UID: \"2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3\") " pod="openshift-image-registry/image-registry-697d97f7c8-kkxjz" Oct 14 07:03:29 crc kubenswrapper[4870]: E1014 07:03:29.369343 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-14 07:03:29.869324994 +0000 UTC m=+145.566685375 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kkxjz" (UID: "2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:03:29 crc kubenswrapper[4870]: I1014 07:03:29.469663 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 07:03:29 crc kubenswrapper[4870]: E1014 07:03:29.470465 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 07:03:29.970429256 +0000 UTC m=+145.667789627 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:03:29 crc kubenswrapper[4870]: I1014 07:03:29.474151 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-55gqt"] Oct 14 07:03:29 crc kubenswrapper[4870]: I1014 07:03:29.480861 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-6vdcc" Oct 14 07:03:29 crc kubenswrapper[4870]: I1014 07:03:29.482051 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-p769c"] Oct 14 07:03:29 crc kubenswrapper[4870]: I1014 07:03:29.484629 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-cvh5p"] Oct 14 07:03:29 crc kubenswrapper[4870]: I1014 07:03:29.486561 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rnf2p" Oct 14 07:03:29 crc kubenswrapper[4870]: I1014 07:03:29.493982 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-5ptfj" Oct 14 07:03:29 crc kubenswrapper[4870]: I1014 07:03:29.496306 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-zpgbg"] Oct 14 07:03:29 crc kubenswrapper[4870]: I1014 07:03:29.503951 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-52qbc" Oct 14 07:03:29 crc kubenswrapper[4870]: I1014 07:03:29.528434 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-69t6j" Oct 14 07:03:29 crc kubenswrapper[4870]: I1014 07:03:29.561300 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-ptjzm" Oct 14 07:03:29 crc kubenswrapper[4870]: I1014 07:03:29.567555 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-d4fhp" Oct 14 07:03:29 crc kubenswrapper[4870]: I1014 07:03:29.573294 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kkxjz\" (UID: \"2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3\") " pod="openshift-image-registry/image-registry-697d97f7c8-kkxjz" Oct 14 07:03:29 crc kubenswrapper[4870]: E1014 07:03:29.574189 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-14 07:03:30.074174382 +0000 UTC m=+145.771534753 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kkxjz" (UID: "2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:03:29 crc kubenswrapper[4870]: I1014 07:03:29.634812 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wjm9z"] Oct 14 07:03:29 crc kubenswrapper[4870]: W1014 07:03:29.655046 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod49dc86fe_e718_4220_81d7_2ebd7fdd6c2f.slice/crio-8bb84f150df9cba1118e6e53b01cda72441c98bf44de715354f789a0caed8165 WatchSource:0}: Error finding container 8bb84f150df9cba1118e6e53b01cda72441c98bf44de715354f789a0caed8165: Status 404 returned error can't find the container with id 8bb84f150df9cba1118e6e53b01cda72441c98bf44de715354f789a0caed8165 Oct 14 07:03:29 crc kubenswrapper[4870]: I1014 07:03:29.675510 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 07:03:29 crc kubenswrapper[4870]: E1014 07:03:29.676342 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 07:03:30.17631516 +0000 UTC m=+145.873675521 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:03:29 crc kubenswrapper[4870]: I1014 07:03:29.676405 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kkxjz\" (UID: \"2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3\") " pod="openshift-image-registry/image-registry-697d97f7c8-kkxjz" Oct 14 07:03:29 crc kubenswrapper[4870]: E1014 07:03:29.676781 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-14 07:03:30.176774821 +0000 UTC m=+145.874135192 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kkxjz" (UID: "2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:03:29 crc kubenswrapper[4870]: I1014 07:03:29.720355 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-7jz27"] Oct 14 07:03:29 crc kubenswrapper[4870]: I1014 07:03:29.726629 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-szwj9"] Oct 14 07:03:29 crc kubenswrapper[4870]: I1014 07:03:29.733129 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-g7crn"] Oct 14 07:03:29 crc kubenswrapper[4870]: I1014 07:03:29.777646 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 07:03:29 crc kubenswrapper[4870]: E1014 07:03:29.778369 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 07:03:30.278341594 +0000 UTC m=+145.975701965 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:03:29 crc kubenswrapper[4870]: W1014 07:03:29.797889 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod222f1962_d6c1_4e69_a3bc_0b0a12703cff.slice/crio-2aa6720926b29c5431a3bb809a04f1c61d12428383d0b97dd5430afb08575f4b WatchSource:0}: Error finding container 2aa6720926b29c5431a3bb809a04f1c61d12428383d0b97dd5430afb08575f4b: Status 404 returned error can't find the container with id 2aa6720926b29c5431a3bb809a04f1c61d12428383d0b97dd5430afb08575f4b Oct 14 07:03:29 crc kubenswrapper[4870]: W1014 07:03:29.825417 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf07c76fa_f754_4175_8bc5_2aae95985b3b.slice/crio-743b6d5785594c542b1932b7b3d9b414ad0cf7d045450796cc4ac7f72d440c00 WatchSource:0}: Error finding container 743b6d5785594c542b1932b7b3d9b414ad0cf7d045450796cc4ac7f72d440c00: Status 404 returned error can't find the container with id 743b6d5785594c542b1932b7b3d9b414ad0cf7d045450796cc4ac7f72d440c00 Oct 14 07:03:29 crc kubenswrapper[4870]: W1014 07:03:29.875966 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb419b072_5f0d_4103_82c6_c296f71eee8a.slice/crio-a77f14f88687cd51ba8d857613ad96caed60a974c0552b6b9a922f84693a5ffd WatchSource:0}: Error finding container a77f14f88687cd51ba8d857613ad96caed60a974c0552b6b9a922f84693a5ffd: Status 404 returned error can't find the container with id a77f14f88687cd51ba8d857613ad96caed60a974c0552b6b9a922f84693a5ffd Oct 14 07:03:29 crc kubenswrapper[4870]: I1014 07:03:29.883309 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kkxjz\" (UID: \"2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3\") " pod="openshift-image-registry/image-registry-697d97f7c8-kkxjz" Oct 14 07:03:29 crc kubenswrapper[4870]: E1014 07:03:29.883876 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-14 07:03:30.383859363 +0000 UTC m=+146.081219734 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kkxjz" (UID: "2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:03:29 crc kubenswrapper[4870]: I1014 07:03:29.918014 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-6z4z4"] Oct 14 07:03:29 crc kubenswrapper[4870]: I1014 07:03:29.939182 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-kqxpq"] Oct 14 07:03:29 crc kubenswrapper[4870]: I1014 07:03:29.951268 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ss574"] Oct 14 07:03:30 crc kubenswrapper[4870]: I1014 07:03:30.004313 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 07:03:30 crc kubenswrapper[4870]: E1014 07:03:30.005303 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 07:03:30.505276738 +0000 UTC m=+146.202637109 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:03:30 crc kubenswrapper[4870]: W1014 07:03:30.046409 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd9cb408e_a00b_4f76_b8fb_bf144bfee740.slice/crio-5b3252b1ad8e355181475011ceab250bd5282cf07eeb2c8f8c9b833be7163248 WatchSource:0}: Error finding container 5b3252b1ad8e355181475011ceab250bd5282cf07eeb2c8f8c9b833be7163248: Status 404 returned error can't find the container with id 5b3252b1ad8e355181475011ceab250bd5282cf07eeb2c8f8c9b833be7163248 Oct 14 07:03:30 crc kubenswrapper[4870]: I1014 07:03:30.085710 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-r9t2k" podStartSLOduration=124.085674088 podStartE2EDuration="2m4.085674088s" podCreationTimestamp="2025-10-14 07:01:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:03:30.081930597 +0000 UTC m=+145.779290968" watchObservedRunningTime="2025-10-14 07:03:30.085674088 +0000 UTC m=+145.783034459" Oct 14 07:03:30 crc kubenswrapper[4870]: I1014 07:03:30.112379 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vc8ks"] Oct 14 07:03:30 crc kubenswrapper[4870]: I1014 07:03:30.112475 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-lbq8g"] Oct 14 07:03:30 crc kubenswrapper[4870]: I1014 07:03:30.117474 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kkxjz\" (UID: \"2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3\") " pod="openshift-image-registry/image-registry-697d97f7c8-kkxjz" Oct 14 07:03:30 crc kubenswrapper[4870]: E1014 07:03:30.119014 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-14 07:03:30.618997306 +0000 UTC m=+146.316357677 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kkxjz" (UID: "2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:03:30 crc kubenswrapper[4870]: I1014 07:03:30.120183 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-mlqqj"] Oct 14 07:03:30 crc kubenswrapper[4870]: I1014 07:03:30.139682 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-2qq5l" podStartSLOduration=124.139657087 podStartE2EDuration="2m4.139657087s" podCreationTimestamp="2025-10-14 07:01:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:03:30.134451261 +0000 UTC m=+145.831811632" watchObservedRunningTime="2025-10-14 07:03:30.139657087 +0000 UTC m=+145.837017468" Oct 14 07:03:30 crc kubenswrapper[4870]: I1014 07:03:30.214796 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-szwj9" event={"ID":"8070623a-4696-490e-af90-5b2155c0506c","Type":"ContainerStarted","Data":"65dfadcc37a956936f9d64f27b68c9aac6823178f1c4b0e06bee8ebc27e0b067"} Oct 14 07:03:30 crc kubenswrapper[4870]: I1014 07:03:30.221400 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 07:03:30 crc kubenswrapper[4870]: E1014 07:03:30.221825 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 07:03:30.72180521 +0000 UTC m=+146.419165581 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:03:30 crc kubenswrapper[4870]: I1014 07:03:30.264127 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-d7cdf"] Oct 14 07:03:30 crc kubenswrapper[4870]: I1014 07:03:30.272115 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-9lgs8" event={"ID":"9b9caf4b-0443-4cfe-b9d3-6d7093d07db7","Type":"ContainerStarted","Data":"2df04bb6d041f0454c32f36b94c88d451eb63ad63a179fa80407527c0536647f"} Oct 14 07:03:30 crc kubenswrapper[4870]: I1014 07:03:30.274161 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-55gqt" event={"ID":"49dc86fe-e718-4220-81d7-2ebd7fdd6c2f","Type":"ContainerStarted","Data":"8bb84f150df9cba1118e6e53b01cda72441c98bf44de715354f789a0caed8165"} Oct 14 07:03:30 crc kubenswrapper[4870]: I1014 07:03:30.280097 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-g7crn" event={"ID":"16295052-1474-4dc9-b1d6-10983e6bfc6f","Type":"ContainerStarted","Data":"f74cae48b61f3526f6a4434c418c8b903efffd1323734b2d12b24851d45ee5a0"} Oct 14 07:03:30 crc kubenswrapper[4870]: I1014 07:03:30.287867 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-cvh5p" event={"ID":"337df1ca-c52e-499c-ab1d-a32aee9c9922","Type":"ContainerStarted","Data":"56d538dc7d9751de7e20f1d5fd5df37610e0bad45ab296979d7169447777f747"} Oct 14 07:03:30 crc kubenswrapper[4870]: I1014 07:03:30.315492 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29340420-t6cgj"] Oct 14 07:03:30 crc kubenswrapper[4870]: I1014 07:03:30.316968 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-q26kj"] Oct 14 07:03:30 crc kubenswrapper[4870]: I1014 07:03:30.326500 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kkxjz\" (UID: \"2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3\") " pod="openshift-image-registry/image-registry-697d97f7c8-kkxjz" Oct 14 07:03:30 crc kubenswrapper[4870]: E1014 07:03:30.326833 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-14 07:03:30.826820067 +0000 UTC m=+146.524180438 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kkxjz" (UID: "2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:03:30 crc kubenswrapper[4870]: I1014 07:03:30.327484 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-cgwzv" event={"ID":"88363344-bc9a-4050-87e2-cf6f07708835","Type":"ContainerStarted","Data":"5b22c5a28105d010f3ad1cdfcc16b200120d0661c9e0482843532e463c6d5ffa"} Oct 14 07:03:30 crc kubenswrapper[4870]: I1014 07:03:30.327535 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-cgwzv" event={"ID":"88363344-bc9a-4050-87e2-cf6f07708835","Type":"ContainerStarted","Data":"fa080008caef84fd2513c326caeb7737b986b7c85f980e23bd143a6d2bd12ec1"} Oct 14 07:03:30 crc kubenswrapper[4870]: I1014 07:03:30.330132 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-hwfsm" event={"ID":"7b7a521b-c795-43cc-9576-5a3be3cb7409","Type":"ContainerStarted","Data":"672526d47e1097673498f09b0e1505e3678cbfca5cfaedd12cd11fdd8eca206a"} Oct 14 07:03:30 crc kubenswrapper[4870]: I1014 07:03:30.330163 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-hwfsm" event={"ID":"7b7a521b-c795-43cc-9576-5a3be3cb7409","Type":"ContainerStarted","Data":"4fc575a3c02fc645d6e14ac1c2a5bae30b18a12ed1a50c7688e191bdb42f6645"} Oct 14 07:03:30 crc kubenswrapper[4870]: I1014 07:03:30.330846 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-kqxpq" event={"ID":"a0207141-7032-4930-a837-4736469c0974","Type":"ContainerStarted","Data":"f5087e09ddcf9a8b1e750ff21d0097811ee7a2985aad1f04e8f7df5bfc19ad57"} Oct 14 07:03:30 crc kubenswrapper[4870]: I1014 07:03:30.332823 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ss574" event={"ID":"d9cb408e-a00b-4f76-b8fb-bf144bfee740","Type":"ContainerStarted","Data":"5b3252b1ad8e355181475011ceab250bd5282cf07eeb2c8f8c9b833be7163248"} Oct 14 07:03:30 crc kubenswrapper[4870]: I1014 07:03:30.334091 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-zpgbg" event={"ID":"222f1962-d6c1-4e69-a3bc-0b0a12703cff","Type":"ContainerStarted","Data":"2aa6720926b29c5431a3bb809a04f1c61d12428383d0b97dd5430afb08575f4b"} Oct 14 07:03:30 crc kubenswrapper[4870]: I1014 07:03:30.335469 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wjm9z" event={"ID":"f07c76fa-f754-4175-8bc5-2aae95985b3b","Type":"ContainerStarted","Data":"743b6d5785594c542b1932b7b3d9b414ad0cf7d045450796cc4ac7f72d440c00"} Oct 14 07:03:30 crc kubenswrapper[4870]: I1014 07:03:30.345398 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-l24lg" event={"ID":"c0cd54e9-09ac-49a6-a774-a9148024b4b9","Type":"ContainerStarted","Data":"b69c464fa90f7ffb7396a685487e1337b3cefc3ba0811363ac9f7df1b93c74b4"} Oct 14 07:03:30 crc kubenswrapper[4870]: I1014 07:03:30.345473 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-l24lg" event={"ID":"c0cd54e9-09ac-49a6-a774-a9148024b4b9","Type":"ContainerStarted","Data":"418206cdc42d2ca021d151c2705191c6bd77ddcd178e4ab575cac9a9337d3fc0"} Oct 14 07:03:30 crc kubenswrapper[4870]: I1014 07:03:30.352166 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-h82f7" event={"ID":"28a417a5-b0b0-465a-ade0-3bf3ceb5ebdd","Type":"ContainerStarted","Data":"dbdfb39654f27eccc5021f9079874e212fab7697bff86240b6360f8cccf4ec48"} Oct 14 07:03:30 crc kubenswrapper[4870]: W1014 07:03:30.362119 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7562fe32_ac93_483f_b017_3eab4d7f94d4.slice/crio-dafd5010890540ccc37181405e1f0df5eee0c233a4b089ca7fada071f80f0a0b WatchSource:0}: Error finding container dafd5010890540ccc37181405e1f0df5eee0c233a4b089ca7fada071f80f0a0b: Status 404 returned error can't find the container with id dafd5010890540ccc37181405e1f0df5eee0c233a4b089ca7fada071f80f0a0b Oct 14 07:03:30 crc kubenswrapper[4870]: I1014 07:03:30.371740 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-wkc9x"] Oct 14 07:03:30 crc kubenswrapper[4870]: I1014 07:03:30.371800 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-klh8n" event={"ID":"1df54538-19c2-472b-b93e-b7f947190db0","Type":"ContainerStarted","Data":"06a2f2401e331f5906c14a86d82535049465cb4e51a3e1b23a026c66e8968ac2"} Oct 14 07:03:30 crc kubenswrapper[4870]: I1014 07:03:30.379524 4870 generic.go:334] "Generic (PLEG): container finished" podID="3638c2bd-8853-410f-bac4-0d46d9dfa8a1" containerID="eaff82536cb78abcfe1cb7a411a04b0b0cdf5c9fbf3d0ad3dd09c6885f845345" exitCode=0 Oct 14 07:03:30 crc kubenswrapper[4870]: I1014 07:03:30.379595 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-q9knv" event={"ID":"3638c2bd-8853-410f-bac4-0d46d9dfa8a1","Type":"ContainerDied","Data":"eaff82536cb78abcfe1cb7a411a04b0b0cdf5c9fbf3d0ad3dd09c6885f845345"} Oct 14 07:03:30 crc kubenswrapper[4870]: I1014 07:03:30.379626 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-q9knv" event={"ID":"3638c2bd-8853-410f-bac4-0d46d9dfa8a1","Type":"ContainerStarted","Data":"73fc477de36251e85594491e1204338b688271b4ed430994dbd33ed1c4b3c6bb"} Oct 14 07:03:30 crc kubenswrapper[4870]: I1014 07:03:30.384731 4870 generic.go:334] "Generic (PLEG): container finished" podID="03d1e16b-e364-4902-9412-0dc138fc76e3" containerID="f7fd134badf268b663a34fa412290cc9039cd6780375a09a3417d7a1d720b78a" exitCode=0 Oct 14 07:03:30 crc kubenswrapper[4870]: I1014 07:03:30.384787 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-n8z6k" event={"ID":"03d1e16b-e364-4902-9412-0dc138fc76e3","Type":"ContainerDied","Data":"f7fd134badf268b663a34fa412290cc9039cd6780375a09a3417d7a1d720b78a"} Oct 14 07:03:30 crc kubenswrapper[4870]: I1014 07:03:30.392212 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-7jz27" event={"ID":"b419b072-5f0d-4103-82c6-c296f71eee8a","Type":"ContainerStarted","Data":"a77f14f88687cd51ba8d857613ad96caed60a974c0552b6b9a922f84693a5ffd"} Oct 14 07:03:30 crc kubenswrapper[4870]: I1014 07:03:30.398817 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-52qbc"] Oct 14 07:03:30 crc kubenswrapper[4870]: I1014 07:03:30.416988 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-6ztwm"] Oct 14 07:03:30 crc kubenswrapper[4870]: I1014 07:03:30.427464 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 07:03:30 crc kubenswrapper[4870]: E1014 07:03:30.433923 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 07:03:30.933888303 +0000 UTC m=+146.631248674 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:03:30 crc kubenswrapper[4870]: I1014 07:03:30.458594 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-5dsrq" event={"ID":"3f44b737-38ee-4c89-85fb-264a66ce8771","Type":"ContainerStarted","Data":"8323356a7abe12ca87e33237676874d7b02b20dcb69cc2f9c92ee29089ad90bb"} Oct 14 07:03:30 crc kubenswrapper[4870]: I1014 07:03:30.458645 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-5dsrq" event={"ID":"3f44b737-38ee-4c89-85fb-264a66ce8771","Type":"ContainerStarted","Data":"a8d0976f75c8a811ad238d4e8c5c2d14a789898e867c800ed5f18dafb7a547d4"} Oct 14 07:03:30 crc kubenswrapper[4870]: I1014 07:03:30.487764 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-p769c" event={"ID":"146160a7-2b1c-43f4-bc42-7b92d9663dcc","Type":"ContainerStarted","Data":"6a2006ccad2af6161a41ba4966bc9647dd19b12f9d00a3b6fd03cc4c7ac4fc5f"} Oct 14 07:03:30 crc kubenswrapper[4870]: I1014 07:03:30.495659 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-x8qtb" event={"ID":"50732192-84e4-41fa-9c91-e5f77d85fc5e","Type":"ContainerStarted","Data":"80aa6bb356d5e8eb5904e69dd4809ac3a3a7f105d22cae7a99b5353b68dcbfb2"} Oct 14 07:03:30 crc kubenswrapper[4870]: I1014 07:03:30.496074 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-x8qtb" Oct 14 07:03:30 crc kubenswrapper[4870]: I1014 07:03:30.503585 4870 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-x8qtb container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.14:6443/healthz\": dial tcp 10.217.0.14:6443: connect: connection refused" start-of-body= Oct 14 07:03:30 crc kubenswrapper[4870]: I1014 07:03:30.503683 4870 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-x8qtb" podUID="50732192-84e4-41fa-9c91-e5f77d85fc5e" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.14:6443/healthz\": dial tcp 10.217.0.14:6443: connect: connection refused" Oct 14 07:03:30 crc kubenswrapper[4870]: I1014 07:03:30.520837 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-6z4z4" event={"ID":"be735c57-60b4-44cb-abf3-888cf00d6c31","Type":"ContainerStarted","Data":"f4029f8caf9466e6bdaa4f3f83547cc86e708099a51b033f780361e05586fd28"} Oct 14 07:03:30 crc kubenswrapper[4870]: I1014 07:03:30.529968 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-4wngz" Oct 14 07:03:30 crc kubenswrapper[4870]: I1014 07:03:30.534388 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kkxjz\" (UID: \"2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3\") " pod="openshift-image-registry/image-registry-697d97f7c8-kkxjz" Oct 14 07:03:30 crc kubenswrapper[4870]: E1014 07:03:30.536941 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-14 07:03:31.036921042 +0000 UTC m=+146.734281423 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kkxjz" (UID: "2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:03:30 crc kubenswrapper[4870]: I1014 07:03:30.566495 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-6vdcc"] Oct 14 07:03:30 crc kubenswrapper[4870]: I1014 07:03:30.635928 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 07:03:30 crc kubenswrapper[4870]: E1014 07:03:30.637772 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 07:03:31.137745928 +0000 UTC m=+146.835106349 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:03:30 crc kubenswrapper[4870]: I1014 07:03:30.676607 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-4wngz" podStartSLOduration=124.67658631 podStartE2EDuration="2m4.67658631s" podCreationTimestamp="2025-10-14 07:01:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:03:30.638302621 +0000 UTC m=+146.335662982" watchObservedRunningTime="2025-10-14 07:03:30.67658631 +0000 UTC m=+146.373946681" Oct 14 07:03:30 crc kubenswrapper[4870]: I1014 07:03:30.742741 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kkxjz\" (UID: \"2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3\") " pod="openshift-image-registry/image-registry-697d97f7c8-kkxjz" Oct 14 07:03:30 crc kubenswrapper[4870]: E1014 07:03:30.743077 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-14 07:03:31.243064542 +0000 UTC m=+146.940424913 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kkxjz" (UID: "2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:03:30 crc kubenswrapper[4870]: I1014 07:03:30.780172 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-nc79f" podStartSLOduration=124.780138181 podStartE2EDuration="2m4.780138181s" podCreationTimestamp="2025-10-14 07:01:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:03:30.748876213 +0000 UTC m=+146.446236574" watchObservedRunningTime="2025-10-14 07:03:30.780138181 +0000 UTC m=+146.477498552" Oct 14 07:03:30 crc kubenswrapper[4870]: I1014 07:03:30.851352 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 07:03:30 crc kubenswrapper[4870]: E1014 07:03:30.851579 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 07:03:31.351536913 +0000 UTC m=+147.048897284 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:03:30 crc kubenswrapper[4870]: I1014 07:03:30.851771 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kkxjz\" (UID: \"2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3\") " pod="openshift-image-registry/image-registry-697d97f7c8-kkxjz" Oct 14 07:03:30 crc kubenswrapper[4870]: E1014 07:03:30.852141 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-14 07:03:31.352123777 +0000 UTC m=+147.049484148 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kkxjz" (UID: "2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:03:30 crc kubenswrapper[4870]: I1014 07:03:30.946711 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-4wddg" podStartSLOduration=124.94667492 podStartE2EDuration="2m4.94667492s" podCreationTimestamp="2025-10-14 07:01:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:03:30.925545318 +0000 UTC m=+146.622905689" watchObservedRunningTime="2025-10-14 07:03:30.94667492 +0000 UTC m=+146.644035301" Oct 14 07:03:30 crc kubenswrapper[4870]: I1014 07:03:30.954055 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 07:03:30 crc kubenswrapper[4870]: I1014 07:03:30.954387 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-69t6j"] Oct 14 07:03:30 crc kubenswrapper[4870]: E1014 07:03:30.954638 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 07:03:31.454617473 +0000 UTC m=+147.151977844 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:03:30 crc kubenswrapper[4870]: I1014 07:03:30.983911 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-2qq5l" Oct 14 07:03:31 crc kubenswrapper[4870]: I1014 07:03:31.057891 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kkxjz\" (UID: \"2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3\") " pod="openshift-image-registry/image-registry-697d97f7c8-kkxjz" Oct 14 07:03:31 crc kubenswrapper[4870]: E1014 07:03:31.058339 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-14 07:03:31.558323778 +0000 UTC m=+147.255684149 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kkxjz" (UID: "2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:03:31 crc kubenswrapper[4870]: I1014 07:03:31.073385 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-q9knv" podStartSLOduration=125.073363052 podStartE2EDuration="2m5.073363052s" podCreationTimestamp="2025-10-14 07:01:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:03:31.069334004 +0000 UTC m=+146.766694375" watchObservedRunningTime="2025-10-14 07:03:31.073363052 +0000 UTC m=+146.770723423" Oct 14 07:03:31 crc kubenswrapper[4870]: I1014 07:03:31.159941 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 07:03:31 crc kubenswrapper[4870]: E1014 07:03:31.160413 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 07:03:31.660398073 +0000 UTC m=+147.357758444 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:03:31 crc kubenswrapper[4870]: I1014 07:03:31.222043 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-cgwzv" podStartSLOduration=126.222007947 podStartE2EDuration="2m6.222007947s" podCreationTimestamp="2025-10-14 07:01:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:03:31.168723665 +0000 UTC m=+146.866084036" watchObservedRunningTime="2025-10-14 07:03:31.222007947 +0000 UTC m=+146.919368318" Oct 14 07:03:31 crc kubenswrapper[4870]: I1014 07:03:31.251570 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-ptjzm"] Oct 14 07:03:31 crc kubenswrapper[4870]: I1014 07:03:31.251977 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-d4fhp"] Oct 14 07:03:31 crc kubenswrapper[4870]: I1014 07:03:31.261962 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kkxjz\" (UID: \"2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3\") " pod="openshift-image-registry/image-registry-697d97f7c8-kkxjz" Oct 14 07:03:31 crc kubenswrapper[4870]: I1014 07:03:31.283630 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rnf2p"] Oct 14 07:03:31 crc kubenswrapper[4870]: E1014 07:03:31.285811 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-14 07:03:31.785780354 +0000 UTC m=+147.483140725 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kkxjz" (UID: "2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:03:31 crc kubenswrapper[4870]: I1014 07:03:31.339161 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-x8qtb" podStartSLOduration=126.339142808 podStartE2EDuration="2m6.339142808s" podCreationTimestamp="2025-10-14 07:01:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:03:31.337045297 +0000 UTC m=+147.034405668" watchObservedRunningTime="2025-10-14 07:03:31.339142808 +0000 UTC m=+147.036503179" Oct 14 07:03:31 crc kubenswrapper[4870]: I1014 07:03:31.361977 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-5ptfj"] Oct 14 07:03:31 crc kubenswrapper[4870]: I1014 07:03:31.363041 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-9lgs8" podStartSLOduration=126.363015617 podStartE2EDuration="2m6.363015617s" podCreationTimestamp="2025-10-14 07:01:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:03:31.362358671 +0000 UTC m=+147.059719042" watchObservedRunningTime="2025-10-14 07:03:31.363015617 +0000 UTC m=+147.060375988" Oct 14 07:03:31 crc kubenswrapper[4870]: I1014 07:03:31.363173 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 07:03:31 crc kubenswrapper[4870]: E1014 07:03:31.363681 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 07:03:31.863661653 +0000 UTC m=+147.561022024 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:03:31 crc kubenswrapper[4870]: I1014 07:03:31.407498 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-5dsrq" podStartSLOduration=125.407482746 podStartE2EDuration="2m5.407482746s" podCreationTimestamp="2025-10-14 07:01:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:03:31.404418842 +0000 UTC m=+147.101779213" watchObservedRunningTime="2025-10-14 07:03:31.407482746 +0000 UTC m=+147.104843117" Oct 14 07:03:31 crc kubenswrapper[4870]: I1014 07:03:31.465857 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kkxjz\" (UID: \"2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3\") " pod="openshift-image-registry/image-registry-697d97f7c8-kkxjz" Oct 14 07:03:31 crc kubenswrapper[4870]: E1014 07:03:31.466221 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-14 07:03:31.96620943 +0000 UTC m=+147.663569801 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kkxjz" (UID: "2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:03:31 crc kubenswrapper[4870]: I1014 07:03:31.528638 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rnf2p" event={"ID":"1d891345-93bd-4726-905f-8a946a4a885f","Type":"ContainerStarted","Data":"725a1d4ee709a05bb75ad19ea98241e0785fd9fe9e05fb7738244edc48b89d53"} Oct 14 07:03:31 crc kubenswrapper[4870]: I1014 07:03:31.535673 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29340420-t6cgj" event={"ID":"737f304d-12ac-4d4b-97fb-e99506dd08a8","Type":"ContainerStarted","Data":"153066511bbafa672f0b023504ef494ec1e5bead898e6bb97e8b93cfb16c221a"} Oct 14 07:03:31 crc kubenswrapper[4870]: I1014 07:03:31.537405 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-q26kj" event={"ID":"30ec1e91-85db-4abb-b3dc-19cdcc8d4234","Type":"ContainerStarted","Data":"f04eaa8694a551e4c63e09af31811e127046454ffcc6be81382e3b1cfc8f7a5b"} Oct 14 07:03:31 crc kubenswrapper[4870]: I1014 07:03:31.578610 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-wkc9x" event={"ID":"0a83e6b8-c65a-42ef-b4e3-8748503bcd84","Type":"ContainerStarted","Data":"95ec28a1f25f2ed617062ad87838bcf47ea80d61a9eb54c1f989e4977302a481"} Oct 14 07:03:31 crc kubenswrapper[4870]: I1014 07:03:31.579603 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 07:03:31 crc kubenswrapper[4870]: E1014 07:03:31.580089 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 07:03:32.080075072 +0000 UTC m=+147.777435563 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:03:31 crc kubenswrapper[4870]: I1014 07:03:31.587883 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-6z4z4" event={"ID":"be735c57-60b4-44cb-abf3-888cf00d6c31","Type":"ContainerStarted","Data":"3a305b276b8cac04a8146a156de2595d9dc94a309f420d415f9b59f4a4ad398e"} Oct 14 07:03:31 crc kubenswrapper[4870]: I1014 07:03:31.609042 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-d4fhp" event={"ID":"9057841b-b3a4-49f2-85a6-812074c635e6","Type":"ContainerStarted","Data":"d2fa0d9c8e25c3d7ee8e84539ab5aec106ff8b30d8347f1af5a72fcb50fb3450"} Oct 14 07:03:31 crc kubenswrapper[4870]: I1014 07:03:31.645463 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vc8ks" event={"ID":"5848d6ea-90e8-4c07-92e0-8849b5b32c67","Type":"ContainerStarted","Data":"7d44c8bce478e3dfc8110576fff6bcbd76cdaa98fca65d556f92dc4181c60157"} Oct 14 07:03:31 crc kubenswrapper[4870]: I1014 07:03:31.650268 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-lbq8g" event={"ID":"a3280409-7095-4407-aa58-04be41daabf7","Type":"ContainerStarted","Data":"516a8be0fd4f71255df4a8c969f7f6e5b37660a51bc2adb8ad63c30dfee4adbc"} Oct 14 07:03:31 crc kubenswrapper[4870]: I1014 07:03:31.654602 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-mlqqj" event={"ID":"7562fe32-ac93-483f-b017-3eab4d7f94d4","Type":"ContainerStarted","Data":"dafd5010890540ccc37181405e1f0df5eee0c233a4b089ca7fada071f80f0a0b"} Oct 14 07:03:31 crc kubenswrapper[4870]: I1014 07:03:31.656090 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-6vdcc" event={"ID":"8f610aec-a4a3-4d54-b805-4f01c0de5dce","Type":"ContainerStarted","Data":"02eb8cf0763979f3b30fe30d0a2a270da81a8afc5b0b2157500f6c0ea557c208"} Oct 14 07:03:31 crc kubenswrapper[4870]: I1014 07:03:31.677980 4870 generic.go:334] "Generic (PLEG): container finished" podID="146160a7-2b1c-43f4-bc42-7b92d9663dcc" containerID="844103a6d60382d7ad285bf40560941896b5ac9cc3514da2d45bae8c6182d8b1" exitCode=0 Oct 14 07:03:31 crc kubenswrapper[4870]: I1014 07:03:31.678290 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-p769c" event={"ID":"146160a7-2b1c-43f4-bc42-7b92d9663dcc","Type":"ContainerDied","Data":"844103a6d60382d7ad285bf40560941896b5ac9cc3514da2d45bae8c6182d8b1"} Oct 14 07:03:31 crc kubenswrapper[4870]: I1014 07:03:31.681380 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kkxjz\" (UID: \"2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3\") " pod="openshift-image-registry/image-registry-697d97f7c8-kkxjz" Oct 14 07:03:31 crc kubenswrapper[4870]: E1014 07:03:31.681822 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-14 07:03:32.181810399 +0000 UTC m=+147.879170770 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kkxjz" (UID: "2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:03:31 crc kubenswrapper[4870]: I1014 07:03:31.706747 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-55gqt" event={"ID":"49dc86fe-e718-4220-81d7-2ebd7fdd6c2f","Type":"ContainerStarted","Data":"0cea61399cf4397b85a9675093666e7ab73ca5de6732a1c1fb42cbdf05c0fbc2"} Oct 14 07:03:31 crc kubenswrapper[4870]: I1014 07:03:31.763869 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-h82f7" event={"ID":"28a417a5-b0b0-465a-ade0-3bf3ceb5ebdd","Type":"ContainerStarted","Data":"9c8dfce986002e6c3d16a71afe0de6c8e1c6410c3ff9b4cd771ed1c0fd772a14"} Oct 14 07:03:31 crc kubenswrapper[4870]: I1014 07:03:31.789705 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-ptjzm" event={"ID":"edd471fb-addd-407d-b7cb-55dbffd7a1ee","Type":"ContainerStarted","Data":"b21841c85ac1755d724c194835091a83feac74d0ee09211d432f73520c15a2f8"} Oct 14 07:03:31 crc kubenswrapper[4870]: I1014 07:03:31.790255 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 07:03:31 crc kubenswrapper[4870]: E1014 07:03:31.791636 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 07:03:32.291619482 +0000 UTC m=+147.988979853 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:03:31 crc kubenswrapper[4870]: I1014 07:03:31.811967 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-h82f7" podStartSLOduration=125.811951366 podStartE2EDuration="2m5.811951366s" podCreationTimestamp="2025-10-14 07:01:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:03:31.811409663 +0000 UTC m=+147.508770034" watchObservedRunningTime="2025-10-14 07:03:31.811951366 +0000 UTC m=+147.509311737" Oct 14 07:03:31 crc kubenswrapper[4870]: I1014 07:03:31.812509 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-55gqt" podStartSLOduration=125.812502469 podStartE2EDuration="2m5.812502469s" podCreationTimestamp="2025-10-14 07:01:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:03:31.772648682 +0000 UTC m=+147.470009283" watchObservedRunningTime="2025-10-14 07:03:31.812502469 +0000 UTC m=+147.509862830" Oct 14 07:03:31 crc kubenswrapper[4870]: I1014 07:03:31.845237 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-kqxpq" event={"ID":"a0207141-7032-4930-a837-4736469c0974","Type":"ContainerStarted","Data":"889285c31850c93e05ea913a258620088dbe41335574e2fa13f2338b027b523a"} Oct 14 07:03:31 crc kubenswrapper[4870]: I1014 07:03:31.846353 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-kqxpq" Oct 14 07:03:31 crc kubenswrapper[4870]: I1014 07:03:31.847233 4870 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-kqxpq container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.21:8443/healthz\": dial tcp 10.217.0.21:8443: connect: connection refused" start-of-body= Oct 14 07:03:31 crc kubenswrapper[4870]: I1014 07:03:31.847303 4870 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-kqxpq" podUID="a0207141-7032-4930-a837-4736469c0974" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.21:8443/healthz\": dial tcp 10.217.0.21:8443: connect: connection refused" Oct 14 07:03:31 crc kubenswrapper[4870]: I1014 07:03:31.849757 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-69t6j" event={"ID":"27bd2071-dddb-48c6-905c-22bcb4e84482","Type":"ContainerStarted","Data":"96fe7903a918d14365bc512f919c888d33ce6c5d7c5701e9e3f54d6a67d26212"} Oct 14 07:03:31 crc kubenswrapper[4870]: I1014 07:03:31.857258 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-5ptfj" event={"ID":"8432ae60-6e63-45cf-b21b-ab1010687654","Type":"ContainerStarted","Data":"7b29bc1a28168e76b88b021ef4b270c185a801eb236d1230401fe0d6c8fea235"} Oct 14 07:03:31 crc kubenswrapper[4870]: I1014 07:03:31.866542 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-d7cdf" event={"ID":"2a72b0a0-287b-42dc-8c39-1976188955de","Type":"ContainerStarted","Data":"8b5b05535a261f74ade1acc6e350a36b71c46f6d5ec90ff41302955f92e2195f"} Oct 14 07:03:31 crc kubenswrapper[4870]: I1014 07:03:31.873332 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-cvh5p" event={"ID":"337df1ca-c52e-499c-ab1d-a32aee9c9922","Type":"ContainerStarted","Data":"5b05aece76a9da1095075bc0e26999116c4581000a919fbc4e9d05a852960061"} Oct 14 07:03:31 crc kubenswrapper[4870]: I1014 07:03:31.876726 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-kqxpq" podStartSLOduration=125.876690936 podStartE2EDuration="2m5.876690936s" podCreationTimestamp="2025-10-14 07:01:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:03:31.872988386 +0000 UTC m=+147.570348757" watchObservedRunningTime="2025-10-14 07:03:31.876690936 +0000 UTC m=+147.574051307" Oct 14 07:03:31 crc kubenswrapper[4870]: I1014 07:03:31.883261 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-6ztwm" event={"ID":"bb92d614-7f4e-470e-a68d-0aff68d3894d","Type":"ContainerStarted","Data":"57bff9e4a40083ec1011c1e2225894fc355a67b95f973c41dbf99fdc5aaeb99d"} Oct 14 07:03:31 crc kubenswrapper[4870]: I1014 07:03:31.891720 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kkxjz\" (UID: \"2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3\") " pod="openshift-image-registry/image-registry-697d97f7c8-kkxjz" Oct 14 07:03:31 crc kubenswrapper[4870]: E1014 07:03:31.893828 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-14 07:03:32.393812531 +0000 UTC m=+148.091172902 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kkxjz" (UID: "2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:03:31 crc kubenswrapper[4870]: I1014 07:03:31.894712 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-7jz27" event={"ID":"b419b072-5f0d-4103-82c6-c296f71eee8a","Type":"ContainerStarted","Data":"85f6abb813dc9e9b36519bc932a6c44f0fd1ab710d7df37a737f3b0870b0b8c3"} Oct 14 07:03:31 crc kubenswrapper[4870]: I1014 07:03:31.895425 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-7jz27" Oct 14 07:03:31 crc kubenswrapper[4870]: I1014 07:03:31.896631 4870 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-7jz27 container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.27:8443/healthz\": dial tcp 10.217.0.27:8443: connect: connection refused" start-of-body= Oct 14 07:03:31 crc kubenswrapper[4870]: I1014 07:03:31.896688 4870 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-7jz27" podUID="b419b072-5f0d-4103-82c6-c296f71eee8a" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.27:8443/healthz\": dial tcp 10.217.0.27:8443: connect: connection refused" Oct 14 07:03:31 crc kubenswrapper[4870]: I1014 07:03:31.898476 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-52qbc" event={"ID":"63176391-0acc-4316-a0fe-520669eb1744","Type":"ContainerStarted","Data":"b8576612fe9eea06a0c2a6415a0f5419d08a3c4f684f2e85b994025005a16584"} Oct 14 07:03:31 crc kubenswrapper[4870]: I1014 07:03:31.905755 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-cvh5p" podStartSLOduration=126.90571298 podStartE2EDuration="2m6.90571298s" podCreationTimestamp="2025-10-14 07:01:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:03:31.903276281 +0000 UTC m=+147.600636652" watchObservedRunningTime="2025-10-14 07:03:31.90571298 +0000 UTC m=+147.603073351" Oct 14 07:03:31 crc kubenswrapper[4870]: I1014 07:03:31.927775 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-x8qtb" Oct 14 07:03:31 crc kubenswrapper[4870]: I1014 07:03:31.947572 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-7jz27" podStartSLOduration=125.947545834 podStartE2EDuration="2m5.947545834s" podCreationTimestamp="2025-10-14 07:01:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:03:31.946864998 +0000 UTC m=+147.644225369" watchObservedRunningTime="2025-10-14 07:03:31.947545834 +0000 UTC m=+147.644906205" Oct 14 07:03:31 crc kubenswrapper[4870]: I1014 07:03:31.994636 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 07:03:31 crc kubenswrapper[4870]: E1014 07:03:31.999340 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 07:03:32.499294459 +0000 UTC m=+148.196655000 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:03:32 crc kubenswrapper[4870]: I1014 07:03:32.084785 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-h82f7" Oct 14 07:03:32 crc kubenswrapper[4870]: I1014 07:03:32.112464 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kkxjz\" (UID: \"2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3\") " pod="openshift-image-registry/image-registry-697d97f7c8-kkxjz" Oct 14 07:03:32 crc kubenswrapper[4870]: I1014 07:03:32.113072 4870 patch_prober.go:28] interesting pod/router-default-5444994796-h82f7 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 14 07:03:32 crc kubenswrapper[4870]: [-]has-synced failed: reason withheld Oct 14 07:03:32 crc kubenswrapper[4870]: [+]process-running ok Oct 14 07:03:32 crc kubenswrapper[4870]: healthz check failed Oct 14 07:03:32 crc kubenswrapper[4870]: I1014 07:03:32.113138 4870 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-h82f7" podUID="28a417a5-b0b0-465a-ade0-3bf3ceb5ebdd" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 14 07:03:32 crc kubenswrapper[4870]: E1014 07:03:32.113682 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-14 07:03:32.613667593 +0000 UTC m=+148.311027954 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kkxjz" (UID: "2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:03:32 crc kubenswrapper[4870]: I1014 07:03:32.214601 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 07:03:32 crc kubenswrapper[4870]: E1014 07:03:32.214801 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 07:03:32.714767275 +0000 UTC m=+148.412127646 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:03:32 crc kubenswrapper[4870]: I1014 07:03:32.215203 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kkxjz\" (UID: \"2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3\") " pod="openshift-image-registry/image-registry-697d97f7c8-kkxjz" Oct 14 07:03:32 crc kubenswrapper[4870]: E1014 07:03:32.215651 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-14 07:03:32.715634716 +0000 UTC m=+148.412995087 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kkxjz" (UID: "2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:03:32 crc kubenswrapper[4870]: I1014 07:03:32.317486 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 07:03:32 crc kubenswrapper[4870]: E1014 07:03:32.318110 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 07:03:32.818086661 +0000 UTC m=+148.515447032 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:03:32 crc kubenswrapper[4870]: I1014 07:03:32.420649 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kkxjz\" (UID: \"2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3\") " pod="openshift-image-registry/image-registry-697d97f7c8-kkxjz" Oct 14 07:03:32 crc kubenswrapper[4870]: E1014 07:03:32.422991 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-14 07:03:32.922972285 +0000 UTC m=+148.620332656 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kkxjz" (UID: "2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:03:32 crc kubenswrapper[4870]: I1014 07:03:32.523353 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 07:03:32 crc kubenswrapper[4870]: E1014 07:03:32.523773 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 07:03:33.023730079 +0000 UTC m=+148.721090450 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:03:32 crc kubenswrapper[4870]: I1014 07:03:32.524201 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kkxjz\" (UID: \"2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3\") " pod="openshift-image-registry/image-registry-697d97f7c8-kkxjz" Oct 14 07:03:32 crc kubenswrapper[4870]: E1014 07:03:32.526108 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-14 07:03:33.026087806 +0000 UTC m=+148.723448177 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kkxjz" (UID: "2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:03:32 crc kubenswrapper[4870]: I1014 07:03:32.632426 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 07:03:32 crc kubenswrapper[4870]: E1014 07:03:32.633019 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 07:03:33.133003019 +0000 UTC m=+148.830363390 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:03:32 crc kubenswrapper[4870]: I1014 07:03:32.737746 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kkxjz\" (UID: \"2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3\") " pod="openshift-image-registry/image-registry-697d97f7c8-kkxjz" Oct 14 07:03:32 crc kubenswrapper[4870]: E1014 07:03:32.738850 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-14 07:03:33.238830736 +0000 UTC m=+148.936191107 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kkxjz" (UID: "2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:03:32 crc kubenswrapper[4870]: I1014 07:03:32.838951 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 07:03:32 crc kubenswrapper[4870]: E1014 07:03:32.839351 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 07:03:33.339333804 +0000 UTC m=+149.036694175 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:03:32 crc kubenswrapper[4870]: I1014 07:03:32.940722 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kkxjz\" (UID: \"2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3\") " pod="openshift-image-registry/image-registry-697d97f7c8-kkxjz" Oct 14 07:03:32 crc kubenswrapper[4870]: E1014 07:03:32.941006 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-14 07:03:33.440994819 +0000 UTC m=+149.138355190 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kkxjz" (UID: "2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:03:32 crc kubenswrapper[4870]: I1014 07:03:32.979294 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-szwj9" event={"ID":"8070623a-4696-490e-af90-5b2155c0506c","Type":"ContainerStarted","Data":"5fcdc22335e4482984cc528863431e9a70f2248ce965d5c8e6cf276410ff62d1"} Oct 14 07:03:33 crc kubenswrapper[4870]: I1014 07:03:33.018820 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-n8z6k" event={"ID":"03d1e16b-e364-4902-9412-0dc138fc76e3","Type":"ContainerStarted","Data":"faa95e0b0342e39f04972dfcad44536e9bb67ccea3c9227bf962a42ea512b5cd"} Oct 14 07:03:33 crc kubenswrapper[4870]: I1014 07:03:33.026622 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-52qbc" event={"ID":"63176391-0acc-4316-a0fe-520669eb1744","Type":"ContainerStarted","Data":"140197042d5f7802bb09a6b212cde482ab87f97cada5650eb2681b497690e4ed"} Oct 14 07:03:33 crc kubenswrapper[4870]: I1014 07:03:33.042130 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 07:03:33 crc kubenswrapper[4870]: E1014 07:03:33.042652 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 07:03:33.542631234 +0000 UTC m=+149.239991605 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:03:33 crc kubenswrapper[4870]: I1014 07:03:33.051429 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-mlqqj" event={"ID":"7562fe32-ac93-483f-b017-3eab4d7f94d4","Type":"ContainerStarted","Data":"f00c9a88430d93c66ed95575406132e56ebce8086d5868151d9d0b5a166a1b97"} Oct 14 07:03:33 crc kubenswrapper[4870]: I1014 07:03:33.069520 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-wkc9x" event={"ID":"0a83e6b8-c65a-42ef-b4e3-8748503bcd84","Type":"ContainerStarted","Data":"6218ddaa5bc756ef7a0d1957e5165e869d53b472f7c4cfdf983a35a0afe418d4"} Oct 14 07:03:33 crc kubenswrapper[4870]: I1014 07:03:33.090866 4870 patch_prober.go:28] interesting pod/router-default-5444994796-h82f7 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 14 07:03:33 crc kubenswrapper[4870]: [-]has-synced failed: reason withheld Oct 14 07:03:33 crc kubenswrapper[4870]: [+]process-running ok Oct 14 07:03:33 crc kubenswrapper[4870]: healthz check failed Oct 14 07:03:33 crc kubenswrapper[4870]: I1014 07:03:33.090967 4870 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-h82f7" podUID="28a417a5-b0b0-465a-ade0-3bf3ceb5ebdd" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 14 07:03:33 crc kubenswrapper[4870]: I1014 07:03:33.098031 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-klh8n" event={"ID":"1df54538-19c2-472b-b93e-b7f947190db0","Type":"ContainerStarted","Data":"a102394d675a0bd9cccd10592418f19f1c49d249c728ba62bc6ffe9accbb5490"} Oct 14 07:03:33 crc kubenswrapper[4870]: I1014 07:03:33.114383 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-p769c" Oct 14 07:03:33 crc kubenswrapper[4870]: I1014 07:03:33.128657 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-l24lg" event={"ID":"c0cd54e9-09ac-49a6-a774-a9148024b4b9","Type":"ContainerStarted","Data":"580d99f3969074fe8717569be49947d7393b9574cd3a18abc5be6a5c56b1e1e0"} Oct 14 07:03:33 crc kubenswrapper[4870]: I1014 07:03:33.139090 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-d7cdf" event={"ID":"2a72b0a0-287b-42dc-8c39-1976188955de","Type":"ContainerStarted","Data":"6e6e4d659e926961d64c80f9a53371a3314a0272c5fb837690fb4e5120701f93"} Oct 14 07:03:33 crc kubenswrapper[4870]: I1014 07:03:33.142290 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-d7cdf" Oct 14 07:03:33 crc kubenswrapper[4870]: I1014 07:03:33.143873 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-wkc9x" podStartSLOduration=127.143859009 podStartE2EDuration="2m7.143859009s" podCreationTimestamp="2025-10-14 07:01:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:03:33.110818108 +0000 UTC m=+148.808178479" watchObservedRunningTime="2025-10-14 07:03:33.143859009 +0000 UTC m=+148.841219380" Oct 14 07:03:33 crc kubenswrapper[4870]: I1014 07:03:33.144503 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kkxjz\" (UID: \"2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3\") " pod="openshift-image-registry/image-registry-697d97f7c8-kkxjz" Oct 14 07:03:33 crc kubenswrapper[4870]: E1014 07:03:33.144777 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-14 07:03:33.644764031 +0000 UTC m=+149.342124402 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kkxjz" (UID: "2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:03:33 crc kubenswrapper[4870]: I1014 07:03:33.154760 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-klh8n" podStartSLOduration=7.154736493 podStartE2EDuration="7.154736493s" podCreationTimestamp="2025-10-14 07:03:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:03:33.142964198 +0000 UTC m=+148.840324569" watchObservedRunningTime="2025-10-14 07:03:33.154736493 +0000 UTC m=+148.852096864" Oct 14 07:03:33 crc kubenswrapper[4870]: I1014 07:03:33.160306 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-q9knv" Oct 14 07:03:33 crc kubenswrapper[4870]: I1014 07:03:33.167415 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-q9knv" Oct 14 07:03:33 crc kubenswrapper[4870]: I1014 07:03:33.171704 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ss574" event={"ID":"d9cb408e-a00b-4f76-b8fb-bf144bfee740","Type":"ContainerStarted","Data":"a5f559965df24782040e3bc4b4d03dd793ba237a9f37ccb21148f03fe3a86c30"} Oct 14 07:03:33 crc kubenswrapper[4870]: I1014 07:03:33.173934 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-g7crn" event={"ID":"16295052-1474-4dc9-b1d6-10983e6bfc6f","Type":"ContainerStarted","Data":"75b16fd03cca8a9519f2c844c02a6143328dbb6804d906e688987d7be8ddd04d"} Oct 14 07:03:33 crc kubenswrapper[4870]: I1014 07:03:33.200985 4870 patch_prober.go:28] interesting pod/downloads-7954f5f757-d7cdf container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.15:8080/\": dial tcp 10.217.0.15:8080: connect: connection refused" start-of-body= Oct 14 07:03:33 crc kubenswrapper[4870]: I1014 07:03:33.201040 4870 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-d7cdf" podUID="2a72b0a0-287b-42dc-8c39-1976188955de" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.15:8080/\": dial tcp 10.217.0.15:8080: connect: connection refused" Oct 14 07:03:33 crc kubenswrapper[4870]: I1014 07:03:33.216756 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-zpgbg" event={"ID":"222f1962-d6c1-4e69-a3bc-0b0a12703cff","Type":"ContainerStarted","Data":"46bd5bac24dcc8f11a7332048506db555080a49a85ff7ad1fa0152abc2261557"} Oct 14 07:03:33 crc kubenswrapper[4870]: I1014 07:03:33.230844 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-l24lg" podStartSLOduration=127.230824419 podStartE2EDuration="2m7.230824419s" podCreationTimestamp="2025-10-14 07:01:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:03:33.225859768 +0000 UTC m=+148.923220139" watchObservedRunningTime="2025-10-14 07:03:33.230824419 +0000 UTC m=+148.928184790" Oct 14 07:03:33 crc kubenswrapper[4870]: I1014 07:03:33.232694 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-q9knv" Oct 14 07:03:33 crc kubenswrapper[4870]: I1014 07:03:33.238891 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29340420-t6cgj" event={"ID":"737f304d-12ac-4d4b-97fb-e99506dd08a8","Type":"ContainerStarted","Data":"54dde5dec3d53d9948e6ae0d3f700606e677c020c9d7e4bd7179a25688038920"} Oct 14 07:03:33 crc kubenswrapper[4870]: I1014 07:03:33.246199 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 07:03:33 crc kubenswrapper[4870]: E1014 07:03:33.247659 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 07:03:33.747634696 +0000 UTC m=+149.444995067 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:03:33 crc kubenswrapper[4870]: I1014 07:03:33.274067 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-d7cdf" podStartSLOduration=127.274049297 podStartE2EDuration="2m7.274049297s" podCreationTimestamp="2025-10-14 07:01:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:03:33.271337231 +0000 UTC m=+148.968697602" watchObservedRunningTime="2025-10-14 07:03:33.274049297 +0000 UTC m=+148.971409668" Oct 14 07:03:33 crc kubenswrapper[4870]: I1014 07:03:33.279915 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-hwfsm" event={"ID":"7b7a521b-c795-43cc-9576-5a3be3cb7409","Type":"ContainerStarted","Data":"78ee71ad26e4d5fa75a48374bf4d62d18694451d6a74dd956bfd2c62bf31b419"} Oct 14 07:03:33 crc kubenswrapper[4870]: I1014 07:03:33.357863 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kkxjz\" (UID: \"2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3\") " pod="openshift-image-registry/image-registry-697d97f7c8-kkxjz" Oct 14 07:03:33 crc kubenswrapper[4870]: E1014 07:03:33.358833 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-14 07:03:33.858818123 +0000 UTC m=+149.556178494 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kkxjz" (UID: "2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:03:33 crc kubenswrapper[4870]: I1014 07:03:33.363915 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-p769c" podStartSLOduration=127.363884346 podStartE2EDuration="2m7.363884346s" podCreationTimestamp="2025-10-14 07:01:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:03:33.359132531 +0000 UTC m=+149.056492902" watchObservedRunningTime="2025-10-14 07:03:33.363884346 +0000 UTC m=+149.061244717" Oct 14 07:03:33 crc kubenswrapper[4870]: I1014 07:03:33.406126 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ss574" podStartSLOduration=127.40610006 podStartE2EDuration="2m7.40610006s" podCreationTimestamp="2025-10-14 07:01:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:03:33.400065083 +0000 UTC m=+149.097425444" watchObservedRunningTime="2025-10-14 07:03:33.40610006 +0000 UTC m=+149.103460431" Oct 14 07:03:33 crc kubenswrapper[4870]: I1014 07:03:33.444221 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vc8ks" podStartSLOduration=127.444203654 podStartE2EDuration="2m7.444203654s" podCreationTimestamp="2025-10-14 07:01:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:03:33.441616641 +0000 UTC m=+149.138977012" watchObservedRunningTime="2025-10-14 07:03:33.444203654 +0000 UTC m=+149.141564025" Oct 14 07:03:33 crc kubenswrapper[4870]: I1014 07:03:33.448904 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-6ztwm" Oct 14 07:03:33 crc kubenswrapper[4870]: I1014 07:03:33.458832 4870 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-6ztwm container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.28:8080/healthz\": dial tcp 10.217.0.28:8080: connect: connection refused" start-of-body= Oct 14 07:03:33 crc kubenswrapper[4870]: I1014 07:03:33.495212 4870 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-6ztwm" podUID="bb92d614-7f4e-470e-a68d-0aff68d3894d" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.28:8080/healthz\": dial tcp 10.217.0.28:8080: connect: connection refused" Oct 14 07:03:33 crc kubenswrapper[4870]: E1014 07:03:33.460884 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 07:03:33.960842017 +0000 UTC m=+149.658202388 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:03:33 crc kubenswrapper[4870]: I1014 07:03:33.460275 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 07:03:33 crc kubenswrapper[4870]: I1014 07:03:33.496020 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kkxjz\" (UID: \"2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3\") " pod="openshift-image-registry/image-registry-697d97f7c8-kkxjz" Oct 14 07:03:33 crc kubenswrapper[4870]: I1014 07:03:33.496362 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-g7crn" podStartSLOduration=127.496351409 podStartE2EDuration="2m7.496351409s" podCreationTimestamp="2025-10-14 07:01:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:03:33.495493678 +0000 UTC m=+149.192854049" watchObservedRunningTime="2025-10-14 07:03:33.496351409 +0000 UTC m=+149.193711780" Oct 14 07:03:33 crc kubenswrapper[4870]: E1014 07:03:33.498010 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-14 07:03:33.997985838 +0000 UTC m=+149.695346209 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kkxjz" (UID: "2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:03:33 crc kubenswrapper[4870]: I1014 07:03:33.520957 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wjm9z" event={"ID":"f07c76fa-f754-4175-8bc5-2aae95985b3b","Type":"ContainerStarted","Data":"4a913f54dc1a9dde9a3657447201a254baaba278204b4e4d1a33e3cf4913281a"} Oct 14 07:03:33 crc kubenswrapper[4870]: I1014 07:03:33.521751 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wjm9z" Oct 14 07:03:33 crc kubenswrapper[4870]: I1014 07:03:33.552873 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-lbq8g" event={"ID":"a3280409-7095-4407-aa58-04be41daabf7","Type":"ContainerStarted","Data":"bdda8583c7935e643651871ac6ae65f9f64c91faf1cf6b2a8d0257337dd3dd71"} Oct 14 07:03:33 crc kubenswrapper[4870]: I1014 07:03:33.553817 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wjm9z" Oct 14 07:03:33 crc kubenswrapper[4870]: I1014 07:03:33.597712 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-q26kj" event={"ID":"30ec1e91-85db-4abb-b3dc-19cdcc8d4234","Type":"ContainerStarted","Data":"c3cd24691a6517280ea2939bf5b52061af7fd00606df29e633178adad7f8d19f"} Oct 14 07:03:33 crc kubenswrapper[4870]: I1014 07:03:33.599548 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 07:03:33 crc kubenswrapper[4870]: E1014 07:03:33.600418 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 07:03:34.100404812 +0000 UTC m=+149.797765183 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:03:33 crc kubenswrapper[4870]: I1014 07:03:33.622640 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-69t6j" event={"ID":"27bd2071-dddb-48c6-905c-22bcb4e84482","Type":"ContainerStarted","Data":"3058fe62f54dc2b7d0187ec6b92d3a4ebe1286fcac58f7255b90fa926e8b8f50"} Oct 14 07:03:33 crc kubenswrapper[4870]: I1014 07:03:33.648831 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-kqxpq" Oct 14 07:03:33 crc kubenswrapper[4870]: I1014 07:03:33.652588 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-7jz27" Oct 14 07:03:33 crc kubenswrapper[4870]: I1014 07:03:33.658494 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-q9knv" Oct 14 07:03:33 crc kubenswrapper[4870]: I1014 07:03:33.701259 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-hwfsm" podStartSLOduration=127.701241248 podStartE2EDuration="2m7.701241248s" podCreationTimestamp="2025-10-14 07:01:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:03:33.699370463 +0000 UTC m=+149.396730834" watchObservedRunningTime="2025-10-14 07:03:33.701241248 +0000 UTC m=+149.398601619" Oct 14 07:03:33 crc kubenswrapper[4870]: I1014 07:03:33.702906 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kkxjz\" (UID: \"2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3\") " pod="openshift-image-registry/image-registry-697d97f7c8-kkxjz" Oct 14 07:03:33 crc kubenswrapper[4870]: E1014 07:03:33.726595 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-14 07:03:34.226540632 +0000 UTC m=+149.923901003 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kkxjz" (UID: "2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:03:33 crc kubenswrapper[4870]: I1014 07:03:33.746105 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29340420-t6cgj" podStartSLOduration=127.746081115 podStartE2EDuration="2m7.746081115s" podCreationTimestamp="2025-10-14 07:01:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:03:33.743963444 +0000 UTC m=+149.441323815" watchObservedRunningTime="2025-10-14 07:03:33.746081115 +0000 UTC m=+149.443441486" Oct 14 07:03:33 crc kubenswrapper[4870]: I1014 07:03:33.778095 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-q26kj" podStartSLOduration=7.778077421 podStartE2EDuration="7.778077421s" podCreationTimestamp="2025-10-14 07:03:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:03:33.777135019 +0000 UTC m=+149.474495390" watchObservedRunningTime="2025-10-14 07:03:33.778077421 +0000 UTC m=+149.475437792" Oct 14 07:03:33 crc kubenswrapper[4870]: I1014 07:03:33.812648 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 07:03:33 crc kubenswrapper[4870]: E1014 07:03:33.813067 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 07:03:34.31304972 +0000 UTC m=+150.010410091 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:03:33 crc kubenswrapper[4870]: I1014 07:03:33.888005 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wjm9z" podStartSLOduration=127.887983227 podStartE2EDuration="2m7.887983227s" podCreationTimestamp="2025-10-14 07:01:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:03:33.843605371 +0000 UTC m=+149.540965752" watchObservedRunningTime="2025-10-14 07:03:33.887983227 +0000 UTC m=+149.585343598" Oct 14 07:03:33 crc kubenswrapper[4870]: I1014 07:03:33.914885 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kkxjz\" (UID: \"2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3\") " pod="openshift-image-registry/image-registry-697d97f7c8-kkxjz" Oct 14 07:03:33 crc kubenswrapper[4870]: E1014 07:03:33.915256 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-14 07:03:34.415244658 +0000 UTC m=+150.112605029 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kkxjz" (UID: "2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:03:34 crc kubenswrapper[4870]: I1014 07:03:34.018973 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 07:03:34 crc kubenswrapper[4870]: E1014 07:03:34.019388 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 07:03:34.519372514 +0000 UTC m=+150.216732885 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:03:34 crc kubenswrapper[4870]: I1014 07:03:34.022507 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-6ztwm" podStartSLOduration=128.02249635 podStartE2EDuration="2m8.02249635s" podCreationTimestamp="2025-10-14 07:01:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:03:34.021892835 +0000 UTC m=+149.719253206" watchObservedRunningTime="2025-10-14 07:03:34.02249635 +0000 UTC m=+149.719856721" Oct 14 07:03:34 crc kubenswrapper[4870]: I1014 07:03:34.022918 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-lbq8g" podStartSLOduration=128.02291198 podStartE2EDuration="2m8.02291198s" podCreationTimestamp="2025-10-14 07:01:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:03:33.975770916 +0000 UTC m=+149.673131307" watchObservedRunningTime="2025-10-14 07:03:34.02291198 +0000 UTC m=+149.720272351" Oct 14 07:03:34 crc kubenswrapper[4870]: I1014 07:03:34.051166 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-69t6j" podStartSLOduration=128.051148104 podStartE2EDuration="2m8.051148104s" podCreationTimestamp="2025-10-14 07:01:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:03:34.050219182 +0000 UTC m=+149.747579553" watchObservedRunningTime="2025-10-14 07:03:34.051148104 +0000 UTC m=+149.748508475" Oct 14 07:03:34 crc kubenswrapper[4870]: I1014 07:03:34.089425 4870 patch_prober.go:28] interesting pod/router-default-5444994796-h82f7 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 14 07:03:34 crc kubenswrapper[4870]: [-]has-synced failed: reason withheld Oct 14 07:03:34 crc kubenswrapper[4870]: [+]process-running ok Oct 14 07:03:34 crc kubenswrapper[4870]: healthz check failed Oct 14 07:03:34 crc kubenswrapper[4870]: I1014 07:03:34.089521 4870 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-h82f7" podUID="28a417a5-b0b0-465a-ade0-3bf3ceb5ebdd" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 14 07:03:34 crc kubenswrapper[4870]: I1014 07:03:34.120123 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 07:03:34 crc kubenswrapper[4870]: I1014 07:03:34.120208 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 07:03:34 crc kubenswrapper[4870]: I1014 07:03:34.120231 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 07:03:34 crc kubenswrapper[4870]: I1014 07:03:34.120271 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 07:03:34 crc kubenswrapper[4870]: I1014 07:03:34.120296 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kkxjz\" (UID: \"2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3\") " pod="openshift-image-registry/image-registry-697d97f7c8-kkxjz" Oct 14 07:03:34 crc kubenswrapper[4870]: E1014 07:03:34.120602 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-14 07:03:34.620590569 +0000 UTC m=+150.317950940 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kkxjz" (UID: "2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:03:34 crc kubenswrapper[4870]: I1014 07:03:34.121370 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 07:03:34 crc kubenswrapper[4870]: I1014 07:03:34.130724 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 07:03:34 crc kubenswrapper[4870]: I1014 07:03:34.131158 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 07:03:34 crc kubenswrapper[4870]: I1014 07:03:34.157820 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 07:03:34 crc kubenswrapper[4870]: I1014 07:03:34.161167 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-jhjmr"] Oct 14 07:03:34 crc kubenswrapper[4870]: I1014 07:03:34.162309 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jhjmr" Oct 14 07:03:34 crc kubenswrapper[4870]: I1014 07:03:34.169906 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Oct 14 07:03:34 crc kubenswrapper[4870]: I1014 07:03:34.179275 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-jhjmr"] Oct 14 07:03:34 crc kubenswrapper[4870]: I1014 07:03:34.221970 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 07:03:34 crc kubenswrapper[4870]: I1014 07:03:34.222462 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/44734ad6-0693-4d02-ba13-441850b5d23b-utilities\") pod \"community-operators-jhjmr\" (UID: \"44734ad6-0693-4d02-ba13-441850b5d23b\") " pod="openshift-marketplace/community-operators-jhjmr" Oct 14 07:03:34 crc kubenswrapper[4870]: I1014 07:03:34.222544 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pcqbb\" (UniqueName: \"kubernetes.io/projected/44734ad6-0693-4d02-ba13-441850b5d23b-kube-api-access-pcqbb\") pod \"community-operators-jhjmr\" (UID: \"44734ad6-0693-4d02-ba13-441850b5d23b\") " pod="openshift-marketplace/community-operators-jhjmr" Oct 14 07:03:34 crc kubenswrapper[4870]: I1014 07:03:34.222587 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/44734ad6-0693-4d02-ba13-441850b5d23b-catalog-content\") pod \"community-operators-jhjmr\" (UID: \"44734ad6-0693-4d02-ba13-441850b5d23b\") " pod="openshift-marketplace/community-operators-jhjmr" Oct 14 07:03:34 crc kubenswrapper[4870]: E1014 07:03:34.222816 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 07:03:34.722791157 +0000 UTC m=+150.420151528 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:03:34 crc kubenswrapper[4870]: I1014 07:03:34.305475 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 07:03:34 crc kubenswrapper[4870]: I1014 07:03:34.315807 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 14 07:03:34 crc kubenswrapper[4870]: I1014 07:03:34.324507 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/44734ad6-0693-4d02-ba13-441850b5d23b-catalog-content\") pod \"community-operators-jhjmr\" (UID: \"44734ad6-0693-4d02-ba13-441850b5d23b\") " pod="openshift-marketplace/community-operators-jhjmr" Oct 14 07:03:34 crc kubenswrapper[4870]: I1014 07:03:34.324610 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kkxjz\" (UID: \"2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3\") " pod="openshift-image-registry/image-registry-697d97f7c8-kkxjz" Oct 14 07:03:34 crc kubenswrapper[4870]: E1014 07:03:34.325035 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-14 07:03:34.825018747 +0000 UTC m=+150.522379118 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kkxjz" (UID: "2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:03:34 crc kubenswrapper[4870]: I1014 07:03:34.325201 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/44734ad6-0693-4d02-ba13-441850b5d23b-catalog-content\") pod \"community-operators-jhjmr\" (UID: \"44734ad6-0693-4d02-ba13-441850b5d23b\") " pod="openshift-marketplace/community-operators-jhjmr" Oct 14 07:03:34 crc kubenswrapper[4870]: I1014 07:03:34.325252 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/44734ad6-0693-4d02-ba13-441850b5d23b-utilities\") pod \"community-operators-jhjmr\" (UID: \"44734ad6-0693-4d02-ba13-441850b5d23b\") " pod="openshift-marketplace/community-operators-jhjmr" Oct 14 07:03:34 crc kubenswrapper[4870]: I1014 07:03:34.325419 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pcqbb\" (UniqueName: \"kubernetes.io/projected/44734ad6-0693-4d02-ba13-441850b5d23b-kube-api-access-pcqbb\") pod \"community-operators-jhjmr\" (UID: \"44734ad6-0693-4d02-ba13-441850b5d23b\") " pod="openshift-marketplace/community-operators-jhjmr" Oct 14 07:03:34 crc kubenswrapper[4870]: I1014 07:03:34.325950 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/44734ad6-0693-4d02-ba13-441850b5d23b-utilities\") pod \"community-operators-jhjmr\" (UID: \"44734ad6-0693-4d02-ba13-441850b5d23b\") " pod="openshift-marketplace/community-operators-jhjmr" Oct 14 07:03:34 crc kubenswrapper[4870]: I1014 07:03:34.331016 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 14 07:03:34 crc kubenswrapper[4870]: I1014 07:03:34.347087 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pcqbb\" (UniqueName: \"kubernetes.io/projected/44734ad6-0693-4d02-ba13-441850b5d23b-kube-api-access-pcqbb\") pod \"community-operators-jhjmr\" (UID: \"44734ad6-0693-4d02-ba13-441850b5d23b\") " pod="openshift-marketplace/community-operators-jhjmr" Oct 14 07:03:34 crc kubenswrapper[4870]: I1014 07:03:34.371096 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-2glwc"] Oct 14 07:03:34 crc kubenswrapper[4870]: I1014 07:03:34.386739 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2glwc" Oct 14 07:03:34 crc kubenswrapper[4870]: I1014 07:03:34.400277 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Oct 14 07:03:34 crc kubenswrapper[4870]: I1014 07:03:34.420314 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-2glwc"] Oct 14 07:03:34 crc kubenswrapper[4870]: I1014 07:03:34.426543 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 07:03:34 crc kubenswrapper[4870]: I1014 07:03:34.426926 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/788a3634-1dce-4c12-979b-235c74c9200b-catalog-content\") pod \"certified-operators-2glwc\" (UID: \"788a3634-1dce-4c12-979b-235c74c9200b\") " pod="openshift-marketplace/certified-operators-2glwc" Oct 14 07:03:34 crc kubenswrapper[4870]: I1014 07:03:34.427022 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k44m6\" (UniqueName: \"kubernetes.io/projected/788a3634-1dce-4c12-979b-235c74c9200b-kube-api-access-k44m6\") pod \"certified-operators-2glwc\" (UID: \"788a3634-1dce-4c12-979b-235c74c9200b\") " pod="openshift-marketplace/certified-operators-2glwc" Oct 14 07:03:34 crc kubenswrapper[4870]: I1014 07:03:34.427114 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/788a3634-1dce-4c12-979b-235c74c9200b-utilities\") pod \"certified-operators-2glwc\" (UID: \"788a3634-1dce-4c12-979b-235c74c9200b\") " pod="openshift-marketplace/certified-operators-2glwc" Oct 14 07:03:34 crc kubenswrapper[4870]: E1014 07:03:34.427377 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 07:03:34.927354669 +0000 UTC m=+150.624715040 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:03:34 crc kubenswrapper[4870]: I1014 07:03:34.511412 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jhjmr" Oct 14 07:03:34 crc kubenswrapper[4870]: I1014 07:03:34.530554 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kkxjz\" (UID: \"2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3\") " pod="openshift-image-registry/image-registry-697d97f7c8-kkxjz" Oct 14 07:03:34 crc kubenswrapper[4870]: I1014 07:03:34.530604 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/788a3634-1dce-4c12-979b-235c74c9200b-utilities\") pod \"certified-operators-2glwc\" (UID: \"788a3634-1dce-4c12-979b-235c74c9200b\") " pod="openshift-marketplace/certified-operators-2glwc" Oct 14 07:03:34 crc kubenswrapper[4870]: I1014 07:03:34.530645 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/788a3634-1dce-4c12-979b-235c74c9200b-catalog-content\") pod \"certified-operators-2glwc\" (UID: \"788a3634-1dce-4c12-979b-235c74c9200b\") " pod="openshift-marketplace/certified-operators-2glwc" Oct 14 07:03:34 crc kubenswrapper[4870]: I1014 07:03:34.530717 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k44m6\" (UniqueName: \"kubernetes.io/projected/788a3634-1dce-4c12-979b-235c74c9200b-kube-api-access-k44m6\") pod \"certified-operators-2glwc\" (UID: \"788a3634-1dce-4c12-979b-235c74c9200b\") " pod="openshift-marketplace/certified-operators-2glwc" Oct 14 07:03:34 crc kubenswrapper[4870]: E1014 07:03:34.531560 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-14 07:03:35.031543826 +0000 UTC m=+150.728904187 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kkxjz" (UID: "2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:03:34 crc kubenswrapper[4870]: I1014 07:03:34.532096 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/788a3634-1dce-4c12-979b-235c74c9200b-utilities\") pod \"certified-operators-2glwc\" (UID: \"788a3634-1dce-4c12-979b-235c74c9200b\") " pod="openshift-marketplace/certified-operators-2glwc" Oct 14 07:03:34 crc kubenswrapper[4870]: I1014 07:03:34.532391 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/788a3634-1dce-4c12-979b-235c74c9200b-catalog-content\") pod \"certified-operators-2glwc\" (UID: \"788a3634-1dce-4c12-979b-235c74c9200b\") " pod="openshift-marketplace/certified-operators-2glwc" Oct 14 07:03:34 crc kubenswrapper[4870]: I1014 07:03:34.586408 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-8qgzx"] Oct 14 07:03:34 crc kubenswrapper[4870]: I1014 07:03:34.601012 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8qgzx" Oct 14 07:03:34 crc kubenswrapper[4870]: I1014 07:03:34.594858 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k44m6\" (UniqueName: \"kubernetes.io/projected/788a3634-1dce-4c12-979b-235c74c9200b-kube-api-access-k44m6\") pod \"certified-operators-2glwc\" (UID: \"788a3634-1dce-4c12-979b-235c74c9200b\") " pod="openshift-marketplace/certified-operators-2glwc" Oct 14 07:03:34 crc kubenswrapper[4870]: I1014 07:03:34.619636 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-8qgzx"] Oct 14 07:03:34 crc kubenswrapper[4870]: I1014 07:03:34.639124 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 07:03:34 crc kubenswrapper[4870]: E1014 07:03:34.639529 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 07:03:35.139514615 +0000 UTC m=+150.836874986 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:03:34 crc kubenswrapper[4870]: I1014 07:03:34.728307 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-52qbc" event={"ID":"63176391-0acc-4316-a0fe-520669eb1744","Type":"ContainerStarted","Data":"ff8db9b02aca50fed364cf0289431a1c25a1a3bb1522e15b74e15b9561c9b576"} Oct 14 07:03:34 crc kubenswrapper[4870]: I1014 07:03:34.729365 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-52qbc" Oct 14 07:03:34 crc kubenswrapper[4870]: I1014 07:03:34.741120 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5827db23-a022-4d5f-9ae0-ef3505610bce-catalog-content\") pod \"community-operators-8qgzx\" (UID: \"5827db23-a022-4d5f-9ae0-ef3505610bce\") " pod="openshift-marketplace/community-operators-8qgzx" Oct 14 07:03:34 crc kubenswrapper[4870]: E1014 07:03:34.742732 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-14 07:03:35.242707876 +0000 UTC m=+150.940068247 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kkxjz" (UID: "2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:03:34 crc kubenswrapper[4870]: I1014 07:03:34.741173 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kkxjz\" (UID: \"2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3\") " pod="openshift-image-registry/image-registry-697d97f7c8-kkxjz" Oct 14 07:03:34 crc kubenswrapper[4870]: I1014 07:03:34.743155 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rkhl8\" (UniqueName: \"kubernetes.io/projected/5827db23-a022-4d5f-9ae0-ef3505610bce-kube-api-access-rkhl8\") pod \"community-operators-8qgzx\" (UID: \"5827db23-a022-4d5f-9ae0-ef3505610bce\") " pod="openshift-marketplace/community-operators-8qgzx" Oct 14 07:03:34 crc kubenswrapper[4870]: I1014 07:03:34.743183 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5827db23-a022-4d5f-9ae0-ef3505610bce-utilities\") pod \"community-operators-8qgzx\" (UID: \"5827db23-a022-4d5f-9ae0-ef3505610bce\") " pod="openshift-marketplace/community-operators-8qgzx" Oct 14 07:03:34 crc kubenswrapper[4870]: I1014 07:03:34.754898 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-6vdcc" event={"ID":"8f610aec-a4a3-4d54-b805-4f01c0de5dce","Type":"ContainerStarted","Data":"ffc596d6c81f53c98e5e866a6f9ac527bd234512798cc1adf3fc521e785b3a8a"} Oct 14 07:03:34 crc kubenswrapper[4870]: I1014 07:03:34.755734 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2glwc" Oct 14 07:03:34 crc kubenswrapper[4870]: I1014 07:03:34.762942 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-lkfh6"] Oct 14 07:03:34 crc kubenswrapper[4870]: I1014 07:03:34.763923 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lkfh6" Oct 14 07:03:34 crc kubenswrapper[4870]: I1014 07:03:34.788982 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-52qbc" podStartSLOduration=128.788954238 podStartE2EDuration="2m8.788954238s" podCreationTimestamp="2025-10-14 07:01:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:03:34.77376907 +0000 UTC m=+150.471129441" watchObservedRunningTime="2025-10-14 07:03:34.788954238 +0000 UTC m=+150.486314609" Oct 14 07:03:34 crc kubenswrapper[4870]: I1014 07:03:34.790246 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-lkfh6"] Oct 14 07:03:34 crc kubenswrapper[4870]: I1014 07:03:34.791914 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rnf2p" event={"ID":"1d891345-93bd-4726-905f-8a946a4a885f","Type":"ContainerStarted","Data":"7795c6ca5cda590ecc442c65b7746fa20e22cc6f1383c9151777ffd6f2cbc35c"} Oct 14 07:03:34 crc kubenswrapper[4870]: I1014 07:03:34.792971 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rnf2p" Oct 14 07:03:34 crc kubenswrapper[4870]: I1014 07:03:34.813277 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-6vdcc" podStartSLOduration=128.813257758 podStartE2EDuration="2m8.813257758s" podCreationTimestamp="2025-10-14 07:01:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:03:34.811387432 +0000 UTC m=+150.508747803" watchObservedRunningTime="2025-10-14 07:03:34.813257758 +0000 UTC m=+150.510618129" Oct 14 07:03:34 crc kubenswrapper[4870]: I1014 07:03:34.834594 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-szwj9" event={"ID":"8070623a-4696-490e-af90-5b2155c0506c","Type":"ContainerStarted","Data":"5dedcb219d0fd136aa126a43ebfdaf97c62b1b272a8341f7208719cfc6575426"} Oct 14 07:03:34 crc kubenswrapper[4870]: I1014 07:03:34.849958 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 07:03:34 crc kubenswrapper[4870]: I1014 07:03:34.850289 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5827db23-a022-4d5f-9ae0-ef3505610bce-catalog-content\") pod \"community-operators-8qgzx\" (UID: \"5827db23-a022-4d5f-9ae0-ef3505610bce\") " pod="openshift-marketplace/community-operators-8qgzx" Oct 14 07:03:34 crc kubenswrapper[4870]: I1014 07:03:34.850321 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/44a6ae4c-9030-4a51-8960-0c0cc6c59b8e-catalog-content\") pod \"certified-operators-lkfh6\" (UID: \"44a6ae4c-9030-4a51-8960-0c0cc6c59b8e\") " pod="openshift-marketplace/certified-operators-lkfh6" Oct 14 07:03:34 crc kubenswrapper[4870]: I1014 07:03:34.850399 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rkhl8\" (UniqueName: \"kubernetes.io/projected/5827db23-a022-4d5f-9ae0-ef3505610bce-kube-api-access-rkhl8\") pod \"community-operators-8qgzx\" (UID: \"5827db23-a022-4d5f-9ae0-ef3505610bce\") " pod="openshift-marketplace/community-operators-8qgzx" Oct 14 07:03:34 crc kubenswrapper[4870]: I1014 07:03:34.850422 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5827db23-a022-4d5f-9ae0-ef3505610bce-utilities\") pod \"community-operators-8qgzx\" (UID: \"5827db23-a022-4d5f-9ae0-ef3505610bce\") " pod="openshift-marketplace/community-operators-8qgzx" Oct 14 07:03:34 crc kubenswrapper[4870]: I1014 07:03:34.850511 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/44a6ae4c-9030-4a51-8960-0c0cc6c59b8e-utilities\") pod \"certified-operators-lkfh6\" (UID: \"44a6ae4c-9030-4a51-8960-0c0cc6c59b8e\") " pod="openshift-marketplace/certified-operators-lkfh6" Oct 14 07:03:34 crc kubenswrapper[4870]: I1014 07:03:34.850535 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cdzjs\" (UniqueName: \"kubernetes.io/projected/44a6ae4c-9030-4a51-8960-0c0cc6c59b8e-kube-api-access-cdzjs\") pod \"certified-operators-lkfh6\" (UID: \"44a6ae4c-9030-4a51-8960-0c0cc6c59b8e\") " pod="openshift-marketplace/certified-operators-lkfh6" Oct 14 07:03:34 crc kubenswrapper[4870]: E1014 07:03:34.851481 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 07:03:35.351466884 +0000 UTC m=+151.048827245 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:03:34 crc kubenswrapper[4870]: I1014 07:03:34.852154 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5827db23-a022-4d5f-9ae0-ef3505610bce-catalog-content\") pod \"community-operators-8qgzx\" (UID: \"5827db23-a022-4d5f-9ae0-ef3505610bce\") " pod="openshift-marketplace/community-operators-8qgzx" Oct 14 07:03:34 crc kubenswrapper[4870]: I1014 07:03:34.853243 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5827db23-a022-4d5f-9ae0-ef3505610bce-utilities\") pod \"community-operators-8qgzx\" (UID: \"5827db23-a022-4d5f-9ae0-ef3505610bce\") " pod="openshift-marketplace/community-operators-8qgzx" Oct 14 07:03:34 crc kubenswrapper[4870]: I1014 07:03:34.866642 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-p769c" event={"ID":"146160a7-2b1c-43f4-bc42-7b92d9663dcc","Type":"ContainerStarted","Data":"083ff12aa32c14c62c2e6d00beffeba16e4c0406a7b6ad54a89aa7dcdf925022"} Oct 14 07:03:34 crc kubenswrapper[4870]: I1014 07:03:34.920915 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rkhl8\" (UniqueName: \"kubernetes.io/projected/5827db23-a022-4d5f-9ae0-ef3505610bce-kube-api-access-rkhl8\") pod \"community-operators-8qgzx\" (UID: \"5827db23-a022-4d5f-9ae0-ef3505610bce\") " pod="openshift-marketplace/community-operators-8qgzx" Oct 14 07:03:34 crc kubenswrapper[4870]: I1014 07:03:34.921004 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-ptjzm" event={"ID":"edd471fb-addd-407d-b7cb-55dbffd7a1ee","Type":"ContainerStarted","Data":"0a957a28d2dc37a0d693be8f7a5472dddc7f943cbb59710743bdd29c9ad0f66e"} Oct 14 07:03:34 crc kubenswrapper[4870]: I1014 07:03:34.947031 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-6z4z4" event={"ID":"be735c57-60b4-44cb-abf3-888cf00d6c31","Type":"ContainerStarted","Data":"48c2803141968fa35988468dcd45061960d8c47ea582ad6076e881eed78c86ec"} Oct 14 07:03:34 crc kubenswrapper[4870]: I1014 07:03:34.954468 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/44a6ae4c-9030-4a51-8960-0c0cc6c59b8e-utilities\") pod \"certified-operators-lkfh6\" (UID: \"44a6ae4c-9030-4a51-8960-0c0cc6c59b8e\") " pod="openshift-marketplace/certified-operators-lkfh6" Oct 14 07:03:34 crc kubenswrapper[4870]: I1014 07:03:34.954503 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cdzjs\" (UniqueName: \"kubernetes.io/projected/44a6ae4c-9030-4a51-8960-0c0cc6c59b8e-kube-api-access-cdzjs\") pod \"certified-operators-lkfh6\" (UID: \"44a6ae4c-9030-4a51-8960-0c0cc6c59b8e\") " pod="openshift-marketplace/certified-operators-lkfh6" Oct 14 07:03:34 crc kubenswrapper[4870]: I1014 07:03:34.954546 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/44a6ae4c-9030-4a51-8960-0c0cc6c59b8e-catalog-content\") pod \"certified-operators-lkfh6\" (UID: \"44a6ae4c-9030-4a51-8960-0c0cc6c59b8e\") " pod="openshift-marketplace/certified-operators-lkfh6" Oct 14 07:03:34 crc kubenswrapper[4870]: I1014 07:03:34.954593 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kkxjz\" (UID: \"2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3\") " pod="openshift-image-registry/image-registry-697d97f7c8-kkxjz" Oct 14 07:03:34 crc kubenswrapper[4870]: I1014 07:03:34.955769 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/44a6ae4c-9030-4a51-8960-0c0cc6c59b8e-utilities\") pod \"certified-operators-lkfh6\" (UID: \"44a6ae4c-9030-4a51-8960-0c0cc6c59b8e\") " pod="openshift-marketplace/certified-operators-lkfh6" Oct 14 07:03:34 crc kubenswrapper[4870]: E1014 07:03:34.957871 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-14 07:03:35.457852214 +0000 UTC m=+151.155212585 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kkxjz" (UID: "2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:03:34 crc kubenswrapper[4870]: I1014 07:03:34.958409 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/44a6ae4c-9030-4a51-8960-0c0cc6c59b8e-catalog-content\") pod \"certified-operators-lkfh6\" (UID: \"44a6ae4c-9030-4a51-8960-0c0cc6c59b8e\") " pod="openshift-marketplace/certified-operators-lkfh6" Oct 14 07:03:34 crc kubenswrapper[4870]: I1014 07:03:34.964128 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-5ptfj" event={"ID":"8432ae60-6e63-45cf-b21b-ab1010687654","Type":"ContainerStarted","Data":"f73c6be05c6c6d0a1f0bfe481eb1afc1f5b47797e263480fb105f3571d9eeafb"} Oct 14 07:03:34 crc kubenswrapper[4870]: I1014 07:03:34.968396 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-szwj9" podStartSLOduration=129.96837001 podStartE2EDuration="2m9.96837001s" podCreationTimestamp="2025-10-14 07:01:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:03:34.956693786 +0000 UTC m=+150.654054147" watchObservedRunningTime="2025-10-14 07:03:34.96837001 +0000 UTC m=+150.665730381" Oct 14 07:03:34 crc kubenswrapper[4870]: I1014 07:03:34.969337 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rnf2p" podStartSLOduration=128.969332533 podStartE2EDuration="2m8.969332533s" podCreationTimestamp="2025-10-14 07:01:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:03:34.920651952 +0000 UTC m=+150.618012323" watchObservedRunningTime="2025-10-14 07:03:34.969332533 +0000 UTC m=+150.666692904" Oct 14 07:03:34 crc kubenswrapper[4870]: I1014 07:03:34.998722 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-d4fhp" event={"ID":"9057841b-b3a4-49f2-85a6-812074c635e6","Type":"ContainerStarted","Data":"af94a856cfb19726f4c0e16330b4469304b44bbc978117ef7f0bfd4d1427c3e8"} Oct 14 07:03:34 crc kubenswrapper[4870]: I1014 07:03:34.998784 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-d4fhp" event={"ID":"9057841b-b3a4-49f2-85a6-812074c635e6","Type":"ContainerStarted","Data":"e27b35b7dcb965e99577922c73c0895c452347d226a5e1ad069f7cbfba9a1083"} Oct 14 07:03:35 crc kubenswrapper[4870]: I1014 07:03:35.003751 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-d4fhp" Oct 14 07:03:35 crc kubenswrapper[4870]: I1014 07:03:35.004225 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8qgzx" Oct 14 07:03:35 crc kubenswrapper[4870]: I1014 07:03:35.022275 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cdzjs\" (UniqueName: \"kubernetes.io/projected/44a6ae4c-9030-4a51-8960-0c0cc6c59b8e-kube-api-access-cdzjs\") pod \"certified-operators-lkfh6\" (UID: \"44a6ae4c-9030-4a51-8960-0c0cc6c59b8e\") " pod="openshift-marketplace/certified-operators-lkfh6" Oct 14 07:03:35 crc kubenswrapper[4870]: I1014 07:03:35.057276 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 07:03:35 crc kubenswrapper[4870]: E1014 07:03:35.059064 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 07:03:35.559027198 +0000 UTC m=+151.256387729 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:03:35 crc kubenswrapper[4870]: I1014 07:03:35.079634 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vc8ks" event={"ID":"5848d6ea-90e8-4c07-92e0-8849b5b32c67","Type":"ContainerStarted","Data":"81f49cba08cb6b5c38299bf9f52df35afa092a236ec212efde1bcf5f103ed37e"} Oct 14 07:03:35 crc kubenswrapper[4870]: I1014 07:03:35.084049 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-5ptfj" podStartSLOduration=129.084037235 podStartE2EDuration="2m9.084037235s" podCreationTimestamp="2025-10-14 07:01:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:03:35.081532374 +0000 UTC m=+150.778892745" watchObservedRunningTime="2025-10-14 07:03:35.084037235 +0000 UTC m=+150.781397606" Oct 14 07:03:35 crc kubenswrapper[4870]: I1014 07:03:35.088352 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-6z4z4" podStartSLOduration=129.088341419 podStartE2EDuration="2m9.088341419s" podCreationTimestamp="2025-10-14 07:01:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:03:35.041323069 +0000 UTC m=+150.738683440" watchObservedRunningTime="2025-10-14 07:03:35.088341419 +0000 UTC m=+150.785701790" Oct 14 07:03:35 crc kubenswrapper[4870]: I1014 07:03:35.097884 4870 patch_prober.go:28] interesting pod/router-default-5444994796-h82f7 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 14 07:03:35 crc kubenswrapper[4870]: [-]has-synced failed: reason withheld Oct 14 07:03:35 crc kubenswrapper[4870]: [+]process-running ok Oct 14 07:03:35 crc kubenswrapper[4870]: healthz check failed Oct 14 07:03:35 crc kubenswrapper[4870]: I1014 07:03:35.097974 4870 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-h82f7" podUID="28a417a5-b0b0-465a-ade0-3bf3ceb5ebdd" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 14 07:03:35 crc kubenswrapper[4870]: I1014 07:03:35.115147 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-d4fhp" podStartSLOduration=9.115096148 podStartE2EDuration="9.115096148s" podCreationTimestamp="2025-10-14 07:03:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:03:35.104969853 +0000 UTC m=+150.802330214" watchObservedRunningTime="2025-10-14 07:03:35.115096148 +0000 UTC m=+150.812456519" Oct 14 07:03:35 crc kubenswrapper[4870]: I1014 07:03:35.116933 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-6ztwm" event={"ID":"bb92d614-7f4e-470e-a68d-0aff68d3894d","Type":"ContainerStarted","Data":"ccfc6039d09655c9466f3f8a70079496876c6ff1dda47f198233250f08226c39"} Oct 14 07:03:35 crc kubenswrapper[4870]: I1014 07:03:35.149886 4870 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-6ztwm container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.28:8080/healthz\": dial tcp 10.217.0.28:8080: connect: connection refused" start-of-body= Oct 14 07:03:35 crc kubenswrapper[4870]: I1014 07:03:35.149952 4870 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-6ztwm" podUID="bb92d614-7f4e-470e-a68d-0aff68d3894d" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.28:8080/healthz\": dial tcp 10.217.0.28:8080: connect: connection refused" Oct 14 07:03:35 crc kubenswrapper[4870]: I1014 07:03:35.150899 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lkfh6" Oct 14 07:03:35 crc kubenswrapper[4870]: I1014 07:03:35.164851 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kkxjz\" (UID: \"2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3\") " pod="openshift-image-registry/image-registry-697d97f7c8-kkxjz" Oct 14 07:03:35 crc kubenswrapper[4870]: E1014 07:03:35.166453 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-14 07:03:35.666423053 +0000 UTC m=+151.363783424 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kkxjz" (UID: "2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:03:35 crc kubenswrapper[4870]: I1014 07:03:35.308413 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 07:03:35 crc kubenswrapper[4870]: E1014 07:03:35.310407 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 07:03:35.810386405 +0000 UTC m=+151.507746776 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:03:35 crc kubenswrapper[4870]: I1014 07:03:35.316137 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-n8z6k" event={"ID":"03d1e16b-e364-4902-9412-0dc138fc76e3","Type":"ContainerStarted","Data":"9618950e80c6fdb85ac72aa5ba6ee45fddc2ea9a96bd59d27f115c54a6788ce0"} Oct 14 07:03:35 crc kubenswrapper[4870]: I1014 07:03:35.369886 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-mlqqj" event={"ID":"7562fe32-ac93-483f-b017-3eab4d7f94d4","Type":"ContainerStarted","Data":"fe58589bdc848d8bb1e55d99e647f678183e6a84d96bd4749f1947802a9499e0"} Oct 14 07:03:35 crc kubenswrapper[4870]: W1014 07:03:35.370643 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9d751cbb_f2e2_430d_9754_c882a5e924a5.slice/crio-e4655831f29c051f306d035a3389a3befeb69cd2beb911087c4296b26a449d14 WatchSource:0}: Error finding container e4655831f29c051f306d035a3389a3befeb69cd2beb911087c4296b26a449d14: Status 404 returned error can't find the container with id e4655831f29c051f306d035a3389a3befeb69cd2beb911087c4296b26a449d14 Oct 14 07:03:35 crc kubenswrapper[4870]: I1014 07:03:35.412502 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kkxjz\" (UID: \"2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3\") " pod="openshift-image-registry/image-registry-697d97f7c8-kkxjz" Oct 14 07:03:35 crc kubenswrapper[4870]: E1014 07:03:35.414559 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-14 07:03:35.914545981 +0000 UTC m=+151.611906352 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kkxjz" (UID: "2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:03:35 crc kubenswrapper[4870]: I1014 07:03:35.418223 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-n8z6k" podStartSLOduration=130.4182009 podStartE2EDuration="2m10.4182009s" podCreationTimestamp="2025-10-14 07:01:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:03:35.417534474 +0000 UTC m=+151.114894845" watchObservedRunningTime="2025-10-14 07:03:35.4182009 +0000 UTC m=+151.115561271" Oct 14 07:03:35 crc kubenswrapper[4870]: I1014 07:03:35.458214 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-69t6j" event={"ID":"27bd2071-dddb-48c6-905c-22bcb4e84482","Type":"ContainerStarted","Data":"050f5abd66158472c93c82626eab678f5294b5362ed1ad7dd91e79b696ea2f8c"} Oct 14 07:03:35 crc kubenswrapper[4870]: I1014 07:03:35.489621 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-zpgbg" event={"ID":"222f1962-d6c1-4e69-a3bc-0b0a12703cff","Type":"ContainerStarted","Data":"a390a99863a622fbb7f6945521bf5c8c53ceba466877d623509c20add1696520"} Oct 14 07:03:35 crc kubenswrapper[4870]: I1014 07:03:35.498583 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-mlqqj" podStartSLOduration=129.498552939 podStartE2EDuration="2m9.498552939s" podCreationTimestamp="2025-10-14 07:01:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:03:35.483944015 +0000 UTC m=+151.181304386" watchObservedRunningTime="2025-10-14 07:03:35.498552939 +0000 UTC m=+151.195913310" Oct 14 07:03:35 crc kubenswrapper[4870]: I1014 07:03:35.500460 4870 patch_prober.go:28] interesting pod/downloads-7954f5f757-d7cdf container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.15:8080/\": dial tcp 10.217.0.15:8080: connect: connection refused" start-of-body= Oct 14 07:03:35 crc kubenswrapper[4870]: I1014 07:03:35.500543 4870 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-d7cdf" podUID="2a72b0a0-287b-42dc-8c39-1976188955de" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.15:8080/\": dial tcp 10.217.0.15:8080: connect: connection refused" Oct 14 07:03:35 crc kubenswrapper[4870]: I1014 07:03:35.515172 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 07:03:35 crc kubenswrapper[4870]: E1014 07:03:35.515370 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 07:03:36.015334386 +0000 UTC m=+151.712694757 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:03:35 crc kubenswrapper[4870]: I1014 07:03:35.515707 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kkxjz\" (UID: \"2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3\") " pod="openshift-image-registry/image-registry-697d97f7c8-kkxjz" Oct 14 07:03:35 crc kubenswrapper[4870]: E1014 07:03:35.516116 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-14 07:03:36.016101594 +0000 UTC m=+151.713461965 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kkxjz" (UID: "2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:03:35 crc kubenswrapper[4870]: I1014 07:03:35.546458 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-jhjmr"] Oct 14 07:03:35 crc kubenswrapper[4870]: W1014 07:03:35.617337 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod44734ad6_0693_4d02_ba13_441850b5d23b.slice/crio-59d0f3754cc6c986e4ec7f17db273e835be751584eb3f08d2227fe3177af5c06 WatchSource:0}: Error finding container 59d0f3754cc6c986e4ec7f17db273e835be751584eb3f08d2227fe3177af5c06: Status 404 returned error can't find the container with id 59d0f3754cc6c986e4ec7f17db273e835be751584eb3f08d2227fe3177af5c06 Oct 14 07:03:35 crc kubenswrapper[4870]: I1014 07:03:35.618330 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 07:03:35 crc kubenswrapper[4870]: E1014 07:03:35.620035 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 07:03:36.120009895 +0000 UTC m=+151.817370266 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:03:35 crc kubenswrapper[4870]: I1014 07:03:35.672925 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-zpgbg" podStartSLOduration=129.672902548 podStartE2EDuration="2m9.672902548s" podCreationTimestamp="2025-10-14 07:01:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:03:35.618757514 +0000 UTC m=+151.316117885" watchObservedRunningTime="2025-10-14 07:03:35.672902548 +0000 UTC m=+151.370262919" Oct 14 07:03:35 crc kubenswrapper[4870]: I1014 07:03:35.675773 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-2glwc"] Oct 14 07:03:35 crc kubenswrapper[4870]: W1014 07:03:35.721151 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod788a3634_1dce_4c12_979b_235c74c9200b.slice/crio-0471c8097386471530cd4bf159903868505a94fb0c0eedbec7efdef724bb269e WatchSource:0}: Error finding container 0471c8097386471530cd4bf159903868505a94fb0c0eedbec7efdef724bb269e: Status 404 returned error can't find the container with id 0471c8097386471530cd4bf159903868505a94fb0c0eedbec7efdef724bb269e Oct 14 07:03:35 crc kubenswrapper[4870]: I1014 07:03:35.722457 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kkxjz\" (UID: \"2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3\") " pod="openshift-image-registry/image-registry-697d97f7c8-kkxjz" Oct 14 07:03:35 crc kubenswrapper[4870]: E1014 07:03:35.722752 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-14 07:03:36.222741456 +0000 UTC m=+151.920101827 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kkxjz" (UID: "2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:03:35 crc kubenswrapper[4870]: I1014 07:03:35.793515 4870 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-rnf2p container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.36:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Oct 14 07:03:35 crc kubenswrapper[4870]: I1014 07:03:35.793581 4870 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rnf2p" podUID="1d891345-93bd-4726-905f-8a946a4a885f" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.36:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Oct 14 07:03:35 crc kubenswrapper[4870]: I1014 07:03:35.823969 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 07:03:35 crc kubenswrapper[4870]: E1014 07:03:35.824705 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 07:03:36.324665798 +0000 UTC m=+152.022026169 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:03:35 crc kubenswrapper[4870]: I1014 07:03:35.921351 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-8qgzx"] Oct 14 07:03:35 crc kubenswrapper[4870]: I1014 07:03:35.929704 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kkxjz\" (UID: \"2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3\") " pod="openshift-image-registry/image-registry-697d97f7c8-kkxjz" Oct 14 07:03:35 crc kubenswrapper[4870]: E1014 07:03:35.930323 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-14 07:03:36.430307051 +0000 UTC m=+152.127667422 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kkxjz" (UID: "2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:03:36 crc kubenswrapper[4870]: I1014 07:03:36.031214 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 07:03:36 crc kubenswrapper[4870]: E1014 07:03:36.032704 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 07:03:36.532668553 +0000 UTC m=+152.230028924 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:03:36 crc kubenswrapper[4870]: I1014 07:03:36.092708 4870 patch_prober.go:28] interesting pod/router-default-5444994796-h82f7 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 14 07:03:36 crc kubenswrapper[4870]: [-]has-synced failed: reason withheld Oct 14 07:03:36 crc kubenswrapper[4870]: [+]process-running ok Oct 14 07:03:36 crc kubenswrapper[4870]: healthz check failed Oct 14 07:03:36 crc kubenswrapper[4870]: I1014 07:03:36.092786 4870 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-h82f7" podUID="28a417a5-b0b0-465a-ade0-3bf3ceb5ebdd" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 14 07:03:36 crc kubenswrapper[4870]: I1014 07:03:36.139378 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kkxjz\" (UID: \"2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3\") " pod="openshift-image-registry/image-registry-697d97f7c8-kkxjz" Oct 14 07:03:36 crc kubenswrapper[4870]: E1014 07:03:36.139773 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-14 07:03:36.639760011 +0000 UTC m=+152.337120382 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kkxjz" (UID: "2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:03:36 crc kubenswrapper[4870]: I1014 07:03:36.139813 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-lkfh6"] Oct 14 07:03:36 crc kubenswrapper[4870]: I1014 07:03:36.158288 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-4mlsh"] Oct 14 07:03:36 crc kubenswrapper[4870]: I1014 07:03:36.159388 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4mlsh" Oct 14 07:03:36 crc kubenswrapper[4870]: E1014 07:03:36.177534 4870 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod737f304d_12ac_4d4b_97fb_e99506dd08a8.slice/crio-conmon-54dde5dec3d53d9948e6ae0d3f700606e677c020c9d7e4bd7179a25688038920.scope\": RecentStats: unable to find data in memory cache]" Oct 14 07:03:36 crc kubenswrapper[4870]: I1014 07:03:36.178167 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Oct 14 07:03:36 crc kubenswrapper[4870]: I1014 07:03:36.201579 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-4mlsh"] Oct 14 07:03:36 crc kubenswrapper[4870]: I1014 07:03:36.241802 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 07:03:36 crc kubenswrapper[4870]: E1014 07:03:36.242147 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 07:03:36.742090162 +0000 UTC m=+152.439450533 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:03:36 crc kubenswrapper[4870]: I1014 07:03:36.242691 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kkxjz\" (UID: \"2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3\") " pod="openshift-image-registry/image-registry-697d97f7c8-kkxjz" Oct 14 07:03:36 crc kubenswrapper[4870]: E1014 07:03:36.243459 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-14 07:03:36.743424815 +0000 UTC m=+152.440785186 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kkxjz" (UID: "2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:03:36 crc kubenswrapper[4870]: I1014 07:03:36.345541 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 07:03:36 crc kubenswrapper[4870]: I1014 07:03:36.345768 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dbace338-bfac-4c71-a6bf-28227f826098-catalog-content\") pod \"redhat-marketplace-4mlsh\" (UID: \"dbace338-bfac-4c71-a6bf-28227f826098\") " pod="openshift-marketplace/redhat-marketplace-4mlsh" Oct 14 07:03:36 crc kubenswrapper[4870]: I1014 07:03:36.345793 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dbace338-bfac-4c71-a6bf-28227f826098-utilities\") pod \"redhat-marketplace-4mlsh\" (UID: \"dbace338-bfac-4c71-a6bf-28227f826098\") " pod="openshift-marketplace/redhat-marketplace-4mlsh" Oct 14 07:03:36 crc kubenswrapper[4870]: I1014 07:03:36.345835 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7c7n6\" (UniqueName: \"kubernetes.io/projected/dbace338-bfac-4c71-a6bf-28227f826098-kube-api-access-7c7n6\") pod \"redhat-marketplace-4mlsh\" (UID: \"dbace338-bfac-4c71-a6bf-28227f826098\") " pod="openshift-marketplace/redhat-marketplace-4mlsh" Oct 14 07:03:36 crc kubenswrapper[4870]: E1014 07:03:36.345935 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 07:03:36.845917811 +0000 UTC m=+152.543278172 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:03:36 crc kubenswrapper[4870]: I1014 07:03:36.429700 4870 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Oct 14 07:03:36 crc kubenswrapper[4870]: I1014 07:03:36.447043 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dbace338-bfac-4c71-a6bf-28227f826098-catalog-content\") pod \"redhat-marketplace-4mlsh\" (UID: \"dbace338-bfac-4c71-a6bf-28227f826098\") " pod="openshift-marketplace/redhat-marketplace-4mlsh" Oct 14 07:03:36 crc kubenswrapper[4870]: I1014 07:03:36.447095 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dbace338-bfac-4c71-a6bf-28227f826098-utilities\") pod \"redhat-marketplace-4mlsh\" (UID: \"dbace338-bfac-4c71-a6bf-28227f826098\") " pod="openshift-marketplace/redhat-marketplace-4mlsh" Oct 14 07:03:36 crc kubenswrapper[4870]: I1014 07:03:36.447134 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kkxjz\" (UID: \"2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3\") " pod="openshift-image-registry/image-registry-697d97f7c8-kkxjz" Oct 14 07:03:36 crc kubenswrapper[4870]: I1014 07:03:36.447163 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7c7n6\" (UniqueName: \"kubernetes.io/projected/dbace338-bfac-4c71-a6bf-28227f826098-kube-api-access-7c7n6\") pod \"redhat-marketplace-4mlsh\" (UID: \"dbace338-bfac-4c71-a6bf-28227f826098\") " pod="openshift-marketplace/redhat-marketplace-4mlsh" Oct 14 07:03:36 crc kubenswrapper[4870]: I1014 07:03:36.447522 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dbace338-bfac-4c71-a6bf-28227f826098-catalog-content\") pod \"redhat-marketplace-4mlsh\" (UID: \"dbace338-bfac-4c71-a6bf-28227f826098\") " pod="openshift-marketplace/redhat-marketplace-4mlsh" Oct 14 07:03:36 crc kubenswrapper[4870]: I1014 07:03:36.447583 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dbace338-bfac-4c71-a6bf-28227f826098-utilities\") pod \"redhat-marketplace-4mlsh\" (UID: \"dbace338-bfac-4c71-a6bf-28227f826098\") " pod="openshift-marketplace/redhat-marketplace-4mlsh" Oct 14 07:03:36 crc kubenswrapper[4870]: E1014 07:03:36.447634 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-14 07:03:36.947611647 +0000 UTC m=+152.644972018 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kkxjz" (UID: "2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:03:36 crc kubenswrapper[4870]: I1014 07:03:36.476802 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7c7n6\" (UniqueName: \"kubernetes.io/projected/dbace338-bfac-4c71-a6bf-28227f826098-kube-api-access-7c7n6\") pod \"redhat-marketplace-4mlsh\" (UID: \"dbace338-bfac-4c71-a6bf-28227f826098\") " pod="openshift-marketplace/redhat-marketplace-4mlsh" Oct 14 07:03:36 crc kubenswrapper[4870]: I1014 07:03:36.496176 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"5ec66bc257a6df2bc11b5a7898a799c33d48449ac4599efe8ee48807b1e7c185"} Oct 14 07:03:36 crc kubenswrapper[4870]: I1014 07:03:36.496470 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"b50fb527f01934a6b771002ed917ffba6109551d02d9e919abd764643254260c"} Oct 14 07:03:36 crc kubenswrapper[4870]: I1014 07:03:36.500700 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"33249ce4edbf4ac6a8fbdcd9b3d97f0035af91676b2134549264032c6b3dfbef"} Oct 14 07:03:36 crc kubenswrapper[4870]: I1014 07:03:36.500754 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"8e06562079021e3d45d606723261a36beee4998475c19c7136680511c6fb2370"} Oct 14 07:03:36 crc kubenswrapper[4870]: I1014 07:03:36.500926 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 07:03:36 crc kubenswrapper[4870]: I1014 07:03:36.503699 4870 generic.go:334] "Generic (PLEG): container finished" podID="5827db23-a022-4d5f-9ae0-ef3505610bce" containerID="9faa5924aeef603e9561d951bb02c850d2bfb078766f3119284528ddd04605d9" exitCode=0 Oct 14 07:03:36 crc kubenswrapper[4870]: I1014 07:03:36.503758 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8qgzx" event={"ID":"5827db23-a022-4d5f-9ae0-ef3505610bce","Type":"ContainerDied","Data":"9faa5924aeef603e9561d951bb02c850d2bfb078766f3119284528ddd04605d9"} Oct 14 07:03:36 crc kubenswrapper[4870]: I1014 07:03:36.503790 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8qgzx" event={"ID":"5827db23-a022-4d5f-9ae0-ef3505610bce","Type":"ContainerStarted","Data":"8e7c0eb7785db8f50e2b2ccb1b227b026ca5cb14a3b49cface9784094c4d64d6"} Oct 14 07:03:36 crc kubenswrapper[4870]: I1014 07:03:36.506150 4870 generic.go:334] "Generic (PLEG): container finished" podID="44734ad6-0693-4d02-ba13-441850b5d23b" containerID="ec89421786ff16532fc30e300ed0f969965315601580c137ae15ad9425a6f677" exitCode=0 Oct 14 07:03:36 crc kubenswrapper[4870]: I1014 07:03:36.506214 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jhjmr" event={"ID":"44734ad6-0693-4d02-ba13-441850b5d23b","Type":"ContainerDied","Data":"ec89421786ff16532fc30e300ed0f969965315601580c137ae15ad9425a6f677"} Oct 14 07:03:36 crc kubenswrapper[4870]: I1014 07:03:36.506232 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jhjmr" event={"ID":"44734ad6-0693-4d02-ba13-441850b5d23b","Type":"ContainerStarted","Data":"59d0f3754cc6c986e4ec7f17db273e835be751584eb3f08d2227fe3177af5c06"} Oct 14 07:03:36 crc kubenswrapper[4870]: I1014 07:03:36.506502 4870 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 14 07:03:36 crc kubenswrapper[4870]: I1014 07:03:36.511288 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"ad77a2f6383f92f4fdaf059d24b7def09c27ea591f4a3035ce5d6104c98307b5"} Oct 14 07:03:36 crc kubenswrapper[4870]: I1014 07:03:36.511475 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"e4655831f29c051f306d035a3389a3befeb69cd2beb911087c4296b26a449d14"} Oct 14 07:03:36 crc kubenswrapper[4870]: I1014 07:03:36.514859 4870 generic.go:334] "Generic (PLEG): container finished" podID="788a3634-1dce-4c12-979b-235c74c9200b" containerID="2d0b0343b0b9d8ef2b1d33fd9bfa0e64399d9d262da5dad799927400e81ab8a4" exitCode=0 Oct 14 07:03:36 crc kubenswrapper[4870]: I1014 07:03:36.515080 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2glwc" event={"ID":"788a3634-1dce-4c12-979b-235c74c9200b","Type":"ContainerDied","Data":"2d0b0343b0b9d8ef2b1d33fd9bfa0e64399d9d262da5dad799927400e81ab8a4"} Oct 14 07:03:36 crc kubenswrapper[4870]: I1014 07:03:36.515164 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2glwc" event={"ID":"788a3634-1dce-4c12-979b-235c74c9200b","Type":"ContainerStarted","Data":"0471c8097386471530cd4bf159903868505a94fb0c0eedbec7efdef724bb269e"} Oct 14 07:03:36 crc kubenswrapper[4870]: I1014 07:03:36.525924 4870 generic.go:334] "Generic (PLEG): container finished" podID="737f304d-12ac-4d4b-97fb-e99506dd08a8" containerID="54dde5dec3d53d9948e6ae0d3f700606e677c020c9d7e4bd7179a25688038920" exitCode=0 Oct 14 07:03:36 crc kubenswrapper[4870]: I1014 07:03:36.526021 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29340420-t6cgj" event={"ID":"737f304d-12ac-4d4b-97fb-e99506dd08a8","Type":"ContainerDied","Data":"54dde5dec3d53d9948e6ae0d3f700606e677c020c9d7e4bd7179a25688038920"} Oct 14 07:03:36 crc kubenswrapper[4870]: I1014 07:03:36.547948 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 07:03:36 crc kubenswrapper[4870]: E1014 07:03:36.548341 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 07:03:37.04832008 +0000 UTC m=+152.745680441 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:03:36 crc kubenswrapper[4870]: I1014 07:03:36.557347 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-ptjzm" event={"ID":"edd471fb-addd-407d-b7cb-55dbffd7a1ee","Type":"ContainerStarted","Data":"df97a8b762155318e5823bbd30115468b3253f51c9c25b5270df0bd55f008a71"} Oct 14 07:03:36 crc kubenswrapper[4870]: I1014 07:03:36.559884 4870 generic.go:334] "Generic (PLEG): container finished" podID="44a6ae4c-9030-4a51-8960-0c0cc6c59b8e" containerID="c87cc557220569277c6a14a11f64301f1124572b2b0c2f867df197992f777f3f" exitCode=0 Oct 14 07:03:36 crc kubenswrapper[4870]: I1014 07:03:36.562960 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lkfh6" event={"ID":"44a6ae4c-9030-4a51-8960-0c0cc6c59b8e","Type":"ContainerDied","Data":"c87cc557220569277c6a14a11f64301f1124572b2b0c2f867df197992f777f3f"} Oct 14 07:03:36 crc kubenswrapper[4870]: I1014 07:03:36.563049 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lkfh6" event={"ID":"44a6ae4c-9030-4a51-8960-0c0cc6c59b8e","Type":"ContainerStarted","Data":"78504c48419a2d7fd9783cd9f6c026eec82a742d5bc29624f4976ab0bad6aac0"} Oct 14 07:03:36 crc kubenswrapper[4870]: I1014 07:03:36.563063 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-6zj22"] Oct 14 07:03:36 crc kubenswrapper[4870]: I1014 07:03:36.571596 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6zj22" Oct 14 07:03:36 crc kubenswrapper[4870]: I1014 07:03:36.584701 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-6ztwm" Oct 14 07:03:36 crc kubenswrapper[4870]: I1014 07:03:36.592622 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rnf2p" Oct 14 07:03:36 crc kubenswrapper[4870]: I1014 07:03:36.597485 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-6zj22"] Oct 14 07:03:36 crc kubenswrapper[4870]: I1014 07:03:36.610763 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-p769c" Oct 14 07:03:36 crc kubenswrapper[4870]: I1014 07:03:36.650304 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kkxjz\" (UID: \"2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3\") " pod="openshift-image-registry/image-registry-697d97f7c8-kkxjz" Oct 14 07:03:36 crc kubenswrapper[4870]: E1014 07:03:36.654359 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-14 07:03:37.154341251 +0000 UTC m=+152.851701622 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kkxjz" (UID: "2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:03:36 crc kubenswrapper[4870]: I1014 07:03:36.673431 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4mlsh" Oct 14 07:03:36 crc kubenswrapper[4870]: I1014 07:03:36.778635 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 07:03:36 crc kubenswrapper[4870]: E1014 07:03:36.779715 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 07:03:37.279694912 +0000 UTC m=+152.977055283 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:03:36 crc kubenswrapper[4870]: I1014 07:03:36.779843 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/36521dfb-3f0f-4a55-a393-bdf3ec7d3213-catalog-content\") pod \"redhat-marketplace-6zj22\" (UID: \"36521dfb-3f0f-4a55-a393-bdf3ec7d3213\") " pod="openshift-marketplace/redhat-marketplace-6zj22" Oct 14 07:03:36 crc kubenswrapper[4870]: I1014 07:03:36.779901 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kkxjz\" (UID: \"2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3\") " pod="openshift-image-registry/image-registry-697d97f7c8-kkxjz" Oct 14 07:03:36 crc kubenswrapper[4870]: I1014 07:03:36.779942 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w9bb8\" (UniqueName: \"kubernetes.io/projected/36521dfb-3f0f-4a55-a393-bdf3ec7d3213-kube-api-access-w9bb8\") pod \"redhat-marketplace-6zj22\" (UID: \"36521dfb-3f0f-4a55-a393-bdf3ec7d3213\") " pod="openshift-marketplace/redhat-marketplace-6zj22" Oct 14 07:03:36 crc kubenswrapper[4870]: I1014 07:03:36.780116 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/36521dfb-3f0f-4a55-a393-bdf3ec7d3213-utilities\") pod \"redhat-marketplace-6zj22\" (UID: \"36521dfb-3f0f-4a55-a393-bdf3ec7d3213\") " pod="openshift-marketplace/redhat-marketplace-6zj22" Oct 14 07:03:36 crc kubenswrapper[4870]: E1014 07:03:36.782334 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-14 07:03:37.282311415 +0000 UTC m=+152.979671776 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kkxjz" (UID: "2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:03:36 crc kubenswrapper[4870]: I1014 07:03:36.881863 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 07:03:36 crc kubenswrapper[4870]: I1014 07:03:36.882628 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/36521dfb-3f0f-4a55-a393-bdf3ec7d3213-utilities\") pod \"redhat-marketplace-6zj22\" (UID: \"36521dfb-3f0f-4a55-a393-bdf3ec7d3213\") " pod="openshift-marketplace/redhat-marketplace-6zj22" Oct 14 07:03:36 crc kubenswrapper[4870]: I1014 07:03:36.882707 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/36521dfb-3f0f-4a55-a393-bdf3ec7d3213-catalog-content\") pod \"redhat-marketplace-6zj22\" (UID: \"36521dfb-3f0f-4a55-a393-bdf3ec7d3213\") " pod="openshift-marketplace/redhat-marketplace-6zj22" Oct 14 07:03:36 crc kubenswrapper[4870]: E1014 07:03:36.882808 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 07:03:37.382751641 +0000 UTC m=+153.080112012 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:03:36 crc kubenswrapper[4870]: I1014 07:03:36.882877 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w9bb8\" (UniqueName: \"kubernetes.io/projected/36521dfb-3f0f-4a55-a393-bdf3ec7d3213-kube-api-access-w9bb8\") pod \"redhat-marketplace-6zj22\" (UID: \"36521dfb-3f0f-4a55-a393-bdf3ec7d3213\") " pod="openshift-marketplace/redhat-marketplace-6zj22" Oct 14 07:03:36 crc kubenswrapper[4870]: I1014 07:03:36.883800 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/36521dfb-3f0f-4a55-a393-bdf3ec7d3213-catalog-content\") pod \"redhat-marketplace-6zj22\" (UID: \"36521dfb-3f0f-4a55-a393-bdf3ec7d3213\") " pod="openshift-marketplace/redhat-marketplace-6zj22" Oct 14 07:03:36 crc kubenswrapper[4870]: I1014 07:03:36.883948 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/36521dfb-3f0f-4a55-a393-bdf3ec7d3213-utilities\") pod \"redhat-marketplace-6zj22\" (UID: \"36521dfb-3f0f-4a55-a393-bdf3ec7d3213\") " pod="openshift-marketplace/redhat-marketplace-6zj22" Oct 14 07:03:36 crc kubenswrapper[4870]: I1014 07:03:36.905249 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w9bb8\" (UniqueName: \"kubernetes.io/projected/36521dfb-3f0f-4a55-a393-bdf3ec7d3213-kube-api-access-w9bb8\") pod \"redhat-marketplace-6zj22\" (UID: \"36521dfb-3f0f-4a55-a393-bdf3ec7d3213\") " pod="openshift-marketplace/redhat-marketplace-6zj22" Oct 14 07:03:36 crc kubenswrapper[4870]: I1014 07:03:36.924720 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6zj22" Oct 14 07:03:36 crc kubenswrapper[4870]: I1014 07:03:36.988212 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kkxjz\" (UID: \"2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3\") " pod="openshift-image-registry/image-registry-697d97f7c8-kkxjz" Oct 14 07:03:36 crc kubenswrapper[4870]: I1014 07:03:36.993286 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-4mlsh"] Oct 14 07:03:36 crc kubenswrapper[4870]: E1014 07:03:36.993371 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-14 07:03:37.493339153 +0000 UTC m=+153.190699524 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kkxjz" (UID: "2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:03:37 crc kubenswrapper[4870]: I1014 07:03:37.105944 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 07:03:37 crc kubenswrapper[4870]: E1014 07:03:37.106156 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 07:03:37.606118028 +0000 UTC m=+153.303478399 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:03:37 crc kubenswrapper[4870]: I1014 07:03:37.106202 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kkxjz\" (UID: \"2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3\") " pod="openshift-image-registry/image-registry-697d97f7c8-kkxjz" Oct 14 07:03:37 crc kubenswrapper[4870]: E1014 07:03:37.106743 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-14 07:03:37.606724983 +0000 UTC m=+153.304085354 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kkxjz" (UID: "2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:03:37 crc kubenswrapper[4870]: I1014 07:03:37.114013 4870 patch_prober.go:28] interesting pod/router-default-5444994796-h82f7 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 14 07:03:37 crc kubenswrapper[4870]: [-]has-synced failed: reason withheld Oct 14 07:03:37 crc kubenswrapper[4870]: [+]process-running ok Oct 14 07:03:37 crc kubenswrapper[4870]: healthz check failed Oct 14 07:03:37 crc kubenswrapper[4870]: I1014 07:03:37.114314 4870 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-h82f7" podUID="28a417a5-b0b0-465a-ade0-3bf3ceb5ebdd" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 14 07:03:37 crc kubenswrapper[4870]: I1014 07:03:37.207212 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 07:03:37 crc kubenswrapper[4870]: E1014 07:03:37.207672 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-14 07:03:37.707650781 +0000 UTC m=+153.405011152 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 14 07:03:37 crc kubenswrapper[4870]: I1014 07:03:37.239302 4870 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-10-14T07:03:36.429972079Z","Handler":null,"Name":""} Oct 14 07:03:37 crc kubenswrapper[4870]: I1014 07:03:37.264799 4870 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Oct 14 07:03:37 crc kubenswrapper[4870]: I1014 07:03:37.264851 4870 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Oct 14 07:03:37 crc kubenswrapper[4870]: I1014 07:03:37.265815 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-6zj22"] Oct 14 07:03:37 crc kubenswrapper[4870]: I1014 07:03:37.312488 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kkxjz\" (UID: \"2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3\") " pod="openshift-image-registry/image-registry-697d97f7c8-kkxjz" Oct 14 07:03:37 crc kubenswrapper[4870]: I1014 07:03:37.315497 4870 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 14 07:03:37 crc kubenswrapper[4870]: I1014 07:03:37.315549 4870 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kkxjz\" (UID: \"2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-kkxjz" Oct 14 07:03:37 crc kubenswrapper[4870]: I1014 07:03:37.350770 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-cv6fw"] Oct 14 07:03:37 crc kubenswrapper[4870]: I1014 07:03:37.352168 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-cv6fw" Oct 14 07:03:37 crc kubenswrapper[4870]: I1014 07:03:37.356813 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Oct 14 07:03:37 crc kubenswrapper[4870]: I1014 07:03:37.365321 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-cv6fw"] Oct 14 07:03:37 crc kubenswrapper[4870]: I1014 07:03:37.396366 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kkxjz\" (UID: \"2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3\") " pod="openshift-image-registry/image-registry-697d97f7c8-kkxjz" Oct 14 07:03:37 crc kubenswrapper[4870]: I1014 07:03:37.415835 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 14 07:03:37 crc kubenswrapper[4870]: I1014 07:03:37.416048 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wwphs\" (UniqueName: \"kubernetes.io/projected/ea907a55-c1d5-4f49-9ec4-c97461eef014-kube-api-access-wwphs\") pod \"redhat-operators-cv6fw\" (UID: \"ea907a55-c1d5-4f49-9ec4-c97461eef014\") " pod="openshift-marketplace/redhat-operators-cv6fw" Oct 14 07:03:37 crc kubenswrapper[4870]: I1014 07:03:37.416081 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ea907a55-c1d5-4f49-9ec4-c97461eef014-catalog-content\") pod \"redhat-operators-cv6fw\" (UID: \"ea907a55-c1d5-4f49-9ec4-c97461eef014\") " pod="openshift-marketplace/redhat-operators-cv6fw" Oct 14 07:03:37 crc kubenswrapper[4870]: I1014 07:03:37.416107 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ea907a55-c1d5-4f49-9ec4-c97461eef014-utilities\") pod \"redhat-operators-cv6fw\" (UID: \"ea907a55-c1d5-4f49-9ec4-c97461eef014\") " pod="openshift-marketplace/redhat-operators-cv6fw" Oct 14 07:03:37 crc kubenswrapper[4870]: I1014 07:03:37.426974 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 14 07:03:37 crc kubenswrapper[4870]: I1014 07:03:37.518364 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wwphs\" (UniqueName: \"kubernetes.io/projected/ea907a55-c1d5-4f49-9ec4-c97461eef014-kube-api-access-wwphs\") pod \"redhat-operators-cv6fw\" (UID: \"ea907a55-c1d5-4f49-9ec4-c97461eef014\") " pod="openshift-marketplace/redhat-operators-cv6fw" Oct 14 07:03:37 crc kubenswrapper[4870]: I1014 07:03:37.518421 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ea907a55-c1d5-4f49-9ec4-c97461eef014-catalog-content\") pod \"redhat-operators-cv6fw\" (UID: \"ea907a55-c1d5-4f49-9ec4-c97461eef014\") " pod="openshift-marketplace/redhat-operators-cv6fw" Oct 14 07:03:37 crc kubenswrapper[4870]: I1014 07:03:37.518468 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ea907a55-c1d5-4f49-9ec4-c97461eef014-utilities\") pod \"redhat-operators-cv6fw\" (UID: \"ea907a55-c1d5-4f49-9ec4-c97461eef014\") " pod="openshift-marketplace/redhat-operators-cv6fw" Oct 14 07:03:37 crc kubenswrapper[4870]: I1014 07:03:37.518916 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ea907a55-c1d5-4f49-9ec4-c97461eef014-utilities\") pod \"redhat-operators-cv6fw\" (UID: \"ea907a55-c1d5-4f49-9ec4-c97461eef014\") " pod="openshift-marketplace/redhat-operators-cv6fw" Oct 14 07:03:37 crc kubenswrapper[4870]: I1014 07:03:37.519591 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ea907a55-c1d5-4f49-9ec4-c97461eef014-catalog-content\") pod \"redhat-operators-cv6fw\" (UID: \"ea907a55-c1d5-4f49-9ec4-c97461eef014\") " pod="openshift-marketplace/redhat-operators-cv6fw" Oct 14 07:03:37 crc kubenswrapper[4870]: I1014 07:03:37.522480 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-v7br2" Oct 14 07:03:37 crc kubenswrapper[4870]: I1014 07:03:37.524248 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-kkxjz" Oct 14 07:03:37 crc kubenswrapper[4870]: I1014 07:03:37.545046 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wwphs\" (UniqueName: \"kubernetes.io/projected/ea907a55-c1d5-4f49-9ec4-c97461eef014-kube-api-access-wwphs\") pod \"redhat-operators-cv6fw\" (UID: \"ea907a55-c1d5-4f49-9ec4-c97461eef014\") " pod="openshift-marketplace/redhat-operators-cv6fw" Oct 14 07:03:37 crc kubenswrapper[4870]: I1014 07:03:37.586337 4870 generic.go:334] "Generic (PLEG): container finished" podID="dbace338-bfac-4c71-a6bf-28227f826098" containerID="3ec3dde006bb0c7e855f4597f765cbcd6f08cf0c765b91b75871349f8e17c8e3" exitCode=0 Oct 14 07:03:37 crc kubenswrapper[4870]: I1014 07:03:37.586711 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4mlsh" event={"ID":"dbace338-bfac-4c71-a6bf-28227f826098","Type":"ContainerDied","Data":"3ec3dde006bb0c7e855f4597f765cbcd6f08cf0c765b91b75871349f8e17c8e3"} Oct 14 07:03:37 crc kubenswrapper[4870]: I1014 07:03:37.586926 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4mlsh" event={"ID":"dbace338-bfac-4c71-a6bf-28227f826098","Type":"ContainerStarted","Data":"469e4a40b9e35d76c5c7dcc65e03d0bf7b9bf1ddadf222e076d51da9f6576edf"} Oct 14 07:03:37 crc kubenswrapper[4870]: I1014 07:03:37.595643 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6zj22" event={"ID":"36521dfb-3f0f-4a55-a393-bdf3ec7d3213","Type":"ContainerStarted","Data":"d9109f8e72de4d832583b628a21e52c9a80d2b4c505348d5b3a36d6d26b903c3"} Oct 14 07:03:37 crc kubenswrapper[4870]: I1014 07:03:37.595686 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6zj22" event={"ID":"36521dfb-3f0f-4a55-a393-bdf3ec7d3213","Type":"ContainerStarted","Data":"4ec1dcf5cc8e3b870865f8c445531382a2f81caa1de49469f13b74cc4e781eaf"} Oct 14 07:03:37 crc kubenswrapper[4870]: I1014 07:03:37.609633 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-ptjzm" event={"ID":"edd471fb-addd-407d-b7cb-55dbffd7a1ee","Type":"ContainerStarted","Data":"d45a35e5767167c1e29748796350c0011eb8ace2b8e97f885bf04771115ed396"} Oct 14 07:03:37 crc kubenswrapper[4870]: I1014 07:03:37.609687 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-ptjzm" event={"ID":"edd471fb-addd-407d-b7cb-55dbffd7a1ee","Type":"ContainerStarted","Data":"36f722fb28e280a05e9fd1acac32b123cf9873fa3b11b1e9f9c5611c342f21d6"} Oct 14 07:03:37 crc kubenswrapper[4870]: I1014 07:03:37.679018 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 14 07:03:37 crc kubenswrapper[4870]: I1014 07:03:37.679802 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 14 07:03:37 crc kubenswrapper[4870]: I1014 07:03:37.692362 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Oct 14 07:03:37 crc kubenswrapper[4870]: I1014 07:03:37.692646 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Oct 14 07:03:37 crc kubenswrapper[4870]: I1014 07:03:37.694029 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 14 07:03:37 crc kubenswrapper[4870]: I1014 07:03:37.699632 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-cv6fw" Oct 14 07:03:37 crc kubenswrapper[4870]: I1014 07:03:37.753190 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-ptjzm" podStartSLOduration=12.753156992 podStartE2EDuration="12.753156992s" podCreationTimestamp="2025-10-14 07:03:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:03:37.712115866 +0000 UTC m=+153.409476227" watchObservedRunningTime="2025-10-14 07:03:37.753156992 +0000 UTC m=+153.450517363" Oct 14 07:03:37 crc kubenswrapper[4870]: I1014 07:03:37.792985 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-kxxk5"] Oct 14 07:03:37 crc kubenswrapper[4870]: I1014 07:03:37.795641 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kxxk5" Oct 14 07:03:37 crc kubenswrapper[4870]: I1014 07:03:37.808270 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-kxxk5"] Oct 14 07:03:37 crc kubenswrapper[4870]: I1014 07:03:37.826866 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6048b73d-400f-4bdd-9643-0564afe48d76-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"6048b73d-400f-4bdd-9643-0564afe48d76\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 14 07:03:37 crc kubenswrapper[4870]: I1014 07:03:37.826916 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/6048b73d-400f-4bdd-9643-0564afe48d76-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"6048b73d-400f-4bdd-9643-0564afe48d76\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 14 07:03:37 crc kubenswrapper[4870]: I1014 07:03:37.936032 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6048b73d-400f-4bdd-9643-0564afe48d76-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"6048b73d-400f-4bdd-9643-0564afe48d76\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 14 07:03:37 crc kubenswrapper[4870]: I1014 07:03:37.936102 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/6048b73d-400f-4bdd-9643-0564afe48d76-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"6048b73d-400f-4bdd-9643-0564afe48d76\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 14 07:03:37 crc kubenswrapper[4870]: I1014 07:03:37.936128 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d56357b5-cdfe-4903-b59a-4ba8a111883f-utilities\") pod \"redhat-operators-kxxk5\" (UID: \"d56357b5-cdfe-4903-b59a-4ba8a111883f\") " pod="openshift-marketplace/redhat-operators-kxxk5" Oct 14 07:03:37 crc kubenswrapper[4870]: I1014 07:03:37.936148 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d56357b5-cdfe-4903-b59a-4ba8a111883f-catalog-content\") pod \"redhat-operators-kxxk5\" (UID: \"d56357b5-cdfe-4903-b59a-4ba8a111883f\") " pod="openshift-marketplace/redhat-operators-kxxk5" Oct 14 07:03:37 crc kubenswrapper[4870]: I1014 07:03:37.936175 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mmllz\" (UniqueName: \"kubernetes.io/projected/d56357b5-cdfe-4903-b59a-4ba8a111883f-kube-api-access-mmllz\") pod \"redhat-operators-kxxk5\" (UID: \"d56357b5-cdfe-4903-b59a-4ba8a111883f\") " pod="openshift-marketplace/redhat-operators-kxxk5" Oct 14 07:03:37 crc kubenswrapper[4870]: I1014 07:03:37.936592 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/6048b73d-400f-4bdd-9643-0564afe48d76-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"6048b73d-400f-4bdd-9643-0564afe48d76\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 14 07:03:37 crc kubenswrapper[4870]: I1014 07:03:37.981119 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6048b73d-400f-4bdd-9643-0564afe48d76-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"6048b73d-400f-4bdd-9643-0564afe48d76\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 14 07:03:38 crc kubenswrapper[4870]: I1014 07:03:38.010453 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-kkxjz"] Oct 14 07:03:38 crc kubenswrapper[4870]: I1014 07:03:38.042040 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d56357b5-cdfe-4903-b59a-4ba8a111883f-utilities\") pod \"redhat-operators-kxxk5\" (UID: \"d56357b5-cdfe-4903-b59a-4ba8a111883f\") " pod="openshift-marketplace/redhat-operators-kxxk5" Oct 14 07:03:38 crc kubenswrapper[4870]: I1014 07:03:38.042095 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d56357b5-cdfe-4903-b59a-4ba8a111883f-catalog-content\") pod \"redhat-operators-kxxk5\" (UID: \"d56357b5-cdfe-4903-b59a-4ba8a111883f\") " pod="openshift-marketplace/redhat-operators-kxxk5" Oct 14 07:03:38 crc kubenswrapper[4870]: I1014 07:03:38.042125 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mmllz\" (UniqueName: \"kubernetes.io/projected/d56357b5-cdfe-4903-b59a-4ba8a111883f-kube-api-access-mmllz\") pod \"redhat-operators-kxxk5\" (UID: \"d56357b5-cdfe-4903-b59a-4ba8a111883f\") " pod="openshift-marketplace/redhat-operators-kxxk5" Oct 14 07:03:38 crc kubenswrapper[4870]: I1014 07:03:38.043011 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d56357b5-cdfe-4903-b59a-4ba8a111883f-utilities\") pod \"redhat-operators-kxxk5\" (UID: \"d56357b5-cdfe-4903-b59a-4ba8a111883f\") " pod="openshift-marketplace/redhat-operators-kxxk5" Oct 14 07:03:38 crc kubenswrapper[4870]: I1014 07:03:38.043120 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d56357b5-cdfe-4903-b59a-4ba8a111883f-catalog-content\") pod \"redhat-operators-kxxk5\" (UID: \"d56357b5-cdfe-4903-b59a-4ba8a111883f\") " pod="openshift-marketplace/redhat-operators-kxxk5" Oct 14 07:03:38 crc kubenswrapper[4870]: I1014 07:03:38.049742 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 14 07:03:38 crc kubenswrapper[4870]: I1014 07:03:38.064687 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mmllz\" (UniqueName: \"kubernetes.io/projected/d56357b5-cdfe-4903-b59a-4ba8a111883f-kube-api-access-mmllz\") pod \"redhat-operators-kxxk5\" (UID: \"d56357b5-cdfe-4903-b59a-4ba8a111883f\") " pod="openshift-marketplace/redhat-operators-kxxk5" Oct 14 07:03:38 crc kubenswrapper[4870]: I1014 07:03:38.088636 4870 patch_prober.go:28] interesting pod/router-default-5444994796-h82f7 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 14 07:03:38 crc kubenswrapper[4870]: [-]has-synced failed: reason withheld Oct 14 07:03:38 crc kubenswrapper[4870]: [+]process-running ok Oct 14 07:03:38 crc kubenswrapper[4870]: healthz check failed Oct 14 07:03:38 crc kubenswrapper[4870]: I1014 07:03:38.088748 4870 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-h82f7" podUID="28a417a5-b0b0-465a-ade0-3bf3ceb5ebdd" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 14 07:03:38 crc kubenswrapper[4870]: I1014 07:03:38.181702 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29340420-t6cgj" Oct 14 07:03:38 crc kubenswrapper[4870]: I1014 07:03:38.199162 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-r9t2k" Oct 14 07:03:38 crc kubenswrapper[4870]: I1014 07:03:38.199219 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-r9t2k" Oct 14 07:03:38 crc kubenswrapper[4870]: I1014 07:03:38.216743 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kxxk5" Oct 14 07:03:38 crc kubenswrapper[4870]: I1014 07:03:38.218346 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-cv6fw"] Oct 14 07:03:38 crc kubenswrapper[4870]: I1014 07:03:38.228257 4870 patch_prober.go:28] interesting pod/console-f9d7485db-r9t2k container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.8:8443/health\": dial tcp 10.217.0.8:8443: connect: connection refused" start-of-body= Oct 14 07:03:38 crc kubenswrapper[4870]: I1014 07:03:38.228351 4870 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-r9t2k" podUID="b38b0739-21a6-4050-af6c-6accabefcc45" containerName="console" probeResult="failure" output="Get \"https://10.217.0.8:8443/health\": dial tcp 10.217.0.8:8443: connect: connection refused" Oct 14 07:03:38 crc kubenswrapper[4870]: W1014 07:03:38.276496 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podea907a55_c1d5_4f49_9ec4_c97461eef014.slice/crio-82b52bb3e730c37cbb2828c0ded83d866a5c7be7a03af12f79c8825fc05292c7 WatchSource:0}: Error finding container 82b52bb3e730c37cbb2828c0ded83d866a5c7be7a03af12f79c8825fc05292c7: Status 404 returned error can't find the container with id 82b52bb3e730c37cbb2828c0ded83d866a5c7be7a03af12f79c8825fc05292c7 Oct 14 07:03:38 crc kubenswrapper[4870]: I1014 07:03:38.350882 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-n8z6k" Oct 14 07:03:38 crc kubenswrapper[4870]: I1014 07:03:38.350925 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-n8z6k" Oct 14 07:03:38 crc kubenswrapper[4870]: I1014 07:03:38.361886 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/737f304d-12ac-4d4b-97fb-e99506dd08a8-config-volume" (OuterVolumeSpecName: "config-volume") pod "737f304d-12ac-4d4b-97fb-e99506dd08a8" (UID: "737f304d-12ac-4d4b-97fb-e99506dd08a8"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:03:38 crc kubenswrapper[4870]: I1014 07:03:38.358405 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/737f304d-12ac-4d4b-97fb-e99506dd08a8-config-volume\") pod \"737f304d-12ac-4d4b-97fb-e99506dd08a8\" (UID: \"737f304d-12ac-4d4b-97fb-e99506dd08a8\") " Oct 14 07:03:38 crc kubenswrapper[4870]: I1014 07:03:38.366012 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/737f304d-12ac-4d4b-97fb-e99506dd08a8-secret-volume\") pod \"737f304d-12ac-4d4b-97fb-e99506dd08a8\" (UID: \"737f304d-12ac-4d4b-97fb-e99506dd08a8\") " Oct 14 07:03:38 crc kubenswrapper[4870]: I1014 07:03:38.366081 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6m4zp\" (UniqueName: \"kubernetes.io/projected/737f304d-12ac-4d4b-97fb-e99506dd08a8-kube-api-access-6m4zp\") pod \"737f304d-12ac-4d4b-97fb-e99506dd08a8\" (UID: \"737f304d-12ac-4d4b-97fb-e99506dd08a8\") " Oct 14 07:03:38 crc kubenswrapper[4870]: I1014 07:03:38.366674 4870 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/737f304d-12ac-4d4b-97fb-e99506dd08a8-config-volume\") on node \"crc\" DevicePath \"\"" Oct 14 07:03:38 crc kubenswrapper[4870]: I1014 07:03:38.374639 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/737f304d-12ac-4d4b-97fb-e99506dd08a8-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "737f304d-12ac-4d4b-97fb-e99506dd08a8" (UID: "737f304d-12ac-4d4b-97fb-e99506dd08a8"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:03:38 crc kubenswrapper[4870]: I1014 07:03:38.375963 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/737f304d-12ac-4d4b-97fb-e99506dd08a8-kube-api-access-6m4zp" (OuterVolumeSpecName: "kube-api-access-6m4zp") pod "737f304d-12ac-4d4b-97fb-e99506dd08a8" (UID: "737f304d-12ac-4d4b-97fb-e99506dd08a8"). InnerVolumeSpecName "kube-api-access-6m4zp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:03:38 crc kubenswrapper[4870]: I1014 07:03:38.392022 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-n8z6k" Oct 14 07:03:38 crc kubenswrapper[4870]: I1014 07:03:38.468008 4870 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/737f304d-12ac-4d4b-97fb-e99506dd08a8-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 14 07:03:38 crc kubenswrapper[4870]: I1014 07:03:38.468493 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6m4zp\" (UniqueName: \"kubernetes.io/projected/737f304d-12ac-4d4b-97fb-e99506dd08a8-kube-api-access-6m4zp\") on node \"crc\" DevicePath \"\"" Oct 14 07:03:38 crc kubenswrapper[4870]: I1014 07:03:38.564743 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 14 07:03:38 crc kubenswrapper[4870]: I1014 07:03:38.622070 4870 generic.go:334] "Generic (PLEG): container finished" podID="36521dfb-3f0f-4a55-a393-bdf3ec7d3213" containerID="d9109f8e72de4d832583b628a21e52c9a80d2b4c505348d5b3a36d6d26b903c3" exitCode=0 Oct 14 07:03:38 crc kubenswrapper[4870]: I1014 07:03:38.622159 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6zj22" event={"ID":"36521dfb-3f0f-4a55-a393-bdf3ec7d3213","Type":"ContainerDied","Data":"d9109f8e72de4d832583b628a21e52c9a80d2b4c505348d5b3a36d6d26b903c3"} Oct 14 07:03:38 crc kubenswrapper[4870]: I1014 07:03:38.626187 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"6048b73d-400f-4bdd-9643-0564afe48d76","Type":"ContainerStarted","Data":"ec4964928d23f5a07402e195df5f1c97b257ff92b8edf96ce8287358e0785504"} Oct 14 07:03:38 crc kubenswrapper[4870]: I1014 07:03:38.629327 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cv6fw" event={"ID":"ea907a55-c1d5-4f49-9ec4-c97461eef014","Type":"ContainerStarted","Data":"82b52bb3e730c37cbb2828c0ded83d866a5c7be7a03af12f79c8825fc05292c7"} Oct 14 07:03:38 crc kubenswrapper[4870]: I1014 07:03:38.632147 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-kkxjz" event={"ID":"2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3","Type":"ContainerStarted","Data":"e42a9ade61575cb390af63785857635f25fc9e9808cf54a155343a03b597b870"} Oct 14 07:03:38 crc kubenswrapper[4870]: I1014 07:03:38.634134 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-kxxk5"] Oct 14 07:03:38 crc kubenswrapper[4870]: I1014 07:03:38.639297 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29340420-t6cgj" Oct 14 07:03:38 crc kubenswrapper[4870]: I1014 07:03:38.639334 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29340420-t6cgj" event={"ID":"737f304d-12ac-4d4b-97fb-e99506dd08a8","Type":"ContainerDied","Data":"153066511bbafa672f0b023504ef494ec1e5bead898e6bb97e8b93cfb16c221a"} Oct 14 07:03:38 crc kubenswrapper[4870]: I1014 07:03:38.639368 4870 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="153066511bbafa672f0b023504ef494ec1e5bead898e6bb97e8b93cfb16c221a" Oct 14 07:03:38 crc kubenswrapper[4870]: I1014 07:03:38.649209 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-n8z6k" Oct 14 07:03:39 crc kubenswrapper[4870]: I1014 07:03:39.055337 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Oct 14 07:03:39 crc kubenswrapper[4870]: I1014 07:03:39.085728 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-h82f7" Oct 14 07:03:39 crc kubenswrapper[4870]: I1014 07:03:39.091453 4870 patch_prober.go:28] interesting pod/router-default-5444994796-h82f7 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 14 07:03:39 crc kubenswrapper[4870]: [-]has-synced failed: reason withheld Oct 14 07:03:39 crc kubenswrapper[4870]: [+]process-running ok Oct 14 07:03:39 crc kubenswrapper[4870]: healthz check failed Oct 14 07:03:39 crc kubenswrapper[4870]: I1014 07:03:39.091526 4870 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-h82f7" podUID="28a417a5-b0b0-465a-ade0-3bf3ceb5ebdd" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 14 07:03:39 crc kubenswrapper[4870]: I1014 07:03:39.261924 4870 patch_prober.go:28] interesting pod/downloads-7954f5f757-d7cdf container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.15:8080/\": dial tcp 10.217.0.15:8080: connect: connection refused" start-of-body= Oct 14 07:03:39 crc kubenswrapper[4870]: I1014 07:03:39.261977 4870 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-d7cdf" podUID="2a72b0a0-287b-42dc-8c39-1976188955de" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.15:8080/\": dial tcp 10.217.0.15:8080: connect: connection refused" Oct 14 07:03:39 crc kubenswrapper[4870]: I1014 07:03:39.262059 4870 patch_prober.go:28] interesting pod/downloads-7954f5f757-d7cdf container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.15:8080/\": dial tcp 10.217.0.15:8080: connect: connection refused" start-of-body= Oct 14 07:03:39 crc kubenswrapper[4870]: I1014 07:03:39.262131 4870 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-d7cdf" podUID="2a72b0a0-287b-42dc-8c39-1976188955de" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.15:8080/\": dial tcp 10.217.0.15:8080: connect: connection refused" Oct 14 07:03:39 crc kubenswrapper[4870]: I1014 07:03:39.666768 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"6048b73d-400f-4bdd-9643-0564afe48d76","Type":"ContainerStarted","Data":"3503395aa79cc9782a978937571babe4f7f5c866b5f97f13808673fa82fddae3"} Oct 14 07:03:39 crc kubenswrapper[4870]: I1014 07:03:39.682806 4870 generic.go:334] "Generic (PLEG): container finished" podID="d56357b5-cdfe-4903-b59a-4ba8a111883f" containerID="4e6fb7da82fe8780bc23b33bf067825c8859856a442b4a009c8e51dba72e22f0" exitCode=0 Oct 14 07:03:39 crc kubenswrapper[4870]: I1014 07:03:39.682988 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kxxk5" event={"ID":"d56357b5-cdfe-4903-b59a-4ba8a111883f","Type":"ContainerDied","Data":"4e6fb7da82fe8780bc23b33bf067825c8859856a442b4a009c8e51dba72e22f0"} Oct 14 07:03:39 crc kubenswrapper[4870]: I1014 07:03:39.683030 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kxxk5" event={"ID":"d56357b5-cdfe-4903-b59a-4ba8a111883f","Type":"ContainerStarted","Data":"a341a74b2579a68d177e5b9b442c80b86df1c5ac5747ac4f0a15956aac4850c7"} Oct 14 07:03:39 crc kubenswrapper[4870]: I1014 07:03:39.683428 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/revision-pruner-9-crc" podStartSLOduration=2.683407276 podStartE2EDuration="2.683407276s" podCreationTimestamp="2025-10-14 07:03:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:03:39.680739132 +0000 UTC m=+155.378099513" watchObservedRunningTime="2025-10-14 07:03:39.683407276 +0000 UTC m=+155.380767647" Oct 14 07:03:39 crc kubenswrapper[4870]: I1014 07:03:39.698362 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-kkxjz" event={"ID":"2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3","Type":"ContainerStarted","Data":"c708c420a912b0bab2c3a334dfe984c05b1212497c3eadbbf22c03db8be5eb03"} Oct 14 07:03:39 crc kubenswrapper[4870]: I1014 07:03:39.698559 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-kkxjz" Oct 14 07:03:39 crc kubenswrapper[4870]: I1014 07:03:39.714844 4870 generic.go:334] "Generic (PLEG): container finished" podID="ea907a55-c1d5-4f49-9ec4-c97461eef014" containerID="92cc2b999f056cbf8f202800d2be8d2fa4535872ec24c9274d8a48c8314ed164" exitCode=0 Oct 14 07:03:39 crc kubenswrapper[4870]: I1014 07:03:39.715058 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cv6fw" event={"ID":"ea907a55-c1d5-4f49-9ec4-c97461eef014","Type":"ContainerDied","Data":"92cc2b999f056cbf8f202800d2be8d2fa4535872ec24c9274d8a48c8314ed164"} Oct 14 07:03:39 crc kubenswrapper[4870]: I1014 07:03:39.727807 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-kkxjz" podStartSLOduration=133.727788063 podStartE2EDuration="2m13.727788063s" podCreationTimestamp="2025-10-14 07:01:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:03:39.725087947 +0000 UTC m=+155.422448318" watchObservedRunningTime="2025-10-14 07:03:39.727788063 +0000 UTC m=+155.425148434" Oct 14 07:03:40 crc kubenswrapper[4870]: I1014 07:03:40.087590 4870 patch_prober.go:28] interesting pod/router-default-5444994796-h82f7 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 14 07:03:40 crc kubenswrapper[4870]: [-]has-synced failed: reason withheld Oct 14 07:03:40 crc kubenswrapper[4870]: [+]process-running ok Oct 14 07:03:40 crc kubenswrapper[4870]: healthz check failed Oct 14 07:03:40 crc kubenswrapper[4870]: I1014 07:03:40.087655 4870 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-h82f7" podUID="28a417a5-b0b0-465a-ade0-3bf3ceb5ebdd" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 14 07:03:40 crc kubenswrapper[4870]: I1014 07:03:40.759241 4870 generic.go:334] "Generic (PLEG): container finished" podID="6048b73d-400f-4bdd-9643-0564afe48d76" containerID="3503395aa79cc9782a978937571babe4f7f5c866b5f97f13808673fa82fddae3" exitCode=0 Oct 14 07:03:40 crc kubenswrapper[4870]: I1014 07:03:40.759362 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"6048b73d-400f-4bdd-9643-0564afe48d76","Type":"ContainerDied","Data":"3503395aa79cc9782a978937571babe4f7f5c866b5f97f13808673fa82fddae3"} Oct 14 07:03:41 crc kubenswrapper[4870]: I1014 07:03:41.089540 4870 patch_prober.go:28] interesting pod/router-default-5444994796-h82f7 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 14 07:03:41 crc kubenswrapper[4870]: [-]has-synced failed: reason withheld Oct 14 07:03:41 crc kubenswrapper[4870]: [+]process-running ok Oct 14 07:03:41 crc kubenswrapper[4870]: healthz check failed Oct 14 07:03:41 crc kubenswrapper[4870]: I1014 07:03:41.089665 4870 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-h82f7" podUID="28a417a5-b0b0-465a-ade0-3bf3ceb5ebdd" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 14 07:03:42 crc kubenswrapper[4870]: I1014 07:03:42.088195 4870 patch_prober.go:28] interesting pod/router-default-5444994796-h82f7 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 14 07:03:42 crc kubenswrapper[4870]: [-]has-synced failed: reason withheld Oct 14 07:03:42 crc kubenswrapper[4870]: [+]process-running ok Oct 14 07:03:42 crc kubenswrapper[4870]: healthz check failed Oct 14 07:03:42 crc kubenswrapper[4870]: I1014 07:03:42.089050 4870 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-h82f7" podUID="28a417a5-b0b0-465a-ade0-3bf3ceb5ebdd" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 14 07:03:42 crc kubenswrapper[4870]: I1014 07:03:42.212623 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 14 07:03:42 crc kubenswrapper[4870]: I1014 07:03:42.263809 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/6048b73d-400f-4bdd-9643-0564afe48d76-kubelet-dir\") pod \"6048b73d-400f-4bdd-9643-0564afe48d76\" (UID: \"6048b73d-400f-4bdd-9643-0564afe48d76\") " Oct 14 07:03:42 crc kubenswrapper[4870]: I1014 07:03:42.263937 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6048b73d-400f-4bdd-9643-0564afe48d76-kube-api-access\") pod \"6048b73d-400f-4bdd-9643-0564afe48d76\" (UID: \"6048b73d-400f-4bdd-9643-0564afe48d76\") " Oct 14 07:03:42 crc kubenswrapper[4870]: I1014 07:03:42.263952 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6048b73d-400f-4bdd-9643-0564afe48d76-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "6048b73d-400f-4bdd-9643-0564afe48d76" (UID: "6048b73d-400f-4bdd-9643-0564afe48d76"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 14 07:03:42 crc kubenswrapper[4870]: I1014 07:03:42.264259 4870 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/6048b73d-400f-4bdd-9643-0564afe48d76-kubelet-dir\") on node \"crc\" DevicePath \"\"" Oct 14 07:03:42 crc kubenswrapper[4870]: I1014 07:03:42.272192 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6048b73d-400f-4bdd-9643-0564afe48d76-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "6048b73d-400f-4bdd-9643-0564afe48d76" (UID: "6048b73d-400f-4bdd-9643-0564afe48d76"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:03:42 crc kubenswrapper[4870]: I1014 07:03:42.365906 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6048b73d-400f-4bdd-9643-0564afe48d76-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 14 07:03:42 crc kubenswrapper[4870]: I1014 07:03:42.455709 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 14 07:03:42 crc kubenswrapper[4870]: E1014 07:03:42.456319 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6048b73d-400f-4bdd-9643-0564afe48d76" containerName="pruner" Oct 14 07:03:42 crc kubenswrapper[4870]: I1014 07:03:42.456394 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="6048b73d-400f-4bdd-9643-0564afe48d76" containerName="pruner" Oct 14 07:03:42 crc kubenswrapper[4870]: E1014 07:03:42.456493 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="737f304d-12ac-4d4b-97fb-e99506dd08a8" containerName="collect-profiles" Oct 14 07:03:42 crc kubenswrapper[4870]: I1014 07:03:42.456551 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="737f304d-12ac-4d4b-97fb-e99506dd08a8" containerName="collect-profiles" Oct 14 07:03:42 crc kubenswrapper[4870]: I1014 07:03:42.456700 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="737f304d-12ac-4d4b-97fb-e99506dd08a8" containerName="collect-profiles" Oct 14 07:03:42 crc kubenswrapper[4870]: I1014 07:03:42.456764 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="6048b73d-400f-4bdd-9643-0564afe48d76" containerName="pruner" Oct 14 07:03:42 crc kubenswrapper[4870]: I1014 07:03:42.457231 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 14 07:03:42 crc kubenswrapper[4870]: I1014 07:03:42.460704 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Oct 14 07:03:42 crc kubenswrapper[4870]: I1014 07:03:42.468735 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Oct 14 07:03:42 crc kubenswrapper[4870]: I1014 07:03:42.489181 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 14 07:03:42 crc kubenswrapper[4870]: I1014 07:03:42.570660 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9e620080-d3ad-49dc-af0f-8dd64485e9f7-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"9e620080-d3ad-49dc-af0f-8dd64485e9f7\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 14 07:03:42 crc kubenswrapper[4870]: I1014 07:03:42.571246 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9e620080-d3ad-49dc-af0f-8dd64485e9f7-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"9e620080-d3ad-49dc-af0f-8dd64485e9f7\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 14 07:03:42 crc kubenswrapper[4870]: I1014 07:03:42.672963 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9e620080-d3ad-49dc-af0f-8dd64485e9f7-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"9e620080-d3ad-49dc-af0f-8dd64485e9f7\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 14 07:03:42 crc kubenswrapper[4870]: I1014 07:03:42.673057 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9e620080-d3ad-49dc-af0f-8dd64485e9f7-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"9e620080-d3ad-49dc-af0f-8dd64485e9f7\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 14 07:03:42 crc kubenswrapper[4870]: I1014 07:03:42.673579 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9e620080-d3ad-49dc-af0f-8dd64485e9f7-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"9e620080-d3ad-49dc-af0f-8dd64485e9f7\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 14 07:03:42 crc kubenswrapper[4870]: I1014 07:03:42.738817 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9e620080-d3ad-49dc-af0f-8dd64485e9f7-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"9e620080-d3ad-49dc-af0f-8dd64485e9f7\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 14 07:03:42 crc kubenswrapper[4870]: I1014 07:03:42.780960 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"6048b73d-400f-4bdd-9643-0564afe48d76","Type":"ContainerDied","Data":"ec4964928d23f5a07402e195df5f1c97b257ff92b8edf96ce8287358e0785504"} Oct 14 07:03:42 crc kubenswrapper[4870]: I1014 07:03:42.781017 4870 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ec4964928d23f5a07402e195df5f1c97b257ff92b8edf96ce8287358e0785504" Oct 14 07:03:42 crc kubenswrapper[4870]: I1014 07:03:42.781098 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 14 07:03:42 crc kubenswrapper[4870]: I1014 07:03:42.786675 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 14 07:03:43 crc kubenswrapper[4870]: I1014 07:03:43.089087 4870 patch_prober.go:28] interesting pod/router-default-5444994796-h82f7 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 14 07:03:43 crc kubenswrapper[4870]: [-]has-synced failed: reason withheld Oct 14 07:03:43 crc kubenswrapper[4870]: [+]process-running ok Oct 14 07:03:43 crc kubenswrapper[4870]: healthz check failed Oct 14 07:03:43 crc kubenswrapper[4870]: I1014 07:03:43.091845 4870 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-h82f7" podUID="28a417a5-b0b0-465a-ade0-3bf3ceb5ebdd" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 14 07:03:43 crc kubenswrapper[4870]: I1014 07:03:43.281160 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 14 07:03:44 crc kubenswrapper[4870]: I1014 07:03:44.088668 4870 patch_prober.go:28] interesting pod/router-default-5444994796-h82f7 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 14 07:03:44 crc kubenswrapper[4870]: [-]has-synced failed: reason withheld Oct 14 07:03:44 crc kubenswrapper[4870]: [+]process-running ok Oct 14 07:03:44 crc kubenswrapper[4870]: healthz check failed Oct 14 07:03:44 crc kubenswrapper[4870]: I1014 07:03:44.088752 4870 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-h82f7" podUID="28a417a5-b0b0-465a-ade0-3bf3ceb5ebdd" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 14 07:03:44 crc kubenswrapper[4870]: I1014 07:03:44.574518 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-d4fhp" Oct 14 07:03:45 crc kubenswrapper[4870]: I1014 07:03:45.087367 4870 patch_prober.go:28] interesting pod/router-default-5444994796-h82f7 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 14 07:03:45 crc kubenswrapper[4870]: [-]has-synced failed: reason withheld Oct 14 07:03:45 crc kubenswrapper[4870]: [+]process-running ok Oct 14 07:03:45 crc kubenswrapper[4870]: healthz check failed Oct 14 07:03:45 crc kubenswrapper[4870]: I1014 07:03:45.087474 4870 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-h82f7" podUID="28a417a5-b0b0-465a-ade0-3bf3ceb5ebdd" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 14 07:03:46 crc kubenswrapper[4870]: I1014 07:03:46.088299 4870 patch_prober.go:28] interesting pod/router-default-5444994796-h82f7 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 14 07:03:46 crc kubenswrapper[4870]: [-]has-synced failed: reason withheld Oct 14 07:03:46 crc kubenswrapper[4870]: [+]process-running ok Oct 14 07:03:46 crc kubenswrapper[4870]: healthz check failed Oct 14 07:03:46 crc kubenswrapper[4870]: I1014 07:03:46.088422 4870 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-h82f7" podUID="28a417a5-b0b0-465a-ade0-3bf3ceb5ebdd" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 14 07:03:47 crc kubenswrapper[4870]: I1014 07:03:47.088601 4870 patch_prober.go:28] interesting pod/router-default-5444994796-h82f7 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 14 07:03:47 crc kubenswrapper[4870]: [-]has-synced failed: reason withheld Oct 14 07:03:47 crc kubenswrapper[4870]: [+]process-running ok Oct 14 07:03:47 crc kubenswrapper[4870]: healthz check failed Oct 14 07:03:47 crc kubenswrapper[4870]: I1014 07:03:47.088698 4870 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-h82f7" podUID="28a417a5-b0b0-465a-ade0-3bf3ceb5ebdd" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 14 07:03:48 crc kubenswrapper[4870]: I1014 07:03:48.087356 4870 patch_prober.go:28] interesting pod/router-default-5444994796-h82f7 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 14 07:03:48 crc kubenswrapper[4870]: [-]has-synced failed: reason withheld Oct 14 07:03:48 crc kubenswrapper[4870]: [+]process-running ok Oct 14 07:03:48 crc kubenswrapper[4870]: healthz check failed Oct 14 07:03:48 crc kubenswrapper[4870]: I1014 07:03:48.088028 4870 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-h82f7" podUID="28a417a5-b0b0-465a-ade0-3bf3ceb5ebdd" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 14 07:03:48 crc kubenswrapper[4870]: I1014 07:03:48.200735 4870 patch_prober.go:28] interesting pod/console-f9d7485db-r9t2k container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.8:8443/health\": dial tcp 10.217.0.8:8443: connect: connection refused" start-of-body= Oct 14 07:03:48 crc kubenswrapper[4870]: I1014 07:03:48.200853 4870 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-r9t2k" podUID="b38b0739-21a6-4050-af6c-6accabefcc45" containerName="console" probeResult="failure" output="Get \"https://10.217.0.8:8443/health\": dial tcp 10.217.0.8:8443: connect: connection refused" Oct 14 07:03:48 crc kubenswrapper[4870]: I1014 07:03:48.791060 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7d7a4f41-388a-4320-b896-43c17ff13da6-metrics-certs\") pod \"network-metrics-daemon-zc5j6\" (UID: \"7d7a4f41-388a-4320-b896-43c17ff13da6\") " pod="openshift-multus/network-metrics-daemon-zc5j6" Oct 14 07:03:48 crc kubenswrapper[4870]: I1014 07:03:48.798689 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7d7a4f41-388a-4320-b896-43c17ff13da6-metrics-certs\") pod \"network-metrics-daemon-zc5j6\" (UID: \"7d7a4f41-388a-4320-b896-43c17ff13da6\") " pod="openshift-multus/network-metrics-daemon-zc5j6" Oct 14 07:03:48 crc kubenswrapper[4870]: I1014 07:03:48.992987 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zc5j6" Oct 14 07:03:49 crc kubenswrapper[4870]: I1014 07:03:49.089685 4870 patch_prober.go:28] interesting pod/router-default-5444994796-h82f7 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 14 07:03:49 crc kubenswrapper[4870]: [-]has-synced failed: reason withheld Oct 14 07:03:49 crc kubenswrapper[4870]: [+]process-running ok Oct 14 07:03:49 crc kubenswrapper[4870]: healthz check failed Oct 14 07:03:49 crc kubenswrapper[4870]: I1014 07:03:49.089780 4870 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-h82f7" podUID="28a417a5-b0b0-465a-ade0-3bf3ceb5ebdd" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 14 07:03:49 crc kubenswrapper[4870]: I1014 07:03:49.277560 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-d7cdf" Oct 14 07:03:50 crc kubenswrapper[4870]: I1014 07:03:50.087224 4870 patch_prober.go:28] interesting pod/router-default-5444994796-h82f7 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 14 07:03:50 crc kubenswrapper[4870]: [-]has-synced failed: reason withheld Oct 14 07:03:50 crc kubenswrapper[4870]: [+]process-running ok Oct 14 07:03:50 crc kubenswrapper[4870]: healthz check failed Oct 14 07:03:50 crc kubenswrapper[4870]: I1014 07:03:50.087277 4870 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-h82f7" podUID="28a417a5-b0b0-465a-ade0-3bf3ceb5ebdd" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 14 07:03:50 crc kubenswrapper[4870]: I1014 07:03:50.869615 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"9e620080-d3ad-49dc-af0f-8dd64485e9f7","Type":"ContainerStarted","Data":"2f9c575519260f3718387babb98881fd952dfbbee1b84cb29ca6ff7d31eca9b1"} Oct 14 07:03:51 crc kubenswrapper[4870]: I1014 07:03:51.086707 4870 patch_prober.go:28] interesting pod/router-default-5444994796-h82f7 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 14 07:03:51 crc kubenswrapper[4870]: [-]has-synced failed: reason withheld Oct 14 07:03:51 crc kubenswrapper[4870]: [+]process-running ok Oct 14 07:03:51 crc kubenswrapper[4870]: healthz check failed Oct 14 07:03:51 crc kubenswrapper[4870]: I1014 07:03:51.086805 4870 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-h82f7" podUID="28a417a5-b0b0-465a-ade0-3bf3ceb5ebdd" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 14 07:03:52 crc kubenswrapper[4870]: I1014 07:03:52.103902 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-h82f7" Oct 14 07:03:52 crc kubenswrapper[4870]: I1014 07:03:52.109214 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-h82f7" Oct 14 07:03:53 crc kubenswrapper[4870]: I1014 07:03:53.950879 4870 patch_prober.go:28] interesting pod/machine-config-daemon-7tvc7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 07:03:53 crc kubenswrapper[4870]: I1014 07:03:53.950940 4870 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 07:03:57 crc kubenswrapper[4870]: I1014 07:03:57.530454 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-kkxjz" Oct 14 07:03:58 crc kubenswrapper[4870]: I1014 07:03:58.203006 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-r9t2k" Oct 14 07:03:58 crc kubenswrapper[4870]: I1014 07:03:58.207698 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-r9t2k" Oct 14 07:04:05 crc kubenswrapper[4870]: E1014 07:04:05.402891 4870 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Oct 14 07:04:05 crc kubenswrapper[4870]: E1014 07:04:05.404306 4870 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-mmllz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-kxxk5_openshift-marketplace(d56357b5-cdfe-4903-b59a-4ba8a111883f): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 14 07:04:05 crc kubenswrapper[4870]: E1014 07:04:05.405680 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-kxxk5" podUID="d56357b5-cdfe-4903-b59a-4ba8a111883f" Oct 14 07:04:06 crc kubenswrapper[4870]: E1014 07:04:06.477850 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-kxxk5" podUID="d56357b5-cdfe-4903-b59a-4ba8a111883f" Oct 14 07:04:06 crc kubenswrapper[4870]: E1014 07:04:06.604649 4870 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Oct 14 07:04:06 crc kubenswrapper[4870]: E1014 07:04:06.605564 4870 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-pcqbb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-jhjmr_openshift-marketplace(44734ad6-0693-4d02-ba13-441850b5d23b): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 14 07:04:06 crc kubenswrapper[4870]: E1014 07:04:06.606786 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-jhjmr" podUID="44734ad6-0693-4d02-ba13-441850b5d23b" Oct 14 07:04:06 crc kubenswrapper[4870]: E1014 07:04:06.629503 4870 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Oct 14 07:04:06 crc kubenswrapper[4870]: E1014 07:04:06.629927 4870 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-7c7n6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-4mlsh_openshift-marketplace(dbace338-bfac-4c71-a6bf-28227f826098): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 14 07:04:06 crc kubenswrapper[4870]: E1014 07:04:06.637079 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-4mlsh" podUID="dbace338-bfac-4c71-a6bf-28227f826098" Oct 14 07:04:06 crc kubenswrapper[4870]: E1014 07:04:06.667536 4870 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Oct 14 07:04:06 crc kubenswrapper[4870]: E1014 07:04:06.667812 4870 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-cdzjs,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-lkfh6_openshift-marketplace(44a6ae4c-9030-4a51-8960-0c0cc6c59b8e): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 14 07:04:06 crc kubenswrapper[4870]: E1014 07:04:06.669008 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-lkfh6" podUID="44a6ae4c-9030-4a51-8960-0c0cc6c59b8e" Oct 14 07:04:06 crc kubenswrapper[4870]: E1014 07:04:06.679043 4870 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Oct 14 07:04:06 crc kubenswrapper[4870]: E1014 07:04:06.679218 4870 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-wwphs,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-cv6fw_openshift-marketplace(ea907a55-c1d5-4f49-9ec4-c97461eef014): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 14 07:04:06 crc kubenswrapper[4870]: E1014 07:04:06.680403 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-cv6fw" podUID="ea907a55-c1d5-4f49-9ec4-c97461eef014" Oct 14 07:04:06 crc kubenswrapper[4870]: I1014 07:04:06.969296 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8qgzx" event={"ID":"5827db23-a022-4d5f-9ae0-ef3505610bce","Type":"ContainerStarted","Data":"73a091b3e3da99fa9b682697887f8c24631ba40dc8eec83748bbcdfb6609a332"} Oct 14 07:04:06 crc kubenswrapper[4870]: I1014 07:04:06.972903 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2glwc" event={"ID":"788a3634-1dce-4c12-979b-235c74c9200b","Type":"ContainerStarted","Data":"2bea0f06b4f2f68aea3adf931c6f311ab677da53ac19e3f4646a02217929d7a6"} Oct 14 07:04:06 crc kubenswrapper[4870]: I1014 07:04:06.981924 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-zc5j6"] Oct 14 07:04:06 crc kubenswrapper[4870]: I1014 07:04:06.983498 4870 generic.go:334] "Generic (PLEG): container finished" podID="36521dfb-3f0f-4a55-a393-bdf3ec7d3213" containerID="b070a0e16ab1bba9643b460e46c67d81f30b3e04a27b310b8900dcac49e67842" exitCode=0 Oct 14 07:04:06 crc kubenswrapper[4870]: I1014 07:04:06.984007 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6zj22" event={"ID":"36521dfb-3f0f-4a55-a393-bdf3ec7d3213","Type":"ContainerDied","Data":"b070a0e16ab1bba9643b460e46c67d81f30b3e04a27b310b8900dcac49e67842"} Oct 14 07:04:06 crc kubenswrapper[4870]: E1014 07:04:06.986935 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-jhjmr" podUID="44734ad6-0693-4d02-ba13-441850b5d23b" Oct 14 07:04:06 crc kubenswrapper[4870]: E1014 07:04:06.986991 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-cv6fw" podUID="ea907a55-c1d5-4f49-9ec4-c97461eef014" Oct 14 07:04:06 crc kubenswrapper[4870]: E1014 07:04:06.987092 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-lkfh6" podUID="44a6ae4c-9030-4a51-8960-0c0cc6c59b8e" Oct 14 07:04:06 crc kubenswrapper[4870]: E1014 07:04:06.987144 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-4mlsh" podUID="dbace338-bfac-4c71-a6bf-28227f826098" Oct 14 07:04:07 crc kubenswrapper[4870]: I1014 07:04:07.996024 4870 generic.go:334] "Generic (PLEG): container finished" podID="9e620080-d3ad-49dc-af0f-8dd64485e9f7" containerID="cbbd45a5773d36fa37f4de1d7627f30d35490a82923ec70bc6578a26696ffce2" exitCode=0 Oct 14 07:04:07 crc kubenswrapper[4870]: I1014 07:04:07.996126 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"9e620080-d3ad-49dc-af0f-8dd64485e9f7","Type":"ContainerDied","Data":"cbbd45a5773d36fa37f4de1d7627f30d35490a82923ec70bc6578a26696ffce2"} Oct 14 07:04:08 crc kubenswrapper[4870]: I1014 07:04:08.001697 4870 generic.go:334] "Generic (PLEG): container finished" podID="788a3634-1dce-4c12-979b-235c74c9200b" containerID="2bea0f06b4f2f68aea3adf931c6f311ab677da53ac19e3f4646a02217929d7a6" exitCode=0 Oct 14 07:04:08 crc kubenswrapper[4870]: I1014 07:04:08.001822 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2glwc" event={"ID":"788a3634-1dce-4c12-979b-235c74c9200b","Type":"ContainerDied","Data":"2bea0f06b4f2f68aea3adf931c6f311ab677da53ac19e3f4646a02217929d7a6"} Oct 14 07:04:08 crc kubenswrapper[4870]: I1014 07:04:08.010101 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6zj22" event={"ID":"36521dfb-3f0f-4a55-a393-bdf3ec7d3213","Type":"ContainerStarted","Data":"01a290c61289191563d98907f667a32e6af1ce0a7fa897a085a70051a6252929"} Oct 14 07:04:08 crc kubenswrapper[4870]: I1014 07:04:08.012881 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-zc5j6" event={"ID":"7d7a4f41-388a-4320-b896-43c17ff13da6","Type":"ContainerStarted","Data":"53caf466c955996532b77c075cc91656525540a3a04e07504d52bbae1291c17e"} Oct 14 07:04:08 crc kubenswrapper[4870]: I1014 07:04:08.013091 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-zc5j6" event={"ID":"7d7a4f41-388a-4320-b896-43c17ff13da6","Type":"ContainerStarted","Data":"05a452cd49f8be1879fb99e56662dccdb0790c2466e7476cec42952e0d12169e"} Oct 14 07:04:08 crc kubenswrapper[4870]: I1014 07:04:08.013111 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-zc5j6" event={"ID":"7d7a4f41-388a-4320-b896-43c17ff13da6","Type":"ContainerStarted","Data":"6aff0d0a3f08fae31d5b25d9487611881f872ef2c676673cdd37bcbeba2c41db"} Oct 14 07:04:08 crc kubenswrapper[4870]: I1014 07:04:08.022604 4870 generic.go:334] "Generic (PLEG): container finished" podID="5827db23-a022-4d5f-9ae0-ef3505610bce" containerID="73a091b3e3da99fa9b682697887f8c24631ba40dc8eec83748bbcdfb6609a332" exitCode=0 Oct 14 07:04:08 crc kubenswrapper[4870]: I1014 07:04:08.022662 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8qgzx" event={"ID":"5827db23-a022-4d5f-9ae0-ef3505610bce","Type":"ContainerDied","Data":"73a091b3e3da99fa9b682697887f8c24631ba40dc8eec83748bbcdfb6609a332"} Oct 14 07:04:08 crc kubenswrapper[4870]: I1014 07:04:08.062589 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-6zj22" podStartSLOduration=3.034590261 podStartE2EDuration="32.06256199s" podCreationTimestamp="2025-10-14 07:03:36 +0000 UTC" firstStartedPulling="2025-10-14 07:03:38.624850433 +0000 UTC m=+154.322210804" lastFinishedPulling="2025-10-14 07:04:07.652822152 +0000 UTC m=+183.350182533" observedRunningTime="2025-10-14 07:04:08.060701185 +0000 UTC m=+183.758061566" watchObservedRunningTime="2025-10-14 07:04:08.06256199 +0000 UTC m=+183.759922371" Oct 14 07:04:08 crc kubenswrapper[4870]: I1014 07:04:08.079610 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-zc5j6" podStartSLOduration=162.079593053 podStartE2EDuration="2m42.079593053s" podCreationTimestamp="2025-10-14 07:01:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:04:08.078406104 +0000 UTC m=+183.775766495" watchObservedRunningTime="2025-10-14 07:04:08.079593053 +0000 UTC m=+183.776953414" Oct 14 07:04:09 crc kubenswrapper[4870]: I1014 07:04:09.032661 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8qgzx" event={"ID":"5827db23-a022-4d5f-9ae0-ef3505610bce","Type":"ContainerStarted","Data":"5dd81be36b471757840ffb2df36af328b125b07813302e7674080bc843e11bac"} Oct 14 07:04:09 crc kubenswrapper[4870]: I1014 07:04:09.044796 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2glwc" event={"ID":"788a3634-1dce-4c12-979b-235c74c9200b","Type":"ContainerStarted","Data":"df2f460a8e291f154c4ad2a01a01f082c7cf05d6e3201aecd6116f18e97a8d2b"} Oct 14 07:04:09 crc kubenswrapper[4870]: I1014 07:04:09.062579 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-8qgzx" podStartSLOduration=2.985999536 podStartE2EDuration="35.062558894s" podCreationTimestamp="2025-10-14 07:03:34 +0000 UTC" firstStartedPulling="2025-10-14 07:03:36.506225629 +0000 UTC m=+152.203585990" lastFinishedPulling="2025-10-14 07:04:08.582784977 +0000 UTC m=+184.280145348" observedRunningTime="2025-10-14 07:04:09.061545269 +0000 UTC m=+184.758905660" watchObservedRunningTime="2025-10-14 07:04:09.062558894 +0000 UTC m=+184.759919265" Oct 14 07:04:09 crc kubenswrapper[4870]: I1014 07:04:09.086685 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-2glwc" podStartSLOduration=3.201133733 podStartE2EDuration="35.086665068s" podCreationTimestamp="2025-10-14 07:03:34 +0000 UTC" firstStartedPulling="2025-10-14 07:03:36.549684013 +0000 UTC m=+152.247044374" lastFinishedPulling="2025-10-14 07:04:08.435215338 +0000 UTC m=+184.132575709" observedRunningTime="2025-10-14 07:04:09.080759475 +0000 UTC m=+184.778119846" watchObservedRunningTime="2025-10-14 07:04:09.086665068 +0000 UTC m=+184.784025439" Oct 14 07:04:09 crc kubenswrapper[4870]: I1014 07:04:09.473059 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 14 07:04:09 crc kubenswrapper[4870]: I1014 07:04:09.516152 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-52qbc" Oct 14 07:04:09 crc kubenswrapper[4870]: I1014 07:04:09.554326 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9e620080-d3ad-49dc-af0f-8dd64485e9f7-kube-api-access\") pod \"9e620080-d3ad-49dc-af0f-8dd64485e9f7\" (UID: \"9e620080-d3ad-49dc-af0f-8dd64485e9f7\") " Oct 14 07:04:09 crc kubenswrapper[4870]: I1014 07:04:09.554483 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9e620080-d3ad-49dc-af0f-8dd64485e9f7-kubelet-dir\") pod \"9e620080-d3ad-49dc-af0f-8dd64485e9f7\" (UID: \"9e620080-d3ad-49dc-af0f-8dd64485e9f7\") " Oct 14 07:04:09 crc kubenswrapper[4870]: I1014 07:04:09.554867 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9e620080-d3ad-49dc-af0f-8dd64485e9f7-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "9e620080-d3ad-49dc-af0f-8dd64485e9f7" (UID: "9e620080-d3ad-49dc-af0f-8dd64485e9f7"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 14 07:04:09 crc kubenswrapper[4870]: I1014 07:04:09.566664 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9e620080-d3ad-49dc-af0f-8dd64485e9f7-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "9e620080-d3ad-49dc-af0f-8dd64485e9f7" (UID: "9e620080-d3ad-49dc-af0f-8dd64485e9f7"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:04:09 crc kubenswrapper[4870]: I1014 07:04:09.656234 4870 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9e620080-d3ad-49dc-af0f-8dd64485e9f7-kubelet-dir\") on node \"crc\" DevicePath \"\"" Oct 14 07:04:09 crc kubenswrapper[4870]: I1014 07:04:09.656377 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9e620080-d3ad-49dc-af0f-8dd64485e9f7-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 14 07:04:10 crc kubenswrapper[4870]: I1014 07:04:10.051549 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"9e620080-d3ad-49dc-af0f-8dd64485e9f7","Type":"ContainerDied","Data":"2f9c575519260f3718387babb98881fd952dfbbee1b84cb29ca6ff7d31eca9b1"} Oct 14 07:04:10 crc kubenswrapper[4870]: I1014 07:04:10.051636 4870 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2f9c575519260f3718387babb98881fd952dfbbee1b84cb29ca6ff7d31eca9b1" Oct 14 07:04:10 crc kubenswrapper[4870]: I1014 07:04:10.051724 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 14 07:04:14 crc kubenswrapper[4870]: I1014 07:04:14.312120 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 14 07:04:14 crc kubenswrapper[4870]: I1014 07:04:14.756308 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-2glwc" Oct 14 07:04:14 crc kubenswrapper[4870]: I1014 07:04:14.756583 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-2glwc" Oct 14 07:04:14 crc kubenswrapper[4870]: I1014 07:04:14.934176 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-2glwc" Oct 14 07:04:15 crc kubenswrapper[4870]: I1014 07:04:15.005364 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-8qgzx" Oct 14 07:04:15 crc kubenswrapper[4870]: I1014 07:04:15.005512 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-8qgzx" Oct 14 07:04:15 crc kubenswrapper[4870]: I1014 07:04:15.065043 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-8qgzx" Oct 14 07:04:15 crc kubenswrapper[4870]: I1014 07:04:15.150546 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-2glwc" Oct 14 07:04:15 crc kubenswrapper[4870]: I1014 07:04:15.160961 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-8qgzx" Oct 14 07:04:16 crc kubenswrapper[4870]: I1014 07:04:16.170350 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-8qgzx"] Oct 14 07:04:16 crc kubenswrapper[4870]: I1014 07:04:16.925332 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-6zj22" Oct 14 07:04:16 crc kubenswrapper[4870]: I1014 07:04:16.925977 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-6zj22" Oct 14 07:04:16 crc kubenswrapper[4870]: I1014 07:04:16.983066 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-6zj22" Oct 14 07:04:17 crc kubenswrapper[4870]: I1014 07:04:17.103834 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-8qgzx" podUID="5827db23-a022-4d5f-9ae0-ef3505610bce" containerName="registry-server" containerID="cri-o://5dd81be36b471757840ffb2df36af328b125b07813302e7674080bc843e11bac" gracePeriod=2 Oct 14 07:04:17 crc kubenswrapper[4870]: I1014 07:04:17.157796 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-6zj22" Oct 14 07:04:17 crc kubenswrapper[4870]: I1014 07:04:17.642161 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8qgzx" Oct 14 07:04:17 crc kubenswrapper[4870]: I1014 07:04:17.795613 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rkhl8\" (UniqueName: \"kubernetes.io/projected/5827db23-a022-4d5f-9ae0-ef3505610bce-kube-api-access-rkhl8\") pod \"5827db23-a022-4d5f-9ae0-ef3505610bce\" (UID: \"5827db23-a022-4d5f-9ae0-ef3505610bce\") " Oct 14 07:04:17 crc kubenswrapper[4870]: I1014 07:04:17.795795 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5827db23-a022-4d5f-9ae0-ef3505610bce-catalog-content\") pod \"5827db23-a022-4d5f-9ae0-ef3505610bce\" (UID: \"5827db23-a022-4d5f-9ae0-ef3505610bce\") " Oct 14 07:04:17 crc kubenswrapper[4870]: I1014 07:04:17.795881 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5827db23-a022-4d5f-9ae0-ef3505610bce-utilities\") pod \"5827db23-a022-4d5f-9ae0-ef3505610bce\" (UID: \"5827db23-a022-4d5f-9ae0-ef3505610bce\") " Oct 14 07:04:17 crc kubenswrapper[4870]: I1014 07:04:17.796981 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5827db23-a022-4d5f-9ae0-ef3505610bce-utilities" (OuterVolumeSpecName: "utilities") pod "5827db23-a022-4d5f-9ae0-ef3505610bce" (UID: "5827db23-a022-4d5f-9ae0-ef3505610bce"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:04:17 crc kubenswrapper[4870]: I1014 07:04:17.803908 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5827db23-a022-4d5f-9ae0-ef3505610bce-kube-api-access-rkhl8" (OuterVolumeSpecName: "kube-api-access-rkhl8") pod "5827db23-a022-4d5f-9ae0-ef3505610bce" (UID: "5827db23-a022-4d5f-9ae0-ef3505610bce"). InnerVolumeSpecName "kube-api-access-rkhl8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:04:17 crc kubenswrapper[4870]: I1014 07:04:17.868520 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5827db23-a022-4d5f-9ae0-ef3505610bce-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5827db23-a022-4d5f-9ae0-ef3505610bce" (UID: "5827db23-a022-4d5f-9ae0-ef3505610bce"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:04:17 crc kubenswrapper[4870]: I1014 07:04:17.897405 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rkhl8\" (UniqueName: \"kubernetes.io/projected/5827db23-a022-4d5f-9ae0-ef3505610bce-kube-api-access-rkhl8\") on node \"crc\" DevicePath \"\"" Oct 14 07:04:17 crc kubenswrapper[4870]: I1014 07:04:17.897476 4870 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5827db23-a022-4d5f-9ae0-ef3505610bce-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 07:04:17 crc kubenswrapper[4870]: I1014 07:04:17.897495 4870 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5827db23-a022-4d5f-9ae0-ef3505610bce-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 07:04:18 crc kubenswrapper[4870]: I1014 07:04:18.113853 4870 generic.go:334] "Generic (PLEG): container finished" podID="5827db23-a022-4d5f-9ae0-ef3505610bce" containerID="5dd81be36b471757840ffb2df36af328b125b07813302e7674080bc843e11bac" exitCode=0 Oct 14 07:04:18 crc kubenswrapper[4870]: I1014 07:04:18.113917 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8qgzx" event={"ID":"5827db23-a022-4d5f-9ae0-ef3505610bce","Type":"ContainerDied","Data":"5dd81be36b471757840ffb2df36af328b125b07813302e7674080bc843e11bac"} Oct 14 07:04:18 crc kubenswrapper[4870]: I1014 07:04:18.113974 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8qgzx" event={"ID":"5827db23-a022-4d5f-9ae0-ef3505610bce","Type":"ContainerDied","Data":"8e7c0eb7785db8f50e2b2ccb1b227b026ca5cb14a3b49cface9784094c4d64d6"} Oct 14 07:04:18 crc kubenswrapper[4870]: I1014 07:04:18.114002 4870 scope.go:117] "RemoveContainer" containerID="5dd81be36b471757840ffb2df36af328b125b07813302e7674080bc843e11bac" Oct 14 07:04:18 crc kubenswrapper[4870]: I1014 07:04:18.114026 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8qgzx" Oct 14 07:04:18 crc kubenswrapper[4870]: I1014 07:04:18.161044 4870 scope.go:117] "RemoveContainer" containerID="73a091b3e3da99fa9b682697887f8c24631ba40dc8eec83748bbcdfb6609a332" Oct 14 07:04:18 crc kubenswrapper[4870]: I1014 07:04:18.179152 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-8qgzx"] Oct 14 07:04:18 crc kubenswrapper[4870]: I1014 07:04:18.184381 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-8qgzx"] Oct 14 07:04:18 crc kubenswrapper[4870]: I1014 07:04:18.198866 4870 scope.go:117] "RemoveContainer" containerID="9faa5924aeef603e9561d951bb02c850d2bfb078766f3119284528ddd04605d9" Oct 14 07:04:18 crc kubenswrapper[4870]: I1014 07:04:18.253911 4870 scope.go:117] "RemoveContainer" containerID="5dd81be36b471757840ffb2df36af328b125b07813302e7674080bc843e11bac" Oct 14 07:04:18 crc kubenswrapper[4870]: E1014 07:04:18.254627 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5dd81be36b471757840ffb2df36af328b125b07813302e7674080bc843e11bac\": container with ID starting with 5dd81be36b471757840ffb2df36af328b125b07813302e7674080bc843e11bac not found: ID does not exist" containerID="5dd81be36b471757840ffb2df36af328b125b07813302e7674080bc843e11bac" Oct 14 07:04:18 crc kubenswrapper[4870]: I1014 07:04:18.254691 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5dd81be36b471757840ffb2df36af328b125b07813302e7674080bc843e11bac"} err="failed to get container status \"5dd81be36b471757840ffb2df36af328b125b07813302e7674080bc843e11bac\": rpc error: code = NotFound desc = could not find container \"5dd81be36b471757840ffb2df36af328b125b07813302e7674080bc843e11bac\": container with ID starting with 5dd81be36b471757840ffb2df36af328b125b07813302e7674080bc843e11bac not found: ID does not exist" Oct 14 07:04:18 crc kubenswrapper[4870]: I1014 07:04:18.254772 4870 scope.go:117] "RemoveContainer" containerID="73a091b3e3da99fa9b682697887f8c24631ba40dc8eec83748bbcdfb6609a332" Oct 14 07:04:18 crc kubenswrapper[4870]: E1014 07:04:18.255426 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"73a091b3e3da99fa9b682697887f8c24631ba40dc8eec83748bbcdfb6609a332\": container with ID starting with 73a091b3e3da99fa9b682697887f8c24631ba40dc8eec83748bbcdfb6609a332 not found: ID does not exist" containerID="73a091b3e3da99fa9b682697887f8c24631ba40dc8eec83748bbcdfb6609a332" Oct 14 07:04:18 crc kubenswrapper[4870]: I1014 07:04:18.255487 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"73a091b3e3da99fa9b682697887f8c24631ba40dc8eec83748bbcdfb6609a332"} err="failed to get container status \"73a091b3e3da99fa9b682697887f8c24631ba40dc8eec83748bbcdfb6609a332\": rpc error: code = NotFound desc = could not find container \"73a091b3e3da99fa9b682697887f8c24631ba40dc8eec83748bbcdfb6609a332\": container with ID starting with 73a091b3e3da99fa9b682697887f8c24631ba40dc8eec83748bbcdfb6609a332 not found: ID does not exist" Oct 14 07:04:18 crc kubenswrapper[4870]: I1014 07:04:18.255514 4870 scope.go:117] "RemoveContainer" containerID="9faa5924aeef603e9561d951bb02c850d2bfb078766f3119284528ddd04605d9" Oct 14 07:04:18 crc kubenswrapper[4870]: E1014 07:04:18.256117 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9faa5924aeef603e9561d951bb02c850d2bfb078766f3119284528ddd04605d9\": container with ID starting with 9faa5924aeef603e9561d951bb02c850d2bfb078766f3119284528ddd04605d9 not found: ID does not exist" containerID="9faa5924aeef603e9561d951bb02c850d2bfb078766f3119284528ddd04605d9" Oct 14 07:04:18 crc kubenswrapper[4870]: I1014 07:04:18.256199 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9faa5924aeef603e9561d951bb02c850d2bfb078766f3119284528ddd04605d9"} err="failed to get container status \"9faa5924aeef603e9561d951bb02c850d2bfb078766f3119284528ddd04605d9\": rpc error: code = NotFound desc = could not find container \"9faa5924aeef603e9561d951bb02c850d2bfb078766f3119284528ddd04605d9\": container with ID starting with 9faa5924aeef603e9561d951bb02c850d2bfb078766f3119284528ddd04605d9 not found: ID does not exist" Oct 14 07:04:19 crc kubenswrapper[4870]: I1014 07:04:19.042836 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5827db23-a022-4d5f-9ae0-ef3505610bce" path="/var/lib/kubelet/pods/5827db23-a022-4d5f-9ae0-ef3505610bce/volumes" Oct 14 07:04:19 crc kubenswrapper[4870]: I1014 07:04:19.123758 4870 generic.go:334] "Generic (PLEG): container finished" podID="d56357b5-cdfe-4903-b59a-4ba8a111883f" containerID="af8039048df663104e531d0872471589a53728e5495094cfdc0dd92a2c172f4b" exitCode=0 Oct 14 07:04:19 crc kubenswrapper[4870]: I1014 07:04:19.123826 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kxxk5" event={"ID":"d56357b5-cdfe-4903-b59a-4ba8a111883f","Type":"ContainerDied","Data":"af8039048df663104e531d0872471589a53728e5495094cfdc0dd92a2c172f4b"} Oct 14 07:04:19 crc kubenswrapper[4870]: I1014 07:04:19.567826 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-6zj22"] Oct 14 07:04:19 crc kubenswrapper[4870]: I1014 07:04:19.568209 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-6zj22" podUID="36521dfb-3f0f-4a55-a393-bdf3ec7d3213" containerName="registry-server" containerID="cri-o://01a290c61289191563d98907f667a32e6af1ce0a7fa897a085a70051a6252929" gracePeriod=2 Oct 14 07:04:20 crc kubenswrapper[4870]: I1014 07:04:20.058762 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6zj22" Oct 14 07:04:20 crc kubenswrapper[4870]: I1014 07:04:20.129862 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/36521dfb-3f0f-4a55-a393-bdf3ec7d3213-catalog-content\") pod \"36521dfb-3f0f-4a55-a393-bdf3ec7d3213\" (UID: \"36521dfb-3f0f-4a55-a393-bdf3ec7d3213\") " Oct 14 07:04:20 crc kubenswrapper[4870]: I1014 07:04:20.130002 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9bb8\" (UniqueName: \"kubernetes.io/projected/36521dfb-3f0f-4a55-a393-bdf3ec7d3213-kube-api-access-w9bb8\") pod \"36521dfb-3f0f-4a55-a393-bdf3ec7d3213\" (UID: \"36521dfb-3f0f-4a55-a393-bdf3ec7d3213\") " Oct 14 07:04:20 crc kubenswrapper[4870]: I1014 07:04:20.130113 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/36521dfb-3f0f-4a55-a393-bdf3ec7d3213-utilities\") pod \"36521dfb-3f0f-4a55-a393-bdf3ec7d3213\" (UID: \"36521dfb-3f0f-4a55-a393-bdf3ec7d3213\") " Oct 14 07:04:20 crc kubenswrapper[4870]: I1014 07:04:20.132259 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/36521dfb-3f0f-4a55-a393-bdf3ec7d3213-utilities" (OuterVolumeSpecName: "utilities") pod "36521dfb-3f0f-4a55-a393-bdf3ec7d3213" (UID: "36521dfb-3f0f-4a55-a393-bdf3ec7d3213"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:04:20 crc kubenswrapper[4870]: I1014 07:04:20.135415 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jhjmr" event={"ID":"44734ad6-0693-4d02-ba13-441850b5d23b","Type":"ContainerStarted","Data":"d892e031f81fb6e419b43be04dfed8455efaa1f7a9e93f8a9cdde09b7ffe30dc"} Oct 14 07:04:20 crc kubenswrapper[4870]: I1014 07:04:20.138601 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/36521dfb-3f0f-4a55-a393-bdf3ec7d3213-kube-api-access-w9bb8" (OuterVolumeSpecName: "kube-api-access-w9bb8") pod "36521dfb-3f0f-4a55-a393-bdf3ec7d3213" (UID: "36521dfb-3f0f-4a55-a393-bdf3ec7d3213"). InnerVolumeSpecName "kube-api-access-w9bb8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:04:20 crc kubenswrapper[4870]: I1014 07:04:20.140160 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kxxk5" event={"ID":"d56357b5-cdfe-4903-b59a-4ba8a111883f","Type":"ContainerStarted","Data":"a78f8c819123955b9f3c42907be7d1693272c44a7868cbffaf7df4ef46675a5b"} Oct 14 07:04:20 crc kubenswrapper[4870]: I1014 07:04:20.144593 4870 generic.go:334] "Generic (PLEG): container finished" podID="36521dfb-3f0f-4a55-a393-bdf3ec7d3213" containerID="01a290c61289191563d98907f667a32e6af1ce0a7fa897a085a70051a6252929" exitCode=0 Oct 14 07:04:20 crc kubenswrapper[4870]: I1014 07:04:20.144668 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6zj22" event={"ID":"36521dfb-3f0f-4a55-a393-bdf3ec7d3213","Type":"ContainerDied","Data":"01a290c61289191563d98907f667a32e6af1ce0a7fa897a085a70051a6252929"} Oct 14 07:04:20 crc kubenswrapper[4870]: I1014 07:04:20.144710 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6zj22" event={"ID":"36521dfb-3f0f-4a55-a393-bdf3ec7d3213","Type":"ContainerDied","Data":"4ec1dcf5cc8e3b870865f8c445531382a2f81caa1de49469f13b74cc4e781eaf"} Oct 14 07:04:20 crc kubenswrapper[4870]: I1014 07:04:20.144713 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6zj22" Oct 14 07:04:20 crc kubenswrapper[4870]: I1014 07:04:20.144731 4870 scope.go:117] "RemoveContainer" containerID="01a290c61289191563d98907f667a32e6af1ce0a7fa897a085a70051a6252929" Oct 14 07:04:20 crc kubenswrapper[4870]: I1014 07:04:20.150411 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/36521dfb-3f0f-4a55-a393-bdf3ec7d3213-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "36521dfb-3f0f-4a55-a393-bdf3ec7d3213" (UID: "36521dfb-3f0f-4a55-a393-bdf3ec7d3213"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:04:20 crc kubenswrapper[4870]: I1014 07:04:20.164529 4870 scope.go:117] "RemoveContainer" containerID="b070a0e16ab1bba9643b460e46c67d81f30b3e04a27b310b8900dcac49e67842" Oct 14 07:04:20 crc kubenswrapper[4870]: I1014 07:04:20.189967 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-kxxk5" podStartSLOduration=3.142948092 podStartE2EDuration="43.189942433s" podCreationTimestamp="2025-10-14 07:03:37 +0000 UTC" firstStartedPulling="2025-10-14 07:03:39.687044225 +0000 UTC m=+155.384404596" lastFinishedPulling="2025-10-14 07:04:19.734038576 +0000 UTC m=+195.431398937" observedRunningTime="2025-10-14 07:04:20.184939932 +0000 UTC m=+195.882300303" watchObservedRunningTime="2025-10-14 07:04:20.189942433 +0000 UTC m=+195.887302804" Oct 14 07:04:20 crc kubenswrapper[4870]: I1014 07:04:20.200858 4870 scope.go:117] "RemoveContainer" containerID="d9109f8e72de4d832583b628a21e52c9a80d2b4c505348d5b3a36d6d26b903c3" Oct 14 07:04:20 crc kubenswrapper[4870]: I1014 07:04:20.216908 4870 scope.go:117] "RemoveContainer" containerID="01a290c61289191563d98907f667a32e6af1ce0a7fa897a085a70051a6252929" Oct 14 07:04:20 crc kubenswrapper[4870]: E1014 07:04:20.218201 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"01a290c61289191563d98907f667a32e6af1ce0a7fa897a085a70051a6252929\": container with ID starting with 01a290c61289191563d98907f667a32e6af1ce0a7fa897a085a70051a6252929 not found: ID does not exist" containerID="01a290c61289191563d98907f667a32e6af1ce0a7fa897a085a70051a6252929" Oct 14 07:04:20 crc kubenswrapper[4870]: I1014 07:04:20.218247 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"01a290c61289191563d98907f667a32e6af1ce0a7fa897a085a70051a6252929"} err="failed to get container status \"01a290c61289191563d98907f667a32e6af1ce0a7fa897a085a70051a6252929\": rpc error: code = NotFound desc = could not find container \"01a290c61289191563d98907f667a32e6af1ce0a7fa897a085a70051a6252929\": container with ID starting with 01a290c61289191563d98907f667a32e6af1ce0a7fa897a085a70051a6252929 not found: ID does not exist" Oct 14 07:04:20 crc kubenswrapper[4870]: I1014 07:04:20.218290 4870 scope.go:117] "RemoveContainer" containerID="b070a0e16ab1bba9643b460e46c67d81f30b3e04a27b310b8900dcac49e67842" Oct 14 07:04:20 crc kubenswrapper[4870]: E1014 07:04:20.218646 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b070a0e16ab1bba9643b460e46c67d81f30b3e04a27b310b8900dcac49e67842\": container with ID starting with b070a0e16ab1bba9643b460e46c67d81f30b3e04a27b310b8900dcac49e67842 not found: ID does not exist" containerID="b070a0e16ab1bba9643b460e46c67d81f30b3e04a27b310b8900dcac49e67842" Oct 14 07:04:20 crc kubenswrapper[4870]: I1014 07:04:20.218692 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b070a0e16ab1bba9643b460e46c67d81f30b3e04a27b310b8900dcac49e67842"} err="failed to get container status \"b070a0e16ab1bba9643b460e46c67d81f30b3e04a27b310b8900dcac49e67842\": rpc error: code = NotFound desc = could not find container \"b070a0e16ab1bba9643b460e46c67d81f30b3e04a27b310b8900dcac49e67842\": container with ID starting with b070a0e16ab1bba9643b460e46c67d81f30b3e04a27b310b8900dcac49e67842 not found: ID does not exist" Oct 14 07:04:20 crc kubenswrapper[4870]: I1014 07:04:20.218709 4870 scope.go:117] "RemoveContainer" containerID="d9109f8e72de4d832583b628a21e52c9a80d2b4c505348d5b3a36d6d26b903c3" Oct 14 07:04:20 crc kubenswrapper[4870]: E1014 07:04:20.218994 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d9109f8e72de4d832583b628a21e52c9a80d2b4c505348d5b3a36d6d26b903c3\": container with ID starting with d9109f8e72de4d832583b628a21e52c9a80d2b4c505348d5b3a36d6d26b903c3 not found: ID does not exist" containerID="d9109f8e72de4d832583b628a21e52c9a80d2b4c505348d5b3a36d6d26b903c3" Oct 14 07:04:20 crc kubenswrapper[4870]: I1014 07:04:20.219023 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d9109f8e72de4d832583b628a21e52c9a80d2b4c505348d5b3a36d6d26b903c3"} err="failed to get container status \"d9109f8e72de4d832583b628a21e52c9a80d2b4c505348d5b3a36d6d26b903c3\": rpc error: code = NotFound desc = could not find container \"d9109f8e72de4d832583b628a21e52c9a80d2b4c505348d5b3a36d6d26b903c3\": container with ID starting with d9109f8e72de4d832583b628a21e52c9a80d2b4c505348d5b3a36d6d26b903c3 not found: ID does not exist" Oct 14 07:04:20 crc kubenswrapper[4870]: I1014 07:04:20.232414 4870 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/36521dfb-3f0f-4a55-a393-bdf3ec7d3213-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 07:04:20 crc kubenswrapper[4870]: I1014 07:04:20.232473 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9bb8\" (UniqueName: \"kubernetes.io/projected/36521dfb-3f0f-4a55-a393-bdf3ec7d3213-kube-api-access-w9bb8\") on node \"crc\" DevicePath \"\"" Oct 14 07:04:20 crc kubenswrapper[4870]: I1014 07:04:20.232489 4870 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/36521dfb-3f0f-4a55-a393-bdf3ec7d3213-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 07:04:20 crc kubenswrapper[4870]: I1014 07:04:20.491930 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-6zj22"] Oct 14 07:04:20 crc kubenswrapper[4870]: I1014 07:04:20.495283 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-6zj22"] Oct 14 07:04:21 crc kubenswrapper[4870]: I1014 07:04:21.056070 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="36521dfb-3f0f-4a55-a393-bdf3ec7d3213" path="/var/lib/kubelet/pods/36521dfb-3f0f-4a55-a393-bdf3ec7d3213/volumes" Oct 14 07:04:21 crc kubenswrapper[4870]: I1014 07:04:21.154519 4870 generic.go:334] "Generic (PLEG): container finished" podID="44734ad6-0693-4d02-ba13-441850b5d23b" containerID="d892e031f81fb6e419b43be04dfed8455efaa1f7a9e93f8a9cdde09b7ffe30dc" exitCode=0 Oct 14 07:04:21 crc kubenswrapper[4870]: I1014 07:04:21.154605 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jhjmr" event={"ID":"44734ad6-0693-4d02-ba13-441850b5d23b","Type":"ContainerDied","Data":"d892e031f81fb6e419b43be04dfed8455efaa1f7a9e93f8a9cdde09b7ffe30dc"} Oct 14 07:04:22 crc kubenswrapper[4870]: I1014 07:04:22.162491 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lkfh6" event={"ID":"44a6ae4c-9030-4a51-8960-0c0cc6c59b8e","Type":"ContainerStarted","Data":"4c00d93fa663939b356aa33a19729c9f0fe545f382bf028defce467930cf5fc0"} Oct 14 07:04:22 crc kubenswrapper[4870]: I1014 07:04:22.165833 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jhjmr" event={"ID":"44734ad6-0693-4d02-ba13-441850b5d23b","Type":"ContainerStarted","Data":"060a45a1e2cef7c66b65624ead641a9c1743e5787bb9393f509e5f4dbb74f58e"} Oct 14 07:04:22 crc kubenswrapper[4870]: I1014 07:04:22.167523 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cv6fw" event={"ID":"ea907a55-c1d5-4f49-9ec4-c97461eef014","Type":"ContainerStarted","Data":"ae472138465eef062cd3eb0d658d045c6847481f1e81fa227d301eca12a74285"} Oct 14 07:04:22 crc kubenswrapper[4870]: I1014 07:04:22.286958 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-jhjmr" podStartSLOduration=3.145215942 podStartE2EDuration="48.286938163s" podCreationTimestamp="2025-10-14 07:03:34 +0000 UTC" firstStartedPulling="2025-10-14 07:03:36.508704449 +0000 UTC m=+152.206064820" lastFinishedPulling="2025-10-14 07:04:21.65042664 +0000 UTC m=+197.347787041" observedRunningTime="2025-10-14 07:04:22.284035851 +0000 UTC m=+197.981396232" watchObservedRunningTime="2025-10-14 07:04:22.286938163 +0000 UTC m=+197.984298544" Oct 14 07:04:23 crc kubenswrapper[4870]: I1014 07:04:23.178313 4870 generic.go:334] "Generic (PLEG): container finished" podID="44a6ae4c-9030-4a51-8960-0c0cc6c59b8e" containerID="4c00d93fa663939b356aa33a19729c9f0fe545f382bf028defce467930cf5fc0" exitCode=0 Oct 14 07:04:23 crc kubenswrapper[4870]: I1014 07:04:23.178378 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lkfh6" event={"ID":"44a6ae4c-9030-4a51-8960-0c0cc6c59b8e","Type":"ContainerDied","Data":"4c00d93fa663939b356aa33a19729c9f0fe545f382bf028defce467930cf5fc0"} Oct 14 07:04:23 crc kubenswrapper[4870]: I1014 07:04:23.181921 4870 generic.go:334] "Generic (PLEG): container finished" podID="dbace338-bfac-4c71-a6bf-28227f826098" containerID="a2cae3b3a1018bfbd9886537f54022c03285be666d1c80dda72afc6c5a12a553" exitCode=0 Oct 14 07:04:23 crc kubenswrapper[4870]: I1014 07:04:23.182005 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4mlsh" event={"ID":"dbace338-bfac-4c71-a6bf-28227f826098","Type":"ContainerDied","Data":"a2cae3b3a1018bfbd9886537f54022c03285be666d1c80dda72afc6c5a12a553"} Oct 14 07:04:23 crc kubenswrapper[4870]: I1014 07:04:23.185828 4870 generic.go:334] "Generic (PLEG): container finished" podID="ea907a55-c1d5-4f49-9ec4-c97461eef014" containerID="ae472138465eef062cd3eb0d658d045c6847481f1e81fa227d301eca12a74285" exitCode=0 Oct 14 07:04:23 crc kubenswrapper[4870]: I1014 07:04:23.185858 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cv6fw" event={"ID":"ea907a55-c1d5-4f49-9ec4-c97461eef014","Type":"ContainerDied","Data":"ae472138465eef062cd3eb0d658d045c6847481f1e81fa227d301eca12a74285"} Oct 14 07:04:23 crc kubenswrapper[4870]: I1014 07:04:23.950725 4870 patch_prober.go:28] interesting pod/machine-config-daemon-7tvc7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 07:04:23 crc kubenswrapper[4870]: I1014 07:04:23.950819 4870 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 07:04:24 crc kubenswrapper[4870]: I1014 07:04:24.192528 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lkfh6" event={"ID":"44a6ae4c-9030-4a51-8960-0c0cc6c59b8e","Type":"ContainerStarted","Data":"6877f0d90eec3c12ee4fa0bbf5b1898fa54252e8355751b8625687bde66e00d0"} Oct 14 07:04:24 crc kubenswrapper[4870]: I1014 07:04:24.195402 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4mlsh" event={"ID":"dbace338-bfac-4c71-a6bf-28227f826098","Type":"ContainerStarted","Data":"23818a41b9012235bf36fd3c9caca4884535c22691ee73d7291e085758a58b2b"} Oct 14 07:04:24 crc kubenswrapper[4870]: I1014 07:04:24.197947 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cv6fw" event={"ID":"ea907a55-c1d5-4f49-9ec4-c97461eef014","Type":"ContainerStarted","Data":"bd3fda2a0804a806b60fb4b84cba455549812fa02c463419f70ece38ebae2eaf"} Oct 14 07:04:24 crc kubenswrapper[4870]: I1014 07:04:24.214859 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-lkfh6" podStartSLOduration=3.073328472 podStartE2EDuration="50.214839585s" podCreationTimestamp="2025-10-14 07:03:34 +0000 UTC" firstStartedPulling="2025-10-14 07:03:36.564544233 +0000 UTC m=+152.261904604" lastFinishedPulling="2025-10-14 07:04:23.706055356 +0000 UTC m=+199.403415717" observedRunningTime="2025-10-14 07:04:24.211643585 +0000 UTC m=+199.909003956" watchObservedRunningTime="2025-10-14 07:04:24.214839585 +0000 UTC m=+199.912199956" Oct 14 07:04:24 crc kubenswrapper[4870]: I1014 07:04:24.237191 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-4mlsh" podStartSLOduration=2.079860088 podStartE2EDuration="48.237165303s" podCreationTimestamp="2025-10-14 07:03:36 +0000 UTC" firstStartedPulling="2025-10-14 07:03:37.591920931 +0000 UTC m=+153.289281302" lastFinishedPulling="2025-10-14 07:04:23.749226146 +0000 UTC m=+199.446586517" observedRunningTime="2025-10-14 07:04:24.234709282 +0000 UTC m=+199.932069653" watchObservedRunningTime="2025-10-14 07:04:24.237165303 +0000 UTC m=+199.934525674" Oct 14 07:04:24 crc kubenswrapper[4870]: I1014 07:04:24.259233 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-cv6fw" podStartSLOduration=3.296859047 podStartE2EDuration="47.259211615s" podCreationTimestamp="2025-10-14 07:03:37 +0000 UTC" firstStartedPulling="2025-10-14 07:03:39.718685132 +0000 UTC m=+155.416045503" lastFinishedPulling="2025-10-14 07:04:23.68103769 +0000 UTC m=+199.378398071" observedRunningTime="2025-10-14 07:04:24.259122922 +0000 UTC m=+199.956483293" watchObservedRunningTime="2025-10-14 07:04:24.259211615 +0000 UTC m=+199.956571986" Oct 14 07:04:24 crc kubenswrapper[4870]: I1014 07:04:24.512477 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-jhjmr" Oct 14 07:04:24 crc kubenswrapper[4870]: I1014 07:04:24.512535 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-jhjmr" Oct 14 07:04:25 crc kubenswrapper[4870]: I1014 07:04:25.151563 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-lkfh6" Oct 14 07:04:25 crc kubenswrapper[4870]: I1014 07:04:25.151752 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-lkfh6" Oct 14 07:04:25 crc kubenswrapper[4870]: I1014 07:04:25.563917 4870 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-jhjmr" podUID="44734ad6-0693-4d02-ba13-441850b5d23b" containerName="registry-server" probeResult="failure" output=< Oct 14 07:04:25 crc kubenswrapper[4870]: timeout: failed to connect service ":50051" within 1s Oct 14 07:04:25 crc kubenswrapper[4870]: > Oct 14 07:04:26 crc kubenswrapper[4870]: I1014 07:04:26.189774 4870 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-lkfh6" podUID="44a6ae4c-9030-4a51-8960-0c0cc6c59b8e" containerName="registry-server" probeResult="failure" output=< Oct 14 07:04:26 crc kubenswrapper[4870]: timeout: failed to connect service ":50051" within 1s Oct 14 07:04:26 crc kubenswrapper[4870]: > Oct 14 07:04:26 crc kubenswrapper[4870]: I1014 07:04:26.674153 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-4mlsh" Oct 14 07:04:26 crc kubenswrapper[4870]: I1014 07:04:26.674250 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-4mlsh" Oct 14 07:04:26 crc kubenswrapper[4870]: I1014 07:04:26.743354 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-4mlsh" Oct 14 07:04:27 crc kubenswrapper[4870]: I1014 07:04:27.700643 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-cv6fw" Oct 14 07:04:27 crc kubenswrapper[4870]: I1014 07:04:27.701507 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-cv6fw" Oct 14 07:04:28 crc kubenswrapper[4870]: I1014 07:04:28.217065 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-kxxk5" Oct 14 07:04:28 crc kubenswrapper[4870]: I1014 07:04:28.217638 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-kxxk5" Oct 14 07:04:28 crc kubenswrapper[4870]: I1014 07:04:28.284055 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-kxxk5" Oct 14 07:04:28 crc kubenswrapper[4870]: I1014 07:04:28.774415 4870 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-cv6fw" podUID="ea907a55-c1d5-4f49-9ec4-c97461eef014" containerName="registry-server" probeResult="failure" output=< Oct 14 07:04:28 crc kubenswrapper[4870]: timeout: failed to connect service ":50051" within 1s Oct 14 07:04:28 crc kubenswrapper[4870]: > Oct 14 07:04:29 crc kubenswrapper[4870]: I1014 07:04:29.277994 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-kxxk5" Oct 14 07:04:30 crc kubenswrapper[4870]: I1014 07:04:30.164702 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-kxxk5"] Oct 14 07:04:31 crc kubenswrapper[4870]: I1014 07:04:31.240427 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-kxxk5" podUID="d56357b5-cdfe-4903-b59a-4ba8a111883f" containerName="registry-server" containerID="cri-o://a78f8c819123955b9f3c42907be7d1693272c44a7868cbffaf7df4ef46675a5b" gracePeriod=2 Oct 14 07:04:31 crc kubenswrapper[4870]: I1014 07:04:31.637158 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kxxk5" Oct 14 07:04:31 crc kubenswrapper[4870]: I1014 07:04:31.727061 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d56357b5-cdfe-4903-b59a-4ba8a111883f-utilities\") pod \"d56357b5-cdfe-4903-b59a-4ba8a111883f\" (UID: \"d56357b5-cdfe-4903-b59a-4ba8a111883f\") " Oct 14 07:04:31 crc kubenswrapper[4870]: I1014 07:04:31.727658 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d56357b5-cdfe-4903-b59a-4ba8a111883f-catalog-content\") pod \"d56357b5-cdfe-4903-b59a-4ba8a111883f\" (UID: \"d56357b5-cdfe-4903-b59a-4ba8a111883f\") " Oct 14 07:04:31 crc kubenswrapper[4870]: I1014 07:04:31.727703 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mmllz\" (UniqueName: \"kubernetes.io/projected/d56357b5-cdfe-4903-b59a-4ba8a111883f-kube-api-access-mmllz\") pod \"d56357b5-cdfe-4903-b59a-4ba8a111883f\" (UID: \"d56357b5-cdfe-4903-b59a-4ba8a111883f\") " Oct 14 07:04:31 crc kubenswrapper[4870]: I1014 07:04:31.728151 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d56357b5-cdfe-4903-b59a-4ba8a111883f-utilities" (OuterVolumeSpecName: "utilities") pod "d56357b5-cdfe-4903-b59a-4ba8a111883f" (UID: "d56357b5-cdfe-4903-b59a-4ba8a111883f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:04:31 crc kubenswrapper[4870]: I1014 07:04:31.736031 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d56357b5-cdfe-4903-b59a-4ba8a111883f-kube-api-access-mmllz" (OuterVolumeSpecName: "kube-api-access-mmllz") pod "d56357b5-cdfe-4903-b59a-4ba8a111883f" (UID: "d56357b5-cdfe-4903-b59a-4ba8a111883f"). InnerVolumeSpecName "kube-api-access-mmllz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:04:31 crc kubenswrapper[4870]: I1014 07:04:31.824202 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d56357b5-cdfe-4903-b59a-4ba8a111883f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d56357b5-cdfe-4903-b59a-4ba8a111883f" (UID: "d56357b5-cdfe-4903-b59a-4ba8a111883f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:04:31 crc kubenswrapper[4870]: I1014 07:04:31.828973 4870 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d56357b5-cdfe-4903-b59a-4ba8a111883f-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 07:04:31 crc kubenswrapper[4870]: I1014 07:04:31.829011 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mmllz\" (UniqueName: \"kubernetes.io/projected/d56357b5-cdfe-4903-b59a-4ba8a111883f-kube-api-access-mmllz\") on node \"crc\" DevicePath \"\"" Oct 14 07:04:31 crc kubenswrapper[4870]: I1014 07:04:31.829023 4870 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d56357b5-cdfe-4903-b59a-4ba8a111883f-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 07:04:32 crc kubenswrapper[4870]: I1014 07:04:32.249621 4870 generic.go:334] "Generic (PLEG): container finished" podID="d56357b5-cdfe-4903-b59a-4ba8a111883f" containerID="a78f8c819123955b9f3c42907be7d1693272c44a7868cbffaf7df4ef46675a5b" exitCode=0 Oct 14 07:04:32 crc kubenswrapper[4870]: I1014 07:04:32.249667 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kxxk5" event={"ID":"d56357b5-cdfe-4903-b59a-4ba8a111883f","Type":"ContainerDied","Data":"a78f8c819123955b9f3c42907be7d1693272c44a7868cbffaf7df4ef46675a5b"} Oct 14 07:04:32 crc kubenswrapper[4870]: I1014 07:04:32.249701 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kxxk5" event={"ID":"d56357b5-cdfe-4903-b59a-4ba8a111883f","Type":"ContainerDied","Data":"a341a74b2579a68d177e5b9b442c80b86df1c5ac5747ac4f0a15956aac4850c7"} Oct 14 07:04:32 crc kubenswrapper[4870]: I1014 07:04:32.249721 4870 scope.go:117] "RemoveContainer" containerID="a78f8c819123955b9f3c42907be7d1693272c44a7868cbffaf7df4ef46675a5b" Oct 14 07:04:32 crc kubenswrapper[4870]: I1014 07:04:32.249737 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kxxk5" Oct 14 07:04:32 crc kubenswrapper[4870]: I1014 07:04:32.279798 4870 scope.go:117] "RemoveContainer" containerID="af8039048df663104e531d0872471589a53728e5495094cfdc0dd92a2c172f4b" Oct 14 07:04:32 crc kubenswrapper[4870]: I1014 07:04:32.298497 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-kxxk5"] Oct 14 07:04:32 crc kubenswrapper[4870]: I1014 07:04:32.305511 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-kxxk5"] Oct 14 07:04:32 crc kubenswrapper[4870]: I1014 07:04:32.321608 4870 scope.go:117] "RemoveContainer" containerID="4e6fb7da82fe8780bc23b33bf067825c8859856a442b4a009c8e51dba72e22f0" Oct 14 07:04:32 crc kubenswrapper[4870]: I1014 07:04:32.344114 4870 scope.go:117] "RemoveContainer" containerID="a78f8c819123955b9f3c42907be7d1693272c44a7868cbffaf7df4ef46675a5b" Oct 14 07:04:32 crc kubenswrapper[4870]: E1014 07:04:32.344905 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a78f8c819123955b9f3c42907be7d1693272c44a7868cbffaf7df4ef46675a5b\": container with ID starting with a78f8c819123955b9f3c42907be7d1693272c44a7868cbffaf7df4ef46675a5b not found: ID does not exist" containerID="a78f8c819123955b9f3c42907be7d1693272c44a7868cbffaf7df4ef46675a5b" Oct 14 07:04:32 crc kubenswrapper[4870]: I1014 07:04:32.344964 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a78f8c819123955b9f3c42907be7d1693272c44a7868cbffaf7df4ef46675a5b"} err="failed to get container status \"a78f8c819123955b9f3c42907be7d1693272c44a7868cbffaf7df4ef46675a5b\": rpc error: code = NotFound desc = could not find container \"a78f8c819123955b9f3c42907be7d1693272c44a7868cbffaf7df4ef46675a5b\": container with ID starting with a78f8c819123955b9f3c42907be7d1693272c44a7868cbffaf7df4ef46675a5b not found: ID does not exist" Oct 14 07:04:32 crc kubenswrapper[4870]: I1014 07:04:32.345009 4870 scope.go:117] "RemoveContainer" containerID="af8039048df663104e531d0872471589a53728e5495094cfdc0dd92a2c172f4b" Oct 14 07:04:32 crc kubenswrapper[4870]: E1014 07:04:32.345452 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"af8039048df663104e531d0872471589a53728e5495094cfdc0dd92a2c172f4b\": container with ID starting with af8039048df663104e531d0872471589a53728e5495094cfdc0dd92a2c172f4b not found: ID does not exist" containerID="af8039048df663104e531d0872471589a53728e5495094cfdc0dd92a2c172f4b" Oct 14 07:04:32 crc kubenswrapper[4870]: I1014 07:04:32.345513 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"af8039048df663104e531d0872471589a53728e5495094cfdc0dd92a2c172f4b"} err="failed to get container status \"af8039048df663104e531d0872471589a53728e5495094cfdc0dd92a2c172f4b\": rpc error: code = NotFound desc = could not find container \"af8039048df663104e531d0872471589a53728e5495094cfdc0dd92a2c172f4b\": container with ID starting with af8039048df663104e531d0872471589a53728e5495094cfdc0dd92a2c172f4b not found: ID does not exist" Oct 14 07:04:32 crc kubenswrapper[4870]: I1014 07:04:32.345557 4870 scope.go:117] "RemoveContainer" containerID="4e6fb7da82fe8780bc23b33bf067825c8859856a442b4a009c8e51dba72e22f0" Oct 14 07:04:32 crc kubenswrapper[4870]: E1014 07:04:32.345971 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4e6fb7da82fe8780bc23b33bf067825c8859856a442b4a009c8e51dba72e22f0\": container with ID starting with 4e6fb7da82fe8780bc23b33bf067825c8859856a442b4a009c8e51dba72e22f0 not found: ID does not exist" containerID="4e6fb7da82fe8780bc23b33bf067825c8859856a442b4a009c8e51dba72e22f0" Oct 14 07:04:32 crc kubenswrapper[4870]: I1014 07:04:32.346014 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4e6fb7da82fe8780bc23b33bf067825c8859856a442b4a009c8e51dba72e22f0"} err="failed to get container status \"4e6fb7da82fe8780bc23b33bf067825c8859856a442b4a009c8e51dba72e22f0\": rpc error: code = NotFound desc = could not find container \"4e6fb7da82fe8780bc23b33bf067825c8859856a442b4a009c8e51dba72e22f0\": container with ID starting with 4e6fb7da82fe8780bc23b33bf067825c8859856a442b4a009c8e51dba72e22f0 not found: ID does not exist" Oct 14 07:04:33 crc kubenswrapper[4870]: I1014 07:04:33.040660 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d56357b5-cdfe-4903-b59a-4ba8a111883f" path="/var/lib/kubelet/pods/d56357b5-cdfe-4903-b59a-4ba8a111883f/volumes" Oct 14 07:04:34 crc kubenswrapper[4870]: I1014 07:04:34.577138 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-jhjmr" Oct 14 07:04:34 crc kubenswrapper[4870]: I1014 07:04:34.630460 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-jhjmr" Oct 14 07:04:35 crc kubenswrapper[4870]: I1014 07:04:35.211842 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-lkfh6" Oct 14 07:04:35 crc kubenswrapper[4870]: I1014 07:04:35.287900 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-lkfh6" Oct 14 07:04:36 crc kubenswrapper[4870]: I1014 07:04:36.719283 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-4mlsh" Oct 14 07:04:37 crc kubenswrapper[4870]: I1014 07:04:37.167228 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-lkfh6"] Oct 14 07:04:37 crc kubenswrapper[4870]: I1014 07:04:37.167512 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-lkfh6" podUID="44a6ae4c-9030-4a51-8960-0c0cc6c59b8e" containerName="registry-server" containerID="cri-o://6877f0d90eec3c12ee4fa0bbf5b1898fa54252e8355751b8625687bde66e00d0" gracePeriod=2 Oct 14 07:04:37 crc kubenswrapper[4870]: I1014 07:04:37.421021 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-x8qtb"] Oct 14 07:04:37 crc kubenswrapper[4870]: I1014 07:04:37.647296 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lkfh6" Oct 14 07:04:37 crc kubenswrapper[4870]: I1014 07:04:37.735657 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/44a6ae4c-9030-4a51-8960-0c0cc6c59b8e-catalog-content\") pod \"44a6ae4c-9030-4a51-8960-0c0cc6c59b8e\" (UID: \"44a6ae4c-9030-4a51-8960-0c0cc6c59b8e\") " Oct 14 07:04:37 crc kubenswrapper[4870]: I1014 07:04:37.735753 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/44a6ae4c-9030-4a51-8960-0c0cc6c59b8e-utilities\") pod \"44a6ae4c-9030-4a51-8960-0c0cc6c59b8e\" (UID: \"44a6ae4c-9030-4a51-8960-0c0cc6c59b8e\") " Oct 14 07:04:37 crc kubenswrapper[4870]: I1014 07:04:37.735856 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cdzjs\" (UniqueName: \"kubernetes.io/projected/44a6ae4c-9030-4a51-8960-0c0cc6c59b8e-kube-api-access-cdzjs\") pod \"44a6ae4c-9030-4a51-8960-0c0cc6c59b8e\" (UID: \"44a6ae4c-9030-4a51-8960-0c0cc6c59b8e\") " Oct 14 07:04:37 crc kubenswrapper[4870]: I1014 07:04:37.737586 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/44a6ae4c-9030-4a51-8960-0c0cc6c59b8e-utilities" (OuterVolumeSpecName: "utilities") pod "44a6ae4c-9030-4a51-8960-0c0cc6c59b8e" (UID: "44a6ae4c-9030-4a51-8960-0c0cc6c59b8e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:04:37 crc kubenswrapper[4870]: I1014 07:04:37.752637 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44a6ae4c-9030-4a51-8960-0c0cc6c59b8e-kube-api-access-cdzjs" (OuterVolumeSpecName: "kube-api-access-cdzjs") pod "44a6ae4c-9030-4a51-8960-0c0cc6c59b8e" (UID: "44a6ae4c-9030-4a51-8960-0c0cc6c59b8e"). InnerVolumeSpecName "kube-api-access-cdzjs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:04:37 crc kubenswrapper[4870]: I1014 07:04:37.765792 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-cv6fw" Oct 14 07:04:37 crc kubenswrapper[4870]: I1014 07:04:37.807467 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/44a6ae4c-9030-4a51-8960-0c0cc6c59b8e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "44a6ae4c-9030-4a51-8960-0c0cc6c59b8e" (UID: "44a6ae4c-9030-4a51-8960-0c0cc6c59b8e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:04:37 crc kubenswrapper[4870]: I1014 07:04:37.838305 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-cv6fw" Oct 14 07:04:37 crc kubenswrapper[4870]: I1014 07:04:37.838713 4870 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/44a6ae4c-9030-4a51-8960-0c0cc6c59b8e-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 07:04:37 crc kubenswrapper[4870]: I1014 07:04:37.839069 4870 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/44a6ae4c-9030-4a51-8960-0c0cc6c59b8e-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 07:04:37 crc kubenswrapper[4870]: I1014 07:04:37.839080 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cdzjs\" (UniqueName: \"kubernetes.io/projected/44a6ae4c-9030-4a51-8960-0c0cc6c59b8e-kube-api-access-cdzjs\") on node \"crc\" DevicePath \"\"" Oct 14 07:04:38 crc kubenswrapper[4870]: I1014 07:04:38.305650 4870 generic.go:334] "Generic (PLEG): container finished" podID="44a6ae4c-9030-4a51-8960-0c0cc6c59b8e" containerID="6877f0d90eec3c12ee4fa0bbf5b1898fa54252e8355751b8625687bde66e00d0" exitCode=0 Oct 14 07:04:38 crc kubenswrapper[4870]: I1014 07:04:38.305746 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lkfh6" Oct 14 07:04:38 crc kubenswrapper[4870]: I1014 07:04:38.305808 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lkfh6" event={"ID":"44a6ae4c-9030-4a51-8960-0c0cc6c59b8e","Type":"ContainerDied","Data":"6877f0d90eec3c12ee4fa0bbf5b1898fa54252e8355751b8625687bde66e00d0"} Oct 14 07:04:38 crc kubenswrapper[4870]: I1014 07:04:38.305894 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lkfh6" event={"ID":"44a6ae4c-9030-4a51-8960-0c0cc6c59b8e","Type":"ContainerDied","Data":"78504c48419a2d7fd9783cd9f6c026eec82a742d5bc29624f4976ab0bad6aac0"} Oct 14 07:04:38 crc kubenswrapper[4870]: I1014 07:04:38.305935 4870 scope.go:117] "RemoveContainer" containerID="6877f0d90eec3c12ee4fa0bbf5b1898fa54252e8355751b8625687bde66e00d0" Oct 14 07:04:38 crc kubenswrapper[4870]: I1014 07:04:38.334125 4870 scope.go:117] "RemoveContainer" containerID="4c00d93fa663939b356aa33a19729c9f0fe545f382bf028defce467930cf5fc0" Oct 14 07:04:38 crc kubenswrapper[4870]: I1014 07:04:38.338067 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-lkfh6"] Oct 14 07:04:38 crc kubenswrapper[4870]: I1014 07:04:38.340201 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-lkfh6"] Oct 14 07:04:38 crc kubenswrapper[4870]: I1014 07:04:38.370817 4870 scope.go:117] "RemoveContainer" containerID="c87cc557220569277c6a14a11f64301f1124572b2b0c2f867df197992f777f3f" Oct 14 07:04:38 crc kubenswrapper[4870]: I1014 07:04:38.385537 4870 scope.go:117] "RemoveContainer" containerID="6877f0d90eec3c12ee4fa0bbf5b1898fa54252e8355751b8625687bde66e00d0" Oct 14 07:04:38 crc kubenswrapper[4870]: E1014 07:04:38.386105 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6877f0d90eec3c12ee4fa0bbf5b1898fa54252e8355751b8625687bde66e00d0\": container with ID starting with 6877f0d90eec3c12ee4fa0bbf5b1898fa54252e8355751b8625687bde66e00d0 not found: ID does not exist" containerID="6877f0d90eec3c12ee4fa0bbf5b1898fa54252e8355751b8625687bde66e00d0" Oct 14 07:04:38 crc kubenswrapper[4870]: I1014 07:04:38.386178 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6877f0d90eec3c12ee4fa0bbf5b1898fa54252e8355751b8625687bde66e00d0"} err="failed to get container status \"6877f0d90eec3c12ee4fa0bbf5b1898fa54252e8355751b8625687bde66e00d0\": rpc error: code = NotFound desc = could not find container \"6877f0d90eec3c12ee4fa0bbf5b1898fa54252e8355751b8625687bde66e00d0\": container with ID starting with 6877f0d90eec3c12ee4fa0bbf5b1898fa54252e8355751b8625687bde66e00d0 not found: ID does not exist" Oct 14 07:04:38 crc kubenswrapper[4870]: I1014 07:04:38.386229 4870 scope.go:117] "RemoveContainer" containerID="4c00d93fa663939b356aa33a19729c9f0fe545f382bf028defce467930cf5fc0" Oct 14 07:04:38 crc kubenswrapper[4870]: E1014 07:04:38.387208 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4c00d93fa663939b356aa33a19729c9f0fe545f382bf028defce467930cf5fc0\": container with ID starting with 4c00d93fa663939b356aa33a19729c9f0fe545f382bf028defce467930cf5fc0 not found: ID does not exist" containerID="4c00d93fa663939b356aa33a19729c9f0fe545f382bf028defce467930cf5fc0" Oct 14 07:04:38 crc kubenswrapper[4870]: I1014 07:04:38.387244 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4c00d93fa663939b356aa33a19729c9f0fe545f382bf028defce467930cf5fc0"} err="failed to get container status \"4c00d93fa663939b356aa33a19729c9f0fe545f382bf028defce467930cf5fc0\": rpc error: code = NotFound desc = could not find container \"4c00d93fa663939b356aa33a19729c9f0fe545f382bf028defce467930cf5fc0\": container with ID starting with 4c00d93fa663939b356aa33a19729c9f0fe545f382bf028defce467930cf5fc0 not found: ID does not exist" Oct 14 07:04:38 crc kubenswrapper[4870]: I1014 07:04:38.387262 4870 scope.go:117] "RemoveContainer" containerID="c87cc557220569277c6a14a11f64301f1124572b2b0c2f867df197992f777f3f" Oct 14 07:04:38 crc kubenswrapper[4870]: E1014 07:04:38.387602 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c87cc557220569277c6a14a11f64301f1124572b2b0c2f867df197992f777f3f\": container with ID starting with c87cc557220569277c6a14a11f64301f1124572b2b0c2f867df197992f777f3f not found: ID does not exist" containerID="c87cc557220569277c6a14a11f64301f1124572b2b0c2f867df197992f777f3f" Oct 14 07:04:38 crc kubenswrapper[4870]: I1014 07:04:38.387642 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c87cc557220569277c6a14a11f64301f1124572b2b0c2f867df197992f777f3f"} err="failed to get container status \"c87cc557220569277c6a14a11f64301f1124572b2b0c2f867df197992f777f3f\": rpc error: code = NotFound desc = could not find container \"c87cc557220569277c6a14a11f64301f1124572b2b0c2f867df197992f777f3f\": container with ID starting with c87cc557220569277c6a14a11f64301f1124572b2b0c2f867df197992f777f3f not found: ID does not exist" Oct 14 07:04:39 crc kubenswrapper[4870]: I1014 07:04:39.041751 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44a6ae4c-9030-4a51-8960-0c0cc6c59b8e" path="/var/lib/kubelet/pods/44a6ae4c-9030-4a51-8960-0c0cc6c59b8e/volumes" Oct 14 07:04:53 crc kubenswrapper[4870]: I1014 07:04:53.951562 4870 patch_prober.go:28] interesting pod/machine-config-daemon-7tvc7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 07:04:53 crc kubenswrapper[4870]: I1014 07:04:53.952585 4870 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 07:04:53 crc kubenswrapper[4870]: I1014 07:04:53.952671 4870 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" Oct 14 07:04:53 crc kubenswrapper[4870]: I1014 07:04:53.953552 4870 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"8004752575df4a1c44efbcafe64799f09126d0f97285d92388c76a0b90e022f8"} pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 14 07:04:53 crc kubenswrapper[4870]: I1014 07:04:53.953676 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerName="machine-config-daemon" containerID="cri-o://8004752575df4a1c44efbcafe64799f09126d0f97285d92388c76a0b90e022f8" gracePeriod=600 Oct 14 07:04:54 crc kubenswrapper[4870]: I1014 07:04:54.417108 4870 generic.go:334] "Generic (PLEG): container finished" podID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerID="8004752575df4a1c44efbcafe64799f09126d0f97285d92388c76a0b90e022f8" exitCode=0 Oct 14 07:04:54 crc kubenswrapper[4870]: I1014 07:04:54.417179 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" event={"ID":"874a5e45-dffd-4d17-b609-b7d3ed2eab07","Type":"ContainerDied","Data":"8004752575df4a1c44efbcafe64799f09126d0f97285d92388c76a0b90e022f8"} Oct 14 07:04:54 crc kubenswrapper[4870]: I1014 07:04:54.418159 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" event={"ID":"874a5e45-dffd-4d17-b609-b7d3ed2eab07","Type":"ContainerStarted","Data":"fc9b159a611a2f80605df8d6e70c1ccf922f37ec81b713e826ba0a3e8621d3fa"} Oct 14 07:05:02 crc kubenswrapper[4870]: I1014 07:05:02.456366 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-x8qtb" podUID="50732192-84e4-41fa-9c91-e5f77d85fc5e" containerName="oauth-openshift" containerID="cri-o://80aa6bb356d5e8eb5904e69dd4809ac3a3a7f105d22cae7a99b5353b68dcbfb2" gracePeriod=15 Oct 14 07:05:02 crc kubenswrapper[4870]: I1014 07:05:02.942515 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-x8qtb" Oct 14 07:05:02 crc kubenswrapper[4870]: I1014 07:05:02.984636 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-58d58b5989-pqlp6"] Oct 14 07:05:02 crc kubenswrapper[4870]: E1014 07:05:02.984877 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d56357b5-cdfe-4903-b59a-4ba8a111883f" containerName="extract-content" Oct 14 07:05:02 crc kubenswrapper[4870]: I1014 07:05:02.984890 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="d56357b5-cdfe-4903-b59a-4ba8a111883f" containerName="extract-content" Oct 14 07:05:02 crc kubenswrapper[4870]: E1014 07:05:02.984900 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5827db23-a022-4d5f-9ae0-ef3505610bce" containerName="extract-content" Oct 14 07:05:02 crc kubenswrapper[4870]: I1014 07:05:02.984908 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="5827db23-a022-4d5f-9ae0-ef3505610bce" containerName="extract-content" Oct 14 07:05:02 crc kubenswrapper[4870]: E1014 07:05:02.984918 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d56357b5-cdfe-4903-b59a-4ba8a111883f" containerName="extract-utilities" Oct 14 07:05:02 crc kubenswrapper[4870]: I1014 07:05:02.984925 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="d56357b5-cdfe-4903-b59a-4ba8a111883f" containerName="extract-utilities" Oct 14 07:05:02 crc kubenswrapper[4870]: E1014 07:05:02.984936 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e620080-d3ad-49dc-af0f-8dd64485e9f7" containerName="pruner" Oct 14 07:05:02 crc kubenswrapper[4870]: I1014 07:05:02.984942 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e620080-d3ad-49dc-af0f-8dd64485e9f7" containerName="pruner" Oct 14 07:05:02 crc kubenswrapper[4870]: E1014 07:05:02.984951 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="36521dfb-3f0f-4a55-a393-bdf3ec7d3213" containerName="extract-utilities" Oct 14 07:05:02 crc kubenswrapper[4870]: I1014 07:05:02.984957 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="36521dfb-3f0f-4a55-a393-bdf3ec7d3213" containerName="extract-utilities" Oct 14 07:05:02 crc kubenswrapper[4870]: E1014 07:05:02.984967 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44a6ae4c-9030-4a51-8960-0c0cc6c59b8e" containerName="extract-utilities" Oct 14 07:05:02 crc kubenswrapper[4870]: I1014 07:05:02.984972 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="44a6ae4c-9030-4a51-8960-0c0cc6c59b8e" containerName="extract-utilities" Oct 14 07:05:02 crc kubenswrapper[4870]: E1014 07:05:02.984979 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44a6ae4c-9030-4a51-8960-0c0cc6c59b8e" containerName="registry-server" Oct 14 07:05:02 crc kubenswrapper[4870]: I1014 07:05:02.984985 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="44a6ae4c-9030-4a51-8960-0c0cc6c59b8e" containerName="registry-server" Oct 14 07:05:02 crc kubenswrapper[4870]: E1014 07:05:02.984997 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="36521dfb-3f0f-4a55-a393-bdf3ec7d3213" containerName="registry-server" Oct 14 07:05:02 crc kubenswrapper[4870]: I1014 07:05:02.985003 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="36521dfb-3f0f-4a55-a393-bdf3ec7d3213" containerName="registry-server" Oct 14 07:05:02 crc kubenswrapper[4870]: E1014 07:05:02.985012 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="36521dfb-3f0f-4a55-a393-bdf3ec7d3213" containerName="extract-content" Oct 14 07:05:02 crc kubenswrapper[4870]: I1014 07:05:02.985017 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="36521dfb-3f0f-4a55-a393-bdf3ec7d3213" containerName="extract-content" Oct 14 07:05:02 crc kubenswrapper[4870]: E1014 07:05:02.985026 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5827db23-a022-4d5f-9ae0-ef3505610bce" containerName="extract-utilities" Oct 14 07:05:02 crc kubenswrapper[4870]: I1014 07:05:02.985032 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="5827db23-a022-4d5f-9ae0-ef3505610bce" containerName="extract-utilities" Oct 14 07:05:02 crc kubenswrapper[4870]: E1014 07:05:02.985041 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d56357b5-cdfe-4903-b59a-4ba8a111883f" containerName="registry-server" Oct 14 07:05:02 crc kubenswrapper[4870]: I1014 07:05:02.985048 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="d56357b5-cdfe-4903-b59a-4ba8a111883f" containerName="registry-server" Oct 14 07:05:02 crc kubenswrapper[4870]: E1014 07:05:02.985058 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5827db23-a022-4d5f-9ae0-ef3505610bce" containerName="registry-server" Oct 14 07:05:02 crc kubenswrapper[4870]: I1014 07:05:02.985064 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="5827db23-a022-4d5f-9ae0-ef3505610bce" containerName="registry-server" Oct 14 07:05:02 crc kubenswrapper[4870]: E1014 07:05:02.985072 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44a6ae4c-9030-4a51-8960-0c0cc6c59b8e" containerName="extract-content" Oct 14 07:05:02 crc kubenswrapper[4870]: I1014 07:05:02.985078 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="44a6ae4c-9030-4a51-8960-0c0cc6c59b8e" containerName="extract-content" Oct 14 07:05:02 crc kubenswrapper[4870]: E1014 07:05:02.985086 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="50732192-84e4-41fa-9c91-e5f77d85fc5e" containerName="oauth-openshift" Oct 14 07:05:02 crc kubenswrapper[4870]: I1014 07:05:02.985092 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="50732192-84e4-41fa-9c91-e5f77d85fc5e" containerName="oauth-openshift" Oct 14 07:05:02 crc kubenswrapper[4870]: I1014 07:05:02.985196 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="d56357b5-cdfe-4903-b59a-4ba8a111883f" containerName="registry-server" Oct 14 07:05:02 crc kubenswrapper[4870]: I1014 07:05:02.985208 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="9e620080-d3ad-49dc-af0f-8dd64485e9f7" containerName="pruner" Oct 14 07:05:02 crc kubenswrapper[4870]: I1014 07:05:02.985218 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="50732192-84e4-41fa-9c91-e5f77d85fc5e" containerName="oauth-openshift" Oct 14 07:05:02 crc kubenswrapper[4870]: I1014 07:05:02.985225 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="44a6ae4c-9030-4a51-8960-0c0cc6c59b8e" containerName="registry-server" Oct 14 07:05:02 crc kubenswrapper[4870]: I1014 07:05:02.985235 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="36521dfb-3f0f-4a55-a393-bdf3ec7d3213" containerName="registry-server" Oct 14 07:05:02 crc kubenswrapper[4870]: I1014 07:05:02.985243 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="5827db23-a022-4d5f-9ae0-ef3505610bce" containerName="registry-server" Oct 14 07:05:02 crc kubenswrapper[4870]: I1014 07:05:02.987964 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-58d58b5989-pqlp6" Oct 14 07:05:03 crc kubenswrapper[4870]: I1014 07:05:03.001324 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-58d58b5989-pqlp6"] Oct 14 07:05:03 crc kubenswrapper[4870]: I1014 07:05:03.015205 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/50732192-84e4-41fa-9c91-e5f77d85fc5e-v4-0-config-system-router-certs\") pod \"50732192-84e4-41fa-9c91-e5f77d85fc5e\" (UID: \"50732192-84e4-41fa-9c91-e5f77d85fc5e\") " Oct 14 07:05:03 crc kubenswrapper[4870]: I1014 07:05:03.015280 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/50732192-84e4-41fa-9c91-e5f77d85fc5e-v4-0-config-system-serving-cert\") pod \"50732192-84e4-41fa-9c91-e5f77d85fc5e\" (UID: \"50732192-84e4-41fa-9c91-e5f77d85fc5e\") " Oct 14 07:05:03 crc kubenswrapper[4870]: I1014 07:05:03.015346 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/50732192-84e4-41fa-9c91-e5f77d85fc5e-audit-policies\") pod \"50732192-84e4-41fa-9c91-e5f77d85fc5e\" (UID: \"50732192-84e4-41fa-9c91-e5f77d85fc5e\") " Oct 14 07:05:03 crc kubenswrapper[4870]: I1014 07:05:03.015397 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/50732192-84e4-41fa-9c91-e5f77d85fc5e-v4-0-config-system-service-ca\") pod \"50732192-84e4-41fa-9c91-e5f77d85fc5e\" (UID: \"50732192-84e4-41fa-9c91-e5f77d85fc5e\") " Oct 14 07:05:03 crc kubenswrapper[4870]: I1014 07:05:03.015499 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/50732192-84e4-41fa-9c91-e5f77d85fc5e-v4-0-config-system-cliconfig\") pod \"50732192-84e4-41fa-9c91-e5f77d85fc5e\" (UID: \"50732192-84e4-41fa-9c91-e5f77d85fc5e\") " Oct 14 07:05:03 crc kubenswrapper[4870]: I1014 07:05:03.015562 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/50732192-84e4-41fa-9c91-e5f77d85fc5e-v4-0-config-user-template-error\") pod \"50732192-84e4-41fa-9c91-e5f77d85fc5e\" (UID: \"50732192-84e4-41fa-9c91-e5f77d85fc5e\") " Oct 14 07:05:03 crc kubenswrapper[4870]: I1014 07:05:03.015626 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/50732192-84e4-41fa-9c91-e5f77d85fc5e-audit-dir\") pod \"50732192-84e4-41fa-9c91-e5f77d85fc5e\" (UID: \"50732192-84e4-41fa-9c91-e5f77d85fc5e\") " Oct 14 07:05:03 crc kubenswrapper[4870]: I1014 07:05:03.015693 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/50732192-84e4-41fa-9c91-e5f77d85fc5e-v4-0-config-system-session\") pod \"50732192-84e4-41fa-9c91-e5f77d85fc5e\" (UID: \"50732192-84e4-41fa-9c91-e5f77d85fc5e\") " Oct 14 07:05:03 crc kubenswrapper[4870]: I1014 07:05:03.015760 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/50732192-84e4-41fa-9c91-e5f77d85fc5e-v4-0-config-user-template-login\") pod \"50732192-84e4-41fa-9c91-e5f77d85fc5e\" (UID: \"50732192-84e4-41fa-9c91-e5f77d85fc5e\") " Oct 14 07:05:03 crc kubenswrapper[4870]: I1014 07:05:03.015820 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/50732192-84e4-41fa-9c91-e5f77d85fc5e-v4-0-config-user-template-provider-selection\") pod \"50732192-84e4-41fa-9c91-e5f77d85fc5e\" (UID: \"50732192-84e4-41fa-9c91-e5f77d85fc5e\") " Oct 14 07:05:03 crc kubenswrapper[4870]: I1014 07:05:03.015872 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/50732192-84e4-41fa-9c91-e5f77d85fc5e-v4-0-config-system-ocp-branding-template\") pod \"50732192-84e4-41fa-9c91-e5f77d85fc5e\" (UID: \"50732192-84e4-41fa-9c91-e5f77d85fc5e\") " Oct 14 07:05:03 crc kubenswrapper[4870]: I1014 07:05:03.015937 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/50732192-84e4-41fa-9c91-e5f77d85fc5e-v4-0-config-user-idp-0-file-data\") pod \"50732192-84e4-41fa-9c91-e5f77d85fc5e\" (UID: \"50732192-84e4-41fa-9c91-e5f77d85fc5e\") " Oct 14 07:05:03 crc kubenswrapper[4870]: I1014 07:05:03.015976 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9plcl\" (UniqueName: \"kubernetes.io/projected/50732192-84e4-41fa-9c91-e5f77d85fc5e-kube-api-access-9plcl\") pod \"50732192-84e4-41fa-9c91-e5f77d85fc5e\" (UID: \"50732192-84e4-41fa-9c91-e5f77d85fc5e\") " Oct 14 07:05:03 crc kubenswrapper[4870]: I1014 07:05:03.016044 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/50732192-84e4-41fa-9c91-e5f77d85fc5e-v4-0-config-system-trusted-ca-bundle\") pod \"50732192-84e4-41fa-9c91-e5f77d85fc5e\" (UID: \"50732192-84e4-41fa-9c91-e5f77d85fc5e\") " Oct 14 07:05:03 crc kubenswrapper[4870]: I1014 07:05:03.016259 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/b35d80c9-08b1-4e0b-89f2-ac018aab5454-v4-0-config-system-session\") pod \"oauth-openshift-58d58b5989-pqlp6\" (UID: \"b35d80c9-08b1-4e0b-89f2-ac018aab5454\") " pod="openshift-authentication/oauth-openshift-58d58b5989-pqlp6" Oct 14 07:05:03 crc kubenswrapper[4870]: I1014 07:05:03.016356 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/b35d80c9-08b1-4e0b-89f2-ac018aab5454-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-58d58b5989-pqlp6\" (UID: \"b35d80c9-08b1-4e0b-89f2-ac018aab5454\") " pod="openshift-authentication/oauth-openshift-58d58b5989-pqlp6" Oct 14 07:05:03 crc kubenswrapper[4870]: I1014 07:05:03.016411 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/b35d80c9-08b1-4e0b-89f2-ac018aab5454-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-58d58b5989-pqlp6\" (UID: \"b35d80c9-08b1-4e0b-89f2-ac018aab5454\") " pod="openshift-authentication/oauth-openshift-58d58b5989-pqlp6" Oct 14 07:05:03 crc kubenswrapper[4870]: I1014 07:05:03.016476 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/b35d80c9-08b1-4e0b-89f2-ac018aab5454-v4-0-config-system-serving-cert\") pod \"oauth-openshift-58d58b5989-pqlp6\" (UID: \"b35d80c9-08b1-4e0b-89f2-ac018aab5454\") " pod="openshift-authentication/oauth-openshift-58d58b5989-pqlp6" Oct 14 07:05:03 crc kubenswrapper[4870]: I1014 07:05:03.016518 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/b35d80c9-08b1-4e0b-89f2-ac018aab5454-v4-0-config-system-service-ca\") pod \"oauth-openshift-58d58b5989-pqlp6\" (UID: \"b35d80c9-08b1-4e0b-89f2-ac018aab5454\") " pod="openshift-authentication/oauth-openshift-58d58b5989-pqlp6" Oct 14 07:05:03 crc kubenswrapper[4870]: I1014 07:05:03.016555 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/b35d80c9-08b1-4e0b-89f2-ac018aab5454-v4-0-config-system-router-certs\") pod \"oauth-openshift-58d58b5989-pqlp6\" (UID: \"b35d80c9-08b1-4e0b-89f2-ac018aab5454\") " pod="openshift-authentication/oauth-openshift-58d58b5989-pqlp6" Oct 14 07:05:03 crc kubenswrapper[4870]: I1014 07:05:03.016590 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4kq78\" (UniqueName: \"kubernetes.io/projected/b35d80c9-08b1-4e0b-89f2-ac018aab5454-kube-api-access-4kq78\") pod \"oauth-openshift-58d58b5989-pqlp6\" (UID: \"b35d80c9-08b1-4e0b-89f2-ac018aab5454\") " pod="openshift-authentication/oauth-openshift-58d58b5989-pqlp6" Oct 14 07:05:03 crc kubenswrapper[4870]: I1014 07:05:03.016633 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/b35d80c9-08b1-4e0b-89f2-ac018aab5454-v4-0-config-user-template-login\") pod \"oauth-openshift-58d58b5989-pqlp6\" (UID: \"b35d80c9-08b1-4e0b-89f2-ac018aab5454\") " pod="openshift-authentication/oauth-openshift-58d58b5989-pqlp6" Oct 14 07:05:03 crc kubenswrapper[4870]: I1014 07:05:03.016676 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/b35d80c9-08b1-4e0b-89f2-ac018aab5454-audit-dir\") pod \"oauth-openshift-58d58b5989-pqlp6\" (UID: \"b35d80c9-08b1-4e0b-89f2-ac018aab5454\") " pod="openshift-authentication/oauth-openshift-58d58b5989-pqlp6" Oct 14 07:05:03 crc kubenswrapper[4870]: I1014 07:05:03.016723 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/b35d80c9-08b1-4e0b-89f2-ac018aab5454-audit-policies\") pod \"oauth-openshift-58d58b5989-pqlp6\" (UID: \"b35d80c9-08b1-4e0b-89f2-ac018aab5454\") " pod="openshift-authentication/oauth-openshift-58d58b5989-pqlp6" Oct 14 07:05:03 crc kubenswrapper[4870]: I1014 07:05:03.016761 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/b35d80c9-08b1-4e0b-89f2-ac018aab5454-v4-0-config-user-template-error\") pod \"oauth-openshift-58d58b5989-pqlp6\" (UID: \"b35d80c9-08b1-4e0b-89f2-ac018aab5454\") " pod="openshift-authentication/oauth-openshift-58d58b5989-pqlp6" Oct 14 07:05:03 crc kubenswrapper[4870]: I1014 07:05:03.016799 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/b35d80c9-08b1-4e0b-89f2-ac018aab5454-v4-0-config-system-cliconfig\") pod \"oauth-openshift-58d58b5989-pqlp6\" (UID: \"b35d80c9-08b1-4e0b-89f2-ac018aab5454\") " pod="openshift-authentication/oauth-openshift-58d58b5989-pqlp6" Oct 14 07:05:03 crc kubenswrapper[4870]: I1014 07:05:03.016834 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/b35d80c9-08b1-4e0b-89f2-ac018aab5454-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-58d58b5989-pqlp6\" (UID: \"b35d80c9-08b1-4e0b-89f2-ac018aab5454\") " pod="openshift-authentication/oauth-openshift-58d58b5989-pqlp6" Oct 14 07:05:03 crc kubenswrapper[4870]: I1014 07:05:03.016914 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b35d80c9-08b1-4e0b-89f2-ac018aab5454-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-58d58b5989-pqlp6\" (UID: \"b35d80c9-08b1-4e0b-89f2-ac018aab5454\") " pod="openshift-authentication/oauth-openshift-58d58b5989-pqlp6" Oct 14 07:05:03 crc kubenswrapper[4870]: I1014 07:05:03.020030 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/50732192-84e4-41fa-9c91-e5f77d85fc5e-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "50732192-84e4-41fa-9c91-e5f77d85fc5e" (UID: "50732192-84e4-41fa-9c91-e5f77d85fc5e"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:05:03 crc kubenswrapper[4870]: I1014 07:05:03.020492 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/50732192-84e4-41fa-9c91-e5f77d85fc5e-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "50732192-84e4-41fa-9c91-e5f77d85fc5e" (UID: "50732192-84e4-41fa-9c91-e5f77d85fc5e"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 14 07:05:03 crc kubenswrapper[4870]: I1014 07:05:03.020864 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/50732192-84e4-41fa-9c91-e5f77d85fc5e-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "50732192-84e4-41fa-9c91-e5f77d85fc5e" (UID: "50732192-84e4-41fa-9c91-e5f77d85fc5e"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:05:03 crc kubenswrapper[4870]: I1014 07:05:03.021070 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/50732192-84e4-41fa-9c91-e5f77d85fc5e-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "50732192-84e4-41fa-9c91-e5f77d85fc5e" (UID: "50732192-84e4-41fa-9c91-e5f77d85fc5e"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:05:03 crc kubenswrapper[4870]: I1014 07:05:03.021493 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/50732192-84e4-41fa-9c91-e5f77d85fc5e-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "50732192-84e4-41fa-9c91-e5f77d85fc5e" (UID: "50732192-84e4-41fa-9c91-e5f77d85fc5e"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:05:03 crc kubenswrapper[4870]: I1014 07:05:03.025422 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/50732192-84e4-41fa-9c91-e5f77d85fc5e-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "50732192-84e4-41fa-9c91-e5f77d85fc5e" (UID: "50732192-84e4-41fa-9c91-e5f77d85fc5e"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:05:03 crc kubenswrapper[4870]: I1014 07:05:03.025828 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/50732192-84e4-41fa-9c91-e5f77d85fc5e-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "50732192-84e4-41fa-9c91-e5f77d85fc5e" (UID: "50732192-84e4-41fa-9c91-e5f77d85fc5e"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:05:03 crc kubenswrapper[4870]: I1014 07:05:03.026237 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/50732192-84e4-41fa-9c91-e5f77d85fc5e-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "50732192-84e4-41fa-9c91-e5f77d85fc5e" (UID: "50732192-84e4-41fa-9c91-e5f77d85fc5e"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:05:03 crc kubenswrapper[4870]: I1014 07:05:03.026510 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/50732192-84e4-41fa-9c91-e5f77d85fc5e-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "50732192-84e4-41fa-9c91-e5f77d85fc5e" (UID: "50732192-84e4-41fa-9c91-e5f77d85fc5e"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:05:03 crc kubenswrapper[4870]: I1014 07:05:03.028484 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/50732192-84e4-41fa-9c91-e5f77d85fc5e-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "50732192-84e4-41fa-9c91-e5f77d85fc5e" (UID: "50732192-84e4-41fa-9c91-e5f77d85fc5e"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:05:03 crc kubenswrapper[4870]: I1014 07:05:03.038521 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/50732192-84e4-41fa-9c91-e5f77d85fc5e-kube-api-access-9plcl" (OuterVolumeSpecName: "kube-api-access-9plcl") pod "50732192-84e4-41fa-9c91-e5f77d85fc5e" (UID: "50732192-84e4-41fa-9c91-e5f77d85fc5e"). InnerVolumeSpecName "kube-api-access-9plcl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:05:03 crc kubenswrapper[4870]: I1014 07:05:03.041265 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/50732192-84e4-41fa-9c91-e5f77d85fc5e-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "50732192-84e4-41fa-9c91-e5f77d85fc5e" (UID: "50732192-84e4-41fa-9c91-e5f77d85fc5e"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:05:03 crc kubenswrapper[4870]: I1014 07:05:03.041885 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/50732192-84e4-41fa-9c91-e5f77d85fc5e-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "50732192-84e4-41fa-9c91-e5f77d85fc5e" (UID: "50732192-84e4-41fa-9c91-e5f77d85fc5e"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:05:03 crc kubenswrapper[4870]: I1014 07:05:03.042216 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/50732192-84e4-41fa-9c91-e5f77d85fc5e-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "50732192-84e4-41fa-9c91-e5f77d85fc5e" (UID: "50732192-84e4-41fa-9c91-e5f77d85fc5e"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:05:03 crc kubenswrapper[4870]: I1014 07:05:03.118551 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b35d80c9-08b1-4e0b-89f2-ac018aab5454-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-58d58b5989-pqlp6\" (UID: \"b35d80c9-08b1-4e0b-89f2-ac018aab5454\") " pod="openshift-authentication/oauth-openshift-58d58b5989-pqlp6" Oct 14 07:05:03 crc kubenswrapper[4870]: I1014 07:05:03.118677 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/b35d80c9-08b1-4e0b-89f2-ac018aab5454-v4-0-config-system-session\") pod \"oauth-openshift-58d58b5989-pqlp6\" (UID: \"b35d80c9-08b1-4e0b-89f2-ac018aab5454\") " pod="openshift-authentication/oauth-openshift-58d58b5989-pqlp6" Oct 14 07:05:03 crc kubenswrapper[4870]: I1014 07:05:03.118898 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/b35d80c9-08b1-4e0b-89f2-ac018aab5454-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-58d58b5989-pqlp6\" (UID: \"b35d80c9-08b1-4e0b-89f2-ac018aab5454\") " pod="openshift-authentication/oauth-openshift-58d58b5989-pqlp6" Oct 14 07:05:03 crc kubenswrapper[4870]: I1014 07:05:03.118937 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/b35d80c9-08b1-4e0b-89f2-ac018aab5454-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-58d58b5989-pqlp6\" (UID: \"b35d80c9-08b1-4e0b-89f2-ac018aab5454\") " pod="openshift-authentication/oauth-openshift-58d58b5989-pqlp6" Oct 14 07:05:03 crc kubenswrapper[4870]: I1014 07:05:03.118968 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/b35d80c9-08b1-4e0b-89f2-ac018aab5454-v4-0-config-system-serving-cert\") pod \"oauth-openshift-58d58b5989-pqlp6\" (UID: \"b35d80c9-08b1-4e0b-89f2-ac018aab5454\") " pod="openshift-authentication/oauth-openshift-58d58b5989-pqlp6" Oct 14 07:05:03 crc kubenswrapper[4870]: I1014 07:05:03.118991 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/b35d80c9-08b1-4e0b-89f2-ac018aab5454-v4-0-config-system-service-ca\") pod \"oauth-openshift-58d58b5989-pqlp6\" (UID: \"b35d80c9-08b1-4e0b-89f2-ac018aab5454\") " pod="openshift-authentication/oauth-openshift-58d58b5989-pqlp6" Oct 14 07:05:03 crc kubenswrapper[4870]: I1014 07:05:03.120712 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/b35d80c9-08b1-4e0b-89f2-ac018aab5454-v4-0-config-system-router-certs\") pod \"oauth-openshift-58d58b5989-pqlp6\" (UID: \"b35d80c9-08b1-4e0b-89f2-ac018aab5454\") " pod="openshift-authentication/oauth-openshift-58d58b5989-pqlp6" Oct 14 07:05:03 crc kubenswrapper[4870]: I1014 07:05:03.121029 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4kq78\" (UniqueName: \"kubernetes.io/projected/b35d80c9-08b1-4e0b-89f2-ac018aab5454-kube-api-access-4kq78\") pod \"oauth-openshift-58d58b5989-pqlp6\" (UID: \"b35d80c9-08b1-4e0b-89f2-ac018aab5454\") " pod="openshift-authentication/oauth-openshift-58d58b5989-pqlp6" Oct 14 07:05:03 crc kubenswrapper[4870]: I1014 07:05:03.121137 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/b35d80c9-08b1-4e0b-89f2-ac018aab5454-v4-0-config-user-template-login\") pod \"oauth-openshift-58d58b5989-pqlp6\" (UID: \"b35d80c9-08b1-4e0b-89f2-ac018aab5454\") " pod="openshift-authentication/oauth-openshift-58d58b5989-pqlp6" Oct 14 07:05:03 crc kubenswrapper[4870]: I1014 07:05:03.121188 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/b35d80c9-08b1-4e0b-89f2-ac018aab5454-audit-dir\") pod \"oauth-openshift-58d58b5989-pqlp6\" (UID: \"b35d80c9-08b1-4e0b-89f2-ac018aab5454\") " pod="openshift-authentication/oauth-openshift-58d58b5989-pqlp6" Oct 14 07:05:03 crc kubenswrapper[4870]: I1014 07:05:03.121256 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/b35d80c9-08b1-4e0b-89f2-ac018aab5454-audit-policies\") pod \"oauth-openshift-58d58b5989-pqlp6\" (UID: \"b35d80c9-08b1-4e0b-89f2-ac018aab5454\") " pod="openshift-authentication/oauth-openshift-58d58b5989-pqlp6" Oct 14 07:05:03 crc kubenswrapper[4870]: I1014 07:05:03.121304 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/b35d80c9-08b1-4e0b-89f2-ac018aab5454-v4-0-config-user-template-error\") pod \"oauth-openshift-58d58b5989-pqlp6\" (UID: \"b35d80c9-08b1-4e0b-89f2-ac018aab5454\") " pod="openshift-authentication/oauth-openshift-58d58b5989-pqlp6" Oct 14 07:05:03 crc kubenswrapper[4870]: I1014 07:05:03.121373 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/b35d80c9-08b1-4e0b-89f2-ac018aab5454-v4-0-config-system-cliconfig\") pod \"oauth-openshift-58d58b5989-pqlp6\" (UID: \"b35d80c9-08b1-4e0b-89f2-ac018aab5454\") " pod="openshift-authentication/oauth-openshift-58d58b5989-pqlp6" Oct 14 07:05:03 crc kubenswrapper[4870]: I1014 07:05:03.121412 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/b35d80c9-08b1-4e0b-89f2-ac018aab5454-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-58d58b5989-pqlp6\" (UID: \"b35d80c9-08b1-4e0b-89f2-ac018aab5454\") " pod="openshift-authentication/oauth-openshift-58d58b5989-pqlp6" Oct 14 07:05:03 crc kubenswrapper[4870]: I1014 07:05:03.122616 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/b35d80c9-08b1-4e0b-89f2-ac018aab5454-v4-0-config-system-service-ca\") pod \"oauth-openshift-58d58b5989-pqlp6\" (UID: \"b35d80c9-08b1-4e0b-89f2-ac018aab5454\") " pod="openshift-authentication/oauth-openshift-58d58b5989-pqlp6" Oct 14 07:05:03 crc kubenswrapper[4870]: I1014 07:05:03.122738 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/b35d80c9-08b1-4e0b-89f2-ac018aab5454-audit-dir\") pod \"oauth-openshift-58d58b5989-pqlp6\" (UID: \"b35d80c9-08b1-4e0b-89f2-ac018aab5454\") " pod="openshift-authentication/oauth-openshift-58d58b5989-pqlp6" Oct 14 07:05:03 crc kubenswrapper[4870]: I1014 07:05:03.123150 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b35d80c9-08b1-4e0b-89f2-ac018aab5454-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-58d58b5989-pqlp6\" (UID: \"b35d80c9-08b1-4e0b-89f2-ac018aab5454\") " pod="openshift-authentication/oauth-openshift-58d58b5989-pqlp6" Oct 14 07:05:03 crc kubenswrapper[4870]: I1014 07:05:03.124947 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/b35d80c9-08b1-4e0b-89f2-ac018aab5454-v4-0-config-system-cliconfig\") pod \"oauth-openshift-58d58b5989-pqlp6\" (UID: \"b35d80c9-08b1-4e0b-89f2-ac018aab5454\") " pod="openshift-authentication/oauth-openshift-58d58b5989-pqlp6" Oct 14 07:05:03 crc kubenswrapper[4870]: I1014 07:05:03.125371 4870 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/50732192-84e4-41fa-9c91-e5f77d85fc5e-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Oct 14 07:05:03 crc kubenswrapper[4870]: I1014 07:05:03.125428 4870 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/50732192-84e4-41fa-9c91-e5f77d85fc5e-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Oct 14 07:05:03 crc kubenswrapper[4870]: I1014 07:05:03.125529 4870 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/50732192-84e4-41fa-9c91-e5f77d85fc5e-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Oct 14 07:05:03 crc kubenswrapper[4870]: I1014 07:05:03.125557 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9plcl\" (UniqueName: \"kubernetes.io/projected/50732192-84e4-41fa-9c91-e5f77d85fc5e-kube-api-access-9plcl\") on node \"crc\" DevicePath \"\"" Oct 14 07:05:03 crc kubenswrapper[4870]: I1014 07:05:03.125587 4870 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/50732192-84e4-41fa-9c91-e5f77d85fc5e-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Oct 14 07:05:03 crc kubenswrapper[4870]: I1014 07:05:03.125609 4870 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/50732192-84e4-41fa-9c91-e5f77d85fc5e-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 07:05:03 crc kubenswrapper[4870]: I1014 07:05:03.125632 4870 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/50732192-84e4-41fa-9c91-e5f77d85fc5e-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 14 07:05:03 crc kubenswrapper[4870]: I1014 07:05:03.125647 4870 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/50732192-84e4-41fa-9c91-e5f77d85fc5e-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Oct 14 07:05:03 crc kubenswrapper[4870]: I1014 07:05:03.125681 4870 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/50732192-84e4-41fa-9c91-e5f77d85fc5e-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 14 07:05:03 crc kubenswrapper[4870]: I1014 07:05:03.125701 4870 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/50732192-84e4-41fa-9c91-e5f77d85fc5e-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Oct 14 07:05:03 crc kubenswrapper[4870]: I1014 07:05:03.125730 4870 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/50732192-84e4-41fa-9c91-e5f77d85fc5e-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Oct 14 07:05:03 crc kubenswrapper[4870]: I1014 07:05:03.125751 4870 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/50732192-84e4-41fa-9c91-e5f77d85fc5e-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Oct 14 07:05:03 crc kubenswrapper[4870]: I1014 07:05:03.125778 4870 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/50732192-84e4-41fa-9c91-e5f77d85fc5e-audit-dir\") on node \"crc\" DevicePath \"\"" Oct 14 07:05:03 crc kubenswrapper[4870]: I1014 07:05:03.125800 4870 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/50732192-84e4-41fa-9c91-e5f77d85fc5e-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Oct 14 07:05:03 crc kubenswrapper[4870]: I1014 07:05:03.126278 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/b35d80c9-08b1-4e0b-89f2-ac018aab5454-audit-policies\") pod \"oauth-openshift-58d58b5989-pqlp6\" (UID: \"b35d80c9-08b1-4e0b-89f2-ac018aab5454\") " pod="openshift-authentication/oauth-openshift-58d58b5989-pqlp6" Oct 14 07:05:03 crc kubenswrapper[4870]: I1014 07:05:03.126904 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/b35d80c9-08b1-4e0b-89f2-ac018aab5454-v4-0-config-system-session\") pod \"oauth-openshift-58d58b5989-pqlp6\" (UID: \"b35d80c9-08b1-4e0b-89f2-ac018aab5454\") " pod="openshift-authentication/oauth-openshift-58d58b5989-pqlp6" Oct 14 07:05:03 crc kubenswrapper[4870]: I1014 07:05:03.126920 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/b35d80c9-08b1-4e0b-89f2-ac018aab5454-v4-0-config-system-serving-cert\") pod \"oauth-openshift-58d58b5989-pqlp6\" (UID: \"b35d80c9-08b1-4e0b-89f2-ac018aab5454\") " pod="openshift-authentication/oauth-openshift-58d58b5989-pqlp6" Oct 14 07:05:03 crc kubenswrapper[4870]: I1014 07:05:03.127208 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/b35d80c9-08b1-4e0b-89f2-ac018aab5454-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-58d58b5989-pqlp6\" (UID: \"b35d80c9-08b1-4e0b-89f2-ac018aab5454\") " pod="openshift-authentication/oauth-openshift-58d58b5989-pqlp6" Oct 14 07:05:03 crc kubenswrapper[4870]: I1014 07:05:03.128513 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/b35d80c9-08b1-4e0b-89f2-ac018aab5454-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-58d58b5989-pqlp6\" (UID: \"b35d80c9-08b1-4e0b-89f2-ac018aab5454\") " pod="openshift-authentication/oauth-openshift-58d58b5989-pqlp6" Oct 14 07:05:03 crc kubenswrapper[4870]: I1014 07:05:03.133040 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/b35d80c9-08b1-4e0b-89f2-ac018aab5454-v4-0-config-system-router-certs\") pod \"oauth-openshift-58d58b5989-pqlp6\" (UID: \"b35d80c9-08b1-4e0b-89f2-ac018aab5454\") " pod="openshift-authentication/oauth-openshift-58d58b5989-pqlp6" Oct 14 07:05:03 crc kubenswrapper[4870]: I1014 07:05:03.133341 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/b35d80c9-08b1-4e0b-89f2-ac018aab5454-v4-0-config-user-template-login\") pod \"oauth-openshift-58d58b5989-pqlp6\" (UID: \"b35d80c9-08b1-4e0b-89f2-ac018aab5454\") " pod="openshift-authentication/oauth-openshift-58d58b5989-pqlp6" Oct 14 07:05:03 crc kubenswrapper[4870]: I1014 07:05:03.137325 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/b35d80c9-08b1-4e0b-89f2-ac018aab5454-v4-0-config-user-template-error\") pod \"oauth-openshift-58d58b5989-pqlp6\" (UID: \"b35d80c9-08b1-4e0b-89f2-ac018aab5454\") " pod="openshift-authentication/oauth-openshift-58d58b5989-pqlp6" Oct 14 07:05:03 crc kubenswrapper[4870]: I1014 07:05:03.144149 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/b35d80c9-08b1-4e0b-89f2-ac018aab5454-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-58d58b5989-pqlp6\" (UID: \"b35d80c9-08b1-4e0b-89f2-ac018aab5454\") " pod="openshift-authentication/oauth-openshift-58d58b5989-pqlp6" Oct 14 07:05:03 crc kubenswrapper[4870]: I1014 07:05:03.145960 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4kq78\" (UniqueName: \"kubernetes.io/projected/b35d80c9-08b1-4e0b-89f2-ac018aab5454-kube-api-access-4kq78\") pod \"oauth-openshift-58d58b5989-pqlp6\" (UID: \"b35d80c9-08b1-4e0b-89f2-ac018aab5454\") " pod="openshift-authentication/oauth-openshift-58d58b5989-pqlp6" Oct 14 07:05:03 crc kubenswrapper[4870]: I1014 07:05:03.322921 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-58d58b5989-pqlp6" Oct 14 07:05:03 crc kubenswrapper[4870]: I1014 07:05:03.482768 4870 generic.go:334] "Generic (PLEG): container finished" podID="50732192-84e4-41fa-9c91-e5f77d85fc5e" containerID="80aa6bb356d5e8eb5904e69dd4809ac3a3a7f105d22cae7a99b5353b68dcbfb2" exitCode=0 Oct 14 07:05:03 crc kubenswrapper[4870]: I1014 07:05:03.482829 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-x8qtb" event={"ID":"50732192-84e4-41fa-9c91-e5f77d85fc5e","Type":"ContainerDied","Data":"80aa6bb356d5e8eb5904e69dd4809ac3a3a7f105d22cae7a99b5353b68dcbfb2"} Oct 14 07:05:03 crc kubenswrapper[4870]: I1014 07:05:03.482872 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-x8qtb" event={"ID":"50732192-84e4-41fa-9c91-e5f77d85fc5e","Type":"ContainerDied","Data":"67de1d7ccbe5e226b042604620c783e69b32f30a5b52159c97f5d396a579cbaf"} Oct 14 07:05:03 crc kubenswrapper[4870]: I1014 07:05:03.482896 4870 scope.go:117] "RemoveContainer" containerID="80aa6bb356d5e8eb5904e69dd4809ac3a3a7f105d22cae7a99b5353b68dcbfb2" Oct 14 07:05:03 crc kubenswrapper[4870]: I1014 07:05:03.483081 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-x8qtb" Oct 14 07:05:03 crc kubenswrapper[4870]: I1014 07:05:03.518087 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-x8qtb"] Oct 14 07:05:03 crc kubenswrapper[4870]: I1014 07:05:03.521710 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-x8qtb"] Oct 14 07:05:03 crc kubenswrapper[4870]: I1014 07:05:03.528208 4870 scope.go:117] "RemoveContainer" containerID="80aa6bb356d5e8eb5904e69dd4809ac3a3a7f105d22cae7a99b5353b68dcbfb2" Oct 14 07:05:03 crc kubenswrapper[4870]: E1014 07:05:03.528958 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"80aa6bb356d5e8eb5904e69dd4809ac3a3a7f105d22cae7a99b5353b68dcbfb2\": container with ID starting with 80aa6bb356d5e8eb5904e69dd4809ac3a3a7f105d22cae7a99b5353b68dcbfb2 not found: ID does not exist" containerID="80aa6bb356d5e8eb5904e69dd4809ac3a3a7f105d22cae7a99b5353b68dcbfb2" Oct 14 07:05:03 crc kubenswrapper[4870]: I1014 07:05:03.529006 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"80aa6bb356d5e8eb5904e69dd4809ac3a3a7f105d22cae7a99b5353b68dcbfb2"} err="failed to get container status \"80aa6bb356d5e8eb5904e69dd4809ac3a3a7f105d22cae7a99b5353b68dcbfb2\": rpc error: code = NotFound desc = could not find container \"80aa6bb356d5e8eb5904e69dd4809ac3a3a7f105d22cae7a99b5353b68dcbfb2\": container with ID starting with 80aa6bb356d5e8eb5904e69dd4809ac3a3a7f105d22cae7a99b5353b68dcbfb2 not found: ID does not exist" Oct 14 07:05:03 crc kubenswrapper[4870]: I1014 07:05:03.614962 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-58d58b5989-pqlp6"] Oct 14 07:05:04 crc kubenswrapper[4870]: I1014 07:05:04.493089 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-58d58b5989-pqlp6" event={"ID":"b35d80c9-08b1-4e0b-89f2-ac018aab5454","Type":"ContainerStarted","Data":"40fbc7aebe564b8a41e4f9459a3365c265fb39372daa49a8f20795ec01470752"} Oct 14 07:05:04 crc kubenswrapper[4870]: I1014 07:05:04.493174 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-58d58b5989-pqlp6" event={"ID":"b35d80c9-08b1-4e0b-89f2-ac018aab5454","Type":"ContainerStarted","Data":"f43004920a34ce677b77d6528c647f81989843b9cb5f8f136e8c4348116fc34e"} Oct 14 07:05:04 crc kubenswrapper[4870]: I1014 07:05:04.493663 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-58d58b5989-pqlp6" Oct 14 07:05:04 crc kubenswrapper[4870]: I1014 07:05:04.502524 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-58d58b5989-pqlp6" Oct 14 07:05:04 crc kubenswrapper[4870]: I1014 07:05:04.581867 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-58d58b5989-pqlp6" podStartSLOduration=27.581832638 podStartE2EDuration="27.581832638s" podCreationTimestamp="2025-10-14 07:04:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:05:04.534860263 +0000 UTC m=+240.232220634" watchObservedRunningTime="2025-10-14 07:05:04.581832638 +0000 UTC m=+240.279193049" Oct 14 07:05:05 crc kubenswrapper[4870]: I1014 07:05:05.047515 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="50732192-84e4-41fa-9c91-e5f77d85fc5e" path="/var/lib/kubelet/pods/50732192-84e4-41fa-9c91-e5f77d85fc5e/volumes" Oct 14 07:05:24 crc kubenswrapper[4870]: I1014 07:05:24.952514 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-2glwc"] Oct 14 07:05:24 crc kubenswrapper[4870]: I1014 07:05:24.954302 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-2glwc" podUID="788a3634-1dce-4c12-979b-235c74c9200b" containerName="registry-server" containerID="cri-o://df2f460a8e291f154c4ad2a01a01f082c7cf05d6e3201aecd6116f18e97a8d2b" gracePeriod=30 Oct 14 07:05:24 crc kubenswrapper[4870]: I1014 07:05:24.958026 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-jhjmr"] Oct 14 07:05:24 crc kubenswrapper[4870]: I1014 07:05:24.958213 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-jhjmr" podUID="44734ad6-0693-4d02-ba13-441850b5d23b" containerName="registry-server" containerID="cri-o://060a45a1e2cef7c66b65624ead641a9c1743e5787bb9393f509e5f4dbb74f58e" gracePeriod=30 Oct 14 07:05:24 crc kubenswrapper[4870]: I1014 07:05:24.978694 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-6ztwm"] Oct 14 07:05:24 crc kubenswrapper[4870]: I1014 07:05:24.979026 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-6ztwm" podUID="bb92d614-7f4e-470e-a68d-0aff68d3894d" containerName="marketplace-operator" containerID="cri-o://ccfc6039d09655c9466f3f8a70079496876c6ff1dda47f198233250f08226c39" gracePeriod=30 Oct 14 07:05:25 crc kubenswrapper[4870]: I1014 07:05:25.005034 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-4mlsh"] Oct 14 07:05:25 crc kubenswrapper[4870]: I1014 07:05:25.005272 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-4mlsh" podUID="dbace338-bfac-4c71-a6bf-28227f826098" containerName="registry-server" containerID="cri-o://23818a41b9012235bf36fd3c9caca4884535c22691ee73d7291e085758a58b2b" gracePeriod=30 Oct 14 07:05:25 crc kubenswrapper[4870]: I1014 07:05:25.010150 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-cv6fw"] Oct 14 07:05:25 crc kubenswrapper[4870]: I1014 07:05:25.010391 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-cv6fw" podUID="ea907a55-c1d5-4f49-9ec4-c97461eef014" containerName="registry-server" containerID="cri-o://bd3fda2a0804a806b60fb4b84cba455549812fa02c463419f70ece38ebae2eaf" gracePeriod=30 Oct 14 07:05:25 crc kubenswrapper[4870]: I1014 07:05:25.014144 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-7kgk8"] Oct 14 07:05:25 crc kubenswrapper[4870]: I1014 07:05:25.014926 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-7kgk8" Oct 14 07:05:25 crc kubenswrapper[4870]: I1014 07:05:25.026301 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-7kgk8"] Oct 14 07:05:25 crc kubenswrapper[4870]: I1014 07:05:25.092887 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1a1da733-c81b-4cdf-be2c-8f1997560247-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-7kgk8\" (UID: \"1a1da733-c81b-4cdf-be2c-8f1997560247\") " pod="openshift-marketplace/marketplace-operator-79b997595-7kgk8" Oct 14 07:05:25 crc kubenswrapper[4870]: I1014 07:05:25.092940 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/1a1da733-c81b-4cdf-be2c-8f1997560247-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-7kgk8\" (UID: \"1a1da733-c81b-4cdf-be2c-8f1997560247\") " pod="openshift-marketplace/marketplace-operator-79b997595-7kgk8" Oct 14 07:05:25 crc kubenswrapper[4870]: I1014 07:05:25.093035 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-82xdg\" (UniqueName: \"kubernetes.io/projected/1a1da733-c81b-4cdf-be2c-8f1997560247-kube-api-access-82xdg\") pod \"marketplace-operator-79b997595-7kgk8\" (UID: \"1a1da733-c81b-4cdf-be2c-8f1997560247\") " pod="openshift-marketplace/marketplace-operator-79b997595-7kgk8" Oct 14 07:05:25 crc kubenswrapper[4870]: I1014 07:05:25.195068 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1a1da733-c81b-4cdf-be2c-8f1997560247-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-7kgk8\" (UID: \"1a1da733-c81b-4cdf-be2c-8f1997560247\") " pod="openshift-marketplace/marketplace-operator-79b997595-7kgk8" Oct 14 07:05:25 crc kubenswrapper[4870]: I1014 07:05:25.195145 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/1a1da733-c81b-4cdf-be2c-8f1997560247-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-7kgk8\" (UID: \"1a1da733-c81b-4cdf-be2c-8f1997560247\") " pod="openshift-marketplace/marketplace-operator-79b997595-7kgk8" Oct 14 07:05:25 crc kubenswrapper[4870]: I1014 07:05:25.195193 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-82xdg\" (UniqueName: \"kubernetes.io/projected/1a1da733-c81b-4cdf-be2c-8f1997560247-kube-api-access-82xdg\") pod \"marketplace-operator-79b997595-7kgk8\" (UID: \"1a1da733-c81b-4cdf-be2c-8f1997560247\") " pod="openshift-marketplace/marketplace-operator-79b997595-7kgk8" Oct 14 07:05:25 crc kubenswrapper[4870]: I1014 07:05:25.197146 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1a1da733-c81b-4cdf-be2c-8f1997560247-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-7kgk8\" (UID: \"1a1da733-c81b-4cdf-be2c-8f1997560247\") " pod="openshift-marketplace/marketplace-operator-79b997595-7kgk8" Oct 14 07:05:25 crc kubenswrapper[4870]: I1014 07:05:25.212052 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/1a1da733-c81b-4cdf-be2c-8f1997560247-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-7kgk8\" (UID: \"1a1da733-c81b-4cdf-be2c-8f1997560247\") " pod="openshift-marketplace/marketplace-operator-79b997595-7kgk8" Oct 14 07:05:25 crc kubenswrapper[4870]: I1014 07:05:25.221647 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-82xdg\" (UniqueName: \"kubernetes.io/projected/1a1da733-c81b-4cdf-be2c-8f1997560247-kube-api-access-82xdg\") pod \"marketplace-operator-79b997595-7kgk8\" (UID: \"1a1da733-c81b-4cdf-be2c-8f1997560247\") " pod="openshift-marketplace/marketplace-operator-79b997595-7kgk8" Oct 14 07:05:25 crc kubenswrapper[4870]: I1014 07:05:25.415471 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-7kgk8" Oct 14 07:05:25 crc kubenswrapper[4870]: I1014 07:05:25.423138 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jhjmr" Oct 14 07:05:25 crc kubenswrapper[4870]: I1014 07:05:25.462056 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2glwc" Oct 14 07:05:25 crc kubenswrapper[4870]: I1014 07:05:25.498259 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/788a3634-1dce-4c12-979b-235c74c9200b-utilities\") pod \"788a3634-1dce-4c12-979b-235c74c9200b\" (UID: \"788a3634-1dce-4c12-979b-235c74c9200b\") " Oct 14 07:05:25 crc kubenswrapper[4870]: I1014 07:05:25.498318 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k44m6\" (UniqueName: \"kubernetes.io/projected/788a3634-1dce-4c12-979b-235c74c9200b-kube-api-access-k44m6\") pod \"788a3634-1dce-4c12-979b-235c74c9200b\" (UID: \"788a3634-1dce-4c12-979b-235c74c9200b\") " Oct 14 07:05:25 crc kubenswrapper[4870]: I1014 07:05:25.498355 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/44734ad6-0693-4d02-ba13-441850b5d23b-catalog-content\") pod \"44734ad6-0693-4d02-ba13-441850b5d23b\" (UID: \"44734ad6-0693-4d02-ba13-441850b5d23b\") " Oct 14 07:05:25 crc kubenswrapper[4870]: I1014 07:05:25.498377 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/788a3634-1dce-4c12-979b-235c74c9200b-catalog-content\") pod \"788a3634-1dce-4c12-979b-235c74c9200b\" (UID: \"788a3634-1dce-4c12-979b-235c74c9200b\") " Oct 14 07:05:25 crc kubenswrapper[4870]: I1014 07:05:25.498500 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcqbb\" (UniqueName: \"kubernetes.io/projected/44734ad6-0693-4d02-ba13-441850b5d23b-kube-api-access-pcqbb\") pod \"44734ad6-0693-4d02-ba13-441850b5d23b\" (UID: \"44734ad6-0693-4d02-ba13-441850b5d23b\") " Oct 14 07:05:25 crc kubenswrapper[4870]: I1014 07:05:25.498533 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/44734ad6-0693-4d02-ba13-441850b5d23b-utilities\") pod \"44734ad6-0693-4d02-ba13-441850b5d23b\" (UID: \"44734ad6-0693-4d02-ba13-441850b5d23b\") " Oct 14 07:05:25 crc kubenswrapper[4870]: I1014 07:05:25.499904 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/788a3634-1dce-4c12-979b-235c74c9200b-utilities" (OuterVolumeSpecName: "utilities") pod "788a3634-1dce-4c12-979b-235c74c9200b" (UID: "788a3634-1dce-4c12-979b-235c74c9200b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:05:25 crc kubenswrapper[4870]: I1014 07:05:25.500123 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/44734ad6-0693-4d02-ba13-441850b5d23b-utilities" (OuterVolumeSpecName: "utilities") pod "44734ad6-0693-4d02-ba13-441850b5d23b" (UID: "44734ad6-0693-4d02-ba13-441850b5d23b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:05:25 crc kubenswrapper[4870]: I1014 07:05:25.503188 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/788a3634-1dce-4c12-979b-235c74c9200b-kube-api-access-k44m6" (OuterVolumeSpecName: "kube-api-access-k44m6") pod "788a3634-1dce-4c12-979b-235c74c9200b" (UID: "788a3634-1dce-4c12-979b-235c74c9200b"). InnerVolumeSpecName "kube-api-access-k44m6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:05:25 crc kubenswrapper[4870]: I1014 07:05:25.504469 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44734ad6-0693-4d02-ba13-441850b5d23b-kube-api-access-pcqbb" (OuterVolumeSpecName: "kube-api-access-pcqbb") pod "44734ad6-0693-4d02-ba13-441850b5d23b" (UID: "44734ad6-0693-4d02-ba13-441850b5d23b"). InnerVolumeSpecName "kube-api-access-pcqbb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:05:25 crc kubenswrapper[4870]: I1014 07:05:25.569888 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/44734ad6-0693-4d02-ba13-441850b5d23b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "44734ad6-0693-4d02-ba13-441850b5d23b" (UID: "44734ad6-0693-4d02-ba13-441850b5d23b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:05:25 crc kubenswrapper[4870]: I1014 07:05:25.592331 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/788a3634-1dce-4c12-979b-235c74c9200b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "788a3634-1dce-4c12-979b-235c74c9200b" (UID: "788a3634-1dce-4c12-979b-235c74c9200b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:05:25 crc kubenswrapper[4870]: I1014 07:05:25.601270 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k44m6\" (UniqueName: \"kubernetes.io/projected/788a3634-1dce-4c12-979b-235c74c9200b-kube-api-access-k44m6\") on node \"crc\" DevicePath \"\"" Oct 14 07:05:25 crc kubenswrapper[4870]: I1014 07:05:25.601310 4870 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/788a3634-1dce-4c12-979b-235c74c9200b-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 07:05:25 crc kubenswrapper[4870]: I1014 07:05:25.601320 4870 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/44734ad6-0693-4d02-ba13-441850b5d23b-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 07:05:25 crc kubenswrapper[4870]: I1014 07:05:25.601331 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcqbb\" (UniqueName: \"kubernetes.io/projected/44734ad6-0693-4d02-ba13-441850b5d23b-kube-api-access-pcqbb\") on node \"crc\" DevicePath \"\"" Oct 14 07:05:25 crc kubenswrapper[4870]: I1014 07:05:25.601341 4870 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/44734ad6-0693-4d02-ba13-441850b5d23b-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 07:05:25 crc kubenswrapper[4870]: I1014 07:05:25.601349 4870 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/788a3634-1dce-4c12-979b-235c74c9200b-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 07:05:25 crc kubenswrapper[4870]: I1014 07:05:25.609214 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4mlsh" Oct 14 07:05:25 crc kubenswrapper[4870]: I1014 07:05:25.628299 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-cv6fw" Oct 14 07:05:25 crc kubenswrapper[4870]: I1014 07:05:25.661960 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-6ztwm" Oct 14 07:05:25 crc kubenswrapper[4870]: I1014 07:05:25.670081 4870 generic.go:334] "Generic (PLEG): container finished" podID="dbace338-bfac-4c71-a6bf-28227f826098" containerID="23818a41b9012235bf36fd3c9caca4884535c22691ee73d7291e085758a58b2b" exitCode=0 Oct 14 07:05:25 crc kubenswrapper[4870]: I1014 07:05:25.670590 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4mlsh" event={"ID":"dbace338-bfac-4c71-a6bf-28227f826098","Type":"ContainerDied","Data":"23818a41b9012235bf36fd3c9caca4884535c22691ee73d7291e085758a58b2b"} Oct 14 07:05:25 crc kubenswrapper[4870]: I1014 07:05:25.670655 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4mlsh" event={"ID":"dbace338-bfac-4c71-a6bf-28227f826098","Type":"ContainerDied","Data":"469e4a40b9e35d76c5c7dcc65e03d0bf7b9bf1ddadf222e076d51da9f6576edf"} Oct 14 07:05:25 crc kubenswrapper[4870]: I1014 07:05:25.670676 4870 scope.go:117] "RemoveContainer" containerID="23818a41b9012235bf36fd3c9caca4884535c22691ee73d7291e085758a58b2b" Oct 14 07:05:25 crc kubenswrapper[4870]: I1014 07:05:25.671160 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4mlsh" Oct 14 07:05:25 crc kubenswrapper[4870]: I1014 07:05:25.674615 4870 generic.go:334] "Generic (PLEG): container finished" podID="44734ad6-0693-4d02-ba13-441850b5d23b" containerID="060a45a1e2cef7c66b65624ead641a9c1743e5787bb9393f509e5f4dbb74f58e" exitCode=0 Oct 14 07:05:25 crc kubenswrapper[4870]: I1014 07:05:25.674667 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jhjmr" event={"ID":"44734ad6-0693-4d02-ba13-441850b5d23b","Type":"ContainerDied","Data":"060a45a1e2cef7c66b65624ead641a9c1743e5787bb9393f509e5f4dbb74f58e"} Oct 14 07:05:25 crc kubenswrapper[4870]: I1014 07:05:25.674686 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jhjmr" event={"ID":"44734ad6-0693-4d02-ba13-441850b5d23b","Type":"ContainerDied","Data":"59d0f3754cc6c986e4ec7f17db273e835be751584eb3f08d2227fe3177af5c06"} Oct 14 07:05:25 crc kubenswrapper[4870]: I1014 07:05:25.674743 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jhjmr" Oct 14 07:05:25 crc kubenswrapper[4870]: I1014 07:05:25.684771 4870 generic.go:334] "Generic (PLEG): container finished" podID="788a3634-1dce-4c12-979b-235c74c9200b" containerID="df2f460a8e291f154c4ad2a01a01f082c7cf05d6e3201aecd6116f18e97a8d2b" exitCode=0 Oct 14 07:05:25 crc kubenswrapper[4870]: I1014 07:05:25.684875 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2glwc" event={"ID":"788a3634-1dce-4c12-979b-235c74c9200b","Type":"ContainerDied","Data":"df2f460a8e291f154c4ad2a01a01f082c7cf05d6e3201aecd6116f18e97a8d2b"} Oct 14 07:05:25 crc kubenswrapper[4870]: I1014 07:05:25.684903 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2glwc" event={"ID":"788a3634-1dce-4c12-979b-235c74c9200b","Type":"ContainerDied","Data":"0471c8097386471530cd4bf159903868505a94fb0c0eedbec7efdef724bb269e"} Oct 14 07:05:25 crc kubenswrapper[4870]: I1014 07:05:25.684862 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2glwc" Oct 14 07:05:25 crc kubenswrapper[4870]: I1014 07:05:25.695732 4870 generic.go:334] "Generic (PLEG): container finished" podID="ea907a55-c1d5-4f49-9ec4-c97461eef014" containerID="bd3fda2a0804a806b60fb4b84cba455549812fa02c463419f70ece38ebae2eaf" exitCode=0 Oct 14 07:05:25 crc kubenswrapper[4870]: I1014 07:05:25.695788 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-cv6fw" Oct 14 07:05:25 crc kubenswrapper[4870]: I1014 07:05:25.695822 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cv6fw" event={"ID":"ea907a55-c1d5-4f49-9ec4-c97461eef014","Type":"ContainerDied","Data":"bd3fda2a0804a806b60fb4b84cba455549812fa02c463419f70ece38ebae2eaf"} Oct 14 07:05:25 crc kubenswrapper[4870]: I1014 07:05:25.695880 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cv6fw" event={"ID":"ea907a55-c1d5-4f49-9ec4-c97461eef014","Type":"ContainerDied","Data":"82b52bb3e730c37cbb2828c0ded83d866a5c7be7a03af12f79c8825fc05292c7"} Oct 14 07:05:25 crc kubenswrapper[4870]: I1014 07:05:25.702353 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ea907a55-c1d5-4f49-9ec4-c97461eef014-utilities\") pod \"ea907a55-c1d5-4f49-9ec4-c97461eef014\" (UID: \"ea907a55-c1d5-4f49-9ec4-c97461eef014\") " Oct 14 07:05:25 crc kubenswrapper[4870]: I1014 07:05:25.702401 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bb92d614-7f4e-470e-a68d-0aff68d3894d-marketplace-trusted-ca\") pod \"bb92d614-7f4e-470e-a68d-0aff68d3894d\" (UID: \"bb92d614-7f4e-470e-a68d-0aff68d3894d\") " Oct 14 07:05:25 crc kubenswrapper[4870]: I1014 07:05:25.702459 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7np8q\" (UniqueName: \"kubernetes.io/projected/bb92d614-7f4e-470e-a68d-0aff68d3894d-kube-api-access-7np8q\") pod \"bb92d614-7f4e-470e-a68d-0aff68d3894d\" (UID: \"bb92d614-7f4e-470e-a68d-0aff68d3894d\") " Oct 14 07:05:25 crc kubenswrapper[4870]: I1014 07:05:25.702516 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dbace338-bfac-4c71-a6bf-28227f826098-utilities\") pod \"dbace338-bfac-4c71-a6bf-28227f826098\" (UID: \"dbace338-bfac-4c71-a6bf-28227f826098\") " Oct 14 07:05:25 crc kubenswrapper[4870]: I1014 07:05:25.702562 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dbace338-bfac-4c71-a6bf-28227f826098-catalog-content\") pod \"dbace338-bfac-4c71-a6bf-28227f826098\" (UID: \"dbace338-bfac-4c71-a6bf-28227f826098\") " Oct 14 07:05:25 crc kubenswrapper[4870]: I1014 07:05:25.703044 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c7n6\" (UniqueName: \"kubernetes.io/projected/dbace338-bfac-4c71-a6bf-28227f826098-kube-api-access-7c7n6\") pod \"dbace338-bfac-4c71-a6bf-28227f826098\" (UID: \"dbace338-bfac-4c71-a6bf-28227f826098\") " Oct 14 07:05:25 crc kubenswrapper[4870]: I1014 07:05:25.703314 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/bb92d614-7f4e-470e-a68d-0aff68d3894d-marketplace-operator-metrics\") pod \"bb92d614-7f4e-470e-a68d-0aff68d3894d\" (UID: \"bb92d614-7f4e-470e-a68d-0aff68d3894d\") " Oct 14 07:05:25 crc kubenswrapper[4870]: I1014 07:05:25.703537 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wwphs\" (UniqueName: \"kubernetes.io/projected/ea907a55-c1d5-4f49-9ec4-c97461eef014-kube-api-access-wwphs\") pod \"ea907a55-c1d5-4f49-9ec4-c97461eef014\" (UID: \"ea907a55-c1d5-4f49-9ec4-c97461eef014\") " Oct 14 07:05:25 crc kubenswrapper[4870]: I1014 07:05:25.703570 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ea907a55-c1d5-4f49-9ec4-c97461eef014-catalog-content\") pod \"ea907a55-c1d5-4f49-9ec4-c97461eef014\" (UID: \"ea907a55-c1d5-4f49-9ec4-c97461eef014\") " Oct 14 07:05:25 crc kubenswrapper[4870]: I1014 07:05:25.703817 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dbace338-bfac-4c71-a6bf-28227f826098-utilities" (OuterVolumeSpecName: "utilities") pod "dbace338-bfac-4c71-a6bf-28227f826098" (UID: "dbace338-bfac-4c71-a6bf-28227f826098"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:05:25 crc kubenswrapper[4870]: I1014 07:05:25.703965 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ea907a55-c1d5-4f49-9ec4-c97461eef014-utilities" (OuterVolumeSpecName: "utilities") pod "ea907a55-c1d5-4f49-9ec4-c97461eef014" (UID: "ea907a55-c1d5-4f49-9ec4-c97461eef014"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:05:25 crc kubenswrapper[4870]: I1014 07:05:25.704278 4870 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ea907a55-c1d5-4f49-9ec4-c97461eef014-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 07:05:25 crc kubenswrapper[4870]: I1014 07:05:25.704383 4870 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dbace338-bfac-4c71-a6bf-28227f826098-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 07:05:25 crc kubenswrapper[4870]: I1014 07:05:25.705621 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bb92d614-7f4e-470e-a68d-0aff68d3894d-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "bb92d614-7f4e-470e-a68d-0aff68d3894d" (UID: "bb92d614-7f4e-470e-a68d-0aff68d3894d"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:05:25 crc kubenswrapper[4870]: I1014 07:05:25.709170 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-6ztwm" Oct 14 07:05:25 crc kubenswrapper[4870]: I1014 07:05:25.709213 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-6ztwm" event={"ID":"bb92d614-7f4e-470e-a68d-0aff68d3894d","Type":"ContainerDied","Data":"ccfc6039d09655c9466f3f8a70079496876c6ff1dda47f198233250f08226c39"} Oct 14 07:05:25 crc kubenswrapper[4870]: I1014 07:05:25.709715 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ea907a55-c1d5-4f49-9ec4-c97461eef014-kube-api-access-wwphs" (OuterVolumeSpecName: "kube-api-access-wwphs") pod "ea907a55-c1d5-4f49-9ec4-c97461eef014" (UID: "ea907a55-c1d5-4f49-9ec4-c97461eef014"). InnerVolumeSpecName "kube-api-access-wwphs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:05:25 crc kubenswrapper[4870]: I1014 07:05:25.709374 4870 generic.go:334] "Generic (PLEG): container finished" podID="bb92d614-7f4e-470e-a68d-0aff68d3894d" containerID="ccfc6039d09655c9466f3f8a70079496876c6ff1dda47f198233250f08226c39" exitCode=0 Oct 14 07:05:25 crc kubenswrapper[4870]: I1014 07:05:25.710107 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-6ztwm" event={"ID":"bb92d614-7f4e-470e-a68d-0aff68d3894d","Type":"ContainerDied","Data":"57bff9e4a40083ec1011c1e2225894fc355a67b95f973c41dbf99fdc5aaeb99d"} Oct 14 07:05:25 crc kubenswrapper[4870]: I1014 07:05:25.710768 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bb92d614-7f4e-470e-a68d-0aff68d3894d-kube-api-access-7np8q" (OuterVolumeSpecName: "kube-api-access-7np8q") pod "bb92d614-7f4e-470e-a68d-0aff68d3894d" (UID: "bb92d614-7f4e-470e-a68d-0aff68d3894d"). InnerVolumeSpecName "kube-api-access-7np8q". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:05:25 crc kubenswrapper[4870]: I1014 07:05:25.714808 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dbace338-bfac-4c71-a6bf-28227f826098-kube-api-access-7c7n6" (OuterVolumeSpecName: "kube-api-access-7c7n6") pod "dbace338-bfac-4c71-a6bf-28227f826098" (UID: "dbace338-bfac-4c71-a6bf-28227f826098"). InnerVolumeSpecName "kube-api-access-7c7n6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:05:25 crc kubenswrapper[4870]: I1014 07:05:25.715412 4870 scope.go:117] "RemoveContainer" containerID="a2cae3b3a1018bfbd9886537f54022c03285be666d1c80dda72afc6c5a12a553" Oct 14 07:05:25 crc kubenswrapper[4870]: I1014 07:05:25.720351 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-2glwc"] Oct 14 07:05:25 crc kubenswrapper[4870]: I1014 07:05:25.729601 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bb92d614-7f4e-470e-a68d-0aff68d3894d-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "bb92d614-7f4e-470e-a68d-0aff68d3894d" (UID: "bb92d614-7f4e-470e-a68d-0aff68d3894d"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:05:25 crc kubenswrapper[4870]: I1014 07:05:25.730511 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-2glwc"] Oct 14 07:05:25 crc kubenswrapper[4870]: I1014 07:05:25.743336 4870 scope.go:117] "RemoveContainer" containerID="3ec3dde006bb0c7e855f4597f765cbcd6f08cf0c765b91b75871349f8e17c8e3" Oct 14 07:05:25 crc kubenswrapper[4870]: I1014 07:05:25.735010 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-jhjmr"] Oct 14 07:05:25 crc kubenswrapper[4870]: I1014 07:05:25.747462 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dbace338-bfac-4c71-a6bf-28227f826098-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "dbace338-bfac-4c71-a6bf-28227f826098" (UID: "dbace338-bfac-4c71-a6bf-28227f826098"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:05:25 crc kubenswrapper[4870]: I1014 07:05:25.757897 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-jhjmr"] Oct 14 07:05:25 crc kubenswrapper[4870]: I1014 07:05:25.768821 4870 scope.go:117] "RemoveContainer" containerID="23818a41b9012235bf36fd3c9caca4884535c22691ee73d7291e085758a58b2b" Oct 14 07:05:25 crc kubenswrapper[4870]: E1014 07:05:25.769305 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"23818a41b9012235bf36fd3c9caca4884535c22691ee73d7291e085758a58b2b\": container with ID starting with 23818a41b9012235bf36fd3c9caca4884535c22691ee73d7291e085758a58b2b not found: ID does not exist" containerID="23818a41b9012235bf36fd3c9caca4884535c22691ee73d7291e085758a58b2b" Oct 14 07:05:25 crc kubenswrapper[4870]: I1014 07:05:25.769342 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"23818a41b9012235bf36fd3c9caca4884535c22691ee73d7291e085758a58b2b"} err="failed to get container status \"23818a41b9012235bf36fd3c9caca4884535c22691ee73d7291e085758a58b2b\": rpc error: code = NotFound desc = could not find container \"23818a41b9012235bf36fd3c9caca4884535c22691ee73d7291e085758a58b2b\": container with ID starting with 23818a41b9012235bf36fd3c9caca4884535c22691ee73d7291e085758a58b2b not found: ID does not exist" Oct 14 07:05:25 crc kubenswrapper[4870]: I1014 07:05:25.769368 4870 scope.go:117] "RemoveContainer" containerID="a2cae3b3a1018bfbd9886537f54022c03285be666d1c80dda72afc6c5a12a553" Oct 14 07:05:25 crc kubenswrapper[4870]: E1014 07:05:25.769843 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a2cae3b3a1018bfbd9886537f54022c03285be666d1c80dda72afc6c5a12a553\": container with ID starting with a2cae3b3a1018bfbd9886537f54022c03285be666d1c80dda72afc6c5a12a553 not found: ID does not exist" containerID="a2cae3b3a1018bfbd9886537f54022c03285be666d1c80dda72afc6c5a12a553" Oct 14 07:05:25 crc kubenswrapper[4870]: I1014 07:05:25.769862 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a2cae3b3a1018bfbd9886537f54022c03285be666d1c80dda72afc6c5a12a553"} err="failed to get container status \"a2cae3b3a1018bfbd9886537f54022c03285be666d1c80dda72afc6c5a12a553\": rpc error: code = NotFound desc = could not find container \"a2cae3b3a1018bfbd9886537f54022c03285be666d1c80dda72afc6c5a12a553\": container with ID starting with a2cae3b3a1018bfbd9886537f54022c03285be666d1c80dda72afc6c5a12a553 not found: ID does not exist" Oct 14 07:05:25 crc kubenswrapper[4870]: I1014 07:05:25.769879 4870 scope.go:117] "RemoveContainer" containerID="3ec3dde006bb0c7e855f4597f765cbcd6f08cf0c765b91b75871349f8e17c8e3" Oct 14 07:05:25 crc kubenswrapper[4870]: E1014 07:05:25.770208 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3ec3dde006bb0c7e855f4597f765cbcd6f08cf0c765b91b75871349f8e17c8e3\": container with ID starting with 3ec3dde006bb0c7e855f4597f765cbcd6f08cf0c765b91b75871349f8e17c8e3 not found: ID does not exist" containerID="3ec3dde006bb0c7e855f4597f765cbcd6f08cf0c765b91b75871349f8e17c8e3" Oct 14 07:05:25 crc kubenswrapper[4870]: I1014 07:05:25.770230 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3ec3dde006bb0c7e855f4597f765cbcd6f08cf0c765b91b75871349f8e17c8e3"} err="failed to get container status \"3ec3dde006bb0c7e855f4597f765cbcd6f08cf0c765b91b75871349f8e17c8e3\": rpc error: code = NotFound desc = could not find container \"3ec3dde006bb0c7e855f4597f765cbcd6f08cf0c765b91b75871349f8e17c8e3\": container with ID starting with 3ec3dde006bb0c7e855f4597f765cbcd6f08cf0c765b91b75871349f8e17c8e3 not found: ID does not exist" Oct 14 07:05:25 crc kubenswrapper[4870]: I1014 07:05:25.770244 4870 scope.go:117] "RemoveContainer" containerID="060a45a1e2cef7c66b65624ead641a9c1743e5787bb9393f509e5f4dbb74f58e" Oct 14 07:05:25 crc kubenswrapper[4870]: I1014 07:05:25.786227 4870 scope.go:117] "RemoveContainer" containerID="d892e031f81fb6e419b43be04dfed8455efaa1f7a9e93f8a9cdde09b7ffe30dc" Oct 14 07:05:25 crc kubenswrapper[4870]: I1014 07:05:25.800500 4870 scope.go:117] "RemoveContainer" containerID="ec89421786ff16532fc30e300ed0f969965315601580c137ae15ad9425a6f677" Oct 14 07:05:25 crc kubenswrapper[4870]: I1014 07:05:25.807814 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c7n6\" (UniqueName: \"kubernetes.io/projected/dbace338-bfac-4c71-a6bf-28227f826098-kube-api-access-7c7n6\") on node \"crc\" DevicePath \"\"" Oct 14 07:05:25 crc kubenswrapper[4870]: I1014 07:05:25.807840 4870 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/bb92d614-7f4e-470e-a68d-0aff68d3894d-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Oct 14 07:05:25 crc kubenswrapper[4870]: I1014 07:05:25.807880 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wwphs\" (UniqueName: \"kubernetes.io/projected/ea907a55-c1d5-4f49-9ec4-c97461eef014-kube-api-access-wwphs\") on node \"crc\" DevicePath \"\"" Oct 14 07:05:25 crc kubenswrapper[4870]: I1014 07:05:25.807890 4870 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bb92d614-7f4e-470e-a68d-0aff68d3894d-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 14 07:05:25 crc kubenswrapper[4870]: I1014 07:05:25.807898 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7np8q\" (UniqueName: \"kubernetes.io/projected/bb92d614-7f4e-470e-a68d-0aff68d3894d-kube-api-access-7np8q\") on node \"crc\" DevicePath \"\"" Oct 14 07:05:25 crc kubenswrapper[4870]: I1014 07:05:25.807911 4870 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dbace338-bfac-4c71-a6bf-28227f826098-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 07:05:25 crc kubenswrapper[4870]: I1014 07:05:25.817859 4870 scope.go:117] "RemoveContainer" containerID="060a45a1e2cef7c66b65624ead641a9c1743e5787bb9393f509e5f4dbb74f58e" Oct 14 07:05:25 crc kubenswrapper[4870]: E1014 07:05:25.818456 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"060a45a1e2cef7c66b65624ead641a9c1743e5787bb9393f509e5f4dbb74f58e\": container with ID starting with 060a45a1e2cef7c66b65624ead641a9c1743e5787bb9393f509e5f4dbb74f58e not found: ID does not exist" containerID="060a45a1e2cef7c66b65624ead641a9c1743e5787bb9393f509e5f4dbb74f58e" Oct 14 07:05:25 crc kubenswrapper[4870]: I1014 07:05:25.818520 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"060a45a1e2cef7c66b65624ead641a9c1743e5787bb9393f509e5f4dbb74f58e"} err="failed to get container status \"060a45a1e2cef7c66b65624ead641a9c1743e5787bb9393f509e5f4dbb74f58e\": rpc error: code = NotFound desc = could not find container \"060a45a1e2cef7c66b65624ead641a9c1743e5787bb9393f509e5f4dbb74f58e\": container with ID starting with 060a45a1e2cef7c66b65624ead641a9c1743e5787bb9393f509e5f4dbb74f58e not found: ID does not exist" Oct 14 07:05:25 crc kubenswrapper[4870]: I1014 07:05:25.818559 4870 scope.go:117] "RemoveContainer" containerID="d892e031f81fb6e419b43be04dfed8455efaa1f7a9e93f8a9cdde09b7ffe30dc" Oct 14 07:05:25 crc kubenswrapper[4870]: E1014 07:05:25.818973 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d892e031f81fb6e419b43be04dfed8455efaa1f7a9e93f8a9cdde09b7ffe30dc\": container with ID starting with d892e031f81fb6e419b43be04dfed8455efaa1f7a9e93f8a9cdde09b7ffe30dc not found: ID does not exist" containerID="d892e031f81fb6e419b43be04dfed8455efaa1f7a9e93f8a9cdde09b7ffe30dc" Oct 14 07:05:25 crc kubenswrapper[4870]: I1014 07:05:25.818999 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d892e031f81fb6e419b43be04dfed8455efaa1f7a9e93f8a9cdde09b7ffe30dc"} err="failed to get container status \"d892e031f81fb6e419b43be04dfed8455efaa1f7a9e93f8a9cdde09b7ffe30dc\": rpc error: code = NotFound desc = could not find container \"d892e031f81fb6e419b43be04dfed8455efaa1f7a9e93f8a9cdde09b7ffe30dc\": container with ID starting with d892e031f81fb6e419b43be04dfed8455efaa1f7a9e93f8a9cdde09b7ffe30dc not found: ID does not exist" Oct 14 07:05:25 crc kubenswrapper[4870]: I1014 07:05:25.819012 4870 scope.go:117] "RemoveContainer" containerID="ec89421786ff16532fc30e300ed0f969965315601580c137ae15ad9425a6f677" Oct 14 07:05:25 crc kubenswrapper[4870]: E1014 07:05:25.819233 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ec89421786ff16532fc30e300ed0f969965315601580c137ae15ad9425a6f677\": container with ID starting with ec89421786ff16532fc30e300ed0f969965315601580c137ae15ad9425a6f677 not found: ID does not exist" containerID="ec89421786ff16532fc30e300ed0f969965315601580c137ae15ad9425a6f677" Oct 14 07:05:25 crc kubenswrapper[4870]: I1014 07:05:25.819257 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ec89421786ff16532fc30e300ed0f969965315601580c137ae15ad9425a6f677"} err="failed to get container status \"ec89421786ff16532fc30e300ed0f969965315601580c137ae15ad9425a6f677\": rpc error: code = NotFound desc = could not find container \"ec89421786ff16532fc30e300ed0f969965315601580c137ae15ad9425a6f677\": container with ID starting with ec89421786ff16532fc30e300ed0f969965315601580c137ae15ad9425a6f677 not found: ID does not exist" Oct 14 07:05:25 crc kubenswrapper[4870]: I1014 07:05:25.819267 4870 scope.go:117] "RemoveContainer" containerID="df2f460a8e291f154c4ad2a01a01f082c7cf05d6e3201aecd6116f18e97a8d2b" Oct 14 07:05:25 crc kubenswrapper[4870]: I1014 07:05:25.831697 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ea907a55-c1d5-4f49-9ec4-c97461eef014-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ea907a55-c1d5-4f49-9ec4-c97461eef014" (UID: "ea907a55-c1d5-4f49-9ec4-c97461eef014"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:05:25 crc kubenswrapper[4870]: I1014 07:05:25.839761 4870 scope.go:117] "RemoveContainer" containerID="2bea0f06b4f2f68aea3adf931c6f311ab677da53ac19e3f4646a02217929d7a6" Oct 14 07:05:25 crc kubenswrapper[4870]: I1014 07:05:25.853750 4870 scope.go:117] "RemoveContainer" containerID="2d0b0343b0b9d8ef2b1d33fd9bfa0e64399d9d262da5dad799927400e81ab8a4" Oct 14 07:05:25 crc kubenswrapper[4870]: I1014 07:05:25.866547 4870 scope.go:117] "RemoveContainer" containerID="df2f460a8e291f154c4ad2a01a01f082c7cf05d6e3201aecd6116f18e97a8d2b" Oct 14 07:05:25 crc kubenswrapper[4870]: E1014 07:05:25.866939 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"df2f460a8e291f154c4ad2a01a01f082c7cf05d6e3201aecd6116f18e97a8d2b\": container with ID starting with df2f460a8e291f154c4ad2a01a01f082c7cf05d6e3201aecd6116f18e97a8d2b not found: ID does not exist" containerID="df2f460a8e291f154c4ad2a01a01f082c7cf05d6e3201aecd6116f18e97a8d2b" Oct 14 07:05:25 crc kubenswrapper[4870]: I1014 07:05:25.866978 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"df2f460a8e291f154c4ad2a01a01f082c7cf05d6e3201aecd6116f18e97a8d2b"} err="failed to get container status \"df2f460a8e291f154c4ad2a01a01f082c7cf05d6e3201aecd6116f18e97a8d2b\": rpc error: code = NotFound desc = could not find container \"df2f460a8e291f154c4ad2a01a01f082c7cf05d6e3201aecd6116f18e97a8d2b\": container with ID starting with df2f460a8e291f154c4ad2a01a01f082c7cf05d6e3201aecd6116f18e97a8d2b not found: ID does not exist" Oct 14 07:05:25 crc kubenswrapper[4870]: I1014 07:05:25.867005 4870 scope.go:117] "RemoveContainer" containerID="2bea0f06b4f2f68aea3adf931c6f311ab677da53ac19e3f4646a02217929d7a6" Oct 14 07:05:25 crc kubenswrapper[4870]: E1014 07:05:25.867541 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2bea0f06b4f2f68aea3adf931c6f311ab677da53ac19e3f4646a02217929d7a6\": container with ID starting with 2bea0f06b4f2f68aea3adf931c6f311ab677da53ac19e3f4646a02217929d7a6 not found: ID does not exist" containerID="2bea0f06b4f2f68aea3adf931c6f311ab677da53ac19e3f4646a02217929d7a6" Oct 14 07:05:25 crc kubenswrapper[4870]: I1014 07:05:25.867669 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2bea0f06b4f2f68aea3adf931c6f311ab677da53ac19e3f4646a02217929d7a6"} err="failed to get container status \"2bea0f06b4f2f68aea3adf931c6f311ab677da53ac19e3f4646a02217929d7a6\": rpc error: code = NotFound desc = could not find container \"2bea0f06b4f2f68aea3adf931c6f311ab677da53ac19e3f4646a02217929d7a6\": container with ID starting with 2bea0f06b4f2f68aea3adf931c6f311ab677da53ac19e3f4646a02217929d7a6 not found: ID does not exist" Oct 14 07:05:25 crc kubenswrapper[4870]: I1014 07:05:25.867770 4870 scope.go:117] "RemoveContainer" containerID="2d0b0343b0b9d8ef2b1d33fd9bfa0e64399d9d262da5dad799927400e81ab8a4" Oct 14 07:05:25 crc kubenswrapper[4870]: E1014 07:05:25.868480 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2d0b0343b0b9d8ef2b1d33fd9bfa0e64399d9d262da5dad799927400e81ab8a4\": container with ID starting with 2d0b0343b0b9d8ef2b1d33fd9bfa0e64399d9d262da5dad799927400e81ab8a4 not found: ID does not exist" containerID="2d0b0343b0b9d8ef2b1d33fd9bfa0e64399d9d262da5dad799927400e81ab8a4" Oct 14 07:05:25 crc kubenswrapper[4870]: I1014 07:05:25.868518 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2d0b0343b0b9d8ef2b1d33fd9bfa0e64399d9d262da5dad799927400e81ab8a4"} err="failed to get container status \"2d0b0343b0b9d8ef2b1d33fd9bfa0e64399d9d262da5dad799927400e81ab8a4\": rpc error: code = NotFound desc = could not find container \"2d0b0343b0b9d8ef2b1d33fd9bfa0e64399d9d262da5dad799927400e81ab8a4\": container with ID starting with 2d0b0343b0b9d8ef2b1d33fd9bfa0e64399d9d262da5dad799927400e81ab8a4 not found: ID does not exist" Oct 14 07:05:25 crc kubenswrapper[4870]: I1014 07:05:25.868547 4870 scope.go:117] "RemoveContainer" containerID="bd3fda2a0804a806b60fb4b84cba455549812fa02c463419f70ece38ebae2eaf" Oct 14 07:05:25 crc kubenswrapper[4870]: I1014 07:05:25.898318 4870 scope.go:117] "RemoveContainer" containerID="ae472138465eef062cd3eb0d658d045c6847481f1e81fa227d301eca12a74285" Oct 14 07:05:25 crc kubenswrapper[4870]: I1014 07:05:25.909641 4870 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ea907a55-c1d5-4f49-9ec4-c97461eef014-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 07:05:25 crc kubenswrapper[4870]: I1014 07:05:25.914554 4870 scope.go:117] "RemoveContainer" containerID="92cc2b999f056cbf8f202800d2be8d2fa4535872ec24c9274d8a48c8314ed164" Oct 14 07:05:25 crc kubenswrapper[4870]: I1014 07:05:25.928793 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-7kgk8"] Oct 14 07:05:25 crc kubenswrapper[4870]: I1014 07:05:25.933981 4870 scope.go:117] "RemoveContainer" containerID="bd3fda2a0804a806b60fb4b84cba455549812fa02c463419f70ece38ebae2eaf" Oct 14 07:05:25 crc kubenswrapper[4870]: E1014 07:05:25.935058 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bd3fda2a0804a806b60fb4b84cba455549812fa02c463419f70ece38ebae2eaf\": container with ID starting with bd3fda2a0804a806b60fb4b84cba455549812fa02c463419f70ece38ebae2eaf not found: ID does not exist" containerID="bd3fda2a0804a806b60fb4b84cba455549812fa02c463419f70ece38ebae2eaf" Oct 14 07:05:25 crc kubenswrapper[4870]: I1014 07:05:25.935183 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bd3fda2a0804a806b60fb4b84cba455549812fa02c463419f70ece38ebae2eaf"} err="failed to get container status \"bd3fda2a0804a806b60fb4b84cba455549812fa02c463419f70ece38ebae2eaf\": rpc error: code = NotFound desc = could not find container \"bd3fda2a0804a806b60fb4b84cba455549812fa02c463419f70ece38ebae2eaf\": container with ID starting with bd3fda2a0804a806b60fb4b84cba455549812fa02c463419f70ece38ebae2eaf not found: ID does not exist" Oct 14 07:05:25 crc kubenswrapper[4870]: I1014 07:05:25.935307 4870 scope.go:117] "RemoveContainer" containerID="ae472138465eef062cd3eb0d658d045c6847481f1e81fa227d301eca12a74285" Oct 14 07:05:25 crc kubenswrapper[4870]: E1014 07:05:25.936372 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ae472138465eef062cd3eb0d658d045c6847481f1e81fa227d301eca12a74285\": container with ID starting with ae472138465eef062cd3eb0d658d045c6847481f1e81fa227d301eca12a74285 not found: ID does not exist" containerID="ae472138465eef062cd3eb0d658d045c6847481f1e81fa227d301eca12a74285" Oct 14 07:05:25 crc kubenswrapper[4870]: I1014 07:05:25.936497 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ae472138465eef062cd3eb0d658d045c6847481f1e81fa227d301eca12a74285"} err="failed to get container status \"ae472138465eef062cd3eb0d658d045c6847481f1e81fa227d301eca12a74285\": rpc error: code = NotFound desc = could not find container \"ae472138465eef062cd3eb0d658d045c6847481f1e81fa227d301eca12a74285\": container with ID starting with ae472138465eef062cd3eb0d658d045c6847481f1e81fa227d301eca12a74285 not found: ID does not exist" Oct 14 07:05:25 crc kubenswrapper[4870]: I1014 07:05:25.936568 4870 scope.go:117] "RemoveContainer" containerID="92cc2b999f056cbf8f202800d2be8d2fa4535872ec24c9274d8a48c8314ed164" Oct 14 07:05:25 crc kubenswrapper[4870]: W1014 07:05:25.938655 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1a1da733_c81b_4cdf_be2c_8f1997560247.slice/crio-f1974ac58d4631850daa16050e4bdea122ad222c827ede394d4abf9f5d4d14ee WatchSource:0}: Error finding container f1974ac58d4631850daa16050e4bdea122ad222c827ede394d4abf9f5d4d14ee: Status 404 returned error can't find the container with id f1974ac58d4631850daa16050e4bdea122ad222c827ede394d4abf9f5d4d14ee Oct 14 07:05:25 crc kubenswrapper[4870]: E1014 07:05:25.942090 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"92cc2b999f056cbf8f202800d2be8d2fa4535872ec24c9274d8a48c8314ed164\": container with ID starting with 92cc2b999f056cbf8f202800d2be8d2fa4535872ec24c9274d8a48c8314ed164 not found: ID does not exist" containerID="92cc2b999f056cbf8f202800d2be8d2fa4535872ec24c9274d8a48c8314ed164" Oct 14 07:05:25 crc kubenswrapper[4870]: I1014 07:05:25.942188 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"92cc2b999f056cbf8f202800d2be8d2fa4535872ec24c9274d8a48c8314ed164"} err="failed to get container status \"92cc2b999f056cbf8f202800d2be8d2fa4535872ec24c9274d8a48c8314ed164\": rpc error: code = NotFound desc = could not find container \"92cc2b999f056cbf8f202800d2be8d2fa4535872ec24c9274d8a48c8314ed164\": container with ID starting with 92cc2b999f056cbf8f202800d2be8d2fa4535872ec24c9274d8a48c8314ed164 not found: ID does not exist" Oct 14 07:05:25 crc kubenswrapper[4870]: I1014 07:05:25.942259 4870 scope.go:117] "RemoveContainer" containerID="ccfc6039d09655c9466f3f8a70079496876c6ff1dda47f198233250f08226c39" Oct 14 07:05:25 crc kubenswrapper[4870]: I1014 07:05:25.958220 4870 scope.go:117] "RemoveContainer" containerID="ccfc6039d09655c9466f3f8a70079496876c6ff1dda47f198233250f08226c39" Oct 14 07:05:25 crc kubenswrapper[4870]: E1014 07:05:25.959383 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ccfc6039d09655c9466f3f8a70079496876c6ff1dda47f198233250f08226c39\": container with ID starting with ccfc6039d09655c9466f3f8a70079496876c6ff1dda47f198233250f08226c39 not found: ID does not exist" containerID="ccfc6039d09655c9466f3f8a70079496876c6ff1dda47f198233250f08226c39" Oct 14 07:05:25 crc kubenswrapper[4870]: I1014 07:05:25.959475 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ccfc6039d09655c9466f3f8a70079496876c6ff1dda47f198233250f08226c39"} err="failed to get container status \"ccfc6039d09655c9466f3f8a70079496876c6ff1dda47f198233250f08226c39\": rpc error: code = NotFound desc = could not find container \"ccfc6039d09655c9466f3f8a70079496876c6ff1dda47f198233250f08226c39\": container with ID starting with ccfc6039d09655c9466f3f8a70079496876c6ff1dda47f198233250f08226c39 not found: ID does not exist" Oct 14 07:05:26 crc kubenswrapper[4870]: I1014 07:05:26.011746 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-4mlsh"] Oct 14 07:05:26 crc kubenswrapper[4870]: I1014 07:05:26.013643 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-4mlsh"] Oct 14 07:05:26 crc kubenswrapper[4870]: I1014 07:05:26.038819 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-cv6fw"] Oct 14 07:05:26 crc kubenswrapper[4870]: I1014 07:05:26.042360 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-cv6fw"] Oct 14 07:05:26 crc kubenswrapper[4870]: I1014 07:05:26.073670 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-6ztwm"] Oct 14 07:05:26 crc kubenswrapper[4870]: I1014 07:05:26.077916 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-6ztwm"] Oct 14 07:05:26 crc kubenswrapper[4870]: I1014 07:05:26.724471 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-7kgk8" event={"ID":"1a1da733-c81b-4cdf-be2c-8f1997560247","Type":"ContainerStarted","Data":"ccb36eb2526a2a1db031f223d69c3ce9266adca0e3b115f6e4026e543e89f53e"} Oct 14 07:05:26 crc kubenswrapper[4870]: I1014 07:05:26.724524 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-7kgk8" event={"ID":"1a1da733-c81b-4cdf-be2c-8f1997560247","Type":"ContainerStarted","Data":"f1974ac58d4631850daa16050e4bdea122ad222c827ede394d4abf9f5d4d14ee"} Oct 14 07:05:26 crc kubenswrapper[4870]: I1014 07:05:26.724767 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-7kgk8" Oct 14 07:05:26 crc kubenswrapper[4870]: I1014 07:05:26.729811 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-7kgk8" Oct 14 07:05:26 crc kubenswrapper[4870]: I1014 07:05:26.756048 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-7kgk8" podStartSLOduration=2.7560209430000002 podStartE2EDuration="2.756020943s" podCreationTimestamp="2025-10-14 07:05:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:05:26.748525145 +0000 UTC m=+262.445885516" watchObservedRunningTime="2025-10-14 07:05:26.756020943 +0000 UTC m=+262.453381354" Oct 14 07:05:27 crc kubenswrapper[4870]: I1014 07:05:27.042215 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44734ad6-0693-4d02-ba13-441850b5d23b" path="/var/lib/kubelet/pods/44734ad6-0693-4d02-ba13-441850b5d23b/volumes" Oct 14 07:05:27 crc kubenswrapper[4870]: I1014 07:05:27.043193 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="788a3634-1dce-4c12-979b-235c74c9200b" path="/var/lib/kubelet/pods/788a3634-1dce-4c12-979b-235c74c9200b/volumes" Oct 14 07:05:27 crc kubenswrapper[4870]: I1014 07:05:27.043801 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bb92d614-7f4e-470e-a68d-0aff68d3894d" path="/var/lib/kubelet/pods/bb92d614-7f4e-470e-a68d-0aff68d3894d/volumes" Oct 14 07:05:27 crc kubenswrapper[4870]: I1014 07:05:27.044235 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dbace338-bfac-4c71-a6bf-28227f826098" path="/var/lib/kubelet/pods/dbace338-bfac-4c71-a6bf-28227f826098/volumes" Oct 14 07:05:27 crc kubenswrapper[4870]: I1014 07:05:27.044808 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ea907a55-c1d5-4f49-9ec4-c97461eef014" path="/var/lib/kubelet/pods/ea907a55-c1d5-4f49-9ec4-c97461eef014/volumes" Oct 14 07:05:27 crc kubenswrapper[4870]: I1014 07:05:27.173974 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-jqsnc"] Oct 14 07:05:27 crc kubenswrapper[4870]: E1014 07:05:27.174314 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea907a55-c1d5-4f49-9ec4-c97461eef014" containerName="extract-utilities" Oct 14 07:05:27 crc kubenswrapper[4870]: I1014 07:05:27.174335 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea907a55-c1d5-4f49-9ec4-c97461eef014" containerName="extract-utilities" Oct 14 07:05:27 crc kubenswrapper[4870]: E1014 07:05:27.174347 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="788a3634-1dce-4c12-979b-235c74c9200b" containerName="extract-content" Oct 14 07:05:27 crc kubenswrapper[4870]: I1014 07:05:27.174355 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="788a3634-1dce-4c12-979b-235c74c9200b" containerName="extract-content" Oct 14 07:05:27 crc kubenswrapper[4870]: E1014 07:05:27.174367 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dbace338-bfac-4c71-a6bf-28227f826098" containerName="extract-utilities" Oct 14 07:05:27 crc kubenswrapper[4870]: I1014 07:05:27.174375 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="dbace338-bfac-4c71-a6bf-28227f826098" containerName="extract-utilities" Oct 14 07:05:27 crc kubenswrapper[4870]: E1014 07:05:27.174385 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="788a3634-1dce-4c12-979b-235c74c9200b" containerName="registry-server" Oct 14 07:05:27 crc kubenswrapper[4870]: I1014 07:05:27.174395 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="788a3634-1dce-4c12-979b-235c74c9200b" containerName="registry-server" Oct 14 07:05:27 crc kubenswrapper[4870]: E1014 07:05:27.174404 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb92d614-7f4e-470e-a68d-0aff68d3894d" containerName="marketplace-operator" Oct 14 07:05:27 crc kubenswrapper[4870]: I1014 07:05:27.174411 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb92d614-7f4e-470e-a68d-0aff68d3894d" containerName="marketplace-operator" Oct 14 07:05:27 crc kubenswrapper[4870]: E1014 07:05:27.174418 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="788a3634-1dce-4c12-979b-235c74c9200b" containerName="extract-utilities" Oct 14 07:05:27 crc kubenswrapper[4870]: I1014 07:05:27.174426 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="788a3634-1dce-4c12-979b-235c74c9200b" containerName="extract-utilities" Oct 14 07:05:27 crc kubenswrapper[4870]: E1014 07:05:27.174454 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea907a55-c1d5-4f49-9ec4-c97461eef014" containerName="extract-content" Oct 14 07:05:27 crc kubenswrapper[4870]: I1014 07:05:27.174461 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea907a55-c1d5-4f49-9ec4-c97461eef014" containerName="extract-content" Oct 14 07:05:27 crc kubenswrapper[4870]: E1014 07:05:27.174470 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44734ad6-0693-4d02-ba13-441850b5d23b" containerName="registry-server" Oct 14 07:05:27 crc kubenswrapper[4870]: I1014 07:05:27.174477 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="44734ad6-0693-4d02-ba13-441850b5d23b" containerName="registry-server" Oct 14 07:05:27 crc kubenswrapper[4870]: E1014 07:05:27.174487 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44734ad6-0693-4d02-ba13-441850b5d23b" containerName="extract-utilities" Oct 14 07:05:27 crc kubenswrapper[4870]: I1014 07:05:27.174494 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="44734ad6-0693-4d02-ba13-441850b5d23b" containerName="extract-utilities" Oct 14 07:05:27 crc kubenswrapper[4870]: E1014 07:05:27.174504 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dbace338-bfac-4c71-a6bf-28227f826098" containerName="registry-server" Oct 14 07:05:27 crc kubenswrapper[4870]: I1014 07:05:27.174511 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="dbace338-bfac-4c71-a6bf-28227f826098" containerName="registry-server" Oct 14 07:05:27 crc kubenswrapper[4870]: E1014 07:05:27.174520 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dbace338-bfac-4c71-a6bf-28227f826098" containerName="extract-content" Oct 14 07:05:27 crc kubenswrapper[4870]: I1014 07:05:27.174527 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="dbace338-bfac-4c71-a6bf-28227f826098" containerName="extract-content" Oct 14 07:05:27 crc kubenswrapper[4870]: E1014 07:05:27.174536 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea907a55-c1d5-4f49-9ec4-c97461eef014" containerName="registry-server" Oct 14 07:05:27 crc kubenswrapper[4870]: I1014 07:05:27.174542 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea907a55-c1d5-4f49-9ec4-c97461eef014" containerName="registry-server" Oct 14 07:05:27 crc kubenswrapper[4870]: E1014 07:05:27.174550 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44734ad6-0693-4d02-ba13-441850b5d23b" containerName="extract-content" Oct 14 07:05:27 crc kubenswrapper[4870]: I1014 07:05:27.174556 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="44734ad6-0693-4d02-ba13-441850b5d23b" containerName="extract-content" Oct 14 07:05:27 crc kubenswrapper[4870]: I1014 07:05:27.174646 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="dbace338-bfac-4c71-a6bf-28227f826098" containerName="registry-server" Oct 14 07:05:27 crc kubenswrapper[4870]: I1014 07:05:27.174658 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="44734ad6-0693-4d02-ba13-441850b5d23b" containerName="registry-server" Oct 14 07:05:27 crc kubenswrapper[4870]: I1014 07:05:27.174668 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="788a3634-1dce-4c12-979b-235c74c9200b" containerName="registry-server" Oct 14 07:05:27 crc kubenswrapper[4870]: I1014 07:05:27.174682 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb92d614-7f4e-470e-a68d-0aff68d3894d" containerName="marketplace-operator" Oct 14 07:05:27 crc kubenswrapper[4870]: I1014 07:05:27.174694 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="ea907a55-c1d5-4f49-9ec4-c97461eef014" containerName="registry-server" Oct 14 07:05:27 crc kubenswrapper[4870]: I1014 07:05:27.175649 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jqsnc" Oct 14 07:05:27 crc kubenswrapper[4870]: I1014 07:05:27.178310 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Oct 14 07:05:27 crc kubenswrapper[4870]: I1014 07:05:27.190395 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-jqsnc"] Oct 14 07:05:27 crc kubenswrapper[4870]: I1014 07:05:27.230082 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/08bc1ff9-0341-4e01-93ec-25d41eebf646-catalog-content\") pod \"redhat-marketplace-jqsnc\" (UID: \"08bc1ff9-0341-4e01-93ec-25d41eebf646\") " pod="openshift-marketplace/redhat-marketplace-jqsnc" Oct 14 07:05:27 crc kubenswrapper[4870]: I1014 07:05:27.230143 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kbvnf\" (UniqueName: \"kubernetes.io/projected/08bc1ff9-0341-4e01-93ec-25d41eebf646-kube-api-access-kbvnf\") pod \"redhat-marketplace-jqsnc\" (UID: \"08bc1ff9-0341-4e01-93ec-25d41eebf646\") " pod="openshift-marketplace/redhat-marketplace-jqsnc" Oct 14 07:05:27 crc kubenswrapper[4870]: I1014 07:05:27.230476 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/08bc1ff9-0341-4e01-93ec-25d41eebf646-utilities\") pod \"redhat-marketplace-jqsnc\" (UID: \"08bc1ff9-0341-4e01-93ec-25d41eebf646\") " pod="openshift-marketplace/redhat-marketplace-jqsnc" Oct 14 07:05:27 crc kubenswrapper[4870]: I1014 07:05:27.332149 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/08bc1ff9-0341-4e01-93ec-25d41eebf646-catalog-content\") pod \"redhat-marketplace-jqsnc\" (UID: \"08bc1ff9-0341-4e01-93ec-25d41eebf646\") " pod="openshift-marketplace/redhat-marketplace-jqsnc" Oct 14 07:05:27 crc kubenswrapper[4870]: I1014 07:05:27.332205 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kbvnf\" (UniqueName: \"kubernetes.io/projected/08bc1ff9-0341-4e01-93ec-25d41eebf646-kube-api-access-kbvnf\") pod \"redhat-marketplace-jqsnc\" (UID: \"08bc1ff9-0341-4e01-93ec-25d41eebf646\") " pod="openshift-marketplace/redhat-marketplace-jqsnc" Oct 14 07:05:27 crc kubenswrapper[4870]: I1014 07:05:27.332243 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/08bc1ff9-0341-4e01-93ec-25d41eebf646-utilities\") pod \"redhat-marketplace-jqsnc\" (UID: \"08bc1ff9-0341-4e01-93ec-25d41eebf646\") " pod="openshift-marketplace/redhat-marketplace-jqsnc" Oct 14 07:05:27 crc kubenswrapper[4870]: I1014 07:05:27.332925 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/08bc1ff9-0341-4e01-93ec-25d41eebf646-utilities\") pod \"redhat-marketplace-jqsnc\" (UID: \"08bc1ff9-0341-4e01-93ec-25d41eebf646\") " pod="openshift-marketplace/redhat-marketplace-jqsnc" Oct 14 07:05:27 crc kubenswrapper[4870]: I1014 07:05:27.333027 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/08bc1ff9-0341-4e01-93ec-25d41eebf646-catalog-content\") pod \"redhat-marketplace-jqsnc\" (UID: \"08bc1ff9-0341-4e01-93ec-25d41eebf646\") " pod="openshift-marketplace/redhat-marketplace-jqsnc" Oct 14 07:05:27 crc kubenswrapper[4870]: I1014 07:05:27.359954 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kbvnf\" (UniqueName: \"kubernetes.io/projected/08bc1ff9-0341-4e01-93ec-25d41eebf646-kube-api-access-kbvnf\") pod \"redhat-marketplace-jqsnc\" (UID: \"08bc1ff9-0341-4e01-93ec-25d41eebf646\") " pod="openshift-marketplace/redhat-marketplace-jqsnc" Oct 14 07:05:27 crc kubenswrapper[4870]: I1014 07:05:27.371097 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-5t25t"] Oct 14 07:05:27 crc kubenswrapper[4870]: I1014 07:05:27.372314 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5t25t" Oct 14 07:05:27 crc kubenswrapper[4870]: I1014 07:05:27.374944 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Oct 14 07:05:27 crc kubenswrapper[4870]: I1014 07:05:27.383202 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-5t25t"] Oct 14 07:05:27 crc kubenswrapper[4870]: I1014 07:05:27.433246 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9dc73b58-877c-4e60-8508-19ffeb7410f4-catalog-content\") pod \"redhat-operators-5t25t\" (UID: \"9dc73b58-877c-4e60-8508-19ffeb7410f4\") " pod="openshift-marketplace/redhat-operators-5t25t" Oct 14 07:05:27 crc kubenswrapper[4870]: I1014 07:05:27.433308 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n4mlw\" (UniqueName: \"kubernetes.io/projected/9dc73b58-877c-4e60-8508-19ffeb7410f4-kube-api-access-n4mlw\") pod \"redhat-operators-5t25t\" (UID: \"9dc73b58-877c-4e60-8508-19ffeb7410f4\") " pod="openshift-marketplace/redhat-operators-5t25t" Oct 14 07:05:27 crc kubenswrapper[4870]: I1014 07:05:27.433340 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9dc73b58-877c-4e60-8508-19ffeb7410f4-utilities\") pod \"redhat-operators-5t25t\" (UID: \"9dc73b58-877c-4e60-8508-19ffeb7410f4\") " pod="openshift-marketplace/redhat-operators-5t25t" Oct 14 07:05:27 crc kubenswrapper[4870]: I1014 07:05:27.501065 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jqsnc" Oct 14 07:05:27 crc kubenswrapper[4870]: I1014 07:05:27.534481 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9dc73b58-877c-4e60-8508-19ffeb7410f4-catalog-content\") pod \"redhat-operators-5t25t\" (UID: \"9dc73b58-877c-4e60-8508-19ffeb7410f4\") " pod="openshift-marketplace/redhat-operators-5t25t" Oct 14 07:05:27 crc kubenswrapper[4870]: I1014 07:05:27.534545 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n4mlw\" (UniqueName: \"kubernetes.io/projected/9dc73b58-877c-4e60-8508-19ffeb7410f4-kube-api-access-n4mlw\") pod \"redhat-operators-5t25t\" (UID: \"9dc73b58-877c-4e60-8508-19ffeb7410f4\") " pod="openshift-marketplace/redhat-operators-5t25t" Oct 14 07:05:27 crc kubenswrapper[4870]: I1014 07:05:27.534585 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9dc73b58-877c-4e60-8508-19ffeb7410f4-utilities\") pod \"redhat-operators-5t25t\" (UID: \"9dc73b58-877c-4e60-8508-19ffeb7410f4\") " pod="openshift-marketplace/redhat-operators-5t25t" Oct 14 07:05:27 crc kubenswrapper[4870]: I1014 07:05:27.536042 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9dc73b58-877c-4e60-8508-19ffeb7410f4-utilities\") pod \"redhat-operators-5t25t\" (UID: \"9dc73b58-877c-4e60-8508-19ffeb7410f4\") " pod="openshift-marketplace/redhat-operators-5t25t" Oct 14 07:05:27 crc kubenswrapper[4870]: I1014 07:05:27.536553 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9dc73b58-877c-4e60-8508-19ffeb7410f4-catalog-content\") pod \"redhat-operators-5t25t\" (UID: \"9dc73b58-877c-4e60-8508-19ffeb7410f4\") " pod="openshift-marketplace/redhat-operators-5t25t" Oct 14 07:05:27 crc kubenswrapper[4870]: I1014 07:05:27.557519 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n4mlw\" (UniqueName: \"kubernetes.io/projected/9dc73b58-877c-4e60-8508-19ffeb7410f4-kube-api-access-n4mlw\") pod \"redhat-operators-5t25t\" (UID: \"9dc73b58-877c-4e60-8508-19ffeb7410f4\") " pod="openshift-marketplace/redhat-operators-5t25t" Oct 14 07:05:27 crc kubenswrapper[4870]: I1014 07:05:27.703327 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5t25t" Oct 14 07:05:27 crc kubenswrapper[4870]: I1014 07:05:27.918136 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-jqsnc"] Oct 14 07:05:28 crc kubenswrapper[4870]: I1014 07:05:28.202242 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-5t25t"] Oct 14 07:05:28 crc kubenswrapper[4870]: I1014 07:05:28.751529 4870 generic.go:334] "Generic (PLEG): container finished" podID="9dc73b58-877c-4e60-8508-19ffeb7410f4" containerID="d1ce24f3ae403610927a482b86d7002ae21f5ddc28d63d1f8d37d00362defa5b" exitCode=0 Oct 14 07:05:28 crc kubenswrapper[4870]: I1014 07:05:28.751601 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5t25t" event={"ID":"9dc73b58-877c-4e60-8508-19ffeb7410f4","Type":"ContainerDied","Data":"d1ce24f3ae403610927a482b86d7002ae21f5ddc28d63d1f8d37d00362defa5b"} Oct 14 07:05:28 crc kubenswrapper[4870]: I1014 07:05:28.752267 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5t25t" event={"ID":"9dc73b58-877c-4e60-8508-19ffeb7410f4","Type":"ContainerStarted","Data":"5d02180070586fe73ce473a01193ebb219a3def5f2a64f52dc9a47a7d79c1aa6"} Oct 14 07:05:28 crc kubenswrapper[4870]: I1014 07:05:28.755156 4870 generic.go:334] "Generic (PLEG): container finished" podID="08bc1ff9-0341-4e01-93ec-25d41eebf646" containerID="746039e32d72e9724cd8ff8fa8efa17097c3091d4ea91c383a5228edd2da51b5" exitCode=0 Oct 14 07:05:28 crc kubenswrapper[4870]: I1014 07:05:28.755224 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jqsnc" event={"ID":"08bc1ff9-0341-4e01-93ec-25d41eebf646","Type":"ContainerDied","Data":"746039e32d72e9724cd8ff8fa8efa17097c3091d4ea91c383a5228edd2da51b5"} Oct 14 07:05:28 crc kubenswrapper[4870]: I1014 07:05:28.755292 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jqsnc" event={"ID":"08bc1ff9-0341-4e01-93ec-25d41eebf646","Type":"ContainerStarted","Data":"ae25eb971a77ae125709bab1f8ba1f1f9d1ff344263078538ef706b362cccb78"} Oct 14 07:05:29 crc kubenswrapper[4870]: I1014 07:05:29.575545 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-prqvc"] Oct 14 07:05:29 crc kubenswrapper[4870]: I1014 07:05:29.577869 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-prqvc" Oct 14 07:05:29 crc kubenswrapper[4870]: I1014 07:05:29.581837 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Oct 14 07:05:29 crc kubenswrapper[4870]: I1014 07:05:29.594520 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-prqvc"] Oct 14 07:05:29 crc kubenswrapper[4870]: I1014 07:05:29.679777 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jpwcz\" (UniqueName: \"kubernetes.io/projected/e9f8d5e1-367a-4cb2-9119-b5e6ae554774-kube-api-access-jpwcz\") pod \"certified-operators-prqvc\" (UID: \"e9f8d5e1-367a-4cb2-9119-b5e6ae554774\") " pod="openshift-marketplace/certified-operators-prqvc" Oct 14 07:05:29 crc kubenswrapper[4870]: I1014 07:05:29.679946 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e9f8d5e1-367a-4cb2-9119-b5e6ae554774-utilities\") pod \"certified-operators-prqvc\" (UID: \"e9f8d5e1-367a-4cb2-9119-b5e6ae554774\") " pod="openshift-marketplace/certified-operators-prqvc" Oct 14 07:05:29 crc kubenswrapper[4870]: I1014 07:05:29.680137 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e9f8d5e1-367a-4cb2-9119-b5e6ae554774-catalog-content\") pod \"certified-operators-prqvc\" (UID: \"e9f8d5e1-367a-4cb2-9119-b5e6ae554774\") " pod="openshift-marketplace/certified-operators-prqvc" Oct 14 07:05:29 crc kubenswrapper[4870]: I1014 07:05:29.767746 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-gx6t2"] Oct 14 07:05:29 crc kubenswrapper[4870]: I1014 07:05:29.769043 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gx6t2" Oct 14 07:05:29 crc kubenswrapper[4870]: I1014 07:05:29.769062 4870 generic.go:334] "Generic (PLEG): container finished" podID="08bc1ff9-0341-4e01-93ec-25d41eebf646" containerID="8bfc88465eca2b45c554fd42e6da937c4798807d6c78eeae926aacbff9699e0c" exitCode=0 Oct 14 07:05:29 crc kubenswrapper[4870]: I1014 07:05:29.769125 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jqsnc" event={"ID":"08bc1ff9-0341-4e01-93ec-25d41eebf646","Type":"ContainerDied","Data":"8bfc88465eca2b45c554fd42e6da937c4798807d6c78eeae926aacbff9699e0c"} Oct 14 07:05:29 crc kubenswrapper[4870]: I1014 07:05:29.772375 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Oct 14 07:05:29 crc kubenswrapper[4870]: I1014 07:05:29.783101 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-gx6t2"] Oct 14 07:05:29 crc kubenswrapper[4870]: I1014 07:05:29.783902 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e9f8d5e1-367a-4cb2-9119-b5e6ae554774-catalog-content\") pod \"certified-operators-prqvc\" (UID: \"e9f8d5e1-367a-4cb2-9119-b5e6ae554774\") " pod="openshift-marketplace/certified-operators-prqvc" Oct 14 07:05:29 crc kubenswrapper[4870]: I1014 07:05:29.783973 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jpwcz\" (UniqueName: \"kubernetes.io/projected/e9f8d5e1-367a-4cb2-9119-b5e6ae554774-kube-api-access-jpwcz\") pod \"certified-operators-prqvc\" (UID: \"e9f8d5e1-367a-4cb2-9119-b5e6ae554774\") " pod="openshift-marketplace/certified-operators-prqvc" Oct 14 07:05:29 crc kubenswrapper[4870]: I1014 07:05:29.784074 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e9f8d5e1-367a-4cb2-9119-b5e6ae554774-utilities\") pod \"certified-operators-prqvc\" (UID: \"e9f8d5e1-367a-4cb2-9119-b5e6ae554774\") " pod="openshift-marketplace/certified-operators-prqvc" Oct 14 07:05:29 crc kubenswrapper[4870]: I1014 07:05:29.785027 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e9f8d5e1-367a-4cb2-9119-b5e6ae554774-utilities\") pod \"certified-operators-prqvc\" (UID: \"e9f8d5e1-367a-4cb2-9119-b5e6ae554774\") " pod="openshift-marketplace/certified-operators-prqvc" Oct 14 07:05:29 crc kubenswrapper[4870]: I1014 07:05:29.785838 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e9f8d5e1-367a-4cb2-9119-b5e6ae554774-catalog-content\") pod \"certified-operators-prqvc\" (UID: \"e9f8d5e1-367a-4cb2-9119-b5e6ae554774\") " pod="openshift-marketplace/certified-operators-prqvc" Oct 14 07:05:29 crc kubenswrapper[4870]: I1014 07:05:29.837784 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jpwcz\" (UniqueName: \"kubernetes.io/projected/e9f8d5e1-367a-4cb2-9119-b5e6ae554774-kube-api-access-jpwcz\") pod \"certified-operators-prqvc\" (UID: \"e9f8d5e1-367a-4cb2-9119-b5e6ae554774\") " pod="openshift-marketplace/certified-operators-prqvc" Oct 14 07:05:29 crc kubenswrapper[4870]: I1014 07:05:29.886142 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ad17975e-43e4-4fda-86a2-eb50ba104055-utilities\") pod \"community-operators-gx6t2\" (UID: \"ad17975e-43e4-4fda-86a2-eb50ba104055\") " pod="openshift-marketplace/community-operators-gx6t2" Oct 14 07:05:29 crc kubenswrapper[4870]: I1014 07:05:29.886572 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q4v8w\" (UniqueName: \"kubernetes.io/projected/ad17975e-43e4-4fda-86a2-eb50ba104055-kube-api-access-q4v8w\") pod \"community-operators-gx6t2\" (UID: \"ad17975e-43e4-4fda-86a2-eb50ba104055\") " pod="openshift-marketplace/community-operators-gx6t2" Oct 14 07:05:29 crc kubenswrapper[4870]: I1014 07:05:29.886646 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ad17975e-43e4-4fda-86a2-eb50ba104055-catalog-content\") pod \"community-operators-gx6t2\" (UID: \"ad17975e-43e4-4fda-86a2-eb50ba104055\") " pod="openshift-marketplace/community-operators-gx6t2" Oct 14 07:05:29 crc kubenswrapper[4870]: I1014 07:05:29.897871 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-prqvc" Oct 14 07:05:29 crc kubenswrapper[4870]: I1014 07:05:29.987768 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ad17975e-43e4-4fda-86a2-eb50ba104055-utilities\") pod \"community-operators-gx6t2\" (UID: \"ad17975e-43e4-4fda-86a2-eb50ba104055\") " pod="openshift-marketplace/community-operators-gx6t2" Oct 14 07:05:29 crc kubenswrapper[4870]: I1014 07:05:29.987814 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q4v8w\" (UniqueName: \"kubernetes.io/projected/ad17975e-43e4-4fda-86a2-eb50ba104055-kube-api-access-q4v8w\") pod \"community-operators-gx6t2\" (UID: \"ad17975e-43e4-4fda-86a2-eb50ba104055\") " pod="openshift-marketplace/community-operators-gx6t2" Oct 14 07:05:29 crc kubenswrapper[4870]: I1014 07:05:29.987868 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ad17975e-43e4-4fda-86a2-eb50ba104055-catalog-content\") pod \"community-operators-gx6t2\" (UID: \"ad17975e-43e4-4fda-86a2-eb50ba104055\") " pod="openshift-marketplace/community-operators-gx6t2" Oct 14 07:05:29 crc kubenswrapper[4870]: I1014 07:05:29.988319 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ad17975e-43e4-4fda-86a2-eb50ba104055-catalog-content\") pod \"community-operators-gx6t2\" (UID: \"ad17975e-43e4-4fda-86a2-eb50ba104055\") " pod="openshift-marketplace/community-operators-gx6t2" Oct 14 07:05:29 crc kubenswrapper[4870]: I1014 07:05:29.988834 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ad17975e-43e4-4fda-86a2-eb50ba104055-utilities\") pod \"community-operators-gx6t2\" (UID: \"ad17975e-43e4-4fda-86a2-eb50ba104055\") " pod="openshift-marketplace/community-operators-gx6t2" Oct 14 07:05:30 crc kubenswrapper[4870]: I1014 07:05:30.012863 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q4v8w\" (UniqueName: \"kubernetes.io/projected/ad17975e-43e4-4fda-86a2-eb50ba104055-kube-api-access-q4v8w\") pod \"community-operators-gx6t2\" (UID: \"ad17975e-43e4-4fda-86a2-eb50ba104055\") " pod="openshift-marketplace/community-operators-gx6t2" Oct 14 07:05:30 crc kubenswrapper[4870]: I1014 07:05:30.103525 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gx6t2" Oct 14 07:05:30 crc kubenswrapper[4870]: I1014 07:05:30.361626 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-prqvc"] Oct 14 07:05:30 crc kubenswrapper[4870]: I1014 07:05:30.518606 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-gx6t2"] Oct 14 07:05:30 crc kubenswrapper[4870]: W1014 07:05:30.585091 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podad17975e_43e4_4fda_86a2_eb50ba104055.slice/crio-e1cc2c228224d281e5dcd32cf752f38f21d8b80f4e483ed92cc34b09f507653e WatchSource:0}: Error finding container e1cc2c228224d281e5dcd32cf752f38f21d8b80f4e483ed92cc34b09f507653e: Status 404 returned error can't find the container with id e1cc2c228224d281e5dcd32cf752f38f21d8b80f4e483ed92cc34b09f507653e Oct 14 07:05:30 crc kubenswrapper[4870]: I1014 07:05:30.780732 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jqsnc" event={"ID":"08bc1ff9-0341-4e01-93ec-25d41eebf646","Type":"ContainerStarted","Data":"331add051ececd6164d3a3effe61d92ba394c7c7186c57ac0ed9fd11b7df42f2"} Oct 14 07:05:30 crc kubenswrapper[4870]: I1014 07:05:30.784052 4870 generic.go:334] "Generic (PLEG): container finished" podID="e9f8d5e1-367a-4cb2-9119-b5e6ae554774" containerID="6f17d50e1db785fa7808c819d3b4ed132c7ba8e6b8905cb4458606d91becabf9" exitCode=0 Oct 14 07:05:30 crc kubenswrapper[4870]: I1014 07:05:30.784148 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-prqvc" event={"ID":"e9f8d5e1-367a-4cb2-9119-b5e6ae554774","Type":"ContainerDied","Data":"6f17d50e1db785fa7808c819d3b4ed132c7ba8e6b8905cb4458606d91becabf9"} Oct 14 07:05:30 crc kubenswrapper[4870]: I1014 07:05:30.784204 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-prqvc" event={"ID":"e9f8d5e1-367a-4cb2-9119-b5e6ae554774","Type":"ContainerStarted","Data":"297c9e12ae12940b640ca790fcb54c4271767f051a67ff3110759dcd07c0f8cb"} Oct 14 07:05:30 crc kubenswrapper[4870]: I1014 07:05:30.796836 4870 generic.go:334] "Generic (PLEG): container finished" podID="9dc73b58-877c-4e60-8508-19ffeb7410f4" containerID="4399dd1c239bb7139648df62002064ff6fce443e8519e90cee4081de1f913d0a" exitCode=0 Oct 14 07:05:30 crc kubenswrapper[4870]: I1014 07:05:30.796973 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5t25t" event={"ID":"9dc73b58-877c-4e60-8508-19ffeb7410f4","Type":"ContainerDied","Data":"4399dd1c239bb7139648df62002064ff6fce443e8519e90cee4081de1f913d0a"} Oct 14 07:05:30 crc kubenswrapper[4870]: I1014 07:05:30.803325 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gx6t2" event={"ID":"ad17975e-43e4-4fda-86a2-eb50ba104055","Type":"ContainerStarted","Data":"dbeb0e2d82b30a81e19d4708dd5f09496b0f79175aa2a1e82c9c488114369d44"} Oct 14 07:05:30 crc kubenswrapper[4870]: I1014 07:05:30.803367 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gx6t2" event={"ID":"ad17975e-43e4-4fda-86a2-eb50ba104055","Type":"ContainerStarted","Data":"e1cc2c228224d281e5dcd32cf752f38f21d8b80f4e483ed92cc34b09f507653e"} Oct 14 07:05:30 crc kubenswrapper[4870]: I1014 07:05:30.810558 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-jqsnc" podStartSLOduration=2.264417686 podStartE2EDuration="3.810539005s" podCreationTimestamp="2025-10-14 07:05:27 +0000 UTC" firstStartedPulling="2025-10-14 07:05:28.759485054 +0000 UTC m=+264.456845425" lastFinishedPulling="2025-10-14 07:05:30.305606373 +0000 UTC m=+266.002966744" observedRunningTime="2025-10-14 07:05:30.801506199 +0000 UTC m=+266.498866590" watchObservedRunningTime="2025-10-14 07:05:30.810539005 +0000 UTC m=+266.507899376" Oct 14 07:05:31 crc kubenswrapper[4870]: I1014 07:05:31.814510 4870 generic.go:334] "Generic (PLEG): container finished" podID="e9f8d5e1-367a-4cb2-9119-b5e6ae554774" containerID="0366c38284490d2eacef29960d73b967994d81a6595874a1d5ae8f6cbb3d9f55" exitCode=0 Oct 14 07:05:31 crc kubenswrapper[4870]: I1014 07:05:31.814655 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-prqvc" event={"ID":"e9f8d5e1-367a-4cb2-9119-b5e6ae554774","Type":"ContainerDied","Data":"0366c38284490d2eacef29960d73b967994d81a6595874a1d5ae8f6cbb3d9f55"} Oct 14 07:05:31 crc kubenswrapper[4870]: I1014 07:05:31.821902 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5t25t" event={"ID":"9dc73b58-877c-4e60-8508-19ffeb7410f4","Type":"ContainerStarted","Data":"056466e4ca8980a8e9136d1cfbe342617256392c6762318dd651a7b3a6a7449b"} Oct 14 07:05:31 crc kubenswrapper[4870]: I1014 07:05:31.826387 4870 generic.go:334] "Generic (PLEG): container finished" podID="ad17975e-43e4-4fda-86a2-eb50ba104055" containerID="dbeb0e2d82b30a81e19d4708dd5f09496b0f79175aa2a1e82c9c488114369d44" exitCode=0 Oct 14 07:05:31 crc kubenswrapper[4870]: I1014 07:05:31.827782 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gx6t2" event={"ID":"ad17975e-43e4-4fda-86a2-eb50ba104055","Type":"ContainerDied","Data":"dbeb0e2d82b30a81e19d4708dd5f09496b0f79175aa2a1e82c9c488114369d44"} Oct 14 07:05:31 crc kubenswrapper[4870]: I1014 07:05:31.827829 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gx6t2" event={"ID":"ad17975e-43e4-4fda-86a2-eb50ba104055","Type":"ContainerStarted","Data":"e02875bb6687d2c29029f9c42337dfd4c0b09c990b10d8862351990437a0a8c5"} Oct 14 07:05:31 crc kubenswrapper[4870]: I1014 07:05:31.861002 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-5t25t" podStartSLOduration=2.148532877 podStartE2EDuration="4.860981285s" podCreationTimestamp="2025-10-14 07:05:27 +0000 UTC" firstStartedPulling="2025-10-14 07:05:28.754408347 +0000 UTC m=+264.451768718" lastFinishedPulling="2025-10-14 07:05:31.466856755 +0000 UTC m=+267.164217126" observedRunningTime="2025-10-14 07:05:31.858820101 +0000 UTC m=+267.556180472" watchObservedRunningTime="2025-10-14 07:05:31.860981285 +0000 UTC m=+267.558341656" Oct 14 07:05:32 crc kubenswrapper[4870]: I1014 07:05:32.840375 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-prqvc" event={"ID":"e9f8d5e1-367a-4cb2-9119-b5e6ae554774","Type":"ContainerStarted","Data":"1f13fb0ab1784450611a87ed8e4a2a42e418297c5c135212914c66fdb944f8f8"} Oct 14 07:05:32 crc kubenswrapper[4870]: I1014 07:05:32.842431 4870 generic.go:334] "Generic (PLEG): container finished" podID="ad17975e-43e4-4fda-86a2-eb50ba104055" containerID="e02875bb6687d2c29029f9c42337dfd4c0b09c990b10d8862351990437a0a8c5" exitCode=0 Oct 14 07:05:32 crc kubenswrapper[4870]: I1014 07:05:32.842517 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gx6t2" event={"ID":"ad17975e-43e4-4fda-86a2-eb50ba104055","Type":"ContainerDied","Data":"e02875bb6687d2c29029f9c42337dfd4c0b09c990b10d8862351990437a0a8c5"} Oct 14 07:05:32 crc kubenswrapper[4870]: I1014 07:05:32.870327 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-prqvc" podStartSLOduration=2.31400833 podStartE2EDuration="3.870304464s" podCreationTimestamp="2025-10-14 07:05:29 +0000 UTC" firstStartedPulling="2025-10-14 07:05:30.785333385 +0000 UTC m=+266.482693746" lastFinishedPulling="2025-10-14 07:05:32.341629489 +0000 UTC m=+268.038989880" observedRunningTime="2025-10-14 07:05:32.867511304 +0000 UTC m=+268.564871665" watchObservedRunningTime="2025-10-14 07:05:32.870304464 +0000 UTC m=+268.567664835" Oct 14 07:05:34 crc kubenswrapper[4870]: I1014 07:05:34.862208 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gx6t2" event={"ID":"ad17975e-43e4-4fda-86a2-eb50ba104055","Type":"ContainerStarted","Data":"39672b19a3ed09209a2959f962d538ebd415f15ec16ce8b7054ec9bafcafe900"} Oct 14 07:05:34 crc kubenswrapper[4870]: I1014 07:05:34.887696 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-gx6t2" podStartSLOduration=3.044723091 podStartE2EDuration="5.887662123s" podCreationTimestamp="2025-10-14 07:05:29 +0000 UTC" firstStartedPulling="2025-10-14 07:05:30.805106289 +0000 UTC m=+266.502466660" lastFinishedPulling="2025-10-14 07:05:33.648045321 +0000 UTC m=+269.345405692" observedRunningTime="2025-10-14 07:05:34.883884339 +0000 UTC m=+270.581244710" watchObservedRunningTime="2025-10-14 07:05:34.887662123 +0000 UTC m=+270.585022494" Oct 14 07:05:37 crc kubenswrapper[4870]: I1014 07:05:37.501745 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-jqsnc" Oct 14 07:05:37 crc kubenswrapper[4870]: I1014 07:05:37.502129 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-jqsnc" Oct 14 07:05:37 crc kubenswrapper[4870]: I1014 07:05:37.579880 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-jqsnc" Oct 14 07:05:37 crc kubenswrapper[4870]: I1014 07:05:37.703910 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-5t25t" Oct 14 07:05:37 crc kubenswrapper[4870]: I1014 07:05:37.706656 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-5t25t" Oct 14 07:05:37 crc kubenswrapper[4870]: I1014 07:05:37.789312 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-5t25t" Oct 14 07:05:37 crc kubenswrapper[4870]: I1014 07:05:37.918588 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-5t25t" Oct 14 07:05:37 crc kubenswrapper[4870]: I1014 07:05:37.950414 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-jqsnc" Oct 14 07:05:39 crc kubenswrapper[4870]: I1014 07:05:39.898631 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-prqvc" Oct 14 07:05:39 crc kubenswrapper[4870]: I1014 07:05:39.899909 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-prqvc" Oct 14 07:05:39 crc kubenswrapper[4870]: I1014 07:05:39.942817 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-prqvc" Oct 14 07:05:40 crc kubenswrapper[4870]: I1014 07:05:40.103690 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-gx6t2" Oct 14 07:05:40 crc kubenswrapper[4870]: I1014 07:05:40.106012 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-gx6t2" Oct 14 07:05:40 crc kubenswrapper[4870]: I1014 07:05:40.163494 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-gx6t2" Oct 14 07:05:40 crc kubenswrapper[4870]: I1014 07:05:40.946821 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-gx6t2" Oct 14 07:05:40 crc kubenswrapper[4870]: I1014 07:05:40.968590 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-prqvc" Oct 14 07:07:23 crc kubenswrapper[4870]: I1014 07:07:23.951614 4870 patch_prober.go:28] interesting pod/machine-config-daemon-7tvc7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 07:07:23 crc kubenswrapper[4870]: I1014 07:07:23.952329 4870 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 07:07:53 crc kubenswrapper[4870]: I1014 07:07:53.951203 4870 patch_prober.go:28] interesting pod/machine-config-daemon-7tvc7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 07:07:53 crc kubenswrapper[4870]: I1014 07:07:53.952535 4870 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 07:08:23 crc kubenswrapper[4870]: I1014 07:08:23.951638 4870 patch_prober.go:28] interesting pod/machine-config-daemon-7tvc7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 07:08:23 crc kubenswrapper[4870]: I1014 07:08:23.952377 4870 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 07:08:23 crc kubenswrapper[4870]: I1014 07:08:23.952519 4870 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" Oct 14 07:08:23 crc kubenswrapper[4870]: I1014 07:08:23.953602 4870 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"fc9b159a611a2f80605df8d6e70c1ccf922f37ec81b713e826ba0a3e8621d3fa"} pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 14 07:08:23 crc kubenswrapper[4870]: I1014 07:08:23.953731 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerName="machine-config-daemon" containerID="cri-o://fc9b159a611a2f80605df8d6e70c1ccf922f37ec81b713e826ba0a3e8621d3fa" gracePeriod=600 Oct 14 07:08:24 crc kubenswrapper[4870]: I1014 07:08:24.244218 4870 generic.go:334] "Generic (PLEG): container finished" podID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerID="fc9b159a611a2f80605df8d6e70c1ccf922f37ec81b713e826ba0a3e8621d3fa" exitCode=0 Oct 14 07:08:24 crc kubenswrapper[4870]: I1014 07:08:24.244339 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" event={"ID":"874a5e45-dffd-4d17-b609-b7d3ed2eab07","Type":"ContainerDied","Data":"fc9b159a611a2f80605df8d6e70c1ccf922f37ec81b713e826ba0a3e8621d3fa"} Oct 14 07:08:24 crc kubenswrapper[4870]: I1014 07:08:24.244860 4870 scope.go:117] "RemoveContainer" containerID="8004752575df4a1c44efbcafe64799f09126d0f97285d92388c76a0b90e022f8" Oct 14 07:08:25 crc kubenswrapper[4870]: I1014 07:08:25.266356 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" event={"ID":"874a5e45-dffd-4d17-b609-b7d3ed2eab07","Type":"ContainerStarted","Data":"f5c863e175088c68c7d18f66c4b86bb42e49b43e1b581fdc20f0b5d02ea623e6"} Oct 14 07:08:54 crc kubenswrapper[4870]: I1014 07:08:54.323917 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-qmfgz"] Oct 14 07:08:54 crc kubenswrapper[4870]: I1014 07:08:54.325277 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-qmfgz" Oct 14 07:08:54 crc kubenswrapper[4870]: I1014 07:08:54.343621 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-qmfgz"] Oct 14 07:08:54 crc kubenswrapper[4870]: I1014 07:08:54.465265 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7ndtl\" (UniqueName: \"kubernetes.io/projected/fa11fb26-ef21-4072-b1a4-0453acb3439b-kube-api-access-7ndtl\") pod \"image-registry-66df7c8f76-qmfgz\" (UID: \"fa11fb26-ef21-4072-b1a4-0453acb3439b\") " pod="openshift-image-registry/image-registry-66df7c8f76-qmfgz" Oct 14 07:08:54 crc kubenswrapper[4870]: I1014 07:08:54.465322 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/fa11fb26-ef21-4072-b1a4-0453acb3439b-trusted-ca\") pod \"image-registry-66df7c8f76-qmfgz\" (UID: \"fa11fb26-ef21-4072-b1a4-0453acb3439b\") " pod="openshift-image-registry/image-registry-66df7c8f76-qmfgz" Oct 14 07:08:54 crc kubenswrapper[4870]: I1014 07:08:54.465361 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/fa11fb26-ef21-4072-b1a4-0453acb3439b-bound-sa-token\") pod \"image-registry-66df7c8f76-qmfgz\" (UID: \"fa11fb26-ef21-4072-b1a4-0453acb3439b\") " pod="openshift-image-registry/image-registry-66df7c8f76-qmfgz" Oct 14 07:08:54 crc kubenswrapper[4870]: I1014 07:08:54.465568 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/fa11fb26-ef21-4072-b1a4-0453acb3439b-registry-certificates\") pod \"image-registry-66df7c8f76-qmfgz\" (UID: \"fa11fb26-ef21-4072-b1a4-0453acb3439b\") " pod="openshift-image-registry/image-registry-66df7c8f76-qmfgz" Oct 14 07:08:54 crc kubenswrapper[4870]: I1014 07:08:54.465632 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/fa11fb26-ef21-4072-b1a4-0453acb3439b-ca-trust-extracted\") pod \"image-registry-66df7c8f76-qmfgz\" (UID: \"fa11fb26-ef21-4072-b1a4-0453acb3439b\") " pod="openshift-image-registry/image-registry-66df7c8f76-qmfgz" Oct 14 07:08:54 crc kubenswrapper[4870]: I1014 07:08:54.465673 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/fa11fb26-ef21-4072-b1a4-0453acb3439b-installation-pull-secrets\") pod \"image-registry-66df7c8f76-qmfgz\" (UID: \"fa11fb26-ef21-4072-b1a4-0453acb3439b\") " pod="openshift-image-registry/image-registry-66df7c8f76-qmfgz" Oct 14 07:08:54 crc kubenswrapper[4870]: I1014 07:08:54.465756 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-qmfgz\" (UID: \"fa11fb26-ef21-4072-b1a4-0453acb3439b\") " pod="openshift-image-registry/image-registry-66df7c8f76-qmfgz" Oct 14 07:08:54 crc kubenswrapper[4870]: I1014 07:08:54.465800 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/fa11fb26-ef21-4072-b1a4-0453acb3439b-registry-tls\") pod \"image-registry-66df7c8f76-qmfgz\" (UID: \"fa11fb26-ef21-4072-b1a4-0453acb3439b\") " pod="openshift-image-registry/image-registry-66df7c8f76-qmfgz" Oct 14 07:08:54 crc kubenswrapper[4870]: I1014 07:08:54.491035 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-qmfgz\" (UID: \"fa11fb26-ef21-4072-b1a4-0453acb3439b\") " pod="openshift-image-registry/image-registry-66df7c8f76-qmfgz" Oct 14 07:08:54 crc kubenswrapper[4870]: I1014 07:08:54.568266 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/fa11fb26-ef21-4072-b1a4-0453acb3439b-registry-tls\") pod \"image-registry-66df7c8f76-qmfgz\" (UID: \"fa11fb26-ef21-4072-b1a4-0453acb3439b\") " pod="openshift-image-registry/image-registry-66df7c8f76-qmfgz" Oct 14 07:08:54 crc kubenswrapper[4870]: I1014 07:08:54.568328 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7ndtl\" (UniqueName: \"kubernetes.io/projected/fa11fb26-ef21-4072-b1a4-0453acb3439b-kube-api-access-7ndtl\") pod \"image-registry-66df7c8f76-qmfgz\" (UID: \"fa11fb26-ef21-4072-b1a4-0453acb3439b\") " pod="openshift-image-registry/image-registry-66df7c8f76-qmfgz" Oct 14 07:08:54 crc kubenswrapper[4870]: I1014 07:08:54.568353 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/fa11fb26-ef21-4072-b1a4-0453acb3439b-trusted-ca\") pod \"image-registry-66df7c8f76-qmfgz\" (UID: \"fa11fb26-ef21-4072-b1a4-0453acb3439b\") " pod="openshift-image-registry/image-registry-66df7c8f76-qmfgz" Oct 14 07:08:54 crc kubenswrapper[4870]: I1014 07:08:54.568386 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/fa11fb26-ef21-4072-b1a4-0453acb3439b-bound-sa-token\") pod \"image-registry-66df7c8f76-qmfgz\" (UID: \"fa11fb26-ef21-4072-b1a4-0453acb3439b\") " pod="openshift-image-registry/image-registry-66df7c8f76-qmfgz" Oct 14 07:08:54 crc kubenswrapper[4870]: I1014 07:08:54.568428 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/fa11fb26-ef21-4072-b1a4-0453acb3439b-registry-certificates\") pod \"image-registry-66df7c8f76-qmfgz\" (UID: \"fa11fb26-ef21-4072-b1a4-0453acb3439b\") " pod="openshift-image-registry/image-registry-66df7c8f76-qmfgz" Oct 14 07:08:54 crc kubenswrapper[4870]: I1014 07:08:54.568477 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/fa11fb26-ef21-4072-b1a4-0453acb3439b-ca-trust-extracted\") pod \"image-registry-66df7c8f76-qmfgz\" (UID: \"fa11fb26-ef21-4072-b1a4-0453acb3439b\") " pod="openshift-image-registry/image-registry-66df7c8f76-qmfgz" Oct 14 07:08:54 crc kubenswrapper[4870]: I1014 07:08:54.568531 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/fa11fb26-ef21-4072-b1a4-0453acb3439b-installation-pull-secrets\") pod \"image-registry-66df7c8f76-qmfgz\" (UID: \"fa11fb26-ef21-4072-b1a4-0453acb3439b\") " pod="openshift-image-registry/image-registry-66df7c8f76-qmfgz" Oct 14 07:08:54 crc kubenswrapper[4870]: I1014 07:08:54.569999 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/fa11fb26-ef21-4072-b1a4-0453acb3439b-ca-trust-extracted\") pod \"image-registry-66df7c8f76-qmfgz\" (UID: \"fa11fb26-ef21-4072-b1a4-0453acb3439b\") " pod="openshift-image-registry/image-registry-66df7c8f76-qmfgz" Oct 14 07:08:54 crc kubenswrapper[4870]: I1014 07:08:54.570579 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/fa11fb26-ef21-4072-b1a4-0453acb3439b-trusted-ca\") pod \"image-registry-66df7c8f76-qmfgz\" (UID: \"fa11fb26-ef21-4072-b1a4-0453acb3439b\") " pod="openshift-image-registry/image-registry-66df7c8f76-qmfgz" Oct 14 07:08:54 crc kubenswrapper[4870]: I1014 07:08:54.570875 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/fa11fb26-ef21-4072-b1a4-0453acb3439b-registry-certificates\") pod \"image-registry-66df7c8f76-qmfgz\" (UID: \"fa11fb26-ef21-4072-b1a4-0453acb3439b\") " pod="openshift-image-registry/image-registry-66df7c8f76-qmfgz" Oct 14 07:08:54 crc kubenswrapper[4870]: I1014 07:08:54.574100 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/fa11fb26-ef21-4072-b1a4-0453acb3439b-registry-tls\") pod \"image-registry-66df7c8f76-qmfgz\" (UID: \"fa11fb26-ef21-4072-b1a4-0453acb3439b\") " pod="openshift-image-registry/image-registry-66df7c8f76-qmfgz" Oct 14 07:08:54 crc kubenswrapper[4870]: I1014 07:08:54.574467 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/fa11fb26-ef21-4072-b1a4-0453acb3439b-installation-pull-secrets\") pod \"image-registry-66df7c8f76-qmfgz\" (UID: \"fa11fb26-ef21-4072-b1a4-0453acb3439b\") " pod="openshift-image-registry/image-registry-66df7c8f76-qmfgz" Oct 14 07:08:54 crc kubenswrapper[4870]: I1014 07:08:54.585654 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7ndtl\" (UniqueName: \"kubernetes.io/projected/fa11fb26-ef21-4072-b1a4-0453acb3439b-kube-api-access-7ndtl\") pod \"image-registry-66df7c8f76-qmfgz\" (UID: \"fa11fb26-ef21-4072-b1a4-0453acb3439b\") " pod="openshift-image-registry/image-registry-66df7c8f76-qmfgz" Oct 14 07:08:54 crc kubenswrapper[4870]: I1014 07:08:54.586045 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/fa11fb26-ef21-4072-b1a4-0453acb3439b-bound-sa-token\") pod \"image-registry-66df7c8f76-qmfgz\" (UID: \"fa11fb26-ef21-4072-b1a4-0453acb3439b\") " pod="openshift-image-registry/image-registry-66df7c8f76-qmfgz" Oct 14 07:08:54 crc kubenswrapper[4870]: I1014 07:08:54.643887 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-qmfgz" Oct 14 07:08:55 crc kubenswrapper[4870]: I1014 07:08:55.071514 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-qmfgz"] Oct 14 07:08:55 crc kubenswrapper[4870]: I1014 07:08:55.482289 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-qmfgz" event={"ID":"fa11fb26-ef21-4072-b1a4-0453acb3439b","Type":"ContainerStarted","Data":"45ec2edd4bd3b0033bed4f8b3a30288fa0dade18e5749a9fba0ade04bfd4967b"} Oct 14 07:08:55 crc kubenswrapper[4870]: I1014 07:08:55.482864 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-qmfgz" Oct 14 07:08:55 crc kubenswrapper[4870]: I1014 07:08:55.482895 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-qmfgz" event={"ID":"fa11fb26-ef21-4072-b1a4-0453acb3439b","Type":"ContainerStarted","Data":"66922ba4f82a6c2edc3f6ef72b9b8291798064bf2efdfaecaa707cef5c9194ae"} Oct 14 07:08:55 crc kubenswrapper[4870]: I1014 07:08:55.525123 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-qmfgz" podStartSLOduration=1.52509261 podStartE2EDuration="1.52509261s" podCreationTimestamp="2025-10-14 07:08:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:08:55.518142149 +0000 UTC m=+471.215502550" watchObservedRunningTime="2025-10-14 07:08:55.52509261 +0000 UTC m=+471.222453011" Oct 14 07:09:14 crc kubenswrapper[4870]: I1014 07:09:14.657130 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-qmfgz" Oct 14 07:09:14 crc kubenswrapper[4870]: I1014 07:09:14.751457 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-kkxjz"] Oct 14 07:09:39 crc kubenswrapper[4870]: I1014 07:09:39.808268 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-kkxjz" podUID="2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3" containerName="registry" containerID="cri-o://c708c420a912b0bab2c3a334dfe984c05b1212497c3eadbbf22c03db8be5eb03" gracePeriod=30 Oct 14 07:09:40 crc kubenswrapper[4870]: I1014 07:09:40.234334 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-kkxjz" Oct 14 07:09:40 crc kubenswrapper[4870]: I1014 07:09:40.323173 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3-registry-tls\") pod \"2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3\" (UID: \"2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3\") " Oct 14 07:09:40 crc kubenswrapper[4870]: I1014 07:09:40.323303 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3-installation-pull-secrets\") pod \"2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3\" (UID: \"2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3\") " Oct 14 07:09:40 crc kubenswrapper[4870]: I1014 07:09:40.323344 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3-ca-trust-extracted\") pod \"2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3\" (UID: \"2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3\") " Oct 14 07:09:40 crc kubenswrapper[4870]: I1014 07:09:40.323383 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3-registry-certificates\") pod \"2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3\" (UID: \"2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3\") " Oct 14 07:09:40 crc kubenswrapper[4870]: I1014 07:09:40.323414 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3-trusted-ca\") pod \"2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3\" (UID: \"2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3\") " Oct 14 07:09:40 crc kubenswrapper[4870]: I1014 07:09:40.323717 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3\" (UID: \"2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3\") " Oct 14 07:09:40 crc kubenswrapper[4870]: I1014 07:09:40.323765 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-24b47\" (UniqueName: \"kubernetes.io/projected/2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3-kube-api-access-24b47\") pod \"2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3\" (UID: \"2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3\") " Oct 14 07:09:40 crc kubenswrapper[4870]: I1014 07:09:40.323799 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3-bound-sa-token\") pod \"2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3\" (UID: \"2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3\") " Oct 14 07:09:40 crc kubenswrapper[4870]: I1014 07:09:40.324523 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3" (UID: "2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:09:40 crc kubenswrapper[4870]: I1014 07:09:40.324593 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3" (UID: "2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:09:40 crc kubenswrapper[4870]: I1014 07:09:40.330177 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3" (UID: "2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:09:40 crc kubenswrapper[4870]: I1014 07:09:40.330750 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3" (UID: "2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:09:40 crc kubenswrapper[4870]: I1014 07:09:40.331309 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3" (UID: "2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:09:40 crc kubenswrapper[4870]: I1014 07:09:40.331821 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3-kube-api-access-24b47" (OuterVolumeSpecName: "kube-api-access-24b47") pod "2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3" (UID: "2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3"). InnerVolumeSpecName "kube-api-access-24b47". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:09:40 crc kubenswrapper[4870]: I1014 07:09:40.341817 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3" (UID: "2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:09:40 crc kubenswrapper[4870]: I1014 07:09:40.342753 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3" (UID: "2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 14 07:09:40 crc kubenswrapper[4870]: I1014 07:09:40.425049 4870 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Oct 14 07:09:40 crc kubenswrapper[4870]: I1014 07:09:40.425121 4870 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Oct 14 07:09:40 crc kubenswrapper[4870]: I1014 07:09:40.425159 4870 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3-registry-certificates\") on node \"crc\" DevicePath \"\"" Oct 14 07:09:40 crc kubenswrapper[4870]: I1014 07:09:40.425186 4870 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 14 07:09:40 crc kubenswrapper[4870]: I1014 07:09:40.425677 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-24b47\" (UniqueName: \"kubernetes.io/projected/2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3-kube-api-access-24b47\") on node \"crc\" DevicePath \"\"" Oct 14 07:09:40 crc kubenswrapper[4870]: I1014 07:09:40.425805 4870 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 14 07:09:40 crc kubenswrapper[4870]: I1014 07:09:40.425874 4870 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3-registry-tls\") on node \"crc\" DevicePath \"\"" Oct 14 07:09:40 crc kubenswrapper[4870]: I1014 07:09:40.828965 4870 generic.go:334] "Generic (PLEG): container finished" podID="2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3" containerID="c708c420a912b0bab2c3a334dfe984c05b1212497c3eadbbf22c03db8be5eb03" exitCode=0 Oct 14 07:09:40 crc kubenswrapper[4870]: I1014 07:09:40.829047 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-kkxjz" event={"ID":"2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3","Type":"ContainerDied","Data":"c708c420a912b0bab2c3a334dfe984c05b1212497c3eadbbf22c03db8be5eb03"} Oct 14 07:09:40 crc kubenswrapper[4870]: I1014 07:09:40.829115 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-kkxjz" event={"ID":"2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3","Type":"ContainerDied","Data":"e42a9ade61575cb390af63785857635f25fc9e9808cf54a155343a03b597b870"} Oct 14 07:09:40 crc kubenswrapper[4870]: I1014 07:09:40.829134 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-kkxjz" Oct 14 07:09:40 crc kubenswrapper[4870]: I1014 07:09:40.829151 4870 scope.go:117] "RemoveContainer" containerID="c708c420a912b0bab2c3a334dfe984c05b1212497c3eadbbf22c03db8be5eb03" Oct 14 07:09:40 crc kubenswrapper[4870]: I1014 07:09:40.850246 4870 scope.go:117] "RemoveContainer" containerID="c708c420a912b0bab2c3a334dfe984c05b1212497c3eadbbf22c03db8be5eb03" Oct 14 07:09:40 crc kubenswrapper[4870]: E1014 07:09:40.851068 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c708c420a912b0bab2c3a334dfe984c05b1212497c3eadbbf22c03db8be5eb03\": container with ID starting with c708c420a912b0bab2c3a334dfe984c05b1212497c3eadbbf22c03db8be5eb03 not found: ID does not exist" containerID="c708c420a912b0bab2c3a334dfe984c05b1212497c3eadbbf22c03db8be5eb03" Oct 14 07:09:40 crc kubenswrapper[4870]: I1014 07:09:40.851122 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c708c420a912b0bab2c3a334dfe984c05b1212497c3eadbbf22c03db8be5eb03"} err="failed to get container status \"c708c420a912b0bab2c3a334dfe984c05b1212497c3eadbbf22c03db8be5eb03\": rpc error: code = NotFound desc = could not find container \"c708c420a912b0bab2c3a334dfe984c05b1212497c3eadbbf22c03db8be5eb03\": container with ID starting with c708c420a912b0bab2c3a334dfe984c05b1212497c3eadbbf22c03db8be5eb03 not found: ID does not exist" Oct 14 07:09:40 crc kubenswrapper[4870]: I1014 07:09:40.874348 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-kkxjz"] Oct 14 07:09:40 crc kubenswrapper[4870]: I1014 07:09:40.877961 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-kkxjz"] Oct 14 07:09:41 crc kubenswrapper[4870]: I1014 07:09:41.046268 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3" path="/var/lib/kubelet/pods/2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3/volumes" Oct 14 07:10:53 crc kubenswrapper[4870]: I1014 07:10:53.951510 4870 patch_prober.go:28] interesting pod/machine-config-daemon-7tvc7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 07:10:53 crc kubenswrapper[4870]: I1014 07:10:53.952177 4870 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 07:11:23 crc kubenswrapper[4870]: I1014 07:11:23.950938 4870 patch_prober.go:28] interesting pod/machine-config-daemon-7tvc7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 07:11:23 crc kubenswrapper[4870]: I1014 07:11:23.951849 4870 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 07:11:53 crc kubenswrapper[4870]: I1014 07:11:53.951285 4870 patch_prober.go:28] interesting pod/machine-config-daemon-7tvc7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 07:11:53 crc kubenswrapper[4870]: I1014 07:11:53.952106 4870 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 07:11:53 crc kubenswrapper[4870]: I1014 07:11:53.952172 4870 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" Oct 14 07:11:53 crc kubenswrapper[4870]: I1014 07:11:53.953031 4870 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f5c863e175088c68c7d18f66c4b86bb42e49b43e1b581fdc20f0b5d02ea623e6"} pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 14 07:11:53 crc kubenswrapper[4870]: I1014 07:11:53.953113 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerName="machine-config-daemon" containerID="cri-o://f5c863e175088c68c7d18f66c4b86bb42e49b43e1b581fdc20f0b5d02ea623e6" gracePeriod=600 Oct 14 07:11:54 crc kubenswrapper[4870]: E1014 07:11:54.107926 4870 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod874a5e45_dffd_4d17_b609_b7d3ed2eab07.slice/crio-conmon-f5c863e175088c68c7d18f66c4b86bb42e49b43e1b581fdc20f0b5d02ea623e6.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod874a5e45_dffd_4d17_b609_b7d3ed2eab07.slice/crio-f5c863e175088c68c7d18f66c4b86bb42e49b43e1b581fdc20f0b5d02ea623e6.scope\": RecentStats: unable to find data in memory cache]" Oct 14 07:11:54 crc kubenswrapper[4870]: I1014 07:11:54.885544 4870 generic.go:334] "Generic (PLEG): container finished" podID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerID="f5c863e175088c68c7d18f66c4b86bb42e49b43e1b581fdc20f0b5d02ea623e6" exitCode=0 Oct 14 07:11:54 crc kubenswrapper[4870]: I1014 07:11:54.885643 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" event={"ID":"874a5e45-dffd-4d17-b609-b7d3ed2eab07","Type":"ContainerDied","Data":"f5c863e175088c68c7d18f66c4b86bb42e49b43e1b581fdc20f0b5d02ea623e6"} Oct 14 07:11:54 crc kubenswrapper[4870]: I1014 07:11:54.885900 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" event={"ID":"874a5e45-dffd-4d17-b609-b7d3ed2eab07","Type":"ContainerStarted","Data":"b5bea711ed09c7366862c920d932bb6dfefddcdd90ae2ab8981cfd9449e357a2"} Oct 14 07:11:54 crc kubenswrapper[4870]: I1014 07:11:54.885925 4870 scope.go:117] "RemoveContainer" containerID="fc9b159a611a2f80605df8d6e70c1ccf922f37ec81b713e826ba0a3e8621d3fa" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.182126 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-v7br2"] Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.183486 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-v7br2" podUID="8dcaef0e-42e3-424b-bbe5-4a023efe912d" containerName="ovn-controller" containerID="cri-o://f3b6d20cf459b1f6d3aff5b9290f9391909baf27ee40c49e2b0cd2d5cfe889ad" gracePeriod=30 Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.184087 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-v7br2" podUID="8dcaef0e-42e3-424b-bbe5-4a023efe912d" containerName="sbdb" containerID="cri-o://9fe3468cdd612fc4378e1f6f55a2a9d00274f7ddb65cb57b2a368dfc065cfb4e" gracePeriod=30 Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.184169 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-v7br2" podUID="8dcaef0e-42e3-424b-bbe5-4a023efe912d" containerName="nbdb" containerID="cri-o://fc9d4db944f837289d1923176d60cfc4a566a282c2d2e3a703e220df884703c7" gracePeriod=30 Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.184230 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-v7br2" podUID="8dcaef0e-42e3-424b-bbe5-4a023efe912d" containerName="northd" containerID="cri-o://d16ed1c65925549dcb69ab02c3013e25e3f123a83179f1af3a8e08bf91a59896" gracePeriod=30 Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.184287 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-v7br2" podUID="8dcaef0e-42e3-424b-bbe5-4a023efe912d" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://e72cff0723e2ce520340244b4e597d8016c58e1b071c4b087f2f6cd718741594" gracePeriod=30 Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.184364 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-v7br2" podUID="8dcaef0e-42e3-424b-bbe5-4a023efe912d" containerName="kube-rbac-proxy-node" containerID="cri-o://71dec55e859b11a081dbb90d7d6cf6b4846cf591382303d8329003ff50c71f2a" gracePeriod=30 Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.184487 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-v7br2" podUID="8dcaef0e-42e3-424b-bbe5-4a023efe912d" containerName="ovn-acl-logging" containerID="cri-o://ed59243b59c39ce82bf9bc8f33c0a8b84854846afff6a89f5d37ffc97a096b7c" gracePeriod=30 Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.226193 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-v7br2" podUID="8dcaef0e-42e3-424b-bbe5-4a023efe912d" containerName="ovnkube-controller" containerID="cri-o://a23b08d1e6780f4c2ce9ff447bd40050abb1e353c16ebccc8c3ecc7bc9a35e52" gracePeriod=30 Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.524739 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-v7br2_8dcaef0e-42e3-424b-bbe5-4a023efe912d/ovnkube-controller/3.log" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.527928 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-v7br2_8dcaef0e-42e3-424b-bbe5-4a023efe912d/ovn-acl-logging/0.log" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.528735 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-v7br2_8dcaef0e-42e3-424b-bbe5-4a023efe912d/ovn-controller/0.log" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.529408 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-v7br2" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.548313 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-fdxjh_8a3a430a-b540-4a8c-adad-b2893dbf9898/kube-multus/2.log" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.548838 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-fdxjh_8a3a430a-b540-4a8c-adad-b2893dbf9898/kube-multus/1.log" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.548947 4870 generic.go:334] "Generic (PLEG): container finished" podID="8a3a430a-b540-4a8c-adad-b2893dbf9898" containerID="33696bfa1ec686bccf5f7def87cbf3dc2186321f9ba1c180148db5604f27cc90" exitCode=2 Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.549034 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-fdxjh" event={"ID":"8a3a430a-b540-4a8c-adad-b2893dbf9898","Type":"ContainerDied","Data":"33696bfa1ec686bccf5f7def87cbf3dc2186321f9ba1c180148db5604f27cc90"} Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.549086 4870 scope.go:117] "RemoveContainer" containerID="88614373bfe4779bbe61cc41cc4c35f419e438bdae3cec4964271bf0f74ad3f5" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.550048 4870 scope.go:117] "RemoveContainer" containerID="33696bfa1ec686bccf5f7def87cbf3dc2186321f9ba1c180148db5604f27cc90" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.554938 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-v7br2_8dcaef0e-42e3-424b-bbe5-4a023efe912d/ovnkube-controller/3.log" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.558104 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-v7br2_8dcaef0e-42e3-424b-bbe5-4a023efe912d/ovn-acl-logging/0.log" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.560089 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-v7br2_8dcaef0e-42e3-424b-bbe5-4a023efe912d/ovn-controller/0.log" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.562889 4870 generic.go:334] "Generic (PLEG): container finished" podID="8dcaef0e-42e3-424b-bbe5-4a023efe912d" containerID="a23b08d1e6780f4c2ce9ff447bd40050abb1e353c16ebccc8c3ecc7bc9a35e52" exitCode=0 Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.562915 4870 generic.go:334] "Generic (PLEG): container finished" podID="8dcaef0e-42e3-424b-bbe5-4a023efe912d" containerID="9fe3468cdd612fc4378e1f6f55a2a9d00274f7ddb65cb57b2a368dfc065cfb4e" exitCode=0 Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.562926 4870 generic.go:334] "Generic (PLEG): container finished" podID="8dcaef0e-42e3-424b-bbe5-4a023efe912d" containerID="fc9d4db944f837289d1923176d60cfc4a566a282c2d2e3a703e220df884703c7" exitCode=0 Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.562934 4870 generic.go:334] "Generic (PLEG): container finished" podID="8dcaef0e-42e3-424b-bbe5-4a023efe912d" containerID="d16ed1c65925549dcb69ab02c3013e25e3f123a83179f1af3a8e08bf91a59896" exitCode=0 Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.562943 4870 generic.go:334] "Generic (PLEG): container finished" podID="8dcaef0e-42e3-424b-bbe5-4a023efe912d" containerID="e72cff0723e2ce520340244b4e597d8016c58e1b071c4b087f2f6cd718741594" exitCode=0 Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.562951 4870 generic.go:334] "Generic (PLEG): container finished" podID="8dcaef0e-42e3-424b-bbe5-4a023efe912d" containerID="71dec55e859b11a081dbb90d7d6cf6b4846cf591382303d8329003ff50c71f2a" exitCode=0 Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.562960 4870 generic.go:334] "Generic (PLEG): container finished" podID="8dcaef0e-42e3-424b-bbe5-4a023efe912d" containerID="ed59243b59c39ce82bf9bc8f33c0a8b84854846afff6a89f5d37ffc97a096b7c" exitCode=143 Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.562969 4870 generic.go:334] "Generic (PLEG): container finished" podID="8dcaef0e-42e3-424b-bbe5-4a023efe912d" containerID="f3b6d20cf459b1f6d3aff5b9290f9391909baf27ee40c49e2b0cd2d5cfe889ad" exitCode=143 Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.562991 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-v7br2" event={"ID":"8dcaef0e-42e3-424b-bbe5-4a023efe912d","Type":"ContainerDied","Data":"a23b08d1e6780f4c2ce9ff447bd40050abb1e353c16ebccc8c3ecc7bc9a35e52"} Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.563020 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-v7br2" event={"ID":"8dcaef0e-42e3-424b-bbe5-4a023efe912d","Type":"ContainerDied","Data":"9fe3468cdd612fc4378e1f6f55a2a9d00274f7ddb65cb57b2a368dfc065cfb4e"} Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.563035 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-v7br2" event={"ID":"8dcaef0e-42e3-424b-bbe5-4a023efe912d","Type":"ContainerDied","Data":"fc9d4db944f837289d1923176d60cfc4a566a282c2d2e3a703e220df884703c7"} Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.563048 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-v7br2" event={"ID":"8dcaef0e-42e3-424b-bbe5-4a023efe912d","Type":"ContainerDied","Data":"d16ed1c65925549dcb69ab02c3013e25e3f123a83179f1af3a8e08bf91a59896"} Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.563059 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-v7br2" event={"ID":"8dcaef0e-42e3-424b-bbe5-4a023efe912d","Type":"ContainerDied","Data":"e72cff0723e2ce520340244b4e597d8016c58e1b071c4b087f2f6cd718741594"} Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.563071 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-v7br2" event={"ID":"8dcaef0e-42e3-424b-bbe5-4a023efe912d","Type":"ContainerDied","Data":"71dec55e859b11a081dbb90d7d6cf6b4846cf591382303d8329003ff50c71f2a"} Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.563083 4870 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a23b08d1e6780f4c2ce9ff447bd40050abb1e353c16ebccc8c3ecc7bc9a35e52"} Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.563095 4870 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ab4295721b6ec6d64040d321899ab478c612e91cbc3b99b33753869578a2d1e8"} Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.563102 4870 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9fe3468cdd612fc4378e1f6f55a2a9d00274f7ddb65cb57b2a368dfc065cfb4e"} Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.563110 4870 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"fc9d4db944f837289d1923176d60cfc4a566a282c2d2e3a703e220df884703c7"} Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.563117 4870 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d16ed1c65925549dcb69ab02c3013e25e3f123a83179f1af3a8e08bf91a59896"} Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.563124 4870 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e72cff0723e2ce520340244b4e597d8016c58e1b071c4b087f2f6cd718741594"} Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.563131 4870 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"71dec55e859b11a081dbb90d7d6cf6b4846cf591382303d8329003ff50c71f2a"} Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.563138 4870 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ed59243b59c39ce82bf9bc8f33c0a8b84854846afff6a89f5d37ffc97a096b7c"} Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.563145 4870 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f3b6d20cf459b1f6d3aff5b9290f9391909baf27ee40c49e2b0cd2d5cfe889ad"} Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.563153 4870 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1ba24c2e067b9f3a14d9d018b66300b54d866530d30174fbdb3c899299575d28"} Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.563162 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-v7br2" event={"ID":"8dcaef0e-42e3-424b-bbe5-4a023efe912d","Type":"ContainerDied","Data":"ed59243b59c39ce82bf9bc8f33c0a8b84854846afff6a89f5d37ffc97a096b7c"} Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.563172 4870 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a23b08d1e6780f4c2ce9ff447bd40050abb1e353c16ebccc8c3ecc7bc9a35e52"} Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.563180 4870 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ab4295721b6ec6d64040d321899ab478c612e91cbc3b99b33753869578a2d1e8"} Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.563186 4870 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9fe3468cdd612fc4378e1f6f55a2a9d00274f7ddb65cb57b2a368dfc065cfb4e"} Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.563193 4870 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"fc9d4db944f837289d1923176d60cfc4a566a282c2d2e3a703e220df884703c7"} Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.563200 4870 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d16ed1c65925549dcb69ab02c3013e25e3f123a83179f1af3a8e08bf91a59896"} Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.563207 4870 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e72cff0723e2ce520340244b4e597d8016c58e1b071c4b087f2f6cd718741594"} Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.563214 4870 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"71dec55e859b11a081dbb90d7d6cf6b4846cf591382303d8329003ff50c71f2a"} Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.563221 4870 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ed59243b59c39ce82bf9bc8f33c0a8b84854846afff6a89f5d37ffc97a096b7c"} Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.563228 4870 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f3b6d20cf459b1f6d3aff5b9290f9391909baf27ee40c49e2b0cd2d5cfe889ad"} Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.563235 4870 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1ba24c2e067b9f3a14d9d018b66300b54d866530d30174fbdb3c899299575d28"} Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.563244 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-v7br2" event={"ID":"8dcaef0e-42e3-424b-bbe5-4a023efe912d","Type":"ContainerDied","Data":"f3b6d20cf459b1f6d3aff5b9290f9391909baf27ee40c49e2b0cd2d5cfe889ad"} Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.563253 4870 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a23b08d1e6780f4c2ce9ff447bd40050abb1e353c16ebccc8c3ecc7bc9a35e52"} Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.563261 4870 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ab4295721b6ec6d64040d321899ab478c612e91cbc3b99b33753869578a2d1e8"} Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.563268 4870 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9fe3468cdd612fc4378e1f6f55a2a9d00274f7ddb65cb57b2a368dfc065cfb4e"} Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.563275 4870 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"fc9d4db944f837289d1923176d60cfc4a566a282c2d2e3a703e220df884703c7"} Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.563282 4870 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d16ed1c65925549dcb69ab02c3013e25e3f123a83179f1af3a8e08bf91a59896"} Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.563289 4870 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e72cff0723e2ce520340244b4e597d8016c58e1b071c4b087f2f6cd718741594"} Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.563296 4870 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"71dec55e859b11a081dbb90d7d6cf6b4846cf591382303d8329003ff50c71f2a"} Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.563303 4870 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ed59243b59c39ce82bf9bc8f33c0a8b84854846afff6a89f5d37ffc97a096b7c"} Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.563310 4870 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f3b6d20cf459b1f6d3aff5b9290f9391909baf27ee40c49e2b0cd2d5cfe889ad"} Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.563316 4870 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1ba24c2e067b9f3a14d9d018b66300b54d866530d30174fbdb3c899299575d28"} Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.563325 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-v7br2" event={"ID":"8dcaef0e-42e3-424b-bbe5-4a023efe912d","Type":"ContainerDied","Data":"919facd0f2f00aaf6e6831e4778b6f30f1b1428e0530e02fb7d55e87d66fb222"} Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.563334 4870 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a23b08d1e6780f4c2ce9ff447bd40050abb1e353c16ebccc8c3ecc7bc9a35e52"} Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.563342 4870 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ab4295721b6ec6d64040d321899ab478c612e91cbc3b99b33753869578a2d1e8"} Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.563348 4870 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9fe3468cdd612fc4378e1f6f55a2a9d00274f7ddb65cb57b2a368dfc065cfb4e"} Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.563355 4870 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"fc9d4db944f837289d1923176d60cfc4a566a282c2d2e3a703e220df884703c7"} Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.563362 4870 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d16ed1c65925549dcb69ab02c3013e25e3f123a83179f1af3a8e08bf91a59896"} Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.563369 4870 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e72cff0723e2ce520340244b4e597d8016c58e1b071c4b087f2f6cd718741594"} Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.563375 4870 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"71dec55e859b11a081dbb90d7d6cf6b4846cf591382303d8329003ff50c71f2a"} Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.563382 4870 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ed59243b59c39ce82bf9bc8f33c0a8b84854846afff6a89f5d37ffc97a096b7c"} Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.563389 4870 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f3b6d20cf459b1f6d3aff5b9290f9391909baf27ee40c49e2b0cd2d5cfe889ad"} Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.563395 4870 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1ba24c2e067b9f3a14d9d018b66300b54d866530d30174fbdb3c899299575d28"} Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.563508 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-v7br2" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.586338 4870 scope.go:117] "RemoveContainer" containerID="a23b08d1e6780f4c2ce9ff447bd40050abb1e353c16ebccc8c3ecc7bc9a35e52" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.599260 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-bxhvl"] Oct 14 07:13:18 crc kubenswrapper[4870]: E1014 07:13:18.599516 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8dcaef0e-42e3-424b-bbe5-4a023efe912d" containerName="nbdb" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.599535 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="8dcaef0e-42e3-424b-bbe5-4a023efe912d" containerName="nbdb" Oct 14 07:13:18 crc kubenswrapper[4870]: E1014 07:13:18.599554 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8dcaef0e-42e3-424b-bbe5-4a023efe912d" containerName="kube-rbac-proxy-node" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.599564 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="8dcaef0e-42e3-424b-bbe5-4a023efe912d" containerName="kube-rbac-proxy-node" Oct 14 07:13:18 crc kubenswrapper[4870]: E1014 07:13:18.599579 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3" containerName="registry" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.599589 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3" containerName="registry" Oct 14 07:13:18 crc kubenswrapper[4870]: E1014 07:13:18.599599 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8dcaef0e-42e3-424b-bbe5-4a023efe912d" containerName="ovnkube-controller" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.599608 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="8dcaef0e-42e3-424b-bbe5-4a023efe912d" containerName="ovnkube-controller" Oct 14 07:13:18 crc kubenswrapper[4870]: E1014 07:13:18.599618 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8dcaef0e-42e3-424b-bbe5-4a023efe912d" containerName="kubecfg-setup" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.599627 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="8dcaef0e-42e3-424b-bbe5-4a023efe912d" containerName="kubecfg-setup" Oct 14 07:13:18 crc kubenswrapper[4870]: E1014 07:13:18.599637 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8dcaef0e-42e3-424b-bbe5-4a023efe912d" containerName="ovnkube-controller" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.599646 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="8dcaef0e-42e3-424b-bbe5-4a023efe912d" containerName="ovnkube-controller" Oct 14 07:13:18 crc kubenswrapper[4870]: E1014 07:13:18.599662 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8dcaef0e-42e3-424b-bbe5-4a023efe912d" containerName="sbdb" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.599670 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="8dcaef0e-42e3-424b-bbe5-4a023efe912d" containerName="sbdb" Oct 14 07:13:18 crc kubenswrapper[4870]: E1014 07:13:18.599679 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8dcaef0e-42e3-424b-bbe5-4a023efe912d" containerName="ovn-acl-logging" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.599688 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="8dcaef0e-42e3-424b-bbe5-4a023efe912d" containerName="ovn-acl-logging" Oct 14 07:13:18 crc kubenswrapper[4870]: E1014 07:13:18.599700 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8dcaef0e-42e3-424b-bbe5-4a023efe912d" containerName="ovnkube-controller" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.599709 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="8dcaef0e-42e3-424b-bbe5-4a023efe912d" containerName="ovnkube-controller" Oct 14 07:13:18 crc kubenswrapper[4870]: E1014 07:13:18.599720 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8dcaef0e-42e3-424b-bbe5-4a023efe912d" containerName="kube-rbac-proxy-ovn-metrics" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.599728 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="8dcaef0e-42e3-424b-bbe5-4a023efe912d" containerName="kube-rbac-proxy-ovn-metrics" Oct 14 07:13:18 crc kubenswrapper[4870]: E1014 07:13:18.599739 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8dcaef0e-42e3-424b-bbe5-4a023efe912d" containerName="ovn-controller" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.599748 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="8dcaef0e-42e3-424b-bbe5-4a023efe912d" containerName="ovn-controller" Oct 14 07:13:18 crc kubenswrapper[4870]: E1014 07:13:18.599757 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8dcaef0e-42e3-424b-bbe5-4a023efe912d" containerName="ovnkube-controller" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.599765 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="8dcaef0e-42e3-424b-bbe5-4a023efe912d" containerName="ovnkube-controller" Oct 14 07:13:18 crc kubenswrapper[4870]: E1014 07:13:18.599776 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8dcaef0e-42e3-424b-bbe5-4a023efe912d" containerName="northd" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.599784 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="8dcaef0e-42e3-424b-bbe5-4a023efe912d" containerName="northd" Oct 14 07:13:18 crc kubenswrapper[4870]: E1014 07:13:18.599795 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8dcaef0e-42e3-424b-bbe5-4a023efe912d" containerName="ovnkube-controller" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.599803 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="8dcaef0e-42e3-424b-bbe5-4a023efe912d" containerName="ovnkube-controller" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.599912 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="8dcaef0e-42e3-424b-bbe5-4a023efe912d" containerName="ovnkube-controller" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.599923 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="8dcaef0e-42e3-424b-bbe5-4a023efe912d" containerName="ovn-controller" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.599936 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="8dcaef0e-42e3-424b-bbe5-4a023efe912d" containerName="ovnkube-controller" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.599946 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="2d2994b3-6bc8-4a3c-8bd6-3aa6d8f0f3d3" containerName="registry" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.599958 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="8dcaef0e-42e3-424b-bbe5-4a023efe912d" containerName="ovnkube-controller" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.599968 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="8dcaef0e-42e3-424b-bbe5-4a023efe912d" containerName="kube-rbac-proxy-ovn-metrics" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.599978 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="8dcaef0e-42e3-424b-bbe5-4a023efe912d" containerName="ovnkube-controller" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.599986 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="8dcaef0e-42e3-424b-bbe5-4a023efe912d" containerName="nbdb" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.599998 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="8dcaef0e-42e3-424b-bbe5-4a023efe912d" containerName="sbdb" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.600007 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="8dcaef0e-42e3-424b-bbe5-4a023efe912d" containerName="northd" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.600017 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="8dcaef0e-42e3-424b-bbe5-4a023efe912d" containerName="kube-rbac-proxy-node" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.600026 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="8dcaef0e-42e3-424b-bbe5-4a023efe912d" containerName="ovn-acl-logging" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.600249 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="8dcaef0e-42e3-424b-bbe5-4a023efe912d" containerName="ovnkube-controller" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.602164 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-bxhvl" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.625094 4870 scope.go:117] "RemoveContainer" containerID="ab4295721b6ec6d64040d321899ab478c612e91cbc3b99b33753869578a2d1e8" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.659552 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/8dcaef0e-42e3-424b-bbe5-4a023efe912d-host-var-lib-cni-networks-ovn-kubernetes\") pod \"8dcaef0e-42e3-424b-bbe5-4a023efe912d\" (UID: \"8dcaef0e-42e3-424b-bbe5-4a023efe912d\") " Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.659623 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/8dcaef0e-42e3-424b-bbe5-4a023efe912d-host-kubelet\") pod \"8dcaef0e-42e3-424b-bbe5-4a023efe912d\" (UID: \"8dcaef0e-42e3-424b-bbe5-4a023efe912d\") " Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.659644 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/8dcaef0e-42e3-424b-bbe5-4a023efe912d-host-run-netns\") pod \"8dcaef0e-42e3-424b-bbe5-4a023efe912d\" (UID: \"8dcaef0e-42e3-424b-bbe5-4a023efe912d\") " Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.659665 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/8dcaef0e-42e3-424b-bbe5-4a023efe912d-node-log\") pod \"8dcaef0e-42e3-424b-bbe5-4a023efe912d\" (UID: \"8dcaef0e-42e3-424b-bbe5-4a023efe912d\") " Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.659681 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/8dcaef0e-42e3-424b-bbe5-4a023efe912d-host-slash\") pod \"8dcaef0e-42e3-424b-bbe5-4a023efe912d\" (UID: \"8dcaef0e-42e3-424b-bbe5-4a023efe912d\") " Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.659701 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/8dcaef0e-42e3-424b-bbe5-4a023efe912d-ovnkube-script-lib\") pod \"8dcaef0e-42e3-424b-bbe5-4a023efe912d\" (UID: \"8dcaef0e-42e3-424b-bbe5-4a023efe912d\") " Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.659730 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/8dcaef0e-42e3-424b-bbe5-4a023efe912d-host-run-ovn-kubernetes\") pod \"8dcaef0e-42e3-424b-bbe5-4a023efe912d\" (UID: \"8dcaef0e-42e3-424b-bbe5-4a023efe912d\") " Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.659746 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/8dcaef0e-42e3-424b-bbe5-4a023efe912d-run-openvswitch\") pod \"8dcaef0e-42e3-424b-bbe5-4a023efe912d\" (UID: \"8dcaef0e-42e3-424b-bbe5-4a023efe912d\") " Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.659774 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/8dcaef0e-42e3-424b-bbe5-4a023efe912d-systemd-units\") pod \"8dcaef0e-42e3-424b-bbe5-4a023efe912d\" (UID: \"8dcaef0e-42e3-424b-bbe5-4a023efe912d\") " Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.659831 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/8dcaef0e-42e3-424b-bbe5-4a023efe912d-host-cni-bin\") pod \"8dcaef0e-42e3-424b-bbe5-4a023efe912d\" (UID: \"8dcaef0e-42e3-424b-bbe5-4a023efe912d\") " Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.659856 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-glrll\" (UniqueName: \"kubernetes.io/projected/8dcaef0e-42e3-424b-bbe5-4a023efe912d-kube-api-access-glrll\") pod \"8dcaef0e-42e3-424b-bbe5-4a023efe912d\" (UID: \"8dcaef0e-42e3-424b-bbe5-4a023efe912d\") " Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.659880 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/8dcaef0e-42e3-424b-bbe5-4a023efe912d-etc-openvswitch\") pod \"8dcaef0e-42e3-424b-bbe5-4a023efe912d\" (UID: \"8dcaef0e-42e3-424b-bbe5-4a023efe912d\") " Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.659904 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/8dcaef0e-42e3-424b-bbe5-4a023efe912d-env-overrides\") pod \"8dcaef0e-42e3-424b-bbe5-4a023efe912d\" (UID: \"8dcaef0e-42e3-424b-bbe5-4a023efe912d\") " Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.659948 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/8dcaef0e-42e3-424b-bbe5-4a023efe912d-run-systemd\") pod \"8dcaef0e-42e3-424b-bbe5-4a023efe912d\" (UID: \"8dcaef0e-42e3-424b-bbe5-4a023efe912d\") " Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.659967 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/8dcaef0e-42e3-424b-bbe5-4a023efe912d-log-socket\") pod \"8dcaef0e-42e3-424b-bbe5-4a023efe912d\" (UID: \"8dcaef0e-42e3-424b-bbe5-4a023efe912d\") " Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.659990 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/8dcaef0e-42e3-424b-bbe5-4a023efe912d-host-cni-netd\") pod \"8dcaef0e-42e3-424b-bbe5-4a023efe912d\" (UID: \"8dcaef0e-42e3-424b-bbe5-4a023efe912d\") " Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.660014 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/8dcaef0e-42e3-424b-bbe5-4a023efe912d-var-lib-openvswitch\") pod \"8dcaef0e-42e3-424b-bbe5-4a023efe912d\" (UID: \"8dcaef0e-42e3-424b-bbe5-4a023efe912d\") " Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.660070 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/8dcaef0e-42e3-424b-bbe5-4a023efe912d-ovn-node-metrics-cert\") pod \"8dcaef0e-42e3-424b-bbe5-4a023efe912d\" (UID: \"8dcaef0e-42e3-424b-bbe5-4a023efe912d\") " Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.660094 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/8dcaef0e-42e3-424b-bbe5-4a023efe912d-ovnkube-config\") pod \"8dcaef0e-42e3-424b-bbe5-4a023efe912d\" (UID: \"8dcaef0e-42e3-424b-bbe5-4a023efe912d\") " Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.660117 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/8dcaef0e-42e3-424b-bbe5-4a023efe912d-run-ovn\") pod \"8dcaef0e-42e3-424b-bbe5-4a023efe912d\" (UID: \"8dcaef0e-42e3-424b-bbe5-4a023efe912d\") " Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.659762 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8dcaef0e-42e3-424b-bbe5-4a023efe912d-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "8dcaef0e-42e3-424b-bbe5-4a023efe912d" (UID: "8dcaef0e-42e3-424b-bbe5-4a023efe912d"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.659796 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8dcaef0e-42e3-424b-bbe5-4a023efe912d-host-slash" (OuterVolumeSpecName: "host-slash") pod "8dcaef0e-42e3-424b-bbe5-4a023efe912d" (UID: "8dcaef0e-42e3-424b-bbe5-4a023efe912d"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.659814 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8dcaef0e-42e3-424b-bbe5-4a023efe912d-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "8dcaef0e-42e3-424b-bbe5-4a023efe912d" (UID: "8dcaef0e-42e3-424b-bbe5-4a023efe912d"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.659835 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8dcaef0e-42e3-424b-bbe5-4a023efe912d-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "8dcaef0e-42e3-424b-bbe5-4a023efe912d" (UID: "8dcaef0e-42e3-424b-bbe5-4a023efe912d"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.659851 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8dcaef0e-42e3-424b-bbe5-4a023efe912d-node-log" (OuterVolumeSpecName: "node-log") pod "8dcaef0e-42e3-424b-bbe5-4a023efe912d" (UID: "8dcaef0e-42e3-424b-bbe5-4a023efe912d"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.659874 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8dcaef0e-42e3-424b-bbe5-4a023efe912d-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "8dcaef0e-42e3-424b-bbe5-4a023efe912d" (UID: "8dcaef0e-42e3-424b-bbe5-4a023efe912d"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.660387 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8dcaef0e-42e3-424b-bbe5-4a023efe912d-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "8dcaef0e-42e3-424b-bbe5-4a023efe912d" (UID: "8dcaef0e-42e3-424b-bbe5-4a023efe912d"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.660453 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8dcaef0e-42e3-424b-bbe5-4a023efe912d-log-socket" (OuterVolumeSpecName: "log-socket") pod "8dcaef0e-42e3-424b-bbe5-4a023efe912d" (UID: "8dcaef0e-42e3-424b-bbe5-4a023efe912d"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.660475 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8dcaef0e-42e3-424b-bbe5-4a023efe912d-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "8dcaef0e-42e3-424b-bbe5-4a023efe912d" (UID: "8dcaef0e-42e3-424b-bbe5-4a023efe912d"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.660589 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8dcaef0e-42e3-424b-bbe5-4a023efe912d-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "8dcaef0e-42e3-424b-bbe5-4a023efe912d" (UID: "8dcaef0e-42e3-424b-bbe5-4a023efe912d"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.660693 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8dcaef0e-42e3-424b-bbe5-4a023efe912d-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "8dcaef0e-42e3-424b-bbe5-4a023efe912d" (UID: "8dcaef0e-42e3-424b-bbe5-4a023efe912d"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.660968 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8dcaef0e-42e3-424b-bbe5-4a023efe912d-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "8dcaef0e-42e3-424b-bbe5-4a023efe912d" (UID: "8dcaef0e-42e3-424b-bbe5-4a023efe912d"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.661009 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8dcaef0e-42e3-424b-bbe5-4a023efe912d-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "8dcaef0e-42e3-424b-bbe5-4a023efe912d" (UID: "8dcaef0e-42e3-424b-bbe5-4a023efe912d"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.661157 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8dcaef0e-42e3-424b-bbe5-4a023efe912d-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "8dcaef0e-42e3-424b-bbe5-4a023efe912d" (UID: "8dcaef0e-42e3-424b-bbe5-4a023efe912d"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.661209 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8dcaef0e-42e3-424b-bbe5-4a023efe912d-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "8dcaef0e-42e3-424b-bbe5-4a023efe912d" (UID: "8dcaef0e-42e3-424b-bbe5-4a023efe912d"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.661287 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8dcaef0e-42e3-424b-bbe5-4a023efe912d-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "8dcaef0e-42e3-424b-bbe5-4a023efe912d" (UID: "8dcaef0e-42e3-424b-bbe5-4a023efe912d"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.661903 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8dcaef0e-42e3-424b-bbe5-4a023efe912d-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "8dcaef0e-42e3-424b-bbe5-4a023efe912d" (UID: "8dcaef0e-42e3-424b-bbe5-4a023efe912d"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.664488 4870 scope.go:117] "RemoveContainer" containerID="9fe3468cdd612fc4378e1f6f55a2a9d00274f7ddb65cb57b2a368dfc065cfb4e" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.669452 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8dcaef0e-42e3-424b-bbe5-4a023efe912d-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "8dcaef0e-42e3-424b-bbe5-4a023efe912d" (UID: "8dcaef0e-42e3-424b-bbe5-4a023efe912d"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.670861 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8dcaef0e-42e3-424b-bbe5-4a023efe912d-kube-api-access-glrll" (OuterVolumeSpecName: "kube-api-access-glrll") pod "8dcaef0e-42e3-424b-bbe5-4a023efe912d" (UID: "8dcaef0e-42e3-424b-bbe5-4a023efe912d"). InnerVolumeSpecName "kube-api-access-glrll". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.679177 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8dcaef0e-42e3-424b-bbe5-4a023efe912d-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "8dcaef0e-42e3-424b-bbe5-4a023efe912d" (UID: "8dcaef0e-42e3-424b-bbe5-4a023efe912d"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.687941 4870 scope.go:117] "RemoveContainer" containerID="fc9d4db944f837289d1923176d60cfc4a566a282c2d2e3a703e220df884703c7" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.703407 4870 scope.go:117] "RemoveContainer" containerID="d16ed1c65925549dcb69ab02c3013e25e3f123a83179f1af3a8e08bf91a59896" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.726993 4870 scope.go:117] "RemoveContainer" containerID="e72cff0723e2ce520340244b4e597d8016c58e1b071c4b087f2f6cd718741594" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.743575 4870 scope.go:117] "RemoveContainer" containerID="71dec55e859b11a081dbb90d7d6cf6b4846cf591382303d8329003ff50c71f2a" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.761326 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/03297609-38cc-4df8-8150-86105829e6c4-run-ovn\") pod \"ovnkube-node-bxhvl\" (UID: \"03297609-38cc-4df8-8150-86105829e6c4\") " pod="openshift-ovn-kubernetes/ovnkube-node-bxhvl" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.761350 4870 scope.go:117] "RemoveContainer" containerID="ed59243b59c39ce82bf9bc8f33c0a8b84854846afff6a89f5d37ffc97a096b7c" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.761372 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/03297609-38cc-4df8-8150-86105829e6c4-run-systemd\") pod \"ovnkube-node-bxhvl\" (UID: \"03297609-38cc-4df8-8150-86105829e6c4\") " pod="openshift-ovn-kubernetes/ovnkube-node-bxhvl" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.761391 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/03297609-38cc-4df8-8150-86105829e6c4-host-kubelet\") pod \"ovnkube-node-bxhvl\" (UID: \"03297609-38cc-4df8-8150-86105829e6c4\") " pod="openshift-ovn-kubernetes/ovnkube-node-bxhvl" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.761411 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/03297609-38cc-4df8-8150-86105829e6c4-host-slash\") pod \"ovnkube-node-bxhvl\" (UID: \"03297609-38cc-4df8-8150-86105829e6c4\") " pod="openshift-ovn-kubernetes/ovnkube-node-bxhvl" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.761425 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/03297609-38cc-4df8-8150-86105829e6c4-ovnkube-config\") pod \"ovnkube-node-bxhvl\" (UID: \"03297609-38cc-4df8-8150-86105829e6c4\") " pod="openshift-ovn-kubernetes/ovnkube-node-bxhvl" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.761473 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4pktf\" (UniqueName: \"kubernetes.io/projected/03297609-38cc-4df8-8150-86105829e6c4-kube-api-access-4pktf\") pod \"ovnkube-node-bxhvl\" (UID: \"03297609-38cc-4df8-8150-86105829e6c4\") " pod="openshift-ovn-kubernetes/ovnkube-node-bxhvl" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.761535 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/03297609-38cc-4df8-8150-86105829e6c4-log-socket\") pod \"ovnkube-node-bxhvl\" (UID: \"03297609-38cc-4df8-8150-86105829e6c4\") " pod="openshift-ovn-kubernetes/ovnkube-node-bxhvl" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.761556 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/03297609-38cc-4df8-8150-86105829e6c4-host-cni-bin\") pod \"ovnkube-node-bxhvl\" (UID: \"03297609-38cc-4df8-8150-86105829e6c4\") " pod="openshift-ovn-kubernetes/ovnkube-node-bxhvl" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.761577 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/03297609-38cc-4df8-8150-86105829e6c4-ovn-node-metrics-cert\") pod \"ovnkube-node-bxhvl\" (UID: \"03297609-38cc-4df8-8150-86105829e6c4\") " pod="openshift-ovn-kubernetes/ovnkube-node-bxhvl" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.761603 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/03297609-38cc-4df8-8150-86105829e6c4-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-bxhvl\" (UID: \"03297609-38cc-4df8-8150-86105829e6c4\") " pod="openshift-ovn-kubernetes/ovnkube-node-bxhvl" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.761619 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/03297609-38cc-4df8-8150-86105829e6c4-systemd-units\") pod \"ovnkube-node-bxhvl\" (UID: \"03297609-38cc-4df8-8150-86105829e6c4\") " pod="openshift-ovn-kubernetes/ovnkube-node-bxhvl" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.761634 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/03297609-38cc-4df8-8150-86105829e6c4-etc-openvswitch\") pod \"ovnkube-node-bxhvl\" (UID: \"03297609-38cc-4df8-8150-86105829e6c4\") " pod="openshift-ovn-kubernetes/ovnkube-node-bxhvl" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.761649 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/03297609-38cc-4df8-8150-86105829e6c4-host-run-ovn-kubernetes\") pod \"ovnkube-node-bxhvl\" (UID: \"03297609-38cc-4df8-8150-86105829e6c4\") " pod="openshift-ovn-kubernetes/ovnkube-node-bxhvl" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.761738 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/03297609-38cc-4df8-8150-86105829e6c4-env-overrides\") pod \"ovnkube-node-bxhvl\" (UID: \"03297609-38cc-4df8-8150-86105829e6c4\") " pod="openshift-ovn-kubernetes/ovnkube-node-bxhvl" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.761758 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/03297609-38cc-4df8-8150-86105829e6c4-run-openvswitch\") pod \"ovnkube-node-bxhvl\" (UID: \"03297609-38cc-4df8-8150-86105829e6c4\") " pod="openshift-ovn-kubernetes/ovnkube-node-bxhvl" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.761776 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/03297609-38cc-4df8-8150-86105829e6c4-node-log\") pod \"ovnkube-node-bxhvl\" (UID: \"03297609-38cc-4df8-8150-86105829e6c4\") " pod="openshift-ovn-kubernetes/ovnkube-node-bxhvl" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.761793 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/03297609-38cc-4df8-8150-86105829e6c4-host-cni-netd\") pod \"ovnkube-node-bxhvl\" (UID: \"03297609-38cc-4df8-8150-86105829e6c4\") " pod="openshift-ovn-kubernetes/ovnkube-node-bxhvl" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.761859 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/03297609-38cc-4df8-8150-86105829e6c4-var-lib-openvswitch\") pod \"ovnkube-node-bxhvl\" (UID: \"03297609-38cc-4df8-8150-86105829e6c4\") " pod="openshift-ovn-kubernetes/ovnkube-node-bxhvl" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.761921 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/03297609-38cc-4df8-8150-86105829e6c4-host-run-netns\") pod \"ovnkube-node-bxhvl\" (UID: \"03297609-38cc-4df8-8150-86105829e6c4\") " pod="openshift-ovn-kubernetes/ovnkube-node-bxhvl" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.761940 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/03297609-38cc-4df8-8150-86105829e6c4-ovnkube-script-lib\") pod \"ovnkube-node-bxhvl\" (UID: \"03297609-38cc-4df8-8150-86105829e6c4\") " pod="openshift-ovn-kubernetes/ovnkube-node-bxhvl" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.761982 4870 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/8dcaef0e-42e3-424b-bbe5-4a023efe912d-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.761992 4870 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/8dcaef0e-42e3-424b-bbe5-4a023efe912d-run-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.762001 4870 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/8dcaef0e-42e3-424b-bbe5-4a023efe912d-systemd-units\") on node \"crc\" DevicePath \"\"" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.762010 4870 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/8dcaef0e-42e3-424b-bbe5-4a023efe912d-host-cni-bin\") on node \"crc\" DevicePath \"\"" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.762018 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-glrll\" (UniqueName: \"kubernetes.io/projected/8dcaef0e-42e3-424b-bbe5-4a023efe912d-kube-api-access-glrll\") on node \"crc\" DevicePath \"\"" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.762026 4870 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/8dcaef0e-42e3-424b-bbe5-4a023efe912d-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.762034 4870 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/8dcaef0e-42e3-424b-bbe5-4a023efe912d-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.762042 4870 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/8dcaef0e-42e3-424b-bbe5-4a023efe912d-log-socket\") on node \"crc\" DevicePath \"\"" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.762050 4870 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/8dcaef0e-42e3-424b-bbe5-4a023efe912d-run-systemd\") on node \"crc\" DevicePath \"\"" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.762058 4870 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/8dcaef0e-42e3-424b-bbe5-4a023efe912d-host-cni-netd\") on node \"crc\" DevicePath \"\"" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.762066 4870 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/8dcaef0e-42e3-424b-bbe5-4a023efe912d-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.762074 4870 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/8dcaef0e-42e3-424b-bbe5-4a023efe912d-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.762082 4870 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/8dcaef0e-42e3-424b-bbe5-4a023efe912d-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.762091 4870 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/8dcaef0e-42e3-424b-bbe5-4a023efe912d-run-ovn\") on node \"crc\" DevicePath \"\"" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.762101 4870 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/8dcaef0e-42e3-424b-bbe5-4a023efe912d-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.762110 4870 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/8dcaef0e-42e3-424b-bbe5-4a023efe912d-host-kubelet\") on node \"crc\" DevicePath \"\"" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.762118 4870 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/8dcaef0e-42e3-424b-bbe5-4a023efe912d-host-run-netns\") on node \"crc\" DevicePath \"\"" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.762126 4870 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/8dcaef0e-42e3-424b-bbe5-4a023efe912d-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.762133 4870 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/8dcaef0e-42e3-424b-bbe5-4a023efe912d-node-log\") on node \"crc\" DevicePath \"\"" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.762140 4870 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/8dcaef0e-42e3-424b-bbe5-4a023efe912d-host-slash\") on node \"crc\" DevicePath \"\"" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.780263 4870 scope.go:117] "RemoveContainer" containerID="f3b6d20cf459b1f6d3aff5b9290f9391909baf27ee40c49e2b0cd2d5cfe889ad" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.802723 4870 scope.go:117] "RemoveContainer" containerID="1ba24c2e067b9f3a14d9d018b66300b54d866530d30174fbdb3c899299575d28" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.822945 4870 scope.go:117] "RemoveContainer" containerID="a23b08d1e6780f4c2ce9ff447bd40050abb1e353c16ebccc8c3ecc7bc9a35e52" Oct 14 07:13:18 crc kubenswrapper[4870]: E1014 07:13:18.823585 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a23b08d1e6780f4c2ce9ff447bd40050abb1e353c16ebccc8c3ecc7bc9a35e52\": container with ID starting with a23b08d1e6780f4c2ce9ff447bd40050abb1e353c16ebccc8c3ecc7bc9a35e52 not found: ID does not exist" containerID="a23b08d1e6780f4c2ce9ff447bd40050abb1e353c16ebccc8c3ecc7bc9a35e52" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.823648 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a23b08d1e6780f4c2ce9ff447bd40050abb1e353c16ebccc8c3ecc7bc9a35e52"} err="failed to get container status \"a23b08d1e6780f4c2ce9ff447bd40050abb1e353c16ebccc8c3ecc7bc9a35e52\": rpc error: code = NotFound desc = could not find container \"a23b08d1e6780f4c2ce9ff447bd40050abb1e353c16ebccc8c3ecc7bc9a35e52\": container with ID starting with a23b08d1e6780f4c2ce9ff447bd40050abb1e353c16ebccc8c3ecc7bc9a35e52 not found: ID does not exist" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.823687 4870 scope.go:117] "RemoveContainer" containerID="ab4295721b6ec6d64040d321899ab478c612e91cbc3b99b33753869578a2d1e8" Oct 14 07:13:18 crc kubenswrapper[4870]: E1014 07:13:18.824323 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ab4295721b6ec6d64040d321899ab478c612e91cbc3b99b33753869578a2d1e8\": container with ID starting with ab4295721b6ec6d64040d321899ab478c612e91cbc3b99b33753869578a2d1e8 not found: ID does not exist" containerID="ab4295721b6ec6d64040d321899ab478c612e91cbc3b99b33753869578a2d1e8" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.824367 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ab4295721b6ec6d64040d321899ab478c612e91cbc3b99b33753869578a2d1e8"} err="failed to get container status \"ab4295721b6ec6d64040d321899ab478c612e91cbc3b99b33753869578a2d1e8\": rpc error: code = NotFound desc = could not find container \"ab4295721b6ec6d64040d321899ab478c612e91cbc3b99b33753869578a2d1e8\": container with ID starting with ab4295721b6ec6d64040d321899ab478c612e91cbc3b99b33753869578a2d1e8 not found: ID does not exist" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.824399 4870 scope.go:117] "RemoveContainer" containerID="9fe3468cdd612fc4378e1f6f55a2a9d00274f7ddb65cb57b2a368dfc065cfb4e" Oct 14 07:13:18 crc kubenswrapper[4870]: E1014 07:13:18.824895 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9fe3468cdd612fc4378e1f6f55a2a9d00274f7ddb65cb57b2a368dfc065cfb4e\": container with ID starting with 9fe3468cdd612fc4378e1f6f55a2a9d00274f7ddb65cb57b2a368dfc065cfb4e not found: ID does not exist" containerID="9fe3468cdd612fc4378e1f6f55a2a9d00274f7ddb65cb57b2a368dfc065cfb4e" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.824942 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9fe3468cdd612fc4378e1f6f55a2a9d00274f7ddb65cb57b2a368dfc065cfb4e"} err="failed to get container status \"9fe3468cdd612fc4378e1f6f55a2a9d00274f7ddb65cb57b2a368dfc065cfb4e\": rpc error: code = NotFound desc = could not find container \"9fe3468cdd612fc4378e1f6f55a2a9d00274f7ddb65cb57b2a368dfc065cfb4e\": container with ID starting with 9fe3468cdd612fc4378e1f6f55a2a9d00274f7ddb65cb57b2a368dfc065cfb4e not found: ID does not exist" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.824969 4870 scope.go:117] "RemoveContainer" containerID="fc9d4db944f837289d1923176d60cfc4a566a282c2d2e3a703e220df884703c7" Oct 14 07:13:18 crc kubenswrapper[4870]: E1014 07:13:18.825337 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fc9d4db944f837289d1923176d60cfc4a566a282c2d2e3a703e220df884703c7\": container with ID starting with fc9d4db944f837289d1923176d60cfc4a566a282c2d2e3a703e220df884703c7 not found: ID does not exist" containerID="fc9d4db944f837289d1923176d60cfc4a566a282c2d2e3a703e220df884703c7" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.825390 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fc9d4db944f837289d1923176d60cfc4a566a282c2d2e3a703e220df884703c7"} err="failed to get container status \"fc9d4db944f837289d1923176d60cfc4a566a282c2d2e3a703e220df884703c7\": rpc error: code = NotFound desc = could not find container \"fc9d4db944f837289d1923176d60cfc4a566a282c2d2e3a703e220df884703c7\": container with ID starting with fc9d4db944f837289d1923176d60cfc4a566a282c2d2e3a703e220df884703c7 not found: ID does not exist" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.825406 4870 scope.go:117] "RemoveContainer" containerID="d16ed1c65925549dcb69ab02c3013e25e3f123a83179f1af3a8e08bf91a59896" Oct 14 07:13:18 crc kubenswrapper[4870]: E1014 07:13:18.825769 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d16ed1c65925549dcb69ab02c3013e25e3f123a83179f1af3a8e08bf91a59896\": container with ID starting with d16ed1c65925549dcb69ab02c3013e25e3f123a83179f1af3a8e08bf91a59896 not found: ID does not exist" containerID="d16ed1c65925549dcb69ab02c3013e25e3f123a83179f1af3a8e08bf91a59896" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.825808 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d16ed1c65925549dcb69ab02c3013e25e3f123a83179f1af3a8e08bf91a59896"} err="failed to get container status \"d16ed1c65925549dcb69ab02c3013e25e3f123a83179f1af3a8e08bf91a59896\": rpc error: code = NotFound desc = could not find container \"d16ed1c65925549dcb69ab02c3013e25e3f123a83179f1af3a8e08bf91a59896\": container with ID starting with d16ed1c65925549dcb69ab02c3013e25e3f123a83179f1af3a8e08bf91a59896 not found: ID does not exist" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.825821 4870 scope.go:117] "RemoveContainer" containerID="e72cff0723e2ce520340244b4e597d8016c58e1b071c4b087f2f6cd718741594" Oct 14 07:13:18 crc kubenswrapper[4870]: E1014 07:13:18.826203 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e72cff0723e2ce520340244b4e597d8016c58e1b071c4b087f2f6cd718741594\": container with ID starting with e72cff0723e2ce520340244b4e597d8016c58e1b071c4b087f2f6cd718741594 not found: ID does not exist" containerID="e72cff0723e2ce520340244b4e597d8016c58e1b071c4b087f2f6cd718741594" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.826242 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e72cff0723e2ce520340244b4e597d8016c58e1b071c4b087f2f6cd718741594"} err="failed to get container status \"e72cff0723e2ce520340244b4e597d8016c58e1b071c4b087f2f6cd718741594\": rpc error: code = NotFound desc = could not find container \"e72cff0723e2ce520340244b4e597d8016c58e1b071c4b087f2f6cd718741594\": container with ID starting with e72cff0723e2ce520340244b4e597d8016c58e1b071c4b087f2f6cd718741594 not found: ID does not exist" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.826254 4870 scope.go:117] "RemoveContainer" containerID="71dec55e859b11a081dbb90d7d6cf6b4846cf591382303d8329003ff50c71f2a" Oct 14 07:13:18 crc kubenswrapper[4870]: E1014 07:13:18.826757 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"71dec55e859b11a081dbb90d7d6cf6b4846cf591382303d8329003ff50c71f2a\": container with ID starting with 71dec55e859b11a081dbb90d7d6cf6b4846cf591382303d8329003ff50c71f2a not found: ID does not exist" containerID="71dec55e859b11a081dbb90d7d6cf6b4846cf591382303d8329003ff50c71f2a" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.826778 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"71dec55e859b11a081dbb90d7d6cf6b4846cf591382303d8329003ff50c71f2a"} err="failed to get container status \"71dec55e859b11a081dbb90d7d6cf6b4846cf591382303d8329003ff50c71f2a\": rpc error: code = NotFound desc = could not find container \"71dec55e859b11a081dbb90d7d6cf6b4846cf591382303d8329003ff50c71f2a\": container with ID starting with 71dec55e859b11a081dbb90d7d6cf6b4846cf591382303d8329003ff50c71f2a not found: ID does not exist" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.826790 4870 scope.go:117] "RemoveContainer" containerID="ed59243b59c39ce82bf9bc8f33c0a8b84854846afff6a89f5d37ffc97a096b7c" Oct 14 07:13:18 crc kubenswrapper[4870]: E1014 07:13:18.827042 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ed59243b59c39ce82bf9bc8f33c0a8b84854846afff6a89f5d37ffc97a096b7c\": container with ID starting with ed59243b59c39ce82bf9bc8f33c0a8b84854846afff6a89f5d37ffc97a096b7c not found: ID does not exist" containerID="ed59243b59c39ce82bf9bc8f33c0a8b84854846afff6a89f5d37ffc97a096b7c" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.827063 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ed59243b59c39ce82bf9bc8f33c0a8b84854846afff6a89f5d37ffc97a096b7c"} err="failed to get container status \"ed59243b59c39ce82bf9bc8f33c0a8b84854846afff6a89f5d37ffc97a096b7c\": rpc error: code = NotFound desc = could not find container \"ed59243b59c39ce82bf9bc8f33c0a8b84854846afff6a89f5d37ffc97a096b7c\": container with ID starting with ed59243b59c39ce82bf9bc8f33c0a8b84854846afff6a89f5d37ffc97a096b7c not found: ID does not exist" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.827076 4870 scope.go:117] "RemoveContainer" containerID="f3b6d20cf459b1f6d3aff5b9290f9391909baf27ee40c49e2b0cd2d5cfe889ad" Oct 14 07:13:18 crc kubenswrapper[4870]: E1014 07:13:18.827339 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f3b6d20cf459b1f6d3aff5b9290f9391909baf27ee40c49e2b0cd2d5cfe889ad\": container with ID starting with f3b6d20cf459b1f6d3aff5b9290f9391909baf27ee40c49e2b0cd2d5cfe889ad not found: ID does not exist" containerID="f3b6d20cf459b1f6d3aff5b9290f9391909baf27ee40c49e2b0cd2d5cfe889ad" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.827390 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f3b6d20cf459b1f6d3aff5b9290f9391909baf27ee40c49e2b0cd2d5cfe889ad"} err="failed to get container status \"f3b6d20cf459b1f6d3aff5b9290f9391909baf27ee40c49e2b0cd2d5cfe889ad\": rpc error: code = NotFound desc = could not find container \"f3b6d20cf459b1f6d3aff5b9290f9391909baf27ee40c49e2b0cd2d5cfe889ad\": container with ID starting with f3b6d20cf459b1f6d3aff5b9290f9391909baf27ee40c49e2b0cd2d5cfe889ad not found: ID does not exist" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.827405 4870 scope.go:117] "RemoveContainer" containerID="1ba24c2e067b9f3a14d9d018b66300b54d866530d30174fbdb3c899299575d28" Oct 14 07:13:18 crc kubenswrapper[4870]: E1014 07:13:18.827720 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1ba24c2e067b9f3a14d9d018b66300b54d866530d30174fbdb3c899299575d28\": container with ID starting with 1ba24c2e067b9f3a14d9d018b66300b54d866530d30174fbdb3c899299575d28 not found: ID does not exist" containerID="1ba24c2e067b9f3a14d9d018b66300b54d866530d30174fbdb3c899299575d28" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.827764 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1ba24c2e067b9f3a14d9d018b66300b54d866530d30174fbdb3c899299575d28"} err="failed to get container status \"1ba24c2e067b9f3a14d9d018b66300b54d866530d30174fbdb3c899299575d28\": rpc error: code = NotFound desc = could not find container \"1ba24c2e067b9f3a14d9d018b66300b54d866530d30174fbdb3c899299575d28\": container with ID starting with 1ba24c2e067b9f3a14d9d018b66300b54d866530d30174fbdb3c899299575d28 not found: ID does not exist" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.827797 4870 scope.go:117] "RemoveContainer" containerID="a23b08d1e6780f4c2ce9ff447bd40050abb1e353c16ebccc8c3ecc7bc9a35e52" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.828183 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a23b08d1e6780f4c2ce9ff447bd40050abb1e353c16ebccc8c3ecc7bc9a35e52"} err="failed to get container status \"a23b08d1e6780f4c2ce9ff447bd40050abb1e353c16ebccc8c3ecc7bc9a35e52\": rpc error: code = NotFound desc = could not find container \"a23b08d1e6780f4c2ce9ff447bd40050abb1e353c16ebccc8c3ecc7bc9a35e52\": container with ID starting with a23b08d1e6780f4c2ce9ff447bd40050abb1e353c16ebccc8c3ecc7bc9a35e52 not found: ID does not exist" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.828222 4870 scope.go:117] "RemoveContainer" containerID="ab4295721b6ec6d64040d321899ab478c612e91cbc3b99b33753869578a2d1e8" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.828619 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ab4295721b6ec6d64040d321899ab478c612e91cbc3b99b33753869578a2d1e8"} err="failed to get container status \"ab4295721b6ec6d64040d321899ab478c612e91cbc3b99b33753869578a2d1e8\": rpc error: code = NotFound desc = could not find container \"ab4295721b6ec6d64040d321899ab478c612e91cbc3b99b33753869578a2d1e8\": container with ID starting with ab4295721b6ec6d64040d321899ab478c612e91cbc3b99b33753869578a2d1e8 not found: ID does not exist" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.828677 4870 scope.go:117] "RemoveContainer" containerID="9fe3468cdd612fc4378e1f6f55a2a9d00274f7ddb65cb57b2a368dfc065cfb4e" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.828925 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9fe3468cdd612fc4378e1f6f55a2a9d00274f7ddb65cb57b2a368dfc065cfb4e"} err="failed to get container status \"9fe3468cdd612fc4378e1f6f55a2a9d00274f7ddb65cb57b2a368dfc065cfb4e\": rpc error: code = NotFound desc = could not find container \"9fe3468cdd612fc4378e1f6f55a2a9d00274f7ddb65cb57b2a368dfc065cfb4e\": container with ID starting with 9fe3468cdd612fc4378e1f6f55a2a9d00274f7ddb65cb57b2a368dfc065cfb4e not found: ID does not exist" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.828947 4870 scope.go:117] "RemoveContainer" containerID="fc9d4db944f837289d1923176d60cfc4a566a282c2d2e3a703e220df884703c7" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.829275 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fc9d4db944f837289d1923176d60cfc4a566a282c2d2e3a703e220df884703c7"} err="failed to get container status \"fc9d4db944f837289d1923176d60cfc4a566a282c2d2e3a703e220df884703c7\": rpc error: code = NotFound desc = could not find container \"fc9d4db944f837289d1923176d60cfc4a566a282c2d2e3a703e220df884703c7\": container with ID starting with fc9d4db944f837289d1923176d60cfc4a566a282c2d2e3a703e220df884703c7 not found: ID does not exist" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.829314 4870 scope.go:117] "RemoveContainer" containerID="d16ed1c65925549dcb69ab02c3013e25e3f123a83179f1af3a8e08bf91a59896" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.829828 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d16ed1c65925549dcb69ab02c3013e25e3f123a83179f1af3a8e08bf91a59896"} err="failed to get container status \"d16ed1c65925549dcb69ab02c3013e25e3f123a83179f1af3a8e08bf91a59896\": rpc error: code = NotFound desc = could not find container \"d16ed1c65925549dcb69ab02c3013e25e3f123a83179f1af3a8e08bf91a59896\": container with ID starting with d16ed1c65925549dcb69ab02c3013e25e3f123a83179f1af3a8e08bf91a59896 not found: ID does not exist" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.829846 4870 scope.go:117] "RemoveContainer" containerID="e72cff0723e2ce520340244b4e597d8016c58e1b071c4b087f2f6cd718741594" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.830158 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e72cff0723e2ce520340244b4e597d8016c58e1b071c4b087f2f6cd718741594"} err="failed to get container status \"e72cff0723e2ce520340244b4e597d8016c58e1b071c4b087f2f6cd718741594\": rpc error: code = NotFound desc = could not find container \"e72cff0723e2ce520340244b4e597d8016c58e1b071c4b087f2f6cd718741594\": container with ID starting with e72cff0723e2ce520340244b4e597d8016c58e1b071c4b087f2f6cd718741594 not found: ID does not exist" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.830179 4870 scope.go:117] "RemoveContainer" containerID="71dec55e859b11a081dbb90d7d6cf6b4846cf591382303d8329003ff50c71f2a" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.830473 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"71dec55e859b11a081dbb90d7d6cf6b4846cf591382303d8329003ff50c71f2a"} err="failed to get container status \"71dec55e859b11a081dbb90d7d6cf6b4846cf591382303d8329003ff50c71f2a\": rpc error: code = NotFound desc = could not find container \"71dec55e859b11a081dbb90d7d6cf6b4846cf591382303d8329003ff50c71f2a\": container with ID starting with 71dec55e859b11a081dbb90d7d6cf6b4846cf591382303d8329003ff50c71f2a not found: ID does not exist" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.830492 4870 scope.go:117] "RemoveContainer" containerID="ed59243b59c39ce82bf9bc8f33c0a8b84854846afff6a89f5d37ffc97a096b7c" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.830786 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ed59243b59c39ce82bf9bc8f33c0a8b84854846afff6a89f5d37ffc97a096b7c"} err="failed to get container status \"ed59243b59c39ce82bf9bc8f33c0a8b84854846afff6a89f5d37ffc97a096b7c\": rpc error: code = NotFound desc = could not find container \"ed59243b59c39ce82bf9bc8f33c0a8b84854846afff6a89f5d37ffc97a096b7c\": container with ID starting with ed59243b59c39ce82bf9bc8f33c0a8b84854846afff6a89f5d37ffc97a096b7c not found: ID does not exist" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.830806 4870 scope.go:117] "RemoveContainer" containerID="f3b6d20cf459b1f6d3aff5b9290f9391909baf27ee40c49e2b0cd2d5cfe889ad" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.831247 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f3b6d20cf459b1f6d3aff5b9290f9391909baf27ee40c49e2b0cd2d5cfe889ad"} err="failed to get container status \"f3b6d20cf459b1f6d3aff5b9290f9391909baf27ee40c49e2b0cd2d5cfe889ad\": rpc error: code = NotFound desc = could not find container \"f3b6d20cf459b1f6d3aff5b9290f9391909baf27ee40c49e2b0cd2d5cfe889ad\": container with ID starting with f3b6d20cf459b1f6d3aff5b9290f9391909baf27ee40c49e2b0cd2d5cfe889ad not found: ID does not exist" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.831293 4870 scope.go:117] "RemoveContainer" containerID="1ba24c2e067b9f3a14d9d018b66300b54d866530d30174fbdb3c899299575d28" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.831551 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1ba24c2e067b9f3a14d9d018b66300b54d866530d30174fbdb3c899299575d28"} err="failed to get container status \"1ba24c2e067b9f3a14d9d018b66300b54d866530d30174fbdb3c899299575d28\": rpc error: code = NotFound desc = could not find container \"1ba24c2e067b9f3a14d9d018b66300b54d866530d30174fbdb3c899299575d28\": container with ID starting with 1ba24c2e067b9f3a14d9d018b66300b54d866530d30174fbdb3c899299575d28 not found: ID does not exist" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.831571 4870 scope.go:117] "RemoveContainer" containerID="a23b08d1e6780f4c2ce9ff447bd40050abb1e353c16ebccc8c3ecc7bc9a35e52" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.831855 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a23b08d1e6780f4c2ce9ff447bd40050abb1e353c16ebccc8c3ecc7bc9a35e52"} err="failed to get container status \"a23b08d1e6780f4c2ce9ff447bd40050abb1e353c16ebccc8c3ecc7bc9a35e52\": rpc error: code = NotFound desc = could not find container \"a23b08d1e6780f4c2ce9ff447bd40050abb1e353c16ebccc8c3ecc7bc9a35e52\": container with ID starting with a23b08d1e6780f4c2ce9ff447bd40050abb1e353c16ebccc8c3ecc7bc9a35e52 not found: ID does not exist" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.831871 4870 scope.go:117] "RemoveContainer" containerID="ab4295721b6ec6d64040d321899ab478c612e91cbc3b99b33753869578a2d1e8" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.832117 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ab4295721b6ec6d64040d321899ab478c612e91cbc3b99b33753869578a2d1e8"} err="failed to get container status \"ab4295721b6ec6d64040d321899ab478c612e91cbc3b99b33753869578a2d1e8\": rpc error: code = NotFound desc = could not find container \"ab4295721b6ec6d64040d321899ab478c612e91cbc3b99b33753869578a2d1e8\": container with ID starting with ab4295721b6ec6d64040d321899ab478c612e91cbc3b99b33753869578a2d1e8 not found: ID does not exist" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.832133 4870 scope.go:117] "RemoveContainer" containerID="9fe3468cdd612fc4378e1f6f55a2a9d00274f7ddb65cb57b2a368dfc065cfb4e" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.832431 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9fe3468cdd612fc4378e1f6f55a2a9d00274f7ddb65cb57b2a368dfc065cfb4e"} err="failed to get container status \"9fe3468cdd612fc4378e1f6f55a2a9d00274f7ddb65cb57b2a368dfc065cfb4e\": rpc error: code = NotFound desc = could not find container \"9fe3468cdd612fc4378e1f6f55a2a9d00274f7ddb65cb57b2a368dfc065cfb4e\": container with ID starting with 9fe3468cdd612fc4378e1f6f55a2a9d00274f7ddb65cb57b2a368dfc065cfb4e not found: ID does not exist" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.832462 4870 scope.go:117] "RemoveContainer" containerID="fc9d4db944f837289d1923176d60cfc4a566a282c2d2e3a703e220df884703c7" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.832782 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fc9d4db944f837289d1923176d60cfc4a566a282c2d2e3a703e220df884703c7"} err="failed to get container status \"fc9d4db944f837289d1923176d60cfc4a566a282c2d2e3a703e220df884703c7\": rpc error: code = NotFound desc = could not find container \"fc9d4db944f837289d1923176d60cfc4a566a282c2d2e3a703e220df884703c7\": container with ID starting with fc9d4db944f837289d1923176d60cfc4a566a282c2d2e3a703e220df884703c7 not found: ID does not exist" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.832802 4870 scope.go:117] "RemoveContainer" containerID="d16ed1c65925549dcb69ab02c3013e25e3f123a83179f1af3a8e08bf91a59896" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.833187 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d16ed1c65925549dcb69ab02c3013e25e3f123a83179f1af3a8e08bf91a59896"} err="failed to get container status \"d16ed1c65925549dcb69ab02c3013e25e3f123a83179f1af3a8e08bf91a59896\": rpc error: code = NotFound desc = could not find container \"d16ed1c65925549dcb69ab02c3013e25e3f123a83179f1af3a8e08bf91a59896\": container with ID starting with d16ed1c65925549dcb69ab02c3013e25e3f123a83179f1af3a8e08bf91a59896 not found: ID does not exist" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.833202 4870 scope.go:117] "RemoveContainer" containerID="e72cff0723e2ce520340244b4e597d8016c58e1b071c4b087f2f6cd718741594" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.833490 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e72cff0723e2ce520340244b4e597d8016c58e1b071c4b087f2f6cd718741594"} err="failed to get container status \"e72cff0723e2ce520340244b4e597d8016c58e1b071c4b087f2f6cd718741594\": rpc error: code = NotFound desc = could not find container \"e72cff0723e2ce520340244b4e597d8016c58e1b071c4b087f2f6cd718741594\": container with ID starting with e72cff0723e2ce520340244b4e597d8016c58e1b071c4b087f2f6cd718741594 not found: ID does not exist" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.833505 4870 scope.go:117] "RemoveContainer" containerID="71dec55e859b11a081dbb90d7d6cf6b4846cf591382303d8329003ff50c71f2a" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.833791 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"71dec55e859b11a081dbb90d7d6cf6b4846cf591382303d8329003ff50c71f2a"} err="failed to get container status \"71dec55e859b11a081dbb90d7d6cf6b4846cf591382303d8329003ff50c71f2a\": rpc error: code = NotFound desc = could not find container \"71dec55e859b11a081dbb90d7d6cf6b4846cf591382303d8329003ff50c71f2a\": container with ID starting with 71dec55e859b11a081dbb90d7d6cf6b4846cf591382303d8329003ff50c71f2a not found: ID does not exist" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.833803 4870 scope.go:117] "RemoveContainer" containerID="ed59243b59c39ce82bf9bc8f33c0a8b84854846afff6a89f5d37ffc97a096b7c" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.834099 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ed59243b59c39ce82bf9bc8f33c0a8b84854846afff6a89f5d37ffc97a096b7c"} err="failed to get container status \"ed59243b59c39ce82bf9bc8f33c0a8b84854846afff6a89f5d37ffc97a096b7c\": rpc error: code = NotFound desc = could not find container \"ed59243b59c39ce82bf9bc8f33c0a8b84854846afff6a89f5d37ffc97a096b7c\": container with ID starting with ed59243b59c39ce82bf9bc8f33c0a8b84854846afff6a89f5d37ffc97a096b7c not found: ID does not exist" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.834113 4870 scope.go:117] "RemoveContainer" containerID="f3b6d20cf459b1f6d3aff5b9290f9391909baf27ee40c49e2b0cd2d5cfe889ad" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.834460 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f3b6d20cf459b1f6d3aff5b9290f9391909baf27ee40c49e2b0cd2d5cfe889ad"} err="failed to get container status \"f3b6d20cf459b1f6d3aff5b9290f9391909baf27ee40c49e2b0cd2d5cfe889ad\": rpc error: code = NotFound desc = could not find container \"f3b6d20cf459b1f6d3aff5b9290f9391909baf27ee40c49e2b0cd2d5cfe889ad\": container with ID starting with f3b6d20cf459b1f6d3aff5b9290f9391909baf27ee40c49e2b0cd2d5cfe889ad not found: ID does not exist" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.834477 4870 scope.go:117] "RemoveContainer" containerID="1ba24c2e067b9f3a14d9d018b66300b54d866530d30174fbdb3c899299575d28" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.834741 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1ba24c2e067b9f3a14d9d018b66300b54d866530d30174fbdb3c899299575d28"} err="failed to get container status \"1ba24c2e067b9f3a14d9d018b66300b54d866530d30174fbdb3c899299575d28\": rpc error: code = NotFound desc = could not find container \"1ba24c2e067b9f3a14d9d018b66300b54d866530d30174fbdb3c899299575d28\": container with ID starting with 1ba24c2e067b9f3a14d9d018b66300b54d866530d30174fbdb3c899299575d28 not found: ID does not exist" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.834754 4870 scope.go:117] "RemoveContainer" containerID="a23b08d1e6780f4c2ce9ff447bd40050abb1e353c16ebccc8c3ecc7bc9a35e52" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.835059 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a23b08d1e6780f4c2ce9ff447bd40050abb1e353c16ebccc8c3ecc7bc9a35e52"} err="failed to get container status \"a23b08d1e6780f4c2ce9ff447bd40050abb1e353c16ebccc8c3ecc7bc9a35e52\": rpc error: code = NotFound desc = could not find container \"a23b08d1e6780f4c2ce9ff447bd40050abb1e353c16ebccc8c3ecc7bc9a35e52\": container with ID starting with a23b08d1e6780f4c2ce9ff447bd40050abb1e353c16ebccc8c3ecc7bc9a35e52 not found: ID does not exist" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.835074 4870 scope.go:117] "RemoveContainer" containerID="ab4295721b6ec6d64040d321899ab478c612e91cbc3b99b33753869578a2d1e8" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.835364 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ab4295721b6ec6d64040d321899ab478c612e91cbc3b99b33753869578a2d1e8"} err="failed to get container status \"ab4295721b6ec6d64040d321899ab478c612e91cbc3b99b33753869578a2d1e8\": rpc error: code = NotFound desc = could not find container \"ab4295721b6ec6d64040d321899ab478c612e91cbc3b99b33753869578a2d1e8\": container with ID starting with ab4295721b6ec6d64040d321899ab478c612e91cbc3b99b33753869578a2d1e8 not found: ID does not exist" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.835379 4870 scope.go:117] "RemoveContainer" containerID="9fe3468cdd612fc4378e1f6f55a2a9d00274f7ddb65cb57b2a368dfc065cfb4e" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.836458 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9fe3468cdd612fc4378e1f6f55a2a9d00274f7ddb65cb57b2a368dfc065cfb4e"} err="failed to get container status \"9fe3468cdd612fc4378e1f6f55a2a9d00274f7ddb65cb57b2a368dfc065cfb4e\": rpc error: code = NotFound desc = could not find container \"9fe3468cdd612fc4378e1f6f55a2a9d00274f7ddb65cb57b2a368dfc065cfb4e\": container with ID starting with 9fe3468cdd612fc4378e1f6f55a2a9d00274f7ddb65cb57b2a368dfc065cfb4e not found: ID does not exist" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.836479 4870 scope.go:117] "RemoveContainer" containerID="fc9d4db944f837289d1923176d60cfc4a566a282c2d2e3a703e220df884703c7" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.836770 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fc9d4db944f837289d1923176d60cfc4a566a282c2d2e3a703e220df884703c7"} err="failed to get container status \"fc9d4db944f837289d1923176d60cfc4a566a282c2d2e3a703e220df884703c7\": rpc error: code = NotFound desc = could not find container \"fc9d4db944f837289d1923176d60cfc4a566a282c2d2e3a703e220df884703c7\": container with ID starting with fc9d4db944f837289d1923176d60cfc4a566a282c2d2e3a703e220df884703c7 not found: ID does not exist" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.836807 4870 scope.go:117] "RemoveContainer" containerID="d16ed1c65925549dcb69ab02c3013e25e3f123a83179f1af3a8e08bf91a59896" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.837077 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d16ed1c65925549dcb69ab02c3013e25e3f123a83179f1af3a8e08bf91a59896"} err="failed to get container status \"d16ed1c65925549dcb69ab02c3013e25e3f123a83179f1af3a8e08bf91a59896\": rpc error: code = NotFound desc = could not find container \"d16ed1c65925549dcb69ab02c3013e25e3f123a83179f1af3a8e08bf91a59896\": container with ID starting with d16ed1c65925549dcb69ab02c3013e25e3f123a83179f1af3a8e08bf91a59896 not found: ID does not exist" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.837102 4870 scope.go:117] "RemoveContainer" containerID="e72cff0723e2ce520340244b4e597d8016c58e1b071c4b087f2f6cd718741594" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.837734 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e72cff0723e2ce520340244b4e597d8016c58e1b071c4b087f2f6cd718741594"} err="failed to get container status \"e72cff0723e2ce520340244b4e597d8016c58e1b071c4b087f2f6cd718741594\": rpc error: code = NotFound desc = could not find container \"e72cff0723e2ce520340244b4e597d8016c58e1b071c4b087f2f6cd718741594\": container with ID starting with e72cff0723e2ce520340244b4e597d8016c58e1b071c4b087f2f6cd718741594 not found: ID does not exist" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.837781 4870 scope.go:117] "RemoveContainer" containerID="71dec55e859b11a081dbb90d7d6cf6b4846cf591382303d8329003ff50c71f2a" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.838104 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"71dec55e859b11a081dbb90d7d6cf6b4846cf591382303d8329003ff50c71f2a"} err="failed to get container status \"71dec55e859b11a081dbb90d7d6cf6b4846cf591382303d8329003ff50c71f2a\": rpc error: code = NotFound desc = could not find container \"71dec55e859b11a081dbb90d7d6cf6b4846cf591382303d8329003ff50c71f2a\": container with ID starting with 71dec55e859b11a081dbb90d7d6cf6b4846cf591382303d8329003ff50c71f2a not found: ID does not exist" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.838178 4870 scope.go:117] "RemoveContainer" containerID="ed59243b59c39ce82bf9bc8f33c0a8b84854846afff6a89f5d37ffc97a096b7c" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.838728 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ed59243b59c39ce82bf9bc8f33c0a8b84854846afff6a89f5d37ffc97a096b7c"} err="failed to get container status \"ed59243b59c39ce82bf9bc8f33c0a8b84854846afff6a89f5d37ffc97a096b7c\": rpc error: code = NotFound desc = could not find container \"ed59243b59c39ce82bf9bc8f33c0a8b84854846afff6a89f5d37ffc97a096b7c\": container with ID starting with ed59243b59c39ce82bf9bc8f33c0a8b84854846afff6a89f5d37ffc97a096b7c not found: ID does not exist" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.838752 4870 scope.go:117] "RemoveContainer" containerID="f3b6d20cf459b1f6d3aff5b9290f9391909baf27ee40c49e2b0cd2d5cfe889ad" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.838967 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f3b6d20cf459b1f6d3aff5b9290f9391909baf27ee40c49e2b0cd2d5cfe889ad"} err="failed to get container status \"f3b6d20cf459b1f6d3aff5b9290f9391909baf27ee40c49e2b0cd2d5cfe889ad\": rpc error: code = NotFound desc = could not find container \"f3b6d20cf459b1f6d3aff5b9290f9391909baf27ee40c49e2b0cd2d5cfe889ad\": container with ID starting with f3b6d20cf459b1f6d3aff5b9290f9391909baf27ee40c49e2b0cd2d5cfe889ad not found: ID does not exist" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.839004 4870 scope.go:117] "RemoveContainer" containerID="1ba24c2e067b9f3a14d9d018b66300b54d866530d30174fbdb3c899299575d28" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.839253 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1ba24c2e067b9f3a14d9d018b66300b54d866530d30174fbdb3c899299575d28"} err="failed to get container status \"1ba24c2e067b9f3a14d9d018b66300b54d866530d30174fbdb3c899299575d28\": rpc error: code = NotFound desc = could not find container \"1ba24c2e067b9f3a14d9d018b66300b54d866530d30174fbdb3c899299575d28\": container with ID starting with 1ba24c2e067b9f3a14d9d018b66300b54d866530d30174fbdb3c899299575d28 not found: ID does not exist" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.863184 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/03297609-38cc-4df8-8150-86105829e6c4-env-overrides\") pod \"ovnkube-node-bxhvl\" (UID: \"03297609-38cc-4df8-8150-86105829e6c4\") " pod="openshift-ovn-kubernetes/ovnkube-node-bxhvl" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.863290 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/03297609-38cc-4df8-8150-86105829e6c4-node-log\") pod \"ovnkube-node-bxhvl\" (UID: \"03297609-38cc-4df8-8150-86105829e6c4\") " pod="openshift-ovn-kubernetes/ovnkube-node-bxhvl" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.863418 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/03297609-38cc-4df8-8150-86105829e6c4-node-log\") pod \"ovnkube-node-bxhvl\" (UID: \"03297609-38cc-4df8-8150-86105829e6c4\") " pod="openshift-ovn-kubernetes/ovnkube-node-bxhvl" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.863573 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/03297609-38cc-4df8-8150-86105829e6c4-run-openvswitch\") pod \"ovnkube-node-bxhvl\" (UID: \"03297609-38cc-4df8-8150-86105829e6c4\") " pod="openshift-ovn-kubernetes/ovnkube-node-bxhvl" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.863660 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/03297609-38cc-4df8-8150-86105829e6c4-run-openvswitch\") pod \"ovnkube-node-bxhvl\" (UID: \"03297609-38cc-4df8-8150-86105829e6c4\") " pod="openshift-ovn-kubernetes/ovnkube-node-bxhvl" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.863804 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/03297609-38cc-4df8-8150-86105829e6c4-host-cni-netd\") pod \"ovnkube-node-bxhvl\" (UID: \"03297609-38cc-4df8-8150-86105829e6c4\") " pod="openshift-ovn-kubernetes/ovnkube-node-bxhvl" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.863745 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/03297609-38cc-4df8-8150-86105829e6c4-host-cni-netd\") pod \"ovnkube-node-bxhvl\" (UID: \"03297609-38cc-4df8-8150-86105829e6c4\") " pod="openshift-ovn-kubernetes/ovnkube-node-bxhvl" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.863942 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/03297609-38cc-4df8-8150-86105829e6c4-env-overrides\") pod \"ovnkube-node-bxhvl\" (UID: \"03297609-38cc-4df8-8150-86105829e6c4\") " pod="openshift-ovn-kubernetes/ovnkube-node-bxhvl" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.864169 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/03297609-38cc-4df8-8150-86105829e6c4-var-lib-openvswitch\") pod \"ovnkube-node-bxhvl\" (UID: \"03297609-38cc-4df8-8150-86105829e6c4\") " pod="openshift-ovn-kubernetes/ovnkube-node-bxhvl" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.864238 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/03297609-38cc-4df8-8150-86105829e6c4-var-lib-openvswitch\") pod \"ovnkube-node-bxhvl\" (UID: \"03297609-38cc-4df8-8150-86105829e6c4\") " pod="openshift-ovn-kubernetes/ovnkube-node-bxhvl" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.864336 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/03297609-38cc-4df8-8150-86105829e6c4-host-run-netns\") pod \"ovnkube-node-bxhvl\" (UID: \"03297609-38cc-4df8-8150-86105829e6c4\") " pod="openshift-ovn-kubernetes/ovnkube-node-bxhvl" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.864387 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/03297609-38cc-4df8-8150-86105829e6c4-host-run-netns\") pod \"ovnkube-node-bxhvl\" (UID: \"03297609-38cc-4df8-8150-86105829e6c4\") " pod="openshift-ovn-kubernetes/ovnkube-node-bxhvl" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.864489 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/03297609-38cc-4df8-8150-86105829e6c4-ovnkube-script-lib\") pod \"ovnkube-node-bxhvl\" (UID: \"03297609-38cc-4df8-8150-86105829e6c4\") " pod="openshift-ovn-kubernetes/ovnkube-node-bxhvl" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.865144 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/03297609-38cc-4df8-8150-86105829e6c4-ovnkube-script-lib\") pod \"ovnkube-node-bxhvl\" (UID: \"03297609-38cc-4df8-8150-86105829e6c4\") " pod="openshift-ovn-kubernetes/ovnkube-node-bxhvl" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.865227 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/03297609-38cc-4df8-8150-86105829e6c4-run-ovn\") pod \"ovnkube-node-bxhvl\" (UID: \"03297609-38cc-4df8-8150-86105829e6c4\") " pod="openshift-ovn-kubernetes/ovnkube-node-bxhvl" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.865310 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/03297609-38cc-4df8-8150-86105829e6c4-run-ovn\") pod \"ovnkube-node-bxhvl\" (UID: \"03297609-38cc-4df8-8150-86105829e6c4\") " pod="openshift-ovn-kubernetes/ovnkube-node-bxhvl" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.865383 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/03297609-38cc-4df8-8150-86105829e6c4-run-systemd\") pod \"ovnkube-node-bxhvl\" (UID: \"03297609-38cc-4df8-8150-86105829e6c4\") " pod="openshift-ovn-kubernetes/ovnkube-node-bxhvl" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.865507 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/03297609-38cc-4df8-8150-86105829e6c4-run-systemd\") pod \"ovnkube-node-bxhvl\" (UID: \"03297609-38cc-4df8-8150-86105829e6c4\") " pod="openshift-ovn-kubernetes/ovnkube-node-bxhvl" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.865586 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/03297609-38cc-4df8-8150-86105829e6c4-host-kubelet\") pod \"ovnkube-node-bxhvl\" (UID: \"03297609-38cc-4df8-8150-86105829e6c4\") " pod="openshift-ovn-kubernetes/ovnkube-node-bxhvl" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.865654 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/03297609-38cc-4df8-8150-86105829e6c4-host-kubelet\") pod \"ovnkube-node-bxhvl\" (UID: \"03297609-38cc-4df8-8150-86105829e6c4\") " pod="openshift-ovn-kubernetes/ovnkube-node-bxhvl" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.865608 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/03297609-38cc-4df8-8150-86105829e6c4-host-slash\") pod \"ovnkube-node-bxhvl\" (UID: \"03297609-38cc-4df8-8150-86105829e6c4\") " pod="openshift-ovn-kubernetes/ovnkube-node-bxhvl" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.865728 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/03297609-38cc-4df8-8150-86105829e6c4-host-slash\") pod \"ovnkube-node-bxhvl\" (UID: \"03297609-38cc-4df8-8150-86105829e6c4\") " pod="openshift-ovn-kubernetes/ovnkube-node-bxhvl" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.865746 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/03297609-38cc-4df8-8150-86105829e6c4-ovnkube-config\") pod \"ovnkube-node-bxhvl\" (UID: \"03297609-38cc-4df8-8150-86105829e6c4\") " pod="openshift-ovn-kubernetes/ovnkube-node-bxhvl" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.865805 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4pktf\" (UniqueName: \"kubernetes.io/projected/03297609-38cc-4df8-8150-86105829e6c4-kube-api-access-4pktf\") pod \"ovnkube-node-bxhvl\" (UID: \"03297609-38cc-4df8-8150-86105829e6c4\") " pod="openshift-ovn-kubernetes/ovnkube-node-bxhvl" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.865836 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/03297609-38cc-4df8-8150-86105829e6c4-log-socket\") pod \"ovnkube-node-bxhvl\" (UID: \"03297609-38cc-4df8-8150-86105829e6c4\") " pod="openshift-ovn-kubernetes/ovnkube-node-bxhvl" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.865862 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/03297609-38cc-4df8-8150-86105829e6c4-host-cni-bin\") pod \"ovnkube-node-bxhvl\" (UID: \"03297609-38cc-4df8-8150-86105829e6c4\") " pod="openshift-ovn-kubernetes/ovnkube-node-bxhvl" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.865885 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/03297609-38cc-4df8-8150-86105829e6c4-ovn-node-metrics-cert\") pod \"ovnkube-node-bxhvl\" (UID: \"03297609-38cc-4df8-8150-86105829e6c4\") " pod="openshift-ovn-kubernetes/ovnkube-node-bxhvl" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.865921 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/03297609-38cc-4df8-8150-86105829e6c4-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-bxhvl\" (UID: \"03297609-38cc-4df8-8150-86105829e6c4\") " pod="openshift-ovn-kubernetes/ovnkube-node-bxhvl" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.865943 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/03297609-38cc-4df8-8150-86105829e6c4-systemd-units\") pod \"ovnkube-node-bxhvl\" (UID: \"03297609-38cc-4df8-8150-86105829e6c4\") " pod="openshift-ovn-kubernetes/ovnkube-node-bxhvl" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.865958 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/03297609-38cc-4df8-8150-86105829e6c4-etc-openvswitch\") pod \"ovnkube-node-bxhvl\" (UID: \"03297609-38cc-4df8-8150-86105829e6c4\") " pod="openshift-ovn-kubernetes/ovnkube-node-bxhvl" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.865956 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/03297609-38cc-4df8-8150-86105829e6c4-log-socket\") pod \"ovnkube-node-bxhvl\" (UID: \"03297609-38cc-4df8-8150-86105829e6c4\") " pod="openshift-ovn-kubernetes/ovnkube-node-bxhvl" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.865976 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/03297609-38cc-4df8-8150-86105829e6c4-host-run-ovn-kubernetes\") pod \"ovnkube-node-bxhvl\" (UID: \"03297609-38cc-4df8-8150-86105829e6c4\") " pod="openshift-ovn-kubernetes/ovnkube-node-bxhvl" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.865999 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/03297609-38cc-4df8-8150-86105829e6c4-host-run-ovn-kubernetes\") pod \"ovnkube-node-bxhvl\" (UID: \"03297609-38cc-4df8-8150-86105829e6c4\") " pod="openshift-ovn-kubernetes/ovnkube-node-bxhvl" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.866075 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/03297609-38cc-4df8-8150-86105829e6c4-etc-openvswitch\") pod \"ovnkube-node-bxhvl\" (UID: \"03297609-38cc-4df8-8150-86105829e6c4\") " pod="openshift-ovn-kubernetes/ovnkube-node-bxhvl" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.866115 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/03297609-38cc-4df8-8150-86105829e6c4-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-bxhvl\" (UID: \"03297609-38cc-4df8-8150-86105829e6c4\") " pod="openshift-ovn-kubernetes/ovnkube-node-bxhvl" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.866148 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/03297609-38cc-4df8-8150-86105829e6c4-systemd-units\") pod \"ovnkube-node-bxhvl\" (UID: \"03297609-38cc-4df8-8150-86105829e6c4\") " pod="openshift-ovn-kubernetes/ovnkube-node-bxhvl" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.866190 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/03297609-38cc-4df8-8150-86105829e6c4-host-cni-bin\") pod \"ovnkube-node-bxhvl\" (UID: \"03297609-38cc-4df8-8150-86105829e6c4\") " pod="openshift-ovn-kubernetes/ovnkube-node-bxhvl" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.867713 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/03297609-38cc-4df8-8150-86105829e6c4-ovnkube-config\") pod \"ovnkube-node-bxhvl\" (UID: \"03297609-38cc-4df8-8150-86105829e6c4\") " pod="openshift-ovn-kubernetes/ovnkube-node-bxhvl" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.870088 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/03297609-38cc-4df8-8150-86105829e6c4-ovn-node-metrics-cert\") pod \"ovnkube-node-bxhvl\" (UID: \"03297609-38cc-4df8-8150-86105829e6c4\") " pod="openshift-ovn-kubernetes/ovnkube-node-bxhvl" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.881675 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4pktf\" (UniqueName: \"kubernetes.io/projected/03297609-38cc-4df8-8150-86105829e6c4-kube-api-access-4pktf\") pod \"ovnkube-node-bxhvl\" (UID: \"03297609-38cc-4df8-8150-86105829e6c4\") " pod="openshift-ovn-kubernetes/ovnkube-node-bxhvl" Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.913099 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-v7br2"] Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.913970 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-v7br2"] Oct 14 07:13:18 crc kubenswrapper[4870]: I1014 07:13:18.939331 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-bxhvl" Oct 14 07:13:18 crc kubenswrapper[4870]: W1014 07:13:18.974469 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod03297609_38cc_4df8_8150_86105829e6c4.slice/crio-af89559afb67e7bdb1f63a1f09d9b343c86f1d2c0aedd9bae6b08f4bbe7f9b2d WatchSource:0}: Error finding container af89559afb67e7bdb1f63a1f09d9b343c86f1d2c0aedd9bae6b08f4bbe7f9b2d: Status 404 returned error can't find the container with id af89559afb67e7bdb1f63a1f09d9b343c86f1d2c0aedd9bae6b08f4bbe7f9b2d Oct 14 07:13:19 crc kubenswrapper[4870]: I1014 07:13:19.047766 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8dcaef0e-42e3-424b-bbe5-4a023efe912d" path="/var/lib/kubelet/pods/8dcaef0e-42e3-424b-bbe5-4a023efe912d/volumes" Oct 14 07:13:19 crc kubenswrapper[4870]: I1014 07:13:19.575419 4870 generic.go:334] "Generic (PLEG): container finished" podID="03297609-38cc-4df8-8150-86105829e6c4" containerID="9b314bc13018c751536530002a1f568dc7d295dcf17f9187fd3f6d08c2a0d085" exitCode=0 Oct 14 07:13:19 crc kubenswrapper[4870]: I1014 07:13:19.575909 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bxhvl" event={"ID":"03297609-38cc-4df8-8150-86105829e6c4","Type":"ContainerDied","Data":"9b314bc13018c751536530002a1f568dc7d295dcf17f9187fd3f6d08c2a0d085"} Oct 14 07:13:19 crc kubenswrapper[4870]: I1014 07:13:19.575981 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bxhvl" event={"ID":"03297609-38cc-4df8-8150-86105829e6c4","Type":"ContainerStarted","Data":"af89559afb67e7bdb1f63a1f09d9b343c86f1d2c0aedd9bae6b08f4bbe7f9b2d"} Oct 14 07:13:19 crc kubenswrapper[4870]: I1014 07:13:19.581843 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-fdxjh_8a3a430a-b540-4a8c-adad-b2893dbf9898/kube-multus/2.log" Oct 14 07:13:19 crc kubenswrapper[4870]: I1014 07:13:19.582022 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-fdxjh" event={"ID":"8a3a430a-b540-4a8c-adad-b2893dbf9898","Type":"ContainerStarted","Data":"13758d0d7140e3ce34f83913b0b02811fb1d032ac438a5dfdf69e1042190278a"} Oct 14 07:13:20 crc kubenswrapper[4870]: I1014 07:13:20.596415 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bxhvl" event={"ID":"03297609-38cc-4df8-8150-86105829e6c4","Type":"ContainerStarted","Data":"ef54c2f99bb25a971a9116d85b3180a3edf52396d1bb14658a9a8bc8bcbeeb47"} Oct 14 07:13:20 crc kubenswrapper[4870]: I1014 07:13:20.597074 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bxhvl" event={"ID":"03297609-38cc-4df8-8150-86105829e6c4","Type":"ContainerStarted","Data":"7dbd68893ebf49b16c0c9688d61867d072d819394f53e12e8ebe93e9ea5cde0c"} Oct 14 07:13:20 crc kubenswrapper[4870]: I1014 07:13:20.597089 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bxhvl" event={"ID":"03297609-38cc-4df8-8150-86105829e6c4","Type":"ContainerStarted","Data":"dd29e7833ac6585c5ec2aceaf00673a9eacd06ab6c248420b52a3bbc268effde"} Oct 14 07:13:20 crc kubenswrapper[4870]: I1014 07:13:20.597097 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bxhvl" event={"ID":"03297609-38cc-4df8-8150-86105829e6c4","Type":"ContainerStarted","Data":"11661931dae422fa75a954a67f1cfc84bc51b6fd0a13c2be8fb2ff989e57d665"} Oct 14 07:13:20 crc kubenswrapper[4870]: I1014 07:13:20.597106 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bxhvl" event={"ID":"03297609-38cc-4df8-8150-86105829e6c4","Type":"ContainerStarted","Data":"3602c30f246697dd0435fa762d1aedb451c53c9c480062e011699201ea59675a"} Oct 14 07:13:20 crc kubenswrapper[4870]: I1014 07:13:20.597114 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bxhvl" event={"ID":"03297609-38cc-4df8-8150-86105829e6c4","Type":"ContainerStarted","Data":"62b82582a2a7e2a8611229865e9f747e8965cddab9246b82d6af62b219a6c78c"} Oct 14 07:13:23 crc kubenswrapper[4870]: I1014 07:13:23.623410 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bxhvl" event={"ID":"03297609-38cc-4df8-8150-86105829e6c4","Type":"ContainerStarted","Data":"db4aa573c22bc6467ad9ba1e4ad5d4ab7bd50fee141e55286e13a88095a2c507"} Oct 14 07:13:25 crc kubenswrapper[4870]: I1014 07:13:25.649144 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bxhvl" event={"ID":"03297609-38cc-4df8-8150-86105829e6c4","Type":"ContainerStarted","Data":"c55ec02379815256769e17f21137fa5aaaac173569b370fe6864e399d1ba6028"} Oct 14 07:13:25 crc kubenswrapper[4870]: I1014 07:13:25.650489 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-bxhvl" Oct 14 07:13:25 crc kubenswrapper[4870]: I1014 07:13:25.650530 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-bxhvl" Oct 14 07:13:25 crc kubenswrapper[4870]: I1014 07:13:25.650544 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-bxhvl" Oct 14 07:13:25 crc kubenswrapper[4870]: I1014 07:13:25.688587 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-bxhvl" Oct 14 07:13:25 crc kubenswrapper[4870]: I1014 07:13:25.692265 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-bxhvl" Oct 14 07:13:25 crc kubenswrapper[4870]: I1014 07:13:25.720125 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-bxhvl" podStartSLOduration=7.720105903 podStartE2EDuration="7.720105903s" podCreationTimestamp="2025-10-14 07:13:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:13:25.686934625 +0000 UTC m=+741.384294996" watchObservedRunningTime="2025-10-14 07:13:25.720105903 +0000 UTC m=+741.417466284" Oct 14 07:13:27 crc kubenswrapper[4870]: I1014 07:13:27.659533 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["crc-storage/crc-storage-crc-d2ngg"] Oct 14 07:13:27 crc kubenswrapper[4870]: I1014 07:13:27.660586 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-d2ngg" Oct 14 07:13:27 crc kubenswrapper[4870]: I1014 07:13:27.668105 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"crc-storage" Oct 14 07:13:27 crc kubenswrapper[4870]: I1014 07:13:27.668394 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"kube-root-ca.crt" Oct 14 07:13:27 crc kubenswrapper[4870]: I1014 07:13:27.668628 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"openshift-service-ca.crt" Oct 14 07:13:27 crc kubenswrapper[4870]: I1014 07:13:27.668841 4870 reflector.go:368] Caches populated for *v1.Secret from object-"crc-storage"/"crc-storage-dockercfg-pm8px" Oct 14 07:13:27 crc kubenswrapper[4870]: I1014 07:13:27.684002 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-d2ngg"] Oct 14 07:13:27 crc kubenswrapper[4870]: I1014 07:13:27.797945 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/436866a2-a717-4431-aefe-56a515024a72-node-mnt\") pod \"crc-storage-crc-d2ngg\" (UID: \"436866a2-a717-4431-aefe-56a515024a72\") " pod="crc-storage/crc-storage-crc-d2ngg" Oct 14 07:13:27 crc kubenswrapper[4870]: I1014 07:13:27.798102 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/436866a2-a717-4431-aefe-56a515024a72-crc-storage\") pod \"crc-storage-crc-d2ngg\" (UID: \"436866a2-a717-4431-aefe-56a515024a72\") " pod="crc-storage/crc-storage-crc-d2ngg" Oct 14 07:13:27 crc kubenswrapper[4870]: I1014 07:13:27.798150 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vprhw\" (UniqueName: \"kubernetes.io/projected/436866a2-a717-4431-aefe-56a515024a72-kube-api-access-vprhw\") pod \"crc-storage-crc-d2ngg\" (UID: \"436866a2-a717-4431-aefe-56a515024a72\") " pod="crc-storage/crc-storage-crc-d2ngg" Oct 14 07:13:27 crc kubenswrapper[4870]: I1014 07:13:27.899358 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/436866a2-a717-4431-aefe-56a515024a72-node-mnt\") pod \"crc-storage-crc-d2ngg\" (UID: \"436866a2-a717-4431-aefe-56a515024a72\") " pod="crc-storage/crc-storage-crc-d2ngg" Oct 14 07:13:27 crc kubenswrapper[4870]: I1014 07:13:27.899428 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/436866a2-a717-4431-aefe-56a515024a72-crc-storage\") pod \"crc-storage-crc-d2ngg\" (UID: \"436866a2-a717-4431-aefe-56a515024a72\") " pod="crc-storage/crc-storage-crc-d2ngg" Oct 14 07:13:27 crc kubenswrapper[4870]: I1014 07:13:27.899479 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vprhw\" (UniqueName: \"kubernetes.io/projected/436866a2-a717-4431-aefe-56a515024a72-kube-api-access-vprhw\") pod \"crc-storage-crc-d2ngg\" (UID: \"436866a2-a717-4431-aefe-56a515024a72\") " pod="crc-storage/crc-storage-crc-d2ngg" Oct 14 07:13:27 crc kubenswrapper[4870]: I1014 07:13:27.899685 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/436866a2-a717-4431-aefe-56a515024a72-node-mnt\") pod \"crc-storage-crc-d2ngg\" (UID: \"436866a2-a717-4431-aefe-56a515024a72\") " pod="crc-storage/crc-storage-crc-d2ngg" Oct 14 07:13:27 crc kubenswrapper[4870]: I1014 07:13:27.900331 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/436866a2-a717-4431-aefe-56a515024a72-crc-storage\") pod \"crc-storage-crc-d2ngg\" (UID: \"436866a2-a717-4431-aefe-56a515024a72\") " pod="crc-storage/crc-storage-crc-d2ngg" Oct 14 07:13:27 crc kubenswrapper[4870]: I1014 07:13:27.920184 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vprhw\" (UniqueName: \"kubernetes.io/projected/436866a2-a717-4431-aefe-56a515024a72-kube-api-access-vprhw\") pod \"crc-storage-crc-d2ngg\" (UID: \"436866a2-a717-4431-aefe-56a515024a72\") " pod="crc-storage/crc-storage-crc-d2ngg" Oct 14 07:13:28 crc kubenswrapper[4870]: I1014 07:13:28.066707 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-d2ngg" Oct 14 07:13:28 crc kubenswrapper[4870]: E1014 07:13:28.101023 4870 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-d2ngg_crc-storage_436866a2-a717-4431-aefe-56a515024a72_0(d4277acb7c708abc7dc0cb6c3aaccc4828749a1b805fa98ec40fa8cf7b85acd7): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 14 07:13:28 crc kubenswrapper[4870]: E1014 07:13:28.101165 4870 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-d2ngg_crc-storage_436866a2-a717-4431-aefe-56a515024a72_0(d4277acb7c708abc7dc0cb6c3aaccc4828749a1b805fa98ec40fa8cf7b85acd7): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="crc-storage/crc-storage-crc-d2ngg" Oct 14 07:13:28 crc kubenswrapper[4870]: E1014 07:13:28.101263 4870 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-d2ngg_crc-storage_436866a2-a717-4431-aefe-56a515024a72_0(d4277acb7c708abc7dc0cb6c3aaccc4828749a1b805fa98ec40fa8cf7b85acd7): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="crc-storage/crc-storage-crc-d2ngg" Oct 14 07:13:28 crc kubenswrapper[4870]: E1014 07:13:28.101376 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"crc-storage-crc-d2ngg_crc-storage(436866a2-a717-4431-aefe-56a515024a72)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"crc-storage-crc-d2ngg_crc-storage(436866a2-a717-4431-aefe-56a515024a72)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-d2ngg_crc-storage_436866a2-a717-4431-aefe-56a515024a72_0(d4277acb7c708abc7dc0cb6c3aaccc4828749a1b805fa98ec40fa8cf7b85acd7): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="crc-storage/crc-storage-crc-d2ngg" podUID="436866a2-a717-4431-aefe-56a515024a72" Oct 14 07:13:28 crc kubenswrapper[4870]: I1014 07:13:28.668705 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-d2ngg" Oct 14 07:13:28 crc kubenswrapper[4870]: I1014 07:13:28.670618 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-d2ngg" Oct 14 07:13:28 crc kubenswrapper[4870]: E1014 07:13:28.711955 4870 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-d2ngg_crc-storage_436866a2-a717-4431-aefe-56a515024a72_0(7debe0b13e57b7843aaf5f895774a8b567d84b1eebd7e81c39ff3460eeb7a7cc): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 14 07:13:28 crc kubenswrapper[4870]: E1014 07:13:28.712075 4870 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-d2ngg_crc-storage_436866a2-a717-4431-aefe-56a515024a72_0(7debe0b13e57b7843aaf5f895774a8b567d84b1eebd7e81c39ff3460eeb7a7cc): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="crc-storage/crc-storage-crc-d2ngg" Oct 14 07:13:28 crc kubenswrapper[4870]: E1014 07:13:28.712293 4870 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-d2ngg_crc-storage_436866a2-a717-4431-aefe-56a515024a72_0(7debe0b13e57b7843aaf5f895774a8b567d84b1eebd7e81c39ff3460eeb7a7cc): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="crc-storage/crc-storage-crc-d2ngg" Oct 14 07:13:28 crc kubenswrapper[4870]: E1014 07:13:28.712388 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"crc-storage-crc-d2ngg_crc-storage(436866a2-a717-4431-aefe-56a515024a72)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"crc-storage-crc-d2ngg_crc-storage(436866a2-a717-4431-aefe-56a515024a72)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-d2ngg_crc-storage_436866a2-a717-4431-aefe-56a515024a72_0(7debe0b13e57b7843aaf5f895774a8b567d84b1eebd7e81c39ff3460eeb7a7cc): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="crc-storage/crc-storage-crc-d2ngg" podUID="436866a2-a717-4431-aefe-56a515024a72" Oct 14 07:13:33 crc kubenswrapper[4870]: I1014 07:13:33.154375 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-4wngz"] Oct 14 07:13:33 crc kubenswrapper[4870]: I1014 07:13:33.155620 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-4wngz" podUID="dce0e109-9f42-4320-b7bb-36dfd8aecb90" containerName="controller-manager" containerID="cri-o://5691df69a48cbc54d505808e821a44380c74c285fce936e31d76f5cf21d595e5" gracePeriod=30 Oct 14 07:13:33 crc kubenswrapper[4870]: I1014 07:13:33.240202 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-7jz27"] Oct 14 07:13:33 crc kubenswrapper[4870]: I1014 07:13:33.240415 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-7jz27" podUID="b419b072-5f0d-4103-82c6-c296f71eee8a" containerName="route-controller-manager" containerID="cri-o://85f6abb813dc9e9b36519bc932a6c44f0fd1ab710d7df37a737f3b0870b0b8c3" gracePeriod=30 Oct 14 07:13:33 crc kubenswrapper[4870]: I1014 07:13:33.509075 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-4wngz" Oct 14 07:13:33 crc kubenswrapper[4870]: I1014 07:13:33.553361 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-7jz27" Oct 14 07:13:33 crc kubenswrapper[4870]: I1014 07:13:33.597570 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/dce0e109-9f42-4320-b7bb-36dfd8aecb90-client-ca\") pod \"dce0e109-9f42-4320-b7bb-36dfd8aecb90\" (UID: \"dce0e109-9f42-4320-b7bb-36dfd8aecb90\") " Oct 14 07:13:33 crc kubenswrapper[4870]: I1014 07:13:33.597737 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d5lhs\" (UniqueName: \"kubernetes.io/projected/dce0e109-9f42-4320-b7bb-36dfd8aecb90-kube-api-access-d5lhs\") pod \"dce0e109-9f42-4320-b7bb-36dfd8aecb90\" (UID: \"dce0e109-9f42-4320-b7bb-36dfd8aecb90\") " Oct 14 07:13:33 crc kubenswrapper[4870]: I1014 07:13:33.597759 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/dce0e109-9f42-4320-b7bb-36dfd8aecb90-proxy-ca-bundles\") pod \"dce0e109-9f42-4320-b7bb-36dfd8aecb90\" (UID: \"dce0e109-9f42-4320-b7bb-36dfd8aecb90\") " Oct 14 07:13:33 crc kubenswrapper[4870]: I1014 07:13:33.597788 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dce0e109-9f42-4320-b7bb-36dfd8aecb90-config\") pod \"dce0e109-9f42-4320-b7bb-36dfd8aecb90\" (UID: \"dce0e109-9f42-4320-b7bb-36dfd8aecb90\") " Oct 14 07:13:33 crc kubenswrapper[4870]: I1014 07:13:33.597819 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dce0e109-9f42-4320-b7bb-36dfd8aecb90-serving-cert\") pod \"dce0e109-9f42-4320-b7bb-36dfd8aecb90\" (UID: \"dce0e109-9f42-4320-b7bb-36dfd8aecb90\") " Oct 14 07:13:33 crc kubenswrapper[4870]: I1014 07:13:33.598572 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dce0e109-9f42-4320-b7bb-36dfd8aecb90-client-ca" (OuterVolumeSpecName: "client-ca") pod "dce0e109-9f42-4320-b7bb-36dfd8aecb90" (UID: "dce0e109-9f42-4320-b7bb-36dfd8aecb90"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:13:33 crc kubenswrapper[4870]: I1014 07:13:33.599105 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dce0e109-9f42-4320-b7bb-36dfd8aecb90-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "dce0e109-9f42-4320-b7bb-36dfd8aecb90" (UID: "dce0e109-9f42-4320-b7bb-36dfd8aecb90"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:13:33 crc kubenswrapper[4870]: I1014 07:13:33.600218 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dce0e109-9f42-4320-b7bb-36dfd8aecb90-config" (OuterVolumeSpecName: "config") pod "dce0e109-9f42-4320-b7bb-36dfd8aecb90" (UID: "dce0e109-9f42-4320-b7bb-36dfd8aecb90"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:13:33 crc kubenswrapper[4870]: I1014 07:13:33.603983 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dce0e109-9f42-4320-b7bb-36dfd8aecb90-kube-api-access-d5lhs" (OuterVolumeSpecName: "kube-api-access-d5lhs") pod "dce0e109-9f42-4320-b7bb-36dfd8aecb90" (UID: "dce0e109-9f42-4320-b7bb-36dfd8aecb90"). InnerVolumeSpecName "kube-api-access-d5lhs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:13:33 crc kubenswrapper[4870]: I1014 07:13:33.605035 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dce0e109-9f42-4320-b7bb-36dfd8aecb90-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "dce0e109-9f42-4320-b7bb-36dfd8aecb90" (UID: "dce0e109-9f42-4320-b7bb-36dfd8aecb90"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:13:33 crc kubenswrapper[4870]: I1014 07:13:33.698860 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b419b072-5f0d-4103-82c6-c296f71eee8a-config\") pod \"b419b072-5f0d-4103-82c6-c296f71eee8a\" (UID: \"b419b072-5f0d-4103-82c6-c296f71eee8a\") " Oct 14 07:13:33 crc kubenswrapper[4870]: I1014 07:13:33.698919 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b419b072-5f0d-4103-82c6-c296f71eee8a-client-ca\") pod \"b419b072-5f0d-4103-82c6-c296f71eee8a\" (UID: \"b419b072-5f0d-4103-82c6-c296f71eee8a\") " Oct 14 07:13:33 crc kubenswrapper[4870]: I1014 07:13:33.698945 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cb6jb\" (UniqueName: \"kubernetes.io/projected/b419b072-5f0d-4103-82c6-c296f71eee8a-kube-api-access-cb6jb\") pod \"b419b072-5f0d-4103-82c6-c296f71eee8a\" (UID: \"b419b072-5f0d-4103-82c6-c296f71eee8a\") " Oct 14 07:13:33 crc kubenswrapper[4870]: I1014 07:13:33.698963 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b419b072-5f0d-4103-82c6-c296f71eee8a-serving-cert\") pod \"b419b072-5f0d-4103-82c6-c296f71eee8a\" (UID: \"b419b072-5f0d-4103-82c6-c296f71eee8a\") " Oct 14 07:13:33 crc kubenswrapper[4870]: I1014 07:13:33.699239 4870 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/dce0e109-9f42-4320-b7bb-36dfd8aecb90-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Oct 14 07:13:33 crc kubenswrapper[4870]: I1014 07:13:33.699253 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d5lhs\" (UniqueName: \"kubernetes.io/projected/dce0e109-9f42-4320-b7bb-36dfd8aecb90-kube-api-access-d5lhs\") on node \"crc\" DevicePath \"\"" Oct 14 07:13:33 crc kubenswrapper[4870]: I1014 07:13:33.699263 4870 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dce0e109-9f42-4320-b7bb-36dfd8aecb90-config\") on node \"crc\" DevicePath \"\"" Oct 14 07:13:33 crc kubenswrapper[4870]: I1014 07:13:33.699272 4870 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dce0e109-9f42-4320-b7bb-36dfd8aecb90-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 14 07:13:33 crc kubenswrapper[4870]: I1014 07:13:33.699279 4870 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/dce0e109-9f42-4320-b7bb-36dfd8aecb90-client-ca\") on node \"crc\" DevicePath \"\"" Oct 14 07:13:33 crc kubenswrapper[4870]: I1014 07:13:33.699848 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b419b072-5f0d-4103-82c6-c296f71eee8a-config" (OuterVolumeSpecName: "config") pod "b419b072-5f0d-4103-82c6-c296f71eee8a" (UID: "b419b072-5f0d-4103-82c6-c296f71eee8a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:13:33 crc kubenswrapper[4870]: I1014 07:13:33.700016 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b419b072-5f0d-4103-82c6-c296f71eee8a-client-ca" (OuterVolumeSpecName: "client-ca") pod "b419b072-5f0d-4103-82c6-c296f71eee8a" (UID: "b419b072-5f0d-4103-82c6-c296f71eee8a"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:13:33 crc kubenswrapper[4870]: I1014 07:13:33.700915 4870 generic.go:334] "Generic (PLEG): container finished" podID="dce0e109-9f42-4320-b7bb-36dfd8aecb90" containerID="5691df69a48cbc54d505808e821a44380c74c285fce936e31d76f5cf21d595e5" exitCode=0 Oct 14 07:13:33 crc kubenswrapper[4870]: I1014 07:13:33.700996 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-4wngz" event={"ID":"dce0e109-9f42-4320-b7bb-36dfd8aecb90","Type":"ContainerDied","Data":"5691df69a48cbc54d505808e821a44380c74c285fce936e31d76f5cf21d595e5"} Oct 14 07:13:33 crc kubenswrapper[4870]: I1014 07:13:33.701031 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-4wngz" event={"ID":"dce0e109-9f42-4320-b7bb-36dfd8aecb90","Type":"ContainerDied","Data":"8e036a0d8d7b4bbcdf48f51122e9b5e7af15d873287975925dd048f5a3f01332"} Oct 14 07:13:33 crc kubenswrapper[4870]: I1014 07:13:33.701053 4870 scope.go:117] "RemoveContainer" containerID="5691df69a48cbc54d505808e821a44380c74c285fce936e31d76f5cf21d595e5" Oct 14 07:13:33 crc kubenswrapper[4870]: I1014 07:13:33.701166 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-4wngz" Oct 14 07:13:33 crc kubenswrapper[4870]: I1014 07:13:33.702183 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b419b072-5f0d-4103-82c6-c296f71eee8a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "b419b072-5f0d-4103-82c6-c296f71eee8a" (UID: "b419b072-5f0d-4103-82c6-c296f71eee8a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:13:33 crc kubenswrapper[4870]: I1014 07:13:33.704385 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b419b072-5f0d-4103-82c6-c296f71eee8a-kube-api-access-cb6jb" (OuterVolumeSpecName: "kube-api-access-cb6jb") pod "b419b072-5f0d-4103-82c6-c296f71eee8a" (UID: "b419b072-5f0d-4103-82c6-c296f71eee8a"). InnerVolumeSpecName "kube-api-access-cb6jb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:13:33 crc kubenswrapper[4870]: I1014 07:13:33.704670 4870 generic.go:334] "Generic (PLEG): container finished" podID="b419b072-5f0d-4103-82c6-c296f71eee8a" containerID="85f6abb813dc9e9b36519bc932a6c44f0fd1ab710d7df37a737f3b0870b0b8c3" exitCode=0 Oct 14 07:13:33 crc kubenswrapper[4870]: I1014 07:13:33.704696 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-7jz27" event={"ID":"b419b072-5f0d-4103-82c6-c296f71eee8a","Type":"ContainerDied","Data":"85f6abb813dc9e9b36519bc932a6c44f0fd1ab710d7df37a737f3b0870b0b8c3"} Oct 14 07:13:33 crc kubenswrapper[4870]: I1014 07:13:33.704712 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-7jz27" event={"ID":"b419b072-5f0d-4103-82c6-c296f71eee8a","Type":"ContainerDied","Data":"a77f14f88687cd51ba8d857613ad96caed60a974c0552b6b9a922f84693a5ffd"} Oct 14 07:13:33 crc kubenswrapper[4870]: I1014 07:13:33.704760 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-7jz27" Oct 14 07:13:33 crc kubenswrapper[4870]: I1014 07:13:33.736223 4870 scope.go:117] "RemoveContainer" containerID="5691df69a48cbc54d505808e821a44380c74c285fce936e31d76f5cf21d595e5" Oct 14 07:13:33 crc kubenswrapper[4870]: E1014 07:13:33.736643 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5691df69a48cbc54d505808e821a44380c74c285fce936e31d76f5cf21d595e5\": container with ID starting with 5691df69a48cbc54d505808e821a44380c74c285fce936e31d76f5cf21d595e5 not found: ID does not exist" containerID="5691df69a48cbc54d505808e821a44380c74c285fce936e31d76f5cf21d595e5" Oct 14 07:13:33 crc kubenswrapper[4870]: I1014 07:13:33.736673 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5691df69a48cbc54d505808e821a44380c74c285fce936e31d76f5cf21d595e5"} err="failed to get container status \"5691df69a48cbc54d505808e821a44380c74c285fce936e31d76f5cf21d595e5\": rpc error: code = NotFound desc = could not find container \"5691df69a48cbc54d505808e821a44380c74c285fce936e31d76f5cf21d595e5\": container with ID starting with 5691df69a48cbc54d505808e821a44380c74c285fce936e31d76f5cf21d595e5 not found: ID does not exist" Oct 14 07:13:33 crc kubenswrapper[4870]: I1014 07:13:33.736694 4870 scope.go:117] "RemoveContainer" containerID="85f6abb813dc9e9b36519bc932a6c44f0fd1ab710d7df37a737f3b0870b0b8c3" Oct 14 07:13:33 crc kubenswrapper[4870]: I1014 07:13:33.739000 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-4wngz"] Oct 14 07:13:33 crc kubenswrapper[4870]: I1014 07:13:33.741695 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-4wngz"] Oct 14 07:13:33 crc kubenswrapper[4870]: I1014 07:13:33.758548 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-7jz27"] Oct 14 07:13:33 crc kubenswrapper[4870]: I1014 07:13:33.760346 4870 scope.go:117] "RemoveContainer" containerID="85f6abb813dc9e9b36519bc932a6c44f0fd1ab710d7df37a737f3b0870b0b8c3" Oct 14 07:13:33 crc kubenswrapper[4870]: I1014 07:13:33.760646 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-7jz27"] Oct 14 07:13:33 crc kubenswrapper[4870]: E1014 07:13:33.761014 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"85f6abb813dc9e9b36519bc932a6c44f0fd1ab710d7df37a737f3b0870b0b8c3\": container with ID starting with 85f6abb813dc9e9b36519bc932a6c44f0fd1ab710d7df37a737f3b0870b0b8c3 not found: ID does not exist" containerID="85f6abb813dc9e9b36519bc932a6c44f0fd1ab710d7df37a737f3b0870b0b8c3" Oct 14 07:13:33 crc kubenswrapper[4870]: I1014 07:13:33.761055 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"85f6abb813dc9e9b36519bc932a6c44f0fd1ab710d7df37a737f3b0870b0b8c3"} err="failed to get container status \"85f6abb813dc9e9b36519bc932a6c44f0fd1ab710d7df37a737f3b0870b0b8c3\": rpc error: code = NotFound desc = could not find container \"85f6abb813dc9e9b36519bc932a6c44f0fd1ab710d7df37a737f3b0870b0b8c3\": container with ID starting with 85f6abb813dc9e9b36519bc932a6c44f0fd1ab710d7df37a737f3b0870b0b8c3 not found: ID does not exist" Oct 14 07:13:33 crc kubenswrapper[4870]: I1014 07:13:33.800071 4870 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b419b072-5f0d-4103-82c6-c296f71eee8a-config\") on node \"crc\" DevicePath \"\"" Oct 14 07:13:33 crc kubenswrapper[4870]: I1014 07:13:33.800108 4870 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b419b072-5f0d-4103-82c6-c296f71eee8a-client-ca\") on node \"crc\" DevicePath \"\"" Oct 14 07:13:33 crc kubenswrapper[4870]: I1014 07:13:33.800123 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cb6jb\" (UniqueName: \"kubernetes.io/projected/b419b072-5f0d-4103-82c6-c296f71eee8a-kube-api-access-cb6jb\") on node \"crc\" DevicePath \"\"" Oct 14 07:13:33 crc kubenswrapper[4870]: I1014 07:13:33.800137 4870 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b419b072-5f0d-4103-82c6-c296f71eee8a-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 14 07:13:34 crc kubenswrapper[4870]: I1014 07:13:34.853948 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-9fc988b76-4gbpp"] Oct 14 07:13:34 crc kubenswrapper[4870]: E1014 07:13:34.854430 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dce0e109-9f42-4320-b7bb-36dfd8aecb90" containerName="controller-manager" Oct 14 07:13:34 crc kubenswrapper[4870]: I1014 07:13:34.854490 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="dce0e109-9f42-4320-b7bb-36dfd8aecb90" containerName="controller-manager" Oct 14 07:13:34 crc kubenswrapper[4870]: E1014 07:13:34.854531 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b419b072-5f0d-4103-82c6-c296f71eee8a" containerName="route-controller-manager" Oct 14 07:13:34 crc kubenswrapper[4870]: I1014 07:13:34.854550 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="b419b072-5f0d-4103-82c6-c296f71eee8a" containerName="route-controller-manager" Oct 14 07:13:34 crc kubenswrapper[4870]: I1014 07:13:34.854794 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="dce0e109-9f42-4320-b7bb-36dfd8aecb90" containerName="controller-manager" Oct 14 07:13:34 crc kubenswrapper[4870]: I1014 07:13:34.854832 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="b419b072-5f0d-4103-82c6-c296f71eee8a" containerName="route-controller-manager" Oct 14 07:13:34 crc kubenswrapper[4870]: I1014 07:13:34.855900 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-9fc988b76-4gbpp" Oct 14 07:13:34 crc kubenswrapper[4870]: I1014 07:13:34.862323 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7cb9656cfd-qwwn4"] Oct 14 07:13:34 crc kubenswrapper[4870]: I1014 07:13:34.862805 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Oct 14 07:13:34 crc kubenswrapper[4870]: I1014 07:13:34.863533 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7cb9656cfd-qwwn4" Oct 14 07:13:34 crc kubenswrapper[4870]: I1014 07:13:34.864323 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Oct 14 07:13:34 crc kubenswrapper[4870]: I1014 07:13:34.867310 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Oct 14 07:13:34 crc kubenswrapper[4870]: I1014 07:13:34.868097 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Oct 14 07:13:34 crc kubenswrapper[4870]: I1014 07:13:34.868130 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Oct 14 07:13:34 crc kubenswrapper[4870]: I1014 07:13:34.868686 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Oct 14 07:13:34 crc kubenswrapper[4870]: I1014 07:13:34.869700 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Oct 14 07:13:34 crc kubenswrapper[4870]: I1014 07:13:34.871729 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Oct 14 07:13:34 crc kubenswrapper[4870]: I1014 07:13:34.872643 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Oct 14 07:13:34 crc kubenswrapper[4870]: I1014 07:13:34.872946 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Oct 14 07:13:34 crc kubenswrapper[4870]: I1014 07:13:34.873668 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Oct 14 07:13:34 crc kubenswrapper[4870]: I1014 07:13:34.881057 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Oct 14 07:13:34 crc kubenswrapper[4870]: I1014 07:13:34.885935 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Oct 14 07:13:34 crc kubenswrapper[4870]: I1014 07:13:34.890361 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-9fc988b76-4gbpp"] Oct 14 07:13:34 crc kubenswrapper[4870]: I1014 07:13:34.895924 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7cb9656cfd-qwwn4"] Oct 14 07:13:35 crc kubenswrapper[4870]: I1014 07:13:35.017548 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5fe046b1-e37b-423b-b9ca-4a8a97b65ca6-config\") pod \"route-controller-manager-7cb9656cfd-qwwn4\" (UID: \"5fe046b1-e37b-423b-b9ca-4a8a97b65ca6\") " pod="openshift-route-controller-manager/route-controller-manager-7cb9656cfd-qwwn4" Oct 14 07:13:35 crc kubenswrapper[4870]: I1014 07:13:35.017621 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/241189b0-55ae-4646-8f6a-314cdd089cc2-client-ca\") pod \"controller-manager-9fc988b76-4gbpp\" (UID: \"241189b0-55ae-4646-8f6a-314cdd089cc2\") " pod="openshift-controller-manager/controller-manager-9fc988b76-4gbpp" Oct 14 07:13:35 crc kubenswrapper[4870]: I1014 07:13:35.017721 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/241189b0-55ae-4646-8f6a-314cdd089cc2-config\") pod \"controller-manager-9fc988b76-4gbpp\" (UID: \"241189b0-55ae-4646-8f6a-314cdd089cc2\") " pod="openshift-controller-manager/controller-manager-9fc988b76-4gbpp" Oct 14 07:13:35 crc kubenswrapper[4870]: I1014 07:13:35.017891 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/241189b0-55ae-4646-8f6a-314cdd089cc2-proxy-ca-bundles\") pod \"controller-manager-9fc988b76-4gbpp\" (UID: \"241189b0-55ae-4646-8f6a-314cdd089cc2\") " pod="openshift-controller-manager/controller-manager-9fc988b76-4gbpp" Oct 14 07:13:35 crc kubenswrapper[4870]: I1014 07:13:35.017984 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mnlst\" (UniqueName: \"kubernetes.io/projected/241189b0-55ae-4646-8f6a-314cdd089cc2-kube-api-access-mnlst\") pod \"controller-manager-9fc988b76-4gbpp\" (UID: \"241189b0-55ae-4646-8f6a-314cdd089cc2\") " pod="openshift-controller-manager/controller-manager-9fc988b76-4gbpp" Oct 14 07:13:35 crc kubenswrapper[4870]: I1014 07:13:35.018071 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5fe046b1-e37b-423b-b9ca-4a8a97b65ca6-serving-cert\") pod \"route-controller-manager-7cb9656cfd-qwwn4\" (UID: \"5fe046b1-e37b-423b-b9ca-4a8a97b65ca6\") " pod="openshift-route-controller-manager/route-controller-manager-7cb9656cfd-qwwn4" Oct 14 07:13:35 crc kubenswrapper[4870]: I1014 07:13:35.018124 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5fe046b1-e37b-423b-b9ca-4a8a97b65ca6-client-ca\") pod \"route-controller-manager-7cb9656cfd-qwwn4\" (UID: \"5fe046b1-e37b-423b-b9ca-4a8a97b65ca6\") " pod="openshift-route-controller-manager/route-controller-manager-7cb9656cfd-qwwn4" Oct 14 07:13:35 crc kubenswrapper[4870]: I1014 07:13:35.018328 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/241189b0-55ae-4646-8f6a-314cdd089cc2-serving-cert\") pod \"controller-manager-9fc988b76-4gbpp\" (UID: \"241189b0-55ae-4646-8f6a-314cdd089cc2\") " pod="openshift-controller-manager/controller-manager-9fc988b76-4gbpp" Oct 14 07:13:35 crc kubenswrapper[4870]: I1014 07:13:35.018373 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4pqm7\" (UniqueName: \"kubernetes.io/projected/5fe046b1-e37b-423b-b9ca-4a8a97b65ca6-kube-api-access-4pqm7\") pod \"route-controller-manager-7cb9656cfd-qwwn4\" (UID: \"5fe046b1-e37b-423b-b9ca-4a8a97b65ca6\") " pod="openshift-route-controller-manager/route-controller-manager-7cb9656cfd-qwwn4" Oct 14 07:13:35 crc kubenswrapper[4870]: I1014 07:13:35.046400 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b419b072-5f0d-4103-82c6-c296f71eee8a" path="/var/lib/kubelet/pods/b419b072-5f0d-4103-82c6-c296f71eee8a/volumes" Oct 14 07:13:35 crc kubenswrapper[4870]: I1014 07:13:35.047594 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dce0e109-9f42-4320-b7bb-36dfd8aecb90" path="/var/lib/kubelet/pods/dce0e109-9f42-4320-b7bb-36dfd8aecb90/volumes" Oct 14 07:13:35 crc kubenswrapper[4870]: I1014 07:13:35.120045 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5fe046b1-e37b-423b-b9ca-4a8a97b65ca6-serving-cert\") pod \"route-controller-manager-7cb9656cfd-qwwn4\" (UID: \"5fe046b1-e37b-423b-b9ca-4a8a97b65ca6\") " pod="openshift-route-controller-manager/route-controller-manager-7cb9656cfd-qwwn4" Oct 14 07:13:35 crc kubenswrapper[4870]: I1014 07:13:35.120099 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5fe046b1-e37b-423b-b9ca-4a8a97b65ca6-client-ca\") pod \"route-controller-manager-7cb9656cfd-qwwn4\" (UID: \"5fe046b1-e37b-423b-b9ca-4a8a97b65ca6\") " pod="openshift-route-controller-manager/route-controller-manager-7cb9656cfd-qwwn4" Oct 14 07:13:35 crc kubenswrapper[4870]: I1014 07:13:35.120151 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/241189b0-55ae-4646-8f6a-314cdd089cc2-serving-cert\") pod \"controller-manager-9fc988b76-4gbpp\" (UID: \"241189b0-55ae-4646-8f6a-314cdd089cc2\") " pod="openshift-controller-manager/controller-manager-9fc988b76-4gbpp" Oct 14 07:13:35 crc kubenswrapper[4870]: I1014 07:13:35.120172 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4pqm7\" (UniqueName: \"kubernetes.io/projected/5fe046b1-e37b-423b-b9ca-4a8a97b65ca6-kube-api-access-4pqm7\") pod \"route-controller-manager-7cb9656cfd-qwwn4\" (UID: \"5fe046b1-e37b-423b-b9ca-4a8a97b65ca6\") " pod="openshift-route-controller-manager/route-controller-manager-7cb9656cfd-qwwn4" Oct 14 07:13:35 crc kubenswrapper[4870]: I1014 07:13:35.120215 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5fe046b1-e37b-423b-b9ca-4a8a97b65ca6-config\") pod \"route-controller-manager-7cb9656cfd-qwwn4\" (UID: \"5fe046b1-e37b-423b-b9ca-4a8a97b65ca6\") " pod="openshift-route-controller-manager/route-controller-manager-7cb9656cfd-qwwn4" Oct 14 07:13:35 crc kubenswrapper[4870]: I1014 07:13:35.120238 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/241189b0-55ae-4646-8f6a-314cdd089cc2-client-ca\") pod \"controller-manager-9fc988b76-4gbpp\" (UID: \"241189b0-55ae-4646-8f6a-314cdd089cc2\") " pod="openshift-controller-manager/controller-manager-9fc988b76-4gbpp" Oct 14 07:13:35 crc kubenswrapper[4870]: I1014 07:13:35.120267 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/241189b0-55ae-4646-8f6a-314cdd089cc2-config\") pod \"controller-manager-9fc988b76-4gbpp\" (UID: \"241189b0-55ae-4646-8f6a-314cdd089cc2\") " pod="openshift-controller-manager/controller-manager-9fc988b76-4gbpp" Oct 14 07:13:35 crc kubenswrapper[4870]: I1014 07:13:35.120295 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/241189b0-55ae-4646-8f6a-314cdd089cc2-proxy-ca-bundles\") pod \"controller-manager-9fc988b76-4gbpp\" (UID: \"241189b0-55ae-4646-8f6a-314cdd089cc2\") " pod="openshift-controller-manager/controller-manager-9fc988b76-4gbpp" Oct 14 07:13:35 crc kubenswrapper[4870]: I1014 07:13:35.120322 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mnlst\" (UniqueName: \"kubernetes.io/projected/241189b0-55ae-4646-8f6a-314cdd089cc2-kube-api-access-mnlst\") pod \"controller-manager-9fc988b76-4gbpp\" (UID: \"241189b0-55ae-4646-8f6a-314cdd089cc2\") " pod="openshift-controller-manager/controller-manager-9fc988b76-4gbpp" Oct 14 07:13:35 crc kubenswrapper[4870]: I1014 07:13:35.121641 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5fe046b1-e37b-423b-b9ca-4a8a97b65ca6-config\") pod \"route-controller-manager-7cb9656cfd-qwwn4\" (UID: \"5fe046b1-e37b-423b-b9ca-4a8a97b65ca6\") " pod="openshift-route-controller-manager/route-controller-manager-7cb9656cfd-qwwn4" Oct 14 07:13:35 crc kubenswrapper[4870]: I1014 07:13:35.122055 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5fe046b1-e37b-423b-b9ca-4a8a97b65ca6-client-ca\") pod \"route-controller-manager-7cb9656cfd-qwwn4\" (UID: \"5fe046b1-e37b-423b-b9ca-4a8a97b65ca6\") " pod="openshift-route-controller-manager/route-controller-manager-7cb9656cfd-qwwn4" Oct 14 07:13:35 crc kubenswrapper[4870]: I1014 07:13:35.122192 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/241189b0-55ae-4646-8f6a-314cdd089cc2-config\") pod \"controller-manager-9fc988b76-4gbpp\" (UID: \"241189b0-55ae-4646-8f6a-314cdd089cc2\") " pod="openshift-controller-manager/controller-manager-9fc988b76-4gbpp" Oct 14 07:13:35 crc kubenswrapper[4870]: I1014 07:13:35.122302 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/241189b0-55ae-4646-8f6a-314cdd089cc2-proxy-ca-bundles\") pod \"controller-manager-9fc988b76-4gbpp\" (UID: \"241189b0-55ae-4646-8f6a-314cdd089cc2\") " pod="openshift-controller-manager/controller-manager-9fc988b76-4gbpp" Oct 14 07:13:35 crc kubenswrapper[4870]: I1014 07:13:35.122390 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/241189b0-55ae-4646-8f6a-314cdd089cc2-client-ca\") pod \"controller-manager-9fc988b76-4gbpp\" (UID: \"241189b0-55ae-4646-8f6a-314cdd089cc2\") " pod="openshift-controller-manager/controller-manager-9fc988b76-4gbpp" Oct 14 07:13:35 crc kubenswrapper[4870]: I1014 07:13:35.133094 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/241189b0-55ae-4646-8f6a-314cdd089cc2-serving-cert\") pod \"controller-manager-9fc988b76-4gbpp\" (UID: \"241189b0-55ae-4646-8f6a-314cdd089cc2\") " pod="openshift-controller-manager/controller-manager-9fc988b76-4gbpp" Oct 14 07:13:35 crc kubenswrapper[4870]: I1014 07:13:35.139430 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5fe046b1-e37b-423b-b9ca-4a8a97b65ca6-serving-cert\") pod \"route-controller-manager-7cb9656cfd-qwwn4\" (UID: \"5fe046b1-e37b-423b-b9ca-4a8a97b65ca6\") " pod="openshift-route-controller-manager/route-controller-manager-7cb9656cfd-qwwn4" Oct 14 07:13:35 crc kubenswrapper[4870]: I1014 07:13:35.153333 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mnlst\" (UniqueName: \"kubernetes.io/projected/241189b0-55ae-4646-8f6a-314cdd089cc2-kube-api-access-mnlst\") pod \"controller-manager-9fc988b76-4gbpp\" (UID: \"241189b0-55ae-4646-8f6a-314cdd089cc2\") " pod="openshift-controller-manager/controller-manager-9fc988b76-4gbpp" Oct 14 07:13:35 crc kubenswrapper[4870]: I1014 07:13:35.155339 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4pqm7\" (UniqueName: \"kubernetes.io/projected/5fe046b1-e37b-423b-b9ca-4a8a97b65ca6-kube-api-access-4pqm7\") pod \"route-controller-manager-7cb9656cfd-qwwn4\" (UID: \"5fe046b1-e37b-423b-b9ca-4a8a97b65ca6\") " pod="openshift-route-controller-manager/route-controller-manager-7cb9656cfd-qwwn4" Oct 14 07:13:35 crc kubenswrapper[4870]: I1014 07:13:35.188920 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-9fc988b76-4gbpp" Oct 14 07:13:35 crc kubenswrapper[4870]: I1014 07:13:35.203347 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7cb9656cfd-qwwn4" Oct 14 07:13:35 crc kubenswrapper[4870]: I1014 07:13:35.438280 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-9fc988b76-4gbpp"] Oct 14 07:13:35 crc kubenswrapper[4870]: I1014 07:13:35.691052 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7cb9656cfd-qwwn4"] Oct 14 07:13:35 crc kubenswrapper[4870]: I1014 07:13:35.717361 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7cb9656cfd-qwwn4" event={"ID":"5fe046b1-e37b-423b-b9ca-4a8a97b65ca6","Type":"ContainerStarted","Data":"55d81118fa27568abad36c43cc4175421c8b934760a5db8b7d597df25c65f2f4"} Oct 14 07:13:35 crc kubenswrapper[4870]: I1014 07:13:35.719169 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-9fc988b76-4gbpp" event={"ID":"241189b0-55ae-4646-8f6a-314cdd089cc2","Type":"ContainerStarted","Data":"6b0cf9cbc492739848600a2fe6ece293033c2b7a919f67965a4f365df7c9a3a3"} Oct 14 07:13:35 crc kubenswrapper[4870]: I1014 07:13:35.719200 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-9fc988b76-4gbpp" event={"ID":"241189b0-55ae-4646-8f6a-314cdd089cc2","Type":"ContainerStarted","Data":"72e042ae9ce537e830b555b42720cb00f8a22f735465242f787efb48f2721edb"} Oct 14 07:13:35 crc kubenswrapper[4870]: I1014 07:13:35.720591 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-9fc988b76-4gbpp" Oct 14 07:13:35 crc kubenswrapper[4870]: I1014 07:13:35.725693 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-9fc988b76-4gbpp" Oct 14 07:13:35 crc kubenswrapper[4870]: I1014 07:13:35.745356 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-9fc988b76-4gbpp" podStartSLOduration=2.745336285 podStartE2EDuration="2.745336285s" podCreationTimestamp="2025-10-14 07:13:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:13:35.74147351 +0000 UTC m=+751.438833871" watchObservedRunningTime="2025-10-14 07:13:35.745336285 +0000 UTC m=+751.442696656" Oct 14 07:13:36 crc kubenswrapper[4870]: I1014 07:13:36.727332 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7cb9656cfd-qwwn4" event={"ID":"5fe046b1-e37b-423b-b9ca-4a8a97b65ca6","Type":"ContainerStarted","Data":"0539e26bf494f08920c2f80ff6f498c1a5a20a8e118f5f436fbbbbdb1a70559b"} Oct 14 07:13:36 crc kubenswrapper[4870]: I1014 07:13:36.727945 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-7cb9656cfd-qwwn4" Oct 14 07:13:36 crc kubenswrapper[4870]: I1014 07:13:36.734479 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-7cb9656cfd-qwwn4" Oct 14 07:13:36 crc kubenswrapper[4870]: I1014 07:13:36.790678 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-7cb9656cfd-qwwn4" podStartSLOduration=3.790658124 podStartE2EDuration="3.790658124s" podCreationTimestamp="2025-10-14 07:13:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:13:36.758015831 +0000 UTC m=+752.455376272" watchObservedRunningTime="2025-10-14 07:13:36.790658124 +0000 UTC m=+752.488018495" Oct 14 07:13:42 crc kubenswrapper[4870]: I1014 07:13:42.033106 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-d2ngg" Oct 14 07:13:42 crc kubenswrapper[4870]: I1014 07:13:42.034082 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-d2ngg" Oct 14 07:13:42 crc kubenswrapper[4870]: I1014 07:13:42.528937 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-d2ngg"] Oct 14 07:13:42 crc kubenswrapper[4870]: I1014 07:13:42.549850 4870 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 14 07:13:42 crc kubenswrapper[4870]: I1014 07:13:42.591472 4870 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 14 07:13:42 crc kubenswrapper[4870]: I1014 07:13:42.780233 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-d2ngg" event={"ID":"436866a2-a717-4431-aefe-56a515024a72","Type":"ContainerStarted","Data":"5339b2ab9e0f6099418b16ca912932d17c4c18ad8e1a5bbfca20fedc34b8267f"} Oct 14 07:13:44 crc kubenswrapper[4870]: I1014 07:13:44.797431 4870 generic.go:334] "Generic (PLEG): container finished" podID="436866a2-a717-4431-aefe-56a515024a72" containerID="2cbfbf3bceef05399e1e3f329b2560558a7255dc007de0432c52c0cece04e00f" exitCode=0 Oct 14 07:13:44 crc kubenswrapper[4870]: I1014 07:13:44.797517 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-d2ngg" event={"ID":"436866a2-a717-4431-aefe-56a515024a72","Type":"ContainerDied","Data":"2cbfbf3bceef05399e1e3f329b2560558a7255dc007de0432c52c0cece04e00f"} Oct 14 07:13:46 crc kubenswrapper[4870]: I1014 07:13:46.199253 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-d2ngg" Oct 14 07:13:46 crc kubenswrapper[4870]: I1014 07:13:46.384280 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vprhw\" (UniqueName: \"kubernetes.io/projected/436866a2-a717-4431-aefe-56a515024a72-kube-api-access-vprhw\") pod \"436866a2-a717-4431-aefe-56a515024a72\" (UID: \"436866a2-a717-4431-aefe-56a515024a72\") " Oct 14 07:13:46 crc kubenswrapper[4870]: I1014 07:13:46.384420 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/436866a2-a717-4431-aefe-56a515024a72-node-mnt\") pod \"436866a2-a717-4431-aefe-56a515024a72\" (UID: \"436866a2-a717-4431-aefe-56a515024a72\") " Oct 14 07:13:46 crc kubenswrapper[4870]: I1014 07:13:46.384539 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/436866a2-a717-4431-aefe-56a515024a72-node-mnt" (OuterVolumeSpecName: "node-mnt") pod "436866a2-a717-4431-aefe-56a515024a72" (UID: "436866a2-a717-4431-aefe-56a515024a72"). InnerVolumeSpecName "node-mnt". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 14 07:13:46 crc kubenswrapper[4870]: I1014 07:13:46.384631 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/436866a2-a717-4431-aefe-56a515024a72-crc-storage\") pod \"436866a2-a717-4431-aefe-56a515024a72\" (UID: \"436866a2-a717-4431-aefe-56a515024a72\") " Oct 14 07:13:46 crc kubenswrapper[4870]: I1014 07:13:46.386068 4870 reconciler_common.go:293] "Volume detached for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/436866a2-a717-4431-aefe-56a515024a72-node-mnt\") on node \"crc\" DevicePath \"\"" Oct 14 07:13:46 crc kubenswrapper[4870]: I1014 07:13:46.390313 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/436866a2-a717-4431-aefe-56a515024a72-kube-api-access-vprhw" (OuterVolumeSpecName: "kube-api-access-vprhw") pod "436866a2-a717-4431-aefe-56a515024a72" (UID: "436866a2-a717-4431-aefe-56a515024a72"). InnerVolumeSpecName "kube-api-access-vprhw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:13:46 crc kubenswrapper[4870]: I1014 07:13:46.408482 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/436866a2-a717-4431-aefe-56a515024a72-crc-storage" (OuterVolumeSpecName: "crc-storage") pod "436866a2-a717-4431-aefe-56a515024a72" (UID: "436866a2-a717-4431-aefe-56a515024a72"). InnerVolumeSpecName "crc-storage". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:13:46 crc kubenswrapper[4870]: I1014 07:13:46.487969 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vprhw\" (UniqueName: \"kubernetes.io/projected/436866a2-a717-4431-aefe-56a515024a72-kube-api-access-vprhw\") on node \"crc\" DevicePath \"\"" Oct 14 07:13:46 crc kubenswrapper[4870]: I1014 07:13:46.488042 4870 reconciler_common.go:293] "Volume detached for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/436866a2-a717-4431-aefe-56a515024a72-crc-storage\") on node \"crc\" DevicePath \"\"" Oct 14 07:13:46 crc kubenswrapper[4870]: I1014 07:13:46.816420 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-d2ngg" event={"ID":"436866a2-a717-4431-aefe-56a515024a72","Type":"ContainerDied","Data":"5339b2ab9e0f6099418b16ca912932d17c4c18ad8e1a5bbfca20fedc34b8267f"} Oct 14 07:13:46 crc kubenswrapper[4870]: I1014 07:13:46.816525 4870 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5339b2ab9e0f6099418b16ca912932d17c4c18ad8e1a5bbfca20fedc34b8267f" Oct 14 07:13:46 crc kubenswrapper[4870]: I1014 07:13:46.816535 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-d2ngg" Oct 14 07:13:48 crc kubenswrapper[4870]: I1014 07:13:48.978806 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-bxhvl" Oct 14 07:13:53 crc kubenswrapper[4870]: I1014 07:13:53.355459 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cmmng9"] Oct 14 07:13:53 crc kubenswrapper[4870]: E1014 07:13:53.355913 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="436866a2-a717-4431-aefe-56a515024a72" containerName="storage" Oct 14 07:13:53 crc kubenswrapper[4870]: I1014 07:13:53.355928 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="436866a2-a717-4431-aefe-56a515024a72" containerName="storage" Oct 14 07:13:53 crc kubenswrapper[4870]: I1014 07:13:53.356046 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="436866a2-a717-4431-aefe-56a515024a72" containerName="storage" Oct 14 07:13:53 crc kubenswrapper[4870]: I1014 07:13:53.356935 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cmmng9" Oct 14 07:13:53 crc kubenswrapper[4870]: I1014 07:13:53.364339 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Oct 14 07:13:53 crc kubenswrapper[4870]: I1014 07:13:53.373614 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cmmng9"] Oct 14 07:13:53 crc kubenswrapper[4870]: I1014 07:13:53.486529 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/54f5c716-17eb-4c40-88c3-95ec36d3f406-bundle\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cmmng9\" (UID: \"54f5c716-17eb-4c40-88c3-95ec36d3f406\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cmmng9" Oct 14 07:13:53 crc kubenswrapper[4870]: I1014 07:13:53.486876 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p7wg5\" (UniqueName: \"kubernetes.io/projected/54f5c716-17eb-4c40-88c3-95ec36d3f406-kube-api-access-p7wg5\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cmmng9\" (UID: \"54f5c716-17eb-4c40-88c3-95ec36d3f406\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cmmng9" Oct 14 07:13:53 crc kubenswrapper[4870]: I1014 07:13:53.486983 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/54f5c716-17eb-4c40-88c3-95ec36d3f406-util\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cmmng9\" (UID: \"54f5c716-17eb-4c40-88c3-95ec36d3f406\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cmmng9" Oct 14 07:13:53 crc kubenswrapper[4870]: I1014 07:13:53.587936 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p7wg5\" (UniqueName: \"kubernetes.io/projected/54f5c716-17eb-4c40-88c3-95ec36d3f406-kube-api-access-p7wg5\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cmmng9\" (UID: \"54f5c716-17eb-4c40-88c3-95ec36d3f406\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cmmng9" Oct 14 07:13:53 crc kubenswrapper[4870]: I1014 07:13:53.587996 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/54f5c716-17eb-4c40-88c3-95ec36d3f406-util\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cmmng9\" (UID: \"54f5c716-17eb-4c40-88c3-95ec36d3f406\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cmmng9" Oct 14 07:13:53 crc kubenswrapper[4870]: I1014 07:13:53.588041 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/54f5c716-17eb-4c40-88c3-95ec36d3f406-bundle\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cmmng9\" (UID: \"54f5c716-17eb-4c40-88c3-95ec36d3f406\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cmmng9" Oct 14 07:13:53 crc kubenswrapper[4870]: I1014 07:13:53.588590 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/54f5c716-17eb-4c40-88c3-95ec36d3f406-bundle\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cmmng9\" (UID: \"54f5c716-17eb-4c40-88c3-95ec36d3f406\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cmmng9" Oct 14 07:13:53 crc kubenswrapper[4870]: I1014 07:13:53.588887 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/54f5c716-17eb-4c40-88c3-95ec36d3f406-util\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cmmng9\" (UID: \"54f5c716-17eb-4c40-88c3-95ec36d3f406\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cmmng9" Oct 14 07:13:53 crc kubenswrapper[4870]: I1014 07:13:53.619391 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p7wg5\" (UniqueName: \"kubernetes.io/projected/54f5c716-17eb-4c40-88c3-95ec36d3f406-kube-api-access-p7wg5\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cmmng9\" (UID: \"54f5c716-17eb-4c40-88c3-95ec36d3f406\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cmmng9" Oct 14 07:13:53 crc kubenswrapper[4870]: I1014 07:13:53.674834 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cmmng9" Oct 14 07:13:54 crc kubenswrapper[4870]: I1014 07:13:54.196971 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cmmng9"] Oct 14 07:13:54 crc kubenswrapper[4870]: W1014 07:13:54.211663 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod54f5c716_17eb_4c40_88c3_95ec36d3f406.slice/crio-1e80fd781ffcc9b8233ec2f8b5479cc024e5afea85217b2cd4e7157acc997e5b WatchSource:0}: Error finding container 1e80fd781ffcc9b8233ec2f8b5479cc024e5afea85217b2cd4e7157acc997e5b: Status 404 returned error can't find the container with id 1e80fd781ffcc9b8233ec2f8b5479cc024e5afea85217b2cd4e7157acc997e5b Oct 14 07:13:54 crc kubenswrapper[4870]: I1014 07:13:54.873345 4870 generic.go:334] "Generic (PLEG): container finished" podID="54f5c716-17eb-4c40-88c3-95ec36d3f406" containerID="fc296b09ac628831d7307395f41a8ed772b0f1ff816bbe4cbeb82a0bb3922683" exitCode=0 Oct 14 07:13:54 crc kubenswrapper[4870]: I1014 07:13:54.873387 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cmmng9" event={"ID":"54f5c716-17eb-4c40-88c3-95ec36d3f406","Type":"ContainerDied","Data":"fc296b09ac628831d7307395f41a8ed772b0f1ff816bbe4cbeb82a0bb3922683"} Oct 14 07:13:54 crc kubenswrapper[4870]: I1014 07:13:54.873411 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cmmng9" event={"ID":"54f5c716-17eb-4c40-88c3-95ec36d3f406","Type":"ContainerStarted","Data":"1e80fd781ffcc9b8233ec2f8b5479cc024e5afea85217b2cd4e7157acc997e5b"} Oct 14 07:13:55 crc kubenswrapper[4870]: I1014 07:13:55.539534 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-xwpt9"] Oct 14 07:13:55 crc kubenswrapper[4870]: I1014 07:13:55.542828 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xwpt9" Oct 14 07:13:55 crc kubenswrapper[4870]: I1014 07:13:55.550663 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-xwpt9"] Oct 14 07:13:55 crc kubenswrapper[4870]: I1014 07:13:55.712679 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ljbtr\" (UniqueName: \"kubernetes.io/projected/aeca7d4f-c3c4-404c-9bdf-a1ed0951ada9-kube-api-access-ljbtr\") pod \"redhat-operators-xwpt9\" (UID: \"aeca7d4f-c3c4-404c-9bdf-a1ed0951ada9\") " pod="openshift-marketplace/redhat-operators-xwpt9" Oct 14 07:13:55 crc kubenswrapper[4870]: I1014 07:13:55.712793 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aeca7d4f-c3c4-404c-9bdf-a1ed0951ada9-catalog-content\") pod \"redhat-operators-xwpt9\" (UID: \"aeca7d4f-c3c4-404c-9bdf-a1ed0951ada9\") " pod="openshift-marketplace/redhat-operators-xwpt9" Oct 14 07:13:55 crc kubenswrapper[4870]: I1014 07:13:55.712951 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aeca7d4f-c3c4-404c-9bdf-a1ed0951ada9-utilities\") pod \"redhat-operators-xwpt9\" (UID: \"aeca7d4f-c3c4-404c-9bdf-a1ed0951ada9\") " pod="openshift-marketplace/redhat-operators-xwpt9" Oct 14 07:13:55 crc kubenswrapper[4870]: I1014 07:13:55.813963 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ljbtr\" (UniqueName: \"kubernetes.io/projected/aeca7d4f-c3c4-404c-9bdf-a1ed0951ada9-kube-api-access-ljbtr\") pod \"redhat-operators-xwpt9\" (UID: \"aeca7d4f-c3c4-404c-9bdf-a1ed0951ada9\") " pod="openshift-marketplace/redhat-operators-xwpt9" Oct 14 07:13:55 crc kubenswrapper[4870]: I1014 07:13:55.814036 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aeca7d4f-c3c4-404c-9bdf-a1ed0951ada9-catalog-content\") pod \"redhat-operators-xwpt9\" (UID: \"aeca7d4f-c3c4-404c-9bdf-a1ed0951ada9\") " pod="openshift-marketplace/redhat-operators-xwpt9" Oct 14 07:13:55 crc kubenswrapper[4870]: I1014 07:13:55.814101 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aeca7d4f-c3c4-404c-9bdf-a1ed0951ada9-utilities\") pod \"redhat-operators-xwpt9\" (UID: \"aeca7d4f-c3c4-404c-9bdf-a1ed0951ada9\") " pod="openshift-marketplace/redhat-operators-xwpt9" Oct 14 07:13:55 crc kubenswrapper[4870]: I1014 07:13:55.814625 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aeca7d4f-c3c4-404c-9bdf-a1ed0951ada9-utilities\") pod \"redhat-operators-xwpt9\" (UID: \"aeca7d4f-c3c4-404c-9bdf-a1ed0951ada9\") " pod="openshift-marketplace/redhat-operators-xwpt9" Oct 14 07:13:55 crc kubenswrapper[4870]: I1014 07:13:55.814868 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aeca7d4f-c3c4-404c-9bdf-a1ed0951ada9-catalog-content\") pod \"redhat-operators-xwpt9\" (UID: \"aeca7d4f-c3c4-404c-9bdf-a1ed0951ada9\") " pod="openshift-marketplace/redhat-operators-xwpt9" Oct 14 07:13:55 crc kubenswrapper[4870]: I1014 07:13:55.832419 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ljbtr\" (UniqueName: \"kubernetes.io/projected/aeca7d4f-c3c4-404c-9bdf-a1ed0951ada9-kube-api-access-ljbtr\") pod \"redhat-operators-xwpt9\" (UID: \"aeca7d4f-c3c4-404c-9bdf-a1ed0951ada9\") " pod="openshift-marketplace/redhat-operators-xwpt9" Oct 14 07:13:55 crc kubenswrapper[4870]: I1014 07:13:55.870335 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xwpt9" Oct 14 07:13:56 crc kubenswrapper[4870]: I1014 07:13:56.261209 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-xwpt9"] Oct 14 07:13:56 crc kubenswrapper[4870]: W1014 07:13:56.296655 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaeca7d4f_c3c4_404c_9bdf_a1ed0951ada9.slice/crio-58b0cd7c623d8eeebb8aba3cfe7b5ffddd5adaa6cea4645a51617fd6556b774c WatchSource:0}: Error finding container 58b0cd7c623d8eeebb8aba3cfe7b5ffddd5adaa6cea4645a51617fd6556b774c: Status 404 returned error can't find the container with id 58b0cd7c623d8eeebb8aba3cfe7b5ffddd5adaa6cea4645a51617fd6556b774c Oct 14 07:13:56 crc kubenswrapper[4870]: I1014 07:13:56.887569 4870 generic.go:334] "Generic (PLEG): container finished" podID="aeca7d4f-c3c4-404c-9bdf-a1ed0951ada9" containerID="899754855bb11ba4691fc9e2d644607368e1d1d256d678f6715ec75365cb503f" exitCode=0 Oct 14 07:13:56 crc kubenswrapper[4870]: I1014 07:13:56.887680 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xwpt9" event={"ID":"aeca7d4f-c3c4-404c-9bdf-a1ed0951ada9","Type":"ContainerDied","Data":"899754855bb11ba4691fc9e2d644607368e1d1d256d678f6715ec75365cb503f"} Oct 14 07:13:56 crc kubenswrapper[4870]: I1014 07:13:56.887723 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xwpt9" event={"ID":"aeca7d4f-c3c4-404c-9bdf-a1ed0951ada9","Type":"ContainerStarted","Data":"58b0cd7c623d8eeebb8aba3cfe7b5ffddd5adaa6cea4645a51617fd6556b774c"} Oct 14 07:13:56 crc kubenswrapper[4870]: I1014 07:13:56.891665 4870 generic.go:334] "Generic (PLEG): container finished" podID="54f5c716-17eb-4c40-88c3-95ec36d3f406" containerID="9d540cc5d81f314a5b16fa1745e0654cd914fe5d9badca7d7853e61ddb93f53f" exitCode=0 Oct 14 07:13:56 crc kubenswrapper[4870]: I1014 07:13:56.891724 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cmmng9" event={"ID":"54f5c716-17eb-4c40-88c3-95ec36d3f406","Type":"ContainerDied","Data":"9d540cc5d81f314a5b16fa1745e0654cd914fe5d9badca7d7853e61ddb93f53f"} Oct 14 07:13:57 crc kubenswrapper[4870]: I1014 07:13:57.901939 4870 generic.go:334] "Generic (PLEG): container finished" podID="54f5c716-17eb-4c40-88c3-95ec36d3f406" containerID="607eccd4c55abd464bc10d48f3c31c3875cb7c607f8224dec618c70bc2146380" exitCode=0 Oct 14 07:13:57 crc kubenswrapper[4870]: I1014 07:13:57.902013 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cmmng9" event={"ID":"54f5c716-17eb-4c40-88c3-95ec36d3f406","Type":"ContainerDied","Data":"607eccd4c55abd464bc10d48f3c31c3875cb7c607f8224dec618c70bc2146380"} Oct 14 07:13:58 crc kubenswrapper[4870]: I1014 07:13:58.933813 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-7468q"] Oct 14 07:13:58 crc kubenswrapper[4870]: I1014 07:13:58.935937 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7468q" Oct 14 07:13:58 crc kubenswrapper[4870]: I1014 07:13:58.941351 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-7468q"] Oct 14 07:13:59 crc kubenswrapper[4870]: I1014 07:13:59.053054 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9574cf23-fd49-4791-ad7d-48e2c4bb69a9-utilities\") pod \"community-operators-7468q\" (UID: \"9574cf23-fd49-4791-ad7d-48e2c4bb69a9\") " pod="openshift-marketplace/community-operators-7468q" Oct 14 07:13:59 crc kubenswrapper[4870]: I1014 07:13:59.053399 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tcdfz\" (UniqueName: \"kubernetes.io/projected/9574cf23-fd49-4791-ad7d-48e2c4bb69a9-kube-api-access-tcdfz\") pod \"community-operators-7468q\" (UID: \"9574cf23-fd49-4791-ad7d-48e2c4bb69a9\") " pod="openshift-marketplace/community-operators-7468q" Oct 14 07:13:59 crc kubenswrapper[4870]: I1014 07:13:59.053513 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9574cf23-fd49-4791-ad7d-48e2c4bb69a9-catalog-content\") pod \"community-operators-7468q\" (UID: \"9574cf23-fd49-4791-ad7d-48e2c4bb69a9\") " pod="openshift-marketplace/community-operators-7468q" Oct 14 07:13:59 crc kubenswrapper[4870]: I1014 07:13:59.154830 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9574cf23-fd49-4791-ad7d-48e2c4bb69a9-catalog-content\") pod \"community-operators-7468q\" (UID: \"9574cf23-fd49-4791-ad7d-48e2c4bb69a9\") " pod="openshift-marketplace/community-operators-7468q" Oct 14 07:13:59 crc kubenswrapper[4870]: I1014 07:13:59.154890 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9574cf23-fd49-4791-ad7d-48e2c4bb69a9-utilities\") pod \"community-operators-7468q\" (UID: \"9574cf23-fd49-4791-ad7d-48e2c4bb69a9\") " pod="openshift-marketplace/community-operators-7468q" Oct 14 07:13:59 crc kubenswrapper[4870]: I1014 07:13:59.154939 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tcdfz\" (UniqueName: \"kubernetes.io/projected/9574cf23-fd49-4791-ad7d-48e2c4bb69a9-kube-api-access-tcdfz\") pod \"community-operators-7468q\" (UID: \"9574cf23-fd49-4791-ad7d-48e2c4bb69a9\") " pod="openshift-marketplace/community-operators-7468q" Oct 14 07:13:59 crc kubenswrapper[4870]: I1014 07:13:59.155471 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9574cf23-fd49-4791-ad7d-48e2c4bb69a9-catalog-content\") pod \"community-operators-7468q\" (UID: \"9574cf23-fd49-4791-ad7d-48e2c4bb69a9\") " pod="openshift-marketplace/community-operators-7468q" Oct 14 07:13:59 crc kubenswrapper[4870]: I1014 07:13:59.155512 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9574cf23-fd49-4791-ad7d-48e2c4bb69a9-utilities\") pod \"community-operators-7468q\" (UID: \"9574cf23-fd49-4791-ad7d-48e2c4bb69a9\") " pod="openshift-marketplace/community-operators-7468q" Oct 14 07:13:59 crc kubenswrapper[4870]: I1014 07:13:59.174790 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tcdfz\" (UniqueName: \"kubernetes.io/projected/9574cf23-fd49-4791-ad7d-48e2c4bb69a9-kube-api-access-tcdfz\") pod \"community-operators-7468q\" (UID: \"9574cf23-fd49-4791-ad7d-48e2c4bb69a9\") " pod="openshift-marketplace/community-operators-7468q" Oct 14 07:13:59 crc kubenswrapper[4870]: I1014 07:13:59.261769 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cmmng9" Oct 14 07:13:59 crc kubenswrapper[4870]: I1014 07:13:59.261903 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7468q" Oct 14 07:13:59 crc kubenswrapper[4870]: I1014 07:13:59.456990 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/54f5c716-17eb-4c40-88c3-95ec36d3f406-util\") pod \"54f5c716-17eb-4c40-88c3-95ec36d3f406\" (UID: \"54f5c716-17eb-4c40-88c3-95ec36d3f406\") " Oct 14 07:13:59 crc kubenswrapper[4870]: I1014 07:13:59.457251 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/54f5c716-17eb-4c40-88c3-95ec36d3f406-bundle\") pod \"54f5c716-17eb-4c40-88c3-95ec36d3f406\" (UID: \"54f5c716-17eb-4c40-88c3-95ec36d3f406\") " Oct 14 07:13:59 crc kubenswrapper[4870]: I1014 07:13:59.457316 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p7wg5\" (UniqueName: \"kubernetes.io/projected/54f5c716-17eb-4c40-88c3-95ec36d3f406-kube-api-access-p7wg5\") pod \"54f5c716-17eb-4c40-88c3-95ec36d3f406\" (UID: \"54f5c716-17eb-4c40-88c3-95ec36d3f406\") " Oct 14 07:13:59 crc kubenswrapper[4870]: I1014 07:13:59.457953 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/54f5c716-17eb-4c40-88c3-95ec36d3f406-bundle" (OuterVolumeSpecName: "bundle") pod "54f5c716-17eb-4c40-88c3-95ec36d3f406" (UID: "54f5c716-17eb-4c40-88c3-95ec36d3f406"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:13:59 crc kubenswrapper[4870]: I1014 07:13:59.462403 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/54f5c716-17eb-4c40-88c3-95ec36d3f406-kube-api-access-p7wg5" (OuterVolumeSpecName: "kube-api-access-p7wg5") pod "54f5c716-17eb-4c40-88c3-95ec36d3f406" (UID: "54f5c716-17eb-4c40-88c3-95ec36d3f406"). InnerVolumeSpecName "kube-api-access-p7wg5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:13:59 crc kubenswrapper[4870]: I1014 07:13:59.472770 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/54f5c716-17eb-4c40-88c3-95ec36d3f406-util" (OuterVolumeSpecName: "util") pod "54f5c716-17eb-4c40-88c3-95ec36d3f406" (UID: "54f5c716-17eb-4c40-88c3-95ec36d3f406"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:13:59 crc kubenswrapper[4870]: I1014 07:13:59.558656 4870 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/54f5c716-17eb-4c40-88c3-95ec36d3f406-util\") on node \"crc\" DevicePath \"\"" Oct 14 07:13:59 crc kubenswrapper[4870]: I1014 07:13:59.558697 4870 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/54f5c716-17eb-4c40-88c3-95ec36d3f406-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 07:13:59 crc kubenswrapper[4870]: I1014 07:13:59.558710 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p7wg5\" (UniqueName: \"kubernetes.io/projected/54f5c716-17eb-4c40-88c3-95ec36d3f406-kube-api-access-p7wg5\") on node \"crc\" DevicePath \"\"" Oct 14 07:13:59 crc kubenswrapper[4870]: I1014 07:13:59.698872 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-7468q"] Oct 14 07:13:59 crc kubenswrapper[4870]: I1014 07:13:59.915190 4870 generic.go:334] "Generic (PLEG): container finished" podID="9574cf23-fd49-4791-ad7d-48e2c4bb69a9" containerID="2fae72da1e50d3e3b2f7f871ec0fd4a5e686be5f1adf9fcfb8676861027afc02" exitCode=0 Oct 14 07:13:59 crc kubenswrapper[4870]: I1014 07:13:59.915269 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7468q" event={"ID":"9574cf23-fd49-4791-ad7d-48e2c4bb69a9","Type":"ContainerDied","Data":"2fae72da1e50d3e3b2f7f871ec0fd4a5e686be5f1adf9fcfb8676861027afc02"} Oct 14 07:13:59 crc kubenswrapper[4870]: I1014 07:13:59.915302 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7468q" event={"ID":"9574cf23-fd49-4791-ad7d-48e2c4bb69a9","Type":"ContainerStarted","Data":"10fe2bdeb01e1cdb40323886ee0e14aff656bacac2afbc659474c1ac27903234"} Oct 14 07:13:59 crc kubenswrapper[4870]: I1014 07:13:59.920520 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cmmng9" event={"ID":"54f5c716-17eb-4c40-88c3-95ec36d3f406","Type":"ContainerDied","Data":"1e80fd781ffcc9b8233ec2f8b5479cc024e5afea85217b2cd4e7157acc997e5b"} Oct 14 07:13:59 crc kubenswrapper[4870]: I1014 07:13:59.920548 4870 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1e80fd781ffcc9b8233ec2f8b5479cc024e5afea85217b2cd4e7157acc997e5b" Oct 14 07:13:59 crc kubenswrapper[4870]: I1014 07:13:59.920625 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cmmng9" Oct 14 07:14:00 crc kubenswrapper[4870]: I1014 07:14:00.929988 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7468q" event={"ID":"9574cf23-fd49-4791-ad7d-48e2c4bb69a9","Type":"ContainerStarted","Data":"bc59f4bdba65ece0ae32bec288269ba68f06cb2c97842be773a606c678e8e95a"} Oct 14 07:14:01 crc kubenswrapper[4870]: I1014 07:14:01.937656 4870 generic.go:334] "Generic (PLEG): container finished" podID="9574cf23-fd49-4791-ad7d-48e2c4bb69a9" containerID="bc59f4bdba65ece0ae32bec288269ba68f06cb2c97842be773a606c678e8e95a" exitCode=0 Oct 14 07:14:01 crc kubenswrapper[4870]: I1014 07:14:01.937704 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7468q" event={"ID":"9574cf23-fd49-4791-ad7d-48e2c4bb69a9","Type":"ContainerDied","Data":"bc59f4bdba65ece0ae32bec288269ba68f06cb2c97842be773a606c678e8e95a"} Oct 14 07:14:02 crc kubenswrapper[4870]: I1014 07:14:02.539383 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-x7gsd"] Oct 14 07:14:02 crc kubenswrapper[4870]: E1014 07:14:02.539863 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54f5c716-17eb-4c40-88c3-95ec36d3f406" containerName="extract" Oct 14 07:14:02 crc kubenswrapper[4870]: I1014 07:14:02.539888 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="54f5c716-17eb-4c40-88c3-95ec36d3f406" containerName="extract" Oct 14 07:14:02 crc kubenswrapper[4870]: E1014 07:14:02.539905 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54f5c716-17eb-4c40-88c3-95ec36d3f406" containerName="util" Oct 14 07:14:02 crc kubenswrapper[4870]: I1014 07:14:02.539917 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="54f5c716-17eb-4c40-88c3-95ec36d3f406" containerName="util" Oct 14 07:14:02 crc kubenswrapper[4870]: E1014 07:14:02.539945 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54f5c716-17eb-4c40-88c3-95ec36d3f406" containerName="pull" Oct 14 07:14:02 crc kubenswrapper[4870]: I1014 07:14:02.539958 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="54f5c716-17eb-4c40-88c3-95ec36d3f406" containerName="pull" Oct 14 07:14:02 crc kubenswrapper[4870]: I1014 07:14:02.540132 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="54f5c716-17eb-4c40-88c3-95ec36d3f406" containerName="extract" Oct 14 07:14:02 crc kubenswrapper[4870]: I1014 07:14:02.559274 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-x7gsd" Oct 14 07:14:02 crc kubenswrapper[4870]: I1014 07:14:02.560952 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-x7gsd"] Oct 14 07:14:02 crc kubenswrapper[4870]: I1014 07:14:02.704090 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/68ce28b7-83f9-4ba8-8a95-b8995f5151ca-catalog-content\") pod \"certified-operators-x7gsd\" (UID: \"68ce28b7-83f9-4ba8-8a95-b8995f5151ca\") " pod="openshift-marketplace/certified-operators-x7gsd" Oct 14 07:14:02 crc kubenswrapper[4870]: I1014 07:14:02.704380 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/68ce28b7-83f9-4ba8-8a95-b8995f5151ca-utilities\") pod \"certified-operators-x7gsd\" (UID: \"68ce28b7-83f9-4ba8-8a95-b8995f5151ca\") " pod="openshift-marketplace/certified-operators-x7gsd" Oct 14 07:14:02 crc kubenswrapper[4870]: I1014 07:14:02.704490 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jt2v7\" (UniqueName: \"kubernetes.io/projected/68ce28b7-83f9-4ba8-8a95-b8995f5151ca-kube-api-access-jt2v7\") pod \"certified-operators-x7gsd\" (UID: \"68ce28b7-83f9-4ba8-8a95-b8995f5151ca\") " pod="openshift-marketplace/certified-operators-x7gsd" Oct 14 07:14:02 crc kubenswrapper[4870]: I1014 07:14:02.809358 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/68ce28b7-83f9-4ba8-8a95-b8995f5151ca-catalog-content\") pod \"certified-operators-x7gsd\" (UID: \"68ce28b7-83f9-4ba8-8a95-b8995f5151ca\") " pod="openshift-marketplace/certified-operators-x7gsd" Oct 14 07:14:02 crc kubenswrapper[4870]: I1014 07:14:02.809425 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/68ce28b7-83f9-4ba8-8a95-b8995f5151ca-utilities\") pod \"certified-operators-x7gsd\" (UID: \"68ce28b7-83f9-4ba8-8a95-b8995f5151ca\") " pod="openshift-marketplace/certified-operators-x7gsd" Oct 14 07:14:02 crc kubenswrapper[4870]: I1014 07:14:02.809511 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jt2v7\" (UniqueName: \"kubernetes.io/projected/68ce28b7-83f9-4ba8-8a95-b8995f5151ca-kube-api-access-jt2v7\") pod \"certified-operators-x7gsd\" (UID: \"68ce28b7-83f9-4ba8-8a95-b8995f5151ca\") " pod="openshift-marketplace/certified-operators-x7gsd" Oct 14 07:14:02 crc kubenswrapper[4870]: I1014 07:14:02.809967 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/68ce28b7-83f9-4ba8-8a95-b8995f5151ca-catalog-content\") pod \"certified-operators-x7gsd\" (UID: \"68ce28b7-83f9-4ba8-8a95-b8995f5151ca\") " pod="openshift-marketplace/certified-operators-x7gsd" Oct 14 07:14:02 crc kubenswrapper[4870]: I1014 07:14:02.810079 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/68ce28b7-83f9-4ba8-8a95-b8995f5151ca-utilities\") pod \"certified-operators-x7gsd\" (UID: \"68ce28b7-83f9-4ba8-8a95-b8995f5151ca\") " pod="openshift-marketplace/certified-operators-x7gsd" Oct 14 07:14:02 crc kubenswrapper[4870]: I1014 07:14:02.832984 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jt2v7\" (UniqueName: \"kubernetes.io/projected/68ce28b7-83f9-4ba8-8a95-b8995f5151ca-kube-api-access-jt2v7\") pod \"certified-operators-x7gsd\" (UID: \"68ce28b7-83f9-4ba8-8a95-b8995f5151ca\") " pod="openshift-marketplace/certified-operators-x7gsd" Oct 14 07:14:02 crc kubenswrapper[4870]: I1014 07:14:02.879290 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-x7gsd" Oct 14 07:14:04 crc kubenswrapper[4870]: I1014 07:14:04.017775 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-858ddd8f98-q77sx"] Oct 14 07:14:04 crc kubenswrapper[4870]: I1014 07:14:04.018608 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-858ddd8f98-q77sx" Oct 14 07:14:04 crc kubenswrapper[4870]: I1014 07:14:04.020500 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-6lvg2" Oct 14 07:14:04 crc kubenswrapper[4870]: I1014 07:14:04.020587 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Oct 14 07:14:04 crc kubenswrapper[4870]: I1014 07:14:04.020619 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Oct 14 07:14:04 crc kubenswrapper[4870]: I1014 07:14:04.031871 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-858ddd8f98-q77sx"] Oct 14 07:14:04 crc kubenswrapper[4870]: I1014 07:14:04.129292 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2dpl9\" (UniqueName: \"kubernetes.io/projected/98eac063-80b2-46f7-8096-d384fc10d632-kube-api-access-2dpl9\") pod \"nmstate-operator-858ddd8f98-q77sx\" (UID: \"98eac063-80b2-46f7-8096-d384fc10d632\") " pod="openshift-nmstate/nmstate-operator-858ddd8f98-q77sx" Oct 14 07:14:04 crc kubenswrapper[4870]: I1014 07:14:04.230642 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2dpl9\" (UniqueName: \"kubernetes.io/projected/98eac063-80b2-46f7-8096-d384fc10d632-kube-api-access-2dpl9\") pod \"nmstate-operator-858ddd8f98-q77sx\" (UID: \"98eac063-80b2-46f7-8096-d384fc10d632\") " pod="openshift-nmstate/nmstate-operator-858ddd8f98-q77sx" Oct 14 07:14:04 crc kubenswrapper[4870]: I1014 07:14:04.246391 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2dpl9\" (UniqueName: \"kubernetes.io/projected/98eac063-80b2-46f7-8096-d384fc10d632-kube-api-access-2dpl9\") pod \"nmstate-operator-858ddd8f98-q77sx\" (UID: \"98eac063-80b2-46f7-8096-d384fc10d632\") " pod="openshift-nmstate/nmstate-operator-858ddd8f98-q77sx" Oct 14 07:14:04 crc kubenswrapper[4870]: I1014 07:14:04.333867 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-858ddd8f98-q77sx" Oct 14 07:14:05 crc kubenswrapper[4870]: I1014 07:14:05.339813 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-858ddd8f98-q77sx"] Oct 14 07:14:05 crc kubenswrapper[4870]: W1014 07:14:05.443507 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod98eac063_80b2_46f7_8096_d384fc10d632.slice/crio-f72ffdab49d4f93a8208d4ea7bdbf14b5aa42f688dbcf96ee368a1ef978da5e5 WatchSource:0}: Error finding container f72ffdab49d4f93a8208d4ea7bdbf14b5aa42f688dbcf96ee368a1ef978da5e5: Status 404 returned error can't find the container with id f72ffdab49d4f93a8208d4ea7bdbf14b5aa42f688dbcf96ee368a1ef978da5e5 Oct 14 07:14:05 crc kubenswrapper[4870]: I1014 07:14:05.454926 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-x7gsd"] Oct 14 07:14:05 crc kubenswrapper[4870]: W1014 07:14:05.478772 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod68ce28b7_83f9_4ba8_8a95_b8995f5151ca.slice/crio-21b97618051e13bd0ecdda8c2bc413790c7e0fddade0a14d133a421e5253656a WatchSource:0}: Error finding container 21b97618051e13bd0ecdda8c2bc413790c7e0fddade0a14d133a421e5253656a: Status 404 returned error can't find the container with id 21b97618051e13bd0ecdda8c2bc413790c7e0fddade0a14d133a421e5253656a Oct 14 07:14:05 crc kubenswrapper[4870]: I1014 07:14:05.966825 4870 generic.go:334] "Generic (PLEG): container finished" podID="68ce28b7-83f9-4ba8-8a95-b8995f5151ca" containerID="fab4b4e6086d944653e286c195f0877c8457965eced0604f885e8a105ed8ace0" exitCode=0 Oct 14 07:14:05 crc kubenswrapper[4870]: I1014 07:14:05.967134 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-x7gsd" event={"ID":"68ce28b7-83f9-4ba8-8a95-b8995f5151ca","Type":"ContainerDied","Data":"fab4b4e6086d944653e286c195f0877c8457965eced0604f885e8a105ed8ace0"} Oct 14 07:14:05 crc kubenswrapper[4870]: I1014 07:14:05.967213 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-x7gsd" event={"ID":"68ce28b7-83f9-4ba8-8a95-b8995f5151ca","Type":"ContainerStarted","Data":"21b97618051e13bd0ecdda8c2bc413790c7e0fddade0a14d133a421e5253656a"} Oct 14 07:14:05 crc kubenswrapper[4870]: I1014 07:14:05.970537 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xwpt9" event={"ID":"aeca7d4f-c3c4-404c-9bdf-a1ed0951ada9","Type":"ContainerStarted","Data":"df1e991ca91d6a83ec52845be3ea04155de224c50912bb7dcbb13d8d2c334147"} Oct 14 07:14:05 crc kubenswrapper[4870]: I1014 07:14:05.972608 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-858ddd8f98-q77sx" event={"ID":"98eac063-80b2-46f7-8096-d384fc10d632","Type":"ContainerStarted","Data":"f72ffdab49d4f93a8208d4ea7bdbf14b5aa42f688dbcf96ee368a1ef978da5e5"} Oct 14 07:14:06 crc kubenswrapper[4870]: I1014 07:14:06.980718 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-x7gsd" event={"ID":"68ce28b7-83f9-4ba8-8a95-b8995f5151ca","Type":"ContainerStarted","Data":"5d60cde3e1e8e54de1128bea6a214678baee822141909f171431934a2e45082d"} Oct 14 07:14:06 crc kubenswrapper[4870]: I1014 07:14:06.983667 4870 generic.go:334] "Generic (PLEG): container finished" podID="aeca7d4f-c3c4-404c-9bdf-a1ed0951ada9" containerID="df1e991ca91d6a83ec52845be3ea04155de224c50912bb7dcbb13d8d2c334147" exitCode=0 Oct 14 07:14:06 crc kubenswrapper[4870]: I1014 07:14:06.984138 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xwpt9" event={"ID":"aeca7d4f-c3c4-404c-9bdf-a1ed0951ada9","Type":"ContainerDied","Data":"df1e991ca91d6a83ec52845be3ea04155de224c50912bb7dcbb13d8d2c334147"} Oct 14 07:14:06 crc kubenswrapper[4870]: I1014 07:14:06.988135 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7468q" event={"ID":"9574cf23-fd49-4791-ad7d-48e2c4bb69a9","Type":"ContainerStarted","Data":"9d6af32291b8bc4830801d4959abf0ee0f4472e274bb56a09ff7fc991b7fe965"} Oct 14 07:14:07 crc kubenswrapper[4870]: I1014 07:14:07.039947 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-7468q" podStartSLOduration=3.147945266 podStartE2EDuration="9.039930172s" podCreationTimestamp="2025-10-14 07:13:58 +0000 UTC" firstStartedPulling="2025-10-14 07:13:59.917734583 +0000 UTC m=+775.615094954" lastFinishedPulling="2025-10-14 07:14:05.809719489 +0000 UTC m=+781.507079860" observedRunningTime="2025-10-14 07:14:07.03748118 +0000 UTC m=+782.734841571" watchObservedRunningTime="2025-10-14 07:14:07.039930172 +0000 UTC m=+782.737290553" Oct 14 07:14:07 crc kubenswrapper[4870]: I1014 07:14:07.996419 4870 generic.go:334] "Generic (PLEG): container finished" podID="68ce28b7-83f9-4ba8-8a95-b8995f5151ca" containerID="5d60cde3e1e8e54de1128bea6a214678baee822141909f171431934a2e45082d" exitCode=0 Oct 14 07:14:07 crc kubenswrapper[4870]: I1014 07:14:07.996476 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-x7gsd" event={"ID":"68ce28b7-83f9-4ba8-8a95-b8995f5151ca","Type":"ContainerDied","Data":"5d60cde3e1e8e54de1128bea6a214678baee822141909f171431934a2e45082d"} Oct 14 07:14:08 crc kubenswrapper[4870]: I1014 07:14:08.001197 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xwpt9" event={"ID":"aeca7d4f-c3c4-404c-9bdf-a1ed0951ada9","Type":"ContainerStarted","Data":"f203f942444d205a40086ae2434be7198113f9d3ff89f9893f0a349715fba6b1"} Oct 14 07:14:08 crc kubenswrapper[4870]: I1014 07:14:08.043438 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-xwpt9" podStartSLOduration=2.536387039 podStartE2EDuration="13.043419451s" podCreationTimestamp="2025-10-14 07:13:55 +0000 UTC" firstStartedPulling="2025-10-14 07:13:56.890376027 +0000 UTC m=+772.587736438" lastFinishedPulling="2025-10-14 07:14:07.397408479 +0000 UTC m=+783.094768850" observedRunningTime="2025-10-14 07:14:08.040064546 +0000 UTC m=+783.737424917" watchObservedRunningTime="2025-10-14 07:14:08.043419451 +0000 UTC m=+783.740779822" Oct 14 07:14:09 crc kubenswrapper[4870]: I1014 07:14:09.006954 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-858ddd8f98-q77sx" event={"ID":"98eac063-80b2-46f7-8096-d384fc10d632","Type":"ContainerStarted","Data":"c150107a03b3ff188a2a028dedd09c915592a681fcdbaf0b6d6b812c8546bf1c"} Oct 14 07:14:09 crc kubenswrapper[4870]: I1014 07:14:09.009403 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-x7gsd" event={"ID":"68ce28b7-83f9-4ba8-8a95-b8995f5151ca","Type":"ContainerStarted","Data":"c9b11a2b2df54d1e96336de480fa60aad3cca93be98e3fd4ed2110b8ffb36483"} Oct 14 07:14:09 crc kubenswrapper[4870]: I1014 07:14:09.032008 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-858ddd8f98-q77sx" podStartSLOduration=2.565186047 podStartE2EDuration="5.031989238s" podCreationTimestamp="2025-10-14 07:14:04 +0000 UTC" firstStartedPulling="2025-10-14 07:14:05.446764371 +0000 UTC m=+781.144124742" lastFinishedPulling="2025-10-14 07:14:07.913567562 +0000 UTC m=+783.610927933" observedRunningTime="2025-10-14 07:14:09.027800761 +0000 UTC m=+784.725161142" watchObservedRunningTime="2025-10-14 07:14:09.031989238 +0000 UTC m=+784.729349609" Oct 14 07:14:09 crc kubenswrapper[4870]: I1014 07:14:09.262355 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-7468q" Oct 14 07:14:09 crc kubenswrapper[4870]: I1014 07:14:09.263218 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-7468q" Oct 14 07:14:09 crc kubenswrapper[4870]: I1014 07:14:09.304382 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-7468q" Oct 14 07:14:09 crc kubenswrapper[4870]: I1014 07:14:09.333529 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-x7gsd" podStartSLOduration=4.880066256 podStartE2EDuration="7.333430853s" podCreationTimestamp="2025-10-14 07:14:02 +0000 UTC" firstStartedPulling="2025-10-14 07:14:05.968545118 +0000 UTC m=+781.665905489" lastFinishedPulling="2025-10-14 07:14:08.421909715 +0000 UTC m=+784.119270086" observedRunningTime="2025-10-14 07:14:09.052888102 +0000 UTC m=+784.750248493" watchObservedRunningTime="2025-10-14 07:14:09.333430853 +0000 UTC m=+785.030791264" Oct 14 07:14:10 crc kubenswrapper[4870]: I1014 07:14:10.053990 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-7468q" Oct 14 07:14:12 crc kubenswrapper[4870]: I1014 07:14:12.326783 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-7468q"] Oct 14 07:14:12 crc kubenswrapper[4870]: I1014 07:14:12.880115 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-x7gsd" Oct 14 07:14:12 crc kubenswrapper[4870]: I1014 07:14:12.880172 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-x7gsd" Oct 14 07:14:12 crc kubenswrapper[4870]: I1014 07:14:12.951063 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-x7gsd" Oct 14 07:14:13 crc kubenswrapper[4870]: I1014 07:14:13.041820 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-7468q" podUID="9574cf23-fd49-4791-ad7d-48e2c4bb69a9" containerName="registry-server" containerID="cri-o://9d6af32291b8bc4830801d4959abf0ee0f4472e274bb56a09ff7fc991b7fe965" gracePeriod=2 Oct 14 07:14:13 crc kubenswrapper[4870]: I1014 07:14:13.112545 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-x7gsd" Oct 14 07:14:13 crc kubenswrapper[4870]: I1014 07:14:13.832239 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-fdff9cb8d-269h6"] Oct 14 07:14:13 crc kubenswrapper[4870]: I1014 07:14:13.833144 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-269h6" Oct 14 07:14:13 crc kubenswrapper[4870]: I1014 07:14:13.834943 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-vbhb4" Oct 14 07:14:13 crc kubenswrapper[4870]: I1014 07:14:13.856278 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-6cdbc54649-5gj9z"] Oct 14 07:14:13 crc kubenswrapper[4870]: I1014 07:14:13.857157 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-5gj9z" Oct 14 07:14:13 crc kubenswrapper[4870]: I1014 07:14:13.858758 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Oct 14 07:14:13 crc kubenswrapper[4870]: I1014 07:14:13.859924 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-fdff9cb8d-269h6"] Oct 14 07:14:13 crc kubenswrapper[4870]: I1014 07:14:13.882511 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-ffdgs"] Oct 14 07:14:13 crc kubenswrapper[4870]: I1014 07:14:13.883414 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-ffdgs" Oct 14 07:14:13 crc kubenswrapper[4870]: I1014 07:14:13.894468 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6cdbc54649-5gj9z"] Oct 14 07:14:13 crc kubenswrapper[4870]: I1014 07:14:13.919754 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-x7gsd"] Oct 14 07:14:13 crc kubenswrapper[4870]: I1014 07:14:13.923788 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/65450853-71f4-44b7-bb21-1e5c650de334-nmstate-lock\") pod \"nmstate-handler-ffdgs\" (UID: \"65450853-71f4-44b7-bb21-1e5c650de334\") " pod="openshift-nmstate/nmstate-handler-ffdgs" Oct 14 07:14:13 crc kubenswrapper[4870]: I1014 07:14:13.923859 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jmpn5\" (UniqueName: \"kubernetes.io/projected/e4f05978-9de8-4cad-89d0-8620fbe004a3-kube-api-access-jmpn5\") pod \"nmstate-metrics-fdff9cb8d-269h6\" (UID: \"e4f05978-9de8-4cad-89d0-8620fbe004a3\") " pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-269h6" Oct 14 07:14:13 crc kubenswrapper[4870]: I1014 07:14:13.923895 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/78703da9-8ebd-4610-9d48-61c17c1eddeb-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-5gj9z\" (UID: \"78703da9-8ebd-4610-9d48-61c17c1eddeb\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-5gj9z" Oct 14 07:14:13 crc kubenswrapper[4870]: I1014 07:14:13.923919 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wbx9k\" (UniqueName: \"kubernetes.io/projected/78703da9-8ebd-4610-9d48-61c17c1eddeb-kube-api-access-wbx9k\") pod \"nmstate-webhook-6cdbc54649-5gj9z\" (UID: \"78703da9-8ebd-4610-9d48-61c17c1eddeb\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-5gj9z" Oct 14 07:14:13 crc kubenswrapper[4870]: I1014 07:14:13.923942 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/65450853-71f4-44b7-bb21-1e5c650de334-dbus-socket\") pod \"nmstate-handler-ffdgs\" (UID: \"65450853-71f4-44b7-bb21-1e5c650de334\") " pod="openshift-nmstate/nmstate-handler-ffdgs" Oct 14 07:14:13 crc kubenswrapper[4870]: I1014 07:14:13.923967 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/65450853-71f4-44b7-bb21-1e5c650de334-ovs-socket\") pod \"nmstate-handler-ffdgs\" (UID: \"65450853-71f4-44b7-bb21-1e5c650de334\") " pod="openshift-nmstate/nmstate-handler-ffdgs" Oct 14 07:14:13 crc kubenswrapper[4870]: I1014 07:14:13.923980 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4fqzl\" (UniqueName: \"kubernetes.io/projected/65450853-71f4-44b7-bb21-1e5c650de334-kube-api-access-4fqzl\") pod \"nmstate-handler-ffdgs\" (UID: \"65450853-71f4-44b7-bb21-1e5c650de334\") " pod="openshift-nmstate/nmstate-handler-ffdgs" Oct 14 07:14:13 crc kubenswrapper[4870]: I1014 07:14:13.986992 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6b874cbd85-w5zvx"] Oct 14 07:14:13 crc kubenswrapper[4870]: I1014 07:14:13.987781 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-w5zvx" Oct 14 07:14:13 crc kubenswrapper[4870]: I1014 07:14:13.989996 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Oct 14 07:14:13 crc kubenswrapper[4870]: I1014 07:14:13.990068 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-x2jdt" Oct 14 07:14:13 crc kubenswrapper[4870]: I1014 07:14:13.990973 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Oct 14 07:14:13 crc kubenswrapper[4870]: I1014 07:14:13.997002 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6b874cbd85-w5zvx"] Oct 14 07:14:14 crc kubenswrapper[4870]: I1014 07:14:14.025038 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/78703da9-8ebd-4610-9d48-61c17c1eddeb-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-5gj9z\" (UID: \"78703da9-8ebd-4610-9d48-61c17c1eddeb\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-5gj9z" Oct 14 07:14:14 crc kubenswrapper[4870]: I1014 07:14:14.025294 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wbx9k\" (UniqueName: \"kubernetes.io/projected/78703da9-8ebd-4610-9d48-61c17c1eddeb-kube-api-access-wbx9k\") pod \"nmstate-webhook-6cdbc54649-5gj9z\" (UID: \"78703da9-8ebd-4610-9d48-61c17c1eddeb\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-5gj9z" Oct 14 07:14:14 crc kubenswrapper[4870]: E1014 07:14:14.025217 4870 secret.go:188] Couldn't get secret openshift-nmstate/openshift-nmstate-webhook: secret "openshift-nmstate-webhook" not found Oct 14 07:14:14 crc kubenswrapper[4870]: E1014 07:14:14.025529 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/78703da9-8ebd-4610-9d48-61c17c1eddeb-tls-key-pair podName:78703da9-8ebd-4610-9d48-61c17c1eddeb nodeName:}" failed. No retries permitted until 2025-10-14 07:14:14.525507832 +0000 UTC m=+790.222868303 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "tls-key-pair" (UniqueName: "kubernetes.io/secret/78703da9-8ebd-4610-9d48-61c17c1eddeb-tls-key-pair") pod "nmstate-webhook-6cdbc54649-5gj9z" (UID: "78703da9-8ebd-4610-9d48-61c17c1eddeb") : secret "openshift-nmstate-webhook" not found Oct 14 07:14:14 crc kubenswrapper[4870]: I1014 07:14:14.025435 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/65450853-71f4-44b7-bb21-1e5c650de334-dbus-socket\") pod \"nmstate-handler-ffdgs\" (UID: \"65450853-71f4-44b7-bb21-1e5c650de334\") " pod="openshift-nmstate/nmstate-handler-ffdgs" Oct 14 07:14:14 crc kubenswrapper[4870]: I1014 07:14:14.025677 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/65450853-71f4-44b7-bb21-1e5c650de334-ovs-socket\") pod \"nmstate-handler-ffdgs\" (UID: \"65450853-71f4-44b7-bb21-1e5c650de334\") " pod="openshift-nmstate/nmstate-handler-ffdgs" Oct 14 07:14:14 crc kubenswrapper[4870]: I1014 07:14:14.025707 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4fqzl\" (UniqueName: \"kubernetes.io/projected/65450853-71f4-44b7-bb21-1e5c650de334-kube-api-access-4fqzl\") pod \"nmstate-handler-ffdgs\" (UID: \"65450853-71f4-44b7-bb21-1e5c650de334\") " pod="openshift-nmstate/nmstate-handler-ffdgs" Oct 14 07:14:14 crc kubenswrapper[4870]: I1014 07:14:14.025765 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/65450853-71f4-44b7-bb21-1e5c650de334-ovs-socket\") pod \"nmstate-handler-ffdgs\" (UID: \"65450853-71f4-44b7-bb21-1e5c650de334\") " pod="openshift-nmstate/nmstate-handler-ffdgs" Oct 14 07:14:14 crc kubenswrapper[4870]: I1014 07:14:14.025843 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/65450853-71f4-44b7-bb21-1e5c650de334-nmstate-lock\") pod \"nmstate-handler-ffdgs\" (UID: \"65450853-71f4-44b7-bb21-1e5c650de334\") " pod="openshift-nmstate/nmstate-handler-ffdgs" Oct 14 07:14:14 crc kubenswrapper[4870]: I1014 07:14:14.025905 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jmpn5\" (UniqueName: \"kubernetes.io/projected/e4f05978-9de8-4cad-89d0-8620fbe004a3-kube-api-access-jmpn5\") pod \"nmstate-metrics-fdff9cb8d-269h6\" (UID: \"e4f05978-9de8-4cad-89d0-8620fbe004a3\") " pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-269h6" Oct 14 07:14:14 crc kubenswrapper[4870]: I1014 07:14:14.025951 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/65450853-71f4-44b7-bb21-1e5c650de334-nmstate-lock\") pod \"nmstate-handler-ffdgs\" (UID: \"65450853-71f4-44b7-bb21-1e5c650de334\") " pod="openshift-nmstate/nmstate-handler-ffdgs" Oct 14 07:14:14 crc kubenswrapper[4870]: I1014 07:14:14.026262 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/65450853-71f4-44b7-bb21-1e5c650de334-dbus-socket\") pod \"nmstate-handler-ffdgs\" (UID: \"65450853-71f4-44b7-bb21-1e5c650de334\") " pod="openshift-nmstate/nmstate-handler-ffdgs" Oct 14 07:14:14 crc kubenswrapper[4870]: I1014 07:14:14.042771 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jmpn5\" (UniqueName: \"kubernetes.io/projected/e4f05978-9de8-4cad-89d0-8620fbe004a3-kube-api-access-jmpn5\") pod \"nmstate-metrics-fdff9cb8d-269h6\" (UID: \"e4f05978-9de8-4cad-89d0-8620fbe004a3\") " pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-269h6" Oct 14 07:14:14 crc kubenswrapper[4870]: I1014 07:14:14.043431 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wbx9k\" (UniqueName: \"kubernetes.io/projected/78703da9-8ebd-4610-9d48-61c17c1eddeb-kube-api-access-wbx9k\") pod \"nmstate-webhook-6cdbc54649-5gj9z\" (UID: \"78703da9-8ebd-4610-9d48-61c17c1eddeb\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-5gj9z" Oct 14 07:14:14 crc kubenswrapper[4870]: I1014 07:14:14.044949 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4fqzl\" (UniqueName: \"kubernetes.io/projected/65450853-71f4-44b7-bb21-1e5c650de334-kube-api-access-4fqzl\") pod \"nmstate-handler-ffdgs\" (UID: \"65450853-71f4-44b7-bb21-1e5c650de334\") " pod="openshift-nmstate/nmstate-handler-ffdgs" Oct 14 07:14:14 crc kubenswrapper[4870]: I1014 07:14:14.133397 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/9292dbae-ed2d-4a25-a889-b951b36614c6-plugin-serving-cert\") pod \"nmstate-console-plugin-6b874cbd85-w5zvx\" (UID: \"9292dbae-ed2d-4a25-a889-b951b36614c6\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-w5zvx" Oct 14 07:14:14 crc kubenswrapper[4870]: I1014 07:14:14.133904 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/9292dbae-ed2d-4a25-a889-b951b36614c6-nginx-conf\") pod \"nmstate-console-plugin-6b874cbd85-w5zvx\" (UID: \"9292dbae-ed2d-4a25-a889-b951b36614c6\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-w5zvx" Oct 14 07:14:14 crc kubenswrapper[4870]: I1014 07:14:14.134129 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m6x4h\" (UniqueName: \"kubernetes.io/projected/9292dbae-ed2d-4a25-a889-b951b36614c6-kube-api-access-m6x4h\") pod \"nmstate-console-plugin-6b874cbd85-w5zvx\" (UID: \"9292dbae-ed2d-4a25-a889-b951b36614c6\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-w5zvx" Oct 14 07:14:14 crc kubenswrapper[4870]: I1014 07:14:14.150804 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-269h6" Oct 14 07:14:14 crc kubenswrapper[4870]: I1014 07:14:14.177990 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-76877f899c-x2gzs"] Oct 14 07:14:14 crc kubenswrapper[4870]: I1014 07:14:14.178632 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-76877f899c-x2gzs" Oct 14 07:14:14 crc kubenswrapper[4870]: I1014 07:14:14.203190 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-ffdgs" Oct 14 07:14:14 crc kubenswrapper[4870]: I1014 07:14:14.203990 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-76877f899c-x2gzs"] Oct 14 07:14:14 crc kubenswrapper[4870]: I1014 07:14:14.235079 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kt6zz\" (UniqueName: \"kubernetes.io/projected/af917993-9d26-4d59-9176-32b4e53e2400-kube-api-access-kt6zz\") pod \"console-76877f899c-x2gzs\" (UID: \"af917993-9d26-4d59-9176-32b4e53e2400\") " pod="openshift-console/console-76877f899c-x2gzs" Oct 14 07:14:14 crc kubenswrapper[4870]: I1014 07:14:14.235125 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/af917993-9d26-4d59-9176-32b4e53e2400-service-ca\") pod \"console-76877f899c-x2gzs\" (UID: \"af917993-9d26-4d59-9176-32b4e53e2400\") " pod="openshift-console/console-76877f899c-x2gzs" Oct 14 07:14:14 crc kubenswrapper[4870]: I1014 07:14:14.235150 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/af917993-9d26-4d59-9176-32b4e53e2400-oauth-serving-cert\") pod \"console-76877f899c-x2gzs\" (UID: \"af917993-9d26-4d59-9176-32b4e53e2400\") " pod="openshift-console/console-76877f899c-x2gzs" Oct 14 07:14:14 crc kubenswrapper[4870]: I1014 07:14:14.235192 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/af917993-9d26-4d59-9176-32b4e53e2400-trusted-ca-bundle\") pod \"console-76877f899c-x2gzs\" (UID: \"af917993-9d26-4d59-9176-32b4e53e2400\") " pod="openshift-console/console-76877f899c-x2gzs" Oct 14 07:14:14 crc kubenswrapper[4870]: I1014 07:14:14.235355 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/9292dbae-ed2d-4a25-a889-b951b36614c6-nginx-conf\") pod \"nmstate-console-plugin-6b874cbd85-w5zvx\" (UID: \"9292dbae-ed2d-4a25-a889-b951b36614c6\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-w5zvx" Oct 14 07:14:14 crc kubenswrapper[4870]: I1014 07:14:14.235455 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/af917993-9d26-4d59-9176-32b4e53e2400-console-config\") pod \"console-76877f899c-x2gzs\" (UID: \"af917993-9d26-4d59-9176-32b4e53e2400\") " pod="openshift-console/console-76877f899c-x2gzs" Oct 14 07:14:14 crc kubenswrapper[4870]: I1014 07:14:14.235484 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/af917993-9d26-4d59-9176-32b4e53e2400-console-serving-cert\") pod \"console-76877f899c-x2gzs\" (UID: \"af917993-9d26-4d59-9176-32b4e53e2400\") " pod="openshift-console/console-76877f899c-x2gzs" Oct 14 07:14:14 crc kubenswrapper[4870]: I1014 07:14:14.235530 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/af917993-9d26-4d59-9176-32b4e53e2400-console-oauth-config\") pod \"console-76877f899c-x2gzs\" (UID: \"af917993-9d26-4d59-9176-32b4e53e2400\") " pod="openshift-console/console-76877f899c-x2gzs" Oct 14 07:14:14 crc kubenswrapper[4870]: I1014 07:14:14.235601 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m6x4h\" (UniqueName: \"kubernetes.io/projected/9292dbae-ed2d-4a25-a889-b951b36614c6-kube-api-access-m6x4h\") pod \"nmstate-console-plugin-6b874cbd85-w5zvx\" (UID: \"9292dbae-ed2d-4a25-a889-b951b36614c6\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-w5zvx" Oct 14 07:14:14 crc kubenswrapper[4870]: I1014 07:14:14.235648 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/9292dbae-ed2d-4a25-a889-b951b36614c6-plugin-serving-cert\") pod \"nmstate-console-plugin-6b874cbd85-w5zvx\" (UID: \"9292dbae-ed2d-4a25-a889-b951b36614c6\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-w5zvx" Oct 14 07:14:14 crc kubenswrapper[4870]: I1014 07:14:14.236241 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/9292dbae-ed2d-4a25-a889-b951b36614c6-nginx-conf\") pod \"nmstate-console-plugin-6b874cbd85-w5zvx\" (UID: \"9292dbae-ed2d-4a25-a889-b951b36614c6\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-w5zvx" Oct 14 07:14:14 crc kubenswrapper[4870]: I1014 07:14:14.245551 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/9292dbae-ed2d-4a25-a889-b951b36614c6-plugin-serving-cert\") pod \"nmstate-console-plugin-6b874cbd85-w5zvx\" (UID: \"9292dbae-ed2d-4a25-a889-b951b36614c6\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-w5zvx" Oct 14 07:14:14 crc kubenswrapper[4870]: I1014 07:14:14.251394 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m6x4h\" (UniqueName: \"kubernetes.io/projected/9292dbae-ed2d-4a25-a889-b951b36614c6-kube-api-access-m6x4h\") pod \"nmstate-console-plugin-6b874cbd85-w5zvx\" (UID: \"9292dbae-ed2d-4a25-a889-b951b36614c6\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-w5zvx" Oct 14 07:14:14 crc kubenswrapper[4870]: I1014 07:14:14.301405 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-w5zvx" Oct 14 07:14:14 crc kubenswrapper[4870]: I1014 07:14:14.336301 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/af917993-9d26-4d59-9176-32b4e53e2400-console-oauth-config\") pod \"console-76877f899c-x2gzs\" (UID: \"af917993-9d26-4d59-9176-32b4e53e2400\") " pod="openshift-console/console-76877f899c-x2gzs" Oct 14 07:14:14 crc kubenswrapper[4870]: I1014 07:14:14.336384 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kt6zz\" (UniqueName: \"kubernetes.io/projected/af917993-9d26-4d59-9176-32b4e53e2400-kube-api-access-kt6zz\") pod \"console-76877f899c-x2gzs\" (UID: \"af917993-9d26-4d59-9176-32b4e53e2400\") " pod="openshift-console/console-76877f899c-x2gzs" Oct 14 07:14:14 crc kubenswrapper[4870]: I1014 07:14:14.336406 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/af917993-9d26-4d59-9176-32b4e53e2400-service-ca\") pod \"console-76877f899c-x2gzs\" (UID: \"af917993-9d26-4d59-9176-32b4e53e2400\") " pod="openshift-console/console-76877f899c-x2gzs" Oct 14 07:14:14 crc kubenswrapper[4870]: I1014 07:14:14.336424 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/af917993-9d26-4d59-9176-32b4e53e2400-oauth-serving-cert\") pod \"console-76877f899c-x2gzs\" (UID: \"af917993-9d26-4d59-9176-32b4e53e2400\") " pod="openshift-console/console-76877f899c-x2gzs" Oct 14 07:14:14 crc kubenswrapper[4870]: I1014 07:14:14.336480 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/af917993-9d26-4d59-9176-32b4e53e2400-trusted-ca-bundle\") pod \"console-76877f899c-x2gzs\" (UID: \"af917993-9d26-4d59-9176-32b4e53e2400\") " pod="openshift-console/console-76877f899c-x2gzs" Oct 14 07:14:14 crc kubenswrapper[4870]: I1014 07:14:14.336520 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/af917993-9d26-4d59-9176-32b4e53e2400-console-config\") pod \"console-76877f899c-x2gzs\" (UID: \"af917993-9d26-4d59-9176-32b4e53e2400\") " pod="openshift-console/console-76877f899c-x2gzs" Oct 14 07:14:14 crc kubenswrapper[4870]: I1014 07:14:14.336537 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/af917993-9d26-4d59-9176-32b4e53e2400-console-serving-cert\") pod \"console-76877f899c-x2gzs\" (UID: \"af917993-9d26-4d59-9176-32b4e53e2400\") " pod="openshift-console/console-76877f899c-x2gzs" Oct 14 07:14:14 crc kubenswrapper[4870]: I1014 07:14:14.337600 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/af917993-9d26-4d59-9176-32b4e53e2400-oauth-serving-cert\") pod \"console-76877f899c-x2gzs\" (UID: \"af917993-9d26-4d59-9176-32b4e53e2400\") " pod="openshift-console/console-76877f899c-x2gzs" Oct 14 07:14:14 crc kubenswrapper[4870]: I1014 07:14:14.337888 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/af917993-9d26-4d59-9176-32b4e53e2400-console-config\") pod \"console-76877f899c-x2gzs\" (UID: \"af917993-9d26-4d59-9176-32b4e53e2400\") " pod="openshift-console/console-76877f899c-x2gzs" Oct 14 07:14:14 crc kubenswrapper[4870]: I1014 07:14:14.337979 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/af917993-9d26-4d59-9176-32b4e53e2400-trusted-ca-bundle\") pod \"console-76877f899c-x2gzs\" (UID: \"af917993-9d26-4d59-9176-32b4e53e2400\") " pod="openshift-console/console-76877f899c-x2gzs" Oct 14 07:14:14 crc kubenswrapper[4870]: I1014 07:14:14.338949 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/af917993-9d26-4d59-9176-32b4e53e2400-service-ca\") pod \"console-76877f899c-x2gzs\" (UID: \"af917993-9d26-4d59-9176-32b4e53e2400\") " pod="openshift-console/console-76877f899c-x2gzs" Oct 14 07:14:14 crc kubenswrapper[4870]: I1014 07:14:14.342552 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/af917993-9d26-4d59-9176-32b4e53e2400-console-oauth-config\") pod \"console-76877f899c-x2gzs\" (UID: \"af917993-9d26-4d59-9176-32b4e53e2400\") " pod="openshift-console/console-76877f899c-x2gzs" Oct 14 07:14:14 crc kubenswrapper[4870]: I1014 07:14:14.352029 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/af917993-9d26-4d59-9176-32b4e53e2400-console-serving-cert\") pod \"console-76877f899c-x2gzs\" (UID: \"af917993-9d26-4d59-9176-32b4e53e2400\") " pod="openshift-console/console-76877f899c-x2gzs" Oct 14 07:14:14 crc kubenswrapper[4870]: I1014 07:14:14.355849 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kt6zz\" (UniqueName: \"kubernetes.io/projected/af917993-9d26-4d59-9176-32b4e53e2400-kube-api-access-kt6zz\") pod \"console-76877f899c-x2gzs\" (UID: \"af917993-9d26-4d59-9176-32b4e53e2400\") " pod="openshift-console/console-76877f899c-x2gzs" Oct 14 07:14:14 crc kubenswrapper[4870]: I1014 07:14:14.531718 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-76877f899c-x2gzs" Oct 14 07:14:14 crc kubenswrapper[4870]: I1014 07:14:14.539056 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/78703da9-8ebd-4610-9d48-61c17c1eddeb-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-5gj9z\" (UID: \"78703da9-8ebd-4610-9d48-61c17c1eddeb\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-5gj9z" Oct 14 07:14:14 crc kubenswrapper[4870]: I1014 07:14:14.542138 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/78703da9-8ebd-4610-9d48-61c17c1eddeb-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-5gj9z\" (UID: \"78703da9-8ebd-4610-9d48-61c17c1eddeb\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-5gj9z" Oct 14 07:14:14 crc kubenswrapper[4870]: I1014 07:14:14.611282 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-fdff9cb8d-269h6"] Oct 14 07:14:14 crc kubenswrapper[4870]: W1014 07:14:14.628281 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode4f05978_9de8_4cad_89d0_8620fbe004a3.slice/crio-ed6f60c40795929414b65a5abe4ee2a3ab37a50e0ed1427d9941511e98f8402c WatchSource:0}: Error finding container ed6f60c40795929414b65a5abe4ee2a3ab37a50e0ed1427d9941511e98f8402c: Status 404 returned error can't find the container with id ed6f60c40795929414b65a5abe4ee2a3ab37a50e0ed1427d9941511e98f8402c Oct 14 07:14:14 crc kubenswrapper[4870]: I1014 07:14:14.695450 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6b874cbd85-w5zvx"] Oct 14 07:14:14 crc kubenswrapper[4870]: W1014 07:14:14.711050 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9292dbae_ed2d_4a25_a889_b951b36614c6.slice/crio-55badf7daf777c5a74d9e14faa9ce0c14f28e9bfa759ade098a882536f7d8ff3 WatchSource:0}: Error finding container 55badf7daf777c5a74d9e14faa9ce0c14f28e9bfa759ade098a882536f7d8ff3: Status 404 returned error can't find the container with id 55badf7daf777c5a74d9e14faa9ce0c14f28e9bfa759ade098a882536f7d8ff3 Oct 14 07:14:14 crc kubenswrapper[4870]: I1014 07:14:14.747196 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-76877f899c-x2gzs"] Oct 14 07:14:14 crc kubenswrapper[4870]: I1014 07:14:14.775928 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-5gj9z" Oct 14 07:14:15 crc kubenswrapper[4870]: I1014 07:14:15.053182 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-76877f899c-x2gzs" event={"ID":"af917993-9d26-4d59-9176-32b4e53e2400","Type":"ContainerStarted","Data":"eb3b91ab4e9c3b95160f176e4fa1a184d624ab4421fb5ec85f506bb81f1268de"} Oct 14 07:14:15 crc kubenswrapper[4870]: I1014 07:14:15.054214 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-w5zvx" event={"ID":"9292dbae-ed2d-4a25-a889-b951b36614c6","Type":"ContainerStarted","Data":"55badf7daf777c5a74d9e14faa9ce0c14f28e9bfa759ade098a882536f7d8ff3"} Oct 14 07:14:15 crc kubenswrapper[4870]: I1014 07:14:15.055205 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-ffdgs" event={"ID":"65450853-71f4-44b7-bb21-1e5c650de334","Type":"ContainerStarted","Data":"6e030f494cb028c03b4ad0cd85cf636cd3830426ee592cb0243e65ca33a1311d"} Oct 14 07:14:15 crc kubenswrapper[4870]: I1014 07:14:15.056099 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-269h6" event={"ID":"e4f05978-9de8-4cad-89d0-8620fbe004a3","Type":"ContainerStarted","Data":"ed6f60c40795929414b65a5abe4ee2a3ab37a50e0ed1427d9941511e98f8402c"} Oct 14 07:14:15 crc kubenswrapper[4870]: I1014 07:14:15.056240 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-x7gsd" podUID="68ce28b7-83f9-4ba8-8a95-b8995f5151ca" containerName="registry-server" containerID="cri-o://c9b11a2b2df54d1e96336de480fa60aad3cca93be98e3fd4ed2110b8ffb36483" gracePeriod=2 Oct 14 07:14:15 crc kubenswrapper[4870]: I1014 07:14:15.173663 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6cdbc54649-5gj9z"] Oct 14 07:14:15 crc kubenswrapper[4870]: I1014 07:14:15.870937 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-xwpt9" Oct 14 07:14:15 crc kubenswrapper[4870]: I1014 07:14:15.871054 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-xwpt9" Oct 14 07:14:15 crc kubenswrapper[4870]: I1014 07:14:15.946793 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-xwpt9" Oct 14 07:14:16 crc kubenswrapper[4870]: I1014 07:14:16.071206 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-76877f899c-x2gzs" event={"ID":"af917993-9d26-4d59-9176-32b4e53e2400","Type":"ContainerStarted","Data":"2e9a145f981e8c87e49d5ccc23289185481682d1917cc9350bdff109d3ab1033"} Oct 14 07:14:16 crc kubenswrapper[4870]: I1014 07:14:16.072925 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-5gj9z" event={"ID":"78703da9-8ebd-4610-9d48-61c17c1eddeb","Type":"ContainerStarted","Data":"f564bfa4b35a8cb2bf6d6a09d43a36a125ccc9712ba9454bc5f759112bd308a4"} Oct 14 07:14:16 crc kubenswrapper[4870]: I1014 07:14:16.078519 4870 generic.go:334] "Generic (PLEG): container finished" podID="68ce28b7-83f9-4ba8-8a95-b8995f5151ca" containerID="c9b11a2b2df54d1e96336de480fa60aad3cca93be98e3fd4ed2110b8ffb36483" exitCode=0 Oct 14 07:14:16 crc kubenswrapper[4870]: I1014 07:14:16.078604 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-x7gsd" event={"ID":"68ce28b7-83f9-4ba8-8a95-b8995f5151ca","Type":"ContainerDied","Data":"c9b11a2b2df54d1e96336de480fa60aad3cca93be98e3fd4ed2110b8ffb36483"} Oct 14 07:14:16 crc kubenswrapper[4870]: I1014 07:14:16.084482 4870 generic.go:334] "Generic (PLEG): container finished" podID="9574cf23-fd49-4791-ad7d-48e2c4bb69a9" containerID="9d6af32291b8bc4830801d4959abf0ee0f4472e274bb56a09ff7fc991b7fe965" exitCode=0 Oct 14 07:14:16 crc kubenswrapper[4870]: I1014 07:14:16.085146 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7468q" event={"ID":"9574cf23-fd49-4791-ad7d-48e2c4bb69a9","Type":"ContainerDied","Data":"9d6af32291b8bc4830801d4959abf0ee0f4472e274bb56a09ff7fc991b7fe965"} Oct 14 07:14:16 crc kubenswrapper[4870]: I1014 07:14:16.156613 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-xwpt9" Oct 14 07:14:16 crc kubenswrapper[4870]: I1014 07:14:16.176570 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-76877f899c-x2gzs" podStartSLOduration=2.176550591 podStartE2EDuration="2.176550591s" podCreationTimestamp="2025-10-14 07:14:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:14:16.107145537 +0000 UTC m=+791.804505998" watchObservedRunningTime="2025-10-14 07:14:16.176550591 +0000 UTC m=+791.873910972" Oct 14 07:14:16 crc kubenswrapper[4870]: I1014 07:14:16.404046 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7468q" Oct 14 07:14:16 crc kubenswrapper[4870]: I1014 07:14:16.570977 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9574cf23-fd49-4791-ad7d-48e2c4bb69a9-utilities\") pod \"9574cf23-fd49-4791-ad7d-48e2c4bb69a9\" (UID: \"9574cf23-fd49-4791-ad7d-48e2c4bb69a9\") " Oct 14 07:14:16 crc kubenswrapper[4870]: I1014 07:14:16.571579 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tcdfz\" (UniqueName: \"kubernetes.io/projected/9574cf23-fd49-4791-ad7d-48e2c4bb69a9-kube-api-access-tcdfz\") pod \"9574cf23-fd49-4791-ad7d-48e2c4bb69a9\" (UID: \"9574cf23-fd49-4791-ad7d-48e2c4bb69a9\") " Oct 14 07:14:16 crc kubenswrapper[4870]: I1014 07:14:16.571669 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9574cf23-fd49-4791-ad7d-48e2c4bb69a9-catalog-content\") pod \"9574cf23-fd49-4791-ad7d-48e2c4bb69a9\" (UID: \"9574cf23-fd49-4791-ad7d-48e2c4bb69a9\") " Oct 14 07:14:16 crc kubenswrapper[4870]: I1014 07:14:16.571679 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9574cf23-fd49-4791-ad7d-48e2c4bb69a9-utilities" (OuterVolumeSpecName: "utilities") pod "9574cf23-fd49-4791-ad7d-48e2c4bb69a9" (UID: "9574cf23-fd49-4791-ad7d-48e2c4bb69a9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:14:16 crc kubenswrapper[4870]: I1014 07:14:16.572043 4870 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9574cf23-fd49-4791-ad7d-48e2c4bb69a9-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 07:14:16 crc kubenswrapper[4870]: I1014 07:14:16.579653 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9574cf23-fd49-4791-ad7d-48e2c4bb69a9-kube-api-access-tcdfz" (OuterVolumeSpecName: "kube-api-access-tcdfz") pod "9574cf23-fd49-4791-ad7d-48e2c4bb69a9" (UID: "9574cf23-fd49-4791-ad7d-48e2c4bb69a9"). InnerVolumeSpecName "kube-api-access-tcdfz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:14:16 crc kubenswrapper[4870]: I1014 07:14:16.673114 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tcdfz\" (UniqueName: \"kubernetes.io/projected/9574cf23-fd49-4791-ad7d-48e2c4bb69a9-kube-api-access-tcdfz\") on node \"crc\" DevicePath \"\"" Oct 14 07:14:17 crc kubenswrapper[4870]: I1014 07:14:17.091362 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7468q" event={"ID":"9574cf23-fd49-4791-ad7d-48e2c4bb69a9","Type":"ContainerDied","Data":"10fe2bdeb01e1cdb40323886ee0e14aff656bacac2afbc659474c1ac27903234"} Oct 14 07:14:17 crc kubenswrapper[4870]: I1014 07:14:17.091404 4870 scope.go:117] "RemoveContainer" containerID="9d6af32291b8bc4830801d4959abf0ee0f4472e274bb56a09ff7fc991b7fe965" Oct 14 07:14:17 crc kubenswrapper[4870]: I1014 07:14:17.091986 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7468q" Oct 14 07:14:17 crc kubenswrapper[4870]: I1014 07:14:17.114240 4870 scope.go:117] "RemoveContainer" containerID="bc59f4bdba65ece0ae32bec288269ba68f06cb2c97842be773a606c678e8e95a" Oct 14 07:14:17 crc kubenswrapper[4870]: I1014 07:14:17.158721 4870 scope.go:117] "RemoveContainer" containerID="2fae72da1e50d3e3b2f7f871ec0fd4a5e686be5f1adf9fcfb8676861027afc02" Oct 14 07:14:17 crc kubenswrapper[4870]: I1014 07:14:17.197654 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-x7gsd" Oct 14 07:14:17 crc kubenswrapper[4870]: I1014 07:14:17.290140 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/68ce28b7-83f9-4ba8-8a95-b8995f5151ca-utilities\") pod \"68ce28b7-83f9-4ba8-8a95-b8995f5151ca\" (UID: \"68ce28b7-83f9-4ba8-8a95-b8995f5151ca\") " Oct 14 07:14:17 crc kubenswrapper[4870]: I1014 07:14:17.291610 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/68ce28b7-83f9-4ba8-8a95-b8995f5151ca-utilities" (OuterVolumeSpecName: "utilities") pod "68ce28b7-83f9-4ba8-8a95-b8995f5151ca" (UID: "68ce28b7-83f9-4ba8-8a95-b8995f5151ca"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:14:17 crc kubenswrapper[4870]: I1014 07:14:17.391081 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jt2v7\" (UniqueName: \"kubernetes.io/projected/68ce28b7-83f9-4ba8-8a95-b8995f5151ca-kube-api-access-jt2v7\") pod \"68ce28b7-83f9-4ba8-8a95-b8995f5151ca\" (UID: \"68ce28b7-83f9-4ba8-8a95-b8995f5151ca\") " Oct 14 07:14:17 crc kubenswrapper[4870]: I1014 07:14:17.391272 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/68ce28b7-83f9-4ba8-8a95-b8995f5151ca-catalog-content\") pod \"68ce28b7-83f9-4ba8-8a95-b8995f5151ca\" (UID: \"68ce28b7-83f9-4ba8-8a95-b8995f5151ca\") " Oct 14 07:14:17 crc kubenswrapper[4870]: I1014 07:14:17.391905 4870 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/68ce28b7-83f9-4ba8-8a95-b8995f5151ca-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 07:14:17 crc kubenswrapper[4870]: I1014 07:14:17.398358 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/68ce28b7-83f9-4ba8-8a95-b8995f5151ca-kube-api-access-jt2v7" (OuterVolumeSpecName: "kube-api-access-jt2v7") pod "68ce28b7-83f9-4ba8-8a95-b8995f5151ca" (UID: "68ce28b7-83f9-4ba8-8a95-b8995f5151ca"). InnerVolumeSpecName "kube-api-access-jt2v7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:14:17 crc kubenswrapper[4870]: I1014 07:14:17.493026 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jt2v7\" (UniqueName: \"kubernetes.io/projected/68ce28b7-83f9-4ba8-8a95-b8995f5151ca-kube-api-access-jt2v7\") on node \"crc\" DevicePath \"\"" Oct 14 07:14:18 crc kubenswrapper[4870]: I1014 07:14:18.104974 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-x7gsd" event={"ID":"68ce28b7-83f9-4ba8-8a95-b8995f5151ca","Type":"ContainerDied","Data":"21b97618051e13bd0ecdda8c2bc413790c7e0fddade0a14d133a421e5253656a"} Oct 14 07:14:18 crc kubenswrapper[4870]: I1014 07:14:18.105721 4870 scope.go:117] "RemoveContainer" containerID="c9b11a2b2df54d1e96336de480fa60aad3cca93be98e3fd4ed2110b8ffb36483" Oct 14 07:14:18 crc kubenswrapper[4870]: I1014 07:14:18.106036 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-x7gsd" Oct 14 07:14:18 crc kubenswrapper[4870]: I1014 07:14:18.130472 4870 scope.go:117] "RemoveContainer" containerID="5d60cde3e1e8e54de1128bea6a214678baee822141909f171431934a2e45082d" Oct 14 07:14:18 crc kubenswrapper[4870]: I1014 07:14:18.158693 4870 scope.go:117] "RemoveContainer" containerID="fab4b4e6086d944653e286c195f0877c8457965eced0604f885e8a105ed8ace0" Oct 14 07:14:18 crc kubenswrapper[4870]: I1014 07:14:18.579981 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-xwpt9"] Oct 14 07:14:18 crc kubenswrapper[4870]: I1014 07:14:18.820737 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/68ce28b7-83f9-4ba8-8a95-b8995f5151ca-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "68ce28b7-83f9-4ba8-8a95-b8995f5151ca" (UID: "68ce28b7-83f9-4ba8-8a95-b8995f5151ca"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:14:18 crc kubenswrapper[4870]: I1014 07:14:18.821214 4870 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/68ce28b7-83f9-4ba8-8a95-b8995f5151ca-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 07:14:18 crc kubenswrapper[4870]: I1014 07:14:18.833970 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9574cf23-fd49-4791-ad7d-48e2c4bb69a9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9574cf23-fd49-4791-ad7d-48e2c4bb69a9" (UID: "9574cf23-fd49-4791-ad7d-48e2c4bb69a9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:14:18 crc kubenswrapper[4870]: I1014 07:14:18.942092 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-5t25t"] Oct 14 07:14:18 crc kubenswrapper[4870]: I1014 07:14:18.942218 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-7468q"] Oct 14 07:14:18 crc kubenswrapper[4870]: I1014 07:14:18.942636 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-5t25t" podUID="9dc73b58-877c-4e60-8508-19ffeb7410f4" containerName="registry-server" containerID="cri-o://056466e4ca8980a8e9136d1cfbe342617256392c6762318dd651a7b3a6a7449b" gracePeriod=2 Oct 14 07:14:18 crc kubenswrapper[4870]: I1014 07:14:18.948318 4870 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9574cf23-fd49-4791-ad7d-48e2c4bb69a9-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 07:14:18 crc kubenswrapper[4870]: I1014 07:14:18.956608 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-7468q"] Oct 14 07:14:19 crc kubenswrapper[4870]: I1014 07:14:19.044839 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9574cf23-fd49-4791-ad7d-48e2c4bb69a9" path="/var/lib/kubelet/pods/9574cf23-fd49-4791-ad7d-48e2c4bb69a9/volumes" Oct 14 07:14:19 crc kubenswrapper[4870]: I1014 07:14:19.066760 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-x7gsd"] Oct 14 07:14:19 crc kubenswrapper[4870]: I1014 07:14:19.070740 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-x7gsd"] Oct 14 07:14:19 crc kubenswrapper[4870]: I1014 07:14:19.123694 4870 generic.go:334] "Generic (PLEG): container finished" podID="9dc73b58-877c-4e60-8508-19ffeb7410f4" containerID="056466e4ca8980a8e9136d1cfbe342617256392c6762318dd651a7b3a6a7449b" exitCode=0 Oct 14 07:14:19 crc kubenswrapper[4870]: I1014 07:14:19.123773 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5t25t" event={"ID":"9dc73b58-877c-4e60-8508-19ffeb7410f4","Type":"ContainerDied","Data":"056466e4ca8980a8e9136d1cfbe342617256392c6762318dd651a7b3a6a7449b"} Oct 14 07:14:20 crc kubenswrapper[4870]: I1014 07:14:20.784741 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5t25t" Oct 14 07:14:20 crc kubenswrapper[4870]: I1014 07:14:20.976562 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n4mlw\" (UniqueName: \"kubernetes.io/projected/9dc73b58-877c-4e60-8508-19ffeb7410f4-kube-api-access-n4mlw\") pod \"9dc73b58-877c-4e60-8508-19ffeb7410f4\" (UID: \"9dc73b58-877c-4e60-8508-19ffeb7410f4\") " Oct 14 07:14:20 crc kubenswrapper[4870]: I1014 07:14:20.977021 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9dc73b58-877c-4e60-8508-19ffeb7410f4-utilities\") pod \"9dc73b58-877c-4e60-8508-19ffeb7410f4\" (UID: \"9dc73b58-877c-4e60-8508-19ffeb7410f4\") " Oct 14 07:14:20 crc kubenswrapper[4870]: I1014 07:14:20.977078 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9dc73b58-877c-4e60-8508-19ffeb7410f4-catalog-content\") pod \"9dc73b58-877c-4e60-8508-19ffeb7410f4\" (UID: \"9dc73b58-877c-4e60-8508-19ffeb7410f4\") " Oct 14 07:14:20 crc kubenswrapper[4870]: I1014 07:14:20.977969 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9dc73b58-877c-4e60-8508-19ffeb7410f4-utilities" (OuterVolumeSpecName: "utilities") pod "9dc73b58-877c-4e60-8508-19ffeb7410f4" (UID: "9dc73b58-877c-4e60-8508-19ffeb7410f4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:14:20 crc kubenswrapper[4870]: I1014 07:14:20.980402 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9dc73b58-877c-4e60-8508-19ffeb7410f4-kube-api-access-n4mlw" (OuterVolumeSpecName: "kube-api-access-n4mlw") pod "9dc73b58-877c-4e60-8508-19ffeb7410f4" (UID: "9dc73b58-877c-4e60-8508-19ffeb7410f4"). InnerVolumeSpecName "kube-api-access-n4mlw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:14:21 crc kubenswrapper[4870]: I1014 07:14:21.041385 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="68ce28b7-83f9-4ba8-8a95-b8995f5151ca" path="/var/lib/kubelet/pods/68ce28b7-83f9-4ba8-8a95-b8995f5151ca/volumes" Oct 14 07:14:21 crc kubenswrapper[4870]: I1014 07:14:21.066837 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9dc73b58-877c-4e60-8508-19ffeb7410f4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9dc73b58-877c-4e60-8508-19ffeb7410f4" (UID: "9dc73b58-877c-4e60-8508-19ffeb7410f4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:14:21 crc kubenswrapper[4870]: I1014 07:14:21.078871 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n4mlw\" (UniqueName: \"kubernetes.io/projected/9dc73b58-877c-4e60-8508-19ffeb7410f4-kube-api-access-n4mlw\") on node \"crc\" DevicePath \"\"" Oct 14 07:14:21 crc kubenswrapper[4870]: I1014 07:14:21.078914 4870 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9dc73b58-877c-4e60-8508-19ffeb7410f4-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 07:14:21 crc kubenswrapper[4870]: I1014 07:14:21.078925 4870 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9dc73b58-877c-4e60-8508-19ffeb7410f4-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 07:14:21 crc kubenswrapper[4870]: I1014 07:14:21.148781 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5t25t" event={"ID":"9dc73b58-877c-4e60-8508-19ffeb7410f4","Type":"ContainerDied","Data":"5d02180070586fe73ce473a01193ebb219a3def5f2a64f52dc9a47a7d79c1aa6"} Oct 14 07:14:21 crc kubenswrapper[4870]: I1014 07:14:21.148835 4870 scope.go:117] "RemoveContainer" containerID="056466e4ca8980a8e9136d1cfbe342617256392c6762318dd651a7b3a6a7449b" Oct 14 07:14:21 crc kubenswrapper[4870]: I1014 07:14:21.148835 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5t25t" Oct 14 07:14:21 crc kubenswrapper[4870]: I1014 07:14:21.176495 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-5t25t"] Oct 14 07:14:21 crc kubenswrapper[4870]: I1014 07:14:21.177279 4870 scope.go:117] "RemoveContainer" containerID="4399dd1c239bb7139648df62002064ff6fce443e8519e90cee4081de1f913d0a" Oct 14 07:14:21 crc kubenswrapper[4870]: I1014 07:14:21.178107 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-5t25t"] Oct 14 07:14:21 crc kubenswrapper[4870]: I1014 07:14:21.217248 4870 scope.go:117] "RemoveContainer" containerID="d1ce24f3ae403610927a482b86d7002ae21f5ddc28d63d1f8d37d00362defa5b" Oct 14 07:14:22 crc kubenswrapper[4870]: I1014 07:14:22.155496 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-w5zvx" event={"ID":"9292dbae-ed2d-4a25-a889-b951b36614c6","Type":"ContainerStarted","Data":"ee4bfd8e9aea93183eea6572eb32d25bdd7b3d3801544fbae0f9893662ce62e0"} Oct 14 07:14:22 crc kubenswrapper[4870]: I1014 07:14:22.157767 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-ffdgs" event={"ID":"65450853-71f4-44b7-bb21-1e5c650de334","Type":"ContainerStarted","Data":"6effda8c778614cbdd7ec5b51553d2024a58da3e391ee21d69d7941c0e936293"} Oct 14 07:14:22 crc kubenswrapper[4870]: I1014 07:14:22.157898 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-ffdgs" Oct 14 07:14:22 crc kubenswrapper[4870]: I1014 07:14:22.158803 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-269h6" event={"ID":"e4f05978-9de8-4cad-89d0-8620fbe004a3","Type":"ContainerStarted","Data":"0228bfaadf67a0af21ab7604a966c94905f75aabff8bd1344bd4abbb809cb192"} Oct 14 07:14:22 crc kubenswrapper[4870]: I1014 07:14:22.160021 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-5gj9z" event={"ID":"78703da9-8ebd-4610-9d48-61c17c1eddeb","Type":"ContainerStarted","Data":"9a428880eb1b1f25e5c38b422ab73e324f75f27f937f95471aa02e65b6a1f7de"} Oct 14 07:14:22 crc kubenswrapper[4870]: I1014 07:14:22.160154 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-5gj9z" Oct 14 07:14:22 crc kubenswrapper[4870]: I1014 07:14:22.175111 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-w5zvx" podStartSLOduration=2.930433272 podStartE2EDuration="9.175091973s" podCreationTimestamp="2025-10-14 07:14:13 +0000 UTC" firstStartedPulling="2025-10-14 07:14:14.714819311 +0000 UTC m=+790.412179682" lastFinishedPulling="2025-10-14 07:14:20.959478002 +0000 UTC m=+796.656838383" observedRunningTime="2025-10-14 07:14:22.171808679 +0000 UTC m=+797.869169050" watchObservedRunningTime="2025-10-14 07:14:22.175091973 +0000 UTC m=+797.872452354" Oct 14 07:14:22 crc kubenswrapper[4870]: I1014 07:14:22.189251 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-5gj9z" podStartSLOduration=3.416144147 podStartE2EDuration="9.189232724s" podCreationTimestamp="2025-10-14 07:14:13 +0000 UTC" firstStartedPulling="2025-10-14 07:14:15.190998232 +0000 UTC m=+790.888358643" lastFinishedPulling="2025-10-14 07:14:20.964086849 +0000 UTC m=+796.661447220" observedRunningTime="2025-10-14 07:14:22.188315581 +0000 UTC m=+797.885675952" watchObservedRunningTime="2025-10-14 07:14:22.189232724 +0000 UTC m=+797.886593095" Oct 14 07:14:22 crc kubenswrapper[4870]: I1014 07:14:22.209274 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-ffdgs" podStartSLOduration=2.498950344 podStartE2EDuration="9.209253396s" podCreationTimestamp="2025-10-14 07:14:13 +0000 UTC" firstStartedPulling="2025-10-14 07:14:14.248846731 +0000 UTC m=+789.946207102" lastFinishedPulling="2025-10-14 07:14:20.959149783 +0000 UTC m=+796.656510154" observedRunningTime="2025-10-14 07:14:22.205690275 +0000 UTC m=+797.903050656" watchObservedRunningTime="2025-10-14 07:14:22.209253396 +0000 UTC m=+797.906613787" Oct 14 07:14:23 crc kubenswrapper[4870]: I1014 07:14:23.043049 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9dc73b58-877c-4e60-8508-19ffeb7410f4" path="/var/lib/kubelet/pods/9dc73b58-877c-4e60-8508-19ffeb7410f4/volumes" Oct 14 07:14:23 crc kubenswrapper[4870]: I1014 07:14:23.951153 4870 patch_prober.go:28] interesting pod/machine-config-daemon-7tvc7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 07:14:23 crc kubenswrapper[4870]: I1014 07:14:23.951214 4870 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 07:14:24 crc kubenswrapper[4870]: I1014 07:14:24.531850 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-76877f899c-x2gzs" Oct 14 07:14:24 crc kubenswrapper[4870]: I1014 07:14:24.531903 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-76877f899c-x2gzs" Oct 14 07:14:24 crc kubenswrapper[4870]: I1014 07:14:24.537649 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-76877f899c-x2gzs" Oct 14 07:14:25 crc kubenswrapper[4870]: I1014 07:14:25.179514 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-269h6" event={"ID":"e4f05978-9de8-4cad-89d0-8620fbe004a3","Type":"ContainerStarted","Data":"eaa3bf2a2687e27f433bfbaddbba37f4a1d3b82b2c2975535a64458087e9636e"} Oct 14 07:14:25 crc kubenswrapper[4870]: I1014 07:14:25.187776 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-76877f899c-x2gzs" Oct 14 07:14:25 crc kubenswrapper[4870]: I1014 07:14:25.198775 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-269h6" podStartSLOduration=2.369971036 podStartE2EDuration="12.198758126s" podCreationTimestamp="2025-10-14 07:14:13 +0000 UTC" firstStartedPulling="2025-10-14 07:14:14.642583384 +0000 UTC m=+790.339943765" lastFinishedPulling="2025-10-14 07:14:24.471370474 +0000 UTC m=+800.168730855" observedRunningTime="2025-10-14 07:14:25.197797612 +0000 UTC m=+800.895157983" watchObservedRunningTime="2025-10-14 07:14:25.198758126 +0000 UTC m=+800.896118487" Oct 14 07:14:25 crc kubenswrapper[4870]: I1014 07:14:25.265526 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-r9t2k"] Oct 14 07:14:29 crc kubenswrapper[4870]: I1014 07:14:29.243888 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-ffdgs" Oct 14 07:14:34 crc kubenswrapper[4870]: I1014 07:14:34.786168 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-5gj9z" Oct 14 07:14:48 crc kubenswrapper[4870]: I1014 07:14:48.124120 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2lsrmg"] Oct 14 07:14:48 crc kubenswrapper[4870]: E1014 07:14:48.124796 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68ce28b7-83f9-4ba8-8a95-b8995f5151ca" containerName="registry-server" Oct 14 07:14:48 crc kubenswrapper[4870]: I1014 07:14:48.124807 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="68ce28b7-83f9-4ba8-8a95-b8995f5151ca" containerName="registry-server" Oct 14 07:14:48 crc kubenswrapper[4870]: E1014 07:14:48.124819 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68ce28b7-83f9-4ba8-8a95-b8995f5151ca" containerName="extract-content" Oct 14 07:14:48 crc kubenswrapper[4870]: I1014 07:14:48.124824 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="68ce28b7-83f9-4ba8-8a95-b8995f5151ca" containerName="extract-content" Oct 14 07:14:48 crc kubenswrapper[4870]: E1014 07:14:48.124834 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68ce28b7-83f9-4ba8-8a95-b8995f5151ca" containerName="extract-utilities" Oct 14 07:14:48 crc kubenswrapper[4870]: I1014 07:14:48.124840 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="68ce28b7-83f9-4ba8-8a95-b8995f5151ca" containerName="extract-utilities" Oct 14 07:14:48 crc kubenswrapper[4870]: E1014 07:14:48.124848 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9574cf23-fd49-4791-ad7d-48e2c4bb69a9" containerName="registry-server" Oct 14 07:14:48 crc kubenswrapper[4870]: I1014 07:14:48.124853 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="9574cf23-fd49-4791-ad7d-48e2c4bb69a9" containerName="registry-server" Oct 14 07:14:48 crc kubenswrapper[4870]: E1014 07:14:48.124862 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9dc73b58-877c-4e60-8508-19ffeb7410f4" containerName="extract-utilities" Oct 14 07:14:48 crc kubenswrapper[4870]: I1014 07:14:48.124867 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="9dc73b58-877c-4e60-8508-19ffeb7410f4" containerName="extract-utilities" Oct 14 07:14:48 crc kubenswrapper[4870]: E1014 07:14:48.124878 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9dc73b58-877c-4e60-8508-19ffeb7410f4" containerName="registry-server" Oct 14 07:14:48 crc kubenswrapper[4870]: I1014 07:14:48.124883 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="9dc73b58-877c-4e60-8508-19ffeb7410f4" containerName="registry-server" Oct 14 07:14:48 crc kubenswrapper[4870]: E1014 07:14:48.124895 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9574cf23-fd49-4791-ad7d-48e2c4bb69a9" containerName="extract-utilities" Oct 14 07:14:48 crc kubenswrapper[4870]: I1014 07:14:48.124901 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="9574cf23-fd49-4791-ad7d-48e2c4bb69a9" containerName="extract-utilities" Oct 14 07:14:48 crc kubenswrapper[4870]: E1014 07:14:48.124907 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9574cf23-fd49-4791-ad7d-48e2c4bb69a9" containerName="extract-content" Oct 14 07:14:48 crc kubenswrapper[4870]: I1014 07:14:48.124912 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="9574cf23-fd49-4791-ad7d-48e2c4bb69a9" containerName="extract-content" Oct 14 07:14:48 crc kubenswrapper[4870]: E1014 07:14:48.124918 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9dc73b58-877c-4e60-8508-19ffeb7410f4" containerName="extract-content" Oct 14 07:14:48 crc kubenswrapper[4870]: I1014 07:14:48.124924 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="9dc73b58-877c-4e60-8508-19ffeb7410f4" containerName="extract-content" Oct 14 07:14:48 crc kubenswrapper[4870]: I1014 07:14:48.125017 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="68ce28b7-83f9-4ba8-8a95-b8995f5151ca" containerName="registry-server" Oct 14 07:14:48 crc kubenswrapper[4870]: I1014 07:14:48.125026 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="9574cf23-fd49-4791-ad7d-48e2c4bb69a9" containerName="registry-server" Oct 14 07:14:48 crc kubenswrapper[4870]: I1014 07:14:48.125032 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="9dc73b58-877c-4e60-8508-19ffeb7410f4" containerName="registry-server" Oct 14 07:14:48 crc kubenswrapper[4870]: I1014 07:14:48.125748 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2lsrmg" Oct 14 07:14:48 crc kubenswrapper[4870]: I1014 07:14:48.127102 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Oct 14 07:14:48 crc kubenswrapper[4870]: I1014 07:14:48.135957 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2lsrmg"] Oct 14 07:14:48 crc kubenswrapper[4870]: I1014 07:14:48.172061 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-222hg\" (UniqueName: \"kubernetes.io/projected/82fe0e08-918e-44f3-a5ac-940278d1e9c3-kube-api-access-222hg\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2lsrmg\" (UID: \"82fe0e08-918e-44f3-a5ac-940278d1e9c3\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2lsrmg" Oct 14 07:14:48 crc kubenswrapper[4870]: I1014 07:14:48.172367 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/82fe0e08-918e-44f3-a5ac-940278d1e9c3-util\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2lsrmg\" (UID: \"82fe0e08-918e-44f3-a5ac-940278d1e9c3\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2lsrmg" Oct 14 07:14:48 crc kubenswrapper[4870]: I1014 07:14:48.172578 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/82fe0e08-918e-44f3-a5ac-940278d1e9c3-bundle\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2lsrmg\" (UID: \"82fe0e08-918e-44f3-a5ac-940278d1e9c3\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2lsrmg" Oct 14 07:14:48 crc kubenswrapper[4870]: I1014 07:14:48.274128 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-222hg\" (UniqueName: \"kubernetes.io/projected/82fe0e08-918e-44f3-a5ac-940278d1e9c3-kube-api-access-222hg\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2lsrmg\" (UID: \"82fe0e08-918e-44f3-a5ac-940278d1e9c3\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2lsrmg" Oct 14 07:14:48 crc kubenswrapper[4870]: I1014 07:14:48.274203 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/82fe0e08-918e-44f3-a5ac-940278d1e9c3-util\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2lsrmg\" (UID: \"82fe0e08-918e-44f3-a5ac-940278d1e9c3\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2lsrmg" Oct 14 07:14:48 crc kubenswrapper[4870]: I1014 07:14:48.274246 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/82fe0e08-918e-44f3-a5ac-940278d1e9c3-bundle\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2lsrmg\" (UID: \"82fe0e08-918e-44f3-a5ac-940278d1e9c3\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2lsrmg" Oct 14 07:14:48 crc kubenswrapper[4870]: I1014 07:14:48.275073 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/82fe0e08-918e-44f3-a5ac-940278d1e9c3-bundle\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2lsrmg\" (UID: \"82fe0e08-918e-44f3-a5ac-940278d1e9c3\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2lsrmg" Oct 14 07:14:48 crc kubenswrapper[4870]: I1014 07:14:48.275309 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/82fe0e08-918e-44f3-a5ac-940278d1e9c3-util\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2lsrmg\" (UID: \"82fe0e08-918e-44f3-a5ac-940278d1e9c3\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2lsrmg" Oct 14 07:14:48 crc kubenswrapper[4870]: I1014 07:14:48.313111 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-222hg\" (UniqueName: \"kubernetes.io/projected/82fe0e08-918e-44f3-a5ac-940278d1e9c3-kube-api-access-222hg\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2lsrmg\" (UID: \"82fe0e08-918e-44f3-a5ac-940278d1e9c3\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2lsrmg" Oct 14 07:14:48 crc kubenswrapper[4870]: I1014 07:14:48.441362 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2lsrmg" Oct 14 07:14:48 crc kubenswrapper[4870]: I1014 07:14:48.886291 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2lsrmg"] Oct 14 07:14:49 crc kubenswrapper[4870]: I1014 07:14:49.353155 4870 generic.go:334] "Generic (PLEG): container finished" podID="82fe0e08-918e-44f3-a5ac-940278d1e9c3" containerID="5ce555e39454dc1fb1dbf9a9abca8de45355cdd85a8bfd959a87c4b431487914" exitCode=0 Oct 14 07:14:49 crc kubenswrapper[4870]: I1014 07:14:49.353276 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2lsrmg" event={"ID":"82fe0e08-918e-44f3-a5ac-940278d1e9c3","Type":"ContainerDied","Data":"5ce555e39454dc1fb1dbf9a9abca8de45355cdd85a8bfd959a87c4b431487914"} Oct 14 07:14:49 crc kubenswrapper[4870]: I1014 07:14:49.355014 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2lsrmg" event={"ID":"82fe0e08-918e-44f3-a5ac-940278d1e9c3","Type":"ContainerStarted","Data":"90aee5a3366ce454333de28831e1dd6767ba234e4556a4fee59c20c3f409bc51"} Oct 14 07:14:50 crc kubenswrapper[4870]: I1014 07:14:50.320480 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-r9t2k" podUID="b38b0739-21a6-4050-af6c-6accabefcc45" containerName="console" containerID="cri-o://326a7f990605a108949aeb6d5053860eea43df99e0119fb3ddc3efb1c5302546" gracePeriod=15 Oct 14 07:14:50 crc kubenswrapper[4870]: I1014 07:14:50.773934 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-r9t2k_b38b0739-21a6-4050-af6c-6accabefcc45/console/0.log" Oct 14 07:14:50 crc kubenswrapper[4870]: I1014 07:14:50.774476 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-r9t2k" Oct 14 07:14:50 crc kubenswrapper[4870]: I1014 07:14:50.909891 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/b38b0739-21a6-4050-af6c-6accabefcc45-console-oauth-config\") pod \"b38b0739-21a6-4050-af6c-6accabefcc45\" (UID: \"b38b0739-21a6-4050-af6c-6accabefcc45\") " Oct 14 07:14:50 crc kubenswrapper[4870]: I1014 07:14:50.909990 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/b38b0739-21a6-4050-af6c-6accabefcc45-service-ca\") pod \"b38b0739-21a6-4050-af6c-6accabefcc45\" (UID: \"b38b0739-21a6-4050-af6c-6accabefcc45\") " Oct 14 07:14:50 crc kubenswrapper[4870]: I1014 07:14:50.910043 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t8p86\" (UniqueName: \"kubernetes.io/projected/b38b0739-21a6-4050-af6c-6accabefcc45-kube-api-access-t8p86\") pod \"b38b0739-21a6-4050-af6c-6accabefcc45\" (UID: \"b38b0739-21a6-4050-af6c-6accabefcc45\") " Oct 14 07:14:50 crc kubenswrapper[4870]: I1014 07:14:50.910115 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/b38b0739-21a6-4050-af6c-6accabefcc45-console-serving-cert\") pod \"b38b0739-21a6-4050-af6c-6accabefcc45\" (UID: \"b38b0739-21a6-4050-af6c-6accabefcc45\") " Oct 14 07:14:50 crc kubenswrapper[4870]: I1014 07:14:50.910192 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b38b0739-21a6-4050-af6c-6accabefcc45-trusted-ca-bundle\") pod \"b38b0739-21a6-4050-af6c-6accabefcc45\" (UID: \"b38b0739-21a6-4050-af6c-6accabefcc45\") " Oct 14 07:14:50 crc kubenswrapper[4870]: I1014 07:14:50.910280 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/b38b0739-21a6-4050-af6c-6accabefcc45-oauth-serving-cert\") pod \"b38b0739-21a6-4050-af6c-6accabefcc45\" (UID: \"b38b0739-21a6-4050-af6c-6accabefcc45\") " Oct 14 07:14:50 crc kubenswrapper[4870]: I1014 07:14:50.910308 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/b38b0739-21a6-4050-af6c-6accabefcc45-console-config\") pod \"b38b0739-21a6-4050-af6c-6accabefcc45\" (UID: \"b38b0739-21a6-4050-af6c-6accabefcc45\") " Oct 14 07:14:50 crc kubenswrapper[4870]: I1014 07:14:50.910931 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b38b0739-21a6-4050-af6c-6accabefcc45-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "b38b0739-21a6-4050-af6c-6accabefcc45" (UID: "b38b0739-21a6-4050-af6c-6accabefcc45"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:14:50 crc kubenswrapper[4870]: I1014 07:14:50.910960 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b38b0739-21a6-4050-af6c-6accabefcc45-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "b38b0739-21a6-4050-af6c-6accabefcc45" (UID: "b38b0739-21a6-4050-af6c-6accabefcc45"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:14:50 crc kubenswrapper[4870]: I1014 07:14:50.911331 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b38b0739-21a6-4050-af6c-6accabefcc45-service-ca" (OuterVolumeSpecName: "service-ca") pod "b38b0739-21a6-4050-af6c-6accabefcc45" (UID: "b38b0739-21a6-4050-af6c-6accabefcc45"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:14:50 crc kubenswrapper[4870]: I1014 07:14:50.911991 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b38b0739-21a6-4050-af6c-6accabefcc45-console-config" (OuterVolumeSpecName: "console-config") pod "b38b0739-21a6-4050-af6c-6accabefcc45" (UID: "b38b0739-21a6-4050-af6c-6accabefcc45"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:14:50 crc kubenswrapper[4870]: I1014 07:14:50.919613 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b38b0739-21a6-4050-af6c-6accabefcc45-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "b38b0739-21a6-4050-af6c-6accabefcc45" (UID: "b38b0739-21a6-4050-af6c-6accabefcc45"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:14:50 crc kubenswrapper[4870]: I1014 07:14:50.919744 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b38b0739-21a6-4050-af6c-6accabefcc45-kube-api-access-t8p86" (OuterVolumeSpecName: "kube-api-access-t8p86") pod "b38b0739-21a6-4050-af6c-6accabefcc45" (UID: "b38b0739-21a6-4050-af6c-6accabefcc45"). InnerVolumeSpecName "kube-api-access-t8p86". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:14:50 crc kubenswrapper[4870]: I1014 07:14:50.920927 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b38b0739-21a6-4050-af6c-6accabefcc45-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "b38b0739-21a6-4050-af6c-6accabefcc45" (UID: "b38b0739-21a6-4050-af6c-6accabefcc45"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:14:51 crc kubenswrapper[4870]: I1014 07:14:51.011410 4870 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/b38b0739-21a6-4050-af6c-6accabefcc45-console-oauth-config\") on node \"crc\" DevicePath \"\"" Oct 14 07:14:51 crc kubenswrapper[4870]: I1014 07:14:51.011497 4870 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/b38b0739-21a6-4050-af6c-6accabefcc45-service-ca\") on node \"crc\" DevicePath \"\"" Oct 14 07:14:51 crc kubenswrapper[4870]: I1014 07:14:51.011517 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t8p86\" (UniqueName: \"kubernetes.io/projected/b38b0739-21a6-4050-af6c-6accabefcc45-kube-api-access-t8p86\") on node \"crc\" DevicePath \"\"" Oct 14 07:14:51 crc kubenswrapper[4870]: I1014 07:14:51.011532 4870 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/b38b0739-21a6-4050-af6c-6accabefcc45-console-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 14 07:14:51 crc kubenswrapper[4870]: I1014 07:14:51.011549 4870 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b38b0739-21a6-4050-af6c-6accabefcc45-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 07:14:51 crc kubenswrapper[4870]: I1014 07:14:51.011564 4870 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/b38b0739-21a6-4050-af6c-6accabefcc45-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 14 07:14:51 crc kubenswrapper[4870]: I1014 07:14:51.011578 4870 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/b38b0739-21a6-4050-af6c-6accabefcc45-console-config\") on node \"crc\" DevicePath \"\"" Oct 14 07:14:51 crc kubenswrapper[4870]: I1014 07:14:51.372838 4870 generic.go:334] "Generic (PLEG): container finished" podID="82fe0e08-918e-44f3-a5ac-940278d1e9c3" containerID="95e24c69a308f93d2278838f7ef152b22ea77646d72ed599aaeb3e6a7231e22f" exitCode=0 Oct 14 07:14:51 crc kubenswrapper[4870]: I1014 07:14:51.372893 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2lsrmg" event={"ID":"82fe0e08-918e-44f3-a5ac-940278d1e9c3","Type":"ContainerDied","Data":"95e24c69a308f93d2278838f7ef152b22ea77646d72ed599aaeb3e6a7231e22f"} Oct 14 07:14:51 crc kubenswrapper[4870]: I1014 07:14:51.378588 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-r9t2k_b38b0739-21a6-4050-af6c-6accabefcc45/console/0.log" Oct 14 07:14:51 crc kubenswrapper[4870]: I1014 07:14:51.378659 4870 generic.go:334] "Generic (PLEG): container finished" podID="b38b0739-21a6-4050-af6c-6accabefcc45" containerID="326a7f990605a108949aeb6d5053860eea43df99e0119fb3ddc3efb1c5302546" exitCode=2 Oct 14 07:14:51 crc kubenswrapper[4870]: I1014 07:14:51.378708 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-r9t2k" event={"ID":"b38b0739-21a6-4050-af6c-6accabefcc45","Type":"ContainerDied","Data":"326a7f990605a108949aeb6d5053860eea43df99e0119fb3ddc3efb1c5302546"} Oct 14 07:14:51 crc kubenswrapper[4870]: I1014 07:14:51.378748 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-r9t2k" event={"ID":"b38b0739-21a6-4050-af6c-6accabefcc45","Type":"ContainerDied","Data":"e1b6a976c48ab09e1c019e72e377ebd7c1e1da64ad1c680caf61df4052a61312"} Oct 14 07:14:51 crc kubenswrapper[4870]: I1014 07:14:51.378778 4870 scope.go:117] "RemoveContainer" containerID="326a7f990605a108949aeb6d5053860eea43df99e0119fb3ddc3efb1c5302546" Oct 14 07:14:51 crc kubenswrapper[4870]: I1014 07:14:51.378716 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-r9t2k" Oct 14 07:14:51 crc kubenswrapper[4870]: I1014 07:14:51.425686 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-r9t2k"] Oct 14 07:14:51 crc kubenswrapper[4870]: I1014 07:14:51.429818 4870 scope.go:117] "RemoveContainer" containerID="326a7f990605a108949aeb6d5053860eea43df99e0119fb3ddc3efb1c5302546" Oct 14 07:14:51 crc kubenswrapper[4870]: E1014 07:14:51.430577 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"326a7f990605a108949aeb6d5053860eea43df99e0119fb3ddc3efb1c5302546\": container with ID starting with 326a7f990605a108949aeb6d5053860eea43df99e0119fb3ddc3efb1c5302546 not found: ID does not exist" containerID="326a7f990605a108949aeb6d5053860eea43df99e0119fb3ddc3efb1c5302546" Oct 14 07:14:51 crc kubenswrapper[4870]: I1014 07:14:51.430745 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"326a7f990605a108949aeb6d5053860eea43df99e0119fb3ddc3efb1c5302546"} err="failed to get container status \"326a7f990605a108949aeb6d5053860eea43df99e0119fb3ddc3efb1c5302546\": rpc error: code = NotFound desc = could not find container \"326a7f990605a108949aeb6d5053860eea43df99e0119fb3ddc3efb1c5302546\": container with ID starting with 326a7f990605a108949aeb6d5053860eea43df99e0119fb3ddc3efb1c5302546 not found: ID does not exist" Oct 14 07:14:51 crc kubenswrapper[4870]: I1014 07:14:51.431955 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-r9t2k"] Oct 14 07:14:52 crc kubenswrapper[4870]: I1014 07:14:52.390708 4870 generic.go:334] "Generic (PLEG): container finished" podID="82fe0e08-918e-44f3-a5ac-940278d1e9c3" containerID="8951b1fed845a9b05e8495ca3a82697fee0cb26dbc9e5e811c45fc30637cf749" exitCode=0 Oct 14 07:14:52 crc kubenswrapper[4870]: I1014 07:14:52.390766 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2lsrmg" event={"ID":"82fe0e08-918e-44f3-a5ac-940278d1e9c3","Type":"ContainerDied","Data":"8951b1fed845a9b05e8495ca3a82697fee0cb26dbc9e5e811c45fc30637cf749"} Oct 14 07:14:53 crc kubenswrapper[4870]: I1014 07:14:53.048139 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b38b0739-21a6-4050-af6c-6accabefcc45" path="/var/lib/kubelet/pods/b38b0739-21a6-4050-af6c-6accabefcc45/volumes" Oct 14 07:14:53 crc kubenswrapper[4870]: I1014 07:14:53.694646 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2lsrmg" Oct 14 07:14:53 crc kubenswrapper[4870]: I1014 07:14:53.849420 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/82fe0e08-918e-44f3-a5ac-940278d1e9c3-bundle\") pod \"82fe0e08-918e-44f3-a5ac-940278d1e9c3\" (UID: \"82fe0e08-918e-44f3-a5ac-940278d1e9c3\") " Oct 14 07:14:53 crc kubenswrapper[4870]: I1014 07:14:53.849507 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/82fe0e08-918e-44f3-a5ac-940278d1e9c3-util\") pod \"82fe0e08-918e-44f3-a5ac-940278d1e9c3\" (UID: \"82fe0e08-918e-44f3-a5ac-940278d1e9c3\") " Oct 14 07:14:53 crc kubenswrapper[4870]: I1014 07:14:53.849672 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-222hg\" (UniqueName: \"kubernetes.io/projected/82fe0e08-918e-44f3-a5ac-940278d1e9c3-kube-api-access-222hg\") pod \"82fe0e08-918e-44f3-a5ac-940278d1e9c3\" (UID: \"82fe0e08-918e-44f3-a5ac-940278d1e9c3\") " Oct 14 07:14:53 crc kubenswrapper[4870]: I1014 07:14:53.851017 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/82fe0e08-918e-44f3-a5ac-940278d1e9c3-bundle" (OuterVolumeSpecName: "bundle") pod "82fe0e08-918e-44f3-a5ac-940278d1e9c3" (UID: "82fe0e08-918e-44f3-a5ac-940278d1e9c3"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:14:53 crc kubenswrapper[4870]: I1014 07:14:53.857145 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/82fe0e08-918e-44f3-a5ac-940278d1e9c3-kube-api-access-222hg" (OuterVolumeSpecName: "kube-api-access-222hg") pod "82fe0e08-918e-44f3-a5ac-940278d1e9c3" (UID: "82fe0e08-918e-44f3-a5ac-940278d1e9c3"). InnerVolumeSpecName "kube-api-access-222hg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:14:53 crc kubenswrapper[4870]: I1014 07:14:53.869786 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/82fe0e08-918e-44f3-a5ac-940278d1e9c3-util" (OuterVolumeSpecName: "util") pod "82fe0e08-918e-44f3-a5ac-940278d1e9c3" (UID: "82fe0e08-918e-44f3-a5ac-940278d1e9c3"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:14:53 crc kubenswrapper[4870]: I1014 07:14:53.950956 4870 patch_prober.go:28] interesting pod/machine-config-daemon-7tvc7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 07:14:53 crc kubenswrapper[4870]: I1014 07:14:53.951022 4870 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 07:14:53 crc kubenswrapper[4870]: I1014 07:14:53.951213 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-222hg\" (UniqueName: \"kubernetes.io/projected/82fe0e08-918e-44f3-a5ac-940278d1e9c3-kube-api-access-222hg\") on node \"crc\" DevicePath \"\"" Oct 14 07:14:53 crc kubenswrapper[4870]: I1014 07:14:53.951244 4870 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/82fe0e08-918e-44f3-a5ac-940278d1e9c3-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 07:14:53 crc kubenswrapper[4870]: I1014 07:14:53.951259 4870 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/82fe0e08-918e-44f3-a5ac-940278d1e9c3-util\") on node \"crc\" DevicePath \"\"" Oct 14 07:14:54 crc kubenswrapper[4870]: I1014 07:14:54.406582 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2lsrmg" event={"ID":"82fe0e08-918e-44f3-a5ac-940278d1e9c3","Type":"ContainerDied","Data":"90aee5a3366ce454333de28831e1dd6767ba234e4556a4fee59c20c3f409bc51"} Oct 14 07:14:54 crc kubenswrapper[4870]: I1014 07:14:54.406652 4870 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="90aee5a3366ce454333de28831e1dd6767ba234e4556a4fee59c20c3f409bc51" Oct 14 07:14:54 crc kubenswrapper[4870]: I1014 07:14:54.406684 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2lsrmg" Oct 14 07:14:55 crc kubenswrapper[4870]: I1014 07:14:55.888101 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-vnhs7"] Oct 14 07:14:55 crc kubenswrapper[4870]: E1014 07:14:55.889425 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82fe0e08-918e-44f3-a5ac-940278d1e9c3" containerName="util" Oct 14 07:14:55 crc kubenswrapper[4870]: I1014 07:14:55.889574 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="82fe0e08-918e-44f3-a5ac-940278d1e9c3" containerName="util" Oct 14 07:14:55 crc kubenswrapper[4870]: E1014 07:14:55.890308 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82fe0e08-918e-44f3-a5ac-940278d1e9c3" containerName="pull" Oct 14 07:14:55 crc kubenswrapper[4870]: I1014 07:14:55.890341 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="82fe0e08-918e-44f3-a5ac-940278d1e9c3" containerName="pull" Oct 14 07:14:55 crc kubenswrapper[4870]: E1014 07:14:55.890391 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b38b0739-21a6-4050-af6c-6accabefcc45" containerName="console" Oct 14 07:14:55 crc kubenswrapper[4870]: I1014 07:14:55.890472 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="b38b0739-21a6-4050-af6c-6accabefcc45" containerName="console" Oct 14 07:14:55 crc kubenswrapper[4870]: E1014 07:14:55.890499 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82fe0e08-918e-44f3-a5ac-940278d1e9c3" containerName="extract" Oct 14 07:14:55 crc kubenswrapper[4870]: I1014 07:14:55.890529 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="82fe0e08-918e-44f3-a5ac-940278d1e9c3" containerName="extract" Oct 14 07:14:55 crc kubenswrapper[4870]: I1014 07:14:55.890732 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="82fe0e08-918e-44f3-a5ac-940278d1e9c3" containerName="extract" Oct 14 07:14:55 crc kubenswrapper[4870]: I1014 07:14:55.890756 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="b38b0739-21a6-4050-af6c-6accabefcc45" containerName="console" Oct 14 07:14:55 crc kubenswrapper[4870]: I1014 07:14:55.893968 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vnhs7" Oct 14 07:14:55 crc kubenswrapper[4870]: I1014 07:14:55.902842 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-vnhs7"] Oct 14 07:14:56 crc kubenswrapper[4870]: I1014 07:14:56.084289 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/00be94d3-e58b-4871-8f4f-97ff5fc158b7-catalog-content\") pod \"redhat-marketplace-vnhs7\" (UID: \"00be94d3-e58b-4871-8f4f-97ff5fc158b7\") " pod="openshift-marketplace/redhat-marketplace-vnhs7" Oct 14 07:14:56 crc kubenswrapper[4870]: I1014 07:14:56.084759 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mcssd\" (UniqueName: \"kubernetes.io/projected/00be94d3-e58b-4871-8f4f-97ff5fc158b7-kube-api-access-mcssd\") pod \"redhat-marketplace-vnhs7\" (UID: \"00be94d3-e58b-4871-8f4f-97ff5fc158b7\") " pod="openshift-marketplace/redhat-marketplace-vnhs7" Oct 14 07:14:56 crc kubenswrapper[4870]: I1014 07:14:56.084977 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/00be94d3-e58b-4871-8f4f-97ff5fc158b7-utilities\") pod \"redhat-marketplace-vnhs7\" (UID: \"00be94d3-e58b-4871-8f4f-97ff5fc158b7\") " pod="openshift-marketplace/redhat-marketplace-vnhs7" Oct 14 07:14:56 crc kubenswrapper[4870]: I1014 07:14:56.185422 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/00be94d3-e58b-4871-8f4f-97ff5fc158b7-catalog-content\") pod \"redhat-marketplace-vnhs7\" (UID: \"00be94d3-e58b-4871-8f4f-97ff5fc158b7\") " pod="openshift-marketplace/redhat-marketplace-vnhs7" Oct 14 07:14:56 crc kubenswrapper[4870]: I1014 07:14:56.185524 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mcssd\" (UniqueName: \"kubernetes.io/projected/00be94d3-e58b-4871-8f4f-97ff5fc158b7-kube-api-access-mcssd\") pod \"redhat-marketplace-vnhs7\" (UID: \"00be94d3-e58b-4871-8f4f-97ff5fc158b7\") " pod="openshift-marketplace/redhat-marketplace-vnhs7" Oct 14 07:14:56 crc kubenswrapper[4870]: I1014 07:14:56.185566 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/00be94d3-e58b-4871-8f4f-97ff5fc158b7-utilities\") pod \"redhat-marketplace-vnhs7\" (UID: \"00be94d3-e58b-4871-8f4f-97ff5fc158b7\") " pod="openshift-marketplace/redhat-marketplace-vnhs7" Oct 14 07:14:56 crc kubenswrapper[4870]: I1014 07:14:56.185898 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/00be94d3-e58b-4871-8f4f-97ff5fc158b7-catalog-content\") pod \"redhat-marketplace-vnhs7\" (UID: \"00be94d3-e58b-4871-8f4f-97ff5fc158b7\") " pod="openshift-marketplace/redhat-marketplace-vnhs7" Oct 14 07:14:56 crc kubenswrapper[4870]: I1014 07:14:56.186038 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/00be94d3-e58b-4871-8f4f-97ff5fc158b7-utilities\") pod \"redhat-marketplace-vnhs7\" (UID: \"00be94d3-e58b-4871-8f4f-97ff5fc158b7\") " pod="openshift-marketplace/redhat-marketplace-vnhs7" Oct 14 07:14:56 crc kubenswrapper[4870]: I1014 07:14:56.204690 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mcssd\" (UniqueName: \"kubernetes.io/projected/00be94d3-e58b-4871-8f4f-97ff5fc158b7-kube-api-access-mcssd\") pod \"redhat-marketplace-vnhs7\" (UID: \"00be94d3-e58b-4871-8f4f-97ff5fc158b7\") " pod="openshift-marketplace/redhat-marketplace-vnhs7" Oct 14 07:14:56 crc kubenswrapper[4870]: I1014 07:14:56.249355 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vnhs7" Oct 14 07:14:56 crc kubenswrapper[4870]: I1014 07:14:56.700121 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-vnhs7"] Oct 14 07:14:57 crc kubenswrapper[4870]: I1014 07:14:57.424035 4870 generic.go:334] "Generic (PLEG): container finished" podID="00be94d3-e58b-4871-8f4f-97ff5fc158b7" containerID="729460503d7cb277fd8658c04d4f9cd53a2598d73dcb80e73f41d4409133cf2d" exitCode=0 Oct 14 07:14:57 crc kubenswrapper[4870]: I1014 07:14:57.424140 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vnhs7" event={"ID":"00be94d3-e58b-4871-8f4f-97ff5fc158b7","Type":"ContainerDied","Data":"729460503d7cb277fd8658c04d4f9cd53a2598d73dcb80e73f41d4409133cf2d"} Oct 14 07:14:57 crc kubenswrapper[4870]: I1014 07:14:57.424497 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vnhs7" event={"ID":"00be94d3-e58b-4871-8f4f-97ff5fc158b7","Type":"ContainerStarted","Data":"0b222e5955c9db35f8b860a4e48a901b8c7d85ee85a5ce85af91bd26be536a6f"} Oct 14 07:14:58 crc kubenswrapper[4870]: I1014 07:14:58.432172 4870 generic.go:334] "Generic (PLEG): container finished" podID="00be94d3-e58b-4871-8f4f-97ff5fc158b7" containerID="7b7e30dee0ab31ad34f62d92a825cd71cb9afc7e6038bd4bd4d92f23eea18f19" exitCode=0 Oct 14 07:14:58 crc kubenswrapper[4870]: I1014 07:14:58.432216 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vnhs7" event={"ID":"00be94d3-e58b-4871-8f4f-97ff5fc158b7","Type":"ContainerDied","Data":"7b7e30dee0ab31ad34f62d92a825cd71cb9afc7e6038bd4bd4d92f23eea18f19"} Oct 14 07:14:59 crc kubenswrapper[4870]: I1014 07:14:59.445620 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vnhs7" event={"ID":"00be94d3-e58b-4871-8f4f-97ff5fc158b7","Type":"ContainerStarted","Data":"9188753dd58084edda9d550b40c978e8337e7873471f0ecdb2ef435ea77c51ee"} Oct 14 07:14:59 crc kubenswrapper[4870]: I1014 07:14:59.461751 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-vnhs7" podStartSLOduration=2.891701948 podStartE2EDuration="4.461733976s" podCreationTimestamp="2025-10-14 07:14:55 +0000 UTC" firstStartedPulling="2025-10-14 07:14:57.426527138 +0000 UTC m=+833.123887519" lastFinishedPulling="2025-10-14 07:14:58.996559176 +0000 UTC m=+834.693919547" observedRunningTime="2025-10-14 07:14:59.460496985 +0000 UTC m=+835.157857376" watchObservedRunningTime="2025-10-14 07:14:59.461733976 +0000 UTC m=+835.159094357" Oct 14 07:15:00 crc kubenswrapper[4870]: I1014 07:15:00.181955 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29340435-g6qxb"] Oct 14 07:15:00 crc kubenswrapper[4870]: I1014 07:15:00.182943 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29340435-g6qxb" Oct 14 07:15:00 crc kubenswrapper[4870]: I1014 07:15:00.184626 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 14 07:15:00 crc kubenswrapper[4870]: I1014 07:15:00.184818 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 14 07:15:00 crc kubenswrapper[4870]: I1014 07:15:00.191607 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29340435-g6qxb"] Oct 14 07:15:00 crc kubenswrapper[4870]: I1014 07:15:00.234569 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mrbvw\" (UniqueName: \"kubernetes.io/projected/89fe6660-6f16-4ddb-9b5c-1582231f5f30-kube-api-access-mrbvw\") pod \"collect-profiles-29340435-g6qxb\" (UID: \"89fe6660-6f16-4ddb-9b5c-1582231f5f30\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340435-g6qxb" Oct 14 07:15:00 crc kubenswrapper[4870]: I1014 07:15:00.234721 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/89fe6660-6f16-4ddb-9b5c-1582231f5f30-secret-volume\") pod \"collect-profiles-29340435-g6qxb\" (UID: \"89fe6660-6f16-4ddb-9b5c-1582231f5f30\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340435-g6qxb" Oct 14 07:15:00 crc kubenswrapper[4870]: I1014 07:15:00.234810 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/89fe6660-6f16-4ddb-9b5c-1582231f5f30-config-volume\") pod \"collect-profiles-29340435-g6qxb\" (UID: \"89fe6660-6f16-4ddb-9b5c-1582231f5f30\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340435-g6qxb" Oct 14 07:15:00 crc kubenswrapper[4870]: I1014 07:15:00.335778 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/89fe6660-6f16-4ddb-9b5c-1582231f5f30-config-volume\") pod \"collect-profiles-29340435-g6qxb\" (UID: \"89fe6660-6f16-4ddb-9b5c-1582231f5f30\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340435-g6qxb" Oct 14 07:15:00 crc kubenswrapper[4870]: I1014 07:15:00.335843 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mrbvw\" (UniqueName: \"kubernetes.io/projected/89fe6660-6f16-4ddb-9b5c-1582231f5f30-kube-api-access-mrbvw\") pod \"collect-profiles-29340435-g6qxb\" (UID: \"89fe6660-6f16-4ddb-9b5c-1582231f5f30\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340435-g6qxb" Oct 14 07:15:00 crc kubenswrapper[4870]: I1014 07:15:00.335894 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/89fe6660-6f16-4ddb-9b5c-1582231f5f30-secret-volume\") pod \"collect-profiles-29340435-g6qxb\" (UID: \"89fe6660-6f16-4ddb-9b5c-1582231f5f30\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340435-g6qxb" Oct 14 07:15:00 crc kubenswrapper[4870]: I1014 07:15:00.336812 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/89fe6660-6f16-4ddb-9b5c-1582231f5f30-config-volume\") pod \"collect-profiles-29340435-g6qxb\" (UID: \"89fe6660-6f16-4ddb-9b5c-1582231f5f30\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340435-g6qxb" Oct 14 07:15:00 crc kubenswrapper[4870]: I1014 07:15:00.347378 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/89fe6660-6f16-4ddb-9b5c-1582231f5f30-secret-volume\") pod \"collect-profiles-29340435-g6qxb\" (UID: \"89fe6660-6f16-4ddb-9b5c-1582231f5f30\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340435-g6qxb" Oct 14 07:15:00 crc kubenswrapper[4870]: I1014 07:15:00.352042 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mrbvw\" (UniqueName: \"kubernetes.io/projected/89fe6660-6f16-4ddb-9b5c-1582231f5f30-kube-api-access-mrbvw\") pod \"collect-profiles-29340435-g6qxb\" (UID: \"89fe6660-6f16-4ddb-9b5c-1582231f5f30\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340435-g6qxb" Oct 14 07:15:00 crc kubenswrapper[4870]: I1014 07:15:00.501291 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29340435-g6qxb" Oct 14 07:15:00 crc kubenswrapper[4870]: I1014 07:15:00.898815 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29340435-g6qxb"] Oct 14 07:15:00 crc kubenswrapper[4870]: W1014 07:15:00.904231 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod89fe6660_6f16_4ddb_9b5c_1582231f5f30.slice/crio-bb4c5a4693eb6c5dbe677b276665adf1672adb43a2f006de26e7428de201e100 WatchSource:0}: Error finding container bb4c5a4693eb6c5dbe677b276665adf1672adb43a2f006de26e7428de201e100: Status 404 returned error can't find the container with id bb4c5a4693eb6c5dbe677b276665adf1672adb43a2f006de26e7428de201e100 Oct 14 07:15:01 crc kubenswrapper[4870]: I1014 07:15:01.457749 4870 generic.go:334] "Generic (PLEG): container finished" podID="89fe6660-6f16-4ddb-9b5c-1582231f5f30" containerID="2d804bb157221ca38727691cccd928f8917bce656da1d8d959887b79b3c20d34" exitCode=0 Oct 14 07:15:01 crc kubenswrapper[4870]: I1014 07:15:01.457797 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29340435-g6qxb" event={"ID":"89fe6660-6f16-4ddb-9b5c-1582231f5f30","Type":"ContainerDied","Data":"2d804bb157221ca38727691cccd928f8917bce656da1d8d959887b79b3c20d34"} Oct 14 07:15:01 crc kubenswrapper[4870]: I1014 07:15:01.457828 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29340435-g6qxb" event={"ID":"89fe6660-6f16-4ddb-9b5c-1582231f5f30","Type":"ContainerStarted","Data":"bb4c5a4693eb6c5dbe677b276665adf1672adb43a2f006de26e7428de201e100"} Oct 14 07:15:02 crc kubenswrapper[4870]: I1014 07:15:02.081668 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-f459659f5-vgrhm"] Oct 14 07:15:02 crc kubenswrapper[4870]: I1014 07:15:02.082586 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-f459659f5-vgrhm" Oct 14 07:15:02 crc kubenswrapper[4870]: I1014 07:15:02.084412 4870 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-7dqmz" Oct 14 07:15:02 crc kubenswrapper[4870]: I1014 07:15:02.085320 4870 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Oct 14 07:15:02 crc kubenswrapper[4870]: I1014 07:15:02.085842 4870 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Oct 14 07:15:02 crc kubenswrapper[4870]: I1014 07:15:02.085867 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Oct 14 07:15:02 crc kubenswrapper[4870]: I1014 07:15:02.086071 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Oct 14 07:15:02 crc kubenswrapper[4870]: I1014 07:15:02.098083 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-f459659f5-vgrhm"] Oct 14 07:15:02 crc kubenswrapper[4870]: I1014 07:15:02.257837 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/1150555d-bbee-49bc-b90c-69e30c971a19-apiservice-cert\") pod \"metallb-operator-controller-manager-f459659f5-vgrhm\" (UID: \"1150555d-bbee-49bc-b90c-69e30c971a19\") " pod="metallb-system/metallb-operator-controller-manager-f459659f5-vgrhm" Oct 14 07:15:02 crc kubenswrapper[4870]: I1014 07:15:02.257936 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vv7xh\" (UniqueName: \"kubernetes.io/projected/1150555d-bbee-49bc-b90c-69e30c971a19-kube-api-access-vv7xh\") pod \"metallb-operator-controller-manager-f459659f5-vgrhm\" (UID: \"1150555d-bbee-49bc-b90c-69e30c971a19\") " pod="metallb-system/metallb-operator-controller-manager-f459659f5-vgrhm" Oct 14 07:15:02 crc kubenswrapper[4870]: I1014 07:15:02.257972 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/1150555d-bbee-49bc-b90c-69e30c971a19-webhook-cert\") pod \"metallb-operator-controller-manager-f459659f5-vgrhm\" (UID: \"1150555d-bbee-49bc-b90c-69e30c971a19\") " pod="metallb-system/metallb-operator-controller-manager-f459659f5-vgrhm" Oct 14 07:15:02 crc kubenswrapper[4870]: I1014 07:15:02.322808 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-7544c98cdc-r5nkv"] Oct 14 07:15:02 crc kubenswrapper[4870]: I1014 07:15:02.323467 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-7544c98cdc-r5nkv" Oct 14 07:15:02 crc kubenswrapper[4870]: I1014 07:15:02.326756 4870 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Oct 14 07:15:02 crc kubenswrapper[4870]: I1014 07:15:02.326982 4870 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Oct 14 07:15:02 crc kubenswrapper[4870]: I1014 07:15:02.327259 4870 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-bxpgc" Oct 14 07:15:02 crc kubenswrapper[4870]: I1014 07:15:02.341597 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-7544c98cdc-r5nkv"] Oct 14 07:15:02 crc kubenswrapper[4870]: I1014 07:15:02.359572 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vv7xh\" (UniqueName: \"kubernetes.io/projected/1150555d-bbee-49bc-b90c-69e30c971a19-kube-api-access-vv7xh\") pod \"metallb-operator-controller-manager-f459659f5-vgrhm\" (UID: \"1150555d-bbee-49bc-b90c-69e30c971a19\") " pod="metallb-system/metallb-operator-controller-manager-f459659f5-vgrhm" Oct 14 07:15:02 crc kubenswrapper[4870]: I1014 07:15:02.359767 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/1150555d-bbee-49bc-b90c-69e30c971a19-webhook-cert\") pod \"metallb-operator-controller-manager-f459659f5-vgrhm\" (UID: \"1150555d-bbee-49bc-b90c-69e30c971a19\") " pod="metallb-system/metallb-operator-controller-manager-f459659f5-vgrhm" Oct 14 07:15:02 crc kubenswrapper[4870]: I1014 07:15:02.360763 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/1150555d-bbee-49bc-b90c-69e30c971a19-apiservice-cert\") pod \"metallb-operator-controller-manager-f459659f5-vgrhm\" (UID: \"1150555d-bbee-49bc-b90c-69e30c971a19\") " pod="metallb-system/metallb-operator-controller-manager-f459659f5-vgrhm" Oct 14 07:15:02 crc kubenswrapper[4870]: I1014 07:15:02.366256 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/1150555d-bbee-49bc-b90c-69e30c971a19-apiservice-cert\") pod \"metallb-operator-controller-manager-f459659f5-vgrhm\" (UID: \"1150555d-bbee-49bc-b90c-69e30c971a19\") " pod="metallb-system/metallb-operator-controller-manager-f459659f5-vgrhm" Oct 14 07:15:02 crc kubenswrapper[4870]: I1014 07:15:02.376404 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/1150555d-bbee-49bc-b90c-69e30c971a19-webhook-cert\") pod \"metallb-operator-controller-manager-f459659f5-vgrhm\" (UID: \"1150555d-bbee-49bc-b90c-69e30c971a19\") " pod="metallb-system/metallb-operator-controller-manager-f459659f5-vgrhm" Oct 14 07:15:02 crc kubenswrapper[4870]: I1014 07:15:02.380614 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vv7xh\" (UniqueName: \"kubernetes.io/projected/1150555d-bbee-49bc-b90c-69e30c971a19-kube-api-access-vv7xh\") pod \"metallb-operator-controller-manager-f459659f5-vgrhm\" (UID: \"1150555d-bbee-49bc-b90c-69e30c971a19\") " pod="metallb-system/metallb-operator-controller-manager-f459659f5-vgrhm" Oct 14 07:15:02 crc kubenswrapper[4870]: I1014 07:15:02.402507 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-f459659f5-vgrhm" Oct 14 07:15:02 crc kubenswrapper[4870]: I1014 07:15:02.461597 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/c1cd26c0-7d04-45fb-b171-d3398fcf8d24-webhook-cert\") pod \"metallb-operator-webhook-server-7544c98cdc-r5nkv\" (UID: \"c1cd26c0-7d04-45fb-b171-d3398fcf8d24\") " pod="metallb-system/metallb-operator-webhook-server-7544c98cdc-r5nkv" Oct 14 07:15:02 crc kubenswrapper[4870]: I1014 07:15:02.461663 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b8nsw\" (UniqueName: \"kubernetes.io/projected/c1cd26c0-7d04-45fb-b171-d3398fcf8d24-kube-api-access-b8nsw\") pod \"metallb-operator-webhook-server-7544c98cdc-r5nkv\" (UID: \"c1cd26c0-7d04-45fb-b171-d3398fcf8d24\") " pod="metallb-system/metallb-operator-webhook-server-7544c98cdc-r5nkv" Oct 14 07:15:02 crc kubenswrapper[4870]: I1014 07:15:02.461687 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/c1cd26c0-7d04-45fb-b171-d3398fcf8d24-apiservice-cert\") pod \"metallb-operator-webhook-server-7544c98cdc-r5nkv\" (UID: \"c1cd26c0-7d04-45fb-b171-d3398fcf8d24\") " pod="metallb-system/metallb-operator-webhook-server-7544c98cdc-r5nkv" Oct 14 07:15:02 crc kubenswrapper[4870]: I1014 07:15:02.572223 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/c1cd26c0-7d04-45fb-b171-d3398fcf8d24-webhook-cert\") pod \"metallb-operator-webhook-server-7544c98cdc-r5nkv\" (UID: \"c1cd26c0-7d04-45fb-b171-d3398fcf8d24\") " pod="metallb-system/metallb-operator-webhook-server-7544c98cdc-r5nkv" Oct 14 07:15:02 crc kubenswrapper[4870]: I1014 07:15:02.572289 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b8nsw\" (UniqueName: \"kubernetes.io/projected/c1cd26c0-7d04-45fb-b171-d3398fcf8d24-kube-api-access-b8nsw\") pod \"metallb-operator-webhook-server-7544c98cdc-r5nkv\" (UID: \"c1cd26c0-7d04-45fb-b171-d3398fcf8d24\") " pod="metallb-system/metallb-operator-webhook-server-7544c98cdc-r5nkv" Oct 14 07:15:02 crc kubenswrapper[4870]: I1014 07:15:02.572320 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/c1cd26c0-7d04-45fb-b171-d3398fcf8d24-apiservice-cert\") pod \"metallb-operator-webhook-server-7544c98cdc-r5nkv\" (UID: \"c1cd26c0-7d04-45fb-b171-d3398fcf8d24\") " pod="metallb-system/metallb-operator-webhook-server-7544c98cdc-r5nkv" Oct 14 07:15:02 crc kubenswrapper[4870]: I1014 07:15:02.578450 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/c1cd26c0-7d04-45fb-b171-d3398fcf8d24-webhook-cert\") pod \"metallb-operator-webhook-server-7544c98cdc-r5nkv\" (UID: \"c1cd26c0-7d04-45fb-b171-d3398fcf8d24\") " pod="metallb-system/metallb-operator-webhook-server-7544c98cdc-r5nkv" Oct 14 07:15:02 crc kubenswrapper[4870]: I1014 07:15:02.579987 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/c1cd26c0-7d04-45fb-b171-d3398fcf8d24-apiservice-cert\") pod \"metallb-operator-webhook-server-7544c98cdc-r5nkv\" (UID: \"c1cd26c0-7d04-45fb-b171-d3398fcf8d24\") " pod="metallb-system/metallb-operator-webhook-server-7544c98cdc-r5nkv" Oct 14 07:15:02 crc kubenswrapper[4870]: I1014 07:15:02.609242 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b8nsw\" (UniqueName: \"kubernetes.io/projected/c1cd26c0-7d04-45fb-b171-d3398fcf8d24-kube-api-access-b8nsw\") pod \"metallb-operator-webhook-server-7544c98cdc-r5nkv\" (UID: \"c1cd26c0-7d04-45fb-b171-d3398fcf8d24\") " pod="metallb-system/metallb-operator-webhook-server-7544c98cdc-r5nkv" Oct 14 07:15:02 crc kubenswrapper[4870]: I1014 07:15:02.636286 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-7544c98cdc-r5nkv" Oct 14 07:15:02 crc kubenswrapper[4870]: I1014 07:15:02.838010 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29340435-g6qxb" Oct 14 07:15:02 crc kubenswrapper[4870]: I1014 07:15:02.902476 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-f459659f5-vgrhm"] Oct 14 07:15:02 crc kubenswrapper[4870]: I1014 07:15:02.978255 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mrbvw\" (UniqueName: \"kubernetes.io/projected/89fe6660-6f16-4ddb-9b5c-1582231f5f30-kube-api-access-mrbvw\") pod \"89fe6660-6f16-4ddb-9b5c-1582231f5f30\" (UID: \"89fe6660-6f16-4ddb-9b5c-1582231f5f30\") " Oct 14 07:15:02 crc kubenswrapper[4870]: I1014 07:15:02.978370 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/89fe6660-6f16-4ddb-9b5c-1582231f5f30-config-volume\") pod \"89fe6660-6f16-4ddb-9b5c-1582231f5f30\" (UID: \"89fe6660-6f16-4ddb-9b5c-1582231f5f30\") " Oct 14 07:15:02 crc kubenswrapper[4870]: I1014 07:15:02.978398 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/89fe6660-6f16-4ddb-9b5c-1582231f5f30-secret-volume\") pod \"89fe6660-6f16-4ddb-9b5c-1582231f5f30\" (UID: \"89fe6660-6f16-4ddb-9b5c-1582231f5f30\") " Oct 14 07:15:02 crc kubenswrapper[4870]: I1014 07:15:02.979085 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/89fe6660-6f16-4ddb-9b5c-1582231f5f30-config-volume" (OuterVolumeSpecName: "config-volume") pod "89fe6660-6f16-4ddb-9b5c-1582231f5f30" (UID: "89fe6660-6f16-4ddb-9b5c-1582231f5f30"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:15:02 crc kubenswrapper[4870]: I1014 07:15:02.981997 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/89fe6660-6f16-4ddb-9b5c-1582231f5f30-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "89fe6660-6f16-4ddb-9b5c-1582231f5f30" (UID: "89fe6660-6f16-4ddb-9b5c-1582231f5f30"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:15:02 crc kubenswrapper[4870]: I1014 07:15:02.982670 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/89fe6660-6f16-4ddb-9b5c-1582231f5f30-kube-api-access-mrbvw" (OuterVolumeSpecName: "kube-api-access-mrbvw") pod "89fe6660-6f16-4ddb-9b5c-1582231f5f30" (UID: "89fe6660-6f16-4ddb-9b5c-1582231f5f30"). InnerVolumeSpecName "kube-api-access-mrbvw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:15:03 crc kubenswrapper[4870]: I1014 07:15:03.075249 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-7544c98cdc-r5nkv"] Oct 14 07:15:03 crc kubenswrapper[4870]: I1014 07:15:03.080035 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mrbvw\" (UniqueName: \"kubernetes.io/projected/89fe6660-6f16-4ddb-9b5c-1582231f5f30-kube-api-access-mrbvw\") on node \"crc\" DevicePath \"\"" Oct 14 07:15:03 crc kubenswrapper[4870]: I1014 07:15:03.080066 4870 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/89fe6660-6f16-4ddb-9b5c-1582231f5f30-config-volume\") on node \"crc\" DevicePath \"\"" Oct 14 07:15:03 crc kubenswrapper[4870]: I1014 07:15:03.080078 4870 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/89fe6660-6f16-4ddb-9b5c-1582231f5f30-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 14 07:15:03 crc kubenswrapper[4870]: W1014 07:15:03.083959 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc1cd26c0_7d04_45fb_b171_d3398fcf8d24.slice/crio-580620f3f4c4af57812c42036e517a21e4545018f2314cb21263cbd19cb2a042 WatchSource:0}: Error finding container 580620f3f4c4af57812c42036e517a21e4545018f2314cb21263cbd19cb2a042: Status 404 returned error can't find the container with id 580620f3f4c4af57812c42036e517a21e4545018f2314cb21263cbd19cb2a042 Oct 14 07:15:03 crc kubenswrapper[4870]: I1014 07:15:03.469052 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-7544c98cdc-r5nkv" event={"ID":"c1cd26c0-7d04-45fb-b171-d3398fcf8d24","Type":"ContainerStarted","Data":"580620f3f4c4af57812c42036e517a21e4545018f2314cb21263cbd19cb2a042"} Oct 14 07:15:03 crc kubenswrapper[4870]: I1014 07:15:03.471467 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29340435-g6qxb" event={"ID":"89fe6660-6f16-4ddb-9b5c-1582231f5f30","Type":"ContainerDied","Data":"bb4c5a4693eb6c5dbe677b276665adf1672adb43a2f006de26e7428de201e100"} Oct 14 07:15:03 crc kubenswrapper[4870]: I1014 07:15:03.471526 4870 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bb4c5a4693eb6c5dbe677b276665adf1672adb43a2f006de26e7428de201e100" Oct 14 07:15:03 crc kubenswrapper[4870]: I1014 07:15:03.471535 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29340435-g6qxb" Oct 14 07:15:03 crc kubenswrapper[4870]: I1014 07:15:03.473529 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-f459659f5-vgrhm" event={"ID":"1150555d-bbee-49bc-b90c-69e30c971a19","Type":"ContainerStarted","Data":"cd88eb55612bac69008bd20636d5f29e4ea1b24c6c16f577c841b4517c6fb0c9"} Oct 14 07:15:06 crc kubenswrapper[4870]: I1014 07:15:06.250193 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-vnhs7" Oct 14 07:15:06 crc kubenswrapper[4870]: I1014 07:15:06.250661 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-vnhs7" Oct 14 07:15:06 crc kubenswrapper[4870]: I1014 07:15:06.292504 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-vnhs7" Oct 14 07:15:06 crc kubenswrapper[4870]: I1014 07:15:06.575052 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-vnhs7" Oct 14 07:15:08 crc kubenswrapper[4870]: I1014 07:15:08.274326 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-vnhs7"] Oct 14 07:15:08 crc kubenswrapper[4870]: I1014 07:15:08.530715 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-f459659f5-vgrhm" event={"ID":"1150555d-bbee-49bc-b90c-69e30c971a19","Type":"ContainerStarted","Data":"dd8e642f91cf7701a99d60ad33c38d9ebcb6850cba6686961b7e68e4b4d7d677"} Oct 14 07:15:08 crc kubenswrapper[4870]: I1014 07:15:08.531579 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-f459659f5-vgrhm" Oct 14 07:15:08 crc kubenswrapper[4870]: I1014 07:15:08.533908 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-7544c98cdc-r5nkv" event={"ID":"c1cd26c0-7d04-45fb-b171-d3398fcf8d24","Type":"ContainerStarted","Data":"854dbe7047a8f86f594df6989cedfd912cc434bfde8fee985d1d5c985560df2a"} Oct 14 07:15:08 crc kubenswrapper[4870]: I1014 07:15:08.533991 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-vnhs7" podUID="00be94d3-e58b-4871-8f4f-97ff5fc158b7" containerName="registry-server" containerID="cri-o://9188753dd58084edda9d550b40c978e8337e7873471f0ecdb2ef435ea77c51ee" gracePeriod=2 Oct 14 07:15:08 crc kubenswrapper[4870]: I1014 07:15:08.555072 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-f459659f5-vgrhm" podStartSLOduration=1.312701046 podStartE2EDuration="6.555057037s" podCreationTimestamp="2025-10-14 07:15:02 +0000 UTC" firstStartedPulling="2025-10-14 07:15:02.917283749 +0000 UTC m=+838.614644120" lastFinishedPulling="2025-10-14 07:15:08.15963974 +0000 UTC m=+843.857000111" observedRunningTime="2025-10-14 07:15:08.552834342 +0000 UTC m=+844.250194713" watchObservedRunningTime="2025-10-14 07:15:08.555057037 +0000 UTC m=+844.252417408" Oct 14 07:15:08 crc kubenswrapper[4870]: I1014 07:15:08.582778 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-7544c98cdc-r5nkv" podStartSLOduration=1.484522058 podStartE2EDuration="6.582762241s" podCreationTimestamp="2025-10-14 07:15:02 +0000 UTC" firstStartedPulling="2025-10-14 07:15:03.090079105 +0000 UTC m=+838.787439476" lastFinishedPulling="2025-10-14 07:15:08.188319288 +0000 UTC m=+843.885679659" observedRunningTime="2025-10-14 07:15:08.580889024 +0000 UTC m=+844.278249395" watchObservedRunningTime="2025-10-14 07:15:08.582762241 +0000 UTC m=+844.280122612" Oct 14 07:15:08 crc kubenswrapper[4870]: I1014 07:15:08.983658 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vnhs7" Oct 14 07:15:09 crc kubenswrapper[4870]: I1014 07:15:09.069944 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mcssd\" (UniqueName: \"kubernetes.io/projected/00be94d3-e58b-4871-8f4f-97ff5fc158b7-kube-api-access-mcssd\") pod \"00be94d3-e58b-4871-8f4f-97ff5fc158b7\" (UID: \"00be94d3-e58b-4871-8f4f-97ff5fc158b7\") " Oct 14 07:15:09 crc kubenswrapper[4870]: I1014 07:15:09.070226 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/00be94d3-e58b-4871-8f4f-97ff5fc158b7-catalog-content\") pod \"00be94d3-e58b-4871-8f4f-97ff5fc158b7\" (UID: \"00be94d3-e58b-4871-8f4f-97ff5fc158b7\") " Oct 14 07:15:09 crc kubenswrapper[4870]: I1014 07:15:09.070274 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/00be94d3-e58b-4871-8f4f-97ff5fc158b7-utilities\") pod \"00be94d3-e58b-4871-8f4f-97ff5fc158b7\" (UID: \"00be94d3-e58b-4871-8f4f-97ff5fc158b7\") " Oct 14 07:15:09 crc kubenswrapper[4870]: I1014 07:15:09.071224 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/00be94d3-e58b-4871-8f4f-97ff5fc158b7-utilities" (OuterVolumeSpecName: "utilities") pod "00be94d3-e58b-4871-8f4f-97ff5fc158b7" (UID: "00be94d3-e58b-4871-8f4f-97ff5fc158b7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:15:09 crc kubenswrapper[4870]: I1014 07:15:09.071842 4870 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/00be94d3-e58b-4871-8f4f-97ff5fc158b7-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 07:15:09 crc kubenswrapper[4870]: I1014 07:15:09.084493 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/00be94d3-e58b-4871-8f4f-97ff5fc158b7-kube-api-access-mcssd" (OuterVolumeSpecName: "kube-api-access-mcssd") pod "00be94d3-e58b-4871-8f4f-97ff5fc158b7" (UID: "00be94d3-e58b-4871-8f4f-97ff5fc158b7"). InnerVolumeSpecName "kube-api-access-mcssd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:15:09 crc kubenswrapper[4870]: I1014 07:15:09.089564 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/00be94d3-e58b-4871-8f4f-97ff5fc158b7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "00be94d3-e58b-4871-8f4f-97ff5fc158b7" (UID: "00be94d3-e58b-4871-8f4f-97ff5fc158b7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:15:09 crc kubenswrapper[4870]: I1014 07:15:09.173627 4870 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/00be94d3-e58b-4871-8f4f-97ff5fc158b7-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 07:15:09 crc kubenswrapper[4870]: I1014 07:15:09.173683 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mcssd\" (UniqueName: \"kubernetes.io/projected/00be94d3-e58b-4871-8f4f-97ff5fc158b7-kube-api-access-mcssd\") on node \"crc\" DevicePath \"\"" Oct 14 07:15:09 crc kubenswrapper[4870]: I1014 07:15:09.540512 4870 generic.go:334] "Generic (PLEG): container finished" podID="00be94d3-e58b-4871-8f4f-97ff5fc158b7" containerID="9188753dd58084edda9d550b40c978e8337e7873471f0ecdb2ef435ea77c51ee" exitCode=0 Oct 14 07:15:09 crc kubenswrapper[4870]: I1014 07:15:09.540551 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vnhs7" event={"ID":"00be94d3-e58b-4871-8f4f-97ff5fc158b7","Type":"ContainerDied","Data":"9188753dd58084edda9d550b40c978e8337e7873471f0ecdb2ef435ea77c51ee"} Oct 14 07:15:09 crc kubenswrapper[4870]: I1014 07:15:09.540607 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vnhs7" event={"ID":"00be94d3-e58b-4871-8f4f-97ff5fc158b7","Type":"ContainerDied","Data":"0b222e5955c9db35f8b860a4e48a901b8c7d85ee85a5ce85af91bd26be536a6f"} Oct 14 07:15:09 crc kubenswrapper[4870]: I1014 07:15:09.540636 4870 scope.go:117] "RemoveContainer" containerID="9188753dd58084edda9d550b40c978e8337e7873471f0ecdb2ef435ea77c51ee" Oct 14 07:15:09 crc kubenswrapper[4870]: I1014 07:15:09.540690 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vnhs7" Oct 14 07:15:09 crc kubenswrapper[4870]: I1014 07:15:09.540709 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-7544c98cdc-r5nkv" Oct 14 07:15:09 crc kubenswrapper[4870]: I1014 07:15:09.554401 4870 scope.go:117] "RemoveContainer" containerID="7b7e30dee0ab31ad34f62d92a825cd71cb9afc7e6038bd4bd4d92f23eea18f19" Oct 14 07:15:09 crc kubenswrapper[4870]: I1014 07:15:09.569691 4870 scope.go:117] "RemoveContainer" containerID="729460503d7cb277fd8658c04d4f9cd53a2598d73dcb80e73f41d4409133cf2d" Oct 14 07:15:09 crc kubenswrapper[4870]: I1014 07:15:09.592392 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-vnhs7"] Oct 14 07:15:09 crc kubenswrapper[4870]: I1014 07:15:09.596203 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-vnhs7"] Oct 14 07:15:09 crc kubenswrapper[4870]: I1014 07:15:09.603962 4870 scope.go:117] "RemoveContainer" containerID="9188753dd58084edda9d550b40c978e8337e7873471f0ecdb2ef435ea77c51ee" Oct 14 07:15:09 crc kubenswrapper[4870]: E1014 07:15:09.604373 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9188753dd58084edda9d550b40c978e8337e7873471f0ecdb2ef435ea77c51ee\": container with ID starting with 9188753dd58084edda9d550b40c978e8337e7873471f0ecdb2ef435ea77c51ee not found: ID does not exist" containerID="9188753dd58084edda9d550b40c978e8337e7873471f0ecdb2ef435ea77c51ee" Oct 14 07:15:09 crc kubenswrapper[4870]: I1014 07:15:09.604406 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9188753dd58084edda9d550b40c978e8337e7873471f0ecdb2ef435ea77c51ee"} err="failed to get container status \"9188753dd58084edda9d550b40c978e8337e7873471f0ecdb2ef435ea77c51ee\": rpc error: code = NotFound desc = could not find container \"9188753dd58084edda9d550b40c978e8337e7873471f0ecdb2ef435ea77c51ee\": container with ID starting with 9188753dd58084edda9d550b40c978e8337e7873471f0ecdb2ef435ea77c51ee not found: ID does not exist" Oct 14 07:15:09 crc kubenswrapper[4870]: I1014 07:15:09.604430 4870 scope.go:117] "RemoveContainer" containerID="7b7e30dee0ab31ad34f62d92a825cd71cb9afc7e6038bd4bd4d92f23eea18f19" Oct 14 07:15:09 crc kubenswrapper[4870]: E1014 07:15:09.604818 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7b7e30dee0ab31ad34f62d92a825cd71cb9afc7e6038bd4bd4d92f23eea18f19\": container with ID starting with 7b7e30dee0ab31ad34f62d92a825cd71cb9afc7e6038bd4bd4d92f23eea18f19 not found: ID does not exist" containerID="7b7e30dee0ab31ad34f62d92a825cd71cb9afc7e6038bd4bd4d92f23eea18f19" Oct 14 07:15:09 crc kubenswrapper[4870]: I1014 07:15:09.604849 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7b7e30dee0ab31ad34f62d92a825cd71cb9afc7e6038bd4bd4d92f23eea18f19"} err="failed to get container status \"7b7e30dee0ab31ad34f62d92a825cd71cb9afc7e6038bd4bd4d92f23eea18f19\": rpc error: code = NotFound desc = could not find container \"7b7e30dee0ab31ad34f62d92a825cd71cb9afc7e6038bd4bd4d92f23eea18f19\": container with ID starting with 7b7e30dee0ab31ad34f62d92a825cd71cb9afc7e6038bd4bd4d92f23eea18f19 not found: ID does not exist" Oct 14 07:15:09 crc kubenswrapper[4870]: I1014 07:15:09.604861 4870 scope.go:117] "RemoveContainer" containerID="729460503d7cb277fd8658c04d4f9cd53a2598d73dcb80e73f41d4409133cf2d" Oct 14 07:15:09 crc kubenswrapper[4870]: E1014 07:15:09.605280 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"729460503d7cb277fd8658c04d4f9cd53a2598d73dcb80e73f41d4409133cf2d\": container with ID starting with 729460503d7cb277fd8658c04d4f9cd53a2598d73dcb80e73f41d4409133cf2d not found: ID does not exist" containerID="729460503d7cb277fd8658c04d4f9cd53a2598d73dcb80e73f41d4409133cf2d" Oct 14 07:15:09 crc kubenswrapper[4870]: I1014 07:15:09.605307 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"729460503d7cb277fd8658c04d4f9cd53a2598d73dcb80e73f41d4409133cf2d"} err="failed to get container status \"729460503d7cb277fd8658c04d4f9cd53a2598d73dcb80e73f41d4409133cf2d\": rpc error: code = NotFound desc = could not find container \"729460503d7cb277fd8658c04d4f9cd53a2598d73dcb80e73f41d4409133cf2d\": container with ID starting with 729460503d7cb277fd8658c04d4f9cd53a2598d73dcb80e73f41d4409133cf2d not found: ID does not exist" Oct 14 07:15:11 crc kubenswrapper[4870]: I1014 07:15:11.045670 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="00be94d3-e58b-4871-8f4f-97ff5fc158b7" path="/var/lib/kubelet/pods/00be94d3-e58b-4871-8f4f-97ff5fc158b7/volumes" Oct 14 07:15:22 crc kubenswrapper[4870]: I1014 07:15:22.643615 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-7544c98cdc-r5nkv" Oct 14 07:15:23 crc kubenswrapper[4870]: I1014 07:15:23.951267 4870 patch_prober.go:28] interesting pod/machine-config-daemon-7tvc7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 07:15:23 crc kubenswrapper[4870]: I1014 07:15:23.951328 4870 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 07:15:23 crc kubenswrapper[4870]: I1014 07:15:23.951378 4870 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" Oct 14 07:15:23 crc kubenswrapper[4870]: I1014 07:15:23.951882 4870 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"b5bea711ed09c7366862c920d932bb6dfefddcdd90ae2ab8981cfd9449e357a2"} pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 14 07:15:23 crc kubenswrapper[4870]: I1014 07:15:23.951940 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerName="machine-config-daemon" containerID="cri-o://b5bea711ed09c7366862c920d932bb6dfefddcdd90ae2ab8981cfd9449e357a2" gracePeriod=600 Oct 14 07:15:24 crc kubenswrapper[4870]: I1014 07:15:24.655356 4870 generic.go:334] "Generic (PLEG): container finished" podID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerID="b5bea711ed09c7366862c920d932bb6dfefddcdd90ae2ab8981cfd9449e357a2" exitCode=0 Oct 14 07:15:24 crc kubenswrapper[4870]: I1014 07:15:24.655410 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" event={"ID":"874a5e45-dffd-4d17-b609-b7d3ed2eab07","Type":"ContainerDied","Data":"b5bea711ed09c7366862c920d932bb6dfefddcdd90ae2ab8981cfd9449e357a2"} Oct 14 07:15:24 crc kubenswrapper[4870]: I1014 07:15:24.656207 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" event={"ID":"874a5e45-dffd-4d17-b609-b7d3ed2eab07","Type":"ContainerStarted","Data":"afa79ed981f5045c93ba7103bf82588637ff210f79db4fa50132dc5fa3e91338"} Oct 14 07:15:24 crc kubenswrapper[4870]: I1014 07:15:24.656242 4870 scope.go:117] "RemoveContainer" containerID="f5c863e175088c68c7d18f66c4b86bb42e49b43e1b581fdc20f0b5d02ea623e6" Oct 14 07:15:42 crc kubenswrapper[4870]: I1014 07:15:42.406896 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-f459659f5-vgrhm" Oct 14 07:15:43 crc kubenswrapper[4870]: I1014 07:15:43.120616 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-64bf5d555-f8wq5"] Oct 14 07:15:43 crc kubenswrapper[4870]: E1014 07:15:43.120959 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00be94d3-e58b-4871-8f4f-97ff5fc158b7" containerName="extract-utilities" Oct 14 07:15:43 crc kubenswrapper[4870]: I1014 07:15:43.120982 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="00be94d3-e58b-4871-8f4f-97ff5fc158b7" containerName="extract-utilities" Oct 14 07:15:43 crc kubenswrapper[4870]: E1014 07:15:43.121007 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00be94d3-e58b-4871-8f4f-97ff5fc158b7" containerName="registry-server" Oct 14 07:15:43 crc kubenswrapper[4870]: I1014 07:15:43.121019 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="00be94d3-e58b-4871-8f4f-97ff5fc158b7" containerName="registry-server" Oct 14 07:15:43 crc kubenswrapper[4870]: E1014 07:15:43.121035 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="89fe6660-6f16-4ddb-9b5c-1582231f5f30" containerName="collect-profiles" Oct 14 07:15:43 crc kubenswrapper[4870]: I1014 07:15:43.121046 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="89fe6660-6f16-4ddb-9b5c-1582231f5f30" containerName="collect-profiles" Oct 14 07:15:43 crc kubenswrapper[4870]: E1014 07:15:43.121070 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00be94d3-e58b-4871-8f4f-97ff5fc158b7" containerName="extract-content" Oct 14 07:15:43 crc kubenswrapper[4870]: I1014 07:15:43.121082 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="00be94d3-e58b-4871-8f4f-97ff5fc158b7" containerName="extract-content" Oct 14 07:15:43 crc kubenswrapper[4870]: I1014 07:15:43.121266 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="00be94d3-e58b-4871-8f4f-97ff5fc158b7" containerName="registry-server" Oct 14 07:15:43 crc kubenswrapper[4870]: I1014 07:15:43.121292 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="89fe6660-6f16-4ddb-9b5c-1582231f5f30" containerName="collect-profiles" Oct 14 07:15:43 crc kubenswrapper[4870]: I1014 07:15:43.121857 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-f8wq5" Oct 14 07:15:43 crc kubenswrapper[4870]: I1014 07:15:43.123679 4870 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Oct 14 07:15:43 crc kubenswrapper[4870]: I1014 07:15:43.124480 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-pbtkp"] Oct 14 07:15:43 crc kubenswrapper[4870]: I1014 07:15:43.124540 4870 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-4fbsh" Oct 14 07:15:43 crc kubenswrapper[4870]: I1014 07:15:43.126526 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-pbtkp" Oct 14 07:15:43 crc kubenswrapper[4870]: I1014 07:15:43.128984 4870 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Oct 14 07:15:43 crc kubenswrapper[4870]: I1014 07:15:43.129031 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Oct 14 07:15:43 crc kubenswrapper[4870]: I1014 07:15:43.138298 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-64bf5d555-f8wq5"] Oct 14 07:15:43 crc kubenswrapper[4870]: I1014 07:15:43.194973 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/2f058085-57d8-48a8-9267-fde99e4c15fb-frr-sockets\") pod \"frr-k8s-pbtkp\" (UID: \"2f058085-57d8-48a8-9267-fde99e4c15fb\") " pod="metallb-system/frr-k8s-pbtkp" Oct 14 07:15:43 crc kubenswrapper[4870]: I1014 07:15:43.195034 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c59sz\" (UniqueName: \"kubernetes.io/projected/2f058085-57d8-48a8-9267-fde99e4c15fb-kube-api-access-c59sz\") pod \"frr-k8s-pbtkp\" (UID: \"2f058085-57d8-48a8-9267-fde99e4c15fb\") " pod="metallb-system/frr-k8s-pbtkp" Oct 14 07:15:43 crc kubenswrapper[4870]: I1014 07:15:43.195178 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/2f058085-57d8-48a8-9267-fde99e4c15fb-frr-startup\") pod \"frr-k8s-pbtkp\" (UID: \"2f058085-57d8-48a8-9267-fde99e4c15fb\") " pod="metallb-system/frr-k8s-pbtkp" Oct 14 07:15:43 crc kubenswrapper[4870]: I1014 07:15:43.195213 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/2f058085-57d8-48a8-9267-fde99e4c15fb-frr-conf\") pod \"frr-k8s-pbtkp\" (UID: \"2f058085-57d8-48a8-9267-fde99e4c15fb\") " pod="metallb-system/frr-k8s-pbtkp" Oct 14 07:15:43 crc kubenswrapper[4870]: I1014 07:15:43.195301 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2f058085-57d8-48a8-9267-fde99e4c15fb-metrics-certs\") pod \"frr-k8s-pbtkp\" (UID: \"2f058085-57d8-48a8-9267-fde99e4c15fb\") " pod="metallb-system/frr-k8s-pbtkp" Oct 14 07:15:43 crc kubenswrapper[4870]: I1014 07:15:43.195382 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zcs96\" (UniqueName: \"kubernetes.io/projected/d0a20afe-74ac-4d3f-80ca-981fb5ce00a8-kube-api-access-zcs96\") pod \"frr-k8s-webhook-server-64bf5d555-f8wq5\" (UID: \"d0a20afe-74ac-4d3f-80ca-981fb5ce00a8\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-f8wq5" Oct 14 07:15:43 crc kubenswrapper[4870]: I1014 07:15:43.195415 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d0a20afe-74ac-4d3f-80ca-981fb5ce00a8-cert\") pod \"frr-k8s-webhook-server-64bf5d555-f8wq5\" (UID: \"d0a20afe-74ac-4d3f-80ca-981fb5ce00a8\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-f8wq5" Oct 14 07:15:43 crc kubenswrapper[4870]: I1014 07:15:43.195453 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/2f058085-57d8-48a8-9267-fde99e4c15fb-metrics\") pod \"frr-k8s-pbtkp\" (UID: \"2f058085-57d8-48a8-9267-fde99e4c15fb\") " pod="metallb-system/frr-k8s-pbtkp" Oct 14 07:15:43 crc kubenswrapper[4870]: I1014 07:15:43.195507 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/2f058085-57d8-48a8-9267-fde99e4c15fb-reloader\") pod \"frr-k8s-pbtkp\" (UID: \"2f058085-57d8-48a8-9267-fde99e4c15fb\") " pod="metallb-system/frr-k8s-pbtkp" Oct 14 07:15:43 crc kubenswrapper[4870]: I1014 07:15:43.200173 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-mtvtd"] Oct 14 07:15:43 crc kubenswrapper[4870]: I1014 07:15:43.201027 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-mtvtd" Oct 14 07:15:43 crc kubenswrapper[4870]: I1014 07:15:43.203032 4870 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-6ldqr" Oct 14 07:15:43 crc kubenswrapper[4870]: I1014 07:15:43.203080 4870 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Oct 14 07:15:43 crc kubenswrapper[4870]: I1014 07:15:43.203368 4870 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Oct 14 07:15:43 crc kubenswrapper[4870]: I1014 07:15:43.203411 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Oct 14 07:15:43 crc kubenswrapper[4870]: I1014 07:15:43.210973 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-68d546b9d8-bs82b"] Oct 14 07:15:43 crc kubenswrapper[4870]: I1014 07:15:43.212190 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-68d546b9d8-bs82b" Oct 14 07:15:43 crc kubenswrapper[4870]: I1014 07:15:43.213495 4870 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Oct 14 07:15:43 crc kubenswrapper[4870]: I1014 07:15:43.223269 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-68d546b9d8-bs82b"] Oct 14 07:15:43 crc kubenswrapper[4870]: I1014 07:15:43.296801 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c59sz\" (UniqueName: \"kubernetes.io/projected/2f058085-57d8-48a8-9267-fde99e4c15fb-kube-api-access-c59sz\") pod \"frr-k8s-pbtkp\" (UID: \"2f058085-57d8-48a8-9267-fde99e4c15fb\") " pod="metallb-system/frr-k8s-pbtkp" Oct 14 07:15:43 crc kubenswrapper[4870]: I1014 07:15:43.296855 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gb8bb\" (UniqueName: \"kubernetes.io/projected/c7f0d440-b7cc-440a-9aac-e19c2bafab73-kube-api-access-gb8bb\") pod \"controller-68d546b9d8-bs82b\" (UID: \"c7f0d440-b7cc-440a-9aac-e19c2bafab73\") " pod="metallb-system/controller-68d546b9d8-bs82b" Oct 14 07:15:43 crc kubenswrapper[4870]: I1014 07:15:43.296893 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/2f058085-57d8-48a8-9267-fde99e4c15fb-frr-startup\") pod \"frr-k8s-pbtkp\" (UID: \"2f058085-57d8-48a8-9267-fde99e4c15fb\") " pod="metallb-system/frr-k8s-pbtkp" Oct 14 07:15:43 crc kubenswrapper[4870]: I1014 07:15:43.296913 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/2f058085-57d8-48a8-9267-fde99e4c15fb-frr-conf\") pod \"frr-k8s-pbtkp\" (UID: \"2f058085-57d8-48a8-9267-fde99e4c15fb\") " pod="metallb-system/frr-k8s-pbtkp" Oct 14 07:15:43 crc kubenswrapper[4870]: I1014 07:15:43.296940 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/010032d8-4503-4fb5-8158-066872d97d52-metallb-excludel2\") pod \"speaker-mtvtd\" (UID: \"010032d8-4503-4fb5-8158-066872d97d52\") " pod="metallb-system/speaker-mtvtd" Oct 14 07:15:43 crc kubenswrapper[4870]: I1014 07:15:43.296957 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c7f0d440-b7cc-440a-9aac-e19c2bafab73-cert\") pod \"controller-68d546b9d8-bs82b\" (UID: \"c7f0d440-b7cc-440a-9aac-e19c2bafab73\") " pod="metallb-system/controller-68d546b9d8-bs82b" Oct 14 07:15:43 crc kubenswrapper[4870]: I1014 07:15:43.296983 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2f058085-57d8-48a8-9267-fde99e4c15fb-metrics-certs\") pod \"frr-k8s-pbtkp\" (UID: \"2f058085-57d8-48a8-9267-fde99e4c15fb\") " pod="metallb-system/frr-k8s-pbtkp" Oct 14 07:15:43 crc kubenswrapper[4870]: I1014 07:15:43.297001 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/010032d8-4503-4fb5-8158-066872d97d52-metrics-certs\") pod \"speaker-mtvtd\" (UID: \"010032d8-4503-4fb5-8158-066872d97d52\") " pod="metallb-system/speaker-mtvtd" Oct 14 07:15:43 crc kubenswrapper[4870]: I1014 07:15:43.297029 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zcs96\" (UniqueName: \"kubernetes.io/projected/d0a20afe-74ac-4d3f-80ca-981fb5ce00a8-kube-api-access-zcs96\") pod \"frr-k8s-webhook-server-64bf5d555-f8wq5\" (UID: \"d0a20afe-74ac-4d3f-80ca-981fb5ce00a8\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-f8wq5" Oct 14 07:15:43 crc kubenswrapper[4870]: I1014 07:15:43.297048 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d0a20afe-74ac-4d3f-80ca-981fb5ce00a8-cert\") pod \"frr-k8s-webhook-server-64bf5d555-f8wq5\" (UID: \"d0a20afe-74ac-4d3f-80ca-981fb5ce00a8\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-f8wq5" Oct 14 07:15:43 crc kubenswrapper[4870]: I1014 07:15:43.297064 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/010032d8-4503-4fb5-8158-066872d97d52-memberlist\") pod \"speaker-mtvtd\" (UID: \"010032d8-4503-4fb5-8158-066872d97d52\") " pod="metallb-system/speaker-mtvtd" Oct 14 07:15:43 crc kubenswrapper[4870]: I1014 07:15:43.297083 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/2f058085-57d8-48a8-9267-fde99e4c15fb-metrics\") pod \"frr-k8s-pbtkp\" (UID: \"2f058085-57d8-48a8-9267-fde99e4c15fb\") " pod="metallb-system/frr-k8s-pbtkp" Oct 14 07:15:43 crc kubenswrapper[4870]: I1014 07:15:43.297097 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c7f0d440-b7cc-440a-9aac-e19c2bafab73-metrics-certs\") pod \"controller-68d546b9d8-bs82b\" (UID: \"c7f0d440-b7cc-440a-9aac-e19c2bafab73\") " pod="metallb-system/controller-68d546b9d8-bs82b" Oct 14 07:15:43 crc kubenswrapper[4870]: I1014 07:15:43.297112 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/2f058085-57d8-48a8-9267-fde99e4c15fb-reloader\") pod \"frr-k8s-pbtkp\" (UID: \"2f058085-57d8-48a8-9267-fde99e4c15fb\") " pod="metallb-system/frr-k8s-pbtkp" Oct 14 07:15:43 crc kubenswrapper[4870]: I1014 07:15:43.297150 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9rxd4\" (UniqueName: \"kubernetes.io/projected/010032d8-4503-4fb5-8158-066872d97d52-kube-api-access-9rxd4\") pod \"speaker-mtvtd\" (UID: \"010032d8-4503-4fb5-8158-066872d97d52\") " pod="metallb-system/speaker-mtvtd" Oct 14 07:15:43 crc kubenswrapper[4870]: I1014 07:15:43.297172 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/2f058085-57d8-48a8-9267-fde99e4c15fb-frr-sockets\") pod \"frr-k8s-pbtkp\" (UID: \"2f058085-57d8-48a8-9267-fde99e4c15fb\") " pod="metallb-system/frr-k8s-pbtkp" Oct 14 07:15:43 crc kubenswrapper[4870]: I1014 07:15:43.297594 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/2f058085-57d8-48a8-9267-fde99e4c15fb-frr-sockets\") pod \"frr-k8s-pbtkp\" (UID: \"2f058085-57d8-48a8-9267-fde99e4c15fb\") " pod="metallb-system/frr-k8s-pbtkp" Oct 14 07:15:43 crc kubenswrapper[4870]: E1014 07:15:43.297721 4870 secret.go:188] Couldn't get secret metallb-system/frr-k8s-certs-secret: secret "frr-k8s-certs-secret" not found Oct 14 07:15:43 crc kubenswrapper[4870]: E1014 07:15:43.297824 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2f058085-57d8-48a8-9267-fde99e4c15fb-metrics-certs podName:2f058085-57d8-48a8-9267-fde99e4c15fb nodeName:}" failed. No retries permitted until 2025-10-14 07:15:43.797793209 +0000 UTC m=+879.495153710 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/2f058085-57d8-48a8-9267-fde99e4c15fb-metrics-certs") pod "frr-k8s-pbtkp" (UID: "2f058085-57d8-48a8-9267-fde99e4c15fb") : secret "frr-k8s-certs-secret" not found Oct 14 07:15:43 crc kubenswrapper[4870]: I1014 07:15:43.297863 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/2f058085-57d8-48a8-9267-fde99e4c15fb-frr-conf\") pod \"frr-k8s-pbtkp\" (UID: \"2f058085-57d8-48a8-9267-fde99e4c15fb\") " pod="metallb-system/frr-k8s-pbtkp" Oct 14 07:15:43 crc kubenswrapper[4870]: E1014 07:15:43.297895 4870 secret.go:188] Couldn't get secret metallb-system/frr-k8s-webhook-server-cert: secret "frr-k8s-webhook-server-cert" not found Oct 14 07:15:43 crc kubenswrapper[4870]: E1014 07:15:43.297934 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d0a20afe-74ac-4d3f-80ca-981fb5ce00a8-cert podName:d0a20afe-74ac-4d3f-80ca-981fb5ce00a8 nodeName:}" failed. No retries permitted until 2025-10-14 07:15:43.797921572 +0000 UTC m=+879.495281943 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/d0a20afe-74ac-4d3f-80ca-981fb5ce00a8-cert") pod "frr-k8s-webhook-server-64bf5d555-f8wq5" (UID: "d0a20afe-74ac-4d3f-80ca-981fb5ce00a8") : secret "frr-k8s-webhook-server-cert" not found Oct 14 07:15:43 crc kubenswrapper[4870]: I1014 07:15:43.298171 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/2f058085-57d8-48a8-9267-fde99e4c15fb-metrics\") pod \"frr-k8s-pbtkp\" (UID: \"2f058085-57d8-48a8-9267-fde99e4c15fb\") " pod="metallb-system/frr-k8s-pbtkp" Oct 14 07:15:43 crc kubenswrapper[4870]: I1014 07:15:43.298742 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/2f058085-57d8-48a8-9267-fde99e4c15fb-frr-startup\") pod \"frr-k8s-pbtkp\" (UID: \"2f058085-57d8-48a8-9267-fde99e4c15fb\") " pod="metallb-system/frr-k8s-pbtkp" Oct 14 07:15:43 crc kubenswrapper[4870]: I1014 07:15:43.299587 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/2f058085-57d8-48a8-9267-fde99e4c15fb-reloader\") pod \"frr-k8s-pbtkp\" (UID: \"2f058085-57d8-48a8-9267-fde99e4c15fb\") " pod="metallb-system/frr-k8s-pbtkp" Oct 14 07:15:43 crc kubenswrapper[4870]: I1014 07:15:43.317745 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zcs96\" (UniqueName: \"kubernetes.io/projected/d0a20afe-74ac-4d3f-80ca-981fb5ce00a8-kube-api-access-zcs96\") pod \"frr-k8s-webhook-server-64bf5d555-f8wq5\" (UID: \"d0a20afe-74ac-4d3f-80ca-981fb5ce00a8\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-f8wq5" Oct 14 07:15:43 crc kubenswrapper[4870]: I1014 07:15:43.317814 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c59sz\" (UniqueName: \"kubernetes.io/projected/2f058085-57d8-48a8-9267-fde99e4c15fb-kube-api-access-c59sz\") pod \"frr-k8s-pbtkp\" (UID: \"2f058085-57d8-48a8-9267-fde99e4c15fb\") " pod="metallb-system/frr-k8s-pbtkp" Oct 14 07:15:43 crc kubenswrapper[4870]: I1014 07:15:43.398542 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/010032d8-4503-4fb5-8158-066872d97d52-metrics-certs\") pod \"speaker-mtvtd\" (UID: \"010032d8-4503-4fb5-8158-066872d97d52\") " pod="metallb-system/speaker-mtvtd" Oct 14 07:15:43 crc kubenswrapper[4870]: I1014 07:15:43.398646 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/010032d8-4503-4fb5-8158-066872d97d52-memberlist\") pod \"speaker-mtvtd\" (UID: \"010032d8-4503-4fb5-8158-066872d97d52\") " pod="metallb-system/speaker-mtvtd" Oct 14 07:15:43 crc kubenswrapper[4870]: I1014 07:15:43.398701 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c7f0d440-b7cc-440a-9aac-e19c2bafab73-metrics-certs\") pod \"controller-68d546b9d8-bs82b\" (UID: \"c7f0d440-b7cc-440a-9aac-e19c2bafab73\") " pod="metallb-system/controller-68d546b9d8-bs82b" Oct 14 07:15:43 crc kubenswrapper[4870]: I1014 07:15:43.398758 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9rxd4\" (UniqueName: \"kubernetes.io/projected/010032d8-4503-4fb5-8158-066872d97d52-kube-api-access-9rxd4\") pod \"speaker-mtvtd\" (UID: \"010032d8-4503-4fb5-8158-066872d97d52\") " pod="metallb-system/speaker-mtvtd" Oct 14 07:15:43 crc kubenswrapper[4870]: I1014 07:15:43.398805 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gb8bb\" (UniqueName: \"kubernetes.io/projected/c7f0d440-b7cc-440a-9aac-e19c2bafab73-kube-api-access-gb8bb\") pod \"controller-68d546b9d8-bs82b\" (UID: \"c7f0d440-b7cc-440a-9aac-e19c2bafab73\") " pod="metallb-system/controller-68d546b9d8-bs82b" Oct 14 07:15:43 crc kubenswrapper[4870]: I1014 07:15:43.398876 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/010032d8-4503-4fb5-8158-066872d97d52-metallb-excludel2\") pod \"speaker-mtvtd\" (UID: \"010032d8-4503-4fb5-8158-066872d97d52\") " pod="metallb-system/speaker-mtvtd" Oct 14 07:15:43 crc kubenswrapper[4870]: I1014 07:15:43.398907 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c7f0d440-b7cc-440a-9aac-e19c2bafab73-cert\") pod \"controller-68d546b9d8-bs82b\" (UID: \"c7f0d440-b7cc-440a-9aac-e19c2bafab73\") " pod="metallb-system/controller-68d546b9d8-bs82b" Oct 14 07:15:43 crc kubenswrapper[4870]: E1014 07:15:43.399130 4870 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Oct 14 07:15:43 crc kubenswrapper[4870]: E1014 07:15:43.399208 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/010032d8-4503-4fb5-8158-066872d97d52-memberlist podName:010032d8-4503-4fb5-8158-066872d97d52 nodeName:}" failed. No retries permitted until 2025-10-14 07:15:43.899185887 +0000 UTC m=+879.596546268 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/010032d8-4503-4fb5-8158-066872d97d52-memberlist") pod "speaker-mtvtd" (UID: "010032d8-4503-4fb5-8158-066872d97d52") : secret "metallb-memberlist" not found Oct 14 07:15:43 crc kubenswrapper[4870]: I1014 07:15:43.400384 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/010032d8-4503-4fb5-8158-066872d97d52-metallb-excludel2\") pod \"speaker-mtvtd\" (UID: \"010032d8-4503-4fb5-8158-066872d97d52\") " pod="metallb-system/speaker-mtvtd" Oct 14 07:15:43 crc kubenswrapper[4870]: I1014 07:15:43.401775 4870 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Oct 14 07:15:43 crc kubenswrapper[4870]: I1014 07:15:43.401963 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/010032d8-4503-4fb5-8158-066872d97d52-metrics-certs\") pod \"speaker-mtvtd\" (UID: \"010032d8-4503-4fb5-8158-066872d97d52\") " pod="metallb-system/speaker-mtvtd" Oct 14 07:15:43 crc kubenswrapper[4870]: I1014 07:15:43.403917 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c7f0d440-b7cc-440a-9aac-e19c2bafab73-metrics-certs\") pod \"controller-68d546b9d8-bs82b\" (UID: \"c7f0d440-b7cc-440a-9aac-e19c2bafab73\") " pod="metallb-system/controller-68d546b9d8-bs82b" Oct 14 07:15:43 crc kubenswrapper[4870]: I1014 07:15:43.413069 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c7f0d440-b7cc-440a-9aac-e19c2bafab73-cert\") pod \"controller-68d546b9d8-bs82b\" (UID: \"c7f0d440-b7cc-440a-9aac-e19c2bafab73\") " pod="metallb-system/controller-68d546b9d8-bs82b" Oct 14 07:15:43 crc kubenswrapper[4870]: I1014 07:15:43.422262 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9rxd4\" (UniqueName: \"kubernetes.io/projected/010032d8-4503-4fb5-8158-066872d97d52-kube-api-access-9rxd4\") pod \"speaker-mtvtd\" (UID: \"010032d8-4503-4fb5-8158-066872d97d52\") " pod="metallb-system/speaker-mtvtd" Oct 14 07:15:43 crc kubenswrapper[4870]: I1014 07:15:43.423652 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gb8bb\" (UniqueName: \"kubernetes.io/projected/c7f0d440-b7cc-440a-9aac-e19c2bafab73-kube-api-access-gb8bb\") pod \"controller-68d546b9d8-bs82b\" (UID: \"c7f0d440-b7cc-440a-9aac-e19c2bafab73\") " pod="metallb-system/controller-68d546b9d8-bs82b" Oct 14 07:15:43 crc kubenswrapper[4870]: I1014 07:15:43.528828 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-68d546b9d8-bs82b" Oct 14 07:15:43 crc kubenswrapper[4870]: I1014 07:15:43.803430 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2f058085-57d8-48a8-9267-fde99e4c15fb-metrics-certs\") pod \"frr-k8s-pbtkp\" (UID: \"2f058085-57d8-48a8-9267-fde99e4c15fb\") " pod="metallb-system/frr-k8s-pbtkp" Oct 14 07:15:43 crc kubenswrapper[4870]: I1014 07:15:43.803822 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d0a20afe-74ac-4d3f-80ca-981fb5ce00a8-cert\") pod \"frr-k8s-webhook-server-64bf5d555-f8wq5\" (UID: \"d0a20afe-74ac-4d3f-80ca-981fb5ce00a8\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-f8wq5" Oct 14 07:15:43 crc kubenswrapper[4870]: I1014 07:15:43.808087 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2f058085-57d8-48a8-9267-fde99e4c15fb-metrics-certs\") pod \"frr-k8s-pbtkp\" (UID: \"2f058085-57d8-48a8-9267-fde99e4c15fb\") " pod="metallb-system/frr-k8s-pbtkp" Oct 14 07:15:43 crc kubenswrapper[4870]: I1014 07:15:43.808105 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d0a20afe-74ac-4d3f-80ca-981fb5ce00a8-cert\") pod \"frr-k8s-webhook-server-64bf5d555-f8wq5\" (UID: \"d0a20afe-74ac-4d3f-80ca-981fb5ce00a8\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-f8wq5" Oct 14 07:15:43 crc kubenswrapper[4870]: I1014 07:15:43.904835 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/010032d8-4503-4fb5-8158-066872d97d52-memberlist\") pod \"speaker-mtvtd\" (UID: \"010032d8-4503-4fb5-8158-066872d97d52\") " pod="metallb-system/speaker-mtvtd" Oct 14 07:15:43 crc kubenswrapper[4870]: E1014 07:15:43.905106 4870 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Oct 14 07:15:43 crc kubenswrapper[4870]: E1014 07:15:43.905239 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/010032d8-4503-4fb5-8158-066872d97d52-memberlist podName:010032d8-4503-4fb5-8158-066872d97d52 nodeName:}" failed. No retries permitted until 2025-10-14 07:15:44.905211873 +0000 UTC m=+880.602572284 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/010032d8-4503-4fb5-8158-066872d97d52-memberlist") pod "speaker-mtvtd" (UID: "010032d8-4503-4fb5-8158-066872d97d52") : secret "metallb-memberlist" not found Oct 14 07:15:44 crc kubenswrapper[4870]: I1014 07:15:44.029835 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-68d546b9d8-bs82b"] Oct 14 07:15:44 crc kubenswrapper[4870]: I1014 07:15:44.045667 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-f8wq5" Oct 14 07:15:44 crc kubenswrapper[4870]: I1014 07:15:44.053884 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-pbtkp" Oct 14 07:15:44 crc kubenswrapper[4870]: I1014 07:15:44.317659 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-64bf5d555-f8wq5"] Oct 14 07:15:44 crc kubenswrapper[4870]: W1014 07:15:44.325359 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd0a20afe_74ac_4d3f_80ca_981fb5ce00a8.slice/crio-96a54c7122eaa089064355c20e5f9bc508785d6e413914192fcf1ef303a177d5 WatchSource:0}: Error finding container 96a54c7122eaa089064355c20e5f9bc508785d6e413914192fcf1ef303a177d5: Status 404 returned error can't find the container with id 96a54c7122eaa089064355c20e5f9bc508785d6e413914192fcf1ef303a177d5 Oct 14 07:15:44 crc kubenswrapper[4870]: I1014 07:15:44.774823 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-f8wq5" event={"ID":"d0a20afe-74ac-4d3f-80ca-981fb5ce00a8","Type":"ContainerStarted","Data":"96a54c7122eaa089064355c20e5f9bc508785d6e413914192fcf1ef303a177d5"} Oct 14 07:15:44 crc kubenswrapper[4870]: I1014 07:15:44.775909 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-pbtkp" event={"ID":"2f058085-57d8-48a8-9267-fde99e4c15fb","Type":"ContainerStarted","Data":"1f3031595ab1b073e596fa43608aa6fe343b8721d961ed16c3b15d779486eb80"} Oct 14 07:15:44 crc kubenswrapper[4870]: I1014 07:15:44.777406 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-68d546b9d8-bs82b" event={"ID":"c7f0d440-b7cc-440a-9aac-e19c2bafab73","Type":"ContainerStarted","Data":"26486b62e8d793ee9da395c39214cd0c23c32bb4be21578fc9736452db4e6ae0"} Oct 14 07:15:44 crc kubenswrapper[4870]: I1014 07:15:44.777634 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-68d546b9d8-bs82b" event={"ID":"c7f0d440-b7cc-440a-9aac-e19c2bafab73","Type":"ContainerStarted","Data":"6016a98819830c42067453d341d4f5c4ff6cedc1f36ee512c8bed3a0c9518770"} Oct 14 07:15:44 crc kubenswrapper[4870]: I1014 07:15:44.777750 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-68d546b9d8-bs82b" event={"ID":"c7f0d440-b7cc-440a-9aac-e19c2bafab73","Type":"ContainerStarted","Data":"18db9ac1bce8d41697b8f7a48aa831f589088a9c004619c1b228632f027eab1a"} Oct 14 07:15:44 crc kubenswrapper[4870]: I1014 07:15:44.777860 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-68d546b9d8-bs82b" Oct 14 07:15:44 crc kubenswrapper[4870]: I1014 07:15:44.919494 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/010032d8-4503-4fb5-8158-066872d97d52-memberlist\") pod \"speaker-mtvtd\" (UID: \"010032d8-4503-4fb5-8158-066872d97d52\") " pod="metallb-system/speaker-mtvtd" Oct 14 07:15:44 crc kubenswrapper[4870]: I1014 07:15:44.933088 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/010032d8-4503-4fb5-8158-066872d97d52-memberlist\") pod \"speaker-mtvtd\" (UID: \"010032d8-4503-4fb5-8158-066872d97d52\") " pod="metallb-system/speaker-mtvtd" Oct 14 07:15:45 crc kubenswrapper[4870]: I1014 07:15:45.022378 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-mtvtd" Oct 14 07:15:45 crc kubenswrapper[4870]: W1014 07:15:45.051678 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod010032d8_4503_4fb5_8158_066872d97d52.slice/crio-1cad843554981f4735afd243cd9c4692187689b604aae76a590471915014fbf6 WatchSource:0}: Error finding container 1cad843554981f4735afd243cd9c4692187689b604aae76a590471915014fbf6: Status 404 returned error can't find the container with id 1cad843554981f4735afd243cd9c4692187689b604aae76a590471915014fbf6 Oct 14 07:15:45 crc kubenswrapper[4870]: I1014 07:15:45.066541 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-68d546b9d8-bs82b" podStartSLOduration=2.0665231 podStartE2EDuration="2.0665231s" podCreationTimestamp="2025-10-14 07:15:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:15:44.801685842 +0000 UTC m=+880.499046223" watchObservedRunningTime="2025-10-14 07:15:45.0665231 +0000 UTC m=+880.763883471" Oct 14 07:15:45 crc kubenswrapper[4870]: I1014 07:15:45.788498 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-mtvtd" event={"ID":"010032d8-4503-4fb5-8158-066872d97d52","Type":"ContainerStarted","Data":"932edc8f4aed178326ce4460d2228fbbf4573048f32cd21b1aa304dc9d323f9e"} Oct 14 07:15:45 crc kubenswrapper[4870]: I1014 07:15:45.788548 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-mtvtd" event={"ID":"010032d8-4503-4fb5-8158-066872d97d52","Type":"ContainerStarted","Data":"f5a0fd4a9274f79ae389fa2405a597ebd97c612dcb7f63243dbfdaf78f794646"} Oct 14 07:15:45 crc kubenswrapper[4870]: I1014 07:15:45.788568 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-mtvtd" event={"ID":"010032d8-4503-4fb5-8158-066872d97d52","Type":"ContainerStarted","Data":"1cad843554981f4735afd243cd9c4692187689b604aae76a590471915014fbf6"} Oct 14 07:15:45 crc kubenswrapper[4870]: I1014 07:15:45.789395 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-mtvtd" Oct 14 07:15:45 crc kubenswrapper[4870]: I1014 07:15:45.807944 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-mtvtd" podStartSLOduration=2.807907037 podStartE2EDuration="2.807907037s" podCreationTimestamp="2025-10-14 07:15:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:15:45.806581064 +0000 UTC m=+881.503941475" watchObservedRunningTime="2025-10-14 07:15:45.807907037 +0000 UTC m=+881.505267438" Oct 14 07:15:51 crc kubenswrapper[4870]: I1014 07:15:51.839078 4870 generic.go:334] "Generic (PLEG): container finished" podID="2f058085-57d8-48a8-9267-fde99e4c15fb" containerID="39ea957434003494623374dfa349686dc983f696b2fc01643e19e3e8d1b4be99" exitCode=0 Oct 14 07:15:51 crc kubenswrapper[4870]: I1014 07:15:51.839149 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-pbtkp" event={"ID":"2f058085-57d8-48a8-9267-fde99e4c15fb","Type":"ContainerDied","Data":"39ea957434003494623374dfa349686dc983f696b2fc01643e19e3e8d1b4be99"} Oct 14 07:15:51 crc kubenswrapper[4870]: I1014 07:15:51.843333 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-f8wq5" event={"ID":"d0a20afe-74ac-4d3f-80ca-981fb5ce00a8","Type":"ContainerStarted","Data":"c8b41591fb08ba2e3e42626b470f6a9317ea484ec0bf7cfa7729d604759cb8a2"} Oct 14 07:15:51 crc kubenswrapper[4870]: I1014 07:15:51.843937 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-f8wq5" Oct 14 07:15:51 crc kubenswrapper[4870]: I1014 07:15:51.901140 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-f8wq5" podStartSLOduration=1.605615155 podStartE2EDuration="8.901118221s" podCreationTimestamp="2025-10-14 07:15:43 +0000 UTC" firstStartedPulling="2025-10-14 07:15:44.327101243 +0000 UTC m=+880.024461634" lastFinishedPulling="2025-10-14 07:15:51.622604329 +0000 UTC m=+887.319964700" observedRunningTime="2025-10-14 07:15:51.897802378 +0000 UTC m=+887.595162759" watchObservedRunningTime="2025-10-14 07:15:51.901118221 +0000 UTC m=+887.598478602" Oct 14 07:15:52 crc kubenswrapper[4870]: I1014 07:15:52.856240 4870 generic.go:334] "Generic (PLEG): container finished" podID="2f058085-57d8-48a8-9267-fde99e4c15fb" containerID="fbe83bd20766dccf6db1cc262898b8493740957114f90026df65233aad0827bf" exitCode=0 Oct 14 07:15:52 crc kubenswrapper[4870]: I1014 07:15:52.856309 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-pbtkp" event={"ID":"2f058085-57d8-48a8-9267-fde99e4c15fb","Type":"ContainerDied","Data":"fbe83bd20766dccf6db1cc262898b8493740957114f90026df65233aad0827bf"} Oct 14 07:15:53 crc kubenswrapper[4870]: I1014 07:15:53.864583 4870 generic.go:334] "Generic (PLEG): container finished" podID="2f058085-57d8-48a8-9267-fde99e4c15fb" containerID="1ac23a1b97e7fde333d5aadfa42fc6342103461f4b8c54c1ee0a59e18167e174" exitCode=0 Oct 14 07:15:53 crc kubenswrapper[4870]: I1014 07:15:53.865301 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-pbtkp" event={"ID":"2f058085-57d8-48a8-9267-fde99e4c15fb","Type":"ContainerDied","Data":"1ac23a1b97e7fde333d5aadfa42fc6342103461f4b8c54c1ee0a59e18167e174"} Oct 14 07:15:54 crc kubenswrapper[4870]: I1014 07:15:54.882668 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-pbtkp" event={"ID":"2f058085-57d8-48a8-9267-fde99e4c15fb","Type":"ContainerStarted","Data":"9dccce6466e0ab1fb1e81c7ac7ccc15dd98551c3310a74f955eada8e12aea071"} Oct 14 07:15:54 crc kubenswrapper[4870]: I1014 07:15:54.883066 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-pbtkp" event={"ID":"2f058085-57d8-48a8-9267-fde99e4c15fb","Type":"ContainerStarted","Data":"2e40d41c9ba647381cfc2f6fe395bbce358fa6886f6026d20f0edfa997a3a42c"} Oct 14 07:15:54 crc kubenswrapper[4870]: I1014 07:15:54.883082 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-pbtkp" event={"ID":"2f058085-57d8-48a8-9267-fde99e4c15fb","Type":"ContainerStarted","Data":"baffd2b93d9079dc63bd5d4dcb86dd5ac5ab196f7b52b660b924e0da8e784ac6"} Oct 14 07:15:54 crc kubenswrapper[4870]: I1014 07:15:54.883095 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-pbtkp" event={"ID":"2f058085-57d8-48a8-9267-fde99e4c15fb","Type":"ContainerStarted","Data":"62e71e8ec76f89a026288e05bde2069c46b1ca68163302b2198e4dadaf74b7ba"} Oct 14 07:15:54 crc kubenswrapper[4870]: I1014 07:15:54.883126 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-pbtkp" event={"ID":"2f058085-57d8-48a8-9267-fde99e4c15fb","Type":"ContainerStarted","Data":"7d0cd5eaf8a95cf275d54dd7b4aeb677158031fa72b79e86e0b0821d07f66082"} Oct 14 07:15:55 crc kubenswrapper[4870]: I1014 07:15:55.027392 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-mtvtd" Oct 14 07:15:55 crc kubenswrapper[4870]: I1014 07:15:55.932293 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-pbtkp" event={"ID":"2f058085-57d8-48a8-9267-fde99e4c15fb","Type":"ContainerStarted","Data":"8d778c5a50f5645fb51b8122b67b071e4481e1d49a03ffd53e073c1e22b33fc2"} Oct 14 07:15:55 crc kubenswrapper[4870]: I1014 07:15:55.932739 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-pbtkp" Oct 14 07:15:55 crc kubenswrapper[4870]: I1014 07:15:55.962815 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-pbtkp" podStartSLOduration=5.586492945 podStartE2EDuration="12.962792194s" podCreationTimestamp="2025-10-14 07:15:43 +0000 UTC" firstStartedPulling="2025-10-14 07:15:44.210377031 +0000 UTC m=+879.907737422" lastFinishedPulling="2025-10-14 07:15:51.5866763 +0000 UTC m=+887.284036671" observedRunningTime="2025-10-14 07:15:55.959175223 +0000 UTC m=+891.656535594" watchObservedRunningTime="2025-10-14 07:15:55.962792194 +0000 UTC m=+891.660152585" Oct 14 07:15:56 crc kubenswrapper[4870]: I1014 07:15:56.476985 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69dkdfp"] Oct 14 07:15:56 crc kubenswrapper[4870]: I1014 07:15:56.479435 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69dkdfp" Oct 14 07:15:56 crc kubenswrapper[4870]: I1014 07:15:56.483593 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Oct 14 07:15:56 crc kubenswrapper[4870]: I1014 07:15:56.493637 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69dkdfp"] Oct 14 07:15:56 crc kubenswrapper[4870]: I1014 07:15:56.639135 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ebb7cfd0-0388-4aca-99f6-7669c6bc8d4e-util\") pod \"695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69dkdfp\" (UID: \"ebb7cfd0-0388-4aca-99f6-7669c6bc8d4e\") " pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69dkdfp" Oct 14 07:15:56 crc kubenswrapper[4870]: I1014 07:15:56.639264 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ebb7cfd0-0388-4aca-99f6-7669c6bc8d4e-bundle\") pod \"695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69dkdfp\" (UID: \"ebb7cfd0-0388-4aca-99f6-7669c6bc8d4e\") " pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69dkdfp" Oct 14 07:15:56 crc kubenswrapper[4870]: I1014 07:15:56.639338 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x5whp\" (UniqueName: \"kubernetes.io/projected/ebb7cfd0-0388-4aca-99f6-7669c6bc8d4e-kube-api-access-x5whp\") pod \"695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69dkdfp\" (UID: \"ebb7cfd0-0388-4aca-99f6-7669c6bc8d4e\") " pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69dkdfp" Oct 14 07:15:56 crc kubenswrapper[4870]: I1014 07:15:56.740585 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ebb7cfd0-0388-4aca-99f6-7669c6bc8d4e-util\") pod \"695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69dkdfp\" (UID: \"ebb7cfd0-0388-4aca-99f6-7669c6bc8d4e\") " pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69dkdfp" Oct 14 07:15:56 crc kubenswrapper[4870]: I1014 07:15:56.740906 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ebb7cfd0-0388-4aca-99f6-7669c6bc8d4e-bundle\") pod \"695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69dkdfp\" (UID: \"ebb7cfd0-0388-4aca-99f6-7669c6bc8d4e\") " pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69dkdfp" Oct 14 07:15:56 crc kubenswrapper[4870]: I1014 07:15:56.741053 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x5whp\" (UniqueName: \"kubernetes.io/projected/ebb7cfd0-0388-4aca-99f6-7669c6bc8d4e-kube-api-access-x5whp\") pod \"695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69dkdfp\" (UID: \"ebb7cfd0-0388-4aca-99f6-7669c6bc8d4e\") " pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69dkdfp" Oct 14 07:15:56 crc kubenswrapper[4870]: I1014 07:15:56.741942 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ebb7cfd0-0388-4aca-99f6-7669c6bc8d4e-bundle\") pod \"695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69dkdfp\" (UID: \"ebb7cfd0-0388-4aca-99f6-7669c6bc8d4e\") " pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69dkdfp" Oct 14 07:15:56 crc kubenswrapper[4870]: I1014 07:15:56.742017 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ebb7cfd0-0388-4aca-99f6-7669c6bc8d4e-util\") pod \"695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69dkdfp\" (UID: \"ebb7cfd0-0388-4aca-99f6-7669c6bc8d4e\") " pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69dkdfp" Oct 14 07:15:56 crc kubenswrapper[4870]: I1014 07:15:56.771399 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x5whp\" (UniqueName: \"kubernetes.io/projected/ebb7cfd0-0388-4aca-99f6-7669c6bc8d4e-kube-api-access-x5whp\") pod \"695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69dkdfp\" (UID: \"ebb7cfd0-0388-4aca-99f6-7669c6bc8d4e\") " pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69dkdfp" Oct 14 07:15:56 crc kubenswrapper[4870]: I1014 07:15:56.798434 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69dkdfp" Oct 14 07:15:57 crc kubenswrapper[4870]: I1014 07:15:57.301458 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69dkdfp"] Oct 14 07:15:57 crc kubenswrapper[4870]: W1014 07:15:57.313670 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podebb7cfd0_0388_4aca_99f6_7669c6bc8d4e.slice/crio-08f5113a171da98d668b4916d6cb75a2135cae3ce7425f4ed9e6de9ba7a088c0 WatchSource:0}: Error finding container 08f5113a171da98d668b4916d6cb75a2135cae3ce7425f4ed9e6de9ba7a088c0: Status 404 returned error can't find the container with id 08f5113a171da98d668b4916d6cb75a2135cae3ce7425f4ed9e6de9ba7a088c0 Oct 14 07:15:57 crc kubenswrapper[4870]: I1014 07:15:57.947246 4870 generic.go:334] "Generic (PLEG): container finished" podID="ebb7cfd0-0388-4aca-99f6-7669c6bc8d4e" containerID="c8b1637800db5ecb56422e011669ad1194df159faba3f82e69108ac18ddb3527" exitCode=0 Oct 14 07:15:57 crc kubenswrapper[4870]: I1014 07:15:57.947294 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69dkdfp" event={"ID":"ebb7cfd0-0388-4aca-99f6-7669c6bc8d4e","Type":"ContainerDied","Data":"c8b1637800db5ecb56422e011669ad1194df159faba3f82e69108ac18ddb3527"} Oct 14 07:15:57 crc kubenswrapper[4870]: I1014 07:15:57.947602 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69dkdfp" event={"ID":"ebb7cfd0-0388-4aca-99f6-7669c6bc8d4e","Type":"ContainerStarted","Data":"08f5113a171da98d668b4916d6cb75a2135cae3ce7425f4ed9e6de9ba7a088c0"} Oct 14 07:15:59 crc kubenswrapper[4870]: I1014 07:15:59.054323 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-pbtkp" Oct 14 07:15:59 crc kubenswrapper[4870]: I1014 07:15:59.109179 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-pbtkp" Oct 14 07:16:01 crc kubenswrapper[4870]: I1014 07:16:01.977605 4870 generic.go:334] "Generic (PLEG): container finished" podID="ebb7cfd0-0388-4aca-99f6-7669c6bc8d4e" containerID="95768f7e7e8b5e25b6725ee1a10536b3141e0e273ae99b932e6c4b64b09818ab" exitCode=0 Oct 14 07:16:01 crc kubenswrapper[4870]: I1014 07:16:01.977726 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69dkdfp" event={"ID":"ebb7cfd0-0388-4aca-99f6-7669c6bc8d4e","Type":"ContainerDied","Data":"95768f7e7e8b5e25b6725ee1a10536b3141e0e273ae99b932e6c4b64b09818ab"} Oct 14 07:16:02 crc kubenswrapper[4870]: I1014 07:16:02.987214 4870 generic.go:334] "Generic (PLEG): container finished" podID="ebb7cfd0-0388-4aca-99f6-7669c6bc8d4e" containerID="310cb2f96f75ffbe351ec9cc4f1f57739041657b4809b9f981d6817c0dcd68e8" exitCode=0 Oct 14 07:16:02 crc kubenswrapper[4870]: I1014 07:16:02.987306 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69dkdfp" event={"ID":"ebb7cfd0-0388-4aca-99f6-7669c6bc8d4e","Type":"ContainerDied","Data":"310cb2f96f75ffbe351ec9cc4f1f57739041657b4809b9f981d6817c0dcd68e8"} Oct 14 07:16:03 crc kubenswrapper[4870]: I1014 07:16:03.537613 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-68d546b9d8-bs82b" Oct 14 07:16:04 crc kubenswrapper[4870]: I1014 07:16:04.057990 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-f8wq5" Oct 14 07:16:04 crc kubenswrapper[4870]: I1014 07:16:04.058712 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-pbtkp" Oct 14 07:16:04 crc kubenswrapper[4870]: I1014 07:16:04.294662 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69dkdfp" Oct 14 07:16:04 crc kubenswrapper[4870]: I1014 07:16:04.371501 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ebb7cfd0-0388-4aca-99f6-7669c6bc8d4e-bundle\") pod \"ebb7cfd0-0388-4aca-99f6-7669c6bc8d4e\" (UID: \"ebb7cfd0-0388-4aca-99f6-7669c6bc8d4e\") " Oct 14 07:16:04 crc kubenswrapper[4870]: I1014 07:16:04.371582 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ebb7cfd0-0388-4aca-99f6-7669c6bc8d4e-util\") pod \"ebb7cfd0-0388-4aca-99f6-7669c6bc8d4e\" (UID: \"ebb7cfd0-0388-4aca-99f6-7669c6bc8d4e\") " Oct 14 07:16:04 crc kubenswrapper[4870]: I1014 07:16:04.371645 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x5whp\" (UniqueName: \"kubernetes.io/projected/ebb7cfd0-0388-4aca-99f6-7669c6bc8d4e-kube-api-access-x5whp\") pod \"ebb7cfd0-0388-4aca-99f6-7669c6bc8d4e\" (UID: \"ebb7cfd0-0388-4aca-99f6-7669c6bc8d4e\") " Oct 14 07:16:04 crc kubenswrapper[4870]: I1014 07:16:04.373236 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ebb7cfd0-0388-4aca-99f6-7669c6bc8d4e-bundle" (OuterVolumeSpecName: "bundle") pod "ebb7cfd0-0388-4aca-99f6-7669c6bc8d4e" (UID: "ebb7cfd0-0388-4aca-99f6-7669c6bc8d4e"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:16:04 crc kubenswrapper[4870]: I1014 07:16:04.380343 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ebb7cfd0-0388-4aca-99f6-7669c6bc8d4e-kube-api-access-x5whp" (OuterVolumeSpecName: "kube-api-access-x5whp") pod "ebb7cfd0-0388-4aca-99f6-7669c6bc8d4e" (UID: "ebb7cfd0-0388-4aca-99f6-7669c6bc8d4e"). InnerVolumeSpecName "kube-api-access-x5whp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:16:04 crc kubenswrapper[4870]: I1014 07:16:04.386651 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ebb7cfd0-0388-4aca-99f6-7669c6bc8d4e-util" (OuterVolumeSpecName: "util") pod "ebb7cfd0-0388-4aca-99f6-7669c6bc8d4e" (UID: "ebb7cfd0-0388-4aca-99f6-7669c6bc8d4e"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:16:04 crc kubenswrapper[4870]: I1014 07:16:04.472788 4870 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ebb7cfd0-0388-4aca-99f6-7669c6bc8d4e-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 07:16:04 crc kubenswrapper[4870]: I1014 07:16:04.472833 4870 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ebb7cfd0-0388-4aca-99f6-7669c6bc8d4e-util\") on node \"crc\" DevicePath \"\"" Oct 14 07:16:04 crc kubenswrapper[4870]: I1014 07:16:04.472848 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x5whp\" (UniqueName: \"kubernetes.io/projected/ebb7cfd0-0388-4aca-99f6-7669c6bc8d4e-kube-api-access-x5whp\") on node \"crc\" DevicePath \"\"" Oct 14 07:16:05 crc kubenswrapper[4870]: I1014 07:16:05.006262 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69dkdfp" event={"ID":"ebb7cfd0-0388-4aca-99f6-7669c6bc8d4e","Type":"ContainerDied","Data":"08f5113a171da98d668b4916d6cb75a2135cae3ce7425f4ed9e6de9ba7a088c0"} Oct 14 07:16:05 crc kubenswrapper[4870]: I1014 07:16:05.006318 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69dkdfp" Oct 14 07:16:05 crc kubenswrapper[4870]: I1014 07:16:05.006333 4870 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="08f5113a171da98d668b4916d6cb75a2135cae3ce7425f4ed9e6de9ba7a088c0" Oct 14 07:16:08 crc kubenswrapper[4870]: E1014 07:16:08.071230 4870 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podebb7cfd0_0388_4aca_99f6_7669c6bc8d4e.slice\": RecentStats: unable to find data in memory cache]" Oct 14 07:16:09 crc kubenswrapper[4870]: I1014 07:16:09.289320 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-57cd46d6d-qp242"] Oct 14 07:16:09 crc kubenswrapper[4870]: E1014 07:16:09.289561 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ebb7cfd0-0388-4aca-99f6-7669c6bc8d4e" containerName="extract" Oct 14 07:16:09 crc kubenswrapper[4870]: I1014 07:16:09.289577 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="ebb7cfd0-0388-4aca-99f6-7669c6bc8d4e" containerName="extract" Oct 14 07:16:09 crc kubenswrapper[4870]: E1014 07:16:09.289590 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ebb7cfd0-0388-4aca-99f6-7669c6bc8d4e" containerName="pull" Oct 14 07:16:09 crc kubenswrapper[4870]: I1014 07:16:09.289598 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="ebb7cfd0-0388-4aca-99f6-7669c6bc8d4e" containerName="pull" Oct 14 07:16:09 crc kubenswrapper[4870]: E1014 07:16:09.289608 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ebb7cfd0-0388-4aca-99f6-7669c6bc8d4e" containerName="util" Oct 14 07:16:09 crc kubenswrapper[4870]: I1014 07:16:09.289616 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="ebb7cfd0-0388-4aca-99f6-7669c6bc8d4e" containerName="util" Oct 14 07:16:09 crc kubenswrapper[4870]: I1014 07:16:09.289711 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="ebb7cfd0-0388-4aca-99f6-7669c6bc8d4e" containerName="extract" Oct 14 07:16:09 crc kubenswrapper[4870]: I1014 07:16:09.290080 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager-operator/cert-manager-operator-controller-manager-57cd46d6d-qp242" Oct 14 07:16:09 crc kubenswrapper[4870]: I1014 07:16:09.291990 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager-operator"/"openshift-service-ca.crt" Oct 14 07:16:09 crc kubenswrapper[4870]: I1014 07:16:09.293455 4870 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager-operator"/"cert-manager-operator-controller-manager-dockercfg-zqqp4" Oct 14 07:16:09 crc kubenswrapper[4870]: I1014 07:16:09.294860 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager-operator"/"kube-root-ca.crt" Oct 14 07:16:09 crc kubenswrapper[4870]: I1014 07:16:09.308205 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-57cd46d6d-qp242"] Oct 14 07:16:09 crc kubenswrapper[4870]: I1014 07:16:09.335232 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zsxxr\" (UniqueName: \"kubernetes.io/projected/b5000042-14eb-4e8d-be3f-a664ff4df0af-kube-api-access-zsxxr\") pod \"cert-manager-operator-controller-manager-57cd46d6d-qp242\" (UID: \"b5000042-14eb-4e8d-be3f-a664ff4df0af\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-57cd46d6d-qp242" Oct 14 07:16:09 crc kubenswrapper[4870]: I1014 07:16:09.436529 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zsxxr\" (UniqueName: \"kubernetes.io/projected/b5000042-14eb-4e8d-be3f-a664ff4df0af-kube-api-access-zsxxr\") pod \"cert-manager-operator-controller-manager-57cd46d6d-qp242\" (UID: \"b5000042-14eb-4e8d-be3f-a664ff4df0af\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-57cd46d6d-qp242" Oct 14 07:16:09 crc kubenswrapper[4870]: I1014 07:16:09.456463 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zsxxr\" (UniqueName: \"kubernetes.io/projected/b5000042-14eb-4e8d-be3f-a664ff4df0af-kube-api-access-zsxxr\") pod \"cert-manager-operator-controller-manager-57cd46d6d-qp242\" (UID: \"b5000042-14eb-4e8d-be3f-a664ff4df0af\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-57cd46d6d-qp242" Oct 14 07:16:09 crc kubenswrapper[4870]: I1014 07:16:09.606200 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager-operator/cert-manager-operator-controller-manager-57cd46d6d-qp242" Oct 14 07:16:09 crc kubenswrapper[4870]: I1014 07:16:09.907592 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-57cd46d6d-qp242"] Oct 14 07:16:10 crc kubenswrapper[4870]: I1014 07:16:10.032568 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager-operator/cert-manager-operator-controller-manager-57cd46d6d-qp242" event={"ID":"b5000042-14eb-4e8d-be3f-a664ff4df0af","Type":"ContainerStarted","Data":"6e5eb957051d7008a95dd50550e1b8d8e5a242ae74506eb49a53dee85aaacc5d"} Oct 14 07:16:17 crc kubenswrapper[4870]: I1014 07:16:17.079506 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager-operator/cert-manager-operator-controller-manager-57cd46d6d-qp242" event={"ID":"b5000042-14eb-4e8d-be3f-a664ff4df0af","Type":"ContainerStarted","Data":"d635c1b49ce298825e3d39a69da5175a1f0010a384f3dfe5961ca8d7e2995f41"} Oct 14 07:16:17 crc kubenswrapper[4870]: I1014 07:16:17.111383 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager-operator/cert-manager-operator-controller-manager-57cd46d6d-qp242" podStartSLOduration=1.60782855 podStartE2EDuration="8.111356794s" podCreationTimestamp="2025-10-14 07:16:09 +0000 UTC" firstStartedPulling="2025-10-14 07:16:09.926564197 +0000 UTC m=+905.623924568" lastFinishedPulling="2025-10-14 07:16:16.430092441 +0000 UTC m=+912.127452812" observedRunningTime="2025-10-14 07:16:17.107926325 +0000 UTC m=+912.805286736" watchObservedRunningTime="2025-10-14 07:16:17.111356794 +0000 UTC m=+912.808717215" Oct 14 07:16:18 crc kubenswrapper[4870]: E1014 07:16:18.219610 4870 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podebb7cfd0_0388_4aca_99f6_7669c6bc8d4e.slice\": RecentStats: unable to find data in memory cache]" Oct 14 07:16:20 crc kubenswrapper[4870]: I1014 07:16:20.438407 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-d969966f-wfx5l"] Oct 14 07:16:20 crc kubenswrapper[4870]: I1014 07:16:20.439664 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-d969966f-wfx5l" Oct 14 07:16:20 crc kubenswrapper[4870]: I1014 07:16:20.443011 4870 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-67k2x" Oct 14 07:16:20 crc kubenswrapper[4870]: I1014 07:16:20.443085 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Oct 14 07:16:20 crc kubenswrapper[4870]: I1014 07:16:20.443261 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Oct 14 07:16:20 crc kubenswrapper[4870]: I1014 07:16:20.456899 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-d969966f-wfx5l"] Oct 14 07:16:20 crc kubenswrapper[4870]: I1014 07:16:20.499593 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/6ae15c75-9b7b-4583-8ac9-f60db231ae06-bound-sa-token\") pod \"cert-manager-webhook-d969966f-wfx5l\" (UID: \"6ae15c75-9b7b-4583-8ac9-f60db231ae06\") " pod="cert-manager/cert-manager-webhook-d969966f-wfx5l" Oct 14 07:16:20 crc kubenswrapper[4870]: I1014 07:16:20.499675 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rzp5v\" (UniqueName: \"kubernetes.io/projected/6ae15c75-9b7b-4583-8ac9-f60db231ae06-kube-api-access-rzp5v\") pod \"cert-manager-webhook-d969966f-wfx5l\" (UID: \"6ae15c75-9b7b-4583-8ac9-f60db231ae06\") " pod="cert-manager/cert-manager-webhook-d969966f-wfx5l" Oct 14 07:16:20 crc kubenswrapper[4870]: I1014 07:16:20.601827 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rzp5v\" (UniqueName: \"kubernetes.io/projected/6ae15c75-9b7b-4583-8ac9-f60db231ae06-kube-api-access-rzp5v\") pod \"cert-manager-webhook-d969966f-wfx5l\" (UID: \"6ae15c75-9b7b-4583-8ac9-f60db231ae06\") " pod="cert-manager/cert-manager-webhook-d969966f-wfx5l" Oct 14 07:16:20 crc kubenswrapper[4870]: I1014 07:16:20.602197 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/6ae15c75-9b7b-4583-8ac9-f60db231ae06-bound-sa-token\") pod \"cert-manager-webhook-d969966f-wfx5l\" (UID: \"6ae15c75-9b7b-4583-8ac9-f60db231ae06\") " pod="cert-manager/cert-manager-webhook-d969966f-wfx5l" Oct 14 07:16:20 crc kubenswrapper[4870]: I1014 07:16:20.631400 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rzp5v\" (UniqueName: \"kubernetes.io/projected/6ae15c75-9b7b-4583-8ac9-f60db231ae06-kube-api-access-rzp5v\") pod \"cert-manager-webhook-d969966f-wfx5l\" (UID: \"6ae15c75-9b7b-4583-8ac9-f60db231ae06\") " pod="cert-manager/cert-manager-webhook-d969966f-wfx5l" Oct 14 07:16:20 crc kubenswrapper[4870]: I1014 07:16:20.631775 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/6ae15c75-9b7b-4583-8ac9-f60db231ae06-bound-sa-token\") pod \"cert-manager-webhook-d969966f-wfx5l\" (UID: \"6ae15c75-9b7b-4583-8ac9-f60db231ae06\") " pod="cert-manager/cert-manager-webhook-d969966f-wfx5l" Oct 14 07:16:20 crc kubenswrapper[4870]: I1014 07:16:20.761666 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-d969966f-wfx5l" Oct 14 07:16:21 crc kubenswrapper[4870]: I1014 07:16:21.180556 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-7d9f95dbf-ddfsr"] Oct 14 07:16:21 crc kubenswrapper[4870]: I1014 07:16:21.181769 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7d9f95dbf-ddfsr" Oct 14 07:16:21 crc kubenswrapper[4870]: I1014 07:16:21.183533 4870 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-z4vrd" Oct 14 07:16:21 crc kubenswrapper[4870]: I1014 07:16:21.188081 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7d9f95dbf-ddfsr"] Oct 14 07:16:21 crc kubenswrapper[4870]: I1014 07:16:21.312640 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/06432b45-ce1d-4ace-b390-476fc76b06dc-bound-sa-token\") pod \"cert-manager-cainjector-7d9f95dbf-ddfsr\" (UID: \"06432b45-ce1d-4ace-b390-476fc76b06dc\") " pod="cert-manager/cert-manager-cainjector-7d9f95dbf-ddfsr" Oct 14 07:16:21 crc kubenswrapper[4870]: I1014 07:16:21.312780 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q7mqb\" (UniqueName: \"kubernetes.io/projected/06432b45-ce1d-4ace-b390-476fc76b06dc-kube-api-access-q7mqb\") pod \"cert-manager-cainjector-7d9f95dbf-ddfsr\" (UID: \"06432b45-ce1d-4ace-b390-476fc76b06dc\") " pod="cert-manager/cert-manager-cainjector-7d9f95dbf-ddfsr" Oct 14 07:16:21 crc kubenswrapper[4870]: I1014 07:16:21.378406 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-d969966f-wfx5l"] Oct 14 07:16:21 crc kubenswrapper[4870]: W1014 07:16:21.385669 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6ae15c75_9b7b_4583_8ac9_f60db231ae06.slice/crio-54dcc24ce4875e6aa1f2dca178ca404efffd22ccd16148c3cb524f868c323b47 WatchSource:0}: Error finding container 54dcc24ce4875e6aa1f2dca178ca404efffd22ccd16148c3cb524f868c323b47: Status 404 returned error can't find the container with id 54dcc24ce4875e6aa1f2dca178ca404efffd22ccd16148c3cb524f868c323b47 Oct 14 07:16:21 crc kubenswrapper[4870]: I1014 07:16:21.413612 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/06432b45-ce1d-4ace-b390-476fc76b06dc-bound-sa-token\") pod \"cert-manager-cainjector-7d9f95dbf-ddfsr\" (UID: \"06432b45-ce1d-4ace-b390-476fc76b06dc\") " pod="cert-manager/cert-manager-cainjector-7d9f95dbf-ddfsr" Oct 14 07:16:21 crc kubenswrapper[4870]: I1014 07:16:21.413683 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q7mqb\" (UniqueName: \"kubernetes.io/projected/06432b45-ce1d-4ace-b390-476fc76b06dc-kube-api-access-q7mqb\") pod \"cert-manager-cainjector-7d9f95dbf-ddfsr\" (UID: \"06432b45-ce1d-4ace-b390-476fc76b06dc\") " pod="cert-manager/cert-manager-cainjector-7d9f95dbf-ddfsr" Oct 14 07:16:21 crc kubenswrapper[4870]: I1014 07:16:21.440753 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/06432b45-ce1d-4ace-b390-476fc76b06dc-bound-sa-token\") pod \"cert-manager-cainjector-7d9f95dbf-ddfsr\" (UID: \"06432b45-ce1d-4ace-b390-476fc76b06dc\") " pod="cert-manager/cert-manager-cainjector-7d9f95dbf-ddfsr" Oct 14 07:16:21 crc kubenswrapper[4870]: I1014 07:16:21.450145 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q7mqb\" (UniqueName: \"kubernetes.io/projected/06432b45-ce1d-4ace-b390-476fc76b06dc-kube-api-access-q7mqb\") pod \"cert-manager-cainjector-7d9f95dbf-ddfsr\" (UID: \"06432b45-ce1d-4ace-b390-476fc76b06dc\") " pod="cert-manager/cert-manager-cainjector-7d9f95dbf-ddfsr" Oct 14 07:16:21 crc kubenswrapper[4870]: I1014 07:16:21.505808 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7d9f95dbf-ddfsr" Oct 14 07:16:21 crc kubenswrapper[4870]: I1014 07:16:21.974612 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7d9f95dbf-ddfsr"] Oct 14 07:16:21 crc kubenswrapper[4870]: W1014 07:16:21.977596 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod06432b45_ce1d_4ace_b390_476fc76b06dc.slice/crio-18dae2a92e34ff25ab8c7eb6b2c036681cf08bf5f6f475390c033d9177af4d77 WatchSource:0}: Error finding container 18dae2a92e34ff25ab8c7eb6b2c036681cf08bf5f6f475390c033d9177af4d77: Status 404 returned error can't find the container with id 18dae2a92e34ff25ab8c7eb6b2c036681cf08bf5f6f475390c033d9177af4d77 Oct 14 07:16:22 crc kubenswrapper[4870]: I1014 07:16:22.111298 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7d9f95dbf-ddfsr" event={"ID":"06432b45-ce1d-4ace-b390-476fc76b06dc","Type":"ContainerStarted","Data":"18dae2a92e34ff25ab8c7eb6b2c036681cf08bf5f6f475390c033d9177af4d77"} Oct 14 07:16:22 crc kubenswrapper[4870]: I1014 07:16:22.113227 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-d969966f-wfx5l" event={"ID":"6ae15c75-9b7b-4583-8ac9-f60db231ae06","Type":"ContainerStarted","Data":"54dcc24ce4875e6aa1f2dca178ca404efffd22ccd16148c3cb524f868c323b47"} Oct 14 07:16:27 crc kubenswrapper[4870]: I1014 07:16:27.154220 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7d9f95dbf-ddfsr" event={"ID":"06432b45-ce1d-4ace-b390-476fc76b06dc","Type":"ContainerStarted","Data":"4735d82af42e14f509314eef064b98455e84a24e527d77e5f20ea7548b3299ab"} Oct 14 07:16:27 crc kubenswrapper[4870]: I1014 07:16:27.156265 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-d969966f-wfx5l" event={"ID":"6ae15c75-9b7b-4583-8ac9-f60db231ae06","Type":"ContainerStarted","Data":"71221974f1a23c84eb172f25d6407d3cb735d250f90df16766224c87ee786182"} Oct 14 07:16:27 crc kubenswrapper[4870]: I1014 07:16:27.156587 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-d969966f-wfx5l" Oct 14 07:16:27 crc kubenswrapper[4870]: I1014 07:16:27.176168 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-7d9f95dbf-ddfsr" podStartSLOduration=2.014942504 podStartE2EDuration="6.176135336s" podCreationTimestamp="2025-10-14 07:16:21 +0000 UTC" firstStartedPulling="2025-10-14 07:16:21.980604451 +0000 UTC m=+917.677964852" lastFinishedPulling="2025-10-14 07:16:26.141797313 +0000 UTC m=+921.839157684" observedRunningTime="2025-10-14 07:16:27.171785663 +0000 UTC m=+922.869146114" watchObservedRunningTime="2025-10-14 07:16:27.176135336 +0000 UTC m=+922.873495747" Oct 14 07:16:27 crc kubenswrapper[4870]: I1014 07:16:27.197589 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-d969966f-wfx5l" podStartSLOduration=2.468699571 podStartE2EDuration="7.197558658s" podCreationTimestamp="2025-10-14 07:16:20 +0000 UTC" firstStartedPulling="2025-10-14 07:16:21.388986699 +0000 UTC m=+917.086347080" lastFinishedPulling="2025-10-14 07:16:26.117845796 +0000 UTC m=+921.815206167" observedRunningTime="2025-10-14 07:16:27.190039334 +0000 UTC m=+922.887399735" watchObservedRunningTime="2025-10-14 07:16:27.197558658 +0000 UTC m=+922.894919079" Oct 14 07:16:28 crc kubenswrapper[4870]: E1014 07:16:28.371931 4870 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podebb7cfd0_0388_4aca_99f6_7669c6bc8d4e.slice\": RecentStats: unable to find data in memory cache]" Oct 14 07:16:35 crc kubenswrapper[4870]: I1014 07:16:35.765276 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-d969966f-wfx5l" Oct 14 07:16:38 crc kubenswrapper[4870]: E1014 07:16:38.539147 4870 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podebb7cfd0_0388_4aca_99f6_7669c6bc8d4e.slice\": RecentStats: unable to find data in memory cache]" Oct 14 07:16:40 crc kubenswrapper[4870]: I1014 07:16:40.140617 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-7d4cc89fcb-gvlll"] Oct 14 07:16:40 crc kubenswrapper[4870]: I1014 07:16:40.141674 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-7d4cc89fcb-gvlll" Oct 14 07:16:40 crc kubenswrapper[4870]: I1014 07:16:40.150663 4870 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-qwcv7" Oct 14 07:16:40 crc kubenswrapper[4870]: I1014 07:16:40.172428 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-7d4cc89fcb-gvlll"] Oct 14 07:16:40 crc kubenswrapper[4870]: I1014 07:16:40.208918 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d8377a33-a023-456d-811a-53be669f1178-bound-sa-token\") pod \"cert-manager-7d4cc89fcb-gvlll\" (UID: \"d8377a33-a023-456d-811a-53be669f1178\") " pod="cert-manager/cert-manager-7d4cc89fcb-gvlll" Oct 14 07:16:40 crc kubenswrapper[4870]: I1014 07:16:40.209131 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wg6zv\" (UniqueName: \"kubernetes.io/projected/d8377a33-a023-456d-811a-53be669f1178-kube-api-access-wg6zv\") pod \"cert-manager-7d4cc89fcb-gvlll\" (UID: \"d8377a33-a023-456d-811a-53be669f1178\") " pod="cert-manager/cert-manager-7d4cc89fcb-gvlll" Oct 14 07:16:40 crc kubenswrapper[4870]: I1014 07:16:40.311595 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wg6zv\" (UniqueName: \"kubernetes.io/projected/d8377a33-a023-456d-811a-53be669f1178-kube-api-access-wg6zv\") pod \"cert-manager-7d4cc89fcb-gvlll\" (UID: \"d8377a33-a023-456d-811a-53be669f1178\") " pod="cert-manager/cert-manager-7d4cc89fcb-gvlll" Oct 14 07:16:40 crc kubenswrapper[4870]: I1014 07:16:40.311817 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d8377a33-a023-456d-811a-53be669f1178-bound-sa-token\") pod \"cert-manager-7d4cc89fcb-gvlll\" (UID: \"d8377a33-a023-456d-811a-53be669f1178\") " pod="cert-manager/cert-manager-7d4cc89fcb-gvlll" Oct 14 07:16:40 crc kubenswrapper[4870]: I1014 07:16:40.348551 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d8377a33-a023-456d-811a-53be669f1178-bound-sa-token\") pod \"cert-manager-7d4cc89fcb-gvlll\" (UID: \"d8377a33-a023-456d-811a-53be669f1178\") " pod="cert-manager/cert-manager-7d4cc89fcb-gvlll" Oct 14 07:16:40 crc kubenswrapper[4870]: I1014 07:16:40.349358 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wg6zv\" (UniqueName: \"kubernetes.io/projected/d8377a33-a023-456d-811a-53be669f1178-kube-api-access-wg6zv\") pod \"cert-manager-7d4cc89fcb-gvlll\" (UID: \"d8377a33-a023-456d-811a-53be669f1178\") " pod="cert-manager/cert-manager-7d4cc89fcb-gvlll" Oct 14 07:16:40 crc kubenswrapper[4870]: I1014 07:16:40.465920 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-7d4cc89fcb-gvlll" Oct 14 07:16:40 crc kubenswrapper[4870]: I1014 07:16:40.891284 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-7d4cc89fcb-gvlll"] Oct 14 07:16:40 crc kubenswrapper[4870]: W1014 07:16:40.899352 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd8377a33_a023_456d_811a_53be669f1178.slice/crio-e02989b842d16423847dd200e6fa89c083a2f50ba51973dd34f897c02a3cdc6a WatchSource:0}: Error finding container e02989b842d16423847dd200e6fa89c083a2f50ba51973dd34f897c02a3cdc6a: Status 404 returned error can't find the container with id e02989b842d16423847dd200e6fa89c083a2f50ba51973dd34f897c02a3cdc6a Oct 14 07:16:41 crc kubenswrapper[4870]: I1014 07:16:41.265347 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-7d4cc89fcb-gvlll" event={"ID":"d8377a33-a023-456d-811a-53be669f1178","Type":"ContainerStarted","Data":"6b470f333bafbbe6c4a8477a6ba95c82e1b6d2aa4f771687e8c290e90967c850"} Oct 14 07:16:41 crc kubenswrapper[4870]: I1014 07:16:41.265423 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-7d4cc89fcb-gvlll" event={"ID":"d8377a33-a023-456d-811a-53be669f1178","Type":"ContainerStarted","Data":"e02989b842d16423847dd200e6fa89c083a2f50ba51973dd34f897c02a3cdc6a"} Oct 14 07:16:41 crc kubenswrapper[4870]: I1014 07:16:41.286830 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-7d4cc89fcb-gvlll" podStartSLOduration=1.2867954240000001 podStartE2EDuration="1.286795424s" podCreationTimestamp="2025-10-14 07:16:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:16:41.279408243 +0000 UTC m=+936.976768624" watchObservedRunningTime="2025-10-14 07:16:41.286795424 +0000 UTC m=+936.984155835" Oct 14 07:16:48 crc kubenswrapper[4870]: E1014 07:16:48.744375 4870 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podebb7cfd0_0388_4aca_99f6_7669c6bc8d4e.slice\": RecentStats: unable to find data in memory cache]" Oct 14 07:16:49 crc kubenswrapper[4870]: I1014 07:16:49.261852 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-7745w"] Oct 14 07:16:49 crc kubenswrapper[4870]: I1014 07:16:49.263021 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-7745w" Oct 14 07:16:49 crc kubenswrapper[4870]: I1014 07:16:49.266017 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Oct 14 07:16:49 crc kubenswrapper[4870]: I1014 07:16:49.267217 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Oct 14 07:16:49 crc kubenswrapper[4870]: I1014 07:16:49.282758 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-7745w"] Oct 14 07:16:49 crc kubenswrapper[4870]: I1014 07:16:49.284863 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-r5ggq" Oct 14 07:16:49 crc kubenswrapper[4870]: I1014 07:16:49.441098 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gxtxm\" (UniqueName: \"kubernetes.io/projected/ee7e419d-de23-461a-8c37-10c89b1a5c13-kube-api-access-gxtxm\") pod \"openstack-operator-index-7745w\" (UID: \"ee7e419d-de23-461a-8c37-10c89b1a5c13\") " pod="openstack-operators/openstack-operator-index-7745w" Oct 14 07:16:49 crc kubenswrapper[4870]: I1014 07:16:49.543004 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gxtxm\" (UniqueName: \"kubernetes.io/projected/ee7e419d-de23-461a-8c37-10c89b1a5c13-kube-api-access-gxtxm\") pod \"openstack-operator-index-7745w\" (UID: \"ee7e419d-de23-461a-8c37-10c89b1a5c13\") " pod="openstack-operators/openstack-operator-index-7745w" Oct 14 07:16:49 crc kubenswrapper[4870]: I1014 07:16:49.561623 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gxtxm\" (UniqueName: \"kubernetes.io/projected/ee7e419d-de23-461a-8c37-10c89b1a5c13-kube-api-access-gxtxm\") pod \"openstack-operator-index-7745w\" (UID: \"ee7e419d-de23-461a-8c37-10c89b1a5c13\") " pod="openstack-operators/openstack-operator-index-7745w" Oct 14 07:16:49 crc kubenswrapper[4870]: I1014 07:16:49.588339 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-7745w" Oct 14 07:16:49 crc kubenswrapper[4870]: I1014 07:16:49.844672 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-7745w"] Oct 14 07:16:50 crc kubenswrapper[4870]: I1014 07:16:50.334114 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-7745w" event={"ID":"ee7e419d-de23-461a-8c37-10c89b1a5c13","Type":"ContainerStarted","Data":"69a79d055144675b3b5279adcfceabce051b2f16561fac2f04eb079d50afdeb3"} Oct 14 07:16:51 crc kubenswrapper[4870]: I1014 07:16:51.343956 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-7745w" event={"ID":"ee7e419d-de23-461a-8c37-10c89b1a5c13","Type":"ContainerStarted","Data":"a83a9d457657911d1b660862271f2501857fc55c89887205ab74f68f5f8d98f5"} Oct 14 07:16:51 crc kubenswrapper[4870]: I1014 07:16:51.364341 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-7745w" podStartSLOduration=1.091949343 podStartE2EDuration="2.364318996s" podCreationTimestamp="2025-10-14 07:16:49 +0000 UTC" firstStartedPulling="2025-10-14 07:16:49.850248318 +0000 UTC m=+945.547608689" lastFinishedPulling="2025-10-14 07:16:51.122617941 +0000 UTC m=+946.819978342" observedRunningTime="2025-10-14 07:16:51.360862937 +0000 UTC m=+947.058223358" watchObservedRunningTime="2025-10-14 07:16:51.364318996 +0000 UTC m=+947.061679377" Oct 14 07:16:51 crc kubenswrapper[4870]: I1014 07:16:51.620832 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-7745w"] Oct 14 07:16:52 crc kubenswrapper[4870]: I1014 07:16:52.228885 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-wfxzz"] Oct 14 07:16:52 crc kubenswrapper[4870]: I1014 07:16:52.229755 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-wfxzz" Oct 14 07:16:52 crc kubenswrapper[4870]: I1014 07:16:52.247331 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-wfxzz"] Oct 14 07:16:52 crc kubenswrapper[4870]: I1014 07:16:52.388353 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fp28d\" (UniqueName: \"kubernetes.io/projected/18d94d36-8f20-4399-b74e-29d831b2ffe8-kube-api-access-fp28d\") pod \"openstack-operator-index-wfxzz\" (UID: \"18d94d36-8f20-4399-b74e-29d831b2ffe8\") " pod="openstack-operators/openstack-operator-index-wfxzz" Oct 14 07:16:52 crc kubenswrapper[4870]: I1014 07:16:52.489822 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fp28d\" (UniqueName: \"kubernetes.io/projected/18d94d36-8f20-4399-b74e-29d831b2ffe8-kube-api-access-fp28d\") pod \"openstack-operator-index-wfxzz\" (UID: \"18d94d36-8f20-4399-b74e-29d831b2ffe8\") " pod="openstack-operators/openstack-operator-index-wfxzz" Oct 14 07:16:52 crc kubenswrapper[4870]: I1014 07:16:52.513582 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fp28d\" (UniqueName: \"kubernetes.io/projected/18d94d36-8f20-4399-b74e-29d831b2ffe8-kube-api-access-fp28d\") pod \"openstack-operator-index-wfxzz\" (UID: \"18d94d36-8f20-4399-b74e-29d831b2ffe8\") " pod="openstack-operators/openstack-operator-index-wfxzz" Oct 14 07:16:52 crc kubenswrapper[4870]: I1014 07:16:52.545612 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-wfxzz" Oct 14 07:16:52 crc kubenswrapper[4870]: I1014 07:16:52.977125 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-wfxzz"] Oct 14 07:16:53 crc kubenswrapper[4870]: I1014 07:16:53.358093 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-7745w" podUID="ee7e419d-de23-461a-8c37-10c89b1a5c13" containerName="registry-server" containerID="cri-o://a83a9d457657911d1b660862271f2501857fc55c89887205ab74f68f5f8d98f5" gracePeriod=2 Oct 14 07:16:53 crc kubenswrapper[4870]: I1014 07:16:53.358314 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-wfxzz" event={"ID":"18d94d36-8f20-4399-b74e-29d831b2ffe8","Type":"ContainerStarted","Data":"2b5bc432b81c2f7d97b1f183075aa8ec995501984532598ddb86d0ce219c2b2c"} Oct 14 07:16:53 crc kubenswrapper[4870]: I1014 07:16:53.721647 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-7745w" Oct 14 07:16:53 crc kubenswrapper[4870]: I1014 07:16:53.912377 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gxtxm\" (UniqueName: \"kubernetes.io/projected/ee7e419d-de23-461a-8c37-10c89b1a5c13-kube-api-access-gxtxm\") pod \"ee7e419d-de23-461a-8c37-10c89b1a5c13\" (UID: \"ee7e419d-de23-461a-8c37-10c89b1a5c13\") " Oct 14 07:16:53 crc kubenswrapper[4870]: I1014 07:16:53.918594 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ee7e419d-de23-461a-8c37-10c89b1a5c13-kube-api-access-gxtxm" (OuterVolumeSpecName: "kube-api-access-gxtxm") pod "ee7e419d-de23-461a-8c37-10c89b1a5c13" (UID: "ee7e419d-de23-461a-8c37-10c89b1a5c13"). InnerVolumeSpecName "kube-api-access-gxtxm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:16:54 crc kubenswrapper[4870]: I1014 07:16:54.014345 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gxtxm\" (UniqueName: \"kubernetes.io/projected/ee7e419d-de23-461a-8c37-10c89b1a5c13-kube-api-access-gxtxm\") on node \"crc\" DevicePath \"\"" Oct 14 07:16:54 crc kubenswrapper[4870]: I1014 07:16:54.369058 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-wfxzz" event={"ID":"18d94d36-8f20-4399-b74e-29d831b2ffe8","Type":"ContainerStarted","Data":"ec5c9ccae82d375ac1e0b670dd79d93b211d21194c26f02e0895eb806dd46064"} Oct 14 07:16:54 crc kubenswrapper[4870]: I1014 07:16:54.372050 4870 generic.go:334] "Generic (PLEG): container finished" podID="ee7e419d-de23-461a-8c37-10c89b1a5c13" containerID="a83a9d457657911d1b660862271f2501857fc55c89887205ab74f68f5f8d98f5" exitCode=0 Oct 14 07:16:54 crc kubenswrapper[4870]: I1014 07:16:54.372103 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-7745w" event={"ID":"ee7e419d-de23-461a-8c37-10c89b1a5c13","Type":"ContainerDied","Data":"a83a9d457657911d1b660862271f2501857fc55c89887205ab74f68f5f8d98f5"} Oct 14 07:16:54 crc kubenswrapper[4870]: I1014 07:16:54.372136 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-7745w" event={"ID":"ee7e419d-de23-461a-8c37-10c89b1a5c13","Type":"ContainerDied","Data":"69a79d055144675b3b5279adcfceabce051b2f16561fac2f04eb079d50afdeb3"} Oct 14 07:16:54 crc kubenswrapper[4870]: I1014 07:16:54.372156 4870 scope.go:117] "RemoveContainer" containerID="a83a9d457657911d1b660862271f2501857fc55c89887205ab74f68f5f8d98f5" Oct 14 07:16:54 crc kubenswrapper[4870]: I1014 07:16:54.372224 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-7745w" Oct 14 07:16:54 crc kubenswrapper[4870]: I1014 07:16:54.389024 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-wfxzz" podStartSLOduration=1.939513527 podStartE2EDuration="2.389001402s" podCreationTimestamp="2025-10-14 07:16:52 +0000 UTC" firstStartedPulling="2025-10-14 07:16:52.985327932 +0000 UTC m=+948.682688343" lastFinishedPulling="2025-10-14 07:16:53.434815827 +0000 UTC m=+949.132176218" observedRunningTime="2025-10-14 07:16:54.387951635 +0000 UTC m=+950.085312006" watchObservedRunningTime="2025-10-14 07:16:54.389001402 +0000 UTC m=+950.086361773" Oct 14 07:16:54 crc kubenswrapper[4870]: I1014 07:16:54.408831 4870 scope.go:117] "RemoveContainer" containerID="a83a9d457657911d1b660862271f2501857fc55c89887205ab74f68f5f8d98f5" Oct 14 07:16:54 crc kubenswrapper[4870]: E1014 07:16:54.414653 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a83a9d457657911d1b660862271f2501857fc55c89887205ab74f68f5f8d98f5\": container with ID starting with a83a9d457657911d1b660862271f2501857fc55c89887205ab74f68f5f8d98f5 not found: ID does not exist" containerID="a83a9d457657911d1b660862271f2501857fc55c89887205ab74f68f5f8d98f5" Oct 14 07:16:54 crc kubenswrapper[4870]: I1014 07:16:54.414710 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a83a9d457657911d1b660862271f2501857fc55c89887205ab74f68f5f8d98f5"} err="failed to get container status \"a83a9d457657911d1b660862271f2501857fc55c89887205ab74f68f5f8d98f5\": rpc error: code = NotFound desc = could not find container \"a83a9d457657911d1b660862271f2501857fc55c89887205ab74f68f5f8d98f5\": container with ID starting with a83a9d457657911d1b660862271f2501857fc55c89887205ab74f68f5f8d98f5 not found: ID does not exist" Oct 14 07:16:54 crc kubenswrapper[4870]: I1014 07:16:54.419652 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-7745w"] Oct 14 07:16:54 crc kubenswrapper[4870]: I1014 07:16:54.422167 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-index-7745w"] Oct 14 07:16:55 crc kubenswrapper[4870]: I1014 07:16:55.044007 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ee7e419d-de23-461a-8c37-10c89b1a5c13" path="/var/lib/kubelet/pods/ee7e419d-de23-461a-8c37-10c89b1a5c13/volumes" Oct 14 07:16:58 crc kubenswrapper[4870]: E1014 07:16:58.885152 4870 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podebb7cfd0_0388_4aca_99f6_7669c6bc8d4e.slice\": RecentStats: unable to find data in memory cache]" Oct 14 07:17:02 crc kubenswrapper[4870]: I1014 07:17:02.547934 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-wfxzz" Oct 14 07:17:02 crc kubenswrapper[4870]: I1014 07:17:02.548004 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-wfxzz" Oct 14 07:17:02 crc kubenswrapper[4870]: I1014 07:17:02.637284 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-wfxzz" Oct 14 07:17:03 crc kubenswrapper[4870]: I1014 07:17:03.476858 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-wfxzz" Oct 14 07:17:09 crc kubenswrapper[4870]: I1014 07:17:09.283374 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/32da80840a2017f27ed4ad61f02adc64a25aa18e8dad0409953372036al2fnh"] Oct 14 07:17:09 crc kubenswrapper[4870]: E1014 07:17:09.284960 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee7e419d-de23-461a-8c37-10c89b1a5c13" containerName="registry-server" Oct 14 07:17:09 crc kubenswrapper[4870]: I1014 07:17:09.284985 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee7e419d-de23-461a-8c37-10c89b1a5c13" containerName="registry-server" Oct 14 07:17:09 crc kubenswrapper[4870]: I1014 07:17:09.285168 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee7e419d-de23-461a-8c37-10c89b1a5c13" containerName="registry-server" Oct 14 07:17:09 crc kubenswrapper[4870]: I1014 07:17:09.286649 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/32da80840a2017f27ed4ad61f02adc64a25aa18e8dad0409953372036al2fnh" Oct 14 07:17:09 crc kubenswrapper[4870]: I1014 07:17:09.293619 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-v2pcx" Oct 14 07:17:09 crc kubenswrapper[4870]: I1014 07:17:09.296392 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/32da80840a2017f27ed4ad61f02adc64a25aa18e8dad0409953372036al2fnh"] Oct 14 07:17:09 crc kubenswrapper[4870]: I1014 07:17:09.466147 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g8gm4\" (UniqueName: \"kubernetes.io/projected/36e53446-018c-48eb-8398-f28a109c9cb0-kube-api-access-g8gm4\") pod \"32da80840a2017f27ed4ad61f02adc64a25aa18e8dad0409953372036al2fnh\" (UID: \"36e53446-018c-48eb-8398-f28a109c9cb0\") " pod="openstack-operators/32da80840a2017f27ed4ad61f02adc64a25aa18e8dad0409953372036al2fnh" Oct 14 07:17:09 crc kubenswrapper[4870]: I1014 07:17:09.466233 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/36e53446-018c-48eb-8398-f28a109c9cb0-util\") pod \"32da80840a2017f27ed4ad61f02adc64a25aa18e8dad0409953372036al2fnh\" (UID: \"36e53446-018c-48eb-8398-f28a109c9cb0\") " pod="openstack-operators/32da80840a2017f27ed4ad61f02adc64a25aa18e8dad0409953372036al2fnh" Oct 14 07:17:09 crc kubenswrapper[4870]: I1014 07:17:09.466268 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/36e53446-018c-48eb-8398-f28a109c9cb0-bundle\") pod \"32da80840a2017f27ed4ad61f02adc64a25aa18e8dad0409953372036al2fnh\" (UID: \"36e53446-018c-48eb-8398-f28a109c9cb0\") " pod="openstack-operators/32da80840a2017f27ed4ad61f02adc64a25aa18e8dad0409953372036al2fnh" Oct 14 07:17:09 crc kubenswrapper[4870]: I1014 07:17:09.567136 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/36e53446-018c-48eb-8398-f28a109c9cb0-util\") pod \"32da80840a2017f27ed4ad61f02adc64a25aa18e8dad0409953372036al2fnh\" (UID: \"36e53446-018c-48eb-8398-f28a109c9cb0\") " pod="openstack-operators/32da80840a2017f27ed4ad61f02adc64a25aa18e8dad0409953372036al2fnh" Oct 14 07:17:09 crc kubenswrapper[4870]: I1014 07:17:09.567205 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/36e53446-018c-48eb-8398-f28a109c9cb0-bundle\") pod \"32da80840a2017f27ed4ad61f02adc64a25aa18e8dad0409953372036al2fnh\" (UID: \"36e53446-018c-48eb-8398-f28a109c9cb0\") " pod="openstack-operators/32da80840a2017f27ed4ad61f02adc64a25aa18e8dad0409953372036al2fnh" Oct 14 07:17:09 crc kubenswrapper[4870]: I1014 07:17:09.567360 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g8gm4\" (UniqueName: \"kubernetes.io/projected/36e53446-018c-48eb-8398-f28a109c9cb0-kube-api-access-g8gm4\") pod \"32da80840a2017f27ed4ad61f02adc64a25aa18e8dad0409953372036al2fnh\" (UID: \"36e53446-018c-48eb-8398-f28a109c9cb0\") " pod="openstack-operators/32da80840a2017f27ed4ad61f02adc64a25aa18e8dad0409953372036al2fnh" Oct 14 07:17:09 crc kubenswrapper[4870]: I1014 07:17:09.568031 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/36e53446-018c-48eb-8398-f28a109c9cb0-bundle\") pod \"32da80840a2017f27ed4ad61f02adc64a25aa18e8dad0409953372036al2fnh\" (UID: \"36e53446-018c-48eb-8398-f28a109c9cb0\") " pod="openstack-operators/32da80840a2017f27ed4ad61f02adc64a25aa18e8dad0409953372036al2fnh" Oct 14 07:17:09 crc kubenswrapper[4870]: I1014 07:17:09.568079 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/36e53446-018c-48eb-8398-f28a109c9cb0-util\") pod \"32da80840a2017f27ed4ad61f02adc64a25aa18e8dad0409953372036al2fnh\" (UID: \"36e53446-018c-48eb-8398-f28a109c9cb0\") " pod="openstack-operators/32da80840a2017f27ed4ad61f02adc64a25aa18e8dad0409953372036al2fnh" Oct 14 07:17:09 crc kubenswrapper[4870]: I1014 07:17:09.593554 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g8gm4\" (UniqueName: \"kubernetes.io/projected/36e53446-018c-48eb-8398-f28a109c9cb0-kube-api-access-g8gm4\") pod \"32da80840a2017f27ed4ad61f02adc64a25aa18e8dad0409953372036al2fnh\" (UID: \"36e53446-018c-48eb-8398-f28a109c9cb0\") " pod="openstack-operators/32da80840a2017f27ed4ad61f02adc64a25aa18e8dad0409953372036al2fnh" Oct 14 07:17:09 crc kubenswrapper[4870]: I1014 07:17:09.616367 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/32da80840a2017f27ed4ad61f02adc64a25aa18e8dad0409953372036al2fnh" Oct 14 07:17:10 crc kubenswrapper[4870]: I1014 07:17:10.087716 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/32da80840a2017f27ed4ad61f02adc64a25aa18e8dad0409953372036al2fnh"] Oct 14 07:17:10 crc kubenswrapper[4870]: W1014 07:17:10.095066 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod36e53446_018c_48eb_8398_f28a109c9cb0.slice/crio-e73ce35282bcb60e0ffe6eea64fd82aa7d68973a43c7000dde9e91aafbbb466a WatchSource:0}: Error finding container e73ce35282bcb60e0ffe6eea64fd82aa7d68973a43c7000dde9e91aafbbb466a: Status 404 returned error can't find the container with id e73ce35282bcb60e0ffe6eea64fd82aa7d68973a43c7000dde9e91aafbbb466a Oct 14 07:17:10 crc kubenswrapper[4870]: I1014 07:17:10.485162 4870 generic.go:334] "Generic (PLEG): container finished" podID="36e53446-018c-48eb-8398-f28a109c9cb0" containerID="8a93ad9ddad60563b438b21ce4655cd9e08e4b6f388e0112fae59c88bd1eb7fb" exitCode=0 Oct 14 07:17:10 crc kubenswrapper[4870]: I1014 07:17:10.485249 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/32da80840a2017f27ed4ad61f02adc64a25aa18e8dad0409953372036al2fnh" event={"ID":"36e53446-018c-48eb-8398-f28a109c9cb0","Type":"ContainerDied","Data":"8a93ad9ddad60563b438b21ce4655cd9e08e4b6f388e0112fae59c88bd1eb7fb"} Oct 14 07:17:10 crc kubenswrapper[4870]: I1014 07:17:10.485641 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/32da80840a2017f27ed4ad61f02adc64a25aa18e8dad0409953372036al2fnh" event={"ID":"36e53446-018c-48eb-8398-f28a109c9cb0","Type":"ContainerStarted","Data":"e73ce35282bcb60e0ffe6eea64fd82aa7d68973a43c7000dde9e91aafbbb466a"} Oct 14 07:17:11 crc kubenswrapper[4870]: I1014 07:17:11.493407 4870 generic.go:334] "Generic (PLEG): container finished" podID="36e53446-018c-48eb-8398-f28a109c9cb0" containerID="2c96f47145a6a48608fb7c3c164efafe91c83feb36d9fce82a88b2628ef52dbe" exitCode=0 Oct 14 07:17:11 crc kubenswrapper[4870]: I1014 07:17:11.493468 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/32da80840a2017f27ed4ad61f02adc64a25aa18e8dad0409953372036al2fnh" event={"ID":"36e53446-018c-48eb-8398-f28a109c9cb0","Type":"ContainerDied","Data":"2c96f47145a6a48608fb7c3c164efafe91c83feb36d9fce82a88b2628ef52dbe"} Oct 14 07:17:12 crc kubenswrapper[4870]: I1014 07:17:12.506589 4870 generic.go:334] "Generic (PLEG): container finished" podID="36e53446-018c-48eb-8398-f28a109c9cb0" containerID="6b8a037adc28995b8f73914a49d20558cf23c88881727ee3d609d0e88568c5a5" exitCode=0 Oct 14 07:17:12 crc kubenswrapper[4870]: I1014 07:17:12.506693 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/32da80840a2017f27ed4ad61f02adc64a25aa18e8dad0409953372036al2fnh" event={"ID":"36e53446-018c-48eb-8398-f28a109c9cb0","Type":"ContainerDied","Data":"6b8a037adc28995b8f73914a49d20558cf23c88881727ee3d609d0e88568c5a5"} Oct 14 07:17:13 crc kubenswrapper[4870]: I1014 07:17:13.895630 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/32da80840a2017f27ed4ad61f02adc64a25aa18e8dad0409953372036al2fnh" Oct 14 07:17:13 crc kubenswrapper[4870]: I1014 07:17:13.935034 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/36e53446-018c-48eb-8398-f28a109c9cb0-bundle\") pod \"36e53446-018c-48eb-8398-f28a109c9cb0\" (UID: \"36e53446-018c-48eb-8398-f28a109c9cb0\") " Oct 14 07:17:13 crc kubenswrapper[4870]: I1014 07:17:13.935203 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/36e53446-018c-48eb-8398-f28a109c9cb0-util\") pod \"36e53446-018c-48eb-8398-f28a109c9cb0\" (UID: \"36e53446-018c-48eb-8398-f28a109c9cb0\") " Oct 14 07:17:13 crc kubenswrapper[4870]: I1014 07:17:13.935296 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g8gm4\" (UniqueName: \"kubernetes.io/projected/36e53446-018c-48eb-8398-f28a109c9cb0-kube-api-access-g8gm4\") pod \"36e53446-018c-48eb-8398-f28a109c9cb0\" (UID: \"36e53446-018c-48eb-8398-f28a109c9cb0\") " Oct 14 07:17:13 crc kubenswrapper[4870]: I1014 07:17:13.936068 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/36e53446-018c-48eb-8398-f28a109c9cb0-bundle" (OuterVolumeSpecName: "bundle") pod "36e53446-018c-48eb-8398-f28a109c9cb0" (UID: "36e53446-018c-48eb-8398-f28a109c9cb0"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:17:13 crc kubenswrapper[4870]: I1014 07:17:13.942125 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/36e53446-018c-48eb-8398-f28a109c9cb0-kube-api-access-g8gm4" (OuterVolumeSpecName: "kube-api-access-g8gm4") pod "36e53446-018c-48eb-8398-f28a109c9cb0" (UID: "36e53446-018c-48eb-8398-f28a109c9cb0"). InnerVolumeSpecName "kube-api-access-g8gm4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:17:13 crc kubenswrapper[4870]: I1014 07:17:13.968615 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/36e53446-018c-48eb-8398-f28a109c9cb0-util" (OuterVolumeSpecName: "util") pod "36e53446-018c-48eb-8398-f28a109c9cb0" (UID: "36e53446-018c-48eb-8398-f28a109c9cb0"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:17:14 crc kubenswrapper[4870]: I1014 07:17:14.036869 4870 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/36e53446-018c-48eb-8398-f28a109c9cb0-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 07:17:14 crc kubenswrapper[4870]: I1014 07:17:14.036907 4870 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/36e53446-018c-48eb-8398-f28a109c9cb0-util\") on node \"crc\" DevicePath \"\"" Oct 14 07:17:14 crc kubenswrapper[4870]: I1014 07:17:14.036920 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g8gm4\" (UniqueName: \"kubernetes.io/projected/36e53446-018c-48eb-8398-f28a109c9cb0-kube-api-access-g8gm4\") on node \"crc\" DevicePath \"\"" Oct 14 07:17:14 crc kubenswrapper[4870]: I1014 07:17:14.530141 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/32da80840a2017f27ed4ad61f02adc64a25aa18e8dad0409953372036al2fnh" event={"ID":"36e53446-018c-48eb-8398-f28a109c9cb0","Type":"ContainerDied","Data":"e73ce35282bcb60e0ffe6eea64fd82aa7d68973a43c7000dde9e91aafbbb466a"} Oct 14 07:17:14 crc kubenswrapper[4870]: I1014 07:17:14.530247 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/32da80840a2017f27ed4ad61f02adc64a25aa18e8dad0409953372036al2fnh" Oct 14 07:17:14 crc kubenswrapper[4870]: I1014 07:17:14.530295 4870 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e73ce35282bcb60e0ffe6eea64fd82aa7d68973a43c7000dde9e91aafbbb466a" Oct 14 07:17:22 crc kubenswrapper[4870]: I1014 07:17:22.164053 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-operator-64895cd698-r8hz4"] Oct 14 07:17:22 crc kubenswrapper[4870]: E1014 07:17:22.164975 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="36e53446-018c-48eb-8398-f28a109c9cb0" containerName="pull" Oct 14 07:17:22 crc kubenswrapper[4870]: I1014 07:17:22.164990 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="36e53446-018c-48eb-8398-f28a109c9cb0" containerName="pull" Oct 14 07:17:22 crc kubenswrapper[4870]: E1014 07:17:22.165007 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="36e53446-018c-48eb-8398-f28a109c9cb0" containerName="util" Oct 14 07:17:22 crc kubenswrapper[4870]: I1014 07:17:22.165014 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="36e53446-018c-48eb-8398-f28a109c9cb0" containerName="util" Oct 14 07:17:22 crc kubenswrapper[4870]: E1014 07:17:22.165034 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="36e53446-018c-48eb-8398-f28a109c9cb0" containerName="extract" Oct 14 07:17:22 crc kubenswrapper[4870]: I1014 07:17:22.165045 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="36e53446-018c-48eb-8398-f28a109c9cb0" containerName="extract" Oct 14 07:17:22 crc kubenswrapper[4870]: I1014 07:17:22.165180 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="36e53446-018c-48eb-8398-f28a109c9cb0" containerName="extract" Oct 14 07:17:22 crc kubenswrapper[4870]: I1014 07:17:22.165932 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-64895cd698-r8hz4" Oct 14 07:17:22 crc kubenswrapper[4870]: I1014 07:17:22.169186 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-operator-dockercfg-m67ck" Oct 14 07:17:22 crc kubenswrapper[4870]: I1014 07:17:22.187092 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-64895cd698-r8hz4"] Oct 14 07:17:22 crc kubenswrapper[4870]: I1014 07:17:22.243236 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w7w2r\" (UniqueName: \"kubernetes.io/projected/f7122ea8-dbd9-4f40-9eca-401b4a189140-kube-api-access-w7w2r\") pod \"openstack-operator-controller-operator-64895cd698-r8hz4\" (UID: \"f7122ea8-dbd9-4f40-9eca-401b4a189140\") " pod="openstack-operators/openstack-operator-controller-operator-64895cd698-r8hz4" Oct 14 07:17:22 crc kubenswrapper[4870]: I1014 07:17:22.344558 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w7w2r\" (UniqueName: \"kubernetes.io/projected/f7122ea8-dbd9-4f40-9eca-401b4a189140-kube-api-access-w7w2r\") pod \"openstack-operator-controller-operator-64895cd698-r8hz4\" (UID: \"f7122ea8-dbd9-4f40-9eca-401b4a189140\") " pod="openstack-operators/openstack-operator-controller-operator-64895cd698-r8hz4" Oct 14 07:17:22 crc kubenswrapper[4870]: I1014 07:17:22.365111 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w7w2r\" (UniqueName: \"kubernetes.io/projected/f7122ea8-dbd9-4f40-9eca-401b4a189140-kube-api-access-w7w2r\") pod \"openstack-operator-controller-operator-64895cd698-r8hz4\" (UID: \"f7122ea8-dbd9-4f40-9eca-401b4a189140\") " pod="openstack-operators/openstack-operator-controller-operator-64895cd698-r8hz4" Oct 14 07:17:22 crc kubenswrapper[4870]: I1014 07:17:22.484473 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-64895cd698-r8hz4" Oct 14 07:17:22 crc kubenswrapper[4870]: W1014 07:17:22.995629 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf7122ea8_dbd9_4f40_9eca_401b4a189140.slice/crio-252d1ea084b6ca1b6ea3b2bf1c11cb00a2283e724d9f805b105ea1aedb506c94 WatchSource:0}: Error finding container 252d1ea084b6ca1b6ea3b2bf1c11cb00a2283e724d9f805b105ea1aedb506c94: Status 404 returned error can't find the container with id 252d1ea084b6ca1b6ea3b2bf1c11cb00a2283e724d9f805b105ea1aedb506c94 Oct 14 07:17:22 crc kubenswrapper[4870]: I1014 07:17:22.996771 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-64895cd698-r8hz4"] Oct 14 07:17:23 crc kubenswrapper[4870]: I1014 07:17:23.608808 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-64895cd698-r8hz4" event={"ID":"f7122ea8-dbd9-4f40-9eca-401b4a189140","Type":"ContainerStarted","Data":"252d1ea084b6ca1b6ea3b2bf1c11cb00a2283e724d9f805b105ea1aedb506c94"} Oct 14 07:17:27 crc kubenswrapper[4870]: I1014 07:17:27.634022 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-64895cd698-r8hz4" event={"ID":"f7122ea8-dbd9-4f40-9eca-401b4a189140","Type":"ContainerStarted","Data":"79bc81fcd95217e8089538037eca59eb0f79b920cbcfd49d0e4e0e87ad4e10e3"} Oct 14 07:17:30 crc kubenswrapper[4870]: I1014 07:17:30.655490 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-64895cd698-r8hz4" event={"ID":"f7122ea8-dbd9-4f40-9eca-401b4a189140","Type":"ContainerStarted","Data":"3334195f5b503ec31f3cb239f0ae9c33c48a01de21ca6fd211b249b03eef0320"} Oct 14 07:17:30 crc kubenswrapper[4870]: I1014 07:17:30.655981 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-64895cd698-r8hz4" Oct 14 07:17:32 crc kubenswrapper[4870]: I1014 07:17:32.488795 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-64895cd698-r8hz4" Oct 14 07:17:32 crc kubenswrapper[4870]: I1014 07:17:32.551503 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-operator-64895cd698-r8hz4" podStartSLOduration=3.980505317 podStartE2EDuration="10.551476324s" podCreationTimestamp="2025-10-14 07:17:22 +0000 UTC" firstStartedPulling="2025-10-14 07:17:22.997373194 +0000 UTC m=+978.694733565" lastFinishedPulling="2025-10-14 07:17:29.568344201 +0000 UTC m=+985.265704572" observedRunningTime="2025-10-14 07:17:30.706918921 +0000 UTC m=+986.404279332" watchObservedRunningTime="2025-10-14 07:17:32.551476324 +0000 UTC m=+988.248836735" Oct 14 07:17:48 crc kubenswrapper[4870]: I1014 07:17:48.886159 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-658bdf4b74-76kw6"] Oct 14 07:17:48 crc kubenswrapper[4870]: I1014 07:17:48.887494 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-658bdf4b74-76kw6" Oct 14 07:17:48 crc kubenswrapper[4870]: I1014 07:17:48.890504 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-j6czk" Oct 14 07:17:48 crc kubenswrapper[4870]: I1014 07:17:48.897984 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-658bdf4b74-76kw6"] Oct 14 07:17:48 crc kubenswrapper[4870]: I1014 07:17:48.909977 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-7b7fb68549-xnk22"] Oct 14 07:17:48 crc kubenswrapper[4870]: I1014 07:17:48.910982 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-7b7fb68549-xnk22" Oct 14 07:17:48 crc kubenswrapper[4870]: I1014 07:17:48.913639 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-g4cjj" Oct 14 07:17:48 crc kubenswrapper[4870]: I1014 07:17:48.926587 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-7b7fb68549-xnk22"] Oct 14 07:17:48 crc kubenswrapper[4870]: I1014 07:17:48.932128 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-85d5d9dd78-ld7xh"] Oct 14 07:17:48 crc kubenswrapper[4870]: I1014 07:17:48.933258 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-85d5d9dd78-ld7xh" Oct 14 07:17:48 crc kubenswrapper[4870]: I1014 07:17:48.939649 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-wmxh9" Oct 14 07:17:48 crc kubenswrapper[4870]: I1014 07:17:48.968506 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-85d5d9dd78-ld7xh"] Oct 14 07:17:48 crc kubenswrapper[4870]: I1014 07:17:48.986082 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-84b9b84486-cl7ft"] Oct 14 07:17:48 crc kubenswrapper[4870]: I1014 07:17:48.987076 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-84b9b84486-cl7ft" Oct 14 07:17:48 crc kubenswrapper[4870]: I1014 07:17:48.993298 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-84b9b84486-cl7ft"] Oct 14 07:17:48 crc kubenswrapper[4870]: I1014 07:17:48.998047 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-tr4hd" Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.002934 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-858f76bbdd-9bdj8"] Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.007432 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-858f76bbdd-9bdj8" Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.010377 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-k2vdn" Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.018118 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-858f76bbdd-9bdj8"] Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.031389 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-7ffbcb7588-g2fdg"] Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.032508 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-7ffbcb7588-g2fdg" Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.034704 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-v7zh5" Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.054251 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c9tdn\" (UniqueName: \"kubernetes.io/projected/9f311a89-2483-42d6-9fe5-57b1900624d2-kube-api-access-c9tdn\") pod \"cinder-operator-controller-manager-7b7fb68549-xnk22\" (UID: \"9f311a89-2483-42d6-9fe5-57b1900624d2\") " pod="openstack-operators/cinder-operator-controller-manager-7b7fb68549-xnk22" Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.055702 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tbk5q\" (UniqueName: \"kubernetes.io/projected/36e6bbc0-f57d-4486-9120-d2cd5a5ce638-kube-api-access-tbk5q\") pod \"barbican-operator-controller-manager-658bdf4b74-76kw6\" (UID: \"36e6bbc0-f57d-4486-9120-d2cd5a5ce638\") " pod="openstack-operators/barbican-operator-controller-manager-658bdf4b74-76kw6" Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.055822 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qf9r7\" (UniqueName: \"kubernetes.io/projected/634060b4-d317-415d-bc9e-aa3ed6503b4f-kube-api-access-qf9r7\") pod \"designate-operator-controller-manager-85d5d9dd78-ld7xh\" (UID: \"634060b4-d317-415d-bc9e-aa3ed6503b4f\") " pod="openstack-operators/designate-operator-controller-manager-85d5d9dd78-ld7xh" Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.057863 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-656bcbd775-lp9f2"] Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.059229 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-656bcbd775-lp9f2" Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.063607 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.063879 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-vm25q" Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.067515 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-9c5c78d49-gjsj6"] Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.068855 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-9c5c78d49-gjsj6" Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.073965 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-2v5ql" Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.092365 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-656bcbd775-lp9f2"] Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.107982 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-7ffbcb7588-g2fdg"] Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.115563 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-9c5c78d49-gjsj6"] Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.129032 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-5f67fbc655-szjkl"] Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.129949 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-5f67fbc655-szjkl" Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.134270 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-65vxp" Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.135364 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-55b6b7c7b8-jbt9q"] Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.136416 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-55b6b7c7b8-jbt9q" Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.142768 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-fvtrr" Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.156594 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-f9fb45f8f-kvmkj"] Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.157998 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-f9fb45f8f-kvmkj" Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.158788 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-62xb7\" (UniqueName: \"kubernetes.io/projected/8ed9c494-ed5c-46d0-9434-1edbfcfba361-kube-api-access-62xb7\") pod \"horizon-operator-controller-manager-7ffbcb7588-g2fdg\" (UID: \"8ed9c494-ed5c-46d0-9434-1edbfcfba361\") " pod="openstack-operators/horizon-operator-controller-manager-7ffbcb7588-g2fdg" Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.158869 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qf9r7\" (UniqueName: \"kubernetes.io/projected/634060b4-d317-415d-bc9e-aa3ed6503b4f-kube-api-access-qf9r7\") pod \"designate-operator-controller-manager-85d5d9dd78-ld7xh\" (UID: \"634060b4-d317-415d-bc9e-aa3ed6503b4f\") " pod="openstack-operators/designate-operator-controller-manager-85d5d9dd78-ld7xh" Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.159788 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r7b5s\" (UniqueName: \"kubernetes.io/projected/519142a1-852b-4410-a511-13e2a36aed77-kube-api-access-r7b5s\") pod \"heat-operator-controller-manager-858f76bbdd-9bdj8\" (UID: \"519142a1-852b-4410-a511-13e2a36aed77\") " pod="openstack-operators/heat-operator-controller-manager-858f76bbdd-9bdj8" Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.159861 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x9xg6\" (UniqueName: \"kubernetes.io/projected/b4a4fd5f-360a-45ca-9684-74fe1c7631a6-kube-api-access-x9xg6\") pod \"glance-operator-controller-manager-84b9b84486-cl7ft\" (UID: \"b4a4fd5f-360a-45ca-9684-74fe1c7631a6\") " pod="openstack-operators/glance-operator-controller-manager-84b9b84486-cl7ft" Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.159897 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c9tdn\" (UniqueName: \"kubernetes.io/projected/9f311a89-2483-42d6-9fe5-57b1900624d2-kube-api-access-c9tdn\") pod \"cinder-operator-controller-manager-7b7fb68549-xnk22\" (UID: \"9f311a89-2483-42d6-9fe5-57b1900624d2\") " pod="openstack-operators/cinder-operator-controller-manager-7b7fb68549-xnk22" Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.159933 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tbk5q\" (UniqueName: \"kubernetes.io/projected/36e6bbc0-f57d-4486-9120-d2cd5a5ce638-kube-api-access-tbk5q\") pod \"barbican-operator-controller-manager-658bdf4b74-76kw6\" (UID: \"36e6bbc0-f57d-4486-9120-d2cd5a5ce638\") " pod="openstack-operators/barbican-operator-controller-manager-658bdf4b74-76kw6" Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.164953 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-pkrqz" Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.172292 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-55b6b7c7b8-jbt9q"] Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.184695 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-5f67fbc655-szjkl"] Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.204772 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qf9r7\" (UniqueName: \"kubernetes.io/projected/634060b4-d317-415d-bc9e-aa3ed6503b4f-kube-api-access-qf9r7\") pod \"designate-operator-controller-manager-85d5d9dd78-ld7xh\" (UID: \"634060b4-d317-415d-bc9e-aa3ed6503b4f\") " pod="openstack-operators/designate-operator-controller-manager-85d5d9dd78-ld7xh" Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.205145 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tbk5q\" (UniqueName: \"kubernetes.io/projected/36e6bbc0-f57d-4486-9120-d2cd5a5ce638-kube-api-access-tbk5q\") pod \"barbican-operator-controller-manager-658bdf4b74-76kw6\" (UID: \"36e6bbc0-f57d-4486-9120-d2cd5a5ce638\") " pod="openstack-operators/barbican-operator-controller-manager-658bdf4b74-76kw6" Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.207179 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-658bdf4b74-76kw6" Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.219152 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c9tdn\" (UniqueName: \"kubernetes.io/projected/9f311a89-2483-42d6-9fe5-57b1900624d2-kube-api-access-c9tdn\") pod \"cinder-operator-controller-manager-7b7fb68549-xnk22\" (UID: \"9f311a89-2483-42d6-9fe5-57b1900624d2\") " pod="openstack-operators/cinder-operator-controller-manager-7b7fb68549-xnk22" Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.231863 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-7b7fb68549-xnk22" Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.232685 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-79d585cb66-67fkk"] Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.239081 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-79d585cb66-67fkk" Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.245992 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-fvgj7" Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.256175 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-85d5d9dd78-ld7xh" Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.264576 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-f9fb45f8f-kvmkj"] Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.272124 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-79d585cb66-67fkk"] Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.273163 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hzm7d\" (UniqueName: \"kubernetes.io/projected/23eadd14-852e-4683-bae5-3720a36d9407-kube-api-access-hzm7d\") pod \"mariadb-operator-controller-manager-f9fb45f8f-kvmkj\" (UID: \"23eadd14-852e-4683-bae5-3720a36d9407\") " pod="openstack-operators/mariadb-operator-controller-manager-f9fb45f8f-kvmkj" Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.273234 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mrgx6\" (UniqueName: \"kubernetes.io/projected/4bc1fceb-2d98-415e-b34e-82c2d6e1430f-kube-api-access-mrgx6\") pod \"manila-operator-controller-manager-5f67fbc655-szjkl\" (UID: \"4bc1fceb-2d98-415e-b34e-82c2d6e1430f\") " pod="openstack-operators/manila-operator-controller-manager-5f67fbc655-szjkl" Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.273383 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x9xg6\" (UniqueName: \"kubernetes.io/projected/b4a4fd5f-360a-45ca-9684-74fe1c7631a6-kube-api-access-x9xg6\") pod \"glance-operator-controller-manager-84b9b84486-cl7ft\" (UID: \"b4a4fd5f-360a-45ca-9684-74fe1c7631a6\") " pod="openstack-operators/glance-operator-controller-manager-84b9b84486-cl7ft" Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.273751 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9vxzk\" (UniqueName: \"kubernetes.io/projected/d4236126-b907-4f39-a491-5006f2a9c301-kube-api-access-9vxzk\") pod \"infra-operator-controller-manager-656bcbd775-lp9f2\" (UID: \"d4236126-b907-4f39-a491-5006f2a9c301\") " pod="openstack-operators/infra-operator-controller-manager-656bcbd775-lp9f2" Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.273902 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-klrlz\" (UniqueName: \"kubernetes.io/projected/586d846b-21c2-40f4-8da4-37392d812645-kube-api-access-klrlz\") pod \"neutron-operator-controller-manager-79d585cb66-67fkk\" (UID: \"586d846b-21c2-40f4-8da4-37392d812645\") " pod="openstack-operators/neutron-operator-controller-manager-79d585cb66-67fkk" Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.274019 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-62xb7\" (UniqueName: \"kubernetes.io/projected/8ed9c494-ed5c-46d0-9434-1edbfcfba361-kube-api-access-62xb7\") pod \"horizon-operator-controller-manager-7ffbcb7588-g2fdg\" (UID: \"8ed9c494-ed5c-46d0-9434-1edbfcfba361\") " pod="openstack-operators/horizon-operator-controller-manager-7ffbcb7588-g2fdg" Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.274059 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-25qll\" (UniqueName: \"kubernetes.io/projected/5ad2a6ec-e2aa-45d9-b1a8-59ed908aaacd-kube-api-access-25qll\") pod \"keystone-operator-controller-manager-55b6b7c7b8-jbt9q\" (UID: \"5ad2a6ec-e2aa-45d9-b1a8-59ed908aaacd\") " pod="openstack-operators/keystone-operator-controller-manager-55b6b7c7b8-jbt9q" Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.274105 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xwtw5\" (UniqueName: \"kubernetes.io/projected/d27f8663-102d-4565-b7d5-00db25a8388a-kube-api-access-xwtw5\") pod \"ironic-operator-controller-manager-9c5c78d49-gjsj6\" (UID: \"d27f8663-102d-4565-b7d5-00db25a8388a\") " pod="openstack-operators/ironic-operator-controller-manager-9c5c78d49-gjsj6" Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.274134 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d4236126-b907-4f39-a491-5006f2a9c301-cert\") pod \"infra-operator-controller-manager-656bcbd775-lp9f2\" (UID: \"d4236126-b907-4f39-a491-5006f2a9c301\") " pod="openstack-operators/infra-operator-controller-manager-656bcbd775-lp9f2" Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.274161 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r7b5s\" (UniqueName: \"kubernetes.io/projected/519142a1-852b-4410-a511-13e2a36aed77-kube-api-access-r7b5s\") pod \"heat-operator-controller-manager-858f76bbdd-9bdj8\" (UID: \"519142a1-852b-4410-a511-13e2a36aed77\") " pod="openstack-operators/heat-operator-controller-manager-858f76bbdd-9bdj8" Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.287855 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-5df598886f-zj4rk"] Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.289247 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-5df598886f-zj4rk" Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.291245 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-tbz5x" Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.293844 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-69fdcfc5f5-2schz"] Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.294592 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-62xb7\" (UniqueName: \"kubernetes.io/projected/8ed9c494-ed5c-46d0-9434-1edbfcfba361-kube-api-access-62xb7\") pod \"horizon-operator-controller-manager-7ffbcb7588-g2fdg\" (UID: \"8ed9c494-ed5c-46d0-9434-1edbfcfba361\") " pod="openstack-operators/horizon-operator-controller-manager-7ffbcb7588-g2fdg" Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.294945 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-69fdcfc5f5-2schz" Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.295340 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x9xg6\" (UniqueName: \"kubernetes.io/projected/b4a4fd5f-360a-45ca-9684-74fe1c7631a6-kube-api-access-x9xg6\") pod \"glance-operator-controller-manager-84b9b84486-cl7ft\" (UID: \"b4a4fd5f-360a-45ca-9684-74fe1c7631a6\") " pod="openstack-operators/glance-operator-controller-manager-84b9b84486-cl7ft" Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.296725 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-hdgpq" Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.301143 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-5df598886f-zj4rk"] Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.303567 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r7b5s\" (UniqueName: \"kubernetes.io/projected/519142a1-852b-4410-a511-13e2a36aed77-kube-api-access-r7b5s\") pod \"heat-operator-controller-manager-858f76bbdd-9bdj8\" (UID: \"519142a1-852b-4410-a511-13e2a36aed77\") " pod="openstack-operators/heat-operator-controller-manager-858f76bbdd-9bdj8" Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.311663 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-69fdcfc5f5-2schz"] Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.313812 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-84b9b84486-cl7ft" Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.319566 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-55b7d4484882chz"] Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.322698 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-55b7d4484882chz" Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.323122 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-79df5fb58c-s54lf"] Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.324642 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-79df5fb58c-s54lf" Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.326475 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-4cxkw" Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.326666 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-5g7tz" Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.326690 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.331661 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-68b6c87b68-dtn8j"] Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.333986 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-68b6c87b68-dtn8j" Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.336219 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-79df5fb58c-s54lf"] Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.337907 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-5g84v" Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.339558 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-68b6c87b68-dtn8j"] Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.343019 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-db6d7f97b-22qw9"] Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.344174 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-db6d7f97b-22qw9" Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.345946 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-9rddf" Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.348248 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-db6d7f97b-22qw9"] Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.350516 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-858f76bbdd-9bdj8" Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.352859 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-55b7d4484882chz"] Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.366274 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-67cfc6749b-cblnh"] Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.367525 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-67cfc6749b-cblnh" Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.369648 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-gnlb7" Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.369820 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-67cfc6749b-cblnh"] Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.375431 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-klrlz\" (UniqueName: \"kubernetes.io/projected/586d846b-21c2-40f4-8da4-37392d812645-kube-api-access-klrlz\") pod \"neutron-operator-controller-manager-79d585cb66-67fkk\" (UID: \"586d846b-21c2-40f4-8da4-37392d812645\") " pod="openstack-operators/neutron-operator-controller-manager-79d585cb66-67fkk" Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.375494 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-25qll\" (UniqueName: \"kubernetes.io/projected/5ad2a6ec-e2aa-45d9-b1a8-59ed908aaacd-kube-api-access-25qll\") pod \"keystone-operator-controller-manager-55b6b7c7b8-jbt9q\" (UID: \"5ad2a6ec-e2aa-45d9-b1a8-59ed908aaacd\") " pod="openstack-operators/keystone-operator-controller-manager-55b6b7c7b8-jbt9q" Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.375528 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xwtw5\" (UniqueName: \"kubernetes.io/projected/d27f8663-102d-4565-b7d5-00db25a8388a-kube-api-access-xwtw5\") pod \"ironic-operator-controller-manager-9c5c78d49-gjsj6\" (UID: \"d27f8663-102d-4565-b7d5-00db25a8388a\") " pod="openstack-operators/ironic-operator-controller-manager-9c5c78d49-gjsj6" Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.375551 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d4236126-b907-4f39-a491-5006f2a9c301-cert\") pod \"infra-operator-controller-manager-656bcbd775-lp9f2\" (UID: \"d4236126-b907-4f39-a491-5006f2a9c301\") " pod="openstack-operators/infra-operator-controller-manager-656bcbd775-lp9f2" Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.375584 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hzm7d\" (UniqueName: \"kubernetes.io/projected/23eadd14-852e-4683-bae5-3720a36d9407-kube-api-access-hzm7d\") pod \"mariadb-operator-controller-manager-f9fb45f8f-kvmkj\" (UID: \"23eadd14-852e-4683-bae5-3720a36d9407\") " pod="openstack-operators/mariadb-operator-controller-manager-f9fb45f8f-kvmkj" Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.375614 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mrgx6\" (UniqueName: \"kubernetes.io/projected/4bc1fceb-2d98-415e-b34e-82c2d6e1430f-kube-api-access-mrgx6\") pod \"manila-operator-controller-manager-5f67fbc655-szjkl\" (UID: \"4bc1fceb-2d98-415e-b34e-82c2d6e1430f\") " pod="openstack-operators/manila-operator-controller-manager-5f67fbc655-szjkl" Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.375654 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9vxzk\" (UniqueName: \"kubernetes.io/projected/d4236126-b907-4f39-a491-5006f2a9c301-kube-api-access-9vxzk\") pod \"infra-operator-controller-manager-656bcbd775-lp9f2\" (UID: \"d4236126-b907-4f39-a491-5006f2a9c301\") " pod="openstack-operators/infra-operator-controller-manager-656bcbd775-lp9f2" Oct 14 07:17:49 crc kubenswrapper[4870]: E1014 07:17:49.378899 4870 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Oct 14 07:17:49 crc kubenswrapper[4870]: E1014 07:17:49.378959 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d4236126-b907-4f39-a491-5006f2a9c301-cert podName:d4236126-b907-4f39-a491-5006f2a9c301 nodeName:}" failed. No retries permitted until 2025-10-14 07:17:49.878939746 +0000 UTC m=+1005.576300117 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/d4236126-b907-4f39-a491-5006f2a9c301-cert") pod "infra-operator-controller-manager-656bcbd775-lp9f2" (UID: "d4236126-b907-4f39-a491-5006f2a9c301") : secret "infra-operator-webhook-server-cert" not found Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.380221 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-7ffbcb7588-g2fdg" Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.407295 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-5458f77c4-n4csp"] Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.409648 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5458f77c4-n4csp" Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.412411 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mrgx6\" (UniqueName: \"kubernetes.io/projected/4bc1fceb-2d98-415e-b34e-82c2d6e1430f-kube-api-access-mrgx6\") pod \"manila-operator-controller-manager-5f67fbc655-szjkl\" (UID: \"4bc1fceb-2d98-415e-b34e-82c2d6e1430f\") " pod="openstack-operators/manila-operator-controller-manager-5f67fbc655-szjkl" Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.416508 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-klrlz\" (UniqueName: \"kubernetes.io/projected/586d846b-21c2-40f4-8da4-37392d812645-kube-api-access-klrlz\") pod \"neutron-operator-controller-manager-79d585cb66-67fkk\" (UID: \"586d846b-21c2-40f4-8da4-37392d812645\") " pod="openstack-operators/neutron-operator-controller-manager-79d585cb66-67fkk" Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.416890 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-lfnxb" Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.421040 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-25qll\" (UniqueName: \"kubernetes.io/projected/5ad2a6ec-e2aa-45d9-b1a8-59ed908aaacd-kube-api-access-25qll\") pod \"keystone-operator-controller-manager-55b6b7c7b8-jbt9q\" (UID: \"5ad2a6ec-e2aa-45d9-b1a8-59ed908aaacd\") " pod="openstack-operators/keystone-operator-controller-manager-55b6b7c7b8-jbt9q" Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.424016 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hzm7d\" (UniqueName: \"kubernetes.io/projected/23eadd14-852e-4683-bae5-3720a36d9407-kube-api-access-hzm7d\") pod \"mariadb-operator-controller-manager-f9fb45f8f-kvmkj\" (UID: \"23eadd14-852e-4683-bae5-3720a36d9407\") " pod="openstack-operators/mariadb-operator-controller-manager-f9fb45f8f-kvmkj" Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.429687 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xwtw5\" (UniqueName: \"kubernetes.io/projected/d27f8663-102d-4565-b7d5-00db25a8388a-kube-api-access-xwtw5\") pod \"ironic-operator-controller-manager-9c5c78d49-gjsj6\" (UID: \"d27f8663-102d-4565-b7d5-00db25a8388a\") " pod="openstack-operators/ironic-operator-controller-manager-9c5c78d49-gjsj6" Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.452355 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-5f67fbc655-szjkl" Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.458037 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9vxzk\" (UniqueName: \"kubernetes.io/projected/d4236126-b907-4f39-a491-5006f2a9c301-kube-api-access-9vxzk\") pod \"infra-operator-controller-manager-656bcbd775-lp9f2\" (UID: \"d4236126-b907-4f39-a491-5006f2a9c301\") " pod="openstack-operators/infra-operator-controller-manager-656bcbd775-lp9f2" Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.469599 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-55b6b7c7b8-jbt9q" Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.479462 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zxw8h\" (UniqueName: \"kubernetes.io/projected/01663eb4-cb0e-41fe-92c9-97e908f4ddff-kube-api-access-zxw8h\") pod \"telemetry-operator-controller-manager-67cfc6749b-cblnh\" (UID: \"01663eb4-cb0e-41fe-92c9-97e908f4ddff\") " pod="openstack-operators/telemetry-operator-controller-manager-67cfc6749b-cblnh" Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.479506 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2nnc5\" (UniqueName: \"kubernetes.io/projected/3278d66b-0ec7-47b4-8499-1671fc13f4ee-kube-api-access-2nnc5\") pod \"openstack-baremetal-operator-controller-manager-55b7d4484882chz\" (UID: \"3278d66b-0ec7-47b4-8499-1671fc13f4ee\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-55b7d4484882chz" Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.479543 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mmq7l\" (UniqueName: \"kubernetes.io/projected/34891227-4e7a-43d5-8dc0-3f56b54bb9fd-kube-api-access-mmq7l\") pod \"octavia-operator-controller-manager-69fdcfc5f5-2schz\" (UID: \"34891227-4e7a-43d5-8dc0-3f56b54bb9fd\") " pod="openstack-operators/octavia-operator-controller-manager-69fdcfc5f5-2schz" Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.479610 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d2pg4\" (UniqueName: \"kubernetes.io/projected/96effd40-d810-4b88-9dec-a6e98993778f-kube-api-access-d2pg4\") pod \"placement-operator-controller-manager-68b6c87b68-dtn8j\" (UID: \"96effd40-d810-4b88-9dec-a6e98993778f\") " pod="openstack-operators/placement-operator-controller-manager-68b6c87b68-dtn8j" Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.479630 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/3278d66b-0ec7-47b4-8499-1671fc13f4ee-cert\") pod \"openstack-baremetal-operator-controller-manager-55b7d4484882chz\" (UID: \"3278d66b-0ec7-47b4-8499-1671fc13f4ee\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-55b7d4484882chz" Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.479649 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qp7tr\" (UniqueName: \"kubernetes.io/projected/c2b45ced-e9c2-4d6f-ac5e-1960e5a9296a-kube-api-access-qp7tr\") pod \"swift-operator-controller-manager-db6d7f97b-22qw9\" (UID: \"c2b45ced-e9c2-4d6f-ac5e-1960e5a9296a\") " pod="openstack-operators/swift-operator-controller-manager-db6d7f97b-22qw9" Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.479674 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h59q2\" (UniqueName: \"kubernetes.io/projected/34297fe0-c74f-4d69-8847-86346bfbea4f-kube-api-access-h59q2\") pod \"nova-operator-controller-manager-5df598886f-zj4rk\" (UID: \"34297fe0-c74f-4d69-8847-86346bfbea4f\") " pod="openstack-operators/nova-operator-controller-manager-5df598886f-zj4rk" Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.479702 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wjkws\" (UniqueName: \"kubernetes.io/projected/9916d05d-8aa7-4e71-9ea6-07f2847fc4e6-kube-api-access-wjkws\") pod \"ovn-operator-controller-manager-79df5fb58c-s54lf\" (UID: \"9916d05d-8aa7-4e71-9ea6-07f2847fc4e6\") " pod="openstack-operators/ovn-operator-controller-manager-79df5fb58c-s54lf" Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.484592 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5458f77c4-n4csp"] Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.490146 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-f9fb45f8f-kvmkj" Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.517000 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-7f554bff7b-68d4q"] Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.518173 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-7f554bff7b-68d4q" Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.519898 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-cxbc9" Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.529766 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-7f554bff7b-68d4q"] Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.580643 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d2pg4\" (UniqueName: \"kubernetes.io/projected/96effd40-d810-4b88-9dec-a6e98993778f-kube-api-access-d2pg4\") pod \"placement-operator-controller-manager-68b6c87b68-dtn8j\" (UID: \"96effd40-d810-4b88-9dec-a6e98993778f\") " pod="openstack-operators/placement-operator-controller-manager-68b6c87b68-dtn8j" Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.580677 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/3278d66b-0ec7-47b4-8499-1671fc13f4ee-cert\") pod \"openstack-baremetal-operator-controller-manager-55b7d4484882chz\" (UID: \"3278d66b-0ec7-47b4-8499-1671fc13f4ee\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-55b7d4484882chz" Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.580701 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qp7tr\" (UniqueName: \"kubernetes.io/projected/c2b45ced-e9c2-4d6f-ac5e-1960e5a9296a-kube-api-access-qp7tr\") pod \"swift-operator-controller-manager-db6d7f97b-22qw9\" (UID: \"c2b45ced-e9c2-4d6f-ac5e-1960e5a9296a\") " pod="openstack-operators/swift-operator-controller-manager-db6d7f97b-22qw9" Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.580736 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h59q2\" (UniqueName: \"kubernetes.io/projected/34297fe0-c74f-4d69-8847-86346bfbea4f-kube-api-access-h59q2\") pod \"nova-operator-controller-manager-5df598886f-zj4rk\" (UID: \"34297fe0-c74f-4d69-8847-86346bfbea4f\") " pod="openstack-operators/nova-operator-controller-manager-5df598886f-zj4rk" Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.580762 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hflh4\" (UniqueName: \"kubernetes.io/projected/f47c02fb-6e94-4f5c-873f-ecaacd522121-kube-api-access-hflh4\") pod \"test-operator-controller-manager-5458f77c4-n4csp\" (UID: \"f47c02fb-6e94-4f5c-873f-ecaacd522121\") " pod="openstack-operators/test-operator-controller-manager-5458f77c4-n4csp" Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.580786 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wjkws\" (UniqueName: \"kubernetes.io/projected/9916d05d-8aa7-4e71-9ea6-07f2847fc4e6-kube-api-access-wjkws\") pod \"ovn-operator-controller-manager-79df5fb58c-s54lf\" (UID: \"9916d05d-8aa7-4e71-9ea6-07f2847fc4e6\") " pod="openstack-operators/ovn-operator-controller-manager-79df5fb58c-s54lf" Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.580814 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zxw8h\" (UniqueName: \"kubernetes.io/projected/01663eb4-cb0e-41fe-92c9-97e908f4ddff-kube-api-access-zxw8h\") pod \"telemetry-operator-controller-manager-67cfc6749b-cblnh\" (UID: \"01663eb4-cb0e-41fe-92c9-97e908f4ddff\") " pod="openstack-operators/telemetry-operator-controller-manager-67cfc6749b-cblnh" Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.580830 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2nnc5\" (UniqueName: \"kubernetes.io/projected/3278d66b-0ec7-47b4-8499-1671fc13f4ee-kube-api-access-2nnc5\") pod \"openstack-baremetal-operator-controller-manager-55b7d4484882chz\" (UID: \"3278d66b-0ec7-47b4-8499-1671fc13f4ee\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-55b7d4484882chz" Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.580857 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mmq7l\" (UniqueName: \"kubernetes.io/projected/34891227-4e7a-43d5-8dc0-3f56b54bb9fd-kube-api-access-mmq7l\") pod \"octavia-operator-controller-manager-69fdcfc5f5-2schz\" (UID: \"34891227-4e7a-43d5-8dc0-3f56b54bb9fd\") " pod="openstack-operators/octavia-operator-controller-manager-69fdcfc5f5-2schz" Oct 14 07:17:49 crc kubenswrapper[4870]: E1014 07:17:49.581163 4870 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Oct 14 07:17:49 crc kubenswrapper[4870]: E1014 07:17:49.581202 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3278d66b-0ec7-47b4-8499-1671fc13f4ee-cert podName:3278d66b-0ec7-47b4-8499-1671fc13f4ee nodeName:}" failed. No retries permitted until 2025-10-14 07:17:50.081188574 +0000 UTC m=+1005.778548945 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/3278d66b-0ec7-47b4-8499-1671fc13f4ee-cert") pod "openstack-baremetal-operator-controller-manager-55b7d4484882chz" (UID: "3278d66b-0ec7-47b4-8499-1671fc13f4ee") : secret "openstack-baremetal-operator-webhook-server-cert" not found Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.604047 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wjkws\" (UniqueName: \"kubernetes.io/projected/9916d05d-8aa7-4e71-9ea6-07f2847fc4e6-kube-api-access-wjkws\") pod \"ovn-operator-controller-manager-79df5fb58c-s54lf\" (UID: \"9916d05d-8aa7-4e71-9ea6-07f2847fc4e6\") " pod="openstack-operators/ovn-operator-controller-manager-79df5fb58c-s54lf" Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.619036 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zxw8h\" (UniqueName: \"kubernetes.io/projected/01663eb4-cb0e-41fe-92c9-97e908f4ddff-kube-api-access-zxw8h\") pod \"telemetry-operator-controller-manager-67cfc6749b-cblnh\" (UID: \"01663eb4-cb0e-41fe-92c9-97e908f4ddff\") " pod="openstack-operators/telemetry-operator-controller-manager-67cfc6749b-cblnh" Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.609210 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mmq7l\" (UniqueName: \"kubernetes.io/projected/34891227-4e7a-43d5-8dc0-3f56b54bb9fd-kube-api-access-mmq7l\") pod \"octavia-operator-controller-manager-69fdcfc5f5-2schz\" (UID: \"34891227-4e7a-43d5-8dc0-3f56b54bb9fd\") " pod="openstack-operators/octavia-operator-controller-manager-69fdcfc5f5-2schz" Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.623341 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2nnc5\" (UniqueName: \"kubernetes.io/projected/3278d66b-0ec7-47b4-8499-1671fc13f4ee-kube-api-access-2nnc5\") pod \"openstack-baremetal-operator-controller-manager-55b7d4484882chz\" (UID: \"3278d66b-0ec7-47b4-8499-1671fc13f4ee\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-55b7d4484882chz" Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.649558 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-7fb8c88b76-jzcb4"] Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.655494 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qp7tr\" (UniqueName: \"kubernetes.io/projected/c2b45ced-e9c2-4d6f-ac5e-1960e5a9296a-kube-api-access-qp7tr\") pod \"swift-operator-controller-manager-db6d7f97b-22qw9\" (UID: \"c2b45ced-e9c2-4d6f-ac5e-1960e5a9296a\") " pod="openstack-operators/swift-operator-controller-manager-db6d7f97b-22qw9" Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.661697 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-7fb8c88b76-jzcb4" Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.667221 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-g7625" Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.669265 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h59q2\" (UniqueName: \"kubernetes.io/projected/34297fe0-c74f-4d69-8847-86346bfbea4f-kube-api-access-h59q2\") pod \"nova-operator-controller-manager-5df598886f-zj4rk\" (UID: \"34297fe0-c74f-4d69-8847-86346bfbea4f\") " pod="openstack-operators/nova-operator-controller-manager-5df598886f-zj4rk" Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.669552 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.670362 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-79d585cb66-67fkk" Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.676011 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-7fb8c88b76-jzcb4"] Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.680471 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d2pg4\" (UniqueName: \"kubernetes.io/projected/96effd40-d810-4b88-9dec-a6e98993778f-kube-api-access-d2pg4\") pod \"placement-operator-controller-manager-68b6c87b68-dtn8j\" (UID: \"96effd40-d810-4b88-9dec-a6e98993778f\") " pod="openstack-operators/placement-operator-controller-manager-68b6c87b68-dtn8j" Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.682642 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-5df598886f-zj4rk" Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.687578 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pnmb7\" (UniqueName: \"kubernetes.io/projected/a1711137-ad44-4da0-ab90-5666fd111419-kube-api-access-pnmb7\") pod \"watcher-operator-controller-manager-7f554bff7b-68d4q\" (UID: \"a1711137-ad44-4da0-ab90-5666fd111419\") " pod="openstack-operators/watcher-operator-controller-manager-7f554bff7b-68d4q" Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.687646 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hflh4\" (UniqueName: \"kubernetes.io/projected/f47c02fb-6e94-4f5c-873f-ecaacd522121-kube-api-access-hflh4\") pod \"test-operator-controller-manager-5458f77c4-n4csp\" (UID: \"f47c02fb-6e94-4f5c-873f-ecaacd522121\") " pod="openstack-operators/test-operator-controller-manager-5458f77c4-n4csp" Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.709659 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-9c5c78d49-gjsj6" Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.711784 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-lgqbd"] Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.713027 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-lgqbd" Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.715479 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-nkbl9" Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.726359 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-69fdcfc5f5-2schz" Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.728803 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hflh4\" (UniqueName: \"kubernetes.io/projected/f47c02fb-6e94-4f5c-873f-ecaacd522121-kube-api-access-hflh4\") pod \"test-operator-controller-manager-5458f77c4-n4csp\" (UID: \"f47c02fb-6e94-4f5c-873f-ecaacd522121\") " pod="openstack-operators/test-operator-controller-manager-5458f77c4-n4csp" Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.735598 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-lgqbd"] Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.773486 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-db6d7f97b-22qw9" Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.780694 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-67cfc6749b-cblnh" Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.789858 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bxspk\" (UniqueName: \"kubernetes.io/projected/de2a07da-ed81-4aac-bf11-da6b54ad095a-kube-api-access-bxspk\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-lgqbd\" (UID: \"de2a07da-ed81-4aac-bf11-da6b54ad095a\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-lgqbd" Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.789940 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/3645fdee-7690-4d37-a0f8-16ea440d9453-cert\") pod \"openstack-operator-controller-manager-7fb8c88b76-jzcb4\" (UID: \"3645fdee-7690-4d37-a0f8-16ea440d9453\") " pod="openstack-operators/openstack-operator-controller-manager-7fb8c88b76-jzcb4" Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.790205 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pnmb7\" (UniqueName: \"kubernetes.io/projected/a1711137-ad44-4da0-ab90-5666fd111419-kube-api-access-pnmb7\") pod \"watcher-operator-controller-manager-7f554bff7b-68d4q\" (UID: \"a1711137-ad44-4da0-ab90-5666fd111419\") " pod="openstack-operators/watcher-operator-controller-manager-7f554bff7b-68d4q" Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.790248 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vhbdh\" (UniqueName: \"kubernetes.io/projected/3645fdee-7690-4d37-a0f8-16ea440d9453-kube-api-access-vhbdh\") pod \"openstack-operator-controller-manager-7fb8c88b76-jzcb4\" (UID: \"3645fdee-7690-4d37-a0f8-16ea440d9453\") " pod="openstack-operators/openstack-operator-controller-manager-7fb8c88b76-jzcb4" Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.800293 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5458f77c4-n4csp" Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.819932 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pnmb7\" (UniqueName: \"kubernetes.io/projected/a1711137-ad44-4da0-ab90-5666fd111419-kube-api-access-pnmb7\") pod \"watcher-operator-controller-manager-7f554bff7b-68d4q\" (UID: \"a1711137-ad44-4da0-ab90-5666fd111419\") " pod="openstack-operators/watcher-operator-controller-manager-7f554bff7b-68d4q" Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.832061 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-7f554bff7b-68d4q" Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.864093 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-79df5fb58c-s54lf" Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.891714 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d4236126-b907-4f39-a491-5006f2a9c301-cert\") pod \"infra-operator-controller-manager-656bcbd775-lp9f2\" (UID: \"d4236126-b907-4f39-a491-5006f2a9c301\") " pod="openstack-operators/infra-operator-controller-manager-656bcbd775-lp9f2" Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.891771 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vhbdh\" (UniqueName: \"kubernetes.io/projected/3645fdee-7690-4d37-a0f8-16ea440d9453-kube-api-access-vhbdh\") pod \"openstack-operator-controller-manager-7fb8c88b76-jzcb4\" (UID: \"3645fdee-7690-4d37-a0f8-16ea440d9453\") " pod="openstack-operators/openstack-operator-controller-manager-7fb8c88b76-jzcb4" Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.891820 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bxspk\" (UniqueName: \"kubernetes.io/projected/de2a07da-ed81-4aac-bf11-da6b54ad095a-kube-api-access-bxspk\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-lgqbd\" (UID: \"de2a07da-ed81-4aac-bf11-da6b54ad095a\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-lgqbd" Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.891841 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/3645fdee-7690-4d37-a0f8-16ea440d9453-cert\") pod \"openstack-operator-controller-manager-7fb8c88b76-jzcb4\" (UID: \"3645fdee-7690-4d37-a0f8-16ea440d9453\") " pod="openstack-operators/openstack-operator-controller-manager-7fb8c88b76-jzcb4" Oct 14 07:17:49 crc kubenswrapper[4870]: E1014 07:17:49.893475 4870 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Oct 14 07:17:49 crc kubenswrapper[4870]: E1014 07:17:49.893519 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3645fdee-7690-4d37-a0f8-16ea440d9453-cert podName:3645fdee-7690-4d37-a0f8-16ea440d9453 nodeName:}" failed. No retries permitted until 2025-10-14 07:17:50.39350524 +0000 UTC m=+1006.090865611 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/3645fdee-7690-4d37-a0f8-16ea440d9453-cert") pod "openstack-operator-controller-manager-7fb8c88b76-jzcb4" (UID: "3645fdee-7690-4d37-a0f8-16ea440d9453") : secret "webhook-server-cert" not found Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.907199 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d4236126-b907-4f39-a491-5006f2a9c301-cert\") pod \"infra-operator-controller-manager-656bcbd775-lp9f2\" (UID: \"d4236126-b907-4f39-a491-5006f2a9c301\") " pod="openstack-operators/infra-operator-controller-manager-656bcbd775-lp9f2" Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.916204 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vhbdh\" (UniqueName: \"kubernetes.io/projected/3645fdee-7690-4d37-a0f8-16ea440d9453-kube-api-access-vhbdh\") pod \"openstack-operator-controller-manager-7fb8c88b76-jzcb4\" (UID: \"3645fdee-7690-4d37-a0f8-16ea440d9453\") " pod="openstack-operators/openstack-operator-controller-manager-7fb8c88b76-jzcb4" Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.928302 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bxspk\" (UniqueName: \"kubernetes.io/projected/de2a07da-ed81-4aac-bf11-da6b54ad095a-kube-api-access-bxspk\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-lgqbd\" (UID: \"de2a07da-ed81-4aac-bf11-da6b54ad095a\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-lgqbd" Oct 14 07:17:49 crc kubenswrapper[4870]: I1014 07:17:49.935960 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-68b6c87b68-dtn8j" Oct 14 07:17:50 crc kubenswrapper[4870]: I1014 07:17:50.001815 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-656bcbd775-lp9f2" Oct 14 07:17:50 crc kubenswrapper[4870]: I1014 07:17:50.095874 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/3278d66b-0ec7-47b4-8499-1671fc13f4ee-cert\") pod \"openstack-baremetal-operator-controller-manager-55b7d4484882chz\" (UID: \"3278d66b-0ec7-47b4-8499-1671fc13f4ee\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-55b7d4484882chz" Oct 14 07:17:50 crc kubenswrapper[4870]: I1014 07:17:50.103091 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/3278d66b-0ec7-47b4-8499-1671fc13f4ee-cert\") pod \"openstack-baremetal-operator-controller-manager-55b7d4484882chz\" (UID: \"3278d66b-0ec7-47b4-8499-1671fc13f4ee\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-55b7d4484882chz" Oct 14 07:17:50 crc kubenswrapper[4870]: I1014 07:17:50.191812 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-lgqbd" Oct 14 07:17:50 crc kubenswrapper[4870]: I1014 07:17:50.195759 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-85d5d9dd78-ld7xh"] Oct 14 07:17:50 crc kubenswrapper[4870]: I1014 07:17:50.354981 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-55b7d4484882chz" Oct 14 07:17:50 crc kubenswrapper[4870]: I1014 07:17:50.404157 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/3645fdee-7690-4d37-a0f8-16ea440d9453-cert\") pod \"openstack-operator-controller-manager-7fb8c88b76-jzcb4\" (UID: \"3645fdee-7690-4d37-a0f8-16ea440d9453\") " pod="openstack-operators/openstack-operator-controller-manager-7fb8c88b76-jzcb4" Oct 14 07:17:50 crc kubenswrapper[4870]: E1014 07:17:50.404358 4870 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Oct 14 07:17:50 crc kubenswrapper[4870]: E1014 07:17:50.404409 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3645fdee-7690-4d37-a0f8-16ea440d9453-cert podName:3645fdee-7690-4d37-a0f8-16ea440d9453 nodeName:}" failed. No retries permitted until 2025-10-14 07:17:51.404394218 +0000 UTC m=+1007.101754589 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/3645fdee-7690-4d37-a0f8-16ea440d9453-cert") pod "openstack-operator-controller-manager-7fb8c88b76-jzcb4" (UID: "3645fdee-7690-4d37-a0f8-16ea440d9453") : secret "webhook-server-cert" not found Oct 14 07:17:50 crc kubenswrapper[4870]: I1014 07:17:50.427754 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-658bdf4b74-76kw6"] Oct 14 07:17:50 crc kubenswrapper[4870]: I1014 07:17:50.896817 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-7ffbcb7588-g2fdg"] Oct 14 07:17:50 crc kubenswrapper[4870]: I1014 07:17:50.909617 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-84b9b84486-cl7ft"] Oct 14 07:17:50 crc kubenswrapper[4870]: I1014 07:17:50.918132 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-85d5d9dd78-ld7xh" event={"ID":"634060b4-d317-415d-bc9e-aa3ed6503b4f","Type":"ContainerStarted","Data":"e4102ab3c559c2ac6722b50226bf547282e632e1289ce99aef79242669c78759"} Oct 14 07:17:50 crc kubenswrapper[4870]: I1014 07:17:50.918498 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-5f67fbc655-szjkl"] Oct 14 07:17:50 crc kubenswrapper[4870]: I1014 07:17:50.923014 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-55b6b7c7b8-jbt9q"] Oct 14 07:17:50 crc kubenswrapper[4870]: I1014 07:17:50.926996 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-658bdf4b74-76kw6" event={"ID":"36e6bbc0-f57d-4486-9120-d2cd5a5ce638","Type":"ContainerStarted","Data":"0778905e3472e4f8cdea2120c029b09bdefda1643b4d53c06031590ef6893ca4"} Oct 14 07:17:50 crc kubenswrapper[4870]: I1014 07:17:50.929652 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-858f76bbdd-9bdj8"] Oct 14 07:17:50 crc kubenswrapper[4870]: I1014 07:17:50.934800 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-7b7fb68549-xnk22"] Oct 14 07:17:50 crc kubenswrapper[4870]: I1014 07:17:50.957631 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-9c5c78d49-gjsj6"] Oct 14 07:17:50 crc kubenswrapper[4870]: W1014 07:17:50.968883 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod23eadd14_852e_4683_bae5_3720a36d9407.slice/crio-75da8f8484a03bbda4202d28165ecdd6483da2da0a7ea1492ea636e82d716210 WatchSource:0}: Error finding container 75da8f8484a03bbda4202d28165ecdd6483da2da0a7ea1492ea636e82d716210: Status 404 returned error can't find the container with id 75da8f8484a03bbda4202d28165ecdd6483da2da0a7ea1492ea636e82d716210 Oct 14 07:17:50 crc kubenswrapper[4870]: W1014 07:17:50.979873 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd27f8663_102d_4565_b7d5_00db25a8388a.slice/crio-0625082843ab92a357a4aa63b720f29efd40d531011415896a763e95ecebddb4 WatchSource:0}: Error finding container 0625082843ab92a357a4aa63b720f29efd40d531011415896a763e95ecebddb4: Status 404 returned error can't find the container with id 0625082843ab92a357a4aa63b720f29efd40d531011415896a763e95ecebddb4 Oct 14 07:17:50 crc kubenswrapper[4870]: W1014 07:17:50.981305 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod34297fe0_c74f_4d69_8847_86346bfbea4f.slice/crio-6af064ea906d91d300ba56c4e386475a92f721c282f358991d95da5ddbf88b74 WatchSource:0}: Error finding container 6af064ea906d91d300ba56c4e386475a92f721c282f358991d95da5ddbf88b74: Status 404 returned error can't find the container with id 6af064ea906d91d300ba56c4e386475a92f721c282f358991d95da5ddbf88b74 Oct 14 07:17:50 crc kubenswrapper[4870]: I1014 07:17:50.984502 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-f9fb45f8f-kvmkj"] Oct 14 07:17:50 crc kubenswrapper[4870]: I1014 07:17:50.992183 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-69fdcfc5f5-2schz"] Oct 14 07:17:50 crc kubenswrapper[4870]: I1014 07:17:50.996718 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-5df598886f-zj4rk"] Oct 14 07:17:51 crc kubenswrapper[4870]: I1014 07:17:51.336736 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5458f77c4-n4csp"] Oct 14 07:17:51 crc kubenswrapper[4870]: I1014 07:17:51.343761 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-7f554bff7b-68d4q"] Oct 14 07:17:51 crc kubenswrapper[4870]: W1014 07:17:51.350624 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda1711137_ad44_4da0_ab90_5666fd111419.slice/crio-dbd7992aac8237a4ba71f957e666da232495b64febba3bd4c80e7c71b39a523d WatchSource:0}: Error finding container dbd7992aac8237a4ba71f957e666da232495b64febba3bd4c80e7c71b39a523d: Status 404 returned error can't find the container with id dbd7992aac8237a4ba71f957e666da232495b64febba3bd4c80e7c71b39a523d Oct 14 07:17:51 crc kubenswrapper[4870]: W1014 07:17:51.352690 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf47c02fb_6e94_4f5c_873f_ecaacd522121.slice/crio-25e1f80edb99216885fc256a03130d09e04229f63c37a191794bd87248198ba2 WatchSource:0}: Error finding container 25e1f80edb99216885fc256a03130d09e04229f63c37a191794bd87248198ba2: Status 404 returned error can't find the container with id 25e1f80edb99216885fc256a03130d09e04229f63c37a191794bd87248198ba2 Oct 14 07:17:51 crc kubenswrapper[4870]: I1014 07:17:51.374481 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-79d585cb66-67fkk"] Oct 14 07:17:51 crc kubenswrapper[4870]: I1014 07:17:51.395746 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-68b6c87b68-dtn8j"] Oct 14 07:17:51 crc kubenswrapper[4870]: I1014 07:17:51.397720 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-lgqbd"] Oct 14 07:17:51 crc kubenswrapper[4870]: E1014 07:17:51.399493 4870 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/telemetry-operator@sha256:abe978f8da75223de5043cca50278ad4e28c8dd309883f502fe1e7a9998733b0,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-zxw8h,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-67cfc6749b-cblnh_openstack-operators(01663eb4-cb0e-41fe-92c9-97e908f4ddff): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 14 07:17:51 crc kubenswrapper[4870]: E1014 07:17:51.399983 4870 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/swift-operator@sha256:4b4a17fe08ce00e375afaaec6a28835f5c1784f03d11c4558376ac04130f3a9e,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-qp7tr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-db6d7f97b-22qw9_openstack-operators(c2b45ced-e9c2-4d6f-ac5e-1960e5a9296a): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 14 07:17:51 crc kubenswrapper[4870]: E1014 07:17:51.400473 4870 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/placement-operator@sha256:d33c1f507e1f5b9a4bf226ad98917e92101ac66b36e19d35cbe04ae7014f6bff,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-d2pg4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-68b6c87b68-dtn8j_openstack-operators(96effd40-d810-4b88-9dec-a6e98993778f): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 14 07:17:51 crc kubenswrapper[4870]: E1014 07:17:51.400696 4870 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-bxspk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-5f97d8c699-lgqbd_openstack-operators(de2a07da-ed81-4aac-bf11-da6b54ad095a): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 14 07:17:51 crc kubenswrapper[4870]: E1014 07:17:51.402218 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-lgqbd" podUID="de2a07da-ed81-4aac-bf11-da6b54ad095a" Oct 14 07:17:51 crc kubenswrapper[4870]: I1014 07:17:51.421690 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/3645fdee-7690-4d37-a0f8-16ea440d9453-cert\") pod \"openstack-operator-controller-manager-7fb8c88b76-jzcb4\" (UID: \"3645fdee-7690-4d37-a0f8-16ea440d9453\") " pod="openstack-operators/openstack-operator-controller-manager-7fb8c88b76-jzcb4" Oct 14 07:17:51 crc kubenswrapper[4870]: I1014 07:17:51.423239 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-db6d7f97b-22qw9"] Oct 14 07:17:51 crc kubenswrapper[4870]: E1014 07:17:51.423855 4870 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/infra-operator@sha256:5cfb2ae1092445950b39dd59caa9a8c9367f42fb8353a8c3848d3bc729f24492,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{600 -3} {} 600m DecimalSI},memory: {{2147483648 0} {} 2Gi BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{536870912 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:cert,ReadOnly:true,MountPath:/tmp/k8s-webhook-server/serving-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-9vxzk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod infra-operator-controller-manager-656bcbd775-lp9f2_openstack-operators(d4236126-b907-4f39-a491-5006f2a9c301): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 14 07:17:51 crc kubenswrapper[4870]: E1014 07:17:51.428634 4870 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/neutron-operator@sha256:33652e75a03a058769019fe8d8c51585a6eeefef5e1ecb96f9965434117954f2,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-klrlz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod neutron-operator-controller-manager-79d585cb66-67fkk_openstack-operators(586d846b-21c2-40f4-8da4-37392d812645): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 14 07:17:51 crc kubenswrapper[4870]: I1014 07:17:51.431019 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/3645fdee-7690-4d37-a0f8-16ea440d9453-cert\") pod \"openstack-operator-controller-manager-7fb8c88b76-jzcb4\" (UID: \"3645fdee-7690-4d37-a0f8-16ea440d9453\") " pod="openstack-operators/openstack-operator-controller-manager-7fb8c88b76-jzcb4" Oct 14 07:17:51 crc kubenswrapper[4870]: I1014 07:17:51.434175 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-67cfc6749b-cblnh"] Oct 14 07:17:51 crc kubenswrapper[4870]: I1014 07:17:51.446522 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-656bcbd775-lp9f2"] Oct 14 07:17:51 crc kubenswrapper[4870]: E1014 07:17:51.450213 4870 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ovn-operator@sha256:315e558023b41ac1aa215082096995a03810c5b42910a33b00427ffcac9c6a14,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-wjkws,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-operator-controller-manager-79df5fb58c-s54lf_openstack-operators(9916d05d-8aa7-4e71-9ea6-07f2847fc4e6): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 14 07:17:51 crc kubenswrapper[4870]: W1014 07:17:51.453648 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3278d66b_0ec7_47b4_8499_1671fc13f4ee.slice/crio-9b683453df0abbe44cc9d84e555c6c73cc85e23af4db1d2388386231d6edd11b WatchSource:0}: Error finding container 9b683453df0abbe44cc9d84e555c6c73cc85e23af4db1d2388386231d6edd11b: Status 404 returned error can't find the container with id 9b683453df0abbe44cc9d84e555c6c73cc85e23af4db1d2388386231d6edd11b Oct 14 07:17:51 crc kubenswrapper[4870]: I1014 07:17:51.453894 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-79df5fb58c-s54lf"] Oct 14 07:17:51 crc kubenswrapper[4870]: E1014 07:17:51.458431 4870 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:a17fc270857869fd1efe5020b2a1cb8c2abbd838f08de88f3a6a59e8754ec351,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:true,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-baremetal-operator-agent@sha256:03b4f3db4b373515f7e4095984b97197c05a14f87b2a0a525eb5d7be1d7bda66,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_ANSIBLEEE_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-ansibleee-runner@sha256:6722a752fb7cbffbae811f6ad6567120fbd4ebbe8c38a83ec2df02850a3276bd,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-api@sha256:6889b3aa62468a9af923ac9e4f0aed418e4edd076d342c81e3a0042c3c8df19d,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_EVALUATOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-evaluator@sha256:bf9ab38bfe20a161e7376392f27674c833eeaaa737c2902d9132d5ee7845ab34,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_LISTENER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-listener@sha256:d879a4fe24c7fd9c2e7f5f91bea440c97fc7fe4891d76f761d45341e1f530b49,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_NOTIFIER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-notifier@sha256:17c8a458ac0f0fe1c1f34fd5d81d0291ed16e7bfa2ced8f2a68b13dcda00702a,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_APACHE_IMAGE_URL_DEFAULT,Value:registry.redhat.io/ubi9/httpd-24@sha256:e91d58021b54c46883595ff66be65882de54abdb3be2ca53c4162b20d18b5f48,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-api@sha256:121a845dacd051814fb4709fc557420363cd923a9cf2b4ed09addd394f83a3f5,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_KEYSTONE_LISTENER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-keystone-listener@sha256:8c9f2c92e8c6fcdbc2b90f8edc5e3a7055f965c93f426a70caaf2dd2808e01a9,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-worker@sha256:f0e68b61dcb34ac3851ff056661b66a5f1095bc862f4391d4abcc36397a9bd49,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_CENTRAL_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-central@sha256:37be59ccd30847c8e5a52ca48fb90e60ed1134aa96770929238f414bb9a96442,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_COMPUTE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-compute@sha256:52c583ac7a1f249e658bb7cbc02ffe863a3e090e3baf5fc253c60e267e73c713,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_IPMI_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-ipmi@sha256:8b755bcae75f21718f07c7740080b034f4f289b859072ec2020fa0fde3f8c4f0,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_MYSQLD_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/prometheus/mysqld-exporter@sha256:7211a617ec657701ca819aa0ba28e1d5750f5bf2c1391b755cc4a48cc360b0fa,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_NOTIFICATION_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-notification@sha256:bcf261baa33be891dfbda582cf679d59a943f9d311dbadbfac35f6bc3233b467,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_SGCORE_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/sg-core@sha256:09b5017c95d7697e66b9c64846bc48ef5826a009cba89b956ec54561e5f4a2d1,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-api@sha256:93b475af15a0d10e95cb17b98927077f05ac24c89472a601d677eb89f82fd429,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_BACKUP_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-backup@sha256:25f32b470fa8a112d2b3378215986adff06180f526a6ac59d892df0712ae4e92,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-scheduler@sha256:0f5f6fe0a4f79c2081e0cf990940ddd8976d105c004956e09debc366cfebb7ff,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_VOLUME_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-volume@sha256:6ed56dd94f79ff155814949be4d76a837be594c341f497191f276ffa5e681aa5,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-api@sha256:bb5fe980813796fabb38d0ee83dc949f775aed360aea1d3a9636c9ae7342c4d2,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_BACKENDBIND9_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-backend-bind9@sha256:46f7c5ad91770e99d707c09bca696e68b6248d33c0da45a89a8e2c26053b1436,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_CENTRAL_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-central@sha256:701f27d93b40becf0d8445a0ec993eb98606d0b4fb60cf21f745d21af6ce146c,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_MDNS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-mdns@sha256:8117021a965e52c9d2bd7b39b7ff7fa5b8217a0cd81c998a48c981191a490b6f,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_PRODUCER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-producer@sha256:1fa006cc027f654e8cc52f2ed908b1fd4de5ddd76acdaf2eff6fced95b6c3237,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_UNBOUND_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-unbound@sha256:2067f20fa52f106035b36c34e1a6378ec78b90b63fac500441b8e0ef86dabf4a,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-worker@sha256:2cb5b13dc65c9407d6221fd2831f310c51fb61d060cdf9b6555c9936fabad079,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_FRR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-frr@sha256:1fee90435ec29f7923062cfca6db89298f69ef7f1864df3e61258715d9d93c74,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_ISCSID_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-iscsid@sha256:124d7cea22be48d4f1a8cfedec66864ccd3bea72d0fbc0d6c8e6bf4a6820e8fe,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_KEPLER_IMAGE_URL_DEFAULT,Value:quay.io/sustainable_computing_io/kepler@sha256:581b65b646301e0fcb07582150ba63438f1353a85bf9acf1eb2acb4ce71c58bd,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_LOGROTATE_CROND_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cron@sha256:f0501133d0a13a33a8492ead797a981cf04bcd6e469de2390b645d763b9ce447,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_MULTIPATHD_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-multipathd@sha256:0f26bfcc3cc838a38a36e11055a96f7d28fb841d04aaf952494f27b1f8919d97,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_DHCP_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent@sha256:ad5f34ab2be053a1b70063f53d30c78ce1d53b7ed26d351d980256aea71d663e,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_METADATA_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn@sha256:081710f3f67a74adb03d6d8f527f6ef01828243c2be24ca57436de2be8618576,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_OVN_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-ovn-agent@sha256:49f308e55764cd3852a3c403107b0e5a91ccc86ec85733d4af341a18cdac4bc1,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_SRIOV_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-sriov-agent@sha256:375faec05ca0a5094bc75c17b0a775afefcc1dc30f5d7eb6ef670fb451a3fd1f,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NODE_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_OVN_BGP_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-bgp-agent@sha256:fe04296ed96d5cae7489761537c530d175e1d306c00daf109bcd6c86bf04767f,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_PODMAN_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_GLANCE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-glance-api@sha256:839f0e593dd6b59e385ec9471f4eeaa34f1c539268588114cbc34cc9a6117835,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-api@sha256:e46413b26dca08cdf44c265f9a0a772943a8113f67eb015185901f02bf3f467f,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_CFNAPI_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-api-cfn@sha256:fa9b8191aa63a4e2ba6b880a2680ce323afcc5ef655bff22d344edbb7bd60eff,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_ENGINE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-engine@sha256:76da897b5ed8dbfb9ddd60237d2bd05d9675a11e8e29054b9313f083503ab8d7,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HORIZON_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-horizon@sha256:a165bb40dcffccf823656b90260c23c472338339306b1942cd0e863ac45df852,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_INFRA_MEMCACHED_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-memcached@sha256:a5425f2dff41e2a3fc8c7168ceadc882dfb68dfd0e82e416bacd6e67604cdbac,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_INFRA_REDIS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-redis@sha256:be240c1d8a08b9461e19d95133dda2d4aa0b37f8a22d1f54b2602a04eff217de,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-api@sha256:0bfbc82dc110a6453af41ead54980dd869774dd754cb9a9a4908c56b02307928,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_CONDUCTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-conductor@sha256:9483e655210e66c5c3325afee77282aecee9c2d747a6a22f1468e5b365b90a27,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_INSPECTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-inspector@sha256:c3a430afd671404cb240b4aca54746b9c4bf93645f94c6613999196ca513913c,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_NEUTRON_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-neutron-agent@sha256:8da4a885ac8adabaeb60feffda093e2670fef1b2b614bf6b04637fe221cebb9a,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_PXE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-pxe@sha256:94e857a625e3cc4d8ec00ba84bdcaf368cb1ecd26c5e256f998664fb063addd5,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_PYTHON_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/ironic-python-agent@sha256:2acb369d62c3152313c3ac1d5b8c79b3e1f7d46ff9c3456e4f7e295a68f27c09,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_KEYSTONE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-keystone@sha256:f7302eb8964de699cf44da13958a8ce3c1c4c05406a6fc58b6cdcb1706b8f439,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_KSM_IMAGE_URL_DEFAULT,Value:registry.k8s.io/kube-state-metrics/kube-state-metrics@sha256:db384bf43222b066c378e77027a675d4cd9911107adba46c2922b3a55e10d6fb,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-api@sha256:25f6226ecbc296f4a9d3b220613400c135c9dc98dc8f06c5b26faf80cb92bb3d,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-scheduler@sha256:27aa5719ddf5a0130152ad03fd2a43118e42f508c83b672e3d8b01767bd66d37,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_SHARE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-share@sha256:cc67888d0b5b52501151b01a8075df22fadccc05ac31adc9c0a0135a8d99c946,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MARIADB_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-mariadb@sha256:91737b02501b73af0aee486b7447b4ae3005c904f31f1a9bf4047d0433586f80,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NET_UTILS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-netutils@sha256:6a72c6ddc7b9da9d91d37e15430acc63b6d9ebe44840d50e7fdc29eb511d3850,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NEUTRON_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:673685cea6ea2dbd78bcb555955c1b9f05ea26018f79ee34494256a5f2d7b74a,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-api@sha256:ba81e563e9ff7186a072aecff403c4e062fbaf7e06283de40fa6a724d9ff25b2,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_COMPUTE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-compute@sha256:33c58faa12b90b6009f89c9c60baeadc1323b62dcb141619a7a11c3c10903560,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_CONDUCTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-conductor@sha256:85219a043b93a0cad67e3c6a21c82ebd0de298be40d2124d678e551aba9cdca3,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_NOVNC_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-novncproxy@sha256:bcac7d43f7ca3e408eaab5af5b59bf83658e69d362b221495e3849c1412e7ac9,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-scheduler@sha256:b1a59a930d7b3f9411be8c55a9661adaae0927aa11d3379a93afa37e3273009f,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-api@sha256:dc4f0d8c3ef7fc6d16df02b5e068a1a561eae9e155f4dda7664a21ded81cedca,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_HEALTHMANAGER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-health-manager@sha256:5d3255457450c6b22246625db777c1f834ef9feacbef7d4bd1ed475a4aa65757,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_HOUSEKEEPING_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-housekeeping@sha256:609883d4f5ee1eeeb9877b1ab3636f325970d67a6e9e0a0674a26e9c12168e5f,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_RSYSLOG_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-rsyslog@sha256:8257d380a013eb290d88cdbb6279e142d293115e844f8554b493a7fd64dbe5db,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-worker@sha256:02f98ecaab5a69b32bc7184051eb1cf4ed84908766c4a78af5e8a25f3398af75,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_CLIENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-openstackclient@sha256:c2ebfcd639773ddc62a37198da44de8fd76348610d91424b880d41a51b702418,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_MUST_GATHER_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-must-gather@sha256:943eee724277e252795909137538a553ef5284c8103ad01b9be7b0138c66d14d,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_NETWORK_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OS_CONTAINER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/edpm-hardened-uefi@sha256:da88acc08ce68a76a93bcc45edcd66cbef80835f4881503a9050ad67739db89b,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_CONTROLLER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-controller@sha256:1f610ed4ebf657334da87dfd95b3dc5299fb3540ec1433ae3db34f0f247d8abf,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_CONTROLLER_OVS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-base@sha256:725e18e616e058867b99eb7f4b26e81374a7c6c52e1f9518392079e568d38f4a,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_NB_DBCLUSTER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-nb-db-server@sha256:b272e3f26af5f9f92234b6e1fd6ce860f46308e5cbe38cdbf631d19499ae115f,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_NORTHD_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-northd@sha256:fcff3afee08fff84ac33b5c25e3ee22c9b05402624163b1b2b9bb2e39ed7205d,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_SB_DBCLUSTER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-sb-db-server@sha256:f42888dd0e4efdc82c334650e97164c6b50e195acb4eee0bc77b35337ccc1328,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_PLACEMENT_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-placement-api@sha256:03b1444ea142e857467c1ee43e8c7eb4b0d57a3fa4dceed0aebbffab9d999c4c,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_RABBITMQ_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-rabbitmq@sha256:97feaea1e66145857f3eb548d741ee56062b97fd3e8f4d136a5ca807c49c0cca,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_ACCOUNT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-account@sha256:0eb4330fbe649411e9de5bc7a80af6050d765283f97d46a725ee49269b0600cf,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_CONTAINER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-container@sha256:b222aa9f56abead1e1b5d8159c0bc3687024d151f53a33cafdf33377feab8f35,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_OBJECT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-object@sha256:74280842b5e733f8456ae433560fc90aca37fd6ccc0c6d902b835d36b978e041,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_PROXY_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-proxy-server@sha256:9b2e302646bbf02dd583d9c340c5e7cdf708241c4c32700ebe7696ee235725bd,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_TEST_TEMPEST_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-tempest-all@sha256:74db5648836429519c8bebf9d1e7ea4dd3961c7b3b83bb3d50dffbcfe4c41a5d,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-api@sha256:062ac305d19a6eec23a678835b5aa3dc3dda82c7470089408f98f5728e90ce5a,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_APPLIER_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-applier@sha256:c83af60e124c22c9e9f2cdf6f1f00bd63a0d9b038dafac29c1c91c5c978923bc,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_DECISION_ENGINE_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-decision-engine@sha256:00176d96a79fb4eda8540cb22f1dce081234c60e9e6a2b034950bc8d6e718b79,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:cert,ReadOnly:true,MountPath:/tmp/k8s-webhook-server/serving-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-2nnc5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstack-baremetal-operator-controller-manager-55b7d4484882chz_openstack-operators(3278d66b-0ec7-47b4-8499-1671fc13f4ee): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 14 07:17:51 crc kubenswrapper[4870]: I1014 07:17:51.461371 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-55b7d4484882chz"] Oct 14 07:17:51 crc kubenswrapper[4870]: E1014 07:17:51.637352 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/swift-operator-controller-manager-db6d7f97b-22qw9" podUID="c2b45ced-e9c2-4d6f-ac5e-1960e5a9296a" Oct 14 07:17:51 crc kubenswrapper[4870]: I1014 07:17:51.662735 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-7fb8c88b76-jzcb4" Oct 14 07:17:51 crc kubenswrapper[4870]: E1014 07:17:51.718065 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/placement-operator-controller-manager-68b6c87b68-dtn8j" podUID="96effd40-d810-4b88-9dec-a6e98993778f" Oct 14 07:17:51 crc kubenswrapper[4870]: E1014 07:17:51.739235 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/infra-operator-controller-manager-656bcbd775-lp9f2" podUID="d4236126-b907-4f39-a491-5006f2a9c301" Oct 14 07:17:51 crc kubenswrapper[4870]: E1014 07:17:51.745360 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/ovn-operator-controller-manager-79df5fb58c-s54lf" podUID="9916d05d-8aa7-4e71-9ea6-07f2847fc4e6" Oct 14 07:17:51 crc kubenswrapper[4870]: E1014 07:17:51.784993 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/telemetry-operator-controller-manager-67cfc6749b-cblnh" podUID="01663eb4-cb0e-41fe-92c9-97e908f4ddff" Oct 14 07:17:51 crc kubenswrapper[4870]: E1014 07:17:51.797118 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/neutron-operator-controller-manager-79d585cb66-67fkk" podUID="586d846b-21c2-40f4-8da4-37392d812645" Oct 14 07:17:51 crc kubenswrapper[4870]: E1014 07:17:51.825333 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-55b7d4484882chz" podUID="3278d66b-0ec7-47b4-8499-1671fc13f4ee" Oct 14 07:17:51 crc kubenswrapper[4870]: I1014 07:17:51.996636 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-7f554bff7b-68d4q" event={"ID":"a1711137-ad44-4da0-ab90-5666fd111419","Type":"ContainerStarted","Data":"dbd7992aac8237a4ba71f957e666da232495b64febba3bd4c80e7c71b39a523d"} Oct 14 07:17:52 crc kubenswrapper[4870]: I1014 07:17:52.003082 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-7ffbcb7588-g2fdg" event={"ID":"8ed9c494-ed5c-46d0-9434-1edbfcfba361","Type":"ContainerStarted","Data":"1f2d9e02651d469e0e9ce50813e9204343e7f74cc004899993a3399d5f4d591c"} Oct 14 07:17:52 crc kubenswrapper[4870]: I1014 07:17:52.006858 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-5df598886f-zj4rk" event={"ID":"34297fe0-c74f-4d69-8847-86346bfbea4f","Type":"ContainerStarted","Data":"6af064ea906d91d300ba56c4e386475a92f721c282f358991d95da5ddbf88b74"} Oct 14 07:17:52 crc kubenswrapper[4870]: I1014 07:17:52.013776 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-656bcbd775-lp9f2" event={"ID":"d4236126-b907-4f39-a491-5006f2a9c301","Type":"ContainerStarted","Data":"1bbc052a864d4630c43ea25aa2907241e1508bcc09758b11abf2ff55d5409597"} Oct 14 07:17:52 crc kubenswrapper[4870]: I1014 07:17:52.013810 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-656bcbd775-lp9f2" event={"ID":"d4236126-b907-4f39-a491-5006f2a9c301","Type":"ContainerStarted","Data":"d5ae36cf0990130ddf9b3969d7eaf462a3dc88744969f66a65af136b1aa372b0"} Oct 14 07:17:52 crc kubenswrapper[4870]: E1014 07:17:52.017494 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/infra-operator@sha256:5cfb2ae1092445950b39dd59caa9a8c9367f42fb8353a8c3848d3bc729f24492\\\"\"" pod="openstack-operators/infra-operator-controller-manager-656bcbd775-lp9f2" podUID="d4236126-b907-4f39-a491-5006f2a9c301" Oct 14 07:17:52 crc kubenswrapper[4870]: I1014 07:17:52.035823 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-9c5c78d49-gjsj6" event={"ID":"d27f8663-102d-4565-b7d5-00db25a8388a","Type":"ContainerStarted","Data":"0625082843ab92a357a4aa63b720f29efd40d531011415896a763e95ecebddb4"} Oct 14 07:17:52 crc kubenswrapper[4870]: I1014 07:17:52.037996 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-7fb8c88b76-jzcb4"] Oct 14 07:17:52 crc kubenswrapper[4870]: I1014 07:17:52.049982 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-79d585cb66-67fkk" event={"ID":"586d846b-21c2-40f4-8da4-37392d812645","Type":"ContainerStarted","Data":"e071d7a404563445c78e609723016eaf6e278d92d7d8fc543bc18912011fd445"} Oct 14 07:17:52 crc kubenswrapper[4870]: I1014 07:17:52.050027 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-79d585cb66-67fkk" event={"ID":"586d846b-21c2-40f4-8da4-37392d812645","Type":"ContainerStarted","Data":"61e3e4e6acfbe078c2d4a87962c61e1ff6371548d5f1655a1e4250ba44881e44"} Oct 14 07:17:52 crc kubenswrapper[4870]: E1014 07:17:52.066274 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/neutron-operator@sha256:33652e75a03a058769019fe8d8c51585a6eeefef5e1ecb96f9965434117954f2\\\"\"" pod="openstack-operators/neutron-operator-controller-manager-79d585cb66-67fkk" podUID="586d846b-21c2-40f4-8da4-37392d812645" Oct 14 07:17:52 crc kubenswrapper[4870]: I1014 07:17:52.066901 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-68b6c87b68-dtn8j" event={"ID":"96effd40-d810-4b88-9dec-a6e98993778f","Type":"ContainerStarted","Data":"74fb49aba0145ef61683cca04ca24e58c8ecce9474feac291fa4127c61d603ce"} Oct 14 07:17:52 crc kubenswrapper[4870]: I1014 07:17:52.066948 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-68b6c87b68-dtn8j" event={"ID":"96effd40-d810-4b88-9dec-a6e98993778f","Type":"ContainerStarted","Data":"393084ff8bd63e877aa72367c016942ca2d7e41643991f03d71b7e5e75419210"} Oct 14 07:17:52 crc kubenswrapper[4870]: E1014 07:17:52.069895 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:d33c1f507e1f5b9a4bf226ad98917e92101ac66b36e19d35cbe04ae7014f6bff\\\"\"" pod="openstack-operators/placement-operator-controller-manager-68b6c87b68-dtn8j" podUID="96effd40-d810-4b88-9dec-a6e98993778f" Oct 14 07:17:52 crc kubenswrapper[4870]: I1014 07:17:52.087572 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-55b6b7c7b8-jbt9q" event={"ID":"5ad2a6ec-e2aa-45d9-b1a8-59ed908aaacd","Type":"ContainerStarted","Data":"121be7fdb48de0713e07d0d50d75b39c024ab34b5560b6e8c92fc9fc98b69286"} Oct 14 07:17:52 crc kubenswrapper[4870]: I1014 07:17:52.092657 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-69fdcfc5f5-2schz" event={"ID":"34891227-4e7a-43d5-8dc0-3f56b54bb9fd","Type":"ContainerStarted","Data":"058e57af6d64145c5b9dd24a7df4ea5faac423f3edfc86a7b35db4ef2ba3e937"} Oct 14 07:17:52 crc kubenswrapper[4870]: I1014 07:17:52.104718 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-f9fb45f8f-kvmkj" event={"ID":"23eadd14-852e-4683-bae5-3720a36d9407","Type":"ContainerStarted","Data":"75da8f8484a03bbda4202d28165ecdd6483da2da0a7ea1492ea636e82d716210"} Oct 14 07:17:52 crc kubenswrapper[4870]: I1014 07:17:52.110271 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5458f77c4-n4csp" event={"ID":"f47c02fb-6e94-4f5c-873f-ecaacd522121","Type":"ContainerStarted","Data":"25e1f80edb99216885fc256a03130d09e04229f63c37a191794bd87248198ba2"} Oct 14 07:17:52 crc kubenswrapper[4870]: I1014 07:17:52.115901 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-55b7d4484882chz" event={"ID":"3278d66b-0ec7-47b4-8499-1671fc13f4ee","Type":"ContainerStarted","Data":"855388c1c9a6d20eeedf50c5989289a7d49c89edc51916f44ac0e52a847d7aa6"} Oct 14 07:17:52 crc kubenswrapper[4870]: I1014 07:17:52.115974 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-55b7d4484882chz" event={"ID":"3278d66b-0ec7-47b4-8499-1671fc13f4ee","Type":"ContainerStarted","Data":"9b683453df0abbe44cc9d84e555c6c73cc85e23af4db1d2388386231d6edd11b"} Oct 14 07:17:52 crc kubenswrapper[4870]: I1014 07:17:52.118227 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-lgqbd" event={"ID":"de2a07da-ed81-4aac-bf11-da6b54ad095a","Type":"ContainerStarted","Data":"726f2d4023c1879e4f5364f2e1f0f15fab8b912210777dede6997b04a917a809"} Oct 14 07:17:52 crc kubenswrapper[4870]: I1014 07:17:52.120224 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-7b7fb68549-xnk22" event={"ID":"9f311a89-2483-42d6-9fe5-57b1900624d2","Type":"ContainerStarted","Data":"a65f823e19fec8febc7f818f093d6a98b6a5ea3da6b207dd6de2ecf958b6c0f2"} Oct 14 07:17:52 crc kubenswrapper[4870]: E1014 07:17:52.120311 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:a17fc270857869fd1efe5020b2a1cb8c2abbd838f08de88f3a6a59e8754ec351\\\"\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-55b7d4484882chz" podUID="3278d66b-0ec7-47b4-8499-1671fc13f4ee" Oct 14 07:17:52 crc kubenswrapper[4870]: I1014 07:17:52.123204 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-79df5fb58c-s54lf" event={"ID":"9916d05d-8aa7-4e71-9ea6-07f2847fc4e6","Type":"ContainerStarted","Data":"c07c3b096fcf9d6541148572481a5f58548f38e2e00773ba4ed3b755b116cb90"} Oct 14 07:17:52 crc kubenswrapper[4870]: I1014 07:17:52.123248 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-79df5fb58c-s54lf" event={"ID":"9916d05d-8aa7-4e71-9ea6-07f2847fc4e6","Type":"ContainerStarted","Data":"bb05ba4ae60491d208003c9ac95b1d34de4675e2979bcd7b1b82d283baefc20c"} Oct 14 07:17:52 crc kubenswrapper[4870]: E1014 07:17:52.124936 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-lgqbd" podUID="de2a07da-ed81-4aac-bf11-da6b54ad095a" Oct 14 07:17:52 crc kubenswrapper[4870]: E1014 07:17:52.125403 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:315e558023b41ac1aa215082096995a03810c5b42910a33b00427ffcac9c6a14\\\"\"" pod="openstack-operators/ovn-operator-controller-manager-79df5fb58c-s54lf" podUID="9916d05d-8aa7-4e71-9ea6-07f2847fc4e6" Oct 14 07:17:52 crc kubenswrapper[4870]: I1014 07:17:52.128266 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-858f76bbdd-9bdj8" event={"ID":"519142a1-852b-4410-a511-13e2a36aed77","Type":"ContainerStarted","Data":"8ac6d6b2cbd7fe3bc55ae423d32aa639f00f843db5654d096bc95c039cdfe0bd"} Oct 14 07:17:52 crc kubenswrapper[4870]: I1014 07:17:52.131818 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-5f67fbc655-szjkl" event={"ID":"4bc1fceb-2d98-415e-b34e-82c2d6e1430f","Type":"ContainerStarted","Data":"92a4531b54c9c50f804214960da28491829a6029e23b50950ec13afaef8d15f4"} Oct 14 07:17:52 crc kubenswrapper[4870]: I1014 07:17:52.157145 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-67cfc6749b-cblnh" event={"ID":"01663eb4-cb0e-41fe-92c9-97e908f4ddff","Type":"ContainerStarted","Data":"096abb6fa991707dff487d31022b7894f2f5d81cd0ce80cff111ed961b4f2d9b"} Oct 14 07:17:52 crc kubenswrapper[4870]: I1014 07:17:52.157206 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-67cfc6749b-cblnh" event={"ID":"01663eb4-cb0e-41fe-92c9-97e908f4ddff","Type":"ContainerStarted","Data":"6689480e5264b1270ba593a10d22fe954c7310694db0508351ba2e7f8f44bb4a"} Oct 14 07:17:52 crc kubenswrapper[4870]: E1014 07:17:52.168981 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:abe978f8da75223de5043cca50278ad4e28c8dd309883f502fe1e7a9998733b0\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-67cfc6749b-cblnh" podUID="01663eb4-cb0e-41fe-92c9-97e908f4ddff" Oct 14 07:17:52 crc kubenswrapper[4870]: I1014 07:17:52.172224 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-84b9b84486-cl7ft" event={"ID":"b4a4fd5f-360a-45ca-9684-74fe1c7631a6","Type":"ContainerStarted","Data":"0e6e43a29a422c84633449187b82742327cc93e5177dd529d507bb5584acc689"} Oct 14 07:17:52 crc kubenswrapper[4870]: I1014 07:17:52.175941 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-db6d7f97b-22qw9" event={"ID":"c2b45ced-e9c2-4d6f-ac5e-1960e5a9296a","Type":"ContainerStarted","Data":"12b5f79f9c319e698702995a459a89309b2bdbb579125517e87b7aaab7611265"} Oct 14 07:17:52 crc kubenswrapper[4870]: I1014 07:17:52.175981 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-db6d7f97b-22qw9" event={"ID":"c2b45ced-e9c2-4d6f-ac5e-1960e5a9296a","Type":"ContainerStarted","Data":"327890c1e4063e3fd22ad82f680fa7206753e2f8cb7b0582b7ecd81cca2ab972"} Oct 14 07:17:52 crc kubenswrapper[4870]: E1014 07:17:52.179691 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:4b4a17fe08ce00e375afaaec6a28835f5c1784f03d11c4558376ac04130f3a9e\\\"\"" pod="openstack-operators/swift-operator-controller-manager-db6d7f97b-22qw9" podUID="c2b45ced-e9c2-4d6f-ac5e-1960e5a9296a" Oct 14 07:17:53 crc kubenswrapper[4870]: I1014 07:17:53.194773 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-7fb8c88b76-jzcb4" event={"ID":"3645fdee-7690-4d37-a0f8-16ea440d9453","Type":"ContainerStarted","Data":"aa9ee78c52a3655b1244e29184fc74fcbcfd363ab1ae5a1c1d7279b7b510e1a3"} Oct 14 07:17:53 crc kubenswrapper[4870]: I1014 07:17:53.195181 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-7fb8c88b76-jzcb4" event={"ID":"3645fdee-7690-4d37-a0f8-16ea440d9453","Type":"ContainerStarted","Data":"463704bae2496288e1325422b567f4d8f10164b30f76819337c269d025d7aaa7"} Oct 14 07:17:53 crc kubenswrapper[4870]: E1014 07:17:53.197829 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/infra-operator@sha256:5cfb2ae1092445950b39dd59caa9a8c9367f42fb8353a8c3848d3bc729f24492\\\"\"" pod="openstack-operators/infra-operator-controller-manager-656bcbd775-lp9f2" podUID="d4236126-b907-4f39-a491-5006f2a9c301" Oct 14 07:17:53 crc kubenswrapper[4870]: E1014 07:17:53.197917 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:abe978f8da75223de5043cca50278ad4e28c8dd309883f502fe1e7a9998733b0\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-67cfc6749b-cblnh" podUID="01663eb4-cb0e-41fe-92c9-97e908f4ddff" Oct 14 07:17:53 crc kubenswrapper[4870]: E1014 07:17:53.197971 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/neutron-operator@sha256:33652e75a03a058769019fe8d8c51585a6eeefef5e1ecb96f9965434117954f2\\\"\"" pod="openstack-operators/neutron-operator-controller-manager-79d585cb66-67fkk" podUID="586d846b-21c2-40f4-8da4-37392d812645" Oct 14 07:17:53 crc kubenswrapper[4870]: E1014 07:17:53.198302 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:d33c1f507e1f5b9a4bf226ad98917e92101ac66b36e19d35cbe04ae7014f6bff\\\"\"" pod="openstack-operators/placement-operator-controller-manager-68b6c87b68-dtn8j" podUID="96effd40-d810-4b88-9dec-a6e98993778f" Oct 14 07:17:53 crc kubenswrapper[4870]: E1014 07:17:53.198360 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:4b4a17fe08ce00e375afaaec6a28835f5c1784f03d11c4558376ac04130f3a9e\\\"\"" pod="openstack-operators/swift-operator-controller-manager-db6d7f97b-22qw9" podUID="c2b45ced-e9c2-4d6f-ac5e-1960e5a9296a" Oct 14 07:17:53 crc kubenswrapper[4870]: E1014 07:17:53.199566 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:315e558023b41ac1aa215082096995a03810c5b42910a33b00427ffcac9c6a14\\\"\"" pod="openstack-operators/ovn-operator-controller-manager-79df5fb58c-s54lf" podUID="9916d05d-8aa7-4e71-9ea6-07f2847fc4e6" Oct 14 07:17:53 crc kubenswrapper[4870]: E1014 07:17:53.201641 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:a17fc270857869fd1efe5020b2a1cb8c2abbd838f08de88f3a6a59e8754ec351\\\"\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-55b7d4484882chz" podUID="3278d66b-0ec7-47b4-8499-1671fc13f4ee" Oct 14 07:17:53 crc kubenswrapper[4870]: E1014 07:17:53.201685 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-lgqbd" podUID="de2a07da-ed81-4aac-bf11-da6b54ad095a" Oct 14 07:17:53 crc kubenswrapper[4870]: I1014 07:17:53.951316 4870 patch_prober.go:28] interesting pod/machine-config-daemon-7tvc7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 07:17:53 crc kubenswrapper[4870]: I1014 07:17:53.951456 4870 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 07:18:00 crc kubenswrapper[4870]: I1014 07:18:00.269755 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-69fdcfc5f5-2schz" event={"ID":"34891227-4e7a-43d5-8dc0-3f56b54bb9fd","Type":"ContainerStarted","Data":"147c3e797835a6466d5bf483fb00c683a168ebf2ce97cc97a2a90d41f3742f04"} Oct 14 07:18:00 crc kubenswrapper[4870]: I1014 07:18:00.271808 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-85d5d9dd78-ld7xh" event={"ID":"634060b4-d317-415d-bc9e-aa3ed6503b4f","Type":"ContainerStarted","Data":"c89783a853f5b7e05a8de73f1f14dba5a5bb4e0ab6223551cb254b1b49fbc0f9"} Oct 14 07:18:00 crc kubenswrapper[4870]: I1014 07:18:00.274049 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-7fb8c88b76-jzcb4" event={"ID":"3645fdee-7690-4d37-a0f8-16ea440d9453","Type":"ContainerStarted","Data":"3e6fc90e881aaba12612e15bb0b701b92739c01985d33888c0b8e20c37b6378d"} Oct 14 07:18:00 crc kubenswrapper[4870]: I1014 07:18:00.276662 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-7fb8c88b76-jzcb4" Oct 14 07:18:00 crc kubenswrapper[4870]: I1014 07:18:00.280805 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-7fb8c88b76-jzcb4" Oct 14 07:18:00 crc kubenswrapper[4870]: I1014 07:18:00.281946 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-7f554bff7b-68d4q" event={"ID":"a1711137-ad44-4da0-ab90-5666fd111419","Type":"ContainerStarted","Data":"2205f5d226567d99495a63dca82f38f12e0fbf0d54aded002e5ee50660a30cd7"} Oct 14 07:18:00 crc kubenswrapper[4870]: I1014 07:18:00.290779 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5458f77c4-n4csp" event={"ID":"f47c02fb-6e94-4f5c-873f-ecaacd522121","Type":"ContainerStarted","Data":"8eb0056a9dfc4d1258e0aae0a4194e7446820c52fdff28788a09a84cddd234b6"} Oct 14 07:18:00 crc kubenswrapper[4870]: I1014 07:18:00.300650 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-658bdf4b74-76kw6" event={"ID":"36e6bbc0-f57d-4486-9120-d2cd5a5ce638","Type":"ContainerStarted","Data":"91e576d237ce14a2f268984518b528b6b02e4d5fb2f829c3c9487e2e5627889f"} Oct 14 07:18:00 crc kubenswrapper[4870]: I1014 07:18:00.304795 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-7b7fb68549-xnk22" event={"ID":"9f311a89-2483-42d6-9fe5-57b1900624d2","Type":"ContainerStarted","Data":"2c1b832901719faf3fd6f11818f5a58bfccf3474f90c651f74172166676c06b6"} Oct 14 07:18:00 crc kubenswrapper[4870]: I1014 07:18:00.312383 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-5f67fbc655-szjkl" event={"ID":"4bc1fceb-2d98-415e-b34e-82c2d6e1430f","Type":"ContainerStarted","Data":"a9425835a0247f16759155732f9102b72ee24e9ec5852fdc5a35d5f3b2177643"} Oct 14 07:18:00 crc kubenswrapper[4870]: I1014 07:18:00.313698 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-7ffbcb7588-g2fdg" event={"ID":"8ed9c494-ed5c-46d0-9434-1edbfcfba361","Type":"ContainerStarted","Data":"5ad5337f79708d38d426461d813087a6a1c39b6d831d1ddc9c6ba0997d3d6feb"} Oct 14 07:18:00 crc kubenswrapper[4870]: I1014 07:18:00.321640 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-858f76bbdd-9bdj8" event={"ID":"519142a1-852b-4410-a511-13e2a36aed77","Type":"ContainerStarted","Data":"ac469fd342c272c3a1ce0ab05413f555effa9658ed57bdf10b9cbfc45e4455f2"} Oct 14 07:18:00 crc kubenswrapper[4870]: I1014 07:18:00.329511 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-7fb8c88b76-jzcb4" podStartSLOduration=11.329497597 podStartE2EDuration="11.329497597s" podCreationTimestamp="2025-10-14 07:17:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:18:00.329281352 +0000 UTC m=+1016.026641723" watchObservedRunningTime="2025-10-14 07:18:00.329497597 +0000 UTC m=+1016.026857968" Oct 14 07:18:00 crc kubenswrapper[4870]: I1014 07:18:00.329970 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-9c5c78d49-gjsj6" event={"ID":"d27f8663-102d-4565-b7d5-00db25a8388a","Type":"ContainerStarted","Data":"33514b0f7974646fb2f3db13b2d582637a8308dd4e7bbd085cb636fb92690016"} Oct 14 07:18:00 crc kubenswrapper[4870]: I1014 07:18:00.346165 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-f9fb45f8f-kvmkj" event={"ID":"23eadd14-852e-4683-bae5-3720a36d9407","Type":"ContainerStarted","Data":"bda3078dcce889095819e51cf8de0b437c97e59620332431fd421afd62906930"} Oct 14 07:18:00 crc kubenswrapper[4870]: I1014 07:18:00.359668 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-5df598886f-zj4rk" event={"ID":"34297fe0-c74f-4d69-8847-86346bfbea4f","Type":"ContainerStarted","Data":"48bfc02836e60269f8b7bb2dd9942879960ff3df3d883a187e2dbb4239cd9b9a"} Oct 14 07:18:01 crc kubenswrapper[4870]: I1014 07:18:01.366988 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-f9fb45f8f-kvmkj" event={"ID":"23eadd14-852e-4683-bae5-3720a36d9407","Type":"ContainerStarted","Data":"10562a44e7bc81d8aefe3b67c21391a2c63f2a734601ff776e8858863d82c0b9"} Oct 14 07:18:01 crc kubenswrapper[4870]: I1014 07:18:01.367356 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-f9fb45f8f-kvmkj" Oct 14 07:18:01 crc kubenswrapper[4870]: I1014 07:18:01.369300 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-85d5d9dd78-ld7xh" event={"ID":"634060b4-d317-415d-bc9e-aa3ed6503b4f","Type":"ContainerStarted","Data":"e5c77bdc12996244acb771da232933833af6513fca5be0d8d4e88aa493308253"} Oct 14 07:18:01 crc kubenswrapper[4870]: I1014 07:18:01.369497 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-85d5d9dd78-ld7xh" Oct 14 07:18:01 crc kubenswrapper[4870]: I1014 07:18:01.371148 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-5df598886f-zj4rk" event={"ID":"34297fe0-c74f-4d69-8847-86346bfbea4f","Type":"ContainerStarted","Data":"ed5d82b9ee20300f9b746fffcf2c4ec5e4dc2287a81e508e959ecd8f1cb58df8"} Oct 14 07:18:01 crc kubenswrapper[4870]: I1014 07:18:01.371253 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-5df598886f-zj4rk" Oct 14 07:18:01 crc kubenswrapper[4870]: I1014 07:18:01.372810 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-7b7fb68549-xnk22" event={"ID":"9f311a89-2483-42d6-9fe5-57b1900624d2","Type":"ContainerStarted","Data":"ab1421e0a8ba667439afca9488bcab0339033ca4df9163de19a284d3e03bc7d0"} Oct 14 07:18:01 crc kubenswrapper[4870]: I1014 07:18:01.373059 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-7b7fb68549-xnk22" Oct 14 07:18:01 crc kubenswrapper[4870]: I1014 07:18:01.374626 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5458f77c4-n4csp" event={"ID":"f47c02fb-6e94-4f5c-873f-ecaacd522121","Type":"ContainerStarted","Data":"47a4f95bb8dcc866722972fddf8a4544e77e15112399fa032627368a7f752258"} Oct 14 07:18:01 crc kubenswrapper[4870]: I1014 07:18:01.374700 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-5458f77c4-n4csp" Oct 14 07:18:01 crc kubenswrapper[4870]: I1014 07:18:01.376331 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-84b9b84486-cl7ft" event={"ID":"b4a4fd5f-360a-45ca-9684-74fe1c7631a6","Type":"ContainerStarted","Data":"58adbb866339fd1202bf5ead8a6f40beffd3b726ee45ad6b5b55a9c2cdc6c0b6"} Oct 14 07:18:01 crc kubenswrapper[4870]: I1014 07:18:01.376463 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-84b9b84486-cl7ft" Oct 14 07:18:01 crc kubenswrapper[4870]: I1014 07:18:01.376543 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-84b9b84486-cl7ft" event={"ID":"b4a4fd5f-360a-45ca-9684-74fe1c7631a6","Type":"ContainerStarted","Data":"2db1a60ce300fe9806dda95e8256c27e8f6d640bdfe1558d1f3095358d9bfa13"} Oct 14 07:18:01 crc kubenswrapper[4870]: I1014 07:18:01.377850 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-9c5c78d49-gjsj6" event={"ID":"d27f8663-102d-4565-b7d5-00db25a8388a","Type":"ContainerStarted","Data":"0879736e22e65f90e86f02db6c704f50c6e47a009b475e3efe7035ab912ccf21"} Oct 14 07:18:01 crc kubenswrapper[4870]: I1014 07:18:01.377984 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-9c5c78d49-gjsj6" Oct 14 07:18:01 crc kubenswrapper[4870]: I1014 07:18:01.379372 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-55b6b7c7b8-jbt9q" event={"ID":"5ad2a6ec-e2aa-45d9-b1a8-59ed908aaacd","Type":"ContainerStarted","Data":"2a4d35726e0462eb871034c961e24fbe58e4e7ab0fbcf077dd76df9ebd6e0d0c"} Oct 14 07:18:01 crc kubenswrapper[4870]: I1014 07:18:01.379397 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-55b6b7c7b8-jbt9q" event={"ID":"5ad2a6ec-e2aa-45d9-b1a8-59ed908aaacd","Type":"ContainerStarted","Data":"b9774b4197393328ce28b85648a645ab542d2aa1abd2241ae0673b4015d4bc49"} Oct 14 07:18:01 crc kubenswrapper[4870]: I1014 07:18:01.379879 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-55b6b7c7b8-jbt9q" Oct 14 07:18:01 crc kubenswrapper[4870]: I1014 07:18:01.381317 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-7f554bff7b-68d4q" event={"ID":"a1711137-ad44-4da0-ab90-5666fd111419","Type":"ContainerStarted","Data":"e857ce932202019df7e03df523a821121b052e422bcdb5150b2e97c4016d8c30"} Oct 14 07:18:01 crc kubenswrapper[4870]: I1014 07:18:01.381473 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-7f554bff7b-68d4q" Oct 14 07:18:01 crc kubenswrapper[4870]: I1014 07:18:01.383560 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-5f67fbc655-szjkl" event={"ID":"4bc1fceb-2d98-415e-b34e-82c2d6e1430f","Type":"ContainerStarted","Data":"1ded58d0f9d262605f2960443335f057cb4cbe47c45db8400e948e4bd5616ec7"} Oct 14 07:18:01 crc kubenswrapper[4870]: I1014 07:18:01.383692 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-5f67fbc655-szjkl" Oct 14 07:18:01 crc kubenswrapper[4870]: I1014 07:18:01.385188 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-69fdcfc5f5-2schz" event={"ID":"34891227-4e7a-43d5-8dc0-3f56b54bb9fd","Type":"ContainerStarted","Data":"224c303e5d48c109ce870b9c43613601cba133db19e80ca0d820a43f1f273b2c"} Oct 14 07:18:01 crc kubenswrapper[4870]: I1014 07:18:01.385374 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-69fdcfc5f5-2schz" Oct 14 07:18:01 crc kubenswrapper[4870]: I1014 07:18:01.386770 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-7ffbcb7588-g2fdg" event={"ID":"8ed9c494-ed5c-46d0-9434-1edbfcfba361","Type":"ContainerStarted","Data":"a208a02d84062e0b7f1d96690844f81448980ad1ab32be9e94c48c19bd7140b6"} Oct 14 07:18:01 crc kubenswrapper[4870]: I1014 07:18:01.386869 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-7ffbcb7588-g2fdg" Oct 14 07:18:01 crc kubenswrapper[4870]: I1014 07:18:01.390848 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-858f76bbdd-9bdj8" event={"ID":"519142a1-852b-4410-a511-13e2a36aed77","Type":"ContainerStarted","Data":"303c34d4c9dda14ebe4c47f600df356ee751ec20e8c12d8f38a161dcc6756014"} Oct 14 07:18:01 crc kubenswrapper[4870]: I1014 07:18:01.391547 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-858f76bbdd-9bdj8" Oct 14 07:18:01 crc kubenswrapper[4870]: I1014 07:18:01.391710 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-f9fb45f8f-kvmkj" podStartSLOduration=3.9481718470000002 podStartE2EDuration="12.391687337s" podCreationTimestamp="2025-10-14 07:17:49 +0000 UTC" firstStartedPulling="2025-10-14 07:17:50.977657816 +0000 UTC m=+1006.675018187" lastFinishedPulling="2025-10-14 07:17:59.421173296 +0000 UTC m=+1015.118533677" observedRunningTime="2025-10-14 07:18:01.388175897 +0000 UTC m=+1017.085536278" watchObservedRunningTime="2025-10-14 07:18:01.391687337 +0000 UTC m=+1017.089047718" Oct 14 07:18:01 crc kubenswrapper[4870]: I1014 07:18:01.393240 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-658bdf4b74-76kw6" event={"ID":"36e6bbc0-f57d-4486-9120-d2cd5a5ce638","Type":"ContainerStarted","Data":"78ae5458613b27c35e862f5b96937b7a58494bf942ffc2515009da1e41870dfc"} Oct 14 07:18:01 crc kubenswrapper[4870]: I1014 07:18:01.408430 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-7b7fb68549-xnk22" podStartSLOduration=4.903078765 podStartE2EDuration="13.408408909s" podCreationTimestamp="2025-10-14 07:17:48 +0000 UTC" firstStartedPulling="2025-10-14 07:17:50.931247279 +0000 UTC m=+1006.628607650" lastFinishedPulling="2025-10-14 07:17:59.436577373 +0000 UTC m=+1015.133937794" observedRunningTime="2025-10-14 07:18:01.401450939 +0000 UTC m=+1017.098811320" watchObservedRunningTime="2025-10-14 07:18:01.408408909 +0000 UTC m=+1017.105769280" Oct 14 07:18:01 crc kubenswrapper[4870]: I1014 07:18:01.430234 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-84b9b84486-cl7ft" podStartSLOduration=4.936013244 podStartE2EDuration="13.430213571s" podCreationTimestamp="2025-10-14 07:17:48 +0000 UTC" firstStartedPulling="2025-10-14 07:17:50.93671675 +0000 UTC m=+1006.634077121" lastFinishedPulling="2025-10-14 07:17:59.430917057 +0000 UTC m=+1015.128277448" observedRunningTime="2025-10-14 07:18:01.419400062 +0000 UTC m=+1017.116760433" watchObservedRunningTime="2025-10-14 07:18:01.430213571 +0000 UTC m=+1017.127573942" Oct 14 07:18:01 crc kubenswrapper[4870]: I1014 07:18:01.440886 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-5458f77c4-n4csp" podStartSLOduration=4.380003807 podStartE2EDuration="12.440869106s" podCreationTimestamp="2025-10-14 07:17:49 +0000 UTC" firstStartedPulling="2025-10-14 07:17:51.36080949 +0000 UTC m=+1007.058169861" lastFinishedPulling="2025-10-14 07:17:59.421674779 +0000 UTC m=+1015.119035160" observedRunningTime="2025-10-14 07:18:01.437326095 +0000 UTC m=+1017.134686466" watchObservedRunningTime="2025-10-14 07:18:01.440869106 +0000 UTC m=+1017.138229477" Oct 14 07:18:01 crc kubenswrapper[4870]: I1014 07:18:01.470060 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-69fdcfc5f5-2schz" podStartSLOduration=4.04208515 podStartE2EDuration="12.470032728s" podCreationTimestamp="2025-10-14 07:17:49 +0000 UTC" firstStartedPulling="2025-10-14 07:17:50.985894579 +0000 UTC m=+1006.683254950" lastFinishedPulling="2025-10-14 07:17:59.413842157 +0000 UTC m=+1015.111202528" observedRunningTime="2025-10-14 07:18:01.461154059 +0000 UTC m=+1017.158514430" watchObservedRunningTime="2025-10-14 07:18:01.470032728 +0000 UTC m=+1017.167393109" Oct 14 07:18:01 crc kubenswrapper[4870]: I1014 07:18:01.508116 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-7ffbcb7588-g2fdg" podStartSLOduration=5.040589001 podStartE2EDuration="13.50809913s" podCreationTimestamp="2025-10-14 07:17:48 +0000 UTC" firstStartedPulling="2025-10-14 07:17:50.941342859 +0000 UTC m=+1006.638703220" lastFinishedPulling="2025-10-14 07:17:59.408852978 +0000 UTC m=+1015.106213349" observedRunningTime="2025-10-14 07:18:01.482789497 +0000 UTC m=+1017.180149868" watchObservedRunningTime="2025-10-14 07:18:01.50809913 +0000 UTC m=+1017.205459501" Oct 14 07:18:01 crc kubenswrapper[4870]: I1014 07:18:01.510725 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-9c5c78d49-gjsj6" podStartSLOduration=5.042445059 podStartE2EDuration="13.510716368s" podCreationTimestamp="2025-10-14 07:17:48 +0000 UTC" firstStartedPulling="2025-10-14 07:17:50.985883358 +0000 UTC m=+1006.683243729" lastFinishedPulling="2025-10-14 07:17:59.454154657 +0000 UTC m=+1015.151515038" observedRunningTime="2025-10-14 07:18:01.508239254 +0000 UTC m=+1017.205599625" watchObservedRunningTime="2025-10-14 07:18:01.510716368 +0000 UTC m=+1017.208076739" Oct 14 07:18:01 crc kubenswrapper[4870]: I1014 07:18:01.523218 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-55b6b7c7b8-jbt9q" podStartSLOduration=4.024468416 podStartE2EDuration="12.52319913s" podCreationTimestamp="2025-10-14 07:17:49 +0000 UTC" firstStartedPulling="2025-10-14 07:17:50.968656094 +0000 UTC m=+1006.666016455" lastFinishedPulling="2025-10-14 07:17:59.467386788 +0000 UTC m=+1015.164747169" observedRunningTime="2025-10-14 07:18:01.521890676 +0000 UTC m=+1017.219251057" watchObservedRunningTime="2025-10-14 07:18:01.52319913 +0000 UTC m=+1017.220559501" Oct 14 07:18:01 crc kubenswrapper[4870]: I1014 07:18:01.538691 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-5f67fbc655-szjkl" podStartSLOduration=4.045811087 podStartE2EDuration="12.538670469s" podCreationTimestamp="2025-10-14 07:17:49 +0000 UTC" firstStartedPulling="2025-10-14 07:17:50.94135454 +0000 UTC m=+1006.638714911" lastFinishedPulling="2025-10-14 07:17:59.434213912 +0000 UTC m=+1015.131574293" observedRunningTime="2025-10-14 07:18:01.535879767 +0000 UTC m=+1017.233240138" watchObservedRunningTime="2025-10-14 07:18:01.538670469 +0000 UTC m=+1017.236030850" Oct 14 07:18:01 crc kubenswrapper[4870]: I1014 07:18:01.558260 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-5df598886f-zj4rk" podStartSLOduration=4.08355797 podStartE2EDuration="12.558226053s" podCreationTimestamp="2025-10-14 07:17:49 +0000 UTC" firstStartedPulling="2025-10-14 07:17:50.982930052 +0000 UTC m=+1006.680290423" lastFinishedPulling="2025-10-14 07:17:59.457598115 +0000 UTC m=+1015.154958506" observedRunningTime="2025-10-14 07:18:01.553684126 +0000 UTC m=+1017.251044507" watchObservedRunningTime="2025-10-14 07:18:01.558226053 +0000 UTC m=+1017.255586424" Oct 14 07:18:01 crc kubenswrapper[4870]: I1014 07:18:01.569070 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-85d5d9dd78-ld7xh" podStartSLOduration=4.5363082630000005 podStartE2EDuration="13.569060383s" podCreationTimestamp="2025-10-14 07:17:48 +0000 UTC" firstStartedPulling="2025-10-14 07:17:50.336605909 +0000 UTC m=+1006.033966280" lastFinishedPulling="2025-10-14 07:17:59.369358029 +0000 UTC m=+1015.066718400" observedRunningTime="2025-10-14 07:18:01.567752399 +0000 UTC m=+1017.265112770" watchObservedRunningTime="2025-10-14 07:18:01.569060383 +0000 UTC m=+1017.266420744" Oct 14 07:18:01 crc kubenswrapper[4870]: I1014 07:18:01.585212 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-7f554bff7b-68d4q" podStartSLOduration=4.488000463 podStartE2EDuration="12.585199939s" podCreationTimestamp="2025-10-14 07:17:49 +0000 UTC" firstStartedPulling="2025-10-14 07:17:51.36041754 +0000 UTC m=+1007.057777911" lastFinishedPulling="2025-10-14 07:17:59.457616976 +0000 UTC m=+1015.154977387" observedRunningTime="2025-10-14 07:18:01.584366518 +0000 UTC m=+1017.281726889" watchObservedRunningTime="2025-10-14 07:18:01.585199939 +0000 UTC m=+1017.282560310" Oct 14 07:18:01 crc kubenswrapper[4870]: I1014 07:18:01.604836 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-658bdf4b74-76kw6" podStartSLOduration=4.703076414 podStartE2EDuration="13.604819785s" podCreationTimestamp="2025-10-14 07:17:48 +0000 UTC" firstStartedPulling="2025-10-14 07:17:50.53046856 +0000 UTC m=+1006.227828931" lastFinishedPulling="2025-10-14 07:17:59.432211921 +0000 UTC m=+1015.129572302" observedRunningTime="2025-10-14 07:18:01.601606192 +0000 UTC m=+1017.298966563" watchObservedRunningTime="2025-10-14 07:18:01.604819785 +0000 UTC m=+1017.302180156" Oct 14 07:18:01 crc kubenswrapper[4870]: I1014 07:18:01.621067 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-858f76bbdd-9bdj8" podStartSLOduration=5.108282689 podStartE2EDuration="13.621047284s" podCreationTimestamp="2025-10-14 07:17:48 +0000 UTC" firstStartedPulling="2025-10-14 07:17:50.941388291 +0000 UTC m=+1006.638748662" lastFinishedPulling="2025-10-14 07:17:59.454152846 +0000 UTC m=+1015.151513257" observedRunningTime="2025-10-14 07:18:01.617198615 +0000 UTC m=+1017.314558986" watchObservedRunningTime="2025-10-14 07:18:01.621047284 +0000 UTC m=+1017.318407655" Oct 14 07:18:02 crc kubenswrapper[4870]: I1014 07:18:02.402921 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-658bdf4b74-76kw6" Oct 14 07:18:06 crc kubenswrapper[4870]: I1014 07:18:06.446131 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-55b7d4484882chz" event={"ID":"3278d66b-0ec7-47b4-8499-1671fc13f4ee","Type":"ContainerStarted","Data":"11b7e177384bea50c2582e751e6009a07477a1891a7ebd8b1a05fa6a6483e9e0"} Oct 14 07:18:06 crc kubenswrapper[4870]: I1014 07:18:06.447033 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-55b7d4484882chz" Oct 14 07:18:06 crc kubenswrapper[4870]: I1014 07:18:06.487874 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-55b7d4484882chz" podStartSLOduration=3.103469068 podStartE2EDuration="17.487850329s" podCreationTimestamp="2025-10-14 07:17:49 +0000 UTC" firstStartedPulling="2025-10-14 07:17:51.457739811 +0000 UTC m=+1007.155100182" lastFinishedPulling="2025-10-14 07:18:05.842121032 +0000 UTC m=+1021.539481443" observedRunningTime="2025-10-14 07:18:06.483593259 +0000 UTC m=+1022.180953660" watchObservedRunningTime="2025-10-14 07:18:06.487850329 +0000 UTC m=+1022.185210710" Oct 14 07:18:08 crc kubenswrapper[4870]: I1014 07:18:08.475985 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-67cfc6749b-cblnh" event={"ID":"01663eb4-cb0e-41fe-92c9-97e908f4ddff","Type":"ContainerStarted","Data":"6a74c42b82f7142bf1dc9a517a8b822c25ba2deb9663a07337b6095069261766"} Oct 14 07:18:08 crc kubenswrapper[4870]: I1014 07:18:08.477630 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-67cfc6749b-cblnh" Oct 14 07:18:08 crc kubenswrapper[4870]: I1014 07:18:08.486510 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-db6d7f97b-22qw9" event={"ID":"c2b45ced-e9c2-4d6f-ac5e-1960e5a9296a","Type":"ContainerStarted","Data":"1876d3fdd23935346005bcf72fc4a30e7ca2902e7d94c1ccac0071482f36f9d8"} Oct 14 07:18:08 crc kubenswrapper[4870]: I1014 07:18:08.487164 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-db6d7f97b-22qw9" Oct 14 07:18:08 crc kubenswrapper[4870]: I1014 07:18:08.500359 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-67cfc6749b-cblnh" podStartSLOduration=2.857346809 podStartE2EDuration="19.500342833s" podCreationTimestamp="2025-10-14 07:17:49 +0000 UTC" firstStartedPulling="2025-10-14 07:17:51.399306293 +0000 UTC m=+1007.096666664" lastFinishedPulling="2025-10-14 07:18:08.042302317 +0000 UTC m=+1023.739662688" observedRunningTime="2025-10-14 07:18:08.497951361 +0000 UTC m=+1024.195311752" watchObservedRunningTime="2025-10-14 07:18:08.500342833 +0000 UTC m=+1024.197703204" Oct 14 07:18:08 crc kubenswrapper[4870]: I1014 07:18:08.515533 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-db6d7f97b-22qw9" podStartSLOduration=2.871386371 podStartE2EDuration="19.515512084s" podCreationTimestamp="2025-10-14 07:17:49 +0000 UTC" firstStartedPulling="2025-10-14 07:17:51.399830437 +0000 UTC m=+1007.097190808" lastFinishedPulling="2025-10-14 07:18:08.04395615 +0000 UTC m=+1023.741316521" observedRunningTime="2025-10-14 07:18:08.510620588 +0000 UTC m=+1024.207980959" watchObservedRunningTime="2025-10-14 07:18:08.515512084 +0000 UTC m=+1024.212872455" Oct 14 07:18:09 crc kubenswrapper[4870]: I1014 07:18:09.214325 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-658bdf4b74-76kw6" Oct 14 07:18:09 crc kubenswrapper[4870]: I1014 07:18:09.235257 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-7b7fb68549-xnk22" Oct 14 07:18:09 crc kubenswrapper[4870]: I1014 07:18:09.263118 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-85d5d9dd78-ld7xh" Oct 14 07:18:09 crc kubenswrapper[4870]: I1014 07:18:09.316830 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-84b9b84486-cl7ft" Oct 14 07:18:09 crc kubenswrapper[4870]: I1014 07:18:09.353381 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-858f76bbdd-9bdj8" Oct 14 07:18:09 crc kubenswrapper[4870]: I1014 07:18:09.383224 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-7ffbcb7588-g2fdg" Oct 14 07:18:09 crc kubenswrapper[4870]: I1014 07:18:09.456104 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-5f67fbc655-szjkl" Oct 14 07:18:09 crc kubenswrapper[4870]: I1014 07:18:09.473765 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-55b6b7c7b8-jbt9q" Oct 14 07:18:09 crc kubenswrapper[4870]: I1014 07:18:09.492804 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-f9fb45f8f-kvmkj" Oct 14 07:18:09 crc kubenswrapper[4870]: I1014 07:18:09.685488 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-5df598886f-zj4rk" Oct 14 07:18:09 crc kubenswrapper[4870]: I1014 07:18:09.716118 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-9c5c78d49-gjsj6" Oct 14 07:18:09 crc kubenswrapper[4870]: I1014 07:18:09.732944 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-69fdcfc5f5-2schz" Oct 14 07:18:09 crc kubenswrapper[4870]: I1014 07:18:09.803701 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-5458f77c4-n4csp" Oct 14 07:18:09 crc kubenswrapper[4870]: I1014 07:18:09.836353 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-7f554bff7b-68d4q" Oct 14 07:18:10 crc kubenswrapper[4870]: I1014 07:18:10.361042 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-55b7d4484882chz" Oct 14 07:18:11 crc kubenswrapper[4870]: I1014 07:18:11.525853 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-lgqbd" event={"ID":"de2a07da-ed81-4aac-bf11-da6b54ad095a","Type":"ContainerStarted","Data":"318f388a12bca3bb2bd43a552846186737f1ff70edaefc997162dfc11332ad74"} Oct 14 07:18:11 crc kubenswrapper[4870]: I1014 07:18:11.529367 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-79df5fb58c-s54lf" event={"ID":"9916d05d-8aa7-4e71-9ea6-07f2847fc4e6","Type":"ContainerStarted","Data":"d698025571747d848765d6d1060c0d04b663329075aba36b278c3db7296c559d"} Oct 14 07:18:11 crc kubenswrapper[4870]: I1014 07:18:11.529757 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-79df5fb58c-s54lf" Oct 14 07:18:11 crc kubenswrapper[4870]: I1014 07:18:11.530920 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-656bcbd775-lp9f2" event={"ID":"d4236126-b907-4f39-a491-5006f2a9c301","Type":"ContainerStarted","Data":"7f4c384873df2b6ffe1d4e6d7f94be4b67f2154cf528d592ab4d495e330eb0f8"} Oct 14 07:18:11 crc kubenswrapper[4870]: I1014 07:18:11.531314 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-656bcbd775-lp9f2" Oct 14 07:18:11 crc kubenswrapper[4870]: I1014 07:18:11.532546 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-79d585cb66-67fkk" event={"ID":"586d846b-21c2-40f4-8da4-37392d812645","Type":"ContainerStarted","Data":"7430bad286a5fa9b470fc7d7b87b35848fc6e6f18707c2bc90ea05b7b9d9537f"} Oct 14 07:18:11 crc kubenswrapper[4870]: I1014 07:18:11.532894 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-79d585cb66-67fkk" Oct 14 07:18:11 crc kubenswrapper[4870]: I1014 07:18:11.534068 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-68b6c87b68-dtn8j" event={"ID":"96effd40-d810-4b88-9dec-a6e98993778f","Type":"ContainerStarted","Data":"9dd5c635c225002c954f5f48b8db155d8d91476cebfc96617f5f608c8ea089b8"} Oct 14 07:18:11 crc kubenswrapper[4870]: I1014 07:18:11.534242 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-68b6c87b68-dtn8j" Oct 14 07:18:11 crc kubenswrapper[4870]: I1014 07:18:11.546309 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-lgqbd" podStartSLOduration=2.770353194 podStartE2EDuration="22.546286167s" podCreationTimestamp="2025-10-14 07:17:49 +0000 UTC" firstStartedPulling="2025-10-14 07:17:51.400581366 +0000 UTC m=+1007.097941737" lastFinishedPulling="2025-10-14 07:18:11.176514329 +0000 UTC m=+1026.873874710" observedRunningTime="2025-10-14 07:18:11.540767225 +0000 UTC m=+1027.238127606" watchObservedRunningTime="2025-10-14 07:18:11.546286167 +0000 UTC m=+1027.243646568" Oct 14 07:18:11 crc kubenswrapper[4870]: I1014 07:18:11.560427 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-68b6c87b68-dtn8j" podStartSLOduration=2.810363477 podStartE2EDuration="22.560406001s" podCreationTimestamp="2025-10-14 07:17:49 +0000 UTC" firstStartedPulling="2025-10-14 07:17:51.400143275 +0000 UTC m=+1007.097503646" lastFinishedPulling="2025-10-14 07:18:11.150185789 +0000 UTC m=+1026.847546170" observedRunningTime="2025-10-14 07:18:11.559247271 +0000 UTC m=+1027.256607652" watchObservedRunningTime="2025-10-14 07:18:11.560406001 +0000 UTC m=+1027.257766392" Oct 14 07:18:11 crc kubenswrapper[4870]: I1014 07:18:11.587381 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-656bcbd775-lp9f2" podStartSLOduration=3.834646723 podStartE2EDuration="23.587357036s" podCreationTimestamp="2025-10-14 07:17:48 +0000 UTC" firstStartedPulling="2025-10-14 07:17:51.423636051 +0000 UTC m=+1007.120996422" lastFinishedPulling="2025-10-14 07:18:11.176346364 +0000 UTC m=+1026.873706735" observedRunningTime="2025-10-14 07:18:11.581739761 +0000 UTC m=+1027.279100172" watchObservedRunningTime="2025-10-14 07:18:11.587357036 +0000 UTC m=+1027.284717437" Oct 14 07:18:11 crc kubenswrapper[4870]: I1014 07:18:11.610910 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-79d585cb66-67fkk" podStartSLOduration=2.889092507 podStartE2EDuration="22.610889313s" podCreationTimestamp="2025-10-14 07:17:49 +0000 UTC" firstStartedPulling="2025-10-14 07:17:51.428411214 +0000 UTC m=+1007.125771585" lastFinishedPulling="2025-10-14 07:18:11.15020802 +0000 UTC m=+1026.847568391" observedRunningTime="2025-10-14 07:18:11.604859727 +0000 UTC m=+1027.302220108" watchObservedRunningTime="2025-10-14 07:18:11.610889313 +0000 UTC m=+1027.308249684" Oct 14 07:18:11 crc kubenswrapper[4870]: I1014 07:18:11.625779 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-79df5fb58c-s54lf" podStartSLOduration=2.925641719 podStartE2EDuration="22.625761196s" podCreationTimestamp="2025-10-14 07:17:49 +0000 UTC" firstStartedPulling="2025-10-14 07:17:51.450059112 +0000 UTC m=+1007.147419483" lastFinishedPulling="2025-10-14 07:18:11.150178589 +0000 UTC m=+1026.847538960" observedRunningTime="2025-10-14 07:18:11.622298977 +0000 UTC m=+1027.319659358" watchObservedRunningTime="2025-10-14 07:18:11.625761196 +0000 UTC m=+1027.323121567" Oct 14 07:18:19 crc kubenswrapper[4870]: I1014 07:18:19.674972 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-79d585cb66-67fkk" Oct 14 07:18:19 crc kubenswrapper[4870]: I1014 07:18:19.778961 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-db6d7f97b-22qw9" Oct 14 07:18:19 crc kubenswrapper[4870]: I1014 07:18:19.785598 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-67cfc6749b-cblnh" Oct 14 07:18:19 crc kubenswrapper[4870]: I1014 07:18:19.869182 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-79df5fb58c-s54lf" Oct 14 07:18:19 crc kubenswrapper[4870]: I1014 07:18:19.949105 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-68b6c87b68-dtn8j" Oct 14 07:18:20 crc kubenswrapper[4870]: I1014 07:18:20.016545 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-656bcbd775-lp9f2" Oct 14 07:18:23 crc kubenswrapper[4870]: I1014 07:18:23.950425 4870 patch_prober.go:28] interesting pod/machine-config-daemon-7tvc7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 07:18:23 crc kubenswrapper[4870]: I1014 07:18:23.950775 4870 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 07:18:33 crc kubenswrapper[4870]: I1014 07:18:33.243791 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5d487d97d7-hd6gw"] Oct 14 07:18:33 crc kubenswrapper[4870]: I1014 07:18:33.245764 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d487d97d7-hd6gw" Oct 14 07:18:33 crc kubenswrapper[4870]: I1014 07:18:33.250482 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-x4b7p" Oct 14 07:18:33 crc kubenswrapper[4870]: I1014 07:18:33.251098 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Oct 14 07:18:33 crc kubenswrapper[4870]: I1014 07:18:33.251532 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Oct 14 07:18:33 crc kubenswrapper[4870]: I1014 07:18:33.255574 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5d487d97d7-hd6gw"] Oct 14 07:18:33 crc kubenswrapper[4870]: I1014 07:18:33.259057 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Oct 14 07:18:33 crc kubenswrapper[4870]: I1014 07:18:33.281362 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-57pz4\" (UniqueName: \"kubernetes.io/projected/84d82363-9888-49ac-8e1d-5f98b834a7e1-kube-api-access-57pz4\") pod \"dnsmasq-dns-5d487d97d7-hd6gw\" (UID: \"84d82363-9888-49ac-8e1d-5f98b834a7e1\") " pod="openstack/dnsmasq-dns-5d487d97d7-hd6gw" Oct 14 07:18:33 crc kubenswrapper[4870]: I1014 07:18:33.281425 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/84d82363-9888-49ac-8e1d-5f98b834a7e1-config\") pod \"dnsmasq-dns-5d487d97d7-hd6gw\" (UID: \"84d82363-9888-49ac-8e1d-5f98b834a7e1\") " pod="openstack/dnsmasq-dns-5d487d97d7-hd6gw" Oct 14 07:18:33 crc kubenswrapper[4870]: I1014 07:18:33.304712 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6948694bd9-5r72k"] Oct 14 07:18:33 crc kubenswrapper[4870]: I1014 07:18:33.306321 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6948694bd9-5r72k" Oct 14 07:18:33 crc kubenswrapper[4870]: I1014 07:18:33.308053 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Oct 14 07:18:33 crc kubenswrapper[4870]: I1014 07:18:33.320707 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6948694bd9-5r72k"] Oct 14 07:18:33 crc kubenswrapper[4870]: I1014 07:18:33.382297 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/84d82363-9888-49ac-8e1d-5f98b834a7e1-config\") pod \"dnsmasq-dns-5d487d97d7-hd6gw\" (UID: \"84d82363-9888-49ac-8e1d-5f98b834a7e1\") " pod="openstack/dnsmasq-dns-5d487d97d7-hd6gw" Oct 14 07:18:33 crc kubenswrapper[4870]: I1014 07:18:33.382362 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5vllq\" (UniqueName: \"kubernetes.io/projected/f92a9598-06e8-4b09-9504-20d07f4f6281-kube-api-access-5vllq\") pod \"dnsmasq-dns-6948694bd9-5r72k\" (UID: \"f92a9598-06e8-4b09-9504-20d07f4f6281\") " pod="openstack/dnsmasq-dns-6948694bd9-5r72k" Oct 14 07:18:33 crc kubenswrapper[4870]: I1014 07:18:33.382423 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f92a9598-06e8-4b09-9504-20d07f4f6281-dns-svc\") pod \"dnsmasq-dns-6948694bd9-5r72k\" (UID: \"f92a9598-06e8-4b09-9504-20d07f4f6281\") " pod="openstack/dnsmasq-dns-6948694bd9-5r72k" Oct 14 07:18:33 crc kubenswrapper[4870]: I1014 07:18:33.382473 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-57pz4\" (UniqueName: \"kubernetes.io/projected/84d82363-9888-49ac-8e1d-5f98b834a7e1-kube-api-access-57pz4\") pod \"dnsmasq-dns-5d487d97d7-hd6gw\" (UID: \"84d82363-9888-49ac-8e1d-5f98b834a7e1\") " pod="openstack/dnsmasq-dns-5d487d97d7-hd6gw" Oct 14 07:18:33 crc kubenswrapper[4870]: I1014 07:18:33.382508 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f92a9598-06e8-4b09-9504-20d07f4f6281-config\") pod \"dnsmasq-dns-6948694bd9-5r72k\" (UID: \"f92a9598-06e8-4b09-9504-20d07f4f6281\") " pod="openstack/dnsmasq-dns-6948694bd9-5r72k" Oct 14 07:18:33 crc kubenswrapper[4870]: I1014 07:18:33.383516 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/84d82363-9888-49ac-8e1d-5f98b834a7e1-config\") pod \"dnsmasq-dns-5d487d97d7-hd6gw\" (UID: \"84d82363-9888-49ac-8e1d-5f98b834a7e1\") " pod="openstack/dnsmasq-dns-5d487d97d7-hd6gw" Oct 14 07:18:33 crc kubenswrapper[4870]: I1014 07:18:33.402419 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-57pz4\" (UniqueName: \"kubernetes.io/projected/84d82363-9888-49ac-8e1d-5f98b834a7e1-kube-api-access-57pz4\") pod \"dnsmasq-dns-5d487d97d7-hd6gw\" (UID: \"84d82363-9888-49ac-8e1d-5f98b834a7e1\") " pod="openstack/dnsmasq-dns-5d487d97d7-hd6gw" Oct 14 07:18:33 crc kubenswrapper[4870]: I1014 07:18:33.483928 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f92a9598-06e8-4b09-9504-20d07f4f6281-dns-svc\") pod \"dnsmasq-dns-6948694bd9-5r72k\" (UID: \"f92a9598-06e8-4b09-9504-20d07f4f6281\") " pod="openstack/dnsmasq-dns-6948694bd9-5r72k" Oct 14 07:18:33 crc kubenswrapper[4870]: I1014 07:18:33.484004 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f92a9598-06e8-4b09-9504-20d07f4f6281-config\") pod \"dnsmasq-dns-6948694bd9-5r72k\" (UID: \"f92a9598-06e8-4b09-9504-20d07f4f6281\") " pod="openstack/dnsmasq-dns-6948694bd9-5r72k" Oct 14 07:18:33 crc kubenswrapper[4870]: I1014 07:18:33.484037 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5vllq\" (UniqueName: \"kubernetes.io/projected/f92a9598-06e8-4b09-9504-20d07f4f6281-kube-api-access-5vllq\") pod \"dnsmasq-dns-6948694bd9-5r72k\" (UID: \"f92a9598-06e8-4b09-9504-20d07f4f6281\") " pod="openstack/dnsmasq-dns-6948694bd9-5r72k" Oct 14 07:18:33 crc kubenswrapper[4870]: I1014 07:18:33.484805 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f92a9598-06e8-4b09-9504-20d07f4f6281-dns-svc\") pod \"dnsmasq-dns-6948694bd9-5r72k\" (UID: \"f92a9598-06e8-4b09-9504-20d07f4f6281\") " pod="openstack/dnsmasq-dns-6948694bd9-5r72k" Oct 14 07:18:33 crc kubenswrapper[4870]: I1014 07:18:33.484965 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f92a9598-06e8-4b09-9504-20d07f4f6281-config\") pod \"dnsmasq-dns-6948694bd9-5r72k\" (UID: \"f92a9598-06e8-4b09-9504-20d07f4f6281\") " pod="openstack/dnsmasq-dns-6948694bd9-5r72k" Oct 14 07:18:33 crc kubenswrapper[4870]: I1014 07:18:33.505123 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5vllq\" (UniqueName: \"kubernetes.io/projected/f92a9598-06e8-4b09-9504-20d07f4f6281-kube-api-access-5vllq\") pod \"dnsmasq-dns-6948694bd9-5r72k\" (UID: \"f92a9598-06e8-4b09-9504-20d07f4f6281\") " pod="openstack/dnsmasq-dns-6948694bd9-5r72k" Oct 14 07:18:33 crc kubenswrapper[4870]: I1014 07:18:33.571790 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d487d97d7-hd6gw" Oct 14 07:18:33 crc kubenswrapper[4870]: I1014 07:18:33.623169 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6948694bd9-5r72k" Oct 14 07:18:34 crc kubenswrapper[4870]: I1014 07:18:34.048351 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5d487d97d7-hd6gw"] Oct 14 07:18:34 crc kubenswrapper[4870]: I1014 07:18:34.168287 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6948694bd9-5r72k"] Oct 14 07:18:34 crc kubenswrapper[4870]: W1014 07:18:34.178178 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf92a9598_06e8_4b09_9504_20d07f4f6281.slice/crio-e46f6f733bc263a47b96e5ddba18087043c08e13d14474feff6bf718f00ed70c WatchSource:0}: Error finding container e46f6f733bc263a47b96e5ddba18087043c08e13d14474feff6bf718f00ed70c: Status 404 returned error can't find the container with id e46f6f733bc263a47b96e5ddba18087043c08e13d14474feff6bf718f00ed70c Oct 14 07:18:34 crc kubenswrapper[4870]: I1014 07:18:34.789432 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d487d97d7-hd6gw" event={"ID":"84d82363-9888-49ac-8e1d-5f98b834a7e1","Type":"ContainerStarted","Data":"a0379a48d121338e28ce393309fc934509f49a899fc7d9ca72f56b49a92f65a1"} Oct 14 07:18:34 crc kubenswrapper[4870]: I1014 07:18:34.791130 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6948694bd9-5r72k" event={"ID":"f92a9598-06e8-4b09-9504-20d07f4f6281","Type":"ContainerStarted","Data":"e46f6f733bc263a47b96e5ddba18087043c08e13d14474feff6bf718f00ed70c"} Oct 14 07:18:35 crc kubenswrapper[4870]: I1014 07:18:35.845142 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5d487d97d7-hd6gw"] Oct 14 07:18:35 crc kubenswrapper[4870]: I1014 07:18:35.886117 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-86f694bf-k7p29"] Oct 14 07:18:35 crc kubenswrapper[4870]: I1014 07:18:35.887590 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86f694bf-k7p29" Oct 14 07:18:35 crc kubenswrapper[4870]: I1014 07:18:35.906526 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-86f694bf-k7p29"] Oct 14 07:18:36 crc kubenswrapper[4870]: I1014 07:18:36.023630 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lfz64\" (UniqueName: \"kubernetes.io/projected/86d52d2f-b41b-4c56-af26-b32e8caaf225-kube-api-access-lfz64\") pod \"dnsmasq-dns-86f694bf-k7p29\" (UID: \"86d52d2f-b41b-4c56-af26-b32e8caaf225\") " pod="openstack/dnsmasq-dns-86f694bf-k7p29" Oct 14 07:18:36 crc kubenswrapper[4870]: I1014 07:18:36.023691 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/86d52d2f-b41b-4c56-af26-b32e8caaf225-dns-svc\") pod \"dnsmasq-dns-86f694bf-k7p29\" (UID: \"86d52d2f-b41b-4c56-af26-b32e8caaf225\") " pod="openstack/dnsmasq-dns-86f694bf-k7p29" Oct 14 07:18:36 crc kubenswrapper[4870]: I1014 07:18:36.023735 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/86d52d2f-b41b-4c56-af26-b32e8caaf225-config\") pod \"dnsmasq-dns-86f694bf-k7p29\" (UID: \"86d52d2f-b41b-4c56-af26-b32e8caaf225\") " pod="openstack/dnsmasq-dns-86f694bf-k7p29" Oct 14 07:18:36 crc kubenswrapper[4870]: I1014 07:18:36.130017 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/86d52d2f-b41b-4c56-af26-b32e8caaf225-dns-svc\") pod \"dnsmasq-dns-86f694bf-k7p29\" (UID: \"86d52d2f-b41b-4c56-af26-b32e8caaf225\") " pod="openstack/dnsmasq-dns-86f694bf-k7p29" Oct 14 07:18:36 crc kubenswrapper[4870]: I1014 07:18:36.130094 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/86d52d2f-b41b-4c56-af26-b32e8caaf225-config\") pod \"dnsmasq-dns-86f694bf-k7p29\" (UID: \"86d52d2f-b41b-4c56-af26-b32e8caaf225\") " pod="openstack/dnsmasq-dns-86f694bf-k7p29" Oct 14 07:18:36 crc kubenswrapper[4870]: I1014 07:18:36.130155 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lfz64\" (UniqueName: \"kubernetes.io/projected/86d52d2f-b41b-4c56-af26-b32e8caaf225-kube-api-access-lfz64\") pod \"dnsmasq-dns-86f694bf-k7p29\" (UID: \"86d52d2f-b41b-4c56-af26-b32e8caaf225\") " pod="openstack/dnsmasq-dns-86f694bf-k7p29" Oct 14 07:18:36 crc kubenswrapper[4870]: I1014 07:18:36.130806 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/86d52d2f-b41b-4c56-af26-b32e8caaf225-dns-svc\") pod \"dnsmasq-dns-86f694bf-k7p29\" (UID: \"86d52d2f-b41b-4c56-af26-b32e8caaf225\") " pod="openstack/dnsmasq-dns-86f694bf-k7p29" Oct 14 07:18:36 crc kubenswrapper[4870]: I1014 07:18:36.131217 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/86d52d2f-b41b-4c56-af26-b32e8caaf225-config\") pod \"dnsmasq-dns-86f694bf-k7p29\" (UID: \"86d52d2f-b41b-4c56-af26-b32e8caaf225\") " pod="openstack/dnsmasq-dns-86f694bf-k7p29" Oct 14 07:18:36 crc kubenswrapper[4870]: I1014 07:18:36.155906 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lfz64\" (UniqueName: \"kubernetes.io/projected/86d52d2f-b41b-4c56-af26-b32e8caaf225-kube-api-access-lfz64\") pod \"dnsmasq-dns-86f694bf-k7p29\" (UID: \"86d52d2f-b41b-4c56-af26-b32e8caaf225\") " pod="openstack/dnsmasq-dns-86f694bf-k7p29" Oct 14 07:18:36 crc kubenswrapper[4870]: I1014 07:18:36.187582 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6948694bd9-5r72k"] Oct 14 07:18:36 crc kubenswrapper[4870]: I1014 07:18:36.208035 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86f694bf-k7p29" Oct 14 07:18:36 crc kubenswrapper[4870]: I1014 07:18:36.209155 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7869c47d6c-44h7s"] Oct 14 07:18:36 crc kubenswrapper[4870]: I1014 07:18:36.210246 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7869c47d6c-44h7s" Oct 14 07:18:36 crc kubenswrapper[4870]: I1014 07:18:36.231617 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/11aab367-2351-4568-9a45-43c79d419bca-dns-svc\") pod \"dnsmasq-dns-7869c47d6c-44h7s\" (UID: \"11aab367-2351-4568-9a45-43c79d419bca\") " pod="openstack/dnsmasq-dns-7869c47d6c-44h7s" Oct 14 07:18:36 crc kubenswrapper[4870]: I1014 07:18:36.231687 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/11aab367-2351-4568-9a45-43c79d419bca-config\") pod \"dnsmasq-dns-7869c47d6c-44h7s\" (UID: \"11aab367-2351-4568-9a45-43c79d419bca\") " pod="openstack/dnsmasq-dns-7869c47d6c-44h7s" Oct 14 07:18:36 crc kubenswrapper[4870]: I1014 07:18:36.231790 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-psk44\" (UniqueName: \"kubernetes.io/projected/11aab367-2351-4568-9a45-43c79d419bca-kube-api-access-psk44\") pod \"dnsmasq-dns-7869c47d6c-44h7s\" (UID: \"11aab367-2351-4568-9a45-43c79d419bca\") " pod="openstack/dnsmasq-dns-7869c47d6c-44h7s" Oct 14 07:18:36 crc kubenswrapper[4870]: I1014 07:18:36.237847 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7869c47d6c-44h7s"] Oct 14 07:18:36 crc kubenswrapper[4870]: I1014 07:18:36.336454 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/11aab367-2351-4568-9a45-43c79d419bca-dns-svc\") pod \"dnsmasq-dns-7869c47d6c-44h7s\" (UID: \"11aab367-2351-4568-9a45-43c79d419bca\") " pod="openstack/dnsmasq-dns-7869c47d6c-44h7s" Oct 14 07:18:36 crc kubenswrapper[4870]: I1014 07:18:36.336793 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/11aab367-2351-4568-9a45-43c79d419bca-config\") pod \"dnsmasq-dns-7869c47d6c-44h7s\" (UID: \"11aab367-2351-4568-9a45-43c79d419bca\") " pod="openstack/dnsmasq-dns-7869c47d6c-44h7s" Oct 14 07:18:36 crc kubenswrapper[4870]: I1014 07:18:36.336834 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-psk44\" (UniqueName: \"kubernetes.io/projected/11aab367-2351-4568-9a45-43c79d419bca-kube-api-access-psk44\") pod \"dnsmasq-dns-7869c47d6c-44h7s\" (UID: \"11aab367-2351-4568-9a45-43c79d419bca\") " pod="openstack/dnsmasq-dns-7869c47d6c-44h7s" Oct 14 07:18:36 crc kubenswrapper[4870]: I1014 07:18:36.337409 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/11aab367-2351-4568-9a45-43c79d419bca-dns-svc\") pod \"dnsmasq-dns-7869c47d6c-44h7s\" (UID: \"11aab367-2351-4568-9a45-43c79d419bca\") " pod="openstack/dnsmasq-dns-7869c47d6c-44h7s" Oct 14 07:18:36 crc kubenswrapper[4870]: I1014 07:18:36.339908 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/11aab367-2351-4568-9a45-43c79d419bca-config\") pod \"dnsmasq-dns-7869c47d6c-44h7s\" (UID: \"11aab367-2351-4568-9a45-43c79d419bca\") " pod="openstack/dnsmasq-dns-7869c47d6c-44h7s" Oct 14 07:18:36 crc kubenswrapper[4870]: I1014 07:18:36.355693 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-psk44\" (UniqueName: \"kubernetes.io/projected/11aab367-2351-4568-9a45-43c79d419bca-kube-api-access-psk44\") pod \"dnsmasq-dns-7869c47d6c-44h7s\" (UID: \"11aab367-2351-4568-9a45-43c79d419bca\") " pod="openstack/dnsmasq-dns-7869c47d6c-44h7s" Oct 14 07:18:36 crc kubenswrapper[4870]: I1014 07:18:36.541624 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7869c47d6c-44h7s" Oct 14 07:18:36 crc kubenswrapper[4870]: I1014 07:18:36.740691 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-86f694bf-k7p29"] Oct 14 07:18:36 crc kubenswrapper[4870]: W1014 07:18:36.775857 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod86d52d2f_b41b_4c56_af26_b32e8caaf225.slice/crio-484a8f4e1c510fa9732ac80d023636d9cd97ecc2ee418590f9c7d2d461321d4e WatchSource:0}: Error finding container 484a8f4e1c510fa9732ac80d023636d9cd97ecc2ee418590f9c7d2d461321d4e: Status 404 returned error can't find the container with id 484a8f4e1c510fa9732ac80d023636d9cd97ecc2ee418590f9c7d2d461321d4e Oct 14 07:18:36 crc kubenswrapper[4870]: I1014 07:18:36.830411 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86f694bf-k7p29" event={"ID":"86d52d2f-b41b-4c56-af26-b32e8caaf225","Type":"ContainerStarted","Data":"484a8f4e1c510fa9732ac80d023636d9cd97ecc2ee418590f9c7d2d461321d4e"} Oct 14 07:18:37 crc kubenswrapper[4870]: I1014 07:18:37.030776 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7869c47d6c-44h7s"] Oct 14 07:18:37 crc kubenswrapper[4870]: I1014 07:18:37.104657 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Oct 14 07:18:37 crc kubenswrapper[4870]: I1014 07:18:37.107953 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 14 07:18:37 crc kubenswrapper[4870]: I1014 07:18:37.110696 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Oct 14 07:18:37 crc kubenswrapper[4870]: I1014 07:18:37.110714 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-b2lr4" Oct 14 07:18:37 crc kubenswrapper[4870]: I1014 07:18:37.112231 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 14 07:18:37 crc kubenswrapper[4870]: I1014 07:18:37.112675 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Oct 14 07:18:37 crc kubenswrapper[4870]: I1014 07:18:37.112724 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Oct 14 07:18:37 crc kubenswrapper[4870]: I1014 07:18:37.112734 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Oct 14 07:18:37 crc kubenswrapper[4870]: I1014 07:18:37.112808 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Oct 14 07:18:37 crc kubenswrapper[4870]: I1014 07:18:37.112730 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Oct 14 07:18:37 crc kubenswrapper[4870]: I1014 07:18:37.147403 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-server-0\" (UID: \"da8a9731-f917-462f-9932-b37b6abb9a64\") " pod="openstack/rabbitmq-server-0" Oct 14 07:18:37 crc kubenswrapper[4870]: I1014 07:18:37.147532 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/da8a9731-f917-462f-9932-b37b6abb9a64-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"da8a9731-f917-462f-9932-b37b6abb9a64\") " pod="openstack/rabbitmq-server-0" Oct 14 07:18:37 crc kubenswrapper[4870]: I1014 07:18:37.147556 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/da8a9731-f917-462f-9932-b37b6abb9a64-server-conf\") pod \"rabbitmq-server-0\" (UID: \"da8a9731-f917-462f-9932-b37b6abb9a64\") " pod="openstack/rabbitmq-server-0" Oct 14 07:18:37 crc kubenswrapper[4870]: I1014 07:18:37.147607 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/da8a9731-f917-462f-9932-b37b6abb9a64-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"da8a9731-f917-462f-9932-b37b6abb9a64\") " pod="openstack/rabbitmq-server-0" Oct 14 07:18:37 crc kubenswrapper[4870]: I1014 07:18:37.147697 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/da8a9731-f917-462f-9932-b37b6abb9a64-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"da8a9731-f917-462f-9932-b37b6abb9a64\") " pod="openstack/rabbitmq-server-0" Oct 14 07:18:37 crc kubenswrapper[4870]: I1014 07:18:37.147729 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/da8a9731-f917-462f-9932-b37b6abb9a64-config-data\") pod \"rabbitmq-server-0\" (UID: \"da8a9731-f917-462f-9932-b37b6abb9a64\") " pod="openstack/rabbitmq-server-0" Oct 14 07:18:37 crc kubenswrapper[4870]: I1014 07:18:37.147785 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/da8a9731-f917-462f-9932-b37b6abb9a64-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"da8a9731-f917-462f-9932-b37b6abb9a64\") " pod="openstack/rabbitmq-server-0" Oct 14 07:18:37 crc kubenswrapper[4870]: I1014 07:18:37.147807 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/da8a9731-f917-462f-9932-b37b6abb9a64-pod-info\") pod \"rabbitmq-server-0\" (UID: \"da8a9731-f917-462f-9932-b37b6abb9a64\") " pod="openstack/rabbitmq-server-0" Oct 14 07:18:37 crc kubenswrapper[4870]: I1014 07:18:37.147848 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nl2bs\" (UniqueName: \"kubernetes.io/projected/da8a9731-f917-462f-9932-b37b6abb9a64-kube-api-access-nl2bs\") pod \"rabbitmq-server-0\" (UID: \"da8a9731-f917-462f-9932-b37b6abb9a64\") " pod="openstack/rabbitmq-server-0" Oct 14 07:18:37 crc kubenswrapper[4870]: I1014 07:18:37.147911 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/da8a9731-f917-462f-9932-b37b6abb9a64-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"da8a9731-f917-462f-9932-b37b6abb9a64\") " pod="openstack/rabbitmq-server-0" Oct 14 07:18:37 crc kubenswrapper[4870]: I1014 07:18:37.147957 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/da8a9731-f917-462f-9932-b37b6abb9a64-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"da8a9731-f917-462f-9932-b37b6abb9a64\") " pod="openstack/rabbitmq-server-0" Oct 14 07:18:37 crc kubenswrapper[4870]: I1014 07:18:37.248877 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/da8a9731-f917-462f-9932-b37b6abb9a64-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"da8a9731-f917-462f-9932-b37b6abb9a64\") " pod="openstack/rabbitmq-server-0" Oct 14 07:18:37 crc kubenswrapper[4870]: I1014 07:18:37.248948 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/da8a9731-f917-462f-9932-b37b6abb9a64-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"da8a9731-f917-462f-9932-b37b6abb9a64\") " pod="openstack/rabbitmq-server-0" Oct 14 07:18:37 crc kubenswrapper[4870]: I1014 07:18:37.248989 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-server-0\" (UID: \"da8a9731-f917-462f-9932-b37b6abb9a64\") " pod="openstack/rabbitmq-server-0" Oct 14 07:18:37 crc kubenswrapper[4870]: I1014 07:18:37.249023 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/da8a9731-f917-462f-9932-b37b6abb9a64-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"da8a9731-f917-462f-9932-b37b6abb9a64\") " pod="openstack/rabbitmq-server-0" Oct 14 07:18:37 crc kubenswrapper[4870]: I1014 07:18:37.249039 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/da8a9731-f917-462f-9932-b37b6abb9a64-server-conf\") pod \"rabbitmq-server-0\" (UID: \"da8a9731-f917-462f-9932-b37b6abb9a64\") " pod="openstack/rabbitmq-server-0" Oct 14 07:18:37 crc kubenswrapper[4870]: I1014 07:18:37.249056 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/da8a9731-f917-462f-9932-b37b6abb9a64-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"da8a9731-f917-462f-9932-b37b6abb9a64\") " pod="openstack/rabbitmq-server-0" Oct 14 07:18:37 crc kubenswrapper[4870]: I1014 07:18:37.249086 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/da8a9731-f917-462f-9932-b37b6abb9a64-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"da8a9731-f917-462f-9932-b37b6abb9a64\") " pod="openstack/rabbitmq-server-0" Oct 14 07:18:37 crc kubenswrapper[4870]: I1014 07:18:37.249111 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/da8a9731-f917-462f-9932-b37b6abb9a64-config-data\") pod \"rabbitmq-server-0\" (UID: \"da8a9731-f917-462f-9932-b37b6abb9a64\") " pod="openstack/rabbitmq-server-0" Oct 14 07:18:37 crc kubenswrapper[4870]: I1014 07:18:37.249132 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/da8a9731-f917-462f-9932-b37b6abb9a64-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"da8a9731-f917-462f-9932-b37b6abb9a64\") " pod="openstack/rabbitmq-server-0" Oct 14 07:18:37 crc kubenswrapper[4870]: I1014 07:18:37.249151 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/da8a9731-f917-462f-9932-b37b6abb9a64-pod-info\") pod \"rabbitmq-server-0\" (UID: \"da8a9731-f917-462f-9932-b37b6abb9a64\") " pod="openstack/rabbitmq-server-0" Oct 14 07:18:37 crc kubenswrapper[4870]: I1014 07:18:37.249167 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nl2bs\" (UniqueName: \"kubernetes.io/projected/da8a9731-f917-462f-9932-b37b6abb9a64-kube-api-access-nl2bs\") pod \"rabbitmq-server-0\" (UID: \"da8a9731-f917-462f-9932-b37b6abb9a64\") " pod="openstack/rabbitmq-server-0" Oct 14 07:18:37 crc kubenswrapper[4870]: I1014 07:18:37.249602 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/da8a9731-f917-462f-9932-b37b6abb9a64-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"da8a9731-f917-462f-9932-b37b6abb9a64\") " pod="openstack/rabbitmq-server-0" Oct 14 07:18:37 crc kubenswrapper[4870]: I1014 07:18:37.250241 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/da8a9731-f917-462f-9932-b37b6abb9a64-config-data\") pod \"rabbitmq-server-0\" (UID: \"da8a9731-f917-462f-9932-b37b6abb9a64\") " pod="openstack/rabbitmq-server-0" Oct 14 07:18:37 crc kubenswrapper[4870]: I1014 07:18:37.250247 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/da8a9731-f917-462f-9932-b37b6abb9a64-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"da8a9731-f917-462f-9932-b37b6abb9a64\") " pod="openstack/rabbitmq-server-0" Oct 14 07:18:37 crc kubenswrapper[4870]: I1014 07:18:37.250352 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/da8a9731-f917-462f-9932-b37b6abb9a64-server-conf\") pod \"rabbitmq-server-0\" (UID: \"da8a9731-f917-462f-9932-b37b6abb9a64\") " pod="openstack/rabbitmq-server-0" Oct 14 07:18:37 crc kubenswrapper[4870]: I1014 07:18:37.250552 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/da8a9731-f917-462f-9932-b37b6abb9a64-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"da8a9731-f917-462f-9932-b37b6abb9a64\") " pod="openstack/rabbitmq-server-0" Oct 14 07:18:37 crc kubenswrapper[4870]: I1014 07:18:37.252109 4870 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-server-0\" (UID: \"da8a9731-f917-462f-9932-b37b6abb9a64\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/rabbitmq-server-0" Oct 14 07:18:37 crc kubenswrapper[4870]: I1014 07:18:37.255291 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/da8a9731-f917-462f-9932-b37b6abb9a64-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"da8a9731-f917-462f-9932-b37b6abb9a64\") " pod="openstack/rabbitmq-server-0" Oct 14 07:18:37 crc kubenswrapper[4870]: I1014 07:18:37.255501 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/da8a9731-f917-462f-9932-b37b6abb9a64-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"da8a9731-f917-462f-9932-b37b6abb9a64\") " pod="openstack/rabbitmq-server-0" Oct 14 07:18:37 crc kubenswrapper[4870]: I1014 07:18:37.261284 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/da8a9731-f917-462f-9932-b37b6abb9a64-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"da8a9731-f917-462f-9932-b37b6abb9a64\") " pod="openstack/rabbitmq-server-0" Oct 14 07:18:37 crc kubenswrapper[4870]: I1014 07:18:37.262157 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/da8a9731-f917-462f-9932-b37b6abb9a64-pod-info\") pod \"rabbitmq-server-0\" (UID: \"da8a9731-f917-462f-9932-b37b6abb9a64\") " pod="openstack/rabbitmq-server-0" Oct 14 07:18:37 crc kubenswrapper[4870]: I1014 07:18:37.265135 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nl2bs\" (UniqueName: \"kubernetes.io/projected/da8a9731-f917-462f-9932-b37b6abb9a64-kube-api-access-nl2bs\") pod \"rabbitmq-server-0\" (UID: \"da8a9731-f917-462f-9932-b37b6abb9a64\") " pod="openstack/rabbitmq-server-0" Oct 14 07:18:37 crc kubenswrapper[4870]: I1014 07:18:37.282673 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-server-0\" (UID: \"da8a9731-f917-462f-9932-b37b6abb9a64\") " pod="openstack/rabbitmq-server-0" Oct 14 07:18:37 crc kubenswrapper[4870]: I1014 07:18:37.364619 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 14 07:18:37 crc kubenswrapper[4870]: I1014 07:18:37.367376 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 14 07:18:37 crc kubenswrapper[4870]: I1014 07:18:37.370489 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Oct 14 07:18:37 crc kubenswrapper[4870]: I1014 07:18:37.370606 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Oct 14 07:18:37 crc kubenswrapper[4870]: I1014 07:18:37.370635 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Oct 14 07:18:37 crc kubenswrapper[4870]: I1014 07:18:37.370717 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Oct 14 07:18:37 crc kubenswrapper[4870]: I1014 07:18:37.370731 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Oct 14 07:18:37 crc kubenswrapper[4870]: I1014 07:18:37.370913 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-wlpqf" Oct 14 07:18:37 crc kubenswrapper[4870]: I1014 07:18:37.371236 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Oct 14 07:18:37 crc kubenswrapper[4870]: I1014 07:18:37.384563 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 14 07:18:37 crc kubenswrapper[4870]: I1014 07:18:37.445157 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 14 07:18:37 crc kubenswrapper[4870]: I1014 07:18:37.557970 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/5212ba20-83c3-41ff-a3d9-5a315f009c1b-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"5212ba20-83c3-41ff-a3d9-5a315f009c1b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 07:18:37 crc kubenswrapper[4870]: I1014 07:18:37.558611 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/5212ba20-83c3-41ff-a3d9-5a315f009c1b-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"5212ba20-83c3-41ff-a3d9-5a315f009c1b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 07:18:37 crc kubenswrapper[4870]: I1014 07:18:37.558641 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/5212ba20-83c3-41ff-a3d9-5a315f009c1b-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"5212ba20-83c3-41ff-a3d9-5a315f009c1b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 07:18:37 crc kubenswrapper[4870]: I1014 07:18:37.558666 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"5212ba20-83c3-41ff-a3d9-5a315f009c1b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 07:18:37 crc kubenswrapper[4870]: I1014 07:18:37.558733 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/5212ba20-83c3-41ff-a3d9-5a315f009c1b-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"5212ba20-83c3-41ff-a3d9-5a315f009c1b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 07:18:37 crc kubenswrapper[4870]: I1014 07:18:37.558816 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5212ba20-83c3-41ff-a3d9-5a315f009c1b-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"5212ba20-83c3-41ff-a3d9-5a315f009c1b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 07:18:37 crc kubenswrapper[4870]: I1014 07:18:37.558838 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/5212ba20-83c3-41ff-a3d9-5a315f009c1b-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"5212ba20-83c3-41ff-a3d9-5a315f009c1b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 07:18:37 crc kubenswrapper[4870]: I1014 07:18:37.558863 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d4r8n\" (UniqueName: \"kubernetes.io/projected/5212ba20-83c3-41ff-a3d9-5a315f009c1b-kube-api-access-d4r8n\") pod \"rabbitmq-cell1-server-0\" (UID: \"5212ba20-83c3-41ff-a3d9-5a315f009c1b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 07:18:37 crc kubenswrapper[4870]: I1014 07:18:37.559166 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/5212ba20-83c3-41ff-a3d9-5a315f009c1b-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"5212ba20-83c3-41ff-a3d9-5a315f009c1b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 07:18:37 crc kubenswrapper[4870]: I1014 07:18:37.559254 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/5212ba20-83c3-41ff-a3d9-5a315f009c1b-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"5212ba20-83c3-41ff-a3d9-5a315f009c1b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 07:18:37 crc kubenswrapper[4870]: I1014 07:18:37.559367 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/5212ba20-83c3-41ff-a3d9-5a315f009c1b-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"5212ba20-83c3-41ff-a3d9-5a315f009c1b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 07:18:37 crc kubenswrapper[4870]: I1014 07:18:37.661030 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/5212ba20-83c3-41ff-a3d9-5a315f009c1b-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"5212ba20-83c3-41ff-a3d9-5a315f009c1b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 07:18:37 crc kubenswrapper[4870]: I1014 07:18:37.661083 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/5212ba20-83c3-41ff-a3d9-5a315f009c1b-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"5212ba20-83c3-41ff-a3d9-5a315f009c1b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 07:18:37 crc kubenswrapper[4870]: I1014 07:18:37.661121 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/5212ba20-83c3-41ff-a3d9-5a315f009c1b-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"5212ba20-83c3-41ff-a3d9-5a315f009c1b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 07:18:37 crc kubenswrapper[4870]: I1014 07:18:37.661160 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/5212ba20-83c3-41ff-a3d9-5a315f009c1b-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"5212ba20-83c3-41ff-a3d9-5a315f009c1b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 07:18:37 crc kubenswrapper[4870]: I1014 07:18:37.661179 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/5212ba20-83c3-41ff-a3d9-5a315f009c1b-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"5212ba20-83c3-41ff-a3d9-5a315f009c1b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 07:18:37 crc kubenswrapper[4870]: I1014 07:18:37.661196 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/5212ba20-83c3-41ff-a3d9-5a315f009c1b-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"5212ba20-83c3-41ff-a3d9-5a315f009c1b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 07:18:37 crc kubenswrapper[4870]: I1014 07:18:37.661215 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"5212ba20-83c3-41ff-a3d9-5a315f009c1b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 07:18:37 crc kubenswrapper[4870]: I1014 07:18:37.661236 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/5212ba20-83c3-41ff-a3d9-5a315f009c1b-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"5212ba20-83c3-41ff-a3d9-5a315f009c1b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 07:18:37 crc kubenswrapper[4870]: I1014 07:18:37.661260 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5212ba20-83c3-41ff-a3d9-5a315f009c1b-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"5212ba20-83c3-41ff-a3d9-5a315f009c1b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 07:18:37 crc kubenswrapper[4870]: I1014 07:18:37.661277 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/5212ba20-83c3-41ff-a3d9-5a315f009c1b-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"5212ba20-83c3-41ff-a3d9-5a315f009c1b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 07:18:37 crc kubenswrapper[4870]: I1014 07:18:37.661294 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d4r8n\" (UniqueName: \"kubernetes.io/projected/5212ba20-83c3-41ff-a3d9-5a315f009c1b-kube-api-access-d4r8n\") pod \"rabbitmq-cell1-server-0\" (UID: \"5212ba20-83c3-41ff-a3d9-5a315f009c1b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 07:18:37 crc kubenswrapper[4870]: I1014 07:18:37.662389 4870 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"5212ba20-83c3-41ff-a3d9-5a315f009c1b\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/rabbitmq-cell1-server-0" Oct 14 07:18:37 crc kubenswrapper[4870]: I1014 07:18:37.662625 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/5212ba20-83c3-41ff-a3d9-5a315f009c1b-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"5212ba20-83c3-41ff-a3d9-5a315f009c1b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 07:18:37 crc kubenswrapper[4870]: I1014 07:18:37.662667 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/5212ba20-83c3-41ff-a3d9-5a315f009c1b-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"5212ba20-83c3-41ff-a3d9-5a315f009c1b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 07:18:37 crc kubenswrapper[4870]: I1014 07:18:37.663285 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5212ba20-83c3-41ff-a3d9-5a315f009c1b-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"5212ba20-83c3-41ff-a3d9-5a315f009c1b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 07:18:37 crc kubenswrapper[4870]: I1014 07:18:37.662402 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/5212ba20-83c3-41ff-a3d9-5a315f009c1b-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"5212ba20-83c3-41ff-a3d9-5a315f009c1b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 07:18:37 crc kubenswrapper[4870]: I1014 07:18:37.663356 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/5212ba20-83c3-41ff-a3d9-5a315f009c1b-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"5212ba20-83c3-41ff-a3d9-5a315f009c1b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 07:18:37 crc kubenswrapper[4870]: I1014 07:18:37.666544 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/5212ba20-83c3-41ff-a3d9-5a315f009c1b-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"5212ba20-83c3-41ff-a3d9-5a315f009c1b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 07:18:37 crc kubenswrapper[4870]: I1014 07:18:37.668304 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/5212ba20-83c3-41ff-a3d9-5a315f009c1b-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"5212ba20-83c3-41ff-a3d9-5a315f009c1b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 07:18:37 crc kubenswrapper[4870]: I1014 07:18:37.668299 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/5212ba20-83c3-41ff-a3d9-5a315f009c1b-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"5212ba20-83c3-41ff-a3d9-5a315f009c1b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 07:18:37 crc kubenswrapper[4870]: I1014 07:18:37.670971 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/5212ba20-83c3-41ff-a3d9-5a315f009c1b-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"5212ba20-83c3-41ff-a3d9-5a315f009c1b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 07:18:37 crc kubenswrapper[4870]: I1014 07:18:37.678947 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d4r8n\" (UniqueName: \"kubernetes.io/projected/5212ba20-83c3-41ff-a3d9-5a315f009c1b-kube-api-access-d4r8n\") pod \"rabbitmq-cell1-server-0\" (UID: \"5212ba20-83c3-41ff-a3d9-5a315f009c1b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 07:18:37 crc kubenswrapper[4870]: I1014 07:18:37.707836 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"5212ba20-83c3-41ff-a3d9-5a315f009c1b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 07:18:37 crc kubenswrapper[4870]: I1014 07:18:37.849533 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7869c47d6c-44h7s" event={"ID":"11aab367-2351-4568-9a45-43c79d419bca","Type":"ContainerStarted","Data":"d45565075ad8f85eaabe03f0f165cb8b812379f94c0a532c4e203f8434352b6c"} Oct 14 07:18:37 crc kubenswrapper[4870]: I1014 07:18:37.986747 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 14 07:18:37 crc kubenswrapper[4870]: I1014 07:18:37.998462 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 14 07:18:38 crc kubenswrapper[4870]: I1014 07:18:38.538754 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Oct 14 07:18:38 crc kubenswrapper[4870]: I1014 07:18:38.540557 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Oct 14 07:18:38 crc kubenswrapper[4870]: I1014 07:18:38.547798 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Oct 14 07:18:38 crc kubenswrapper[4870]: I1014 07:18:38.548070 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Oct 14 07:18:38 crc kubenswrapper[4870]: I1014 07:18:38.548156 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Oct 14 07:18:38 crc kubenswrapper[4870]: I1014 07:18:38.548207 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Oct 14 07:18:38 crc kubenswrapper[4870]: I1014 07:18:38.548155 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-s4pf5" Oct 14 07:18:38 crc kubenswrapper[4870]: I1014 07:18:38.548738 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Oct 14 07:18:38 crc kubenswrapper[4870]: I1014 07:18:38.555597 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Oct 14 07:18:38 crc kubenswrapper[4870]: I1014 07:18:38.674949 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/3de2167a-7663-4b00-9743-229ff1864cb8-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"3de2167a-7663-4b00-9743-229ff1864cb8\") " pod="openstack/openstack-galera-0" Oct 14 07:18:38 crc kubenswrapper[4870]: I1014 07:18:38.674988 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/3de2167a-7663-4b00-9743-229ff1864cb8-secrets\") pod \"openstack-galera-0\" (UID: \"3de2167a-7663-4b00-9743-229ff1864cb8\") " pod="openstack/openstack-galera-0" Oct 14 07:18:38 crc kubenswrapper[4870]: I1014 07:18:38.675020 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-galera-0\" (UID: \"3de2167a-7663-4b00-9743-229ff1864cb8\") " pod="openstack/openstack-galera-0" Oct 14 07:18:38 crc kubenswrapper[4870]: I1014 07:18:38.675056 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/3de2167a-7663-4b00-9743-229ff1864cb8-kolla-config\") pod \"openstack-galera-0\" (UID: \"3de2167a-7663-4b00-9743-229ff1864cb8\") " pod="openstack/openstack-galera-0" Oct 14 07:18:38 crc kubenswrapper[4870]: I1014 07:18:38.675070 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3de2167a-7663-4b00-9743-229ff1864cb8-operator-scripts\") pod \"openstack-galera-0\" (UID: \"3de2167a-7663-4b00-9743-229ff1864cb8\") " pod="openstack/openstack-galera-0" Oct 14 07:18:38 crc kubenswrapper[4870]: I1014 07:18:38.675099 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/3de2167a-7663-4b00-9743-229ff1864cb8-config-data-generated\") pod \"openstack-galera-0\" (UID: \"3de2167a-7663-4b00-9743-229ff1864cb8\") " pod="openstack/openstack-galera-0" Oct 14 07:18:38 crc kubenswrapper[4870]: I1014 07:18:38.675124 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3de2167a-7663-4b00-9743-229ff1864cb8-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"3de2167a-7663-4b00-9743-229ff1864cb8\") " pod="openstack/openstack-galera-0" Oct 14 07:18:38 crc kubenswrapper[4870]: I1014 07:18:38.675150 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bz55d\" (UniqueName: \"kubernetes.io/projected/3de2167a-7663-4b00-9743-229ff1864cb8-kube-api-access-bz55d\") pod \"openstack-galera-0\" (UID: \"3de2167a-7663-4b00-9743-229ff1864cb8\") " pod="openstack/openstack-galera-0" Oct 14 07:18:38 crc kubenswrapper[4870]: I1014 07:18:38.675169 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/3de2167a-7663-4b00-9743-229ff1864cb8-config-data-default\") pod \"openstack-galera-0\" (UID: \"3de2167a-7663-4b00-9743-229ff1864cb8\") " pod="openstack/openstack-galera-0" Oct 14 07:18:38 crc kubenswrapper[4870]: I1014 07:18:38.776910 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/3de2167a-7663-4b00-9743-229ff1864cb8-secrets\") pod \"openstack-galera-0\" (UID: \"3de2167a-7663-4b00-9743-229ff1864cb8\") " pod="openstack/openstack-galera-0" Oct 14 07:18:38 crc kubenswrapper[4870]: I1014 07:18:38.776949 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/3de2167a-7663-4b00-9743-229ff1864cb8-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"3de2167a-7663-4b00-9743-229ff1864cb8\") " pod="openstack/openstack-galera-0" Oct 14 07:18:38 crc kubenswrapper[4870]: I1014 07:18:38.776985 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-galera-0\" (UID: \"3de2167a-7663-4b00-9743-229ff1864cb8\") " pod="openstack/openstack-galera-0" Oct 14 07:18:38 crc kubenswrapper[4870]: I1014 07:18:38.777026 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/3de2167a-7663-4b00-9743-229ff1864cb8-kolla-config\") pod \"openstack-galera-0\" (UID: \"3de2167a-7663-4b00-9743-229ff1864cb8\") " pod="openstack/openstack-galera-0" Oct 14 07:18:38 crc kubenswrapper[4870]: I1014 07:18:38.777044 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3de2167a-7663-4b00-9743-229ff1864cb8-operator-scripts\") pod \"openstack-galera-0\" (UID: \"3de2167a-7663-4b00-9743-229ff1864cb8\") " pod="openstack/openstack-galera-0" Oct 14 07:18:38 crc kubenswrapper[4870]: I1014 07:18:38.777079 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/3de2167a-7663-4b00-9743-229ff1864cb8-config-data-generated\") pod \"openstack-galera-0\" (UID: \"3de2167a-7663-4b00-9743-229ff1864cb8\") " pod="openstack/openstack-galera-0" Oct 14 07:18:38 crc kubenswrapper[4870]: I1014 07:18:38.777135 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3de2167a-7663-4b00-9743-229ff1864cb8-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"3de2167a-7663-4b00-9743-229ff1864cb8\") " pod="openstack/openstack-galera-0" Oct 14 07:18:38 crc kubenswrapper[4870]: I1014 07:18:38.777170 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bz55d\" (UniqueName: \"kubernetes.io/projected/3de2167a-7663-4b00-9743-229ff1864cb8-kube-api-access-bz55d\") pod \"openstack-galera-0\" (UID: \"3de2167a-7663-4b00-9743-229ff1864cb8\") " pod="openstack/openstack-galera-0" Oct 14 07:18:38 crc kubenswrapper[4870]: I1014 07:18:38.777505 4870 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-galera-0\" (UID: \"3de2167a-7663-4b00-9743-229ff1864cb8\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/openstack-galera-0" Oct 14 07:18:38 crc kubenswrapper[4870]: I1014 07:18:38.778551 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/3de2167a-7663-4b00-9743-229ff1864cb8-kolla-config\") pod \"openstack-galera-0\" (UID: \"3de2167a-7663-4b00-9743-229ff1864cb8\") " pod="openstack/openstack-galera-0" Oct 14 07:18:38 crc kubenswrapper[4870]: I1014 07:18:38.780226 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3de2167a-7663-4b00-9743-229ff1864cb8-operator-scripts\") pod \"openstack-galera-0\" (UID: \"3de2167a-7663-4b00-9743-229ff1864cb8\") " pod="openstack/openstack-galera-0" Oct 14 07:18:38 crc kubenswrapper[4870]: I1014 07:18:38.780822 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/3de2167a-7663-4b00-9743-229ff1864cb8-config-data-default\") pod \"openstack-galera-0\" (UID: \"3de2167a-7663-4b00-9743-229ff1864cb8\") " pod="openstack/openstack-galera-0" Oct 14 07:18:38 crc kubenswrapper[4870]: I1014 07:18:38.782343 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/3de2167a-7663-4b00-9743-229ff1864cb8-config-data-default\") pod \"openstack-galera-0\" (UID: \"3de2167a-7663-4b00-9743-229ff1864cb8\") " pod="openstack/openstack-galera-0" Oct 14 07:18:38 crc kubenswrapper[4870]: I1014 07:18:38.783182 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/3de2167a-7663-4b00-9743-229ff1864cb8-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"3de2167a-7663-4b00-9743-229ff1864cb8\") " pod="openstack/openstack-galera-0" Oct 14 07:18:38 crc kubenswrapper[4870]: I1014 07:18:38.783598 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3de2167a-7663-4b00-9743-229ff1864cb8-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"3de2167a-7663-4b00-9743-229ff1864cb8\") " pod="openstack/openstack-galera-0" Oct 14 07:18:38 crc kubenswrapper[4870]: I1014 07:18:38.784771 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/3de2167a-7663-4b00-9743-229ff1864cb8-config-data-generated\") pod \"openstack-galera-0\" (UID: \"3de2167a-7663-4b00-9743-229ff1864cb8\") " pod="openstack/openstack-galera-0" Oct 14 07:18:38 crc kubenswrapper[4870]: I1014 07:18:38.785856 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/3de2167a-7663-4b00-9743-229ff1864cb8-secrets\") pod \"openstack-galera-0\" (UID: \"3de2167a-7663-4b00-9743-229ff1864cb8\") " pod="openstack/openstack-galera-0" Oct 14 07:18:38 crc kubenswrapper[4870]: I1014 07:18:38.796222 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bz55d\" (UniqueName: \"kubernetes.io/projected/3de2167a-7663-4b00-9743-229ff1864cb8-kube-api-access-bz55d\") pod \"openstack-galera-0\" (UID: \"3de2167a-7663-4b00-9743-229ff1864cb8\") " pod="openstack/openstack-galera-0" Oct 14 07:18:38 crc kubenswrapper[4870]: I1014 07:18:38.819756 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-galera-0\" (UID: \"3de2167a-7663-4b00-9743-229ff1864cb8\") " pod="openstack/openstack-galera-0" Oct 14 07:18:38 crc kubenswrapper[4870]: I1014 07:18:38.869782 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Oct 14 07:18:39 crc kubenswrapper[4870]: I1014 07:18:39.925097 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 14 07:18:39 crc kubenswrapper[4870]: I1014 07:18:39.926863 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Oct 14 07:18:39 crc kubenswrapper[4870]: I1014 07:18:39.928851 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-ttmdp" Oct 14 07:18:39 crc kubenswrapper[4870]: I1014 07:18:39.929036 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Oct 14 07:18:39 crc kubenswrapper[4870]: I1014 07:18:39.937406 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 14 07:18:39 crc kubenswrapper[4870]: I1014 07:18:39.937598 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Oct 14 07:18:39 crc kubenswrapper[4870]: I1014 07:18:39.937944 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Oct 14 07:18:40 crc kubenswrapper[4870]: I1014 07:18:40.099116 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-cell1-galera-0\" (UID: \"4d74ef50-1861-4d70-b781-f92d5f4b35ba\") " pod="openstack/openstack-cell1-galera-0" Oct 14 07:18:40 crc kubenswrapper[4870]: I1014 07:18:40.099234 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/4d74ef50-1861-4d70-b781-f92d5f4b35ba-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"4d74ef50-1861-4d70-b781-f92d5f4b35ba\") " pod="openstack/openstack-cell1-galera-0" Oct 14 07:18:40 crc kubenswrapper[4870]: I1014 07:18:40.099267 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d74ef50-1861-4d70-b781-f92d5f4b35ba-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"4d74ef50-1861-4d70-b781-f92d5f4b35ba\") " pod="openstack/openstack-cell1-galera-0" Oct 14 07:18:40 crc kubenswrapper[4870]: I1014 07:18:40.099329 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/4d74ef50-1861-4d70-b781-f92d5f4b35ba-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"4d74ef50-1861-4d70-b781-f92d5f4b35ba\") " pod="openstack/openstack-cell1-galera-0" Oct 14 07:18:40 crc kubenswrapper[4870]: I1014 07:18:40.099354 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/4d74ef50-1861-4d70-b781-f92d5f4b35ba-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"4d74ef50-1861-4d70-b781-f92d5f4b35ba\") " pod="openstack/openstack-cell1-galera-0" Oct 14 07:18:40 crc kubenswrapper[4870]: I1014 07:18:40.099388 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8mr4z\" (UniqueName: \"kubernetes.io/projected/4d74ef50-1861-4d70-b781-f92d5f4b35ba-kube-api-access-8mr4z\") pod \"openstack-cell1-galera-0\" (UID: \"4d74ef50-1861-4d70-b781-f92d5f4b35ba\") " pod="openstack/openstack-cell1-galera-0" Oct 14 07:18:40 crc kubenswrapper[4870]: I1014 07:18:40.099417 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/4d74ef50-1861-4d70-b781-f92d5f4b35ba-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"4d74ef50-1861-4d70-b781-f92d5f4b35ba\") " pod="openstack/openstack-cell1-galera-0" Oct 14 07:18:40 crc kubenswrapper[4870]: I1014 07:18:40.099495 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4d74ef50-1861-4d70-b781-f92d5f4b35ba-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"4d74ef50-1861-4d70-b781-f92d5f4b35ba\") " pod="openstack/openstack-cell1-galera-0" Oct 14 07:18:40 crc kubenswrapper[4870]: I1014 07:18:40.099559 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/4d74ef50-1861-4d70-b781-f92d5f4b35ba-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"4d74ef50-1861-4d70-b781-f92d5f4b35ba\") " pod="openstack/openstack-cell1-galera-0" Oct 14 07:18:40 crc kubenswrapper[4870]: I1014 07:18:40.201578 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/4d74ef50-1861-4d70-b781-f92d5f4b35ba-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"4d74ef50-1861-4d70-b781-f92d5f4b35ba\") " pod="openstack/openstack-cell1-galera-0" Oct 14 07:18:40 crc kubenswrapper[4870]: I1014 07:18:40.201643 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/4d74ef50-1861-4d70-b781-f92d5f4b35ba-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"4d74ef50-1861-4d70-b781-f92d5f4b35ba\") " pod="openstack/openstack-cell1-galera-0" Oct 14 07:18:40 crc kubenswrapper[4870]: I1014 07:18:40.201673 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8mr4z\" (UniqueName: \"kubernetes.io/projected/4d74ef50-1861-4d70-b781-f92d5f4b35ba-kube-api-access-8mr4z\") pod \"openstack-cell1-galera-0\" (UID: \"4d74ef50-1861-4d70-b781-f92d5f4b35ba\") " pod="openstack/openstack-cell1-galera-0" Oct 14 07:18:40 crc kubenswrapper[4870]: I1014 07:18:40.201712 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/4d74ef50-1861-4d70-b781-f92d5f4b35ba-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"4d74ef50-1861-4d70-b781-f92d5f4b35ba\") " pod="openstack/openstack-cell1-galera-0" Oct 14 07:18:40 crc kubenswrapper[4870]: I1014 07:18:40.202408 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/4d74ef50-1861-4d70-b781-f92d5f4b35ba-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"4d74ef50-1861-4d70-b781-f92d5f4b35ba\") " pod="openstack/openstack-cell1-galera-0" Oct 14 07:18:40 crc kubenswrapper[4870]: I1014 07:18:40.202547 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4d74ef50-1861-4d70-b781-f92d5f4b35ba-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"4d74ef50-1861-4d70-b781-f92d5f4b35ba\") " pod="openstack/openstack-cell1-galera-0" Oct 14 07:18:40 crc kubenswrapper[4870]: I1014 07:18:40.202600 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/4d74ef50-1861-4d70-b781-f92d5f4b35ba-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"4d74ef50-1861-4d70-b781-f92d5f4b35ba\") " pod="openstack/openstack-cell1-galera-0" Oct 14 07:18:40 crc kubenswrapper[4870]: I1014 07:18:40.202652 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-cell1-galera-0\" (UID: \"4d74ef50-1861-4d70-b781-f92d5f4b35ba\") " pod="openstack/openstack-cell1-galera-0" Oct 14 07:18:40 crc kubenswrapper[4870]: I1014 07:18:40.202696 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/4d74ef50-1861-4d70-b781-f92d5f4b35ba-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"4d74ef50-1861-4d70-b781-f92d5f4b35ba\") " pod="openstack/openstack-cell1-galera-0" Oct 14 07:18:40 crc kubenswrapper[4870]: I1014 07:18:40.202724 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d74ef50-1861-4d70-b781-f92d5f4b35ba-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"4d74ef50-1861-4d70-b781-f92d5f4b35ba\") " pod="openstack/openstack-cell1-galera-0" Oct 14 07:18:40 crc kubenswrapper[4870]: I1014 07:18:40.202969 4870 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-cell1-galera-0\" (UID: \"4d74ef50-1861-4d70-b781-f92d5f4b35ba\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/openstack-cell1-galera-0" Oct 14 07:18:40 crc kubenswrapper[4870]: I1014 07:18:40.203609 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/4d74ef50-1861-4d70-b781-f92d5f4b35ba-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"4d74ef50-1861-4d70-b781-f92d5f4b35ba\") " pod="openstack/openstack-cell1-galera-0" Oct 14 07:18:40 crc kubenswrapper[4870]: I1014 07:18:40.203700 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4d74ef50-1861-4d70-b781-f92d5f4b35ba-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"4d74ef50-1861-4d70-b781-f92d5f4b35ba\") " pod="openstack/openstack-cell1-galera-0" Oct 14 07:18:40 crc kubenswrapper[4870]: I1014 07:18:40.203917 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/4d74ef50-1861-4d70-b781-f92d5f4b35ba-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"4d74ef50-1861-4d70-b781-f92d5f4b35ba\") " pod="openstack/openstack-cell1-galera-0" Oct 14 07:18:40 crc kubenswrapper[4870]: I1014 07:18:40.206114 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/4d74ef50-1861-4d70-b781-f92d5f4b35ba-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"4d74ef50-1861-4d70-b781-f92d5f4b35ba\") " pod="openstack/openstack-cell1-galera-0" Oct 14 07:18:40 crc kubenswrapper[4870]: I1014 07:18:40.212948 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/4d74ef50-1861-4d70-b781-f92d5f4b35ba-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"4d74ef50-1861-4d70-b781-f92d5f4b35ba\") " pod="openstack/openstack-cell1-galera-0" Oct 14 07:18:40 crc kubenswrapper[4870]: I1014 07:18:40.213147 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d74ef50-1861-4d70-b781-f92d5f4b35ba-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"4d74ef50-1861-4d70-b781-f92d5f4b35ba\") " pod="openstack/openstack-cell1-galera-0" Oct 14 07:18:40 crc kubenswrapper[4870]: I1014 07:18:40.219351 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8mr4z\" (UniqueName: \"kubernetes.io/projected/4d74ef50-1861-4d70-b781-f92d5f4b35ba-kube-api-access-8mr4z\") pod \"openstack-cell1-galera-0\" (UID: \"4d74ef50-1861-4d70-b781-f92d5f4b35ba\") " pod="openstack/openstack-cell1-galera-0" Oct 14 07:18:40 crc kubenswrapper[4870]: I1014 07:18:40.226480 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-cell1-galera-0\" (UID: \"4d74ef50-1861-4d70-b781-f92d5f4b35ba\") " pod="openstack/openstack-cell1-galera-0" Oct 14 07:18:40 crc kubenswrapper[4870]: I1014 07:18:40.295998 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Oct 14 07:18:40 crc kubenswrapper[4870]: I1014 07:18:40.305185 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Oct 14 07:18:40 crc kubenswrapper[4870]: I1014 07:18:40.306645 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Oct 14 07:18:40 crc kubenswrapper[4870]: I1014 07:18:40.313647 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Oct 14 07:18:40 crc kubenswrapper[4870]: I1014 07:18:40.313694 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Oct 14 07:18:40 crc kubenswrapper[4870]: I1014 07:18:40.313721 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Oct 14 07:18:40 crc kubenswrapper[4870]: I1014 07:18:40.313890 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-kmmhp" Oct 14 07:18:40 crc kubenswrapper[4870]: I1014 07:18:40.405191 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/3bd1e649-3d09-4cda-921f-ac9a84a5066e-kolla-config\") pod \"memcached-0\" (UID: \"3bd1e649-3d09-4cda-921f-ac9a84a5066e\") " pod="openstack/memcached-0" Oct 14 07:18:40 crc kubenswrapper[4870]: I1014 07:18:40.405238 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3bd1e649-3d09-4cda-921f-ac9a84a5066e-config-data\") pod \"memcached-0\" (UID: \"3bd1e649-3d09-4cda-921f-ac9a84a5066e\") " pod="openstack/memcached-0" Oct 14 07:18:40 crc kubenswrapper[4870]: I1014 07:18:40.405265 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3bd1e649-3d09-4cda-921f-ac9a84a5066e-combined-ca-bundle\") pod \"memcached-0\" (UID: \"3bd1e649-3d09-4cda-921f-ac9a84a5066e\") " pod="openstack/memcached-0" Oct 14 07:18:40 crc kubenswrapper[4870]: I1014 07:18:40.405300 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/3bd1e649-3d09-4cda-921f-ac9a84a5066e-memcached-tls-certs\") pod \"memcached-0\" (UID: \"3bd1e649-3d09-4cda-921f-ac9a84a5066e\") " pod="openstack/memcached-0" Oct 14 07:18:40 crc kubenswrapper[4870]: I1014 07:18:40.405330 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bgmmx\" (UniqueName: \"kubernetes.io/projected/3bd1e649-3d09-4cda-921f-ac9a84a5066e-kube-api-access-bgmmx\") pod \"memcached-0\" (UID: \"3bd1e649-3d09-4cda-921f-ac9a84a5066e\") " pod="openstack/memcached-0" Oct 14 07:18:40 crc kubenswrapper[4870]: I1014 07:18:40.507025 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/3bd1e649-3d09-4cda-921f-ac9a84a5066e-kolla-config\") pod \"memcached-0\" (UID: \"3bd1e649-3d09-4cda-921f-ac9a84a5066e\") " pod="openstack/memcached-0" Oct 14 07:18:40 crc kubenswrapper[4870]: I1014 07:18:40.507076 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3bd1e649-3d09-4cda-921f-ac9a84a5066e-config-data\") pod \"memcached-0\" (UID: \"3bd1e649-3d09-4cda-921f-ac9a84a5066e\") " pod="openstack/memcached-0" Oct 14 07:18:40 crc kubenswrapper[4870]: I1014 07:18:40.507103 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3bd1e649-3d09-4cda-921f-ac9a84a5066e-combined-ca-bundle\") pod \"memcached-0\" (UID: \"3bd1e649-3d09-4cda-921f-ac9a84a5066e\") " pod="openstack/memcached-0" Oct 14 07:18:40 crc kubenswrapper[4870]: I1014 07:18:40.507139 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/3bd1e649-3d09-4cda-921f-ac9a84a5066e-memcached-tls-certs\") pod \"memcached-0\" (UID: \"3bd1e649-3d09-4cda-921f-ac9a84a5066e\") " pod="openstack/memcached-0" Oct 14 07:18:40 crc kubenswrapper[4870]: I1014 07:18:40.507169 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bgmmx\" (UniqueName: \"kubernetes.io/projected/3bd1e649-3d09-4cda-921f-ac9a84a5066e-kube-api-access-bgmmx\") pod \"memcached-0\" (UID: \"3bd1e649-3d09-4cda-921f-ac9a84a5066e\") " pod="openstack/memcached-0" Oct 14 07:18:40 crc kubenswrapper[4870]: I1014 07:18:40.508133 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/3bd1e649-3d09-4cda-921f-ac9a84a5066e-kolla-config\") pod \"memcached-0\" (UID: \"3bd1e649-3d09-4cda-921f-ac9a84a5066e\") " pod="openstack/memcached-0" Oct 14 07:18:40 crc kubenswrapper[4870]: I1014 07:18:40.508563 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3bd1e649-3d09-4cda-921f-ac9a84a5066e-config-data\") pod \"memcached-0\" (UID: \"3bd1e649-3d09-4cda-921f-ac9a84a5066e\") " pod="openstack/memcached-0" Oct 14 07:18:40 crc kubenswrapper[4870]: I1014 07:18:40.526226 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/3bd1e649-3d09-4cda-921f-ac9a84a5066e-memcached-tls-certs\") pod \"memcached-0\" (UID: \"3bd1e649-3d09-4cda-921f-ac9a84a5066e\") " pod="openstack/memcached-0" Oct 14 07:18:40 crc kubenswrapper[4870]: I1014 07:18:40.529378 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bgmmx\" (UniqueName: \"kubernetes.io/projected/3bd1e649-3d09-4cda-921f-ac9a84a5066e-kube-api-access-bgmmx\") pod \"memcached-0\" (UID: \"3bd1e649-3d09-4cda-921f-ac9a84a5066e\") " pod="openstack/memcached-0" Oct 14 07:18:40 crc kubenswrapper[4870]: I1014 07:18:40.530389 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3bd1e649-3d09-4cda-921f-ac9a84a5066e-combined-ca-bundle\") pod \"memcached-0\" (UID: \"3bd1e649-3d09-4cda-921f-ac9a84a5066e\") " pod="openstack/memcached-0" Oct 14 07:18:40 crc kubenswrapper[4870]: I1014 07:18:40.627969 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Oct 14 07:18:42 crc kubenswrapper[4870]: I1014 07:18:42.621167 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Oct 14 07:18:42 crc kubenswrapper[4870]: I1014 07:18:42.622533 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 14 07:18:42 crc kubenswrapper[4870]: I1014 07:18:42.628095 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-lcmwg" Oct 14 07:18:42 crc kubenswrapper[4870]: I1014 07:18:42.632587 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 14 07:18:42 crc kubenswrapper[4870]: I1014 07:18:42.746501 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nmlrt\" (UniqueName: \"kubernetes.io/projected/b2d073e2-e7bb-44c9-b074-c39afeae16b8-kube-api-access-nmlrt\") pod \"kube-state-metrics-0\" (UID: \"b2d073e2-e7bb-44c9-b074-c39afeae16b8\") " pod="openstack/kube-state-metrics-0" Oct 14 07:18:42 crc kubenswrapper[4870]: I1014 07:18:42.848616 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nmlrt\" (UniqueName: \"kubernetes.io/projected/b2d073e2-e7bb-44c9-b074-c39afeae16b8-kube-api-access-nmlrt\") pod \"kube-state-metrics-0\" (UID: \"b2d073e2-e7bb-44c9-b074-c39afeae16b8\") " pod="openstack/kube-state-metrics-0" Oct 14 07:18:42 crc kubenswrapper[4870]: I1014 07:18:42.869767 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nmlrt\" (UniqueName: \"kubernetes.io/projected/b2d073e2-e7bb-44c9-b074-c39afeae16b8-kube-api-access-nmlrt\") pod \"kube-state-metrics-0\" (UID: \"b2d073e2-e7bb-44c9-b074-c39afeae16b8\") " pod="openstack/kube-state-metrics-0" Oct 14 07:18:42 crc kubenswrapper[4870]: I1014 07:18:42.980316 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 14 07:18:43 crc kubenswrapper[4870]: I1014 07:18:43.069966 4870 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 14 07:18:43 crc kubenswrapper[4870]: I1014 07:18:43.908135 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"da8a9731-f917-462f-9932-b37b6abb9a64","Type":"ContainerStarted","Data":"b2190e6313a8dbb2a9d5a3f14b18f0e56fba0be535cabc899a22fee9bf2d18bf"} Oct 14 07:18:45 crc kubenswrapper[4870]: I1014 07:18:45.977808 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Oct 14 07:18:46 crc kubenswrapper[4870]: I1014 07:18:46.052202 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 14 07:18:46 crc kubenswrapper[4870]: I1014 07:18:46.053538 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Oct 14 07:18:46 crc kubenswrapper[4870]: I1014 07:18:46.055400 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Oct 14 07:18:46 crc kubenswrapper[4870]: I1014 07:18:46.055430 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Oct 14 07:18:46 crc kubenswrapper[4870]: I1014 07:18:46.055619 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Oct 14 07:18:46 crc kubenswrapper[4870]: I1014 07:18:46.055638 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Oct 14 07:18:46 crc kubenswrapper[4870]: I1014 07:18:46.060830 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-4k7t6" Oct 14 07:18:46 crc kubenswrapper[4870]: I1014 07:18:46.064146 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 14 07:18:46 crc kubenswrapper[4870]: I1014 07:18:46.108073 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/09f89569-6fe3-4b3f-9394-a990ad8cde30-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"09f89569-6fe3-4b3f-9394-a990ad8cde30\") " pod="openstack/ovsdbserver-nb-0" Oct 14 07:18:46 crc kubenswrapper[4870]: I1014 07:18:46.108115 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/09f89569-6fe3-4b3f-9394-a990ad8cde30-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"09f89569-6fe3-4b3f-9394-a990ad8cde30\") " pod="openstack/ovsdbserver-nb-0" Oct 14 07:18:46 crc kubenswrapper[4870]: I1014 07:18:46.108140 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"ovsdbserver-nb-0\" (UID: \"09f89569-6fe3-4b3f-9394-a990ad8cde30\") " pod="openstack/ovsdbserver-nb-0" Oct 14 07:18:46 crc kubenswrapper[4870]: I1014 07:18:46.108181 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/09f89569-6fe3-4b3f-9394-a990ad8cde30-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"09f89569-6fe3-4b3f-9394-a990ad8cde30\") " pod="openstack/ovsdbserver-nb-0" Oct 14 07:18:46 crc kubenswrapper[4870]: I1014 07:18:46.108246 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8hks2\" (UniqueName: \"kubernetes.io/projected/09f89569-6fe3-4b3f-9394-a990ad8cde30-kube-api-access-8hks2\") pod \"ovsdbserver-nb-0\" (UID: \"09f89569-6fe3-4b3f-9394-a990ad8cde30\") " pod="openstack/ovsdbserver-nb-0" Oct 14 07:18:46 crc kubenswrapper[4870]: I1014 07:18:46.108303 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09f89569-6fe3-4b3f-9394-a990ad8cde30-config\") pod \"ovsdbserver-nb-0\" (UID: \"09f89569-6fe3-4b3f-9394-a990ad8cde30\") " pod="openstack/ovsdbserver-nb-0" Oct 14 07:18:46 crc kubenswrapper[4870]: I1014 07:18:46.108329 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/09f89569-6fe3-4b3f-9394-a990ad8cde30-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"09f89569-6fe3-4b3f-9394-a990ad8cde30\") " pod="openstack/ovsdbserver-nb-0" Oct 14 07:18:46 crc kubenswrapper[4870]: I1014 07:18:46.108345 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/09f89569-6fe3-4b3f-9394-a990ad8cde30-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"09f89569-6fe3-4b3f-9394-a990ad8cde30\") " pod="openstack/ovsdbserver-nb-0" Oct 14 07:18:46 crc kubenswrapper[4870]: I1014 07:18:46.209516 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/09f89569-6fe3-4b3f-9394-a990ad8cde30-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"09f89569-6fe3-4b3f-9394-a990ad8cde30\") " pod="openstack/ovsdbserver-nb-0" Oct 14 07:18:46 crc kubenswrapper[4870]: I1014 07:18:46.209561 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/09f89569-6fe3-4b3f-9394-a990ad8cde30-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"09f89569-6fe3-4b3f-9394-a990ad8cde30\") " pod="openstack/ovsdbserver-nb-0" Oct 14 07:18:46 crc kubenswrapper[4870]: I1014 07:18:46.209600 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"ovsdbserver-nb-0\" (UID: \"09f89569-6fe3-4b3f-9394-a990ad8cde30\") " pod="openstack/ovsdbserver-nb-0" Oct 14 07:18:46 crc kubenswrapper[4870]: I1014 07:18:46.209634 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/09f89569-6fe3-4b3f-9394-a990ad8cde30-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"09f89569-6fe3-4b3f-9394-a990ad8cde30\") " pod="openstack/ovsdbserver-nb-0" Oct 14 07:18:46 crc kubenswrapper[4870]: I1014 07:18:46.209680 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8hks2\" (UniqueName: \"kubernetes.io/projected/09f89569-6fe3-4b3f-9394-a990ad8cde30-kube-api-access-8hks2\") pod \"ovsdbserver-nb-0\" (UID: \"09f89569-6fe3-4b3f-9394-a990ad8cde30\") " pod="openstack/ovsdbserver-nb-0" Oct 14 07:18:46 crc kubenswrapper[4870]: I1014 07:18:46.209723 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09f89569-6fe3-4b3f-9394-a990ad8cde30-config\") pod \"ovsdbserver-nb-0\" (UID: \"09f89569-6fe3-4b3f-9394-a990ad8cde30\") " pod="openstack/ovsdbserver-nb-0" Oct 14 07:18:46 crc kubenswrapper[4870]: I1014 07:18:46.209759 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/09f89569-6fe3-4b3f-9394-a990ad8cde30-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"09f89569-6fe3-4b3f-9394-a990ad8cde30\") " pod="openstack/ovsdbserver-nb-0" Oct 14 07:18:46 crc kubenswrapper[4870]: I1014 07:18:46.209785 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/09f89569-6fe3-4b3f-9394-a990ad8cde30-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"09f89569-6fe3-4b3f-9394-a990ad8cde30\") " pod="openstack/ovsdbserver-nb-0" Oct 14 07:18:46 crc kubenswrapper[4870]: I1014 07:18:46.210081 4870 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"ovsdbserver-nb-0\" (UID: \"09f89569-6fe3-4b3f-9394-a990ad8cde30\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/ovsdbserver-nb-0" Oct 14 07:18:46 crc kubenswrapper[4870]: I1014 07:18:46.211129 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/09f89569-6fe3-4b3f-9394-a990ad8cde30-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"09f89569-6fe3-4b3f-9394-a990ad8cde30\") " pod="openstack/ovsdbserver-nb-0" Oct 14 07:18:46 crc kubenswrapper[4870]: I1014 07:18:46.211476 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09f89569-6fe3-4b3f-9394-a990ad8cde30-config\") pod \"ovsdbserver-nb-0\" (UID: \"09f89569-6fe3-4b3f-9394-a990ad8cde30\") " pod="openstack/ovsdbserver-nb-0" Oct 14 07:18:46 crc kubenswrapper[4870]: I1014 07:18:46.212047 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/09f89569-6fe3-4b3f-9394-a990ad8cde30-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"09f89569-6fe3-4b3f-9394-a990ad8cde30\") " pod="openstack/ovsdbserver-nb-0" Oct 14 07:18:46 crc kubenswrapper[4870]: I1014 07:18:46.216770 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/09f89569-6fe3-4b3f-9394-a990ad8cde30-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"09f89569-6fe3-4b3f-9394-a990ad8cde30\") " pod="openstack/ovsdbserver-nb-0" Oct 14 07:18:46 crc kubenswrapper[4870]: I1014 07:18:46.216902 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/09f89569-6fe3-4b3f-9394-a990ad8cde30-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"09f89569-6fe3-4b3f-9394-a990ad8cde30\") " pod="openstack/ovsdbserver-nb-0" Oct 14 07:18:46 crc kubenswrapper[4870]: I1014 07:18:46.217009 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/09f89569-6fe3-4b3f-9394-a990ad8cde30-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"09f89569-6fe3-4b3f-9394-a990ad8cde30\") " pod="openstack/ovsdbserver-nb-0" Oct 14 07:18:46 crc kubenswrapper[4870]: I1014 07:18:46.230814 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8hks2\" (UniqueName: \"kubernetes.io/projected/09f89569-6fe3-4b3f-9394-a990ad8cde30-kube-api-access-8hks2\") pod \"ovsdbserver-nb-0\" (UID: \"09f89569-6fe3-4b3f-9394-a990ad8cde30\") " pod="openstack/ovsdbserver-nb-0" Oct 14 07:18:46 crc kubenswrapper[4870]: I1014 07:18:46.257466 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"ovsdbserver-nb-0\" (UID: \"09f89569-6fe3-4b3f-9394-a990ad8cde30\") " pod="openstack/ovsdbserver-nb-0" Oct 14 07:18:46 crc kubenswrapper[4870]: I1014 07:18:46.297787 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-rn98c"] Oct 14 07:18:46 crc kubenswrapper[4870]: I1014 07:18:46.299539 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-rn98c" Oct 14 07:18:46 crc kubenswrapper[4870]: I1014 07:18:46.305417 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Oct 14 07:18:46 crc kubenswrapper[4870]: I1014 07:18:46.305785 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-cjx4k" Oct 14 07:18:46 crc kubenswrapper[4870]: I1014 07:18:46.305975 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Oct 14 07:18:46 crc kubenswrapper[4870]: I1014 07:18:46.315604 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-6482n"] Oct 14 07:18:46 crc kubenswrapper[4870]: I1014 07:18:46.317474 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-6482n" Oct 14 07:18:46 crc kubenswrapper[4870]: I1014 07:18:46.327196 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-rn98c"] Oct 14 07:18:46 crc kubenswrapper[4870]: I1014 07:18:46.368120 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-6482n"] Oct 14 07:18:46 crc kubenswrapper[4870]: I1014 07:18:46.384561 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Oct 14 07:18:46 crc kubenswrapper[4870]: I1014 07:18:46.414160 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2jg2f\" (UniqueName: \"kubernetes.io/projected/503a07f8-2c05-458a-80ff-1abfe973dbd5-kube-api-access-2jg2f\") pod \"ovn-controller-rn98c\" (UID: \"503a07f8-2c05-458a-80ff-1abfe973dbd5\") " pod="openstack/ovn-controller-rn98c" Oct 14 07:18:46 crc kubenswrapper[4870]: I1014 07:18:46.414218 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/91a2e1de-104f-4e14-9192-c6587ed13be1-var-lib\") pod \"ovn-controller-ovs-6482n\" (UID: \"91a2e1de-104f-4e14-9192-c6587ed13be1\") " pod="openstack/ovn-controller-ovs-6482n" Oct 14 07:18:46 crc kubenswrapper[4870]: I1014 07:18:46.414249 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/91a2e1de-104f-4e14-9192-c6587ed13be1-var-log\") pod \"ovn-controller-ovs-6482n\" (UID: \"91a2e1de-104f-4e14-9192-c6587ed13be1\") " pod="openstack/ovn-controller-ovs-6482n" Oct 14 07:18:46 crc kubenswrapper[4870]: I1014 07:18:46.414272 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/91a2e1de-104f-4e14-9192-c6587ed13be1-etc-ovs\") pod \"ovn-controller-ovs-6482n\" (UID: \"91a2e1de-104f-4e14-9192-c6587ed13be1\") " pod="openstack/ovn-controller-ovs-6482n" Oct 14 07:18:46 crc kubenswrapper[4870]: I1014 07:18:46.414311 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ldpld\" (UniqueName: \"kubernetes.io/projected/91a2e1de-104f-4e14-9192-c6587ed13be1-kube-api-access-ldpld\") pod \"ovn-controller-ovs-6482n\" (UID: \"91a2e1de-104f-4e14-9192-c6587ed13be1\") " pod="openstack/ovn-controller-ovs-6482n" Oct 14 07:18:46 crc kubenswrapper[4870]: I1014 07:18:46.414334 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/503a07f8-2c05-458a-80ff-1abfe973dbd5-scripts\") pod \"ovn-controller-rn98c\" (UID: \"503a07f8-2c05-458a-80ff-1abfe973dbd5\") " pod="openstack/ovn-controller-rn98c" Oct 14 07:18:46 crc kubenswrapper[4870]: I1014 07:18:46.414347 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/503a07f8-2c05-458a-80ff-1abfe973dbd5-combined-ca-bundle\") pod \"ovn-controller-rn98c\" (UID: \"503a07f8-2c05-458a-80ff-1abfe973dbd5\") " pod="openstack/ovn-controller-rn98c" Oct 14 07:18:46 crc kubenswrapper[4870]: I1014 07:18:46.414367 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/503a07f8-2c05-458a-80ff-1abfe973dbd5-var-run\") pod \"ovn-controller-rn98c\" (UID: \"503a07f8-2c05-458a-80ff-1abfe973dbd5\") " pod="openstack/ovn-controller-rn98c" Oct 14 07:18:46 crc kubenswrapper[4870]: I1014 07:18:46.414385 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/91a2e1de-104f-4e14-9192-c6587ed13be1-var-run\") pod \"ovn-controller-ovs-6482n\" (UID: \"91a2e1de-104f-4e14-9192-c6587ed13be1\") " pod="openstack/ovn-controller-ovs-6482n" Oct 14 07:18:46 crc kubenswrapper[4870]: I1014 07:18:46.414400 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/91a2e1de-104f-4e14-9192-c6587ed13be1-scripts\") pod \"ovn-controller-ovs-6482n\" (UID: \"91a2e1de-104f-4e14-9192-c6587ed13be1\") " pod="openstack/ovn-controller-ovs-6482n" Oct 14 07:18:46 crc kubenswrapper[4870]: I1014 07:18:46.414413 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/503a07f8-2c05-458a-80ff-1abfe973dbd5-var-log-ovn\") pod \"ovn-controller-rn98c\" (UID: \"503a07f8-2c05-458a-80ff-1abfe973dbd5\") " pod="openstack/ovn-controller-rn98c" Oct 14 07:18:46 crc kubenswrapper[4870]: I1014 07:18:46.414431 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/503a07f8-2c05-458a-80ff-1abfe973dbd5-ovn-controller-tls-certs\") pod \"ovn-controller-rn98c\" (UID: \"503a07f8-2c05-458a-80ff-1abfe973dbd5\") " pod="openstack/ovn-controller-rn98c" Oct 14 07:18:46 crc kubenswrapper[4870]: I1014 07:18:46.414474 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/503a07f8-2c05-458a-80ff-1abfe973dbd5-var-run-ovn\") pod \"ovn-controller-rn98c\" (UID: \"503a07f8-2c05-458a-80ff-1abfe973dbd5\") " pod="openstack/ovn-controller-rn98c" Oct 14 07:18:46 crc kubenswrapper[4870]: I1014 07:18:46.515804 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/91a2e1de-104f-4e14-9192-c6587ed13be1-var-log\") pod \"ovn-controller-ovs-6482n\" (UID: \"91a2e1de-104f-4e14-9192-c6587ed13be1\") " pod="openstack/ovn-controller-ovs-6482n" Oct 14 07:18:46 crc kubenswrapper[4870]: I1014 07:18:46.515853 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/91a2e1de-104f-4e14-9192-c6587ed13be1-etc-ovs\") pod \"ovn-controller-ovs-6482n\" (UID: \"91a2e1de-104f-4e14-9192-c6587ed13be1\") " pod="openstack/ovn-controller-ovs-6482n" Oct 14 07:18:46 crc kubenswrapper[4870]: I1014 07:18:46.515888 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ldpld\" (UniqueName: \"kubernetes.io/projected/91a2e1de-104f-4e14-9192-c6587ed13be1-kube-api-access-ldpld\") pod \"ovn-controller-ovs-6482n\" (UID: \"91a2e1de-104f-4e14-9192-c6587ed13be1\") " pod="openstack/ovn-controller-ovs-6482n" Oct 14 07:18:46 crc kubenswrapper[4870]: I1014 07:18:46.515917 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/503a07f8-2c05-458a-80ff-1abfe973dbd5-scripts\") pod \"ovn-controller-rn98c\" (UID: \"503a07f8-2c05-458a-80ff-1abfe973dbd5\") " pod="openstack/ovn-controller-rn98c" Oct 14 07:18:46 crc kubenswrapper[4870]: I1014 07:18:46.515934 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/503a07f8-2c05-458a-80ff-1abfe973dbd5-combined-ca-bundle\") pod \"ovn-controller-rn98c\" (UID: \"503a07f8-2c05-458a-80ff-1abfe973dbd5\") " pod="openstack/ovn-controller-rn98c" Oct 14 07:18:46 crc kubenswrapper[4870]: I1014 07:18:46.515954 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/503a07f8-2c05-458a-80ff-1abfe973dbd5-var-run\") pod \"ovn-controller-rn98c\" (UID: \"503a07f8-2c05-458a-80ff-1abfe973dbd5\") " pod="openstack/ovn-controller-rn98c" Oct 14 07:18:46 crc kubenswrapper[4870]: I1014 07:18:46.515971 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/91a2e1de-104f-4e14-9192-c6587ed13be1-var-run\") pod \"ovn-controller-ovs-6482n\" (UID: \"91a2e1de-104f-4e14-9192-c6587ed13be1\") " pod="openstack/ovn-controller-ovs-6482n" Oct 14 07:18:46 crc kubenswrapper[4870]: I1014 07:18:46.515987 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/91a2e1de-104f-4e14-9192-c6587ed13be1-scripts\") pod \"ovn-controller-ovs-6482n\" (UID: \"91a2e1de-104f-4e14-9192-c6587ed13be1\") " pod="openstack/ovn-controller-ovs-6482n" Oct 14 07:18:46 crc kubenswrapper[4870]: I1014 07:18:46.516003 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/503a07f8-2c05-458a-80ff-1abfe973dbd5-var-log-ovn\") pod \"ovn-controller-rn98c\" (UID: \"503a07f8-2c05-458a-80ff-1abfe973dbd5\") " pod="openstack/ovn-controller-rn98c" Oct 14 07:18:46 crc kubenswrapper[4870]: I1014 07:18:46.516022 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/503a07f8-2c05-458a-80ff-1abfe973dbd5-ovn-controller-tls-certs\") pod \"ovn-controller-rn98c\" (UID: \"503a07f8-2c05-458a-80ff-1abfe973dbd5\") " pod="openstack/ovn-controller-rn98c" Oct 14 07:18:46 crc kubenswrapper[4870]: I1014 07:18:46.516039 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/503a07f8-2c05-458a-80ff-1abfe973dbd5-var-run-ovn\") pod \"ovn-controller-rn98c\" (UID: \"503a07f8-2c05-458a-80ff-1abfe973dbd5\") " pod="openstack/ovn-controller-rn98c" Oct 14 07:18:46 crc kubenswrapper[4870]: I1014 07:18:46.516077 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2jg2f\" (UniqueName: \"kubernetes.io/projected/503a07f8-2c05-458a-80ff-1abfe973dbd5-kube-api-access-2jg2f\") pod \"ovn-controller-rn98c\" (UID: \"503a07f8-2c05-458a-80ff-1abfe973dbd5\") " pod="openstack/ovn-controller-rn98c" Oct 14 07:18:46 crc kubenswrapper[4870]: I1014 07:18:46.516107 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/91a2e1de-104f-4e14-9192-c6587ed13be1-var-lib\") pod \"ovn-controller-ovs-6482n\" (UID: \"91a2e1de-104f-4e14-9192-c6587ed13be1\") " pod="openstack/ovn-controller-ovs-6482n" Oct 14 07:18:46 crc kubenswrapper[4870]: I1014 07:18:46.516655 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/91a2e1de-104f-4e14-9192-c6587ed13be1-var-lib\") pod \"ovn-controller-ovs-6482n\" (UID: \"91a2e1de-104f-4e14-9192-c6587ed13be1\") " pod="openstack/ovn-controller-ovs-6482n" Oct 14 07:18:46 crc kubenswrapper[4870]: I1014 07:18:46.516783 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/91a2e1de-104f-4e14-9192-c6587ed13be1-var-log\") pod \"ovn-controller-ovs-6482n\" (UID: \"91a2e1de-104f-4e14-9192-c6587ed13be1\") " pod="openstack/ovn-controller-ovs-6482n" Oct 14 07:18:46 crc kubenswrapper[4870]: I1014 07:18:46.516895 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/91a2e1de-104f-4e14-9192-c6587ed13be1-etc-ovs\") pod \"ovn-controller-ovs-6482n\" (UID: \"91a2e1de-104f-4e14-9192-c6587ed13be1\") " pod="openstack/ovn-controller-ovs-6482n" Oct 14 07:18:46 crc kubenswrapper[4870]: I1014 07:18:46.517838 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/503a07f8-2c05-458a-80ff-1abfe973dbd5-var-run\") pod \"ovn-controller-rn98c\" (UID: \"503a07f8-2c05-458a-80ff-1abfe973dbd5\") " pod="openstack/ovn-controller-rn98c" Oct 14 07:18:46 crc kubenswrapper[4870]: I1014 07:18:46.517964 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/91a2e1de-104f-4e14-9192-c6587ed13be1-var-run\") pod \"ovn-controller-ovs-6482n\" (UID: \"91a2e1de-104f-4e14-9192-c6587ed13be1\") " pod="openstack/ovn-controller-ovs-6482n" Oct 14 07:18:46 crc kubenswrapper[4870]: I1014 07:18:46.518088 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/503a07f8-2c05-458a-80ff-1abfe973dbd5-var-log-ovn\") pod \"ovn-controller-rn98c\" (UID: \"503a07f8-2c05-458a-80ff-1abfe973dbd5\") " pod="openstack/ovn-controller-rn98c" Oct 14 07:18:46 crc kubenswrapper[4870]: I1014 07:18:46.518302 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/503a07f8-2c05-458a-80ff-1abfe973dbd5-var-run-ovn\") pod \"ovn-controller-rn98c\" (UID: \"503a07f8-2c05-458a-80ff-1abfe973dbd5\") " pod="openstack/ovn-controller-rn98c" Oct 14 07:18:46 crc kubenswrapper[4870]: I1014 07:18:46.520301 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/91a2e1de-104f-4e14-9192-c6587ed13be1-scripts\") pod \"ovn-controller-ovs-6482n\" (UID: \"91a2e1de-104f-4e14-9192-c6587ed13be1\") " pod="openstack/ovn-controller-ovs-6482n" Oct 14 07:18:46 crc kubenswrapper[4870]: I1014 07:18:46.520390 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/503a07f8-2c05-458a-80ff-1abfe973dbd5-scripts\") pod \"ovn-controller-rn98c\" (UID: \"503a07f8-2c05-458a-80ff-1abfe973dbd5\") " pod="openstack/ovn-controller-rn98c" Oct 14 07:18:46 crc kubenswrapper[4870]: I1014 07:18:46.521159 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/503a07f8-2c05-458a-80ff-1abfe973dbd5-combined-ca-bundle\") pod \"ovn-controller-rn98c\" (UID: \"503a07f8-2c05-458a-80ff-1abfe973dbd5\") " pod="openstack/ovn-controller-rn98c" Oct 14 07:18:46 crc kubenswrapper[4870]: I1014 07:18:46.522382 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/503a07f8-2c05-458a-80ff-1abfe973dbd5-ovn-controller-tls-certs\") pod \"ovn-controller-rn98c\" (UID: \"503a07f8-2c05-458a-80ff-1abfe973dbd5\") " pod="openstack/ovn-controller-rn98c" Oct 14 07:18:46 crc kubenswrapper[4870]: I1014 07:18:46.539981 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2jg2f\" (UniqueName: \"kubernetes.io/projected/503a07f8-2c05-458a-80ff-1abfe973dbd5-kube-api-access-2jg2f\") pod \"ovn-controller-rn98c\" (UID: \"503a07f8-2c05-458a-80ff-1abfe973dbd5\") " pod="openstack/ovn-controller-rn98c" Oct 14 07:18:46 crc kubenswrapper[4870]: I1014 07:18:46.540466 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ldpld\" (UniqueName: \"kubernetes.io/projected/91a2e1de-104f-4e14-9192-c6587ed13be1-kube-api-access-ldpld\") pod \"ovn-controller-ovs-6482n\" (UID: \"91a2e1de-104f-4e14-9192-c6587ed13be1\") " pod="openstack/ovn-controller-ovs-6482n" Oct 14 07:18:46 crc kubenswrapper[4870]: I1014 07:18:46.621777 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-rn98c" Oct 14 07:18:46 crc kubenswrapper[4870]: I1014 07:18:46.636326 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-6482n" Oct 14 07:18:49 crc kubenswrapper[4870]: I1014 07:18:49.552638 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 14 07:18:49 crc kubenswrapper[4870]: I1014 07:18:49.556242 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Oct 14 07:18:49 crc kubenswrapper[4870]: I1014 07:18:49.562476 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-dgjk6" Oct 14 07:18:49 crc kubenswrapper[4870]: I1014 07:18:49.562586 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Oct 14 07:18:49 crc kubenswrapper[4870]: I1014 07:18:49.562627 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Oct 14 07:18:49 crc kubenswrapper[4870]: I1014 07:18:49.563601 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Oct 14 07:18:49 crc kubenswrapper[4870]: I1014 07:18:49.565743 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 14 07:18:49 crc kubenswrapper[4870]: I1014 07:18:49.672055 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/36717039-e3e2-4b74-b612-655556122574-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"36717039-e3e2-4b74-b612-655556122574\") " pod="openstack/ovsdbserver-sb-0" Oct 14 07:18:49 crc kubenswrapper[4870]: I1014 07:18:49.672129 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"ovsdbserver-sb-0\" (UID: \"36717039-e3e2-4b74-b612-655556122574\") " pod="openstack/ovsdbserver-sb-0" Oct 14 07:18:49 crc kubenswrapper[4870]: I1014 07:18:49.672163 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/36717039-e3e2-4b74-b612-655556122574-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"36717039-e3e2-4b74-b612-655556122574\") " pod="openstack/ovsdbserver-sb-0" Oct 14 07:18:49 crc kubenswrapper[4870]: I1014 07:18:49.672198 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2gt69\" (UniqueName: \"kubernetes.io/projected/36717039-e3e2-4b74-b612-655556122574-kube-api-access-2gt69\") pod \"ovsdbserver-sb-0\" (UID: \"36717039-e3e2-4b74-b612-655556122574\") " pod="openstack/ovsdbserver-sb-0" Oct 14 07:18:49 crc kubenswrapper[4870]: I1014 07:18:49.672232 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/36717039-e3e2-4b74-b612-655556122574-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"36717039-e3e2-4b74-b612-655556122574\") " pod="openstack/ovsdbserver-sb-0" Oct 14 07:18:49 crc kubenswrapper[4870]: I1014 07:18:49.672301 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/36717039-e3e2-4b74-b612-655556122574-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"36717039-e3e2-4b74-b612-655556122574\") " pod="openstack/ovsdbserver-sb-0" Oct 14 07:18:49 crc kubenswrapper[4870]: I1014 07:18:49.672500 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36717039-e3e2-4b74-b612-655556122574-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"36717039-e3e2-4b74-b612-655556122574\") " pod="openstack/ovsdbserver-sb-0" Oct 14 07:18:49 crc kubenswrapper[4870]: I1014 07:18:49.672572 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/36717039-e3e2-4b74-b612-655556122574-config\") pod \"ovsdbserver-sb-0\" (UID: \"36717039-e3e2-4b74-b612-655556122574\") " pod="openstack/ovsdbserver-sb-0" Oct 14 07:18:49 crc kubenswrapper[4870]: I1014 07:18:49.775056 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/36717039-e3e2-4b74-b612-655556122574-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"36717039-e3e2-4b74-b612-655556122574\") " pod="openstack/ovsdbserver-sb-0" Oct 14 07:18:49 crc kubenswrapper[4870]: I1014 07:18:49.775258 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36717039-e3e2-4b74-b612-655556122574-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"36717039-e3e2-4b74-b612-655556122574\") " pod="openstack/ovsdbserver-sb-0" Oct 14 07:18:49 crc kubenswrapper[4870]: I1014 07:18:49.775318 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/36717039-e3e2-4b74-b612-655556122574-config\") pod \"ovsdbserver-sb-0\" (UID: \"36717039-e3e2-4b74-b612-655556122574\") " pod="openstack/ovsdbserver-sb-0" Oct 14 07:18:49 crc kubenswrapper[4870]: I1014 07:18:49.775393 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/36717039-e3e2-4b74-b612-655556122574-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"36717039-e3e2-4b74-b612-655556122574\") " pod="openstack/ovsdbserver-sb-0" Oct 14 07:18:49 crc kubenswrapper[4870]: I1014 07:18:49.775505 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"ovsdbserver-sb-0\" (UID: \"36717039-e3e2-4b74-b612-655556122574\") " pod="openstack/ovsdbserver-sb-0" Oct 14 07:18:49 crc kubenswrapper[4870]: I1014 07:18:49.775632 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/36717039-e3e2-4b74-b612-655556122574-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"36717039-e3e2-4b74-b612-655556122574\") " pod="openstack/ovsdbserver-sb-0" Oct 14 07:18:49 crc kubenswrapper[4870]: I1014 07:18:49.775833 4870 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"ovsdbserver-sb-0\" (UID: \"36717039-e3e2-4b74-b612-655556122574\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/ovsdbserver-sb-0" Oct 14 07:18:49 crc kubenswrapper[4870]: I1014 07:18:49.776207 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/36717039-e3e2-4b74-b612-655556122574-config\") pod \"ovsdbserver-sb-0\" (UID: \"36717039-e3e2-4b74-b612-655556122574\") " pod="openstack/ovsdbserver-sb-0" Oct 14 07:18:49 crc kubenswrapper[4870]: I1014 07:18:49.776577 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2gt69\" (UniqueName: \"kubernetes.io/projected/36717039-e3e2-4b74-b612-655556122574-kube-api-access-2gt69\") pod \"ovsdbserver-sb-0\" (UID: \"36717039-e3e2-4b74-b612-655556122574\") " pod="openstack/ovsdbserver-sb-0" Oct 14 07:18:49 crc kubenswrapper[4870]: I1014 07:18:49.776652 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/36717039-e3e2-4b74-b612-655556122574-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"36717039-e3e2-4b74-b612-655556122574\") " pod="openstack/ovsdbserver-sb-0" Oct 14 07:18:49 crc kubenswrapper[4870]: I1014 07:18:49.777309 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/36717039-e3e2-4b74-b612-655556122574-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"36717039-e3e2-4b74-b612-655556122574\") " pod="openstack/ovsdbserver-sb-0" Oct 14 07:18:49 crc kubenswrapper[4870]: I1014 07:18:49.777560 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/36717039-e3e2-4b74-b612-655556122574-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"36717039-e3e2-4b74-b612-655556122574\") " pod="openstack/ovsdbserver-sb-0" Oct 14 07:18:49 crc kubenswrapper[4870]: I1014 07:18:49.782047 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/36717039-e3e2-4b74-b612-655556122574-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"36717039-e3e2-4b74-b612-655556122574\") " pod="openstack/ovsdbserver-sb-0" Oct 14 07:18:49 crc kubenswrapper[4870]: I1014 07:18:49.782208 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36717039-e3e2-4b74-b612-655556122574-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"36717039-e3e2-4b74-b612-655556122574\") " pod="openstack/ovsdbserver-sb-0" Oct 14 07:18:49 crc kubenswrapper[4870]: I1014 07:18:49.793396 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/36717039-e3e2-4b74-b612-655556122574-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"36717039-e3e2-4b74-b612-655556122574\") " pod="openstack/ovsdbserver-sb-0" Oct 14 07:18:49 crc kubenswrapper[4870]: I1014 07:18:49.803858 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2gt69\" (UniqueName: \"kubernetes.io/projected/36717039-e3e2-4b74-b612-655556122574-kube-api-access-2gt69\") pod \"ovsdbserver-sb-0\" (UID: \"36717039-e3e2-4b74-b612-655556122574\") " pod="openstack/ovsdbserver-sb-0" Oct 14 07:18:49 crc kubenswrapper[4870]: I1014 07:18:49.808725 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"ovsdbserver-sb-0\" (UID: \"36717039-e3e2-4b74-b612-655556122574\") " pod="openstack/ovsdbserver-sb-0" Oct 14 07:18:49 crc kubenswrapper[4870]: I1014 07:18:49.886583 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Oct 14 07:18:52 crc kubenswrapper[4870]: I1014 07:18:52.988282 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"3de2167a-7663-4b00-9743-229ff1864cb8","Type":"ContainerStarted","Data":"e83ba8ed593070f7993afb4580d91a1da037ccd817bcfc2a6029ab4e2b0edf86"} Oct 14 07:18:53 crc kubenswrapper[4870]: E1014 07:18:53.563561 4870 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:673685cea6ea2dbd78bcb555955c1b9f05ea26018f79ee34494256a5f2d7b74a" Oct 14 07:18:53 crc kubenswrapper[4870]: E1014 07:18:53.563791 4870 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:673685cea6ea2dbd78bcb555955c1b9f05ea26018f79ee34494256a5f2d7b74a,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-5vllq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-6948694bd9-5r72k_openstack(f92a9598-06e8-4b09-9504-20d07f4f6281): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 14 07:18:53 crc kubenswrapper[4870]: E1014 07:18:53.564965 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-6948694bd9-5r72k" podUID="f92a9598-06e8-4b09-9504-20d07f4f6281" Oct 14 07:18:53 crc kubenswrapper[4870]: E1014 07:18:53.598370 4870 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:673685cea6ea2dbd78bcb555955c1b9f05ea26018f79ee34494256a5f2d7b74a" Oct 14 07:18:53 crc kubenswrapper[4870]: E1014 07:18:53.598596 4870 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:673685cea6ea2dbd78bcb555955c1b9f05ea26018f79ee34494256a5f2d7b74a,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n68chd6h679hbfh55fhc6h5ffh5d8h94h56ch589hb4hc5h57bh677hcdh655h8dh667h675h654h66ch567h8fh659h5b4h675h566h55bh54h67dh6dq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-lfz64,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-86f694bf-k7p29_openstack(86d52d2f-b41b-4c56-af26-b32e8caaf225): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 14 07:18:53 crc kubenswrapper[4870]: E1014 07:18:53.598904 4870 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:673685cea6ea2dbd78bcb555955c1b9f05ea26018f79ee34494256a5f2d7b74a" Oct 14 07:18:53 crc kubenswrapper[4870]: E1014 07:18:53.599088 4870 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:673685cea6ea2dbd78bcb555955c1b9f05ea26018f79ee34494256a5f2d7b74a,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-57pz4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-5d487d97d7-hd6gw_openstack(84d82363-9888-49ac-8e1d-5f98b834a7e1): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 14 07:18:53 crc kubenswrapper[4870]: E1014 07:18:53.600076 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-86f694bf-k7p29" podUID="86d52d2f-b41b-4c56-af26-b32e8caaf225" Oct 14 07:18:53 crc kubenswrapper[4870]: E1014 07:18:53.601186 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-5d487d97d7-hd6gw" podUID="84d82363-9888-49ac-8e1d-5f98b834a7e1" Oct 14 07:18:53 crc kubenswrapper[4870]: E1014 07:18:53.678346 4870 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:673685cea6ea2dbd78bcb555955c1b9f05ea26018f79ee34494256a5f2d7b74a" Oct 14 07:18:53 crc kubenswrapper[4870]: E1014 07:18:53.678967 4870 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:673685cea6ea2dbd78bcb555955c1b9f05ea26018f79ee34494256a5f2d7b74a,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n659h4h664hbh658h587h67ch89h587h8fh679hc6hf9h55fh644h5d5h698h68dh5cdh5ffh669h54ch9h689hb8hd4h5bfhd8h5d7h5fh665h574q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-psk44,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-7869c47d6c-44h7s_openstack(11aab367-2351-4568-9a45-43c79d419bca): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 14 07:18:53 crc kubenswrapper[4870]: E1014 07:18:53.680499 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-7869c47d6c-44h7s" podUID="11aab367-2351-4568-9a45-43c79d419bca" Oct 14 07:18:53 crc kubenswrapper[4870]: I1014 07:18:53.951583 4870 patch_prober.go:28] interesting pod/machine-config-daemon-7tvc7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 07:18:53 crc kubenswrapper[4870]: I1014 07:18:53.951635 4870 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 07:18:53 crc kubenswrapper[4870]: I1014 07:18:53.951684 4870 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" Oct 14 07:18:53 crc kubenswrapper[4870]: I1014 07:18:53.952384 4870 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"afa79ed981f5045c93ba7103bf82588637ff210f79db4fa50132dc5fa3e91338"} pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 14 07:18:53 crc kubenswrapper[4870]: I1014 07:18:53.952500 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerName="machine-config-daemon" containerID="cri-o://afa79ed981f5045c93ba7103bf82588637ff210f79db4fa50132dc5fa3e91338" gracePeriod=600 Oct 14 07:18:53 crc kubenswrapper[4870]: E1014 07:18:53.997147 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:673685cea6ea2dbd78bcb555955c1b9f05ea26018f79ee34494256a5f2d7b74a\\\"\"" pod="openstack/dnsmasq-dns-7869c47d6c-44h7s" podUID="11aab367-2351-4568-9a45-43c79d419bca" Oct 14 07:18:53 crc kubenswrapper[4870]: E1014 07:18:53.998069 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:673685cea6ea2dbd78bcb555955c1b9f05ea26018f79ee34494256a5f2d7b74a\\\"\"" pod="openstack/dnsmasq-dns-86f694bf-k7p29" podUID="86d52d2f-b41b-4c56-af26-b32e8caaf225" Oct 14 07:18:54 crc kubenswrapper[4870]: I1014 07:18:54.044782 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 14 07:18:54 crc kubenswrapper[4870]: I1014 07:18:54.057102 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Oct 14 07:18:54 crc kubenswrapper[4870]: I1014 07:18:54.143366 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 14 07:18:54 crc kubenswrapper[4870]: I1014 07:18:54.151277 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 14 07:18:54 crc kubenswrapper[4870]: I1014 07:18:54.252075 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-rn98c"] Oct 14 07:18:54 crc kubenswrapper[4870]: W1014 07:18:54.260465 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod503a07f8_2c05_458a_80ff_1abfe973dbd5.slice/crio-39ecc2d2eb7b47544395be53fdac9994576cc65c656cd6d71aaa3811d2ed728b WatchSource:0}: Error finding container 39ecc2d2eb7b47544395be53fdac9994576cc65c656cd6d71aaa3811d2ed728b: Status 404 returned error can't find the container with id 39ecc2d2eb7b47544395be53fdac9994576cc65c656cd6d71aaa3811d2ed728b Oct 14 07:18:54 crc kubenswrapper[4870]: I1014 07:18:54.281141 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 14 07:18:54 crc kubenswrapper[4870]: W1014 07:18:54.284370 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod09f89569_6fe3_4b3f_9394_a990ad8cde30.slice/crio-bc1f5cf5d7dba356af68fb612233bc6c1440a741d5141cff777e43bb708f2fb5 WatchSource:0}: Error finding container bc1f5cf5d7dba356af68fb612233bc6c1440a741d5141cff777e43bb708f2fb5: Status 404 returned error can't find the container with id bc1f5cf5d7dba356af68fb612233bc6c1440a741d5141cff777e43bb708f2fb5 Oct 14 07:18:54 crc kubenswrapper[4870]: I1014 07:18:54.446889 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-6482n"] Oct 14 07:18:54 crc kubenswrapper[4870]: I1014 07:18:54.482555 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6948694bd9-5r72k" Oct 14 07:18:54 crc kubenswrapper[4870]: I1014 07:18:54.514665 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 14 07:18:54 crc kubenswrapper[4870]: I1014 07:18:54.521855 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d487d97d7-hd6gw" Oct 14 07:18:54 crc kubenswrapper[4870]: I1014 07:18:54.576336 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f92a9598-06e8-4b09-9504-20d07f4f6281-dns-svc\") pod \"f92a9598-06e8-4b09-9504-20d07f4f6281\" (UID: \"f92a9598-06e8-4b09-9504-20d07f4f6281\") " Oct 14 07:18:54 crc kubenswrapper[4870]: I1014 07:18:54.576524 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/84d82363-9888-49ac-8e1d-5f98b834a7e1-config\") pod \"84d82363-9888-49ac-8e1d-5f98b834a7e1\" (UID: \"84d82363-9888-49ac-8e1d-5f98b834a7e1\") " Oct 14 07:18:54 crc kubenswrapper[4870]: I1014 07:18:54.576558 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5vllq\" (UniqueName: \"kubernetes.io/projected/f92a9598-06e8-4b09-9504-20d07f4f6281-kube-api-access-5vllq\") pod \"f92a9598-06e8-4b09-9504-20d07f4f6281\" (UID: \"f92a9598-06e8-4b09-9504-20d07f4f6281\") " Oct 14 07:18:54 crc kubenswrapper[4870]: I1014 07:18:54.576591 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-57pz4\" (UniqueName: \"kubernetes.io/projected/84d82363-9888-49ac-8e1d-5f98b834a7e1-kube-api-access-57pz4\") pod \"84d82363-9888-49ac-8e1d-5f98b834a7e1\" (UID: \"84d82363-9888-49ac-8e1d-5f98b834a7e1\") " Oct 14 07:18:54 crc kubenswrapper[4870]: I1014 07:18:54.576626 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f92a9598-06e8-4b09-9504-20d07f4f6281-config\") pod \"f92a9598-06e8-4b09-9504-20d07f4f6281\" (UID: \"f92a9598-06e8-4b09-9504-20d07f4f6281\") " Oct 14 07:18:54 crc kubenswrapper[4870]: I1014 07:18:54.576990 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f92a9598-06e8-4b09-9504-20d07f4f6281-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "f92a9598-06e8-4b09-9504-20d07f4f6281" (UID: "f92a9598-06e8-4b09-9504-20d07f4f6281"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:18:54 crc kubenswrapper[4870]: I1014 07:18:54.577910 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/84d82363-9888-49ac-8e1d-5f98b834a7e1-config" (OuterVolumeSpecName: "config") pod "84d82363-9888-49ac-8e1d-5f98b834a7e1" (UID: "84d82363-9888-49ac-8e1d-5f98b834a7e1"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:18:54 crc kubenswrapper[4870]: I1014 07:18:54.578217 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f92a9598-06e8-4b09-9504-20d07f4f6281-config" (OuterVolumeSpecName: "config") pod "f92a9598-06e8-4b09-9504-20d07f4f6281" (UID: "f92a9598-06e8-4b09-9504-20d07f4f6281"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:18:54 crc kubenswrapper[4870]: I1014 07:18:54.582973 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f92a9598-06e8-4b09-9504-20d07f4f6281-kube-api-access-5vllq" (OuterVolumeSpecName: "kube-api-access-5vllq") pod "f92a9598-06e8-4b09-9504-20d07f4f6281" (UID: "f92a9598-06e8-4b09-9504-20d07f4f6281"). InnerVolumeSpecName "kube-api-access-5vllq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:18:54 crc kubenswrapper[4870]: I1014 07:18:54.583827 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/84d82363-9888-49ac-8e1d-5f98b834a7e1-kube-api-access-57pz4" (OuterVolumeSpecName: "kube-api-access-57pz4") pod "84d82363-9888-49ac-8e1d-5f98b834a7e1" (UID: "84d82363-9888-49ac-8e1d-5f98b834a7e1"). InnerVolumeSpecName "kube-api-access-57pz4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:18:54 crc kubenswrapper[4870]: I1014 07:18:54.678134 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-57pz4\" (UniqueName: \"kubernetes.io/projected/84d82363-9888-49ac-8e1d-5f98b834a7e1-kube-api-access-57pz4\") on node \"crc\" DevicePath \"\"" Oct 14 07:18:54 crc kubenswrapper[4870]: I1014 07:18:54.678178 4870 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f92a9598-06e8-4b09-9504-20d07f4f6281-config\") on node \"crc\" DevicePath \"\"" Oct 14 07:18:54 crc kubenswrapper[4870]: I1014 07:18:54.678192 4870 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f92a9598-06e8-4b09-9504-20d07f4f6281-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 14 07:18:54 crc kubenswrapper[4870]: I1014 07:18:54.678205 4870 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/84d82363-9888-49ac-8e1d-5f98b834a7e1-config\") on node \"crc\" DevicePath \"\"" Oct 14 07:18:54 crc kubenswrapper[4870]: I1014 07:18:54.678218 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5vllq\" (UniqueName: \"kubernetes.io/projected/f92a9598-06e8-4b09-9504-20d07f4f6281-kube-api-access-5vllq\") on node \"crc\" DevicePath \"\"" Oct 14 07:18:55 crc kubenswrapper[4870]: I1014 07:18:55.007488 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"5212ba20-83c3-41ff-a3d9-5a315f009c1b","Type":"ContainerStarted","Data":"d98e1ffb219c955289fa2faf4924c6941b1140c0dffec8bed2ba720a0fc61177"} Oct 14 07:18:55 crc kubenswrapper[4870]: I1014 07:18:55.008396 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-6482n" event={"ID":"91a2e1de-104f-4e14-9192-c6587ed13be1","Type":"ContainerStarted","Data":"ca705e1a5041f27d32a37ea1dfaf6496c090f4c3a3d80f271e9fc688a3c8e78f"} Oct 14 07:18:55 crc kubenswrapper[4870]: I1014 07:18:55.010013 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"36717039-e3e2-4b74-b612-655556122574","Type":"ContainerStarted","Data":"b62fd674ae95524d37f10dfead22b74ed9112cfa1180b36f995c9dd46c127b3e"} Oct 14 07:18:55 crc kubenswrapper[4870]: I1014 07:18:55.011740 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6948694bd9-5r72k" event={"ID":"f92a9598-06e8-4b09-9504-20d07f4f6281","Type":"ContainerDied","Data":"e46f6f733bc263a47b96e5ddba18087043c08e13d14474feff6bf718f00ed70c"} Oct 14 07:18:55 crc kubenswrapper[4870]: I1014 07:18:55.011765 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6948694bd9-5r72k" Oct 14 07:18:55 crc kubenswrapper[4870]: I1014 07:18:55.013731 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"09f89569-6fe3-4b3f-9394-a990ad8cde30","Type":"ContainerStarted","Data":"bc1f5cf5d7dba356af68fb612233bc6c1440a741d5141cff777e43bb708f2fb5"} Oct 14 07:18:55 crc kubenswrapper[4870]: I1014 07:18:55.015607 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"4d74ef50-1861-4d70-b781-f92d5f4b35ba","Type":"ContainerStarted","Data":"474bb87795c2229a29716f4b23006219a79aa8b329f0418dda86379779caeef2"} Oct 14 07:18:55 crc kubenswrapper[4870]: I1014 07:18:55.017416 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"b2d073e2-e7bb-44c9-b074-c39afeae16b8","Type":"ContainerStarted","Data":"bd238754e5c63ab7ff61dc40da6abf66375324653fb9c0c31ea19bac500142ed"} Oct 14 07:18:55 crc kubenswrapper[4870]: I1014 07:18:55.019529 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"3bd1e649-3d09-4cda-921f-ac9a84a5066e","Type":"ContainerStarted","Data":"052bde8548ecfddf46ee9484a1311dff89a89809aee66740b22b81c7a218cd25"} Oct 14 07:18:55 crc kubenswrapper[4870]: I1014 07:18:55.021015 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-rn98c" event={"ID":"503a07f8-2c05-458a-80ff-1abfe973dbd5","Type":"ContainerStarted","Data":"39ecc2d2eb7b47544395be53fdac9994576cc65c656cd6d71aaa3811d2ed728b"} Oct 14 07:18:55 crc kubenswrapper[4870]: I1014 07:18:55.025758 4870 generic.go:334] "Generic (PLEG): container finished" podID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerID="afa79ed981f5045c93ba7103bf82588637ff210f79db4fa50132dc5fa3e91338" exitCode=0 Oct 14 07:18:55 crc kubenswrapper[4870]: I1014 07:18:55.025802 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" event={"ID":"874a5e45-dffd-4d17-b609-b7d3ed2eab07","Type":"ContainerDied","Data":"afa79ed981f5045c93ba7103bf82588637ff210f79db4fa50132dc5fa3e91338"} Oct 14 07:18:55 crc kubenswrapper[4870]: I1014 07:18:55.025874 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" event={"ID":"874a5e45-dffd-4d17-b609-b7d3ed2eab07","Type":"ContainerStarted","Data":"5389ca0979ade5a4bccec9faa4aa51cda4fcec373b2752e9cf9faa85bf74bfdf"} Oct 14 07:18:55 crc kubenswrapper[4870]: I1014 07:18:55.025953 4870 scope.go:117] "RemoveContainer" containerID="b5bea711ed09c7366862c920d932bb6dfefddcdd90ae2ab8981cfd9449e357a2" Oct 14 07:18:55 crc kubenswrapper[4870]: I1014 07:18:55.030570 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d487d97d7-hd6gw" event={"ID":"84d82363-9888-49ac-8e1d-5f98b834a7e1","Type":"ContainerDied","Data":"a0379a48d121338e28ce393309fc934509f49a899fc7d9ca72f56b49a92f65a1"} Oct 14 07:18:55 crc kubenswrapper[4870]: I1014 07:18:55.030660 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d487d97d7-hd6gw" Oct 14 07:18:55 crc kubenswrapper[4870]: I1014 07:18:55.366345 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6948694bd9-5r72k"] Oct 14 07:18:55 crc kubenswrapper[4870]: I1014 07:18:55.384018 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6948694bd9-5r72k"] Oct 14 07:18:55 crc kubenswrapper[4870]: I1014 07:18:55.395309 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5d487d97d7-hd6gw"] Oct 14 07:18:55 crc kubenswrapper[4870]: I1014 07:18:55.401253 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5d487d97d7-hd6gw"] Oct 14 07:18:57 crc kubenswrapper[4870]: I1014 07:18:57.055554 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="84d82363-9888-49ac-8e1d-5f98b834a7e1" path="/var/lib/kubelet/pods/84d82363-9888-49ac-8e1d-5f98b834a7e1/volumes" Oct 14 07:18:57 crc kubenswrapper[4870]: I1014 07:18:57.056690 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f92a9598-06e8-4b09-9504-20d07f4f6281" path="/var/lib/kubelet/pods/f92a9598-06e8-4b09-9504-20d07f4f6281/volumes" Oct 14 07:18:58 crc kubenswrapper[4870]: I1014 07:18:58.066098 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"5212ba20-83c3-41ff-a3d9-5a315f009c1b","Type":"ContainerStarted","Data":"80776853a799d9822a89d9d86e1010ddb56ba32acd904f622c45923cef6744ac"} Oct 14 07:18:58 crc kubenswrapper[4870]: I1014 07:18:58.067542 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"da8a9731-f917-462f-9932-b37b6abb9a64","Type":"ContainerStarted","Data":"fdc546ad7b75936b8391d1c25cb200fd0e53c0d3629443c510ae84c5a357cc6e"} Oct 14 07:19:03 crc kubenswrapper[4870]: I1014 07:19:03.124916 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"09f89569-6fe3-4b3f-9394-a990ad8cde30","Type":"ContainerStarted","Data":"76d17f5aef5402fa0f6135921f3ba0aeb9f5640152e827b4385348efdbe620d3"} Oct 14 07:19:03 crc kubenswrapper[4870]: I1014 07:19:03.126812 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"4d74ef50-1861-4d70-b781-f92d5f4b35ba","Type":"ContainerStarted","Data":"13f6a7296cd23cf0aa5a1d5c03164c9801caa1ced55914f7f42966f05f4d9aa7"} Oct 14 07:19:03 crc kubenswrapper[4870]: I1014 07:19:03.135098 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"36717039-e3e2-4b74-b612-655556122574","Type":"ContainerStarted","Data":"afc051c5e25bf94e2cbc69eabd5bf3dc5066c8c7364b3f767d2bb33960df9191"} Oct 14 07:19:03 crc kubenswrapper[4870]: I1014 07:19:03.138869 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"3bd1e649-3d09-4cda-921f-ac9a84a5066e","Type":"ContainerStarted","Data":"cd6dabf959a5bb827e348f8580f0325c6287615f71bc7279bb2e2744f5cf4f7c"} Oct 14 07:19:03 crc kubenswrapper[4870]: I1014 07:19:03.139549 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Oct 14 07:19:03 crc kubenswrapper[4870]: I1014 07:19:03.141339 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"3de2167a-7663-4b00-9743-229ff1864cb8","Type":"ContainerStarted","Data":"eb1654e601e030081e70bbe69e7b024b64dd889e97e317b911a3c3468c7fcfb8"} Oct 14 07:19:03 crc kubenswrapper[4870]: I1014 07:19:03.183349 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=15.598325466 podStartE2EDuration="23.183324229s" podCreationTimestamp="2025-10-14 07:18:40 +0000 UTC" firstStartedPulling="2025-10-14 07:18:54.085976606 +0000 UTC m=+1069.783336977" lastFinishedPulling="2025-10-14 07:19:01.670975369 +0000 UTC m=+1077.368335740" observedRunningTime="2025-10-14 07:19:03.17981453 +0000 UTC m=+1078.877174901" watchObservedRunningTime="2025-10-14 07:19:03.183324229 +0000 UTC m=+1078.880684600" Oct 14 07:19:04 crc kubenswrapper[4870]: I1014 07:19:04.151852 4870 generic.go:334] "Generic (PLEG): container finished" podID="91a2e1de-104f-4e14-9192-c6587ed13be1" containerID="31bee80630d9bb4bd377e738c7f57f0caeb9aafbcc1210717378d78bab78ce72" exitCode=0 Oct 14 07:19:04 crc kubenswrapper[4870]: I1014 07:19:04.151917 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-6482n" event={"ID":"91a2e1de-104f-4e14-9192-c6587ed13be1","Type":"ContainerDied","Data":"31bee80630d9bb4bd377e738c7f57f0caeb9aafbcc1210717378d78bab78ce72"} Oct 14 07:19:04 crc kubenswrapper[4870]: I1014 07:19:04.156695 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"b2d073e2-e7bb-44c9-b074-c39afeae16b8","Type":"ContainerStarted","Data":"ffd73f77d45e4b6cd97b896c63da4c555a8d71621706a5dd9aa9fade449b5ce5"} Oct 14 07:19:04 crc kubenswrapper[4870]: I1014 07:19:04.156815 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Oct 14 07:19:04 crc kubenswrapper[4870]: I1014 07:19:04.159098 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-rn98c" event={"ID":"503a07f8-2c05-458a-80ff-1abfe973dbd5","Type":"ContainerStarted","Data":"124ba16adf714b9b67c2030611bc8afc10924c09b62003e8aa1a0045f6015014"} Oct 14 07:19:04 crc kubenswrapper[4870]: I1014 07:19:04.196313 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=13.335716339 podStartE2EDuration="22.196291888s" podCreationTimestamp="2025-10-14 07:18:42 +0000 UTC" firstStartedPulling="2025-10-14 07:18:54.074577958 +0000 UTC m=+1069.771938329" lastFinishedPulling="2025-10-14 07:19:02.935153497 +0000 UTC m=+1078.632513878" observedRunningTime="2025-10-14 07:19:04.19280205 +0000 UTC m=+1079.890162421" watchObservedRunningTime="2025-10-14 07:19:04.196291888 +0000 UTC m=+1079.893652269" Oct 14 07:19:04 crc kubenswrapper[4870]: I1014 07:19:04.217618 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-rn98c" podStartSLOduration=10.47321056 podStartE2EDuration="18.217594315s" podCreationTimestamp="2025-10-14 07:18:46 +0000 UTC" firstStartedPulling="2025-10-14 07:18:54.262865569 +0000 UTC m=+1069.960225930" lastFinishedPulling="2025-10-14 07:19:02.007249314 +0000 UTC m=+1077.704609685" observedRunningTime="2025-10-14 07:19:04.212069646 +0000 UTC m=+1079.909430037" watchObservedRunningTime="2025-10-14 07:19:04.217594315 +0000 UTC m=+1079.914954696" Oct 14 07:19:05 crc kubenswrapper[4870]: I1014 07:19:05.174942 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-6482n" event={"ID":"91a2e1de-104f-4e14-9192-c6587ed13be1","Type":"ContainerStarted","Data":"6f14317c64b4cb248d4254d308d86fadbdb28d2daee87abbf75414c219c2c477"} Oct 14 07:19:05 crc kubenswrapper[4870]: I1014 07:19:05.175473 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-rn98c" Oct 14 07:19:06 crc kubenswrapper[4870]: I1014 07:19:06.190844 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-6482n" event={"ID":"91a2e1de-104f-4e14-9192-c6587ed13be1","Type":"ContainerStarted","Data":"72ead03af69bb2d536002fe16f85a88974968529ec05c0e9e78254a3f796b690"} Oct 14 07:19:06 crc kubenswrapper[4870]: I1014 07:19:06.191429 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-6482n" Oct 14 07:19:06 crc kubenswrapper[4870]: I1014 07:19:06.191557 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-6482n" Oct 14 07:19:06 crc kubenswrapper[4870]: I1014 07:19:06.215028 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-6482n" podStartSLOduration=12.996228511 podStartE2EDuration="20.215007824s" podCreationTimestamp="2025-10-14 07:18:46 +0000 UTC" firstStartedPulling="2025-10-14 07:18:54.452197776 +0000 UTC m=+1070.149558147" lastFinishedPulling="2025-10-14 07:19:01.670977099 +0000 UTC m=+1077.368337460" observedRunningTime="2025-10-14 07:19:06.210137831 +0000 UTC m=+1081.907498202" watchObservedRunningTime="2025-10-14 07:19:06.215007824 +0000 UTC m=+1081.912368195" Oct 14 07:19:07 crc kubenswrapper[4870]: I1014 07:19:07.202539 4870 generic.go:334] "Generic (PLEG): container finished" podID="3de2167a-7663-4b00-9743-229ff1864cb8" containerID="eb1654e601e030081e70bbe69e7b024b64dd889e97e317b911a3c3468c7fcfb8" exitCode=0 Oct 14 07:19:07 crc kubenswrapper[4870]: I1014 07:19:07.202660 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"3de2167a-7663-4b00-9743-229ff1864cb8","Type":"ContainerDied","Data":"eb1654e601e030081e70bbe69e7b024b64dd889e97e317b911a3c3468c7fcfb8"} Oct 14 07:19:07 crc kubenswrapper[4870]: I1014 07:19:07.209430 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"09f89569-6fe3-4b3f-9394-a990ad8cde30","Type":"ContainerStarted","Data":"b03ec54549f3acea91303d6cc01b2627a9ebd5ec3e78dcb4527b100efe1ebd9c"} Oct 14 07:19:07 crc kubenswrapper[4870]: I1014 07:19:07.212379 4870 generic.go:334] "Generic (PLEG): container finished" podID="4d74ef50-1861-4d70-b781-f92d5f4b35ba" containerID="13f6a7296cd23cf0aa5a1d5c03164c9801caa1ced55914f7f42966f05f4d9aa7" exitCode=0 Oct 14 07:19:07 crc kubenswrapper[4870]: I1014 07:19:07.212989 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"4d74ef50-1861-4d70-b781-f92d5f4b35ba","Type":"ContainerDied","Data":"13f6a7296cd23cf0aa5a1d5c03164c9801caa1ced55914f7f42966f05f4d9aa7"} Oct 14 07:19:07 crc kubenswrapper[4870]: I1014 07:19:07.216358 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"36717039-e3e2-4b74-b612-655556122574","Type":"ContainerStarted","Data":"660968003babd11d1e6d54c59b0e91ea2368857634bd407e25468d0645470948"} Oct 14 07:19:07 crc kubenswrapper[4870]: I1014 07:19:07.310342 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=10.457709478 podStartE2EDuration="22.310322781s" podCreationTimestamp="2025-10-14 07:18:45 +0000 UTC" firstStartedPulling="2025-10-14 07:18:54.286878595 +0000 UTC m=+1069.984238966" lastFinishedPulling="2025-10-14 07:19:06.139491868 +0000 UTC m=+1081.836852269" observedRunningTime="2025-10-14 07:19:07.305875079 +0000 UTC m=+1083.003235490" watchObservedRunningTime="2025-10-14 07:19:07.310322781 +0000 UTC m=+1083.007683152" Oct 14 07:19:07 crc kubenswrapper[4870]: I1014 07:19:07.346592 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=7.747842137 podStartE2EDuration="19.346567085s" podCreationTimestamp="2025-10-14 07:18:48 +0000 UTC" firstStartedPulling="2025-10-14 07:18:54.523477704 +0000 UTC m=+1070.220838075" lastFinishedPulling="2025-10-14 07:19:06.122202652 +0000 UTC m=+1081.819563023" observedRunningTime="2025-10-14 07:19:07.338846881 +0000 UTC m=+1083.036207282" watchObservedRunningTime="2025-10-14 07:19:07.346567085 +0000 UTC m=+1083.043927476" Oct 14 07:19:07 crc kubenswrapper[4870]: I1014 07:19:07.386163 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Oct 14 07:19:07 crc kubenswrapper[4870]: I1014 07:19:07.448284 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Oct 14 07:19:07 crc kubenswrapper[4870]: I1014 07:19:07.886845 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Oct 14 07:19:07 crc kubenswrapper[4870]: I1014 07:19:07.936404 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Oct 14 07:19:08 crc kubenswrapper[4870]: I1014 07:19:08.229566 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"3de2167a-7663-4b00-9743-229ff1864cb8","Type":"ContainerStarted","Data":"1b53e989d30152b230c176967064509ca1fafba96b986cc96bc23bdcafabf49f"} Oct 14 07:19:08 crc kubenswrapper[4870]: I1014 07:19:08.233945 4870 generic.go:334] "Generic (PLEG): container finished" podID="11aab367-2351-4568-9a45-43c79d419bca" containerID="d6c766e837161620bf402ef73ee0dcf72e7ec5f987d78eb89535c0a55743f119" exitCode=0 Oct 14 07:19:08 crc kubenswrapper[4870]: I1014 07:19:08.234039 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7869c47d6c-44h7s" event={"ID":"11aab367-2351-4568-9a45-43c79d419bca","Type":"ContainerDied","Data":"d6c766e837161620bf402ef73ee0dcf72e7ec5f987d78eb89535c0a55743f119"} Oct 14 07:19:08 crc kubenswrapper[4870]: I1014 07:19:08.239229 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"4d74ef50-1861-4d70-b781-f92d5f4b35ba","Type":"ContainerStarted","Data":"9389d35044ef25b0e127f63d901f00d3d48e58573671b63fa03de1238628ba21"} Oct 14 07:19:08 crc kubenswrapper[4870]: I1014 07:19:08.241040 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Oct 14 07:19:08 crc kubenswrapper[4870]: I1014 07:19:08.241533 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Oct 14 07:19:08 crc kubenswrapper[4870]: I1014 07:19:08.283070 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=22.844061712 podStartE2EDuration="31.283042444s" podCreationTimestamp="2025-10-14 07:18:37 +0000 UTC" firstStartedPulling="2025-10-14 07:18:52.633116147 +0000 UTC m=+1068.330476558" lastFinishedPulling="2025-10-14 07:19:01.072096909 +0000 UTC m=+1076.769457290" observedRunningTime="2025-10-14 07:19:08.269332728 +0000 UTC m=+1083.966693129" watchObservedRunningTime="2025-10-14 07:19:08.283042444 +0000 UTC m=+1083.980402825" Oct 14 07:19:08 crc kubenswrapper[4870]: I1014 07:19:08.300953 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=22.810041278 podStartE2EDuration="30.300929886s" podCreationTimestamp="2025-10-14 07:18:38 +0000 UTC" firstStartedPulling="2025-10-14 07:18:54.180343157 +0000 UTC m=+1069.877703528" lastFinishedPulling="2025-10-14 07:19:01.671231735 +0000 UTC m=+1077.368592136" observedRunningTime="2025-10-14 07:19:08.299194642 +0000 UTC m=+1083.996555023" watchObservedRunningTime="2025-10-14 07:19:08.300929886 +0000 UTC m=+1083.998290267" Oct 14 07:19:08 crc kubenswrapper[4870]: I1014 07:19:08.311709 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Oct 14 07:19:08 crc kubenswrapper[4870]: I1014 07:19:08.315245 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Oct 14 07:19:08 crc kubenswrapper[4870]: I1014 07:19:08.515567 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-86f694bf-k7p29"] Oct 14 07:19:08 crc kubenswrapper[4870]: I1014 07:19:08.559936 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-746b7bb85-l46gj"] Oct 14 07:19:08 crc kubenswrapper[4870]: I1014 07:19:08.561227 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-746b7bb85-l46gj" Oct 14 07:19:08 crc kubenswrapper[4870]: I1014 07:19:08.565514 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Oct 14 07:19:08 crc kubenswrapper[4870]: I1014 07:19:08.570826 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-9dqlp"] Oct 14 07:19:08 crc kubenswrapper[4870]: I1014 07:19:08.571833 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-9dqlp" Oct 14 07:19:08 crc kubenswrapper[4870]: I1014 07:19:08.578321 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Oct 14 07:19:08 crc kubenswrapper[4870]: I1014 07:19:08.581761 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-746b7bb85-l46gj"] Oct 14 07:19:08 crc kubenswrapper[4870]: I1014 07:19:08.592653 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-9dqlp"] Oct 14 07:19:08 crc kubenswrapper[4870]: I1014 07:19:08.628903 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/4c8d0269-7501-4c96-9f9e-30bea4b0fb92-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-9dqlp\" (UID: \"4c8d0269-7501-4c96-9f9e-30bea4b0fb92\") " pod="openstack/ovn-controller-metrics-9dqlp" Oct 14 07:19:08 crc kubenswrapper[4870]: I1014 07:19:08.628964 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4c8d0269-7501-4c96-9f9e-30bea4b0fb92-config\") pod \"ovn-controller-metrics-9dqlp\" (UID: \"4c8d0269-7501-4c96-9f9e-30bea4b0fb92\") " pod="openstack/ovn-controller-metrics-9dqlp" Oct 14 07:19:08 crc kubenswrapper[4870]: I1014 07:19:08.628987 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1142fdf2-8706-464b-9607-7b1fd093d7cf-dns-svc\") pod \"dnsmasq-dns-746b7bb85-l46gj\" (UID: \"1142fdf2-8706-464b-9607-7b1fd093d7cf\") " pod="openstack/dnsmasq-dns-746b7bb85-l46gj" Oct 14 07:19:08 crc kubenswrapper[4870]: I1014 07:19:08.629019 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/4c8d0269-7501-4c96-9f9e-30bea4b0fb92-ovs-rundir\") pod \"ovn-controller-metrics-9dqlp\" (UID: \"4c8d0269-7501-4c96-9f9e-30bea4b0fb92\") " pod="openstack/ovn-controller-metrics-9dqlp" Oct 14 07:19:08 crc kubenswrapper[4870]: I1014 07:19:08.629059 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c8d0269-7501-4c96-9f9e-30bea4b0fb92-combined-ca-bundle\") pod \"ovn-controller-metrics-9dqlp\" (UID: \"4c8d0269-7501-4c96-9f9e-30bea4b0fb92\") " pod="openstack/ovn-controller-metrics-9dqlp" Oct 14 07:19:08 crc kubenswrapper[4870]: I1014 07:19:08.629080 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1142fdf2-8706-464b-9607-7b1fd093d7cf-ovsdbserver-nb\") pod \"dnsmasq-dns-746b7bb85-l46gj\" (UID: \"1142fdf2-8706-464b-9607-7b1fd093d7cf\") " pod="openstack/dnsmasq-dns-746b7bb85-l46gj" Oct 14 07:19:08 crc kubenswrapper[4870]: I1014 07:19:08.629104 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1142fdf2-8706-464b-9607-7b1fd093d7cf-config\") pod \"dnsmasq-dns-746b7bb85-l46gj\" (UID: \"1142fdf2-8706-464b-9607-7b1fd093d7cf\") " pod="openstack/dnsmasq-dns-746b7bb85-l46gj" Oct 14 07:19:08 crc kubenswrapper[4870]: I1014 07:19:08.629126 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8t6l9\" (UniqueName: \"kubernetes.io/projected/4c8d0269-7501-4c96-9f9e-30bea4b0fb92-kube-api-access-8t6l9\") pod \"ovn-controller-metrics-9dqlp\" (UID: \"4c8d0269-7501-4c96-9f9e-30bea4b0fb92\") " pod="openstack/ovn-controller-metrics-9dqlp" Oct 14 07:19:08 crc kubenswrapper[4870]: I1014 07:19:08.629142 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/4c8d0269-7501-4c96-9f9e-30bea4b0fb92-ovn-rundir\") pod \"ovn-controller-metrics-9dqlp\" (UID: \"4c8d0269-7501-4c96-9f9e-30bea4b0fb92\") " pod="openstack/ovn-controller-metrics-9dqlp" Oct 14 07:19:08 crc kubenswrapper[4870]: I1014 07:19:08.629168 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jlf6g\" (UniqueName: \"kubernetes.io/projected/1142fdf2-8706-464b-9607-7b1fd093d7cf-kube-api-access-jlf6g\") pod \"dnsmasq-dns-746b7bb85-l46gj\" (UID: \"1142fdf2-8706-464b-9607-7b1fd093d7cf\") " pod="openstack/dnsmasq-dns-746b7bb85-l46gj" Oct 14 07:19:08 crc kubenswrapper[4870]: I1014 07:19:08.685841 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7869c47d6c-44h7s"] Oct 14 07:19:08 crc kubenswrapper[4870]: I1014 07:19:08.718141 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Oct 14 07:19:08 crc kubenswrapper[4870]: I1014 07:19:08.730474 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Oct 14 07:19:08 crc kubenswrapper[4870]: I1014 07:19:08.733528 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Oct 14 07:19:08 crc kubenswrapper[4870]: I1014 07:19:08.733699 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-4qt4g" Oct 14 07:19:08 crc kubenswrapper[4870]: I1014 07:19:08.734285 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Oct 14 07:19:08 crc kubenswrapper[4870]: I1014 07:19:08.736211 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7764bd4845-4zlrf"] Oct 14 07:19:08 crc kubenswrapper[4870]: I1014 07:19:08.738179 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7764bd4845-4zlrf" Oct 14 07:19:08 crc kubenswrapper[4870]: I1014 07:19:08.738902 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1142fdf2-8706-464b-9607-7b1fd093d7cf-config\") pod \"dnsmasq-dns-746b7bb85-l46gj\" (UID: \"1142fdf2-8706-464b-9607-7b1fd093d7cf\") " pod="openstack/dnsmasq-dns-746b7bb85-l46gj" Oct 14 07:19:08 crc kubenswrapper[4870]: I1014 07:19:08.738946 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8t6l9\" (UniqueName: \"kubernetes.io/projected/4c8d0269-7501-4c96-9f9e-30bea4b0fb92-kube-api-access-8t6l9\") pod \"ovn-controller-metrics-9dqlp\" (UID: \"4c8d0269-7501-4c96-9f9e-30bea4b0fb92\") " pod="openstack/ovn-controller-metrics-9dqlp" Oct 14 07:19:08 crc kubenswrapper[4870]: I1014 07:19:08.738968 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/4c8d0269-7501-4c96-9f9e-30bea4b0fb92-ovn-rundir\") pod \"ovn-controller-metrics-9dqlp\" (UID: \"4c8d0269-7501-4c96-9f9e-30bea4b0fb92\") " pod="openstack/ovn-controller-metrics-9dqlp" Oct 14 07:19:08 crc kubenswrapper[4870]: I1014 07:19:08.739022 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jlf6g\" (UniqueName: \"kubernetes.io/projected/1142fdf2-8706-464b-9607-7b1fd093d7cf-kube-api-access-jlf6g\") pod \"dnsmasq-dns-746b7bb85-l46gj\" (UID: \"1142fdf2-8706-464b-9607-7b1fd093d7cf\") " pod="openstack/dnsmasq-dns-746b7bb85-l46gj" Oct 14 07:19:08 crc kubenswrapper[4870]: I1014 07:19:08.739085 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/4c8d0269-7501-4c96-9f9e-30bea4b0fb92-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-9dqlp\" (UID: \"4c8d0269-7501-4c96-9f9e-30bea4b0fb92\") " pod="openstack/ovn-controller-metrics-9dqlp" Oct 14 07:19:08 crc kubenswrapper[4870]: I1014 07:19:08.739148 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4c8d0269-7501-4c96-9f9e-30bea4b0fb92-config\") pod \"ovn-controller-metrics-9dqlp\" (UID: \"4c8d0269-7501-4c96-9f9e-30bea4b0fb92\") " pod="openstack/ovn-controller-metrics-9dqlp" Oct 14 07:19:08 crc kubenswrapper[4870]: I1014 07:19:08.739179 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1142fdf2-8706-464b-9607-7b1fd093d7cf-dns-svc\") pod \"dnsmasq-dns-746b7bb85-l46gj\" (UID: \"1142fdf2-8706-464b-9607-7b1fd093d7cf\") " pod="openstack/dnsmasq-dns-746b7bb85-l46gj" Oct 14 07:19:08 crc kubenswrapper[4870]: I1014 07:19:08.739238 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/4c8d0269-7501-4c96-9f9e-30bea4b0fb92-ovs-rundir\") pod \"ovn-controller-metrics-9dqlp\" (UID: \"4c8d0269-7501-4c96-9f9e-30bea4b0fb92\") " pod="openstack/ovn-controller-metrics-9dqlp" Oct 14 07:19:08 crc kubenswrapper[4870]: I1014 07:19:08.739310 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c8d0269-7501-4c96-9f9e-30bea4b0fb92-combined-ca-bundle\") pod \"ovn-controller-metrics-9dqlp\" (UID: \"4c8d0269-7501-4c96-9f9e-30bea4b0fb92\") " pod="openstack/ovn-controller-metrics-9dqlp" Oct 14 07:19:08 crc kubenswrapper[4870]: I1014 07:19:08.739344 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1142fdf2-8706-464b-9607-7b1fd093d7cf-ovsdbserver-nb\") pod \"dnsmasq-dns-746b7bb85-l46gj\" (UID: \"1142fdf2-8706-464b-9607-7b1fd093d7cf\") " pod="openstack/dnsmasq-dns-746b7bb85-l46gj" Oct 14 07:19:08 crc kubenswrapper[4870]: I1014 07:19:08.748785 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Oct 14 07:19:08 crc kubenswrapper[4870]: I1014 07:19:08.749241 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1142fdf2-8706-464b-9607-7b1fd093d7cf-config\") pod \"dnsmasq-dns-746b7bb85-l46gj\" (UID: \"1142fdf2-8706-464b-9607-7b1fd093d7cf\") " pod="openstack/dnsmasq-dns-746b7bb85-l46gj" Oct 14 07:19:08 crc kubenswrapper[4870]: I1014 07:19:08.749374 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/4c8d0269-7501-4c96-9f9e-30bea4b0fb92-ovn-rundir\") pod \"ovn-controller-metrics-9dqlp\" (UID: \"4c8d0269-7501-4c96-9f9e-30bea4b0fb92\") " pod="openstack/ovn-controller-metrics-9dqlp" Oct 14 07:19:08 crc kubenswrapper[4870]: I1014 07:19:08.749532 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/4c8d0269-7501-4c96-9f9e-30bea4b0fb92-ovs-rundir\") pod \"ovn-controller-metrics-9dqlp\" (UID: \"4c8d0269-7501-4c96-9f9e-30bea4b0fb92\") " pod="openstack/ovn-controller-metrics-9dqlp" Oct 14 07:19:08 crc kubenswrapper[4870]: I1014 07:19:08.749614 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Oct 14 07:19:08 crc kubenswrapper[4870]: I1014 07:19:08.750158 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4c8d0269-7501-4c96-9f9e-30bea4b0fb92-config\") pod \"ovn-controller-metrics-9dqlp\" (UID: \"4c8d0269-7501-4c96-9f9e-30bea4b0fb92\") " pod="openstack/ovn-controller-metrics-9dqlp" Oct 14 07:19:08 crc kubenswrapper[4870]: I1014 07:19:08.750305 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1142fdf2-8706-464b-9607-7b1fd093d7cf-ovsdbserver-nb\") pod \"dnsmasq-dns-746b7bb85-l46gj\" (UID: \"1142fdf2-8706-464b-9607-7b1fd093d7cf\") " pod="openstack/dnsmasq-dns-746b7bb85-l46gj" Oct 14 07:19:08 crc kubenswrapper[4870]: I1014 07:19:08.750650 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1142fdf2-8706-464b-9607-7b1fd093d7cf-dns-svc\") pod \"dnsmasq-dns-746b7bb85-l46gj\" (UID: \"1142fdf2-8706-464b-9607-7b1fd093d7cf\") " pod="openstack/dnsmasq-dns-746b7bb85-l46gj" Oct 14 07:19:08 crc kubenswrapper[4870]: I1014 07:19:08.755545 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Oct 14 07:19:08 crc kubenswrapper[4870]: I1014 07:19:08.758219 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c8d0269-7501-4c96-9f9e-30bea4b0fb92-combined-ca-bundle\") pod \"ovn-controller-metrics-9dqlp\" (UID: \"4c8d0269-7501-4c96-9f9e-30bea4b0fb92\") " pod="openstack/ovn-controller-metrics-9dqlp" Oct 14 07:19:08 crc kubenswrapper[4870]: I1014 07:19:08.761009 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/4c8d0269-7501-4c96-9f9e-30bea4b0fb92-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-9dqlp\" (UID: \"4c8d0269-7501-4c96-9f9e-30bea4b0fb92\") " pod="openstack/ovn-controller-metrics-9dqlp" Oct 14 07:19:08 crc kubenswrapper[4870]: I1014 07:19:08.764635 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7764bd4845-4zlrf"] Oct 14 07:19:08 crc kubenswrapper[4870]: I1014 07:19:08.771815 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jlf6g\" (UniqueName: \"kubernetes.io/projected/1142fdf2-8706-464b-9607-7b1fd093d7cf-kube-api-access-jlf6g\") pod \"dnsmasq-dns-746b7bb85-l46gj\" (UID: \"1142fdf2-8706-464b-9607-7b1fd093d7cf\") " pod="openstack/dnsmasq-dns-746b7bb85-l46gj" Oct 14 07:19:08 crc kubenswrapper[4870]: I1014 07:19:08.780304 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8t6l9\" (UniqueName: \"kubernetes.io/projected/4c8d0269-7501-4c96-9f9e-30bea4b0fb92-kube-api-access-8t6l9\") pod \"ovn-controller-metrics-9dqlp\" (UID: \"4c8d0269-7501-4c96-9f9e-30bea4b0fb92\") " pod="openstack/ovn-controller-metrics-9dqlp" Oct 14 07:19:08 crc kubenswrapper[4870]: I1014 07:19:08.843195 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d28jg\" (UniqueName: \"kubernetes.io/projected/416d7b39-4541-44a8-a55f-924bc86fee32-kube-api-access-d28jg\") pod \"ovn-northd-0\" (UID: \"416d7b39-4541-44a8-a55f-924bc86fee32\") " pod="openstack/ovn-northd-0" Oct 14 07:19:08 crc kubenswrapper[4870]: I1014 07:19:08.843253 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/416d7b39-4541-44a8-a55f-924bc86fee32-scripts\") pod \"ovn-northd-0\" (UID: \"416d7b39-4541-44a8-a55f-924bc86fee32\") " pod="openstack/ovn-northd-0" Oct 14 07:19:08 crc kubenswrapper[4870]: I1014 07:19:08.843273 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/416d7b39-4541-44a8-a55f-924bc86fee32-config\") pod \"ovn-northd-0\" (UID: \"416d7b39-4541-44a8-a55f-924bc86fee32\") " pod="openstack/ovn-northd-0" Oct 14 07:19:08 crc kubenswrapper[4870]: I1014 07:19:08.843303 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/416d7b39-4541-44a8-a55f-924bc86fee32-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"416d7b39-4541-44a8-a55f-924bc86fee32\") " pod="openstack/ovn-northd-0" Oct 14 07:19:08 crc kubenswrapper[4870]: I1014 07:19:08.843483 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kbrgv\" (UniqueName: \"kubernetes.io/projected/a67380d5-9cd0-4443-b45b-44663fddd756-kube-api-access-kbrgv\") pod \"dnsmasq-dns-7764bd4845-4zlrf\" (UID: \"a67380d5-9cd0-4443-b45b-44663fddd756\") " pod="openstack/dnsmasq-dns-7764bd4845-4zlrf" Oct 14 07:19:08 crc kubenswrapper[4870]: I1014 07:19:08.843588 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/416d7b39-4541-44a8-a55f-924bc86fee32-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"416d7b39-4541-44a8-a55f-924bc86fee32\") " pod="openstack/ovn-northd-0" Oct 14 07:19:08 crc kubenswrapper[4870]: I1014 07:19:08.843628 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a67380d5-9cd0-4443-b45b-44663fddd756-config\") pod \"dnsmasq-dns-7764bd4845-4zlrf\" (UID: \"a67380d5-9cd0-4443-b45b-44663fddd756\") " pod="openstack/dnsmasq-dns-7764bd4845-4zlrf" Oct 14 07:19:08 crc kubenswrapper[4870]: I1014 07:19:08.843681 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a67380d5-9cd0-4443-b45b-44663fddd756-dns-svc\") pod \"dnsmasq-dns-7764bd4845-4zlrf\" (UID: \"a67380d5-9cd0-4443-b45b-44663fddd756\") " pod="openstack/dnsmasq-dns-7764bd4845-4zlrf" Oct 14 07:19:08 crc kubenswrapper[4870]: I1014 07:19:08.843697 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a67380d5-9cd0-4443-b45b-44663fddd756-ovsdbserver-sb\") pod \"dnsmasq-dns-7764bd4845-4zlrf\" (UID: \"a67380d5-9cd0-4443-b45b-44663fddd756\") " pod="openstack/dnsmasq-dns-7764bd4845-4zlrf" Oct 14 07:19:08 crc kubenswrapper[4870]: I1014 07:19:08.843713 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/416d7b39-4541-44a8-a55f-924bc86fee32-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"416d7b39-4541-44a8-a55f-924bc86fee32\") " pod="openstack/ovn-northd-0" Oct 14 07:19:08 crc kubenswrapper[4870]: I1014 07:19:08.843751 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a67380d5-9cd0-4443-b45b-44663fddd756-ovsdbserver-nb\") pod \"dnsmasq-dns-7764bd4845-4zlrf\" (UID: \"a67380d5-9cd0-4443-b45b-44663fddd756\") " pod="openstack/dnsmasq-dns-7764bd4845-4zlrf" Oct 14 07:19:08 crc kubenswrapper[4870]: I1014 07:19:08.843776 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/416d7b39-4541-44a8-a55f-924bc86fee32-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"416d7b39-4541-44a8-a55f-924bc86fee32\") " pod="openstack/ovn-northd-0" Oct 14 07:19:08 crc kubenswrapper[4870]: I1014 07:19:08.871100 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Oct 14 07:19:08 crc kubenswrapper[4870]: I1014 07:19:08.871143 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Oct 14 07:19:08 crc kubenswrapper[4870]: I1014 07:19:08.885374 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-746b7bb85-l46gj" Oct 14 07:19:08 crc kubenswrapper[4870]: I1014 07:19:08.896693 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-9dqlp" Oct 14 07:19:08 crc kubenswrapper[4870]: I1014 07:19:08.945726 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/416d7b39-4541-44a8-a55f-924bc86fee32-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"416d7b39-4541-44a8-a55f-924bc86fee32\") " pod="openstack/ovn-northd-0" Oct 14 07:19:08 crc kubenswrapper[4870]: I1014 07:19:08.945775 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a67380d5-9cd0-4443-b45b-44663fddd756-config\") pod \"dnsmasq-dns-7764bd4845-4zlrf\" (UID: \"a67380d5-9cd0-4443-b45b-44663fddd756\") " pod="openstack/dnsmasq-dns-7764bd4845-4zlrf" Oct 14 07:19:08 crc kubenswrapper[4870]: I1014 07:19:08.945805 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a67380d5-9cd0-4443-b45b-44663fddd756-dns-svc\") pod \"dnsmasq-dns-7764bd4845-4zlrf\" (UID: \"a67380d5-9cd0-4443-b45b-44663fddd756\") " pod="openstack/dnsmasq-dns-7764bd4845-4zlrf" Oct 14 07:19:08 crc kubenswrapper[4870]: I1014 07:19:08.945820 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a67380d5-9cd0-4443-b45b-44663fddd756-ovsdbserver-sb\") pod \"dnsmasq-dns-7764bd4845-4zlrf\" (UID: \"a67380d5-9cd0-4443-b45b-44663fddd756\") " pod="openstack/dnsmasq-dns-7764bd4845-4zlrf" Oct 14 07:19:08 crc kubenswrapper[4870]: I1014 07:19:08.945839 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/416d7b39-4541-44a8-a55f-924bc86fee32-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"416d7b39-4541-44a8-a55f-924bc86fee32\") " pod="openstack/ovn-northd-0" Oct 14 07:19:08 crc kubenswrapper[4870]: I1014 07:19:08.945866 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a67380d5-9cd0-4443-b45b-44663fddd756-ovsdbserver-nb\") pod \"dnsmasq-dns-7764bd4845-4zlrf\" (UID: \"a67380d5-9cd0-4443-b45b-44663fddd756\") " pod="openstack/dnsmasq-dns-7764bd4845-4zlrf" Oct 14 07:19:08 crc kubenswrapper[4870]: I1014 07:19:08.945888 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/416d7b39-4541-44a8-a55f-924bc86fee32-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"416d7b39-4541-44a8-a55f-924bc86fee32\") " pod="openstack/ovn-northd-0" Oct 14 07:19:08 crc kubenswrapper[4870]: I1014 07:19:08.945925 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d28jg\" (UniqueName: \"kubernetes.io/projected/416d7b39-4541-44a8-a55f-924bc86fee32-kube-api-access-d28jg\") pod \"ovn-northd-0\" (UID: \"416d7b39-4541-44a8-a55f-924bc86fee32\") " pod="openstack/ovn-northd-0" Oct 14 07:19:08 crc kubenswrapper[4870]: I1014 07:19:08.945943 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/416d7b39-4541-44a8-a55f-924bc86fee32-scripts\") pod \"ovn-northd-0\" (UID: \"416d7b39-4541-44a8-a55f-924bc86fee32\") " pod="openstack/ovn-northd-0" Oct 14 07:19:08 crc kubenswrapper[4870]: I1014 07:19:08.945958 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/416d7b39-4541-44a8-a55f-924bc86fee32-config\") pod \"ovn-northd-0\" (UID: \"416d7b39-4541-44a8-a55f-924bc86fee32\") " pod="openstack/ovn-northd-0" Oct 14 07:19:08 crc kubenswrapper[4870]: I1014 07:19:08.945977 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/416d7b39-4541-44a8-a55f-924bc86fee32-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"416d7b39-4541-44a8-a55f-924bc86fee32\") " pod="openstack/ovn-northd-0" Oct 14 07:19:08 crc kubenswrapper[4870]: I1014 07:19:08.946021 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kbrgv\" (UniqueName: \"kubernetes.io/projected/a67380d5-9cd0-4443-b45b-44663fddd756-kube-api-access-kbrgv\") pod \"dnsmasq-dns-7764bd4845-4zlrf\" (UID: \"a67380d5-9cd0-4443-b45b-44663fddd756\") " pod="openstack/dnsmasq-dns-7764bd4845-4zlrf" Oct 14 07:19:08 crc kubenswrapper[4870]: I1014 07:19:08.947803 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/416d7b39-4541-44a8-a55f-924bc86fee32-scripts\") pod \"ovn-northd-0\" (UID: \"416d7b39-4541-44a8-a55f-924bc86fee32\") " pod="openstack/ovn-northd-0" Oct 14 07:19:08 crc kubenswrapper[4870]: I1014 07:19:08.947871 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a67380d5-9cd0-4443-b45b-44663fddd756-dns-svc\") pod \"dnsmasq-dns-7764bd4845-4zlrf\" (UID: \"a67380d5-9cd0-4443-b45b-44663fddd756\") " pod="openstack/dnsmasq-dns-7764bd4845-4zlrf" Oct 14 07:19:08 crc kubenswrapper[4870]: I1014 07:19:08.948747 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a67380d5-9cd0-4443-b45b-44663fddd756-config\") pod \"dnsmasq-dns-7764bd4845-4zlrf\" (UID: \"a67380d5-9cd0-4443-b45b-44663fddd756\") " pod="openstack/dnsmasq-dns-7764bd4845-4zlrf" Oct 14 07:19:08 crc kubenswrapper[4870]: I1014 07:19:08.949074 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a67380d5-9cd0-4443-b45b-44663fddd756-ovsdbserver-sb\") pod \"dnsmasq-dns-7764bd4845-4zlrf\" (UID: \"a67380d5-9cd0-4443-b45b-44663fddd756\") " pod="openstack/dnsmasq-dns-7764bd4845-4zlrf" Oct 14 07:19:08 crc kubenswrapper[4870]: I1014 07:19:08.949584 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/416d7b39-4541-44a8-a55f-924bc86fee32-config\") pod \"ovn-northd-0\" (UID: \"416d7b39-4541-44a8-a55f-924bc86fee32\") " pod="openstack/ovn-northd-0" Oct 14 07:19:08 crc kubenswrapper[4870]: I1014 07:19:08.949834 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/416d7b39-4541-44a8-a55f-924bc86fee32-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"416d7b39-4541-44a8-a55f-924bc86fee32\") " pod="openstack/ovn-northd-0" Oct 14 07:19:08 crc kubenswrapper[4870]: I1014 07:19:08.950420 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/416d7b39-4541-44a8-a55f-924bc86fee32-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"416d7b39-4541-44a8-a55f-924bc86fee32\") " pod="openstack/ovn-northd-0" Oct 14 07:19:08 crc kubenswrapper[4870]: I1014 07:19:08.953540 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a67380d5-9cd0-4443-b45b-44663fddd756-ovsdbserver-nb\") pod \"dnsmasq-dns-7764bd4845-4zlrf\" (UID: \"a67380d5-9cd0-4443-b45b-44663fddd756\") " pod="openstack/dnsmasq-dns-7764bd4845-4zlrf" Oct 14 07:19:08 crc kubenswrapper[4870]: I1014 07:19:08.953678 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/416d7b39-4541-44a8-a55f-924bc86fee32-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"416d7b39-4541-44a8-a55f-924bc86fee32\") " pod="openstack/ovn-northd-0" Oct 14 07:19:08 crc kubenswrapper[4870]: I1014 07:19:08.953808 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/416d7b39-4541-44a8-a55f-924bc86fee32-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"416d7b39-4541-44a8-a55f-924bc86fee32\") " pod="openstack/ovn-northd-0" Oct 14 07:19:08 crc kubenswrapper[4870]: I1014 07:19:08.963641 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kbrgv\" (UniqueName: \"kubernetes.io/projected/a67380d5-9cd0-4443-b45b-44663fddd756-kube-api-access-kbrgv\") pod \"dnsmasq-dns-7764bd4845-4zlrf\" (UID: \"a67380d5-9cd0-4443-b45b-44663fddd756\") " pod="openstack/dnsmasq-dns-7764bd4845-4zlrf" Oct 14 07:19:08 crc kubenswrapper[4870]: I1014 07:19:08.964561 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86f694bf-k7p29" Oct 14 07:19:08 crc kubenswrapper[4870]: I1014 07:19:08.987072 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d28jg\" (UniqueName: \"kubernetes.io/projected/416d7b39-4541-44a8-a55f-924bc86fee32-kube-api-access-d28jg\") pod \"ovn-northd-0\" (UID: \"416d7b39-4541-44a8-a55f-924bc86fee32\") " pod="openstack/ovn-northd-0" Oct 14 07:19:09 crc kubenswrapper[4870]: I1014 07:19:09.049905 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lfz64\" (UniqueName: \"kubernetes.io/projected/86d52d2f-b41b-4c56-af26-b32e8caaf225-kube-api-access-lfz64\") pod \"86d52d2f-b41b-4c56-af26-b32e8caaf225\" (UID: \"86d52d2f-b41b-4c56-af26-b32e8caaf225\") " Oct 14 07:19:09 crc kubenswrapper[4870]: I1014 07:19:09.050085 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/86d52d2f-b41b-4c56-af26-b32e8caaf225-dns-svc\") pod \"86d52d2f-b41b-4c56-af26-b32e8caaf225\" (UID: \"86d52d2f-b41b-4c56-af26-b32e8caaf225\") " Oct 14 07:19:09 crc kubenswrapper[4870]: I1014 07:19:09.050115 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/86d52d2f-b41b-4c56-af26-b32e8caaf225-config\") pod \"86d52d2f-b41b-4c56-af26-b32e8caaf225\" (UID: \"86d52d2f-b41b-4c56-af26-b32e8caaf225\") " Oct 14 07:19:09 crc kubenswrapper[4870]: I1014 07:19:09.051499 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/86d52d2f-b41b-4c56-af26-b32e8caaf225-config" (OuterVolumeSpecName: "config") pod "86d52d2f-b41b-4c56-af26-b32e8caaf225" (UID: "86d52d2f-b41b-4c56-af26-b32e8caaf225"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:19:09 crc kubenswrapper[4870]: I1014 07:19:09.052644 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/86d52d2f-b41b-4c56-af26-b32e8caaf225-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "86d52d2f-b41b-4c56-af26-b32e8caaf225" (UID: "86d52d2f-b41b-4c56-af26-b32e8caaf225"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:19:09 crc kubenswrapper[4870]: I1014 07:19:09.054651 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/86d52d2f-b41b-4c56-af26-b32e8caaf225-kube-api-access-lfz64" (OuterVolumeSpecName: "kube-api-access-lfz64") pod "86d52d2f-b41b-4c56-af26-b32e8caaf225" (UID: "86d52d2f-b41b-4c56-af26-b32e8caaf225"). InnerVolumeSpecName "kube-api-access-lfz64". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:19:09 crc kubenswrapper[4870]: I1014 07:19:09.133416 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Oct 14 07:19:09 crc kubenswrapper[4870]: I1014 07:19:09.140110 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7764bd4845-4zlrf" Oct 14 07:19:09 crc kubenswrapper[4870]: I1014 07:19:09.151940 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lfz64\" (UniqueName: \"kubernetes.io/projected/86d52d2f-b41b-4c56-af26-b32e8caaf225-kube-api-access-lfz64\") on node \"crc\" DevicePath \"\"" Oct 14 07:19:09 crc kubenswrapper[4870]: I1014 07:19:09.151987 4870 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/86d52d2f-b41b-4c56-af26-b32e8caaf225-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 14 07:19:09 crc kubenswrapper[4870]: I1014 07:19:09.151997 4870 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/86d52d2f-b41b-4c56-af26-b32e8caaf225-config\") on node \"crc\" DevicePath \"\"" Oct 14 07:19:09 crc kubenswrapper[4870]: I1014 07:19:09.259516 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86f694bf-k7p29" event={"ID":"86d52d2f-b41b-4c56-af26-b32e8caaf225","Type":"ContainerDied","Data":"484a8f4e1c510fa9732ac80d023636d9cd97ecc2ee418590f9c7d2d461321d4e"} Oct 14 07:19:09 crc kubenswrapper[4870]: I1014 07:19:09.259631 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86f694bf-k7p29" Oct 14 07:19:09 crc kubenswrapper[4870]: I1014 07:19:09.268469 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7869c47d6c-44h7s" event={"ID":"11aab367-2351-4568-9a45-43c79d419bca","Type":"ContainerStarted","Data":"59f292e567ffd2aaa8466b5b0d9838137e443ce7f26c627f41dd0166a49bdf23"} Oct 14 07:19:09 crc kubenswrapper[4870]: I1014 07:19:09.269095 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7869c47d6c-44h7s" podUID="11aab367-2351-4568-9a45-43c79d419bca" containerName="dnsmasq-dns" containerID="cri-o://59f292e567ffd2aaa8466b5b0d9838137e443ce7f26c627f41dd0166a49bdf23" gracePeriod=10 Oct 14 07:19:09 crc kubenswrapper[4870]: I1014 07:19:09.269263 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7869c47d6c-44h7s" Oct 14 07:19:09 crc kubenswrapper[4870]: I1014 07:19:09.294656 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7869c47d6c-44h7s" podStartSLOduration=2.860475776 podStartE2EDuration="33.294636608s" podCreationTimestamp="2025-10-14 07:18:36 +0000 UTC" firstStartedPulling="2025-10-14 07:18:37.058386437 +0000 UTC m=+1052.755746808" lastFinishedPulling="2025-10-14 07:19:07.492547259 +0000 UTC m=+1083.189907640" observedRunningTime="2025-10-14 07:19:09.288947915 +0000 UTC m=+1084.986308286" watchObservedRunningTime="2025-10-14 07:19:09.294636608 +0000 UTC m=+1084.991996979" Oct 14 07:19:09 crc kubenswrapper[4870]: I1014 07:19:09.335088 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-86f694bf-k7p29"] Oct 14 07:19:09 crc kubenswrapper[4870]: I1014 07:19:09.341289 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-86f694bf-k7p29"] Oct 14 07:19:09 crc kubenswrapper[4870]: I1014 07:19:09.379972 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-746b7bb85-l46gj"] Oct 14 07:19:09 crc kubenswrapper[4870]: W1014 07:19:09.393807 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1142fdf2_8706_464b_9607_7b1fd093d7cf.slice/crio-65a1296dc09c948f28550ebfb2cd92e0576cccbcc210ddd6df4c2208fe155912 WatchSource:0}: Error finding container 65a1296dc09c948f28550ebfb2cd92e0576cccbcc210ddd6df4c2208fe155912: Status 404 returned error can't find the container with id 65a1296dc09c948f28550ebfb2cd92e0576cccbcc210ddd6df4c2208fe155912 Oct 14 07:19:09 crc kubenswrapper[4870]: I1014 07:19:09.440226 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-9dqlp"] Oct 14 07:19:09 crc kubenswrapper[4870]: W1014 07:19:09.496797 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4c8d0269_7501_4c96_9f9e_30bea4b0fb92.slice/crio-b0345ab44f21f08a1613972c5c05f4e2a7385879b829d80e6f6d9c546fe68b40 WatchSource:0}: Error finding container b0345ab44f21f08a1613972c5c05f4e2a7385879b829d80e6f6d9c546fe68b40: Status 404 returned error can't find the container with id b0345ab44f21f08a1613972c5c05f4e2a7385879b829d80e6f6d9c546fe68b40 Oct 14 07:19:09 crc kubenswrapper[4870]: I1014 07:19:09.632299 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Oct 14 07:19:09 crc kubenswrapper[4870]: I1014 07:19:09.639673 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7764bd4845-4zlrf"] Oct 14 07:19:09 crc kubenswrapper[4870]: I1014 07:19:09.810774 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7869c47d6c-44h7s" Oct 14 07:19:09 crc kubenswrapper[4870]: I1014 07:19:09.866102 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-psk44\" (UniqueName: \"kubernetes.io/projected/11aab367-2351-4568-9a45-43c79d419bca-kube-api-access-psk44\") pod \"11aab367-2351-4568-9a45-43c79d419bca\" (UID: \"11aab367-2351-4568-9a45-43c79d419bca\") " Oct 14 07:19:09 crc kubenswrapper[4870]: I1014 07:19:09.866263 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/11aab367-2351-4568-9a45-43c79d419bca-dns-svc\") pod \"11aab367-2351-4568-9a45-43c79d419bca\" (UID: \"11aab367-2351-4568-9a45-43c79d419bca\") " Oct 14 07:19:09 crc kubenswrapper[4870]: I1014 07:19:09.866385 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/11aab367-2351-4568-9a45-43c79d419bca-config\") pod \"11aab367-2351-4568-9a45-43c79d419bca\" (UID: \"11aab367-2351-4568-9a45-43c79d419bca\") " Oct 14 07:19:09 crc kubenswrapper[4870]: I1014 07:19:09.873592 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/11aab367-2351-4568-9a45-43c79d419bca-kube-api-access-psk44" (OuterVolumeSpecName: "kube-api-access-psk44") pod "11aab367-2351-4568-9a45-43c79d419bca" (UID: "11aab367-2351-4568-9a45-43c79d419bca"). InnerVolumeSpecName "kube-api-access-psk44". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:19:09 crc kubenswrapper[4870]: I1014 07:19:09.913983 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/11aab367-2351-4568-9a45-43c79d419bca-config" (OuterVolumeSpecName: "config") pod "11aab367-2351-4568-9a45-43c79d419bca" (UID: "11aab367-2351-4568-9a45-43c79d419bca"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:19:09 crc kubenswrapper[4870]: I1014 07:19:09.915520 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/11aab367-2351-4568-9a45-43c79d419bca-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "11aab367-2351-4568-9a45-43c79d419bca" (UID: "11aab367-2351-4568-9a45-43c79d419bca"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:19:09 crc kubenswrapper[4870]: I1014 07:19:09.967903 4870 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/11aab367-2351-4568-9a45-43c79d419bca-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 14 07:19:09 crc kubenswrapper[4870]: I1014 07:19:09.968159 4870 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/11aab367-2351-4568-9a45-43c79d419bca-config\") on node \"crc\" DevicePath \"\"" Oct 14 07:19:09 crc kubenswrapper[4870]: I1014 07:19:09.968169 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-psk44\" (UniqueName: \"kubernetes.io/projected/11aab367-2351-4568-9a45-43c79d419bca-kube-api-access-psk44\") on node \"crc\" DevicePath \"\"" Oct 14 07:19:10 crc kubenswrapper[4870]: I1014 07:19:10.276861 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-9dqlp" event={"ID":"4c8d0269-7501-4c96-9f9e-30bea4b0fb92","Type":"ContainerStarted","Data":"f1b865fdd1614804c6e4ae198a96114c844ab93d4e346b066d14705872321268"} Oct 14 07:19:10 crc kubenswrapper[4870]: I1014 07:19:10.276903 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-9dqlp" event={"ID":"4c8d0269-7501-4c96-9f9e-30bea4b0fb92","Type":"ContainerStarted","Data":"b0345ab44f21f08a1613972c5c05f4e2a7385879b829d80e6f6d9c546fe68b40"} Oct 14 07:19:10 crc kubenswrapper[4870]: I1014 07:19:10.281359 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"416d7b39-4541-44a8-a55f-924bc86fee32","Type":"ContainerStarted","Data":"ae426546eb62a779272b920709b17c66107242f70b6d596ad360a33101291fe0"} Oct 14 07:19:10 crc kubenswrapper[4870]: I1014 07:19:10.283220 4870 generic.go:334] "Generic (PLEG): container finished" podID="11aab367-2351-4568-9a45-43c79d419bca" containerID="59f292e567ffd2aaa8466b5b0d9838137e443ce7f26c627f41dd0166a49bdf23" exitCode=0 Oct 14 07:19:10 crc kubenswrapper[4870]: I1014 07:19:10.283265 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7869c47d6c-44h7s" event={"ID":"11aab367-2351-4568-9a45-43c79d419bca","Type":"ContainerDied","Data":"59f292e567ffd2aaa8466b5b0d9838137e443ce7f26c627f41dd0166a49bdf23"} Oct 14 07:19:10 crc kubenswrapper[4870]: I1014 07:19:10.283283 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7869c47d6c-44h7s" event={"ID":"11aab367-2351-4568-9a45-43c79d419bca","Type":"ContainerDied","Data":"d45565075ad8f85eaabe03f0f165cb8b812379f94c0a532c4e203f8434352b6c"} Oct 14 07:19:10 crc kubenswrapper[4870]: I1014 07:19:10.283322 4870 scope.go:117] "RemoveContainer" containerID="59f292e567ffd2aaa8466b5b0d9838137e443ce7f26c627f41dd0166a49bdf23" Oct 14 07:19:10 crc kubenswrapper[4870]: I1014 07:19:10.283568 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7869c47d6c-44h7s" Oct 14 07:19:10 crc kubenswrapper[4870]: I1014 07:19:10.285868 4870 generic.go:334] "Generic (PLEG): container finished" podID="1142fdf2-8706-464b-9607-7b1fd093d7cf" containerID="0c0369e5032cbb5a6790a98dd7a1df20a98e32c588b729d7406be04b578ed108" exitCode=0 Oct 14 07:19:10 crc kubenswrapper[4870]: I1014 07:19:10.285950 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-746b7bb85-l46gj" event={"ID":"1142fdf2-8706-464b-9607-7b1fd093d7cf","Type":"ContainerDied","Data":"0c0369e5032cbb5a6790a98dd7a1df20a98e32c588b729d7406be04b578ed108"} Oct 14 07:19:10 crc kubenswrapper[4870]: I1014 07:19:10.285982 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-746b7bb85-l46gj" event={"ID":"1142fdf2-8706-464b-9607-7b1fd093d7cf","Type":"ContainerStarted","Data":"65a1296dc09c948f28550ebfb2cd92e0576cccbcc210ddd6df4c2208fe155912"} Oct 14 07:19:10 crc kubenswrapper[4870]: I1014 07:19:10.289449 4870 generic.go:334] "Generic (PLEG): container finished" podID="a67380d5-9cd0-4443-b45b-44663fddd756" containerID="d614d9d6e11a79690035da28f7f8cbdd7407e7f268ed5312d39fce5ec71890e3" exitCode=0 Oct 14 07:19:10 crc kubenswrapper[4870]: I1014 07:19:10.289517 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7764bd4845-4zlrf" event={"ID":"a67380d5-9cd0-4443-b45b-44663fddd756","Type":"ContainerDied","Data":"d614d9d6e11a79690035da28f7f8cbdd7407e7f268ed5312d39fce5ec71890e3"} Oct 14 07:19:10 crc kubenswrapper[4870]: I1014 07:19:10.289569 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7764bd4845-4zlrf" event={"ID":"a67380d5-9cd0-4443-b45b-44663fddd756","Type":"ContainerStarted","Data":"ec1a938e2190cea94baf85f78c82af874b1811a42007f42d6a014e6a398dcdae"} Oct 14 07:19:10 crc kubenswrapper[4870]: I1014 07:19:10.295230 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-9dqlp" podStartSLOduration=2.295212695 podStartE2EDuration="2.295212695s" podCreationTimestamp="2025-10-14 07:19:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:19:10.293912892 +0000 UTC m=+1085.991273283" watchObservedRunningTime="2025-10-14 07:19:10.295212695 +0000 UTC m=+1085.992573066" Oct 14 07:19:10 crc kubenswrapper[4870]: I1014 07:19:10.297193 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Oct 14 07:19:10 crc kubenswrapper[4870]: I1014 07:19:10.298009 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Oct 14 07:19:10 crc kubenswrapper[4870]: I1014 07:19:10.447083 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7869c47d6c-44h7s"] Oct 14 07:19:10 crc kubenswrapper[4870]: I1014 07:19:10.467344 4870 scope.go:117] "RemoveContainer" containerID="d6c766e837161620bf402ef73ee0dcf72e7ec5f987d78eb89535c0a55743f119" Oct 14 07:19:10 crc kubenswrapper[4870]: I1014 07:19:10.486267 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7869c47d6c-44h7s"] Oct 14 07:19:10 crc kubenswrapper[4870]: I1014 07:19:10.491887 4870 scope.go:117] "RemoveContainer" containerID="59f292e567ffd2aaa8466b5b0d9838137e443ce7f26c627f41dd0166a49bdf23" Oct 14 07:19:10 crc kubenswrapper[4870]: E1014 07:19:10.492215 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"59f292e567ffd2aaa8466b5b0d9838137e443ce7f26c627f41dd0166a49bdf23\": container with ID starting with 59f292e567ffd2aaa8466b5b0d9838137e443ce7f26c627f41dd0166a49bdf23 not found: ID does not exist" containerID="59f292e567ffd2aaa8466b5b0d9838137e443ce7f26c627f41dd0166a49bdf23" Oct 14 07:19:10 crc kubenswrapper[4870]: I1014 07:19:10.492259 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"59f292e567ffd2aaa8466b5b0d9838137e443ce7f26c627f41dd0166a49bdf23"} err="failed to get container status \"59f292e567ffd2aaa8466b5b0d9838137e443ce7f26c627f41dd0166a49bdf23\": rpc error: code = NotFound desc = could not find container \"59f292e567ffd2aaa8466b5b0d9838137e443ce7f26c627f41dd0166a49bdf23\": container with ID starting with 59f292e567ffd2aaa8466b5b0d9838137e443ce7f26c627f41dd0166a49bdf23 not found: ID does not exist" Oct 14 07:19:10 crc kubenswrapper[4870]: I1014 07:19:10.492294 4870 scope.go:117] "RemoveContainer" containerID="d6c766e837161620bf402ef73ee0dcf72e7ec5f987d78eb89535c0a55743f119" Oct 14 07:19:10 crc kubenswrapper[4870]: E1014 07:19:10.492825 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d6c766e837161620bf402ef73ee0dcf72e7ec5f987d78eb89535c0a55743f119\": container with ID starting with d6c766e837161620bf402ef73ee0dcf72e7ec5f987d78eb89535c0a55743f119 not found: ID does not exist" containerID="d6c766e837161620bf402ef73ee0dcf72e7ec5f987d78eb89535c0a55743f119" Oct 14 07:19:10 crc kubenswrapper[4870]: I1014 07:19:10.492854 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d6c766e837161620bf402ef73ee0dcf72e7ec5f987d78eb89535c0a55743f119"} err="failed to get container status \"d6c766e837161620bf402ef73ee0dcf72e7ec5f987d78eb89535c0a55743f119\": rpc error: code = NotFound desc = could not find container \"d6c766e837161620bf402ef73ee0dcf72e7ec5f987d78eb89535c0a55743f119\": container with ID starting with d6c766e837161620bf402ef73ee0dcf72e7ec5f987d78eb89535c0a55743f119 not found: ID does not exist" Oct 14 07:19:10 crc kubenswrapper[4870]: I1014 07:19:10.630273 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Oct 14 07:19:11 crc kubenswrapper[4870]: I1014 07:19:11.044661 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="11aab367-2351-4568-9a45-43c79d419bca" path="/var/lib/kubelet/pods/11aab367-2351-4568-9a45-43c79d419bca/volumes" Oct 14 07:19:11 crc kubenswrapper[4870]: I1014 07:19:11.045466 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="86d52d2f-b41b-4c56-af26-b32e8caaf225" path="/var/lib/kubelet/pods/86d52d2f-b41b-4c56-af26-b32e8caaf225/volumes" Oct 14 07:19:11 crc kubenswrapper[4870]: I1014 07:19:11.300252 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-746b7bb85-l46gj" event={"ID":"1142fdf2-8706-464b-9607-7b1fd093d7cf","Type":"ContainerStarted","Data":"4dfd29d1bf7a2d5610f6a802f7d18bd7bcfd0364ecf1fcbe7b15495aa87493e1"} Oct 14 07:19:11 crc kubenswrapper[4870]: I1014 07:19:11.301514 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-746b7bb85-l46gj" Oct 14 07:19:11 crc kubenswrapper[4870]: I1014 07:19:11.303240 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7764bd4845-4zlrf" event={"ID":"a67380d5-9cd0-4443-b45b-44663fddd756","Type":"ContainerStarted","Data":"44771a13a098ac9d67781af2c4e40602aeb29022af9b243ae7996df7a250b70b"} Oct 14 07:19:11 crc kubenswrapper[4870]: I1014 07:19:11.303671 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7764bd4845-4zlrf" Oct 14 07:19:11 crc kubenswrapper[4870]: I1014 07:19:11.306242 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"416d7b39-4541-44a8-a55f-924bc86fee32","Type":"ContainerStarted","Data":"11e57f381235088d13225057331ed61d6e2ab6f98595ff6fe4992063044159e4"} Oct 14 07:19:11 crc kubenswrapper[4870]: I1014 07:19:11.306269 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"416d7b39-4541-44a8-a55f-924bc86fee32","Type":"ContainerStarted","Data":"457905ca1130c3804c068cf6fc44521ad064a3fb34449d666cf858f14e2feace"} Oct 14 07:19:11 crc kubenswrapper[4870]: I1014 07:19:11.306305 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Oct 14 07:19:11 crc kubenswrapper[4870]: I1014 07:19:11.337598 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-746b7bb85-l46gj" podStartSLOduration=3.337582716 podStartE2EDuration="3.337582716s" podCreationTimestamp="2025-10-14 07:19:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:19:11.333583525 +0000 UTC m=+1087.030943886" watchObservedRunningTime="2025-10-14 07:19:11.337582716 +0000 UTC m=+1087.034943087" Oct 14 07:19:11 crc kubenswrapper[4870]: I1014 07:19:11.356150 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7764bd4845-4zlrf" podStartSLOduration=3.356132574 podStartE2EDuration="3.356132574s" podCreationTimestamp="2025-10-14 07:19:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:19:11.355422426 +0000 UTC m=+1087.052782797" watchObservedRunningTime="2025-10-14 07:19:11.356132574 +0000 UTC m=+1087.053492945" Oct 14 07:19:11 crc kubenswrapper[4870]: I1014 07:19:11.379668 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=2.139726062 podStartE2EDuration="3.379649628s" podCreationTimestamp="2025-10-14 07:19:08 +0000 UTC" firstStartedPulling="2025-10-14 07:19:09.66249364 +0000 UTC m=+1085.359854011" lastFinishedPulling="2025-10-14 07:19:10.902417216 +0000 UTC m=+1086.599777577" observedRunningTime="2025-10-14 07:19:11.374241361 +0000 UTC m=+1087.071601732" watchObservedRunningTime="2025-10-14 07:19:11.379649628 +0000 UTC m=+1087.077009999" Oct 14 07:19:12 crc kubenswrapper[4870]: I1014 07:19:12.424542 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Oct 14 07:19:12 crc kubenswrapper[4870]: I1014 07:19:12.477939 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Oct 14 07:19:12 crc kubenswrapper[4870]: I1014 07:19:12.937178 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-746b7bb85-l46gj"] Oct 14 07:19:12 crc kubenswrapper[4870]: I1014 07:19:12.963575 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-549b48ff7f-vhsh8"] Oct 14 07:19:12 crc kubenswrapper[4870]: E1014 07:19:12.964028 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="11aab367-2351-4568-9a45-43c79d419bca" containerName="dnsmasq-dns" Oct 14 07:19:12 crc kubenswrapper[4870]: I1014 07:19:12.964046 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="11aab367-2351-4568-9a45-43c79d419bca" containerName="dnsmasq-dns" Oct 14 07:19:12 crc kubenswrapper[4870]: E1014 07:19:12.964070 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="11aab367-2351-4568-9a45-43c79d419bca" containerName="init" Oct 14 07:19:12 crc kubenswrapper[4870]: I1014 07:19:12.964083 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="11aab367-2351-4568-9a45-43c79d419bca" containerName="init" Oct 14 07:19:12 crc kubenswrapper[4870]: I1014 07:19:12.964355 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="11aab367-2351-4568-9a45-43c79d419bca" containerName="dnsmasq-dns" Oct 14 07:19:12 crc kubenswrapper[4870]: I1014 07:19:12.965597 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-549b48ff7f-vhsh8" Oct 14 07:19:12 crc kubenswrapper[4870]: I1014 07:19:12.979295 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-549b48ff7f-vhsh8"] Oct 14 07:19:12 crc kubenswrapper[4870]: I1014 07:19:12.986914 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Oct 14 07:19:13 crc kubenswrapper[4870]: I1014 07:19:13.140562 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/17f2363b-7ac4-4e5d-92e0-0869ab39c98e-ovsdbserver-sb\") pod \"dnsmasq-dns-549b48ff7f-vhsh8\" (UID: \"17f2363b-7ac4-4e5d-92e0-0869ab39c98e\") " pod="openstack/dnsmasq-dns-549b48ff7f-vhsh8" Oct 14 07:19:13 crc kubenswrapper[4870]: I1014 07:19:13.140627 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/17f2363b-7ac4-4e5d-92e0-0869ab39c98e-dns-svc\") pod \"dnsmasq-dns-549b48ff7f-vhsh8\" (UID: \"17f2363b-7ac4-4e5d-92e0-0869ab39c98e\") " pod="openstack/dnsmasq-dns-549b48ff7f-vhsh8" Oct 14 07:19:13 crc kubenswrapper[4870]: I1014 07:19:13.140672 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/17f2363b-7ac4-4e5d-92e0-0869ab39c98e-config\") pod \"dnsmasq-dns-549b48ff7f-vhsh8\" (UID: \"17f2363b-7ac4-4e5d-92e0-0869ab39c98e\") " pod="openstack/dnsmasq-dns-549b48ff7f-vhsh8" Oct 14 07:19:13 crc kubenswrapper[4870]: I1014 07:19:13.140876 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/17f2363b-7ac4-4e5d-92e0-0869ab39c98e-ovsdbserver-nb\") pod \"dnsmasq-dns-549b48ff7f-vhsh8\" (UID: \"17f2363b-7ac4-4e5d-92e0-0869ab39c98e\") " pod="openstack/dnsmasq-dns-549b48ff7f-vhsh8" Oct 14 07:19:13 crc kubenswrapper[4870]: I1014 07:19:13.140991 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ssrhl\" (UniqueName: \"kubernetes.io/projected/17f2363b-7ac4-4e5d-92e0-0869ab39c98e-kube-api-access-ssrhl\") pod \"dnsmasq-dns-549b48ff7f-vhsh8\" (UID: \"17f2363b-7ac4-4e5d-92e0-0869ab39c98e\") " pod="openstack/dnsmasq-dns-549b48ff7f-vhsh8" Oct 14 07:19:13 crc kubenswrapper[4870]: I1014 07:19:13.241975 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/17f2363b-7ac4-4e5d-92e0-0869ab39c98e-ovsdbserver-nb\") pod \"dnsmasq-dns-549b48ff7f-vhsh8\" (UID: \"17f2363b-7ac4-4e5d-92e0-0869ab39c98e\") " pod="openstack/dnsmasq-dns-549b48ff7f-vhsh8" Oct 14 07:19:13 crc kubenswrapper[4870]: I1014 07:19:13.242045 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ssrhl\" (UniqueName: \"kubernetes.io/projected/17f2363b-7ac4-4e5d-92e0-0869ab39c98e-kube-api-access-ssrhl\") pod \"dnsmasq-dns-549b48ff7f-vhsh8\" (UID: \"17f2363b-7ac4-4e5d-92e0-0869ab39c98e\") " pod="openstack/dnsmasq-dns-549b48ff7f-vhsh8" Oct 14 07:19:13 crc kubenswrapper[4870]: I1014 07:19:13.242070 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/17f2363b-7ac4-4e5d-92e0-0869ab39c98e-ovsdbserver-sb\") pod \"dnsmasq-dns-549b48ff7f-vhsh8\" (UID: \"17f2363b-7ac4-4e5d-92e0-0869ab39c98e\") " pod="openstack/dnsmasq-dns-549b48ff7f-vhsh8" Oct 14 07:19:13 crc kubenswrapper[4870]: I1014 07:19:13.242107 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/17f2363b-7ac4-4e5d-92e0-0869ab39c98e-dns-svc\") pod \"dnsmasq-dns-549b48ff7f-vhsh8\" (UID: \"17f2363b-7ac4-4e5d-92e0-0869ab39c98e\") " pod="openstack/dnsmasq-dns-549b48ff7f-vhsh8" Oct 14 07:19:13 crc kubenswrapper[4870]: I1014 07:19:13.242145 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/17f2363b-7ac4-4e5d-92e0-0869ab39c98e-config\") pod \"dnsmasq-dns-549b48ff7f-vhsh8\" (UID: \"17f2363b-7ac4-4e5d-92e0-0869ab39c98e\") " pod="openstack/dnsmasq-dns-549b48ff7f-vhsh8" Oct 14 07:19:13 crc kubenswrapper[4870]: I1014 07:19:13.243212 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/17f2363b-7ac4-4e5d-92e0-0869ab39c98e-dns-svc\") pod \"dnsmasq-dns-549b48ff7f-vhsh8\" (UID: \"17f2363b-7ac4-4e5d-92e0-0869ab39c98e\") " pod="openstack/dnsmasq-dns-549b48ff7f-vhsh8" Oct 14 07:19:13 crc kubenswrapper[4870]: I1014 07:19:13.243214 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/17f2363b-7ac4-4e5d-92e0-0869ab39c98e-ovsdbserver-nb\") pod \"dnsmasq-dns-549b48ff7f-vhsh8\" (UID: \"17f2363b-7ac4-4e5d-92e0-0869ab39c98e\") " pod="openstack/dnsmasq-dns-549b48ff7f-vhsh8" Oct 14 07:19:13 crc kubenswrapper[4870]: I1014 07:19:13.243280 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/17f2363b-7ac4-4e5d-92e0-0869ab39c98e-config\") pod \"dnsmasq-dns-549b48ff7f-vhsh8\" (UID: \"17f2363b-7ac4-4e5d-92e0-0869ab39c98e\") " pod="openstack/dnsmasq-dns-549b48ff7f-vhsh8" Oct 14 07:19:13 crc kubenswrapper[4870]: I1014 07:19:13.243293 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/17f2363b-7ac4-4e5d-92e0-0869ab39c98e-ovsdbserver-sb\") pod \"dnsmasq-dns-549b48ff7f-vhsh8\" (UID: \"17f2363b-7ac4-4e5d-92e0-0869ab39c98e\") " pod="openstack/dnsmasq-dns-549b48ff7f-vhsh8" Oct 14 07:19:13 crc kubenswrapper[4870]: I1014 07:19:13.261328 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ssrhl\" (UniqueName: \"kubernetes.io/projected/17f2363b-7ac4-4e5d-92e0-0869ab39c98e-kube-api-access-ssrhl\") pod \"dnsmasq-dns-549b48ff7f-vhsh8\" (UID: \"17f2363b-7ac4-4e5d-92e0-0869ab39c98e\") " pod="openstack/dnsmasq-dns-549b48ff7f-vhsh8" Oct 14 07:19:13 crc kubenswrapper[4870]: I1014 07:19:13.294494 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-549b48ff7f-vhsh8" Oct 14 07:19:13 crc kubenswrapper[4870]: W1014 07:19:13.717397 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod17f2363b_7ac4_4e5d_92e0_0869ab39c98e.slice/crio-d3928912276e8b221ca8e94547860c384bd7ea04738a14a504f17ed450a446b6 WatchSource:0}: Error finding container d3928912276e8b221ca8e94547860c384bd7ea04738a14a504f17ed450a446b6: Status 404 returned error can't find the container with id d3928912276e8b221ca8e94547860c384bd7ea04738a14a504f17ed450a446b6 Oct 14 07:19:13 crc kubenswrapper[4870]: I1014 07:19:13.719245 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-549b48ff7f-vhsh8"] Oct 14 07:19:14 crc kubenswrapper[4870]: I1014 07:19:14.030150 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Oct 14 07:19:14 crc kubenswrapper[4870]: I1014 07:19:14.035654 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Oct 14 07:19:14 crc kubenswrapper[4870]: I1014 07:19:14.039043 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Oct 14 07:19:14 crc kubenswrapper[4870]: I1014 07:19:14.039237 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Oct 14 07:19:14 crc kubenswrapper[4870]: I1014 07:19:14.039886 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-xcxk4" Oct 14 07:19:14 crc kubenswrapper[4870]: I1014 07:19:14.039922 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Oct 14 07:19:14 crc kubenswrapper[4870]: I1014 07:19:14.068783 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Oct 14 07:19:14 crc kubenswrapper[4870]: I1014 07:19:14.157674 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gsdsq\" (UniqueName: \"kubernetes.io/projected/2bd1f41f-3f1a-4ca7-8789-429104ce2120-kube-api-access-gsdsq\") pod \"swift-storage-0\" (UID: \"2bd1f41f-3f1a-4ca7-8789-429104ce2120\") " pod="openstack/swift-storage-0" Oct 14 07:19:14 crc kubenswrapper[4870]: I1014 07:19:14.157969 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"swift-storage-0\" (UID: \"2bd1f41f-3f1a-4ca7-8789-429104ce2120\") " pod="openstack/swift-storage-0" Oct 14 07:19:14 crc kubenswrapper[4870]: I1014 07:19:14.158083 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/2bd1f41f-3f1a-4ca7-8789-429104ce2120-cache\") pod \"swift-storage-0\" (UID: \"2bd1f41f-3f1a-4ca7-8789-429104ce2120\") " pod="openstack/swift-storage-0" Oct 14 07:19:14 crc kubenswrapper[4870]: I1014 07:19:14.158112 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/2bd1f41f-3f1a-4ca7-8789-429104ce2120-etc-swift\") pod \"swift-storage-0\" (UID: \"2bd1f41f-3f1a-4ca7-8789-429104ce2120\") " pod="openstack/swift-storage-0" Oct 14 07:19:14 crc kubenswrapper[4870]: I1014 07:19:14.158134 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/2bd1f41f-3f1a-4ca7-8789-429104ce2120-lock\") pod \"swift-storage-0\" (UID: \"2bd1f41f-3f1a-4ca7-8789-429104ce2120\") " pod="openstack/swift-storage-0" Oct 14 07:19:14 crc kubenswrapper[4870]: I1014 07:19:14.260485 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/2bd1f41f-3f1a-4ca7-8789-429104ce2120-cache\") pod \"swift-storage-0\" (UID: \"2bd1f41f-3f1a-4ca7-8789-429104ce2120\") " pod="openstack/swift-storage-0" Oct 14 07:19:14 crc kubenswrapper[4870]: I1014 07:19:14.260574 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/2bd1f41f-3f1a-4ca7-8789-429104ce2120-etc-swift\") pod \"swift-storage-0\" (UID: \"2bd1f41f-3f1a-4ca7-8789-429104ce2120\") " pod="openstack/swift-storage-0" Oct 14 07:19:14 crc kubenswrapper[4870]: I1014 07:19:14.260624 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/2bd1f41f-3f1a-4ca7-8789-429104ce2120-lock\") pod \"swift-storage-0\" (UID: \"2bd1f41f-3f1a-4ca7-8789-429104ce2120\") " pod="openstack/swift-storage-0" Oct 14 07:19:14 crc kubenswrapper[4870]: I1014 07:19:14.260733 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gsdsq\" (UniqueName: \"kubernetes.io/projected/2bd1f41f-3f1a-4ca7-8789-429104ce2120-kube-api-access-gsdsq\") pod \"swift-storage-0\" (UID: \"2bd1f41f-3f1a-4ca7-8789-429104ce2120\") " pod="openstack/swift-storage-0" Oct 14 07:19:14 crc kubenswrapper[4870]: I1014 07:19:14.260781 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"swift-storage-0\" (UID: \"2bd1f41f-3f1a-4ca7-8789-429104ce2120\") " pod="openstack/swift-storage-0" Oct 14 07:19:14 crc kubenswrapper[4870]: E1014 07:19:14.260876 4870 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 14 07:19:14 crc kubenswrapper[4870]: E1014 07:19:14.260962 4870 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 14 07:19:14 crc kubenswrapper[4870]: I1014 07:19:14.261307 4870 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"swift-storage-0\" (UID: \"2bd1f41f-3f1a-4ca7-8789-429104ce2120\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/swift-storage-0" Oct 14 07:19:14 crc kubenswrapper[4870]: E1014 07:19:14.261511 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/2bd1f41f-3f1a-4ca7-8789-429104ce2120-etc-swift podName:2bd1f41f-3f1a-4ca7-8789-429104ce2120 nodeName:}" failed. No retries permitted until 2025-10-14 07:19:14.76104341 +0000 UTC m=+1090.458403821 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/2bd1f41f-3f1a-4ca7-8789-429104ce2120-etc-swift") pod "swift-storage-0" (UID: "2bd1f41f-3f1a-4ca7-8789-429104ce2120") : configmap "swift-ring-files" not found Oct 14 07:19:14 crc kubenswrapper[4870]: I1014 07:19:14.261546 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/2bd1f41f-3f1a-4ca7-8789-429104ce2120-lock\") pod \"swift-storage-0\" (UID: \"2bd1f41f-3f1a-4ca7-8789-429104ce2120\") " pod="openstack/swift-storage-0" Oct 14 07:19:14 crc kubenswrapper[4870]: I1014 07:19:14.261556 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/2bd1f41f-3f1a-4ca7-8789-429104ce2120-cache\") pod \"swift-storage-0\" (UID: \"2bd1f41f-3f1a-4ca7-8789-429104ce2120\") " pod="openstack/swift-storage-0" Oct 14 07:19:14 crc kubenswrapper[4870]: I1014 07:19:14.294743 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gsdsq\" (UniqueName: \"kubernetes.io/projected/2bd1f41f-3f1a-4ca7-8789-429104ce2120-kube-api-access-gsdsq\") pod \"swift-storage-0\" (UID: \"2bd1f41f-3f1a-4ca7-8789-429104ce2120\") " pod="openstack/swift-storage-0" Oct 14 07:19:14 crc kubenswrapper[4870]: I1014 07:19:14.303237 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"swift-storage-0\" (UID: \"2bd1f41f-3f1a-4ca7-8789-429104ce2120\") " pod="openstack/swift-storage-0" Oct 14 07:19:14 crc kubenswrapper[4870]: I1014 07:19:14.336874 4870 generic.go:334] "Generic (PLEG): container finished" podID="17f2363b-7ac4-4e5d-92e0-0869ab39c98e" containerID="cf76b6c52610aa2316d561fd70d09c1ffa1e569cd5a36ff134d2a023b2154d84" exitCode=0 Oct 14 07:19:14 crc kubenswrapper[4870]: I1014 07:19:14.336959 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-549b48ff7f-vhsh8" event={"ID":"17f2363b-7ac4-4e5d-92e0-0869ab39c98e","Type":"ContainerDied","Data":"cf76b6c52610aa2316d561fd70d09c1ffa1e569cd5a36ff134d2a023b2154d84"} Oct 14 07:19:14 crc kubenswrapper[4870]: I1014 07:19:14.337042 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-549b48ff7f-vhsh8" event={"ID":"17f2363b-7ac4-4e5d-92e0-0869ab39c98e","Type":"ContainerStarted","Data":"d3928912276e8b221ca8e94547860c384bd7ea04738a14a504f17ed450a446b6"} Oct 14 07:19:14 crc kubenswrapper[4870]: I1014 07:19:14.337233 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-746b7bb85-l46gj" podUID="1142fdf2-8706-464b-9607-7b1fd093d7cf" containerName="dnsmasq-dns" containerID="cri-o://4dfd29d1bf7a2d5610f6a802f7d18bd7bcfd0364ecf1fcbe7b15495aa87493e1" gracePeriod=10 Oct 14 07:19:14 crc kubenswrapper[4870]: I1014 07:19:14.541470 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-lqpbp"] Oct 14 07:19:14 crc kubenswrapper[4870]: I1014 07:19:14.543419 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-lqpbp" Oct 14 07:19:14 crc kubenswrapper[4870]: I1014 07:19:14.546458 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Oct 14 07:19:14 crc kubenswrapper[4870]: I1014 07:19:14.546636 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Oct 14 07:19:14 crc kubenswrapper[4870]: I1014 07:19:14.547218 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Oct 14 07:19:14 crc kubenswrapper[4870]: I1014 07:19:14.549416 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-lqpbp"] Oct 14 07:19:14 crc kubenswrapper[4870]: I1014 07:19:14.671761 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/73ffa9f3-6390-4ed0-b0ba-5a3e2a44ae1f-scripts\") pod \"swift-ring-rebalance-lqpbp\" (UID: \"73ffa9f3-6390-4ed0-b0ba-5a3e2a44ae1f\") " pod="openstack/swift-ring-rebalance-lqpbp" Oct 14 07:19:14 crc kubenswrapper[4870]: I1014 07:19:14.672146 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/73ffa9f3-6390-4ed0-b0ba-5a3e2a44ae1f-swiftconf\") pod \"swift-ring-rebalance-lqpbp\" (UID: \"73ffa9f3-6390-4ed0-b0ba-5a3e2a44ae1f\") " pod="openstack/swift-ring-rebalance-lqpbp" Oct 14 07:19:14 crc kubenswrapper[4870]: I1014 07:19:14.672177 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/73ffa9f3-6390-4ed0-b0ba-5a3e2a44ae1f-combined-ca-bundle\") pod \"swift-ring-rebalance-lqpbp\" (UID: \"73ffa9f3-6390-4ed0-b0ba-5a3e2a44ae1f\") " pod="openstack/swift-ring-rebalance-lqpbp" Oct 14 07:19:14 crc kubenswrapper[4870]: I1014 07:19:14.672206 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/73ffa9f3-6390-4ed0-b0ba-5a3e2a44ae1f-ring-data-devices\") pod \"swift-ring-rebalance-lqpbp\" (UID: \"73ffa9f3-6390-4ed0-b0ba-5a3e2a44ae1f\") " pod="openstack/swift-ring-rebalance-lqpbp" Oct 14 07:19:14 crc kubenswrapper[4870]: I1014 07:19:14.672236 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/73ffa9f3-6390-4ed0-b0ba-5a3e2a44ae1f-etc-swift\") pod \"swift-ring-rebalance-lqpbp\" (UID: \"73ffa9f3-6390-4ed0-b0ba-5a3e2a44ae1f\") " pod="openstack/swift-ring-rebalance-lqpbp" Oct 14 07:19:14 crc kubenswrapper[4870]: I1014 07:19:14.672270 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gwhpv\" (UniqueName: \"kubernetes.io/projected/73ffa9f3-6390-4ed0-b0ba-5a3e2a44ae1f-kube-api-access-gwhpv\") pod \"swift-ring-rebalance-lqpbp\" (UID: \"73ffa9f3-6390-4ed0-b0ba-5a3e2a44ae1f\") " pod="openstack/swift-ring-rebalance-lqpbp" Oct 14 07:19:14 crc kubenswrapper[4870]: I1014 07:19:14.672296 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/73ffa9f3-6390-4ed0-b0ba-5a3e2a44ae1f-dispersionconf\") pod \"swift-ring-rebalance-lqpbp\" (UID: \"73ffa9f3-6390-4ed0-b0ba-5a3e2a44ae1f\") " pod="openstack/swift-ring-rebalance-lqpbp" Oct 14 07:19:14 crc kubenswrapper[4870]: I1014 07:19:14.752901 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-746b7bb85-l46gj" Oct 14 07:19:14 crc kubenswrapper[4870]: I1014 07:19:14.773415 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/73ffa9f3-6390-4ed0-b0ba-5a3e2a44ae1f-scripts\") pod \"swift-ring-rebalance-lqpbp\" (UID: \"73ffa9f3-6390-4ed0-b0ba-5a3e2a44ae1f\") " pod="openstack/swift-ring-rebalance-lqpbp" Oct 14 07:19:14 crc kubenswrapper[4870]: I1014 07:19:14.773503 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/2bd1f41f-3f1a-4ca7-8789-429104ce2120-etc-swift\") pod \"swift-storage-0\" (UID: \"2bd1f41f-3f1a-4ca7-8789-429104ce2120\") " pod="openstack/swift-storage-0" Oct 14 07:19:14 crc kubenswrapper[4870]: I1014 07:19:14.773584 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/73ffa9f3-6390-4ed0-b0ba-5a3e2a44ae1f-swiftconf\") pod \"swift-ring-rebalance-lqpbp\" (UID: \"73ffa9f3-6390-4ed0-b0ba-5a3e2a44ae1f\") " pod="openstack/swift-ring-rebalance-lqpbp" Oct 14 07:19:14 crc kubenswrapper[4870]: I1014 07:19:14.773608 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/73ffa9f3-6390-4ed0-b0ba-5a3e2a44ae1f-combined-ca-bundle\") pod \"swift-ring-rebalance-lqpbp\" (UID: \"73ffa9f3-6390-4ed0-b0ba-5a3e2a44ae1f\") " pod="openstack/swift-ring-rebalance-lqpbp" Oct 14 07:19:14 crc kubenswrapper[4870]: I1014 07:19:14.773631 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/73ffa9f3-6390-4ed0-b0ba-5a3e2a44ae1f-ring-data-devices\") pod \"swift-ring-rebalance-lqpbp\" (UID: \"73ffa9f3-6390-4ed0-b0ba-5a3e2a44ae1f\") " pod="openstack/swift-ring-rebalance-lqpbp" Oct 14 07:19:14 crc kubenswrapper[4870]: I1014 07:19:14.773663 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/73ffa9f3-6390-4ed0-b0ba-5a3e2a44ae1f-etc-swift\") pod \"swift-ring-rebalance-lqpbp\" (UID: \"73ffa9f3-6390-4ed0-b0ba-5a3e2a44ae1f\") " pod="openstack/swift-ring-rebalance-lqpbp" Oct 14 07:19:14 crc kubenswrapper[4870]: E1014 07:19:14.773672 4870 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 14 07:19:14 crc kubenswrapper[4870]: E1014 07:19:14.774016 4870 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 14 07:19:14 crc kubenswrapper[4870]: I1014 07:19:14.774340 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/73ffa9f3-6390-4ed0-b0ba-5a3e2a44ae1f-etc-swift\") pod \"swift-ring-rebalance-lqpbp\" (UID: \"73ffa9f3-6390-4ed0-b0ba-5a3e2a44ae1f\") " pod="openstack/swift-ring-rebalance-lqpbp" Oct 14 07:19:14 crc kubenswrapper[4870]: I1014 07:19:14.773699 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gwhpv\" (UniqueName: \"kubernetes.io/projected/73ffa9f3-6390-4ed0-b0ba-5a3e2a44ae1f-kube-api-access-gwhpv\") pod \"swift-ring-rebalance-lqpbp\" (UID: \"73ffa9f3-6390-4ed0-b0ba-5a3e2a44ae1f\") " pod="openstack/swift-ring-rebalance-lqpbp" Oct 14 07:19:14 crc kubenswrapper[4870]: I1014 07:19:14.774453 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/73ffa9f3-6390-4ed0-b0ba-5a3e2a44ae1f-dispersionconf\") pod \"swift-ring-rebalance-lqpbp\" (UID: \"73ffa9f3-6390-4ed0-b0ba-5a3e2a44ae1f\") " pod="openstack/swift-ring-rebalance-lqpbp" Oct 14 07:19:14 crc kubenswrapper[4870]: I1014 07:19:14.774533 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/73ffa9f3-6390-4ed0-b0ba-5a3e2a44ae1f-scripts\") pod \"swift-ring-rebalance-lqpbp\" (UID: \"73ffa9f3-6390-4ed0-b0ba-5a3e2a44ae1f\") " pod="openstack/swift-ring-rebalance-lqpbp" Oct 14 07:19:14 crc kubenswrapper[4870]: I1014 07:19:14.774734 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/73ffa9f3-6390-4ed0-b0ba-5a3e2a44ae1f-ring-data-devices\") pod \"swift-ring-rebalance-lqpbp\" (UID: \"73ffa9f3-6390-4ed0-b0ba-5a3e2a44ae1f\") " pod="openstack/swift-ring-rebalance-lqpbp" Oct 14 07:19:14 crc kubenswrapper[4870]: E1014 07:19:14.774878 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/2bd1f41f-3f1a-4ca7-8789-429104ce2120-etc-swift podName:2bd1f41f-3f1a-4ca7-8789-429104ce2120 nodeName:}" failed. No retries permitted until 2025-10-14 07:19:15.774847935 +0000 UTC m=+1091.472208346 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/2bd1f41f-3f1a-4ca7-8789-429104ce2120-etc-swift") pod "swift-storage-0" (UID: "2bd1f41f-3f1a-4ca7-8789-429104ce2120") : configmap "swift-ring-files" not found Oct 14 07:19:14 crc kubenswrapper[4870]: I1014 07:19:14.779674 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/73ffa9f3-6390-4ed0-b0ba-5a3e2a44ae1f-swiftconf\") pod \"swift-ring-rebalance-lqpbp\" (UID: \"73ffa9f3-6390-4ed0-b0ba-5a3e2a44ae1f\") " pod="openstack/swift-ring-rebalance-lqpbp" Oct 14 07:19:14 crc kubenswrapper[4870]: I1014 07:19:14.780128 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/73ffa9f3-6390-4ed0-b0ba-5a3e2a44ae1f-combined-ca-bundle\") pod \"swift-ring-rebalance-lqpbp\" (UID: \"73ffa9f3-6390-4ed0-b0ba-5a3e2a44ae1f\") " pod="openstack/swift-ring-rebalance-lqpbp" Oct 14 07:19:14 crc kubenswrapper[4870]: I1014 07:19:14.790397 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/73ffa9f3-6390-4ed0-b0ba-5a3e2a44ae1f-dispersionconf\") pod \"swift-ring-rebalance-lqpbp\" (UID: \"73ffa9f3-6390-4ed0-b0ba-5a3e2a44ae1f\") " pod="openstack/swift-ring-rebalance-lqpbp" Oct 14 07:19:14 crc kubenswrapper[4870]: I1014 07:19:14.801949 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gwhpv\" (UniqueName: \"kubernetes.io/projected/73ffa9f3-6390-4ed0-b0ba-5a3e2a44ae1f-kube-api-access-gwhpv\") pod \"swift-ring-rebalance-lqpbp\" (UID: \"73ffa9f3-6390-4ed0-b0ba-5a3e2a44ae1f\") " pod="openstack/swift-ring-rebalance-lqpbp" Oct 14 07:19:14 crc kubenswrapper[4870]: I1014 07:19:14.875018 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-lqpbp" Oct 14 07:19:14 crc kubenswrapper[4870]: I1014 07:19:14.875174 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1142fdf2-8706-464b-9607-7b1fd093d7cf-config\") pod \"1142fdf2-8706-464b-9607-7b1fd093d7cf\" (UID: \"1142fdf2-8706-464b-9607-7b1fd093d7cf\") " Oct 14 07:19:14 crc kubenswrapper[4870]: I1014 07:19:14.875309 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1142fdf2-8706-464b-9607-7b1fd093d7cf-dns-svc\") pod \"1142fdf2-8706-464b-9607-7b1fd093d7cf\" (UID: \"1142fdf2-8706-464b-9607-7b1fd093d7cf\") " Oct 14 07:19:14 crc kubenswrapper[4870]: I1014 07:19:14.875409 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jlf6g\" (UniqueName: \"kubernetes.io/projected/1142fdf2-8706-464b-9607-7b1fd093d7cf-kube-api-access-jlf6g\") pod \"1142fdf2-8706-464b-9607-7b1fd093d7cf\" (UID: \"1142fdf2-8706-464b-9607-7b1fd093d7cf\") " Oct 14 07:19:14 crc kubenswrapper[4870]: I1014 07:19:14.875546 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1142fdf2-8706-464b-9607-7b1fd093d7cf-ovsdbserver-nb\") pod \"1142fdf2-8706-464b-9607-7b1fd093d7cf\" (UID: \"1142fdf2-8706-464b-9607-7b1fd093d7cf\") " Oct 14 07:19:14 crc kubenswrapper[4870]: I1014 07:19:14.880424 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1142fdf2-8706-464b-9607-7b1fd093d7cf-kube-api-access-jlf6g" (OuterVolumeSpecName: "kube-api-access-jlf6g") pod "1142fdf2-8706-464b-9607-7b1fd093d7cf" (UID: "1142fdf2-8706-464b-9607-7b1fd093d7cf"). InnerVolumeSpecName "kube-api-access-jlf6g". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:19:14 crc kubenswrapper[4870]: I1014 07:19:14.919637 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1142fdf2-8706-464b-9607-7b1fd093d7cf-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "1142fdf2-8706-464b-9607-7b1fd093d7cf" (UID: "1142fdf2-8706-464b-9607-7b1fd093d7cf"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:19:14 crc kubenswrapper[4870]: I1014 07:19:14.947527 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1142fdf2-8706-464b-9607-7b1fd093d7cf-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "1142fdf2-8706-464b-9607-7b1fd093d7cf" (UID: "1142fdf2-8706-464b-9607-7b1fd093d7cf"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:19:14 crc kubenswrapper[4870]: I1014 07:19:14.947859 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1142fdf2-8706-464b-9607-7b1fd093d7cf-config" (OuterVolumeSpecName: "config") pod "1142fdf2-8706-464b-9607-7b1fd093d7cf" (UID: "1142fdf2-8706-464b-9607-7b1fd093d7cf"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:19:14 crc kubenswrapper[4870]: I1014 07:19:14.974220 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Oct 14 07:19:14 crc kubenswrapper[4870]: I1014 07:19:14.980389 4870 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1142fdf2-8706-464b-9607-7b1fd093d7cf-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 14 07:19:14 crc kubenswrapper[4870]: I1014 07:19:14.980965 4870 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1142fdf2-8706-464b-9607-7b1fd093d7cf-config\") on node \"crc\" DevicePath \"\"" Oct 14 07:19:14 crc kubenswrapper[4870]: I1014 07:19:14.981267 4870 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1142fdf2-8706-464b-9607-7b1fd093d7cf-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 14 07:19:14 crc kubenswrapper[4870]: I1014 07:19:14.981283 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jlf6g\" (UniqueName: \"kubernetes.io/projected/1142fdf2-8706-464b-9607-7b1fd093d7cf-kube-api-access-jlf6g\") on node \"crc\" DevicePath \"\"" Oct 14 07:19:15 crc kubenswrapper[4870]: I1014 07:19:15.063134 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Oct 14 07:19:15 crc kubenswrapper[4870]: I1014 07:19:15.327559 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-lqpbp"] Oct 14 07:19:15 crc kubenswrapper[4870]: W1014 07:19:15.333842 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod73ffa9f3_6390_4ed0_b0ba_5a3e2a44ae1f.slice/crio-5406ff5ba7fe072aac7d4a55694936460e4bf0eda7d0f06b2e50df806b27b7cb WatchSource:0}: Error finding container 5406ff5ba7fe072aac7d4a55694936460e4bf0eda7d0f06b2e50df806b27b7cb: Status 404 returned error can't find the container with id 5406ff5ba7fe072aac7d4a55694936460e4bf0eda7d0f06b2e50df806b27b7cb Oct 14 07:19:15 crc kubenswrapper[4870]: I1014 07:19:15.354342 4870 generic.go:334] "Generic (PLEG): container finished" podID="1142fdf2-8706-464b-9607-7b1fd093d7cf" containerID="4dfd29d1bf7a2d5610f6a802f7d18bd7bcfd0364ecf1fcbe7b15495aa87493e1" exitCode=0 Oct 14 07:19:15 crc kubenswrapper[4870]: I1014 07:19:15.354415 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-746b7bb85-l46gj" event={"ID":"1142fdf2-8706-464b-9607-7b1fd093d7cf","Type":"ContainerDied","Data":"4dfd29d1bf7a2d5610f6a802f7d18bd7bcfd0364ecf1fcbe7b15495aa87493e1"} Oct 14 07:19:15 crc kubenswrapper[4870]: I1014 07:19:15.354474 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-746b7bb85-l46gj" event={"ID":"1142fdf2-8706-464b-9607-7b1fd093d7cf","Type":"ContainerDied","Data":"65a1296dc09c948f28550ebfb2cd92e0576cccbcc210ddd6df4c2208fe155912"} Oct 14 07:19:15 crc kubenswrapper[4870]: I1014 07:19:15.354497 4870 scope.go:117] "RemoveContainer" containerID="4dfd29d1bf7a2d5610f6a802f7d18bd7bcfd0364ecf1fcbe7b15495aa87493e1" Oct 14 07:19:15 crc kubenswrapper[4870]: I1014 07:19:15.354654 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-746b7bb85-l46gj" Oct 14 07:19:15 crc kubenswrapper[4870]: I1014 07:19:15.361780 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-lqpbp" event={"ID":"73ffa9f3-6390-4ed0-b0ba-5a3e2a44ae1f","Type":"ContainerStarted","Data":"5406ff5ba7fe072aac7d4a55694936460e4bf0eda7d0f06b2e50df806b27b7cb"} Oct 14 07:19:15 crc kubenswrapper[4870]: I1014 07:19:15.369550 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-549b48ff7f-vhsh8" event={"ID":"17f2363b-7ac4-4e5d-92e0-0869ab39c98e","Type":"ContainerStarted","Data":"c3d069abc429cccb2df1fa8e60ce1cfc0cc23b619edfc9bd8c05a1e88471f1bc"} Oct 14 07:19:15 crc kubenswrapper[4870]: I1014 07:19:15.369589 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-549b48ff7f-vhsh8" Oct 14 07:19:15 crc kubenswrapper[4870]: I1014 07:19:15.391028 4870 scope.go:117] "RemoveContainer" containerID="0c0369e5032cbb5a6790a98dd7a1df20a98e32c588b729d7406be04b578ed108" Oct 14 07:19:15 crc kubenswrapper[4870]: I1014 07:19:15.396294 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-549b48ff7f-vhsh8" podStartSLOduration=3.396274585 podStartE2EDuration="3.396274585s" podCreationTimestamp="2025-10-14 07:19:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:19:15.387932954 +0000 UTC m=+1091.085293325" watchObservedRunningTime="2025-10-14 07:19:15.396274585 +0000 UTC m=+1091.093634956" Oct 14 07:19:15 crc kubenswrapper[4870]: I1014 07:19:15.415000 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-746b7bb85-l46gj"] Oct 14 07:19:15 crc kubenswrapper[4870]: I1014 07:19:15.415177 4870 scope.go:117] "RemoveContainer" containerID="4dfd29d1bf7a2d5610f6a802f7d18bd7bcfd0364ecf1fcbe7b15495aa87493e1" Oct 14 07:19:15 crc kubenswrapper[4870]: E1014 07:19:15.415669 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4dfd29d1bf7a2d5610f6a802f7d18bd7bcfd0364ecf1fcbe7b15495aa87493e1\": container with ID starting with 4dfd29d1bf7a2d5610f6a802f7d18bd7bcfd0364ecf1fcbe7b15495aa87493e1 not found: ID does not exist" containerID="4dfd29d1bf7a2d5610f6a802f7d18bd7bcfd0364ecf1fcbe7b15495aa87493e1" Oct 14 07:19:15 crc kubenswrapper[4870]: I1014 07:19:15.415718 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4dfd29d1bf7a2d5610f6a802f7d18bd7bcfd0364ecf1fcbe7b15495aa87493e1"} err="failed to get container status \"4dfd29d1bf7a2d5610f6a802f7d18bd7bcfd0364ecf1fcbe7b15495aa87493e1\": rpc error: code = NotFound desc = could not find container \"4dfd29d1bf7a2d5610f6a802f7d18bd7bcfd0364ecf1fcbe7b15495aa87493e1\": container with ID starting with 4dfd29d1bf7a2d5610f6a802f7d18bd7bcfd0364ecf1fcbe7b15495aa87493e1 not found: ID does not exist" Oct 14 07:19:15 crc kubenswrapper[4870]: I1014 07:19:15.415744 4870 scope.go:117] "RemoveContainer" containerID="0c0369e5032cbb5a6790a98dd7a1df20a98e32c588b729d7406be04b578ed108" Oct 14 07:19:15 crc kubenswrapper[4870]: E1014 07:19:15.416115 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0c0369e5032cbb5a6790a98dd7a1df20a98e32c588b729d7406be04b578ed108\": container with ID starting with 0c0369e5032cbb5a6790a98dd7a1df20a98e32c588b729d7406be04b578ed108 not found: ID does not exist" containerID="0c0369e5032cbb5a6790a98dd7a1df20a98e32c588b729d7406be04b578ed108" Oct 14 07:19:15 crc kubenswrapper[4870]: I1014 07:19:15.416179 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0c0369e5032cbb5a6790a98dd7a1df20a98e32c588b729d7406be04b578ed108"} err="failed to get container status \"0c0369e5032cbb5a6790a98dd7a1df20a98e32c588b729d7406be04b578ed108\": rpc error: code = NotFound desc = could not find container \"0c0369e5032cbb5a6790a98dd7a1df20a98e32c588b729d7406be04b578ed108\": container with ID starting with 0c0369e5032cbb5a6790a98dd7a1df20a98e32c588b729d7406be04b578ed108 not found: ID does not exist" Oct 14 07:19:15 crc kubenswrapper[4870]: I1014 07:19:15.418172 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-746b7bb85-l46gj"] Oct 14 07:19:15 crc kubenswrapper[4870]: I1014 07:19:15.794638 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/2bd1f41f-3f1a-4ca7-8789-429104ce2120-etc-swift\") pod \"swift-storage-0\" (UID: \"2bd1f41f-3f1a-4ca7-8789-429104ce2120\") " pod="openstack/swift-storage-0" Oct 14 07:19:15 crc kubenswrapper[4870]: E1014 07:19:15.794951 4870 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 14 07:19:15 crc kubenswrapper[4870]: E1014 07:19:15.794971 4870 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 14 07:19:15 crc kubenswrapper[4870]: E1014 07:19:15.795023 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/2bd1f41f-3f1a-4ca7-8789-429104ce2120-etc-swift podName:2bd1f41f-3f1a-4ca7-8789-429104ce2120 nodeName:}" failed. No retries permitted until 2025-10-14 07:19:17.795004416 +0000 UTC m=+1093.492364797 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/2bd1f41f-3f1a-4ca7-8789-429104ce2120-etc-swift") pod "swift-storage-0" (UID: "2bd1f41f-3f1a-4ca7-8789-429104ce2120") : configmap "swift-ring-files" not found Oct 14 07:19:17 crc kubenswrapper[4870]: I1014 07:19:17.045580 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1142fdf2-8706-464b-9607-7b1fd093d7cf" path="/var/lib/kubelet/pods/1142fdf2-8706-464b-9607-7b1fd093d7cf/volumes" Oct 14 07:19:17 crc kubenswrapper[4870]: I1014 07:19:17.830929 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/2bd1f41f-3f1a-4ca7-8789-429104ce2120-etc-swift\") pod \"swift-storage-0\" (UID: \"2bd1f41f-3f1a-4ca7-8789-429104ce2120\") " pod="openstack/swift-storage-0" Oct 14 07:19:17 crc kubenswrapper[4870]: E1014 07:19:17.831541 4870 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 14 07:19:17 crc kubenswrapper[4870]: E1014 07:19:17.831567 4870 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 14 07:19:17 crc kubenswrapper[4870]: E1014 07:19:17.831626 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/2bd1f41f-3f1a-4ca7-8789-429104ce2120-etc-swift podName:2bd1f41f-3f1a-4ca7-8789-429104ce2120 nodeName:}" failed. No retries permitted until 2025-10-14 07:19:21.831607893 +0000 UTC m=+1097.528968284 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/2bd1f41f-3f1a-4ca7-8789-429104ce2120-etc-swift") pod "swift-storage-0" (UID: "2bd1f41f-3f1a-4ca7-8789-429104ce2120") : configmap "swift-ring-files" not found Oct 14 07:19:19 crc kubenswrapper[4870]: I1014 07:19:19.142602 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7764bd4845-4zlrf" Oct 14 07:19:19 crc kubenswrapper[4870]: I1014 07:19:19.406700 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-lqpbp" event={"ID":"73ffa9f3-6390-4ed0-b0ba-5a3e2a44ae1f","Type":"ContainerStarted","Data":"365d48ec2f5364c81df1e76cb37b8b718e730cf770131932caead0d2a5c0f467"} Oct 14 07:19:19 crc kubenswrapper[4870]: I1014 07:19:19.431648 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-lqpbp" podStartSLOduration=2.276426623 podStartE2EDuration="5.431631245s" podCreationTimestamp="2025-10-14 07:19:14 +0000 UTC" firstStartedPulling="2025-10-14 07:19:15.336790664 +0000 UTC m=+1091.034151035" lastFinishedPulling="2025-10-14 07:19:18.491995286 +0000 UTC m=+1094.189355657" observedRunningTime="2025-10-14 07:19:19.426810343 +0000 UTC m=+1095.124170714" watchObservedRunningTime="2025-10-14 07:19:19.431631245 +0000 UTC m=+1095.128991616" Oct 14 07:19:20 crc kubenswrapper[4870]: I1014 07:19:20.314157 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-4dkbn"] Oct 14 07:19:20 crc kubenswrapper[4870]: E1014 07:19:20.314901 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1142fdf2-8706-464b-9607-7b1fd093d7cf" containerName="init" Oct 14 07:19:20 crc kubenswrapper[4870]: I1014 07:19:20.314918 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="1142fdf2-8706-464b-9607-7b1fd093d7cf" containerName="init" Oct 14 07:19:20 crc kubenswrapper[4870]: E1014 07:19:20.314952 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1142fdf2-8706-464b-9607-7b1fd093d7cf" containerName="dnsmasq-dns" Oct 14 07:19:20 crc kubenswrapper[4870]: I1014 07:19:20.314961 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="1142fdf2-8706-464b-9607-7b1fd093d7cf" containerName="dnsmasq-dns" Oct 14 07:19:20 crc kubenswrapper[4870]: I1014 07:19:20.315186 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="1142fdf2-8706-464b-9607-7b1fd093d7cf" containerName="dnsmasq-dns" Oct 14 07:19:20 crc kubenswrapper[4870]: I1014 07:19:20.315937 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-4dkbn" Oct 14 07:19:20 crc kubenswrapper[4870]: I1014 07:19:20.327468 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-4dkbn"] Oct 14 07:19:20 crc kubenswrapper[4870]: I1014 07:19:20.485072 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qwkjm\" (UniqueName: \"kubernetes.io/projected/25f5ffd0-3c55-46ee-ade3-ec4cd9471e21-kube-api-access-qwkjm\") pod \"keystone-db-create-4dkbn\" (UID: \"25f5ffd0-3c55-46ee-ade3-ec4cd9471e21\") " pod="openstack/keystone-db-create-4dkbn" Oct 14 07:19:20 crc kubenswrapper[4870]: I1014 07:19:20.530357 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-nzvfg"] Oct 14 07:19:20 crc kubenswrapper[4870]: I1014 07:19:20.532037 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-nzvfg" Oct 14 07:19:20 crc kubenswrapper[4870]: I1014 07:19:20.540524 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-nzvfg"] Oct 14 07:19:20 crc kubenswrapper[4870]: I1014 07:19:20.587132 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qwkjm\" (UniqueName: \"kubernetes.io/projected/25f5ffd0-3c55-46ee-ade3-ec4cd9471e21-kube-api-access-qwkjm\") pod \"keystone-db-create-4dkbn\" (UID: \"25f5ffd0-3c55-46ee-ade3-ec4cd9471e21\") " pod="openstack/keystone-db-create-4dkbn" Oct 14 07:19:20 crc kubenswrapper[4870]: I1014 07:19:20.611353 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qwkjm\" (UniqueName: \"kubernetes.io/projected/25f5ffd0-3c55-46ee-ade3-ec4cd9471e21-kube-api-access-qwkjm\") pod \"keystone-db-create-4dkbn\" (UID: \"25f5ffd0-3c55-46ee-ade3-ec4cd9471e21\") " pod="openstack/keystone-db-create-4dkbn" Oct 14 07:19:20 crc kubenswrapper[4870]: I1014 07:19:20.644076 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-4dkbn" Oct 14 07:19:20 crc kubenswrapper[4870]: I1014 07:19:20.688802 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9l54v\" (UniqueName: \"kubernetes.io/projected/835bb9d2-3b63-4fa5-bda6-9cc6a9a899a2-kube-api-access-9l54v\") pod \"placement-db-create-nzvfg\" (UID: \"835bb9d2-3b63-4fa5-bda6-9cc6a9a899a2\") " pod="openstack/placement-db-create-nzvfg" Oct 14 07:19:20 crc kubenswrapper[4870]: I1014 07:19:20.784699 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-f6n86"] Oct 14 07:19:20 crc kubenswrapper[4870]: I1014 07:19:20.786835 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-f6n86" Oct 14 07:19:20 crc kubenswrapper[4870]: I1014 07:19:20.790687 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9l54v\" (UniqueName: \"kubernetes.io/projected/835bb9d2-3b63-4fa5-bda6-9cc6a9a899a2-kube-api-access-9l54v\") pod \"placement-db-create-nzvfg\" (UID: \"835bb9d2-3b63-4fa5-bda6-9cc6a9a899a2\") " pod="openstack/placement-db-create-nzvfg" Oct 14 07:19:20 crc kubenswrapper[4870]: I1014 07:19:20.798123 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-f6n86"] Oct 14 07:19:20 crc kubenswrapper[4870]: I1014 07:19:20.811281 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9l54v\" (UniqueName: \"kubernetes.io/projected/835bb9d2-3b63-4fa5-bda6-9cc6a9a899a2-kube-api-access-9l54v\") pod \"placement-db-create-nzvfg\" (UID: \"835bb9d2-3b63-4fa5-bda6-9cc6a9a899a2\") " pod="openstack/placement-db-create-nzvfg" Oct 14 07:19:20 crc kubenswrapper[4870]: I1014 07:19:20.849812 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-nzvfg" Oct 14 07:19:20 crc kubenswrapper[4870]: I1014 07:19:20.892618 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wd6bp\" (UniqueName: \"kubernetes.io/projected/cd4cf561-b2b4-486b-b317-bd88b71badd9-kube-api-access-wd6bp\") pod \"glance-db-create-f6n86\" (UID: \"cd4cf561-b2b4-486b-b317-bd88b71badd9\") " pod="openstack/glance-db-create-f6n86" Oct 14 07:19:20 crc kubenswrapper[4870]: I1014 07:19:20.993786 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wd6bp\" (UniqueName: \"kubernetes.io/projected/cd4cf561-b2b4-486b-b317-bd88b71badd9-kube-api-access-wd6bp\") pod \"glance-db-create-f6n86\" (UID: \"cd4cf561-b2b4-486b-b317-bd88b71badd9\") " pod="openstack/glance-db-create-f6n86" Oct 14 07:19:21 crc kubenswrapper[4870]: I1014 07:19:21.013312 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wd6bp\" (UniqueName: \"kubernetes.io/projected/cd4cf561-b2b4-486b-b317-bd88b71badd9-kube-api-access-wd6bp\") pod \"glance-db-create-f6n86\" (UID: \"cd4cf561-b2b4-486b-b317-bd88b71badd9\") " pod="openstack/glance-db-create-f6n86" Oct 14 07:19:21 crc kubenswrapper[4870]: I1014 07:19:21.107815 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-f6n86" Oct 14 07:19:21 crc kubenswrapper[4870]: I1014 07:19:21.123220 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-4dkbn"] Oct 14 07:19:21 crc kubenswrapper[4870]: W1014 07:19:21.140784 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod25f5ffd0_3c55_46ee_ade3_ec4cd9471e21.slice/crio-fb73b2ee7af0c79ed7904e3e29c0d47808c2afac58abd694d97e9a16014fc1d0 WatchSource:0}: Error finding container fb73b2ee7af0c79ed7904e3e29c0d47808c2afac58abd694d97e9a16014fc1d0: Status 404 returned error can't find the container with id fb73b2ee7af0c79ed7904e3e29c0d47808c2afac58abd694d97e9a16014fc1d0 Oct 14 07:19:21 crc kubenswrapper[4870]: I1014 07:19:21.315281 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-nzvfg"] Oct 14 07:19:21 crc kubenswrapper[4870]: W1014 07:19:21.318875 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod835bb9d2_3b63_4fa5_bda6_9cc6a9a899a2.slice/crio-e5cd462957537fad952f36e02700fe1f4f2270bd80bd8f73defbfa6d8f38b673 WatchSource:0}: Error finding container e5cd462957537fad952f36e02700fe1f4f2270bd80bd8f73defbfa6d8f38b673: Status 404 returned error can't find the container with id e5cd462957537fad952f36e02700fe1f4f2270bd80bd8f73defbfa6d8f38b673 Oct 14 07:19:21 crc kubenswrapper[4870]: I1014 07:19:21.422914 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-4dkbn" event={"ID":"25f5ffd0-3c55-46ee-ade3-ec4cd9471e21","Type":"ContainerStarted","Data":"eba786db77e4785126a35b98eed013099cfce39a7e939fc837fb1037d741e55c"} Oct 14 07:19:21 crc kubenswrapper[4870]: I1014 07:19:21.422982 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-4dkbn" event={"ID":"25f5ffd0-3c55-46ee-ade3-ec4cd9471e21","Type":"ContainerStarted","Data":"fb73b2ee7af0c79ed7904e3e29c0d47808c2afac58abd694d97e9a16014fc1d0"} Oct 14 07:19:21 crc kubenswrapper[4870]: I1014 07:19:21.424390 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-nzvfg" event={"ID":"835bb9d2-3b63-4fa5-bda6-9cc6a9a899a2","Type":"ContainerStarted","Data":"e5cd462957537fad952f36e02700fe1f4f2270bd80bd8f73defbfa6d8f38b673"} Oct 14 07:19:21 crc kubenswrapper[4870]: I1014 07:19:21.451623 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-create-4dkbn" podStartSLOduration=1.451597431 podStartE2EDuration="1.451597431s" podCreationTimestamp="2025-10-14 07:19:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:19:21.442346677 +0000 UTC m=+1097.139707058" watchObservedRunningTime="2025-10-14 07:19:21.451597431 +0000 UTC m=+1097.148957842" Oct 14 07:19:21 crc kubenswrapper[4870]: I1014 07:19:21.625801 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-f6n86"] Oct 14 07:19:21 crc kubenswrapper[4870]: W1014 07:19:21.647948 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcd4cf561_b2b4_486b_b317_bd88b71badd9.slice/crio-0fb25087f506075020a5ce946f1d3fe048514e21bc00cf9988ca478f45307151 WatchSource:0}: Error finding container 0fb25087f506075020a5ce946f1d3fe048514e21bc00cf9988ca478f45307151: Status 404 returned error can't find the container with id 0fb25087f506075020a5ce946f1d3fe048514e21bc00cf9988ca478f45307151 Oct 14 07:19:21 crc kubenswrapper[4870]: I1014 07:19:21.910179 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/2bd1f41f-3f1a-4ca7-8789-429104ce2120-etc-swift\") pod \"swift-storage-0\" (UID: \"2bd1f41f-3f1a-4ca7-8789-429104ce2120\") " pod="openstack/swift-storage-0" Oct 14 07:19:21 crc kubenswrapper[4870]: E1014 07:19:21.910610 4870 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 14 07:19:21 crc kubenswrapper[4870]: E1014 07:19:21.910686 4870 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 14 07:19:21 crc kubenswrapper[4870]: E1014 07:19:21.911108 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/2bd1f41f-3f1a-4ca7-8789-429104ce2120-etc-swift podName:2bd1f41f-3f1a-4ca7-8789-429104ce2120 nodeName:}" failed. No retries permitted until 2025-10-14 07:19:29.910786927 +0000 UTC m=+1105.608147328 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/2bd1f41f-3f1a-4ca7-8789-429104ce2120-etc-swift") pod "swift-storage-0" (UID: "2bd1f41f-3f1a-4ca7-8789-429104ce2120") : configmap "swift-ring-files" not found Oct 14 07:19:22 crc kubenswrapper[4870]: I1014 07:19:22.449017 4870 generic.go:334] "Generic (PLEG): container finished" podID="25f5ffd0-3c55-46ee-ade3-ec4cd9471e21" containerID="eba786db77e4785126a35b98eed013099cfce39a7e939fc837fb1037d741e55c" exitCode=0 Oct 14 07:19:22 crc kubenswrapper[4870]: I1014 07:19:22.449125 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-4dkbn" event={"ID":"25f5ffd0-3c55-46ee-ade3-ec4cd9471e21","Type":"ContainerDied","Data":"eba786db77e4785126a35b98eed013099cfce39a7e939fc837fb1037d741e55c"} Oct 14 07:19:22 crc kubenswrapper[4870]: I1014 07:19:22.452656 4870 generic.go:334] "Generic (PLEG): container finished" podID="835bb9d2-3b63-4fa5-bda6-9cc6a9a899a2" containerID="1b44e2314629f4fd5b7db5c31f15920edf5685a4b13b8bd037c33027e60486b4" exitCode=0 Oct 14 07:19:22 crc kubenswrapper[4870]: I1014 07:19:22.452781 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-nzvfg" event={"ID":"835bb9d2-3b63-4fa5-bda6-9cc6a9a899a2","Type":"ContainerDied","Data":"1b44e2314629f4fd5b7db5c31f15920edf5685a4b13b8bd037c33027e60486b4"} Oct 14 07:19:22 crc kubenswrapper[4870]: I1014 07:19:22.455270 4870 generic.go:334] "Generic (PLEG): container finished" podID="cd4cf561-b2b4-486b-b317-bd88b71badd9" containerID="130ba2c834779a188207288977bbddd0143d006f944bd94f346093f2993c5ff4" exitCode=0 Oct 14 07:19:22 crc kubenswrapper[4870]: I1014 07:19:22.455342 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-f6n86" event={"ID":"cd4cf561-b2b4-486b-b317-bd88b71badd9","Type":"ContainerDied","Data":"130ba2c834779a188207288977bbddd0143d006f944bd94f346093f2993c5ff4"} Oct 14 07:19:22 crc kubenswrapper[4870]: I1014 07:19:22.455556 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-f6n86" event={"ID":"cd4cf561-b2b4-486b-b317-bd88b71badd9","Type":"ContainerStarted","Data":"0fb25087f506075020a5ce946f1d3fe048514e21bc00cf9988ca478f45307151"} Oct 14 07:19:23 crc kubenswrapper[4870]: I1014 07:19:23.296823 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-549b48ff7f-vhsh8" Oct 14 07:19:23 crc kubenswrapper[4870]: I1014 07:19:23.380606 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7764bd4845-4zlrf"] Oct 14 07:19:23 crc kubenswrapper[4870]: I1014 07:19:23.381076 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7764bd4845-4zlrf" podUID="a67380d5-9cd0-4443-b45b-44663fddd756" containerName="dnsmasq-dns" containerID="cri-o://44771a13a098ac9d67781af2c4e40602aeb29022af9b243ae7996df7a250b70b" gracePeriod=10 Oct 14 07:19:23 crc kubenswrapper[4870]: I1014 07:19:23.835578 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-nzvfg" Oct 14 07:19:23 crc kubenswrapper[4870]: I1014 07:19:23.950951 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9l54v\" (UniqueName: \"kubernetes.io/projected/835bb9d2-3b63-4fa5-bda6-9cc6a9a899a2-kube-api-access-9l54v\") pod \"835bb9d2-3b63-4fa5-bda6-9cc6a9a899a2\" (UID: \"835bb9d2-3b63-4fa5-bda6-9cc6a9a899a2\") " Oct 14 07:19:23 crc kubenswrapper[4870]: I1014 07:19:23.958911 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/835bb9d2-3b63-4fa5-bda6-9cc6a9a899a2-kube-api-access-9l54v" (OuterVolumeSpecName: "kube-api-access-9l54v") pod "835bb9d2-3b63-4fa5-bda6-9cc6a9a899a2" (UID: "835bb9d2-3b63-4fa5-bda6-9cc6a9a899a2"). InnerVolumeSpecName "kube-api-access-9l54v". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:19:24 crc kubenswrapper[4870]: I1014 07:19:24.045527 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-4dkbn" Oct 14 07:19:24 crc kubenswrapper[4870]: I1014 07:19:24.053971 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9l54v\" (UniqueName: \"kubernetes.io/projected/835bb9d2-3b63-4fa5-bda6-9cc6a9a899a2-kube-api-access-9l54v\") on node \"crc\" DevicePath \"\"" Oct 14 07:19:24 crc kubenswrapper[4870]: I1014 07:19:24.055293 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7764bd4845-4zlrf" Oct 14 07:19:24 crc kubenswrapper[4870]: I1014 07:19:24.077400 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-f6n86" Oct 14 07:19:24 crc kubenswrapper[4870]: I1014 07:19:24.154973 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wd6bp\" (UniqueName: \"kubernetes.io/projected/cd4cf561-b2b4-486b-b317-bd88b71badd9-kube-api-access-wd6bp\") pod \"cd4cf561-b2b4-486b-b317-bd88b71badd9\" (UID: \"cd4cf561-b2b4-486b-b317-bd88b71badd9\") " Oct 14 07:19:24 crc kubenswrapper[4870]: I1014 07:19:24.155027 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kbrgv\" (UniqueName: \"kubernetes.io/projected/a67380d5-9cd0-4443-b45b-44663fddd756-kube-api-access-kbrgv\") pod \"a67380d5-9cd0-4443-b45b-44663fddd756\" (UID: \"a67380d5-9cd0-4443-b45b-44663fddd756\") " Oct 14 07:19:24 crc kubenswrapper[4870]: I1014 07:19:24.155048 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a67380d5-9cd0-4443-b45b-44663fddd756-ovsdbserver-nb\") pod \"a67380d5-9cd0-4443-b45b-44663fddd756\" (UID: \"a67380d5-9cd0-4443-b45b-44663fddd756\") " Oct 14 07:19:24 crc kubenswrapper[4870]: I1014 07:19:24.155121 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a67380d5-9cd0-4443-b45b-44663fddd756-dns-svc\") pod \"a67380d5-9cd0-4443-b45b-44663fddd756\" (UID: \"a67380d5-9cd0-4443-b45b-44663fddd756\") " Oct 14 07:19:24 crc kubenswrapper[4870]: I1014 07:19:24.155145 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a67380d5-9cd0-4443-b45b-44663fddd756-ovsdbserver-sb\") pod \"a67380d5-9cd0-4443-b45b-44663fddd756\" (UID: \"a67380d5-9cd0-4443-b45b-44663fddd756\") " Oct 14 07:19:24 crc kubenswrapper[4870]: I1014 07:19:24.155192 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qwkjm\" (UniqueName: \"kubernetes.io/projected/25f5ffd0-3c55-46ee-ade3-ec4cd9471e21-kube-api-access-qwkjm\") pod \"25f5ffd0-3c55-46ee-ade3-ec4cd9471e21\" (UID: \"25f5ffd0-3c55-46ee-ade3-ec4cd9471e21\") " Oct 14 07:19:24 crc kubenswrapper[4870]: I1014 07:19:24.155216 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a67380d5-9cd0-4443-b45b-44663fddd756-config\") pod \"a67380d5-9cd0-4443-b45b-44663fddd756\" (UID: \"a67380d5-9cd0-4443-b45b-44663fddd756\") " Oct 14 07:19:24 crc kubenswrapper[4870]: I1014 07:19:24.158624 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25f5ffd0-3c55-46ee-ade3-ec4cd9471e21-kube-api-access-qwkjm" (OuterVolumeSpecName: "kube-api-access-qwkjm") pod "25f5ffd0-3c55-46ee-ade3-ec4cd9471e21" (UID: "25f5ffd0-3c55-46ee-ade3-ec4cd9471e21"). InnerVolumeSpecName "kube-api-access-qwkjm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:19:24 crc kubenswrapper[4870]: I1014 07:19:24.160289 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd4cf561-b2b4-486b-b317-bd88b71badd9-kube-api-access-wd6bp" (OuterVolumeSpecName: "kube-api-access-wd6bp") pod "cd4cf561-b2b4-486b-b317-bd88b71badd9" (UID: "cd4cf561-b2b4-486b-b317-bd88b71badd9"). InnerVolumeSpecName "kube-api-access-wd6bp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:19:24 crc kubenswrapper[4870]: I1014 07:19:24.165866 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a67380d5-9cd0-4443-b45b-44663fddd756-kube-api-access-kbrgv" (OuterVolumeSpecName: "kube-api-access-kbrgv") pod "a67380d5-9cd0-4443-b45b-44663fddd756" (UID: "a67380d5-9cd0-4443-b45b-44663fddd756"). InnerVolumeSpecName "kube-api-access-kbrgv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:19:24 crc kubenswrapper[4870]: I1014 07:19:24.189003 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Oct 14 07:19:24 crc kubenswrapper[4870]: I1014 07:19:24.195565 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a67380d5-9cd0-4443-b45b-44663fddd756-config" (OuterVolumeSpecName: "config") pod "a67380d5-9cd0-4443-b45b-44663fddd756" (UID: "a67380d5-9cd0-4443-b45b-44663fddd756"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:19:24 crc kubenswrapper[4870]: I1014 07:19:24.203715 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a67380d5-9cd0-4443-b45b-44663fddd756-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "a67380d5-9cd0-4443-b45b-44663fddd756" (UID: "a67380d5-9cd0-4443-b45b-44663fddd756"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:19:24 crc kubenswrapper[4870]: I1014 07:19:24.207328 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a67380d5-9cd0-4443-b45b-44663fddd756-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "a67380d5-9cd0-4443-b45b-44663fddd756" (UID: "a67380d5-9cd0-4443-b45b-44663fddd756"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:19:24 crc kubenswrapper[4870]: I1014 07:19:24.233923 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a67380d5-9cd0-4443-b45b-44663fddd756-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "a67380d5-9cd0-4443-b45b-44663fddd756" (UID: "a67380d5-9cd0-4443-b45b-44663fddd756"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:19:24 crc kubenswrapper[4870]: I1014 07:19:24.257908 4870 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a67380d5-9cd0-4443-b45b-44663fddd756-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 14 07:19:24 crc kubenswrapper[4870]: I1014 07:19:24.257953 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qwkjm\" (UniqueName: \"kubernetes.io/projected/25f5ffd0-3c55-46ee-ade3-ec4cd9471e21-kube-api-access-qwkjm\") on node \"crc\" DevicePath \"\"" Oct 14 07:19:24 crc kubenswrapper[4870]: I1014 07:19:24.257966 4870 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a67380d5-9cd0-4443-b45b-44663fddd756-config\") on node \"crc\" DevicePath \"\"" Oct 14 07:19:24 crc kubenswrapper[4870]: I1014 07:19:24.257975 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wd6bp\" (UniqueName: \"kubernetes.io/projected/cd4cf561-b2b4-486b-b317-bd88b71badd9-kube-api-access-wd6bp\") on node \"crc\" DevicePath \"\"" Oct 14 07:19:24 crc kubenswrapper[4870]: I1014 07:19:24.257984 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kbrgv\" (UniqueName: \"kubernetes.io/projected/a67380d5-9cd0-4443-b45b-44663fddd756-kube-api-access-kbrgv\") on node \"crc\" DevicePath \"\"" Oct 14 07:19:24 crc kubenswrapper[4870]: I1014 07:19:24.257992 4870 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a67380d5-9cd0-4443-b45b-44663fddd756-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 14 07:19:24 crc kubenswrapper[4870]: I1014 07:19:24.258000 4870 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a67380d5-9cd0-4443-b45b-44663fddd756-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 14 07:19:24 crc kubenswrapper[4870]: I1014 07:19:24.473632 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-f6n86" event={"ID":"cd4cf561-b2b4-486b-b317-bd88b71badd9","Type":"ContainerDied","Data":"0fb25087f506075020a5ce946f1d3fe048514e21bc00cf9988ca478f45307151"} Oct 14 07:19:24 crc kubenswrapper[4870]: I1014 07:19:24.473668 4870 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0fb25087f506075020a5ce946f1d3fe048514e21bc00cf9988ca478f45307151" Oct 14 07:19:24 crc kubenswrapper[4870]: I1014 07:19:24.473671 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-f6n86" Oct 14 07:19:24 crc kubenswrapper[4870]: I1014 07:19:24.476005 4870 generic.go:334] "Generic (PLEG): container finished" podID="a67380d5-9cd0-4443-b45b-44663fddd756" containerID="44771a13a098ac9d67781af2c4e40602aeb29022af9b243ae7996df7a250b70b" exitCode=0 Oct 14 07:19:24 crc kubenswrapper[4870]: I1014 07:19:24.476097 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7764bd4845-4zlrf" event={"ID":"a67380d5-9cd0-4443-b45b-44663fddd756","Type":"ContainerDied","Data":"44771a13a098ac9d67781af2c4e40602aeb29022af9b243ae7996df7a250b70b"} Oct 14 07:19:24 crc kubenswrapper[4870]: I1014 07:19:24.476117 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7764bd4845-4zlrf" Oct 14 07:19:24 crc kubenswrapper[4870]: I1014 07:19:24.476154 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7764bd4845-4zlrf" event={"ID":"a67380d5-9cd0-4443-b45b-44663fddd756","Type":"ContainerDied","Data":"ec1a938e2190cea94baf85f78c82af874b1811a42007f42d6a014e6a398dcdae"} Oct 14 07:19:24 crc kubenswrapper[4870]: I1014 07:19:24.476184 4870 scope.go:117] "RemoveContainer" containerID="44771a13a098ac9d67781af2c4e40602aeb29022af9b243ae7996df7a250b70b" Oct 14 07:19:24 crc kubenswrapper[4870]: I1014 07:19:24.477989 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-4dkbn" event={"ID":"25f5ffd0-3c55-46ee-ade3-ec4cd9471e21","Type":"ContainerDied","Data":"fb73b2ee7af0c79ed7904e3e29c0d47808c2afac58abd694d97e9a16014fc1d0"} Oct 14 07:19:24 crc kubenswrapper[4870]: I1014 07:19:24.478038 4870 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fb73b2ee7af0c79ed7904e3e29c0d47808c2afac58abd694d97e9a16014fc1d0" Oct 14 07:19:24 crc kubenswrapper[4870]: I1014 07:19:24.478098 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-4dkbn" Oct 14 07:19:24 crc kubenswrapper[4870]: I1014 07:19:24.480731 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-nzvfg" event={"ID":"835bb9d2-3b63-4fa5-bda6-9cc6a9a899a2","Type":"ContainerDied","Data":"e5cd462957537fad952f36e02700fe1f4f2270bd80bd8f73defbfa6d8f38b673"} Oct 14 07:19:24 crc kubenswrapper[4870]: I1014 07:19:24.480763 4870 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e5cd462957537fad952f36e02700fe1f4f2270bd80bd8f73defbfa6d8f38b673" Oct 14 07:19:24 crc kubenswrapper[4870]: I1014 07:19:24.480825 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-nzvfg" Oct 14 07:19:24 crc kubenswrapper[4870]: I1014 07:19:24.500107 4870 scope.go:117] "RemoveContainer" containerID="d614d9d6e11a79690035da28f7f8cbdd7407e7f268ed5312d39fce5ec71890e3" Oct 14 07:19:24 crc kubenswrapper[4870]: I1014 07:19:24.590964 4870 scope.go:117] "RemoveContainer" containerID="44771a13a098ac9d67781af2c4e40602aeb29022af9b243ae7996df7a250b70b" Oct 14 07:19:24 crc kubenswrapper[4870]: E1014 07:19:24.592230 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"44771a13a098ac9d67781af2c4e40602aeb29022af9b243ae7996df7a250b70b\": container with ID starting with 44771a13a098ac9d67781af2c4e40602aeb29022af9b243ae7996df7a250b70b not found: ID does not exist" containerID="44771a13a098ac9d67781af2c4e40602aeb29022af9b243ae7996df7a250b70b" Oct 14 07:19:24 crc kubenswrapper[4870]: I1014 07:19:24.592278 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"44771a13a098ac9d67781af2c4e40602aeb29022af9b243ae7996df7a250b70b"} err="failed to get container status \"44771a13a098ac9d67781af2c4e40602aeb29022af9b243ae7996df7a250b70b\": rpc error: code = NotFound desc = could not find container \"44771a13a098ac9d67781af2c4e40602aeb29022af9b243ae7996df7a250b70b\": container with ID starting with 44771a13a098ac9d67781af2c4e40602aeb29022af9b243ae7996df7a250b70b not found: ID does not exist" Oct 14 07:19:24 crc kubenswrapper[4870]: I1014 07:19:24.592309 4870 scope.go:117] "RemoveContainer" containerID="d614d9d6e11a79690035da28f7f8cbdd7407e7f268ed5312d39fce5ec71890e3" Oct 14 07:19:24 crc kubenswrapper[4870]: E1014 07:19:24.592790 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d614d9d6e11a79690035da28f7f8cbdd7407e7f268ed5312d39fce5ec71890e3\": container with ID starting with d614d9d6e11a79690035da28f7f8cbdd7407e7f268ed5312d39fce5ec71890e3 not found: ID does not exist" containerID="d614d9d6e11a79690035da28f7f8cbdd7407e7f268ed5312d39fce5ec71890e3" Oct 14 07:19:24 crc kubenswrapper[4870]: I1014 07:19:24.592818 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d614d9d6e11a79690035da28f7f8cbdd7407e7f268ed5312d39fce5ec71890e3"} err="failed to get container status \"d614d9d6e11a79690035da28f7f8cbdd7407e7f268ed5312d39fce5ec71890e3\": rpc error: code = NotFound desc = could not find container \"d614d9d6e11a79690035da28f7f8cbdd7407e7f268ed5312d39fce5ec71890e3\": container with ID starting with d614d9d6e11a79690035da28f7f8cbdd7407e7f268ed5312d39fce5ec71890e3 not found: ID does not exist" Oct 14 07:19:24 crc kubenswrapper[4870]: I1014 07:19:24.593587 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7764bd4845-4zlrf"] Oct 14 07:19:24 crc kubenswrapper[4870]: I1014 07:19:24.600845 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7764bd4845-4zlrf"] Oct 14 07:19:25 crc kubenswrapper[4870]: I1014 07:19:25.047197 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a67380d5-9cd0-4443-b45b-44663fddd756" path="/var/lib/kubelet/pods/a67380d5-9cd0-4443-b45b-44663fddd756/volumes" Oct 14 07:19:25 crc kubenswrapper[4870]: I1014 07:19:25.490135 4870 generic.go:334] "Generic (PLEG): container finished" podID="73ffa9f3-6390-4ed0-b0ba-5a3e2a44ae1f" containerID="365d48ec2f5364c81df1e76cb37b8b718e730cf770131932caead0d2a5c0f467" exitCode=0 Oct 14 07:19:25 crc kubenswrapper[4870]: I1014 07:19:25.490236 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-lqpbp" event={"ID":"73ffa9f3-6390-4ed0-b0ba-5a3e2a44ae1f","Type":"ContainerDied","Data":"365d48ec2f5364c81df1e76cb37b8b718e730cf770131932caead0d2a5c0f467"} Oct 14 07:19:26 crc kubenswrapper[4870]: I1014 07:19:26.947882 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-lqpbp" Oct 14 07:19:27 crc kubenswrapper[4870]: I1014 07:19:27.123571 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/73ffa9f3-6390-4ed0-b0ba-5a3e2a44ae1f-etc-swift\") pod \"73ffa9f3-6390-4ed0-b0ba-5a3e2a44ae1f\" (UID: \"73ffa9f3-6390-4ed0-b0ba-5a3e2a44ae1f\") " Oct 14 07:19:27 crc kubenswrapper[4870]: I1014 07:19:27.123632 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/73ffa9f3-6390-4ed0-b0ba-5a3e2a44ae1f-combined-ca-bundle\") pod \"73ffa9f3-6390-4ed0-b0ba-5a3e2a44ae1f\" (UID: \"73ffa9f3-6390-4ed0-b0ba-5a3e2a44ae1f\") " Oct 14 07:19:27 crc kubenswrapper[4870]: I1014 07:19:27.123673 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/73ffa9f3-6390-4ed0-b0ba-5a3e2a44ae1f-scripts\") pod \"73ffa9f3-6390-4ed0-b0ba-5a3e2a44ae1f\" (UID: \"73ffa9f3-6390-4ed0-b0ba-5a3e2a44ae1f\") " Oct 14 07:19:27 crc kubenswrapper[4870]: I1014 07:19:27.123716 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/73ffa9f3-6390-4ed0-b0ba-5a3e2a44ae1f-swiftconf\") pod \"73ffa9f3-6390-4ed0-b0ba-5a3e2a44ae1f\" (UID: \"73ffa9f3-6390-4ed0-b0ba-5a3e2a44ae1f\") " Oct 14 07:19:27 crc kubenswrapper[4870]: I1014 07:19:27.123826 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/73ffa9f3-6390-4ed0-b0ba-5a3e2a44ae1f-ring-data-devices\") pod \"73ffa9f3-6390-4ed0-b0ba-5a3e2a44ae1f\" (UID: \"73ffa9f3-6390-4ed0-b0ba-5a3e2a44ae1f\") " Oct 14 07:19:27 crc kubenswrapper[4870]: I1014 07:19:27.123934 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gwhpv\" (UniqueName: \"kubernetes.io/projected/73ffa9f3-6390-4ed0-b0ba-5a3e2a44ae1f-kube-api-access-gwhpv\") pod \"73ffa9f3-6390-4ed0-b0ba-5a3e2a44ae1f\" (UID: \"73ffa9f3-6390-4ed0-b0ba-5a3e2a44ae1f\") " Oct 14 07:19:27 crc kubenswrapper[4870]: I1014 07:19:27.123996 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/73ffa9f3-6390-4ed0-b0ba-5a3e2a44ae1f-dispersionconf\") pod \"73ffa9f3-6390-4ed0-b0ba-5a3e2a44ae1f\" (UID: \"73ffa9f3-6390-4ed0-b0ba-5a3e2a44ae1f\") " Oct 14 07:19:27 crc kubenswrapper[4870]: I1014 07:19:27.126409 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/73ffa9f3-6390-4ed0-b0ba-5a3e2a44ae1f-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "73ffa9f3-6390-4ed0-b0ba-5a3e2a44ae1f" (UID: "73ffa9f3-6390-4ed0-b0ba-5a3e2a44ae1f"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:19:27 crc kubenswrapper[4870]: I1014 07:19:27.127036 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/73ffa9f3-6390-4ed0-b0ba-5a3e2a44ae1f-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "73ffa9f3-6390-4ed0-b0ba-5a3e2a44ae1f" (UID: "73ffa9f3-6390-4ed0-b0ba-5a3e2a44ae1f"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:19:27 crc kubenswrapper[4870]: I1014 07:19:27.133768 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/73ffa9f3-6390-4ed0-b0ba-5a3e2a44ae1f-kube-api-access-gwhpv" (OuterVolumeSpecName: "kube-api-access-gwhpv") pod "73ffa9f3-6390-4ed0-b0ba-5a3e2a44ae1f" (UID: "73ffa9f3-6390-4ed0-b0ba-5a3e2a44ae1f"). InnerVolumeSpecName "kube-api-access-gwhpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:19:27 crc kubenswrapper[4870]: I1014 07:19:27.145347 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/73ffa9f3-6390-4ed0-b0ba-5a3e2a44ae1f-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "73ffa9f3-6390-4ed0-b0ba-5a3e2a44ae1f" (UID: "73ffa9f3-6390-4ed0-b0ba-5a3e2a44ae1f"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:19:27 crc kubenswrapper[4870]: I1014 07:19:27.153026 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/73ffa9f3-6390-4ed0-b0ba-5a3e2a44ae1f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "73ffa9f3-6390-4ed0-b0ba-5a3e2a44ae1f" (UID: "73ffa9f3-6390-4ed0-b0ba-5a3e2a44ae1f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:19:27 crc kubenswrapper[4870]: I1014 07:19:27.177544 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/73ffa9f3-6390-4ed0-b0ba-5a3e2a44ae1f-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "73ffa9f3-6390-4ed0-b0ba-5a3e2a44ae1f" (UID: "73ffa9f3-6390-4ed0-b0ba-5a3e2a44ae1f"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:19:27 crc kubenswrapper[4870]: I1014 07:19:27.184096 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/73ffa9f3-6390-4ed0-b0ba-5a3e2a44ae1f-scripts" (OuterVolumeSpecName: "scripts") pod "73ffa9f3-6390-4ed0-b0ba-5a3e2a44ae1f" (UID: "73ffa9f3-6390-4ed0-b0ba-5a3e2a44ae1f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:19:27 crc kubenswrapper[4870]: I1014 07:19:27.227558 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gwhpv\" (UniqueName: \"kubernetes.io/projected/73ffa9f3-6390-4ed0-b0ba-5a3e2a44ae1f-kube-api-access-gwhpv\") on node \"crc\" DevicePath \"\"" Oct 14 07:19:27 crc kubenswrapper[4870]: I1014 07:19:27.227827 4870 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/73ffa9f3-6390-4ed0-b0ba-5a3e2a44ae1f-dispersionconf\") on node \"crc\" DevicePath \"\"" Oct 14 07:19:27 crc kubenswrapper[4870]: I1014 07:19:27.228107 4870 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/73ffa9f3-6390-4ed0-b0ba-5a3e2a44ae1f-etc-swift\") on node \"crc\" DevicePath \"\"" Oct 14 07:19:27 crc kubenswrapper[4870]: I1014 07:19:27.228279 4870 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/73ffa9f3-6390-4ed0-b0ba-5a3e2a44ae1f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 07:19:27 crc kubenswrapper[4870]: I1014 07:19:27.228517 4870 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/73ffa9f3-6390-4ed0-b0ba-5a3e2a44ae1f-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 07:19:27 crc kubenswrapper[4870]: I1014 07:19:27.228673 4870 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/73ffa9f3-6390-4ed0-b0ba-5a3e2a44ae1f-swiftconf\") on node \"crc\" DevicePath \"\"" Oct 14 07:19:27 crc kubenswrapper[4870]: I1014 07:19:27.228811 4870 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/73ffa9f3-6390-4ed0-b0ba-5a3e2a44ae1f-ring-data-devices\") on node \"crc\" DevicePath \"\"" Oct 14 07:19:27 crc kubenswrapper[4870]: I1014 07:19:27.519878 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-lqpbp" event={"ID":"73ffa9f3-6390-4ed0-b0ba-5a3e2a44ae1f","Type":"ContainerDied","Data":"5406ff5ba7fe072aac7d4a55694936460e4bf0eda7d0f06b2e50df806b27b7cb"} Oct 14 07:19:27 crc kubenswrapper[4870]: I1014 07:19:27.519938 4870 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5406ff5ba7fe072aac7d4a55694936460e4bf0eda7d0f06b2e50df806b27b7cb" Oct 14 07:19:27 crc kubenswrapper[4870]: I1014 07:19:27.520019 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-lqpbp" Oct 14 07:19:29 crc kubenswrapper[4870]: I1014 07:19:29.540162 4870 generic.go:334] "Generic (PLEG): container finished" podID="5212ba20-83c3-41ff-a3d9-5a315f009c1b" containerID="80776853a799d9822a89d9d86e1010ddb56ba32acd904f622c45923cef6744ac" exitCode=0 Oct 14 07:19:29 crc kubenswrapper[4870]: I1014 07:19:29.540215 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"5212ba20-83c3-41ff-a3d9-5a315f009c1b","Type":"ContainerDied","Data":"80776853a799d9822a89d9d86e1010ddb56ba32acd904f622c45923cef6744ac"} Oct 14 07:19:29 crc kubenswrapper[4870]: I1014 07:19:29.980740 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/2bd1f41f-3f1a-4ca7-8789-429104ce2120-etc-swift\") pod \"swift-storage-0\" (UID: \"2bd1f41f-3f1a-4ca7-8789-429104ce2120\") " pod="openstack/swift-storage-0" Oct 14 07:19:29 crc kubenswrapper[4870]: I1014 07:19:29.987802 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/2bd1f41f-3f1a-4ca7-8789-429104ce2120-etc-swift\") pod \"swift-storage-0\" (UID: \"2bd1f41f-3f1a-4ca7-8789-429104ce2120\") " pod="openstack/swift-storage-0" Oct 14 07:19:30 crc kubenswrapper[4870]: I1014 07:19:30.251924 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Oct 14 07:19:30 crc kubenswrapper[4870]: I1014 07:19:30.364073 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-220b-account-create-gthws"] Oct 14 07:19:30 crc kubenswrapper[4870]: E1014 07:19:30.364519 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a67380d5-9cd0-4443-b45b-44663fddd756" containerName="init" Oct 14 07:19:30 crc kubenswrapper[4870]: I1014 07:19:30.364548 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="a67380d5-9cd0-4443-b45b-44663fddd756" containerName="init" Oct 14 07:19:30 crc kubenswrapper[4870]: E1014 07:19:30.364569 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="25f5ffd0-3c55-46ee-ade3-ec4cd9471e21" containerName="mariadb-database-create" Oct 14 07:19:30 crc kubenswrapper[4870]: I1014 07:19:30.364577 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="25f5ffd0-3c55-46ee-ade3-ec4cd9471e21" containerName="mariadb-database-create" Oct 14 07:19:30 crc kubenswrapper[4870]: E1014 07:19:30.364600 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="835bb9d2-3b63-4fa5-bda6-9cc6a9a899a2" containerName="mariadb-database-create" Oct 14 07:19:30 crc kubenswrapper[4870]: I1014 07:19:30.364608 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="835bb9d2-3b63-4fa5-bda6-9cc6a9a899a2" containerName="mariadb-database-create" Oct 14 07:19:30 crc kubenswrapper[4870]: E1014 07:19:30.364619 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73ffa9f3-6390-4ed0-b0ba-5a3e2a44ae1f" containerName="swift-ring-rebalance" Oct 14 07:19:30 crc kubenswrapper[4870]: I1014 07:19:30.364627 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="73ffa9f3-6390-4ed0-b0ba-5a3e2a44ae1f" containerName="swift-ring-rebalance" Oct 14 07:19:30 crc kubenswrapper[4870]: E1014 07:19:30.364648 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a67380d5-9cd0-4443-b45b-44663fddd756" containerName="dnsmasq-dns" Oct 14 07:19:30 crc kubenswrapper[4870]: I1014 07:19:30.364658 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="a67380d5-9cd0-4443-b45b-44663fddd756" containerName="dnsmasq-dns" Oct 14 07:19:30 crc kubenswrapper[4870]: E1014 07:19:30.364683 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd4cf561-b2b4-486b-b317-bd88b71badd9" containerName="mariadb-database-create" Oct 14 07:19:30 crc kubenswrapper[4870]: I1014 07:19:30.364694 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd4cf561-b2b4-486b-b317-bd88b71badd9" containerName="mariadb-database-create" Oct 14 07:19:30 crc kubenswrapper[4870]: I1014 07:19:30.365015 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="835bb9d2-3b63-4fa5-bda6-9cc6a9a899a2" containerName="mariadb-database-create" Oct 14 07:19:30 crc kubenswrapper[4870]: I1014 07:19:30.365044 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="cd4cf561-b2b4-486b-b317-bd88b71badd9" containerName="mariadb-database-create" Oct 14 07:19:30 crc kubenswrapper[4870]: I1014 07:19:30.365058 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="a67380d5-9cd0-4443-b45b-44663fddd756" containerName="dnsmasq-dns" Oct 14 07:19:30 crc kubenswrapper[4870]: I1014 07:19:30.365074 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="25f5ffd0-3c55-46ee-ade3-ec4cd9471e21" containerName="mariadb-database-create" Oct 14 07:19:30 crc kubenswrapper[4870]: I1014 07:19:30.365092 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="73ffa9f3-6390-4ed0-b0ba-5a3e2a44ae1f" containerName="swift-ring-rebalance" Oct 14 07:19:30 crc kubenswrapper[4870]: I1014 07:19:30.365804 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-220b-account-create-gthws" Oct 14 07:19:30 crc kubenswrapper[4870]: I1014 07:19:30.367699 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Oct 14 07:19:30 crc kubenswrapper[4870]: I1014 07:19:30.377089 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-220b-account-create-gthws"] Oct 14 07:19:30 crc kubenswrapper[4870]: I1014 07:19:30.489489 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hdc2h\" (UniqueName: \"kubernetes.io/projected/d318a57f-fb03-4112-b027-b8eaf943b00d-kube-api-access-hdc2h\") pod \"keystone-220b-account-create-gthws\" (UID: \"d318a57f-fb03-4112-b027-b8eaf943b00d\") " pod="openstack/keystone-220b-account-create-gthws" Oct 14 07:19:30 crc kubenswrapper[4870]: I1014 07:19:30.551208 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"5212ba20-83c3-41ff-a3d9-5a315f009c1b","Type":"ContainerStarted","Data":"eeb00ec87974e56f742f7ae182732d1dfbbc2c6523987de46327c67f0a729b18"} Oct 14 07:19:30 crc kubenswrapper[4870]: I1014 07:19:30.551429 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Oct 14 07:19:30 crc kubenswrapper[4870]: I1014 07:19:30.554338 4870 generic.go:334] "Generic (PLEG): container finished" podID="da8a9731-f917-462f-9932-b37b6abb9a64" containerID="fdc546ad7b75936b8391d1c25cb200fd0e53c0d3629443c510ae84c5a357cc6e" exitCode=0 Oct 14 07:19:30 crc kubenswrapper[4870]: I1014 07:19:30.554376 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"da8a9731-f917-462f-9932-b37b6abb9a64","Type":"ContainerDied","Data":"fdc546ad7b75936b8391d1c25cb200fd0e53c0d3629443c510ae84c5a357cc6e"} Oct 14 07:19:30 crc kubenswrapper[4870]: I1014 07:19:30.586535 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=54.586514389 podStartE2EDuration="54.586514389s" podCreationTimestamp="2025-10-14 07:18:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:19:30.573647161 +0000 UTC m=+1106.271007542" watchObservedRunningTime="2025-10-14 07:19:30.586514389 +0000 UTC m=+1106.283874760" Oct 14 07:19:30 crc kubenswrapper[4870]: I1014 07:19:30.590967 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hdc2h\" (UniqueName: \"kubernetes.io/projected/d318a57f-fb03-4112-b027-b8eaf943b00d-kube-api-access-hdc2h\") pod \"keystone-220b-account-create-gthws\" (UID: \"d318a57f-fb03-4112-b027-b8eaf943b00d\") " pod="openstack/keystone-220b-account-create-gthws" Oct 14 07:19:30 crc kubenswrapper[4870]: I1014 07:19:30.616146 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hdc2h\" (UniqueName: \"kubernetes.io/projected/d318a57f-fb03-4112-b027-b8eaf943b00d-kube-api-access-hdc2h\") pod \"keystone-220b-account-create-gthws\" (UID: \"d318a57f-fb03-4112-b027-b8eaf943b00d\") " pod="openstack/keystone-220b-account-create-gthws" Oct 14 07:19:30 crc kubenswrapper[4870]: I1014 07:19:30.655955 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-1a80-account-create-rcm8s"] Oct 14 07:19:30 crc kubenswrapper[4870]: I1014 07:19:30.656954 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-1a80-account-create-rcm8s" Oct 14 07:19:30 crc kubenswrapper[4870]: I1014 07:19:30.659604 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Oct 14 07:19:30 crc kubenswrapper[4870]: I1014 07:19:30.666232 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-1a80-account-create-rcm8s"] Oct 14 07:19:30 crc kubenswrapper[4870]: I1014 07:19:30.690525 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-220b-account-create-gthws" Oct 14 07:19:30 crc kubenswrapper[4870]: I1014 07:19:30.714190 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Oct 14 07:19:30 crc kubenswrapper[4870]: W1014 07:19:30.720483 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2bd1f41f_3f1a_4ca7_8789_429104ce2120.slice/crio-b3bced4776c0abee54d19d97669ffda4f01fe4106bd9452fcfe3165b4e59267e WatchSource:0}: Error finding container b3bced4776c0abee54d19d97669ffda4f01fe4106bd9452fcfe3165b4e59267e: Status 404 returned error can't find the container with id b3bced4776c0abee54d19d97669ffda4f01fe4106bd9452fcfe3165b4e59267e Oct 14 07:19:30 crc kubenswrapper[4870]: I1014 07:19:30.793893 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dl8tt\" (UniqueName: \"kubernetes.io/projected/3c2004c3-445d-4495-b24d-862f402ff5a6-kube-api-access-dl8tt\") pod \"placement-1a80-account-create-rcm8s\" (UID: \"3c2004c3-445d-4495-b24d-862f402ff5a6\") " pod="openstack/placement-1a80-account-create-rcm8s" Oct 14 07:19:30 crc kubenswrapper[4870]: I1014 07:19:30.864376 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-1aaa-account-create-6t7t7"] Oct 14 07:19:30 crc kubenswrapper[4870]: I1014 07:19:30.865340 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-1aaa-account-create-6t7t7" Oct 14 07:19:30 crc kubenswrapper[4870]: I1014 07:19:30.868013 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Oct 14 07:19:30 crc kubenswrapper[4870]: I1014 07:19:30.884282 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-1aaa-account-create-6t7t7"] Oct 14 07:19:30 crc kubenswrapper[4870]: I1014 07:19:30.907271 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dl8tt\" (UniqueName: \"kubernetes.io/projected/3c2004c3-445d-4495-b24d-862f402ff5a6-kube-api-access-dl8tt\") pod \"placement-1a80-account-create-rcm8s\" (UID: \"3c2004c3-445d-4495-b24d-862f402ff5a6\") " pod="openstack/placement-1a80-account-create-rcm8s" Oct 14 07:19:30 crc kubenswrapper[4870]: I1014 07:19:30.929368 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dl8tt\" (UniqueName: \"kubernetes.io/projected/3c2004c3-445d-4495-b24d-862f402ff5a6-kube-api-access-dl8tt\") pod \"placement-1a80-account-create-rcm8s\" (UID: \"3c2004c3-445d-4495-b24d-862f402ff5a6\") " pod="openstack/placement-1a80-account-create-rcm8s" Oct 14 07:19:31 crc kubenswrapper[4870]: I1014 07:19:31.009248 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7p5f7\" (UniqueName: \"kubernetes.io/projected/0c16dc18-0486-420e-86c8-ed2374a9a705-kube-api-access-7p5f7\") pod \"glance-1aaa-account-create-6t7t7\" (UID: \"0c16dc18-0486-420e-86c8-ed2374a9a705\") " pod="openstack/glance-1aaa-account-create-6t7t7" Oct 14 07:19:31 crc kubenswrapper[4870]: I1014 07:19:31.065819 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-1a80-account-create-rcm8s" Oct 14 07:19:31 crc kubenswrapper[4870]: I1014 07:19:31.111102 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7p5f7\" (UniqueName: \"kubernetes.io/projected/0c16dc18-0486-420e-86c8-ed2374a9a705-kube-api-access-7p5f7\") pod \"glance-1aaa-account-create-6t7t7\" (UID: \"0c16dc18-0486-420e-86c8-ed2374a9a705\") " pod="openstack/glance-1aaa-account-create-6t7t7" Oct 14 07:19:31 crc kubenswrapper[4870]: I1014 07:19:31.128841 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7p5f7\" (UniqueName: \"kubernetes.io/projected/0c16dc18-0486-420e-86c8-ed2374a9a705-kube-api-access-7p5f7\") pod \"glance-1aaa-account-create-6t7t7\" (UID: \"0c16dc18-0486-420e-86c8-ed2374a9a705\") " pod="openstack/glance-1aaa-account-create-6t7t7" Oct 14 07:19:31 crc kubenswrapper[4870]: I1014 07:19:31.178853 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-220b-account-create-gthws"] Oct 14 07:19:31 crc kubenswrapper[4870]: I1014 07:19:31.185489 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-1aaa-account-create-6t7t7" Oct 14 07:19:31 crc kubenswrapper[4870]: I1014 07:19:31.565003 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"da8a9731-f917-462f-9932-b37b6abb9a64","Type":"ContainerStarted","Data":"bc4b5271d735615160d1289dff31839e1ba9c4495e333ee1f50cab05851fd959"} Oct 14 07:19:31 crc kubenswrapper[4870]: I1014 07:19:31.566535 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Oct 14 07:19:31 crc kubenswrapper[4870]: I1014 07:19:31.569237 4870 generic.go:334] "Generic (PLEG): container finished" podID="d318a57f-fb03-4112-b027-b8eaf943b00d" containerID="f712d6e12449176eab14639b7624cd8368d3985bad7d851e1a3f05c35886b81a" exitCode=0 Oct 14 07:19:31 crc kubenswrapper[4870]: I1014 07:19:31.569293 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-220b-account-create-gthws" event={"ID":"d318a57f-fb03-4112-b027-b8eaf943b00d","Type":"ContainerDied","Data":"f712d6e12449176eab14639b7624cd8368d3985bad7d851e1a3f05c35886b81a"} Oct 14 07:19:31 crc kubenswrapper[4870]: I1014 07:19:31.569308 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-220b-account-create-gthws" event={"ID":"d318a57f-fb03-4112-b027-b8eaf943b00d","Type":"ContainerStarted","Data":"90d44bdea59708190fe170257835f60c4a7f633b4b6f1d4e3775eaec3a93bcb4"} Oct 14 07:19:31 crc kubenswrapper[4870]: I1014 07:19:31.571248 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"2bd1f41f-3f1a-4ca7-8789-429104ce2120","Type":"ContainerStarted","Data":"b3bced4776c0abee54d19d97669ffda4f01fe4106bd9452fcfe3165b4e59267e"} Oct 14 07:19:31 crc kubenswrapper[4870]: I1014 07:19:31.594438 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=44.991972141 podStartE2EDuration="55.59441965s" podCreationTimestamp="2025-10-14 07:18:36 +0000 UTC" firstStartedPulling="2025-10-14 07:18:43.069662894 +0000 UTC m=+1058.767023275" lastFinishedPulling="2025-10-14 07:18:53.672110413 +0000 UTC m=+1069.369470784" observedRunningTime="2025-10-14 07:19:31.585286864 +0000 UTC m=+1107.282647235" watchObservedRunningTime="2025-10-14 07:19:31.59441965 +0000 UTC m=+1107.291780021" Oct 14 07:19:31 crc kubenswrapper[4870]: I1014 07:19:31.633791 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-1a80-account-create-rcm8s"] Oct 14 07:19:31 crc kubenswrapper[4870]: W1014 07:19:31.637517 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3c2004c3_445d_4495_b24d_862f402ff5a6.slice/crio-d99638d4ce866f2c194bd88c41448b5f4b3afac6075185b1f25720d324e564d5 WatchSource:0}: Error finding container d99638d4ce866f2c194bd88c41448b5f4b3afac6075185b1f25720d324e564d5: Status 404 returned error can't find the container with id d99638d4ce866f2c194bd88c41448b5f4b3afac6075185b1f25720d324e564d5 Oct 14 07:19:31 crc kubenswrapper[4870]: I1014 07:19:31.697134 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-1aaa-account-create-6t7t7"] Oct 14 07:19:31 crc kubenswrapper[4870]: W1014 07:19:31.883815 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0c16dc18_0486_420e_86c8_ed2374a9a705.slice/crio-5a9500467dfce86242c725ed352d82ee738550e7bc32a02518f77839d5475f8a WatchSource:0}: Error finding container 5a9500467dfce86242c725ed352d82ee738550e7bc32a02518f77839d5475f8a: Status 404 returned error can't find the container with id 5a9500467dfce86242c725ed352d82ee738550e7bc32a02518f77839d5475f8a Oct 14 07:19:32 crc kubenswrapper[4870]: E1014 07:19:32.260170 4870 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3c2004c3_445d_4495_b24d_862f402ff5a6.slice/crio-conmon-c5e07d714b9d6ff354b5240abbc6e77cc9c929d1f020b6844549c47e3eded85e.scope\": RecentStats: unable to find data in memory cache]" Oct 14 07:19:32 crc kubenswrapper[4870]: I1014 07:19:32.581044 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"2bd1f41f-3f1a-4ca7-8789-429104ce2120","Type":"ContainerStarted","Data":"edba2f1cac0476dc717b946f00e8879272b5577916eb07fb08304de266852951"} Oct 14 07:19:32 crc kubenswrapper[4870]: I1014 07:19:32.582640 4870 generic.go:334] "Generic (PLEG): container finished" podID="3c2004c3-445d-4495-b24d-862f402ff5a6" containerID="c5e07d714b9d6ff354b5240abbc6e77cc9c929d1f020b6844549c47e3eded85e" exitCode=0 Oct 14 07:19:32 crc kubenswrapper[4870]: I1014 07:19:32.582701 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-1a80-account-create-rcm8s" event={"ID":"3c2004c3-445d-4495-b24d-862f402ff5a6","Type":"ContainerDied","Data":"c5e07d714b9d6ff354b5240abbc6e77cc9c929d1f020b6844549c47e3eded85e"} Oct 14 07:19:32 crc kubenswrapper[4870]: I1014 07:19:32.582726 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-1a80-account-create-rcm8s" event={"ID":"3c2004c3-445d-4495-b24d-862f402ff5a6","Type":"ContainerStarted","Data":"d99638d4ce866f2c194bd88c41448b5f4b3afac6075185b1f25720d324e564d5"} Oct 14 07:19:32 crc kubenswrapper[4870]: I1014 07:19:32.585522 4870 generic.go:334] "Generic (PLEG): container finished" podID="0c16dc18-0486-420e-86c8-ed2374a9a705" containerID="cea067cfa733de231a12525a23273b289acdfd787a45cbbf5c5d50bd4fd9998e" exitCode=0 Oct 14 07:19:32 crc kubenswrapper[4870]: I1014 07:19:32.585635 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-1aaa-account-create-6t7t7" event={"ID":"0c16dc18-0486-420e-86c8-ed2374a9a705","Type":"ContainerDied","Data":"cea067cfa733de231a12525a23273b289acdfd787a45cbbf5c5d50bd4fd9998e"} Oct 14 07:19:32 crc kubenswrapper[4870]: I1014 07:19:32.585734 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-1aaa-account-create-6t7t7" event={"ID":"0c16dc18-0486-420e-86c8-ed2374a9a705","Type":"ContainerStarted","Data":"5a9500467dfce86242c725ed352d82ee738550e7bc32a02518f77839d5475f8a"} Oct 14 07:19:32 crc kubenswrapper[4870]: I1014 07:19:32.990605 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-220b-account-create-gthws" Oct 14 07:19:33 crc kubenswrapper[4870]: I1014 07:19:33.045549 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hdc2h\" (UniqueName: \"kubernetes.io/projected/d318a57f-fb03-4112-b027-b8eaf943b00d-kube-api-access-hdc2h\") pod \"d318a57f-fb03-4112-b027-b8eaf943b00d\" (UID: \"d318a57f-fb03-4112-b027-b8eaf943b00d\") " Oct 14 07:19:33 crc kubenswrapper[4870]: I1014 07:19:33.053948 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d318a57f-fb03-4112-b027-b8eaf943b00d-kube-api-access-hdc2h" (OuterVolumeSpecName: "kube-api-access-hdc2h") pod "d318a57f-fb03-4112-b027-b8eaf943b00d" (UID: "d318a57f-fb03-4112-b027-b8eaf943b00d"). InnerVolumeSpecName "kube-api-access-hdc2h". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:19:33 crc kubenswrapper[4870]: I1014 07:19:33.146834 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hdc2h\" (UniqueName: \"kubernetes.io/projected/d318a57f-fb03-4112-b027-b8eaf943b00d-kube-api-access-hdc2h\") on node \"crc\" DevicePath \"\"" Oct 14 07:19:33 crc kubenswrapper[4870]: I1014 07:19:33.599236 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-220b-account-create-gthws" Oct 14 07:19:33 crc kubenswrapper[4870]: I1014 07:19:33.599668 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-220b-account-create-gthws" event={"ID":"d318a57f-fb03-4112-b027-b8eaf943b00d","Type":"ContainerDied","Data":"90d44bdea59708190fe170257835f60c4a7f633b4b6f1d4e3775eaec3a93bcb4"} Oct 14 07:19:33 crc kubenswrapper[4870]: I1014 07:19:33.599730 4870 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="90d44bdea59708190fe170257835f60c4a7f633b4b6f1d4e3775eaec3a93bcb4" Oct 14 07:19:34 crc kubenswrapper[4870]: I1014 07:19:34.023820 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-1aaa-account-create-6t7t7" Oct 14 07:19:34 crc kubenswrapper[4870]: I1014 07:19:34.039283 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-1a80-account-create-rcm8s" Oct 14 07:19:34 crc kubenswrapper[4870]: I1014 07:19:34.063639 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dl8tt\" (UniqueName: \"kubernetes.io/projected/3c2004c3-445d-4495-b24d-862f402ff5a6-kube-api-access-dl8tt\") pod \"3c2004c3-445d-4495-b24d-862f402ff5a6\" (UID: \"3c2004c3-445d-4495-b24d-862f402ff5a6\") " Oct 14 07:19:34 crc kubenswrapper[4870]: I1014 07:19:34.063776 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7p5f7\" (UniqueName: \"kubernetes.io/projected/0c16dc18-0486-420e-86c8-ed2374a9a705-kube-api-access-7p5f7\") pod \"0c16dc18-0486-420e-86c8-ed2374a9a705\" (UID: \"0c16dc18-0486-420e-86c8-ed2374a9a705\") " Oct 14 07:19:34 crc kubenswrapper[4870]: I1014 07:19:34.071698 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0c16dc18-0486-420e-86c8-ed2374a9a705-kube-api-access-7p5f7" (OuterVolumeSpecName: "kube-api-access-7p5f7") pod "0c16dc18-0486-420e-86c8-ed2374a9a705" (UID: "0c16dc18-0486-420e-86c8-ed2374a9a705"). InnerVolumeSpecName "kube-api-access-7p5f7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:19:34 crc kubenswrapper[4870]: I1014 07:19:34.072294 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3c2004c3-445d-4495-b24d-862f402ff5a6-kube-api-access-dl8tt" (OuterVolumeSpecName: "kube-api-access-dl8tt") pod "3c2004c3-445d-4495-b24d-862f402ff5a6" (UID: "3c2004c3-445d-4495-b24d-862f402ff5a6"). InnerVolumeSpecName "kube-api-access-dl8tt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:19:34 crc kubenswrapper[4870]: I1014 07:19:34.167457 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dl8tt\" (UniqueName: \"kubernetes.io/projected/3c2004c3-445d-4495-b24d-862f402ff5a6-kube-api-access-dl8tt\") on node \"crc\" DevicePath \"\"" Oct 14 07:19:34 crc kubenswrapper[4870]: I1014 07:19:34.167708 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7p5f7\" (UniqueName: \"kubernetes.io/projected/0c16dc18-0486-420e-86c8-ed2374a9a705-kube-api-access-7p5f7\") on node \"crc\" DevicePath \"\"" Oct 14 07:19:34 crc kubenswrapper[4870]: I1014 07:19:34.609607 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"2bd1f41f-3f1a-4ca7-8789-429104ce2120","Type":"ContainerStarted","Data":"e3fe8e57380100f6a293cf390d246d76bd9db2c91dc07edc65f2b23cfa5d4871"} Oct 14 07:19:34 crc kubenswrapper[4870]: I1014 07:19:34.609660 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"2bd1f41f-3f1a-4ca7-8789-429104ce2120","Type":"ContainerStarted","Data":"427d65d1b5cbeeeb2e36fcfaa93892d1227204dbfad3f98d9cee60708c1f3a0d"} Oct 14 07:19:34 crc kubenswrapper[4870]: I1014 07:19:34.611312 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-1a80-account-create-rcm8s" event={"ID":"3c2004c3-445d-4495-b24d-862f402ff5a6","Type":"ContainerDied","Data":"d99638d4ce866f2c194bd88c41448b5f4b3afac6075185b1f25720d324e564d5"} Oct 14 07:19:34 crc kubenswrapper[4870]: I1014 07:19:34.611357 4870 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d99638d4ce866f2c194bd88c41448b5f4b3afac6075185b1f25720d324e564d5" Oct 14 07:19:34 crc kubenswrapper[4870]: I1014 07:19:34.611314 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-1a80-account-create-rcm8s" Oct 14 07:19:34 crc kubenswrapper[4870]: I1014 07:19:34.622473 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-1aaa-account-create-6t7t7" event={"ID":"0c16dc18-0486-420e-86c8-ed2374a9a705","Type":"ContainerDied","Data":"5a9500467dfce86242c725ed352d82ee738550e7bc32a02518f77839d5475f8a"} Oct 14 07:19:34 crc kubenswrapper[4870]: I1014 07:19:34.622534 4870 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5a9500467dfce86242c725ed352d82ee738550e7bc32a02518f77839d5475f8a" Oct 14 07:19:34 crc kubenswrapper[4870]: I1014 07:19:34.622664 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-1aaa-account-create-6t7t7" Oct 14 07:19:35 crc kubenswrapper[4870]: I1014 07:19:35.636492 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"2bd1f41f-3f1a-4ca7-8789-429104ce2120","Type":"ContainerStarted","Data":"177c2ea0ed77d9c56284c56d85f024d847d0abd99647a2f3eedfd9a880d3b512"} Oct 14 07:19:36 crc kubenswrapper[4870]: I1014 07:19:36.064281 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-v2d79"] Oct 14 07:19:36 crc kubenswrapper[4870]: E1014 07:19:36.064626 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c16dc18-0486-420e-86c8-ed2374a9a705" containerName="mariadb-account-create" Oct 14 07:19:36 crc kubenswrapper[4870]: I1014 07:19:36.064643 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c16dc18-0486-420e-86c8-ed2374a9a705" containerName="mariadb-account-create" Oct 14 07:19:36 crc kubenswrapper[4870]: E1014 07:19:36.064666 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c2004c3-445d-4495-b24d-862f402ff5a6" containerName="mariadb-account-create" Oct 14 07:19:36 crc kubenswrapper[4870]: I1014 07:19:36.064673 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c2004c3-445d-4495-b24d-862f402ff5a6" containerName="mariadb-account-create" Oct 14 07:19:36 crc kubenswrapper[4870]: E1014 07:19:36.064692 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d318a57f-fb03-4112-b027-b8eaf943b00d" containerName="mariadb-account-create" Oct 14 07:19:36 crc kubenswrapper[4870]: I1014 07:19:36.064698 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="d318a57f-fb03-4112-b027-b8eaf943b00d" containerName="mariadb-account-create" Oct 14 07:19:36 crc kubenswrapper[4870]: I1014 07:19:36.064916 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="0c16dc18-0486-420e-86c8-ed2374a9a705" containerName="mariadb-account-create" Oct 14 07:19:36 crc kubenswrapper[4870]: I1014 07:19:36.064942 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="3c2004c3-445d-4495-b24d-862f402ff5a6" containerName="mariadb-account-create" Oct 14 07:19:36 crc kubenswrapper[4870]: I1014 07:19:36.064952 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="d318a57f-fb03-4112-b027-b8eaf943b00d" containerName="mariadb-account-create" Oct 14 07:19:36 crc kubenswrapper[4870]: I1014 07:19:36.065514 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-v2d79" Oct 14 07:19:36 crc kubenswrapper[4870]: I1014 07:19:36.067637 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-n9lf9" Oct 14 07:19:36 crc kubenswrapper[4870]: I1014 07:19:36.067813 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Oct 14 07:19:36 crc kubenswrapper[4870]: I1014 07:19:36.073426 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-v2d79"] Oct 14 07:19:36 crc kubenswrapper[4870]: I1014 07:19:36.097187 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x97n2\" (UniqueName: \"kubernetes.io/projected/85aaeb67-e809-416d-b4e0-eb7605cdd307-kube-api-access-x97n2\") pod \"glance-db-sync-v2d79\" (UID: \"85aaeb67-e809-416d-b4e0-eb7605cdd307\") " pod="openstack/glance-db-sync-v2d79" Oct 14 07:19:36 crc kubenswrapper[4870]: I1014 07:19:36.097284 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85aaeb67-e809-416d-b4e0-eb7605cdd307-combined-ca-bundle\") pod \"glance-db-sync-v2d79\" (UID: \"85aaeb67-e809-416d-b4e0-eb7605cdd307\") " pod="openstack/glance-db-sync-v2d79" Oct 14 07:19:36 crc kubenswrapper[4870]: I1014 07:19:36.097311 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/85aaeb67-e809-416d-b4e0-eb7605cdd307-db-sync-config-data\") pod \"glance-db-sync-v2d79\" (UID: \"85aaeb67-e809-416d-b4e0-eb7605cdd307\") " pod="openstack/glance-db-sync-v2d79" Oct 14 07:19:36 crc kubenswrapper[4870]: I1014 07:19:36.097358 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/85aaeb67-e809-416d-b4e0-eb7605cdd307-config-data\") pod \"glance-db-sync-v2d79\" (UID: \"85aaeb67-e809-416d-b4e0-eb7605cdd307\") " pod="openstack/glance-db-sync-v2d79" Oct 14 07:19:36 crc kubenswrapper[4870]: I1014 07:19:36.199336 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85aaeb67-e809-416d-b4e0-eb7605cdd307-combined-ca-bundle\") pod \"glance-db-sync-v2d79\" (UID: \"85aaeb67-e809-416d-b4e0-eb7605cdd307\") " pod="openstack/glance-db-sync-v2d79" Oct 14 07:19:36 crc kubenswrapper[4870]: I1014 07:19:36.200156 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/85aaeb67-e809-416d-b4e0-eb7605cdd307-db-sync-config-data\") pod \"glance-db-sync-v2d79\" (UID: \"85aaeb67-e809-416d-b4e0-eb7605cdd307\") " pod="openstack/glance-db-sync-v2d79" Oct 14 07:19:36 crc kubenswrapper[4870]: I1014 07:19:36.200212 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/85aaeb67-e809-416d-b4e0-eb7605cdd307-config-data\") pod \"glance-db-sync-v2d79\" (UID: \"85aaeb67-e809-416d-b4e0-eb7605cdd307\") " pod="openstack/glance-db-sync-v2d79" Oct 14 07:19:36 crc kubenswrapper[4870]: I1014 07:19:36.200323 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x97n2\" (UniqueName: \"kubernetes.io/projected/85aaeb67-e809-416d-b4e0-eb7605cdd307-kube-api-access-x97n2\") pod \"glance-db-sync-v2d79\" (UID: \"85aaeb67-e809-416d-b4e0-eb7605cdd307\") " pod="openstack/glance-db-sync-v2d79" Oct 14 07:19:36 crc kubenswrapper[4870]: I1014 07:19:36.207201 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/85aaeb67-e809-416d-b4e0-eb7605cdd307-db-sync-config-data\") pod \"glance-db-sync-v2d79\" (UID: \"85aaeb67-e809-416d-b4e0-eb7605cdd307\") " pod="openstack/glance-db-sync-v2d79" Oct 14 07:19:36 crc kubenswrapper[4870]: I1014 07:19:36.207348 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85aaeb67-e809-416d-b4e0-eb7605cdd307-combined-ca-bundle\") pod \"glance-db-sync-v2d79\" (UID: \"85aaeb67-e809-416d-b4e0-eb7605cdd307\") " pod="openstack/glance-db-sync-v2d79" Oct 14 07:19:36 crc kubenswrapper[4870]: I1014 07:19:36.208342 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/85aaeb67-e809-416d-b4e0-eb7605cdd307-config-data\") pod \"glance-db-sync-v2d79\" (UID: \"85aaeb67-e809-416d-b4e0-eb7605cdd307\") " pod="openstack/glance-db-sync-v2d79" Oct 14 07:19:36 crc kubenswrapper[4870]: I1014 07:19:36.224319 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x97n2\" (UniqueName: \"kubernetes.io/projected/85aaeb67-e809-416d-b4e0-eb7605cdd307-kube-api-access-x97n2\") pod \"glance-db-sync-v2d79\" (UID: \"85aaeb67-e809-416d-b4e0-eb7605cdd307\") " pod="openstack/glance-db-sync-v2d79" Oct 14 07:19:36 crc kubenswrapper[4870]: I1014 07:19:36.390493 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-v2d79" Oct 14 07:19:36 crc kubenswrapper[4870]: I1014 07:19:36.670992 4870 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-rn98c" podUID="503a07f8-2c05-458a-80ff-1abfe973dbd5" containerName="ovn-controller" probeResult="failure" output=< Oct 14 07:19:36 crc kubenswrapper[4870]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Oct 14 07:19:36 crc kubenswrapper[4870]: > Oct 14 07:19:36 crc kubenswrapper[4870]: I1014 07:19:36.699551 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"2bd1f41f-3f1a-4ca7-8789-429104ce2120","Type":"ContainerStarted","Data":"418f1f077bbfc2949a0abda0f2ddd66da4131527026acec0d19a4c08dec4dc66"} Oct 14 07:19:36 crc kubenswrapper[4870]: I1014 07:19:36.699616 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"2bd1f41f-3f1a-4ca7-8789-429104ce2120","Type":"ContainerStarted","Data":"642a18f2173ab4124b7442763988bbe3a2c0fdd80e87caad5e48f172fd55efc7"} Oct 14 07:19:36 crc kubenswrapper[4870]: I1014 07:19:36.704822 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-6482n" Oct 14 07:19:36 crc kubenswrapper[4870]: I1014 07:19:36.771978 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-6482n" Oct 14 07:19:36 crc kubenswrapper[4870]: I1014 07:19:36.836802 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-v2d79"] Oct 14 07:19:36 crc kubenswrapper[4870]: W1014 07:19:36.862608 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod85aaeb67_e809_416d_b4e0_eb7605cdd307.slice/crio-b45d93a9f7cd85e1331326b27328fbb7859749e534da22dfff45f8a22d445a33 WatchSource:0}: Error finding container b45d93a9f7cd85e1331326b27328fbb7859749e534da22dfff45f8a22d445a33: Status 404 returned error can't find the container with id b45d93a9f7cd85e1331326b27328fbb7859749e534da22dfff45f8a22d445a33 Oct 14 07:19:36 crc kubenswrapper[4870]: I1014 07:19:36.994273 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-rn98c-config-2s96z"] Oct 14 07:19:36 crc kubenswrapper[4870]: I1014 07:19:36.995289 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-rn98c-config-2s96z" Oct 14 07:19:36 crc kubenswrapper[4870]: I1014 07:19:36.997580 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Oct 14 07:19:37 crc kubenswrapper[4870]: I1014 07:19:37.011132 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-rn98c-config-2s96z"] Oct 14 07:19:37 crc kubenswrapper[4870]: I1014 07:19:37.121717 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/631522fc-993b-496a-8ebb-03b289e5cc7a-var-run\") pod \"ovn-controller-rn98c-config-2s96z\" (UID: \"631522fc-993b-496a-8ebb-03b289e5cc7a\") " pod="openstack/ovn-controller-rn98c-config-2s96z" Oct 14 07:19:37 crc kubenswrapper[4870]: I1014 07:19:37.121773 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/631522fc-993b-496a-8ebb-03b289e5cc7a-var-run-ovn\") pod \"ovn-controller-rn98c-config-2s96z\" (UID: \"631522fc-993b-496a-8ebb-03b289e5cc7a\") " pod="openstack/ovn-controller-rn98c-config-2s96z" Oct 14 07:19:37 crc kubenswrapper[4870]: I1014 07:19:37.121804 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mmcrt\" (UniqueName: \"kubernetes.io/projected/631522fc-993b-496a-8ebb-03b289e5cc7a-kube-api-access-mmcrt\") pod \"ovn-controller-rn98c-config-2s96z\" (UID: \"631522fc-993b-496a-8ebb-03b289e5cc7a\") " pod="openstack/ovn-controller-rn98c-config-2s96z" Oct 14 07:19:37 crc kubenswrapper[4870]: I1014 07:19:37.121980 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/631522fc-993b-496a-8ebb-03b289e5cc7a-var-log-ovn\") pod \"ovn-controller-rn98c-config-2s96z\" (UID: \"631522fc-993b-496a-8ebb-03b289e5cc7a\") " pod="openstack/ovn-controller-rn98c-config-2s96z" Oct 14 07:19:37 crc kubenswrapper[4870]: I1014 07:19:37.122253 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/631522fc-993b-496a-8ebb-03b289e5cc7a-scripts\") pod \"ovn-controller-rn98c-config-2s96z\" (UID: \"631522fc-993b-496a-8ebb-03b289e5cc7a\") " pod="openstack/ovn-controller-rn98c-config-2s96z" Oct 14 07:19:37 crc kubenswrapper[4870]: I1014 07:19:37.122309 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/631522fc-993b-496a-8ebb-03b289e5cc7a-additional-scripts\") pod \"ovn-controller-rn98c-config-2s96z\" (UID: \"631522fc-993b-496a-8ebb-03b289e5cc7a\") " pod="openstack/ovn-controller-rn98c-config-2s96z" Oct 14 07:19:37 crc kubenswrapper[4870]: I1014 07:19:37.223876 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/631522fc-993b-496a-8ebb-03b289e5cc7a-scripts\") pod \"ovn-controller-rn98c-config-2s96z\" (UID: \"631522fc-993b-496a-8ebb-03b289e5cc7a\") " pod="openstack/ovn-controller-rn98c-config-2s96z" Oct 14 07:19:37 crc kubenswrapper[4870]: I1014 07:19:37.224211 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/631522fc-993b-496a-8ebb-03b289e5cc7a-additional-scripts\") pod \"ovn-controller-rn98c-config-2s96z\" (UID: \"631522fc-993b-496a-8ebb-03b289e5cc7a\") " pod="openstack/ovn-controller-rn98c-config-2s96z" Oct 14 07:19:37 crc kubenswrapper[4870]: I1014 07:19:37.224256 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/631522fc-993b-496a-8ebb-03b289e5cc7a-var-run\") pod \"ovn-controller-rn98c-config-2s96z\" (UID: \"631522fc-993b-496a-8ebb-03b289e5cc7a\") " pod="openstack/ovn-controller-rn98c-config-2s96z" Oct 14 07:19:37 crc kubenswrapper[4870]: I1014 07:19:37.224288 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/631522fc-993b-496a-8ebb-03b289e5cc7a-var-run-ovn\") pod \"ovn-controller-rn98c-config-2s96z\" (UID: \"631522fc-993b-496a-8ebb-03b289e5cc7a\") " pod="openstack/ovn-controller-rn98c-config-2s96z" Oct 14 07:19:37 crc kubenswrapper[4870]: I1014 07:19:37.224308 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mmcrt\" (UniqueName: \"kubernetes.io/projected/631522fc-993b-496a-8ebb-03b289e5cc7a-kube-api-access-mmcrt\") pod \"ovn-controller-rn98c-config-2s96z\" (UID: \"631522fc-993b-496a-8ebb-03b289e5cc7a\") " pod="openstack/ovn-controller-rn98c-config-2s96z" Oct 14 07:19:37 crc kubenswrapper[4870]: I1014 07:19:37.224617 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/631522fc-993b-496a-8ebb-03b289e5cc7a-var-run\") pod \"ovn-controller-rn98c-config-2s96z\" (UID: \"631522fc-993b-496a-8ebb-03b289e5cc7a\") " pod="openstack/ovn-controller-rn98c-config-2s96z" Oct 14 07:19:37 crc kubenswrapper[4870]: I1014 07:19:37.224641 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/631522fc-993b-496a-8ebb-03b289e5cc7a-var-run-ovn\") pod \"ovn-controller-rn98c-config-2s96z\" (UID: \"631522fc-993b-496a-8ebb-03b289e5cc7a\") " pod="openstack/ovn-controller-rn98c-config-2s96z" Oct 14 07:19:37 crc kubenswrapper[4870]: I1014 07:19:37.224683 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/631522fc-993b-496a-8ebb-03b289e5cc7a-var-log-ovn\") pod \"ovn-controller-rn98c-config-2s96z\" (UID: \"631522fc-993b-496a-8ebb-03b289e5cc7a\") " pod="openstack/ovn-controller-rn98c-config-2s96z" Oct 14 07:19:37 crc kubenswrapper[4870]: I1014 07:19:37.224630 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/631522fc-993b-496a-8ebb-03b289e5cc7a-var-log-ovn\") pod \"ovn-controller-rn98c-config-2s96z\" (UID: \"631522fc-993b-496a-8ebb-03b289e5cc7a\") " pod="openstack/ovn-controller-rn98c-config-2s96z" Oct 14 07:19:37 crc kubenswrapper[4870]: I1014 07:19:37.225393 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/631522fc-993b-496a-8ebb-03b289e5cc7a-additional-scripts\") pod \"ovn-controller-rn98c-config-2s96z\" (UID: \"631522fc-993b-496a-8ebb-03b289e5cc7a\") " pod="openstack/ovn-controller-rn98c-config-2s96z" Oct 14 07:19:37 crc kubenswrapper[4870]: I1014 07:19:37.226437 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/631522fc-993b-496a-8ebb-03b289e5cc7a-scripts\") pod \"ovn-controller-rn98c-config-2s96z\" (UID: \"631522fc-993b-496a-8ebb-03b289e5cc7a\") " pod="openstack/ovn-controller-rn98c-config-2s96z" Oct 14 07:19:37 crc kubenswrapper[4870]: I1014 07:19:37.242197 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mmcrt\" (UniqueName: \"kubernetes.io/projected/631522fc-993b-496a-8ebb-03b289e5cc7a-kube-api-access-mmcrt\") pod \"ovn-controller-rn98c-config-2s96z\" (UID: \"631522fc-993b-496a-8ebb-03b289e5cc7a\") " pod="openstack/ovn-controller-rn98c-config-2s96z" Oct 14 07:19:37 crc kubenswrapper[4870]: I1014 07:19:37.310118 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-rn98c-config-2s96z" Oct 14 07:19:37 crc kubenswrapper[4870]: I1014 07:19:37.714191 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"2bd1f41f-3f1a-4ca7-8789-429104ce2120","Type":"ContainerStarted","Data":"42954c6bfcca40f75f1c6f4bcad9b47bf39a17a66a9a6f9bbf941bbed4d28ebd"} Oct 14 07:19:37 crc kubenswrapper[4870]: I1014 07:19:37.714562 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"2bd1f41f-3f1a-4ca7-8789-429104ce2120","Type":"ContainerStarted","Data":"a2214a07cbc03e9d4c3657396572e211461dec75c50cb838dd53f64e5d4ae7de"} Oct 14 07:19:37 crc kubenswrapper[4870]: I1014 07:19:37.716656 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-v2d79" event={"ID":"85aaeb67-e809-416d-b4e0-eb7605cdd307","Type":"ContainerStarted","Data":"b45d93a9f7cd85e1331326b27328fbb7859749e534da22dfff45f8a22d445a33"} Oct 14 07:19:37 crc kubenswrapper[4870]: I1014 07:19:37.785924 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-rn98c-config-2s96z"] Oct 14 07:19:38 crc kubenswrapper[4870]: I1014 07:19:38.727140 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-rn98c-config-2s96z" event={"ID":"631522fc-993b-496a-8ebb-03b289e5cc7a","Type":"ContainerStarted","Data":"1c949b8d27a9dfc9f3814669313a8b8a65011e66ad70b42109a72ddb8e1e9aba"} Oct 14 07:19:39 crc kubenswrapper[4870]: I1014 07:19:39.738228 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"2bd1f41f-3f1a-4ca7-8789-429104ce2120","Type":"ContainerStarted","Data":"18b08dbc7de7f34cc2bbb65969ca2e9d3e2da4bddab756fdee2d7db4e53e9180"} Oct 14 07:19:39 crc kubenswrapper[4870]: I1014 07:19:39.738666 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"2bd1f41f-3f1a-4ca7-8789-429104ce2120","Type":"ContainerStarted","Data":"b51d4fe9abeae800a8b66afb49b1fafee55d1cdb96718d65b7d5b35f6fd7519b"} Oct 14 07:19:39 crc kubenswrapper[4870]: I1014 07:19:39.738679 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"2bd1f41f-3f1a-4ca7-8789-429104ce2120","Type":"ContainerStarted","Data":"5330b69d8d67676f3cc5cd3efe7cfe42a11bfa0006bfe019a9c536b30e4c1f0d"} Oct 14 07:19:39 crc kubenswrapper[4870]: I1014 07:19:39.741056 4870 generic.go:334] "Generic (PLEG): container finished" podID="631522fc-993b-496a-8ebb-03b289e5cc7a" containerID="3e6bcfe5922f64d7458739850f7cc5cf193c595b1c6d384c0bbf95f39c7959bc" exitCode=0 Oct 14 07:19:39 crc kubenswrapper[4870]: I1014 07:19:39.741131 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-rn98c-config-2s96z" event={"ID":"631522fc-993b-496a-8ebb-03b289e5cc7a","Type":"ContainerDied","Data":"3e6bcfe5922f64d7458739850f7cc5cf193c595b1c6d384c0bbf95f39c7959bc"} Oct 14 07:19:40 crc kubenswrapper[4870]: I1014 07:19:40.774522 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"2bd1f41f-3f1a-4ca7-8789-429104ce2120","Type":"ContainerStarted","Data":"9104317bae5cf155b62976217e7c03cd64c95837b39e557bc2162f991c3e3920"} Oct 14 07:19:40 crc kubenswrapper[4870]: I1014 07:19:40.774559 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"2bd1f41f-3f1a-4ca7-8789-429104ce2120","Type":"ContainerStarted","Data":"eaac2b7e339adcaf70b58432d22ea22656cd6aebcd6871bb19adeae4be5f56f0"} Oct 14 07:19:40 crc kubenswrapper[4870]: I1014 07:19:40.774569 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"2bd1f41f-3f1a-4ca7-8789-429104ce2120","Type":"ContainerStarted","Data":"7c8fe31733cd77d0b1ee88ab091697a9851a822a29c50d287246ced8eca4d502"} Oct 14 07:19:40 crc kubenswrapper[4870]: I1014 07:19:40.774577 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"2bd1f41f-3f1a-4ca7-8789-429104ce2120","Type":"ContainerStarted","Data":"ad9dccd6af895adbf1e3c3fcd2eca40527602d0f3bfc6d6d7ea2730f4a267119"} Oct 14 07:19:40 crc kubenswrapper[4870]: I1014 07:19:40.831120 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=20.387830913 podStartE2EDuration="28.831101504s" podCreationTimestamp="2025-10-14 07:19:12 +0000 UTC" firstStartedPulling="2025-10-14 07:19:30.723283713 +0000 UTC m=+1106.420644084" lastFinishedPulling="2025-10-14 07:19:39.166554304 +0000 UTC m=+1114.863914675" observedRunningTime="2025-10-14 07:19:40.815942719 +0000 UTC m=+1116.513303090" watchObservedRunningTime="2025-10-14 07:19:40.831101504 +0000 UTC m=+1116.528461875" Oct 14 07:19:41 crc kubenswrapper[4870]: I1014 07:19:41.120665 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7cc794b75c-d2b9c"] Oct 14 07:19:41 crc kubenswrapper[4870]: I1014 07:19:41.122093 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7cc794b75c-d2b9c" Oct 14 07:19:41 crc kubenswrapper[4870]: I1014 07:19:41.124477 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Oct 14 07:19:41 crc kubenswrapper[4870]: I1014 07:19:41.132111 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7cc794b75c-d2b9c"] Oct 14 07:19:41 crc kubenswrapper[4870]: I1014 07:19:41.220398 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/69a02fec-1745-4827-9c90-d9dffb829a66-ovsdbserver-sb\") pod \"dnsmasq-dns-7cc794b75c-d2b9c\" (UID: \"69a02fec-1745-4827-9c90-d9dffb829a66\") " pod="openstack/dnsmasq-dns-7cc794b75c-d2b9c" Oct 14 07:19:41 crc kubenswrapper[4870]: I1014 07:19:41.220516 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/69a02fec-1745-4827-9c90-d9dffb829a66-config\") pod \"dnsmasq-dns-7cc794b75c-d2b9c\" (UID: \"69a02fec-1745-4827-9c90-d9dffb829a66\") " pod="openstack/dnsmasq-dns-7cc794b75c-d2b9c" Oct 14 07:19:41 crc kubenswrapper[4870]: I1014 07:19:41.220548 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/69a02fec-1745-4827-9c90-d9dffb829a66-dns-swift-storage-0\") pod \"dnsmasq-dns-7cc794b75c-d2b9c\" (UID: \"69a02fec-1745-4827-9c90-d9dffb829a66\") " pod="openstack/dnsmasq-dns-7cc794b75c-d2b9c" Oct 14 07:19:41 crc kubenswrapper[4870]: I1014 07:19:41.220635 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/69a02fec-1745-4827-9c90-d9dffb829a66-ovsdbserver-nb\") pod \"dnsmasq-dns-7cc794b75c-d2b9c\" (UID: \"69a02fec-1745-4827-9c90-d9dffb829a66\") " pod="openstack/dnsmasq-dns-7cc794b75c-d2b9c" Oct 14 07:19:41 crc kubenswrapper[4870]: I1014 07:19:41.220661 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/69a02fec-1745-4827-9c90-d9dffb829a66-dns-svc\") pod \"dnsmasq-dns-7cc794b75c-d2b9c\" (UID: \"69a02fec-1745-4827-9c90-d9dffb829a66\") " pod="openstack/dnsmasq-dns-7cc794b75c-d2b9c" Oct 14 07:19:41 crc kubenswrapper[4870]: I1014 07:19:41.220695 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-njwsb\" (UniqueName: \"kubernetes.io/projected/69a02fec-1745-4827-9c90-d9dffb829a66-kube-api-access-njwsb\") pod \"dnsmasq-dns-7cc794b75c-d2b9c\" (UID: \"69a02fec-1745-4827-9c90-d9dffb829a66\") " pod="openstack/dnsmasq-dns-7cc794b75c-d2b9c" Oct 14 07:19:41 crc kubenswrapper[4870]: I1014 07:19:41.321992 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/69a02fec-1745-4827-9c90-d9dffb829a66-config\") pod \"dnsmasq-dns-7cc794b75c-d2b9c\" (UID: \"69a02fec-1745-4827-9c90-d9dffb829a66\") " pod="openstack/dnsmasq-dns-7cc794b75c-d2b9c" Oct 14 07:19:41 crc kubenswrapper[4870]: I1014 07:19:41.322062 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/69a02fec-1745-4827-9c90-d9dffb829a66-dns-swift-storage-0\") pod \"dnsmasq-dns-7cc794b75c-d2b9c\" (UID: \"69a02fec-1745-4827-9c90-d9dffb829a66\") " pod="openstack/dnsmasq-dns-7cc794b75c-d2b9c" Oct 14 07:19:41 crc kubenswrapper[4870]: I1014 07:19:41.322158 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/69a02fec-1745-4827-9c90-d9dffb829a66-ovsdbserver-nb\") pod \"dnsmasq-dns-7cc794b75c-d2b9c\" (UID: \"69a02fec-1745-4827-9c90-d9dffb829a66\") " pod="openstack/dnsmasq-dns-7cc794b75c-d2b9c" Oct 14 07:19:41 crc kubenswrapper[4870]: I1014 07:19:41.322181 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/69a02fec-1745-4827-9c90-d9dffb829a66-dns-svc\") pod \"dnsmasq-dns-7cc794b75c-d2b9c\" (UID: \"69a02fec-1745-4827-9c90-d9dffb829a66\") " pod="openstack/dnsmasq-dns-7cc794b75c-d2b9c" Oct 14 07:19:41 crc kubenswrapper[4870]: I1014 07:19:41.322215 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-njwsb\" (UniqueName: \"kubernetes.io/projected/69a02fec-1745-4827-9c90-d9dffb829a66-kube-api-access-njwsb\") pod \"dnsmasq-dns-7cc794b75c-d2b9c\" (UID: \"69a02fec-1745-4827-9c90-d9dffb829a66\") " pod="openstack/dnsmasq-dns-7cc794b75c-d2b9c" Oct 14 07:19:41 crc kubenswrapper[4870]: I1014 07:19:41.322249 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/69a02fec-1745-4827-9c90-d9dffb829a66-ovsdbserver-sb\") pod \"dnsmasq-dns-7cc794b75c-d2b9c\" (UID: \"69a02fec-1745-4827-9c90-d9dffb829a66\") " pod="openstack/dnsmasq-dns-7cc794b75c-d2b9c" Oct 14 07:19:41 crc kubenswrapper[4870]: I1014 07:19:41.323329 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/69a02fec-1745-4827-9c90-d9dffb829a66-ovsdbserver-sb\") pod \"dnsmasq-dns-7cc794b75c-d2b9c\" (UID: \"69a02fec-1745-4827-9c90-d9dffb829a66\") " pod="openstack/dnsmasq-dns-7cc794b75c-d2b9c" Oct 14 07:19:41 crc kubenswrapper[4870]: I1014 07:19:41.324012 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/69a02fec-1745-4827-9c90-d9dffb829a66-config\") pod \"dnsmasq-dns-7cc794b75c-d2b9c\" (UID: \"69a02fec-1745-4827-9c90-d9dffb829a66\") " pod="openstack/dnsmasq-dns-7cc794b75c-d2b9c" Oct 14 07:19:41 crc kubenswrapper[4870]: I1014 07:19:41.324322 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/69a02fec-1745-4827-9c90-d9dffb829a66-ovsdbserver-nb\") pod \"dnsmasq-dns-7cc794b75c-d2b9c\" (UID: \"69a02fec-1745-4827-9c90-d9dffb829a66\") " pod="openstack/dnsmasq-dns-7cc794b75c-d2b9c" Oct 14 07:19:41 crc kubenswrapper[4870]: I1014 07:19:41.324720 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/69a02fec-1745-4827-9c90-d9dffb829a66-dns-svc\") pod \"dnsmasq-dns-7cc794b75c-d2b9c\" (UID: \"69a02fec-1745-4827-9c90-d9dffb829a66\") " pod="openstack/dnsmasq-dns-7cc794b75c-d2b9c" Oct 14 07:19:41 crc kubenswrapper[4870]: I1014 07:19:41.324992 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/69a02fec-1745-4827-9c90-d9dffb829a66-dns-swift-storage-0\") pod \"dnsmasq-dns-7cc794b75c-d2b9c\" (UID: \"69a02fec-1745-4827-9c90-d9dffb829a66\") " pod="openstack/dnsmasq-dns-7cc794b75c-d2b9c" Oct 14 07:19:41 crc kubenswrapper[4870]: I1014 07:19:41.352650 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-njwsb\" (UniqueName: \"kubernetes.io/projected/69a02fec-1745-4827-9c90-d9dffb829a66-kube-api-access-njwsb\") pod \"dnsmasq-dns-7cc794b75c-d2b9c\" (UID: \"69a02fec-1745-4827-9c90-d9dffb829a66\") " pod="openstack/dnsmasq-dns-7cc794b75c-d2b9c" Oct 14 07:19:41 crc kubenswrapper[4870]: I1014 07:19:41.495042 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7cc794b75c-d2b9c" Oct 14 07:19:41 crc kubenswrapper[4870]: I1014 07:19:41.677127 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-rn98c" Oct 14 07:19:41 crc kubenswrapper[4870]: I1014 07:19:41.989304 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-rn98c-config-2s96z" Oct 14 07:19:42 crc kubenswrapper[4870]: I1014 07:19:42.034625 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/631522fc-993b-496a-8ebb-03b289e5cc7a-var-log-ovn\") pod \"631522fc-993b-496a-8ebb-03b289e5cc7a\" (UID: \"631522fc-993b-496a-8ebb-03b289e5cc7a\") " Oct 14 07:19:42 crc kubenswrapper[4870]: I1014 07:19:42.034697 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/631522fc-993b-496a-8ebb-03b289e5cc7a-var-run-ovn\") pod \"631522fc-993b-496a-8ebb-03b289e5cc7a\" (UID: \"631522fc-993b-496a-8ebb-03b289e5cc7a\") " Oct 14 07:19:42 crc kubenswrapper[4870]: I1014 07:19:42.034776 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/631522fc-993b-496a-8ebb-03b289e5cc7a-var-run\") pod \"631522fc-993b-496a-8ebb-03b289e5cc7a\" (UID: \"631522fc-993b-496a-8ebb-03b289e5cc7a\") " Oct 14 07:19:42 crc kubenswrapper[4870]: I1014 07:19:42.034827 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/631522fc-993b-496a-8ebb-03b289e5cc7a-scripts\") pod \"631522fc-993b-496a-8ebb-03b289e5cc7a\" (UID: \"631522fc-993b-496a-8ebb-03b289e5cc7a\") " Oct 14 07:19:42 crc kubenswrapper[4870]: I1014 07:19:42.034864 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mmcrt\" (UniqueName: \"kubernetes.io/projected/631522fc-993b-496a-8ebb-03b289e5cc7a-kube-api-access-mmcrt\") pod \"631522fc-993b-496a-8ebb-03b289e5cc7a\" (UID: \"631522fc-993b-496a-8ebb-03b289e5cc7a\") " Oct 14 07:19:42 crc kubenswrapper[4870]: I1014 07:19:42.034959 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/631522fc-993b-496a-8ebb-03b289e5cc7a-additional-scripts\") pod \"631522fc-993b-496a-8ebb-03b289e5cc7a\" (UID: \"631522fc-993b-496a-8ebb-03b289e5cc7a\") " Oct 14 07:19:42 crc kubenswrapper[4870]: I1014 07:19:42.035460 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/631522fc-993b-496a-8ebb-03b289e5cc7a-var-run" (OuterVolumeSpecName: "var-run") pod "631522fc-993b-496a-8ebb-03b289e5cc7a" (UID: "631522fc-993b-496a-8ebb-03b289e5cc7a"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 14 07:19:42 crc kubenswrapper[4870]: I1014 07:19:42.035502 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/631522fc-993b-496a-8ebb-03b289e5cc7a-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "631522fc-993b-496a-8ebb-03b289e5cc7a" (UID: "631522fc-993b-496a-8ebb-03b289e5cc7a"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 14 07:19:42 crc kubenswrapper[4870]: I1014 07:19:42.035524 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/631522fc-993b-496a-8ebb-03b289e5cc7a-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "631522fc-993b-496a-8ebb-03b289e5cc7a" (UID: "631522fc-993b-496a-8ebb-03b289e5cc7a"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 14 07:19:42 crc kubenswrapper[4870]: I1014 07:19:42.036165 4870 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/631522fc-993b-496a-8ebb-03b289e5cc7a-var-log-ovn\") on node \"crc\" DevicePath \"\"" Oct 14 07:19:42 crc kubenswrapper[4870]: I1014 07:19:42.036186 4870 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/631522fc-993b-496a-8ebb-03b289e5cc7a-var-run-ovn\") on node \"crc\" DevicePath \"\"" Oct 14 07:19:42 crc kubenswrapper[4870]: I1014 07:19:42.036196 4870 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/631522fc-993b-496a-8ebb-03b289e5cc7a-var-run\") on node \"crc\" DevicePath \"\"" Oct 14 07:19:42 crc kubenswrapper[4870]: I1014 07:19:42.036201 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/631522fc-993b-496a-8ebb-03b289e5cc7a-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "631522fc-993b-496a-8ebb-03b289e5cc7a" (UID: "631522fc-993b-496a-8ebb-03b289e5cc7a"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:19:42 crc kubenswrapper[4870]: I1014 07:19:42.036585 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/631522fc-993b-496a-8ebb-03b289e5cc7a-scripts" (OuterVolumeSpecName: "scripts") pod "631522fc-993b-496a-8ebb-03b289e5cc7a" (UID: "631522fc-993b-496a-8ebb-03b289e5cc7a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:19:42 crc kubenswrapper[4870]: I1014 07:19:42.044033 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/631522fc-993b-496a-8ebb-03b289e5cc7a-kube-api-access-mmcrt" (OuterVolumeSpecName: "kube-api-access-mmcrt") pod "631522fc-993b-496a-8ebb-03b289e5cc7a" (UID: "631522fc-993b-496a-8ebb-03b289e5cc7a"). InnerVolumeSpecName "kube-api-access-mmcrt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:19:42 crc kubenswrapper[4870]: I1014 07:19:42.137395 4870 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/631522fc-993b-496a-8ebb-03b289e5cc7a-additional-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 07:19:42 crc kubenswrapper[4870]: I1014 07:19:42.137429 4870 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/631522fc-993b-496a-8ebb-03b289e5cc7a-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 07:19:42 crc kubenswrapper[4870]: I1014 07:19:42.137458 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mmcrt\" (UniqueName: \"kubernetes.io/projected/631522fc-993b-496a-8ebb-03b289e5cc7a-kube-api-access-mmcrt\") on node \"crc\" DevicePath \"\"" Oct 14 07:19:42 crc kubenswrapper[4870]: I1014 07:19:42.423684 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7cc794b75c-d2b9c"] Oct 14 07:19:42 crc kubenswrapper[4870]: I1014 07:19:42.793692 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-rn98c-config-2s96z" event={"ID":"631522fc-993b-496a-8ebb-03b289e5cc7a","Type":"ContainerDied","Data":"1c949b8d27a9dfc9f3814669313a8b8a65011e66ad70b42109a72ddb8e1e9aba"} Oct 14 07:19:42 crc kubenswrapper[4870]: I1014 07:19:42.794076 4870 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1c949b8d27a9dfc9f3814669313a8b8a65011e66ad70b42109a72ddb8e1e9aba" Oct 14 07:19:42 crc kubenswrapper[4870]: I1014 07:19:42.793954 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-rn98c-config-2s96z" Oct 14 07:19:42 crc kubenswrapper[4870]: I1014 07:19:42.796964 4870 generic.go:334] "Generic (PLEG): container finished" podID="69a02fec-1745-4827-9c90-d9dffb829a66" containerID="d4bce05dddb71b43960de2db8b34835c15e33a352960018f56e19358841c968e" exitCode=0 Oct 14 07:19:42 crc kubenswrapper[4870]: I1014 07:19:42.797004 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cc794b75c-d2b9c" event={"ID":"69a02fec-1745-4827-9c90-d9dffb829a66","Type":"ContainerDied","Data":"d4bce05dddb71b43960de2db8b34835c15e33a352960018f56e19358841c968e"} Oct 14 07:19:42 crc kubenswrapper[4870]: I1014 07:19:42.797030 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cc794b75c-d2b9c" event={"ID":"69a02fec-1745-4827-9c90-d9dffb829a66","Type":"ContainerStarted","Data":"8c2d981927ad6fd023274e61249b87cbf6a8f5b6a813de27473ea342707764b2"} Oct 14 07:19:43 crc kubenswrapper[4870]: I1014 07:19:43.127584 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-rn98c-config-2s96z"] Oct 14 07:19:43 crc kubenswrapper[4870]: I1014 07:19:43.133733 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-rn98c-config-2s96z"] Oct 14 07:19:43 crc kubenswrapper[4870]: I1014 07:19:43.209918 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-rn98c-config-gnlxw"] Oct 14 07:19:43 crc kubenswrapper[4870]: E1014 07:19:43.210320 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="631522fc-993b-496a-8ebb-03b289e5cc7a" containerName="ovn-config" Oct 14 07:19:43 crc kubenswrapper[4870]: I1014 07:19:43.210345 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="631522fc-993b-496a-8ebb-03b289e5cc7a" containerName="ovn-config" Oct 14 07:19:43 crc kubenswrapper[4870]: I1014 07:19:43.210599 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="631522fc-993b-496a-8ebb-03b289e5cc7a" containerName="ovn-config" Oct 14 07:19:43 crc kubenswrapper[4870]: I1014 07:19:43.211295 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-rn98c-config-gnlxw" Oct 14 07:19:43 crc kubenswrapper[4870]: I1014 07:19:43.214533 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Oct 14 07:19:43 crc kubenswrapper[4870]: I1014 07:19:43.235049 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-rn98c-config-gnlxw"] Oct 14 07:19:43 crc kubenswrapper[4870]: I1014 07:19:43.270804 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/7c8980f9-662b-46a7-a017-c1f038c3566c-additional-scripts\") pod \"ovn-controller-rn98c-config-gnlxw\" (UID: \"7c8980f9-662b-46a7-a017-c1f038c3566c\") " pod="openstack/ovn-controller-rn98c-config-gnlxw" Oct 14 07:19:43 crc kubenswrapper[4870]: I1014 07:19:43.270929 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/7c8980f9-662b-46a7-a017-c1f038c3566c-var-run\") pod \"ovn-controller-rn98c-config-gnlxw\" (UID: \"7c8980f9-662b-46a7-a017-c1f038c3566c\") " pod="openstack/ovn-controller-rn98c-config-gnlxw" Oct 14 07:19:43 crc kubenswrapper[4870]: I1014 07:19:43.270954 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/7c8980f9-662b-46a7-a017-c1f038c3566c-var-run-ovn\") pod \"ovn-controller-rn98c-config-gnlxw\" (UID: \"7c8980f9-662b-46a7-a017-c1f038c3566c\") " pod="openstack/ovn-controller-rn98c-config-gnlxw" Oct 14 07:19:43 crc kubenswrapper[4870]: I1014 07:19:43.271010 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7c8980f9-662b-46a7-a017-c1f038c3566c-scripts\") pod \"ovn-controller-rn98c-config-gnlxw\" (UID: \"7c8980f9-662b-46a7-a017-c1f038c3566c\") " pod="openstack/ovn-controller-rn98c-config-gnlxw" Oct 14 07:19:43 crc kubenswrapper[4870]: I1014 07:19:43.271067 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nxltb\" (UniqueName: \"kubernetes.io/projected/7c8980f9-662b-46a7-a017-c1f038c3566c-kube-api-access-nxltb\") pod \"ovn-controller-rn98c-config-gnlxw\" (UID: \"7c8980f9-662b-46a7-a017-c1f038c3566c\") " pod="openstack/ovn-controller-rn98c-config-gnlxw" Oct 14 07:19:43 crc kubenswrapper[4870]: I1014 07:19:43.271096 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/7c8980f9-662b-46a7-a017-c1f038c3566c-var-log-ovn\") pod \"ovn-controller-rn98c-config-gnlxw\" (UID: \"7c8980f9-662b-46a7-a017-c1f038c3566c\") " pod="openstack/ovn-controller-rn98c-config-gnlxw" Oct 14 07:19:43 crc kubenswrapper[4870]: I1014 07:19:43.372603 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/7c8980f9-662b-46a7-a017-c1f038c3566c-additional-scripts\") pod \"ovn-controller-rn98c-config-gnlxw\" (UID: \"7c8980f9-662b-46a7-a017-c1f038c3566c\") " pod="openstack/ovn-controller-rn98c-config-gnlxw" Oct 14 07:19:43 crc kubenswrapper[4870]: I1014 07:19:43.372705 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/7c8980f9-662b-46a7-a017-c1f038c3566c-var-run\") pod \"ovn-controller-rn98c-config-gnlxw\" (UID: \"7c8980f9-662b-46a7-a017-c1f038c3566c\") " pod="openstack/ovn-controller-rn98c-config-gnlxw" Oct 14 07:19:43 crc kubenswrapper[4870]: I1014 07:19:43.372733 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/7c8980f9-662b-46a7-a017-c1f038c3566c-var-run-ovn\") pod \"ovn-controller-rn98c-config-gnlxw\" (UID: \"7c8980f9-662b-46a7-a017-c1f038c3566c\") " pod="openstack/ovn-controller-rn98c-config-gnlxw" Oct 14 07:19:43 crc kubenswrapper[4870]: I1014 07:19:43.372777 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7c8980f9-662b-46a7-a017-c1f038c3566c-scripts\") pod \"ovn-controller-rn98c-config-gnlxw\" (UID: \"7c8980f9-662b-46a7-a017-c1f038c3566c\") " pod="openstack/ovn-controller-rn98c-config-gnlxw" Oct 14 07:19:43 crc kubenswrapper[4870]: I1014 07:19:43.372821 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nxltb\" (UniqueName: \"kubernetes.io/projected/7c8980f9-662b-46a7-a017-c1f038c3566c-kube-api-access-nxltb\") pod \"ovn-controller-rn98c-config-gnlxw\" (UID: \"7c8980f9-662b-46a7-a017-c1f038c3566c\") " pod="openstack/ovn-controller-rn98c-config-gnlxw" Oct 14 07:19:43 crc kubenswrapper[4870]: I1014 07:19:43.372845 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/7c8980f9-662b-46a7-a017-c1f038c3566c-var-log-ovn\") pod \"ovn-controller-rn98c-config-gnlxw\" (UID: \"7c8980f9-662b-46a7-a017-c1f038c3566c\") " pod="openstack/ovn-controller-rn98c-config-gnlxw" Oct 14 07:19:43 crc kubenswrapper[4870]: I1014 07:19:43.373192 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/7c8980f9-662b-46a7-a017-c1f038c3566c-var-run\") pod \"ovn-controller-rn98c-config-gnlxw\" (UID: \"7c8980f9-662b-46a7-a017-c1f038c3566c\") " pod="openstack/ovn-controller-rn98c-config-gnlxw" Oct 14 07:19:43 crc kubenswrapper[4870]: I1014 07:19:43.373241 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/7c8980f9-662b-46a7-a017-c1f038c3566c-var-run-ovn\") pod \"ovn-controller-rn98c-config-gnlxw\" (UID: \"7c8980f9-662b-46a7-a017-c1f038c3566c\") " pod="openstack/ovn-controller-rn98c-config-gnlxw" Oct 14 07:19:43 crc kubenswrapper[4870]: I1014 07:19:43.373695 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/7c8980f9-662b-46a7-a017-c1f038c3566c-additional-scripts\") pod \"ovn-controller-rn98c-config-gnlxw\" (UID: \"7c8980f9-662b-46a7-a017-c1f038c3566c\") " pod="openstack/ovn-controller-rn98c-config-gnlxw" Oct 14 07:19:43 crc kubenswrapper[4870]: I1014 07:19:43.375071 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7c8980f9-662b-46a7-a017-c1f038c3566c-scripts\") pod \"ovn-controller-rn98c-config-gnlxw\" (UID: \"7c8980f9-662b-46a7-a017-c1f038c3566c\") " pod="openstack/ovn-controller-rn98c-config-gnlxw" Oct 14 07:19:43 crc kubenswrapper[4870]: I1014 07:19:43.378147 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/7c8980f9-662b-46a7-a017-c1f038c3566c-var-log-ovn\") pod \"ovn-controller-rn98c-config-gnlxw\" (UID: \"7c8980f9-662b-46a7-a017-c1f038c3566c\") " pod="openstack/ovn-controller-rn98c-config-gnlxw" Oct 14 07:19:43 crc kubenswrapper[4870]: I1014 07:19:43.391558 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nxltb\" (UniqueName: \"kubernetes.io/projected/7c8980f9-662b-46a7-a017-c1f038c3566c-kube-api-access-nxltb\") pod \"ovn-controller-rn98c-config-gnlxw\" (UID: \"7c8980f9-662b-46a7-a017-c1f038c3566c\") " pod="openstack/ovn-controller-rn98c-config-gnlxw" Oct 14 07:19:43 crc kubenswrapper[4870]: I1014 07:19:43.533712 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-rn98c-config-gnlxw" Oct 14 07:19:43 crc kubenswrapper[4870]: I1014 07:19:43.811216 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cc794b75c-d2b9c" event={"ID":"69a02fec-1745-4827-9c90-d9dffb829a66","Type":"ContainerStarted","Data":"53081407fd75b324098b0ed4b06bd603bd305531ac1290feb40806718fcc0af4"} Oct 14 07:19:43 crc kubenswrapper[4870]: I1014 07:19:43.811769 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7cc794b75c-d2b9c" Oct 14 07:19:43 crc kubenswrapper[4870]: I1014 07:19:43.863382 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7cc794b75c-d2b9c" podStartSLOduration=2.863362388 podStartE2EDuration="2.863362388s" podCreationTimestamp="2025-10-14 07:19:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:19:43.861754728 +0000 UTC m=+1119.559115099" watchObservedRunningTime="2025-10-14 07:19:43.863362388 +0000 UTC m=+1119.560722759" Oct 14 07:19:44 crc kubenswrapper[4870]: I1014 07:19:44.025594 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-rn98c-config-gnlxw"] Oct 14 07:19:45 crc kubenswrapper[4870]: I1014 07:19:45.046087 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="631522fc-993b-496a-8ebb-03b289e5cc7a" path="/var/lib/kubelet/pods/631522fc-993b-496a-8ebb-03b289e5cc7a/volumes" Oct 14 07:19:47 crc kubenswrapper[4870]: I1014 07:19:47.450884 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Oct 14 07:19:47 crc kubenswrapper[4870]: I1014 07:19:47.898507 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-vvgl7"] Oct 14 07:19:47 crc kubenswrapper[4870]: I1014 07:19:47.899722 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-vvgl7" Oct 14 07:19:47 crc kubenswrapper[4870]: I1014 07:19:47.907504 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-vvgl7"] Oct 14 07:19:47 crc kubenswrapper[4870]: I1014 07:19:47.988769 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-wg6ln"] Oct 14 07:19:47 crc kubenswrapper[4870]: I1014 07:19:47.989875 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jwvv4\" (UniqueName: \"kubernetes.io/projected/67d40a7a-1bb5-457c-9738-b4a05378fee4-kube-api-access-jwvv4\") pod \"cinder-db-create-vvgl7\" (UID: \"67d40a7a-1bb5-457c-9738-b4a05378fee4\") " pod="openstack/cinder-db-create-vvgl7" Oct 14 07:19:47 crc kubenswrapper[4870]: I1014 07:19:47.989908 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-wg6ln" Oct 14 07:19:48 crc kubenswrapper[4870]: I1014 07:19:48.003261 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Oct 14 07:19:48 crc kubenswrapper[4870]: I1014 07:19:48.009245 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-wg6ln"] Oct 14 07:19:48 crc kubenswrapper[4870]: I1014 07:19:48.091032 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s7tlt\" (UniqueName: \"kubernetes.io/projected/ae296f6e-96fb-4e1d-bb24-4f4143558ee2-kube-api-access-s7tlt\") pod \"barbican-db-create-wg6ln\" (UID: \"ae296f6e-96fb-4e1d-bb24-4f4143558ee2\") " pod="openstack/barbican-db-create-wg6ln" Oct 14 07:19:48 crc kubenswrapper[4870]: I1014 07:19:48.091104 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jwvv4\" (UniqueName: \"kubernetes.io/projected/67d40a7a-1bb5-457c-9738-b4a05378fee4-kube-api-access-jwvv4\") pod \"cinder-db-create-vvgl7\" (UID: \"67d40a7a-1bb5-457c-9738-b4a05378fee4\") " pod="openstack/cinder-db-create-vvgl7" Oct 14 07:19:48 crc kubenswrapper[4870]: I1014 07:19:48.116265 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jwvv4\" (UniqueName: \"kubernetes.io/projected/67d40a7a-1bb5-457c-9738-b4a05378fee4-kube-api-access-jwvv4\") pod \"cinder-db-create-vvgl7\" (UID: \"67d40a7a-1bb5-457c-9738-b4a05378fee4\") " pod="openstack/cinder-db-create-vvgl7" Oct 14 07:19:48 crc kubenswrapper[4870]: I1014 07:19:48.145287 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-mwp7r"] Oct 14 07:19:48 crc kubenswrapper[4870]: I1014 07:19:48.146700 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-mwp7r" Oct 14 07:19:48 crc kubenswrapper[4870]: I1014 07:19:48.148695 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 14 07:19:48 crc kubenswrapper[4870]: I1014 07:19:48.149560 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 14 07:19:48 crc kubenswrapper[4870]: I1014 07:19:48.151841 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-cqggg" Oct 14 07:19:48 crc kubenswrapper[4870]: I1014 07:19:48.151966 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 14 07:19:48 crc kubenswrapper[4870]: I1014 07:19:48.169266 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-mwp7r"] Oct 14 07:19:48 crc kubenswrapper[4870]: I1014 07:19:48.192687 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s7tlt\" (UniqueName: \"kubernetes.io/projected/ae296f6e-96fb-4e1d-bb24-4f4143558ee2-kube-api-access-s7tlt\") pod \"barbican-db-create-wg6ln\" (UID: \"ae296f6e-96fb-4e1d-bb24-4f4143558ee2\") " pod="openstack/barbican-db-create-wg6ln" Oct 14 07:19:48 crc kubenswrapper[4870]: I1014 07:19:48.209193 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-m2g9b"] Oct 14 07:19:48 crc kubenswrapper[4870]: I1014 07:19:48.210341 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-m2g9b" Oct 14 07:19:48 crc kubenswrapper[4870]: I1014 07:19:48.218762 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-vvgl7" Oct 14 07:19:48 crc kubenswrapper[4870]: I1014 07:19:48.223100 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s7tlt\" (UniqueName: \"kubernetes.io/projected/ae296f6e-96fb-4e1d-bb24-4f4143558ee2-kube-api-access-s7tlt\") pod \"barbican-db-create-wg6ln\" (UID: \"ae296f6e-96fb-4e1d-bb24-4f4143558ee2\") " pod="openstack/barbican-db-create-wg6ln" Oct 14 07:19:48 crc kubenswrapper[4870]: I1014 07:19:48.226842 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-m2g9b"] Oct 14 07:19:48 crc kubenswrapper[4870]: I1014 07:19:48.294483 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-72bs9\" (UniqueName: \"kubernetes.io/projected/79dfe32a-c048-41d8-aead-de78ae909cd8-kube-api-access-72bs9\") pod \"keystone-db-sync-mwp7r\" (UID: \"79dfe32a-c048-41d8-aead-de78ae909cd8\") " pod="openstack/keystone-db-sync-mwp7r" Oct 14 07:19:48 crc kubenswrapper[4870]: I1014 07:19:48.294837 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/79dfe32a-c048-41d8-aead-de78ae909cd8-config-data\") pod \"keystone-db-sync-mwp7r\" (UID: \"79dfe32a-c048-41d8-aead-de78ae909cd8\") " pod="openstack/keystone-db-sync-mwp7r" Oct 14 07:19:48 crc kubenswrapper[4870]: I1014 07:19:48.294890 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79dfe32a-c048-41d8-aead-de78ae909cd8-combined-ca-bundle\") pod \"keystone-db-sync-mwp7r\" (UID: \"79dfe32a-c048-41d8-aead-de78ae909cd8\") " pod="openstack/keystone-db-sync-mwp7r" Oct 14 07:19:48 crc kubenswrapper[4870]: I1014 07:19:48.294909 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ltkkf\" (UniqueName: \"kubernetes.io/projected/7170a9ca-4322-4d2c-9550-ee587ec981db-kube-api-access-ltkkf\") pod \"neutron-db-create-m2g9b\" (UID: \"7170a9ca-4322-4d2c-9550-ee587ec981db\") " pod="openstack/neutron-db-create-m2g9b" Oct 14 07:19:48 crc kubenswrapper[4870]: I1014 07:19:48.307525 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-wg6ln" Oct 14 07:19:48 crc kubenswrapper[4870]: I1014 07:19:48.396122 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/79dfe32a-c048-41d8-aead-de78ae909cd8-config-data\") pod \"keystone-db-sync-mwp7r\" (UID: \"79dfe32a-c048-41d8-aead-de78ae909cd8\") " pod="openstack/keystone-db-sync-mwp7r" Oct 14 07:19:48 crc kubenswrapper[4870]: I1014 07:19:48.396196 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79dfe32a-c048-41d8-aead-de78ae909cd8-combined-ca-bundle\") pod \"keystone-db-sync-mwp7r\" (UID: \"79dfe32a-c048-41d8-aead-de78ae909cd8\") " pod="openstack/keystone-db-sync-mwp7r" Oct 14 07:19:48 crc kubenswrapper[4870]: I1014 07:19:48.396219 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ltkkf\" (UniqueName: \"kubernetes.io/projected/7170a9ca-4322-4d2c-9550-ee587ec981db-kube-api-access-ltkkf\") pod \"neutron-db-create-m2g9b\" (UID: \"7170a9ca-4322-4d2c-9550-ee587ec981db\") " pod="openstack/neutron-db-create-m2g9b" Oct 14 07:19:48 crc kubenswrapper[4870]: I1014 07:19:48.396263 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-72bs9\" (UniqueName: \"kubernetes.io/projected/79dfe32a-c048-41d8-aead-de78ae909cd8-kube-api-access-72bs9\") pod \"keystone-db-sync-mwp7r\" (UID: \"79dfe32a-c048-41d8-aead-de78ae909cd8\") " pod="openstack/keystone-db-sync-mwp7r" Oct 14 07:19:48 crc kubenswrapper[4870]: I1014 07:19:48.400095 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79dfe32a-c048-41d8-aead-de78ae909cd8-combined-ca-bundle\") pod \"keystone-db-sync-mwp7r\" (UID: \"79dfe32a-c048-41d8-aead-de78ae909cd8\") " pod="openstack/keystone-db-sync-mwp7r" Oct 14 07:19:48 crc kubenswrapper[4870]: I1014 07:19:48.402205 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/79dfe32a-c048-41d8-aead-de78ae909cd8-config-data\") pod \"keystone-db-sync-mwp7r\" (UID: \"79dfe32a-c048-41d8-aead-de78ae909cd8\") " pod="openstack/keystone-db-sync-mwp7r" Oct 14 07:19:48 crc kubenswrapper[4870]: I1014 07:19:48.410577 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-72bs9\" (UniqueName: \"kubernetes.io/projected/79dfe32a-c048-41d8-aead-de78ae909cd8-kube-api-access-72bs9\") pod \"keystone-db-sync-mwp7r\" (UID: \"79dfe32a-c048-41d8-aead-de78ae909cd8\") " pod="openstack/keystone-db-sync-mwp7r" Oct 14 07:19:48 crc kubenswrapper[4870]: I1014 07:19:48.412515 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ltkkf\" (UniqueName: \"kubernetes.io/projected/7170a9ca-4322-4d2c-9550-ee587ec981db-kube-api-access-ltkkf\") pod \"neutron-db-create-m2g9b\" (UID: \"7170a9ca-4322-4d2c-9550-ee587ec981db\") " pod="openstack/neutron-db-create-m2g9b" Oct 14 07:19:48 crc kubenswrapper[4870]: I1014 07:19:48.476678 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-mwp7r" Oct 14 07:19:48 crc kubenswrapper[4870]: I1014 07:19:48.565580 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-m2g9b" Oct 14 07:19:51 crc kubenswrapper[4870]: I1014 07:19:51.497674 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7cc794b75c-d2b9c" Oct 14 07:19:51 crc kubenswrapper[4870]: I1014 07:19:51.563383 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-549b48ff7f-vhsh8"] Oct 14 07:19:51 crc kubenswrapper[4870]: I1014 07:19:51.563610 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-549b48ff7f-vhsh8" podUID="17f2363b-7ac4-4e5d-92e0-0869ab39c98e" containerName="dnsmasq-dns" containerID="cri-o://c3d069abc429cccb2df1fa8e60ce1cfc0cc23b619edfc9bd8c05a1e88471f1bc" gracePeriod=10 Oct 14 07:19:51 crc kubenswrapper[4870]: I1014 07:19:51.879950 4870 generic.go:334] "Generic (PLEG): container finished" podID="17f2363b-7ac4-4e5d-92e0-0869ab39c98e" containerID="c3d069abc429cccb2df1fa8e60ce1cfc0cc23b619edfc9bd8c05a1e88471f1bc" exitCode=0 Oct 14 07:19:51 crc kubenswrapper[4870]: I1014 07:19:51.880003 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-549b48ff7f-vhsh8" event={"ID":"17f2363b-7ac4-4e5d-92e0-0869ab39c98e","Type":"ContainerDied","Data":"c3d069abc429cccb2df1fa8e60ce1cfc0cc23b619edfc9bd8c05a1e88471f1bc"} Oct 14 07:19:52 crc kubenswrapper[4870]: I1014 07:19:52.453162 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-549b48ff7f-vhsh8" Oct 14 07:19:52 crc kubenswrapper[4870]: I1014 07:19:52.566181 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/17f2363b-7ac4-4e5d-92e0-0869ab39c98e-dns-svc\") pod \"17f2363b-7ac4-4e5d-92e0-0869ab39c98e\" (UID: \"17f2363b-7ac4-4e5d-92e0-0869ab39c98e\") " Oct 14 07:19:52 crc kubenswrapper[4870]: I1014 07:19:52.566312 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/17f2363b-7ac4-4e5d-92e0-0869ab39c98e-config\") pod \"17f2363b-7ac4-4e5d-92e0-0869ab39c98e\" (UID: \"17f2363b-7ac4-4e5d-92e0-0869ab39c98e\") " Oct 14 07:19:52 crc kubenswrapper[4870]: I1014 07:19:52.566400 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ssrhl\" (UniqueName: \"kubernetes.io/projected/17f2363b-7ac4-4e5d-92e0-0869ab39c98e-kube-api-access-ssrhl\") pod \"17f2363b-7ac4-4e5d-92e0-0869ab39c98e\" (UID: \"17f2363b-7ac4-4e5d-92e0-0869ab39c98e\") " Oct 14 07:19:52 crc kubenswrapper[4870]: I1014 07:19:52.567477 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/17f2363b-7ac4-4e5d-92e0-0869ab39c98e-ovsdbserver-sb\") pod \"17f2363b-7ac4-4e5d-92e0-0869ab39c98e\" (UID: \"17f2363b-7ac4-4e5d-92e0-0869ab39c98e\") " Oct 14 07:19:52 crc kubenswrapper[4870]: I1014 07:19:52.567544 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/17f2363b-7ac4-4e5d-92e0-0869ab39c98e-ovsdbserver-nb\") pod \"17f2363b-7ac4-4e5d-92e0-0869ab39c98e\" (UID: \"17f2363b-7ac4-4e5d-92e0-0869ab39c98e\") " Oct 14 07:19:52 crc kubenswrapper[4870]: I1014 07:19:52.576411 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/17f2363b-7ac4-4e5d-92e0-0869ab39c98e-kube-api-access-ssrhl" (OuterVolumeSpecName: "kube-api-access-ssrhl") pod "17f2363b-7ac4-4e5d-92e0-0869ab39c98e" (UID: "17f2363b-7ac4-4e5d-92e0-0869ab39c98e"). InnerVolumeSpecName "kube-api-access-ssrhl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:19:52 crc kubenswrapper[4870]: I1014 07:19:52.639044 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/17f2363b-7ac4-4e5d-92e0-0869ab39c98e-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "17f2363b-7ac4-4e5d-92e0-0869ab39c98e" (UID: "17f2363b-7ac4-4e5d-92e0-0869ab39c98e"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:19:52 crc kubenswrapper[4870]: I1014 07:19:52.648487 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/17f2363b-7ac4-4e5d-92e0-0869ab39c98e-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "17f2363b-7ac4-4e5d-92e0-0869ab39c98e" (UID: "17f2363b-7ac4-4e5d-92e0-0869ab39c98e"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:19:52 crc kubenswrapper[4870]: I1014 07:19:52.663795 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/17f2363b-7ac4-4e5d-92e0-0869ab39c98e-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "17f2363b-7ac4-4e5d-92e0-0869ab39c98e" (UID: "17f2363b-7ac4-4e5d-92e0-0869ab39c98e"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:19:52 crc kubenswrapper[4870]: I1014 07:19:52.669469 4870 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/17f2363b-7ac4-4e5d-92e0-0869ab39c98e-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 14 07:19:52 crc kubenswrapper[4870]: I1014 07:19:52.669612 4870 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/17f2363b-7ac4-4e5d-92e0-0869ab39c98e-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 14 07:19:52 crc kubenswrapper[4870]: I1014 07:19:52.669681 4870 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/17f2363b-7ac4-4e5d-92e0-0869ab39c98e-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 14 07:19:52 crc kubenswrapper[4870]: I1014 07:19:52.669749 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ssrhl\" (UniqueName: \"kubernetes.io/projected/17f2363b-7ac4-4e5d-92e0-0869ab39c98e-kube-api-access-ssrhl\") on node \"crc\" DevicePath \"\"" Oct 14 07:19:52 crc kubenswrapper[4870]: I1014 07:19:52.674343 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/17f2363b-7ac4-4e5d-92e0-0869ab39c98e-config" (OuterVolumeSpecName: "config") pod "17f2363b-7ac4-4e5d-92e0-0869ab39c98e" (UID: "17f2363b-7ac4-4e5d-92e0-0869ab39c98e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:19:52 crc kubenswrapper[4870]: I1014 07:19:52.771562 4870 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/17f2363b-7ac4-4e5d-92e0-0869ab39c98e-config\") on node \"crc\" DevicePath \"\"" Oct 14 07:19:52 crc kubenswrapper[4870]: I1014 07:19:52.803154 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-mwp7r"] Oct 14 07:19:52 crc kubenswrapper[4870]: I1014 07:19:52.936681 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-549b48ff7f-vhsh8" event={"ID":"17f2363b-7ac4-4e5d-92e0-0869ab39c98e","Type":"ContainerDied","Data":"d3928912276e8b221ca8e94547860c384bd7ea04738a14a504f17ed450a446b6"} Oct 14 07:19:52 crc kubenswrapper[4870]: I1014 07:19:52.936998 4870 scope.go:117] "RemoveContainer" containerID="c3d069abc429cccb2df1fa8e60ce1cfc0cc23b619edfc9bd8c05a1e88471f1bc" Oct 14 07:19:52 crc kubenswrapper[4870]: I1014 07:19:52.936688 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-549b48ff7f-vhsh8" Oct 14 07:19:52 crc kubenswrapper[4870]: I1014 07:19:52.939160 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-mwp7r" event={"ID":"79dfe32a-c048-41d8-aead-de78ae909cd8","Type":"ContainerStarted","Data":"b784bab685dc64999ccfe19511bcc3547f14731e33636b5c09b1a7d72e78377f"} Oct 14 07:19:52 crc kubenswrapper[4870]: I1014 07:19:52.940774 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-m2g9b"] Oct 14 07:19:52 crc kubenswrapper[4870]: I1014 07:19:52.941597 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-rn98c-config-gnlxw" event={"ID":"7c8980f9-662b-46a7-a017-c1f038c3566c","Type":"ContainerStarted","Data":"a83b32049ab0d1907715a8d4bb87cfcec31636757720f25ebb523546eaa5e523"} Oct 14 07:19:52 crc kubenswrapper[4870]: I1014 07:19:52.941627 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-rn98c-config-gnlxw" event={"ID":"7c8980f9-662b-46a7-a017-c1f038c3566c","Type":"ContainerStarted","Data":"ef2d3620001a907c5f01d6580517392bb2041dff4bcfd916cbe14f55899b4f60"} Oct 14 07:19:52 crc kubenswrapper[4870]: I1014 07:19:52.969593 4870 scope.go:117] "RemoveContainer" containerID="cf76b6c52610aa2316d561fd70d09c1ffa1e569cd5a36ff134d2a023b2154d84" Oct 14 07:19:52 crc kubenswrapper[4870]: I1014 07:19:52.970226 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-rn98c-config-gnlxw" podStartSLOduration=9.970208349 podStartE2EDuration="9.970208349s" podCreationTimestamp="2025-10-14 07:19:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:19:52.961159065 +0000 UTC m=+1128.658519436" watchObservedRunningTime="2025-10-14 07:19:52.970208349 +0000 UTC m=+1128.667568720" Oct 14 07:19:53 crc kubenswrapper[4870]: I1014 07:19:53.016818 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-wg6ln"] Oct 14 07:19:53 crc kubenswrapper[4870]: W1014 07:19:53.020173 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podae296f6e_96fb_4e1d_bb24_4f4143558ee2.slice/crio-a383d73a3a76c2f176be006386c833220e641fe6f3eb3b9e15908e3c79ef4995 WatchSource:0}: Error finding container a383d73a3a76c2f176be006386c833220e641fe6f3eb3b9e15908e3c79ef4995: Status 404 returned error can't find the container with id a383d73a3a76c2f176be006386c833220e641fe6f3eb3b9e15908e3c79ef4995 Oct 14 07:19:53 crc kubenswrapper[4870]: W1014 07:19:53.020984 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod67d40a7a_1bb5_457c_9738_b4a05378fee4.slice/crio-d8791b25293a379e095c6f6de3e8979fda7b8b319c8c6cbbc45a0df44ef633b2 WatchSource:0}: Error finding container d8791b25293a379e095c6f6de3e8979fda7b8b319c8c6cbbc45a0df44ef633b2: Status 404 returned error can't find the container with id d8791b25293a379e095c6f6de3e8979fda7b8b319c8c6cbbc45a0df44ef633b2 Oct 14 07:19:53 crc kubenswrapper[4870]: I1014 07:19:53.021820 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-549b48ff7f-vhsh8"] Oct 14 07:19:53 crc kubenswrapper[4870]: I1014 07:19:53.026621 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-549b48ff7f-vhsh8"] Oct 14 07:19:53 crc kubenswrapper[4870]: I1014 07:19:53.032460 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-vvgl7"] Oct 14 07:19:53 crc kubenswrapper[4870]: I1014 07:19:53.050677 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="17f2363b-7ac4-4e5d-92e0-0869ab39c98e" path="/var/lib/kubelet/pods/17f2363b-7ac4-4e5d-92e0-0869ab39c98e/volumes" Oct 14 07:19:53 crc kubenswrapper[4870]: I1014 07:19:53.951127 4870 generic.go:334] "Generic (PLEG): container finished" podID="7170a9ca-4322-4d2c-9550-ee587ec981db" containerID="1d7dbc12c4434da4d09988b87f042f614c4f32064d8308852b271f35c34e44d1" exitCode=0 Oct 14 07:19:53 crc kubenswrapper[4870]: I1014 07:19:53.951450 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-m2g9b" event={"ID":"7170a9ca-4322-4d2c-9550-ee587ec981db","Type":"ContainerDied","Data":"1d7dbc12c4434da4d09988b87f042f614c4f32064d8308852b271f35c34e44d1"} Oct 14 07:19:53 crc kubenswrapper[4870]: I1014 07:19:53.951477 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-m2g9b" event={"ID":"7170a9ca-4322-4d2c-9550-ee587ec981db","Type":"ContainerStarted","Data":"c0e90da599fd014addae80bd4e7fad254a7e9a91b40af3d0dc7a62a6a1041718"} Oct 14 07:19:53 crc kubenswrapper[4870]: I1014 07:19:53.954450 4870 generic.go:334] "Generic (PLEG): container finished" podID="ae296f6e-96fb-4e1d-bb24-4f4143558ee2" containerID="2c964df0978df33d463f733f9d5d9c9b3b2b0107067baa2d37c1979f7552ff66" exitCode=0 Oct 14 07:19:53 crc kubenswrapper[4870]: I1014 07:19:53.954502 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-wg6ln" event={"ID":"ae296f6e-96fb-4e1d-bb24-4f4143558ee2","Type":"ContainerDied","Data":"2c964df0978df33d463f733f9d5d9c9b3b2b0107067baa2d37c1979f7552ff66"} Oct 14 07:19:53 crc kubenswrapper[4870]: I1014 07:19:53.954519 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-wg6ln" event={"ID":"ae296f6e-96fb-4e1d-bb24-4f4143558ee2","Type":"ContainerStarted","Data":"a383d73a3a76c2f176be006386c833220e641fe6f3eb3b9e15908e3c79ef4995"} Oct 14 07:19:53 crc kubenswrapper[4870]: I1014 07:19:53.957249 4870 generic.go:334] "Generic (PLEG): container finished" podID="7c8980f9-662b-46a7-a017-c1f038c3566c" containerID="a83b32049ab0d1907715a8d4bb87cfcec31636757720f25ebb523546eaa5e523" exitCode=0 Oct 14 07:19:53 crc kubenswrapper[4870]: I1014 07:19:53.957313 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-rn98c-config-gnlxw" event={"ID":"7c8980f9-662b-46a7-a017-c1f038c3566c","Type":"ContainerDied","Data":"a83b32049ab0d1907715a8d4bb87cfcec31636757720f25ebb523546eaa5e523"} Oct 14 07:19:53 crc kubenswrapper[4870]: I1014 07:19:53.958948 4870 generic.go:334] "Generic (PLEG): container finished" podID="67d40a7a-1bb5-457c-9738-b4a05378fee4" containerID="b1e3d02e0b9f1f551752996b576a2919f1585fce83c7b33a6bc6758f9bf28cc1" exitCode=0 Oct 14 07:19:53 crc kubenswrapper[4870]: I1014 07:19:53.958983 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-vvgl7" event={"ID":"67d40a7a-1bb5-457c-9738-b4a05378fee4","Type":"ContainerDied","Data":"b1e3d02e0b9f1f551752996b576a2919f1585fce83c7b33a6bc6758f9bf28cc1"} Oct 14 07:19:53 crc kubenswrapper[4870]: I1014 07:19:53.959030 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-vvgl7" event={"ID":"67d40a7a-1bb5-457c-9738-b4a05378fee4","Type":"ContainerStarted","Data":"d8791b25293a379e095c6f6de3e8979fda7b8b319c8c6cbbc45a0df44ef633b2"} Oct 14 07:19:53 crc kubenswrapper[4870]: I1014 07:19:53.960594 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-v2d79" event={"ID":"85aaeb67-e809-416d-b4e0-eb7605cdd307","Type":"ContainerStarted","Data":"6f74f7987dbeaadf7d1897858c5cff4d3ddcf7dce7dbc12f36e6c1f68826ee1d"} Oct 14 07:19:53 crc kubenswrapper[4870]: I1014 07:19:53.996115 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-v2d79" podStartSLOduration=2.449188704 podStartE2EDuration="17.996097865s" podCreationTimestamp="2025-10-14 07:19:36 +0000 UTC" firstStartedPulling="2025-10-14 07:19:36.868190641 +0000 UTC m=+1112.565551012" lastFinishedPulling="2025-10-14 07:19:52.415099792 +0000 UTC m=+1128.112460173" observedRunningTime="2025-10-14 07:19:53.989414039 +0000 UTC m=+1129.686774410" watchObservedRunningTime="2025-10-14 07:19:53.996097865 +0000 UTC m=+1129.693458236" Oct 14 07:19:58 crc kubenswrapper[4870]: I1014 07:19:58.012141 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-rn98c-config-gnlxw" Oct 14 07:19:58 crc kubenswrapper[4870]: I1014 07:19:58.016863 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-m2g9b" event={"ID":"7170a9ca-4322-4d2c-9550-ee587ec981db","Type":"ContainerDied","Data":"c0e90da599fd014addae80bd4e7fad254a7e9a91b40af3d0dc7a62a6a1041718"} Oct 14 07:19:58 crc kubenswrapper[4870]: I1014 07:19:58.016895 4870 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c0e90da599fd014addae80bd4e7fad254a7e9a91b40af3d0dc7a62a6a1041718" Oct 14 07:19:58 crc kubenswrapper[4870]: I1014 07:19:58.021474 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-vvgl7" Oct 14 07:19:58 crc kubenswrapper[4870]: I1014 07:19:58.021736 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-wg6ln" event={"ID":"ae296f6e-96fb-4e1d-bb24-4f4143558ee2","Type":"ContainerDied","Data":"a383d73a3a76c2f176be006386c833220e641fe6f3eb3b9e15908e3c79ef4995"} Oct 14 07:19:58 crc kubenswrapper[4870]: I1014 07:19:58.021766 4870 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a383d73a3a76c2f176be006386c833220e641fe6f3eb3b9e15908e3c79ef4995" Oct 14 07:19:58 crc kubenswrapper[4870]: I1014 07:19:58.023222 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-rn98c-config-gnlxw" event={"ID":"7c8980f9-662b-46a7-a017-c1f038c3566c","Type":"ContainerDied","Data":"ef2d3620001a907c5f01d6580517392bb2041dff4bcfd916cbe14f55899b4f60"} Oct 14 07:19:58 crc kubenswrapper[4870]: I1014 07:19:58.023250 4870 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ef2d3620001a907c5f01d6580517392bb2041dff4bcfd916cbe14f55899b4f60" Oct 14 07:19:58 crc kubenswrapper[4870]: I1014 07:19:58.023297 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-rn98c-config-gnlxw" Oct 14 07:19:58 crc kubenswrapper[4870]: I1014 07:19:58.024073 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-m2g9b" Oct 14 07:19:58 crc kubenswrapper[4870]: I1014 07:19:58.024483 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-vvgl7" event={"ID":"67d40a7a-1bb5-457c-9738-b4a05378fee4","Type":"ContainerDied","Data":"d8791b25293a379e095c6f6de3e8979fda7b8b319c8c6cbbc45a0df44ef633b2"} Oct 14 07:19:58 crc kubenswrapper[4870]: I1014 07:19:58.024508 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-vvgl7" Oct 14 07:19:58 crc kubenswrapper[4870]: I1014 07:19:58.024520 4870 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d8791b25293a379e095c6f6de3e8979fda7b8b319c8c6cbbc45a0df44ef633b2" Oct 14 07:19:58 crc kubenswrapper[4870]: I1014 07:19:58.031179 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-wg6ln" Oct 14 07:19:58 crc kubenswrapper[4870]: I1014 07:19:58.172129 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s7tlt\" (UniqueName: \"kubernetes.io/projected/ae296f6e-96fb-4e1d-bb24-4f4143558ee2-kube-api-access-s7tlt\") pod \"ae296f6e-96fb-4e1d-bb24-4f4143558ee2\" (UID: \"ae296f6e-96fb-4e1d-bb24-4f4143558ee2\") " Oct 14 07:19:58 crc kubenswrapper[4870]: I1014 07:19:58.172226 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/7c8980f9-662b-46a7-a017-c1f038c3566c-additional-scripts\") pod \"7c8980f9-662b-46a7-a017-c1f038c3566c\" (UID: \"7c8980f9-662b-46a7-a017-c1f038c3566c\") " Oct 14 07:19:58 crc kubenswrapper[4870]: I1014 07:19:58.172270 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/7c8980f9-662b-46a7-a017-c1f038c3566c-var-run-ovn\") pod \"7c8980f9-662b-46a7-a017-c1f038c3566c\" (UID: \"7c8980f9-662b-46a7-a017-c1f038c3566c\") " Oct 14 07:19:58 crc kubenswrapper[4870]: I1014 07:19:58.172356 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jwvv4\" (UniqueName: \"kubernetes.io/projected/67d40a7a-1bb5-457c-9738-b4a05378fee4-kube-api-access-jwvv4\") pod \"67d40a7a-1bb5-457c-9738-b4a05378fee4\" (UID: \"67d40a7a-1bb5-457c-9738-b4a05378fee4\") " Oct 14 07:19:58 crc kubenswrapper[4870]: I1014 07:19:58.172417 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/7c8980f9-662b-46a7-a017-c1f038c3566c-var-run\") pod \"7c8980f9-662b-46a7-a017-c1f038c3566c\" (UID: \"7c8980f9-662b-46a7-a017-c1f038c3566c\") " Oct 14 07:19:58 crc kubenswrapper[4870]: I1014 07:19:58.172453 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7c8980f9-662b-46a7-a017-c1f038c3566c-scripts\") pod \"7c8980f9-662b-46a7-a017-c1f038c3566c\" (UID: \"7c8980f9-662b-46a7-a017-c1f038c3566c\") " Oct 14 07:19:58 crc kubenswrapper[4870]: I1014 07:19:58.172478 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ltkkf\" (UniqueName: \"kubernetes.io/projected/7170a9ca-4322-4d2c-9550-ee587ec981db-kube-api-access-ltkkf\") pod \"7170a9ca-4322-4d2c-9550-ee587ec981db\" (UID: \"7170a9ca-4322-4d2c-9550-ee587ec981db\") " Oct 14 07:19:58 crc kubenswrapper[4870]: I1014 07:19:58.172506 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/7c8980f9-662b-46a7-a017-c1f038c3566c-var-log-ovn\") pod \"7c8980f9-662b-46a7-a017-c1f038c3566c\" (UID: \"7c8980f9-662b-46a7-a017-c1f038c3566c\") " Oct 14 07:19:58 crc kubenswrapper[4870]: I1014 07:19:58.172507 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7c8980f9-662b-46a7-a017-c1f038c3566c-var-run" (OuterVolumeSpecName: "var-run") pod "7c8980f9-662b-46a7-a017-c1f038c3566c" (UID: "7c8980f9-662b-46a7-a017-c1f038c3566c"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 14 07:19:58 crc kubenswrapper[4870]: I1014 07:19:58.172566 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nxltb\" (UniqueName: \"kubernetes.io/projected/7c8980f9-662b-46a7-a017-c1f038c3566c-kube-api-access-nxltb\") pod \"7c8980f9-662b-46a7-a017-c1f038c3566c\" (UID: \"7c8980f9-662b-46a7-a017-c1f038c3566c\") " Oct 14 07:19:58 crc kubenswrapper[4870]: I1014 07:19:58.172811 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7c8980f9-662b-46a7-a017-c1f038c3566c-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "7c8980f9-662b-46a7-a017-c1f038c3566c" (UID: "7c8980f9-662b-46a7-a017-c1f038c3566c"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 14 07:19:58 crc kubenswrapper[4870]: I1014 07:19:58.172988 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7c8980f9-662b-46a7-a017-c1f038c3566c-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "7c8980f9-662b-46a7-a017-c1f038c3566c" (UID: "7c8980f9-662b-46a7-a017-c1f038c3566c"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:19:58 crc kubenswrapper[4870]: I1014 07:19:58.173292 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7c8980f9-662b-46a7-a017-c1f038c3566c-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "7c8980f9-662b-46a7-a017-c1f038c3566c" (UID: "7c8980f9-662b-46a7-a017-c1f038c3566c"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 14 07:19:58 crc kubenswrapper[4870]: I1014 07:19:58.174253 4870 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/7c8980f9-662b-46a7-a017-c1f038c3566c-var-run-ovn\") on node \"crc\" DevicePath \"\"" Oct 14 07:19:58 crc kubenswrapper[4870]: I1014 07:19:58.174277 4870 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/7c8980f9-662b-46a7-a017-c1f038c3566c-var-run\") on node \"crc\" DevicePath \"\"" Oct 14 07:19:58 crc kubenswrapper[4870]: I1014 07:19:58.174287 4870 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/7c8980f9-662b-46a7-a017-c1f038c3566c-var-log-ovn\") on node \"crc\" DevicePath \"\"" Oct 14 07:19:58 crc kubenswrapper[4870]: I1014 07:19:58.174297 4870 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/7c8980f9-662b-46a7-a017-c1f038c3566c-additional-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 07:19:58 crc kubenswrapper[4870]: I1014 07:19:58.174867 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7c8980f9-662b-46a7-a017-c1f038c3566c-scripts" (OuterVolumeSpecName: "scripts") pod "7c8980f9-662b-46a7-a017-c1f038c3566c" (UID: "7c8980f9-662b-46a7-a017-c1f038c3566c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:19:58 crc kubenswrapper[4870]: I1014 07:19:58.176583 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7c8980f9-662b-46a7-a017-c1f038c3566c-kube-api-access-nxltb" (OuterVolumeSpecName: "kube-api-access-nxltb") pod "7c8980f9-662b-46a7-a017-c1f038c3566c" (UID: "7c8980f9-662b-46a7-a017-c1f038c3566c"). InnerVolumeSpecName "kube-api-access-nxltb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:19:58 crc kubenswrapper[4870]: I1014 07:19:58.176739 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/67d40a7a-1bb5-457c-9738-b4a05378fee4-kube-api-access-jwvv4" (OuterVolumeSpecName: "kube-api-access-jwvv4") pod "67d40a7a-1bb5-457c-9738-b4a05378fee4" (UID: "67d40a7a-1bb5-457c-9738-b4a05378fee4"). InnerVolumeSpecName "kube-api-access-jwvv4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:19:58 crc kubenswrapper[4870]: I1014 07:19:58.176847 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7170a9ca-4322-4d2c-9550-ee587ec981db-kube-api-access-ltkkf" (OuterVolumeSpecName: "kube-api-access-ltkkf") pod "7170a9ca-4322-4d2c-9550-ee587ec981db" (UID: "7170a9ca-4322-4d2c-9550-ee587ec981db"). InnerVolumeSpecName "kube-api-access-ltkkf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:19:58 crc kubenswrapper[4870]: I1014 07:19:58.178344 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ae296f6e-96fb-4e1d-bb24-4f4143558ee2-kube-api-access-s7tlt" (OuterVolumeSpecName: "kube-api-access-s7tlt") pod "ae296f6e-96fb-4e1d-bb24-4f4143558ee2" (UID: "ae296f6e-96fb-4e1d-bb24-4f4143558ee2"). InnerVolumeSpecName "kube-api-access-s7tlt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:19:58 crc kubenswrapper[4870]: I1014 07:19:58.276093 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jwvv4\" (UniqueName: \"kubernetes.io/projected/67d40a7a-1bb5-457c-9738-b4a05378fee4-kube-api-access-jwvv4\") on node \"crc\" DevicePath \"\"" Oct 14 07:19:58 crc kubenswrapper[4870]: I1014 07:19:58.276126 4870 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7c8980f9-662b-46a7-a017-c1f038c3566c-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 07:19:58 crc kubenswrapper[4870]: I1014 07:19:58.276136 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ltkkf\" (UniqueName: \"kubernetes.io/projected/7170a9ca-4322-4d2c-9550-ee587ec981db-kube-api-access-ltkkf\") on node \"crc\" DevicePath \"\"" Oct 14 07:19:58 crc kubenswrapper[4870]: I1014 07:19:58.276145 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nxltb\" (UniqueName: \"kubernetes.io/projected/7c8980f9-662b-46a7-a017-c1f038c3566c-kube-api-access-nxltb\") on node \"crc\" DevicePath \"\"" Oct 14 07:19:58 crc kubenswrapper[4870]: I1014 07:19:58.276153 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s7tlt\" (UniqueName: \"kubernetes.io/projected/ae296f6e-96fb-4e1d-bb24-4f4143558ee2-kube-api-access-s7tlt\") on node \"crc\" DevicePath \"\"" Oct 14 07:19:59 crc kubenswrapper[4870]: I1014 07:19:59.035896 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-wg6ln" Oct 14 07:19:59 crc kubenswrapper[4870]: I1014 07:19:59.039609 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-m2g9b" Oct 14 07:19:59 crc kubenswrapper[4870]: I1014 07:19:59.062559 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-mwp7r" event={"ID":"79dfe32a-c048-41d8-aead-de78ae909cd8","Type":"ContainerStarted","Data":"9e34fe756d7db40d6ed7af9609859cb38ea6bbd896e9819ab8153c17bb627435"} Oct 14 07:19:59 crc kubenswrapper[4870]: I1014 07:19:59.080700 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-mwp7r" podStartSLOduration=6.085617549 podStartE2EDuration="11.080672373s" podCreationTimestamp="2025-10-14 07:19:48 +0000 UTC" firstStartedPulling="2025-10-14 07:19:52.889598694 +0000 UTC m=+1128.586959065" lastFinishedPulling="2025-10-14 07:19:57.884653518 +0000 UTC m=+1133.582013889" observedRunningTime="2025-10-14 07:19:59.066837251 +0000 UTC m=+1134.764197632" watchObservedRunningTime="2025-10-14 07:19:59.080672373 +0000 UTC m=+1134.778032744" Oct 14 07:19:59 crc kubenswrapper[4870]: I1014 07:19:59.147286 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-rn98c-config-gnlxw"] Oct 14 07:19:59 crc kubenswrapper[4870]: I1014 07:19:59.161631 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-rn98c-config-gnlxw"] Oct 14 07:20:00 crc kubenswrapper[4870]: I1014 07:20:00.049592 4870 generic.go:334] "Generic (PLEG): container finished" podID="85aaeb67-e809-416d-b4e0-eb7605cdd307" containerID="6f74f7987dbeaadf7d1897858c5cff4d3ddcf7dce7dbc12f36e6c1f68826ee1d" exitCode=0 Oct 14 07:20:00 crc kubenswrapper[4870]: I1014 07:20:00.049686 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-v2d79" event={"ID":"85aaeb67-e809-416d-b4e0-eb7605cdd307","Type":"ContainerDied","Data":"6f74f7987dbeaadf7d1897858c5cff4d3ddcf7dce7dbc12f36e6c1f68826ee1d"} Oct 14 07:20:01 crc kubenswrapper[4870]: I1014 07:20:01.060383 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7c8980f9-662b-46a7-a017-c1f038c3566c" path="/var/lib/kubelet/pods/7c8980f9-662b-46a7-a017-c1f038c3566c/volumes" Oct 14 07:20:01 crc kubenswrapper[4870]: I1014 07:20:01.539426 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-v2d79" Oct 14 07:20:01 crc kubenswrapper[4870]: I1014 07:20:01.636402 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/85aaeb67-e809-416d-b4e0-eb7605cdd307-db-sync-config-data\") pod \"85aaeb67-e809-416d-b4e0-eb7605cdd307\" (UID: \"85aaeb67-e809-416d-b4e0-eb7605cdd307\") " Oct 14 07:20:01 crc kubenswrapper[4870]: I1014 07:20:01.636722 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x97n2\" (UniqueName: \"kubernetes.io/projected/85aaeb67-e809-416d-b4e0-eb7605cdd307-kube-api-access-x97n2\") pod \"85aaeb67-e809-416d-b4e0-eb7605cdd307\" (UID: \"85aaeb67-e809-416d-b4e0-eb7605cdd307\") " Oct 14 07:20:01 crc kubenswrapper[4870]: I1014 07:20:01.636773 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/85aaeb67-e809-416d-b4e0-eb7605cdd307-config-data\") pod \"85aaeb67-e809-416d-b4e0-eb7605cdd307\" (UID: \"85aaeb67-e809-416d-b4e0-eb7605cdd307\") " Oct 14 07:20:01 crc kubenswrapper[4870]: I1014 07:20:01.636798 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85aaeb67-e809-416d-b4e0-eb7605cdd307-combined-ca-bundle\") pod \"85aaeb67-e809-416d-b4e0-eb7605cdd307\" (UID: \"85aaeb67-e809-416d-b4e0-eb7605cdd307\") " Oct 14 07:20:01 crc kubenswrapper[4870]: I1014 07:20:01.643979 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/85aaeb67-e809-416d-b4e0-eb7605cdd307-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "85aaeb67-e809-416d-b4e0-eb7605cdd307" (UID: "85aaeb67-e809-416d-b4e0-eb7605cdd307"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:20:01 crc kubenswrapper[4870]: I1014 07:20:01.644594 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/85aaeb67-e809-416d-b4e0-eb7605cdd307-kube-api-access-x97n2" (OuterVolumeSpecName: "kube-api-access-x97n2") pod "85aaeb67-e809-416d-b4e0-eb7605cdd307" (UID: "85aaeb67-e809-416d-b4e0-eb7605cdd307"). InnerVolumeSpecName "kube-api-access-x97n2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:20:01 crc kubenswrapper[4870]: I1014 07:20:01.664473 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/85aaeb67-e809-416d-b4e0-eb7605cdd307-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "85aaeb67-e809-416d-b4e0-eb7605cdd307" (UID: "85aaeb67-e809-416d-b4e0-eb7605cdd307"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:20:01 crc kubenswrapper[4870]: I1014 07:20:01.693138 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/85aaeb67-e809-416d-b4e0-eb7605cdd307-config-data" (OuterVolumeSpecName: "config-data") pod "85aaeb67-e809-416d-b4e0-eb7605cdd307" (UID: "85aaeb67-e809-416d-b4e0-eb7605cdd307"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:20:01 crc kubenswrapper[4870]: I1014 07:20:01.738542 4870 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/85aaeb67-e809-416d-b4e0-eb7605cdd307-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 07:20:01 crc kubenswrapper[4870]: I1014 07:20:01.738580 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x97n2\" (UniqueName: \"kubernetes.io/projected/85aaeb67-e809-416d-b4e0-eb7605cdd307-kube-api-access-x97n2\") on node \"crc\" DevicePath \"\"" Oct 14 07:20:01 crc kubenswrapper[4870]: I1014 07:20:01.738595 4870 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/85aaeb67-e809-416d-b4e0-eb7605cdd307-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 07:20:01 crc kubenswrapper[4870]: I1014 07:20:01.738604 4870 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85aaeb67-e809-416d-b4e0-eb7605cdd307-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 07:20:02 crc kubenswrapper[4870]: I1014 07:20:02.086494 4870 generic.go:334] "Generic (PLEG): container finished" podID="79dfe32a-c048-41d8-aead-de78ae909cd8" containerID="9e34fe756d7db40d6ed7af9609859cb38ea6bbd896e9819ab8153c17bb627435" exitCode=0 Oct 14 07:20:02 crc kubenswrapper[4870]: I1014 07:20:02.086620 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-mwp7r" event={"ID":"79dfe32a-c048-41d8-aead-de78ae909cd8","Type":"ContainerDied","Data":"9e34fe756d7db40d6ed7af9609859cb38ea6bbd896e9819ab8153c17bb627435"} Oct 14 07:20:02 crc kubenswrapper[4870]: I1014 07:20:02.088794 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-v2d79" event={"ID":"85aaeb67-e809-416d-b4e0-eb7605cdd307","Type":"ContainerDied","Data":"b45d93a9f7cd85e1331326b27328fbb7859749e534da22dfff45f8a22d445a33"} Oct 14 07:20:02 crc kubenswrapper[4870]: I1014 07:20:02.088839 4870 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b45d93a9f7cd85e1331326b27328fbb7859749e534da22dfff45f8a22d445a33" Oct 14 07:20:02 crc kubenswrapper[4870]: I1014 07:20:02.088893 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-v2d79" Oct 14 07:20:02 crc kubenswrapper[4870]: I1014 07:20:02.577651 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-cb67568b9-c75vp"] Oct 14 07:20:02 crc kubenswrapper[4870]: E1014 07:20:02.578045 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85aaeb67-e809-416d-b4e0-eb7605cdd307" containerName="glance-db-sync" Oct 14 07:20:02 crc kubenswrapper[4870]: I1014 07:20:02.578061 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="85aaeb67-e809-416d-b4e0-eb7605cdd307" containerName="glance-db-sync" Oct 14 07:20:02 crc kubenswrapper[4870]: E1014 07:20:02.578080 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae296f6e-96fb-4e1d-bb24-4f4143558ee2" containerName="mariadb-database-create" Oct 14 07:20:02 crc kubenswrapper[4870]: I1014 07:20:02.578088 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae296f6e-96fb-4e1d-bb24-4f4143558ee2" containerName="mariadb-database-create" Oct 14 07:20:02 crc kubenswrapper[4870]: E1014 07:20:02.578103 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c8980f9-662b-46a7-a017-c1f038c3566c" containerName="ovn-config" Oct 14 07:20:02 crc kubenswrapper[4870]: I1014 07:20:02.578109 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c8980f9-662b-46a7-a017-c1f038c3566c" containerName="ovn-config" Oct 14 07:20:02 crc kubenswrapper[4870]: E1014 07:20:02.578123 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17f2363b-7ac4-4e5d-92e0-0869ab39c98e" containerName="dnsmasq-dns" Oct 14 07:20:02 crc kubenswrapper[4870]: I1014 07:20:02.578128 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="17f2363b-7ac4-4e5d-92e0-0869ab39c98e" containerName="dnsmasq-dns" Oct 14 07:20:02 crc kubenswrapper[4870]: E1014 07:20:02.578141 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7170a9ca-4322-4d2c-9550-ee587ec981db" containerName="mariadb-database-create" Oct 14 07:20:02 crc kubenswrapper[4870]: I1014 07:20:02.578151 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="7170a9ca-4322-4d2c-9550-ee587ec981db" containerName="mariadb-database-create" Oct 14 07:20:02 crc kubenswrapper[4870]: E1014 07:20:02.578165 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17f2363b-7ac4-4e5d-92e0-0869ab39c98e" containerName="init" Oct 14 07:20:02 crc kubenswrapper[4870]: I1014 07:20:02.578171 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="17f2363b-7ac4-4e5d-92e0-0869ab39c98e" containerName="init" Oct 14 07:20:02 crc kubenswrapper[4870]: E1014 07:20:02.578187 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67d40a7a-1bb5-457c-9738-b4a05378fee4" containerName="mariadb-database-create" Oct 14 07:20:02 crc kubenswrapper[4870]: I1014 07:20:02.578193 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="67d40a7a-1bb5-457c-9738-b4a05378fee4" containerName="mariadb-database-create" Oct 14 07:20:02 crc kubenswrapper[4870]: I1014 07:20:02.578339 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="85aaeb67-e809-416d-b4e0-eb7605cdd307" containerName="glance-db-sync" Oct 14 07:20:02 crc kubenswrapper[4870]: I1014 07:20:02.578353 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="17f2363b-7ac4-4e5d-92e0-0869ab39c98e" containerName="dnsmasq-dns" Oct 14 07:20:02 crc kubenswrapper[4870]: I1014 07:20:02.578364 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="ae296f6e-96fb-4e1d-bb24-4f4143558ee2" containerName="mariadb-database-create" Oct 14 07:20:02 crc kubenswrapper[4870]: I1014 07:20:02.578376 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="7170a9ca-4322-4d2c-9550-ee587ec981db" containerName="mariadb-database-create" Oct 14 07:20:02 crc kubenswrapper[4870]: I1014 07:20:02.578385 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="67d40a7a-1bb5-457c-9738-b4a05378fee4" containerName="mariadb-database-create" Oct 14 07:20:02 crc kubenswrapper[4870]: I1014 07:20:02.578397 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="7c8980f9-662b-46a7-a017-c1f038c3566c" containerName="ovn-config" Oct 14 07:20:02 crc kubenswrapper[4870]: I1014 07:20:02.579221 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cb67568b9-c75vp" Oct 14 07:20:02 crc kubenswrapper[4870]: I1014 07:20:02.587021 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-cb67568b9-c75vp"] Oct 14 07:20:02 crc kubenswrapper[4870]: I1014 07:20:02.654375 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-plgv5\" (UniqueName: \"kubernetes.io/projected/3bb9dec0-1879-419a-ab06-4fbdf87f75af-kube-api-access-plgv5\") pod \"dnsmasq-dns-cb67568b9-c75vp\" (UID: \"3bb9dec0-1879-419a-ab06-4fbdf87f75af\") " pod="openstack/dnsmasq-dns-cb67568b9-c75vp" Oct 14 07:20:02 crc kubenswrapper[4870]: I1014 07:20:02.654498 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3bb9dec0-1879-419a-ab06-4fbdf87f75af-config\") pod \"dnsmasq-dns-cb67568b9-c75vp\" (UID: \"3bb9dec0-1879-419a-ab06-4fbdf87f75af\") " pod="openstack/dnsmasq-dns-cb67568b9-c75vp" Oct 14 07:20:02 crc kubenswrapper[4870]: I1014 07:20:02.654523 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3bb9dec0-1879-419a-ab06-4fbdf87f75af-dns-swift-storage-0\") pod \"dnsmasq-dns-cb67568b9-c75vp\" (UID: \"3bb9dec0-1879-419a-ab06-4fbdf87f75af\") " pod="openstack/dnsmasq-dns-cb67568b9-c75vp" Oct 14 07:20:02 crc kubenswrapper[4870]: I1014 07:20:02.654550 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3bb9dec0-1879-419a-ab06-4fbdf87f75af-ovsdbserver-sb\") pod \"dnsmasq-dns-cb67568b9-c75vp\" (UID: \"3bb9dec0-1879-419a-ab06-4fbdf87f75af\") " pod="openstack/dnsmasq-dns-cb67568b9-c75vp" Oct 14 07:20:02 crc kubenswrapper[4870]: I1014 07:20:02.654569 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3bb9dec0-1879-419a-ab06-4fbdf87f75af-ovsdbserver-nb\") pod \"dnsmasq-dns-cb67568b9-c75vp\" (UID: \"3bb9dec0-1879-419a-ab06-4fbdf87f75af\") " pod="openstack/dnsmasq-dns-cb67568b9-c75vp" Oct 14 07:20:02 crc kubenswrapper[4870]: I1014 07:20:02.654586 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3bb9dec0-1879-419a-ab06-4fbdf87f75af-dns-svc\") pod \"dnsmasq-dns-cb67568b9-c75vp\" (UID: \"3bb9dec0-1879-419a-ab06-4fbdf87f75af\") " pod="openstack/dnsmasq-dns-cb67568b9-c75vp" Oct 14 07:20:02 crc kubenswrapper[4870]: I1014 07:20:02.755889 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3bb9dec0-1879-419a-ab06-4fbdf87f75af-config\") pod \"dnsmasq-dns-cb67568b9-c75vp\" (UID: \"3bb9dec0-1879-419a-ab06-4fbdf87f75af\") " pod="openstack/dnsmasq-dns-cb67568b9-c75vp" Oct 14 07:20:02 crc kubenswrapper[4870]: I1014 07:20:02.756192 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3bb9dec0-1879-419a-ab06-4fbdf87f75af-dns-swift-storage-0\") pod \"dnsmasq-dns-cb67568b9-c75vp\" (UID: \"3bb9dec0-1879-419a-ab06-4fbdf87f75af\") " pod="openstack/dnsmasq-dns-cb67568b9-c75vp" Oct 14 07:20:02 crc kubenswrapper[4870]: I1014 07:20:02.756222 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3bb9dec0-1879-419a-ab06-4fbdf87f75af-ovsdbserver-sb\") pod \"dnsmasq-dns-cb67568b9-c75vp\" (UID: \"3bb9dec0-1879-419a-ab06-4fbdf87f75af\") " pod="openstack/dnsmasq-dns-cb67568b9-c75vp" Oct 14 07:20:02 crc kubenswrapper[4870]: I1014 07:20:02.756240 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3bb9dec0-1879-419a-ab06-4fbdf87f75af-ovsdbserver-nb\") pod \"dnsmasq-dns-cb67568b9-c75vp\" (UID: \"3bb9dec0-1879-419a-ab06-4fbdf87f75af\") " pod="openstack/dnsmasq-dns-cb67568b9-c75vp" Oct 14 07:20:02 crc kubenswrapper[4870]: I1014 07:20:02.756261 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3bb9dec0-1879-419a-ab06-4fbdf87f75af-dns-svc\") pod \"dnsmasq-dns-cb67568b9-c75vp\" (UID: \"3bb9dec0-1879-419a-ab06-4fbdf87f75af\") " pod="openstack/dnsmasq-dns-cb67568b9-c75vp" Oct 14 07:20:02 crc kubenswrapper[4870]: I1014 07:20:02.756322 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-plgv5\" (UniqueName: \"kubernetes.io/projected/3bb9dec0-1879-419a-ab06-4fbdf87f75af-kube-api-access-plgv5\") pod \"dnsmasq-dns-cb67568b9-c75vp\" (UID: \"3bb9dec0-1879-419a-ab06-4fbdf87f75af\") " pod="openstack/dnsmasq-dns-cb67568b9-c75vp" Oct 14 07:20:02 crc kubenswrapper[4870]: I1014 07:20:02.757255 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3bb9dec0-1879-419a-ab06-4fbdf87f75af-dns-swift-storage-0\") pod \"dnsmasq-dns-cb67568b9-c75vp\" (UID: \"3bb9dec0-1879-419a-ab06-4fbdf87f75af\") " pod="openstack/dnsmasq-dns-cb67568b9-c75vp" Oct 14 07:20:02 crc kubenswrapper[4870]: I1014 07:20:02.757345 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3bb9dec0-1879-419a-ab06-4fbdf87f75af-config\") pod \"dnsmasq-dns-cb67568b9-c75vp\" (UID: \"3bb9dec0-1879-419a-ab06-4fbdf87f75af\") " pod="openstack/dnsmasq-dns-cb67568b9-c75vp" Oct 14 07:20:02 crc kubenswrapper[4870]: I1014 07:20:02.757364 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3bb9dec0-1879-419a-ab06-4fbdf87f75af-ovsdbserver-sb\") pod \"dnsmasq-dns-cb67568b9-c75vp\" (UID: \"3bb9dec0-1879-419a-ab06-4fbdf87f75af\") " pod="openstack/dnsmasq-dns-cb67568b9-c75vp" Oct 14 07:20:02 crc kubenswrapper[4870]: I1014 07:20:02.757927 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3bb9dec0-1879-419a-ab06-4fbdf87f75af-ovsdbserver-nb\") pod \"dnsmasq-dns-cb67568b9-c75vp\" (UID: \"3bb9dec0-1879-419a-ab06-4fbdf87f75af\") " pod="openstack/dnsmasq-dns-cb67568b9-c75vp" Oct 14 07:20:02 crc kubenswrapper[4870]: I1014 07:20:02.758410 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3bb9dec0-1879-419a-ab06-4fbdf87f75af-dns-svc\") pod \"dnsmasq-dns-cb67568b9-c75vp\" (UID: \"3bb9dec0-1879-419a-ab06-4fbdf87f75af\") " pod="openstack/dnsmasq-dns-cb67568b9-c75vp" Oct 14 07:20:02 crc kubenswrapper[4870]: I1014 07:20:02.783730 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-plgv5\" (UniqueName: \"kubernetes.io/projected/3bb9dec0-1879-419a-ab06-4fbdf87f75af-kube-api-access-plgv5\") pod \"dnsmasq-dns-cb67568b9-c75vp\" (UID: \"3bb9dec0-1879-419a-ab06-4fbdf87f75af\") " pod="openstack/dnsmasq-dns-cb67568b9-c75vp" Oct 14 07:20:02 crc kubenswrapper[4870]: I1014 07:20:02.897103 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cb67568b9-c75vp" Oct 14 07:20:03 crc kubenswrapper[4870]: I1014 07:20:03.368001 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-cb67568b9-c75vp"] Oct 14 07:20:03 crc kubenswrapper[4870]: I1014 07:20:03.412848 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-mwp7r" Oct 14 07:20:03 crc kubenswrapper[4870]: I1014 07:20:03.578375 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/79dfe32a-c048-41d8-aead-de78ae909cd8-config-data\") pod \"79dfe32a-c048-41d8-aead-de78ae909cd8\" (UID: \"79dfe32a-c048-41d8-aead-de78ae909cd8\") " Oct 14 07:20:03 crc kubenswrapper[4870]: I1014 07:20:03.578517 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79dfe32a-c048-41d8-aead-de78ae909cd8-combined-ca-bundle\") pod \"79dfe32a-c048-41d8-aead-de78ae909cd8\" (UID: \"79dfe32a-c048-41d8-aead-de78ae909cd8\") " Oct 14 07:20:03 crc kubenswrapper[4870]: I1014 07:20:03.578569 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-72bs9\" (UniqueName: \"kubernetes.io/projected/79dfe32a-c048-41d8-aead-de78ae909cd8-kube-api-access-72bs9\") pod \"79dfe32a-c048-41d8-aead-de78ae909cd8\" (UID: \"79dfe32a-c048-41d8-aead-de78ae909cd8\") " Oct 14 07:20:03 crc kubenswrapper[4870]: I1014 07:20:03.594795 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/79dfe32a-c048-41d8-aead-de78ae909cd8-kube-api-access-72bs9" (OuterVolumeSpecName: "kube-api-access-72bs9") pod "79dfe32a-c048-41d8-aead-de78ae909cd8" (UID: "79dfe32a-c048-41d8-aead-de78ae909cd8"). InnerVolumeSpecName "kube-api-access-72bs9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:20:03 crc kubenswrapper[4870]: I1014 07:20:03.626056 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/79dfe32a-c048-41d8-aead-de78ae909cd8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "79dfe32a-c048-41d8-aead-de78ae909cd8" (UID: "79dfe32a-c048-41d8-aead-de78ae909cd8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:20:03 crc kubenswrapper[4870]: I1014 07:20:03.643579 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/79dfe32a-c048-41d8-aead-de78ae909cd8-config-data" (OuterVolumeSpecName: "config-data") pod "79dfe32a-c048-41d8-aead-de78ae909cd8" (UID: "79dfe32a-c048-41d8-aead-de78ae909cd8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:20:03 crc kubenswrapper[4870]: I1014 07:20:03.680729 4870 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79dfe32a-c048-41d8-aead-de78ae909cd8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 07:20:03 crc kubenswrapper[4870]: I1014 07:20:03.680762 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-72bs9\" (UniqueName: \"kubernetes.io/projected/79dfe32a-c048-41d8-aead-de78ae909cd8-kube-api-access-72bs9\") on node \"crc\" DevicePath \"\"" Oct 14 07:20:03 crc kubenswrapper[4870]: I1014 07:20:03.680776 4870 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/79dfe32a-c048-41d8-aead-de78ae909cd8-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 07:20:04 crc kubenswrapper[4870]: I1014 07:20:04.110908 4870 generic.go:334] "Generic (PLEG): container finished" podID="3bb9dec0-1879-419a-ab06-4fbdf87f75af" containerID="eae46e7838df613139fc0c847c23394c91965a2d5b00fea5ae6a4be757b4138a" exitCode=0 Oct 14 07:20:04 crc kubenswrapper[4870]: I1014 07:20:04.110962 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cb67568b9-c75vp" event={"ID":"3bb9dec0-1879-419a-ab06-4fbdf87f75af","Type":"ContainerDied","Data":"eae46e7838df613139fc0c847c23394c91965a2d5b00fea5ae6a4be757b4138a"} Oct 14 07:20:04 crc kubenswrapper[4870]: I1014 07:20:04.111347 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cb67568b9-c75vp" event={"ID":"3bb9dec0-1879-419a-ab06-4fbdf87f75af","Type":"ContainerStarted","Data":"dbc1f6adb0e4395d7859cff4dd784d12ee5763e3bd8807bd26f1be8f4fd87153"} Oct 14 07:20:04 crc kubenswrapper[4870]: I1014 07:20:04.113888 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-mwp7r" event={"ID":"79dfe32a-c048-41d8-aead-de78ae909cd8","Type":"ContainerDied","Data":"b784bab685dc64999ccfe19511bcc3547f14731e33636b5c09b1a7d72e78377f"} Oct 14 07:20:04 crc kubenswrapper[4870]: I1014 07:20:04.113915 4870 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b784bab685dc64999ccfe19511bcc3547f14731e33636b5c09b1a7d72e78377f" Oct 14 07:20:04 crc kubenswrapper[4870]: I1014 07:20:04.113975 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-mwp7r" Oct 14 07:20:04 crc kubenswrapper[4870]: I1014 07:20:04.392402 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-cb67568b9-c75vp"] Oct 14 07:20:04 crc kubenswrapper[4870]: I1014 07:20:04.402146 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-s2xm6"] Oct 14 07:20:04 crc kubenswrapper[4870]: E1014 07:20:04.402534 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="79dfe32a-c048-41d8-aead-de78ae909cd8" containerName="keystone-db-sync" Oct 14 07:20:04 crc kubenswrapper[4870]: I1014 07:20:04.402546 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="79dfe32a-c048-41d8-aead-de78ae909cd8" containerName="keystone-db-sync" Oct 14 07:20:04 crc kubenswrapper[4870]: I1014 07:20:04.402742 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="79dfe32a-c048-41d8-aead-de78ae909cd8" containerName="keystone-db-sync" Oct 14 07:20:04 crc kubenswrapper[4870]: I1014 07:20:04.403322 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-s2xm6" Oct 14 07:20:04 crc kubenswrapper[4870]: I1014 07:20:04.408050 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 14 07:20:04 crc kubenswrapper[4870]: I1014 07:20:04.408269 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-cqggg" Oct 14 07:20:04 crc kubenswrapper[4870]: I1014 07:20:04.408365 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 14 07:20:04 crc kubenswrapper[4870]: I1014 07:20:04.408836 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 14 07:20:04 crc kubenswrapper[4870]: I1014 07:20:04.424604 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-s2xm6"] Oct 14 07:20:04 crc kubenswrapper[4870]: I1014 07:20:04.467612 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5885bcf447-68v4h"] Oct 14 07:20:04 crc kubenswrapper[4870]: I1014 07:20:04.469869 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5885bcf447-68v4h" Oct 14 07:20:04 crc kubenswrapper[4870]: I1014 07:20:04.476409 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5885bcf447-68v4h"] Oct 14 07:20:04 crc kubenswrapper[4870]: I1014 07:20:04.597380 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d94da551-da6b-48b1-ac7d-a78f06715651-dns-swift-storage-0\") pod \"dnsmasq-dns-5885bcf447-68v4h\" (UID: \"d94da551-da6b-48b1-ac7d-a78f06715651\") " pod="openstack/dnsmasq-dns-5885bcf447-68v4h" Oct 14 07:20:04 crc kubenswrapper[4870]: I1014 07:20:04.597450 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d5f827b-9c3f-4d17-b6a6-506fe3d2e06f-combined-ca-bundle\") pod \"keystone-bootstrap-s2xm6\" (UID: \"6d5f827b-9c3f-4d17-b6a6-506fe3d2e06f\") " pod="openstack/keystone-bootstrap-s2xm6" Oct 14 07:20:04 crc kubenswrapper[4870]: I1014 07:20:04.597479 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d94da551-da6b-48b1-ac7d-a78f06715651-ovsdbserver-sb\") pod \"dnsmasq-dns-5885bcf447-68v4h\" (UID: \"d94da551-da6b-48b1-ac7d-a78f06715651\") " pod="openstack/dnsmasq-dns-5885bcf447-68v4h" Oct 14 07:20:04 crc kubenswrapper[4870]: I1014 07:20:04.597565 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6d5f827b-9c3f-4d17-b6a6-506fe3d2e06f-fernet-keys\") pod \"keystone-bootstrap-s2xm6\" (UID: \"6d5f827b-9c3f-4d17-b6a6-506fe3d2e06f\") " pod="openstack/keystone-bootstrap-s2xm6" Oct 14 07:20:04 crc kubenswrapper[4870]: I1014 07:20:04.597589 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d94da551-da6b-48b1-ac7d-a78f06715651-ovsdbserver-nb\") pod \"dnsmasq-dns-5885bcf447-68v4h\" (UID: \"d94da551-da6b-48b1-ac7d-a78f06715651\") " pod="openstack/dnsmasq-dns-5885bcf447-68v4h" Oct 14 07:20:04 crc kubenswrapper[4870]: I1014 07:20:04.597645 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6d5f827b-9c3f-4d17-b6a6-506fe3d2e06f-scripts\") pod \"keystone-bootstrap-s2xm6\" (UID: \"6d5f827b-9c3f-4d17-b6a6-506fe3d2e06f\") " pod="openstack/keystone-bootstrap-s2xm6" Oct 14 07:20:04 crc kubenswrapper[4870]: I1014 07:20:04.597664 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6d5f827b-9c3f-4d17-b6a6-506fe3d2e06f-config-data\") pod \"keystone-bootstrap-s2xm6\" (UID: \"6d5f827b-9c3f-4d17-b6a6-506fe3d2e06f\") " pod="openstack/keystone-bootstrap-s2xm6" Oct 14 07:20:04 crc kubenswrapper[4870]: I1014 07:20:04.597684 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-llvg4\" (UniqueName: \"kubernetes.io/projected/d94da551-da6b-48b1-ac7d-a78f06715651-kube-api-access-llvg4\") pod \"dnsmasq-dns-5885bcf447-68v4h\" (UID: \"d94da551-da6b-48b1-ac7d-a78f06715651\") " pod="openstack/dnsmasq-dns-5885bcf447-68v4h" Oct 14 07:20:04 crc kubenswrapper[4870]: I1014 07:20:04.597704 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d94da551-da6b-48b1-ac7d-a78f06715651-dns-svc\") pod \"dnsmasq-dns-5885bcf447-68v4h\" (UID: \"d94da551-da6b-48b1-ac7d-a78f06715651\") " pod="openstack/dnsmasq-dns-5885bcf447-68v4h" Oct 14 07:20:04 crc kubenswrapper[4870]: I1014 07:20:04.597736 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bvx4m\" (UniqueName: \"kubernetes.io/projected/6d5f827b-9c3f-4d17-b6a6-506fe3d2e06f-kube-api-access-bvx4m\") pod \"keystone-bootstrap-s2xm6\" (UID: \"6d5f827b-9c3f-4d17-b6a6-506fe3d2e06f\") " pod="openstack/keystone-bootstrap-s2xm6" Oct 14 07:20:04 crc kubenswrapper[4870]: I1014 07:20:04.597768 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/6d5f827b-9c3f-4d17-b6a6-506fe3d2e06f-credential-keys\") pod \"keystone-bootstrap-s2xm6\" (UID: \"6d5f827b-9c3f-4d17-b6a6-506fe3d2e06f\") " pod="openstack/keystone-bootstrap-s2xm6" Oct 14 07:20:04 crc kubenswrapper[4870]: I1014 07:20:04.597785 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d94da551-da6b-48b1-ac7d-a78f06715651-config\") pod \"dnsmasq-dns-5885bcf447-68v4h\" (UID: \"d94da551-da6b-48b1-ac7d-a78f06715651\") " pod="openstack/dnsmasq-dns-5885bcf447-68v4h" Oct 14 07:20:04 crc kubenswrapper[4870]: I1014 07:20:04.667459 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 14 07:20:04 crc kubenswrapper[4870]: I1014 07:20:04.669469 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 14 07:20:04 crc kubenswrapper[4870]: I1014 07:20:04.671537 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 14 07:20:04 crc kubenswrapper[4870]: I1014 07:20:04.671703 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 14 07:20:04 crc kubenswrapper[4870]: I1014 07:20:04.691070 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5885bcf447-68v4h"] Oct 14 07:20:04 crc kubenswrapper[4870]: E1014 07:20:04.691665 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[config dns-svc dns-swift-storage-0 kube-api-access-llvg4 ovsdbserver-nb ovsdbserver-sb], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/dnsmasq-dns-5885bcf447-68v4h" podUID="d94da551-da6b-48b1-ac7d-a78f06715651" Oct 14 07:20:04 crc kubenswrapper[4870]: I1014 07:20:04.696735 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-nmtf9"] Oct 14 07:20:04 crc kubenswrapper[4870]: I1014 07:20:04.697996 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-nmtf9" Oct 14 07:20:04 crc kubenswrapper[4870]: I1014 07:20:04.699718 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d94da551-da6b-48b1-ac7d-a78f06715651-dns-swift-storage-0\") pod \"dnsmasq-dns-5885bcf447-68v4h\" (UID: \"d94da551-da6b-48b1-ac7d-a78f06715651\") " pod="openstack/dnsmasq-dns-5885bcf447-68v4h" Oct 14 07:20:04 crc kubenswrapper[4870]: I1014 07:20:04.699773 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d5f827b-9c3f-4d17-b6a6-506fe3d2e06f-combined-ca-bundle\") pod \"keystone-bootstrap-s2xm6\" (UID: \"6d5f827b-9c3f-4d17-b6a6-506fe3d2e06f\") " pod="openstack/keystone-bootstrap-s2xm6" Oct 14 07:20:04 crc kubenswrapper[4870]: I1014 07:20:04.699800 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d94da551-da6b-48b1-ac7d-a78f06715651-ovsdbserver-sb\") pod \"dnsmasq-dns-5885bcf447-68v4h\" (UID: \"d94da551-da6b-48b1-ac7d-a78f06715651\") " pod="openstack/dnsmasq-dns-5885bcf447-68v4h" Oct 14 07:20:04 crc kubenswrapper[4870]: I1014 07:20:04.699844 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6d5f827b-9c3f-4d17-b6a6-506fe3d2e06f-fernet-keys\") pod \"keystone-bootstrap-s2xm6\" (UID: \"6d5f827b-9c3f-4d17-b6a6-506fe3d2e06f\") " pod="openstack/keystone-bootstrap-s2xm6" Oct 14 07:20:04 crc kubenswrapper[4870]: I1014 07:20:04.699865 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d94da551-da6b-48b1-ac7d-a78f06715651-ovsdbserver-nb\") pod \"dnsmasq-dns-5885bcf447-68v4h\" (UID: \"d94da551-da6b-48b1-ac7d-a78f06715651\") " pod="openstack/dnsmasq-dns-5885bcf447-68v4h" Oct 14 07:20:04 crc kubenswrapper[4870]: I1014 07:20:04.699894 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6d5f827b-9c3f-4d17-b6a6-506fe3d2e06f-scripts\") pod \"keystone-bootstrap-s2xm6\" (UID: \"6d5f827b-9c3f-4d17-b6a6-506fe3d2e06f\") " pod="openstack/keystone-bootstrap-s2xm6" Oct 14 07:20:04 crc kubenswrapper[4870]: I1014 07:20:04.699910 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6d5f827b-9c3f-4d17-b6a6-506fe3d2e06f-config-data\") pod \"keystone-bootstrap-s2xm6\" (UID: \"6d5f827b-9c3f-4d17-b6a6-506fe3d2e06f\") " pod="openstack/keystone-bootstrap-s2xm6" Oct 14 07:20:04 crc kubenswrapper[4870]: I1014 07:20:04.699929 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-llvg4\" (UniqueName: \"kubernetes.io/projected/d94da551-da6b-48b1-ac7d-a78f06715651-kube-api-access-llvg4\") pod \"dnsmasq-dns-5885bcf447-68v4h\" (UID: \"d94da551-da6b-48b1-ac7d-a78f06715651\") " pod="openstack/dnsmasq-dns-5885bcf447-68v4h" Oct 14 07:20:04 crc kubenswrapper[4870]: I1014 07:20:04.699945 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d94da551-da6b-48b1-ac7d-a78f06715651-dns-svc\") pod \"dnsmasq-dns-5885bcf447-68v4h\" (UID: \"d94da551-da6b-48b1-ac7d-a78f06715651\") " pod="openstack/dnsmasq-dns-5885bcf447-68v4h" Oct 14 07:20:04 crc kubenswrapper[4870]: I1014 07:20:04.699978 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bvx4m\" (UniqueName: \"kubernetes.io/projected/6d5f827b-9c3f-4d17-b6a6-506fe3d2e06f-kube-api-access-bvx4m\") pod \"keystone-bootstrap-s2xm6\" (UID: \"6d5f827b-9c3f-4d17-b6a6-506fe3d2e06f\") " pod="openstack/keystone-bootstrap-s2xm6" Oct 14 07:20:04 crc kubenswrapper[4870]: I1014 07:20:04.699997 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/6d5f827b-9c3f-4d17-b6a6-506fe3d2e06f-credential-keys\") pod \"keystone-bootstrap-s2xm6\" (UID: \"6d5f827b-9c3f-4d17-b6a6-506fe3d2e06f\") " pod="openstack/keystone-bootstrap-s2xm6" Oct 14 07:20:04 crc kubenswrapper[4870]: I1014 07:20:04.700014 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d94da551-da6b-48b1-ac7d-a78f06715651-config\") pod \"dnsmasq-dns-5885bcf447-68v4h\" (UID: \"d94da551-da6b-48b1-ac7d-a78f06715651\") " pod="openstack/dnsmasq-dns-5885bcf447-68v4h" Oct 14 07:20:04 crc kubenswrapper[4870]: I1014 07:20:04.701196 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d94da551-da6b-48b1-ac7d-a78f06715651-config\") pod \"dnsmasq-dns-5885bcf447-68v4h\" (UID: \"d94da551-da6b-48b1-ac7d-a78f06715651\") " pod="openstack/dnsmasq-dns-5885bcf447-68v4h" Oct 14 07:20:04 crc kubenswrapper[4870]: I1014 07:20:04.708107 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d94da551-da6b-48b1-ac7d-a78f06715651-dns-svc\") pod \"dnsmasq-dns-5885bcf447-68v4h\" (UID: \"d94da551-da6b-48b1-ac7d-a78f06715651\") " pod="openstack/dnsmasq-dns-5885bcf447-68v4h" Oct 14 07:20:04 crc kubenswrapper[4870]: I1014 07:20:04.708470 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d94da551-da6b-48b1-ac7d-a78f06715651-ovsdbserver-sb\") pod \"dnsmasq-dns-5885bcf447-68v4h\" (UID: \"d94da551-da6b-48b1-ac7d-a78f06715651\") " pod="openstack/dnsmasq-dns-5885bcf447-68v4h" Oct 14 07:20:04 crc kubenswrapper[4870]: I1014 07:20:04.708936 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6d5f827b-9c3f-4d17-b6a6-506fe3d2e06f-scripts\") pod \"keystone-bootstrap-s2xm6\" (UID: \"6d5f827b-9c3f-4d17-b6a6-506fe3d2e06f\") " pod="openstack/keystone-bootstrap-s2xm6" Oct 14 07:20:04 crc kubenswrapper[4870]: I1014 07:20:04.710946 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d94da551-da6b-48b1-ac7d-a78f06715651-dns-swift-storage-0\") pod \"dnsmasq-dns-5885bcf447-68v4h\" (UID: \"d94da551-da6b-48b1-ac7d-a78f06715651\") " pod="openstack/dnsmasq-dns-5885bcf447-68v4h" Oct 14 07:20:04 crc kubenswrapper[4870]: I1014 07:20:04.711900 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Oct 14 07:20:04 crc kubenswrapper[4870]: I1014 07:20:04.712211 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Oct 14 07:20:04 crc kubenswrapper[4870]: I1014 07:20:04.717351 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-zsgq8" Oct 14 07:20:04 crc kubenswrapper[4870]: I1014 07:20:04.722589 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d94da551-da6b-48b1-ac7d-a78f06715651-ovsdbserver-nb\") pod \"dnsmasq-dns-5885bcf447-68v4h\" (UID: \"d94da551-da6b-48b1-ac7d-a78f06715651\") " pod="openstack/dnsmasq-dns-5885bcf447-68v4h" Oct 14 07:20:04 crc kubenswrapper[4870]: I1014 07:20:04.751543 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6d5f827b-9c3f-4d17-b6a6-506fe3d2e06f-config-data\") pod \"keystone-bootstrap-s2xm6\" (UID: \"6d5f827b-9c3f-4d17-b6a6-506fe3d2e06f\") " pod="openstack/keystone-bootstrap-s2xm6" Oct 14 07:20:04 crc kubenswrapper[4870]: I1014 07:20:04.754360 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 14 07:20:04 crc kubenswrapper[4870]: I1014 07:20:04.763635 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7bcdb8c767-9pwfh"] Oct 14 07:20:04 crc kubenswrapper[4870]: I1014 07:20:04.766107 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7bcdb8c767-9pwfh" Oct 14 07:20:04 crc kubenswrapper[4870]: I1014 07:20:04.766320 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-llvg4\" (UniqueName: \"kubernetes.io/projected/d94da551-da6b-48b1-ac7d-a78f06715651-kube-api-access-llvg4\") pod \"dnsmasq-dns-5885bcf447-68v4h\" (UID: \"d94da551-da6b-48b1-ac7d-a78f06715651\") " pod="openstack/dnsmasq-dns-5885bcf447-68v4h" Oct 14 07:20:04 crc kubenswrapper[4870]: I1014 07:20:04.766101 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6d5f827b-9c3f-4d17-b6a6-506fe3d2e06f-fernet-keys\") pod \"keystone-bootstrap-s2xm6\" (UID: \"6d5f827b-9c3f-4d17-b6a6-506fe3d2e06f\") " pod="openstack/keystone-bootstrap-s2xm6" Oct 14 07:20:04 crc kubenswrapper[4870]: I1014 07:20:04.766684 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/6d5f827b-9c3f-4d17-b6a6-506fe3d2e06f-credential-keys\") pod \"keystone-bootstrap-s2xm6\" (UID: \"6d5f827b-9c3f-4d17-b6a6-506fe3d2e06f\") " pod="openstack/keystone-bootstrap-s2xm6" Oct 14 07:20:04 crc kubenswrapper[4870]: I1014 07:20:04.772715 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d5f827b-9c3f-4d17-b6a6-506fe3d2e06f-combined-ca-bundle\") pod \"keystone-bootstrap-s2xm6\" (UID: \"6d5f827b-9c3f-4d17-b6a6-506fe3d2e06f\") " pod="openstack/keystone-bootstrap-s2xm6" Oct 14 07:20:04 crc kubenswrapper[4870]: I1014 07:20:04.789816 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-nmtf9"] Oct 14 07:20:04 crc kubenswrapper[4870]: I1014 07:20:04.807258 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a31f5df-eae7-4087-a8c8-09c18f5ee9f0-config-data\") pod \"ceilometer-0\" (UID: \"0a31f5df-eae7-4087-a8c8-09c18f5ee9f0\") " pod="openstack/ceilometer-0" Oct 14 07:20:04 crc kubenswrapper[4870]: I1014 07:20:04.807305 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4flk9\" (UniqueName: \"kubernetes.io/projected/9dd93b33-f231-4339-9db0-0507cde1dcf5-kube-api-access-4flk9\") pod \"placement-db-sync-nmtf9\" (UID: \"9dd93b33-f231-4339-9db0-0507cde1dcf5\") " pod="openstack/placement-db-sync-nmtf9" Oct 14 07:20:04 crc kubenswrapper[4870]: I1014 07:20:04.807326 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9dd93b33-f231-4339-9db0-0507cde1dcf5-scripts\") pod \"placement-db-sync-nmtf9\" (UID: \"9dd93b33-f231-4339-9db0-0507cde1dcf5\") " pod="openstack/placement-db-sync-nmtf9" Oct 14 07:20:04 crc kubenswrapper[4870]: I1014 07:20:04.807353 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0a31f5df-eae7-4087-a8c8-09c18f5ee9f0-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0a31f5df-eae7-4087-a8c8-09c18f5ee9f0\") " pod="openstack/ceilometer-0" Oct 14 07:20:04 crc kubenswrapper[4870]: I1014 07:20:04.807554 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9dd93b33-f231-4339-9db0-0507cde1dcf5-config-data\") pod \"placement-db-sync-nmtf9\" (UID: \"9dd93b33-f231-4339-9db0-0507cde1dcf5\") " pod="openstack/placement-db-sync-nmtf9" Oct 14 07:20:04 crc kubenswrapper[4870]: I1014 07:20:04.807602 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9dd93b33-f231-4339-9db0-0507cde1dcf5-logs\") pod \"placement-db-sync-nmtf9\" (UID: \"9dd93b33-f231-4339-9db0-0507cde1dcf5\") " pod="openstack/placement-db-sync-nmtf9" Oct 14 07:20:04 crc kubenswrapper[4870]: I1014 07:20:04.807657 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0a31f5df-eae7-4087-a8c8-09c18f5ee9f0-run-httpd\") pod \"ceilometer-0\" (UID: \"0a31f5df-eae7-4087-a8c8-09c18f5ee9f0\") " pod="openstack/ceilometer-0" Oct 14 07:20:04 crc kubenswrapper[4870]: I1014 07:20:04.807716 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0a31f5df-eae7-4087-a8c8-09c18f5ee9f0-log-httpd\") pod \"ceilometer-0\" (UID: \"0a31f5df-eae7-4087-a8c8-09c18f5ee9f0\") " pod="openstack/ceilometer-0" Oct 14 07:20:04 crc kubenswrapper[4870]: I1014 07:20:04.807777 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kdzq6\" (UniqueName: \"kubernetes.io/projected/0a31f5df-eae7-4087-a8c8-09c18f5ee9f0-kube-api-access-kdzq6\") pod \"ceilometer-0\" (UID: \"0a31f5df-eae7-4087-a8c8-09c18f5ee9f0\") " pod="openstack/ceilometer-0" Oct 14 07:20:04 crc kubenswrapper[4870]: I1014 07:20:04.807841 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9dd93b33-f231-4339-9db0-0507cde1dcf5-combined-ca-bundle\") pod \"placement-db-sync-nmtf9\" (UID: \"9dd93b33-f231-4339-9db0-0507cde1dcf5\") " pod="openstack/placement-db-sync-nmtf9" Oct 14 07:20:04 crc kubenswrapper[4870]: I1014 07:20:04.807864 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a31f5df-eae7-4087-a8c8-09c18f5ee9f0-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0a31f5df-eae7-4087-a8c8-09c18f5ee9f0\") " pod="openstack/ceilometer-0" Oct 14 07:20:04 crc kubenswrapper[4870]: I1014 07:20:04.807969 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0a31f5df-eae7-4087-a8c8-09c18f5ee9f0-scripts\") pod \"ceilometer-0\" (UID: \"0a31f5df-eae7-4087-a8c8-09c18f5ee9f0\") " pod="openstack/ceilometer-0" Oct 14 07:20:04 crc kubenswrapper[4870]: I1014 07:20:04.815012 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bvx4m\" (UniqueName: \"kubernetes.io/projected/6d5f827b-9c3f-4d17-b6a6-506fe3d2e06f-kube-api-access-bvx4m\") pod \"keystone-bootstrap-s2xm6\" (UID: \"6d5f827b-9c3f-4d17-b6a6-506fe3d2e06f\") " pod="openstack/keystone-bootstrap-s2xm6" Oct 14 07:20:04 crc kubenswrapper[4870]: I1014 07:20:04.826825 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7bcdb8c767-9pwfh"] Oct 14 07:20:04 crc kubenswrapper[4870]: I1014 07:20:04.909474 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0a31f5df-eae7-4087-a8c8-09c18f5ee9f0-run-httpd\") pod \"ceilometer-0\" (UID: \"0a31f5df-eae7-4087-a8c8-09c18f5ee9f0\") " pod="openstack/ceilometer-0" Oct 14 07:20:04 crc kubenswrapper[4870]: I1014 07:20:04.909548 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0a31f5df-eae7-4087-a8c8-09c18f5ee9f0-log-httpd\") pod \"ceilometer-0\" (UID: \"0a31f5df-eae7-4087-a8c8-09c18f5ee9f0\") " pod="openstack/ceilometer-0" Oct 14 07:20:04 crc kubenswrapper[4870]: I1014 07:20:04.909576 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/32e2388e-2670-44d3-bfe1-6ce4c126eb18-config\") pod \"dnsmasq-dns-7bcdb8c767-9pwfh\" (UID: \"32e2388e-2670-44d3-bfe1-6ce4c126eb18\") " pod="openstack/dnsmasq-dns-7bcdb8c767-9pwfh" Oct 14 07:20:04 crc kubenswrapper[4870]: I1014 07:20:04.909615 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kdzq6\" (UniqueName: \"kubernetes.io/projected/0a31f5df-eae7-4087-a8c8-09c18f5ee9f0-kube-api-access-kdzq6\") pod \"ceilometer-0\" (UID: \"0a31f5df-eae7-4087-a8c8-09c18f5ee9f0\") " pod="openstack/ceilometer-0" Oct 14 07:20:04 crc kubenswrapper[4870]: I1014 07:20:04.909653 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9dd93b33-f231-4339-9db0-0507cde1dcf5-combined-ca-bundle\") pod \"placement-db-sync-nmtf9\" (UID: \"9dd93b33-f231-4339-9db0-0507cde1dcf5\") " pod="openstack/placement-db-sync-nmtf9" Oct 14 07:20:04 crc kubenswrapper[4870]: I1014 07:20:04.909673 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a31f5df-eae7-4087-a8c8-09c18f5ee9f0-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0a31f5df-eae7-4087-a8c8-09c18f5ee9f0\") " pod="openstack/ceilometer-0" Oct 14 07:20:04 crc kubenswrapper[4870]: I1014 07:20:04.909701 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dzqs2\" (UniqueName: \"kubernetes.io/projected/32e2388e-2670-44d3-bfe1-6ce4c126eb18-kube-api-access-dzqs2\") pod \"dnsmasq-dns-7bcdb8c767-9pwfh\" (UID: \"32e2388e-2670-44d3-bfe1-6ce4c126eb18\") " pod="openstack/dnsmasq-dns-7bcdb8c767-9pwfh" Oct 14 07:20:04 crc kubenswrapper[4870]: I1014 07:20:04.909746 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/32e2388e-2670-44d3-bfe1-6ce4c126eb18-ovsdbserver-sb\") pod \"dnsmasq-dns-7bcdb8c767-9pwfh\" (UID: \"32e2388e-2670-44d3-bfe1-6ce4c126eb18\") " pod="openstack/dnsmasq-dns-7bcdb8c767-9pwfh" Oct 14 07:20:04 crc kubenswrapper[4870]: I1014 07:20:04.909778 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0a31f5df-eae7-4087-a8c8-09c18f5ee9f0-scripts\") pod \"ceilometer-0\" (UID: \"0a31f5df-eae7-4087-a8c8-09c18f5ee9f0\") " pod="openstack/ceilometer-0" Oct 14 07:20:04 crc kubenswrapper[4870]: I1014 07:20:04.909816 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a31f5df-eae7-4087-a8c8-09c18f5ee9f0-config-data\") pod \"ceilometer-0\" (UID: \"0a31f5df-eae7-4087-a8c8-09c18f5ee9f0\") " pod="openstack/ceilometer-0" Oct 14 07:20:04 crc kubenswrapper[4870]: I1014 07:20:04.909838 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4flk9\" (UniqueName: \"kubernetes.io/projected/9dd93b33-f231-4339-9db0-0507cde1dcf5-kube-api-access-4flk9\") pod \"placement-db-sync-nmtf9\" (UID: \"9dd93b33-f231-4339-9db0-0507cde1dcf5\") " pod="openstack/placement-db-sync-nmtf9" Oct 14 07:20:04 crc kubenswrapper[4870]: I1014 07:20:04.909859 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9dd93b33-f231-4339-9db0-0507cde1dcf5-scripts\") pod \"placement-db-sync-nmtf9\" (UID: \"9dd93b33-f231-4339-9db0-0507cde1dcf5\") " pod="openstack/placement-db-sync-nmtf9" Oct 14 07:20:04 crc kubenswrapper[4870]: I1014 07:20:04.909887 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0a31f5df-eae7-4087-a8c8-09c18f5ee9f0-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0a31f5df-eae7-4087-a8c8-09c18f5ee9f0\") " pod="openstack/ceilometer-0" Oct 14 07:20:04 crc kubenswrapper[4870]: I1014 07:20:04.909909 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/32e2388e-2670-44d3-bfe1-6ce4c126eb18-dns-swift-storage-0\") pod \"dnsmasq-dns-7bcdb8c767-9pwfh\" (UID: \"32e2388e-2670-44d3-bfe1-6ce4c126eb18\") " pod="openstack/dnsmasq-dns-7bcdb8c767-9pwfh" Oct 14 07:20:04 crc kubenswrapper[4870]: I1014 07:20:04.909932 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9dd93b33-f231-4339-9db0-0507cde1dcf5-config-data\") pod \"placement-db-sync-nmtf9\" (UID: \"9dd93b33-f231-4339-9db0-0507cde1dcf5\") " pod="openstack/placement-db-sync-nmtf9" Oct 14 07:20:04 crc kubenswrapper[4870]: I1014 07:20:04.909955 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9dd93b33-f231-4339-9db0-0507cde1dcf5-logs\") pod \"placement-db-sync-nmtf9\" (UID: \"9dd93b33-f231-4339-9db0-0507cde1dcf5\") " pod="openstack/placement-db-sync-nmtf9" Oct 14 07:20:04 crc kubenswrapper[4870]: I1014 07:20:04.909975 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/32e2388e-2670-44d3-bfe1-6ce4c126eb18-dns-svc\") pod \"dnsmasq-dns-7bcdb8c767-9pwfh\" (UID: \"32e2388e-2670-44d3-bfe1-6ce4c126eb18\") " pod="openstack/dnsmasq-dns-7bcdb8c767-9pwfh" Oct 14 07:20:04 crc kubenswrapper[4870]: I1014 07:20:04.909997 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/32e2388e-2670-44d3-bfe1-6ce4c126eb18-ovsdbserver-nb\") pod \"dnsmasq-dns-7bcdb8c767-9pwfh\" (UID: \"32e2388e-2670-44d3-bfe1-6ce4c126eb18\") " pod="openstack/dnsmasq-dns-7bcdb8c767-9pwfh" Oct 14 07:20:04 crc kubenswrapper[4870]: I1014 07:20:04.910482 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0a31f5df-eae7-4087-a8c8-09c18f5ee9f0-run-httpd\") pod \"ceilometer-0\" (UID: \"0a31f5df-eae7-4087-a8c8-09c18f5ee9f0\") " pod="openstack/ceilometer-0" Oct 14 07:20:04 crc kubenswrapper[4870]: I1014 07:20:04.910707 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0a31f5df-eae7-4087-a8c8-09c18f5ee9f0-log-httpd\") pod \"ceilometer-0\" (UID: \"0a31f5df-eae7-4087-a8c8-09c18f5ee9f0\") " pod="openstack/ceilometer-0" Oct 14 07:20:04 crc kubenswrapper[4870]: I1014 07:20:04.918534 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a31f5df-eae7-4087-a8c8-09c18f5ee9f0-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0a31f5df-eae7-4087-a8c8-09c18f5ee9f0\") " pod="openstack/ceilometer-0" Oct 14 07:20:04 crc kubenswrapper[4870]: I1014 07:20:04.923515 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9dd93b33-f231-4339-9db0-0507cde1dcf5-combined-ca-bundle\") pod \"placement-db-sync-nmtf9\" (UID: \"9dd93b33-f231-4339-9db0-0507cde1dcf5\") " pod="openstack/placement-db-sync-nmtf9" Oct 14 07:20:04 crc kubenswrapper[4870]: I1014 07:20:04.924042 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9dd93b33-f231-4339-9db0-0507cde1dcf5-logs\") pod \"placement-db-sync-nmtf9\" (UID: \"9dd93b33-f231-4339-9db0-0507cde1dcf5\") " pod="openstack/placement-db-sync-nmtf9" Oct 14 07:20:04 crc kubenswrapper[4870]: I1014 07:20:04.924667 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0a31f5df-eae7-4087-a8c8-09c18f5ee9f0-scripts\") pod \"ceilometer-0\" (UID: \"0a31f5df-eae7-4087-a8c8-09c18f5ee9f0\") " pod="openstack/ceilometer-0" Oct 14 07:20:04 crc kubenswrapper[4870]: I1014 07:20:04.926967 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9dd93b33-f231-4339-9db0-0507cde1dcf5-scripts\") pod \"placement-db-sync-nmtf9\" (UID: \"9dd93b33-f231-4339-9db0-0507cde1dcf5\") " pod="openstack/placement-db-sync-nmtf9" Oct 14 07:20:04 crc kubenswrapper[4870]: I1014 07:20:04.931484 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0a31f5df-eae7-4087-a8c8-09c18f5ee9f0-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0a31f5df-eae7-4087-a8c8-09c18f5ee9f0\") " pod="openstack/ceilometer-0" Oct 14 07:20:04 crc kubenswrapper[4870]: I1014 07:20:04.931816 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4flk9\" (UniqueName: \"kubernetes.io/projected/9dd93b33-f231-4339-9db0-0507cde1dcf5-kube-api-access-4flk9\") pod \"placement-db-sync-nmtf9\" (UID: \"9dd93b33-f231-4339-9db0-0507cde1dcf5\") " pod="openstack/placement-db-sync-nmtf9" Oct 14 07:20:04 crc kubenswrapper[4870]: I1014 07:20:04.932168 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a31f5df-eae7-4087-a8c8-09c18f5ee9f0-config-data\") pod \"ceilometer-0\" (UID: \"0a31f5df-eae7-4087-a8c8-09c18f5ee9f0\") " pod="openstack/ceilometer-0" Oct 14 07:20:04 crc kubenswrapper[4870]: I1014 07:20:04.933642 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9dd93b33-f231-4339-9db0-0507cde1dcf5-config-data\") pod \"placement-db-sync-nmtf9\" (UID: \"9dd93b33-f231-4339-9db0-0507cde1dcf5\") " pod="openstack/placement-db-sync-nmtf9" Oct 14 07:20:04 crc kubenswrapper[4870]: I1014 07:20:04.950035 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kdzq6\" (UniqueName: \"kubernetes.io/projected/0a31f5df-eae7-4087-a8c8-09c18f5ee9f0-kube-api-access-kdzq6\") pod \"ceilometer-0\" (UID: \"0a31f5df-eae7-4087-a8c8-09c18f5ee9f0\") " pod="openstack/ceilometer-0" Oct 14 07:20:04 crc kubenswrapper[4870]: I1014 07:20:04.986046 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 14 07:20:05 crc kubenswrapper[4870]: I1014 07:20:05.011421 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/32e2388e-2670-44d3-bfe1-6ce4c126eb18-dns-svc\") pod \"dnsmasq-dns-7bcdb8c767-9pwfh\" (UID: \"32e2388e-2670-44d3-bfe1-6ce4c126eb18\") " pod="openstack/dnsmasq-dns-7bcdb8c767-9pwfh" Oct 14 07:20:05 crc kubenswrapper[4870]: I1014 07:20:05.011481 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/32e2388e-2670-44d3-bfe1-6ce4c126eb18-ovsdbserver-nb\") pod \"dnsmasq-dns-7bcdb8c767-9pwfh\" (UID: \"32e2388e-2670-44d3-bfe1-6ce4c126eb18\") " pod="openstack/dnsmasq-dns-7bcdb8c767-9pwfh" Oct 14 07:20:05 crc kubenswrapper[4870]: I1014 07:20:05.011528 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/32e2388e-2670-44d3-bfe1-6ce4c126eb18-config\") pod \"dnsmasq-dns-7bcdb8c767-9pwfh\" (UID: \"32e2388e-2670-44d3-bfe1-6ce4c126eb18\") " pod="openstack/dnsmasq-dns-7bcdb8c767-9pwfh" Oct 14 07:20:05 crc kubenswrapper[4870]: I1014 07:20:05.011594 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dzqs2\" (UniqueName: \"kubernetes.io/projected/32e2388e-2670-44d3-bfe1-6ce4c126eb18-kube-api-access-dzqs2\") pod \"dnsmasq-dns-7bcdb8c767-9pwfh\" (UID: \"32e2388e-2670-44d3-bfe1-6ce4c126eb18\") " pod="openstack/dnsmasq-dns-7bcdb8c767-9pwfh" Oct 14 07:20:05 crc kubenswrapper[4870]: I1014 07:20:05.011634 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/32e2388e-2670-44d3-bfe1-6ce4c126eb18-ovsdbserver-sb\") pod \"dnsmasq-dns-7bcdb8c767-9pwfh\" (UID: \"32e2388e-2670-44d3-bfe1-6ce4c126eb18\") " pod="openstack/dnsmasq-dns-7bcdb8c767-9pwfh" Oct 14 07:20:05 crc kubenswrapper[4870]: I1014 07:20:05.011680 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/32e2388e-2670-44d3-bfe1-6ce4c126eb18-dns-swift-storage-0\") pod \"dnsmasq-dns-7bcdb8c767-9pwfh\" (UID: \"32e2388e-2670-44d3-bfe1-6ce4c126eb18\") " pod="openstack/dnsmasq-dns-7bcdb8c767-9pwfh" Oct 14 07:20:05 crc kubenswrapper[4870]: I1014 07:20:05.012516 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/32e2388e-2670-44d3-bfe1-6ce4c126eb18-dns-swift-storage-0\") pod \"dnsmasq-dns-7bcdb8c767-9pwfh\" (UID: \"32e2388e-2670-44d3-bfe1-6ce4c126eb18\") " pod="openstack/dnsmasq-dns-7bcdb8c767-9pwfh" Oct 14 07:20:05 crc kubenswrapper[4870]: I1014 07:20:05.012533 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/32e2388e-2670-44d3-bfe1-6ce4c126eb18-ovsdbserver-nb\") pod \"dnsmasq-dns-7bcdb8c767-9pwfh\" (UID: \"32e2388e-2670-44d3-bfe1-6ce4c126eb18\") " pod="openstack/dnsmasq-dns-7bcdb8c767-9pwfh" Oct 14 07:20:05 crc kubenswrapper[4870]: I1014 07:20:05.012578 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/32e2388e-2670-44d3-bfe1-6ce4c126eb18-dns-svc\") pod \"dnsmasq-dns-7bcdb8c767-9pwfh\" (UID: \"32e2388e-2670-44d3-bfe1-6ce4c126eb18\") " pod="openstack/dnsmasq-dns-7bcdb8c767-9pwfh" Oct 14 07:20:05 crc kubenswrapper[4870]: I1014 07:20:05.013168 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/32e2388e-2670-44d3-bfe1-6ce4c126eb18-config\") pod \"dnsmasq-dns-7bcdb8c767-9pwfh\" (UID: \"32e2388e-2670-44d3-bfe1-6ce4c126eb18\") " pod="openstack/dnsmasq-dns-7bcdb8c767-9pwfh" Oct 14 07:20:05 crc kubenswrapper[4870]: I1014 07:20:05.014009 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/32e2388e-2670-44d3-bfe1-6ce4c126eb18-ovsdbserver-sb\") pod \"dnsmasq-dns-7bcdb8c767-9pwfh\" (UID: \"32e2388e-2670-44d3-bfe1-6ce4c126eb18\") " pod="openstack/dnsmasq-dns-7bcdb8c767-9pwfh" Oct 14 07:20:05 crc kubenswrapper[4870]: I1014 07:20:05.027363 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dzqs2\" (UniqueName: \"kubernetes.io/projected/32e2388e-2670-44d3-bfe1-6ce4c126eb18-kube-api-access-dzqs2\") pod \"dnsmasq-dns-7bcdb8c767-9pwfh\" (UID: \"32e2388e-2670-44d3-bfe1-6ce4c126eb18\") " pod="openstack/dnsmasq-dns-7bcdb8c767-9pwfh" Oct 14 07:20:05 crc kubenswrapper[4870]: I1014 07:20:05.047954 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-zsgq8" Oct 14 07:20:05 crc kubenswrapper[4870]: I1014 07:20:05.050494 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-cqggg" Oct 14 07:20:05 crc kubenswrapper[4870]: I1014 07:20:05.059276 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-s2xm6" Oct 14 07:20:05 crc kubenswrapper[4870]: I1014 07:20:05.062539 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-nmtf9" Oct 14 07:20:05 crc kubenswrapper[4870]: I1014 07:20:05.130785 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7bcdb8c767-9pwfh" Oct 14 07:20:05 crc kubenswrapper[4870]: I1014 07:20:05.168616 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5885bcf447-68v4h" Oct 14 07:20:05 crc kubenswrapper[4870]: I1014 07:20:05.168719 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cb67568b9-c75vp" event={"ID":"3bb9dec0-1879-419a-ab06-4fbdf87f75af","Type":"ContainerStarted","Data":"47d72ed3aece21a46c275a0da7700dacf90fa3ce5ad56bf25032383083c026bc"} Oct 14 07:20:05 crc kubenswrapper[4870]: I1014 07:20:05.168786 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-cb67568b9-c75vp" podUID="3bb9dec0-1879-419a-ab06-4fbdf87f75af" containerName="dnsmasq-dns" containerID="cri-o://47d72ed3aece21a46c275a0da7700dacf90fa3ce5ad56bf25032383083c026bc" gracePeriod=10 Oct 14 07:20:05 crc kubenswrapper[4870]: I1014 07:20:05.182910 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5885bcf447-68v4h" Oct 14 07:20:05 crc kubenswrapper[4870]: I1014 07:20:05.206709 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-cb67568b9-c75vp" podStartSLOduration=3.206691422 podStartE2EDuration="3.206691422s" podCreationTimestamp="2025-10-14 07:20:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:20:05.195002693 +0000 UTC m=+1140.892363064" watchObservedRunningTime="2025-10-14 07:20:05.206691422 +0000 UTC m=+1140.904051793" Oct 14 07:20:05 crc kubenswrapper[4870]: I1014 07:20:05.332992 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d94da551-da6b-48b1-ac7d-a78f06715651-config\") pod \"d94da551-da6b-48b1-ac7d-a78f06715651\" (UID: \"d94da551-da6b-48b1-ac7d-a78f06715651\") " Oct 14 07:20:05 crc kubenswrapper[4870]: I1014 07:20:05.333062 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d94da551-da6b-48b1-ac7d-a78f06715651-dns-swift-storage-0\") pod \"d94da551-da6b-48b1-ac7d-a78f06715651\" (UID: \"d94da551-da6b-48b1-ac7d-a78f06715651\") " Oct 14 07:20:05 crc kubenswrapper[4870]: I1014 07:20:05.333112 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d94da551-da6b-48b1-ac7d-a78f06715651-dns-svc\") pod \"d94da551-da6b-48b1-ac7d-a78f06715651\" (UID: \"d94da551-da6b-48b1-ac7d-a78f06715651\") " Oct 14 07:20:05 crc kubenswrapper[4870]: I1014 07:20:05.333174 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-llvg4\" (UniqueName: \"kubernetes.io/projected/d94da551-da6b-48b1-ac7d-a78f06715651-kube-api-access-llvg4\") pod \"d94da551-da6b-48b1-ac7d-a78f06715651\" (UID: \"d94da551-da6b-48b1-ac7d-a78f06715651\") " Oct 14 07:20:05 crc kubenswrapper[4870]: I1014 07:20:05.333250 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d94da551-da6b-48b1-ac7d-a78f06715651-ovsdbserver-nb\") pod \"d94da551-da6b-48b1-ac7d-a78f06715651\" (UID: \"d94da551-da6b-48b1-ac7d-a78f06715651\") " Oct 14 07:20:05 crc kubenswrapper[4870]: I1014 07:20:05.333273 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d94da551-da6b-48b1-ac7d-a78f06715651-ovsdbserver-sb\") pod \"d94da551-da6b-48b1-ac7d-a78f06715651\" (UID: \"d94da551-da6b-48b1-ac7d-a78f06715651\") " Oct 14 07:20:05 crc kubenswrapper[4870]: I1014 07:20:05.333978 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d94da551-da6b-48b1-ac7d-a78f06715651-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "d94da551-da6b-48b1-ac7d-a78f06715651" (UID: "d94da551-da6b-48b1-ac7d-a78f06715651"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:20:05 crc kubenswrapper[4870]: I1014 07:20:05.334270 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d94da551-da6b-48b1-ac7d-a78f06715651-config" (OuterVolumeSpecName: "config") pod "d94da551-da6b-48b1-ac7d-a78f06715651" (UID: "d94da551-da6b-48b1-ac7d-a78f06715651"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:20:05 crc kubenswrapper[4870]: I1014 07:20:05.334609 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d94da551-da6b-48b1-ac7d-a78f06715651-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "d94da551-da6b-48b1-ac7d-a78f06715651" (UID: "d94da551-da6b-48b1-ac7d-a78f06715651"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:20:05 crc kubenswrapper[4870]: I1014 07:20:05.334982 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d94da551-da6b-48b1-ac7d-a78f06715651-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "d94da551-da6b-48b1-ac7d-a78f06715651" (UID: "d94da551-da6b-48b1-ac7d-a78f06715651"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:20:05 crc kubenswrapper[4870]: I1014 07:20:05.335422 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d94da551-da6b-48b1-ac7d-a78f06715651-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "d94da551-da6b-48b1-ac7d-a78f06715651" (UID: "d94da551-da6b-48b1-ac7d-a78f06715651"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:20:05 crc kubenswrapper[4870]: I1014 07:20:05.342260 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d94da551-da6b-48b1-ac7d-a78f06715651-kube-api-access-llvg4" (OuterVolumeSpecName: "kube-api-access-llvg4") pod "d94da551-da6b-48b1-ac7d-a78f06715651" (UID: "d94da551-da6b-48b1-ac7d-a78f06715651"). InnerVolumeSpecName "kube-api-access-llvg4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:20:05 crc kubenswrapper[4870]: I1014 07:20:05.438551 4870 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d94da551-da6b-48b1-ac7d-a78f06715651-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 14 07:20:05 crc kubenswrapper[4870]: I1014 07:20:05.438585 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-llvg4\" (UniqueName: \"kubernetes.io/projected/d94da551-da6b-48b1-ac7d-a78f06715651-kube-api-access-llvg4\") on node \"crc\" DevicePath \"\"" Oct 14 07:20:05 crc kubenswrapper[4870]: I1014 07:20:05.438597 4870 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d94da551-da6b-48b1-ac7d-a78f06715651-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 14 07:20:05 crc kubenswrapper[4870]: I1014 07:20:05.438606 4870 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d94da551-da6b-48b1-ac7d-a78f06715651-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 14 07:20:05 crc kubenswrapper[4870]: I1014 07:20:05.438614 4870 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d94da551-da6b-48b1-ac7d-a78f06715651-config\") on node \"crc\" DevicePath \"\"" Oct 14 07:20:05 crc kubenswrapper[4870]: I1014 07:20:05.438622 4870 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d94da551-da6b-48b1-ac7d-a78f06715651-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 14 07:20:05 crc kubenswrapper[4870]: I1014 07:20:05.526927 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 14 07:20:05 crc kubenswrapper[4870]: I1014 07:20:05.541234 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 14 07:20:05 crc kubenswrapper[4870]: I1014 07:20:05.547219 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 14 07:20:05 crc kubenswrapper[4870]: I1014 07:20:05.550586 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 14 07:20:05 crc kubenswrapper[4870]: I1014 07:20:05.552280 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-n9lf9" Oct 14 07:20:05 crc kubenswrapper[4870]: I1014 07:20:05.555024 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Oct 14 07:20:05 crc kubenswrapper[4870]: I1014 07:20:05.572320 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 14 07:20:05 crc kubenswrapper[4870]: I1014 07:20:05.605108 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 14 07:20:05 crc kubenswrapper[4870]: I1014 07:20:05.607555 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 14 07:20:05 crc kubenswrapper[4870]: I1014 07:20:05.610230 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 14 07:20:05 crc kubenswrapper[4870]: I1014 07:20:05.619098 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 14 07:20:05 crc kubenswrapper[4870]: I1014 07:20:05.640480 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-s2xm6"] Oct 14 07:20:05 crc kubenswrapper[4870]: I1014 07:20:05.641310 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"d1368adf-8a2b-41ff-987d-806f65c1b128\") " pod="openstack/glance-default-external-api-0" Oct 14 07:20:05 crc kubenswrapper[4870]: I1014 07:20:05.641359 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d1368adf-8a2b-41ff-987d-806f65c1b128-scripts\") pod \"glance-default-external-api-0\" (UID: \"d1368adf-8a2b-41ff-987d-806f65c1b128\") " pod="openstack/glance-default-external-api-0" Oct 14 07:20:05 crc kubenswrapper[4870]: I1014 07:20:05.641392 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d1368adf-8a2b-41ff-987d-806f65c1b128-config-data\") pod \"glance-default-external-api-0\" (UID: \"d1368adf-8a2b-41ff-987d-806f65c1b128\") " pod="openstack/glance-default-external-api-0" Oct 14 07:20:05 crc kubenswrapper[4870]: I1014 07:20:05.641416 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d1368adf-8a2b-41ff-987d-806f65c1b128-logs\") pod \"glance-default-external-api-0\" (UID: \"d1368adf-8a2b-41ff-987d-806f65c1b128\") " pod="openstack/glance-default-external-api-0" Oct 14 07:20:05 crc kubenswrapper[4870]: I1014 07:20:05.641493 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d1368adf-8a2b-41ff-987d-806f65c1b128-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"d1368adf-8a2b-41ff-987d-806f65c1b128\") " pod="openstack/glance-default-external-api-0" Oct 14 07:20:05 crc kubenswrapper[4870]: I1014 07:20:05.641526 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l8g2b\" (UniqueName: \"kubernetes.io/projected/d1368adf-8a2b-41ff-987d-806f65c1b128-kube-api-access-l8g2b\") pod \"glance-default-external-api-0\" (UID: \"d1368adf-8a2b-41ff-987d-806f65c1b128\") " pod="openstack/glance-default-external-api-0" Oct 14 07:20:05 crc kubenswrapper[4870]: I1014 07:20:05.641547 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d1368adf-8a2b-41ff-987d-806f65c1b128-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"d1368adf-8a2b-41ff-987d-806f65c1b128\") " pod="openstack/glance-default-external-api-0" Oct 14 07:20:05 crc kubenswrapper[4870]: I1014 07:20:05.653675 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-nmtf9"] Oct 14 07:20:05 crc kubenswrapper[4870]: I1014 07:20:05.751309 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/99fa3d0e-8c4f-4893-b2fa-040837f04767-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"99fa3d0e-8c4f-4893-b2fa-040837f04767\") " pod="openstack/glance-default-internal-api-0" Oct 14 07:20:05 crc kubenswrapper[4870]: I1014 07:20:05.751671 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"d1368adf-8a2b-41ff-987d-806f65c1b128\") " pod="openstack/glance-default-external-api-0" Oct 14 07:20:05 crc kubenswrapper[4870]: I1014 07:20:05.751696 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"99fa3d0e-8c4f-4893-b2fa-040837f04767\") " pod="openstack/glance-default-internal-api-0" Oct 14 07:20:05 crc kubenswrapper[4870]: I1014 07:20:05.751733 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d1368adf-8a2b-41ff-987d-806f65c1b128-scripts\") pod \"glance-default-external-api-0\" (UID: \"d1368adf-8a2b-41ff-987d-806f65c1b128\") " pod="openstack/glance-default-external-api-0" Oct 14 07:20:05 crc kubenswrapper[4870]: I1014 07:20:05.751761 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d1368adf-8a2b-41ff-987d-806f65c1b128-config-data\") pod \"glance-default-external-api-0\" (UID: \"d1368adf-8a2b-41ff-987d-806f65c1b128\") " pod="openstack/glance-default-external-api-0" Oct 14 07:20:05 crc kubenswrapper[4870]: I1014 07:20:05.751786 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z6l65\" (UniqueName: \"kubernetes.io/projected/99fa3d0e-8c4f-4893-b2fa-040837f04767-kube-api-access-z6l65\") pod \"glance-default-internal-api-0\" (UID: \"99fa3d0e-8c4f-4893-b2fa-040837f04767\") " pod="openstack/glance-default-internal-api-0" Oct 14 07:20:05 crc kubenswrapper[4870]: I1014 07:20:05.751806 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d1368adf-8a2b-41ff-987d-806f65c1b128-logs\") pod \"glance-default-external-api-0\" (UID: \"d1368adf-8a2b-41ff-987d-806f65c1b128\") " pod="openstack/glance-default-external-api-0" Oct 14 07:20:05 crc kubenswrapper[4870]: I1014 07:20:05.751828 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/99fa3d0e-8c4f-4893-b2fa-040837f04767-scripts\") pod \"glance-default-internal-api-0\" (UID: \"99fa3d0e-8c4f-4893-b2fa-040837f04767\") " pod="openstack/glance-default-internal-api-0" Oct 14 07:20:05 crc kubenswrapper[4870]: I1014 07:20:05.751846 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d1368adf-8a2b-41ff-987d-806f65c1b128-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"d1368adf-8a2b-41ff-987d-806f65c1b128\") " pod="openstack/glance-default-external-api-0" Oct 14 07:20:05 crc kubenswrapper[4870]: I1014 07:20:05.751879 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l8g2b\" (UniqueName: \"kubernetes.io/projected/d1368adf-8a2b-41ff-987d-806f65c1b128-kube-api-access-l8g2b\") pod \"glance-default-external-api-0\" (UID: \"d1368adf-8a2b-41ff-987d-806f65c1b128\") " pod="openstack/glance-default-external-api-0" Oct 14 07:20:05 crc kubenswrapper[4870]: I1014 07:20:05.751911 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/99fa3d0e-8c4f-4893-b2fa-040837f04767-logs\") pod \"glance-default-internal-api-0\" (UID: \"99fa3d0e-8c4f-4893-b2fa-040837f04767\") " pod="openstack/glance-default-internal-api-0" Oct 14 07:20:05 crc kubenswrapper[4870]: I1014 07:20:05.751930 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d1368adf-8a2b-41ff-987d-806f65c1b128-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"d1368adf-8a2b-41ff-987d-806f65c1b128\") " pod="openstack/glance-default-external-api-0" Oct 14 07:20:05 crc kubenswrapper[4870]: I1014 07:20:05.751950 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/99fa3d0e-8c4f-4893-b2fa-040837f04767-config-data\") pod \"glance-default-internal-api-0\" (UID: \"99fa3d0e-8c4f-4893-b2fa-040837f04767\") " pod="openstack/glance-default-internal-api-0" Oct 14 07:20:05 crc kubenswrapper[4870]: I1014 07:20:05.751982 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99fa3d0e-8c4f-4893-b2fa-040837f04767-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"99fa3d0e-8c4f-4893-b2fa-040837f04767\") " pod="openstack/glance-default-internal-api-0" Oct 14 07:20:05 crc kubenswrapper[4870]: I1014 07:20:05.752287 4870 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"d1368adf-8a2b-41ff-987d-806f65c1b128\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/glance-default-external-api-0" Oct 14 07:20:05 crc kubenswrapper[4870]: I1014 07:20:05.770190 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d1368adf-8a2b-41ff-987d-806f65c1b128-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"d1368adf-8a2b-41ff-987d-806f65c1b128\") " pod="openstack/glance-default-external-api-0" Oct 14 07:20:05 crc kubenswrapper[4870]: I1014 07:20:05.772298 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d1368adf-8a2b-41ff-987d-806f65c1b128-logs\") pod \"glance-default-external-api-0\" (UID: \"d1368adf-8a2b-41ff-987d-806f65c1b128\") " pod="openstack/glance-default-external-api-0" Oct 14 07:20:05 crc kubenswrapper[4870]: I1014 07:20:05.780734 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d1368adf-8a2b-41ff-987d-806f65c1b128-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"d1368adf-8a2b-41ff-987d-806f65c1b128\") " pod="openstack/glance-default-external-api-0" Oct 14 07:20:05 crc kubenswrapper[4870]: I1014 07:20:05.786683 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d1368adf-8a2b-41ff-987d-806f65c1b128-config-data\") pod \"glance-default-external-api-0\" (UID: \"d1368adf-8a2b-41ff-987d-806f65c1b128\") " pod="openstack/glance-default-external-api-0" Oct 14 07:20:05 crc kubenswrapper[4870]: I1014 07:20:05.804141 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d1368adf-8a2b-41ff-987d-806f65c1b128-scripts\") pod \"glance-default-external-api-0\" (UID: \"d1368adf-8a2b-41ff-987d-806f65c1b128\") " pod="openstack/glance-default-external-api-0" Oct 14 07:20:05 crc kubenswrapper[4870]: I1014 07:20:05.804170 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l8g2b\" (UniqueName: \"kubernetes.io/projected/d1368adf-8a2b-41ff-987d-806f65c1b128-kube-api-access-l8g2b\") pod \"glance-default-external-api-0\" (UID: \"d1368adf-8a2b-41ff-987d-806f65c1b128\") " pod="openstack/glance-default-external-api-0" Oct 14 07:20:05 crc kubenswrapper[4870]: I1014 07:20:05.833677 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"d1368adf-8a2b-41ff-987d-806f65c1b128\") " pod="openstack/glance-default-external-api-0" Oct 14 07:20:05 crc kubenswrapper[4870]: I1014 07:20:05.856282 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/99fa3d0e-8c4f-4893-b2fa-040837f04767-logs\") pod \"glance-default-internal-api-0\" (UID: \"99fa3d0e-8c4f-4893-b2fa-040837f04767\") " pod="openstack/glance-default-internal-api-0" Oct 14 07:20:05 crc kubenswrapper[4870]: I1014 07:20:05.856333 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/99fa3d0e-8c4f-4893-b2fa-040837f04767-config-data\") pod \"glance-default-internal-api-0\" (UID: \"99fa3d0e-8c4f-4893-b2fa-040837f04767\") " pod="openstack/glance-default-internal-api-0" Oct 14 07:20:05 crc kubenswrapper[4870]: I1014 07:20:05.856362 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99fa3d0e-8c4f-4893-b2fa-040837f04767-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"99fa3d0e-8c4f-4893-b2fa-040837f04767\") " pod="openstack/glance-default-internal-api-0" Oct 14 07:20:05 crc kubenswrapper[4870]: I1014 07:20:05.856400 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/99fa3d0e-8c4f-4893-b2fa-040837f04767-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"99fa3d0e-8c4f-4893-b2fa-040837f04767\") " pod="openstack/glance-default-internal-api-0" Oct 14 07:20:05 crc kubenswrapper[4870]: I1014 07:20:05.856455 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"99fa3d0e-8c4f-4893-b2fa-040837f04767\") " pod="openstack/glance-default-internal-api-0" Oct 14 07:20:05 crc kubenswrapper[4870]: I1014 07:20:05.856519 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z6l65\" (UniqueName: \"kubernetes.io/projected/99fa3d0e-8c4f-4893-b2fa-040837f04767-kube-api-access-z6l65\") pod \"glance-default-internal-api-0\" (UID: \"99fa3d0e-8c4f-4893-b2fa-040837f04767\") " pod="openstack/glance-default-internal-api-0" Oct 14 07:20:05 crc kubenswrapper[4870]: I1014 07:20:05.856554 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/99fa3d0e-8c4f-4893-b2fa-040837f04767-scripts\") pod \"glance-default-internal-api-0\" (UID: \"99fa3d0e-8c4f-4893-b2fa-040837f04767\") " pod="openstack/glance-default-internal-api-0" Oct 14 07:20:05 crc kubenswrapper[4870]: I1014 07:20:05.856852 4870 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"99fa3d0e-8c4f-4893-b2fa-040837f04767\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/glance-default-internal-api-0" Oct 14 07:20:05 crc kubenswrapper[4870]: I1014 07:20:05.862901 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/99fa3d0e-8c4f-4893-b2fa-040837f04767-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"99fa3d0e-8c4f-4893-b2fa-040837f04767\") " pod="openstack/glance-default-internal-api-0" Oct 14 07:20:05 crc kubenswrapper[4870]: I1014 07:20:05.862910 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/99fa3d0e-8c4f-4893-b2fa-040837f04767-logs\") pod \"glance-default-internal-api-0\" (UID: \"99fa3d0e-8c4f-4893-b2fa-040837f04767\") " pod="openstack/glance-default-internal-api-0" Oct 14 07:20:05 crc kubenswrapper[4870]: I1014 07:20:05.871754 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/99fa3d0e-8c4f-4893-b2fa-040837f04767-config-data\") pod \"glance-default-internal-api-0\" (UID: \"99fa3d0e-8c4f-4893-b2fa-040837f04767\") " pod="openstack/glance-default-internal-api-0" Oct 14 07:20:05 crc kubenswrapper[4870]: I1014 07:20:05.873490 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99fa3d0e-8c4f-4893-b2fa-040837f04767-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"99fa3d0e-8c4f-4893-b2fa-040837f04767\") " pod="openstack/glance-default-internal-api-0" Oct 14 07:20:05 crc kubenswrapper[4870]: I1014 07:20:05.873958 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 14 07:20:05 crc kubenswrapper[4870]: I1014 07:20:05.944727 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7bcdb8c767-9pwfh"] Oct 14 07:20:05 crc kubenswrapper[4870]: I1014 07:20:05.946252 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z6l65\" (UniqueName: \"kubernetes.io/projected/99fa3d0e-8c4f-4893-b2fa-040837f04767-kube-api-access-z6l65\") pod \"glance-default-internal-api-0\" (UID: \"99fa3d0e-8c4f-4893-b2fa-040837f04767\") " pod="openstack/glance-default-internal-api-0" Oct 14 07:20:05 crc kubenswrapper[4870]: I1014 07:20:05.948948 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/99fa3d0e-8c4f-4893-b2fa-040837f04767-scripts\") pod \"glance-default-internal-api-0\" (UID: \"99fa3d0e-8c4f-4893-b2fa-040837f04767\") " pod="openstack/glance-default-internal-api-0" Oct 14 07:20:05 crc kubenswrapper[4870]: I1014 07:20:05.969011 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"99fa3d0e-8c4f-4893-b2fa-040837f04767\") " pod="openstack/glance-default-internal-api-0" Oct 14 07:20:06 crc kubenswrapper[4870]: I1014 07:20:06.103587 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cb67568b9-c75vp" Oct 14 07:20:06 crc kubenswrapper[4870]: I1014 07:20:06.161425 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3bb9dec0-1879-419a-ab06-4fbdf87f75af-config\") pod \"3bb9dec0-1879-419a-ab06-4fbdf87f75af\" (UID: \"3bb9dec0-1879-419a-ab06-4fbdf87f75af\") " Oct 14 07:20:06 crc kubenswrapper[4870]: I1014 07:20:06.161490 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-plgv5\" (UniqueName: \"kubernetes.io/projected/3bb9dec0-1879-419a-ab06-4fbdf87f75af-kube-api-access-plgv5\") pod \"3bb9dec0-1879-419a-ab06-4fbdf87f75af\" (UID: \"3bb9dec0-1879-419a-ab06-4fbdf87f75af\") " Oct 14 07:20:06 crc kubenswrapper[4870]: I1014 07:20:06.161518 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3bb9dec0-1879-419a-ab06-4fbdf87f75af-ovsdbserver-sb\") pod \"3bb9dec0-1879-419a-ab06-4fbdf87f75af\" (UID: \"3bb9dec0-1879-419a-ab06-4fbdf87f75af\") " Oct 14 07:20:06 crc kubenswrapper[4870]: I1014 07:20:06.161571 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3bb9dec0-1879-419a-ab06-4fbdf87f75af-dns-svc\") pod \"3bb9dec0-1879-419a-ab06-4fbdf87f75af\" (UID: \"3bb9dec0-1879-419a-ab06-4fbdf87f75af\") " Oct 14 07:20:06 crc kubenswrapper[4870]: I1014 07:20:06.161596 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3bb9dec0-1879-419a-ab06-4fbdf87f75af-dns-swift-storage-0\") pod \"3bb9dec0-1879-419a-ab06-4fbdf87f75af\" (UID: \"3bb9dec0-1879-419a-ab06-4fbdf87f75af\") " Oct 14 07:20:06 crc kubenswrapper[4870]: I1014 07:20:06.161676 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3bb9dec0-1879-419a-ab06-4fbdf87f75af-ovsdbserver-nb\") pod \"3bb9dec0-1879-419a-ab06-4fbdf87f75af\" (UID: \"3bb9dec0-1879-419a-ab06-4fbdf87f75af\") " Oct 14 07:20:06 crc kubenswrapper[4870]: I1014 07:20:06.191420 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3bb9dec0-1879-419a-ab06-4fbdf87f75af-kube-api-access-plgv5" (OuterVolumeSpecName: "kube-api-access-plgv5") pod "3bb9dec0-1879-419a-ab06-4fbdf87f75af" (UID: "3bb9dec0-1879-419a-ab06-4fbdf87f75af"). InnerVolumeSpecName "kube-api-access-plgv5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:20:06 crc kubenswrapper[4870]: I1014 07:20:06.217700 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-s2xm6" event={"ID":"6d5f827b-9c3f-4d17-b6a6-506fe3d2e06f","Type":"ContainerStarted","Data":"18ea48f83f0697a7940d336acee85144473be5195048c766e3c19cdbc705c74d"} Oct 14 07:20:06 crc kubenswrapper[4870]: I1014 07:20:06.245775 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 14 07:20:06 crc kubenswrapper[4870]: I1014 07:20:06.246540 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7bcdb8c767-9pwfh" event={"ID":"32e2388e-2670-44d3-bfe1-6ce4c126eb18","Type":"ContainerStarted","Data":"7127f9e04251e62c25047d8f4a4b0eaa991272ab883a8e44991d474cf2b19e1e"} Oct 14 07:20:06 crc kubenswrapper[4870]: I1014 07:20:06.248905 4870 generic.go:334] "Generic (PLEG): container finished" podID="3bb9dec0-1879-419a-ab06-4fbdf87f75af" containerID="47d72ed3aece21a46c275a0da7700dacf90fa3ce5ad56bf25032383083c026bc" exitCode=0 Oct 14 07:20:06 crc kubenswrapper[4870]: I1014 07:20:06.248968 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cb67568b9-c75vp" event={"ID":"3bb9dec0-1879-419a-ab06-4fbdf87f75af","Type":"ContainerDied","Data":"47d72ed3aece21a46c275a0da7700dacf90fa3ce5ad56bf25032383083c026bc"} Oct 14 07:20:06 crc kubenswrapper[4870]: I1014 07:20:06.248987 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cb67568b9-c75vp" event={"ID":"3bb9dec0-1879-419a-ab06-4fbdf87f75af","Type":"ContainerDied","Data":"dbc1f6adb0e4395d7859cff4dd784d12ee5763e3bd8807bd26f1be8f4fd87153"} Oct 14 07:20:06 crc kubenswrapper[4870]: I1014 07:20:06.249003 4870 scope.go:117] "RemoveContainer" containerID="47d72ed3aece21a46c275a0da7700dacf90fa3ce5ad56bf25032383083c026bc" Oct 14 07:20:06 crc kubenswrapper[4870]: I1014 07:20:06.248997 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cb67568b9-c75vp" Oct 14 07:20:06 crc kubenswrapper[4870]: I1014 07:20:06.251581 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0a31f5df-eae7-4087-a8c8-09c18f5ee9f0","Type":"ContainerStarted","Data":"18d9eed8f3dddc49cbc986fb2c0eaf635de30dcdadeaba0a0fea12c3c5b66991"} Oct 14 07:20:06 crc kubenswrapper[4870]: I1014 07:20:06.255801 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5885bcf447-68v4h" Oct 14 07:20:06 crc kubenswrapper[4870]: I1014 07:20:06.255797 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-nmtf9" event={"ID":"9dd93b33-f231-4339-9db0-0507cde1dcf5","Type":"ContainerStarted","Data":"45ff0c95eb09cc7c56cbb64dd645167ccb647aa60075375000048be59874f93b"} Oct 14 07:20:06 crc kubenswrapper[4870]: I1014 07:20:06.264218 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-plgv5\" (UniqueName: \"kubernetes.io/projected/3bb9dec0-1879-419a-ab06-4fbdf87f75af-kube-api-access-plgv5\") on node \"crc\" DevicePath \"\"" Oct 14 07:20:06 crc kubenswrapper[4870]: I1014 07:20:06.306782 4870 scope.go:117] "RemoveContainer" containerID="eae46e7838df613139fc0c847c23394c91965a2d5b00fea5ae6a4be757b4138a" Oct 14 07:20:06 crc kubenswrapper[4870]: I1014 07:20:06.326402 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5885bcf447-68v4h"] Oct 14 07:20:06 crc kubenswrapper[4870]: I1014 07:20:06.336175 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3bb9dec0-1879-419a-ab06-4fbdf87f75af-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "3bb9dec0-1879-419a-ab06-4fbdf87f75af" (UID: "3bb9dec0-1879-419a-ab06-4fbdf87f75af"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:20:06 crc kubenswrapper[4870]: I1014 07:20:06.353509 4870 scope.go:117] "RemoveContainer" containerID="47d72ed3aece21a46c275a0da7700dacf90fa3ce5ad56bf25032383083c026bc" Oct 14 07:20:06 crc kubenswrapper[4870]: E1014 07:20:06.353971 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"47d72ed3aece21a46c275a0da7700dacf90fa3ce5ad56bf25032383083c026bc\": container with ID starting with 47d72ed3aece21a46c275a0da7700dacf90fa3ce5ad56bf25032383083c026bc not found: ID does not exist" containerID="47d72ed3aece21a46c275a0da7700dacf90fa3ce5ad56bf25032383083c026bc" Oct 14 07:20:06 crc kubenswrapper[4870]: I1014 07:20:06.354006 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"47d72ed3aece21a46c275a0da7700dacf90fa3ce5ad56bf25032383083c026bc"} err="failed to get container status \"47d72ed3aece21a46c275a0da7700dacf90fa3ce5ad56bf25032383083c026bc\": rpc error: code = NotFound desc = could not find container \"47d72ed3aece21a46c275a0da7700dacf90fa3ce5ad56bf25032383083c026bc\": container with ID starting with 47d72ed3aece21a46c275a0da7700dacf90fa3ce5ad56bf25032383083c026bc not found: ID does not exist" Oct 14 07:20:06 crc kubenswrapper[4870]: I1014 07:20:06.354040 4870 scope.go:117] "RemoveContainer" containerID="eae46e7838df613139fc0c847c23394c91965a2d5b00fea5ae6a4be757b4138a" Oct 14 07:20:06 crc kubenswrapper[4870]: E1014 07:20:06.354316 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eae46e7838df613139fc0c847c23394c91965a2d5b00fea5ae6a4be757b4138a\": container with ID starting with eae46e7838df613139fc0c847c23394c91965a2d5b00fea5ae6a4be757b4138a not found: ID does not exist" containerID="eae46e7838df613139fc0c847c23394c91965a2d5b00fea5ae6a4be757b4138a" Oct 14 07:20:06 crc kubenswrapper[4870]: I1014 07:20:06.354345 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eae46e7838df613139fc0c847c23394c91965a2d5b00fea5ae6a4be757b4138a"} err="failed to get container status \"eae46e7838df613139fc0c847c23394c91965a2d5b00fea5ae6a4be757b4138a\": rpc error: code = NotFound desc = could not find container \"eae46e7838df613139fc0c847c23394c91965a2d5b00fea5ae6a4be757b4138a\": container with ID starting with eae46e7838df613139fc0c847c23394c91965a2d5b00fea5ae6a4be757b4138a not found: ID does not exist" Oct 14 07:20:06 crc kubenswrapper[4870]: I1014 07:20:06.355479 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5885bcf447-68v4h"] Oct 14 07:20:06 crc kubenswrapper[4870]: I1014 07:20:06.356124 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3bb9dec0-1879-419a-ab06-4fbdf87f75af-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "3bb9dec0-1879-419a-ab06-4fbdf87f75af" (UID: "3bb9dec0-1879-419a-ab06-4fbdf87f75af"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:20:06 crc kubenswrapper[4870]: I1014 07:20:06.365825 4870 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3bb9dec0-1879-419a-ab06-4fbdf87f75af-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 14 07:20:06 crc kubenswrapper[4870]: I1014 07:20:06.365981 4870 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3bb9dec0-1879-419a-ab06-4fbdf87f75af-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 14 07:20:06 crc kubenswrapper[4870]: I1014 07:20:06.365835 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3bb9dec0-1879-419a-ab06-4fbdf87f75af-config" (OuterVolumeSpecName: "config") pod "3bb9dec0-1879-419a-ab06-4fbdf87f75af" (UID: "3bb9dec0-1879-419a-ab06-4fbdf87f75af"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:20:06 crc kubenswrapper[4870]: I1014 07:20:06.367561 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3bb9dec0-1879-419a-ab06-4fbdf87f75af-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "3bb9dec0-1879-419a-ab06-4fbdf87f75af" (UID: "3bb9dec0-1879-419a-ab06-4fbdf87f75af"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:20:06 crc kubenswrapper[4870]: I1014 07:20:06.372617 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3bb9dec0-1879-419a-ab06-4fbdf87f75af-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "3bb9dec0-1879-419a-ab06-4fbdf87f75af" (UID: "3bb9dec0-1879-419a-ab06-4fbdf87f75af"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:20:06 crc kubenswrapper[4870]: I1014 07:20:06.470466 4870 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3bb9dec0-1879-419a-ab06-4fbdf87f75af-config\") on node \"crc\" DevicePath \"\"" Oct 14 07:20:06 crc kubenswrapper[4870]: I1014 07:20:06.470505 4870 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3bb9dec0-1879-419a-ab06-4fbdf87f75af-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 14 07:20:06 crc kubenswrapper[4870]: I1014 07:20:06.470519 4870 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3bb9dec0-1879-419a-ab06-4fbdf87f75af-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 14 07:20:06 crc kubenswrapper[4870]: I1014 07:20:06.473974 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 14 07:20:06 crc kubenswrapper[4870]: I1014 07:20:06.589753 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-cb67568b9-c75vp"] Oct 14 07:20:06 crc kubenswrapper[4870]: I1014 07:20:06.598086 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-cb67568b9-c75vp"] Oct 14 07:20:06 crc kubenswrapper[4870]: I1014 07:20:06.944985 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 14 07:20:07 crc kubenswrapper[4870]: I1014 07:20:07.050697 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3bb9dec0-1879-419a-ab06-4fbdf87f75af" path="/var/lib/kubelet/pods/3bb9dec0-1879-419a-ab06-4fbdf87f75af/volumes" Oct 14 07:20:07 crc kubenswrapper[4870]: I1014 07:20:07.051573 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d94da551-da6b-48b1-ac7d-a78f06715651" path="/var/lib/kubelet/pods/d94da551-da6b-48b1-ac7d-a78f06715651/volumes" Oct 14 07:20:07 crc kubenswrapper[4870]: I1014 07:20:07.272408 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"99fa3d0e-8c4f-4893-b2fa-040837f04767","Type":"ContainerStarted","Data":"f1a7b415ab6c6183f0cc4d7292a82a53b740e339e880159e90839c2922b8df52"} Oct 14 07:20:07 crc kubenswrapper[4870]: I1014 07:20:07.274092 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d1368adf-8a2b-41ff-987d-806f65c1b128","Type":"ContainerStarted","Data":"c1c7e6b321e0f5505182cedf957020d5a101cfb5be8ac4c352ed3f33595ba3ac"} Oct 14 07:20:07 crc kubenswrapper[4870]: I1014 07:20:07.274115 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d1368adf-8a2b-41ff-987d-806f65c1b128","Type":"ContainerStarted","Data":"158f8751d0b545f4bec27b23ef9f7e90138acfdf3daae3391cc214f60a503c92"} Oct 14 07:20:07 crc kubenswrapper[4870]: I1014 07:20:07.276460 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-s2xm6" event={"ID":"6d5f827b-9c3f-4d17-b6a6-506fe3d2e06f","Type":"ContainerStarted","Data":"4ffe098a8e8e4c9ba69a5357443ce3b51e9f90cd094ae3dbb2018a4ad085b163"} Oct 14 07:20:07 crc kubenswrapper[4870]: I1014 07:20:07.280396 4870 generic.go:334] "Generic (PLEG): container finished" podID="32e2388e-2670-44d3-bfe1-6ce4c126eb18" containerID="84dddb0e2b529d2169d4dcc0f253839f94043cb699b4ebe505daff1715e343ff" exitCode=0 Oct 14 07:20:07 crc kubenswrapper[4870]: I1014 07:20:07.280480 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7bcdb8c767-9pwfh" event={"ID":"32e2388e-2670-44d3-bfe1-6ce4c126eb18","Type":"ContainerDied","Data":"84dddb0e2b529d2169d4dcc0f253839f94043cb699b4ebe505daff1715e343ff"} Oct 14 07:20:07 crc kubenswrapper[4870]: I1014 07:20:07.299365 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-s2xm6" podStartSLOduration=3.299341365 podStartE2EDuration="3.299341365s" podCreationTimestamp="2025-10-14 07:20:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:20:07.291829749 +0000 UTC m=+1142.989190130" watchObservedRunningTime="2025-10-14 07:20:07.299341365 +0000 UTC m=+1142.996701736" Oct 14 07:20:08 crc kubenswrapper[4870]: I1014 07:20:08.003565 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-5c98-account-create-9bjhh"] Oct 14 07:20:08 crc kubenswrapper[4870]: E1014 07:20:08.004687 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3bb9dec0-1879-419a-ab06-4fbdf87f75af" containerName="dnsmasq-dns" Oct 14 07:20:08 crc kubenswrapper[4870]: I1014 07:20:08.004702 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="3bb9dec0-1879-419a-ab06-4fbdf87f75af" containerName="dnsmasq-dns" Oct 14 07:20:08 crc kubenswrapper[4870]: E1014 07:20:08.004817 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3bb9dec0-1879-419a-ab06-4fbdf87f75af" containerName="init" Oct 14 07:20:08 crc kubenswrapper[4870]: I1014 07:20:08.004825 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="3bb9dec0-1879-419a-ab06-4fbdf87f75af" containerName="init" Oct 14 07:20:08 crc kubenswrapper[4870]: I1014 07:20:08.005874 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="3bb9dec0-1879-419a-ab06-4fbdf87f75af" containerName="dnsmasq-dns" Oct 14 07:20:08 crc kubenswrapper[4870]: I1014 07:20:08.007011 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-5c98-account-create-9bjhh" Oct 14 07:20:08 crc kubenswrapper[4870]: I1014 07:20:08.027552 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Oct 14 07:20:08 crc kubenswrapper[4870]: I1014 07:20:08.056063 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-5c98-account-create-9bjhh"] Oct 14 07:20:08 crc kubenswrapper[4870]: I1014 07:20:08.099618 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-6547-account-create-vqn99"] Oct 14 07:20:08 crc kubenswrapper[4870]: I1014 07:20:08.100749 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-6547-account-create-vqn99" Oct 14 07:20:08 crc kubenswrapper[4870]: I1014 07:20:08.105053 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Oct 14 07:20:08 crc kubenswrapper[4870]: I1014 07:20:08.105919 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2kgdh\" (UniqueName: \"kubernetes.io/projected/5056171a-6fc0-4d48-9ff0-e275137253f3-kube-api-access-2kgdh\") pod \"cinder-5c98-account-create-9bjhh\" (UID: \"5056171a-6fc0-4d48-9ff0-e275137253f3\") " pod="openstack/cinder-5c98-account-create-9bjhh" Oct 14 07:20:08 crc kubenswrapper[4870]: I1014 07:20:08.105950 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-6547-account-create-vqn99"] Oct 14 07:20:08 crc kubenswrapper[4870]: I1014 07:20:08.207546 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2kgdh\" (UniqueName: \"kubernetes.io/projected/5056171a-6fc0-4d48-9ff0-e275137253f3-kube-api-access-2kgdh\") pod \"cinder-5c98-account-create-9bjhh\" (UID: \"5056171a-6fc0-4d48-9ff0-e275137253f3\") " pod="openstack/cinder-5c98-account-create-9bjhh" Oct 14 07:20:08 crc kubenswrapper[4870]: I1014 07:20:08.207923 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-92mbl\" (UniqueName: \"kubernetes.io/projected/625bb722-0865-47f2-b680-ab642ee6c9e5-kube-api-access-92mbl\") pod \"barbican-6547-account-create-vqn99\" (UID: \"625bb722-0865-47f2-b680-ab642ee6c9e5\") " pod="openstack/barbican-6547-account-create-vqn99" Oct 14 07:20:08 crc kubenswrapper[4870]: I1014 07:20:08.251247 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2kgdh\" (UniqueName: \"kubernetes.io/projected/5056171a-6fc0-4d48-9ff0-e275137253f3-kube-api-access-2kgdh\") pod \"cinder-5c98-account-create-9bjhh\" (UID: \"5056171a-6fc0-4d48-9ff0-e275137253f3\") " pod="openstack/cinder-5c98-account-create-9bjhh" Oct 14 07:20:08 crc kubenswrapper[4870]: I1014 07:20:08.302093 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"99fa3d0e-8c4f-4893-b2fa-040837f04767","Type":"ContainerStarted","Data":"3294d0dc2862fd311c8ecc3a5047b4c9f36b8302327a52846a6bcd07e3372697"} Oct 14 07:20:08 crc kubenswrapper[4870]: I1014 07:20:08.306299 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7bcdb8c767-9pwfh" event={"ID":"32e2388e-2670-44d3-bfe1-6ce4c126eb18","Type":"ContainerStarted","Data":"ea7ea208868fa97909c709c50154f181f12abb3d19b900341576486b790d35d4"} Oct 14 07:20:08 crc kubenswrapper[4870]: I1014 07:20:08.306408 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7bcdb8c767-9pwfh" Oct 14 07:20:08 crc kubenswrapper[4870]: I1014 07:20:08.309498 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-92mbl\" (UniqueName: \"kubernetes.io/projected/625bb722-0865-47f2-b680-ab642ee6c9e5-kube-api-access-92mbl\") pod \"barbican-6547-account-create-vqn99\" (UID: \"625bb722-0865-47f2-b680-ab642ee6c9e5\") " pod="openstack/barbican-6547-account-create-vqn99" Oct 14 07:20:08 crc kubenswrapper[4870]: I1014 07:20:08.333944 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7bcdb8c767-9pwfh" podStartSLOduration=4.333928776 podStartE2EDuration="4.333928776s" podCreationTimestamp="2025-10-14 07:20:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:20:08.329273131 +0000 UTC m=+1144.026633502" watchObservedRunningTime="2025-10-14 07:20:08.333928776 +0000 UTC m=+1144.031289147" Oct 14 07:20:08 crc kubenswrapper[4870]: I1014 07:20:08.334251 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-5c98-account-create-9bjhh" Oct 14 07:20:08 crc kubenswrapper[4870]: I1014 07:20:08.336321 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-92mbl\" (UniqueName: \"kubernetes.io/projected/625bb722-0865-47f2-b680-ab642ee6c9e5-kube-api-access-92mbl\") pod \"barbican-6547-account-create-vqn99\" (UID: \"625bb722-0865-47f2-b680-ab642ee6c9e5\") " pod="openstack/barbican-6547-account-create-vqn99" Oct 14 07:20:08 crc kubenswrapper[4870]: I1014 07:20:08.395068 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-cedd-account-create-gknx9"] Oct 14 07:20:08 crc kubenswrapper[4870]: I1014 07:20:08.397399 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-cedd-account-create-gknx9" Oct 14 07:20:08 crc kubenswrapper[4870]: I1014 07:20:08.399863 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Oct 14 07:20:08 crc kubenswrapper[4870]: I1014 07:20:08.409963 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-cedd-account-create-gknx9"] Oct 14 07:20:08 crc kubenswrapper[4870]: I1014 07:20:08.424716 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-6547-account-create-vqn99" Oct 14 07:20:08 crc kubenswrapper[4870]: I1014 07:20:08.512820 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jvr77\" (UniqueName: \"kubernetes.io/projected/f5e5bd22-1692-477a-a375-3eecad66c746-kube-api-access-jvr77\") pod \"neutron-cedd-account-create-gknx9\" (UID: \"f5e5bd22-1692-477a-a375-3eecad66c746\") " pod="openstack/neutron-cedd-account-create-gknx9" Oct 14 07:20:08 crc kubenswrapper[4870]: I1014 07:20:08.588671 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 14 07:20:08 crc kubenswrapper[4870]: I1014 07:20:08.615370 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jvr77\" (UniqueName: \"kubernetes.io/projected/f5e5bd22-1692-477a-a375-3eecad66c746-kube-api-access-jvr77\") pod \"neutron-cedd-account-create-gknx9\" (UID: \"f5e5bd22-1692-477a-a375-3eecad66c746\") " pod="openstack/neutron-cedd-account-create-gknx9" Oct 14 07:20:08 crc kubenswrapper[4870]: I1014 07:20:08.661057 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jvr77\" (UniqueName: \"kubernetes.io/projected/f5e5bd22-1692-477a-a375-3eecad66c746-kube-api-access-jvr77\") pod \"neutron-cedd-account-create-gknx9\" (UID: \"f5e5bd22-1692-477a-a375-3eecad66c746\") " pod="openstack/neutron-cedd-account-create-gknx9" Oct 14 07:20:08 crc kubenswrapper[4870]: I1014 07:20:08.670961 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 14 07:20:08 crc kubenswrapper[4870]: I1014 07:20:08.736834 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 14 07:20:08 crc kubenswrapper[4870]: I1014 07:20:08.780686 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-cedd-account-create-gknx9" Oct 14 07:20:09 crc kubenswrapper[4870]: I1014 07:20:09.073864 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-6547-account-create-vqn99"] Oct 14 07:20:09 crc kubenswrapper[4870]: I1014 07:20:09.138377 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-5c98-account-create-9bjhh"] Oct 14 07:20:09 crc kubenswrapper[4870]: I1014 07:20:09.315274 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"99fa3d0e-8c4f-4893-b2fa-040837f04767","Type":"ContainerStarted","Data":"5d6e51d53c3a4412c7e245a2bc027b8b0885f7aa76c0eba1a2678cb04ac541aa"} Oct 14 07:20:09 crc kubenswrapper[4870]: I1014 07:20:09.315378 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="99fa3d0e-8c4f-4893-b2fa-040837f04767" containerName="glance-log" containerID="cri-o://3294d0dc2862fd311c8ecc3a5047b4c9f36b8302327a52846a6bcd07e3372697" gracePeriod=30 Oct 14 07:20:09 crc kubenswrapper[4870]: I1014 07:20:09.315537 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="99fa3d0e-8c4f-4893-b2fa-040837f04767" containerName="glance-httpd" containerID="cri-o://5d6e51d53c3a4412c7e245a2bc027b8b0885f7aa76c0eba1a2678cb04ac541aa" gracePeriod=30 Oct 14 07:20:09 crc kubenswrapper[4870]: I1014 07:20:09.323030 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d1368adf-8a2b-41ff-987d-806f65c1b128","Type":"ContainerStarted","Data":"4101a1878aa0de0b33fecc28940880f2820ff64654370da2a14b7694f6d7d146"} Oct 14 07:20:09 crc kubenswrapper[4870]: I1014 07:20:09.323261 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="d1368adf-8a2b-41ff-987d-806f65c1b128" containerName="glance-log" containerID="cri-o://c1c7e6b321e0f5505182cedf957020d5a101cfb5be8ac4c352ed3f33595ba3ac" gracePeriod=30 Oct 14 07:20:09 crc kubenswrapper[4870]: I1014 07:20:09.323464 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="d1368adf-8a2b-41ff-987d-806f65c1b128" containerName="glance-httpd" containerID="cri-o://4101a1878aa0de0b33fecc28940880f2820ff64654370da2a14b7694f6d7d146" gracePeriod=30 Oct 14 07:20:09 crc kubenswrapper[4870]: I1014 07:20:09.335421 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=5.335408918 podStartE2EDuration="5.335408918s" podCreationTimestamp="2025-10-14 07:20:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:20:09.335263805 +0000 UTC m=+1145.032624176" watchObservedRunningTime="2025-10-14 07:20:09.335408918 +0000 UTC m=+1145.032769289" Oct 14 07:20:09 crc kubenswrapper[4870]: I1014 07:20:09.367024 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=5.36700749 podStartE2EDuration="5.36700749s" podCreationTimestamp="2025-10-14 07:20:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:20:09.365106693 +0000 UTC m=+1145.062467064" watchObservedRunningTime="2025-10-14 07:20:09.36700749 +0000 UTC m=+1145.064367861" Oct 14 07:20:10 crc kubenswrapper[4870]: I1014 07:20:10.346129 4870 generic.go:334] "Generic (PLEG): container finished" podID="6d5f827b-9c3f-4d17-b6a6-506fe3d2e06f" containerID="4ffe098a8e8e4c9ba69a5357443ce3b51e9f90cd094ae3dbb2018a4ad085b163" exitCode=0 Oct 14 07:20:10 crc kubenswrapper[4870]: I1014 07:20:10.346212 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-s2xm6" event={"ID":"6d5f827b-9c3f-4d17-b6a6-506fe3d2e06f","Type":"ContainerDied","Data":"4ffe098a8e8e4c9ba69a5357443ce3b51e9f90cd094ae3dbb2018a4ad085b163"} Oct 14 07:20:10 crc kubenswrapper[4870]: I1014 07:20:10.351015 4870 generic.go:334] "Generic (PLEG): container finished" podID="99fa3d0e-8c4f-4893-b2fa-040837f04767" containerID="5d6e51d53c3a4412c7e245a2bc027b8b0885f7aa76c0eba1a2678cb04ac541aa" exitCode=0 Oct 14 07:20:10 crc kubenswrapper[4870]: I1014 07:20:10.351038 4870 generic.go:334] "Generic (PLEG): container finished" podID="99fa3d0e-8c4f-4893-b2fa-040837f04767" containerID="3294d0dc2862fd311c8ecc3a5047b4c9f36b8302327a52846a6bcd07e3372697" exitCode=143 Oct 14 07:20:10 crc kubenswrapper[4870]: I1014 07:20:10.351094 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"99fa3d0e-8c4f-4893-b2fa-040837f04767","Type":"ContainerDied","Data":"5d6e51d53c3a4412c7e245a2bc027b8b0885f7aa76c0eba1a2678cb04ac541aa"} Oct 14 07:20:10 crc kubenswrapper[4870]: I1014 07:20:10.351117 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"99fa3d0e-8c4f-4893-b2fa-040837f04767","Type":"ContainerDied","Data":"3294d0dc2862fd311c8ecc3a5047b4c9f36b8302327a52846a6bcd07e3372697"} Oct 14 07:20:10 crc kubenswrapper[4870]: I1014 07:20:10.353693 4870 generic.go:334] "Generic (PLEG): container finished" podID="d1368adf-8a2b-41ff-987d-806f65c1b128" containerID="4101a1878aa0de0b33fecc28940880f2820ff64654370da2a14b7694f6d7d146" exitCode=0 Oct 14 07:20:10 crc kubenswrapper[4870]: I1014 07:20:10.353722 4870 generic.go:334] "Generic (PLEG): container finished" podID="d1368adf-8a2b-41ff-987d-806f65c1b128" containerID="c1c7e6b321e0f5505182cedf957020d5a101cfb5be8ac4c352ed3f33595ba3ac" exitCode=143 Oct 14 07:20:10 crc kubenswrapper[4870]: I1014 07:20:10.353746 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d1368adf-8a2b-41ff-987d-806f65c1b128","Type":"ContainerDied","Data":"4101a1878aa0de0b33fecc28940880f2820ff64654370da2a14b7694f6d7d146"} Oct 14 07:20:10 crc kubenswrapper[4870]: I1014 07:20:10.353769 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d1368adf-8a2b-41ff-987d-806f65c1b128","Type":"ContainerDied","Data":"c1c7e6b321e0f5505182cedf957020d5a101cfb5be8ac4c352ed3f33595ba3ac"} Oct 14 07:20:13 crc kubenswrapper[4870]: W1014 07:20:13.746927 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5056171a_6fc0_4d48_9ff0_e275137253f3.slice/crio-d8080344442285e06403465da4075fc5b7f452b9e48700d7b54d57f1bc6cfafe WatchSource:0}: Error finding container d8080344442285e06403465da4075fc5b7f452b9e48700d7b54d57f1bc6cfafe: Status 404 returned error can't find the container with id d8080344442285e06403465da4075fc5b7f452b9e48700d7b54d57f1bc6cfafe Oct 14 07:20:13 crc kubenswrapper[4870]: I1014 07:20:13.933203 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-s2xm6" Oct 14 07:20:13 crc kubenswrapper[4870]: I1014 07:20:13.938573 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 14 07:20:14 crc kubenswrapper[4870]: I1014 07:20:14.026616 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bvx4m\" (UniqueName: \"kubernetes.io/projected/6d5f827b-9c3f-4d17-b6a6-506fe3d2e06f-kube-api-access-bvx4m\") pod \"6d5f827b-9c3f-4d17-b6a6-506fe3d2e06f\" (UID: \"6d5f827b-9c3f-4d17-b6a6-506fe3d2e06f\") " Oct 14 07:20:14 crc kubenswrapper[4870]: I1014 07:20:14.026982 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"99fa3d0e-8c4f-4893-b2fa-040837f04767\" (UID: \"99fa3d0e-8c4f-4893-b2fa-040837f04767\") " Oct 14 07:20:14 crc kubenswrapper[4870]: I1014 07:20:14.027051 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/99fa3d0e-8c4f-4893-b2fa-040837f04767-logs\") pod \"99fa3d0e-8c4f-4893-b2fa-040837f04767\" (UID: \"99fa3d0e-8c4f-4893-b2fa-040837f04767\") " Oct 14 07:20:14 crc kubenswrapper[4870]: I1014 07:20:14.027100 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d5f827b-9c3f-4d17-b6a6-506fe3d2e06f-combined-ca-bundle\") pod \"6d5f827b-9c3f-4d17-b6a6-506fe3d2e06f\" (UID: \"6d5f827b-9c3f-4d17-b6a6-506fe3d2e06f\") " Oct 14 07:20:14 crc kubenswrapper[4870]: I1014 07:20:14.027138 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6d5f827b-9c3f-4d17-b6a6-506fe3d2e06f-config-data\") pod \"6d5f827b-9c3f-4d17-b6a6-506fe3d2e06f\" (UID: \"6d5f827b-9c3f-4d17-b6a6-506fe3d2e06f\") " Oct 14 07:20:14 crc kubenswrapper[4870]: I1014 07:20:14.027159 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/99fa3d0e-8c4f-4893-b2fa-040837f04767-httpd-run\") pod \"99fa3d0e-8c4f-4893-b2fa-040837f04767\" (UID: \"99fa3d0e-8c4f-4893-b2fa-040837f04767\") " Oct 14 07:20:14 crc kubenswrapper[4870]: I1014 07:20:14.027187 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/99fa3d0e-8c4f-4893-b2fa-040837f04767-scripts\") pod \"99fa3d0e-8c4f-4893-b2fa-040837f04767\" (UID: \"99fa3d0e-8c4f-4893-b2fa-040837f04767\") " Oct 14 07:20:14 crc kubenswrapper[4870]: I1014 07:20:14.027222 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6d5f827b-9c3f-4d17-b6a6-506fe3d2e06f-scripts\") pod \"6d5f827b-9c3f-4d17-b6a6-506fe3d2e06f\" (UID: \"6d5f827b-9c3f-4d17-b6a6-506fe3d2e06f\") " Oct 14 07:20:14 crc kubenswrapper[4870]: I1014 07:20:14.027260 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/6d5f827b-9c3f-4d17-b6a6-506fe3d2e06f-credential-keys\") pod \"6d5f827b-9c3f-4d17-b6a6-506fe3d2e06f\" (UID: \"6d5f827b-9c3f-4d17-b6a6-506fe3d2e06f\") " Oct 14 07:20:14 crc kubenswrapper[4870]: I1014 07:20:14.027284 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99fa3d0e-8c4f-4893-b2fa-040837f04767-combined-ca-bundle\") pod \"99fa3d0e-8c4f-4893-b2fa-040837f04767\" (UID: \"99fa3d0e-8c4f-4893-b2fa-040837f04767\") " Oct 14 07:20:14 crc kubenswrapper[4870]: I1014 07:20:14.027327 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/99fa3d0e-8c4f-4893-b2fa-040837f04767-config-data\") pod \"99fa3d0e-8c4f-4893-b2fa-040837f04767\" (UID: \"99fa3d0e-8c4f-4893-b2fa-040837f04767\") " Oct 14 07:20:14 crc kubenswrapper[4870]: I1014 07:20:14.027349 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z6l65\" (UniqueName: \"kubernetes.io/projected/99fa3d0e-8c4f-4893-b2fa-040837f04767-kube-api-access-z6l65\") pod \"99fa3d0e-8c4f-4893-b2fa-040837f04767\" (UID: \"99fa3d0e-8c4f-4893-b2fa-040837f04767\") " Oct 14 07:20:14 crc kubenswrapper[4870]: I1014 07:20:14.027371 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6d5f827b-9c3f-4d17-b6a6-506fe3d2e06f-fernet-keys\") pod \"6d5f827b-9c3f-4d17-b6a6-506fe3d2e06f\" (UID: \"6d5f827b-9c3f-4d17-b6a6-506fe3d2e06f\") " Oct 14 07:20:14 crc kubenswrapper[4870]: I1014 07:20:14.031552 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/99fa3d0e-8c4f-4893-b2fa-040837f04767-logs" (OuterVolumeSpecName: "logs") pod "99fa3d0e-8c4f-4893-b2fa-040837f04767" (UID: "99fa3d0e-8c4f-4893-b2fa-040837f04767"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:20:14 crc kubenswrapper[4870]: I1014 07:20:14.031771 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/99fa3d0e-8c4f-4893-b2fa-040837f04767-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "99fa3d0e-8c4f-4893-b2fa-040837f04767" (UID: "99fa3d0e-8c4f-4893-b2fa-040837f04767"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:20:14 crc kubenswrapper[4870]: I1014 07:20:14.031962 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6d5f827b-9c3f-4d17-b6a6-506fe3d2e06f-kube-api-access-bvx4m" (OuterVolumeSpecName: "kube-api-access-bvx4m") pod "6d5f827b-9c3f-4d17-b6a6-506fe3d2e06f" (UID: "6d5f827b-9c3f-4d17-b6a6-506fe3d2e06f"). InnerVolumeSpecName "kube-api-access-bvx4m". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:20:14 crc kubenswrapper[4870]: I1014 07:20:14.037861 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6d5f827b-9c3f-4d17-b6a6-506fe3d2e06f-scripts" (OuterVolumeSpecName: "scripts") pod "6d5f827b-9c3f-4d17-b6a6-506fe3d2e06f" (UID: "6d5f827b-9c3f-4d17-b6a6-506fe3d2e06f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:20:14 crc kubenswrapper[4870]: I1014 07:20:14.038771 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6d5f827b-9c3f-4d17-b6a6-506fe3d2e06f-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "6d5f827b-9c3f-4d17-b6a6-506fe3d2e06f" (UID: "6d5f827b-9c3f-4d17-b6a6-506fe3d2e06f"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:20:14 crc kubenswrapper[4870]: I1014 07:20:14.044538 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/99fa3d0e-8c4f-4893-b2fa-040837f04767-scripts" (OuterVolumeSpecName: "scripts") pod "99fa3d0e-8c4f-4893-b2fa-040837f04767" (UID: "99fa3d0e-8c4f-4893-b2fa-040837f04767"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:20:14 crc kubenswrapper[4870]: I1014 07:20:14.044763 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/99fa3d0e-8c4f-4893-b2fa-040837f04767-kube-api-access-z6l65" (OuterVolumeSpecName: "kube-api-access-z6l65") pod "99fa3d0e-8c4f-4893-b2fa-040837f04767" (UID: "99fa3d0e-8c4f-4893-b2fa-040837f04767"). InnerVolumeSpecName "kube-api-access-z6l65". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:20:14 crc kubenswrapper[4870]: I1014 07:20:14.045135 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6d5f827b-9c3f-4d17-b6a6-506fe3d2e06f-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "6d5f827b-9c3f-4d17-b6a6-506fe3d2e06f" (UID: "6d5f827b-9c3f-4d17-b6a6-506fe3d2e06f"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:20:14 crc kubenswrapper[4870]: I1014 07:20:14.047885 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage06-crc" (OuterVolumeSpecName: "glance") pod "99fa3d0e-8c4f-4893-b2fa-040837f04767" (UID: "99fa3d0e-8c4f-4893-b2fa-040837f04767"). InnerVolumeSpecName "local-storage06-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 14 07:20:14 crc kubenswrapper[4870]: I1014 07:20:14.072731 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6d5f827b-9c3f-4d17-b6a6-506fe3d2e06f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6d5f827b-9c3f-4d17-b6a6-506fe3d2e06f" (UID: "6d5f827b-9c3f-4d17-b6a6-506fe3d2e06f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:20:14 crc kubenswrapper[4870]: I1014 07:20:14.099643 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/99fa3d0e-8c4f-4893-b2fa-040837f04767-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "99fa3d0e-8c4f-4893-b2fa-040837f04767" (UID: "99fa3d0e-8c4f-4893-b2fa-040837f04767"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:20:14 crc kubenswrapper[4870]: I1014 07:20:14.129868 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bvx4m\" (UniqueName: \"kubernetes.io/projected/6d5f827b-9c3f-4d17-b6a6-506fe3d2e06f-kube-api-access-bvx4m\") on node \"crc\" DevicePath \"\"" Oct 14 07:20:14 crc kubenswrapper[4870]: I1014 07:20:14.129906 4870 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" " Oct 14 07:20:14 crc kubenswrapper[4870]: I1014 07:20:14.129937 4870 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/99fa3d0e-8c4f-4893-b2fa-040837f04767-logs\") on node \"crc\" DevicePath \"\"" Oct 14 07:20:14 crc kubenswrapper[4870]: I1014 07:20:14.129945 4870 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d5f827b-9c3f-4d17-b6a6-506fe3d2e06f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 07:20:14 crc kubenswrapper[4870]: I1014 07:20:14.129953 4870 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/99fa3d0e-8c4f-4893-b2fa-040837f04767-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 14 07:20:14 crc kubenswrapper[4870]: I1014 07:20:14.129963 4870 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/99fa3d0e-8c4f-4893-b2fa-040837f04767-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 07:20:14 crc kubenswrapper[4870]: I1014 07:20:14.129970 4870 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6d5f827b-9c3f-4d17-b6a6-506fe3d2e06f-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 07:20:14 crc kubenswrapper[4870]: I1014 07:20:14.129977 4870 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/6d5f827b-9c3f-4d17-b6a6-506fe3d2e06f-credential-keys\") on node \"crc\" DevicePath \"\"" Oct 14 07:20:14 crc kubenswrapper[4870]: I1014 07:20:14.129985 4870 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99fa3d0e-8c4f-4893-b2fa-040837f04767-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 07:20:14 crc kubenswrapper[4870]: I1014 07:20:14.130013 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z6l65\" (UniqueName: \"kubernetes.io/projected/99fa3d0e-8c4f-4893-b2fa-040837f04767-kube-api-access-z6l65\") on node \"crc\" DevicePath \"\"" Oct 14 07:20:14 crc kubenswrapper[4870]: I1014 07:20:14.130021 4870 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6d5f827b-9c3f-4d17-b6a6-506fe3d2e06f-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 14 07:20:14 crc kubenswrapper[4870]: I1014 07:20:14.214733 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6d5f827b-9c3f-4d17-b6a6-506fe3d2e06f-config-data" (OuterVolumeSpecName: "config-data") pod "6d5f827b-9c3f-4d17-b6a6-506fe3d2e06f" (UID: "6d5f827b-9c3f-4d17-b6a6-506fe3d2e06f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:20:14 crc kubenswrapper[4870]: I1014 07:20:14.231812 4870 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6d5f827b-9c3f-4d17-b6a6-506fe3d2e06f-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 07:20:14 crc kubenswrapper[4870]: I1014 07:20:14.242306 4870 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage06-crc" (UniqueName: "kubernetes.io/local-volume/local-storage06-crc") on node "crc" Oct 14 07:20:14 crc kubenswrapper[4870]: I1014 07:20:14.265961 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/99fa3d0e-8c4f-4893-b2fa-040837f04767-config-data" (OuterVolumeSpecName: "config-data") pod "99fa3d0e-8c4f-4893-b2fa-040837f04767" (UID: "99fa3d0e-8c4f-4893-b2fa-040837f04767"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:20:14 crc kubenswrapper[4870]: I1014 07:20:14.337312 4870 reconciler_common.go:293] "Volume detached for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" DevicePath \"\"" Oct 14 07:20:14 crc kubenswrapper[4870]: I1014 07:20:14.337346 4870 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/99fa3d0e-8c4f-4893-b2fa-040837f04767-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 07:20:14 crc kubenswrapper[4870]: I1014 07:20:14.382404 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 14 07:20:14 crc kubenswrapper[4870]: I1014 07:20:14.389013 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0a31f5df-eae7-4087-a8c8-09c18f5ee9f0","Type":"ContainerStarted","Data":"fb6bda07666ba00316d5344c82fd8d09dfbee1b602c4c88217bbfce61566b775"} Oct 14 07:20:14 crc kubenswrapper[4870]: I1014 07:20:14.390224 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-nmtf9" event={"ID":"9dd93b33-f231-4339-9db0-0507cde1dcf5","Type":"ContainerStarted","Data":"8f57f5ca62644a1c87bd38a50bcf2f2378a79660bfe53b23e9830b3861c114f8"} Oct 14 07:20:14 crc kubenswrapper[4870]: I1014 07:20:14.393149 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d1368adf-8a2b-41ff-987d-806f65c1b128","Type":"ContainerDied","Data":"158f8751d0b545f4bec27b23ef9f7e90138acfdf3daae3391cc214f60a503c92"} Oct 14 07:20:14 crc kubenswrapper[4870]: I1014 07:20:14.393190 4870 scope.go:117] "RemoveContainer" containerID="4101a1878aa0de0b33fecc28940880f2820ff64654370da2a14b7694f6d7d146" Oct 14 07:20:14 crc kubenswrapper[4870]: I1014 07:20:14.393308 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 14 07:20:14 crc kubenswrapper[4870]: I1014 07:20:14.395607 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-s2xm6" event={"ID":"6d5f827b-9c3f-4d17-b6a6-506fe3d2e06f","Type":"ContainerDied","Data":"18ea48f83f0697a7940d336acee85144473be5195048c766e3c19cdbc705c74d"} Oct 14 07:20:14 crc kubenswrapper[4870]: I1014 07:20:14.395636 4870 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="18ea48f83f0697a7940d336acee85144473be5195048c766e3c19cdbc705c74d" Oct 14 07:20:14 crc kubenswrapper[4870]: I1014 07:20:14.395684 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-s2xm6" Oct 14 07:20:14 crc kubenswrapper[4870]: I1014 07:20:14.412411 4870 generic.go:334] "Generic (PLEG): container finished" podID="625bb722-0865-47f2-b680-ab642ee6c9e5" containerID="87f2dc9b9b112ef49726f92a6ed22f035a2c6c59983215e902ee16f2b25877d9" exitCode=0 Oct 14 07:20:14 crc kubenswrapper[4870]: I1014 07:20:14.412482 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-6547-account-create-vqn99" event={"ID":"625bb722-0865-47f2-b680-ab642ee6c9e5","Type":"ContainerDied","Data":"87f2dc9b9b112ef49726f92a6ed22f035a2c6c59983215e902ee16f2b25877d9"} Oct 14 07:20:14 crc kubenswrapper[4870]: I1014 07:20:14.412506 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-6547-account-create-vqn99" event={"ID":"625bb722-0865-47f2-b680-ab642ee6c9e5","Type":"ContainerStarted","Data":"88ef43717fb61c09aad5b7f62d7199e1c16aa7a44c802f9ceda0b4bc9630d483"} Oct 14 07:20:14 crc kubenswrapper[4870]: I1014 07:20:14.430302 4870 generic.go:334] "Generic (PLEG): container finished" podID="5056171a-6fc0-4d48-9ff0-e275137253f3" containerID="250338fb343035b6cec2629aa8ef485b90fd7a39a4592f8aae30cde99b224717" exitCode=0 Oct 14 07:20:14 crc kubenswrapper[4870]: I1014 07:20:14.430391 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-5c98-account-create-9bjhh" event={"ID":"5056171a-6fc0-4d48-9ff0-e275137253f3","Type":"ContainerDied","Data":"250338fb343035b6cec2629aa8ef485b90fd7a39a4592f8aae30cde99b224717"} Oct 14 07:20:14 crc kubenswrapper[4870]: I1014 07:20:14.430417 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-5c98-account-create-9bjhh" event={"ID":"5056171a-6fc0-4d48-9ff0-e275137253f3","Type":"ContainerStarted","Data":"d8080344442285e06403465da4075fc5b7f452b9e48700d7b54d57f1bc6cfafe"} Oct 14 07:20:14 crc kubenswrapper[4870]: I1014 07:20:14.438953 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l8g2b\" (UniqueName: \"kubernetes.io/projected/d1368adf-8a2b-41ff-987d-806f65c1b128-kube-api-access-l8g2b\") pod \"d1368adf-8a2b-41ff-987d-806f65c1b128\" (UID: \"d1368adf-8a2b-41ff-987d-806f65c1b128\") " Oct 14 07:20:14 crc kubenswrapper[4870]: I1014 07:20:14.439026 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"d1368adf-8a2b-41ff-987d-806f65c1b128\" (UID: \"d1368adf-8a2b-41ff-987d-806f65c1b128\") " Oct 14 07:20:14 crc kubenswrapper[4870]: I1014 07:20:14.439077 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d1368adf-8a2b-41ff-987d-806f65c1b128-httpd-run\") pod \"d1368adf-8a2b-41ff-987d-806f65c1b128\" (UID: \"d1368adf-8a2b-41ff-987d-806f65c1b128\") " Oct 14 07:20:14 crc kubenswrapper[4870]: I1014 07:20:14.439133 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d1368adf-8a2b-41ff-987d-806f65c1b128-scripts\") pod \"d1368adf-8a2b-41ff-987d-806f65c1b128\" (UID: \"d1368adf-8a2b-41ff-987d-806f65c1b128\") " Oct 14 07:20:14 crc kubenswrapper[4870]: I1014 07:20:14.439157 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d1368adf-8a2b-41ff-987d-806f65c1b128-config-data\") pod \"d1368adf-8a2b-41ff-987d-806f65c1b128\" (UID: \"d1368adf-8a2b-41ff-987d-806f65c1b128\") " Oct 14 07:20:14 crc kubenswrapper[4870]: I1014 07:20:14.439255 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d1368adf-8a2b-41ff-987d-806f65c1b128-combined-ca-bundle\") pod \"d1368adf-8a2b-41ff-987d-806f65c1b128\" (UID: \"d1368adf-8a2b-41ff-987d-806f65c1b128\") " Oct 14 07:20:14 crc kubenswrapper[4870]: I1014 07:20:14.439302 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d1368adf-8a2b-41ff-987d-806f65c1b128-logs\") pod \"d1368adf-8a2b-41ff-987d-806f65c1b128\" (UID: \"d1368adf-8a2b-41ff-987d-806f65c1b128\") " Oct 14 07:20:14 crc kubenswrapper[4870]: I1014 07:20:14.439535 4870 scope.go:117] "RemoveContainer" containerID="c1c7e6b321e0f5505182cedf957020d5a101cfb5be8ac4c352ed3f33595ba3ac" Oct 14 07:20:14 crc kubenswrapper[4870]: I1014 07:20:14.439956 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d1368adf-8a2b-41ff-987d-806f65c1b128-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "d1368adf-8a2b-41ff-987d-806f65c1b128" (UID: "d1368adf-8a2b-41ff-987d-806f65c1b128"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:20:14 crc kubenswrapper[4870]: I1014 07:20:14.439974 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d1368adf-8a2b-41ff-987d-806f65c1b128-logs" (OuterVolumeSpecName: "logs") pod "d1368adf-8a2b-41ff-987d-806f65c1b128" (UID: "d1368adf-8a2b-41ff-987d-806f65c1b128"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:20:14 crc kubenswrapper[4870]: I1014 07:20:14.449232 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"99fa3d0e-8c4f-4893-b2fa-040837f04767","Type":"ContainerDied","Data":"f1a7b415ab6c6183f0cc4d7292a82a53b740e339e880159e90839c2922b8df52"} Oct 14 07:20:14 crc kubenswrapper[4870]: I1014 07:20:14.449357 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 14 07:20:14 crc kubenswrapper[4870]: I1014 07:20:14.452930 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d1368adf-8a2b-41ff-987d-806f65c1b128-scripts" (OuterVolumeSpecName: "scripts") pod "d1368adf-8a2b-41ff-987d-806f65c1b128" (UID: "d1368adf-8a2b-41ff-987d-806f65c1b128"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:20:14 crc kubenswrapper[4870]: I1014 07:20:14.454655 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "glance") pod "d1368adf-8a2b-41ff-987d-806f65c1b128" (UID: "d1368adf-8a2b-41ff-987d-806f65c1b128"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 14 07:20:14 crc kubenswrapper[4870]: I1014 07:20:14.471253 4870 scope.go:117] "RemoveContainer" containerID="5d6e51d53c3a4412c7e245a2bc027b8b0885f7aa76c0eba1a2678cb04ac541aa" Oct 14 07:20:14 crc kubenswrapper[4870]: I1014 07:20:14.478925 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d1368adf-8a2b-41ff-987d-806f65c1b128-kube-api-access-l8g2b" (OuterVolumeSpecName: "kube-api-access-l8g2b") pod "d1368adf-8a2b-41ff-987d-806f65c1b128" (UID: "d1368adf-8a2b-41ff-987d-806f65c1b128"). InnerVolumeSpecName "kube-api-access-l8g2b". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:20:14 crc kubenswrapper[4870]: I1014 07:20:14.522758 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-nmtf9" podStartSLOduration=2.239229053 podStartE2EDuration="10.522739839s" podCreationTimestamp="2025-10-14 07:20:04 +0000 UTC" firstStartedPulling="2025-10-14 07:20:05.654187756 +0000 UTC m=+1141.351548127" lastFinishedPulling="2025-10-14 07:20:13.937698542 +0000 UTC m=+1149.635058913" observedRunningTime="2025-10-14 07:20:14.520045493 +0000 UTC m=+1150.217405864" watchObservedRunningTime="2025-10-14 07:20:14.522739839 +0000 UTC m=+1150.220100210" Oct 14 07:20:14 crc kubenswrapper[4870]: I1014 07:20:14.536260 4870 scope.go:117] "RemoveContainer" containerID="3294d0dc2862fd311c8ecc3a5047b4c9f36b8302327a52846a6bcd07e3372697" Oct 14 07:20:14 crc kubenswrapper[4870]: I1014 07:20:14.542878 4870 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d1368adf-8a2b-41ff-987d-806f65c1b128-logs\") on node \"crc\" DevicePath \"\"" Oct 14 07:20:14 crc kubenswrapper[4870]: I1014 07:20:14.542917 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l8g2b\" (UniqueName: \"kubernetes.io/projected/d1368adf-8a2b-41ff-987d-806f65c1b128-kube-api-access-l8g2b\") on node \"crc\" DevicePath \"\"" Oct 14 07:20:14 crc kubenswrapper[4870]: I1014 07:20:14.542943 4870 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Oct 14 07:20:14 crc kubenswrapper[4870]: I1014 07:20:14.542954 4870 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d1368adf-8a2b-41ff-987d-806f65c1b128-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 14 07:20:14 crc kubenswrapper[4870]: I1014 07:20:14.542963 4870 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d1368adf-8a2b-41ff-987d-806f65c1b128-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 07:20:14 crc kubenswrapper[4870]: I1014 07:20:14.553083 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-cedd-account-create-gknx9"] Oct 14 07:20:14 crc kubenswrapper[4870]: I1014 07:20:14.560357 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d1368adf-8a2b-41ff-987d-806f65c1b128-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d1368adf-8a2b-41ff-987d-806f65c1b128" (UID: "d1368adf-8a2b-41ff-987d-806f65c1b128"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:20:14 crc kubenswrapper[4870]: I1014 07:20:14.608574 4870 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Oct 14 07:20:14 crc kubenswrapper[4870]: I1014 07:20:14.610794 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 14 07:20:14 crc kubenswrapper[4870]: I1014 07:20:14.622877 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 14 07:20:14 crc kubenswrapper[4870]: I1014 07:20:14.629817 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 14 07:20:14 crc kubenswrapper[4870]: E1014 07:20:14.630552 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d5f827b-9c3f-4d17-b6a6-506fe3d2e06f" containerName="keystone-bootstrap" Oct 14 07:20:14 crc kubenswrapper[4870]: I1014 07:20:14.630637 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d5f827b-9c3f-4d17-b6a6-506fe3d2e06f" containerName="keystone-bootstrap" Oct 14 07:20:14 crc kubenswrapper[4870]: E1014 07:20:14.630725 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99fa3d0e-8c4f-4893-b2fa-040837f04767" containerName="glance-httpd" Oct 14 07:20:14 crc kubenswrapper[4870]: I1014 07:20:14.630799 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="99fa3d0e-8c4f-4893-b2fa-040837f04767" containerName="glance-httpd" Oct 14 07:20:14 crc kubenswrapper[4870]: E1014 07:20:14.630854 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1368adf-8a2b-41ff-987d-806f65c1b128" containerName="glance-log" Oct 14 07:20:14 crc kubenswrapper[4870]: I1014 07:20:14.630916 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1368adf-8a2b-41ff-987d-806f65c1b128" containerName="glance-log" Oct 14 07:20:14 crc kubenswrapper[4870]: E1014 07:20:14.631004 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99fa3d0e-8c4f-4893-b2fa-040837f04767" containerName="glance-log" Oct 14 07:20:14 crc kubenswrapper[4870]: I1014 07:20:14.631067 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="99fa3d0e-8c4f-4893-b2fa-040837f04767" containerName="glance-log" Oct 14 07:20:14 crc kubenswrapper[4870]: E1014 07:20:14.631149 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1368adf-8a2b-41ff-987d-806f65c1b128" containerName="glance-httpd" Oct 14 07:20:14 crc kubenswrapper[4870]: I1014 07:20:14.631198 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1368adf-8a2b-41ff-987d-806f65c1b128" containerName="glance-httpd" Oct 14 07:20:14 crc kubenswrapper[4870]: I1014 07:20:14.631492 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="99fa3d0e-8c4f-4893-b2fa-040837f04767" containerName="glance-httpd" Oct 14 07:20:14 crc kubenswrapper[4870]: I1014 07:20:14.631569 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="d1368adf-8a2b-41ff-987d-806f65c1b128" containerName="glance-log" Oct 14 07:20:14 crc kubenswrapper[4870]: I1014 07:20:14.631733 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="6d5f827b-9c3f-4d17-b6a6-506fe3d2e06f" containerName="keystone-bootstrap" Oct 14 07:20:14 crc kubenswrapper[4870]: I1014 07:20:14.631806 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="99fa3d0e-8c4f-4893-b2fa-040837f04767" containerName="glance-log" Oct 14 07:20:14 crc kubenswrapper[4870]: I1014 07:20:14.631880 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="d1368adf-8a2b-41ff-987d-806f65c1b128" containerName="glance-httpd" Oct 14 07:20:14 crc kubenswrapper[4870]: I1014 07:20:14.633049 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 14 07:20:14 crc kubenswrapper[4870]: I1014 07:20:14.641328 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Oct 14 07:20:14 crc kubenswrapper[4870]: I1014 07:20:14.641568 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 14 07:20:14 crc kubenswrapper[4870]: I1014 07:20:14.662594 4870 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d1368adf-8a2b-41ff-987d-806f65c1b128-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 07:20:14 crc kubenswrapper[4870]: I1014 07:20:14.662628 4870 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Oct 14 07:20:14 crc kubenswrapper[4870]: I1014 07:20:14.662782 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d1368adf-8a2b-41ff-987d-806f65c1b128-config-data" (OuterVolumeSpecName: "config-data") pod "d1368adf-8a2b-41ff-987d-806f65c1b128" (UID: "d1368adf-8a2b-41ff-987d-806f65c1b128"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:20:14 crc kubenswrapper[4870]: I1014 07:20:14.676797 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 14 07:20:14 crc kubenswrapper[4870]: I1014 07:20:14.736827 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 14 07:20:14 crc kubenswrapper[4870]: I1014 07:20:14.756217 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 14 07:20:14 crc kubenswrapper[4870]: I1014 07:20:14.764447 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"385ca977-7d50-426e-86e9-6a9416ad872c\") " pod="openstack/glance-default-internal-api-0" Oct 14 07:20:14 crc kubenswrapper[4870]: I1014 07:20:14.764502 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/385ca977-7d50-426e-86e9-6a9416ad872c-scripts\") pod \"glance-default-internal-api-0\" (UID: \"385ca977-7d50-426e-86e9-6a9416ad872c\") " pod="openstack/glance-default-internal-api-0" Oct 14 07:20:14 crc kubenswrapper[4870]: I1014 07:20:14.764561 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/385ca977-7d50-426e-86e9-6a9416ad872c-config-data\") pod \"glance-default-internal-api-0\" (UID: \"385ca977-7d50-426e-86e9-6a9416ad872c\") " pod="openstack/glance-default-internal-api-0" Oct 14 07:20:14 crc kubenswrapper[4870]: I1014 07:20:14.764594 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/385ca977-7d50-426e-86e9-6a9416ad872c-logs\") pod \"glance-default-internal-api-0\" (UID: \"385ca977-7d50-426e-86e9-6a9416ad872c\") " pod="openstack/glance-default-internal-api-0" Oct 14 07:20:14 crc kubenswrapper[4870]: I1014 07:20:14.764625 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/385ca977-7d50-426e-86e9-6a9416ad872c-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"385ca977-7d50-426e-86e9-6a9416ad872c\") " pod="openstack/glance-default-internal-api-0" Oct 14 07:20:14 crc kubenswrapper[4870]: I1014 07:20:14.764655 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w8zs9\" (UniqueName: \"kubernetes.io/projected/385ca977-7d50-426e-86e9-6a9416ad872c-kube-api-access-w8zs9\") pod \"glance-default-internal-api-0\" (UID: \"385ca977-7d50-426e-86e9-6a9416ad872c\") " pod="openstack/glance-default-internal-api-0" Oct 14 07:20:14 crc kubenswrapper[4870]: I1014 07:20:14.764675 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/385ca977-7d50-426e-86e9-6a9416ad872c-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"385ca977-7d50-426e-86e9-6a9416ad872c\") " pod="openstack/glance-default-internal-api-0" Oct 14 07:20:14 crc kubenswrapper[4870]: I1014 07:20:14.764698 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/385ca977-7d50-426e-86e9-6a9416ad872c-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"385ca977-7d50-426e-86e9-6a9416ad872c\") " pod="openstack/glance-default-internal-api-0" Oct 14 07:20:14 crc kubenswrapper[4870]: I1014 07:20:14.764743 4870 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d1368adf-8a2b-41ff-987d-806f65c1b128-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 07:20:14 crc kubenswrapper[4870]: I1014 07:20:14.768085 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 14 07:20:14 crc kubenswrapper[4870]: I1014 07:20:14.770074 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 14 07:20:14 crc kubenswrapper[4870]: I1014 07:20:14.772040 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 14 07:20:14 crc kubenswrapper[4870]: I1014 07:20:14.773955 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Oct 14 07:20:14 crc kubenswrapper[4870]: I1014 07:20:14.781103 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 14 07:20:14 crc kubenswrapper[4870]: I1014 07:20:14.866958 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"452a6ce6-d9dc-4455-9304-414701c8895f\") " pod="openstack/glance-default-external-api-0" Oct 14 07:20:14 crc kubenswrapper[4870]: I1014 07:20:14.867019 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"385ca977-7d50-426e-86e9-6a9416ad872c\") " pod="openstack/glance-default-internal-api-0" Oct 14 07:20:14 crc kubenswrapper[4870]: I1014 07:20:14.867066 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/385ca977-7d50-426e-86e9-6a9416ad872c-scripts\") pod \"glance-default-internal-api-0\" (UID: \"385ca977-7d50-426e-86e9-6a9416ad872c\") " pod="openstack/glance-default-internal-api-0" Oct 14 07:20:14 crc kubenswrapper[4870]: I1014 07:20:14.867100 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/452a6ce6-d9dc-4455-9304-414701c8895f-logs\") pod \"glance-default-external-api-0\" (UID: \"452a6ce6-d9dc-4455-9304-414701c8895f\") " pod="openstack/glance-default-external-api-0" Oct 14 07:20:14 crc kubenswrapper[4870]: I1014 07:20:14.867124 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/452a6ce6-d9dc-4455-9304-414701c8895f-config-data\") pod \"glance-default-external-api-0\" (UID: \"452a6ce6-d9dc-4455-9304-414701c8895f\") " pod="openstack/glance-default-external-api-0" Oct 14 07:20:14 crc kubenswrapper[4870]: I1014 07:20:14.867183 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/385ca977-7d50-426e-86e9-6a9416ad872c-config-data\") pod \"glance-default-internal-api-0\" (UID: \"385ca977-7d50-426e-86e9-6a9416ad872c\") " pod="openstack/glance-default-internal-api-0" Oct 14 07:20:14 crc kubenswrapper[4870]: I1014 07:20:14.867204 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/452a6ce6-d9dc-4455-9304-414701c8895f-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"452a6ce6-d9dc-4455-9304-414701c8895f\") " pod="openstack/glance-default-external-api-0" Oct 14 07:20:14 crc kubenswrapper[4870]: I1014 07:20:14.867235 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-98996\" (UniqueName: \"kubernetes.io/projected/452a6ce6-d9dc-4455-9304-414701c8895f-kube-api-access-98996\") pod \"glance-default-external-api-0\" (UID: \"452a6ce6-d9dc-4455-9304-414701c8895f\") " pod="openstack/glance-default-external-api-0" Oct 14 07:20:14 crc kubenswrapper[4870]: I1014 07:20:14.867266 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/452a6ce6-d9dc-4455-9304-414701c8895f-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"452a6ce6-d9dc-4455-9304-414701c8895f\") " pod="openstack/glance-default-external-api-0" Oct 14 07:20:14 crc kubenswrapper[4870]: I1014 07:20:14.867294 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/385ca977-7d50-426e-86e9-6a9416ad872c-logs\") pod \"glance-default-internal-api-0\" (UID: \"385ca977-7d50-426e-86e9-6a9416ad872c\") " pod="openstack/glance-default-internal-api-0" Oct 14 07:20:14 crc kubenswrapper[4870]: I1014 07:20:14.867310 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/385ca977-7d50-426e-86e9-6a9416ad872c-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"385ca977-7d50-426e-86e9-6a9416ad872c\") " pod="openstack/glance-default-internal-api-0" Oct 14 07:20:14 crc kubenswrapper[4870]: I1014 07:20:14.867337 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/452a6ce6-d9dc-4455-9304-414701c8895f-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"452a6ce6-d9dc-4455-9304-414701c8895f\") " pod="openstack/glance-default-external-api-0" Oct 14 07:20:14 crc kubenswrapper[4870]: I1014 07:20:14.867374 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w8zs9\" (UniqueName: \"kubernetes.io/projected/385ca977-7d50-426e-86e9-6a9416ad872c-kube-api-access-w8zs9\") pod \"glance-default-internal-api-0\" (UID: \"385ca977-7d50-426e-86e9-6a9416ad872c\") " pod="openstack/glance-default-internal-api-0" Oct 14 07:20:14 crc kubenswrapper[4870]: I1014 07:20:14.867399 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/452a6ce6-d9dc-4455-9304-414701c8895f-scripts\") pod \"glance-default-external-api-0\" (UID: \"452a6ce6-d9dc-4455-9304-414701c8895f\") " pod="openstack/glance-default-external-api-0" Oct 14 07:20:14 crc kubenswrapper[4870]: I1014 07:20:14.867422 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/385ca977-7d50-426e-86e9-6a9416ad872c-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"385ca977-7d50-426e-86e9-6a9416ad872c\") " pod="openstack/glance-default-internal-api-0" Oct 14 07:20:14 crc kubenswrapper[4870]: I1014 07:20:14.867471 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/385ca977-7d50-426e-86e9-6a9416ad872c-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"385ca977-7d50-426e-86e9-6a9416ad872c\") " pod="openstack/glance-default-internal-api-0" Oct 14 07:20:14 crc kubenswrapper[4870]: I1014 07:20:14.868173 4870 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"385ca977-7d50-426e-86e9-6a9416ad872c\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/glance-default-internal-api-0" Oct 14 07:20:14 crc kubenswrapper[4870]: I1014 07:20:14.868190 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/385ca977-7d50-426e-86e9-6a9416ad872c-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"385ca977-7d50-426e-86e9-6a9416ad872c\") " pod="openstack/glance-default-internal-api-0" Oct 14 07:20:14 crc kubenswrapper[4870]: I1014 07:20:14.868473 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/385ca977-7d50-426e-86e9-6a9416ad872c-logs\") pod \"glance-default-internal-api-0\" (UID: \"385ca977-7d50-426e-86e9-6a9416ad872c\") " pod="openstack/glance-default-internal-api-0" Oct 14 07:20:14 crc kubenswrapper[4870]: I1014 07:20:14.872546 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/385ca977-7d50-426e-86e9-6a9416ad872c-config-data\") pod \"glance-default-internal-api-0\" (UID: \"385ca977-7d50-426e-86e9-6a9416ad872c\") " pod="openstack/glance-default-internal-api-0" Oct 14 07:20:14 crc kubenswrapper[4870]: I1014 07:20:14.872859 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/385ca977-7d50-426e-86e9-6a9416ad872c-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"385ca977-7d50-426e-86e9-6a9416ad872c\") " pod="openstack/glance-default-internal-api-0" Oct 14 07:20:14 crc kubenswrapper[4870]: I1014 07:20:14.873040 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/385ca977-7d50-426e-86e9-6a9416ad872c-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"385ca977-7d50-426e-86e9-6a9416ad872c\") " pod="openstack/glance-default-internal-api-0" Oct 14 07:20:14 crc kubenswrapper[4870]: I1014 07:20:14.889387 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w8zs9\" (UniqueName: \"kubernetes.io/projected/385ca977-7d50-426e-86e9-6a9416ad872c-kube-api-access-w8zs9\") pod \"glance-default-internal-api-0\" (UID: \"385ca977-7d50-426e-86e9-6a9416ad872c\") " pod="openstack/glance-default-internal-api-0" Oct 14 07:20:14 crc kubenswrapper[4870]: I1014 07:20:14.901104 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/385ca977-7d50-426e-86e9-6a9416ad872c-scripts\") pod \"glance-default-internal-api-0\" (UID: \"385ca977-7d50-426e-86e9-6a9416ad872c\") " pod="openstack/glance-default-internal-api-0" Oct 14 07:20:14 crc kubenswrapper[4870]: I1014 07:20:14.924396 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"385ca977-7d50-426e-86e9-6a9416ad872c\") " pod="openstack/glance-default-internal-api-0" Oct 14 07:20:14 crc kubenswrapper[4870]: I1014 07:20:14.970367 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/452a6ce6-d9dc-4455-9304-414701c8895f-config-data\") pod \"glance-default-external-api-0\" (UID: \"452a6ce6-d9dc-4455-9304-414701c8895f\") " pod="openstack/glance-default-external-api-0" Oct 14 07:20:14 crc kubenswrapper[4870]: I1014 07:20:14.970576 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/452a6ce6-d9dc-4455-9304-414701c8895f-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"452a6ce6-d9dc-4455-9304-414701c8895f\") " pod="openstack/glance-default-external-api-0" Oct 14 07:20:14 crc kubenswrapper[4870]: I1014 07:20:14.970607 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-98996\" (UniqueName: \"kubernetes.io/projected/452a6ce6-d9dc-4455-9304-414701c8895f-kube-api-access-98996\") pod \"glance-default-external-api-0\" (UID: \"452a6ce6-d9dc-4455-9304-414701c8895f\") " pod="openstack/glance-default-external-api-0" Oct 14 07:20:14 crc kubenswrapper[4870]: I1014 07:20:14.970633 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/452a6ce6-d9dc-4455-9304-414701c8895f-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"452a6ce6-d9dc-4455-9304-414701c8895f\") " pod="openstack/glance-default-external-api-0" Oct 14 07:20:14 crc kubenswrapper[4870]: I1014 07:20:14.970663 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/452a6ce6-d9dc-4455-9304-414701c8895f-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"452a6ce6-d9dc-4455-9304-414701c8895f\") " pod="openstack/glance-default-external-api-0" Oct 14 07:20:14 crc kubenswrapper[4870]: I1014 07:20:14.970699 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/452a6ce6-d9dc-4455-9304-414701c8895f-scripts\") pod \"glance-default-external-api-0\" (UID: \"452a6ce6-d9dc-4455-9304-414701c8895f\") " pod="openstack/glance-default-external-api-0" Oct 14 07:20:14 crc kubenswrapper[4870]: I1014 07:20:14.970790 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"452a6ce6-d9dc-4455-9304-414701c8895f\") " pod="openstack/glance-default-external-api-0" Oct 14 07:20:14 crc kubenswrapper[4870]: I1014 07:20:14.970845 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/452a6ce6-d9dc-4455-9304-414701c8895f-logs\") pod \"glance-default-external-api-0\" (UID: \"452a6ce6-d9dc-4455-9304-414701c8895f\") " pod="openstack/glance-default-external-api-0" Oct 14 07:20:14 crc kubenswrapper[4870]: I1014 07:20:14.971329 4870 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"452a6ce6-d9dc-4455-9304-414701c8895f\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/glance-default-external-api-0" Oct 14 07:20:14 crc kubenswrapper[4870]: I1014 07:20:14.973740 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/452a6ce6-d9dc-4455-9304-414701c8895f-logs\") pod \"glance-default-external-api-0\" (UID: \"452a6ce6-d9dc-4455-9304-414701c8895f\") " pod="openstack/glance-default-external-api-0" Oct 14 07:20:14 crc kubenswrapper[4870]: I1014 07:20:14.980154 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/452a6ce6-d9dc-4455-9304-414701c8895f-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"452a6ce6-d9dc-4455-9304-414701c8895f\") " pod="openstack/glance-default-external-api-0" Oct 14 07:20:14 crc kubenswrapper[4870]: I1014 07:20:14.982034 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/452a6ce6-d9dc-4455-9304-414701c8895f-scripts\") pod \"glance-default-external-api-0\" (UID: \"452a6ce6-d9dc-4455-9304-414701c8895f\") " pod="openstack/glance-default-external-api-0" Oct 14 07:20:14 crc kubenswrapper[4870]: I1014 07:20:14.983222 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/452a6ce6-d9dc-4455-9304-414701c8895f-config-data\") pod \"glance-default-external-api-0\" (UID: \"452a6ce6-d9dc-4455-9304-414701c8895f\") " pod="openstack/glance-default-external-api-0" Oct 14 07:20:14 crc kubenswrapper[4870]: I1014 07:20:14.985473 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 14 07:20:14 crc kubenswrapper[4870]: I1014 07:20:14.986542 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/452a6ce6-d9dc-4455-9304-414701c8895f-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"452a6ce6-d9dc-4455-9304-414701c8895f\") " pod="openstack/glance-default-external-api-0" Oct 14 07:20:14 crc kubenswrapper[4870]: I1014 07:20:14.988937 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/452a6ce6-d9dc-4455-9304-414701c8895f-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"452a6ce6-d9dc-4455-9304-414701c8895f\") " pod="openstack/glance-default-external-api-0" Oct 14 07:20:14 crc kubenswrapper[4870]: I1014 07:20:14.990266 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-98996\" (UniqueName: \"kubernetes.io/projected/452a6ce6-d9dc-4455-9304-414701c8895f-kube-api-access-98996\") pod \"glance-default-external-api-0\" (UID: \"452a6ce6-d9dc-4455-9304-414701c8895f\") " pod="openstack/glance-default-external-api-0" Oct 14 07:20:15 crc kubenswrapper[4870]: I1014 07:20:15.041785 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"452a6ce6-d9dc-4455-9304-414701c8895f\") " pod="openstack/glance-default-external-api-0" Oct 14 07:20:15 crc kubenswrapper[4870]: I1014 07:20:15.068735 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="99fa3d0e-8c4f-4893-b2fa-040837f04767" path="/var/lib/kubelet/pods/99fa3d0e-8c4f-4893-b2fa-040837f04767/volumes" Oct 14 07:20:15 crc kubenswrapper[4870]: I1014 07:20:15.069753 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d1368adf-8a2b-41ff-987d-806f65c1b128" path="/var/lib/kubelet/pods/d1368adf-8a2b-41ff-987d-806f65c1b128/volumes" Oct 14 07:20:15 crc kubenswrapper[4870]: I1014 07:20:15.070328 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-s2xm6"] Oct 14 07:20:15 crc kubenswrapper[4870]: I1014 07:20:15.073796 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-s2xm6"] Oct 14 07:20:15 crc kubenswrapper[4870]: I1014 07:20:15.133275 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7bcdb8c767-9pwfh" Oct 14 07:20:15 crc kubenswrapper[4870]: I1014 07:20:15.150082 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 14 07:20:15 crc kubenswrapper[4870]: I1014 07:20:15.201773 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-cbhx4"] Oct 14 07:20:15 crc kubenswrapper[4870]: I1014 07:20:15.203594 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-cbhx4" Oct 14 07:20:15 crc kubenswrapper[4870]: I1014 07:20:15.209128 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 14 07:20:15 crc kubenswrapper[4870]: I1014 07:20:15.209356 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-cqggg" Oct 14 07:20:15 crc kubenswrapper[4870]: I1014 07:20:15.209968 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 14 07:20:15 crc kubenswrapper[4870]: I1014 07:20:15.211365 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 14 07:20:15 crc kubenswrapper[4870]: I1014 07:20:15.260721 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-cbhx4"] Oct 14 07:20:15 crc kubenswrapper[4870]: I1014 07:20:15.274545 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kxlb4\" (UniqueName: \"kubernetes.io/projected/fe66a2e1-364a-4cb9-b857-c208c7d1b108-kube-api-access-kxlb4\") pod \"keystone-bootstrap-cbhx4\" (UID: \"fe66a2e1-364a-4cb9-b857-c208c7d1b108\") " pod="openstack/keystone-bootstrap-cbhx4" Oct 14 07:20:15 crc kubenswrapper[4870]: I1014 07:20:15.274785 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/fe66a2e1-364a-4cb9-b857-c208c7d1b108-credential-keys\") pod \"keystone-bootstrap-cbhx4\" (UID: \"fe66a2e1-364a-4cb9-b857-c208c7d1b108\") " pod="openstack/keystone-bootstrap-cbhx4" Oct 14 07:20:15 crc kubenswrapper[4870]: I1014 07:20:15.274822 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fe66a2e1-364a-4cb9-b857-c208c7d1b108-scripts\") pod \"keystone-bootstrap-cbhx4\" (UID: \"fe66a2e1-364a-4cb9-b857-c208c7d1b108\") " pod="openstack/keystone-bootstrap-cbhx4" Oct 14 07:20:15 crc kubenswrapper[4870]: I1014 07:20:15.274859 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fe66a2e1-364a-4cb9-b857-c208c7d1b108-config-data\") pod \"keystone-bootstrap-cbhx4\" (UID: \"fe66a2e1-364a-4cb9-b857-c208c7d1b108\") " pod="openstack/keystone-bootstrap-cbhx4" Oct 14 07:20:15 crc kubenswrapper[4870]: I1014 07:20:15.274875 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/fe66a2e1-364a-4cb9-b857-c208c7d1b108-fernet-keys\") pod \"keystone-bootstrap-cbhx4\" (UID: \"fe66a2e1-364a-4cb9-b857-c208c7d1b108\") " pod="openstack/keystone-bootstrap-cbhx4" Oct 14 07:20:15 crc kubenswrapper[4870]: I1014 07:20:15.274897 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe66a2e1-364a-4cb9-b857-c208c7d1b108-combined-ca-bundle\") pod \"keystone-bootstrap-cbhx4\" (UID: \"fe66a2e1-364a-4cb9-b857-c208c7d1b108\") " pod="openstack/keystone-bootstrap-cbhx4" Oct 14 07:20:15 crc kubenswrapper[4870]: I1014 07:20:15.276457 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7cc794b75c-d2b9c"] Oct 14 07:20:15 crc kubenswrapper[4870]: I1014 07:20:15.276803 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7cc794b75c-d2b9c" podUID="69a02fec-1745-4827-9c90-d9dffb829a66" containerName="dnsmasq-dns" containerID="cri-o://53081407fd75b324098b0ed4b06bd603bd305531ac1290feb40806718fcc0af4" gracePeriod=10 Oct 14 07:20:15 crc kubenswrapper[4870]: I1014 07:20:15.376451 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kxlb4\" (UniqueName: \"kubernetes.io/projected/fe66a2e1-364a-4cb9-b857-c208c7d1b108-kube-api-access-kxlb4\") pod \"keystone-bootstrap-cbhx4\" (UID: \"fe66a2e1-364a-4cb9-b857-c208c7d1b108\") " pod="openstack/keystone-bootstrap-cbhx4" Oct 14 07:20:15 crc kubenswrapper[4870]: I1014 07:20:15.376503 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/fe66a2e1-364a-4cb9-b857-c208c7d1b108-credential-keys\") pod \"keystone-bootstrap-cbhx4\" (UID: \"fe66a2e1-364a-4cb9-b857-c208c7d1b108\") " pod="openstack/keystone-bootstrap-cbhx4" Oct 14 07:20:15 crc kubenswrapper[4870]: I1014 07:20:15.376538 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fe66a2e1-364a-4cb9-b857-c208c7d1b108-scripts\") pod \"keystone-bootstrap-cbhx4\" (UID: \"fe66a2e1-364a-4cb9-b857-c208c7d1b108\") " pod="openstack/keystone-bootstrap-cbhx4" Oct 14 07:20:15 crc kubenswrapper[4870]: I1014 07:20:15.376580 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fe66a2e1-364a-4cb9-b857-c208c7d1b108-config-data\") pod \"keystone-bootstrap-cbhx4\" (UID: \"fe66a2e1-364a-4cb9-b857-c208c7d1b108\") " pod="openstack/keystone-bootstrap-cbhx4" Oct 14 07:20:15 crc kubenswrapper[4870]: I1014 07:20:15.376596 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/fe66a2e1-364a-4cb9-b857-c208c7d1b108-fernet-keys\") pod \"keystone-bootstrap-cbhx4\" (UID: \"fe66a2e1-364a-4cb9-b857-c208c7d1b108\") " pod="openstack/keystone-bootstrap-cbhx4" Oct 14 07:20:15 crc kubenswrapper[4870]: I1014 07:20:15.378566 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe66a2e1-364a-4cb9-b857-c208c7d1b108-combined-ca-bundle\") pod \"keystone-bootstrap-cbhx4\" (UID: \"fe66a2e1-364a-4cb9-b857-c208c7d1b108\") " pod="openstack/keystone-bootstrap-cbhx4" Oct 14 07:20:15 crc kubenswrapper[4870]: I1014 07:20:15.383584 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fe66a2e1-364a-4cb9-b857-c208c7d1b108-scripts\") pod \"keystone-bootstrap-cbhx4\" (UID: \"fe66a2e1-364a-4cb9-b857-c208c7d1b108\") " pod="openstack/keystone-bootstrap-cbhx4" Oct 14 07:20:15 crc kubenswrapper[4870]: I1014 07:20:15.384454 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe66a2e1-364a-4cb9-b857-c208c7d1b108-combined-ca-bundle\") pod \"keystone-bootstrap-cbhx4\" (UID: \"fe66a2e1-364a-4cb9-b857-c208c7d1b108\") " pod="openstack/keystone-bootstrap-cbhx4" Oct 14 07:20:15 crc kubenswrapper[4870]: I1014 07:20:15.388068 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/fe66a2e1-364a-4cb9-b857-c208c7d1b108-fernet-keys\") pod \"keystone-bootstrap-cbhx4\" (UID: \"fe66a2e1-364a-4cb9-b857-c208c7d1b108\") " pod="openstack/keystone-bootstrap-cbhx4" Oct 14 07:20:15 crc kubenswrapper[4870]: I1014 07:20:15.388961 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/fe66a2e1-364a-4cb9-b857-c208c7d1b108-credential-keys\") pod \"keystone-bootstrap-cbhx4\" (UID: \"fe66a2e1-364a-4cb9-b857-c208c7d1b108\") " pod="openstack/keystone-bootstrap-cbhx4" Oct 14 07:20:15 crc kubenswrapper[4870]: I1014 07:20:15.403691 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fe66a2e1-364a-4cb9-b857-c208c7d1b108-config-data\") pod \"keystone-bootstrap-cbhx4\" (UID: \"fe66a2e1-364a-4cb9-b857-c208c7d1b108\") " pod="openstack/keystone-bootstrap-cbhx4" Oct 14 07:20:15 crc kubenswrapper[4870]: I1014 07:20:15.408484 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kxlb4\" (UniqueName: \"kubernetes.io/projected/fe66a2e1-364a-4cb9-b857-c208c7d1b108-kube-api-access-kxlb4\") pod \"keystone-bootstrap-cbhx4\" (UID: \"fe66a2e1-364a-4cb9-b857-c208c7d1b108\") " pod="openstack/keystone-bootstrap-cbhx4" Oct 14 07:20:15 crc kubenswrapper[4870]: I1014 07:20:15.460334 4870 generic.go:334] "Generic (PLEG): container finished" podID="f5e5bd22-1692-477a-a375-3eecad66c746" containerID="474c728d027f33d4775c88671e52d61b3ef2e1359014dd049a62df531455d2e4" exitCode=0 Oct 14 07:20:15 crc kubenswrapper[4870]: I1014 07:20:15.460482 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-cedd-account-create-gknx9" event={"ID":"f5e5bd22-1692-477a-a375-3eecad66c746","Type":"ContainerDied","Data":"474c728d027f33d4775c88671e52d61b3ef2e1359014dd049a62df531455d2e4"} Oct 14 07:20:15 crc kubenswrapper[4870]: I1014 07:20:15.460538 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-cedd-account-create-gknx9" event={"ID":"f5e5bd22-1692-477a-a375-3eecad66c746","Type":"ContainerStarted","Data":"7ccd8ae947ffe6a30545392fb1fe21a2d1b9c90abf105e1d605dbaa67d72480e"} Oct 14 07:20:15 crc kubenswrapper[4870]: I1014 07:20:15.462032 4870 generic.go:334] "Generic (PLEG): container finished" podID="69a02fec-1745-4827-9c90-d9dffb829a66" containerID="53081407fd75b324098b0ed4b06bd603bd305531ac1290feb40806718fcc0af4" exitCode=0 Oct 14 07:20:15 crc kubenswrapper[4870]: I1014 07:20:15.462081 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cc794b75c-d2b9c" event={"ID":"69a02fec-1745-4827-9c90-d9dffb829a66","Type":"ContainerDied","Data":"53081407fd75b324098b0ed4b06bd603bd305531ac1290feb40806718fcc0af4"} Oct 14 07:20:15 crc kubenswrapper[4870]: I1014 07:20:15.552722 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-cbhx4" Oct 14 07:20:15 crc kubenswrapper[4870]: I1014 07:20:15.664882 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 14 07:20:15 crc kubenswrapper[4870]: I1014 07:20:15.938822 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 14 07:20:16 crc kubenswrapper[4870]: I1014 07:20:16.226779 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-cbhx4"] Oct 14 07:20:16 crc kubenswrapper[4870]: W1014 07:20:16.405935 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfe66a2e1_364a_4cb9_b857_c208c7d1b108.slice/crio-af949307c1c161f67b24e0326ac3a341612238074f32d7b2d73053eed3c5b401 WatchSource:0}: Error finding container af949307c1c161f67b24e0326ac3a341612238074f32d7b2d73053eed3c5b401: Status 404 returned error can't find the container with id af949307c1c161f67b24e0326ac3a341612238074f32d7b2d73053eed3c5b401 Oct 14 07:20:16 crc kubenswrapper[4870]: I1014 07:20:16.480291 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cc794b75c-d2b9c" event={"ID":"69a02fec-1745-4827-9c90-d9dffb829a66","Type":"ContainerDied","Data":"8c2d981927ad6fd023274e61249b87cbf6a8f5b6a813de27473ea342707764b2"} Oct 14 07:20:16 crc kubenswrapper[4870]: I1014 07:20:16.480689 4870 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8c2d981927ad6fd023274e61249b87cbf6a8f5b6a813de27473ea342707764b2" Oct 14 07:20:16 crc kubenswrapper[4870]: I1014 07:20:16.481965 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-cbhx4" event={"ID":"fe66a2e1-364a-4cb9-b857-c208c7d1b108","Type":"ContainerStarted","Data":"af949307c1c161f67b24e0326ac3a341612238074f32d7b2d73053eed3c5b401"} Oct 14 07:20:16 crc kubenswrapper[4870]: I1014 07:20:16.483826 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-6547-account-create-vqn99" event={"ID":"625bb722-0865-47f2-b680-ab642ee6c9e5","Type":"ContainerDied","Data":"88ef43717fb61c09aad5b7f62d7199e1c16aa7a44c802f9ceda0b4bc9630d483"} Oct 14 07:20:16 crc kubenswrapper[4870]: I1014 07:20:16.483884 4870 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="88ef43717fb61c09aad5b7f62d7199e1c16aa7a44c802f9ceda0b4bc9630d483" Oct 14 07:20:16 crc kubenswrapper[4870]: I1014 07:20:16.487003 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-5c98-account-create-9bjhh" event={"ID":"5056171a-6fc0-4d48-9ff0-e275137253f3","Type":"ContainerDied","Data":"d8080344442285e06403465da4075fc5b7f452b9e48700d7b54d57f1bc6cfafe"} Oct 14 07:20:16 crc kubenswrapper[4870]: I1014 07:20:16.487027 4870 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d8080344442285e06403465da4075fc5b7f452b9e48700d7b54d57f1bc6cfafe" Oct 14 07:20:16 crc kubenswrapper[4870]: I1014 07:20:16.488148 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"385ca977-7d50-426e-86e9-6a9416ad872c","Type":"ContainerStarted","Data":"4bdba3786e90a36656b26936418c10c2f99a9784397e74e7361942e8723a91ce"} Oct 14 07:20:16 crc kubenswrapper[4870]: I1014 07:20:16.488216 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"385ca977-7d50-426e-86e9-6a9416ad872c","Type":"ContainerStarted","Data":"a61556670dfd1432f61da5383f63bfcbcca889d5c7b1e8b257bf7a66ee23267e"} Oct 14 07:20:16 crc kubenswrapper[4870]: I1014 07:20:16.494544 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"452a6ce6-d9dc-4455-9304-414701c8895f","Type":"ContainerStarted","Data":"32f0b83cd662611a3d7e7503c4a4b8bd18f13afd5fd1f80d21f562d4375742fd"} Oct 14 07:20:16 crc kubenswrapper[4870]: I1014 07:20:16.529126 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-5c98-account-create-9bjhh" Oct 14 07:20:16 crc kubenswrapper[4870]: I1014 07:20:16.533891 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7cc794b75c-d2b9c" Oct 14 07:20:16 crc kubenswrapper[4870]: I1014 07:20:16.574399 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-6547-account-create-vqn99" Oct 14 07:20:16 crc kubenswrapper[4870]: I1014 07:20:16.608954 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/69a02fec-1745-4827-9c90-d9dffb829a66-ovsdbserver-sb\") pod \"69a02fec-1745-4827-9c90-d9dffb829a66\" (UID: \"69a02fec-1745-4827-9c90-d9dffb829a66\") " Oct 14 07:20:16 crc kubenswrapper[4870]: I1014 07:20:16.609039 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/69a02fec-1745-4827-9c90-d9dffb829a66-ovsdbserver-nb\") pod \"69a02fec-1745-4827-9c90-d9dffb829a66\" (UID: \"69a02fec-1745-4827-9c90-d9dffb829a66\") " Oct 14 07:20:16 crc kubenswrapper[4870]: I1014 07:20:16.609111 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-92mbl\" (UniqueName: \"kubernetes.io/projected/625bb722-0865-47f2-b680-ab642ee6c9e5-kube-api-access-92mbl\") pod \"625bb722-0865-47f2-b680-ab642ee6c9e5\" (UID: \"625bb722-0865-47f2-b680-ab642ee6c9e5\") " Oct 14 07:20:16 crc kubenswrapper[4870]: I1014 07:20:16.609165 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2kgdh\" (UniqueName: \"kubernetes.io/projected/5056171a-6fc0-4d48-9ff0-e275137253f3-kube-api-access-2kgdh\") pod \"5056171a-6fc0-4d48-9ff0-e275137253f3\" (UID: \"5056171a-6fc0-4d48-9ff0-e275137253f3\") " Oct 14 07:20:16 crc kubenswrapper[4870]: I1014 07:20:16.609212 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-njwsb\" (UniqueName: \"kubernetes.io/projected/69a02fec-1745-4827-9c90-d9dffb829a66-kube-api-access-njwsb\") pod \"69a02fec-1745-4827-9c90-d9dffb829a66\" (UID: \"69a02fec-1745-4827-9c90-d9dffb829a66\") " Oct 14 07:20:16 crc kubenswrapper[4870]: I1014 07:20:16.609236 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/69a02fec-1745-4827-9c90-d9dffb829a66-dns-swift-storage-0\") pod \"69a02fec-1745-4827-9c90-d9dffb829a66\" (UID: \"69a02fec-1745-4827-9c90-d9dffb829a66\") " Oct 14 07:20:16 crc kubenswrapper[4870]: I1014 07:20:16.609301 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/69a02fec-1745-4827-9c90-d9dffb829a66-dns-svc\") pod \"69a02fec-1745-4827-9c90-d9dffb829a66\" (UID: \"69a02fec-1745-4827-9c90-d9dffb829a66\") " Oct 14 07:20:16 crc kubenswrapper[4870]: I1014 07:20:16.609337 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/69a02fec-1745-4827-9c90-d9dffb829a66-config\") pod \"69a02fec-1745-4827-9c90-d9dffb829a66\" (UID: \"69a02fec-1745-4827-9c90-d9dffb829a66\") " Oct 14 07:20:16 crc kubenswrapper[4870]: I1014 07:20:16.618834 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/625bb722-0865-47f2-b680-ab642ee6c9e5-kube-api-access-92mbl" (OuterVolumeSpecName: "kube-api-access-92mbl") pod "625bb722-0865-47f2-b680-ab642ee6c9e5" (UID: "625bb722-0865-47f2-b680-ab642ee6c9e5"). InnerVolumeSpecName "kube-api-access-92mbl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:20:16 crc kubenswrapper[4870]: I1014 07:20:16.628782 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5056171a-6fc0-4d48-9ff0-e275137253f3-kube-api-access-2kgdh" (OuterVolumeSpecName: "kube-api-access-2kgdh") pod "5056171a-6fc0-4d48-9ff0-e275137253f3" (UID: "5056171a-6fc0-4d48-9ff0-e275137253f3"). InnerVolumeSpecName "kube-api-access-2kgdh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:20:16 crc kubenswrapper[4870]: I1014 07:20:16.643769 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/69a02fec-1745-4827-9c90-d9dffb829a66-kube-api-access-njwsb" (OuterVolumeSpecName: "kube-api-access-njwsb") pod "69a02fec-1745-4827-9c90-d9dffb829a66" (UID: "69a02fec-1745-4827-9c90-d9dffb829a66"). InnerVolumeSpecName "kube-api-access-njwsb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:20:16 crc kubenswrapper[4870]: I1014 07:20:16.711285 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-92mbl\" (UniqueName: \"kubernetes.io/projected/625bb722-0865-47f2-b680-ab642ee6c9e5-kube-api-access-92mbl\") on node \"crc\" DevicePath \"\"" Oct 14 07:20:16 crc kubenswrapper[4870]: I1014 07:20:16.711312 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2kgdh\" (UniqueName: \"kubernetes.io/projected/5056171a-6fc0-4d48-9ff0-e275137253f3-kube-api-access-2kgdh\") on node \"crc\" DevicePath \"\"" Oct 14 07:20:16 crc kubenswrapper[4870]: I1014 07:20:16.711322 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-njwsb\" (UniqueName: \"kubernetes.io/projected/69a02fec-1745-4827-9c90-d9dffb829a66-kube-api-access-njwsb\") on node \"crc\" DevicePath \"\"" Oct 14 07:20:16 crc kubenswrapper[4870]: I1014 07:20:16.738642 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/69a02fec-1745-4827-9c90-d9dffb829a66-config" (OuterVolumeSpecName: "config") pod "69a02fec-1745-4827-9c90-d9dffb829a66" (UID: "69a02fec-1745-4827-9c90-d9dffb829a66"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:20:16 crc kubenswrapper[4870]: I1014 07:20:16.843944 4870 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/69a02fec-1745-4827-9c90-d9dffb829a66-config\") on node \"crc\" DevicePath \"\"" Oct 14 07:20:16 crc kubenswrapper[4870]: I1014 07:20:16.844906 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/69a02fec-1745-4827-9c90-d9dffb829a66-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "69a02fec-1745-4827-9c90-d9dffb829a66" (UID: "69a02fec-1745-4827-9c90-d9dffb829a66"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:20:16 crc kubenswrapper[4870]: I1014 07:20:16.845087 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/69a02fec-1745-4827-9c90-d9dffb829a66-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "69a02fec-1745-4827-9c90-d9dffb829a66" (UID: "69a02fec-1745-4827-9c90-d9dffb829a66"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:20:16 crc kubenswrapper[4870]: I1014 07:20:16.848328 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/69a02fec-1745-4827-9c90-d9dffb829a66-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "69a02fec-1745-4827-9c90-d9dffb829a66" (UID: "69a02fec-1745-4827-9c90-d9dffb829a66"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:20:16 crc kubenswrapper[4870]: I1014 07:20:16.887456 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/69a02fec-1745-4827-9c90-d9dffb829a66-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "69a02fec-1745-4827-9c90-d9dffb829a66" (UID: "69a02fec-1745-4827-9c90-d9dffb829a66"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:20:16 crc kubenswrapper[4870]: I1014 07:20:16.945983 4870 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/69a02fec-1745-4827-9c90-d9dffb829a66-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 14 07:20:16 crc kubenswrapper[4870]: I1014 07:20:16.946016 4870 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/69a02fec-1745-4827-9c90-d9dffb829a66-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 14 07:20:16 crc kubenswrapper[4870]: I1014 07:20:16.946030 4870 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/69a02fec-1745-4827-9c90-d9dffb829a66-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 14 07:20:16 crc kubenswrapper[4870]: I1014 07:20:16.946040 4870 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/69a02fec-1745-4827-9c90-d9dffb829a66-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 14 07:20:16 crc kubenswrapper[4870]: I1014 07:20:16.966665 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-cedd-account-create-gknx9" Oct 14 07:20:17 crc kubenswrapper[4870]: I1014 07:20:17.046556 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6d5f827b-9c3f-4d17-b6a6-506fe3d2e06f" path="/var/lib/kubelet/pods/6d5f827b-9c3f-4d17-b6a6-506fe3d2e06f/volumes" Oct 14 07:20:17 crc kubenswrapper[4870]: I1014 07:20:17.148582 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jvr77\" (UniqueName: \"kubernetes.io/projected/f5e5bd22-1692-477a-a375-3eecad66c746-kube-api-access-jvr77\") pod \"f5e5bd22-1692-477a-a375-3eecad66c746\" (UID: \"f5e5bd22-1692-477a-a375-3eecad66c746\") " Oct 14 07:20:17 crc kubenswrapper[4870]: I1014 07:20:17.162623 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f5e5bd22-1692-477a-a375-3eecad66c746-kube-api-access-jvr77" (OuterVolumeSpecName: "kube-api-access-jvr77") pod "f5e5bd22-1692-477a-a375-3eecad66c746" (UID: "f5e5bd22-1692-477a-a375-3eecad66c746"). InnerVolumeSpecName "kube-api-access-jvr77". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:20:17 crc kubenswrapper[4870]: I1014 07:20:17.250540 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jvr77\" (UniqueName: \"kubernetes.io/projected/f5e5bd22-1692-477a-a375-3eecad66c746-kube-api-access-jvr77\") on node \"crc\" DevicePath \"\"" Oct 14 07:20:17 crc kubenswrapper[4870]: I1014 07:20:17.508958 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"385ca977-7d50-426e-86e9-6a9416ad872c","Type":"ContainerStarted","Data":"f8730cc9a043783b0ae420ff4c40073296e6d3942a94ac10a57a106576eecfda"} Oct 14 07:20:17 crc kubenswrapper[4870]: I1014 07:20:17.521939 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"452a6ce6-d9dc-4455-9304-414701c8895f","Type":"ContainerStarted","Data":"1fde5cdab42d5388f18054e6d6d66a520f1f153b7c41418b4d919ba2fd643643"} Oct 14 07:20:17 crc kubenswrapper[4870]: I1014 07:20:17.534154 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0a31f5df-eae7-4087-a8c8-09c18f5ee9f0","Type":"ContainerStarted","Data":"b3061a7cada98f02a1bef9477945ae37abbb17c577fe0bba622c83a78792b37d"} Oct 14 07:20:17 crc kubenswrapper[4870]: I1014 07:20:17.535845 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-cbhx4" event={"ID":"fe66a2e1-364a-4cb9-b857-c208c7d1b108","Type":"ContainerStarted","Data":"a82c41053072124e1b79993189c4bbf5862bb71cd34f3bf9755173e9a3637eb1"} Oct 14 07:20:17 crc kubenswrapper[4870]: I1014 07:20:17.539158 4870 generic.go:334] "Generic (PLEG): container finished" podID="9dd93b33-f231-4339-9db0-0507cde1dcf5" containerID="8f57f5ca62644a1c87bd38a50bcf2f2378a79660bfe53b23e9830b3861c114f8" exitCode=0 Oct 14 07:20:17 crc kubenswrapper[4870]: I1014 07:20:17.539227 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-nmtf9" event={"ID":"9dd93b33-f231-4339-9db0-0507cde1dcf5","Type":"ContainerDied","Data":"8f57f5ca62644a1c87bd38a50bcf2f2378a79660bfe53b23e9830b3861c114f8"} Oct 14 07:20:17 crc kubenswrapper[4870]: I1014 07:20:17.560824 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-6547-account-create-vqn99" Oct 14 07:20:17 crc kubenswrapper[4870]: I1014 07:20:17.562045 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-cedd-account-create-gknx9" Oct 14 07:20:17 crc kubenswrapper[4870]: I1014 07:20:17.567892 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-5c98-account-create-9bjhh" Oct 14 07:20:17 crc kubenswrapper[4870]: I1014 07:20:17.568136 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-cedd-account-create-gknx9" event={"ID":"f5e5bd22-1692-477a-a375-3eecad66c746","Type":"ContainerDied","Data":"7ccd8ae947ffe6a30545392fb1fe21a2d1b9c90abf105e1d605dbaa67d72480e"} Oct 14 07:20:17 crc kubenswrapper[4870]: I1014 07:20:17.568178 4870 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7ccd8ae947ffe6a30545392fb1fe21a2d1b9c90abf105e1d605dbaa67d72480e" Oct 14 07:20:17 crc kubenswrapper[4870]: I1014 07:20:17.567898 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7cc794b75c-d2b9c" Oct 14 07:20:17 crc kubenswrapper[4870]: I1014 07:20:17.571107 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=3.571091161 podStartE2EDuration="3.571091161s" podCreationTimestamp="2025-10-14 07:20:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:20:17.551931897 +0000 UTC m=+1153.249292328" watchObservedRunningTime="2025-10-14 07:20:17.571091161 +0000 UTC m=+1153.268451532" Oct 14 07:20:17 crc kubenswrapper[4870]: I1014 07:20:17.601112 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-cbhx4" podStartSLOduration=2.601092644 podStartE2EDuration="2.601092644s" podCreationTimestamp="2025-10-14 07:20:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:20:17.596740346 +0000 UTC m=+1153.294100717" watchObservedRunningTime="2025-10-14 07:20:17.601092644 +0000 UTC m=+1153.298453005" Oct 14 07:20:17 crc kubenswrapper[4870]: I1014 07:20:17.631365 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7cc794b75c-d2b9c"] Oct 14 07:20:17 crc kubenswrapper[4870]: I1014 07:20:17.637761 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7cc794b75c-d2b9c"] Oct 14 07:20:18 crc kubenswrapper[4870]: I1014 07:20:18.279109 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-h587c"] Oct 14 07:20:18 crc kubenswrapper[4870]: E1014 07:20:18.280472 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="69a02fec-1745-4827-9c90-d9dffb829a66" containerName="init" Oct 14 07:20:18 crc kubenswrapper[4870]: I1014 07:20:18.280492 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="69a02fec-1745-4827-9c90-d9dffb829a66" containerName="init" Oct 14 07:20:18 crc kubenswrapper[4870]: E1014 07:20:18.280508 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="69a02fec-1745-4827-9c90-d9dffb829a66" containerName="dnsmasq-dns" Oct 14 07:20:18 crc kubenswrapper[4870]: I1014 07:20:18.280517 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="69a02fec-1745-4827-9c90-d9dffb829a66" containerName="dnsmasq-dns" Oct 14 07:20:18 crc kubenswrapper[4870]: E1014 07:20:18.280563 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5056171a-6fc0-4d48-9ff0-e275137253f3" containerName="mariadb-account-create" Oct 14 07:20:18 crc kubenswrapper[4870]: I1014 07:20:18.280570 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="5056171a-6fc0-4d48-9ff0-e275137253f3" containerName="mariadb-account-create" Oct 14 07:20:18 crc kubenswrapper[4870]: E1014 07:20:18.280583 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="625bb722-0865-47f2-b680-ab642ee6c9e5" containerName="mariadb-account-create" Oct 14 07:20:18 crc kubenswrapper[4870]: I1014 07:20:18.280589 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="625bb722-0865-47f2-b680-ab642ee6c9e5" containerName="mariadb-account-create" Oct 14 07:20:18 crc kubenswrapper[4870]: E1014 07:20:18.280610 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5e5bd22-1692-477a-a375-3eecad66c746" containerName="mariadb-account-create" Oct 14 07:20:18 crc kubenswrapper[4870]: I1014 07:20:18.280616 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5e5bd22-1692-477a-a375-3eecad66c746" containerName="mariadb-account-create" Oct 14 07:20:18 crc kubenswrapper[4870]: I1014 07:20:18.280876 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="69a02fec-1745-4827-9c90-d9dffb829a66" containerName="dnsmasq-dns" Oct 14 07:20:18 crc kubenswrapper[4870]: I1014 07:20:18.280891 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="625bb722-0865-47f2-b680-ab642ee6c9e5" containerName="mariadb-account-create" Oct 14 07:20:18 crc kubenswrapper[4870]: I1014 07:20:18.280907 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="5056171a-6fc0-4d48-9ff0-e275137253f3" containerName="mariadb-account-create" Oct 14 07:20:18 crc kubenswrapper[4870]: I1014 07:20:18.280936 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="f5e5bd22-1692-477a-a375-3eecad66c746" containerName="mariadb-account-create" Oct 14 07:20:18 crc kubenswrapper[4870]: I1014 07:20:18.281940 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-h587c" Oct 14 07:20:18 crc kubenswrapper[4870]: I1014 07:20:18.285797 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Oct 14 07:20:18 crc kubenswrapper[4870]: I1014 07:20:18.288027 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-h587c"] Oct 14 07:20:18 crc kubenswrapper[4870]: I1014 07:20:18.289356 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-ccbhx" Oct 14 07:20:18 crc kubenswrapper[4870]: I1014 07:20:18.289577 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Oct 14 07:20:18 crc kubenswrapper[4870]: I1014 07:20:18.380206 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-56zkt\" (UniqueName: \"kubernetes.io/projected/000463bc-f3ac-40b7-90c8-32366e327a19-kube-api-access-56zkt\") pod \"cinder-db-sync-h587c\" (UID: \"000463bc-f3ac-40b7-90c8-32366e327a19\") " pod="openstack/cinder-db-sync-h587c" Oct 14 07:20:18 crc kubenswrapper[4870]: I1014 07:20:18.380258 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/000463bc-f3ac-40b7-90c8-32366e327a19-db-sync-config-data\") pod \"cinder-db-sync-h587c\" (UID: \"000463bc-f3ac-40b7-90c8-32366e327a19\") " pod="openstack/cinder-db-sync-h587c" Oct 14 07:20:18 crc kubenswrapper[4870]: I1014 07:20:18.380425 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/000463bc-f3ac-40b7-90c8-32366e327a19-combined-ca-bundle\") pod \"cinder-db-sync-h587c\" (UID: \"000463bc-f3ac-40b7-90c8-32366e327a19\") " pod="openstack/cinder-db-sync-h587c" Oct 14 07:20:18 crc kubenswrapper[4870]: I1014 07:20:18.380580 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/000463bc-f3ac-40b7-90c8-32366e327a19-config-data\") pod \"cinder-db-sync-h587c\" (UID: \"000463bc-f3ac-40b7-90c8-32366e327a19\") " pod="openstack/cinder-db-sync-h587c" Oct 14 07:20:18 crc kubenswrapper[4870]: I1014 07:20:18.380634 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/000463bc-f3ac-40b7-90c8-32366e327a19-scripts\") pod \"cinder-db-sync-h587c\" (UID: \"000463bc-f3ac-40b7-90c8-32366e327a19\") " pod="openstack/cinder-db-sync-h587c" Oct 14 07:20:18 crc kubenswrapper[4870]: I1014 07:20:18.380678 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/000463bc-f3ac-40b7-90c8-32366e327a19-etc-machine-id\") pod \"cinder-db-sync-h587c\" (UID: \"000463bc-f3ac-40b7-90c8-32366e327a19\") " pod="openstack/cinder-db-sync-h587c" Oct 14 07:20:18 crc kubenswrapper[4870]: I1014 07:20:18.482595 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/000463bc-f3ac-40b7-90c8-32366e327a19-config-data\") pod \"cinder-db-sync-h587c\" (UID: \"000463bc-f3ac-40b7-90c8-32366e327a19\") " pod="openstack/cinder-db-sync-h587c" Oct 14 07:20:18 crc kubenswrapper[4870]: I1014 07:20:18.482661 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/000463bc-f3ac-40b7-90c8-32366e327a19-scripts\") pod \"cinder-db-sync-h587c\" (UID: \"000463bc-f3ac-40b7-90c8-32366e327a19\") " pod="openstack/cinder-db-sync-h587c" Oct 14 07:20:18 crc kubenswrapper[4870]: I1014 07:20:18.482690 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/000463bc-f3ac-40b7-90c8-32366e327a19-etc-machine-id\") pod \"cinder-db-sync-h587c\" (UID: \"000463bc-f3ac-40b7-90c8-32366e327a19\") " pod="openstack/cinder-db-sync-h587c" Oct 14 07:20:18 crc kubenswrapper[4870]: I1014 07:20:18.482758 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-56zkt\" (UniqueName: \"kubernetes.io/projected/000463bc-f3ac-40b7-90c8-32366e327a19-kube-api-access-56zkt\") pod \"cinder-db-sync-h587c\" (UID: \"000463bc-f3ac-40b7-90c8-32366e327a19\") " pod="openstack/cinder-db-sync-h587c" Oct 14 07:20:18 crc kubenswrapper[4870]: I1014 07:20:18.482802 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/000463bc-f3ac-40b7-90c8-32366e327a19-db-sync-config-data\") pod \"cinder-db-sync-h587c\" (UID: \"000463bc-f3ac-40b7-90c8-32366e327a19\") " pod="openstack/cinder-db-sync-h587c" Oct 14 07:20:18 crc kubenswrapper[4870]: I1014 07:20:18.482843 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/000463bc-f3ac-40b7-90c8-32366e327a19-combined-ca-bundle\") pod \"cinder-db-sync-h587c\" (UID: \"000463bc-f3ac-40b7-90c8-32366e327a19\") " pod="openstack/cinder-db-sync-h587c" Oct 14 07:20:18 crc kubenswrapper[4870]: I1014 07:20:18.482941 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/000463bc-f3ac-40b7-90c8-32366e327a19-etc-machine-id\") pod \"cinder-db-sync-h587c\" (UID: \"000463bc-f3ac-40b7-90c8-32366e327a19\") " pod="openstack/cinder-db-sync-h587c" Oct 14 07:20:18 crc kubenswrapper[4870]: I1014 07:20:18.489386 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/000463bc-f3ac-40b7-90c8-32366e327a19-scripts\") pod \"cinder-db-sync-h587c\" (UID: \"000463bc-f3ac-40b7-90c8-32366e327a19\") " pod="openstack/cinder-db-sync-h587c" Oct 14 07:20:18 crc kubenswrapper[4870]: I1014 07:20:18.490271 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/000463bc-f3ac-40b7-90c8-32366e327a19-db-sync-config-data\") pod \"cinder-db-sync-h587c\" (UID: \"000463bc-f3ac-40b7-90c8-32366e327a19\") " pod="openstack/cinder-db-sync-h587c" Oct 14 07:20:18 crc kubenswrapper[4870]: I1014 07:20:18.494548 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/000463bc-f3ac-40b7-90c8-32366e327a19-config-data\") pod \"cinder-db-sync-h587c\" (UID: \"000463bc-f3ac-40b7-90c8-32366e327a19\") " pod="openstack/cinder-db-sync-h587c" Oct 14 07:20:18 crc kubenswrapper[4870]: I1014 07:20:18.501871 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/000463bc-f3ac-40b7-90c8-32366e327a19-combined-ca-bundle\") pod \"cinder-db-sync-h587c\" (UID: \"000463bc-f3ac-40b7-90c8-32366e327a19\") " pod="openstack/cinder-db-sync-h587c" Oct 14 07:20:18 crc kubenswrapper[4870]: I1014 07:20:18.505090 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-56zkt\" (UniqueName: \"kubernetes.io/projected/000463bc-f3ac-40b7-90c8-32366e327a19-kube-api-access-56zkt\") pod \"cinder-db-sync-h587c\" (UID: \"000463bc-f3ac-40b7-90c8-32366e327a19\") " pod="openstack/cinder-db-sync-h587c" Oct 14 07:20:18 crc kubenswrapper[4870]: I1014 07:20:18.576084 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"452a6ce6-d9dc-4455-9304-414701c8895f","Type":"ContainerStarted","Data":"5db90be32cea50de9fa48a1c7946854f77a0364acaf6f2929af1ec8f226fb41f"} Oct 14 07:20:18 crc kubenswrapper[4870]: I1014 07:20:18.576128 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-97xhm"] Oct 14 07:20:18 crc kubenswrapper[4870]: I1014 07:20:18.577728 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-97xhm" Oct 14 07:20:18 crc kubenswrapper[4870]: I1014 07:20:18.580933 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-2s5xt" Oct 14 07:20:18 crc kubenswrapper[4870]: I1014 07:20:18.582424 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Oct 14 07:20:18 crc kubenswrapper[4870]: I1014 07:20:18.586030 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-97xhm"] Oct 14 07:20:18 crc kubenswrapper[4870]: I1014 07:20:18.610642 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-h587c" Oct 14 07:20:18 crc kubenswrapper[4870]: I1014 07:20:18.616466 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=4.616429299 podStartE2EDuration="4.616429299s" podCreationTimestamp="2025-10-14 07:20:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:20:18.597866719 +0000 UTC m=+1154.295227090" watchObservedRunningTime="2025-10-14 07:20:18.616429299 +0000 UTC m=+1154.313789670" Oct 14 07:20:18 crc kubenswrapper[4870]: I1014 07:20:18.667694 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-vtmvg"] Oct 14 07:20:18 crc kubenswrapper[4870]: I1014 07:20:18.669739 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-vtmvg" Oct 14 07:20:18 crc kubenswrapper[4870]: I1014 07:20:18.673582 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-5whvl" Oct 14 07:20:18 crc kubenswrapper[4870]: I1014 07:20:18.674197 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Oct 14 07:20:18 crc kubenswrapper[4870]: I1014 07:20:18.675634 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Oct 14 07:20:18 crc kubenswrapper[4870]: I1014 07:20:18.695777 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/6e87f821-3bc7-4b88-9c72-ccd80b4b19e9-db-sync-config-data\") pod \"barbican-db-sync-97xhm\" (UID: \"6e87f821-3bc7-4b88-9c72-ccd80b4b19e9\") " pod="openstack/barbican-db-sync-97xhm" Oct 14 07:20:18 crc kubenswrapper[4870]: I1014 07:20:18.695909 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e87f821-3bc7-4b88-9c72-ccd80b4b19e9-combined-ca-bundle\") pod \"barbican-db-sync-97xhm\" (UID: \"6e87f821-3bc7-4b88-9c72-ccd80b4b19e9\") " pod="openstack/barbican-db-sync-97xhm" Oct 14 07:20:18 crc kubenswrapper[4870]: I1014 07:20:18.695933 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mfv7v\" (UniqueName: \"kubernetes.io/projected/6e87f821-3bc7-4b88-9c72-ccd80b4b19e9-kube-api-access-mfv7v\") pod \"barbican-db-sync-97xhm\" (UID: \"6e87f821-3bc7-4b88-9c72-ccd80b4b19e9\") " pod="openstack/barbican-db-sync-97xhm" Oct 14 07:20:18 crc kubenswrapper[4870]: I1014 07:20:18.730163 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-vtmvg"] Oct 14 07:20:18 crc kubenswrapper[4870]: I1014 07:20:18.801140 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/6e87f821-3bc7-4b88-9c72-ccd80b4b19e9-db-sync-config-data\") pod \"barbican-db-sync-97xhm\" (UID: \"6e87f821-3bc7-4b88-9c72-ccd80b4b19e9\") " pod="openstack/barbican-db-sync-97xhm" Oct 14 07:20:18 crc kubenswrapper[4870]: I1014 07:20:18.801252 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e87f821-3bc7-4b88-9c72-ccd80b4b19e9-combined-ca-bundle\") pod \"barbican-db-sync-97xhm\" (UID: \"6e87f821-3bc7-4b88-9c72-ccd80b4b19e9\") " pod="openstack/barbican-db-sync-97xhm" Oct 14 07:20:18 crc kubenswrapper[4870]: I1014 07:20:18.801283 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mfv7v\" (UniqueName: \"kubernetes.io/projected/6e87f821-3bc7-4b88-9c72-ccd80b4b19e9-kube-api-access-mfv7v\") pod \"barbican-db-sync-97xhm\" (UID: \"6e87f821-3bc7-4b88-9c72-ccd80b4b19e9\") " pod="openstack/barbican-db-sync-97xhm" Oct 14 07:20:18 crc kubenswrapper[4870]: I1014 07:20:18.801313 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/685539fb-44a2-4f75-b418-65383d02152e-config\") pod \"neutron-db-sync-vtmvg\" (UID: \"685539fb-44a2-4f75-b418-65383d02152e\") " pod="openstack/neutron-db-sync-vtmvg" Oct 14 07:20:18 crc kubenswrapper[4870]: I1014 07:20:18.801348 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7vrfg\" (UniqueName: \"kubernetes.io/projected/685539fb-44a2-4f75-b418-65383d02152e-kube-api-access-7vrfg\") pod \"neutron-db-sync-vtmvg\" (UID: \"685539fb-44a2-4f75-b418-65383d02152e\") " pod="openstack/neutron-db-sync-vtmvg" Oct 14 07:20:18 crc kubenswrapper[4870]: I1014 07:20:18.801392 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/685539fb-44a2-4f75-b418-65383d02152e-combined-ca-bundle\") pod \"neutron-db-sync-vtmvg\" (UID: \"685539fb-44a2-4f75-b418-65383d02152e\") " pod="openstack/neutron-db-sync-vtmvg" Oct 14 07:20:18 crc kubenswrapper[4870]: I1014 07:20:18.808963 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/6e87f821-3bc7-4b88-9c72-ccd80b4b19e9-db-sync-config-data\") pod \"barbican-db-sync-97xhm\" (UID: \"6e87f821-3bc7-4b88-9c72-ccd80b4b19e9\") " pod="openstack/barbican-db-sync-97xhm" Oct 14 07:20:18 crc kubenswrapper[4870]: I1014 07:20:18.816556 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e87f821-3bc7-4b88-9c72-ccd80b4b19e9-combined-ca-bundle\") pod \"barbican-db-sync-97xhm\" (UID: \"6e87f821-3bc7-4b88-9c72-ccd80b4b19e9\") " pod="openstack/barbican-db-sync-97xhm" Oct 14 07:20:18 crc kubenswrapper[4870]: I1014 07:20:18.841965 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mfv7v\" (UniqueName: \"kubernetes.io/projected/6e87f821-3bc7-4b88-9c72-ccd80b4b19e9-kube-api-access-mfv7v\") pod \"barbican-db-sync-97xhm\" (UID: \"6e87f821-3bc7-4b88-9c72-ccd80b4b19e9\") " pod="openstack/barbican-db-sync-97xhm" Oct 14 07:20:18 crc kubenswrapper[4870]: I1014 07:20:18.904624 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/685539fb-44a2-4f75-b418-65383d02152e-combined-ca-bundle\") pod \"neutron-db-sync-vtmvg\" (UID: \"685539fb-44a2-4f75-b418-65383d02152e\") " pod="openstack/neutron-db-sync-vtmvg" Oct 14 07:20:18 crc kubenswrapper[4870]: I1014 07:20:18.906171 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/685539fb-44a2-4f75-b418-65383d02152e-config\") pod \"neutron-db-sync-vtmvg\" (UID: \"685539fb-44a2-4f75-b418-65383d02152e\") " pod="openstack/neutron-db-sync-vtmvg" Oct 14 07:20:18 crc kubenswrapper[4870]: I1014 07:20:18.906244 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7vrfg\" (UniqueName: \"kubernetes.io/projected/685539fb-44a2-4f75-b418-65383d02152e-kube-api-access-7vrfg\") pod \"neutron-db-sync-vtmvg\" (UID: \"685539fb-44a2-4f75-b418-65383d02152e\") " pod="openstack/neutron-db-sync-vtmvg" Oct 14 07:20:18 crc kubenswrapper[4870]: I1014 07:20:18.911222 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-97xhm" Oct 14 07:20:18 crc kubenswrapper[4870]: I1014 07:20:18.913288 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/685539fb-44a2-4f75-b418-65383d02152e-config\") pod \"neutron-db-sync-vtmvg\" (UID: \"685539fb-44a2-4f75-b418-65383d02152e\") " pod="openstack/neutron-db-sync-vtmvg" Oct 14 07:20:18 crc kubenswrapper[4870]: I1014 07:20:18.917186 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/685539fb-44a2-4f75-b418-65383d02152e-combined-ca-bundle\") pod \"neutron-db-sync-vtmvg\" (UID: \"685539fb-44a2-4f75-b418-65383d02152e\") " pod="openstack/neutron-db-sync-vtmvg" Oct 14 07:20:18 crc kubenswrapper[4870]: I1014 07:20:18.927825 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7vrfg\" (UniqueName: \"kubernetes.io/projected/685539fb-44a2-4f75-b418-65383d02152e-kube-api-access-7vrfg\") pod \"neutron-db-sync-vtmvg\" (UID: \"685539fb-44a2-4f75-b418-65383d02152e\") " pod="openstack/neutron-db-sync-vtmvg" Oct 14 07:20:19 crc kubenswrapper[4870]: I1014 07:20:19.031245 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-nmtf9" Oct 14 07:20:19 crc kubenswrapper[4870]: I1014 07:20:19.035585 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-vtmvg" Oct 14 07:20:19 crc kubenswrapper[4870]: I1014 07:20:19.048115 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="69a02fec-1745-4827-9c90-d9dffb829a66" path="/var/lib/kubelet/pods/69a02fec-1745-4827-9c90-d9dffb829a66/volumes" Oct 14 07:20:19 crc kubenswrapper[4870]: I1014 07:20:19.109258 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9dd93b33-f231-4339-9db0-0507cde1dcf5-logs\") pod \"9dd93b33-f231-4339-9db0-0507cde1dcf5\" (UID: \"9dd93b33-f231-4339-9db0-0507cde1dcf5\") " Oct 14 07:20:19 crc kubenswrapper[4870]: I1014 07:20:19.109375 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9dd93b33-f231-4339-9db0-0507cde1dcf5-combined-ca-bundle\") pod \"9dd93b33-f231-4339-9db0-0507cde1dcf5\" (UID: \"9dd93b33-f231-4339-9db0-0507cde1dcf5\") " Oct 14 07:20:19 crc kubenswrapper[4870]: I1014 07:20:19.109430 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9dd93b33-f231-4339-9db0-0507cde1dcf5-scripts\") pod \"9dd93b33-f231-4339-9db0-0507cde1dcf5\" (UID: \"9dd93b33-f231-4339-9db0-0507cde1dcf5\") " Oct 14 07:20:19 crc kubenswrapper[4870]: I1014 07:20:19.109478 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4flk9\" (UniqueName: \"kubernetes.io/projected/9dd93b33-f231-4339-9db0-0507cde1dcf5-kube-api-access-4flk9\") pod \"9dd93b33-f231-4339-9db0-0507cde1dcf5\" (UID: \"9dd93b33-f231-4339-9db0-0507cde1dcf5\") " Oct 14 07:20:19 crc kubenswrapper[4870]: I1014 07:20:19.109495 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9dd93b33-f231-4339-9db0-0507cde1dcf5-config-data\") pod \"9dd93b33-f231-4339-9db0-0507cde1dcf5\" (UID: \"9dd93b33-f231-4339-9db0-0507cde1dcf5\") " Oct 14 07:20:19 crc kubenswrapper[4870]: I1014 07:20:19.109911 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9dd93b33-f231-4339-9db0-0507cde1dcf5-logs" (OuterVolumeSpecName: "logs") pod "9dd93b33-f231-4339-9db0-0507cde1dcf5" (UID: "9dd93b33-f231-4339-9db0-0507cde1dcf5"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:20:19 crc kubenswrapper[4870]: I1014 07:20:19.110008 4870 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9dd93b33-f231-4339-9db0-0507cde1dcf5-logs\") on node \"crc\" DevicePath \"\"" Oct 14 07:20:19 crc kubenswrapper[4870]: I1014 07:20:19.114122 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9dd93b33-f231-4339-9db0-0507cde1dcf5-kube-api-access-4flk9" (OuterVolumeSpecName: "kube-api-access-4flk9") pod "9dd93b33-f231-4339-9db0-0507cde1dcf5" (UID: "9dd93b33-f231-4339-9db0-0507cde1dcf5"). InnerVolumeSpecName "kube-api-access-4flk9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:20:19 crc kubenswrapper[4870]: I1014 07:20:19.116575 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9dd93b33-f231-4339-9db0-0507cde1dcf5-scripts" (OuterVolumeSpecName: "scripts") pod "9dd93b33-f231-4339-9db0-0507cde1dcf5" (UID: "9dd93b33-f231-4339-9db0-0507cde1dcf5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:20:19 crc kubenswrapper[4870]: I1014 07:20:19.155253 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9dd93b33-f231-4339-9db0-0507cde1dcf5-config-data" (OuterVolumeSpecName: "config-data") pod "9dd93b33-f231-4339-9db0-0507cde1dcf5" (UID: "9dd93b33-f231-4339-9db0-0507cde1dcf5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:20:19 crc kubenswrapper[4870]: I1014 07:20:19.168668 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9dd93b33-f231-4339-9db0-0507cde1dcf5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9dd93b33-f231-4339-9db0-0507cde1dcf5" (UID: "9dd93b33-f231-4339-9db0-0507cde1dcf5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:20:19 crc kubenswrapper[4870]: I1014 07:20:19.212054 4870 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9dd93b33-f231-4339-9db0-0507cde1dcf5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 07:20:19 crc kubenswrapper[4870]: I1014 07:20:19.212090 4870 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9dd93b33-f231-4339-9db0-0507cde1dcf5-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 07:20:19 crc kubenswrapper[4870]: I1014 07:20:19.212101 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4flk9\" (UniqueName: \"kubernetes.io/projected/9dd93b33-f231-4339-9db0-0507cde1dcf5-kube-api-access-4flk9\") on node \"crc\" DevicePath \"\"" Oct 14 07:20:19 crc kubenswrapper[4870]: I1014 07:20:19.212112 4870 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9dd93b33-f231-4339-9db0-0507cde1dcf5-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 07:20:19 crc kubenswrapper[4870]: I1014 07:20:19.279166 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-h587c"] Oct 14 07:20:19 crc kubenswrapper[4870]: W1014 07:20:19.291416 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod000463bc_f3ac_40b7_90c8_32366e327a19.slice/crio-7c2eae11e0d67fb773baaf6186f3925313a3290f9b8bf53581a83059721ef23e WatchSource:0}: Error finding container 7c2eae11e0d67fb773baaf6186f3925313a3290f9b8bf53581a83059721ef23e: Status 404 returned error can't find the container with id 7c2eae11e0d67fb773baaf6186f3925313a3290f9b8bf53581a83059721ef23e Oct 14 07:20:19 crc kubenswrapper[4870]: I1014 07:20:19.468148 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-97xhm"] Oct 14 07:20:19 crc kubenswrapper[4870]: W1014 07:20:19.476244 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6e87f821_3bc7_4b88_9c72_ccd80b4b19e9.slice/crio-4e29b89487e5efa0447752c0a7feaef2666e899bf747ab5311fe6af99d9aa374 WatchSource:0}: Error finding container 4e29b89487e5efa0447752c0a7feaef2666e899bf747ab5311fe6af99d9aa374: Status 404 returned error can't find the container with id 4e29b89487e5efa0447752c0a7feaef2666e899bf747ab5311fe6af99d9aa374 Oct 14 07:20:19 crc kubenswrapper[4870]: I1014 07:20:19.595329 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-vtmvg"] Oct 14 07:20:19 crc kubenswrapper[4870]: W1014 07:20:19.609770 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod685539fb_44a2_4f75_b418_65383d02152e.slice/crio-5af8a8a8e81f5a73f71a578bdc2386e1bf1644cbcca6656c79c291aff9014c4e WatchSource:0}: Error finding container 5af8a8a8e81f5a73f71a578bdc2386e1bf1644cbcca6656c79c291aff9014c4e: Status 404 returned error can't find the container with id 5af8a8a8e81f5a73f71a578bdc2386e1bf1644cbcca6656c79c291aff9014c4e Oct 14 07:20:19 crc kubenswrapper[4870]: I1014 07:20:19.610140 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-nmtf9" event={"ID":"9dd93b33-f231-4339-9db0-0507cde1dcf5","Type":"ContainerDied","Data":"45ff0c95eb09cc7c56cbb64dd645167ccb647aa60075375000048be59874f93b"} Oct 14 07:20:19 crc kubenswrapper[4870]: I1014 07:20:19.610214 4870 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="45ff0c95eb09cc7c56cbb64dd645167ccb647aa60075375000048be59874f93b" Oct 14 07:20:19 crc kubenswrapper[4870]: I1014 07:20:19.610322 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-nmtf9" Oct 14 07:20:19 crc kubenswrapper[4870]: I1014 07:20:19.614511 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-h587c" event={"ID":"000463bc-f3ac-40b7-90c8-32366e327a19","Type":"ContainerStarted","Data":"7c2eae11e0d67fb773baaf6186f3925313a3290f9b8bf53581a83059721ef23e"} Oct 14 07:20:19 crc kubenswrapper[4870]: I1014 07:20:19.619121 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-97xhm" event={"ID":"6e87f821-3bc7-4b88-9c72-ccd80b4b19e9","Type":"ContainerStarted","Data":"4e29b89487e5efa0447752c0a7feaef2666e899bf747ab5311fe6af99d9aa374"} Oct 14 07:20:19 crc kubenswrapper[4870]: I1014 07:20:19.758078 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-597ff968b-msxbn"] Oct 14 07:20:19 crc kubenswrapper[4870]: E1014 07:20:19.759002 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9dd93b33-f231-4339-9db0-0507cde1dcf5" containerName="placement-db-sync" Oct 14 07:20:19 crc kubenswrapper[4870]: I1014 07:20:19.759026 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="9dd93b33-f231-4339-9db0-0507cde1dcf5" containerName="placement-db-sync" Oct 14 07:20:19 crc kubenswrapper[4870]: I1014 07:20:19.759301 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="9dd93b33-f231-4339-9db0-0507cde1dcf5" containerName="placement-db-sync" Oct 14 07:20:19 crc kubenswrapper[4870]: I1014 07:20:19.760315 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-597ff968b-msxbn" Oct 14 07:20:19 crc kubenswrapper[4870]: I1014 07:20:19.765734 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Oct 14 07:20:19 crc kubenswrapper[4870]: I1014 07:20:19.765774 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Oct 14 07:20:19 crc kubenswrapper[4870]: I1014 07:20:19.766038 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-zsgq8" Oct 14 07:20:19 crc kubenswrapper[4870]: I1014 07:20:19.767648 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Oct 14 07:20:19 crc kubenswrapper[4870]: I1014 07:20:19.767907 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Oct 14 07:20:19 crc kubenswrapper[4870]: I1014 07:20:19.779993 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-597ff968b-msxbn"] Oct 14 07:20:19 crc kubenswrapper[4870]: I1014 07:20:19.832042 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2130e3f2-cbea-46e1-b5d3-daefdd972935-scripts\") pod \"placement-597ff968b-msxbn\" (UID: \"2130e3f2-cbea-46e1-b5d3-daefdd972935\") " pod="openstack/placement-597ff968b-msxbn" Oct 14 07:20:19 crc kubenswrapper[4870]: I1014 07:20:19.832080 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2130e3f2-cbea-46e1-b5d3-daefdd972935-logs\") pod \"placement-597ff968b-msxbn\" (UID: \"2130e3f2-cbea-46e1-b5d3-daefdd972935\") " pod="openstack/placement-597ff968b-msxbn" Oct 14 07:20:19 crc kubenswrapper[4870]: I1014 07:20:19.832119 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2130e3f2-cbea-46e1-b5d3-daefdd972935-public-tls-certs\") pod \"placement-597ff968b-msxbn\" (UID: \"2130e3f2-cbea-46e1-b5d3-daefdd972935\") " pod="openstack/placement-597ff968b-msxbn" Oct 14 07:20:19 crc kubenswrapper[4870]: I1014 07:20:19.832147 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2130e3f2-cbea-46e1-b5d3-daefdd972935-combined-ca-bundle\") pod \"placement-597ff968b-msxbn\" (UID: \"2130e3f2-cbea-46e1-b5d3-daefdd972935\") " pod="openstack/placement-597ff968b-msxbn" Oct 14 07:20:19 crc kubenswrapper[4870]: I1014 07:20:19.832207 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2130e3f2-cbea-46e1-b5d3-daefdd972935-config-data\") pod \"placement-597ff968b-msxbn\" (UID: \"2130e3f2-cbea-46e1-b5d3-daefdd972935\") " pod="openstack/placement-597ff968b-msxbn" Oct 14 07:20:19 crc kubenswrapper[4870]: I1014 07:20:19.832226 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8rt82\" (UniqueName: \"kubernetes.io/projected/2130e3f2-cbea-46e1-b5d3-daefdd972935-kube-api-access-8rt82\") pod \"placement-597ff968b-msxbn\" (UID: \"2130e3f2-cbea-46e1-b5d3-daefdd972935\") " pod="openstack/placement-597ff968b-msxbn" Oct 14 07:20:19 crc kubenswrapper[4870]: I1014 07:20:19.832250 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2130e3f2-cbea-46e1-b5d3-daefdd972935-internal-tls-certs\") pod \"placement-597ff968b-msxbn\" (UID: \"2130e3f2-cbea-46e1-b5d3-daefdd972935\") " pod="openstack/placement-597ff968b-msxbn" Oct 14 07:20:19 crc kubenswrapper[4870]: I1014 07:20:19.933995 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2130e3f2-cbea-46e1-b5d3-daefdd972935-internal-tls-certs\") pod \"placement-597ff968b-msxbn\" (UID: \"2130e3f2-cbea-46e1-b5d3-daefdd972935\") " pod="openstack/placement-597ff968b-msxbn" Oct 14 07:20:19 crc kubenswrapper[4870]: I1014 07:20:19.934118 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2130e3f2-cbea-46e1-b5d3-daefdd972935-logs\") pod \"placement-597ff968b-msxbn\" (UID: \"2130e3f2-cbea-46e1-b5d3-daefdd972935\") " pod="openstack/placement-597ff968b-msxbn" Oct 14 07:20:19 crc kubenswrapper[4870]: I1014 07:20:19.934150 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2130e3f2-cbea-46e1-b5d3-daefdd972935-scripts\") pod \"placement-597ff968b-msxbn\" (UID: \"2130e3f2-cbea-46e1-b5d3-daefdd972935\") " pod="openstack/placement-597ff968b-msxbn" Oct 14 07:20:19 crc kubenswrapper[4870]: I1014 07:20:19.934202 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2130e3f2-cbea-46e1-b5d3-daefdd972935-public-tls-certs\") pod \"placement-597ff968b-msxbn\" (UID: \"2130e3f2-cbea-46e1-b5d3-daefdd972935\") " pod="openstack/placement-597ff968b-msxbn" Oct 14 07:20:19 crc kubenswrapper[4870]: I1014 07:20:19.934270 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2130e3f2-cbea-46e1-b5d3-daefdd972935-combined-ca-bundle\") pod \"placement-597ff968b-msxbn\" (UID: \"2130e3f2-cbea-46e1-b5d3-daefdd972935\") " pod="openstack/placement-597ff968b-msxbn" Oct 14 07:20:19 crc kubenswrapper[4870]: I1014 07:20:19.934758 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2130e3f2-cbea-46e1-b5d3-daefdd972935-logs\") pod \"placement-597ff968b-msxbn\" (UID: \"2130e3f2-cbea-46e1-b5d3-daefdd972935\") " pod="openstack/placement-597ff968b-msxbn" Oct 14 07:20:19 crc kubenswrapper[4870]: I1014 07:20:19.935645 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2130e3f2-cbea-46e1-b5d3-daefdd972935-config-data\") pod \"placement-597ff968b-msxbn\" (UID: \"2130e3f2-cbea-46e1-b5d3-daefdd972935\") " pod="openstack/placement-597ff968b-msxbn" Oct 14 07:20:19 crc kubenswrapper[4870]: I1014 07:20:19.935683 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8rt82\" (UniqueName: \"kubernetes.io/projected/2130e3f2-cbea-46e1-b5d3-daefdd972935-kube-api-access-8rt82\") pod \"placement-597ff968b-msxbn\" (UID: \"2130e3f2-cbea-46e1-b5d3-daefdd972935\") " pod="openstack/placement-597ff968b-msxbn" Oct 14 07:20:19 crc kubenswrapper[4870]: I1014 07:20:19.939689 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2130e3f2-cbea-46e1-b5d3-daefdd972935-combined-ca-bundle\") pod \"placement-597ff968b-msxbn\" (UID: \"2130e3f2-cbea-46e1-b5d3-daefdd972935\") " pod="openstack/placement-597ff968b-msxbn" Oct 14 07:20:19 crc kubenswrapper[4870]: I1014 07:20:19.940223 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2130e3f2-cbea-46e1-b5d3-daefdd972935-internal-tls-certs\") pod \"placement-597ff968b-msxbn\" (UID: \"2130e3f2-cbea-46e1-b5d3-daefdd972935\") " pod="openstack/placement-597ff968b-msxbn" Oct 14 07:20:19 crc kubenswrapper[4870]: I1014 07:20:19.941039 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2130e3f2-cbea-46e1-b5d3-daefdd972935-public-tls-certs\") pod \"placement-597ff968b-msxbn\" (UID: \"2130e3f2-cbea-46e1-b5d3-daefdd972935\") " pod="openstack/placement-597ff968b-msxbn" Oct 14 07:20:19 crc kubenswrapper[4870]: I1014 07:20:19.941133 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2130e3f2-cbea-46e1-b5d3-daefdd972935-scripts\") pod \"placement-597ff968b-msxbn\" (UID: \"2130e3f2-cbea-46e1-b5d3-daefdd972935\") " pod="openstack/placement-597ff968b-msxbn" Oct 14 07:20:19 crc kubenswrapper[4870]: I1014 07:20:19.956218 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2130e3f2-cbea-46e1-b5d3-daefdd972935-config-data\") pod \"placement-597ff968b-msxbn\" (UID: \"2130e3f2-cbea-46e1-b5d3-daefdd972935\") " pod="openstack/placement-597ff968b-msxbn" Oct 14 07:20:19 crc kubenswrapper[4870]: I1014 07:20:19.956651 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8rt82\" (UniqueName: \"kubernetes.io/projected/2130e3f2-cbea-46e1-b5d3-daefdd972935-kube-api-access-8rt82\") pod \"placement-597ff968b-msxbn\" (UID: \"2130e3f2-cbea-46e1-b5d3-daefdd972935\") " pod="openstack/placement-597ff968b-msxbn" Oct 14 07:20:20 crc kubenswrapper[4870]: I1014 07:20:20.118322 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-597ff968b-msxbn" Oct 14 07:20:20 crc kubenswrapper[4870]: I1014 07:20:20.628101 4870 generic.go:334] "Generic (PLEG): container finished" podID="fe66a2e1-364a-4cb9-b857-c208c7d1b108" containerID="a82c41053072124e1b79993189c4bbf5862bb71cd34f3bf9755173e9a3637eb1" exitCode=0 Oct 14 07:20:20 crc kubenswrapper[4870]: I1014 07:20:20.628403 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-cbhx4" event={"ID":"fe66a2e1-364a-4cb9-b857-c208c7d1b108","Type":"ContainerDied","Data":"a82c41053072124e1b79993189c4bbf5862bb71cd34f3bf9755173e9a3637eb1"} Oct 14 07:20:20 crc kubenswrapper[4870]: I1014 07:20:20.631139 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-vtmvg" event={"ID":"685539fb-44a2-4f75-b418-65383d02152e","Type":"ContainerStarted","Data":"28a8e8e086b6d6c6e4aa03f701ff3d36e5350745361841619c40f9371ec5f443"} Oct 14 07:20:20 crc kubenswrapper[4870]: I1014 07:20:20.631165 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-vtmvg" event={"ID":"685539fb-44a2-4f75-b418-65383d02152e","Type":"ContainerStarted","Data":"5af8a8a8e81f5a73f71a578bdc2386e1bf1644cbcca6656c79c291aff9014c4e"} Oct 14 07:20:20 crc kubenswrapper[4870]: I1014 07:20:20.676521 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-vtmvg" podStartSLOduration=2.676502145 podStartE2EDuration="2.676502145s" podCreationTimestamp="2025-10-14 07:20:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:20:20.671381579 +0000 UTC m=+1156.368741940" watchObservedRunningTime="2025-10-14 07:20:20.676502145 +0000 UTC m=+1156.373862516" Oct 14 07:20:21 crc kubenswrapper[4870]: I1014 07:20:21.497228 4870 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-7cc794b75c-d2b9c" podUID="69a02fec-1745-4827-9c90-d9dffb829a66" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.126:5353: i/o timeout" Oct 14 07:20:22 crc kubenswrapper[4870]: I1014 07:20:22.796063 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-cbhx4" Oct 14 07:20:22 crc kubenswrapper[4870]: I1014 07:20:22.894335 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fe66a2e1-364a-4cb9-b857-c208c7d1b108-config-data\") pod \"fe66a2e1-364a-4cb9-b857-c208c7d1b108\" (UID: \"fe66a2e1-364a-4cb9-b857-c208c7d1b108\") " Oct 14 07:20:22 crc kubenswrapper[4870]: I1014 07:20:22.894428 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/fe66a2e1-364a-4cb9-b857-c208c7d1b108-credential-keys\") pod \"fe66a2e1-364a-4cb9-b857-c208c7d1b108\" (UID: \"fe66a2e1-364a-4cb9-b857-c208c7d1b108\") " Oct 14 07:20:22 crc kubenswrapper[4870]: I1014 07:20:22.894493 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe66a2e1-364a-4cb9-b857-c208c7d1b108-combined-ca-bundle\") pod \"fe66a2e1-364a-4cb9-b857-c208c7d1b108\" (UID: \"fe66a2e1-364a-4cb9-b857-c208c7d1b108\") " Oct 14 07:20:22 crc kubenswrapper[4870]: I1014 07:20:22.894519 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fe66a2e1-364a-4cb9-b857-c208c7d1b108-scripts\") pod \"fe66a2e1-364a-4cb9-b857-c208c7d1b108\" (UID: \"fe66a2e1-364a-4cb9-b857-c208c7d1b108\") " Oct 14 07:20:22 crc kubenswrapper[4870]: I1014 07:20:22.894598 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/fe66a2e1-364a-4cb9-b857-c208c7d1b108-fernet-keys\") pod \"fe66a2e1-364a-4cb9-b857-c208c7d1b108\" (UID: \"fe66a2e1-364a-4cb9-b857-c208c7d1b108\") " Oct 14 07:20:22 crc kubenswrapper[4870]: I1014 07:20:22.894686 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kxlb4\" (UniqueName: \"kubernetes.io/projected/fe66a2e1-364a-4cb9-b857-c208c7d1b108-kube-api-access-kxlb4\") pod \"fe66a2e1-364a-4cb9-b857-c208c7d1b108\" (UID: \"fe66a2e1-364a-4cb9-b857-c208c7d1b108\") " Oct 14 07:20:22 crc kubenswrapper[4870]: I1014 07:20:22.900728 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fe66a2e1-364a-4cb9-b857-c208c7d1b108-kube-api-access-kxlb4" (OuterVolumeSpecName: "kube-api-access-kxlb4") pod "fe66a2e1-364a-4cb9-b857-c208c7d1b108" (UID: "fe66a2e1-364a-4cb9-b857-c208c7d1b108"). InnerVolumeSpecName "kube-api-access-kxlb4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:20:22 crc kubenswrapper[4870]: I1014 07:20:22.905699 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fe66a2e1-364a-4cb9-b857-c208c7d1b108-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "fe66a2e1-364a-4cb9-b857-c208c7d1b108" (UID: "fe66a2e1-364a-4cb9-b857-c208c7d1b108"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:20:22 crc kubenswrapper[4870]: I1014 07:20:22.908858 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fe66a2e1-364a-4cb9-b857-c208c7d1b108-scripts" (OuterVolumeSpecName: "scripts") pod "fe66a2e1-364a-4cb9-b857-c208c7d1b108" (UID: "fe66a2e1-364a-4cb9-b857-c208c7d1b108"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:20:22 crc kubenswrapper[4870]: I1014 07:20:22.914162 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fe66a2e1-364a-4cb9-b857-c208c7d1b108-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "fe66a2e1-364a-4cb9-b857-c208c7d1b108" (UID: "fe66a2e1-364a-4cb9-b857-c208c7d1b108"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:20:22 crc kubenswrapper[4870]: I1014 07:20:22.941153 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fe66a2e1-364a-4cb9-b857-c208c7d1b108-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fe66a2e1-364a-4cb9-b857-c208c7d1b108" (UID: "fe66a2e1-364a-4cb9-b857-c208c7d1b108"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:20:22 crc kubenswrapper[4870]: I1014 07:20:22.953053 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fe66a2e1-364a-4cb9-b857-c208c7d1b108-config-data" (OuterVolumeSpecName: "config-data") pod "fe66a2e1-364a-4cb9-b857-c208c7d1b108" (UID: "fe66a2e1-364a-4cb9-b857-c208c7d1b108"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:20:22 crc kubenswrapper[4870]: I1014 07:20:22.996783 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kxlb4\" (UniqueName: \"kubernetes.io/projected/fe66a2e1-364a-4cb9-b857-c208c7d1b108-kube-api-access-kxlb4\") on node \"crc\" DevicePath \"\"" Oct 14 07:20:22 crc kubenswrapper[4870]: I1014 07:20:22.997207 4870 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fe66a2e1-364a-4cb9-b857-c208c7d1b108-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 07:20:22 crc kubenswrapper[4870]: I1014 07:20:22.997240 4870 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/fe66a2e1-364a-4cb9-b857-c208c7d1b108-credential-keys\") on node \"crc\" DevicePath \"\"" Oct 14 07:20:22 crc kubenswrapper[4870]: I1014 07:20:22.997250 4870 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fe66a2e1-364a-4cb9-b857-c208c7d1b108-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 07:20:22 crc kubenswrapper[4870]: I1014 07:20:22.997259 4870 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe66a2e1-364a-4cb9-b857-c208c7d1b108-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 07:20:22 crc kubenswrapper[4870]: I1014 07:20:22.997299 4870 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/fe66a2e1-364a-4cb9-b857-c208c7d1b108-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 14 07:20:23 crc kubenswrapper[4870]: I1014 07:20:23.227558 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-597ff968b-msxbn"] Oct 14 07:20:23 crc kubenswrapper[4870]: W1014 07:20:23.241359 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2130e3f2_cbea_46e1_b5d3_daefdd972935.slice/crio-a137c2c8f449d4fac987a138026d8768098fc2fa5e342a9df21e1f2d3ff32046 WatchSource:0}: Error finding container a137c2c8f449d4fac987a138026d8768098fc2fa5e342a9df21e1f2d3ff32046: Status 404 returned error can't find the container with id a137c2c8f449d4fac987a138026d8768098fc2fa5e342a9df21e1f2d3ff32046 Oct 14 07:20:23 crc kubenswrapper[4870]: I1014 07:20:23.674430 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-cbhx4" Oct 14 07:20:23 crc kubenswrapper[4870]: I1014 07:20:23.675171 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-cbhx4" event={"ID":"fe66a2e1-364a-4cb9-b857-c208c7d1b108","Type":"ContainerDied","Data":"af949307c1c161f67b24e0326ac3a341612238074f32d7b2d73053eed3c5b401"} Oct 14 07:20:23 crc kubenswrapper[4870]: I1014 07:20:23.675229 4870 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="af949307c1c161f67b24e0326ac3a341612238074f32d7b2d73053eed3c5b401" Oct 14 07:20:23 crc kubenswrapper[4870]: I1014 07:20:23.678377 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-597ff968b-msxbn" event={"ID":"2130e3f2-cbea-46e1-b5d3-daefdd972935","Type":"ContainerStarted","Data":"4339244c7626324d778551bfb8f861a2c704b3e5aa7199c34ba643e475017090"} Oct 14 07:20:23 crc kubenswrapper[4870]: I1014 07:20:23.678426 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-597ff968b-msxbn" event={"ID":"2130e3f2-cbea-46e1-b5d3-daefdd972935","Type":"ContainerStarted","Data":"f4ef48299206032c6fdb947f48913c54fe5fa0a39a9b5ffe9e3ea9fd8ac49f8e"} Oct 14 07:20:23 crc kubenswrapper[4870]: I1014 07:20:23.678473 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-597ff968b-msxbn" event={"ID":"2130e3f2-cbea-46e1-b5d3-daefdd972935","Type":"ContainerStarted","Data":"a137c2c8f449d4fac987a138026d8768098fc2fa5e342a9df21e1f2d3ff32046"} Oct 14 07:20:23 crc kubenswrapper[4870]: I1014 07:20:23.679926 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-597ff968b-msxbn" Oct 14 07:20:23 crc kubenswrapper[4870]: I1014 07:20:23.679974 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-597ff968b-msxbn" Oct 14 07:20:23 crc kubenswrapper[4870]: I1014 07:20:23.683869 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0a31f5df-eae7-4087-a8c8-09c18f5ee9f0","Type":"ContainerStarted","Data":"c9de159d85eb50d1ce07f0bac74b7ca00d3e33977898a2e94dc9152d2951d08b"} Oct 14 07:20:23 crc kubenswrapper[4870]: I1014 07:20:23.711029 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-597ff968b-msxbn" podStartSLOduration=4.710994305 podStartE2EDuration="4.710994305s" podCreationTimestamp="2025-10-14 07:20:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:20:23.69542195 +0000 UTC m=+1159.392782361" watchObservedRunningTime="2025-10-14 07:20:23.710994305 +0000 UTC m=+1159.408354716" Oct 14 07:20:24 crc kubenswrapper[4870]: I1014 07:20:24.007521 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-865c567797-jljrm"] Oct 14 07:20:24 crc kubenswrapper[4870]: E1014 07:20:24.007965 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fe66a2e1-364a-4cb9-b857-c208c7d1b108" containerName="keystone-bootstrap" Oct 14 07:20:24 crc kubenswrapper[4870]: I1014 07:20:24.007985 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="fe66a2e1-364a-4cb9-b857-c208c7d1b108" containerName="keystone-bootstrap" Oct 14 07:20:24 crc kubenswrapper[4870]: I1014 07:20:24.008255 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="fe66a2e1-364a-4cb9-b857-c208c7d1b108" containerName="keystone-bootstrap" Oct 14 07:20:24 crc kubenswrapper[4870]: I1014 07:20:24.008934 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-865c567797-jljrm" Oct 14 07:20:24 crc kubenswrapper[4870]: I1014 07:20:24.012652 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 14 07:20:24 crc kubenswrapper[4870]: I1014 07:20:24.012667 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 14 07:20:24 crc kubenswrapper[4870]: I1014 07:20:24.012805 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 14 07:20:24 crc kubenswrapper[4870]: I1014 07:20:24.012969 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-cqggg" Oct 14 07:20:24 crc kubenswrapper[4870]: I1014 07:20:24.013146 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Oct 14 07:20:24 crc kubenswrapper[4870]: I1014 07:20:24.014294 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Oct 14 07:20:24 crc kubenswrapper[4870]: I1014 07:20:24.019577 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-865c567797-jljrm"] Oct 14 07:20:24 crc kubenswrapper[4870]: I1014 07:20:24.121376 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vhds6\" (UniqueName: \"kubernetes.io/projected/60de9d04-7c24-42a4-9d5e-74b611487b69-kube-api-access-vhds6\") pod \"keystone-865c567797-jljrm\" (UID: \"60de9d04-7c24-42a4-9d5e-74b611487b69\") " pod="openstack/keystone-865c567797-jljrm" Oct 14 07:20:24 crc kubenswrapper[4870]: I1014 07:20:24.121517 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/60de9d04-7c24-42a4-9d5e-74b611487b69-scripts\") pod \"keystone-865c567797-jljrm\" (UID: \"60de9d04-7c24-42a4-9d5e-74b611487b69\") " pod="openstack/keystone-865c567797-jljrm" Oct 14 07:20:24 crc kubenswrapper[4870]: I1014 07:20:24.121566 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/60de9d04-7c24-42a4-9d5e-74b611487b69-internal-tls-certs\") pod \"keystone-865c567797-jljrm\" (UID: \"60de9d04-7c24-42a4-9d5e-74b611487b69\") " pod="openstack/keystone-865c567797-jljrm" Oct 14 07:20:24 crc kubenswrapper[4870]: I1014 07:20:24.122825 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/60de9d04-7c24-42a4-9d5e-74b611487b69-config-data\") pod \"keystone-865c567797-jljrm\" (UID: \"60de9d04-7c24-42a4-9d5e-74b611487b69\") " pod="openstack/keystone-865c567797-jljrm" Oct 14 07:20:24 crc kubenswrapper[4870]: I1014 07:20:24.122888 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/60de9d04-7c24-42a4-9d5e-74b611487b69-credential-keys\") pod \"keystone-865c567797-jljrm\" (UID: \"60de9d04-7c24-42a4-9d5e-74b611487b69\") " pod="openstack/keystone-865c567797-jljrm" Oct 14 07:20:24 crc kubenswrapper[4870]: I1014 07:20:24.123017 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60de9d04-7c24-42a4-9d5e-74b611487b69-combined-ca-bundle\") pod \"keystone-865c567797-jljrm\" (UID: \"60de9d04-7c24-42a4-9d5e-74b611487b69\") " pod="openstack/keystone-865c567797-jljrm" Oct 14 07:20:24 crc kubenswrapper[4870]: I1014 07:20:24.123074 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/60de9d04-7c24-42a4-9d5e-74b611487b69-fernet-keys\") pod \"keystone-865c567797-jljrm\" (UID: \"60de9d04-7c24-42a4-9d5e-74b611487b69\") " pod="openstack/keystone-865c567797-jljrm" Oct 14 07:20:24 crc kubenswrapper[4870]: I1014 07:20:24.123107 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/60de9d04-7c24-42a4-9d5e-74b611487b69-public-tls-certs\") pod \"keystone-865c567797-jljrm\" (UID: \"60de9d04-7c24-42a4-9d5e-74b611487b69\") " pod="openstack/keystone-865c567797-jljrm" Oct 14 07:20:24 crc kubenswrapper[4870]: I1014 07:20:24.224754 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/60de9d04-7c24-42a4-9d5e-74b611487b69-fernet-keys\") pod \"keystone-865c567797-jljrm\" (UID: \"60de9d04-7c24-42a4-9d5e-74b611487b69\") " pod="openstack/keystone-865c567797-jljrm" Oct 14 07:20:24 crc kubenswrapper[4870]: I1014 07:20:24.225110 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/60de9d04-7c24-42a4-9d5e-74b611487b69-public-tls-certs\") pod \"keystone-865c567797-jljrm\" (UID: \"60de9d04-7c24-42a4-9d5e-74b611487b69\") " pod="openstack/keystone-865c567797-jljrm" Oct 14 07:20:24 crc kubenswrapper[4870]: I1014 07:20:24.225156 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vhds6\" (UniqueName: \"kubernetes.io/projected/60de9d04-7c24-42a4-9d5e-74b611487b69-kube-api-access-vhds6\") pod \"keystone-865c567797-jljrm\" (UID: \"60de9d04-7c24-42a4-9d5e-74b611487b69\") " pod="openstack/keystone-865c567797-jljrm" Oct 14 07:20:24 crc kubenswrapper[4870]: I1014 07:20:24.225191 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/60de9d04-7c24-42a4-9d5e-74b611487b69-scripts\") pod \"keystone-865c567797-jljrm\" (UID: \"60de9d04-7c24-42a4-9d5e-74b611487b69\") " pod="openstack/keystone-865c567797-jljrm" Oct 14 07:20:24 crc kubenswrapper[4870]: I1014 07:20:24.225213 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/60de9d04-7c24-42a4-9d5e-74b611487b69-internal-tls-certs\") pod \"keystone-865c567797-jljrm\" (UID: \"60de9d04-7c24-42a4-9d5e-74b611487b69\") " pod="openstack/keystone-865c567797-jljrm" Oct 14 07:20:24 crc kubenswrapper[4870]: I1014 07:20:24.225271 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/60de9d04-7c24-42a4-9d5e-74b611487b69-config-data\") pod \"keystone-865c567797-jljrm\" (UID: \"60de9d04-7c24-42a4-9d5e-74b611487b69\") " pod="openstack/keystone-865c567797-jljrm" Oct 14 07:20:24 crc kubenswrapper[4870]: I1014 07:20:24.225293 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/60de9d04-7c24-42a4-9d5e-74b611487b69-credential-keys\") pod \"keystone-865c567797-jljrm\" (UID: \"60de9d04-7c24-42a4-9d5e-74b611487b69\") " pod="openstack/keystone-865c567797-jljrm" Oct 14 07:20:24 crc kubenswrapper[4870]: I1014 07:20:24.225333 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60de9d04-7c24-42a4-9d5e-74b611487b69-combined-ca-bundle\") pod \"keystone-865c567797-jljrm\" (UID: \"60de9d04-7c24-42a4-9d5e-74b611487b69\") " pod="openstack/keystone-865c567797-jljrm" Oct 14 07:20:24 crc kubenswrapper[4870]: I1014 07:20:24.231026 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/60de9d04-7c24-42a4-9d5e-74b611487b69-credential-keys\") pod \"keystone-865c567797-jljrm\" (UID: \"60de9d04-7c24-42a4-9d5e-74b611487b69\") " pod="openstack/keystone-865c567797-jljrm" Oct 14 07:20:24 crc kubenswrapper[4870]: I1014 07:20:24.231421 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/60de9d04-7c24-42a4-9d5e-74b611487b69-internal-tls-certs\") pod \"keystone-865c567797-jljrm\" (UID: \"60de9d04-7c24-42a4-9d5e-74b611487b69\") " pod="openstack/keystone-865c567797-jljrm" Oct 14 07:20:24 crc kubenswrapper[4870]: I1014 07:20:24.231512 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/60de9d04-7c24-42a4-9d5e-74b611487b69-public-tls-certs\") pod \"keystone-865c567797-jljrm\" (UID: \"60de9d04-7c24-42a4-9d5e-74b611487b69\") " pod="openstack/keystone-865c567797-jljrm" Oct 14 07:20:24 crc kubenswrapper[4870]: I1014 07:20:24.231776 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/60de9d04-7c24-42a4-9d5e-74b611487b69-scripts\") pod \"keystone-865c567797-jljrm\" (UID: \"60de9d04-7c24-42a4-9d5e-74b611487b69\") " pod="openstack/keystone-865c567797-jljrm" Oct 14 07:20:24 crc kubenswrapper[4870]: I1014 07:20:24.236933 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/60de9d04-7c24-42a4-9d5e-74b611487b69-fernet-keys\") pod \"keystone-865c567797-jljrm\" (UID: \"60de9d04-7c24-42a4-9d5e-74b611487b69\") " pod="openstack/keystone-865c567797-jljrm" Oct 14 07:20:24 crc kubenswrapper[4870]: I1014 07:20:24.238067 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/60de9d04-7c24-42a4-9d5e-74b611487b69-config-data\") pod \"keystone-865c567797-jljrm\" (UID: \"60de9d04-7c24-42a4-9d5e-74b611487b69\") " pod="openstack/keystone-865c567797-jljrm" Oct 14 07:20:24 crc kubenswrapper[4870]: I1014 07:20:24.238789 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60de9d04-7c24-42a4-9d5e-74b611487b69-combined-ca-bundle\") pod \"keystone-865c567797-jljrm\" (UID: \"60de9d04-7c24-42a4-9d5e-74b611487b69\") " pod="openstack/keystone-865c567797-jljrm" Oct 14 07:20:24 crc kubenswrapper[4870]: I1014 07:20:24.252916 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vhds6\" (UniqueName: \"kubernetes.io/projected/60de9d04-7c24-42a4-9d5e-74b611487b69-kube-api-access-vhds6\") pod \"keystone-865c567797-jljrm\" (UID: \"60de9d04-7c24-42a4-9d5e-74b611487b69\") " pod="openstack/keystone-865c567797-jljrm" Oct 14 07:20:24 crc kubenswrapper[4870]: I1014 07:20:24.356853 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-865c567797-jljrm" Oct 14 07:20:24 crc kubenswrapper[4870]: I1014 07:20:24.987291 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 14 07:20:24 crc kubenswrapper[4870]: I1014 07:20:24.987362 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 14 07:20:25 crc kubenswrapper[4870]: I1014 07:20:25.055934 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 14 07:20:25 crc kubenswrapper[4870]: I1014 07:20:25.056102 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 14 07:20:25 crc kubenswrapper[4870]: I1014 07:20:25.151021 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 14 07:20:25 crc kubenswrapper[4870]: I1014 07:20:25.152256 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 14 07:20:25 crc kubenswrapper[4870]: I1014 07:20:25.193769 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 14 07:20:25 crc kubenswrapper[4870]: I1014 07:20:25.194919 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 14 07:20:25 crc kubenswrapper[4870]: I1014 07:20:25.700752 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 14 07:20:25 crc kubenswrapper[4870]: I1014 07:20:25.701023 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 14 07:20:25 crc kubenswrapper[4870]: I1014 07:20:25.701090 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 14 07:20:25 crc kubenswrapper[4870]: I1014 07:20:25.701107 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 14 07:20:27 crc kubenswrapper[4870]: I1014 07:20:27.571407 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-865c567797-jljrm"] Oct 14 07:20:27 crc kubenswrapper[4870]: I1014 07:20:27.575175 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 14 07:20:27 crc kubenswrapper[4870]: I1014 07:20:27.575201 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 14 07:20:27 crc kubenswrapper[4870]: I1014 07:20:27.731296 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-97xhm" event={"ID":"6e87f821-3bc7-4b88-9c72-ccd80b4b19e9","Type":"ContainerStarted","Data":"afdfea867b44e3792f20e1c1ecf3ae23b03aa598321ac8c6d8a8225512ee0c5a"} Oct 14 07:20:27 crc kubenswrapper[4870]: I1014 07:20:27.734526 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-865c567797-jljrm" event={"ID":"60de9d04-7c24-42a4-9d5e-74b611487b69","Type":"ContainerStarted","Data":"bb99a031932012bc36d4d56aba1c27bfd8482f6a5153da2e1843ed879054a61a"} Oct 14 07:20:27 crc kubenswrapper[4870]: I1014 07:20:27.757418 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-97xhm" podStartSLOduration=2.047045739 podStartE2EDuration="9.757395354s" podCreationTimestamp="2025-10-14 07:20:18 +0000 UTC" firstStartedPulling="2025-10-14 07:20:19.479058425 +0000 UTC m=+1155.176418806" lastFinishedPulling="2025-10-14 07:20:27.18940805 +0000 UTC m=+1162.886768421" observedRunningTime="2025-10-14 07:20:27.749946439 +0000 UTC m=+1163.447306810" watchObservedRunningTime="2025-10-14 07:20:27.757395354 +0000 UTC m=+1163.454755725" Oct 14 07:20:27 crc kubenswrapper[4870]: I1014 07:20:27.764549 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 14 07:20:27 crc kubenswrapper[4870]: I1014 07:20:27.765137 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 14 07:20:30 crc kubenswrapper[4870]: I1014 07:20:30.778202 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-865c567797-jljrm" event={"ID":"60de9d04-7c24-42a4-9d5e-74b611487b69","Type":"ContainerStarted","Data":"d8615c16655ea7e7d454161d16b61973188b63bffc966ed4b3805db817746641"} Oct 14 07:20:31 crc kubenswrapper[4870]: I1014 07:20:31.785263 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-865c567797-jljrm" Oct 14 07:20:31 crc kubenswrapper[4870]: I1014 07:20:31.815249 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-865c567797-jljrm" podStartSLOduration=8.815228502 podStartE2EDuration="8.815228502s" podCreationTimestamp="2025-10-14 07:20:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:20:31.807813314 +0000 UTC m=+1167.505173685" watchObservedRunningTime="2025-10-14 07:20:31.815228502 +0000 UTC m=+1167.512588873" Oct 14 07:20:33 crc kubenswrapper[4870]: I1014 07:20:33.807614 4870 generic.go:334] "Generic (PLEG): container finished" podID="6e87f821-3bc7-4b88-9c72-ccd80b4b19e9" containerID="afdfea867b44e3792f20e1c1ecf3ae23b03aa598321ac8c6d8a8225512ee0c5a" exitCode=0 Oct 14 07:20:33 crc kubenswrapper[4870]: I1014 07:20:33.807764 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-97xhm" event={"ID":"6e87f821-3bc7-4b88-9c72-ccd80b4b19e9","Type":"ContainerDied","Data":"afdfea867b44e3792f20e1c1ecf3ae23b03aa598321ac8c6d8a8225512ee0c5a"} Oct 14 07:20:38 crc kubenswrapper[4870]: I1014 07:20:38.862530 4870 generic.go:334] "Generic (PLEG): container finished" podID="685539fb-44a2-4f75-b418-65383d02152e" containerID="28a8e8e086b6d6c6e4aa03f701ff3d36e5350745361841619c40f9371ec5f443" exitCode=0 Oct 14 07:20:38 crc kubenswrapper[4870]: I1014 07:20:38.862576 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-vtmvg" event={"ID":"685539fb-44a2-4f75-b418-65383d02152e","Type":"ContainerDied","Data":"28a8e8e086b6d6c6e4aa03f701ff3d36e5350745361841619c40f9371ec5f443"} Oct 14 07:20:42 crc kubenswrapper[4870]: I1014 07:20:42.100694 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-vtmvg" Oct 14 07:20:42 crc kubenswrapper[4870]: I1014 07:20:42.109483 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-97xhm" Oct 14 07:20:42 crc kubenswrapper[4870]: I1014 07:20:42.185535 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/685539fb-44a2-4f75-b418-65383d02152e-combined-ca-bundle\") pod \"685539fb-44a2-4f75-b418-65383d02152e\" (UID: \"685539fb-44a2-4f75-b418-65383d02152e\") " Oct 14 07:20:42 crc kubenswrapper[4870]: I1014 07:20:42.185792 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mfv7v\" (UniqueName: \"kubernetes.io/projected/6e87f821-3bc7-4b88-9c72-ccd80b4b19e9-kube-api-access-mfv7v\") pod \"6e87f821-3bc7-4b88-9c72-ccd80b4b19e9\" (UID: \"6e87f821-3bc7-4b88-9c72-ccd80b4b19e9\") " Oct 14 07:20:42 crc kubenswrapper[4870]: I1014 07:20:42.185892 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e87f821-3bc7-4b88-9c72-ccd80b4b19e9-combined-ca-bundle\") pod \"6e87f821-3bc7-4b88-9c72-ccd80b4b19e9\" (UID: \"6e87f821-3bc7-4b88-9c72-ccd80b4b19e9\") " Oct 14 07:20:42 crc kubenswrapper[4870]: I1014 07:20:42.186035 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7vrfg\" (UniqueName: \"kubernetes.io/projected/685539fb-44a2-4f75-b418-65383d02152e-kube-api-access-7vrfg\") pod \"685539fb-44a2-4f75-b418-65383d02152e\" (UID: \"685539fb-44a2-4f75-b418-65383d02152e\") " Oct 14 07:20:42 crc kubenswrapper[4870]: I1014 07:20:42.186127 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/685539fb-44a2-4f75-b418-65383d02152e-config\") pod \"685539fb-44a2-4f75-b418-65383d02152e\" (UID: \"685539fb-44a2-4f75-b418-65383d02152e\") " Oct 14 07:20:42 crc kubenswrapper[4870]: I1014 07:20:42.186254 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/6e87f821-3bc7-4b88-9c72-ccd80b4b19e9-db-sync-config-data\") pod \"6e87f821-3bc7-4b88-9c72-ccd80b4b19e9\" (UID: \"6e87f821-3bc7-4b88-9c72-ccd80b4b19e9\") " Oct 14 07:20:42 crc kubenswrapper[4870]: I1014 07:20:42.192932 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6e87f821-3bc7-4b88-9c72-ccd80b4b19e9-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "6e87f821-3bc7-4b88-9c72-ccd80b4b19e9" (UID: "6e87f821-3bc7-4b88-9c72-ccd80b4b19e9"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:20:42 crc kubenswrapper[4870]: I1014 07:20:42.192976 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6e87f821-3bc7-4b88-9c72-ccd80b4b19e9-kube-api-access-mfv7v" (OuterVolumeSpecName: "kube-api-access-mfv7v") pod "6e87f821-3bc7-4b88-9c72-ccd80b4b19e9" (UID: "6e87f821-3bc7-4b88-9c72-ccd80b4b19e9"). InnerVolumeSpecName "kube-api-access-mfv7v". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:20:42 crc kubenswrapper[4870]: I1014 07:20:42.193821 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/685539fb-44a2-4f75-b418-65383d02152e-kube-api-access-7vrfg" (OuterVolumeSpecName: "kube-api-access-7vrfg") pod "685539fb-44a2-4f75-b418-65383d02152e" (UID: "685539fb-44a2-4f75-b418-65383d02152e"). InnerVolumeSpecName "kube-api-access-7vrfg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:20:42 crc kubenswrapper[4870]: I1014 07:20:42.222308 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6e87f821-3bc7-4b88-9c72-ccd80b4b19e9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6e87f821-3bc7-4b88-9c72-ccd80b4b19e9" (UID: "6e87f821-3bc7-4b88-9c72-ccd80b4b19e9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:20:42 crc kubenswrapper[4870]: I1014 07:20:42.241901 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/685539fb-44a2-4f75-b418-65383d02152e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "685539fb-44a2-4f75-b418-65383d02152e" (UID: "685539fb-44a2-4f75-b418-65383d02152e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:20:42 crc kubenswrapper[4870]: I1014 07:20:42.249669 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/685539fb-44a2-4f75-b418-65383d02152e-config" (OuterVolumeSpecName: "config") pod "685539fb-44a2-4f75-b418-65383d02152e" (UID: "685539fb-44a2-4f75-b418-65383d02152e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:20:42 crc kubenswrapper[4870]: I1014 07:20:42.288189 4870 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/685539fb-44a2-4f75-b418-65383d02152e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 07:20:42 crc kubenswrapper[4870]: I1014 07:20:42.288236 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mfv7v\" (UniqueName: \"kubernetes.io/projected/6e87f821-3bc7-4b88-9c72-ccd80b4b19e9-kube-api-access-mfv7v\") on node \"crc\" DevicePath \"\"" Oct 14 07:20:42 crc kubenswrapper[4870]: I1014 07:20:42.288253 4870 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e87f821-3bc7-4b88-9c72-ccd80b4b19e9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 07:20:42 crc kubenswrapper[4870]: I1014 07:20:42.288265 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7vrfg\" (UniqueName: \"kubernetes.io/projected/685539fb-44a2-4f75-b418-65383d02152e-kube-api-access-7vrfg\") on node \"crc\" DevicePath \"\"" Oct 14 07:20:42 crc kubenswrapper[4870]: I1014 07:20:42.288279 4870 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/685539fb-44a2-4f75-b418-65383d02152e-config\") on node \"crc\" DevicePath \"\"" Oct 14 07:20:42 crc kubenswrapper[4870]: I1014 07:20:42.288291 4870 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/6e87f821-3bc7-4b88-9c72-ccd80b4b19e9-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 07:20:42 crc kubenswrapper[4870]: I1014 07:20:42.896924 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-vtmvg" Oct 14 07:20:42 crc kubenswrapper[4870]: I1014 07:20:42.896929 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-vtmvg" event={"ID":"685539fb-44a2-4f75-b418-65383d02152e","Type":"ContainerDied","Data":"5af8a8a8e81f5a73f71a578bdc2386e1bf1644cbcca6656c79c291aff9014c4e"} Oct 14 07:20:42 crc kubenswrapper[4870]: I1014 07:20:42.897012 4870 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5af8a8a8e81f5a73f71a578bdc2386e1bf1644cbcca6656c79c291aff9014c4e" Oct 14 07:20:42 crc kubenswrapper[4870]: I1014 07:20:42.898416 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-97xhm" event={"ID":"6e87f821-3bc7-4b88-9c72-ccd80b4b19e9","Type":"ContainerDied","Data":"4e29b89487e5efa0447752c0a7feaef2666e899bf747ab5311fe6af99d9aa374"} Oct 14 07:20:42 crc kubenswrapper[4870]: I1014 07:20:42.898530 4870 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4e29b89487e5efa0447752c0a7feaef2666e899bf747ab5311fe6af99d9aa374" Oct 14 07:20:42 crc kubenswrapper[4870]: I1014 07:20:42.898591 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-97xhm" Oct 14 07:20:43 crc kubenswrapper[4870]: E1014 07:20:43.239981 4870 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api@sha256:93b475af15a0d10e95cb17b98927077f05ac24c89472a601d677eb89f82fd429" Oct 14 07:20:43 crc kubenswrapper[4870]: E1014 07:20:43.240148 4870 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api@sha256:93b475af15a0d10e95cb17b98927077f05ac24c89472a601d677eb89f82fd429,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-56zkt,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-h587c_openstack(000463bc-f3ac-40b7-90c8-32366e327a19): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 14 07:20:43 crc kubenswrapper[4870]: E1014 07:20:43.241537 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-h587c" podUID="000463bc-f3ac-40b7-90c8-32366e327a19" Oct 14 07:20:43 crc kubenswrapper[4870]: I1014 07:20:43.322955 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7fdcbfbb99-5pnss"] Oct 14 07:20:43 crc kubenswrapper[4870]: E1014 07:20:43.323368 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="685539fb-44a2-4f75-b418-65383d02152e" containerName="neutron-db-sync" Oct 14 07:20:43 crc kubenswrapper[4870]: I1014 07:20:43.323388 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="685539fb-44a2-4f75-b418-65383d02152e" containerName="neutron-db-sync" Oct 14 07:20:43 crc kubenswrapper[4870]: E1014 07:20:43.323424 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e87f821-3bc7-4b88-9c72-ccd80b4b19e9" containerName="barbican-db-sync" Oct 14 07:20:43 crc kubenswrapper[4870]: I1014 07:20:43.323448 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e87f821-3bc7-4b88-9c72-ccd80b4b19e9" containerName="barbican-db-sync" Oct 14 07:20:43 crc kubenswrapper[4870]: I1014 07:20:43.323652 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="685539fb-44a2-4f75-b418-65383d02152e" containerName="neutron-db-sync" Oct 14 07:20:43 crc kubenswrapper[4870]: I1014 07:20:43.323676 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="6e87f821-3bc7-4b88-9c72-ccd80b4b19e9" containerName="barbican-db-sync" Oct 14 07:20:43 crc kubenswrapper[4870]: I1014 07:20:43.325156 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fdcbfbb99-5pnss" Oct 14 07:20:43 crc kubenswrapper[4870]: I1014 07:20:43.366971 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7fdcbfbb99-5pnss"] Oct 14 07:20:43 crc kubenswrapper[4870]: I1014 07:20:43.422741 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/47ef233a-2847-4bcc-88cb-2be6e2f13e0e-dns-swift-storage-0\") pod \"dnsmasq-dns-7fdcbfbb99-5pnss\" (UID: \"47ef233a-2847-4bcc-88cb-2be6e2f13e0e\") " pod="openstack/dnsmasq-dns-7fdcbfbb99-5pnss" Oct 14 07:20:43 crc kubenswrapper[4870]: I1014 07:20:43.422822 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-57gqf\" (UniqueName: \"kubernetes.io/projected/47ef233a-2847-4bcc-88cb-2be6e2f13e0e-kube-api-access-57gqf\") pod \"dnsmasq-dns-7fdcbfbb99-5pnss\" (UID: \"47ef233a-2847-4bcc-88cb-2be6e2f13e0e\") " pod="openstack/dnsmasq-dns-7fdcbfbb99-5pnss" Oct 14 07:20:43 crc kubenswrapper[4870]: I1014 07:20:43.422861 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/47ef233a-2847-4bcc-88cb-2be6e2f13e0e-ovsdbserver-nb\") pod \"dnsmasq-dns-7fdcbfbb99-5pnss\" (UID: \"47ef233a-2847-4bcc-88cb-2be6e2f13e0e\") " pod="openstack/dnsmasq-dns-7fdcbfbb99-5pnss" Oct 14 07:20:43 crc kubenswrapper[4870]: I1014 07:20:43.422916 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/47ef233a-2847-4bcc-88cb-2be6e2f13e0e-ovsdbserver-sb\") pod \"dnsmasq-dns-7fdcbfbb99-5pnss\" (UID: \"47ef233a-2847-4bcc-88cb-2be6e2f13e0e\") " pod="openstack/dnsmasq-dns-7fdcbfbb99-5pnss" Oct 14 07:20:43 crc kubenswrapper[4870]: I1014 07:20:43.422980 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/47ef233a-2847-4bcc-88cb-2be6e2f13e0e-config\") pod \"dnsmasq-dns-7fdcbfbb99-5pnss\" (UID: \"47ef233a-2847-4bcc-88cb-2be6e2f13e0e\") " pod="openstack/dnsmasq-dns-7fdcbfbb99-5pnss" Oct 14 07:20:43 crc kubenswrapper[4870]: I1014 07:20:43.423008 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/47ef233a-2847-4bcc-88cb-2be6e2f13e0e-dns-svc\") pod \"dnsmasq-dns-7fdcbfbb99-5pnss\" (UID: \"47ef233a-2847-4bcc-88cb-2be6e2f13e0e\") " pod="openstack/dnsmasq-dns-7fdcbfbb99-5pnss" Oct 14 07:20:43 crc kubenswrapper[4870]: I1014 07:20:43.428815 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-85f775bc47-d95qn"] Oct 14 07:20:43 crc kubenswrapper[4870]: I1014 07:20:43.430264 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-85f775bc47-d95qn" Oct 14 07:20:43 crc kubenswrapper[4870]: I1014 07:20:43.436701 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Oct 14 07:20:43 crc kubenswrapper[4870]: I1014 07:20:43.437145 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-2s5xt" Oct 14 07:20:43 crc kubenswrapper[4870]: I1014 07:20:43.437273 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Oct 14 07:20:43 crc kubenswrapper[4870]: I1014 07:20:43.447220 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-85f775bc47-d95qn"] Oct 14 07:20:43 crc kubenswrapper[4870]: I1014 07:20:43.468190 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-57d8bd87b6-vnpct"] Oct 14 07:20:43 crc kubenswrapper[4870]: I1014 07:20:43.469894 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-57d8bd87b6-vnpct" Oct 14 07:20:43 crc kubenswrapper[4870]: I1014 07:20:43.474099 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Oct 14 07:20:43 crc kubenswrapper[4870]: I1014 07:20:43.478550 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-57d8bd87b6-vnpct"] Oct 14 07:20:43 crc kubenswrapper[4870]: I1014 07:20:43.528457 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5bb3fde-cfaf-4d09-ad31-e23fd830eb91-combined-ca-bundle\") pod \"barbican-worker-85f775bc47-d95qn\" (UID: \"c5bb3fde-cfaf-4d09-ad31-e23fd830eb91\") " pod="openstack/barbican-worker-85f775bc47-d95qn" Oct 14 07:20:43 crc kubenswrapper[4870]: I1014 07:20:43.528768 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/47ef233a-2847-4bcc-88cb-2be6e2f13e0e-dns-swift-storage-0\") pod \"dnsmasq-dns-7fdcbfbb99-5pnss\" (UID: \"47ef233a-2847-4bcc-88cb-2be6e2f13e0e\") " pod="openstack/dnsmasq-dns-7fdcbfbb99-5pnss" Oct 14 07:20:43 crc kubenswrapper[4870]: I1014 07:20:43.528919 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-57gqf\" (UniqueName: \"kubernetes.io/projected/47ef233a-2847-4bcc-88cb-2be6e2f13e0e-kube-api-access-57gqf\") pod \"dnsmasq-dns-7fdcbfbb99-5pnss\" (UID: \"47ef233a-2847-4bcc-88cb-2be6e2f13e0e\") " pod="openstack/dnsmasq-dns-7fdcbfbb99-5pnss" Oct 14 07:20:43 crc kubenswrapper[4870]: I1014 07:20:43.529066 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/47ef233a-2847-4bcc-88cb-2be6e2f13e0e-ovsdbserver-nb\") pod \"dnsmasq-dns-7fdcbfbb99-5pnss\" (UID: \"47ef233a-2847-4bcc-88cb-2be6e2f13e0e\") " pod="openstack/dnsmasq-dns-7fdcbfbb99-5pnss" Oct 14 07:20:43 crc kubenswrapper[4870]: I1014 07:20:43.529161 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c5bb3fde-cfaf-4d09-ad31-e23fd830eb91-config-data\") pod \"barbican-worker-85f775bc47-d95qn\" (UID: \"c5bb3fde-cfaf-4d09-ad31-e23fd830eb91\") " pod="openstack/barbican-worker-85f775bc47-d95qn" Oct 14 07:20:43 crc kubenswrapper[4870]: I1014 07:20:43.529295 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/47ef233a-2847-4bcc-88cb-2be6e2f13e0e-ovsdbserver-sb\") pod \"dnsmasq-dns-7fdcbfbb99-5pnss\" (UID: \"47ef233a-2847-4bcc-88cb-2be6e2f13e0e\") " pod="openstack/dnsmasq-dns-7fdcbfbb99-5pnss" Oct 14 07:20:43 crc kubenswrapper[4870]: I1014 07:20:43.529465 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zxtlp\" (UniqueName: \"kubernetes.io/projected/c5bb3fde-cfaf-4d09-ad31-e23fd830eb91-kube-api-access-zxtlp\") pod \"barbican-worker-85f775bc47-d95qn\" (UID: \"c5bb3fde-cfaf-4d09-ad31-e23fd830eb91\") " pod="openstack/barbican-worker-85f775bc47-d95qn" Oct 14 07:20:43 crc kubenswrapper[4870]: I1014 07:20:43.529556 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/47ef233a-2847-4bcc-88cb-2be6e2f13e0e-config\") pod \"dnsmasq-dns-7fdcbfbb99-5pnss\" (UID: \"47ef233a-2847-4bcc-88cb-2be6e2f13e0e\") " pod="openstack/dnsmasq-dns-7fdcbfbb99-5pnss" Oct 14 07:20:43 crc kubenswrapper[4870]: I1014 07:20:43.530896 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/47ef233a-2847-4bcc-88cb-2be6e2f13e0e-ovsdbserver-sb\") pod \"dnsmasq-dns-7fdcbfbb99-5pnss\" (UID: \"47ef233a-2847-4bcc-88cb-2be6e2f13e0e\") " pod="openstack/dnsmasq-dns-7fdcbfbb99-5pnss" Oct 14 07:20:43 crc kubenswrapper[4870]: I1014 07:20:43.530990 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/47ef233a-2847-4bcc-88cb-2be6e2f13e0e-dns-svc\") pod \"dnsmasq-dns-7fdcbfbb99-5pnss\" (UID: \"47ef233a-2847-4bcc-88cb-2be6e2f13e0e\") " pod="openstack/dnsmasq-dns-7fdcbfbb99-5pnss" Oct 14 07:20:43 crc kubenswrapper[4870]: I1014 07:20:43.531063 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c5bb3fde-cfaf-4d09-ad31-e23fd830eb91-logs\") pod \"barbican-worker-85f775bc47-d95qn\" (UID: \"c5bb3fde-cfaf-4d09-ad31-e23fd830eb91\") " pod="openstack/barbican-worker-85f775bc47-d95qn" Oct 14 07:20:43 crc kubenswrapper[4870]: I1014 07:20:43.531100 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c5bb3fde-cfaf-4d09-ad31-e23fd830eb91-config-data-custom\") pod \"barbican-worker-85f775bc47-d95qn\" (UID: \"c5bb3fde-cfaf-4d09-ad31-e23fd830eb91\") " pod="openstack/barbican-worker-85f775bc47-d95qn" Oct 14 07:20:43 crc kubenswrapper[4870]: I1014 07:20:43.531809 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/47ef233a-2847-4bcc-88cb-2be6e2f13e0e-ovsdbserver-nb\") pod \"dnsmasq-dns-7fdcbfbb99-5pnss\" (UID: \"47ef233a-2847-4bcc-88cb-2be6e2f13e0e\") " pod="openstack/dnsmasq-dns-7fdcbfbb99-5pnss" Oct 14 07:20:43 crc kubenswrapper[4870]: I1014 07:20:43.532009 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/47ef233a-2847-4bcc-88cb-2be6e2f13e0e-dns-svc\") pod \"dnsmasq-dns-7fdcbfbb99-5pnss\" (UID: \"47ef233a-2847-4bcc-88cb-2be6e2f13e0e\") " pod="openstack/dnsmasq-dns-7fdcbfbb99-5pnss" Oct 14 07:20:43 crc kubenswrapper[4870]: I1014 07:20:43.533046 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/47ef233a-2847-4bcc-88cb-2be6e2f13e0e-config\") pod \"dnsmasq-dns-7fdcbfbb99-5pnss\" (UID: \"47ef233a-2847-4bcc-88cb-2be6e2f13e0e\") " pod="openstack/dnsmasq-dns-7fdcbfbb99-5pnss" Oct 14 07:20:43 crc kubenswrapper[4870]: I1014 07:20:43.538037 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/47ef233a-2847-4bcc-88cb-2be6e2f13e0e-dns-swift-storage-0\") pod \"dnsmasq-dns-7fdcbfbb99-5pnss\" (UID: \"47ef233a-2847-4bcc-88cb-2be6e2f13e0e\") " pod="openstack/dnsmasq-dns-7fdcbfbb99-5pnss" Oct 14 07:20:43 crc kubenswrapper[4870]: I1014 07:20:43.565181 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-57gqf\" (UniqueName: \"kubernetes.io/projected/47ef233a-2847-4bcc-88cb-2be6e2f13e0e-kube-api-access-57gqf\") pod \"dnsmasq-dns-7fdcbfbb99-5pnss\" (UID: \"47ef233a-2847-4bcc-88cb-2be6e2f13e0e\") " pod="openstack/dnsmasq-dns-7fdcbfbb99-5pnss" Oct 14 07:20:43 crc kubenswrapper[4870]: I1014 07:20:43.598102 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7fdcbfbb99-5pnss"] Oct 14 07:20:43 crc kubenswrapper[4870]: I1014 07:20:43.599055 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fdcbfbb99-5pnss" Oct 14 07:20:43 crc kubenswrapper[4870]: I1014 07:20:43.620578 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-5dd746bf8d-mvg6f"] Oct 14 07:20:43 crc kubenswrapper[4870]: I1014 07:20:43.622257 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5dd746bf8d-mvg6f" Oct 14 07:20:43 crc kubenswrapper[4870]: I1014 07:20:43.630319 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Oct 14 07:20:43 crc kubenswrapper[4870]: I1014 07:20:43.630707 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-5whvl" Oct 14 07:20:43 crc kubenswrapper[4870]: I1014 07:20:43.630866 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Oct 14 07:20:43 crc kubenswrapper[4870]: I1014 07:20:43.631154 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Oct 14 07:20:43 crc kubenswrapper[4870]: I1014 07:20:43.637142 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0a2e4c80-d44f-40c0-9c04-74e9ce98ae36-config-data-custom\") pod \"barbican-keystone-listener-57d8bd87b6-vnpct\" (UID: \"0a2e4c80-d44f-40c0-9c04-74e9ce98ae36\") " pod="openstack/barbican-keystone-listener-57d8bd87b6-vnpct" Oct 14 07:20:43 crc kubenswrapper[4870]: I1014 07:20:43.637215 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zxtlp\" (UniqueName: \"kubernetes.io/projected/c5bb3fde-cfaf-4d09-ad31-e23fd830eb91-kube-api-access-zxtlp\") pod \"barbican-worker-85f775bc47-d95qn\" (UID: \"c5bb3fde-cfaf-4d09-ad31-e23fd830eb91\") " pod="openstack/barbican-worker-85f775bc47-d95qn" Oct 14 07:20:43 crc kubenswrapper[4870]: I1014 07:20:43.637263 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c5bb3fde-cfaf-4d09-ad31-e23fd830eb91-logs\") pod \"barbican-worker-85f775bc47-d95qn\" (UID: \"c5bb3fde-cfaf-4d09-ad31-e23fd830eb91\") " pod="openstack/barbican-worker-85f775bc47-d95qn" Oct 14 07:20:43 crc kubenswrapper[4870]: I1014 07:20:43.637281 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c5bb3fde-cfaf-4d09-ad31-e23fd830eb91-config-data-custom\") pod \"barbican-worker-85f775bc47-d95qn\" (UID: \"c5bb3fde-cfaf-4d09-ad31-e23fd830eb91\") " pod="openstack/barbican-worker-85f775bc47-d95qn" Oct 14 07:20:43 crc kubenswrapper[4870]: I1014 07:20:43.637309 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a2e4c80-d44f-40c0-9c04-74e9ce98ae36-config-data\") pod \"barbican-keystone-listener-57d8bd87b6-vnpct\" (UID: \"0a2e4c80-d44f-40c0-9c04-74e9ce98ae36\") " pod="openstack/barbican-keystone-listener-57d8bd87b6-vnpct" Oct 14 07:20:43 crc kubenswrapper[4870]: I1014 07:20:43.637327 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5bb3fde-cfaf-4d09-ad31-e23fd830eb91-combined-ca-bundle\") pod \"barbican-worker-85f775bc47-d95qn\" (UID: \"c5bb3fde-cfaf-4d09-ad31-e23fd830eb91\") " pod="openstack/barbican-worker-85f775bc47-d95qn" Oct 14 07:20:43 crc kubenswrapper[4870]: I1014 07:20:43.637345 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a2e4c80-d44f-40c0-9c04-74e9ce98ae36-combined-ca-bundle\") pod \"barbican-keystone-listener-57d8bd87b6-vnpct\" (UID: \"0a2e4c80-d44f-40c0-9c04-74e9ce98ae36\") " pod="openstack/barbican-keystone-listener-57d8bd87b6-vnpct" Oct 14 07:20:43 crc kubenswrapper[4870]: I1014 07:20:43.637362 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0a2e4c80-d44f-40c0-9c04-74e9ce98ae36-logs\") pod \"barbican-keystone-listener-57d8bd87b6-vnpct\" (UID: \"0a2e4c80-d44f-40c0-9c04-74e9ce98ae36\") " pod="openstack/barbican-keystone-listener-57d8bd87b6-vnpct" Oct 14 07:20:43 crc kubenswrapper[4870]: I1014 07:20:43.637403 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hbl6f\" (UniqueName: \"kubernetes.io/projected/0a2e4c80-d44f-40c0-9c04-74e9ce98ae36-kube-api-access-hbl6f\") pod \"barbican-keystone-listener-57d8bd87b6-vnpct\" (UID: \"0a2e4c80-d44f-40c0-9c04-74e9ce98ae36\") " pod="openstack/barbican-keystone-listener-57d8bd87b6-vnpct" Oct 14 07:20:43 crc kubenswrapper[4870]: I1014 07:20:43.637457 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c5bb3fde-cfaf-4d09-ad31-e23fd830eb91-config-data\") pod \"barbican-worker-85f775bc47-d95qn\" (UID: \"c5bb3fde-cfaf-4d09-ad31-e23fd830eb91\") " pod="openstack/barbican-worker-85f775bc47-d95qn" Oct 14 07:20:43 crc kubenswrapper[4870]: I1014 07:20:43.644739 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-5dd746bf8d-mvg6f"] Oct 14 07:20:43 crc kubenswrapper[4870]: I1014 07:20:43.651642 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c5bb3fde-cfaf-4d09-ad31-e23fd830eb91-logs\") pod \"barbican-worker-85f775bc47-d95qn\" (UID: \"c5bb3fde-cfaf-4d09-ad31-e23fd830eb91\") " pod="openstack/barbican-worker-85f775bc47-d95qn" Oct 14 07:20:43 crc kubenswrapper[4870]: I1014 07:20:43.658193 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c5bb3fde-cfaf-4d09-ad31-e23fd830eb91-config-data-custom\") pod \"barbican-worker-85f775bc47-d95qn\" (UID: \"c5bb3fde-cfaf-4d09-ad31-e23fd830eb91\") " pod="openstack/barbican-worker-85f775bc47-d95qn" Oct 14 07:20:43 crc kubenswrapper[4870]: I1014 07:20:43.660518 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-85fd94d64c-nr8rq"] Oct 14 07:20:43 crc kubenswrapper[4870]: I1014 07:20:43.661015 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c5bb3fde-cfaf-4d09-ad31-e23fd830eb91-config-data\") pod \"barbican-worker-85f775bc47-d95qn\" (UID: \"c5bb3fde-cfaf-4d09-ad31-e23fd830eb91\") " pod="openstack/barbican-worker-85f775bc47-d95qn" Oct 14 07:20:43 crc kubenswrapper[4870]: I1014 07:20:43.662364 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85fd94d64c-nr8rq" Oct 14 07:20:43 crc kubenswrapper[4870]: I1014 07:20:43.669293 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-85fd94d64c-nr8rq"] Oct 14 07:20:43 crc kubenswrapper[4870]: I1014 07:20:43.675989 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-8446b89fb-22zh4"] Oct 14 07:20:43 crc kubenswrapper[4870]: I1014 07:20:43.678506 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-8446b89fb-22zh4" Oct 14 07:20:43 crc kubenswrapper[4870]: I1014 07:20:43.678805 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5bb3fde-cfaf-4d09-ad31-e23fd830eb91-combined-ca-bundle\") pod \"barbican-worker-85f775bc47-d95qn\" (UID: \"c5bb3fde-cfaf-4d09-ad31-e23fd830eb91\") " pod="openstack/barbican-worker-85f775bc47-d95qn" Oct 14 07:20:43 crc kubenswrapper[4870]: I1014 07:20:43.684923 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Oct 14 07:20:43 crc kubenswrapper[4870]: I1014 07:20:43.685270 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zxtlp\" (UniqueName: \"kubernetes.io/projected/c5bb3fde-cfaf-4d09-ad31-e23fd830eb91-kube-api-access-zxtlp\") pod \"barbican-worker-85f775bc47-d95qn\" (UID: \"c5bb3fde-cfaf-4d09-ad31-e23fd830eb91\") " pod="openstack/barbican-worker-85f775bc47-d95qn" Oct 14 07:20:43 crc kubenswrapper[4870]: I1014 07:20:43.685570 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-8446b89fb-22zh4"] Oct 14 07:20:43 crc kubenswrapper[4870]: I1014 07:20:43.738827 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d42ee9f9-390b-437c-a48a-ff7ead370f61-dns-svc\") pod \"dnsmasq-dns-85fd94d64c-nr8rq\" (UID: \"d42ee9f9-390b-437c-a48a-ff7ead370f61\") " pod="openstack/dnsmasq-dns-85fd94d64c-nr8rq" Oct 14 07:20:43 crc kubenswrapper[4870]: I1014 07:20:43.739113 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a2e4c80-d44f-40c0-9c04-74e9ce98ae36-config-data\") pod \"barbican-keystone-listener-57d8bd87b6-vnpct\" (UID: \"0a2e4c80-d44f-40c0-9c04-74e9ce98ae36\") " pod="openstack/barbican-keystone-listener-57d8bd87b6-vnpct" Oct 14 07:20:43 crc kubenswrapper[4870]: I1014 07:20:43.739136 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a2e4c80-d44f-40c0-9c04-74e9ce98ae36-combined-ca-bundle\") pod \"barbican-keystone-listener-57d8bd87b6-vnpct\" (UID: \"0a2e4c80-d44f-40c0-9c04-74e9ce98ae36\") " pod="openstack/barbican-keystone-listener-57d8bd87b6-vnpct" Oct 14 07:20:43 crc kubenswrapper[4870]: I1014 07:20:43.739156 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0a2e4c80-d44f-40c0-9c04-74e9ce98ae36-logs\") pod \"barbican-keystone-listener-57d8bd87b6-vnpct\" (UID: \"0a2e4c80-d44f-40c0-9c04-74e9ce98ae36\") " pod="openstack/barbican-keystone-listener-57d8bd87b6-vnpct" Oct 14 07:20:43 crc kubenswrapper[4870]: I1014 07:20:43.739202 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hbl6f\" (UniqueName: \"kubernetes.io/projected/0a2e4c80-d44f-40c0-9c04-74e9ce98ae36-kube-api-access-hbl6f\") pod \"barbican-keystone-listener-57d8bd87b6-vnpct\" (UID: \"0a2e4c80-d44f-40c0-9c04-74e9ce98ae36\") " pod="openstack/barbican-keystone-listener-57d8bd87b6-vnpct" Oct 14 07:20:43 crc kubenswrapper[4870]: I1014 07:20:43.739231 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gmddt\" (UniqueName: \"kubernetes.io/projected/8962df13-02a8-4049-bf25-767413eaa98f-kube-api-access-gmddt\") pod \"neutron-5dd746bf8d-mvg6f\" (UID: \"8962df13-02a8-4049-bf25-767413eaa98f\") " pod="openstack/neutron-5dd746bf8d-mvg6f" Oct 14 07:20:43 crc kubenswrapper[4870]: I1014 07:20:43.739249 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kbwxw\" (UniqueName: \"kubernetes.io/projected/d42ee9f9-390b-437c-a48a-ff7ead370f61-kube-api-access-kbwxw\") pod \"dnsmasq-dns-85fd94d64c-nr8rq\" (UID: \"d42ee9f9-390b-437c-a48a-ff7ead370f61\") " pod="openstack/dnsmasq-dns-85fd94d64c-nr8rq" Oct 14 07:20:43 crc kubenswrapper[4870]: I1014 07:20:43.739285 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/8962df13-02a8-4049-bf25-767413eaa98f-httpd-config\") pod \"neutron-5dd746bf8d-mvg6f\" (UID: \"8962df13-02a8-4049-bf25-767413eaa98f\") " pod="openstack/neutron-5dd746bf8d-mvg6f" Oct 14 07:20:43 crc kubenswrapper[4870]: I1014 07:20:43.739307 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d42ee9f9-390b-437c-a48a-ff7ead370f61-ovsdbserver-nb\") pod \"dnsmasq-dns-85fd94d64c-nr8rq\" (UID: \"d42ee9f9-390b-437c-a48a-ff7ead370f61\") " pod="openstack/dnsmasq-dns-85fd94d64c-nr8rq" Oct 14 07:20:43 crc kubenswrapper[4870]: I1014 07:20:43.739326 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/8962df13-02a8-4049-bf25-767413eaa98f-ovndb-tls-certs\") pod \"neutron-5dd746bf8d-mvg6f\" (UID: \"8962df13-02a8-4049-bf25-767413eaa98f\") " pod="openstack/neutron-5dd746bf8d-mvg6f" Oct 14 07:20:43 crc kubenswrapper[4870]: I1014 07:20:43.739346 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0a2e4c80-d44f-40c0-9c04-74e9ce98ae36-config-data-custom\") pod \"barbican-keystone-listener-57d8bd87b6-vnpct\" (UID: \"0a2e4c80-d44f-40c0-9c04-74e9ce98ae36\") " pod="openstack/barbican-keystone-listener-57d8bd87b6-vnpct" Oct 14 07:20:43 crc kubenswrapper[4870]: I1014 07:20:43.739377 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8962df13-02a8-4049-bf25-767413eaa98f-combined-ca-bundle\") pod \"neutron-5dd746bf8d-mvg6f\" (UID: \"8962df13-02a8-4049-bf25-767413eaa98f\") " pod="openstack/neutron-5dd746bf8d-mvg6f" Oct 14 07:20:43 crc kubenswrapper[4870]: I1014 07:20:43.739405 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d42ee9f9-390b-437c-a48a-ff7ead370f61-ovsdbserver-sb\") pod \"dnsmasq-dns-85fd94d64c-nr8rq\" (UID: \"d42ee9f9-390b-437c-a48a-ff7ead370f61\") " pod="openstack/dnsmasq-dns-85fd94d64c-nr8rq" Oct 14 07:20:43 crc kubenswrapper[4870]: I1014 07:20:43.739423 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d42ee9f9-390b-437c-a48a-ff7ead370f61-config\") pod \"dnsmasq-dns-85fd94d64c-nr8rq\" (UID: \"d42ee9f9-390b-437c-a48a-ff7ead370f61\") " pod="openstack/dnsmasq-dns-85fd94d64c-nr8rq" Oct 14 07:20:43 crc kubenswrapper[4870]: I1014 07:20:43.740223 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d42ee9f9-390b-437c-a48a-ff7ead370f61-dns-swift-storage-0\") pod \"dnsmasq-dns-85fd94d64c-nr8rq\" (UID: \"d42ee9f9-390b-437c-a48a-ff7ead370f61\") " pod="openstack/dnsmasq-dns-85fd94d64c-nr8rq" Oct 14 07:20:43 crc kubenswrapper[4870]: I1014 07:20:43.740254 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/8962df13-02a8-4049-bf25-767413eaa98f-config\") pod \"neutron-5dd746bf8d-mvg6f\" (UID: \"8962df13-02a8-4049-bf25-767413eaa98f\") " pod="openstack/neutron-5dd746bf8d-mvg6f" Oct 14 07:20:43 crc kubenswrapper[4870]: I1014 07:20:43.741279 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0a2e4c80-d44f-40c0-9c04-74e9ce98ae36-logs\") pod \"barbican-keystone-listener-57d8bd87b6-vnpct\" (UID: \"0a2e4c80-d44f-40c0-9c04-74e9ce98ae36\") " pod="openstack/barbican-keystone-listener-57d8bd87b6-vnpct" Oct 14 07:20:43 crc kubenswrapper[4870]: I1014 07:20:43.747572 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a2e4c80-d44f-40c0-9c04-74e9ce98ae36-combined-ca-bundle\") pod \"barbican-keystone-listener-57d8bd87b6-vnpct\" (UID: \"0a2e4c80-d44f-40c0-9c04-74e9ce98ae36\") " pod="openstack/barbican-keystone-listener-57d8bd87b6-vnpct" Oct 14 07:20:43 crc kubenswrapper[4870]: I1014 07:20:43.747763 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0a2e4c80-d44f-40c0-9c04-74e9ce98ae36-config-data-custom\") pod \"barbican-keystone-listener-57d8bd87b6-vnpct\" (UID: \"0a2e4c80-d44f-40c0-9c04-74e9ce98ae36\") " pod="openstack/barbican-keystone-listener-57d8bd87b6-vnpct" Oct 14 07:20:43 crc kubenswrapper[4870]: I1014 07:20:43.748057 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a2e4c80-d44f-40c0-9c04-74e9ce98ae36-config-data\") pod \"barbican-keystone-listener-57d8bd87b6-vnpct\" (UID: \"0a2e4c80-d44f-40c0-9c04-74e9ce98ae36\") " pod="openstack/barbican-keystone-listener-57d8bd87b6-vnpct" Oct 14 07:20:43 crc kubenswrapper[4870]: I1014 07:20:43.760232 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hbl6f\" (UniqueName: \"kubernetes.io/projected/0a2e4c80-d44f-40c0-9c04-74e9ce98ae36-kube-api-access-hbl6f\") pod \"barbican-keystone-listener-57d8bd87b6-vnpct\" (UID: \"0a2e4c80-d44f-40c0-9c04-74e9ce98ae36\") " pod="openstack/barbican-keystone-listener-57d8bd87b6-vnpct" Oct 14 07:20:43 crc kubenswrapper[4870]: I1014 07:20:43.773915 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-85f775bc47-d95qn" Oct 14 07:20:43 crc kubenswrapper[4870]: I1014 07:20:43.803290 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-57d8bd87b6-vnpct" Oct 14 07:20:43 crc kubenswrapper[4870]: I1014 07:20:43.841170 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3822d444-e96e-4883-883f-d74e44b17a03-combined-ca-bundle\") pod \"barbican-api-8446b89fb-22zh4\" (UID: \"3822d444-e96e-4883-883f-d74e44b17a03\") " pod="openstack/barbican-api-8446b89fb-22zh4" Oct 14 07:20:43 crc kubenswrapper[4870]: I1014 07:20:43.841229 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8962df13-02a8-4049-bf25-767413eaa98f-combined-ca-bundle\") pod \"neutron-5dd746bf8d-mvg6f\" (UID: \"8962df13-02a8-4049-bf25-767413eaa98f\") " pod="openstack/neutron-5dd746bf8d-mvg6f" Oct 14 07:20:43 crc kubenswrapper[4870]: I1014 07:20:43.841261 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3822d444-e96e-4883-883f-d74e44b17a03-config-data-custom\") pod \"barbican-api-8446b89fb-22zh4\" (UID: \"3822d444-e96e-4883-883f-d74e44b17a03\") " pod="openstack/barbican-api-8446b89fb-22zh4" Oct 14 07:20:43 crc kubenswrapper[4870]: I1014 07:20:43.841302 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d42ee9f9-390b-437c-a48a-ff7ead370f61-ovsdbserver-sb\") pod \"dnsmasq-dns-85fd94d64c-nr8rq\" (UID: \"d42ee9f9-390b-437c-a48a-ff7ead370f61\") " pod="openstack/dnsmasq-dns-85fd94d64c-nr8rq" Oct 14 07:20:43 crc kubenswrapper[4870]: I1014 07:20:43.841329 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d42ee9f9-390b-437c-a48a-ff7ead370f61-config\") pod \"dnsmasq-dns-85fd94d64c-nr8rq\" (UID: \"d42ee9f9-390b-437c-a48a-ff7ead370f61\") " pod="openstack/dnsmasq-dns-85fd94d64c-nr8rq" Oct 14 07:20:43 crc kubenswrapper[4870]: I1014 07:20:43.841344 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d42ee9f9-390b-437c-a48a-ff7ead370f61-dns-swift-storage-0\") pod \"dnsmasq-dns-85fd94d64c-nr8rq\" (UID: \"d42ee9f9-390b-437c-a48a-ff7ead370f61\") " pod="openstack/dnsmasq-dns-85fd94d64c-nr8rq" Oct 14 07:20:43 crc kubenswrapper[4870]: I1014 07:20:43.841363 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/8962df13-02a8-4049-bf25-767413eaa98f-config\") pod \"neutron-5dd746bf8d-mvg6f\" (UID: \"8962df13-02a8-4049-bf25-767413eaa98f\") " pod="openstack/neutron-5dd746bf8d-mvg6f" Oct 14 07:20:43 crc kubenswrapper[4870]: I1014 07:20:43.841383 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3822d444-e96e-4883-883f-d74e44b17a03-config-data\") pod \"barbican-api-8446b89fb-22zh4\" (UID: \"3822d444-e96e-4883-883f-d74e44b17a03\") " pod="openstack/barbican-api-8446b89fb-22zh4" Oct 14 07:20:43 crc kubenswrapper[4870]: I1014 07:20:43.841412 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d42ee9f9-390b-437c-a48a-ff7ead370f61-dns-svc\") pod \"dnsmasq-dns-85fd94d64c-nr8rq\" (UID: \"d42ee9f9-390b-437c-a48a-ff7ead370f61\") " pod="openstack/dnsmasq-dns-85fd94d64c-nr8rq" Oct 14 07:20:43 crc kubenswrapper[4870]: I1014 07:20:43.841490 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gmddt\" (UniqueName: \"kubernetes.io/projected/8962df13-02a8-4049-bf25-767413eaa98f-kube-api-access-gmddt\") pod \"neutron-5dd746bf8d-mvg6f\" (UID: \"8962df13-02a8-4049-bf25-767413eaa98f\") " pod="openstack/neutron-5dd746bf8d-mvg6f" Oct 14 07:20:43 crc kubenswrapper[4870]: I1014 07:20:43.841510 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kbwxw\" (UniqueName: \"kubernetes.io/projected/d42ee9f9-390b-437c-a48a-ff7ead370f61-kube-api-access-kbwxw\") pod \"dnsmasq-dns-85fd94d64c-nr8rq\" (UID: \"d42ee9f9-390b-437c-a48a-ff7ead370f61\") " pod="openstack/dnsmasq-dns-85fd94d64c-nr8rq" Oct 14 07:20:43 crc kubenswrapper[4870]: I1014 07:20:43.841544 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3822d444-e96e-4883-883f-d74e44b17a03-logs\") pod \"barbican-api-8446b89fb-22zh4\" (UID: \"3822d444-e96e-4883-883f-d74e44b17a03\") " pod="openstack/barbican-api-8446b89fb-22zh4" Oct 14 07:20:43 crc kubenswrapper[4870]: I1014 07:20:43.841623 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/8962df13-02a8-4049-bf25-767413eaa98f-httpd-config\") pod \"neutron-5dd746bf8d-mvg6f\" (UID: \"8962df13-02a8-4049-bf25-767413eaa98f\") " pod="openstack/neutron-5dd746bf8d-mvg6f" Oct 14 07:20:43 crc kubenswrapper[4870]: I1014 07:20:43.841642 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d42ee9f9-390b-437c-a48a-ff7ead370f61-ovsdbserver-nb\") pod \"dnsmasq-dns-85fd94d64c-nr8rq\" (UID: \"d42ee9f9-390b-437c-a48a-ff7ead370f61\") " pod="openstack/dnsmasq-dns-85fd94d64c-nr8rq" Oct 14 07:20:43 crc kubenswrapper[4870]: I1014 07:20:43.841667 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/8962df13-02a8-4049-bf25-767413eaa98f-ovndb-tls-certs\") pod \"neutron-5dd746bf8d-mvg6f\" (UID: \"8962df13-02a8-4049-bf25-767413eaa98f\") " pod="openstack/neutron-5dd746bf8d-mvg6f" Oct 14 07:20:43 crc kubenswrapper[4870]: I1014 07:20:43.841690 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wqrzq\" (UniqueName: \"kubernetes.io/projected/3822d444-e96e-4883-883f-d74e44b17a03-kube-api-access-wqrzq\") pod \"barbican-api-8446b89fb-22zh4\" (UID: \"3822d444-e96e-4883-883f-d74e44b17a03\") " pod="openstack/barbican-api-8446b89fb-22zh4" Oct 14 07:20:43 crc kubenswrapper[4870]: I1014 07:20:43.842718 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d42ee9f9-390b-437c-a48a-ff7ead370f61-dns-swift-storage-0\") pod \"dnsmasq-dns-85fd94d64c-nr8rq\" (UID: \"d42ee9f9-390b-437c-a48a-ff7ead370f61\") " pod="openstack/dnsmasq-dns-85fd94d64c-nr8rq" Oct 14 07:20:43 crc kubenswrapper[4870]: I1014 07:20:43.844783 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d42ee9f9-390b-437c-a48a-ff7ead370f61-dns-svc\") pod \"dnsmasq-dns-85fd94d64c-nr8rq\" (UID: \"d42ee9f9-390b-437c-a48a-ff7ead370f61\") " pod="openstack/dnsmasq-dns-85fd94d64c-nr8rq" Oct 14 07:20:43 crc kubenswrapper[4870]: I1014 07:20:43.845979 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d42ee9f9-390b-437c-a48a-ff7ead370f61-ovsdbserver-nb\") pod \"dnsmasq-dns-85fd94d64c-nr8rq\" (UID: \"d42ee9f9-390b-437c-a48a-ff7ead370f61\") " pod="openstack/dnsmasq-dns-85fd94d64c-nr8rq" Oct 14 07:20:43 crc kubenswrapper[4870]: I1014 07:20:43.846180 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d42ee9f9-390b-437c-a48a-ff7ead370f61-config\") pod \"dnsmasq-dns-85fd94d64c-nr8rq\" (UID: \"d42ee9f9-390b-437c-a48a-ff7ead370f61\") " pod="openstack/dnsmasq-dns-85fd94d64c-nr8rq" Oct 14 07:20:43 crc kubenswrapper[4870]: I1014 07:20:43.846350 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d42ee9f9-390b-437c-a48a-ff7ead370f61-ovsdbserver-sb\") pod \"dnsmasq-dns-85fd94d64c-nr8rq\" (UID: \"d42ee9f9-390b-437c-a48a-ff7ead370f61\") " pod="openstack/dnsmasq-dns-85fd94d64c-nr8rq" Oct 14 07:20:43 crc kubenswrapper[4870]: I1014 07:20:43.849818 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/8962df13-02a8-4049-bf25-767413eaa98f-ovndb-tls-certs\") pod \"neutron-5dd746bf8d-mvg6f\" (UID: \"8962df13-02a8-4049-bf25-767413eaa98f\") " pod="openstack/neutron-5dd746bf8d-mvg6f" Oct 14 07:20:43 crc kubenswrapper[4870]: I1014 07:20:43.853080 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/8962df13-02a8-4049-bf25-767413eaa98f-httpd-config\") pod \"neutron-5dd746bf8d-mvg6f\" (UID: \"8962df13-02a8-4049-bf25-767413eaa98f\") " pod="openstack/neutron-5dd746bf8d-mvg6f" Oct 14 07:20:43 crc kubenswrapper[4870]: I1014 07:20:43.853546 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8962df13-02a8-4049-bf25-767413eaa98f-combined-ca-bundle\") pod \"neutron-5dd746bf8d-mvg6f\" (UID: \"8962df13-02a8-4049-bf25-767413eaa98f\") " pod="openstack/neutron-5dd746bf8d-mvg6f" Oct 14 07:20:43 crc kubenswrapper[4870]: I1014 07:20:43.854193 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/8962df13-02a8-4049-bf25-767413eaa98f-config\") pod \"neutron-5dd746bf8d-mvg6f\" (UID: \"8962df13-02a8-4049-bf25-767413eaa98f\") " pod="openstack/neutron-5dd746bf8d-mvg6f" Oct 14 07:20:43 crc kubenswrapper[4870]: I1014 07:20:43.858298 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gmddt\" (UniqueName: \"kubernetes.io/projected/8962df13-02a8-4049-bf25-767413eaa98f-kube-api-access-gmddt\") pod \"neutron-5dd746bf8d-mvg6f\" (UID: \"8962df13-02a8-4049-bf25-767413eaa98f\") " pod="openstack/neutron-5dd746bf8d-mvg6f" Oct 14 07:20:43 crc kubenswrapper[4870]: I1014 07:20:43.858413 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kbwxw\" (UniqueName: \"kubernetes.io/projected/d42ee9f9-390b-437c-a48a-ff7ead370f61-kube-api-access-kbwxw\") pod \"dnsmasq-dns-85fd94d64c-nr8rq\" (UID: \"d42ee9f9-390b-437c-a48a-ff7ead370f61\") " pod="openstack/dnsmasq-dns-85fd94d64c-nr8rq" Oct 14 07:20:43 crc kubenswrapper[4870]: I1014 07:20:43.941416 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0a31f5df-eae7-4087-a8c8-09c18f5ee9f0" containerName="ceilometer-central-agent" containerID="cri-o://fb6bda07666ba00316d5344c82fd8d09dfbee1b602c4c88217bbfce61566b775" gracePeriod=30 Oct 14 07:20:43 crc kubenswrapper[4870]: I1014 07:20:43.942020 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0a31f5df-eae7-4087-a8c8-09c18f5ee9f0","Type":"ContainerStarted","Data":"8920a6cc7a6c6a67b4c85711f9833baeda2f81771718cba0bb026538e14812f8"} Oct 14 07:20:43 crc kubenswrapper[4870]: I1014 07:20:43.942241 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 14 07:20:43 crc kubenswrapper[4870]: I1014 07:20:43.942282 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0a31f5df-eae7-4087-a8c8-09c18f5ee9f0" containerName="proxy-httpd" containerID="cri-o://8920a6cc7a6c6a67b4c85711f9833baeda2f81771718cba0bb026538e14812f8" gracePeriod=30 Oct 14 07:20:43 crc kubenswrapper[4870]: I1014 07:20:43.942324 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0a31f5df-eae7-4087-a8c8-09c18f5ee9f0" containerName="sg-core" containerID="cri-o://c9de159d85eb50d1ce07f0bac74b7ca00d3e33977898a2e94dc9152d2951d08b" gracePeriod=30 Oct 14 07:20:43 crc kubenswrapper[4870]: I1014 07:20:43.942360 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0a31f5df-eae7-4087-a8c8-09c18f5ee9f0" containerName="ceilometer-notification-agent" containerID="cri-o://b3061a7cada98f02a1bef9477945ae37abbb17c577fe0bba622c83a78792b37d" gracePeriod=30 Oct 14 07:20:43 crc kubenswrapper[4870]: I1014 07:20:43.944073 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3822d444-e96e-4883-883f-d74e44b17a03-logs\") pod \"barbican-api-8446b89fb-22zh4\" (UID: \"3822d444-e96e-4883-883f-d74e44b17a03\") " pod="openstack/barbican-api-8446b89fb-22zh4" Oct 14 07:20:43 crc kubenswrapper[4870]: I1014 07:20:43.944121 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wqrzq\" (UniqueName: \"kubernetes.io/projected/3822d444-e96e-4883-883f-d74e44b17a03-kube-api-access-wqrzq\") pod \"barbican-api-8446b89fb-22zh4\" (UID: \"3822d444-e96e-4883-883f-d74e44b17a03\") " pod="openstack/barbican-api-8446b89fb-22zh4" Oct 14 07:20:43 crc kubenswrapper[4870]: I1014 07:20:43.944146 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3822d444-e96e-4883-883f-d74e44b17a03-combined-ca-bundle\") pod \"barbican-api-8446b89fb-22zh4\" (UID: \"3822d444-e96e-4883-883f-d74e44b17a03\") " pod="openstack/barbican-api-8446b89fb-22zh4" Oct 14 07:20:43 crc kubenswrapper[4870]: I1014 07:20:43.944164 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3822d444-e96e-4883-883f-d74e44b17a03-config-data-custom\") pod \"barbican-api-8446b89fb-22zh4\" (UID: \"3822d444-e96e-4883-883f-d74e44b17a03\") " pod="openstack/barbican-api-8446b89fb-22zh4" Oct 14 07:20:43 crc kubenswrapper[4870]: I1014 07:20:43.944202 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3822d444-e96e-4883-883f-d74e44b17a03-config-data\") pod \"barbican-api-8446b89fb-22zh4\" (UID: \"3822d444-e96e-4883-883f-d74e44b17a03\") " pod="openstack/barbican-api-8446b89fb-22zh4" Oct 14 07:20:43 crc kubenswrapper[4870]: E1014 07:20:43.947841 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api@sha256:93b475af15a0d10e95cb17b98927077f05ac24c89472a601d677eb89f82fd429\\\"\"" pod="openstack/cinder-db-sync-h587c" podUID="000463bc-f3ac-40b7-90c8-32366e327a19" Oct 14 07:20:43 crc kubenswrapper[4870]: I1014 07:20:43.948490 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3822d444-e96e-4883-883f-d74e44b17a03-logs\") pod \"barbican-api-8446b89fb-22zh4\" (UID: \"3822d444-e96e-4883-883f-d74e44b17a03\") " pod="openstack/barbican-api-8446b89fb-22zh4" Oct 14 07:20:43 crc kubenswrapper[4870]: I1014 07:20:43.949097 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3822d444-e96e-4883-883f-d74e44b17a03-combined-ca-bundle\") pod \"barbican-api-8446b89fb-22zh4\" (UID: \"3822d444-e96e-4883-883f-d74e44b17a03\") " pod="openstack/barbican-api-8446b89fb-22zh4" Oct 14 07:20:43 crc kubenswrapper[4870]: I1014 07:20:43.953231 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3822d444-e96e-4883-883f-d74e44b17a03-config-data-custom\") pod \"barbican-api-8446b89fb-22zh4\" (UID: \"3822d444-e96e-4883-883f-d74e44b17a03\") " pod="openstack/barbican-api-8446b89fb-22zh4" Oct 14 07:20:43 crc kubenswrapper[4870]: I1014 07:20:43.968097 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3822d444-e96e-4883-883f-d74e44b17a03-config-data\") pod \"barbican-api-8446b89fb-22zh4\" (UID: \"3822d444-e96e-4883-883f-d74e44b17a03\") " pod="openstack/barbican-api-8446b89fb-22zh4" Oct 14 07:20:43 crc kubenswrapper[4870]: I1014 07:20:43.975509 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wqrzq\" (UniqueName: \"kubernetes.io/projected/3822d444-e96e-4883-883f-d74e44b17a03-kube-api-access-wqrzq\") pod \"barbican-api-8446b89fb-22zh4\" (UID: \"3822d444-e96e-4883-883f-d74e44b17a03\") " pod="openstack/barbican-api-8446b89fb-22zh4" Oct 14 07:20:44 crc kubenswrapper[4870]: I1014 07:20:44.000223 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5dd746bf8d-mvg6f" Oct 14 07:20:44 crc kubenswrapper[4870]: I1014 07:20:44.010116 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85fd94d64c-nr8rq" Oct 14 07:20:44 crc kubenswrapper[4870]: I1014 07:20:44.012741 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.291575396 podStartE2EDuration="40.012706328s" podCreationTimestamp="2025-10-14 07:20:04 +0000 UTC" firstStartedPulling="2025-10-14 07:20:05.516851197 +0000 UTC m=+1141.214211568" lastFinishedPulling="2025-10-14 07:20:43.237982119 +0000 UTC m=+1178.935342500" observedRunningTime="2025-10-14 07:20:43.991268979 +0000 UTC m=+1179.688629350" watchObservedRunningTime="2025-10-14 07:20:44.012706328 +0000 UTC m=+1179.710066699" Oct 14 07:20:44 crc kubenswrapper[4870]: I1014 07:20:44.021684 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-8446b89fb-22zh4" Oct 14 07:20:44 crc kubenswrapper[4870]: I1014 07:20:44.135221 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7fdcbfbb99-5pnss"] Oct 14 07:20:44 crc kubenswrapper[4870]: I1014 07:20:44.294320 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-85f775bc47-d95qn"] Oct 14 07:20:44 crc kubenswrapper[4870]: W1014 07:20:44.308535 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc5bb3fde_cfaf_4d09_ad31_e23fd830eb91.slice/crio-f01c5670c5028a8cacb48b675dfcae6eaa4e1828d52432656e0cc9a9b15bd160 WatchSource:0}: Error finding container f01c5670c5028a8cacb48b675dfcae6eaa4e1828d52432656e0cc9a9b15bd160: Status 404 returned error can't find the container with id f01c5670c5028a8cacb48b675dfcae6eaa4e1828d52432656e0cc9a9b15bd160 Oct 14 07:20:44 crc kubenswrapper[4870]: I1014 07:20:44.374538 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-57d8bd87b6-vnpct"] Oct 14 07:20:44 crc kubenswrapper[4870]: I1014 07:20:44.649732 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-5dd746bf8d-mvg6f"] Oct 14 07:20:44 crc kubenswrapper[4870]: I1014 07:20:44.662785 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-8446b89fb-22zh4"] Oct 14 07:20:44 crc kubenswrapper[4870]: W1014 07:20:44.672145 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8962df13_02a8_4049_bf25_767413eaa98f.slice/crio-fc215b00e15e2e22b7b6f6158ff8bbac03f014ae65563e668ff453f53f611ded WatchSource:0}: Error finding container fc215b00e15e2e22b7b6f6158ff8bbac03f014ae65563e668ff453f53f611ded: Status 404 returned error can't find the container with id fc215b00e15e2e22b7b6f6158ff8bbac03f014ae65563e668ff453f53f611ded Oct 14 07:20:44 crc kubenswrapper[4870]: W1014 07:20:44.672396 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3822d444_e96e_4883_883f_d74e44b17a03.slice/crio-7b0d4385c2f1ebcf7d14f05f88d8a6a7602a4af150408e2e9868df5c97819b2d WatchSource:0}: Error finding container 7b0d4385c2f1ebcf7d14f05f88d8a6a7602a4af150408e2e9868df5c97819b2d: Status 404 returned error can't find the container with id 7b0d4385c2f1ebcf7d14f05f88d8a6a7602a4af150408e2e9868df5c97819b2d Oct 14 07:20:44 crc kubenswrapper[4870]: I1014 07:20:44.716078 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-85fd94d64c-nr8rq"] Oct 14 07:20:45 crc kubenswrapper[4870]: I1014 07:20:45.045060 4870 generic.go:334] "Generic (PLEG): container finished" podID="0a31f5df-eae7-4087-a8c8-09c18f5ee9f0" containerID="8920a6cc7a6c6a67b4c85711f9833baeda2f81771718cba0bb026538e14812f8" exitCode=0 Oct 14 07:20:45 crc kubenswrapper[4870]: I1014 07:20:45.045092 4870 generic.go:334] "Generic (PLEG): container finished" podID="0a31f5df-eae7-4087-a8c8-09c18f5ee9f0" containerID="c9de159d85eb50d1ce07f0bac74b7ca00d3e33977898a2e94dc9152d2951d08b" exitCode=2 Oct 14 07:20:45 crc kubenswrapper[4870]: I1014 07:20:45.045099 4870 generic.go:334] "Generic (PLEG): container finished" podID="0a31f5df-eae7-4087-a8c8-09c18f5ee9f0" containerID="b3061a7cada98f02a1bef9477945ae37abbb17c577fe0bba622c83a78792b37d" exitCode=0 Oct 14 07:20:45 crc kubenswrapper[4870]: I1014 07:20:45.045107 4870 generic.go:334] "Generic (PLEG): container finished" podID="0a31f5df-eae7-4087-a8c8-09c18f5ee9f0" containerID="fb6bda07666ba00316d5344c82fd8d09dfbee1b602c4c88217bbfce61566b775" exitCode=0 Oct 14 07:20:45 crc kubenswrapper[4870]: I1014 07:20:45.048791 4870 generic.go:334] "Generic (PLEG): container finished" podID="47ef233a-2847-4bcc-88cb-2be6e2f13e0e" containerID="b77d2d9a84171292181b6e5008e009f4a2b0c6e6686439d27da0bd3148d96769" exitCode=0 Oct 14 07:20:45 crc kubenswrapper[4870]: I1014 07:20:45.077519 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-85f775bc47-d95qn" event={"ID":"c5bb3fde-cfaf-4d09-ad31-e23fd830eb91","Type":"ContainerStarted","Data":"f01c5670c5028a8cacb48b675dfcae6eaa4e1828d52432656e0cc9a9b15bd160"} Oct 14 07:20:45 crc kubenswrapper[4870]: I1014 07:20:45.077934 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0a31f5df-eae7-4087-a8c8-09c18f5ee9f0","Type":"ContainerDied","Data":"8920a6cc7a6c6a67b4c85711f9833baeda2f81771718cba0bb026538e14812f8"} Oct 14 07:20:45 crc kubenswrapper[4870]: I1014 07:20:45.077949 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0a31f5df-eae7-4087-a8c8-09c18f5ee9f0","Type":"ContainerDied","Data":"c9de159d85eb50d1ce07f0bac74b7ca00d3e33977898a2e94dc9152d2951d08b"} Oct 14 07:20:45 crc kubenswrapper[4870]: I1014 07:20:45.077960 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0a31f5df-eae7-4087-a8c8-09c18f5ee9f0","Type":"ContainerDied","Data":"b3061a7cada98f02a1bef9477945ae37abbb17c577fe0bba622c83a78792b37d"} Oct 14 07:20:45 crc kubenswrapper[4870]: I1014 07:20:45.077971 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0a31f5df-eae7-4087-a8c8-09c18f5ee9f0","Type":"ContainerDied","Data":"fb6bda07666ba00316d5344c82fd8d09dfbee1b602c4c88217bbfce61566b775"} Oct 14 07:20:45 crc kubenswrapper[4870]: I1014 07:20:45.077981 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-8446b89fb-22zh4" event={"ID":"3822d444-e96e-4883-883f-d74e44b17a03","Type":"ContainerStarted","Data":"7b0d4385c2f1ebcf7d14f05f88d8a6a7602a4af150408e2e9868df5c97819b2d"} Oct 14 07:20:45 crc kubenswrapper[4870]: I1014 07:20:45.077992 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85fd94d64c-nr8rq" event={"ID":"d42ee9f9-390b-437c-a48a-ff7ead370f61","Type":"ContainerStarted","Data":"7d4427e577f38e681045883f59f7b2021995b5081d3fd7ec45ee57d5b7a3ee25"} Oct 14 07:20:45 crc kubenswrapper[4870]: I1014 07:20:45.078004 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-57d8bd87b6-vnpct" event={"ID":"0a2e4c80-d44f-40c0-9c04-74e9ce98ae36","Type":"ContainerStarted","Data":"8fe0611de8a26b46ab89d8cd485aaf180738d54c520c20fdca54ea25c93e6929"} Oct 14 07:20:45 crc kubenswrapper[4870]: I1014 07:20:45.078014 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5dd746bf8d-mvg6f" event={"ID":"8962df13-02a8-4049-bf25-767413eaa98f","Type":"ContainerStarted","Data":"fc215b00e15e2e22b7b6f6158ff8bbac03f014ae65563e668ff453f53f611ded"} Oct 14 07:20:45 crc kubenswrapper[4870]: I1014 07:20:45.078024 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fdcbfbb99-5pnss" event={"ID":"47ef233a-2847-4bcc-88cb-2be6e2f13e0e","Type":"ContainerDied","Data":"b77d2d9a84171292181b6e5008e009f4a2b0c6e6686439d27da0bd3148d96769"} Oct 14 07:20:45 crc kubenswrapper[4870]: I1014 07:20:45.078036 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fdcbfbb99-5pnss" event={"ID":"47ef233a-2847-4bcc-88cb-2be6e2f13e0e","Type":"ContainerStarted","Data":"4976b590b654320647eea545dade7c64d73c21029577605f8fbde976d40d6c60"} Oct 14 07:20:45 crc kubenswrapper[4870]: I1014 07:20:45.217913 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 14 07:20:45 crc kubenswrapper[4870]: I1014 07:20:45.293279 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a31f5df-eae7-4087-a8c8-09c18f5ee9f0-config-data\") pod \"0a31f5df-eae7-4087-a8c8-09c18f5ee9f0\" (UID: \"0a31f5df-eae7-4087-a8c8-09c18f5ee9f0\") " Oct 14 07:20:45 crc kubenswrapper[4870]: I1014 07:20:45.293394 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0a31f5df-eae7-4087-a8c8-09c18f5ee9f0-sg-core-conf-yaml\") pod \"0a31f5df-eae7-4087-a8c8-09c18f5ee9f0\" (UID: \"0a31f5df-eae7-4087-a8c8-09c18f5ee9f0\") " Oct 14 07:20:45 crc kubenswrapper[4870]: I1014 07:20:45.293415 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a31f5df-eae7-4087-a8c8-09c18f5ee9f0-combined-ca-bundle\") pod \"0a31f5df-eae7-4087-a8c8-09c18f5ee9f0\" (UID: \"0a31f5df-eae7-4087-a8c8-09c18f5ee9f0\") " Oct 14 07:20:45 crc kubenswrapper[4870]: I1014 07:20:45.293573 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0a31f5df-eae7-4087-a8c8-09c18f5ee9f0-log-httpd\") pod \"0a31f5df-eae7-4087-a8c8-09c18f5ee9f0\" (UID: \"0a31f5df-eae7-4087-a8c8-09c18f5ee9f0\") " Oct 14 07:20:45 crc kubenswrapper[4870]: I1014 07:20:45.293672 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0a31f5df-eae7-4087-a8c8-09c18f5ee9f0-scripts\") pod \"0a31f5df-eae7-4087-a8c8-09c18f5ee9f0\" (UID: \"0a31f5df-eae7-4087-a8c8-09c18f5ee9f0\") " Oct 14 07:20:45 crc kubenswrapper[4870]: I1014 07:20:45.293736 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kdzq6\" (UniqueName: \"kubernetes.io/projected/0a31f5df-eae7-4087-a8c8-09c18f5ee9f0-kube-api-access-kdzq6\") pod \"0a31f5df-eae7-4087-a8c8-09c18f5ee9f0\" (UID: \"0a31f5df-eae7-4087-a8c8-09c18f5ee9f0\") " Oct 14 07:20:45 crc kubenswrapper[4870]: I1014 07:20:45.293807 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0a31f5df-eae7-4087-a8c8-09c18f5ee9f0-run-httpd\") pod \"0a31f5df-eae7-4087-a8c8-09c18f5ee9f0\" (UID: \"0a31f5df-eae7-4087-a8c8-09c18f5ee9f0\") " Oct 14 07:20:45 crc kubenswrapper[4870]: I1014 07:20:45.295243 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0a31f5df-eae7-4087-a8c8-09c18f5ee9f0-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "0a31f5df-eae7-4087-a8c8-09c18f5ee9f0" (UID: "0a31f5df-eae7-4087-a8c8-09c18f5ee9f0"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:20:45 crc kubenswrapper[4870]: I1014 07:20:45.298262 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0a31f5df-eae7-4087-a8c8-09c18f5ee9f0-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "0a31f5df-eae7-4087-a8c8-09c18f5ee9f0" (UID: "0a31f5df-eae7-4087-a8c8-09c18f5ee9f0"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:20:45 crc kubenswrapper[4870]: I1014 07:20:45.307890 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a31f5df-eae7-4087-a8c8-09c18f5ee9f0-scripts" (OuterVolumeSpecName: "scripts") pod "0a31f5df-eae7-4087-a8c8-09c18f5ee9f0" (UID: "0a31f5df-eae7-4087-a8c8-09c18f5ee9f0"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:20:45 crc kubenswrapper[4870]: I1014 07:20:45.324476 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0a31f5df-eae7-4087-a8c8-09c18f5ee9f0-kube-api-access-kdzq6" (OuterVolumeSpecName: "kube-api-access-kdzq6") pod "0a31f5df-eae7-4087-a8c8-09c18f5ee9f0" (UID: "0a31f5df-eae7-4087-a8c8-09c18f5ee9f0"). InnerVolumeSpecName "kube-api-access-kdzq6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:20:45 crc kubenswrapper[4870]: I1014 07:20:45.342601 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a31f5df-eae7-4087-a8c8-09c18f5ee9f0-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "0a31f5df-eae7-4087-a8c8-09c18f5ee9f0" (UID: "0a31f5df-eae7-4087-a8c8-09c18f5ee9f0"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:20:45 crc kubenswrapper[4870]: I1014 07:20:45.397210 4870 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0a31f5df-eae7-4087-a8c8-09c18f5ee9f0-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 14 07:20:45 crc kubenswrapper[4870]: I1014 07:20:45.397555 4870 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0a31f5df-eae7-4087-a8c8-09c18f5ee9f0-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 14 07:20:45 crc kubenswrapper[4870]: I1014 07:20:45.397564 4870 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0a31f5df-eae7-4087-a8c8-09c18f5ee9f0-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 07:20:45 crc kubenswrapper[4870]: I1014 07:20:45.397573 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kdzq6\" (UniqueName: \"kubernetes.io/projected/0a31f5df-eae7-4087-a8c8-09c18f5ee9f0-kube-api-access-kdzq6\") on node \"crc\" DevicePath \"\"" Oct 14 07:20:45 crc kubenswrapper[4870]: I1014 07:20:45.397584 4870 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0a31f5df-eae7-4087-a8c8-09c18f5ee9f0-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 14 07:20:45 crc kubenswrapper[4870]: I1014 07:20:45.407517 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a31f5df-eae7-4087-a8c8-09c18f5ee9f0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0a31f5df-eae7-4087-a8c8-09c18f5ee9f0" (UID: "0a31f5df-eae7-4087-a8c8-09c18f5ee9f0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:20:45 crc kubenswrapper[4870]: I1014 07:20:45.439330 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a31f5df-eae7-4087-a8c8-09c18f5ee9f0-config-data" (OuterVolumeSpecName: "config-data") pod "0a31f5df-eae7-4087-a8c8-09c18f5ee9f0" (UID: "0a31f5df-eae7-4087-a8c8-09c18f5ee9f0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:20:45 crc kubenswrapper[4870]: I1014 07:20:45.499000 4870 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a31f5df-eae7-4087-a8c8-09c18f5ee9f0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 07:20:45 crc kubenswrapper[4870]: I1014 07:20:45.499040 4870 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a31f5df-eae7-4087-a8c8-09c18f5ee9f0-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 07:20:45 crc kubenswrapper[4870]: I1014 07:20:45.575495 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fdcbfbb99-5pnss" Oct 14 07:20:45 crc kubenswrapper[4870]: I1014 07:20:45.702386 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/47ef233a-2847-4bcc-88cb-2be6e2f13e0e-ovsdbserver-sb\") pod \"47ef233a-2847-4bcc-88cb-2be6e2f13e0e\" (UID: \"47ef233a-2847-4bcc-88cb-2be6e2f13e0e\") " Oct 14 07:20:45 crc kubenswrapper[4870]: I1014 07:20:45.702423 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/47ef233a-2847-4bcc-88cb-2be6e2f13e0e-dns-svc\") pod \"47ef233a-2847-4bcc-88cb-2be6e2f13e0e\" (UID: \"47ef233a-2847-4bcc-88cb-2be6e2f13e0e\") " Oct 14 07:20:45 crc kubenswrapper[4870]: I1014 07:20:45.702492 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/47ef233a-2847-4bcc-88cb-2be6e2f13e0e-ovsdbserver-nb\") pod \"47ef233a-2847-4bcc-88cb-2be6e2f13e0e\" (UID: \"47ef233a-2847-4bcc-88cb-2be6e2f13e0e\") " Oct 14 07:20:45 crc kubenswrapper[4870]: I1014 07:20:45.702534 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-57gqf\" (UniqueName: \"kubernetes.io/projected/47ef233a-2847-4bcc-88cb-2be6e2f13e0e-kube-api-access-57gqf\") pod \"47ef233a-2847-4bcc-88cb-2be6e2f13e0e\" (UID: \"47ef233a-2847-4bcc-88cb-2be6e2f13e0e\") " Oct 14 07:20:45 crc kubenswrapper[4870]: I1014 07:20:45.703111 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/47ef233a-2847-4bcc-88cb-2be6e2f13e0e-dns-swift-storage-0\") pod \"47ef233a-2847-4bcc-88cb-2be6e2f13e0e\" (UID: \"47ef233a-2847-4bcc-88cb-2be6e2f13e0e\") " Oct 14 07:20:45 crc kubenswrapper[4870]: I1014 07:20:45.703146 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/47ef233a-2847-4bcc-88cb-2be6e2f13e0e-config\") pod \"47ef233a-2847-4bcc-88cb-2be6e2f13e0e\" (UID: \"47ef233a-2847-4bcc-88cb-2be6e2f13e0e\") " Oct 14 07:20:45 crc kubenswrapper[4870]: I1014 07:20:45.707246 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/47ef233a-2847-4bcc-88cb-2be6e2f13e0e-kube-api-access-57gqf" (OuterVolumeSpecName: "kube-api-access-57gqf") pod "47ef233a-2847-4bcc-88cb-2be6e2f13e0e" (UID: "47ef233a-2847-4bcc-88cb-2be6e2f13e0e"). InnerVolumeSpecName "kube-api-access-57gqf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:20:45 crc kubenswrapper[4870]: I1014 07:20:45.730741 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/47ef233a-2847-4bcc-88cb-2be6e2f13e0e-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "47ef233a-2847-4bcc-88cb-2be6e2f13e0e" (UID: "47ef233a-2847-4bcc-88cb-2be6e2f13e0e"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:20:45 crc kubenswrapper[4870]: I1014 07:20:45.735953 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/47ef233a-2847-4bcc-88cb-2be6e2f13e0e-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "47ef233a-2847-4bcc-88cb-2be6e2f13e0e" (UID: "47ef233a-2847-4bcc-88cb-2be6e2f13e0e"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:20:45 crc kubenswrapper[4870]: I1014 07:20:45.746672 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/47ef233a-2847-4bcc-88cb-2be6e2f13e0e-config" (OuterVolumeSpecName: "config") pod "47ef233a-2847-4bcc-88cb-2be6e2f13e0e" (UID: "47ef233a-2847-4bcc-88cb-2be6e2f13e0e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:20:45 crc kubenswrapper[4870]: I1014 07:20:45.750595 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/47ef233a-2847-4bcc-88cb-2be6e2f13e0e-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "47ef233a-2847-4bcc-88cb-2be6e2f13e0e" (UID: "47ef233a-2847-4bcc-88cb-2be6e2f13e0e"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:20:45 crc kubenswrapper[4870]: I1014 07:20:45.775156 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/47ef233a-2847-4bcc-88cb-2be6e2f13e0e-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "47ef233a-2847-4bcc-88cb-2be6e2f13e0e" (UID: "47ef233a-2847-4bcc-88cb-2be6e2f13e0e"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:20:45 crc kubenswrapper[4870]: I1014 07:20:45.804848 4870 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/47ef233a-2847-4bcc-88cb-2be6e2f13e0e-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 14 07:20:45 crc kubenswrapper[4870]: I1014 07:20:45.804886 4870 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/47ef233a-2847-4bcc-88cb-2be6e2f13e0e-config\") on node \"crc\" DevicePath \"\"" Oct 14 07:20:45 crc kubenswrapper[4870]: I1014 07:20:45.804899 4870 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/47ef233a-2847-4bcc-88cb-2be6e2f13e0e-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 14 07:20:45 crc kubenswrapper[4870]: I1014 07:20:45.804910 4870 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/47ef233a-2847-4bcc-88cb-2be6e2f13e0e-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 14 07:20:45 crc kubenswrapper[4870]: I1014 07:20:45.804921 4870 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/47ef233a-2847-4bcc-88cb-2be6e2f13e0e-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 14 07:20:45 crc kubenswrapper[4870]: I1014 07:20:45.804933 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-57gqf\" (UniqueName: \"kubernetes.io/projected/47ef233a-2847-4bcc-88cb-2be6e2f13e0e-kube-api-access-57gqf\") on node \"crc\" DevicePath \"\"" Oct 14 07:20:46 crc kubenswrapper[4870]: I1014 07:20:46.058342 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5dd746bf8d-mvg6f" event={"ID":"8962df13-02a8-4049-bf25-767413eaa98f","Type":"ContainerStarted","Data":"fd406aca28373d20897ebf5edf751ce84f312fa9b73009dd19544478e754ccde"} Oct 14 07:20:46 crc kubenswrapper[4870]: I1014 07:20:46.058881 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5dd746bf8d-mvg6f" event={"ID":"8962df13-02a8-4049-bf25-767413eaa98f","Type":"ContainerStarted","Data":"b1d406e0278bc21b8489162b2029d7503109f564a8f097477cfabec3a829c8eb"} Oct 14 07:20:46 crc kubenswrapper[4870]: I1014 07:20:46.058966 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-5dd746bf8d-mvg6f" Oct 14 07:20:46 crc kubenswrapper[4870]: I1014 07:20:46.060690 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fdcbfbb99-5pnss" event={"ID":"47ef233a-2847-4bcc-88cb-2be6e2f13e0e","Type":"ContainerDied","Data":"4976b590b654320647eea545dade7c64d73c21029577605f8fbde976d40d6c60"} Oct 14 07:20:46 crc kubenswrapper[4870]: I1014 07:20:46.060719 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fdcbfbb99-5pnss" Oct 14 07:20:46 crc kubenswrapper[4870]: I1014 07:20:46.060733 4870 scope.go:117] "RemoveContainer" containerID="b77d2d9a84171292181b6e5008e009f4a2b0c6e6686439d27da0bd3148d96769" Oct 14 07:20:46 crc kubenswrapper[4870]: I1014 07:20:46.065707 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0a31f5df-eae7-4087-a8c8-09c18f5ee9f0","Type":"ContainerDied","Data":"18d9eed8f3dddc49cbc986fb2c0eaf635de30dcdadeaba0a0fea12c3c5b66991"} Oct 14 07:20:46 crc kubenswrapper[4870]: I1014 07:20:46.065770 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 14 07:20:46 crc kubenswrapper[4870]: I1014 07:20:46.076217 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-8446b89fb-22zh4" event={"ID":"3822d444-e96e-4883-883f-d74e44b17a03","Type":"ContainerStarted","Data":"ec21fa31a0e69180b4e6653848787be77a45780d3320a010fc55c4f44c1180eb"} Oct 14 07:20:46 crc kubenswrapper[4870]: I1014 07:20:46.076283 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-8446b89fb-22zh4" event={"ID":"3822d444-e96e-4883-883f-d74e44b17a03","Type":"ContainerStarted","Data":"1e7415ee401824cb497f36cc1a9431304cd42432efb6947ac813932dbbf40fbd"} Oct 14 07:20:46 crc kubenswrapper[4870]: I1014 07:20:46.076355 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-8446b89fb-22zh4" Oct 14 07:20:46 crc kubenswrapper[4870]: I1014 07:20:46.076414 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-8446b89fb-22zh4" Oct 14 07:20:46 crc kubenswrapper[4870]: I1014 07:20:46.077756 4870 generic.go:334] "Generic (PLEG): container finished" podID="d42ee9f9-390b-437c-a48a-ff7ead370f61" containerID="9a50e7bdf997907fd2e3f2bef9580f2c3d7493ec4d3de99188be8d372a77a601" exitCode=0 Oct 14 07:20:46 crc kubenswrapper[4870]: I1014 07:20:46.077797 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85fd94d64c-nr8rq" event={"ID":"d42ee9f9-390b-437c-a48a-ff7ead370f61","Type":"ContainerDied","Data":"9a50e7bdf997907fd2e3f2bef9580f2c3d7493ec4d3de99188be8d372a77a601"} Oct 14 07:20:46 crc kubenswrapper[4870]: I1014 07:20:46.087869 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-5dd746bf8d-mvg6f" podStartSLOduration=3.087849002 podStartE2EDuration="3.087849002s" podCreationTimestamp="2025-10-14 07:20:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:20:46.087352472 +0000 UTC m=+1181.784712843" watchObservedRunningTime="2025-10-14 07:20:46.087849002 +0000 UTC m=+1181.785209383" Oct 14 07:20:46 crc kubenswrapper[4870]: I1014 07:20:46.112240 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-8446b89fb-22zh4" podStartSLOduration=3.112220469 podStartE2EDuration="3.112220469s" podCreationTimestamp="2025-10-14 07:20:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:20:46.101059566 +0000 UTC m=+1181.798419937" watchObservedRunningTime="2025-10-14 07:20:46.112220469 +0000 UTC m=+1181.809580840" Oct 14 07:20:46 crc kubenswrapper[4870]: I1014 07:20:46.206540 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7fdcbfbb99-5pnss"] Oct 14 07:20:46 crc kubenswrapper[4870]: I1014 07:20:46.213595 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7fdcbfbb99-5pnss"] Oct 14 07:20:46 crc kubenswrapper[4870]: I1014 07:20:46.237999 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 14 07:20:46 crc kubenswrapper[4870]: I1014 07:20:46.253505 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 14 07:20:46 crc kubenswrapper[4870]: I1014 07:20:46.260485 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 14 07:20:46 crc kubenswrapper[4870]: E1014 07:20:46.260928 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a31f5df-eae7-4087-a8c8-09c18f5ee9f0" containerName="ceilometer-central-agent" Oct 14 07:20:46 crc kubenswrapper[4870]: I1014 07:20:46.260948 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a31f5df-eae7-4087-a8c8-09c18f5ee9f0" containerName="ceilometer-central-agent" Oct 14 07:20:46 crc kubenswrapper[4870]: E1014 07:20:46.260968 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="47ef233a-2847-4bcc-88cb-2be6e2f13e0e" containerName="init" Oct 14 07:20:46 crc kubenswrapper[4870]: I1014 07:20:46.260974 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="47ef233a-2847-4bcc-88cb-2be6e2f13e0e" containerName="init" Oct 14 07:20:46 crc kubenswrapper[4870]: E1014 07:20:46.261001 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a31f5df-eae7-4087-a8c8-09c18f5ee9f0" containerName="proxy-httpd" Oct 14 07:20:46 crc kubenswrapper[4870]: I1014 07:20:46.261008 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a31f5df-eae7-4087-a8c8-09c18f5ee9f0" containerName="proxy-httpd" Oct 14 07:20:46 crc kubenswrapper[4870]: E1014 07:20:46.261018 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a31f5df-eae7-4087-a8c8-09c18f5ee9f0" containerName="sg-core" Oct 14 07:20:46 crc kubenswrapper[4870]: I1014 07:20:46.261023 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a31f5df-eae7-4087-a8c8-09c18f5ee9f0" containerName="sg-core" Oct 14 07:20:46 crc kubenswrapper[4870]: E1014 07:20:46.261031 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a31f5df-eae7-4087-a8c8-09c18f5ee9f0" containerName="ceilometer-notification-agent" Oct 14 07:20:46 crc kubenswrapper[4870]: I1014 07:20:46.261037 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a31f5df-eae7-4087-a8c8-09c18f5ee9f0" containerName="ceilometer-notification-agent" Oct 14 07:20:46 crc kubenswrapper[4870]: I1014 07:20:46.261203 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="0a31f5df-eae7-4087-a8c8-09c18f5ee9f0" containerName="ceilometer-notification-agent" Oct 14 07:20:46 crc kubenswrapper[4870]: I1014 07:20:46.261225 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="0a31f5df-eae7-4087-a8c8-09c18f5ee9f0" containerName="proxy-httpd" Oct 14 07:20:46 crc kubenswrapper[4870]: I1014 07:20:46.261237 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="0a31f5df-eae7-4087-a8c8-09c18f5ee9f0" containerName="ceilometer-central-agent" Oct 14 07:20:46 crc kubenswrapper[4870]: I1014 07:20:46.261248 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="47ef233a-2847-4bcc-88cb-2be6e2f13e0e" containerName="init" Oct 14 07:20:46 crc kubenswrapper[4870]: I1014 07:20:46.261256 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="0a31f5df-eae7-4087-a8c8-09c18f5ee9f0" containerName="sg-core" Oct 14 07:20:46 crc kubenswrapper[4870]: I1014 07:20:46.272022 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 14 07:20:46 crc kubenswrapper[4870]: I1014 07:20:46.277400 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 14 07:20:46 crc kubenswrapper[4870]: I1014 07:20:46.288708 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 14 07:20:46 crc kubenswrapper[4870]: I1014 07:20:46.290202 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 14 07:20:46 crc kubenswrapper[4870]: I1014 07:20:46.327032 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a148cb0b-68ec-4e80-90c9-4937ac3e2823-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a148cb0b-68ec-4e80-90c9-4937ac3e2823\") " pod="openstack/ceilometer-0" Oct 14 07:20:46 crc kubenswrapper[4870]: I1014 07:20:46.327108 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a148cb0b-68ec-4e80-90c9-4937ac3e2823-config-data\") pod \"ceilometer-0\" (UID: \"a148cb0b-68ec-4e80-90c9-4937ac3e2823\") " pod="openstack/ceilometer-0" Oct 14 07:20:46 crc kubenswrapper[4870]: I1014 07:20:46.327207 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a148cb0b-68ec-4e80-90c9-4937ac3e2823-scripts\") pod \"ceilometer-0\" (UID: \"a148cb0b-68ec-4e80-90c9-4937ac3e2823\") " pod="openstack/ceilometer-0" Oct 14 07:20:46 crc kubenswrapper[4870]: I1014 07:20:46.327228 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a148cb0b-68ec-4e80-90c9-4937ac3e2823-run-httpd\") pod \"ceilometer-0\" (UID: \"a148cb0b-68ec-4e80-90c9-4937ac3e2823\") " pod="openstack/ceilometer-0" Oct 14 07:20:46 crc kubenswrapper[4870]: I1014 07:20:46.327259 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c55nd\" (UniqueName: \"kubernetes.io/projected/a148cb0b-68ec-4e80-90c9-4937ac3e2823-kube-api-access-c55nd\") pod \"ceilometer-0\" (UID: \"a148cb0b-68ec-4e80-90c9-4937ac3e2823\") " pod="openstack/ceilometer-0" Oct 14 07:20:46 crc kubenswrapper[4870]: I1014 07:20:46.327289 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a148cb0b-68ec-4e80-90c9-4937ac3e2823-log-httpd\") pod \"ceilometer-0\" (UID: \"a148cb0b-68ec-4e80-90c9-4937ac3e2823\") " pod="openstack/ceilometer-0" Oct 14 07:20:46 crc kubenswrapper[4870]: I1014 07:20:46.327310 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a148cb0b-68ec-4e80-90c9-4937ac3e2823-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a148cb0b-68ec-4e80-90c9-4937ac3e2823\") " pod="openstack/ceilometer-0" Oct 14 07:20:46 crc kubenswrapper[4870]: I1014 07:20:46.415645 4870 scope.go:117] "RemoveContainer" containerID="8920a6cc7a6c6a67b4c85711f9833baeda2f81771718cba0bb026538e14812f8" Oct 14 07:20:46 crc kubenswrapper[4870]: I1014 07:20:46.430309 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a148cb0b-68ec-4e80-90c9-4937ac3e2823-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a148cb0b-68ec-4e80-90c9-4937ac3e2823\") " pod="openstack/ceilometer-0" Oct 14 07:20:46 crc kubenswrapper[4870]: I1014 07:20:46.430391 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a148cb0b-68ec-4e80-90c9-4937ac3e2823-config-data\") pod \"ceilometer-0\" (UID: \"a148cb0b-68ec-4e80-90c9-4937ac3e2823\") " pod="openstack/ceilometer-0" Oct 14 07:20:46 crc kubenswrapper[4870]: I1014 07:20:46.430475 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a148cb0b-68ec-4e80-90c9-4937ac3e2823-scripts\") pod \"ceilometer-0\" (UID: \"a148cb0b-68ec-4e80-90c9-4937ac3e2823\") " pod="openstack/ceilometer-0" Oct 14 07:20:46 crc kubenswrapper[4870]: I1014 07:20:46.430495 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a148cb0b-68ec-4e80-90c9-4937ac3e2823-run-httpd\") pod \"ceilometer-0\" (UID: \"a148cb0b-68ec-4e80-90c9-4937ac3e2823\") " pod="openstack/ceilometer-0" Oct 14 07:20:46 crc kubenswrapper[4870]: I1014 07:20:46.430553 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c55nd\" (UniqueName: \"kubernetes.io/projected/a148cb0b-68ec-4e80-90c9-4937ac3e2823-kube-api-access-c55nd\") pod \"ceilometer-0\" (UID: \"a148cb0b-68ec-4e80-90c9-4937ac3e2823\") " pod="openstack/ceilometer-0" Oct 14 07:20:46 crc kubenswrapper[4870]: I1014 07:20:46.430590 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a148cb0b-68ec-4e80-90c9-4937ac3e2823-log-httpd\") pod \"ceilometer-0\" (UID: \"a148cb0b-68ec-4e80-90c9-4937ac3e2823\") " pod="openstack/ceilometer-0" Oct 14 07:20:46 crc kubenswrapper[4870]: I1014 07:20:46.430613 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a148cb0b-68ec-4e80-90c9-4937ac3e2823-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a148cb0b-68ec-4e80-90c9-4937ac3e2823\") " pod="openstack/ceilometer-0" Oct 14 07:20:46 crc kubenswrapper[4870]: I1014 07:20:46.431182 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a148cb0b-68ec-4e80-90c9-4937ac3e2823-run-httpd\") pod \"ceilometer-0\" (UID: \"a148cb0b-68ec-4e80-90c9-4937ac3e2823\") " pod="openstack/ceilometer-0" Oct 14 07:20:46 crc kubenswrapper[4870]: I1014 07:20:46.431640 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a148cb0b-68ec-4e80-90c9-4937ac3e2823-log-httpd\") pod \"ceilometer-0\" (UID: \"a148cb0b-68ec-4e80-90c9-4937ac3e2823\") " pod="openstack/ceilometer-0" Oct 14 07:20:46 crc kubenswrapper[4870]: I1014 07:20:46.433958 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a148cb0b-68ec-4e80-90c9-4937ac3e2823-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a148cb0b-68ec-4e80-90c9-4937ac3e2823\") " pod="openstack/ceilometer-0" Oct 14 07:20:46 crc kubenswrapper[4870]: I1014 07:20:46.435589 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a148cb0b-68ec-4e80-90c9-4937ac3e2823-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a148cb0b-68ec-4e80-90c9-4937ac3e2823\") " pod="openstack/ceilometer-0" Oct 14 07:20:46 crc kubenswrapper[4870]: I1014 07:20:46.436718 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a148cb0b-68ec-4e80-90c9-4937ac3e2823-config-data\") pod \"ceilometer-0\" (UID: \"a148cb0b-68ec-4e80-90c9-4937ac3e2823\") " pod="openstack/ceilometer-0" Oct 14 07:20:46 crc kubenswrapper[4870]: I1014 07:20:46.438245 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a148cb0b-68ec-4e80-90c9-4937ac3e2823-scripts\") pod \"ceilometer-0\" (UID: \"a148cb0b-68ec-4e80-90c9-4937ac3e2823\") " pod="openstack/ceilometer-0" Oct 14 07:20:46 crc kubenswrapper[4870]: I1014 07:20:46.456068 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c55nd\" (UniqueName: \"kubernetes.io/projected/a148cb0b-68ec-4e80-90c9-4937ac3e2823-kube-api-access-c55nd\") pod \"ceilometer-0\" (UID: \"a148cb0b-68ec-4e80-90c9-4937ac3e2823\") " pod="openstack/ceilometer-0" Oct 14 07:20:46 crc kubenswrapper[4870]: I1014 07:20:46.602481 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-546b769ccc-fdhsk"] Oct 14 07:20:46 crc kubenswrapper[4870]: I1014 07:20:46.604814 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-546b769ccc-fdhsk" Oct 14 07:20:46 crc kubenswrapper[4870]: I1014 07:20:46.606882 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Oct 14 07:20:46 crc kubenswrapper[4870]: I1014 07:20:46.607006 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Oct 14 07:20:46 crc kubenswrapper[4870]: I1014 07:20:46.630317 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 14 07:20:46 crc kubenswrapper[4870]: I1014 07:20:46.634726 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/0ba0e2ed-57d0-4535-883e-10b9b985b3d3-config\") pod \"neutron-546b769ccc-fdhsk\" (UID: \"0ba0e2ed-57d0-4535-883e-10b9b985b3d3\") " pod="openstack/neutron-546b769ccc-fdhsk" Oct 14 07:20:46 crc kubenswrapper[4870]: I1014 07:20:46.634781 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0ba0e2ed-57d0-4535-883e-10b9b985b3d3-public-tls-certs\") pod \"neutron-546b769ccc-fdhsk\" (UID: \"0ba0e2ed-57d0-4535-883e-10b9b985b3d3\") " pod="openstack/neutron-546b769ccc-fdhsk" Oct 14 07:20:46 crc kubenswrapper[4870]: I1014 07:20:46.634880 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/0ba0e2ed-57d0-4535-883e-10b9b985b3d3-httpd-config\") pod \"neutron-546b769ccc-fdhsk\" (UID: \"0ba0e2ed-57d0-4535-883e-10b9b985b3d3\") " pod="openstack/neutron-546b769ccc-fdhsk" Oct 14 07:20:46 crc kubenswrapper[4870]: I1014 07:20:46.634916 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ba0e2ed-57d0-4535-883e-10b9b985b3d3-combined-ca-bundle\") pod \"neutron-546b769ccc-fdhsk\" (UID: \"0ba0e2ed-57d0-4535-883e-10b9b985b3d3\") " pod="openstack/neutron-546b769ccc-fdhsk" Oct 14 07:20:46 crc kubenswrapper[4870]: I1014 07:20:46.634948 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/0ba0e2ed-57d0-4535-883e-10b9b985b3d3-ovndb-tls-certs\") pod \"neutron-546b769ccc-fdhsk\" (UID: \"0ba0e2ed-57d0-4535-883e-10b9b985b3d3\") " pod="openstack/neutron-546b769ccc-fdhsk" Oct 14 07:20:46 crc kubenswrapper[4870]: I1014 07:20:46.634970 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0ba0e2ed-57d0-4535-883e-10b9b985b3d3-internal-tls-certs\") pod \"neutron-546b769ccc-fdhsk\" (UID: \"0ba0e2ed-57d0-4535-883e-10b9b985b3d3\") " pod="openstack/neutron-546b769ccc-fdhsk" Oct 14 07:20:46 crc kubenswrapper[4870]: I1014 07:20:46.635009 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-szj76\" (UniqueName: \"kubernetes.io/projected/0ba0e2ed-57d0-4535-883e-10b9b985b3d3-kube-api-access-szj76\") pod \"neutron-546b769ccc-fdhsk\" (UID: \"0ba0e2ed-57d0-4535-883e-10b9b985b3d3\") " pod="openstack/neutron-546b769ccc-fdhsk" Oct 14 07:20:46 crc kubenswrapper[4870]: I1014 07:20:46.637405 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-546b769ccc-fdhsk"] Oct 14 07:20:46 crc kubenswrapper[4870]: I1014 07:20:46.736338 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/0ba0e2ed-57d0-4535-883e-10b9b985b3d3-ovndb-tls-certs\") pod \"neutron-546b769ccc-fdhsk\" (UID: \"0ba0e2ed-57d0-4535-883e-10b9b985b3d3\") " pod="openstack/neutron-546b769ccc-fdhsk" Oct 14 07:20:46 crc kubenswrapper[4870]: I1014 07:20:46.736708 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0ba0e2ed-57d0-4535-883e-10b9b985b3d3-internal-tls-certs\") pod \"neutron-546b769ccc-fdhsk\" (UID: \"0ba0e2ed-57d0-4535-883e-10b9b985b3d3\") " pod="openstack/neutron-546b769ccc-fdhsk" Oct 14 07:20:46 crc kubenswrapper[4870]: I1014 07:20:46.736746 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-szj76\" (UniqueName: \"kubernetes.io/projected/0ba0e2ed-57d0-4535-883e-10b9b985b3d3-kube-api-access-szj76\") pod \"neutron-546b769ccc-fdhsk\" (UID: \"0ba0e2ed-57d0-4535-883e-10b9b985b3d3\") " pod="openstack/neutron-546b769ccc-fdhsk" Oct 14 07:20:46 crc kubenswrapper[4870]: I1014 07:20:46.736825 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/0ba0e2ed-57d0-4535-883e-10b9b985b3d3-config\") pod \"neutron-546b769ccc-fdhsk\" (UID: \"0ba0e2ed-57d0-4535-883e-10b9b985b3d3\") " pod="openstack/neutron-546b769ccc-fdhsk" Oct 14 07:20:46 crc kubenswrapper[4870]: I1014 07:20:46.736847 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0ba0e2ed-57d0-4535-883e-10b9b985b3d3-public-tls-certs\") pod \"neutron-546b769ccc-fdhsk\" (UID: \"0ba0e2ed-57d0-4535-883e-10b9b985b3d3\") " pod="openstack/neutron-546b769ccc-fdhsk" Oct 14 07:20:46 crc kubenswrapper[4870]: I1014 07:20:46.736914 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/0ba0e2ed-57d0-4535-883e-10b9b985b3d3-httpd-config\") pod \"neutron-546b769ccc-fdhsk\" (UID: \"0ba0e2ed-57d0-4535-883e-10b9b985b3d3\") " pod="openstack/neutron-546b769ccc-fdhsk" Oct 14 07:20:46 crc kubenswrapper[4870]: I1014 07:20:46.736939 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ba0e2ed-57d0-4535-883e-10b9b985b3d3-combined-ca-bundle\") pod \"neutron-546b769ccc-fdhsk\" (UID: \"0ba0e2ed-57d0-4535-883e-10b9b985b3d3\") " pod="openstack/neutron-546b769ccc-fdhsk" Oct 14 07:20:46 crc kubenswrapper[4870]: I1014 07:20:46.741454 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/0ba0e2ed-57d0-4535-883e-10b9b985b3d3-httpd-config\") pod \"neutron-546b769ccc-fdhsk\" (UID: \"0ba0e2ed-57d0-4535-883e-10b9b985b3d3\") " pod="openstack/neutron-546b769ccc-fdhsk" Oct 14 07:20:46 crc kubenswrapper[4870]: I1014 07:20:46.742118 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0ba0e2ed-57d0-4535-883e-10b9b985b3d3-internal-tls-certs\") pod \"neutron-546b769ccc-fdhsk\" (UID: \"0ba0e2ed-57d0-4535-883e-10b9b985b3d3\") " pod="openstack/neutron-546b769ccc-fdhsk" Oct 14 07:20:46 crc kubenswrapper[4870]: I1014 07:20:46.742494 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ba0e2ed-57d0-4535-883e-10b9b985b3d3-combined-ca-bundle\") pod \"neutron-546b769ccc-fdhsk\" (UID: \"0ba0e2ed-57d0-4535-883e-10b9b985b3d3\") " pod="openstack/neutron-546b769ccc-fdhsk" Oct 14 07:20:46 crc kubenswrapper[4870]: I1014 07:20:46.743693 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/0ba0e2ed-57d0-4535-883e-10b9b985b3d3-config\") pod \"neutron-546b769ccc-fdhsk\" (UID: \"0ba0e2ed-57d0-4535-883e-10b9b985b3d3\") " pod="openstack/neutron-546b769ccc-fdhsk" Oct 14 07:20:46 crc kubenswrapper[4870]: I1014 07:20:46.747020 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/0ba0e2ed-57d0-4535-883e-10b9b985b3d3-ovndb-tls-certs\") pod \"neutron-546b769ccc-fdhsk\" (UID: \"0ba0e2ed-57d0-4535-883e-10b9b985b3d3\") " pod="openstack/neutron-546b769ccc-fdhsk" Oct 14 07:20:46 crc kubenswrapper[4870]: I1014 07:20:46.750733 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0ba0e2ed-57d0-4535-883e-10b9b985b3d3-public-tls-certs\") pod \"neutron-546b769ccc-fdhsk\" (UID: \"0ba0e2ed-57d0-4535-883e-10b9b985b3d3\") " pod="openstack/neutron-546b769ccc-fdhsk" Oct 14 07:20:46 crc kubenswrapper[4870]: I1014 07:20:46.760122 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-szj76\" (UniqueName: \"kubernetes.io/projected/0ba0e2ed-57d0-4535-883e-10b9b985b3d3-kube-api-access-szj76\") pod \"neutron-546b769ccc-fdhsk\" (UID: \"0ba0e2ed-57d0-4535-883e-10b9b985b3d3\") " pod="openstack/neutron-546b769ccc-fdhsk" Oct 14 07:20:46 crc kubenswrapper[4870]: I1014 07:20:46.937651 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-546b769ccc-fdhsk" Oct 14 07:20:46 crc kubenswrapper[4870]: I1014 07:20:46.940972 4870 scope.go:117] "RemoveContainer" containerID="c9de159d85eb50d1ce07f0bac74b7ca00d3e33977898a2e94dc9152d2951d08b" Oct 14 07:20:47 crc kubenswrapper[4870]: I1014 07:20:47.018478 4870 scope.go:117] "RemoveContainer" containerID="b3061a7cada98f02a1bef9477945ae37abbb17c577fe0bba622c83a78792b37d" Oct 14 07:20:47 crc kubenswrapper[4870]: I1014 07:20:47.048125 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0a31f5df-eae7-4087-a8c8-09c18f5ee9f0" path="/var/lib/kubelet/pods/0a31f5df-eae7-4087-a8c8-09c18f5ee9f0/volumes" Oct 14 07:20:47 crc kubenswrapper[4870]: I1014 07:20:47.050057 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="47ef233a-2847-4bcc-88cb-2be6e2f13e0e" path="/var/lib/kubelet/pods/47ef233a-2847-4bcc-88cb-2be6e2f13e0e/volumes" Oct 14 07:20:47 crc kubenswrapper[4870]: I1014 07:20:47.076718 4870 scope.go:117] "RemoveContainer" containerID="fb6bda07666ba00316d5344c82fd8d09dfbee1b602c4c88217bbfce61566b775" Oct 14 07:20:47 crc kubenswrapper[4870]: I1014 07:20:47.535677 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 14 07:20:47 crc kubenswrapper[4870]: I1014 07:20:47.674806 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-546b769ccc-fdhsk"] Oct 14 07:20:47 crc kubenswrapper[4870]: W1014 07:20:47.678426 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0ba0e2ed_57d0_4535_883e_10b9b985b3d3.slice/crio-1c553e152dd7faad019d3848c24b46d39b8e66a6593592e2f0e2fb4653ee3b84 WatchSource:0}: Error finding container 1c553e152dd7faad019d3848c24b46d39b8e66a6593592e2f0e2fb4653ee3b84: Status 404 returned error can't find the container with id 1c553e152dd7faad019d3848c24b46d39b8e66a6593592e2f0e2fb4653ee3b84 Oct 14 07:20:48 crc kubenswrapper[4870]: I1014 07:20:48.111966 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-546b769ccc-fdhsk" event={"ID":"0ba0e2ed-57d0-4535-883e-10b9b985b3d3","Type":"ContainerStarted","Data":"f84e5ed457b1301ae67b84003d43c8e27a59446623074e29e6e2859280f620fd"} Oct 14 07:20:48 crc kubenswrapper[4870]: I1014 07:20:48.112242 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-546b769ccc-fdhsk" event={"ID":"0ba0e2ed-57d0-4535-883e-10b9b985b3d3","Type":"ContainerStarted","Data":"b6b74fcdf90322787a4403ce2a20ab64b512335fa185f5d3caa8a7cd3a4837f4"} Oct 14 07:20:48 crc kubenswrapper[4870]: I1014 07:20:48.112254 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-546b769ccc-fdhsk" event={"ID":"0ba0e2ed-57d0-4535-883e-10b9b985b3d3","Type":"ContainerStarted","Data":"1c553e152dd7faad019d3848c24b46d39b8e66a6593592e2f0e2fb4653ee3b84"} Oct 14 07:20:48 crc kubenswrapper[4870]: I1014 07:20:48.112464 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-546b769ccc-fdhsk" Oct 14 07:20:48 crc kubenswrapper[4870]: I1014 07:20:48.114185 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-85f775bc47-d95qn" event={"ID":"c5bb3fde-cfaf-4d09-ad31-e23fd830eb91","Type":"ContainerStarted","Data":"94a43ed16da62727b0d42c318ec266880c3131b81bd05fdd1dfddbb2a180533e"} Oct 14 07:20:48 crc kubenswrapper[4870]: I1014 07:20:48.114217 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-85f775bc47-d95qn" event={"ID":"c5bb3fde-cfaf-4d09-ad31-e23fd830eb91","Type":"ContainerStarted","Data":"a77bf6d6eb57cf53a83d8f0ff81a500083d29df6ab1d56a72558d63fae3b342f"} Oct 14 07:20:48 crc kubenswrapper[4870]: I1014 07:20:48.117605 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a148cb0b-68ec-4e80-90c9-4937ac3e2823","Type":"ContainerStarted","Data":"bdbf5a7358035c14b4b3d0dfa179494ba80a24fdcd2325b7cd0d218eab1b4828"} Oct 14 07:20:48 crc kubenswrapper[4870]: I1014 07:20:48.120369 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85fd94d64c-nr8rq" event={"ID":"d42ee9f9-390b-437c-a48a-ff7ead370f61","Type":"ContainerStarted","Data":"02d14106fdda254dcaf2de464737ed7252a713b703f968764194d49711c7bb76"} Oct 14 07:20:48 crc kubenswrapper[4870]: I1014 07:20:48.122548 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-57d8bd87b6-vnpct" event={"ID":"0a2e4c80-d44f-40c0-9c04-74e9ce98ae36","Type":"ContainerStarted","Data":"ea29b92302ab943ca5bc51b26bb9eba8dd8e4aa271ebf5c213d4ae71f3523daf"} Oct 14 07:20:48 crc kubenswrapper[4870]: I1014 07:20:48.122579 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-57d8bd87b6-vnpct" event={"ID":"0a2e4c80-d44f-40c0-9c04-74e9ce98ae36","Type":"ContainerStarted","Data":"bc61afdd96aedc563124f07cd981291ea8c01a0a034c78fba2e57e484564a0ee"} Oct 14 07:20:48 crc kubenswrapper[4870]: I1014 07:20:48.138376 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-546b769ccc-fdhsk" podStartSLOduration=2.138350753 podStartE2EDuration="2.138350753s" podCreationTimestamp="2025-10-14 07:20:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:20:48.136559947 +0000 UTC m=+1183.833920318" watchObservedRunningTime="2025-10-14 07:20:48.138350753 +0000 UTC m=+1183.835711144" Oct 14 07:20:48 crc kubenswrapper[4870]: I1014 07:20:48.179010 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-85f775bc47-d95qn" podStartSLOduration=2.449652045 podStartE2EDuration="5.178987986s" podCreationTimestamp="2025-10-14 07:20:43 +0000 UTC" firstStartedPulling="2025-10-14 07:20:44.314771681 +0000 UTC m=+1180.012132052" lastFinishedPulling="2025-10-14 07:20:47.044107622 +0000 UTC m=+1182.741467993" observedRunningTime="2025-10-14 07:20:48.152968145 +0000 UTC m=+1183.850328506" watchObservedRunningTime="2025-10-14 07:20:48.178987986 +0000 UTC m=+1183.876348357" Oct 14 07:20:48 crc kubenswrapper[4870]: I1014 07:20:48.185164 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-85fd94d64c-nr8rq" podStartSLOduration=5.185142909 podStartE2EDuration="5.185142909s" podCreationTimestamp="2025-10-14 07:20:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:20:48.18221085 +0000 UTC m=+1183.879571221" watchObservedRunningTime="2025-10-14 07:20:48.185142909 +0000 UTC m=+1183.882503280" Oct 14 07:20:48 crc kubenswrapper[4870]: I1014 07:20:48.201789 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-57d8bd87b6-vnpct" podStartSLOduration=2.5356648059999998 podStartE2EDuration="5.201764252s" podCreationTimestamp="2025-10-14 07:20:43 +0000 UTC" firstStartedPulling="2025-10-14 07:20:44.3861988 +0000 UTC m=+1180.083559171" lastFinishedPulling="2025-10-14 07:20:47.052298246 +0000 UTC m=+1182.749658617" observedRunningTime="2025-10-14 07:20:48.197293552 +0000 UTC m=+1183.894653923" watchObservedRunningTime="2025-10-14 07:20:48.201764252 +0000 UTC m=+1183.899124623" Oct 14 07:20:49 crc kubenswrapper[4870]: I1014 07:20:49.010842 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-85fd94d64c-nr8rq" Oct 14 07:20:49 crc kubenswrapper[4870]: I1014 07:20:49.161931 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a148cb0b-68ec-4e80-90c9-4937ac3e2823","Type":"ContainerStarted","Data":"99aa20a03e300b003a0916165f520ef17e1468aea37b7d719db40a6850e99f66"} Oct 14 07:20:49 crc kubenswrapper[4870]: I1014 07:20:49.813881 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-79cc757548-zxf4g"] Oct 14 07:20:49 crc kubenswrapper[4870]: I1014 07:20:49.816423 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-79cc757548-zxf4g" Oct 14 07:20:49 crc kubenswrapper[4870]: I1014 07:20:49.819172 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Oct 14 07:20:49 crc kubenswrapper[4870]: I1014 07:20:49.819316 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Oct 14 07:20:49 crc kubenswrapper[4870]: I1014 07:20:49.827252 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-79cc757548-zxf4g"] Oct 14 07:20:49 crc kubenswrapper[4870]: I1014 07:20:49.916988 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cae5b655-5d7c-4063-90ad-4d9913792d5b-logs\") pod \"barbican-api-79cc757548-zxf4g\" (UID: \"cae5b655-5d7c-4063-90ad-4d9913792d5b\") " pod="openstack/barbican-api-79cc757548-zxf4g" Oct 14 07:20:49 crc kubenswrapper[4870]: I1014 07:20:49.917056 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cae5b655-5d7c-4063-90ad-4d9913792d5b-config-data\") pod \"barbican-api-79cc757548-zxf4g\" (UID: \"cae5b655-5d7c-4063-90ad-4d9913792d5b\") " pod="openstack/barbican-api-79cc757548-zxf4g" Oct 14 07:20:49 crc kubenswrapper[4870]: I1014 07:20:49.917160 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cae5b655-5d7c-4063-90ad-4d9913792d5b-public-tls-certs\") pod \"barbican-api-79cc757548-zxf4g\" (UID: \"cae5b655-5d7c-4063-90ad-4d9913792d5b\") " pod="openstack/barbican-api-79cc757548-zxf4g" Oct 14 07:20:49 crc kubenswrapper[4870]: I1014 07:20:49.917342 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zgsbd\" (UniqueName: \"kubernetes.io/projected/cae5b655-5d7c-4063-90ad-4d9913792d5b-kube-api-access-zgsbd\") pod \"barbican-api-79cc757548-zxf4g\" (UID: \"cae5b655-5d7c-4063-90ad-4d9913792d5b\") " pod="openstack/barbican-api-79cc757548-zxf4g" Oct 14 07:20:49 crc kubenswrapper[4870]: I1014 07:20:49.917373 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/cae5b655-5d7c-4063-90ad-4d9913792d5b-internal-tls-certs\") pod \"barbican-api-79cc757548-zxf4g\" (UID: \"cae5b655-5d7c-4063-90ad-4d9913792d5b\") " pod="openstack/barbican-api-79cc757548-zxf4g" Oct 14 07:20:49 crc kubenswrapper[4870]: I1014 07:20:49.917398 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cae5b655-5d7c-4063-90ad-4d9913792d5b-combined-ca-bundle\") pod \"barbican-api-79cc757548-zxf4g\" (UID: \"cae5b655-5d7c-4063-90ad-4d9913792d5b\") " pod="openstack/barbican-api-79cc757548-zxf4g" Oct 14 07:20:49 crc kubenswrapper[4870]: I1014 07:20:49.917551 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cae5b655-5d7c-4063-90ad-4d9913792d5b-config-data-custom\") pod \"barbican-api-79cc757548-zxf4g\" (UID: \"cae5b655-5d7c-4063-90ad-4d9913792d5b\") " pod="openstack/barbican-api-79cc757548-zxf4g" Oct 14 07:20:50 crc kubenswrapper[4870]: I1014 07:20:50.018540 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cae5b655-5d7c-4063-90ad-4d9913792d5b-config-data\") pod \"barbican-api-79cc757548-zxf4g\" (UID: \"cae5b655-5d7c-4063-90ad-4d9913792d5b\") " pod="openstack/barbican-api-79cc757548-zxf4g" Oct 14 07:20:50 crc kubenswrapper[4870]: I1014 07:20:50.018586 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cae5b655-5d7c-4063-90ad-4d9913792d5b-public-tls-certs\") pod \"barbican-api-79cc757548-zxf4g\" (UID: \"cae5b655-5d7c-4063-90ad-4d9913792d5b\") " pod="openstack/barbican-api-79cc757548-zxf4g" Oct 14 07:20:50 crc kubenswrapper[4870]: I1014 07:20:50.018638 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zgsbd\" (UniqueName: \"kubernetes.io/projected/cae5b655-5d7c-4063-90ad-4d9913792d5b-kube-api-access-zgsbd\") pod \"barbican-api-79cc757548-zxf4g\" (UID: \"cae5b655-5d7c-4063-90ad-4d9913792d5b\") " pod="openstack/barbican-api-79cc757548-zxf4g" Oct 14 07:20:50 crc kubenswrapper[4870]: I1014 07:20:50.018661 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/cae5b655-5d7c-4063-90ad-4d9913792d5b-internal-tls-certs\") pod \"barbican-api-79cc757548-zxf4g\" (UID: \"cae5b655-5d7c-4063-90ad-4d9913792d5b\") " pod="openstack/barbican-api-79cc757548-zxf4g" Oct 14 07:20:50 crc kubenswrapper[4870]: I1014 07:20:50.018677 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cae5b655-5d7c-4063-90ad-4d9913792d5b-combined-ca-bundle\") pod \"barbican-api-79cc757548-zxf4g\" (UID: \"cae5b655-5d7c-4063-90ad-4d9913792d5b\") " pod="openstack/barbican-api-79cc757548-zxf4g" Oct 14 07:20:50 crc kubenswrapper[4870]: I1014 07:20:50.018700 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cae5b655-5d7c-4063-90ad-4d9913792d5b-config-data-custom\") pod \"barbican-api-79cc757548-zxf4g\" (UID: \"cae5b655-5d7c-4063-90ad-4d9913792d5b\") " pod="openstack/barbican-api-79cc757548-zxf4g" Oct 14 07:20:50 crc kubenswrapper[4870]: I1014 07:20:50.018743 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cae5b655-5d7c-4063-90ad-4d9913792d5b-logs\") pod \"barbican-api-79cc757548-zxf4g\" (UID: \"cae5b655-5d7c-4063-90ad-4d9913792d5b\") " pod="openstack/barbican-api-79cc757548-zxf4g" Oct 14 07:20:50 crc kubenswrapper[4870]: I1014 07:20:50.019176 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cae5b655-5d7c-4063-90ad-4d9913792d5b-logs\") pod \"barbican-api-79cc757548-zxf4g\" (UID: \"cae5b655-5d7c-4063-90ad-4d9913792d5b\") " pod="openstack/barbican-api-79cc757548-zxf4g" Oct 14 07:20:50 crc kubenswrapper[4870]: I1014 07:20:50.024239 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/cae5b655-5d7c-4063-90ad-4d9913792d5b-internal-tls-certs\") pod \"barbican-api-79cc757548-zxf4g\" (UID: \"cae5b655-5d7c-4063-90ad-4d9913792d5b\") " pod="openstack/barbican-api-79cc757548-zxf4g" Oct 14 07:20:50 crc kubenswrapper[4870]: I1014 07:20:50.025486 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cae5b655-5d7c-4063-90ad-4d9913792d5b-public-tls-certs\") pod \"barbican-api-79cc757548-zxf4g\" (UID: \"cae5b655-5d7c-4063-90ad-4d9913792d5b\") " pod="openstack/barbican-api-79cc757548-zxf4g" Oct 14 07:20:50 crc kubenswrapper[4870]: I1014 07:20:50.026453 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cae5b655-5d7c-4063-90ad-4d9913792d5b-config-data\") pod \"barbican-api-79cc757548-zxf4g\" (UID: \"cae5b655-5d7c-4063-90ad-4d9913792d5b\") " pod="openstack/barbican-api-79cc757548-zxf4g" Oct 14 07:20:50 crc kubenswrapper[4870]: I1014 07:20:50.028919 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cae5b655-5d7c-4063-90ad-4d9913792d5b-config-data-custom\") pod \"barbican-api-79cc757548-zxf4g\" (UID: \"cae5b655-5d7c-4063-90ad-4d9913792d5b\") " pod="openstack/barbican-api-79cc757548-zxf4g" Oct 14 07:20:50 crc kubenswrapper[4870]: I1014 07:20:50.030568 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cae5b655-5d7c-4063-90ad-4d9913792d5b-combined-ca-bundle\") pod \"barbican-api-79cc757548-zxf4g\" (UID: \"cae5b655-5d7c-4063-90ad-4d9913792d5b\") " pod="openstack/barbican-api-79cc757548-zxf4g" Oct 14 07:20:50 crc kubenswrapper[4870]: I1014 07:20:50.038921 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zgsbd\" (UniqueName: \"kubernetes.io/projected/cae5b655-5d7c-4063-90ad-4d9913792d5b-kube-api-access-zgsbd\") pod \"barbican-api-79cc757548-zxf4g\" (UID: \"cae5b655-5d7c-4063-90ad-4d9913792d5b\") " pod="openstack/barbican-api-79cc757548-zxf4g" Oct 14 07:20:50 crc kubenswrapper[4870]: I1014 07:20:50.148006 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-79cc757548-zxf4g" Oct 14 07:20:50 crc kubenswrapper[4870]: I1014 07:20:50.200186 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a148cb0b-68ec-4e80-90c9-4937ac3e2823","Type":"ContainerStarted","Data":"47b79d1a4d8054d0c304a894efd7d793f07b4275c55e04e12239669c544d33f6"} Oct 14 07:20:50 crc kubenswrapper[4870]: I1014 07:20:50.743368 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-79cc757548-zxf4g"] Oct 14 07:20:50 crc kubenswrapper[4870]: W1014 07:20:50.750597 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcae5b655_5d7c_4063_90ad_4d9913792d5b.slice/crio-1348cc0e9dccef391f47727a7b0673d69bdf58a16020dc4ffad5b01888b723b3 WatchSource:0}: Error finding container 1348cc0e9dccef391f47727a7b0673d69bdf58a16020dc4ffad5b01888b723b3: Status 404 returned error can't find the container with id 1348cc0e9dccef391f47727a7b0673d69bdf58a16020dc4ffad5b01888b723b3 Oct 14 07:20:51 crc kubenswrapper[4870]: I1014 07:20:51.212190 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-79cc757548-zxf4g" event={"ID":"cae5b655-5d7c-4063-90ad-4d9913792d5b","Type":"ContainerStarted","Data":"b97a3d2764961f388fdc1a6459d91ca5f1248920ce92c572b651657c109e7fa4"} Oct 14 07:20:51 crc kubenswrapper[4870]: I1014 07:20:51.212613 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-79cc757548-zxf4g" event={"ID":"cae5b655-5d7c-4063-90ad-4d9913792d5b","Type":"ContainerStarted","Data":"a507ea103a81fde95c96ade7a313c18b4faefbe53e02c99e5cd596e0c8dcb53e"} Oct 14 07:20:51 crc kubenswrapper[4870]: I1014 07:20:51.212636 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-79cc757548-zxf4g" Oct 14 07:20:51 crc kubenswrapper[4870]: I1014 07:20:51.212657 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-79cc757548-zxf4g" event={"ID":"cae5b655-5d7c-4063-90ad-4d9913792d5b","Type":"ContainerStarted","Data":"1348cc0e9dccef391f47727a7b0673d69bdf58a16020dc4ffad5b01888b723b3"} Oct 14 07:20:51 crc kubenswrapper[4870]: I1014 07:20:51.212671 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-79cc757548-zxf4g" Oct 14 07:20:51 crc kubenswrapper[4870]: I1014 07:20:51.213895 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a148cb0b-68ec-4e80-90c9-4937ac3e2823","Type":"ContainerStarted","Data":"44d88789123eae3dc981cd0bab8f55c3e851b41100c3d03e03d237349a7d49cb"} Oct 14 07:20:51 crc kubenswrapper[4870]: I1014 07:20:51.237713 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-79cc757548-zxf4g" podStartSLOduration=2.237691755 podStartE2EDuration="2.237691755s" podCreationTimestamp="2025-10-14 07:20:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:20:51.229745747 +0000 UTC m=+1186.927106138" watchObservedRunningTime="2025-10-14 07:20:51.237691755 +0000 UTC m=+1186.935052126" Oct 14 07:20:51 crc kubenswrapper[4870]: I1014 07:20:51.792052 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-597ff968b-msxbn" Oct 14 07:20:51 crc kubenswrapper[4870]: I1014 07:20:51.915597 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-597ff968b-msxbn" Oct 14 07:20:52 crc kubenswrapper[4870]: I1014 07:20:52.225053 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a148cb0b-68ec-4e80-90c9-4937ac3e2823","Type":"ContainerStarted","Data":"94f837271ef58fef81c8ae6563f0c235e72d9d2177b233a60e2b3a9d8dffd104"} Oct 14 07:20:52 crc kubenswrapper[4870]: I1014 07:20:52.225696 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 14 07:20:52 crc kubenswrapper[4870]: I1014 07:20:52.246384 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.084042266 podStartE2EDuration="6.246363495s" podCreationTimestamp="2025-10-14 07:20:46 +0000 UTC" firstStartedPulling="2025-10-14 07:20:47.548842469 +0000 UTC m=+1183.246202840" lastFinishedPulling="2025-10-14 07:20:51.711163698 +0000 UTC m=+1187.408524069" observedRunningTime="2025-10-14 07:20:52.241426436 +0000 UTC m=+1187.938786837" watchObservedRunningTime="2025-10-14 07:20:52.246363495 +0000 UTC m=+1187.943723866" Oct 14 07:20:54 crc kubenswrapper[4870]: I1014 07:20:54.011637 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-85fd94d64c-nr8rq" Oct 14 07:20:54 crc kubenswrapper[4870]: I1014 07:20:54.133683 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7bcdb8c767-9pwfh"] Oct 14 07:20:54 crc kubenswrapper[4870]: I1014 07:20:54.134031 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7bcdb8c767-9pwfh" podUID="32e2388e-2670-44d3-bfe1-6ce4c126eb18" containerName="dnsmasq-dns" containerID="cri-o://ea7ea208868fa97909c709c50154f181f12abb3d19b900341576486b790d35d4" gracePeriod=10 Oct 14 07:20:54 crc kubenswrapper[4870]: E1014 07:20:54.318803 4870 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod32e2388e_2670_44d3_bfe1_6ce4c126eb18.slice/crio-conmon-ea7ea208868fa97909c709c50154f181f12abb3d19b900341576486b790d35d4.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod32e2388e_2670_44d3_bfe1_6ce4c126eb18.slice/crio-ea7ea208868fa97909c709c50154f181f12abb3d19b900341576486b790d35d4.scope\": RecentStats: unable to find data in memory cache]" Oct 14 07:20:54 crc kubenswrapper[4870]: I1014 07:20:54.704584 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7bcdb8c767-9pwfh" Oct 14 07:20:54 crc kubenswrapper[4870]: I1014 07:20:54.723877 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dzqs2\" (UniqueName: \"kubernetes.io/projected/32e2388e-2670-44d3-bfe1-6ce4c126eb18-kube-api-access-dzqs2\") pod \"32e2388e-2670-44d3-bfe1-6ce4c126eb18\" (UID: \"32e2388e-2670-44d3-bfe1-6ce4c126eb18\") " Oct 14 07:20:54 crc kubenswrapper[4870]: I1014 07:20:54.724105 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/32e2388e-2670-44d3-bfe1-6ce4c126eb18-ovsdbserver-nb\") pod \"32e2388e-2670-44d3-bfe1-6ce4c126eb18\" (UID: \"32e2388e-2670-44d3-bfe1-6ce4c126eb18\") " Oct 14 07:20:54 crc kubenswrapper[4870]: I1014 07:20:54.724133 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/32e2388e-2670-44d3-bfe1-6ce4c126eb18-ovsdbserver-sb\") pod \"32e2388e-2670-44d3-bfe1-6ce4c126eb18\" (UID: \"32e2388e-2670-44d3-bfe1-6ce4c126eb18\") " Oct 14 07:20:54 crc kubenswrapper[4870]: I1014 07:20:54.724159 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/32e2388e-2670-44d3-bfe1-6ce4c126eb18-dns-svc\") pod \"32e2388e-2670-44d3-bfe1-6ce4c126eb18\" (UID: \"32e2388e-2670-44d3-bfe1-6ce4c126eb18\") " Oct 14 07:20:54 crc kubenswrapper[4870]: I1014 07:20:54.724193 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/32e2388e-2670-44d3-bfe1-6ce4c126eb18-config\") pod \"32e2388e-2670-44d3-bfe1-6ce4c126eb18\" (UID: \"32e2388e-2670-44d3-bfe1-6ce4c126eb18\") " Oct 14 07:20:54 crc kubenswrapper[4870]: I1014 07:20:54.724262 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/32e2388e-2670-44d3-bfe1-6ce4c126eb18-dns-swift-storage-0\") pod \"32e2388e-2670-44d3-bfe1-6ce4c126eb18\" (UID: \"32e2388e-2670-44d3-bfe1-6ce4c126eb18\") " Oct 14 07:20:54 crc kubenswrapper[4870]: I1014 07:20:54.730691 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/32e2388e-2670-44d3-bfe1-6ce4c126eb18-kube-api-access-dzqs2" (OuterVolumeSpecName: "kube-api-access-dzqs2") pod "32e2388e-2670-44d3-bfe1-6ce4c126eb18" (UID: "32e2388e-2670-44d3-bfe1-6ce4c126eb18"). InnerVolumeSpecName "kube-api-access-dzqs2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:20:54 crc kubenswrapper[4870]: I1014 07:20:54.786175 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/32e2388e-2670-44d3-bfe1-6ce4c126eb18-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "32e2388e-2670-44d3-bfe1-6ce4c126eb18" (UID: "32e2388e-2670-44d3-bfe1-6ce4c126eb18"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:20:54 crc kubenswrapper[4870]: I1014 07:20:54.789247 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/32e2388e-2670-44d3-bfe1-6ce4c126eb18-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "32e2388e-2670-44d3-bfe1-6ce4c126eb18" (UID: "32e2388e-2670-44d3-bfe1-6ce4c126eb18"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:20:54 crc kubenswrapper[4870]: I1014 07:20:54.823021 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/32e2388e-2670-44d3-bfe1-6ce4c126eb18-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "32e2388e-2670-44d3-bfe1-6ce4c126eb18" (UID: "32e2388e-2670-44d3-bfe1-6ce4c126eb18"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:20:54 crc kubenswrapper[4870]: I1014 07:20:54.823988 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/32e2388e-2670-44d3-bfe1-6ce4c126eb18-config" (OuterVolumeSpecName: "config") pod "32e2388e-2670-44d3-bfe1-6ce4c126eb18" (UID: "32e2388e-2670-44d3-bfe1-6ce4c126eb18"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:20:54 crc kubenswrapper[4870]: I1014 07:20:54.826716 4870 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/32e2388e-2670-44d3-bfe1-6ce4c126eb18-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 14 07:20:54 crc kubenswrapper[4870]: I1014 07:20:54.826742 4870 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/32e2388e-2670-44d3-bfe1-6ce4c126eb18-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 14 07:20:54 crc kubenswrapper[4870]: I1014 07:20:54.826753 4870 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/32e2388e-2670-44d3-bfe1-6ce4c126eb18-config\") on node \"crc\" DevicePath \"\"" Oct 14 07:20:54 crc kubenswrapper[4870]: I1014 07:20:54.826762 4870 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/32e2388e-2670-44d3-bfe1-6ce4c126eb18-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 14 07:20:54 crc kubenswrapper[4870]: I1014 07:20:54.826772 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dzqs2\" (UniqueName: \"kubernetes.io/projected/32e2388e-2670-44d3-bfe1-6ce4c126eb18-kube-api-access-dzqs2\") on node \"crc\" DevicePath \"\"" Oct 14 07:20:54 crc kubenswrapper[4870]: I1014 07:20:54.834097 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/32e2388e-2670-44d3-bfe1-6ce4c126eb18-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "32e2388e-2670-44d3-bfe1-6ce4c126eb18" (UID: "32e2388e-2670-44d3-bfe1-6ce4c126eb18"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:20:54 crc kubenswrapper[4870]: I1014 07:20:54.928368 4870 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/32e2388e-2670-44d3-bfe1-6ce4c126eb18-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 14 07:20:55 crc kubenswrapper[4870]: I1014 07:20:55.253621 4870 generic.go:334] "Generic (PLEG): container finished" podID="32e2388e-2670-44d3-bfe1-6ce4c126eb18" containerID="ea7ea208868fa97909c709c50154f181f12abb3d19b900341576486b790d35d4" exitCode=0 Oct 14 07:20:55 crc kubenswrapper[4870]: I1014 07:20:55.253675 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7bcdb8c767-9pwfh" event={"ID":"32e2388e-2670-44d3-bfe1-6ce4c126eb18","Type":"ContainerDied","Data":"ea7ea208868fa97909c709c50154f181f12abb3d19b900341576486b790d35d4"} Oct 14 07:20:55 crc kubenswrapper[4870]: I1014 07:20:55.253704 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7bcdb8c767-9pwfh" Oct 14 07:20:55 crc kubenswrapper[4870]: I1014 07:20:55.253731 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7bcdb8c767-9pwfh" event={"ID":"32e2388e-2670-44d3-bfe1-6ce4c126eb18","Type":"ContainerDied","Data":"7127f9e04251e62c25047d8f4a4b0eaa991272ab883a8e44991d474cf2b19e1e"} Oct 14 07:20:55 crc kubenswrapper[4870]: I1014 07:20:55.253754 4870 scope.go:117] "RemoveContainer" containerID="ea7ea208868fa97909c709c50154f181f12abb3d19b900341576486b790d35d4" Oct 14 07:20:55 crc kubenswrapper[4870]: I1014 07:20:55.356803 4870 scope.go:117] "RemoveContainer" containerID="84dddb0e2b529d2169d4dcc0f253839f94043cb699b4ebe505daff1715e343ff" Oct 14 07:20:55 crc kubenswrapper[4870]: I1014 07:20:55.365540 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7bcdb8c767-9pwfh"] Oct 14 07:20:55 crc kubenswrapper[4870]: I1014 07:20:55.375114 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7bcdb8c767-9pwfh"] Oct 14 07:20:55 crc kubenswrapper[4870]: I1014 07:20:55.381821 4870 scope.go:117] "RemoveContainer" containerID="ea7ea208868fa97909c709c50154f181f12abb3d19b900341576486b790d35d4" Oct 14 07:20:55 crc kubenswrapper[4870]: E1014 07:20:55.382602 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ea7ea208868fa97909c709c50154f181f12abb3d19b900341576486b790d35d4\": container with ID starting with ea7ea208868fa97909c709c50154f181f12abb3d19b900341576486b790d35d4 not found: ID does not exist" containerID="ea7ea208868fa97909c709c50154f181f12abb3d19b900341576486b790d35d4" Oct 14 07:20:55 crc kubenswrapper[4870]: I1014 07:20:55.382646 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ea7ea208868fa97909c709c50154f181f12abb3d19b900341576486b790d35d4"} err="failed to get container status \"ea7ea208868fa97909c709c50154f181f12abb3d19b900341576486b790d35d4\": rpc error: code = NotFound desc = could not find container \"ea7ea208868fa97909c709c50154f181f12abb3d19b900341576486b790d35d4\": container with ID starting with ea7ea208868fa97909c709c50154f181f12abb3d19b900341576486b790d35d4 not found: ID does not exist" Oct 14 07:20:55 crc kubenswrapper[4870]: I1014 07:20:55.382676 4870 scope.go:117] "RemoveContainer" containerID="84dddb0e2b529d2169d4dcc0f253839f94043cb699b4ebe505daff1715e343ff" Oct 14 07:20:55 crc kubenswrapper[4870]: E1014 07:20:55.382997 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"84dddb0e2b529d2169d4dcc0f253839f94043cb699b4ebe505daff1715e343ff\": container with ID starting with 84dddb0e2b529d2169d4dcc0f253839f94043cb699b4ebe505daff1715e343ff not found: ID does not exist" containerID="84dddb0e2b529d2169d4dcc0f253839f94043cb699b4ebe505daff1715e343ff" Oct 14 07:20:55 crc kubenswrapper[4870]: I1014 07:20:55.383029 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"84dddb0e2b529d2169d4dcc0f253839f94043cb699b4ebe505daff1715e343ff"} err="failed to get container status \"84dddb0e2b529d2169d4dcc0f253839f94043cb699b4ebe505daff1715e343ff\": rpc error: code = NotFound desc = could not find container \"84dddb0e2b529d2169d4dcc0f253839f94043cb699b4ebe505daff1715e343ff\": container with ID starting with 84dddb0e2b529d2169d4dcc0f253839f94043cb699b4ebe505daff1715e343ff not found: ID does not exist" Oct 14 07:20:55 crc kubenswrapper[4870]: I1014 07:20:55.615101 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-8446b89fb-22zh4" Oct 14 07:20:55 crc kubenswrapper[4870]: I1014 07:20:55.681800 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-8446b89fb-22zh4" Oct 14 07:20:56 crc kubenswrapper[4870]: I1014 07:20:56.227657 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-865c567797-jljrm" Oct 14 07:20:56 crc kubenswrapper[4870]: I1014 07:20:56.857921 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-79cc757548-zxf4g" Oct 14 07:20:57 crc kubenswrapper[4870]: I1014 07:20:57.051585 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="32e2388e-2670-44d3-bfe1-6ce4c126eb18" path="/var/lib/kubelet/pods/32e2388e-2670-44d3-bfe1-6ce4c126eb18/volumes" Oct 14 07:20:57 crc kubenswrapper[4870]: I1014 07:20:57.193692 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Oct 14 07:20:57 crc kubenswrapper[4870]: E1014 07:20:57.194054 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="32e2388e-2670-44d3-bfe1-6ce4c126eb18" containerName="init" Oct 14 07:20:57 crc kubenswrapper[4870]: I1014 07:20:57.194069 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="32e2388e-2670-44d3-bfe1-6ce4c126eb18" containerName="init" Oct 14 07:20:57 crc kubenswrapper[4870]: E1014 07:20:57.194113 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="32e2388e-2670-44d3-bfe1-6ce4c126eb18" containerName="dnsmasq-dns" Oct 14 07:20:57 crc kubenswrapper[4870]: I1014 07:20:57.194120 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="32e2388e-2670-44d3-bfe1-6ce4c126eb18" containerName="dnsmasq-dns" Oct 14 07:20:57 crc kubenswrapper[4870]: I1014 07:20:57.194312 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="32e2388e-2670-44d3-bfe1-6ce4c126eb18" containerName="dnsmasq-dns" Oct 14 07:20:57 crc kubenswrapper[4870]: I1014 07:20:57.195075 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 14 07:20:57 crc kubenswrapper[4870]: I1014 07:20:57.198356 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Oct 14 07:20:57 crc kubenswrapper[4870]: I1014 07:20:57.198470 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-svrzz" Oct 14 07:20:57 crc kubenswrapper[4870]: I1014 07:20:57.198686 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Oct 14 07:20:57 crc kubenswrapper[4870]: I1014 07:20:57.216231 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 14 07:20:57 crc kubenswrapper[4870]: I1014 07:20:57.368376 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/7806f6b7-5364-41e9-99af-4d201e72535b-openstack-config-secret\") pod \"openstackclient\" (UID: \"7806f6b7-5364-41e9-99af-4d201e72535b\") " pod="openstack/openstackclient" Oct 14 07:20:57 crc kubenswrapper[4870]: I1014 07:20:57.368426 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l6fdl\" (UniqueName: \"kubernetes.io/projected/7806f6b7-5364-41e9-99af-4d201e72535b-kube-api-access-l6fdl\") pod \"openstackclient\" (UID: \"7806f6b7-5364-41e9-99af-4d201e72535b\") " pod="openstack/openstackclient" Oct 14 07:20:57 crc kubenswrapper[4870]: I1014 07:20:57.369346 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7806f6b7-5364-41e9-99af-4d201e72535b-combined-ca-bundle\") pod \"openstackclient\" (UID: \"7806f6b7-5364-41e9-99af-4d201e72535b\") " pod="openstack/openstackclient" Oct 14 07:20:57 crc kubenswrapper[4870]: I1014 07:20:57.369466 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/7806f6b7-5364-41e9-99af-4d201e72535b-openstack-config\") pod \"openstackclient\" (UID: \"7806f6b7-5364-41e9-99af-4d201e72535b\") " pod="openstack/openstackclient" Oct 14 07:20:57 crc kubenswrapper[4870]: I1014 07:20:57.471341 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7806f6b7-5364-41e9-99af-4d201e72535b-combined-ca-bundle\") pod \"openstackclient\" (UID: \"7806f6b7-5364-41e9-99af-4d201e72535b\") " pod="openstack/openstackclient" Oct 14 07:20:57 crc kubenswrapper[4870]: I1014 07:20:57.471391 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/7806f6b7-5364-41e9-99af-4d201e72535b-openstack-config\") pod \"openstackclient\" (UID: \"7806f6b7-5364-41e9-99af-4d201e72535b\") " pod="openstack/openstackclient" Oct 14 07:20:57 crc kubenswrapper[4870]: I1014 07:20:57.471492 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/7806f6b7-5364-41e9-99af-4d201e72535b-openstack-config-secret\") pod \"openstackclient\" (UID: \"7806f6b7-5364-41e9-99af-4d201e72535b\") " pod="openstack/openstackclient" Oct 14 07:20:57 crc kubenswrapper[4870]: I1014 07:20:57.471519 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l6fdl\" (UniqueName: \"kubernetes.io/projected/7806f6b7-5364-41e9-99af-4d201e72535b-kube-api-access-l6fdl\") pod \"openstackclient\" (UID: \"7806f6b7-5364-41e9-99af-4d201e72535b\") " pod="openstack/openstackclient" Oct 14 07:20:57 crc kubenswrapper[4870]: I1014 07:20:57.472350 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/7806f6b7-5364-41e9-99af-4d201e72535b-openstack-config\") pod \"openstackclient\" (UID: \"7806f6b7-5364-41e9-99af-4d201e72535b\") " pod="openstack/openstackclient" Oct 14 07:20:57 crc kubenswrapper[4870]: E1014 07:20:57.473696 4870 projected.go:194] Error preparing data for projected volume kube-api-access-l6fdl for pod openstack/openstackclient: failed to fetch token: serviceaccounts "openstackclient-openstackclient" is forbidden: User "system:node:crc" cannot create resource "serviceaccounts/token" in API group "" in the namespace "openstack": no relationship found between node 'crc' and this object Oct 14 07:20:57 crc kubenswrapper[4870]: E1014 07:20:57.473815 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/7806f6b7-5364-41e9-99af-4d201e72535b-kube-api-access-l6fdl podName:7806f6b7-5364-41e9-99af-4d201e72535b nodeName:}" failed. No retries permitted until 2025-10-14 07:20:57.973788911 +0000 UTC m=+1193.671149312 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-l6fdl" (UniqueName: "kubernetes.io/projected/7806f6b7-5364-41e9-99af-4d201e72535b-kube-api-access-l6fdl") pod "openstackclient" (UID: "7806f6b7-5364-41e9-99af-4d201e72535b") : failed to fetch token: serviceaccounts "openstackclient-openstackclient" is forbidden: User "system:node:crc" cannot create resource "serviceaccounts/token" in API group "" in the namespace "openstack": no relationship found between node 'crc' and this object Oct 14 07:20:57 crc kubenswrapper[4870]: I1014 07:20:57.477862 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/7806f6b7-5364-41e9-99af-4d201e72535b-openstack-config-secret\") pod \"openstackclient\" (UID: \"7806f6b7-5364-41e9-99af-4d201e72535b\") " pod="openstack/openstackclient" Oct 14 07:20:57 crc kubenswrapper[4870]: I1014 07:20:57.478589 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7806f6b7-5364-41e9-99af-4d201e72535b-combined-ca-bundle\") pod \"openstackclient\" (UID: \"7806f6b7-5364-41e9-99af-4d201e72535b\") " pod="openstack/openstackclient" Oct 14 07:20:57 crc kubenswrapper[4870]: I1014 07:20:57.478628 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Oct 14 07:20:57 crc kubenswrapper[4870]: E1014 07:20:57.479237 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[kube-api-access-l6fdl], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/openstackclient" podUID="7806f6b7-5364-41e9-99af-4d201e72535b" Oct 14 07:20:57 crc kubenswrapper[4870]: I1014 07:20:57.485099 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Oct 14 07:20:57 crc kubenswrapper[4870]: I1014 07:20:57.531409 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Oct 14 07:20:57 crc kubenswrapper[4870]: I1014 07:20:57.532500 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 14 07:20:57 crc kubenswrapper[4870]: I1014 07:20:57.544130 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 14 07:20:57 crc kubenswrapper[4870]: I1014 07:20:57.573716 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0965b7e6-2aa0-4940-a130-324cfb08de5a-combined-ca-bundle\") pod \"openstackclient\" (UID: \"0965b7e6-2aa0-4940-a130-324cfb08de5a\") " pod="openstack/openstackclient" Oct 14 07:20:57 crc kubenswrapper[4870]: I1014 07:20:57.574011 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6zpqx\" (UniqueName: \"kubernetes.io/projected/0965b7e6-2aa0-4940-a130-324cfb08de5a-kube-api-access-6zpqx\") pod \"openstackclient\" (UID: \"0965b7e6-2aa0-4940-a130-324cfb08de5a\") " pod="openstack/openstackclient" Oct 14 07:20:57 crc kubenswrapper[4870]: I1014 07:20:57.574053 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/0965b7e6-2aa0-4940-a130-324cfb08de5a-openstack-config\") pod \"openstackclient\" (UID: \"0965b7e6-2aa0-4940-a130-324cfb08de5a\") " pod="openstack/openstackclient" Oct 14 07:20:57 crc kubenswrapper[4870]: I1014 07:20:57.574092 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/0965b7e6-2aa0-4940-a130-324cfb08de5a-openstack-config-secret\") pod \"openstackclient\" (UID: \"0965b7e6-2aa0-4940-a130-324cfb08de5a\") " pod="openstack/openstackclient" Oct 14 07:20:57 crc kubenswrapper[4870]: I1014 07:20:57.675764 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0965b7e6-2aa0-4940-a130-324cfb08de5a-combined-ca-bundle\") pod \"openstackclient\" (UID: \"0965b7e6-2aa0-4940-a130-324cfb08de5a\") " pod="openstack/openstackclient" Oct 14 07:20:57 crc kubenswrapper[4870]: I1014 07:20:57.675891 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6zpqx\" (UniqueName: \"kubernetes.io/projected/0965b7e6-2aa0-4940-a130-324cfb08de5a-kube-api-access-6zpqx\") pod \"openstackclient\" (UID: \"0965b7e6-2aa0-4940-a130-324cfb08de5a\") " pod="openstack/openstackclient" Oct 14 07:20:57 crc kubenswrapper[4870]: I1014 07:20:57.676386 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/0965b7e6-2aa0-4940-a130-324cfb08de5a-openstack-config\") pod \"openstackclient\" (UID: \"0965b7e6-2aa0-4940-a130-324cfb08de5a\") " pod="openstack/openstackclient" Oct 14 07:20:57 crc kubenswrapper[4870]: I1014 07:20:57.677381 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/0965b7e6-2aa0-4940-a130-324cfb08de5a-openstack-config\") pod \"openstackclient\" (UID: \"0965b7e6-2aa0-4940-a130-324cfb08de5a\") " pod="openstack/openstackclient" Oct 14 07:20:57 crc kubenswrapper[4870]: I1014 07:20:57.677562 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/0965b7e6-2aa0-4940-a130-324cfb08de5a-openstack-config-secret\") pod \"openstackclient\" (UID: \"0965b7e6-2aa0-4940-a130-324cfb08de5a\") " pod="openstack/openstackclient" Oct 14 07:20:57 crc kubenswrapper[4870]: I1014 07:20:57.687983 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/0965b7e6-2aa0-4940-a130-324cfb08de5a-openstack-config-secret\") pod \"openstackclient\" (UID: \"0965b7e6-2aa0-4940-a130-324cfb08de5a\") " pod="openstack/openstackclient" Oct 14 07:20:57 crc kubenswrapper[4870]: I1014 07:20:57.688249 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0965b7e6-2aa0-4940-a130-324cfb08de5a-combined-ca-bundle\") pod \"openstackclient\" (UID: \"0965b7e6-2aa0-4940-a130-324cfb08de5a\") " pod="openstack/openstackclient" Oct 14 07:20:57 crc kubenswrapper[4870]: I1014 07:20:57.691997 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6zpqx\" (UniqueName: \"kubernetes.io/projected/0965b7e6-2aa0-4940-a130-324cfb08de5a-kube-api-access-6zpqx\") pod \"openstackclient\" (UID: \"0965b7e6-2aa0-4940-a130-324cfb08de5a\") " pod="openstack/openstackclient" Oct 14 07:20:57 crc kubenswrapper[4870]: I1014 07:20:57.878096 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 14 07:20:57 crc kubenswrapper[4870]: I1014 07:20:57.983861 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l6fdl\" (UniqueName: \"kubernetes.io/projected/7806f6b7-5364-41e9-99af-4d201e72535b-kube-api-access-l6fdl\") pod \"openstackclient\" (UID: \"7806f6b7-5364-41e9-99af-4d201e72535b\") " pod="openstack/openstackclient" Oct 14 07:20:57 crc kubenswrapper[4870]: E1014 07:20:57.986403 4870 projected.go:194] Error preparing data for projected volume kube-api-access-l6fdl for pod openstack/openstackclient: failed to fetch token: serviceaccounts "openstackclient-openstackclient" is forbidden: the UID in the bound object reference (7806f6b7-5364-41e9-99af-4d201e72535b) does not match the UID in record. The object might have been deleted and then recreated Oct 14 07:20:57 crc kubenswrapper[4870]: E1014 07:20:57.986485 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/7806f6b7-5364-41e9-99af-4d201e72535b-kube-api-access-l6fdl podName:7806f6b7-5364-41e9-99af-4d201e72535b nodeName:}" failed. No retries permitted until 2025-10-14 07:20:58.986467577 +0000 UTC m=+1194.683827948 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-l6fdl" (UniqueName: "kubernetes.io/projected/7806f6b7-5364-41e9-99af-4d201e72535b-kube-api-access-l6fdl") pod "openstackclient" (UID: "7806f6b7-5364-41e9-99af-4d201e72535b") : failed to fetch token: serviceaccounts "openstackclient-openstackclient" is forbidden: the UID in the bound object reference (7806f6b7-5364-41e9-99af-4d201e72535b) does not match the UID in record. The object might have been deleted and then recreated Oct 14 07:20:58 crc kubenswrapper[4870]: I1014 07:20:58.317405 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-h587c" event={"ID":"000463bc-f3ac-40b7-90c8-32366e327a19","Type":"ContainerStarted","Data":"35a065d725ec0b4b5f843ba5a5779e23bb304a93a7feb8c329fffbfe341e4930"} Oct 14 07:20:58 crc kubenswrapper[4870]: I1014 07:20:58.317461 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 14 07:20:58 crc kubenswrapper[4870]: I1014 07:20:58.331010 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 14 07:20:58 crc kubenswrapper[4870]: I1014 07:20:58.341891 4870 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="7806f6b7-5364-41e9-99af-4d201e72535b" podUID="0965b7e6-2aa0-4940-a130-324cfb08de5a" Oct 14 07:20:58 crc kubenswrapper[4870]: I1014 07:20:58.343875 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-h587c" podStartSLOduration=3.115148199 podStartE2EDuration="40.343859267s" podCreationTimestamp="2025-10-14 07:20:18 +0000 UTC" firstStartedPulling="2025-10-14 07:20:19.294150129 +0000 UTC m=+1154.991510500" lastFinishedPulling="2025-10-14 07:20:56.522861197 +0000 UTC m=+1192.220221568" observedRunningTime="2025-10-14 07:20:58.339271065 +0000 UTC m=+1194.036631436" watchObservedRunningTime="2025-10-14 07:20:58.343859267 +0000 UTC m=+1194.041219628" Oct 14 07:20:58 crc kubenswrapper[4870]: I1014 07:20:58.350918 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-79cc757548-zxf4g" Oct 14 07:20:58 crc kubenswrapper[4870]: I1014 07:20:58.412500 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 14 07:20:58 crc kubenswrapper[4870]: I1014 07:20:58.422271 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-8446b89fb-22zh4"] Oct 14 07:20:58 crc kubenswrapper[4870]: I1014 07:20:58.422682 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-8446b89fb-22zh4" podUID="3822d444-e96e-4883-883f-d74e44b17a03" containerName="barbican-api-log" containerID="cri-o://1e7415ee401824cb497f36cc1a9431304cd42432efb6947ac813932dbbf40fbd" gracePeriod=30 Oct 14 07:20:58 crc kubenswrapper[4870]: I1014 07:20:58.423142 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-8446b89fb-22zh4" podUID="3822d444-e96e-4883-883f-d74e44b17a03" containerName="barbican-api" containerID="cri-o://ec21fa31a0e69180b4e6653848787be77a45780d3320a010fc55c4f44c1180eb" gracePeriod=30 Oct 14 07:20:58 crc kubenswrapper[4870]: W1014 07:20:58.428716 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0965b7e6_2aa0_4940_a130_324cfb08de5a.slice/crio-1ea11c2d10d0c4f1305b5868b7069805ca78f6f3b3938ddf514f5eedae74064a WatchSource:0}: Error finding container 1ea11c2d10d0c4f1305b5868b7069805ca78f6f3b3938ddf514f5eedae74064a: Status 404 returned error can't find the container with id 1ea11c2d10d0c4f1305b5868b7069805ca78f6f3b3938ddf514f5eedae74064a Oct 14 07:20:58 crc kubenswrapper[4870]: I1014 07:20:58.493281 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/7806f6b7-5364-41e9-99af-4d201e72535b-openstack-config-secret\") pod \"7806f6b7-5364-41e9-99af-4d201e72535b\" (UID: \"7806f6b7-5364-41e9-99af-4d201e72535b\") " Oct 14 07:20:58 crc kubenswrapper[4870]: I1014 07:20:58.493410 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/7806f6b7-5364-41e9-99af-4d201e72535b-openstack-config\") pod \"7806f6b7-5364-41e9-99af-4d201e72535b\" (UID: \"7806f6b7-5364-41e9-99af-4d201e72535b\") " Oct 14 07:20:58 crc kubenswrapper[4870]: I1014 07:20:58.493485 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7806f6b7-5364-41e9-99af-4d201e72535b-combined-ca-bundle\") pod \"7806f6b7-5364-41e9-99af-4d201e72535b\" (UID: \"7806f6b7-5364-41e9-99af-4d201e72535b\") " Oct 14 07:20:58 crc kubenswrapper[4870]: I1014 07:20:58.494489 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l6fdl\" (UniqueName: \"kubernetes.io/projected/7806f6b7-5364-41e9-99af-4d201e72535b-kube-api-access-l6fdl\") on node \"crc\" DevicePath \"\"" Oct 14 07:20:58 crc kubenswrapper[4870]: I1014 07:20:58.495008 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7806f6b7-5364-41e9-99af-4d201e72535b-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "7806f6b7-5364-41e9-99af-4d201e72535b" (UID: "7806f6b7-5364-41e9-99af-4d201e72535b"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:20:58 crc kubenswrapper[4870]: I1014 07:20:58.500584 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7806f6b7-5364-41e9-99af-4d201e72535b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7806f6b7-5364-41e9-99af-4d201e72535b" (UID: "7806f6b7-5364-41e9-99af-4d201e72535b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:20:58 crc kubenswrapper[4870]: I1014 07:20:58.503566 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7806f6b7-5364-41e9-99af-4d201e72535b-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "7806f6b7-5364-41e9-99af-4d201e72535b" (UID: "7806f6b7-5364-41e9-99af-4d201e72535b"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:20:58 crc kubenswrapper[4870]: I1014 07:20:58.596331 4870 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/7806f6b7-5364-41e9-99af-4d201e72535b-openstack-config\") on node \"crc\" DevicePath \"\"" Oct 14 07:20:58 crc kubenswrapper[4870]: I1014 07:20:58.596368 4870 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7806f6b7-5364-41e9-99af-4d201e72535b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 07:20:58 crc kubenswrapper[4870]: I1014 07:20:58.596379 4870 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/7806f6b7-5364-41e9-99af-4d201e72535b-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Oct 14 07:20:59 crc kubenswrapper[4870]: I1014 07:20:59.051402 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7806f6b7-5364-41e9-99af-4d201e72535b" path="/var/lib/kubelet/pods/7806f6b7-5364-41e9-99af-4d201e72535b/volumes" Oct 14 07:20:59 crc kubenswrapper[4870]: I1014 07:20:59.330072 4870 generic.go:334] "Generic (PLEG): container finished" podID="3822d444-e96e-4883-883f-d74e44b17a03" containerID="1e7415ee401824cb497f36cc1a9431304cd42432efb6947ac813932dbbf40fbd" exitCode=143 Oct 14 07:20:59 crc kubenswrapper[4870]: I1014 07:20:59.330127 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-8446b89fb-22zh4" event={"ID":"3822d444-e96e-4883-883f-d74e44b17a03","Type":"ContainerDied","Data":"1e7415ee401824cb497f36cc1a9431304cd42432efb6947ac813932dbbf40fbd"} Oct 14 07:20:59 crc kubenswrapper[4870]: I1014 07:20:59.332075 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"0965b7e6-2aa0-4940-a130-324cfb08de5a","Type":"ContainerStarted","Data":"1ea11c2d10d0c4f1305b5868b7069805ca78f6f3b3938ddf514f5eedae74064a"} Oct 14 07:20:59 crc kubenswrapper[4870]: I1014 07:20:59.332079 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 14 07:20:59 crc kubenswrapper[4870]: I1014 07:20:59.338703 4870 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="7806f6b7-5364-41e9-99af-4d201e72535b" podUID="0965b7e6-2aa0-4940-a130-324cfb08de5a" Oct 14 07:21:01 crc kubenswrapper[4870]: I1014 07:21:01.643320 4870 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-8446b89fb-22zh4" podUID="3822d444-e96e-4883-883f-d74e44b17a03" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.156:9311/healthcheck\": read tcp 10.217.0.2:48570->10.217.0.156:9311: read: connection reset by peer" Oct 14 07:21:01 crc kubenswrapper[4870]: I1014 07:21:01.643413 4870 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-8446b89fb-22zh4" podUID="3822d444-e96e-4883-883f-d74e44b17a03" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.156:9311/healthcheck\": read tcp 10.217.0.2:48560->10.217.0.156:9311: read: connection reset by peer" Oct 14 07:21:02 crc kubenswrapper[4870]: I1014 07:21:02.110467 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-8446b89fb-22zh4" Oct 14 07:21:02 crc kubenswrapper[4870]: I1014 07:21:02.290547 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3822d444-e96e-4883-883f-d74e44b17a03-logs\") pod \"3822d444-e96e-4883-883f-d74e44b17a03\" (UID: \"3822d444-e96e-4883-883f-d74e44b17a03\") " Oct 14 07:21:02 crc kubenswrapper[4870]: I1014 07:21:02.290780 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wqrzq\" (UniqueName: \"kubernetes.io/projected/3822d444-e96e-4883-883f-d74e44b17a03-kube-api-access-wqrzq\") pod \"3822d444-e96e-4883-883f-d74e44b17a03\" (UID: \"3822d444-e96e-4883-883f-d74e44b17a03\") " Oct 14 07:21:02 crc kubenswrapper[4870]: I1014 07:21:02.291021 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3822d444-e96e-4883-883f-d74e44b17a03-combined-ca-bundle\") pod \"3822d444-e96e-4883-883f-d74e44b17a03\" (UID: \"3822d444-e96e-4883-883f-d74e44b17a03\") " Oct 14 07:21:02 crc kubenswrapper[4870]: I1014 07:21:02.291050 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3822d444-e96e-4883-883f-d74e44b17a03-logs" (OuterVolumeSpecName: "logs") pod "3822d444-e96e-4883-883f-d74e44b17a03" (UID: "3822d444-e96e-4883-883f-d74e44b17a03"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:21:02 crc kubenswrapper[4870]: I1014 07:21:02.291172 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3822d444-e96e-4883-883f-d74e44b17a03-config-data\") pod \"3822d444-e96e-4883-883f-d74e44b17a03\" (UID: \"3822d444-e96e-4883-883f-d74e44b17a03\") " Oct 14 07:21:02 crc kubenswrapper[4870]: I1014 07:21:02.291358 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3822d444-e96e-4883-883f-d74e44b17a03-config-data-custom\") pod \"3822d444-e96e-4883-883f-d74e44b17a03\" (UID: \"3822d444-e96e-4883-883f-d74e44b17a03\") " Oct 14 07:21:02 crc kubenswrapper[4870]: I1014 07:21:02.292099 4870 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3822d444-e96e-4883-883f-d74e44b17a03-logs\") on node \"crc\" DevicePath \"\"" Oct 14 07:21:02 crc kubenswrapper[4870]: I1014 07:21:02.296171 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3822d444-e96e-4883-883f-d74e44b17a03-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "3822d444-e96e-4883-883f-d74e44b17a03" (UID: "3822d444-e96e-4883-883f-d74e44b17a03"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:21:02 crc kubenswrapper[4870]: I1014 07:21:02.297909 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3822d444-e96e-4883-883f-d74e44b17a03-kube-api-access-wqrzq" (OuterVolumeSpecName: "kube-api-access-wqrzq") pod "3822d444-e96e-4883-883f-d74e44b17a03" (UID: "3822d444-e96e-4883-883f-d74e44b17a03"). InnerVolumeSpecName "kube-api-access-wqrzq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:21:02 crc kubenswrapper[4870]: I1014 07:21:02.334659 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3822d444-e96e-4883-883f-d74e44b17a03-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3822d444-e96e-4883-883f-d74e44b17a03" (UID: "3822d444-e96e-4883-883f-d74e44b17a03"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:21:02 crc kubenswrapper[4870]: I1014 07:21:02.341351 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3822d444-e96e-4883-883f-d74e44b17a03-config-data" (OuterVolumeSpecName: "config-data") pod "3822d444-e96e-4883-883f-d74e44b17a03" (UID: "3822d444-e96e-4883-883f-d74e44b17a03"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:21:02 crc kubenswrapper[4870]: I1014 07:21:02.366907 4870 generic.go:334] "Generic (PLEG): container finished" podID="3822d444-e96e-4883-883f-d74e44b17a03" containerID="ec21fa31a0e69180b4e6653848787be77a45780d3320a010fc55c4f44c1180eb" exitCode=0 Oct 14 07:21:02 crc kubenswrapper[4870]: I1014 07:21:02.366991 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-8446b89fb-22zh4" event={"ID":"3822d444-e96e-4883-883f-d74e44b17a03","Type":"ContainerDied","Data":"ec21fa31a0e69180b4e6653848787be77a45780d3320a010fc55c4f44c1180eb"} Oct 14 07:21:02 crc kubenswrapper[4870]: I1014 07:21:02.367063 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-8446b89fb-22zh4" event={"ID":"3822d444-e96e-4883-883f-d74e44b17a03","Type":"ContainerDied","Data":"7b0d4385c2f1ebcf7d14f05f88d8a6a7602a4af150408e2e9868df5c97819b2d"} Oct 14 07:21:02 crc kubenswrapper[4870]: I1014 07:21:02.367084 4870 scope.go:117] "RemoveContainer" containerID="ec21fa31a0e69180b4e6653848787be77a45780d3320a010fc55c4f44c1180eb" Oct 14 07:21:02 crc kubenswrapper[4870]: I1014 07:21:02.367263 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-8446b89fb-22zh4" Oct 14 07:21:02 crc kubenswrapper[4870]: I1014 07:21:02.368401 4870 generic.go:334] "Generic (PLEG): container finished" podID="000463bc-f3ac-40b7-90c8-32366e327a19" containerID="35a065d725ec0b4b5f843ba5a5779e23bb304a93a7feb8c329fffbfe341e4930" exitCode=0 Oct 14 07:21:02 crc kubenswrapper[4870]: I1014 07:21:02.368486 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-h587c" event={"ID":"000463bc-f3ac-40b7-90c8-32366e327a19","Type":"ContainerDied","Data":"35a065d725ec0b4b5f843ba5a5779e23bb304a93a7feb8c329fffbfe341e4930"} Oct 14 07:21:02 crc kubenswrapper[4870]: I1014 07:21:02.393581 4870 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3822d444-e96e-4883-883f-d74e44b17a03-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 14 07:21:02 crc kubenswrapper[4870]: I1014 07:21:02.393613 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wqrzq\" (UniqueName: \"kubernetes.io/projected/3822d444-e96e-4883-883f-d74e44b17a03-kube-api-access-wqrzq\") on node \"crc\" DevicePath \"\"" Oct 14 07:21:02 crc kubenswrapper[4870]: I1014 07:21:02.393625 4870 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3822d444-e96e-4883-883f-d74e44b17a03-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 07:21:02 crc kubenswrapper[4870]: I1014 07:21:02.393634 4870 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3822d444-e96e-4883-883f-d74e44b17a03-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 07:21:02 crc kubenswrapper[4870]: I1014 07:21:02.436721 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-8446b89fb-22zh4"] Oct 14 07:21:02 crc kubenswrapper[4870]: I1014 07:21:02.446866 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-8446b89fb-22zh4"] Oct 14 07:21:02 crc kubenswrapper[4870]: I1014 07:21:02.453942 4870 scope.go:117] "RemoveContainer" containerID="1e7415ee401824cb497f36cc1a9431304cd42432efb6947ac813932dbbf40fbd" Oct 14 07:21:02 crc kubenswrapper[4870]: I1014 07:21:02.475348 4870 scope.go:117] "RemoveContainer" containerID="ec21fa31a0e69180b4e6653848787be77a45780d3320a010fc55c4f44c1180eb" Oct 14 07:21:02 crc kubenswrapper[4870]: E1014 07:21:02.479417 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ec21fa31a0e69180b4e6653848787be77a45780d3320a010fc55c4f44c1180eb\": container with ID starting with ec21fa31a0e69180b4e6653848787be77a45780d3320a010fc55c4f44c1180eb not found: ID does not exist" containerID="ec21fa31a0e69180b4e6653848787be77a45780d3320a010fc55c4f44c1180eb" Oct 14 07:21:02 crc kubenswrapper[4870]: I1014 07:21:02.479471 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ec21fa31a0e69180b4e6653848787be77a45780d3320a010fc55c4f44c1180eb"} err="failed to get container status \"ec21fa31a0e69180b4e6653848787be77a45780d3320a010fc55c4f44c1180eb\": rpc error: code = NotFound desc = could not find container \"ec21fa31a0e69180b4e6653848787be77a45780d3320a010fc55c4f44c1180eb\": container with ID starting with ec21fa31a0e69180b4e6653848787be77a45780d3320a010fc55c4f44c1180eb not found: ID does not exist" Oct 14 07:21:02 crc kubenswrapper[4870]: I1014 07:21:02.486830 4870 scope.go:117] "RemoveContainer" containerID="1e7415ee401824cb497f36cc1a9431304cd42432efb6947ac813932dbbf40fbd" Oct 14 07:21:02 crc kubenswrapper[4870]: E1014 07:21:02.487989 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1e7415ee401824cb497f36cc1a9431304cd42432efb6947ac813932dbbf40fbd\": container with ID starting with 1e7415ee401824cb497f36cc1a9431304cd42432efb6947ac813932dbbf40fbd not found: ID does not exist" containerID="1e7415ee401824cb497f36cc1a9431304cd42432efb6947ac813932dbbf40fbd" Oct 14 07:21:02 crc kubenswrapper[4870]: I1014 07:21:02.488041 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1e7415ee401824cb497f36cc1a9431304cd42432efb6947ac813932dbbf40fbd"} err="failed to get container status \"1e7415ee401824cb497f36cc1a9431304cd42432efb6947ac813932dbbf40fbd\": rpc error: code = NotFound desc = could not find container \"1e7415ee401824cb497f36cc1a9431304cd42432efb6947ac813932dbbf40fbd\": container with ID starting with 1e7415ee401824cb497f36cc1a9431304cd42432efb6947ac813932dbbf40fbd not found: ID does not exist" Oct 14 07:21:02 crc kubenswrapper[4870]: I1014 07:21:02.507191 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-5485b69bc9-ftmmq"] Oct 14 07:21:02 crc kubenswrapper[4870]: E1014 07:21:02.507554 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3822d444-e96e-4883-883f-d74e44b17a03" containerName="barbican-api" Oct 14 07:21:02 crc kubenswrapper[4870]: I1014 07:21:02.507569 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="3822d444-e96e-4883-883f-d74e44b17a03" containerName="barbican-api" Oct 14 07:21:02 crc kubenswrapper[4870]: E1014 07:21:02.507589 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3822d444-e96e-4883-883f-d74e44b17a03" containerName="barbican-api-log" Oct 14 07:21:02 crc kubenswrapper[4870]: I1014 07:21:02.507596 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="3822d444-e96e-4883-883f-d74e44b17a03" containerName="barbican-api-log" Oct 14 07:21:02 crc kubenswrapper[4870]: I1014 07:21:02.507782 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="3822d444-e96e-4883-883f-d74e44b17a03" containerName="barbican-api-log" Oct 14 07:21:02 crc kubenswrapper[4870]: I1014 07:21:02.507800 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="3822d444-e96e-4883-883f-d74e44b17a03" containerName="barbican-api" Oct 14 07:21:02 crc kubenswrapper[4870]: I1014 07:21:02.508675 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-5485b69bc9-ftmmq" Oct 14 07:21:02 crc kubenswrapper[4870]: I1014 07:21:02.517109 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Oct 14 07:21:02 crc kubenswrapper[4870]: I1014 07:21:02.517174 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Oct 14 07:21:02 crc kubenswrapper[4870]: I1014 07:21:02.518146 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Oct 14 07:21:02 crc kubenswrapper[4870]: I1014 07:21:02.528557 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-5485b69bc9-ftmmq"] Oct 14 07:21:02 crc kubenswrapper[4870]: I1014 07:21:02.698130 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3927fb78-4f30-4371-8fbb-62fd276c0460-config-data\") pod \"swift-proxy-5485b69bc9-ftmmq\" (UID: \"3927fb78-4f30-4371-8fbb-62fd276c0460\") " pod="openstack/swift-proxy-5485b69bc9-ftmmq" Oct 14 07:21:02 crc kubenswrapper[4870]: I1014 07:21:02.698181 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3927fb78-4f30-4371-8fbb-62fd276c0460-log-httpd\") pod \"swift-proxy-5485b69bc9-ftmmq\" (UID: \"3927fb78-4f30-4371-8fbb-62fd276c0460\") " pod="openstack/swift-proxy-5485b69bc9-ftmmq" Oct 14 07:21:02 crc kubenswrapper[4870]: I1014 07:21:02.698202 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3927fb78-4f30-4371-8fbb-62fd276c0460-internal-tls-certs\") pod \"swift-proxy-5485b69bc9-ftmmq\" (UID: \"3927fb78-4f30-4371-8fbb-62fd276c0460\") " pod="openstack/swift-proxy-5485b69bc9-ftmmq" Oct 14 07:21:02 crc kubenswrapper[4870]: I1014 07:21:02.698244 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9m4qf\" (UniqueName: \"kubernetes.io/projected/3927fb78-4f30-4371-8fbb-62fd276c0460-kube-api-access-9m4qf\") pod \"swift-proxy-5485b69bc9-ftmmq\" (UID: \"3927fb78-4f30-4371-8fbb-62fd276c0460\") " pod="openstack/swift-proxy-5485b69bc9-ftmmq" Oct 14 07:21:02 crc kubenswrapper[4870]: I1014 07:21:02.698285 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3927fb78-4f30-4371-8fbb-62fd276c0460-public-tls-certs\") pod \"swift-proxy-5485b69bc9-ftmmq\" (UID: \"3927fb78-4f30-4371-8fbb-62fd276c0460\") " pod="openstack/swift-proxy-5485b69bc9-ftmmq" Oct 14 07:21:02 crc kubenswrapper[4870]: I1014 07:21:02.698304 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3927fb78-4f30-4371-8fbb-62fd276c0460-combined-ca-bundle\") pod \"swift-proxy-5485b69bc9-ftmmq\" (UID: \"3927fb78-4f30-4371-8fbb-62fd276c0460\") " pod="openstack/swift-proxy-5485b69bc9-ftmmq" Oct 14 07:21:02 crc kubenswrapper[4870]: I1014 07:21:02.698329 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/3927fb78-4f30-4371-8fbb-62fd276c0460-etc-swift\") pod \"swift-proxy-5485b69bc9-ftmmq\" (UID: \"3927fb78-4f30-4371-8fbb-62fd276c0460\") " pod="openstack/swift-proxy-5485b69bc9-ftmmq" Oct 14 07:21:02 crc kubenswrapper[4870]: I1014 07:21:02.698352 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3927fb78-4f30-4371-8fbb-62fd276c0460-run-httpd\") pod \"swift-proxy-5485b69bc9-ftmmq\" (UID: \"3927fb78-4f30-4371-8fbb-62fd276c0460\") " pod="openstack/swift-proxy-5485b69bc9-ftmmq" Oct 14 07:21:02 crc kubenswrapper[4870]: I1014 07:21:02.800415 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3927fb78-4f30-4371-8fbb-62fd276c0460-config-data\") pod \"swift-proxy-5485b69bc9-ftmmq\" (UID: \"3927fb78-4f30-4371-8fbb-62fd276c0460\") " pod="openstack/swift-proxy-5485b69bc9-ftmmq" Oct 14 07:21:02 crc kubenswrapper[4870]: I1014 07:21:02.800552 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3927fb78-4f30-4371-8fbb-62fd276c0460-log-httpd\") pod \"swift-proxy-5485b69bc9-ftmmq\" (UID: \"3927fb78-4f30-4371-8fbb-62fd276c0460\") " pod="openstack/swift-proxy-5485b69bc9-ftmmq" Oct 14 07:21:02 crc kubenswrapper[4870]: I1014 07:21:02.800600 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3927fb78-4f30-4371-8fbb-62fd276c0460-internal-tls-certs\") pod \"swift-proxy-5485b69bc9-ftmmq\" (UID: \"3927fb78-4f30-4371-8fbb-62fd276c0460\") " pod="openstack/swift-proxy-5485b69bc9-ftmmq" Oct 14 07:21:02 crc kubenswrapper[4870]: I1014 07:21:02.800707 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9m4qf\" (UniqueName: \"kubernetes.io/projected/3927fb78-4f30-4371-8fbb-62fd276c0460-kube-api-access-9m4qf\") pod \"swift-proxy-5485b69bc9-ftmmq\" (UID: \"3927fb78-4f30-4371-8fbb-62fd276c0460\") " pod="openstack/swift-proxy-5485b69bc9-ftmmq" Oct 14 07:21:02 crc kubenswrapper[4870]: I1014 07:21:02.800760 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3927fb78-4f30-4371-8fbb-62fd276c0460-public-tls-certs\") pod \"swift-proxy-5485b69bc9-ftmmq\" (UID: \"3927fb78-4f30-4371-8fbb-62fd276c0460\") " pod="openstack/swift-proxy-5485b69bc9-ftmmq" Oct 14 07:21:02 crc kubenswrapper[4870]: I1014 07:21:02.800794 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3927fb78-4f30-4371-8fbb-62fd276c0460-combined-ca-bundle\") pod \"swift-proxy-5485b69bc9-ftmmq\" (UID: \"3927fb78-4f30-4371-8fbb-62fd276c0460\") " pod="openstack/swift-proxy-5485b69bc9-ftmmq" Oct 14 07:21:02 crc kubenswrapper[4870]: I1014 07:21:02.800856 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/3927fb78-4f30-4371-8fbb-62fd276c0460-etc-swift\") pod \"swift-proxy-5485b69bc9-ftmmq\" (UID: \"3927fb78-4f30-4371-8fbb-62fd276c0460\") " pod="openstack/swift-proxy-5485b69bc9-ftmmq" Oct 14 07:21:02 crc kubenswrapper[4870]: I1014 07:21:02.800903 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3927fb78-4f30-4371-8fbb-62fd276c0460-run-httpd\") pod \"swift-proxy-5485b69bc9-ftmmq\" (UID: \"3927fb78-4f30-4371-8fbb-62fd276c0460\") " pod="openstack/swift-proxy-5485b69bc9-ftmmq" Oct 14 07:21:02 crc kubenswrapper[4870]: I1014 07:21:02.801706 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3927fb78-4f30-4371-8fbb-62fd276c0460-run-httpd\") pod \"swift-proxy-5485b69bc9-ftmmq\" (UID: \"3927fb78-4f30-4371-8fbb-62fd276c0460\") " pod="openstack/swift-proxy-5485b69bc9-ftmmq" Oct 14 07:21:02 crc kubenswrapper[4870]: I1014 07:21:02.801983 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3927fb78-4f30-4371-8fbb-62fd276c0460-log-httpd\") pod \"swift-proxy-5485b69bc9-ftmmq\" (UID: \"3927fb78-4f30-4371-8fbb-62fd276c0460\") " pod="openstack/swift-proxy-5485b69bc9-ftmmq" Oct 14 07:21:02 crc kubenswrapper[4870]: I1014 07:21:02.805534 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3927fb78-4f30-4371-8fbb-62fd276c0460-config-data\") pod \"swift-proxy-5485b69bc9-ftmmq\" (UID: \"3927fb78-4f30-4371-8fbb-62fd276c0460\") " pod="openstack/swift-proxy-5485b69bc9-ftmmq" Oct 14 07:21:02 crc kubenswrapper[4870]: I1014 07:21:02.805572 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3927fb78-4f30-4371-8fbb-62fd276c0460-internal-tls-certs\") pod \"swift-proxy-5485b69bc9-ftmmq\" (UID: \"3927fb78-4f30-4371-8fbb-62fd276c0460\") " pod="openstack/swift-proxy-5485b69bc9-ftmmq" Oct 14 07:21:02 crc kubenswrapper[4870]: I1014 07:21:02.805547 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/3927fb78-4f30-4371-8fbb-62fd276c0460-etc-swift\") pod \"swift-proxy-5485b69bc9-ftmmq\" (UID: \"3927fb78-4f30-4371-8fbb-62fd276c0460\") " pod="openstack/swift-proxy-5485b69bc9-ftmmq" Oct 14 07:21:02 crc kubenswrapper[4870]: I1014 07:21:02.805846 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3927fb78-4f30-4371-8fbb-62fd276c0460-public-tls-certs\") pod \"swift-proxy-5485b69bc9-ftmmq\" (UID: \"3927fb78-4f30-4371-8fbb-62fd276c0460\") " pod="openstack/swift-proxy-5485b69bc9-ftmmq" Oct 14 07:21:02 crc kubenswrapper[4870]: I1014 07:21:02.810362 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3927fb78-4f30-4371-8fbb-62fd276c0460-combined-ca-bundle\") pod \"swift-proxy-5485b69bc9-ftmmq\" (UID: \"3927fb78-4f30-4371-8fbb-62fd276c0460\") " pod="openstack/swift-proxy-5485b69bc9-ftmmq" Oct 14 07:21:02 crc kubenswrapper[4870]: I1014 07:21:02.838353 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9m4qf\" (UniqueName: \"kubernetes.io/projected/3927fb78-4f30-4371-8fbb-62fd276c0460-kube-api-access-9m4qf\") pod \"swift-proxy-5485b69bc9-ftmmq\" (UID: \"3927fb78-4f30-4371-8fbb-62fd276c0460\") " pod="openstack/swift-proxy-5485b69bc9-ftmmq" Oct 14 07:21:02 crc kubenswrapper[4870]: I1014 07:21:02.849933 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-5485b69bc9-ftmmq" Oct 14 07:21:03 crc kubenswrapper[4870]: I1014 07:21:03.062618 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3822d444-e96e-4883-883f-d74e44b17a03" path="/var/lib/kubelet/pods/3822d444-e96e-4883-883f-d74e44b17a03/volumes" Oct 14 07:21:03 crc kubenswrapper[4870]: I1014 07:21:03.201173 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 14 07:21:03 crc kubenswrapper[4870]: I1014 07:21:03.201593 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a148cb0b-68ec-4e80-90c9-4937ac3e2823" containerName="proxy-httpd" containerID="cri-o://94f837271ef58fef81c8ae6563f0c235e72d9d2177b233a60e2b3a9d8dffd104" gracePeriod=30 Oct 14 07:21:03 crc kubenswrapper[4870]: I1014 07:21:03.201699 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a148cb0b-68ec-4e80-90c9-4937ac3e2823" containerName="sg-core" containerID="cri-o://44d88789123eae3dc981cd0bab8f55c3e851b41100c3d03e03d237349a7d49cb" gracePeriod=30 Oct 14 07:21:03 crc kubenswrapper[4870]: I1014 07:21:03.201877 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a148cb0b-68ec-4e80-90c9-4937ac3e2823" containerName="ceilometer-notification-agent" containerID="cri-o://47b79d1a4d8054d0c304a894efd7d793f07b4275c55e04e12239669c544d33f6" gracePeriod=30 Oct 14 07:21:03 crc kubenswrapper[4870]: I1014 07:21:03.203082 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a148cb0b-68ec-4e80-90c9-4937ac3e2823" containerName="ceilometer-central-agent" containerID="cri-o://99aa20a03e300b003a0916165f520ef17e1468aea37b7d719db40a6850e99f66" gracePeriod=30 Oct 14 07:21:03 crc kubenswrapper[4870]: I1014 07:21:03.215776 4870 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="a148cb0b-68ec-4e80-90c9-4937ac3e2823" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.0.157:3000/\": EOF" Oct 14 07:21:03 crc kubenswrapper[4870]: I1014 07:21:03.396705 4870 generic.go:334] "Generic (PLEG): container finished" podID="a148cb0b-68ec-4e80-90c9-4937ac3e2823" containerID="44d88789123eae3dc981cd0bab8f55c3e851b41100c3d03e03d237349a7d49cb" exitCode=2 Oct 14 07:21:03 crc kubenswrapper[4870]: I1014 07:21:03.396763 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a148cb0b-68ec-4e80-90c9-4937ac3e2823","Type":"ContainerDied","Data":"44d88789123eae3dc981cd0bab8f55c3e851b41100c3d03e03d237349a7d49cb"} Oct 14 07:21:04 crc kubenswrapper[4870]: I1014 07:21:04.414230 4870 generic.go:334] "Generic (PLEG): container finished" podID="a148cb0b-68ec-4e80-90c9-4937ac3e2823" containerID="94f837271ef58fef81c8ae6563f0c235e72d9d2177b233a60e2b3a9d8dffd104" exitCode=0 Oct 14 07:21:04 crc kubenswrapper[4870]: I1014 07:21:04.414639 4870 generic.go:334] "Generic (PLEG): container finished" podID="a148cb0b-68ec-4e80-90c9-4937ac3e2823" containerID="99aa20a03e300b003a0916165f520ef17e1468aea37b7d719db40a6850e99f66" exitCode=0 Oct 14 07:21:04 crc kubenswrapper[4870]: I1014 07:21:04.414663 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a148cb0b-68ec-4e80-90c9-4937ac3e2823","Type":"ContainerDied","Data":"94f837271ef58fef81c8ae6563f0c235e72d9d2177b233a60e2b3a9d8dffd104"} Oct 14 07:21:04 crc kubenswrapper[4870]: I1014 07:21:04.414689 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a148cb0b-68ec-4e80-90c9-4937ac3e2823","Type":"ContainerDied","Data":"99aa20a03e300b003a0916165f520ef17e1468aea37b7d719db40a6850e99f66"} Oct 14 07:21:04 crc kubenswrapper[4870]: I1014 07:21:04.941331 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-hjj9z"] Oct 14 07:21:04 crc kubenswrapper[4870]: I1014 07:21:04.942712 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-hjj9z" Oct 14 07:21:04 crc kubenswrapper[4870]: I1014 07:21:04.952346 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-hjj9z"] Oct 14 07:21:05 crc kubenswrapper[4870]: I1014 07:21:05.044858 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dslpw\" (UniqueName: \"kubernetes.io/projected/52ebf07b-1f8d-4005-b8bc-14e250c921d4-kube-api-access-dslpw\") pod \"nova-api-db-create-hjj9z\" (UID: \"52ebf07b-1f8d-4005-b8bc-14e250c921d4\") " pod="openstack/nova-api-db-create-hjj9z" Oct 14 07:21:05 crc kubenswrapper[4870]: I1014 07:21:05.048374 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-snx2f"] Oct 14 07:21:05 crc kubenswrapper[4870]: I1014 07:21:05.049413 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-snx2f" Oct 14 07:21:05 crc kubenswrapper[4870]: I1014 07:21:05.056215 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-snx2f"] Oct 14 07:21:05 crc kubenswrapper[4870]: I1014 07:21:05.135329 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-29752"] Oct 14 07:21:05 crc kubenswrapper[4870]: I1014 07:21:05.136802 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-29752" Oct 14 07:21:05 crc kubenswrapper[4870]: I1014 07:21:05.150661 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-29752"] Oct 14 07:21:05 crc kubenswrapper[4870]: I1014 07:21:05.150961 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dslpw\" (UniqueName: \"kubernetes.io/projected/52ebf07b-1f8d-4005-b8bc-14e250c921d4-kube-api-access-dslpw\") pod \"nova-api-db-create-hjj9z\" (UID: \"52ebf07b-1f8d-4005-b8bc-14e250c921d4\") " pod="openstack/nova-api-db-create-hjj9z" Oct 14 07:21:05 crc kubenswrapper[4870]: I1014 07:21:05.151055 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6sjvq\" (UniqueName: \"kubernetes.io/projected/3ea2c25b-4ae5-473c-9480-ec8444bbee56-kube-api-access-6sjvq\") pod \"nova-cell0-db-create-snx2f\" (UID: \"3ea2c25b-4ae5-473c-9480-ec8444bbee56\") " pod="openstack/nova-cell0-db-create-snx2f" Oct 14 07:21:05 crc kubenswrapper[4870]: I1014 07:21:05.151129 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-248jv\" (UniqueName: \"kubernetes.io/projected/bf415c18-d584-4cf1-ae86-3bebdc74cad1-kube-api-access-248jv\") pod \"nova-cell1-db-create-29752\" (UID: \"bf415c18-d584-4cf1-ae86-3bebdc74cad1\") " pod="openstack/nova-cell1-db-create-29752" Oct 14 07:21:05 crc kubenswrapper[4870]: I1014 07:21:05.169327 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dslpw\" (UniqueName: \"kubernetes.io/projected/52ebf07b-1f8d-4005-b8bc-14e250c921d4-kube-api-access-dslpw\") pod \"nova-api-db-create-hjj9z\" (UID: \"52ebf07b-1f8d-4005-b8bc-14e250c921d4\") " pod="openstack/nova-api-db-create-hjj9z" Oct 14 07:21:05 crc kubenswrapper[4870]: I1014 07:21:05.253461 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6sjvq\" (UniqueName: \"kubernetes.io/projected/3ea2c25b-4ae5-473c-9480-ec8444bbee56-kube-api-access-6sjvq\") pod \"nova-cell0-db-create-snx2f\" (UID: \"3ea2c25b-4ae5-473c-9480-ec8444bbee56\") " pod="openstack/nova-cell0-db-create-snx2f" Oct 14 07:21:05 crc kubenswrapper[4870]: I1014 07:21:05.253606 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-248jv\" (UniqueName: \"kubernetes.io/projected/bf415c18-d584-4cf1-ae86-3bebdc74cad1-kube-api-access-248jv\") pod \"nova-cell1-db-create-29752\" (UID: \"bf415c18-d584-4cf1-ae86-3bebdc74cad1\") " pod="openstack/nova-cell1-db-create-29752" Oct 14 07:21:05 crc kubenswrapper[4870]: I1014 07:21:05.269689 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6sjvq\" (UniqueName: \"kubernetes.io/projected/3ea2c25b-4ae5-473c-9480-ec8444bbee56-kube-api-access-6sjvq\") pod \"nova-cell0-db-create-snx2f\" (UID: \"3ea2c25b-4ae5-473c-9480-ec8444bbee56\") " pod="openstack/nova-cell0-db-create-snx2f" Oct 14 07:21:05 crc kubenswrapper[4870]: I1014 07:21:05.270852 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-248jv\" (UniqueName: \"kubernetes.io/projected/bf415c18-d584-4cf1-ae86-3bebdc74cad1-kube-api-access-248jv\") pod \"nova-cell1-db-create-29752\" (UID: \"bf415c18-d584-4cf1-ae86-3bebdc74cad1\") " pod="openstack/nova-cell1-db-create-29752" Oct 14 07:21:05 crc kubenswrapper[4870]: I1014 07:21:05.289600 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-hjj9z" Oct 14 07:21:05 crc kubenswrapper[4870]: I1014 07:21:05.378668 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-snx2f" Oct 14 07:21:05 crc kubenswrapper[4870]: I1014 07:21:05.463531 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-29752" Oct 14 07:21:06 crc kubenswrapper[4870]: I1014 07:21:06.448989 4870 generic.go:334] "Generic (PLEG): container finished" podID="a148cb0b-68ec-4e80-90c9-4937ac3e2823" containerID="47b79d1a4d8054d0c304a894efd7d793f07b4275c55e04e12239669c544d33f6" exitCode=0 Oct 14 07:21:06 crc kubenswrapper[4870]: I1014 07:21:06.449036 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a148cb0b-68ec-4e80-90c9-4937ac3e2823","Type":"ContainerDied","Data":"47b79d1a4d8054d0c304a894efd7d793f07b4275c55e04e12239669c544d33f6"} Oct 14 07:21:07 crc kubenswrapper[4870]: I1014 07:21:07.780760 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-h587c" Oct 14 07:21:07 crc kubenswrapper[4870]: I1014 07:21:07.799136 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/000463bc-f3ac-40b7-90c8-32366e327a19-db-sync-config-data\") pod \"000463bc-f3ac-40b7-90c8-32366e327a19\" (UID: \"000463bc-f3ac-40b7-90c8-32366e327a19\") " Oct 14 07:21:07 crc kubenswrapper[4870]: I1014 07:21:07.799457 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/000463bc-f3ac-40b7-90c8-32366e327a19-combined-ca-bundle\") pod \"000463bc-f3ac-40b7-90c8-32366e327a19\" (UID: \"000463bc-f3ac-40b7-90c8-32366e327a19\") " Oct 14 07:21:07 crc kubenswrapper[4870]: I1014 07:21:07.799503 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/000463bc-f3ac-40b7-90c8-32366e327a19-etc-machine-id\") pod \"000463bc-f3ac-40b7-90c8-32366e327a19\" (UID: \"000463bc-f3ac-40b7-90c8-32366e327a19\") " Oct 14 07:21:07 crc kubenswrapper[4870]: I1014 07:21:07.799527 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/000463bc-f3ac-40b7-90c8-32366e327a19-scripts\") pod \"000463bc-f3ac-40b7-90c8-32366e327a19\" (UID: \"000463bc-f3ac-40b7-90c8-32366e327a19\") " Oct 14 07:21:07 crc kubenswrapper[4870]: I1014 07:21:07.799609 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-56zkt\" (UniqueName: \"kubernetes.io/projected/000463bc-f3ac-40b7-90c8-32366e327a19-kube-api-access-56zkt\") pod \"000463bc-f3ac-40b7-90c8-32366e327a19\" (UID: \"000463bc-f3ac-40b7-90c8-32366e327a19\") " Oct 14 07:21:07 crc kubenswrapper[4870]: I1014 07:21:07.799658 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/000463bc-f3ac-40b7-90c8-32366e327a19-config-data\") pod \"000463bc-f3ac-40b7-90c8-32366e327a19\" (UID: \"000463bc-f3ac-40b7-90c8-32366e327a19\") " Oct 14 07:21:07 crc kubenswrapper[4870]: I1014 07:21:07.800801 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/000463bc-f3ac-40b7-90c8-32366e327a19-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "000463bc-f3ac-40b7-90c8-32366e327a19" (UID: "000463bc-f3ac-40b7-90c8-32366e327a19"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 14 07:21:07 crc kubenswrapper[4870]: I1014 07:21:07.811048 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/000463bc-f3ac-40b7-90c8-32366e327a19-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "000463bc-f3ac-40b7-90c8-32366e327a19" (UID: "000463bc-f3ac-40b7-90c8-32366e327a19"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:21:07 crc kubenswrapper[4870]: I1014 07:21:07.811067 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/000463bc-f3ac-40b7-90c8-32366e327a19-kube-api-access-56zkt" (OuterVolumeSpecName: "kube-api-access-56zkt") pod "000463bc-f3ac-40b7-90c8-32366e327a19" (UID: "000463bc-f3ac-40b7-90c8-32366e327a19"). InnerVolumeSpecName "kube-api-access-56zkt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:21:07 crc kubenswrapper[4870]: I1014 07:21:07.816626 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/000463bc-f3ac-40b7-90c8-32366e327a19-scripts" (OuterVolumeSpecName: "scripts") pod "000463bc-f3ac-40b7-90c8-32366e327a19" (UID: "000463bc-f3ac-40b7-90c8-32366e327a19"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:21:07 crc kubenswrapper[4870]: I1014 07:21:07.859544 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/000463bc-f3ac-40b7-90c8-32366e327a19-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "000463bc-f3ac-40b7-90c8-32366e327a19" (UID: "000463bc-f3ac-40b7-90c8-32366e327a19"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:21:07 crc kubenswrapper[4870]: I1014 07:21:07.887696 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/000463bc-f3ac-40b7-90c8-32366e327a19-config-data" (OuterVolumeSpecName: "config-data") pod "000463bc-f3ac-40b7-90c8-32366e327a19" (UID: "000463bc-f3ac-40b7-90c8-32366e327a19"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:21:07 crc kubenswrapper[4870]: I1014 07:21:07.902319 4870 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/000463bc-f3ac-40b7-90c8-32366e327a19-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 07:21:07 crc kubenswrapper[4870]: I1014 07:21:07.902351 4870 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/000463bc-f3ac-40b7-90c8-32366e327a19-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 07:21:07 crc kubenswrapper[4870]: I1014 07:21:07.902360 4870 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/000463bc-f3ac-40b7-90c8-32366e327a19-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 14 07:21:07 crc kubenswrapper[4870]: I1014 07:21:07.902369 4870 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/000463bc-f3ac-40b7-90c8-32366e327a19-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 07:21:07 crc kubenswrapper[4870]: I1014 07:21:07.902377 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-56zkt\" (UniqueName: \"kubernetes.io/projected/000463bc-f3ac-40b7-90c8-32366e327a19-kube-api-access-56zkt\") on node \"crc\" DevicePath \"\"" Oct 14 07:21:07 crc kubenswrapper[4870]: I1014 07:21:07.902388 4870 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/000463bc-f3ac-40b7-90c8-32366e327a19-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 07:21:08 crc kubenswrapper[4870]: I1014 07:21:08.056260 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 14 07:21:08 crc kubenswrapper[4870]: I1014 07:21:08.206520 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c55nd\" (UniqueName: \"kubernetes.io/projected/a148cb0b-68ec-4e80-90c9-4937ac3e2823-kube-api-access-c55nd\") pod \"a148cb0b-68ec-4e80-90c9-4937ac3e2823\" (UID: \"a148cb0b-68ec-4e80-90c9-4937ac3e2823\") " Oct 14 07:21:08 crc kubenswrapper[4870]: I1014 07:21:08.206577 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a148cb0b-68ec-4e80-90c9-4937ac3e2823-config-data\") pod \"a148cb0b-68ec-4e80-90c9-4937ac3e2823\" (UID: \"a148cb0b-68ec-4e80-90c9-4937ac3e2823\") " Oct 14 07:21:08 crc kubenswrapper[4870]: I1014 07:21:08.206691 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a148cb0b-68ec-4e80-90c9-4937ac3e2823-scripts\") pod \"a148cb0b-68ec-4e80-90c9-4937ac3e2823\" (UID: \"a148cb0b-68ec-4e80-90c9-4937ac3e2823\") " Oct 14 07:21:08 crc kubenswrapper[4870]: I1014 07:21:08.206742 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a148cb0b-68ec-4e80-90c9-4937ac3e2823-run-httpd\") pod \"a148cb0b-68ec-4e80-90c9-4937ac3e2823\" (UID: \"a148cb0b-68ec-4e80-90c9-4937ac3e2823\") " Oct 14 07:21:08 crc kubenswrapper[4870]: I1014 07:21:08.206863 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a148cb0b-68ec-4e80-90c9-4937ac3e2823-combined-ca-bundle\") pod \"a148cb0b-68ec-4e80-90c9-4937ac3e2823\" (UID: \"a148cb0b-68ec-4e80-90c9-4937ac3e2823\") " Oct 14 07:21:08 crc kubenswrapper[4870]: I1014 07:21:08.206889 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a148cb0b-68ec-4e80-90c9-4937ac3e2823-sg-core-conf-yaml\") pod \"a148cb0b-68ec-4e80-90c9-4937ac3e2823\" (UID: \"a148cb0b-68ec-4e80-90c9-4937ac3e2823\") " Oct 14 07:21:08 crc kubenswrapper[4870]: I1014 07:21:08.206920 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a148cb0b-68ec-4e80-90c9-4937ac3e2823-log-httpd\") pod \"a148cb0b-68ec-4e80-90c9-4937ac3e2823\" (UID: \"a148cb0b-68ec-4e80-90c9-4937ac3e2823\") " Oct 14 07:21:08 crc kubenswrapper[4870]: I1014 07:21:08.207978 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a148cb0b-68ec-4e80-90c9-4937ac3e2823-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "a148cb0b-68ec-4e80-90c9-4937ac3e2823" (UID: "a148cb0b-68ec-4e80-90c9-4937ac3e2823"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:21:08 crc kubenswrapper[4870]: I1014 07:21:08.208429 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a148cb0b-68ec-4e80-90c9-4937ac3e2823-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "a148cb0b-68ec-4e80-90c9-4937ac3e2823" (UID: "a148cb0b-68ec-4e80-90c9-4937ac3e2823"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:21:08 crc kubenswrapper[4870]: I1014 07:21:08.212373 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a148cb0b-68ec-4e80-90c9-4937ac3e2823-kube-api-access-c55nd" (OuterVolumeSpecName: "kube-api-access-c55nd") pod "a148cb0b-68ec-4e80-90c9-4937ac3e2823" (UID: "a148cb0b-68ec-4e80-90c9-4937ac3e2823"). InnerVolumeSpecName "kube-api-access-c55nd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:21:08 crc kubenswrapper[4870]: I1014 07:21:08.212835 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a148cb0b-68ec-4e80-90c9-4937ac3e2823-scripts" (OuterVolumeSpecName: "scripts") pod "a148cb0b-68ec-4e80-90c9-4937ac3e2823" (UID: "a148cb0b-68ec-4e80-90c9-4937ac3e2823"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:21:08 crc kubenswrapper[4870]: I1014 07:21:08.238148 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a148cb0b-68ec-4e80-90c9-4937ac3e2823-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "a148cb0b-68ec-4e80-90c9-4937ac3e2823" (UID: "a148cb0b-68ec-4e80-90c9-4937ac3e2823"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:21:08 crc kubenswrapper[4870]: I1014 07:21:08.283554 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a148cb0b-68ec-4e80-90c9-4937ac3e2823-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a148cb0b-68ec-4e80-90c9-4937ac3e2823" (UID: "a148cb0b-68ec-4e80-90c9-4937ac3e2823"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:21:08 crc kubenswrapper[4870]: I1014 07:21:08.309070 4870 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a148cb0b-68ec-4e80-90c9-4937ac3e2823-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 14 07:21:08 crc kubenswrapper[4870]: I1014 07:21:08.309112 4870 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a148cb0b-68ec-4e80-90c9-4937ac3e2823-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 07:21:08 crc kubenswrapper[4870]: I1014 07:21:08.309126 4870 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a148cb0b-68ec-4e80-90c9-4937ac3e2823-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 14 07:21:08 crc kubenswrapper[4870]: I1014 07:21:08.309141 4870 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a148cb0b-68ec-4e80-90c9-4937ac3e2823-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 14 07:21:08 crc kubenswrapper[4870]: I1014 07:21:08.309154 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c55nd\" (UniqueName: \"kubernetes.io/projected/a148cb0b-68ec-4e80-90c9-4937ac3e2823-kube-api-access-c55nd\") on node \"crc\" DevicePath \"\"" Oct 14 07:21:08 crc kubenswrapper[4870]: I1014 07:21:08.309167 4870 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a148cb0b-68ec-4e80-90c9-4937ac3e2823-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 07:21:08 crc kubenswrapper[4870]: I1014 07:21:08.318075 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a148cb0b-68ec-4e80-90c9-4937ac3e2823-config-data" (OuterVolumeSpecName: "config-data") pod "a148cb0b-68ec-4e80-90c9-4937ac3e2823" (UID: "a148cb0b-68ec-4e80-90c9-4937ac3e2823"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:21:08 crc kubenswrapper[4870]: I1014 07:21:08.391107 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-5485b69bc9-ftmmq"] Oct 14 07:21:08 crc kubenswrapper[4870]: I1014 07:21:08.410889 4870 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a148cb0b-68ec-4e80-90c9-4937ac3e2823-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 07:21:08 crc kubenswrapper[4870]: I1014 07:21:08.416635 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-snx2f"] Oct 14 07:21:08 crc kubenswrapper[4870]: W1014 07:21:08.432652 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbf415c18_d584_4cf1_ae86_3bebdc74cad1.slice/crio-5c953af1359978ee56dd076d26829a8cd8698204c204b3cb112538cb378d406f WatchSource:0}: Error finding container 5c953af1359978ee56dd076d26829a8cd8698204c204b3cb112538cb378d406f: Status 404 returned error can't find the container with id 5c953af1359978ee56dd076d26829a8cd8698204c204b3cb112538cb378d406f Oct 14 07:21:08 crc kubenswrapper[4870]: I1014 07:21:08.435084 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-29752"] Oct 14 07:21:08 crc kubenswrapper[4870]: I1014 07:21:08.443691 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-hjj9z"] Oct 14 07:21:08 crc kubenswrapper[4870]: W1014 07:21:08.454627 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod52ebf07b_1f8d_4005_b8bc_14e250c921d4.slice/crio-7ccfa3b2cb94cb3c6f508623520f00d23df84eb8625198d5ea014296740c1fdb WatchSource:0}: Error finding container 7ccfa3b2cb94cb3c6f508623520f00d23df84eb8625198d5ea014296740c1fdb: Status 404 returned error can't find the container with id 7ccfa3b2cb94cb3c6f508623520f00d23df84eb8625198d5ea014296740c1fdb Oct 14 07:21:08 crc kubenswrapper[4870]: I1014 07:21:08.480993 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-h587c" Oct 14 07:21:08 crc kubenswrapper[4870]: I1014 07:21:08.481819 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-h587c" event={"ID":"000463bc-f3ac-40b7-90c8-32366e327a19","Type":"ContainerDied","Data":"7c2eae11e0d67fb773baaf6186f3925313a3290f9b8bf53581a83059721ef23e"} Oct 14 07:21:08 crc kubenswrapper[4870]: I1014 07:21:08.481912 4870 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7c2eae11e0d67fb773baaf6186f3925313a3290f9b8bf53581a83059721ef23e" Oct 14 07:21:08 crc kubenswrapper[4870]: I1014 07:21:08.484076 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-5485b69bc9-ftmmq" event={"ID":"3927fb78-4f30-4371-8fbb-62fd276c0460","Type":"ContainerStarted","Data":"428db2cf2ded2f12b61c1ef1b5e8692a727c90d13e0dd04dc1e44a5dff104a1c"} Oct 14 07:21:08 crc kubenswrapper[4870]: I1014 07:21:08.488909 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-hjj9z" event={"ID":"52ebf07b-1f8d-4005-b8bc-14e250c921d4","Type":"ContainerStarted","Data":"7ccfa3b2cb94cb3c6f508623520f00d23df84eb8625198d5ea014296740c1fdb"} Oct 14 07:21:08 crc kubenswrapper[4870]: I1014 07:21:08.492586 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a148cb0b-68ec-4e80-90c9-4937ac3e2823","Type":"ContainerDied","Data":"bdbf5a7358035c14b4b3d0dfa179494ba80a24fdcd2325b7cd0d218eab1b4828"} Oct 14 07:21:08 crc kubenswrapper[4870]: I1014 07:21:08.492610 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 14 07:21:08 crc kubenswrapper[4870]: I1014 07:21:08.492634 4870 scope.go:117] "RemoveContainer" containerID="94f837271ef58fef81c8ae6563f0c235e72d9d2177b233a60e2b3a9d8dffd104" Oct 14 07:21:08 crc kubenswrapper[4870]: I1014 07:21:08.501593 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-snx2f" event={"ID":"3ea2c25b-4ae5-473c-9480-ec8444bbee56","Type":"ContainerStarted","Data":"79f643ac455da16662a323680280f334a2933997e2fce29fc4d04415a16d8838"} Oct 14 07:21:08 crc kubenswrapper[4870]: I1014 07:21:08.505071 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"0965b7e6-2aa0-4940-a130-324cfb08de5a","Type":"ContainerStarted","Data":"044ea035cd6094777207ef8f180d32297801ac2dd26be47e987ae4b7318523d6"} Oct 14 07:21:08 crc kubenswrapper[4870]: I1014 07:21:08.510687 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-29752" event={"ID":"bf415c18-d584-4cf1-ae86-3bebdc74cad1","Type":"ContainerStarted","Data":"5c953af1359978ee56dd076d26829a8cd8698204c204b3cb112538cb378d406f"} Oct 14 07:21:08 crc kubenswrapper[4870]: I1014 07:21:08.521715 4870 scope.go:117] "RemoveContainer" containerID="44d88789123eae3dc981cd0bab8f55c3e851b41100c3d03e03d237349a7d49cb" Oct 14 07:21:08 crc kubenswrapper[4870]: I1014 07:21:08.530254 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=2.153703888 podStartE2EDuration="11.530223309s" podCreationTimestamp="2025-10-14 07:20:57 +0000 UTC" firstStartedPulling="2025-10-14 07:20:58.430659213 +0000 UTC m=+1194.128019584" lastFinishedPulling="2025-10-14 07:21:07.807178634 +0000 UTC m=+1203.504539005" observedRunningTime="2025-10-14 07:21:08.519943013 +0000 UTC m=+1204.217303384" watchObservedRunningTime="2025-10-14 07:21:08.530223309 +0000 UTC m=+1204.227583680" Oct 14 07:21:08 crc kubenswrapper[4870]: I1014 07:21:08.568208 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 14 07:21:08 crc kubenswrapper[4870]: I1014 07:21:08.572904 4870 scope.go:117] "RemoveContainer" containerID="47b79d1a4d8054d0c304a894efd7d793f07b4275c55e04e12239669c544d33f6" Oct 14 07:21:08 crc kubenswrapper[4870]: I1014 07:21:08.577239 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 14 07:21:08 crc kubenswrapper[4870]: I1014 07:21:08.589577 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 14 07:21:08 crc kubenswrapper[4870]: E1014 07:21:08.591959 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a148cb0b-68ec-4e80-90c9-4937ac3e2823" containerName="ceilometer-central-agent" Oct 14 07:21:08 crc kubenswrapper[4870]: I1014 07:21:08.591991 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="a148cb0b-68ec-4e80-90c9-4937ac3e2823" containerName="ceilometer-central-agent" Oct 14 07:21:08 crc kubenswrapper[4870]: E1014 07:21:08.592008 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="000463bc-f3ac-40b7-90c8-32366e327a19" containerName="cinder-db-sync" Oct 14 07:21:08 crc kubenswrapper[4870]: I1014 07:21:08.592015 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="000463bc-f3ac-40b7-90c8-32366e327a19" containerName="cinder-db-sync" Oct 14 07:21:08 crc kubenswrapper[4870]: E1014 07:21:08.592028 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a148cb0b-68ec-4e80-90c9-4937ac3e2823" containerName="proxy-httpd" Oct 14 07:21:08 crc kubenswrapper[4870]: I1014 07:21:08.592035 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="a148cb0b-68ec-4e80-90c9-4937ac3e2823" containerName="proxy-httpd" Oct 14 07:21:08 crc kubenswrapper[4870]: E1014 07:21:08.592049 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a148cb0b-68ec-4e80-90c9-4937ac3e2823" containerName="sg-core" Oct 14 07:21:08 crc kubenswrapper[4870]: I1014 07:21:08.592055 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="a148cb0b-68ec-4e80-90c9-4937ac3e2823" containerName="sg-core" Oct 14 07:21:08 crc kubenswrapper[4870]: E1014 07:21:08.592071 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a148cb0b-68ec-4e80-90c9-4937ac3e2823" containerName="ceilometer-notification-agent" Oct 14 07:21:08 crc kubenswrapper[4870]: I1014 07:21:08.592077 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="a148cb0b-68ec-4e80-90c9-4937ac3e2823" containerName="ceilometer-notification-agent" Oct 14 07:21:08 crc kubenswrapper[4870]: I1014 07:21:08.592278 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="000463bc-f3ac-40b7-90c8-32366e327a19" containerName="cinder-db-sync" Oct 14 07:21:08 crc kubenswrapper[4870]: I1014 07:21:08.592295 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="a148cb0b-68ec-4e80-90c9-4937ac3e2823" containerName="ceilometer-notification-agent" Oct 14 07:21:08 crc kubenswrapper[4870]: I1014 07:21:08.592305 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="a148cb0b-68ec-4e80-90c9-4937ac3e2823" containerName="sg-core" Oct 14 07:21:08 crc kubenswrapper[4870]: I1014 07:21:08.592318 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="a148cb0b-68ec-4e80-90c9-4937ac3e2823" containerName="proxy-httpd" Oct 14 07:21:08 crc kubenswrapper[4870]: I1014 07:21:08.592327 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="a148cb0b-68ec-4e80-90c9-4937ac3e2823" containerName="ceilometer-central-agent" Oct 14 07:21:08 crc kubenswrapper[4870]: I1014 07:21:08.594100 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 14 07:21:08 crc kubenswrapper[4870]: I1014 07:21:08.596633 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 14 07:21:08 crc kubenswrapper[4870]: I1014 07:21:08.597057 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 14 07:21:08 crc kubenswrapper[4870]: I1014 07:21:08.617664 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 14 07:21:08 crc kubenswrapper[4870]: I1014 07:21:08.663096 4870 scope.go:117] "RemoveContainer" containerID="99aa20a03e300b003a0916165f520ef17e1468aea37b7d719db40a6850e99f66" Oct 14 07:21:08 crc kubenswrapper[4870]: I1014 07:21:08.729049 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2ad27e3d-56a2-46e2-a6b0-28b03c5564e8-run-httpd\") pod \"ceilometer-0\" (UID: \"2ad27e3d-56a2-46e2-a6b0-28b03c5564e8\") " pod="openstack/ceilometer-0" Oct 14 07:21:08 crc kubenswrapper[4870]: I1014 07:21:08.729139 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2ad27e3d-56a2-46e2-a6b0-28b03c5564e8-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2ad27e3d-56a2-46e2-a6b0-28b03c5564e8\") " pod="openstack/ceilometer-0" Oct 14 07:21:08 crc kubenswrapper[4870]: I1014 07:21:08.729158 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2ad27e3d-56a2-46e2-a6b0-28b03c5564e8-scripts\") pod \"ceilometer-0\" (UID: \"2ad27e3d-56a2-46e2-a6b0-28b03c5564e8\") " pod="openstack/ceilometer-0" Oct 14 07:21:08 crc kubenswrapper[4870]: I1014 07:21:08.729181 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n652l\" (UniqueName: \"kubernetes.io/projected/2ad27e3d-56a2-46e2-a6b0-28b03c5564e8-kube-api-access-n652l\") pod \"ceilometer-0\" (UID: \"2ad27e3d-56a2-46e2-a6b0-28b03c5564e8\") " pod="openstack/ceilometer-0" Oct 14 07:21:08 crc kubenswrapper[4870]: I1014 07:21:08.729205 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2ad27e3d-56a2-46e2-a6b0-28b03c5564e8-config-data\") pod \"ceilometer-0\" (UID: \"2ad27e3d-56a2-46e2-a6b0-28b03c5564e8\") " pod="openstack/ceilometer-0" Oct 14 07:21:08 crc kubenswrapper[4870]: I1014 07:21:08.729414 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2ad27e3d-56a2-46e2-a6b0-28b03c5564e8-log-httpd\") pod \"ceilometer-0\" (UID: \"2ad27e3d-56a2-46e2-a6b0-28b03c5564e8\") " pod="openstack/ceilometer-0" Oct 14 07:21:08 crc kubenswrapper[4870]: I1014 07:21:08.729482 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ad27e3d-56a2-46e2-a6b0-28b03c5564e8-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2ad27e3d-56a2-46e2-a6b0-28b03c5564e8\") " pod="openstack/ceilometer-0" Oct 14 07:21:08 crc kubenswrapper[4870]: I1014 07:21:08.831500 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2ad27e3d-56a2-46e2-a6b0-28b03c5564e8-run-httpd\") pod \"ceilometer-0\" (UID: \"2ad27e3d-56a2-46e2-a6b0-28b03c5564e8\") " pod="openstack/ceilometer-0" Oct 14 07:21:08 crc kubenswrapper[4870]: I1014 07:21:08.831663 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2ad27e3d-56a2-46e2-a6b0-28b03c5564e8-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2ad27e3d-56a2-46e2-a6b0-28b03c5564e8\") " pod="openstack/ceilometer-0" Oct 14 07:21:08 crc kubenswrapper[4870]: I1014 07:21:08.831697 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2ad27e3d-56a2-46e2-a6b0-28b03c5564e8-scripts\") pod \"ceilometer-0\" (UID: \"2ad27e3d-56a2-46e2-a6b0-28b03c5564e8\") " pod="openstack/ceilometer-0" Oct 14 07:21:08 crc kubenswrapper[4870]: I1014 07:21:08.831739 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n652l\" (UniqueName: \"kubernetes.io/projected/2ad27e3d-56a2-46e2-a6b0-28b03c5564e8-kube-api-access-n652l\") pod \"ceilometer-0\" (UID: \"2ad27e3d-56a2-46e2-a6b0-28b03c5564e8\") " pod="openstack/ceilometer-0" Oct 14 07:21:08 crc kubenswrapper[4870]: I1014 07:21:08.831822 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2ad27e3d-56a2-46e2-a6b0-28b03c5564e8-config-data\") pod \"ceilometer-0\" (UID: \"2ad27e3d-56a2-46e2-a6b0-28b03c5564e8\") " pod="openstack/ceilometer-0" Oct 14 07:21:08 crc kubenswrapper[4870]: I1014 07:21:08.831882 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2ad27e3d-56a2-46e2-a6b0-28b03c5564e8-log-httpd\") pod \"ceilometer-0\" (UID: \"2ad27e3d-56a2-46e2-a6b0-28b03c5564e8\") " pod="openstack/ceilometer-0" Oct 14 07:21:08 crc kubenswrapper[4870]: I1014 07:21:08.831915 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ad27e3d-56a2-46e2-a6b0-28b03c5564e8-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2ad27e3d-56a2-46e2-a6b0-28b03c5564e8\") " pod="openstack/ceilometer-0" Oct 14 07:21:08 crc kubenswrapper[4870]: I1014 07:21:08.831975 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2ad27e3d-56a2-46e2-a6b0-28b03c5564e8-run-httpd\") pod \"ceilometer-0\" (UID: \"2ad27e3d-56a2-46e2-a6b0-28b03c5564e8\") " pod="openstack/ceilometer-0" Oct 14 07:21:08 crc kubenswrapper[4870]: I1014 07:21:08.833147 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2ad27e3d-56a2-46e2-a6b0-28b03c5564e8-log-httpd\") pod \"ceilometer-0\" (UID: \"2ad27e3d-56a2-46e2-a6b0-28b03c5564e8\") " pod="openstack/ceilometer-0" Oct 14 07:21:08 crc kubenswrapper[4870]: I1014 07:21:08.840044 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2ad27e3d-56a2-46e2-a6b0-28b03c5564e8-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2ad27e3d-56a2-46e2-a6b0-28b03c5564e8\") " pod="openstack/ceilometer-0" Oct 14 07:21:08 crc kubenswrapper[4870]: I1014 07:21:08.840147 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2ad27e3d-56a2-46e2-a6b0-28b03c5564e8-scripts\") pod \"ceilometer-0\" (UID: \"2ad27e3d-56a2-46e2-a6b0-28b03c5564e8\") " pod="openstack/ceilometer-0" Oct 14 07:21:08 crc kubenswrapper[4870]: I1014 07:21:08.840469 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ad27e3d-56a2-46e2-a6b0-28b03c5564e8-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2ad27e3d-56a2-46e2-a6b0-28b03c5564e8\") " pod="openstack/ceilometer-0" Oct 14 07:21:08 crc kubenswrapper[4870]: I1014 07:21:08.840709 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2ad27e3d-56a2-46e2-a6b0-28b03c5564e8-config-data\") pod \"ceilometer-0\" (UID: \"2ad27e3d-56a2-46e2-a6b0-28b03c5564e8\") " pod="openstack/ceilometer-0" Oct 14 07:21:08 crc kubenswrapper[4870]: I1014 07:21:08.849974 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n652l\" (UniqueName: \"kubernetes.io/projected/2ad27e3d-56a2-46e2-a6b0-28b03c5564e8-kube-api-access-n652l\") pod \"ceilometer-0\" (UID: \"2ad27e3d-56a2-46e2-a6b0-28b03c5564e8\") " pod="openstack/ceilometer-0" Oct 14 07:21:08 crc kubenswrapper[4870]: I1014 07:21:08.920185 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 14 07:21:09 crc kubenswrapper[4870]: I1014 07:21:09.051299 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a148cb0b-68ec-4e80-90c9-4937ac3e2823" path="/var/lib/kubelet/pods/a148cb0b-68ec-4e80-90c9-4937ac3e2823/volumes" Oct 14 07:21:09 crc kubenswrapper[4870]: I1014 07:21:09.052397 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Oct 14 07:21:09 crc kubenswrapper[4870]: I1014 07:21:09.059351 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 14 07:21:09 crc kubenswrapper[4870]: I1014 07:21:09.064326 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Oct 14 07:21:09 crc kubenswrapper[4870]: I1014 07:21:09.065569 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 14 07:21:09 crc kubenswrapper[4870]: I1014 07:21:09.070986 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Oct 14 07:21:09 crc kubenswrapper[4870]: I1014 07:21:09.071172 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Oct 14 07:21:09 crc kubenswrapper[4870]: I1014 07:21:09.071323 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-ccbhx" Oct 14 07:21:09 crc kubenswrapper[4870]: I1014 07:21:09.140654 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5b7d94f5b5-4qfsd"] Oct 14 07:21:09 crc kubenswrapper[4870]: I1014 07:21:09.162665 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f04fb303-8db7-4d2c-a864-817494a7222b-config-data\") pod \"cinder-scheduler-0\" (UID: \"f04fb303-8db7-4d2c-a864-817494a7222b\") " pod="openstack/cinder-scheduler-0" Oct 14 07:21:09 crc kubenswrapper[4870]: I1014 07:21:09.162762 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f04fb303-8db7-4d2c-a864-817494a7222b-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"f04fb303-8db7-4d2c-a864-817494a7222b\") " pod="openstack/cinder-scheduler-0" Oct 14 07:21:09 crc kubenswrapper[4870]: I1014 07:21:09.162846 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f04fb303-8db7-4d2c-a864-817494a7222b-scripts\") pod \"cinder-scheduler-0\" (UID: \"f04fb303-8db7-4d2c-a864-817494a7222b\") " pod="openstack/cinder-scheduler-0" Oct 14 07:21:09 crc kubenswrapper[4870]: I1014 07:21:09.162878 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f04fb303-8db7-4d2c-a864-817494a7222b-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"f04fb303-8db7-4d2c-a864-817494a7222b\") " pod="openstack/cinder-scheduler-0" Oct 14 07:21:09 crc kubenswrapper[4870]: I1014 07:21:09.162985 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2ctzq\" (UniqueName: \"kubernetes.io/projected/f04fb303-8db7-4d2c-a864-817494a7222b-kube-api-access-2ctzq\") pod \"cinder-scheduler-0\" (UID: \"f04fb303-8db7-4d2c-a864-817494a7222b\") " pod="openstack/cinder-scheduler-0" Oct 14 07:21:09 crc kubenswrapper[4870]: I1014 07:21:09.163022 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f04fb303-8db7-4d2c-a864-817494a7222b-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"f04fb303-8db7-4d2c-a864-817494a7222b\") " pod="openstack/cinder-scheduler-0" Oct 14 07:21:09 crc kubenswrapper[4870]: I1014 07:21:09.191264 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b7d94f5b5-4qfsd" Oct 14 07:21:09 crc kubenswrapper[4870]: I1014 07:21:09.229880 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5b7d94f5b5-4qfsd"] Oct 14 07:21:09 crc kubenswrapper[4870]: I1014 07:21:09.268109 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f04fb303-8db7-4d2c-a864-817494a7222b-config-data\") pod \"cinder-scheduler-0\" (UID: \"f04fb303-8db7-4d2c-a864-817494a7222b\") " pod="openstack/cinder-scheduler-0" Oct 14 07:21:09 crc kubenswrapper[4870]: I1014 07:21:09.268368 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f04fb303-8db7-4d2c-a864-817494a7222b-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"f04fb303-8db7-4d2c-a864-817494a7222b\") " pod="openstack/cinder-scheduler-0" Oct 14 07:21:09 crc kubenswrapper[4870]: I1014 07:21:09.268484 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tdcpf\" (UniqueName: \"kubernetes.io/projected/0dba2f3b-4ddd-4a1f-8593-1f98dc4479b1-kube-api-access-tdcpf\") pod \"dnsmasq-dns-5b7d94f5b5-4qfsd\" (UID: \"0dba2f3b-4ddd-4a1f-8593-1f98dc4479b1\") " pod="openstack/dnsmasq-dns-5b7d94f5b5-4qfsd" Oct 14 07:21:09 crc kubenswrapper[4870]: I1014 07:21:09.268562 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0dba2f3b-4ddd-4a1f-8593-1f98dc4479b1-ovsdbserver-sb\") pod \"dnsmasq-dns-5b7d94f5b5-4qfsd\" (UID: \"0dba2f3b-4ddd-4a1f-8593-1f98dc4479b1\") " pod="openstack/dnsmasq-dns-5b7d94f5b5-4qfsd" Oct 14 07:21:09 crc kubenswrapper[4870]: I1014 07:21:09.268636 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0dba2f3b-4ddd-4a1f-8593-1f98dc4479b1-config\") pod \"dnsmasq-dns-5b7d94f5b5-4qfsd\" (UID: \"0dba2f3b-4ddd-4a1f-8593-1f98dc4479b1\") " pod="openstack/dnsmasq-dns-5b7d94f5b5-4qfsd" Oct 14 07:21:09 crc kubenswrapper[4870]: I1014 07:21:09.268727 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f04fb303-8db7-4d2c-a864-817494a7222b-scripts\") pod \"cinder-scheduler-0\" (UID: \"f04fb303-8db7-4d2c-a864-817494a7222b\") " pod="openstack/cinder-scheduler-0" Oct 14 07:21:09 crc kubenswrapper[4870]: I1014 07:21:09.268799 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f04fb303-8db7-4d2c-a864-817494a7222b-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"f04fb303-8db7-4d2c-a864-817494a7222b\") " pod="openstack/cinder-scheduler-0" Oct 14 07:21:09 crc kubenswrapper[4870]: I1014 07:21:09.268886 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0dba2f3b-4ddd-4a1f-8593-1f98dc4479b1-dns-svc\") pod \"dnsmasq-dns-5b7d94f5b5-4qfsd\" (UID: \"0dba2f3b-4ddd-4a1f-8593-1f98dc4479b1\") " pod="openstack/dnsmasq-dns-5b7d94f5b5-4qfsd" Oct 14 07:21:09 crc kubenswrapper[4870]: I1014 07:21:09.268957 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0dba2f3b-4ddd-4a1f-8593-1f98dc4479b1-ovsdbserver-nb\") pod \"dnsmasq-dns-5b7d94f5b5-4qfsd\" (UID: \"0dba2f3b-4ddd-4a1f-8593-1f98dc4479b1\") " pod="openstack/dnsmasq-dns-5b7d94f5b5-4qfsd" Oct 14 07:21:09 crc kubenswrapper[4870]: I1014 07:21:09.269029 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0dba2f3b-4ddd-4a1f-8593-1f98dc4479b1-dns-swift-storage-0\") pod \"dnsmasq-dns-5b7d94f5b5-4qfsd\" (UID: \"0dba2f3b-4ddd-4a1f-8593-1f98dc4479b1\") " pod="openstack/dnsmasq-dns-5b7d94f5b5-4qfsd" Oct 14 07:21:09 crc kubenswrapper[4870]: I1014 07:21:09.269116 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2ctzq\" (UniqueName: \"kubernetes.io/projected/f04fb303-8db7-4d2c-a864-817494a7222b-kube-api-access-2ctzq\") pod \"cinder-scheduler-0\" (UID: \"f04fb303-8db7-4d2c-a864-817494a7222b\") " pod="openstack/cinder-scheduler-0" Oct 14 07:21:09 crc kubenswrapper[4870]: I1014 07:21:09.269192 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f04fb303-8db7-4d2c-a864-817494a7222b-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"f04fb303-8db7-4d2c-a864-817494a7222b\") " pod="openstack/cinder-scheduler-0" Oct 14 07:21:09 crc kubenswrapper[4870]: I1014 07:21:09.274061 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f04fb303-8db7-4d2c-a864-817494a7222b-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"f04fb303-8db7-4d2c-a864-817494a7222b\") " pod="openstack/cinder-scheduler-0" Oct 14 07:21:09 crc kubenswrapper[4870]: I1014 07:21:09.274318 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f04fb303-8db7-4d2c-a864-817494a7222b-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"f04fb303-8db7-4d2c-a864-817494a7222b\") " pod="openstack/cinder-scheduler-0" Oct 14 07:21:09 crc kubenswrapper[4870]: I1014 07:21:09.282450 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f04fb303-8db7-4d2c-a864-817494a7222b-config-data\") pod \"cinder-scheduler-0\" (UID: \"f04fb303-8db7-4d2c-a864-817494a7222b\") " pod="openstack/cinder-scheduler-0" Oct 14 07:21:09 crc kubenswrapper[4870]: I1014 07:21:09.282880 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f04fb303-8db7-4d2c-a864-817494a7222b-scripts\") pod \"cinder-scheduler-0\" (UID: \"f04fb303-8db7-4d2c-a864-817494a7222b\") " pod="openstack/cinder-scheduler-0" Oct 14 07:21:09 crc kubenswrapper[4870]: I1014 07:21:09.289051 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f04fb303-8db7-4d2c-a864-817494a7222b-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"f04fb303-8db7-4d2c-a864-817494a7222b\") " pod="openstack/cinder-scheduler-0" Oct 14 07:21:09 crc kubenswrapper[4870]: I1014 07:21:09.297857 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2ctzq\" (UniqueName: \"kubernetes.io/projected/f04fb303-8db7-4d2c-a864-817494a7222b-kube-api-access-2ctzq\") pod \"cinder-scheduler-0\" (UID: \"f04fb303-8db7-4d2c-a864-817494a7222b\") " pod="openstack/cinder-scheduler-0" Oct 14 07:21:09 crc kubenswrapper[4870]: I1014 07:21:09.300582 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Oct 14 07:21:09 crc kubenswrapper[4870]: I1014 07:21:09.302155 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 14 07:21:09 crc kubenswrapper[4870]: I1014 07:21:09.304726 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Oct 14 07:21:09 crc kubenswrapper[4870]: I1014 07:21:09.310966 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 14 07:21:09 crc kubenswrapper[4870]: I1014 07:21:09.371462 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5p44w\" (UniqueName: \"kubernetes.io/projected/aad458d7-8d76-421e-8036-22510ec4f99f-kube-api-access-5p44w\") pod \"cinder-api-0\" (UID: \"aad458d7-8d76-421e-8036-22510ec4f99f\") " pod="openstack/cinder-api-0" Oct 14 07:21:09 crc kubenswrapper[4870]: I1014 07:21:09.371537 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aad458d7-8d76-421e-8036-22510ec4f99f-scripts\") pod \"cinder-api-0\" (UID: \"aad458d7-8d76-421e-8036-22510ec4f99f\") " pod="openstack/cinder-api-0" Oct 14 07:21:09 crc kubenswrapper[4870]: I1014 07:21:09.371582 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aad458d7-8d76-421e-8036-22510ec4f99f-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"aad458d7-8d76-421e-8036-22510ec4f99f\") " pod="openstack/cinder-api-0" Oct 14 07:21:09 crc kubenswrapper[4870]: I1014 07:21:09.371618 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tdcpf\" (UniqueName: \"kubernetes.io/projected/0dba2f3b-4ddd-4a1f-8593-1f98dc4479b1-kube-api-access-tdcpf\") pod \"dnsmasq-dns-5b7d94f5b5-4qfsd\" (UID: \"0dba2f3b-4ddd-4a1f-8593-1f98dc4479b1\") " pod="openstack/dnsmasq-dns-5b7d94f5b5-4qfsd" Oct 14 07:21:09 crc kubenswrapper[4870]: I1014 07:21:09.371638 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/aad458d7-8d76-421e-8036-22510ec4f99f-logs\") pod \"cinder-api-0\" (UID: \"aad458d7-8d76-421e-8036-22510ec4f99f\") " pod="openstack/cinder-api-0" Oct 14 07:21:09 crc kubenswrapper[4870]: I1014 07:21:09.371658 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0dba2f3b-4ddd-4a1f-8593-1f98dc4479b1-ovsdbserver-sb\") pod \"dnsmasq-dns-5b7d94f5b5-4qfsd\" (UID: \"0dba2f3b-4ddd-4a1f-8593-1f98dc4479b1\") " pod="openstack/dnsmasq-dns-5b7d94f5b5-4qfsd" Oct 14 07:21:09 crc kubenswrapper[4870]: I1014 07:21:09.371676 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0dba2f3b-4ddd-4a1f-8593-1f98dc4479b1-config\") pod \"dnsmasq-dns-5b7d94f5b5-4qfsd\" (UID: \"0dba2f3b-4ddd-4a1f-8593-1f98dc4479b1\") " pod="openstack/dnsmasq-dns-5b7d94f5b5-4qfsd" Oct 14 07:21:09 crc kubenswrapper[4870]: I1014 07:21:09.371717 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/aad458d7-8d76-421e-8036-22510ec4f99f-etc-machine-id\") pod \"cinder-api-0\" (UID: \"aad458d7-8d76-421e-8036-22510ec4f99f\") " pod="openstack/cinder-api-0" Oct 14 07:21:09 crc kubenswrapper[4870]: I1014 07:21:09.371737 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0dba2f3b-4ddd-4a1f-8593-1f98dc4479b1-dns-svc\") pod \"dnsmasq-dns-5b7d94f5b5-4qfsd\" (UID: \"0dba2f3b-4ddd-4a1f-8593-1f98dc4479b1\") " pod="openstack/dnsmasq-dns-5b7d94f5b5-4qfsd" Oct 14 07:21:09 crc kubenswrapper[4870]: I1014 07:21:09.371758 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0dba2f3b-4ddd-4a1f-8593-1f98dc4479b1-ovsdbserver-nb\") pod \"dnsmasq-dns-5b7d94f5b5-4qfsd\" (UID: \"0dba2f3b-4ddd-4a1f-8593-1f98dc4479b1\") " pod="openstack/dnsmasq-dns-5b7d94f5b5-4qfsd" Oct 14 07:21:09 crc kubenswrapper[4870]: I1014 07:21:09.371790 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0dba2f3b-4ddd-4a1f-8593-1f98dc4479b1-dns-swift-storage-0\") pod \"dnsmasq-dns-5b7d94f5b5-4qfsd\" (UID: \"0dba2f3b-4ddd-4a1f-8593-1f98dc4479b1\") " pod="openstack/dnsmasq-dns-5b7d94f5b5-4qfsd" Oct 14 07:21:09 crc kubenswrapper[4870]: I1014 07:21:09.371808 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aad458d7-8d76-421e-8036-22510ec4f99f-config-data\") pod \"cinder-api-0\" (UID: \"aad458d7-8d76-421e-8036-22510ec4f99f\") " pod="openstack/cinder-api-0" Oct 14 07:21:09 crc kubenswrapper[4870]: I1014 07:21:09.371834 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/aad458d7-8d76-421e-8036-22510ec4f99f-config-data-custom\") pod \"cinder-api-0\" (UID: \"aad458d7-8d76-421e-8036-22510ec4f99f\") " pod="openstack/cinder-api-0" Oct 14 07:21:09 crc kubenswrapper[4870]: I1014 07:21:09.373320 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0dba2f3b-4ddd-4a1f-8593-1f98dc4479b1-ovsdbserver-sb\") pod \"dnsmasq-dns-5b7d94f5b5-4qfsd\" (UID: \"0dba2f3b-4ddd-4a1f-8593-1f98dc4479b1\") " pod="openstack/dnsmasq-dns-5b7d94f5b5-4qfsd" Oct 14 07:21:09 crc kubenswrapper[4870]: I1014 07:21:09.373873 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0dba2f3b-4ddd-4a1f-8593-1f98dc4479b1-config\") pod \"dnsmasq-dns-5b7d94f5b5-4qfsd\" (UID: \"0dba2f3b-4ddd-4a1f-8593-1f98dc4479b1\") " pod="openstack/dnsmasq-dns-5b7d94f5b5-4qfsd" Oct 14 07:21:09 crc kubenswrapper[4870]: I1014 07:21:09.374399 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0dba2f3b-4ddd-4a1f-8593-1f98dc4479b1-dns-svc\") pod \"dnsmasq-dns-5b7d94f5b5-4qfsd\" (UID: \"0dba2f3b-4ddd-4a1f-8593-1f98dc4479b1\") " pod="openstack/dnsmasq-dns-5b7d94f5b5-4qfsd" Oct 14 07:21:09 crc kubenswrapper[4870]: I1014 07:21:09.374909 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0dba2f3b-4ddd-4a1f-8593-1f98dc4479b1-ovsdbserver-nb\") pod \"dnsmasq-dns-5b7d94f5b5-4qfsd\" (UID: \"0dba2f3b-4ddd-4a1f-8593-1f98dc4479b1\") " pod="openstack/dnsmasq-dns-5b7d94f5b5-4qfsd" Oct 14 07:21:09 crc kubenswrapper[4870]: I1014 07:21:09.375478 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0dba2f3b-4ddd-4a1f-8593-1f98dc4479b1-dns-swift-storage-0\") pod \"dnsmasq-dns-5b7d94f5b5-4qfsd\" (UID: \"0dba2f3b-4ddd-4a1f-8593-1f98dc4479b1\") " pod="openstack/dnsmasq-dns-5b7d94f5b5-4qfsd" Oct 14 07:21:09 crc kubenswrapper[4870]: I1014 07:21:09.400059 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tdcpf\" (UniqueName: \"kubernetes.io/projected/0dba2f3b-4ddd-4a1f-8593-1f98dc4479b1-kube-api-access-tdcpf\") pod \"dnsmasq-dns-5b7d94f5b5-4qfsd\" (UID: \"0dba2f3b-4ddd-4a1f-8593-1f98dc4479b1\") " pod="openstack/dnsmasq-dns-5b7d94f5b5-4qfsd" Oct 14 07:21:09 crc kubenswrapper[4870]: I1014 07:21:09.411702 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 14 07:21:09 crc kubenswrapper[4870]: I1014 07:21:09.473335 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/aad458d7-8d76-421e-8036-22510ec4f99f-etc-machine-id\") pod \"cinder-api-0\" (UID: \"aad458d7-8d76-421e-8036-22510ec4f99f\") " pod="openstack/cinder-api-0" Oct 14 07:21:09 crc kubenswrapper[4870]: I1014 07:21:09.473784 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aad458d7-8d76-421e-8036-22510ec4f99f-config-data\") pod \"cinder-api-0\" (UID: \"aad458d7-8d76-421e-8036-22510ec4f99f\") " pod="openstack/cinder-api-0" Oct 14 07:21:09 crc kubenswrapper[4870]: I1014 07:21:09.473805 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/aad458d7-8d76-421e-8036-22510ec4f99f-config-data-custom\") pod \"cinder-api-0\" (UID: \"aad458d7-8d76-421e-8036-22510ec4f99f\") " pod="openstack/cinder-api-0" Oct 14 07:21:09 crc kubenswrapper[4870]: I1014 07:21:09.473856 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5p44w\" (UniqueName: \"kubernetes.io/projected/aad458d7-8d76-421e-8036-22510ec4f99f-kube-api-access-5p44w\") pod \"cinder-api-0\" (UID: \"aad458d7-8d76-421e-8036-22510ec4f99f\") " pod="openstack/cinder-api-0" Oct 14 07:21:09 crc kubenswrapper[4870]: I1014 07:21:09.473888 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aad458d7-8d76-421e-8036-22510ec4f99f-scripts\") pod \"cinder-api-0\" (UID: \"aad458d7-8d76-421e-8036-22510ec4f99f\") " pod="openstack/cinder-api-0" Oct 14 07:21:09 crc kubenswrapper[4870]: I1014 07:21:09.473931 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aad458d7-8d76-421e-8036-22510ec4f99f-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"aad458d7-8d76-421e-8036-22510ec4f99f\") " pod="openstack/cinder-api-0" Oct 14 07:21:09 crc kubenswrapper[4870]: I1014 07:21:09.473960 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/aad458d7-8d76-421e-8036-22510ec4f99f-logs\") pod \"cinder-api-0\" (UID: \"aad458d7-8d76-421e-8036-22510ec4f99f\") " pod="openstack/cinder-api-0" Oct 14 07:21:09 crc kubenswrapper[4870]: I1014 07:21:09.474342 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/aad458d7-8d76-421e-8036-22510ec4f99f-logs\") pod \"cinder-api-0\" (UID: \"aad458d7-8d76-421e-8036-22510ec4f99f\") " pod="openstack/cinder-api-0" Oct 14 07:21:09 crc kubenswrapper[4870]: I1014 07:21:09.474391 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/aad458d7-8d76-421e-8036-22510ec4f99f-etc-machine-id\") pod \"cinder-api-0\" (UID: \"aad458d7-8d76-421e-8036-22510ec4f99f\") " pod="openstack/cinder-api-0" Oct 14 07:21:09 crc kubenswrapper[4870]: I1014 07:21:09.501532 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aad458d7-8d76-421e-8036-22510ec4f99f-config-data\") pod \"cinder-api-0\" (UID: \"aad458d7-8d76-421e-8036-22510ec4f99f\") " pod="openstack/cinder-api-0" Oct 14 07:21:09 crc kubenswrapper[4870]: I1014 07:21:09.505375 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aad458d7-8d76-421e-8036-22510ec4f99f-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"aad458d7-8d76-421e-8036-22510ec4f99f\") " pod="openstack/cinder-api-0" Oct 14 07:21:09 crc kubenswrapper[4870]: I1014 07:21:09.513057 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aad458d7-8d76-421e-8036-22510ec4f99f-scripts\") pod \"cinder-api-0\" (UID: \"aad458d7-8d76-421e-8036-22510ec4f99f\") " pod="openstack/cinder-api-0" Oct 14 07:21:09 crc kubenswrapper[4870]: I1014 07:21:09.513202 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/aad458d7-8d76-421e-8036-22510ec4f99f-config-data-custom\") pod \"cinder-api-0\" (UID: \"aad458d7-8d76-421e-8036-22510ec4f99f\") " pod="openstack/cinder-api-0" Oct 14 07:21:09 crc kubenswrapper[4870]: I1014 07:21:09.517376 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5p44w\" (UniqueName: \"kubernetes.io/projected/aad458d7-8d76-421e-8036-22510ec4f99f-kube-api-access-5p44w\") pod \"cinder-api-0\" (UID: \"aad458d7-8d76-421e-8036-22510ec4f99f\") " pod="openstack/cinder-api-0" Oct 14 07:21:09 crc kubenswrapper[4870]: I1014 07:21:09.522890 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b7d94f5b5-4qfsd" Oct 14 07:21:09 crc kubenswrapper[4870]: I1014 07:21:09.582005 4870 generic.go:334] "Generic (PLEG): container finished" podID="bf415c18-d584-4cf1-ae86-3bebdc74cad1" containerID="77a9d26b6f11833d6d27a01cdd11e90dfaece26453d11c1389ea5966482e89ea" exitCode=0 Oct 14 07:21:09 crc kubenswrapper[4870]: I1014 07:21:09.582105 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-29752" event={"ID":"bf415c18-d584-4cf1-ae86-3bebdc74cad1","Type":"ContainerDied","Data":"77a9d26b6f11833d6d27a01cdd11e90dfaece26453d11c1389ea5966482e89ea"} Oct 14 07:21:09 crc kubenswrapper[4870]: I1014 07:21:09.588559 4870 generic.go:334] "Generic (PLEG): container finished" podID="3ea2c25b-4ae5-473c-9480-ec8444bbee56" containerID="00b9207820bcb2e8026c0e8f4ff4fff1bbbb4e4d5e1181da574ef68c63fba739" exitCode=0 Oct 14 07:21:09 crc kubenswrapper[4870]: I1014 07:21:09.588615 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-snx2f" event={"ID":"3ea2c25b-4ae5-473c-9480-ec8444bbee56","Type":"ContainerDied","Data":"00b9207820bcb2e8026c0e8f4ff4fff1bbbb4e4d5e1181da574ef68c63fba739"} Oct 14 07:21:09 crc kubenswrapper[4870]: I1014 07:21:09.624807 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-5485b69bc9-ftmmq" event={"ID":"3927fb78-4f30-4371-8fbb-62fd276c0460","Type":"ContainerStarted","Data":"009e0a5765edbfd4635dd4efcc0b112f3d5e0691352f5b4e2d147aac46326707"} Oct 14 07:21:09 crc kubenswrapper[4870]: I1014 07:21:09.624851 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-5485b69bc9-ftmmq" event={"ID":"3927fb78-4f30-4371-8fbb-62fd276c0460","Type":"ContainerStarted","Data":"58c55f50ceab03915ab9a353ebd2b0d046666be23591bd23a6a60026b468abe2"} Oct 14 07:21:09 crc kubenswrapper[4870]: I1014 07:21:09.625677 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-5485b69bc9-ftmmq" Oct 14 07:21:09 crc kubenswrapper[4870]: I1014 07:21:09.625697 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-5485b69bc9-ftmmq" Oct 14 07:21:09 crc kubenswrapper[4870]: I1014 07:21:09.632145 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 14 07:21:09 crc kubenswrapper[4870]: I1014 07:21:09.670748 4870 generic.go:334] "Generic (PLEG): container finished" podID="52ebf07b-1f8d-4005-b8bc-14e250c921d4" containerID="4a86c8a92708d5909a487ab3cdcb494f586b0ebda954a88a7d9c31a60d00a222" exitCode=0 Oct 14 07:21:09 crc kubenswrapper[4870]: I1014 07:21:09.670866 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-hjj9z" event={"ID":"52ebf07b-1f8d-4005-b8bc-14e250c921d4","Type":"ContainerDied","Data":"4a86c8a92708d5909a487ab3cdcb494f586b0ebda954a88a7d9c31a60d00a222"} Oct 14 07:21:09 crc kubenswrapper[4870]: I1014 07:21:09.740005 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 14 07:21:09 crc kubenswrapper[4870]: I1014 07:21:09.788981 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-5485b69bc9-ftmmq" podStartSLOduration=7.7889623409999995 podStartE2EDuration="7.788962341s" podCreationTimestamp="2025-10-14 07:21:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:21:09.692072092 +0000 UTC m=+1205.389432463" watchObservedRunningTime="2025-10-14 07:21:09.788962341 +0000 UTC m=+1205.486322712" Oct 14 07:21:10 crc kubenswrapper[4870]: I1014 07:21:10.073283 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 14 07:21:10 crc kubenswrapper[4870]: I1014 07:21:10.099273 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 14 07:21:10 crc kubenswrapper[4870]: I1014 07:21:10.358589 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5b7d94f5b5-4qfsd"] Oct 14 07:21:10 crc kubenswrapper[4870]: I1014 07:21:10.730457 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"aad458d7-8d76-421e-8036-22510ec4f99f","Type":"ContainerStarted","Data":"874215bc748e209b91a016089693c2fa19f3277ce49ad8378723989b791c3cc9"} Oct 14 07:21:10 crc kubenswrapper[4870]: I1014 07:21:10.732002 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2ad27e3d-56a2-46e2-a6b0-28b03c5564e8","Type":"ContainerStarted","Data":"02b1c08e5167b9003065eac770e59225bc0ec6c556f92b0afa65f42b1ae6f384"} Oct 14 07:21:10 crc kubenswrapper[4870]: I1014 07:21:10.739472 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b7d94f5b5-4qfsd" event={"ID":"0dba2f3b-4ddd-4a1f-8593-1f98dc4479b1","Type":"ContainerDied","Data":"083d3c82a5af91bb717224dcca9b7742e957f81b1ee53ab1a10a591aa705f77c"} Oct 14 07:21:10 crc kubenswrapper[4870]: I1014 07:21:10.739425 4870 generic.go:334] "Generic (PLEG): container finished" podID="0dba2f3b-4ddd-4a1f-8593-1f98dc4479b1" containerID="083d3c82a5af91bb717224dcca9b7742e957f81b1ee53ab1a10a591aa705f77c" exitCode=0 Oct 14 07:21:10 crc kubenswrapper[4870]: I1014 07:21:10.739680 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b7d94f5b5-4qfsd" event={"ID":"0dba2f3b-4ddd-4a1f-8593-1f98dc4479b1","Type":"ContainerStarted","Data":"458ddb1619a03b4bad71231c7d62bfef1e7348a6b0df30d3b5a0cbd0daefffe5"} Oct 14 07:21:10 crc kubenswrapper[4870]: I1014 07:21:10.742863 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"f04fb303-8db7-4d2c-a864-817494a7222b","Type":"ContainerStarted","Data":"0793837b9e19c6e7c28ba4e548221f38ff6f35c3ed269be3912feeca82aa06e6"} Oct 14 07:21:11 crc kubenswrapper[4870]: I1014 07:21:11.075816 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 14 07:21:11 crc kubenswrapper[4870]: I1014 07:21:11.316647 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-29752" Oct 14 07:21:11 crc kubenswrapper[4870]: I1014 07:21:11.366042 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-hjj9z" Oct 14 07:21:11 crc kubenswrapper[4870]: I1014 07:21:11.373997 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-snx2f" Oct 14 07:21:11 crc kubenswrapper[4870]: I1014 07:21:11.429299 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-248jv\" (UniqueName: \"kubernetes.io/projected/bf415c18-d584-4cf1-ae86-3bebdc74cad1-kube-api-access-248jv\") pod \"bf415c18-d584-4cf1-ae86-3bebdc74cad1\" (UID: \"bf415c18-d584-4cf1-ae86-3bebdc74cad1\") " Oct 14 07:21:11 crc kubenswrapper[4870]: I1014 07:21:11.435099 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf415c18-d584-4cf1-ae86-3bebdc74cad1-kube-api-access-248jv" (OuterVolumeSpecName: "kube-api-access-248jv") pod "bf415c18-d584-4cf1-ae86-3bebdc74cad1" (UID: "bf415c18-d584-4cf1-ae86-3bebdc74cad1"). InnerVolumeSpecName "kube-api-access-248jv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:21:11 crc kubenswrapper[4870]: I1014 07:21:11.531368 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dslpw\" (UniqueName: \"kubernetes.io/projected/52ebf07b-1f8d-4005-b8bc-14e250c921d4-kube-api-access-dslpw\") pod \"52ebf07b-1f8d-4005-b8bc-14e250c921d4\" (UID: \"52ebf07b-1f8d-4005-b8bc-14e250c921d4\") " Oct 14 07:21:11 crc kubenswrapper[4870]: I1014 07:21:11.531668 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6sjvq\" (UniqueName: \"kubernetes.io/projected/3ea2c25b-4ae5-473c-9480-ec8444bbee56-kube-api-access-6sjvq\") pod \"3ea2c25b-4ae5-473c-9480-ec8444bbee56\" (UID: \"3ea2c25b-4ae5-473c-9480-ec8444bbee56\") " Oct 14 07:21:11 crc kubenswrapper[4870]: I1014 07:21:11.532026 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-248jv\" (UniqueName: \"kubernetes.io/projected/bf415c18-d584-4cf1-ae86-3bebdc74cad1-kube-api-access-248jv\") on node \"crc\" DevicePath \"\"" Oct 14 07:21:11 crc kubenswrapper[4870]: I1014 07:21:11.535843 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ea2c25b-4ae5-473c-9480-ec8444bbee56-kube-api-access-6sjvq" (OuterVolumeSpecName: "kube-api-access-6sjvq") pod "3ea2c25b-4ae5-473c-9480-ec8444bbee56" (UID: "3ea2c25b-4ae5-473c-9480-ec8444bbee56"). InnerVolumeSpecName "kube-api-access-6sjvq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:21:11 crc kubenswrapper[4870]: I1014 07:21:11.537680 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/52ebf07b-1f8d-4005-b8bc-14e250c921d4-kube-api-access-dslpw" (OuterVolumeSpecName: "kube-api-access-dslpw") pod "52ebf07b-1f8d-4005-b8bc-14e250c921d4" (UID: "52ebf07b-1f8d-4005-b8bc-14e250c921d4"). InnerVolumeSpecName "kube-api-access-dslpw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:21:11 crc kubenswrapper[4870]: I1014 07:21:11.635414 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6sjvq\" (UniqueName: \"kubernetes.io/projected/3ea2c25b-4ae5-473c-9480-ec8444bbee56-kube-api-access-6sjvq\") on node \"crc\" DevicePath \"\"" Oct 14 07:21:11 crc kubenswrapper[4870]: I1014 07:21:11.635468 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dslpw\" (UniqueName: \"kubernetes.io/projected/52ebf07b-1f8d-4005-b8bc-14e250c921d4-kube-api-access-dslpw\") on node \"crc\" DevicePath \"\"" Oct 14 07:21:11 crc kubenswrapper[4870]: I1014 07:21:11.756531 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-hjj9z" event={"ID":"52ebf07b-1f8d-4005-b8bc-14e250c921d4","Type":"ContainerDied","Data":"7ccfa3b2cb94cb3c6f508623520f00d23df84eb8625198d5ea014296740c1fdb"} Oct 14 07:21:11 crc kubenswrapper[4870]: I1014 07:21:11.756566 4870 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7ccfa3b2cb94cb3c6f508623520f00d23df84eb8625198d5ea014296740c1fdb" Oct 14 07:21:11 crc kubenswrapper[4870]: I1014 07:21:11.756596 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-hjj9z" Oct 14 07:21:11 crc kubenswrapper[4870]: I1014 07:21:11.758951 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"aad458d7-8d76-421e-8036-22510ec4f99f","Type":"ContainerStarted","Data":"fe6816deb9fced577462a66fb5e66f61dda52413027365f0789d845f41569f93"} Oct 14 07:21:11 crc kubenswrapper[4870]: I1014 07:21:11.767274 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-29752" event={"ID":"bf415c18-d584-4cf1-ae86-3bebdc74cad1","Type":"ContainerDied","Data":"5c953af1359978ee56dd076d26829a8cd8698204c204b3cb112538cb378d406f"} Oct 14 07:21:11 crc kubenswrapper[4870]: I1014 07:21:11.767311 4870 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5c953af1359978ee56dd076d26829a8cd8698204c204b3cb112538cb378d406f" Oct 14 07:21:11 crc kubenswrapper[4870]: I1014 07:21:11.767402 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-29752" Oct 14 07:21:11 crc kubenswrapper[4870]: I1014 07:21:11.783595 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-snx2f" event={"ID":"3ea2c25b-4ae5-473c-9480-ec8444bbee56","Type":"ContainerDied","Data":"79f643ac455da16662a323680280f334a2933997e2fce29fc4d04415a16d8838"} Oct 14 07:21:11 crc kubenswrapper[4870]: I1014 07:21:11.783636 4870 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="79f643ac455da16662a323680280f334a2933997e2fce29fc4d04415a16d8838" Oct 14 07:21:11 crc kubenswrapper[4870]: I1014 07:21:11.783718 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-snx2f" Oct 14 07:21:11 crc kubenswrapper[4870]: I1014 07:21:11.786683 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2ad27e3d-56a2-46e2-a6b0-28b03c5564e8","Type":"ContainerStarted","Data":"1a50e3f02e6ba0d60699638881971dbdb3a577fe8453cf5cc9f75e2f8c7f18f6"} Oct 14 07:21:11 crc kubenswrapper[4870]: I1014 07:21:11.790716 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b7d94f5b5-4qfsd" event={"ID":"0dba2f3b-4ddd-4a1f-8593-1f98dc4479b1","Type":"ContainerStarted","Data":"9593b500b23a3c6e6593a46342c30a20d936527ddde619eb3f787d9d70f26504"} Oct 14 07:21:11 crc kubenswrapper[4870]: I1014 07:21:11.790795 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5b7d94f5b5-4qfsd" Oct 14 07:21:11 crc kubenswrapper[4870]: I1014 07:21:11.809072 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5b7d94f5b5-4qfsd" podStartSLOduration=2.809052403 podStartE2EDuration="2.809052403s" podCreationTimestamp="2025-10-14 07:21:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:21:11.807228627 +0000 UTC m=+1207.504588998" watchObservedRunningTime="2025-10-14 07:21:11.809052403 +0000 UTC m=+1207.506412764" Oct 14 07:21:12 crc kubenswrapper[4870]: I1014 07:21:12.804839 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2ad27e3d-56a2-46e2-a6b0-28b03c5564e8","Type":"ContainerStarted","Data":"7168b957cd8420ae3c3849b4234cb79f25064e5611eb0588571632765eb702da"} Oct 14 07:21:12 crc kubenswrapper[4870]: I1014 07:21:12.805538 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2ad27e3d-56a2-46e2-a6b0-28b03c5564e8","Type":"ContainerStarted","Data":"9a9f91f2620a6d42aea6617c0d5bb4f928c39b2d338a18d3aa25af1a81239973"} Oct 14 07:21:12 crc kubenswrapper[4870]: I1014 07:21:12.807404 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"f04fb303-8db7-4d2c-a864-817494a7222b","Type":"ContainerStarted","Data":"ed5d51b7faa1c7111cd571af1833b5840ccac99e3be071ef421e46b99f9cb4b7"} Oct 14 07:21:12 crc kubenswrapper[4870]: I1014 07:21:12.807582 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"f04fb303-8db7-4d2c-a864-817494a7222b","Type":"ContainerStarted","Data":"b8b3775a268dcc4ebc06b8c9a67360595bbc18617ec1fd7cabd8c847229a0730"} Oct 14 07:21:12 crc kubenswrapper[4870]: I1014 07:21:12.809623 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"aad458d7-8d76-421e-8036-22510ec4f99f","Type":"ContainerStarted","Data":"5bc7af17b13bb1325161fd975c8be997797e1c5f96e03be7254d2fd72702fc7d"} Oct 14 07:21:12 crc kubenswrapper[4870]: I1014 07:21:12.835704 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=2.7203774 podStartE2EDuration="3.835681021s" podCreationTimestamp="2025-10-14 07:21:09 +0000 UTC" firstStartedPulling="2025-10-14 07:21:10.087375371 +0000 UTC m=+1205.784735742" lastFinishedPulling="2025-10-14 07:21:11.202679002 +0000 UTC m=+1206.900039363" observedRunningTime="2025-10-14 07:21:12.822992868 +0000 UTC m=+1208.520353239" watchObservedRunningTime="2025-10-14 07:21:12.835681021 +0000 UTC m=+1208.533041412" Oct 14 07:21:12 crc kubenswrapper[4870]: I1014 07:21:12.841675 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=3.8416518809999998 podStartE2EDuration="3.841651881s" podCreationTimestamp="2025-10-14 07:21:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:21:12.84011687 +0000 UTC m=+1208.537477261" watchObservedRunningTime="2025-10-14 07:21:12.841651881 +0000 UTC m=+1208.539012252" Oct 14 07:21:12 crc kubenswrapper[4870]: I1014 07:21:12.986520 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Oct 14 07:21:13 crc kubenswrapper[4870]: I1014 07:21:13.817449 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Oct 14 07:21:14 crc kubenswrapper[4870]: I1014 07:21:14.010334 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-5dd746bf8d-mvg6f" Oct 14 07:21:14 crc kubenswrapper[4870]: I1014 07:21:14.413168 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Oct 14 07:21:14 crc kubenswrapper[4870]: I1014 07:21:14.828422 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2ad27e3d-56a2-46e2-a6b0-28b03c5564e8" containerName="ceilometer-central-agent" containerID="cri-o://1a50e3f02e6ba0d60699638881971dbdb3a577fe8453cf5cc9f75e2f8c7f18f6" gracePeriod=30 Oct 14 07:21:14 crc kubenswrapper[4870]: I1014 07:21:14.828702 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2ad27e3d-56a2-46e2-a6b0-28b03c5564e8","Type":"ContainerStarted","Data":"d3452d3be77f8c486d5a42f9714710b02119f45ae585f3ffbc177ff38f4a717a"} Oct 14 07:21:14 crc kubenswrapper[4870]: I1014 07:21:14.828820 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="aad458d7-8d76-421e-8036-22510ec4f99f" containerName="cinder-api-log" containerID="cri-o://fe6816deb9fced577462a66fb5e66f61dda52413027365f0789d845f41569f93" gracePeriod=30 Oct 14 07:21:14 crc kubenswrapper[4870]: I1014 07:21:14.828900 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 14 07:21:14 crc kubenswrapper[4870]: I1014 07:21:14.829158 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2ad27e3d-56a2-46e2-a6b0-28b03c5564e8" containerName="proxy-httpd" containerID="cri-o://d3452d3be77f8c486d5a42f9714710b02119f45ae585f3ffbc177ff38f4a717a" gracePeriod=30 Oct 14 07:21:14 crc kubenswrapper[4870]: I1014 07:21:14.829210 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2ad27e3d-56a2-46e2-a6b0-28b03c5564e8" containerName="sg-core" containerID="cri-o://7168b957cd8420ae3c3849b4234cb79f25064e5611eb0588571632765eb702da" gracePeriod=30 Oct 14 07:21:14 crc kubenswrapper[4870]: I1014 07:21:14.829241 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2ad27e3d-56a2-46e2-a6b0-28b03c5564e8" containerName="ceilometer-notification-agent" containerID="cri-o://9a9f91f2620a6d42aea6617c0d5bb4f928c39b2d338a18d3aa25af1a81239973" gracePeriod=30 Oct 14 07:21:14 crc kubenswrapper[4870]: I1014 07:21:14.829290 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="aad458d7-8d76-421e-8036-22510ec4f99f" containerName="cinder-api" containerID="cri-o://5bc7af17b13bb1325161fd975c8be997797e1c5f96e03be7254d2fd72702fc7d" gracePeriod=30 Oct 14 07:21:14 crc kubenswrapper[4870]: I1014 07:21:14.857325 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.566158807 podStartE2EDuration="6.857308784s" podCreationTimestamp="2025-10-14 07:21:08 +0000 UTC" firstStartedPulling="2025-10-14 07:21:09.708610443 +0000 UTC m=+1205.405970814" lastFinishedPulling="2025-10-14 07:21:13.99976042 +0000 UTC m=+1209.697120791" observedRunningTime="2025-10-14 07:21:14.854467868 +0000 UTC m=+1210.551828239" watchObservedRunningTime="2025-10-14 07:21:14.857308784 +0000 UTC m=+1210.554669155" Oct 14 07:21:15 crc kubenswrapper[4870]: I1014 07:21:15.080629 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-e66c-account-create-wks4d"] Oct 14 07:21:15 crc kubenswrapper[4870]: E1014 07:21:15.081480 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="52ebf07b-1f8d-4005-b8bc-14e250c921d4" containerName="mariadb-database-create" Oct 14 07:21:15 crc kubenswrapper[4870]: I1014 07:21:15.081584 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="52ebf07b-1f8d-4005-b8bc-14e250c921d4" containerName="mariadb-database-create" Oct 14 07:21:15 crc kubenswrapper[4870]: E1014 07:21:15.081654 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf415c18-d584-4cf1-ae86-3bebdc74cad1" containerName="mariadb-database-create" Oct 14 07:21:15 crc kubenswrapper[4870]: I1014 07:21:15.081709 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf415c18-d584-4cf1-ae86-3bebdc74cad1" containerName="mariadb-database-create" Oct 14 07:21:15 crc kubenswrapper[4870]: E1014 07:21:15.081771 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3ea2c25b-4ae5-473c-9480-ec8444bbee56" containerName="mariadb-database-create" Oct 14 07:21:15 crc kubenswrapper[4870]: I1014 07:21:15.081824 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="3ea2c25b-4ae5-473c-9480-ec8444bbee56" containerName="mariadb-database-create" Oct 14 07:21:15 crc kubenswrapper[4870]: I1014 07:21:15.082055 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="bf415c18-d584-4cf1-ae86-3bebdc74cad1" containerName="mariadb-database-create" Oct 14 07:21:15 crc kubenswrapper[4870]: I1014 07:21:15.082117 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="52ebf07b-1f8d-4005-b8bc-14e250c921d4" containerName="mariadb-database-create" Oct 14 07:21:15 crc kubenswrapper[4870]: I1014 07:21:15.082183 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="3ea2c25b-4ae5-473c-9480-ec8444bbee56" containerName="mariadb-database-create" Oct 14 07:21:15 crc kubenswrapper[4870]: I1014 07:21:15.082772 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-e66c-account-create-wks4d" Oct 14 07:21:15 crc kubenswrapper[4870]: I1014 07:21:15.087384 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Oct 14 07:21:15 crc kubenswrapper[4870]: I1014 07:21:15.100618 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-e66c-account-create-wks4d"] Oct 14 07:21:15 crc kubenswrapper[4870]: I1014 07:21:15.208727 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6lwdb\" (UniqueName: \"kubernetes.io/projected/a4ae5421-6680-4e79-ac45-578fdbbc70f2-kube-api-access-6lwdb\") pod \"nova-api-e66c-account-create-wks4d\" (UID: \"a4ae5421-6680-4e79-ac45-578fdbbc70f2\") " pod="openstack/nova-api-e66c-account-create-wks4d" Oct 14 07:21:15 crc kubenswrapper[4870]: I1014 07:21:15.293990 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-ce16-account-create-nqg26"] Oct 14 07:21:15 crc kubenswrapper[4870]: I1014 07:21:15.296029 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-ce16-account-create-nqg26" Oct 14 07:21:15 crc kubenswrapper[4870]: I1014 07:21:15.298338 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Oct 14 07:21:15 crc kubenswrapper[4870]: I1014 07:21:15.312529 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6lwdb\" (UniqueName: \"kubernetes.io/projected/a4ae5421-6680-4e79-ac45-578fdbbc70f2-kube-api-access-6lwdb\") pod \"nova-api-e66c-account-create-wks4d\" (UID: \"a4ae5421-6680-4e79-ac45-578fdbbc70f2\") " pod="openstack/nova-api-e66c-account-create-wks4d" Oct 14 07:21:15 crc kubenswrapper[4870]: I1014 07:21:15.323138 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-ce16-account-create-nqg26"] Oct 14 07:21:15 crc kubenswrapper[4870]: I1014 07:21:15.358518 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6lwdb\" (UniqueName: \"kubernetes.io/projected/a4ae5421-6680-4e79-ac45-578fdbbc70f2-kube-api-access-6lwdb\") pod \"nova-api-e66c-account-create-wks4d\" (UID: \"a4ae5421-6680-4e79-ac45-578fdbbc70f2\") " pod="openstack/nova-api-e66c-account-create-wks4d" Oct 14 07:21:15 crc kubenswrapper[4870]: I1014 07:21:15.417192 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6vnv8\" (UniqueName: \"kubernetes.io/projected/93d44382-d93e-4db6-b267-181f5e02682b-kube-api-access-6vnv8\") pod \"nova-cell0-ce16-account-create-nqg26\" (UID: \"93d44382-d93e-4db6-b267-181f5e02682b\") " pod="openstack/nova-cell0-ce16-account-create-nqg26" Oct 14 07:21:15 crc kubenswrapper[4870]: I1014 07:21:15.438167 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-e66c-account-create-wks4d" Oct 14 07:21:15 crc kubenswrapper[4870]: I1014 07:21:15.492194 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-52fb-account-create-8kmf7"] Oct 14 07:21:15 crc kubenswrapper[4870]: I1014 07:21:15.493397 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-52fb-account-create-8kmf7" Oct 14 07:21:15 crc kubenswrapper[4870]: I1014 07:21:15.499639 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Oct 14 07:21:15 crc kubenswrapper[4870]: I1014 07:21:15.509120 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-52fb-account-create-8kmf7"] Oct 14 07:21:15 crc kubenswrapper[4870]: I1014 07:21:15.518737 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6vnv8\" (UniqueName: \"kubernetes.io/projected/93d44382-d93e-4db6-b267-181f5e02682b-kube-api-access-6vnv8\") pod \"nova-cell0-ce16-account-create-nqg26\" (UID: \"93d44382-d93e-4db6-b267-181f5e02682b\") " pod="openstack/nova-cell0-ce16-account-create-nqg26" Oct 14 07:21:15 crc kubenswrapper[4870]: I1014 07:21:15.537418 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6vnv8\" (UniqueName: \"kubernetes.io/projected/93d44382-d93e-4db6-b267-181f5e02682b-kube-api-access-6vnv8\") pod \"nova-cell0-ce16-account-create-nqg26\" (UID: \"93d44382-d93e-4db6-b267-181f5e02682b\") " pod="openstack/nova-cell0-ce16-account-create-nqg26" Oct 14 07:21:15 crc kubenswrapper[4870]: I1014 07:21:15.539722 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 14 07:21:15 crc kubenswrapper[4870]: I1014 07:21:15.620517 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ql9vm\" (UniqueName: \"kubernetes.io/projected/93414260-afe1-4af0-8673-d51ec4591776-kube-api-access-ql9vm\") pod \"nova-cell1-52fb-account-create-8kmf7\" (UID: \"93414260-afe1-4af0-8673-d51ec4591776\") " pod="openstack/nova-cell1-52fb-account-create-8kmf7" Oct 14 07:21:15 crc kubenswrapper[4870]: I1014 07:21:15.721784 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aad458d7-8d76-421e-8036-22510ec4f99f-scripts\") pod \"aad458d7-8d76-421e-8036-22510ec4f99f\" (UID: \"aad458d7-8d76-421e-8036-22510ec4f99f\") " Oct 14 07:21:15 crc kubenswrapper[4870]: I1014 07:21:15.721855 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5p44w\" (UniqueName: \"kubernetes.io/projected/aad458d7-8d76-421e-8036-22510ec4f99f-kube-api-access-5p44w\") pod \"aad458d7-8d76-421e-8036-22510ec4f99f\" (UID: \"aad458d7-8d76-421e-8036-22510ec4f99f\") " Oct 14 07:21:15 crc kubenswrapper[4870]: I1014 07:21:15.721901 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aad458d7-8d76-421e-8036-22510ec4f99f-config-data\") pod \"aad458d7-8d76-421e-8036-22510ec4f99f\" (UID: \"aad458d7-8d76-421e-8036-22510ec4f99f\") " Oct 14 07:21:15 crc kubenswrapper[4870]: I1014 07:21:15.721925 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aad458d7-8d76-421e-8036-22510ec4f99f-combined-ca-bundle\") pod \"aad458d7-8d76-421e-8036-22510ec4f99f\" (UID: \"aad458d7-8d76-421e-8036-22510ec4f99f\") " Oct 14 07:21:15 crc kubenswrapper[4870]: I1014 07:21:15.722096 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/aad458d7-8d76-421e-8036-22510ec4f99f-etc-machine-id\") pod \"aad458d7-8d76-421e-8036-22510ec4f99f\" (UID: \"aad458d7-8d76-421e-8036-22510ec4f99f\") " Oct 14 07:21:15 crc kubenswrapper[4870]: I1014 07:21:15.722127 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/aad458d7-8d76-421e-8036-22510ec4f99f-config-data-custom\") pod \"aad458d7-8d76-421e-8036-22510ec4f99f\" (UID: \"aad458d7-8d76-421e-8036-22510ec4f99f\") " Oct 14 07:21:15 crc kubenswrapper[4870]: I1014 07:21:15.722155 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/aad458d7-8d76-421e-8036-22510ec4f99f-logs\") pod \"aad458d7-8d76-421e-8036-22510ec4f99f\" (UID: \"aad458d7-8d76-421e-8036-22510ec4f99f\") " Oct 14 07:21:15 crc kubenswrapper[4870]: I1014 07:21:15.722553 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ql9vm\" (UniqueName: \"kubernetes.io/projected/93414260-afe1-4af0-8673-d51ec4591776-kube-api-access-ql9vm\") pod \"nova-cell1-52fb-account-create-8kmf7\" (UID: \"93414260-afe1-4af0-8673-d51ec4591776\") " pod="openstack/nova-cell1-52fb-account-create-8kmf7" Oct 14 07:21:15 crc kubenswrapper[4870]: I1014 07:21:15.726273 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/aad458d7-8d76-421e-8036-22510ec4f99f-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "aad458d7-8d76-421e-8036-22510ec4f99f" (UID: "aad458d7-8d76-421e-8036-22510ec4f99f"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 14 07:21:15 crc kubenswrapper[4870]: I1014 07:21:15.728104 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aad458d7-8d76-421e-8036-22510ec4f99f-logs" (OuterVolumeSpecName: "logs") pod "aad458d7-8d76-421e-8036-22510ec4f99f" (UID: "aad458d7-8d76-421e-8036-22510ec4f99f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:21:15 crc kubenswrapper[4870]: I1014 07:21:15.728127 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aad458d7-8d76-421e-8036-22510ec4f99f-kube-api-access-5p44w" (OuterVolumeSpecName: "kube-api-access-5p44w") pod "aad458d7-8d76-421e-8036-22510ec4f99f" (UID: "aad458d7-8d76-421e-8036-22510ec4f99f"). InnerVolumeSpecName "kube-api-access-5p44w". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:21:15 crc kubenswrapper[4870]: I1014 07:21:15.731925 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aad458d7-8d76-421e-8036-22510ec4f99f-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "aad458d7-8d76-421e-8036-22510ec4f99f" (UID: "aad458d7-8d76-421e-8036-22510ec4f99f"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:21:15 crc kubenswrapper[4870]: I1014 07:21:15.734632 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aad458d7-8d76-421e-8036-22510ec4f99f-scripts" (OuterVolumeSpecName: "scripts") pod "aad458d7-8d76-421e-8036-22510ec4f99f" (UID: "aad458d7-8d76-421e-8036-22510ec4f99f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:21:15 crc kubenswrapper[4870]: I1014 07:21:15.738942 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ql9vm\" (UniqueName: \"kubernetes.io/projected/93414260-afe1-4af0-8673-d51ec4591776-kube-api-access-ql9vm\") pod \"nova-cell1-52fb-account-create-8kmf7\" (UID: \"93414260-afe1-4af0-8673-d51ec4591776\") " pod="openstack/nova-cell1-52fb-account-create-8kmf7" Oct 14 07:21:15 crc kubenswrapper[4870]: I1014 07:21:15.741014 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-ce16-account-create-nqg26" Oct 14 07:21:15 crc kubenswrapper[4870]: I1014 07:21:15.781995 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 14 07:21:15 crc kubenswrapper[4870]: I1014 07:21:15.790534 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aad458d7-8d76-421e-8036-22510ec4f99f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "aad458d7-8d76-421e-8036-22510ec4f99f" (UID: "aad458d7-8d76-421e-8036-22510ec4f99f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:21:15 crc kubenswrapper[4870]: I1014 07:21:15.822092 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-52fb-account-create-8kmf7" Oct 14 07:21:15 crc kubenswrapper[4870]: I1014 07:21:15.824331 4870 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aad458d7-8d76-421e-8036-22510ec4f99f-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 07:21:15 crc kubenswrapper[4870]: I1014 07:21:15.824358 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5p44w\" (UniqueName: \"kubernetes.io/projected/aad458d7-8d76-421e-8036-22510ec4f99f-kube-api-access-5p44w\") on node \"crc\" DevicePath \"\"" Oct 14 07:21:15 crc kubenswrapper[4870]: I1014 07:21:15.824373 4870 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aad458d7-8d76-421e-8036-22510ec4f99f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 07:21:15 crc kubenswrapper[4870]: I1014 07:21:15.824388 4870 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/aad458d7-8d76-421e-8036-22510ec4f99f-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 14 07:21:15 crc kubenswrapper[4870]: I1014 07:21:15.824417 4870 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/aad458d7-8d76-421e-8036-22510ec4f99f-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 14 07:21:15 crc kubenswrapper[4870]: I1014 07:21:15.824427 4870 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/aad458d7-8d76-421e-8036-22510ec4f99f-logs\") on node \"crc\" DevicePath \"\"" Oct 14 07:21:15 crc kubenswrapper[4870]: I1014 07:21:15.838537 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aad458d7-8d76-421e-8036-22510ec4f99f-config-data" (OuterVolumeSpecName: "config-data") pod "aad458d7-8d76-421e-8036-22510ec4f99f" (UID: "aad458d7-8d76-421e-8036-22510ec4f99f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:21:15 crc kubenswrapper[4870]: I1014 07:21:15.848237 4870 generic.go:334] "Generic (PLEG): container finished" podID="aad458d7-8d76-421e-8036-22510ec4f99f" containerID="5bc7af17b13bb1325161fd975c8be997797e1c5f96e03be7254d2fd72702fc7d" exitCode=0 Oct 14 07:21:15 crc kubenswrapper[4870]: I1014 07:21:15.848275 4870 generic.go:334] "Generic (PLEG): container finished" podID="aad458d7-8d76-421e-8036-22510ec4f99f" containerID="fe6816deb9fced577462a66fb5e66f61dda52413027365f0789d845f41569f93" exitCode=143 Oct 14 07:21:15 crc kubenswrapper[4870]: I1014 07:21:15.848313 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"aad458d7-8d76-421e-8036-22510ec4f99f","Type":"ContainerDied","Data":"5bc7af17b13bb1325161fd975c8be997797e1c5f96e03be7254d2fd72702fc7d"} Oct 14 07:21:15 crc kubenswrapper[4870]: I1014 07:21:15.848344 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"aad458d7-8d76-421e-8036-22510ec4f99f","Type":"ContainerDied","Data":"fe6816deb9fced577462a66fb5e66f61dda52413027365f0789d845f41569f93"} Oct 14 07:21:15 crc kubenswrapper[4870]: I1014 07:21:15.848356 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"aad458d7-8d76-421e-8036-22510ec4f99f","Type":"ContainerDied","Data":"874215bc748e209b91a016089693c2fa19f3277ce49ad8378723989b791c3cc9"} Oct 14 07:21:15 crc kubenswrapper[4870]: I1014 07:21:15.848372 4870 scope.go:117] "RemoveContainer" containerID="5bc7af17b13bb1325161fd975c8be997797e1c5f96e03be7254d2fd72702fc7d" Oct 14 07:21:15 crc kubenswrapper[4870]: I1014 07:21:15.848549 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 14 07:21:15 crc kubenswrapper[4870]: I1014 07:21:15.883699 4870 generic.go:334] "Generic (PLEG): container finished" podID="2ad27e3d-56a2-46e2-a6b0-28b03c5564e8" containerID="d3452d3be77f8c486d5a42f9714710b02119f45ae585f3ffbc177ff38f4a717a" exitCode=0 Oct 14 07:21:15 crc kubenswrapper[4870]: I1014 07:21:15.884421 4870 generic.go:334] "Generic (PLEG): container finished" podID="2ad27e3d-56a2-46e2-a6b0-28b03c5564e8" containerID="7168b957cd8420ae3c3849b4234cb79f25064e5611eb0588571632765eb702da" exitCode=2 Oct 14 07:21:15 crc kubenswrapper[4870]: I1014 07:21:15.884451 4870 generic.go:334] "Generic (PLEG): container finished" podID="2ad27e3d-56a2-46e2-a6b0-28b03c5564e8" containerID="9a9f91f2620a6d42aea6617c0d5bb4f928c39b2d338a18d3aa25af1a81239973" exitCode=0 Oct 14 07:21:15 crc kubenswrapper[4870]: I1014 07:21:15.884458 4870 generic.go:334] "Generic (PLEG): container finished" podID="2ad27e3d-56a2-46e2-a6b0-28b03c5564e8" containerID="1a50e3f02e6ba0d60699638881971dbdb3a577fe8453cf5cc9f75e2f8c7f18f6" exitCode=0 Oct 14 07:21:15 crc kubenswrapper[4870]: I1014 07:21:15.885281 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2ad27e3d-56a2-46e2-a6b0-28b03c5564e8","Type":"ContainerDied","Data":"d3452d3be77f8c486d5a42f9714710b02119f45ae585f3ffbc177ff38f4a717a"} Oct 14 07:21:15 crc kubenswrapper[4870]: I1014 07:21:15.885331 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2ad27e3d-56a2-46e2-a6b0-28b03c5564e8","Type":"ContainerDied","Data":"7168b957cd8420ae3c3849b4234cb79f25064e5611eb0588571632765eb702da"} Oct 14 07:21:15 crc kubenswrapper[4870]: I1014 07:21:15.885345 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2ad27e3d-56a2-46e2-a6b0-28b03c5564e8","Type":"ContainerDied","Data":"9a9f91f2620a6d42aea6617c0d5bb4f928c39b2d338a18d3aa25af1a81239973"} Oct 14 07:21:15 crc kubenswrapper[4870]: I1014 07:21:15.885356 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2ad27e3d-56a2-46e2-a6b0-28b03c5564e8","Type":"ContainerDied","Data":"1a50e3f02e6ba0d60699638881971dbdb3a577fe8453cf5cc9f75e2f8c7f18f6"} Oct 14 07:21:15 crc kubenswrapper[4870]: I1014 07:21:15.885365 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2ad27e3d-56a2-46e2-a6b0-28b03c5564e8","Type":"ContainerDied","Data":"02b1c08e5167b9003065eac770e59225bc0ec6c556f92b0afa65f42b1ae6f384"} Oct 14 07:21:15 crc kubenswrapper[4870]: I1014 07:21:15.885609 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 14 07:21:15 crc kubenswrapper[4870]: I1014 07:21:15.907674 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Oct 14 07:21:15 crc kubenswrapper[4870]: I1014 07:21:15.925351 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2ad27e3d-56a2-46e2-a6b0-28b03c5564e8-log-httpd\") pod \"2ad27e3d-56a2-46e2-a6b0-28b03c5564e8\" (UID: \"2ad27e3d-56a2-46e2-a6b0-28b03c5564e8\") " Oct 14 07:21:15 crc kubenswrapper[4870]: I1014 07:21:15.925497 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2ad27e3d-56a2-46e2-a6b0-28b03c5564e8-scripts\") pod \"2ad27e3d-56a2-46e2-a6b0-28b03c5564e8\" (UID: \"2ad27e3d-56a2-46e2-a6b0-28b03c5564e8\") " Oct 14 07:21:15 crc kubenswrapper[4870]: I1014 07:21:15.925606 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n652l\" (UniqueName: \"kubernetes.io/projected/2ad27e3d-56a2-46e2-a6b0-28b03c5564e8-kube-api-access-n652l\") pod \"2ad27e3d-56a2-46e2-a6b0-28b03c5564e8\" (UID: \"2ad27e3d-56a2-46e2-a6b0-28b03c5564e8\") " Oct 14 07:21:15 crc kubenswrapper[4870]: I1014 07:21:15.925636 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2ad27e3d-56a2-46e2-a6b0-28b03c5564e8-sg-core-conf-yaml\") pod \"2ad27e3d-56a2-46e2-a6b0-28b03c5564e8\" (UID: \"2ad27e3d-56a2-46e2-a6b0-28b03c5564e8\") " Oct 14 07:21:15 crc kubenswrapper[4870]: I1014 07:21:15.925720 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2ad27e3d-56a2-46e2-a6b0-28b03c5564e8-run-httpd\") pod \"2ad27e3d-56a2-46e2-a6b0-28b03c5564e8\" (UID: \"2ad27e3d-56a2-46e2-a6b0-28b03c5564e8\") " Oct 14 07:21:15 crc kubenswrapper[4870]: I1014 07:21:15.925750 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2ad27e3d-56a2-46e2-a6b0-28b03c5564e8-config-data\") pod \"2ad27e3d-56a2-46e2-a6b0-28b03c5564e8\" (UID: \"2ad27e3d-56a2-46e2-a6b0-28b03c5564e8\") " Oct 14 07:21:15 crc kubenswrapper[4870]: I1014 07:21:15.925785 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ad27e3d-56a2-46e2-a6b0-28b03c5564e8-combined-ca-bundle\") pod \"2ad27e3d-56a2-46e2-a6b0-28b03c5564e8\" (UID: \"2ad27e3d-56a2-46e2-a6b0-28b03c5564e8\") " Oct 14 07:21:15 crc kubenswrapper[4870]: I1014 07:21:15.925875 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2ad27e3d-56a2-46e2-a6b0-28b03c5564e8-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "2ad27e3d-56a2-46e2-a6b0-28b03c5564e8" (UID: "2ad27e3d-56a2-46e2-a6b0-28b03c5564e8"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:21:15 crc kubenswrapper[4870]: I1014 07:21:15.926108 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2ad27e3d-56a2-46e2-a6b0-28b03c5564e8-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "2ad27e3d-56a2-46e2-a6b0-28b03c5564e8" (UID: "2ad27e3d-56a2-46e2-a6b0-28b03c5564e8"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:21:15 crc kubenswrapper[4870]: I1014 07:21:15.926133 4870 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aad458d7-8d76-421e-8036-22510ec4f99f-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 07:21:15 crc kubenswrapper[4870]: I1014 07:21:15.926148 4870 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2ad27e3d-56a2-46e2-a6b0-28b03c5564e8-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 14 07:21:15 crc kubenswrapper[4870]: I1014 07:21:15.928415 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Oct 14 07:21:15 crc kubenswrapper[4870]: I1014 07:21:15.965656 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2ad27e3d-56a2-46e2-a6b0-28b03c5564e8-scripts" (OuterVolumeSpecName: "scripts") pod "2ad27e3d-56a2-46e2-a6b0-28b03c5564e8" (UID: "2ad27e3d-56a2-46e2-a6b0-28b03c5564e8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:21:15 crc kubenswrapper[4870]: I1014 07:21:15.966726 4870 scope.go:117] "RemoveContainer" containerID="fe6816deb9fced577462a66fb5e66f61dda52413027365f0789d845f41569f93" Oct 14 07:21:15 crc kubenswrapper[4870]: I1014 07:21:15.996345 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2ad27e3d-56a2-46e2-a6b0-28b03c5564e8-kube-api-access-n652l" (OuterVolumeSpecName: "kube-api-access-n652l") pod "2ad27e3d-56a2-46e2-a6b0-28b03c5564e8" (UID: "2ad27e3d-56a2-46e2-a6b0-28b03c5564e8"). InnerVolumeSpecName "kube-api-access-n652l". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:21:16 crc kubenswrapper[4870]: I1014 07:21:16.002773 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Oct 14 07:21:16 crc kubenswrapper[4870]: E1014 07:21:16.003396 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ad27e3d-56a2-46e2-a6b0-28b03c5564e8" containerName="proxy-httpd" Oct 14 07:21:16 crc kubenswrapper[4870]: I1014 07:21:16.003420 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ad27e3d-56a2-46e2-a6b0-28b03c5564e8" containerName="proxy-httpd" Oct 14 07:21:16 crc kubenswrapper[4870]: E1014 07:21:16.003477 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ad27e3d-56a2-46e2-a6b0-28b03c5564e8" containerName="ceilometer-notification-agent" Oct 14 07:21:16 crc kubenswrapper[4870]: I1014 07:21:16.003485 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ad27e3d-56a2-46e2-a6b0-28b03c5564e8" containerName="ceilometer-notification-agent" Oct 14 07:21:16 crc kubenswrapper[4870]: E1014 07:21:16.003518 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aad458d7-8d76-421e-8036-22510ec4f99f" containerName="cinder-api-log" Oct 14 07:21:16 crc kubenswrapper[4870]: I1014 07:21:16.003526 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="aad458d7-8d76-421e-8036-22510ec4f99f" containerName="cinder-api-log" Oct 14 07:21:16 crc kubenswrapper[4870]: E1014 07:21:16.003554 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aad458d7-8d76-421e-8036-22510ec4f99f" containerName="cinder-api" Oct 14 07:21:16 crc kubenswrapper[4870]: I1014 07:21:16.003563 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="aad458d7-8d76-421e-8036-22510ec4f99f" containerName="cinder-api" Oct 14 07:21:16 crc kubenswrapper[4870]: E1014 07:21:16.003578 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ad27e3d-56a2-46e2-a6b0-28b03c5564e8" containerName="ceilometer-central-agent" Oct 14 07:21:16 crc kubenswrapper[4870]: I1014 07:21:16.003586 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ad27e3d-56a2-46e2-a6b0-28b03c5564e8" containerName="ceilometer-central-agent" Oct 14 07:21:16 crc kubenswrapper[4870]: E1014 07:21:16.003608 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ad27e3d-56a2-46e2-a6b0-28b03c5564e8" containerName="sg-core" Oct 14 07:21:16 crc kubenswrapper[4870]: I1014 07:21:16.003615 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ad27e3d-56a2-46e2-a6b0-28b03c5564e8" containerName="sg-core" Oct 14 07:21:16 crc kubenswrapper[4870]: I1014 07:21:16.003914 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="2ad27e3d-56a2-46e2-a6b0-28b03c5564e8" containerName="ceilometer-notification-agent" Oct 14 07:21:16 crc kubenswrapper[4870]: I1014 07:21:16.003936 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="aad458d7-8d76-421e-8036-22510ec4f99f" containerName="cinder-api" Oct 14 07:21:16 crc kubenswrapper[4870]: I1014 07:21:16.003958 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="aad458d7-8d76-421e-8036-22510ec4f99f" containerName="cinder-api-log" Oct 14 07:21:16 crc kubenswrapper[4870]: I1014 07:21:16.003975 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="2ad27e3d-56a2-46e2-a6b0-28b03c5564e8" containerName="ceilometer-central-agent" Oct 14 07:21:16 crc kubenswrapper[4870]: I1014 07:21:16.003995 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="2ad27e3d-56a2-46e2-a6b0-28b03c5564e8" containerName="proxy-httpd" Oct 14 07:21:16 crc kubenswrapper[4870]: I1014 07:21:16.004003 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="2ad27e3d-56a2-46e2-a6b0-28b03c5564e8" containerName="sg-core" Oct 14 07:21:16 crc kubenswrapper[4870]: I1014 07:21:16.005422 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 14 07:21:16 crc kubenswrapper[4870]: I1014 07:21:16.008811 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Oct 14 07:21:16 crc kubenswrapper[4870]: I1014 07:21:16.008949 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Oct 14 07:21:16 crc kubenswrapper[4870]: I1014 07:21:16.009180 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Oct 14 07:21:16 crc kubenswrapper[4870]: I1014 07:21:16.011156 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2ad27e3d-56a2-46e2-a6b0-28b03c5564e8-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "2ad27e3d-56a2-46e2-a6b0-28b03c5564e8" (UID: "2ad27e3d-56a2-46e2-a6b0-28b03c5564e8"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:21:16 crc kubenswrapper[4870]: I1014 07:21:16.036174 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 14 07:21:16 crc kubenswrapper[4870]: I1014 07:21:16.040394 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n652l\" (UniqueName: \"kubernetes.io/projected/2ad27e3d-56a2-46e2-a6b0-28b03c5564e8-kube-api-access-n652l\") on node \"crc\" DevicePath \"\"" Oct 14 07:21:16 crc kubenswrapper[4870]: I1014 07:21:16.040421 4870 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2ad27e3d-56a2-46e2-a6b0-28b03c5564e8-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 14 07:21:16 crc kubenswrapper[4870]: I1014 07:21:16.040448 4870 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2ad27e3d-56a2-46e2-a6b0-28b03c5564e8-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 14 07:21:16 crc kubenswrapper[4870]: I1014 07:21:16.040461 4870 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2ad27e3d-56a2-46e2-a6b0-28b03c5564e8-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 07:21:16 crc kubenswrapper[4870]: I1014 07:21:16.046228 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-e66c-account-create-wks4d"] Oct 14 07:21:16 crc kubenswrapper[4870]: I1014 07:21:16.049792 4870 scope.go:117] "RemoveContainer" containerID="5bc7af17b13bb1325161fd975c8be997797e1c5f96e03be7254d2fd72702fc7d" Oct 14 07:21:16 crc kubenswrapper[4870]: E1014 07:21:16.062412 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5bc7af17b13bb1325161fd975c8be997797e1c5f96e03be7254d2fd72702fc7d\": container with ID starting with 5bc7af17b13bb1325161fd975c8be997797e1c5f96e03be7254d2fd72702fc7d not found: ID does not exist" containerID="5bc7af17b13bb1325161fd975c8be997797e1c5f96e03be7254d2fd72702fc7d" Oct 14 07:21:16 crc kubenswrapper[4870]: I1014 07:21:16.062483 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5bc7af17b13bb1325161fd975c8be997797e1c5f96e03be7254d2fd72702fc7d"} err="failed to get container status \"5bc7af17b13bb1325161fd975c8be997797e1c5f96e03be7254d2fd72702fc7d\": rpc error: code = NotFound desc = could not find container \"5bc7af17b13bb1325161fd975c8be997797e1c5f96e03be7254d2fd72702fc7d\": container with ID starting with 5bc7af17b13bb1325161fd975c8be997797e1c5f96e03be7254d2fd72702fc7d not found: ID does not exist" Oct 14 07:21:16 crc kubenswrapper[4870]: I1014 07:21:16.062519 4870 scope.go:117] "RemoveContainer" containerID="fe6816deb9fced577462a66fb5e66f61dda52413027365f0789d845f41569f93" Oct 14 07:21:16 crc kubenswrapper[4870]: E1014 07:21:16.064753 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fe6816deb9fced577462a66fb5e66f61dda52413027365f0789d845f41569f93\": container with ID starting with fe6816deb9fced577462a66fb5e66f61dda52413027365f0789d845f41569f93 not found: ID does not exist" containerID="fe6816deb9fced577462a66fb5e66f61dda52413027365f0789d845f41569f93" Oct 14 07:21:16 crc kubenswrapper[4870]: I1014 07:21:16.064785 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fe6816deb9fced577462a66fb5e66f61dda52413027365f0789d845f41569f93"} err="failed to get container status \"fe6816deb9fced577462a66fb5e66f61dda52413027365f0789d845f41569f93\": rpc error: code = NotFound desc = could not find container \"fe6816deb9fced577462a66fb5e66f61dda52413027365f0789d845f41569f93\": container with ID starting with fe6816deb9fced577462a66fb5e66f61dda52413027365f0789d845f41569f93 not found: ID does not exist" Oct 14 07:21:16 crc kubenswrapper[4870]: I1014 07:21:16.064864 4870 scope.go:117] "RemoveContainer" containerID="5bc7af17b13bb1325161fd975c8be997797e1c5f96e03be7254d2fd72702fc7d" Oct 14 07:21:16 crc kubenswrapper[4870]: I1014 07:21:16.065397 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5bc7af17b13bb1325161fd975c8be997797e1c5f96e03be7254d2fd72702fc7d"} err="failed to get container status \"5bc7af17b13bb1325161fd975c8be997797e1c5f96e03be7254d2fd72702fc7d\": rpc error: code = NotFound desc = could not find container \"5bc7af17b13bb1325161fd975c8be997797e1c5f96e03be7254d2fd72702fc7d\": container with ID starting with 5bc7af17b13bb1325161fd975c8be997797e1c5f96e03be7254d2fd72702fc7d not found: ID does not exist" Oct 14 07:21:16 crc kubenswrapper[4870]: I1014 07:21:16.065498 4870 scope.go:117] "RemoveContainer" containerID="fe6816deb9fced577462a66fb5e66f61dda52413027365f0789d845f41569f93" Oct 14 07:21:16 crc kubenswrapper[4870]: I1014 07:21:16.071709 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fe6816deb9fced577462a66fb5e66f61dda52413027365f0789d845f41569f93"} err="failed to get container status \"fe6816deb9fced577462a66fb5e66f61dda52413027365f0789d845f41569f93\": rpc error: code = NotFound desc = could not find container \"fe6816deb9fced577462a66fb5e66f61dda52413027365f0789d845f41569f93\": container with ID starting with fe6816deb9fced577462a66fb5e66f61dda52413027365f0789d845f41569f93 not found: ID does not exist" Oct 14 07:21:16 crc kubenswrapper[4870]: I1014 07:21:16.071760 4870 scope.go:117] "RemoveContainer" containerID="d3452d3be77f8c486d5a42f9714710b02119f45ae585f3ffbc177ff38f4a717a" Oct 14 07:21:16 crc kubenswrapper[4870]: I1014 07:21:16.083116 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2ad27e3d-56a2-46e2-a6b0-28b03c5564e8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2ad27e3d-56a2-46e2-a6b0-28b03c5564e8" (UID: "2ad27e3d-56a2-46e2-a6b0-28b03c5564e8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:21:16 crc kubenswrapper[4870]: I1014 07:21:16.087999 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2ad27e3d-56a2-46e2-a6b0-28b03c5564e8-config-data" (OuterVolumeSpecName: "config-data") pod "2ad27e3d-56a2-46e2-a6b0-28b03c5564e8" (UID: "2ad27e3d-56a2-46e2-a6b0-28b03c5564e8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:21:16 crc kubenswrapper[4870]: I1014 07:21:16.092790 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-ce16-account-create-nqg26"] Oct 14 07:21:16 crc kubenswrapper[4870]: I1014 07:21:16.146363 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4dhzj\" (UniqueName: \"kubernetes.io/projected/51de8090-c48e-498c-99ab-794a7527252c-kube-api-access-4dhzj\") pod \"cinder-api-0\" (UID: \"51de8090-c48e-498c-99ab-794a7527252c\") " pod="openstack/cinder-api-0" Oct 14 07:21:16 crc kubenswrapper[4870]: I1014 07:21:16.147551 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/51de8090-c48e-498c-99ab-794a7527252c-config-data-custom\") pod \"cinder-api-0\" (UID: \"51de8090-c48e-498c-99ab-794a7527252c\") " pod="openstack/cinder-api-0" Oct 14 07:21:16 crc kubenswrapper[4870]: I1014 07:21:16.147580 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/51de8090-c48e-498c-99ab-794a7527252c-logs\") pod \"cinder-api-0\" (UID: \"51de8090-c48e-498c-99ab-794a7527252c\") " pod="openstack/cinder-api-0" Oct 14 07:21:16 crc kubenswrapper[4870]: I1014 07:21:16.147624 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/51de8090-c48e-498c-99ab-794a7527252c-etc-machine-id\") pod \"cinder-api-0\" (UID: \"51de8090-c48e-498c-99ab-794a7527252c\") " pod="openstack/cinder-api-0" Oct 14 07:21:16 crc kubenswrapper[4870]: I1014 07:21:16.147676 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51de8090-c48e-498c-99ab-794a7527252c-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"51de8090-c48e-498c-99ab-794a7527252c\") " pod="openstack/cinder-api-0" Oct 14 07:21:16 crc kubenswrapper[4870]: I1014 07:21:16.147696 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/51de8090-c48e-498c-99ab-794a7527252c-public-tls-certs\") pod \"cinder-api-0\" (UID: \"51de8090-c48e-498c-99ab-794a7527252c\") " pod="openstack/cinder-api-0" Oct 14 07:21:16 crc kubenswrapper[4870]: I1014 07:21:16.147716 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/51de8090-c48e-498c-99ab-794a7527252c-scripts\") pod \"cinder-api-0\" (UID: \"51de8090-c48e-498c-99ab-794a7527252c\") " pod="openstack/cinder-api-0" Oct 14 07:21:16 crc kubenswrapper[4870]: I1014 07:21:16.147806 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/51de8090-c48e-498c-99ab-794a7527252c-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"51de8090-c48e-498c-99ab-794a7527252c\") " pod="openstack/cinder-api-0" Oct 14 07:21:16 crc kubenswrapper[4870]: I1014 07:21:16.147850 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/51de8090-c48e-498c-99ab-794a7527252c-config-data\") pod \"cinder-api-0\" (UID: \"51de8090-c48e-498c-99ab-794a7527252c\") " pod="openstack/cinder-api-0" Oct 14 07:21:16 crc kubenswrapper[4870]: I1014 07:21:16.147955 4870 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2ad27e3d-56a2-46e2-a6b0-28b03c5564e8-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 07:21:16 crc kubenswrapper[4870]: I1014 07:21:16.147974 4870 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ad27e3d-56a2-46e2-a6b0-28b03c5564e8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 07:21:16 crc kubenswrapper[4870]: I1014 07:21:16.249422 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/51de8090-c48e-498c-99ab-794a7527252c-config-data\") pod \"cinder-api-0\" (UID: \"51de8090-c48e-498c-99ab-794a7527252c\") " pod="openstack/cinder-api-0" Oct 14 07:21:16 crc kubenswrapper[4870]: I1014 07:21:16.249648 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4dhzj\" (UniqueName: \"kubernetes.io/projected/51de8090-c48e-498c-99ab-794a7527252c-kube-api-access-4dhzj\") pod \"cinder-api-0\" (UID: \"51de8090-c48e-498c-99ab-794a7527252c\") " pod="openstack/cinder-api-0" Oct 14 07:21:16 crc kubenswrapper[4870]: I1014 07:21:16.249677 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/51de8090-c48e-498c-99ab-794a7527252c-config-data-custom\") pod \"cinder-api-0\" (UID: \"51de8090-c48e-498c-99ab-794a7527252c\") " pod="openstack/cinder-api-0" Oct 14 07:21:16 crc kubenswrapper[4870]: I1014 07:21:16.249701 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/51de8090-c48e-498c-99ab-794a7527252c-logs\") pod \"cinder-api-0\" (UID: \"51de8090-c48e-498c-99ab-794a7527252c\") " pod="openstack/cinder-api-0" Oct 14 07:21:16 crc kubenswrapper[4870]: I1014 07:21:16.249752 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/51de8090-c48e-498c-99ab-794a7527252c-etc-machine-id\") pod \"cinder-api-0\" (UID: \"51de8090-c48e-498c-99ab-794a7527252c\") " pod="openstack/cinder-api-0" Oct 14 07:21:16 crc kubenswrapper[4870]: I1014 07:21:16.249817 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51de8090-c48e-498c-99ab-794a7527252c-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"51de8090-c48e-498c-99ab-794a7527252c\") " pod="openstack/cinder-api-0" Oct 14 07:21:16 crc kubenswrapper[4870]: I1014 07:21:16.249843 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/51de8090-c48e-498c-99ab-794a7527252c-public-tls-certs\") pod \"cinder-api-0\" (UID: \"51de8090-c48e-498c-99ab-794a7527252c\") " pod="openstack/cinder-api-0" Oct 14 07:21:16 crc kubenswrapper[4870]: I1014 07:21:16.249870 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/51de8090-c48e-498c-99ab-794a7527252c-scripts\") pod \"cinder-api-0\" (UID: \"51de8090-c48e-498c-99ab-794a7527252c\") " pod="openstack/cinder-api-0" Oct 14 07:21:16 crc kubenswrapper[4870]: I1014 07:21:16.249917 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/51de8090-c48e-498c-99ab-794a7527252c-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"51de8090-c48e-498c-99ab-794a7527252c\") " pod="openstack/cinder-api-0" Oct 14 07:21:16 crc kubenswrapper[4870]: I1014 07:21:16.251148 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/51de8090-c48e-498c-99ab-794a7527252c-logs\") pod \"cinder-api-0\" (UID: \"51de8090-c48e-498c-99ab-794a7527252c\") " pod="openstack/cinder-api-0" Oct 14 07:21:16 crc kubenswrapper[4870]: I1014 07:21:16.251639 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/51de8090-c48e-498c-99ab-794a7527252c-etc-machine-id\") pod \"cinder-api-0\" (UID: \"51de8090-c48e-498c-99ab-794a7527252c\") " pod="openstack/cinder-api-0" Oct 14 07:21:16 crc kubenswrapper[4870]: I1014 07:21:16.261343 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/51de8090-c48e-498c-99ab-794a7527252c-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"51de8090-c48e-498c-99ab-794a7527252c\") " pod="openstack/cinder-api-0" Oct 14 07:21:16 crc kubenswrapper[4870]: I1014 07:21:16.261938 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/51de8090-c48e-498c-99ab-794a7527252c-config-data\") pod \"cinder-api-0\" (UID: \"51de8090-c48e-498c-99ab-794a7527252c\") " pod="openstack/cinder-api-0" Oct 14 07:21:16 crc kubenswrapper[4870]: I1014 07:21:16.263093 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/51de8090-c48e-498c-99ab-794a7527252c-config-data-custom\") pod \"cinder-api-0\" (UID: \"51de8090-c48e-498c-99ab-794a7527252c\") " pod="openstack/cinder-api-0" Oct 14 07:21:16 crc kubenswrapper[4870]: I1014 07:21:16.263756 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/51de8090-c48e-498c-99ab-794a7527252c-scripts\") pod \"cinder-api-0\" (UID: \"51de8090-c48e-498c-99ab-794a7527252c\") " pod="openstack/cinder-api-0" Oct 14 07:21:16 crc kubenswrapper[4870]: I1014 07:21:16.264211 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/51de8090-c48e-498c-99ab-794a7527252c-public-tls-certs\") pod \"cinder-api-0\" (UID: \"51de8090-c48e-498c-99ab-794a7527252c\") " pod="openstack/cinder-api-0" Oct 14 07:21:16 crc kubenswrapper[4870]: I1014 07:21:16.268771 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4dhzj\" (UniqueName: \"kubernetes.io/projected/51de8090-c48e-498c-99ab-794a7527252c-kube-api-access-4dhzj\") pod \"cinder-api-0\" (UID: \"51de8090-c48e-498c-99ab-794a7527252c\") " pod="openstack/cinder-api-0" Oct 14 07:21:16 crc kubenswrapper[4870]: I1014 07:21:16.275404 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51de8090-c48e-498c-99ab-794a7527252c-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"51de8090-c48e-498c-99ab-794a7527252c\") " pod="openstack/cinder-api-0" Oct 14 07:21:16 crc kubenswrapper[4870]: I1014 07:21:16.346434 4870 scope.go:117] "RemoveContainer" containerID="7168b957cd8420ae3c3849b4234cb79f25064e5611eb0588571632765eb702da" Oct 14 07:21:16 crc kubenswrapper[4870]: I1014 07:21:16.349385 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 14 07:21:16 crc kubenswrapper[4870]: I1014 07:21:16.377821 4870 scope.go:117] "RemoveContainer" containerID="9a9f91f2620a6d42aea6617c0d5bb4f928c39b2d338a18d3aa25af1a81239973" Oct 14 07:21:16 crc kubenswrapper[4870]: I1014 07:21:16.388248 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 14 07:21:16 crc kubenswrapper[4870]: I1014 07:21:16.400139 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 14 07:21:16 crc kubenswrapper[4870]: I1014 07:21:16.414795 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 14 07:21:16 crc kubenswrapper[4870]: I1014 07:21:16.418117 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 14 07:21:16 crc kubenswrapper[4870]: I1014 07:21:16.420722 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 14 07:21:16 crc kubenswrapper[4870]: I1014 07:21:16.420763 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 14 07:21:16 crc kubenswrapper[4870]: I1014 07:21:16.421255 4870 scope.go:117] "RemoveContainer" containerID="1a50e3f02e6ba0d60699638881971dbdb3a577fe8453cf5cc9f75e2f8c7f18f6" Oct 14 07:21:16 crc kubenswrapper[4870]: I1014 07:21:16.423774 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 14 07:21:16 crc kubenswrapper[4870]: I1014 07:21:16.466608 4870 scope.go:117] "RemoveContainer" containerID="d3452d3be77f8c486d5a42f9714710b02119f45ae585f3ffbc177ff38f4a717a" Oct 14 07:21:16 crc kubenswrapper[4870]: E1014 07:21:16.467935 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d3452d3be77f8c486d5a42f9714710b02119f45ae585f3ffbc177ff38f4a717a\": container with ID starting with d3452d3be77f8c486d5a42f9714710b02119f45ae585f3ffbc177ff38f4a717a not found: ID does not exist" containerID="d3452d3be77f8c486d5a42f9714710b02119f45ae585f3ffbc177ff38f4a717a" Oct 14 07:21:16 crc kubenswrapper[4870]: I1014 07:21:16.468003 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d3452d3be77f8c486d5a42f9714710b02119f45ae585f3ffbc177ff38f4a717a"} err="failed to get container status \"d3452d3be77f8c486d5a42f9714710b02119f45ae585f3ffbc177ff38f4a717a\": rpc error: code = NotFound desc = could not find container \"d3452d3be77f8c486d5a42f9714710b02119f45ae585f3ffbc177ff38f4a717a\": container with ID starting with d3452d3be77f8c486d5a42f9714710b02119f45ae585f3ffbc177ff38f4a717a not found: ID does not exist" Oct 14 07:21:16 crc kubenswrapper[4870]: I1014 07:21:16.468051 4870 scope.go:117] "RemoveContainer" containerID="7168b957cd8420ae3c3849b4234cb79f25064e5611eb0588571632765eb702da" Oct 14 07:21:16 crc kubenswrapper[4870]: E1014 07:21:16.469421 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7168b957cd8420ae3c3849b4234cb79f25064e5611eb0588571632765eb702da\": container with ID starting with 7168b957cd8420ae3c3849b4234cb79f25064e5611eb0588571632765eb702da not found: ID does not exist" containerID="7168b957cd8420ae3c3849b4234cb79f25064e5611eb0588571632765eb702da" Oct 14 07:21:16 crc kubenswrapper[4870]: I1014 07:21:16.469477 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7168b957cd8420ae3c3849b4234cb79f25064e5611eb0588571632765eb702da"} err="failed to get container status \"7168b957cd8420ae3c3849b4234cb79f25064e5611eb0588571632765eb702da\": rpc error: code = NotFound desc = could not find container \"7168b957cd8420ae3c3849b4234cb79f25064e5611eb0588571632765eb702da\": container with ID starting with 7168b957cd8420ae3c3849b4234cb79f25064e5611eb0588571632765eb702da not found: ID does not exist" Oct 14 07:21:16 crc kubenswrapper[4870]: I1014 07:21:16.469493 4870 scope.go:117] "RemoveContainer" containerID="9a9f91f2620a6d42aea6617c0d5bb4f928c39b2d338a18d3aa25af1a81239973" Oct 14 07:21:16 crc kubenswrapper[4870]: E1014 07:21:16.469785 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9a9f91f2620a6d42aea6617c0d5bb4f928c39b2d338a18d3aa25af1a81239973\": container with ID starting with 9a9f91f2620a6d42aea6617c0d5bb4f928c39b2d338a18d3aa25af1a81239973 not found: ID does not exist" containerID="9a9f91f2620a6d42aea6617c0d5bb4f928c39b2d338a18d3aa25af1a81239973" Oct 14 07:21:16 crc kubenswrapper[4870]: I1014 07:21:16.469816 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9a9f91f2620a6d42aea6617c0d5bb4f928c39b2d338a18d3aa25af1a81239973"} err="failed to get container status \"9a9f91f2620a6d42aea6617c0d5bb4f928c39b2d338a18d3aa25af1a81239973\": rpc error: code = NotFound desc = could not find container \"9a9f91f2620a6d42aea6617c0d5bb4f928c39b2d338a18d3aa25af1a81239973\": container with ID starting with 9a9f91f2620a6d42aea6617c0d5bb4f928c39b2d338a18d3aa25af1a81239973 not found: ID does not exist" Oct 14 07:21:16 crc kubenswrapper[4870]: I1014 07:21:16.469834 4870 scope.go:117] "RemoveContainer" containerID="1a50e3f02e6ba0d60699638881971dbdb3a577fe8453cf5cc9f75e2f8c7f18f6" Oct 14 07:21:16 crc kubenswrapper[4870]: E1014 07:21:16.470120 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1a50e3f02e6ba0d60699638881971dbdb3a577fe8453cf5cc9f75e2f8c7f18f6\": container with ID starting with 1a50e3f02e6ba0d60699638881971dbdb3a577fe8453cf5cc9f75e2f8c7f18f6 not found: ID does not exist" containerID="1a50e3f02e6ba0d60699638881971dbdb3a577fe8453cf5cc9f75e2f8c7f18f6" Oct 14 07:21:16 crc kubenswrapper[4870]: I1014 07:21:16.470147 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1a50e3f02e6ba0d60699638881971dbdb3a577fe8453cf5cc9f75e2f8c7f18f6"} err="failed to get container status \"1a50e3f02e6ba0d60699638881971dbdb3a577fe8453cf5cc9f75e2f8c7f18f6\": rpc error: code = NotFound desc = could not find container \"1a50e3f02e6ba0d60699638881971dbdb3a577fe8453cf5cc9f75e2f8c7f18f6\": container with ID starting with 1a50e3f02e6ba0d60699638881971dbdb3a577fe8453cf5cc9f75e2f8c7f18f6 not found: ID does not exist" Oct 14 07:21:16 crc kubenswrapper[4870]: I1014 07:21:16.470164 4870 scope.go:117] "RemoveContainer" containerID="d3452d3be77f8c486d5a42f9714710b02119f45ae585f3ffbc177ff38f4a717a" Oct 14 07:21:16 crc kubenswrapper[4870]: I1014 07:21:16.470892 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d3452d3be77f8c486d5a42f9714710b02119f45ae585f3ffbc177ff38f4a717a"} err="failed to get container status \"d3452d3be77f8c486d5a42f9714710b02119f45ae585f3ffbc177ff38f4a717a\": rpc error: code = NotFound desc = could not find container \"d3452d3be77f8c486d5a42f9714710b02119f45ae585f3ffbc177ff38f4a717a\": container with ID starting with d3452d3be77f8c486d5a42f9714710b02119f45ae585f3ffbc177ff38f4a717a not found: ID does not exist" Oct 14 07:21:16 crc kubenswrapper[4870]: I1014 07:21:16.470915 4870 scope.go:117] "RemoveContainer" containerID="7168b957cd8420ae3c3849b4234cb79f25064e5611eb0588571632765eb702da" Oct 14 07:21:16 crc kubenswrapper[4870]: I1014 07:21:16.472527 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7168b957cd8420ae3c3849b4234cb79f25064e5611eb0588571632765eb702da"} err="failed to get container status \"7168b957cd8420ae3c3849b4234cb79f25064e5611eb0588571632765eb702da\": rpc error: code = NotFound desc = could not find container \"7168b957cd8420ae3c3849b4234cb79f25064e5611eb0588571632765eb702da\": container with ID starting with 7168b957cd8420ae3c3849b4234cb79f25064e5611eb0588571632765eb702da not found: ID does not exist" Oct 14 07:21:16 crc kubenswrapper[4870]: I1014 07:21:16.472549 4870 scope.go:117] "RemoveContainer" containerID="9a9f91f2620a6d42aea6617c0d5bb4f928c39b2d338a18d3aa25af1a81239973" Oct 14 07:21:16 crc kubenswrapper[4870]: I1014 07:21:16.472860 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9a9f91f2620a6d42aea6617c0d5bb4f928c39b2d338a18d3aa25af1a81239973"} err="failed to get container status \"9a9f91f2620a6d42aea6617c0d5bb4f928c39b2d338a18d3aa25af1a81239973\": rpc error: code = NotFound desc = could not find container \"9a9f91f2620a6d42aea6617c0d5bb4f928c39b2d338a18d3aa25af1a81239973\": container with ID starting with 9a9f91f2620a6d42aea6617c0d5bb4f928c39b2d338a18d3aa25af1a81239973 not found: ID does not exist" Oct 14 07:21:16 crc kubenswrapper[4870]: I1014 07:21:16.472880 4870 scope.go:117] "RemoveContainer" containerID="1a50e3f02e6ba0d60699638881971dbdb3a577fe8453cf5cc9f75e2f8c7f18f6" Oct 14 07:21:16 crc kubenswrapper[4870]: I1014 07:21:16.473270 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1a50e3f02e6ba0d60699638881971dbdb3a577fe8453cf5cc9f75e2f8c7f18f6"} err="failed to get container status \"1a50e3f02e6ba0d60699638881971dbdb3a577fe8453cf5cc9f75e2f8c7f18f6\": rpc error: code = NotFound desc = could not find container \"1a50e3f02e6ba0d60699638881971dbdb3a577fe8453cf5cc9f75e2f8c7f18f6\": container with ID starting with 1a50e3f02e6ba0d60699638881971dbdb3a577fe8453cf5cc9f75e2f8c7f18f6 not found: ID does not exist" Oct 14 07:21:16 crc kubenswrapper[4870]: I1014 07:21:16.473287 4870 scope.go:117] "RemoveContainer" containerID="d3452d3be77f8c486d5a42f9714710b02119f45ae585f3ffbc177ff38f4a717a" Oct 14 07:21:16 crc kubenswrapper[4870]: I1014 07:21:16.474308 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d3452d3be77f8c486d5a42f9714710b02119f45ae585f3ffbc177ff38f4a717a"} err="failed to get container status \"d3452d3be77f8c486d5a42f9714710b02119f45ae585f3ffbc177ff38f4a717a\": rpc error: code = NotFound desc = could not find container \"d3452d3be77f8c486d5a42f9714710b02119f45ae585f3ffbc177ff38f4a717a\": container with ID starting with d3452d3be77f8c486d5a42f9714710b02119f45ae585f3ffbc177ff38f4a717a not found: ID does not exist" Oct 14 07:21:16 crc kubenswrapper[4870]: I1014 07:21:16.474331 4870 scope.go:117] "RemoveContainer" containerID="7168b957cd8420ae3c3849b4234cb79f25064e5611eb0588571632765eb702da" Oct 14 07:21:16 crc kubenswrapper[4870]: I1014 07:21:16.477945 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7168b957cd8420ae3c3849b4234cb79f25064e5611eb0588571632765eb702da"} err="failed to get container status \"7168b957cd8420ae3c3849b4234cb79f25064e5611eb0588571632765eb702da\": rpc error: code = NotFound desc = could not find container \"7168b957cd8420ae3c3849b4234cb79f25064e5611eb0588571632765eb702da\": container with ID starting with 7168b957cd8420ae3c3849b4234cb79f25064e5611eb0588571632765eb702da not found: ID does not exist" Oct 14 07:21:16 crc kubenswrapper[4870]: I1014 07:21:16.477984 4870 scope.go:117] "RemoveContainer" containerID="9a9f91f2620a6d42aea6617c0d5bb4f928c39b2d338a18d3aa25af1a81239973" Oct 14 07:21:16 crc kubenswrapper[4870]: I1014 07:21:16.478476 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9a9f91f2620a6d42aea6617c0d5bb4f928c39b2d338a18d3aa25af1a81239973"} err="failed to get container status \"9a9f91f2620a6d42aea6617c0d5bb4f928c39b2d338a18d3aa25af1a81239973\": rpc error: code = NotFound desc = could not find container \"9a9f91f2620a6d42aea6617c0d5bb4f928c39b2d338a18d3aa25af1a81239973\": container with ID starting with 9a9f91f2620a6d42aea6617c0d5bb4f928c39b2d338a18d3aa25af1a81239973 not found: ID does not exist" Oct 14 07:21:16 crc kubenswrapper[4870]: I1014 07:21:16.478517 4870 scope.go:117] "RemoveContainer" containerID="1a50e3f02e6ba0d60699638881971dbdb3a577fe8453cf5cc9f75e2f8c7f18f6" Oct 14 07:21:16 crc kubenswrapper[4870]: I1014 07:21:16.478756 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1a50e3f02e6ba0d60699638881971dbdb3a577fe8453cf5cc9f75e2f8c7f18f6"} err="failed to get container status \"1a50e3f02e6ba0d60699638881971dbdb3a577fe8453cf5cc9f75e2f8c7f18f6\": rpc error: code = NotFound desc = could not find container \"1a50e3f02e6ba0d60699638881971dbdb3a577fe8453cf5cc9f75e2f8c7f18f6\": container with ID starting with 1a50e3f02e6ba0d60699638881971dbdb3a577fe8453cf5cc9f75e2f8c7f18f6 not found: ID does not exist" Oct 14 07:21:16 crc kubenswrapper[4870]: I1014 07:21:16.478776 4870 scope.go:117] "RemoveContainer" containerID="d3452d3be77f8c486d5a42f9714710b02119f45ae585f3ffbc177ff38f4a717a" Oct 14 07:21:16 crc kubenswrapper[4870]: I1014 07:21:16.478979 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d3452d3be77f8c486d5a42f9714710b02119f45ae585f3ffbc177ff38f4a717a"} err="failed to get container status \"d3452d3be77f8c486d5a42f9714710b02119f45ae585f3ffbc177ff38f4a717a\": rpc error: code = NotFound desc = could not find container \"d3452d3be77f8c486d5a42f9714710b02119f45ae585f3ffbc177ff38f4a717a\": container with ID starting with d3452d3be77f8c486d5a42f9714710b02119f45ae585f3ffbc177ff38f4a717a not found: ID does not exist" Oct 14 07:21:16 crc kubenswrapper[4870]: I1014 07:21:16.478996 4870 scope.go:117] "RemoveContainer" containerID="7168b957cd8420ae3c3849b4234cb79f25064e5611eb0588571632765eb702da" Oct 14 07:21:16 crc kubenswrapper[4870]: I1014 07:21:16.479194 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7168b957cd8420ae3c3849b4234cb79f25064e5611eb0588571632765eb702da"} err="failed to get container status \"7168b957cd8420ae3c3849b4234cb79f25064e5611eb0588571632765eb702da\": rpc error: code = NotFound desc = could not find container \"7168b957cd8420ae3c3849b4234cb79f25064e5611eb0588571632765eb702da\": container with ID starting with 7168b957cd8420ae3c3849b4234cb79f25064e5611eb0588571632765eb702da not found: ID does not exist" Oct 14 07:21:16 crc kubenswrapper[4870]: I1014 07:21:16.479210 4870 scope.go:117] "RemoveContainer" containerID="9a9f91f2620a6d42aea6617c0d5bb4f928c39b2d338a18d3aa25af1a81239973" Oct 14 07:21:16 crc kubenswrapper[4870]: I1014 07:21:16.479419 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9a9f91f2620a6d42aea6617c0d5bb4f928c39b2d338a18d3aa25af1a81239973"} err="failed to get container status \"9a9f91f2620a6d42aea6617c0d5bb4f928c39b2d338a18d3aa25af1a81239973\": rpc error: code = NotFound desc = could not find container \"9a9f91f2620a6d42aea6617c0d5bb4f928c39b2d338a18d3aa25af1a81239973\": container with ID starting with 9a9f91f2620a6d42aea6617c0d5bb4f928c39b2d338a18d3aa25af1a81239973 not found: ID does not exist" Oct 14 07:21:16 crc kubenswrapper[4870]: I1014 07:21:16.479469 4870 scope.go:117] "RemoveContainer" containerID="1a50e3f02e6ba0d60699638881971dbdb3a577fe8453cf5cc9f75e2f8c7f18f6" Oct 14 07:21:16 crc kubenswrapper[4870]: I1014 07:21:16.479717 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1a50e3f02e6ba0d60699638881971dbdb3a577fe8453cf5cc9f75e2f8c7f18f6"} err="failed to get container status \"1a50e3f02e6ba0d60699638881971dbdb3a577fe8453cf5cc9f75e2f8c7f18f6\": rpc error: code = NotFound desc = could not find container \"1a50e3f02e6ba0d60699638881971dbdb3a577fe8453cf5cc9f75e2f8c7f18f6\": container with ID starting with 1a50e3f02e6ba0d60699638881971dbdb3a577fe8453cf5cc9f75e2f8c7f18f6 not found: ID does not exist" Oct 14 07:21:16 crc kubenswrapper[4870]: I1014 07:21:16.482021 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-52fb-account-create-8kmf7"] Oct 14 07:21:16 crc kubenswrapper[4870]: I1014 07:21:16.555559 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x6hww\" (UniqueName: \"kubernetes.io/projected/4d00ede0-cd53-4791-9325-c712ee891839-kube-api-access-x6hww\") pod \"ceilometer-0\" (UID: \"4d00ede0-cd53-4791-9325-c712ee891839\") " pod="openstack/ceilometer-0" Oct 14 07:21:16 crc kubenswrapper[4870]: I1014 07:21:16.556060 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4d00ede0-cd53-4791-9325-c712ee891839-log-httpd\") pod \"ceilometer-0\" (UID: \"4d00ede0-cd53-4791-9325-c712ee891839\") " pod="openstack/ceilometer-0" Oct 14 07:21:16 crc kubenswrapper[4870]: I1014 07:21:16.556098 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4d00ede0-cd53-4791-9325-c712ee891839-run-httpd\") pod \"ceilometer-0\" (UID: \"4d00ede0-cd53-4791-9325-c712ee891839\") " pod="openstack/ceilometer-0" Oct 14 07:21:16 crc kubenswrapper[4870]: I1014 07:21:16.556128 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4d00ede0-cd53-4791-9325-c712ee891839-config-data\") pod \"ceilometer-0\" (UID: \"4d00ede0-cd53-4791-9325-c712ee891839\") " pod="openstack/ceilometer-0" Oct 14 07:21:16 crc kubenswrapper[4870]: I1014 07:21:16.556162 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4d00ede0-cd53-4791-9325-c712ee891839-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4d00ede0-cd53-4791-9325-c712ee891839\") " pod="openstack/ceilometer-0" Oct 14 07:21:16 crc kubenswrapper[4870]: I1014 07:21:16.556214 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4d00ede0-cd53-4791-9325-c712ee891839-scripts\") pod \"ceilometer-0\" (UID: \"4d00ede0-cd53-4791-9325-c712ee891839\") " pod="openstack/ceilometer-0" Oct 14 07:21:16 crc kubenswrapper[4870]: I1014 07:21:16.556243 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d00ede0-cd53-4791-9325-c712ee891839-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4d00ede0-cd53-4791-9325-c712ee891839\") " pod="openstack/ceilometer-0" Oct 14 07:21:16 crc kubenswrapper[4870]: I1014 07:21:16.657521 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x6hww\" (UniqueName: \"kubernetes.io/projected/4d00ede0-cd53-4791-9325-c712ee891839-kube-api-access-x6hww\") pod \"ceilometer-0\" (UID: \"4d00ede0-cd53-4791-9325-c712ee891839\") " pod="openstack/ceilometer-0" Oct 14 07:21:16 crc kubenswrapper[4870]: I1014 07:21:16.657962 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4d00ede0-cd53-4791-9325-c712ee891839-log-httpd\") pod \"ceilometer-0\" (UID: \"4d00ede0-cd53-4791-9325-c712ee891839\") " pod="openstack/ceilometer-0" Oct 14 07:21:16 crc kubenswrapper[4870]: I1014 07:21:16.657988 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4d00ede0-cd53-4791-9325-c712ee891839-run-httpd\") pod \"ceilometer-0\" (UID: \"4d00ede0-cd53-4791-9325-c712ee891839\") " pod="openstack/ceilometer-0" Oct 14 07:21:16 crc kubenswrapper[4870]: I1014 07:21:16.658004 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4d00ede0-cd53-4791-9325-c712ee891839-config-data\") pod \"ceilometer-0\" (UID: \"4d00ede0-cd53-4791-9325-c712ee891839\") " pod="openstack/ceilometer-0" Oct 14 07:21:16 crc kubenswrapper[4870]: I1014 07:21:16.658028 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4d00ede0-cd53-4791-9325-c712ee891839-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4d00ede0-cd53-4791-9325-c712ee891839\") " pod="openstack/ceilometer-0" Oct 14 07:21:16 crc kubenswrapper[4870]: I1014 07:21:16.658068 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4d00ede0-cd53-4791-9325-c712ee891839-scripts\") pod \"ceilometer-0\" (UID: \"4d00ede0-cd53-4791-9325-c712ee891839\") " pod="openstack/ceilometer-0" Oct 14 07:21:16 crc kubenswrapper[4870]: I1014 07:21:16.658087 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d00ede0-cd53-4791-9325-c712ee891839-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4d00ede0-cd53-4791-9325-c712ee891839\") " pod="openstack/ceilometer-0" Oct 14 07:21:16 crc kubenswrapper[4870]: I1014 07:21:16.664133 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4d00ede0-cd53-4791-9325-c712ee891839-log-httpd\") pod \"ceilometer-0\" (UID: \"4d00ede0-cd53-4791-9325-c712ee891839\") " pod="openstack/ceilometer-0" Oct 14 07:21:16 crc kubenswrapper[4870]: I1014 07:21:16.666661 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4d00ede0-cd53-4791-9325-c712ee891839-run-httpd\") pod \"ceilometer-0\" (UID: \"4d00ede0-cd53-4791-9325-c712ee891839\") " pod="openstack/ceilometer-0" Oct 14 07:21:16 crc kubenswrapper[4870]: I1014 07:21:16.673732 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d00ede0-cd53-4791-9325-c712ee891839-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4d00ede0-cd53-4791-9325-c712ee891839\") " pod="openstack/ceilometer-0" Oct 14 07:21:16 crc kubenswrapper[4870]: I1014 07:21:16.692038 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4d00ede0-cd53-4791-9325-c712ee891839-scripts\") pod \"ceilometer-0\" (UID: \"4d00ede0-cd53-4791-9325-c712ee891839\") " pod="openstack/ceilometer-0" Oct 14 07:21:16 crc kubenswrapper[4870]: I1014 07:21:16.722306 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4d00ede0-cd53-4791-9325-c712ee891839-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4d00ede0-cd53-4791-9325-c712ee891839\") " pod="openstack/ceilometer-0" Oct 14 07:21:16 crc kubenswrapper[4870]: I1014 07:21:16.722567 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x6hww\" (UniqueName: \"kubernetes.io/projected/4d00ede0-cd53-4791-9325-c712ee891839-kube-api-access-x6hww\") pod \"ceilometer-0\" (UID: \"4d00ede0-cd53-4791-9325-c712ee891839\") " pod="openstack/ceilometer-0" Oct 14 07:21:16 crc kubenswrapper[4870]: I1014 07:21:16.724643 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4d00ede0-cd53-4791-9325-c712ee891839-config-data\") pod \"ceilometer-0\" (UID: \"4d00ede0-cd53-4791-9325-c712ee891839\") " pod="openstack/ceilometer-0" Oct 14 07:21:16 crc kubenswrapper[4870]: I1014 07:21:16.738305 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 14 07:21:16 crc kubenswrapper[4870]: I1014 07:21:16.903363 4870 generic.go:334] "Generic (PLEG): container finished" podID="a4ae5421-6680-4e79-ac45-578fdbbc70f2" containerID="3027ab58c61f91be86ef6147ca252c055e21923601e19c2de2f6b7121ee15ada" exitCode=0 Oct 14 07:21:16 crc kubenswrapper[4870]: I1014 07:21:16.903997 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-e66c-account-create-wks4d" event={"ID":"a4ae5421-6680-4e79-ac45-578fdbbc70f2","Type":"ContainerDied","Data":"3027ab58c61f91be86ef6147ca252c055e21923601e19c2de2f6b7121ee15ada"} Oct 14 07:21:16 crc kubenswrapper[4870]: I1014 07:21:16.904039 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-e66c-account-create-wks4d" event={"ID":"a4ae5421-6680-4e79-ac45-578fdbbc70f2","Type":"ContainerStarted","Data":"4ca74c15648617dbfc577dc39194d27b466c9c93c0be1d32fe5cc56febd2996b"} Oct 14 07:21:16 crc kubenswrapper[4870]: I1014 07:21:16.907992 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-52fb-account-create-8kmf7" event={"ID":"93414260-afe1-4af0-8673-d51ec4591776","Type":"ContainerStarted","Data":"8e85e09a8e31af355d315694c672c0246f87999b118cfc6dc5c46c1e00ad0b77"} Oct 14 07:21:16 crc kubenswrapper[4870]: I1014 07:21:16.908035 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-52fb-account-create-8kmf7" event={"ID":"93414260-afe1-4af0-8673-d51ec4591776","Type":"ContainerStarted","Data":"26546c12e0ead0a17576e49c495925e1ba273a833094e4b36f2a78f6502170bf"} Oct 14 07:21:16 crc kubenswrapper[4870]: I1014 07:21:16.909916 4870 generic.go:334] "Generic (PLEG): container finished" podID="93d44382-d93e-4db6-b267-181f5e02682b" containerID="1de36f0eb4ed54f334543bf349d349ea3bac6126555500cdd2af3c26b73dffe7" exitCode=0 Oct 14 07:21:16 crc kubenswrapper[4870]: I1014 07:21:16.909986 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-ce16-account-create-nqg26" event={"ID":"93d44382-d93e-4db6-b267-181f5e02682b","Type":"ContainerDied","Data":"1de36f0eb4ed54f334543bf349d349ea3bac6126555500cdd2af3c26b73dffe7"} Oct 14 07:21:16 crc kubenswrapper[4870]: I1014 07:21:16.910014 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-ce16-account-create-nqg26" event={"ID":"93d44382-d93e-4db6-b267-181f5e02682b","Type":"ContainerStarted","Data":"d3e1f4a624045ee06c5248b6bd44323705ceadbbec8da204e29759acd7595fd9"} Oct 14 07:21:16 crc kubenswrapper[4870]: I1014 07:21:16.957233 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-546b769ccc-fdhsk" Oct 14 07:21:16 crc kubenswrapper[4870]: W1014 07:21:16.979754 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod51de8090_c48e_498c_99ab_794a7527252c.slice/crio-c95d4c3fdb901eaa3800b8cbe01a1d7f2794649a3a2e3635950989ee7a523852 WatchSource:0}: Error finding container c95d4c3fdb901eaa3800b8cbe01a1d7f2794649a3a2e3635950989ee7a523852: Status 404 returned error can't find the container with id c95d4c3fdb901eaa3800b8cbe01a1d7f2794649a3a2e3635950989ee7a523852 Oct 14 07:21:16 crc kubenswrapper[4870]: I1014 07:21:16.996110 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 14 07:21:17 crc kubenswrapper[4870]: I1014 07:21:17.054165 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2ad27e3d-56a2-46e2-a6b0-28b03c5564e8" path="/var/lib/kubelet/pods/2ad27e3d-56a2-46e2-a6b0-28b03c5564e8/volumes" Oct 14 07:21:17 crc kubenswrapper[4870]: I1014 07:21:17.054954 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aad458d7-8d76-421e-8036-22510ec4f99f" path="/var/lib/kubelet/pods/aad458d7-8d76-421e-8036-22510ec4f99f/volumes" Oct 14 07:21:17 crc kubenswrapper[4870]: I1014 07:21:17.057987 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-5dd746bf8d-mvg6f"] Oct 14 07:21:17 crc kubenswrapper[4870]: I1014 07:21:17.058256 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-5dd746bf8d-mvg6f" podUID="8962df13-02a8-4049-bf25-767413eaa98f" containerName="neutron-api" containerID="cri-o://b1d406e0278bc21b8489162b2029d7503109f564a8f097477cfabec3a829c8eb" gracePeriod=30 Oct 14 07:21:17 crc kubenswrapper[4870]: I1014 07:21:17.058399 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-5dd746bf8d-mvg6f" podUID="8962df13-02a8-4049-bf25-767413eaa98f" containerName="neutron-httpd" containerID="cri-o://fd406aca28373d20897ebf5edf751ce84f312fa9b73009dd19544478e754ccde" gracePeriod=30 Oct 14 07:21:17 crc kubenswrapper[4870]: I1014 07:21:17.274335 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 14 07:21:17 crc kubenswrapper[4870]: I1014 07:21:17.858574 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-5485b69bc9-ftmmq" Oct 14 07:21:17 crc kubenswrapper[4870]: I1014 07:21:17.860192 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-5485b69bc9-ftmmq" Oct 14 07:21:17 crc kubenswrapper[4870]: I1014 07:21:17.932913 4870 generic.go:334] "Generic (PLEG): container finished" podID="93414260-afe1-4af0-8673-d51ec4591776" containerID="8e85e09a8e31af355d315694c672c0246f87999b118cfc6dc5c46c1e00ad0b77" exitCode=0 Oct 14 07:21:17 crc kubenswrapper[4870]: I1014 07:21:17.933025 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-52fb-account-create-8kmf7" event={"ID":"93414260-afe1-4af0-8673-d51ec4591776","Type":"ContainerDied","Data":"8e85e09a8e31af355d315694c672c0246f87999b118cfc6dc5c46c1e00ad0b77"} Oct 14 07:21:17 crc kubenswrapper[4870]: I1014 07:21:17.942686 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4d00ede0-cd53-4791-9325-c712ee891839","Type":"ContainerStarted","Data":"18af3f55a518be26512e94d22bb6a5861f9b0a595aab3ee40db0994485db6334"} Oct 14 07:21:17 crc kubenswrapper[4870]: I1014 07:21:17.965959 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"51de8090-c48e-498c-99ab-794a7527252c","Type":"ContainerStarted","Data":"deba864879defc602edfa6ada18f77ffb9b44bf36a5292d8a32f7922bcd29311"} Oct 14 07:21:17 crc kubenswrapper[4870]: I1014 07:21:17.966019 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"51de8090-c48e-498c-99ab-794a7527252c","Type":"ContainerStarted","Data":"c95d4c3fdb901eaa3800b8cbe01a1d7f2794649a3a2e3635950989ee7a523852"} Oct 14 07:21:17 crc kubenswrapper[4870]: I1014 07:21:17.984479 4870 generic.go:334] "Generic (PLEG): container finished" podID="8962df13-02a8-4049-bf25-767413eaa98f" containerID="fd406aca28373d20897ebf5edf751ce84f312fa9b73009dd19544478e754ccde" exitCode=0 Oct 14 07:21:17 crc kubenswrapper[4870]: I1014 07:21:17.984712 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5dd746bf8d-mvg6f" event={"ID":"8962df13-02a8-4049-bf25-767413eaa98f","Type":"ContainerDied","Data":"fd406aca28373d20897ebf5edf751ce84f312fa9b73009dd19544478e754ccde"} Oct 14 07:21:18 crc kubenswrapper[4870]: I1014 07:21:18.433087 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-52fb-account-create-8kmf7" Oct 14 07:21:18 crc kubenswrapper[4870]: I1014 07:21:18.435997 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 14 07:21:18 crc kubenswrapper[4870]: I1014 07:21:18.594172 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-ce16-account-create-nqg26" Oct 14 07:21:18 crc kubenswrapper[4870]: I1014 07:21:18.599114 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-e66c-account-create-wks4d" Oct 14 07:21:18 crc kubenswrapper[4870]: I1014 07:21:18.617392 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ql9vm\" (UniqueName: \"kubernetes.io/projected/93414260-afe1-4af0-8673-d51ec4591776-kube-api-access-ql9vm\") pod \"93414260-afe1-4af0-8673-d51ec4591776\" (UID: \"93414260-afe1-4af0-8673-d51ec4591776\") " Oct 14 07:21:18 crc kubenswrapper[4870]: I1014 07:21:18.633864 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/93414260-afe1-4af0-8673-d51ec4591776-kube-api-access-ql9vm" (OuterVolumeSpecName: "kube-api-access-ql9vm") pod "93414260-afe1-4af0-8673-d51ec4591776" (UID: "93414260-afe1-4af0-8673-d51ec4591776"). InnerVolumeSpecName "kube-api-access-ql9vm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:21:18 crc kubenswrapper[4870]: I1014 07:21:18.718999 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6lwdb\" (UniqueName: \"kubernetes.io/projected/a4ae5421-6680-4e79-ac45-578fdbbc70f2-kube-api-access-6lwdb\") pod \"a4ae5421-6680-4e79-ac45-578fdbbc70f2\" (UID: \"a4ae5421-6680-4e79-ac45-578fdbbc70f2\") " Oct 14 07:21:18 crc kubenswrapper[4870]: I1014 07:21:18.719378 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6vnv8\" (UniqueName: \"kubernetes.io/projected/93d44382-d93e-4db6-b267-181f5e02682b-kube-api-access-6vnv8\") pod \"93d44382-d93e-4db6-b267-181f5e02682b\" (UID: \"93d44382-d93e-4db6-b267-181f5e02682b\") " Oct 14 07:21:18 crc kubenswrapper[4870]: I1014 07:21:18.719842 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ql9vm\" (UniqueName: \"kubernetes.io/projected/93414260-afe1-4af0-8673-d51ec4591776-kube-api-access-ql9vm\") on node \"crc\" DevicePath \"\"" Oct 14 07:21:18 crc kubenswrapper[4870]: I1014 07:21:18.722598 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a4ae5421-6680-4e79-ac45-578fdbbc70f2-kube-api-access-6lwdb" (OuterVolumeSpecName: "kube-api-access-6lwdb") pod "a4ae5421-6680-4e79-ac45-578fdbbc70f2" (UID: "a4ae5421-6680-4e79-ac45-578fdbbc70f2"). InnerVolumeSpecName "kube-api-access-6lwdb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:21:18 crc kubenswrapper[4870]: I1014 07:21:18.728382 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/93d44382-d93e-4db6-b267-181f5e02682b-kube-api-access-6vnv8" (OuterVolumeSpecName: "kube-api-access-6vnv8") pod "93d44382-d93e-4db6-b267-181f5e02682b" (UID: "93d44382-d93e-4db6-b267-181f5e02682b"). InnerVolumeSpecName "kube-api-access-6vnv8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:21:18 crc kubenswrapper[4870]: I1014 07:21:18.821427 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6lwdb\" (UniqueName: \"kubernetes.io/projected/a4ae5421-6680-4e79-ac45-578fdbbc70f2-kube-api-access-6lwdb\") on node \"crc\" DevicePath \"\"" Oct 14 07:21:18 crc kubenswrapper[4870]: I1014 07:21:18.821473 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6vnv8\" (UniqueName: \"kubernetes.io/projected/93d44382-d93e-4db6-b267-181f5e02682b-kube-api-access-6vnv8\") on node \"crc\" DevicePath \"\"" Oct 14 07:21:18 crc kubenswrapper[4870]: I1014 07:21:18.996415 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-e66c-account-create-wks4d" event={"ID":"a4ae5421-6680-4e79-ac45-578fdbbc70f2","Type":"ContainerDied","Data":"4ca74c15648617dbfc577dc39194d27b466c9c93c0be1d32fe5cc56febd2996b"} Oct 14 07:21:18 crc kubenswrapper[4870]: I1014 07:21:18.996484 4870 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4ca74c15648617dbfc577dc39194d27b466c9c93c0be1d32fe5cc56febd2996b" Oct 14 07:21:18 crc kubenswrapper[4870]: I1014 07:21:18.996429 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-e66c-account-create-wks4d" Oct 14 07:21:18 crc kubenswrapper[4870]: I1014 07:21:18.999642 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-52fb-account-create-8kmf7" event={"ID":"93414260-afe1-4af0-8673-d51ec4591776","Type":"ContainerDied","Data":"26546c12e0ead0a17576e49c495925e1ba273a833094e4b36f2a78f6502170bf"} Oct 14 07:21:18 crc kubenswrapper[4870]: I1014 07:21:18.999680 4870 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="26546c12e0ead0a17576e49c495925e1ba273a833094e4b36f2a78f6502170bf" Oct 14 07:21:18 crc kubenswrapper[4870]: I1014 07:21:18.999728 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-52fb-account-create-8kmf7" Oct 14 07:21:19 crc kubenswrapper[4870]: I1014 07:21:19.001641 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-ce16-account-create-nqg26" event={"ID":"93d44382-d93e-4db6-b267-181f5e02682b","Type":"ContainerDied","Data":"d3e1f4a624045ee06c5248b6bd44323705ceadbbec8da204e29759acd7595fd9"} Oct 14 07:21:19 crc kubenswrapper[4870]: I1014 07:21:19.001673 4870 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d3e1f4a624045ee06c5248b6bd44323705ceadbbec8da204e29759acd7595fd9" Oct 14 07:21:19 crc kubenswrapper[4870]: I1014 07:21:19.001727 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-ce16-account-create-nqg26" Oct 14 07:21:19 crc kubenswrapper[4870]: I1014 07:21:19.005478 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4d00ede0-cd53-4791-9325-c712ee891839","Type":"ContainerStarted","Data":"2bcc81a3a9d5c820c0bcd54b459b4f42446c2e0c7a6b7503e11f9c3e6552fadb"} Oct 14 07:21:19 crc kubenswrapper[4870]: I1014 07:21:19.008874 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"51de8090-c48e-498c-99ab-794a7527252c","Type":"ContainerStarted","Data":"571dfedb34f9303684e8f8fe09485d339b6c363eaa75f6950ae538f2526d0129"} Oct 14 07:21:19 crc kubenswrapper[4870]: I1014 07:21:19.010369 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Oct 14 07:21:19 crc kubenswrapper[4870]: I1014 07:21:19.032731 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=4.032715451 podStartE2EDuration="4.032715451s" podCreationTimestamp="2025-10-14 07:21:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:21:19.026798623 +0000 UTC m=+1214.724159004" watchObservedRunningTime="2025-10-14 07:21:19.032715451 +0000 UTC m=+1214.730075822" Oct 14 07:21:19 crc kubenswrapper[4870]: I1014 07:21:19.524657 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5b7d94f5b5-4qfsd" Oct 14 07:21:19 crc kubenswrapper[4870]: I1014 07:21:19.600718 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-85fd94d64c-nr8rq"] Oct 14 07:21:19 crc kubenswrapper[4870]: I1014 07:21:19.600959 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-85fd94d64c-nr8rq" podUID="d42ee9f9-390b-437c-a48a-ff7ead370f61" containerName="dnsmasq-dns" containerID="cri-o://02d14106fdda254dcaf2de464737ed7252a713b703f968764194d49711c7bb76" gracePeriod=10 Oct 14 07:21:19 crc kubenswrapper[4870]: I1014 07:21:19.769600 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Oct 14 07:21:19 crc kubenswrapper[4870]: I1014 07:21:19.810135 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 14 07:21:20 crc kubenswrapper[4870]: I1014 07:21:20.020123 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4d00ede0-cd53-4791-9325-c712ee891839","Type":"ContainerStarted","Data":"ccd5820b235e529b2ee667f2bc0cb4aaa6d87fc0297a583d5753340e0d346237"} Oct 14 07:21:20 crc kubenswrapper[4870]: I1014 07:21:20.022015 4870 generic.go:334] "Generic (PLEG): container finished" podID="d42ee9f9-390b-437c-a48a-ff7ead370f61" containerID="02d14106fdda254dcaf2de464737ed7252a713b703f968764194d49711c7bb76" exitCode=0 Oct 14 07:21:20 crc kubenswrapper[4870]: I1014 07:21:20.022090 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85fd94d64c-nr8rq" event={"ID":"d42ee9f9-390b-437c-a48a-ff7ead370f61","Type":"ContainerDied","Data":"02d14106fdda254dcaf2de464737ed7252a713b703f968764194d49711c7bb76"} Oct 14 07:21:20 crc kubenswrapper[4870]: I1014 07:21:20.022210 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="f04fb303-8db7-4d2c-a864-817494a7222b" containerName="cinder-scheduler" containerID="cri-o://b8b3775a268dcc4ebc06b8c9a67360595bbc18617ec1fd7cabd8c847229a0730" gracePeriod=30 Oct 14 07:21:20 crc kubenswrapper[4870]: I1014 07:21:20.022328 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="f04fb303-8db7-4d2c-a864-817494a7222b" containerName="probe" containerID="cri-o://ed5d51b7faa1c7111cd571af1833b5840ccac99e3be071ef421e46b99f9cb4b7" gracePeriod=30 Oct 14 07:21:20 crc kubenswrapper[4870]: I1014 07:21:20.494215 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-7jfpt"] Oct 14 07:21:20 crc kubenswrapper[4870]: E1014 07:21:20.494776 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a4ae5421-6680-4e79-ac45-578fdbbc70f2" containerName="mariadb-account-create" Oct 14 07:21:20 crc kubenswrapper[4870]: I1014 07:21:20.494803 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="a4ae5421-6680-4e79-ac45-578fdbbc70f2" containerName="mariadb-account-create" Oct 14 07:21:20 crc kubenswrapper[4870]: E1014 07:21:20.494825 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93d44382-d93e-4db6-b267-181f5e02682b" containerName="mariadb-account-create" Oct 14 07:21:20 crc kubenswrapper[4870]: I1014 07:21:20.494836 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="93d44382-d93e-4db6-b267-181f5e02682b" containerName="mariadb-account-create" Oct 14 07:21:20 crc kubenswrapper[4870]: E1014 07:21:20.494844 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93414260-afe1-4af0-8673-d51ec4591776" containerName="mariadb-account-create" Oct 14 07:21:20 crc kubenswrapper[4870]: I1014 07:21:20.494852 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="93414260-afe1-4af0-8673-d51ec4591776" containerName="mariadb-account-create" Oct 14 07:21:20 crc kubenswrapper[4870]: I1014 07:21:20.495068 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="a4ae5421-6680-4e79-ac45-578fdbbc70f2" containerName="mariadb-account-create" Oct 14 07:21:20 crc kubenswrapper[4870]: I1014 07:21:20.495090 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="93d44382-d93e-4db6-b267-181f5e02682b" containerName="mariadb-account-create" Oct 14 07:21:20 crc kubenswrapper[4870]: I1014 07:21:20.495108 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="93414260-afe1-4af0-8673-d51ec4591776" containerName="mariadb-account-create" Oct 14 07:21:20 crc kubenswrapper[4870]: I1014 07:21:20.495885 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-7jfpt" Oct 14 07:21:20 crc kubenswrapper[4870]: I1014 07:21:20.498506 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Oct 14 07:21:20 crc kubenswrapper[4870]: I1014 07:21:20.498696 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Oct 14 07:21:20 crc kubenswrapper[4870]: I1014 07:21:20.502820 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-7jfpt"] Oct 14 07:21:20 crc kubenswrapper[4870]: I1014 07:21:20.503238 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-bbjpq" Oct 14 07:21:20 crc kubenswrapper[4870]: I1014 07:21:20.657657 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bdb3c945-edde-4d43-b3a5-234aaabf00d5-scripts\") pod \"nova-cell0-conductor-db-sync-7jfpt\" (UID: \"bdb3c945-edde-4d43-b3a5-234aaabf00d5\") " pod="openstack/nova-cell0-conductor-db-sync-7jfpt" Oct 14 07:21:20 crc kubenswrapper[4870]: I1014 07:21:20.657732 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bdb3c945-edde-4d43-b3a5-234aaabf00d5-config-data\") pod \"nova-cell0-conductor-db-sync-7jfpt\" (UID: \"bdb3c945-edde-4d43-b3a5-234aaabf00d5\") " pod="openstack/nova-cell0-conductor-db-sync-7jfpt" Oct 14 07:21:20 crc kubenswrapper[4870]: I1014 07:21:20.657788 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bdb3c945-edde-4d43-b3a5-234aaabf00d5-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-7jfpt\" (UID: \"bdb3c945-edde-4d43-b3a5-234aaabf00d5\") " pod="openstack/nova-cell0-conductor-db-sync-7jfpt" Oct 14 07:21:20 crc kubenswrapper[4870]: I1014 07:21:20.657861 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-trbf7\" (UniqueName: \"kubernetes.io/projected/bdb3c945-edde-4d43-b3a5-234aaabf00d5-kube-api-access-trbf7\") pod \"nova-cell0-conductor-db-sync-7jfpt\" (UID: \"bdb3c945-edde-4d43-b3a5-234aaabf00d5\") " pod="openstack/nova-cell0-conductor-db-sync-7jfpt" Oct 14 07:21:20 crc kubenswrapper[4870]: I1014 07:21:20.759415 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-trbf7\" (UniqueName: \"kubernetes.io/projected/bdb3c945-edde-4d43-b3a5-234aaabf00d5-kube-api-access-trbf7\") pod \"nova-cell0-conductor-db-sync-7jfpt\" (UID: \"bdb3c945-edde-4d43-b3a5-234aaabf00d5\") " pod="openstack/nova-cell0-conductor-db-sync-7jfpt" Oct 14 07:21:20 crc kubenswrapper[4870]: I1014 07:21:20.759528 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bdb3c945-edde-4d43-b3a5-234aaabf00d5-scripts\") pod \"nova-cell0-conductor-db-sync-7jfpt\" (UID: \"bdb3c945-edde-4d43-b3a5-234aaabf00d5\") " pod="openstack/nova-cell0-conductor-db-sync-7jfpt" Oct 14 07:21:20 crc kubenswrapper[4870]: I1014 07:21:20.759576 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bdb3c945-edde-4d43-b3a5-234aaabf00d5-config-data\") pod \"nova-cell0-conductor-db-sync-7jfpt\" (UID: \"bdb3c945-edde-4d43-b3a5-234aaabf00d5\") " pod="openstack/nova-cell0-conductor-db-sync-7jfpt" Oct 14 07:21:20 crc kubenswrapper[4870]: I1014 07:21:20.759623 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bdb3c945-edde-4d43-b3a5-234aaabf00d5-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-7jfpt\" (UID: \"bdb3c945-edde-4d43-b3a5-234aaabf00d5\") " pod="openstack/nova-cell0-conductor-db-sync-7jfpt" Oct 14 07:21:20 crc kubenswrapper[4870]: I1014 07:21:20.765948 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bdb3c945-edde-4d43-b3a5-234aaabf00d5-scripts\") pod \"nova-cell0-conductor-db-sync-7jfpt\" (UID: \"bdb3c945-edde-4d43-b3a5-234aaabf00d5\") " pod="openstack/nova-cell0-conductor-db-sync-7jfpt" Oct 14 07:21:20 crc kubenswrapper[4870]: I1014 07:21:20.769998 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bdb3c945-edde-4d43-b3a5-234aaabf00d5-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-7jfpt\" (UID: \"bdb3c945-edde-4d43-b3a5-234aaabf00d5\") " pod="openstack/nova-cell0-conductor-db-sync-7jfpt" Oct 14 07:21:20 crc kubenswrapper[4870]: I1014 07:21:20.777549 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bdb3c945-edde-4d43-b3a5-234aaabf00d5-config-data\") pod \"nova-cell0-conductor-db-sync-7jfpt\" (UID: \"bdb3c945-edde-4d43-b3a5-234aaabf00d5\") " pod="openstack/nova-cell0-conductor-db-sync-7jfpt" Oct 14 07:21:20 crc kubenswrapper[4870]: I1014 07:21:20.789885 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-trbf7\" (UniqueName: \"kubernetes.io/projected/bdb3c945-edde-4d43-b3a5-234aaabf00d5-kube-api-access-trbf7\") pod \"nova-cell0-conductor-db-sync-7jfpt\" (UID: \"bdb3c945-edde-4d43-b3a5-234aaabf00d5\") " pod="openstack/nova-cell0-conductor-db-sync-7jfpt" Oct 14 07:21:20 crc kubenswrapper[4870]: I1014 07:21:20.818081 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-7jfpt" Oct 14 07:21:21 crc kubenswrapper[4870]: I1014 07:21:21.063972 4870 generic.go:334] "Generic (PLEG): container finished" podID="f04fb303-8db7-4d2c-a864-817494a7222b" containerID="ed5d51b7faa1c7111cd571af1833b5840ccac99e3be071ef421e46b99f9cb4b7" exitCode=0 Oct 14 07:21:21 crc kubenswrapper[4870]: I1014 07:21:21.064179 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"f04fb303-8db7-4d2c-a864-817494a7222b","Type":"ContainerDied","Data":"ed5d51b7faa1c7111cd571af1833b5840ccac99e3be071ef421e46b99f9cb4b7"} Oct 14 07:21:21 crc kubenswrapper[4870]: I1014 07:21:21.380770 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-7jfpt"] Oct 14 07:21:21 crc kubenswrapper[4870]: I1014 07:21:21.660111 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85fd94d64c-nr8rq" Oct 14 07:21:21 crc kubenswrapper[4870]: I1014 07:21:21.816009 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d42ee9f9-390b-437c-a48a-ff7ead370f61-config\") pod \"d42ee9f9-390b-437c-a48a-ff7ead370f61\" (UID: \"d42ee9f9-390b-437c-a48a-ff7ead370f61\") " Oct 14 07:21:21 crc kubenswrapper[4870]: I1014 07:21:21.816365 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d42ee9f9-390b-437c-a48a-ff7ead370f61-dns-swift-storage-0\") pod \"d42ee9f9-390b-437c-a48a-ff7ead370f61\" (UID: \"d42ee9f9-390b-437c-a48a-ff7ead370f61\") " Oct 14 07:21:21 crc kubenswrapper[4870]: I1014 07:21:21.816446 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kbwxw\" (UniqueName: \"kubernetes.io/projected/d42ee9f9-390b-437c-a48a-ff7ead370f61-kube-api-access-kbwxw\") pod \"d42ee9f9-390b-437c-a48a-ff7ead370f61\" (UID: \"d42ee9f9-390b-437c-a48a-ff7ead370f61\") " Oct 14 07:21:21 crc kubenswrapper[4870]: I1014 07:21:21.816507 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d42ee9f9-390b-437c-a48a-ff7ead370f61-ovsdbserver-sb\") pod \"d42ee9f9-390b-437c-a48a-ff7ead370f61\" (UID: \"d42ee9f9-390b-437c-a48a-ff7ead370f61\") " Oct 14 07:21:21 crc kubenswrapper[4870]: I1014 07:21:21.816526 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d42ee9f9-390b-437c-a48a-ff7ead370f61-ovsdbserver-nb\") pod \"d42ee9f9-390b-437c-a48a-ff7ead370f61\" (UID: \"d42ee9f9-390b-437c-a48a-ff7ead370f61\") " Oct 14 07:21:21 crc kubenswrapper[4870]: I1014 07:21:21.816622 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d42ee9f9-390b-437c-a48a-ff7ead370f61-dns-svc\") pod \"d42ee9f9-390b-437c-a48a-ff7ead370f61\" (UID: \"d42ee9f9-390b-437c-a48a-ff7ead370f61\") " Oct 14 07:21:21 crc kubenswrapper[4870]: I1014 07:21:21.834785 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d42ee9f9-390b-437c-a48a-ff7ead370f61-kube-api-access-kbwxw" (OuterVolumeSpecName: "kube-api-access-kbwxw") pod "d42ee9f9-390b-437c-a48a-ff7ead370f61" (UID: "d42ee9f9-390b-437c-a48a-ff7ead370f61"). InnerVolumeSpecName "kube-api-access-kbwxw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:21:21 crc kubenswrapper[4870]: I1014 07:21:21.888252 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d42ee9f9-390b-437c-a48a-ff7ead370f61-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "d42ee9f9-390b-437c-a48a-ff7ead370f61" (UID: "d42ee9f9-390b-437c-a48a-ff7ead370f61"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:21:21 crc kubenswrapper[4870]: I1014 07:21:21.894423 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d42ee9f9-390b-437c-a48a-ff7ead370f61-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "d42ee9f9-390b-437c-a48a-ff7ead370f61" (UID: "d42ee9f9-390b-437c-a48a-ff7ead370f61"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:21:21 crc kubenswrapper[4870]: I1014 07:21:21.906671 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d42ee9f9-390b-437c-a48a-ff7ead370f61-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "d42ee9f9-390b-437c-a48a-ff7ead370f61" (UID: "d42ee9f9-390b-437c-a48a-ff7ead370f61"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:21:21 crc kubenswrapper[4870]: I1014 07:21:21.912766 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d42ee9f9-390b-437c-a48a-ff7ead370f61-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "d42ee9f9-390b-437c-a48a-ff7ead370f61" (UID: "d42ee9f9-390b-437c-a48a-ff7ead370f61"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:21:21 crc kubenswrapper[4870]: I1014 07:21:21.920927 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kbwxw\" (UniqueName: \"kubernetes.io/projected/d42ee9f9-390b-437c-a48a-ff7ead370f61-kube-api-access-kbwxw\") on node \"crc\" DevicePath \"\"" Oct 14 07:21:21 crc kubenswrapper[4870]: I1014 07:21:21.920957 4870 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d42ee9f9-390b-437c-a48a-ff7ead370f61-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 14 07:21:21 crc kubenswrapper[4870]: I1014 07:21:21.920966 4870 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d42ee9f9-390b-437c-a48a-ff7ead370f61-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 14 07:21:21 crc kubenswrapper[4870]: I1014 07:21:21.920974 4870 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d42ee9f9-390b-437c-a48a-ff7ead370f61-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 14 07:21:21 crc kubenswrapper[4870]: I1014 07:21:21.920983 4870 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d42ee9f9-390b-437c-a48a-ff7ead370f61-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 14 07:21:21 crc kubenswrapper[4870]: I1014 07:21:21.926885 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d42ee9f9-390b-437c-a48a-ff7ead370f61-config" (OuterVolumeSpecName: "config") pod "d42ee9f9-390b-437c-a48a-ff7ead370f61" (UID: "d42ee9f9-390b-437c-a48a-ff7ead370f61"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:21:22 crc kubenswrapper[4870]: I1014 07:21:22.023137 4870 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d42ee9f9-390b-437c-a48a-ff7ead370f61-config\") on node \"crc\" DevicePath \"\"" Oct 14 07:21:22 crc kubenswrapper[4870]: I1014 07:21:22.074186 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4d00ede0-cd53-4791-9325-c712ee891839","Type":"ContainerStarted","Data":"c07b4185cb742a8d604acc1b5ef508fa09c68ed0229dc00df2a8b8a8be770694"} Oct 14 07:21:22 crc kubenswrapper[4870]: I1014 07:21:22.076682 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85fd94d64c-nr8rq" event={"ID":"d42ee9f9-390b-437c-a48a-ff7ead370f61","Type":"ContainerDied","Data":"7d4427e577f38e681045883f59f7b2021995b5081d3fd7ec45ee57d5b7a3ee25"} Oct 14 07:21:22 crc kubenswrapper[4870]: I1014 07:21:22.076736 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85fd94d64c-nr8rq" Oct 14 07:21:22 crc kubenswrapper[4870]: I1014 07:21:22.076814 4870 scope.go:117] "RemoveContainer" containerID="02d14106fdda254dcaf2de464737ed7252a713b703f968764194d49711c7bb76" Oct 14 07:21:22 crc kubenswrapper[4870]: I1014 07:21:22.078203 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-7jfpt" event={"ID":"bdb3c945-edde-4d43-b3a5-234aaabf00d5","Type":"ContainerStarted","Data":"3b32a42b8d5e3833de5b1a277eafebef3875d31df1829e6d4edf1bee07df70a2"} Oct 14 07:21:22 crc kubenswrapper[4870]: I1014 07:21:22.107326 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-85fd94d64c-nr8rq"] Oct 14 07:21:22 crc kubenswrapper[4870]: I1014 07:21:22.110719 4870 scope.go:117] "RemoveContainer" containerID="9a50e7bdf997907fd2e3f2bef9580f2c3d7493ec4d3de99188be8d372a77a601" Oct 14 07:21:22 crc kubenswrapper[4870]: I1014 07:21:22.123082 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-85fd94d64c-nr8rq"] Oct 14 07:21:23 crc kubenswrapper[4870]: I1014 07:21:23.061037 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d42ee9f9-390b-437c-a48a-ff7ead370f61" path="/var/lib/kubelet/pods/d42ee9f9-390b-437c-a48a-ff7ead370f61/volumes" Oct 14 07:21:23 crc kubenswrapper[4870]: I1014 07:21:23.113667 4870 generic.go:334] "Generic (PLEG): container finished" podID="8962df13-02a8-4049-bf25-767413eaa98f" containerID="b1d406e0278bc21b8489162b2029d7503109f564a8f097477cfabec3a829c8eb" exitCode=0 Oct 14 07:21:23 crc kubenswrapper[4870]: I1014 07:21:23.113748 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5dd746bf8d-mvg6f" event={"ID":"8962df13-02a8-4049-bf25-767413eaa98f","Type":"ContainerDied","Data":"b1d406e0278bc21b8489162b2029d7503109f564a8f097477cfabec3a829c8eb"} Oct 14 07:21:23 crc kubenswrapper[4870]: I1014 07:21:23.247713 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5dd746bf8d-mvg6f" Oct 14 07:21:23 crc kubenswrapper[4870]: I1014 07:21:23.347353 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gmddt\" (UniqueName: \"kubernetes.io/projected/8962df13-02a8-4049-bf25-767413eaa98f-kube-api-access-gmddt\") pod \"8962df13-02a8-4049-bf25-767413eaa98f\" (UID: \"8962df13-02a8-4049-bf25-767413eaa98f\") " Oct 14 07:21:23 crc kubenswrapper[4870]: I1014 07:21:23.347424 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/8962df13-02a8-4049-bf25-767413eaa98f-ovndb-tls-certs\") pod \"8962df13-02a8-4049-bf25-767413eaa98f\" (UID: \"8962df13-02a8-4049-bf25-767413eaa98f\") " Oct 14 07:21:23 crc kubenswrapper[4870]: I1014 07:21:23.347518 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/8962df13-02a8-4049-bf25-767413eaa98f-httpd-config\") pod \"8962df13-02a8-4049-bf25-767413eaa98f\" (UID: \"8962df13-02a8-4049-bf25-767413eaa98f\") " Oct 14 07:21:23 crc kubenswrapper[4870]: I1014 07:21:23.347544 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/8962df13-02a8-4049-bf25-767413eaa98f-config\") pod \"8962df13-02a8-4049-bf25-767413eaa98f\" (UID: \"8962df13-02a8-4049-bf25-767413eaa98f\") " Oct 14 07:21:23 crc kubenswrapper[4870]: I1014 07:21:23.347587 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8962df13-02a8-4049-bf25-767413eaa98f-combined-ca-bundle\") pod \"8962df13-02a8-4049-bf25-767413eaa98f\" (UID: \"8962df13-02a8-4049-bf25-767413eaa98f\") " Oct 14 07:21:23 crc kubenswrapper[4870]: I1014 07:21:23.352463 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8962df13-02a8-4049-bf25-767413eaa98f-kube-api-access-gmddt" (OuterVolumeSpecName: "kube-api-access-gmddt") pod "8962df13-02a8-4049-bf25-767413eaa98f" (UID: "8962df13-02a8-4049-bf25-767413eaa98f"). InnerVolumeSpecName "kube-api-access-gmddt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:21:23 crc kubenswrapper[4870]: I1014 07:21:23.353527 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8962df13-02a8-4049-bf25-767413eaa98f-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "8962df13-02a8-4049-bf25-767413eaa98f" (UID: "8962df13-02a8-4049-bf25-767413eaa98f"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:21:23 crc kubenswrapper[4870]: I1014 07:21:23.427052 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8962df13-02a8-4049-bf25-767413eaa98f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8962df13-02a8-4049-bf25-767413eaa98f" (UID: "8962df13-02a8-4049-bf25-767413eaa98f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:21:23 crc kubenswrapper[4870]: I1014 07:21:23.429009 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8962df13-02a8-4049-bf25-767413eaa98f-config" (OuterVolumeSpecName: "config") pod "8962df13-02a8-4049-bf25-767413eaa98f" (UID: "8962df13-02a8-4049-bf25-767413eaa98f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:21:23 crc kubenswrapper[4870]: I1014 07:21:23.453299 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gmddt\" (UniqueName: \"kubernetes.io/projected/8962df13-02a8-4049-bf25-767413eaa98f-kube-api-access-gmddt\") on node \"crc\" DevicePath \"\"" Oct 14 07:21:23 crc kubenswrapper[4870]: I1014 07:21:23.453340 4870 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/8962df13-02a8-4049-bf25-767413eaa98f-httpd-config\") on node \"crc\" DevicePath \"\"" Oct 14 07:21:23 crc kubenswrapper[4870]: I1014 07:21:23.453350 4870 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/8962df13-02a8-4049-bf25-767413eaa98f-config\") on node \"crc\" DevicePath \"\"" Oct 14 07:21:23 crc kubenswrapper[4870]: I1014 07:21:23.453359 4870 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8962df13-02a8-4049-bf25-767413eaa98f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 07:21:23 crc kubenswrapper[4870]: I1014 07:21:23.459296 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 14 07:21:23 crc kubenswrapper[4870]: I1014 07:21:23.460114 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="452a6ce6-d9dc-4455-9304-414701c8895f" containerName="glance-log" containerID="cri-o://1fde5cdab42d5388f18054e6d6d66a520f1f153b7c41418b4d919ba2fd643643" gracePeriod=30 Oct 14 07:21:23 crc kubenswrapper[4870]: I1014 07:21:23.460313 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="452a6ce6-d9dc-4455-9304-414701c8895f" containerName="glance-httpd" containerID="cri-o://5db90be32cea50de9fa48a1c7946854f77a0364acaf6f2929af1ec8f226fb41f" gracePeriod=30 Oct 14 07:21:23 crc kubenswrapper[4870]: I1014 07:21:23.508060 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8962df13-02a8-4049-bf25-767413eaa98f-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "8962df13-02a8-4049-bf25-767413eaa98f" (UID: "8962df13-02a8-4049-bf25-767413eaa98f"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:21:23 crc kubenswrapper[4870]: I1014 07:21:23.516476 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 14 07:21:23 crc kubenswrapper[4870]: I1014 07:21:23.555680 4870 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/8962df13-02a8-4049-bf25-767413eaa98f-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 14 07:21:23 crc kubenswrapper[4870]: I1014 07:21:23.657054 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f04fb303-8db7-4d2c-a864-817494a7222b-combined-ca-bundle\") pod \"f04fb303-8db7-4d2c-a864-817494a7222b\" (UID: \"f04fb303-8db7-4d2c-a864-817494a7222b\") " Oct 14 07:21:23 crc kubenswrapper[4870]: I1014 07:21:23.657142 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f04fb303-8db7-4d2c-a864-817494a7222b-etc-machine-id\") pod \"f04fb303-8db7-4d2c-a864-817494a7222b\" (UID: \"f04fb303-8db7-4d2c-a864-817494a7222b\") " Oct 14 07:21:23 crc kubenswrapper[4870]: I1014 07:21:23.657214 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f04fb303-8db7-4d2c-a864-817494a7222b-config-data-custom\") pod \"f04fb303-8db7-4d2c-a864-817494a7222b\" (UID: \"f04fb303-8db7-4d2c-a864-817494a7222b\") " Oct 14 07:21:23 crc kubenswrapper[4870]: I1014 07:21:23.657244 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2ctzq\" (UniqueName: \"kubernetes.io/projected/f04fb303-8db7-4d2c-a864-817494a7222b-kube-api-access-2ctzq\") pod \"f04fb303-8db7-4d2c-a864-817494a7222b\" (UID: \"f04fb303-8db7-4d2c-a864-817494a7222b\") " Oct 14 07:21:23 crc kubenswrapper[4870]: I1014 07:21:23.657349 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f04fb303-8db7-4d2c-a864-817494a7222b-config-data\") pod \"f04fb303-8db7-4d2c-a864-817494a7222b\" (UID: \"f04fb303-8db7-4d2c-a864-817494a7222b\") " Oct 14 07:21:23 crc kubenswrapper[4870]: I1014 07:21:23.657409 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f04fb303-8db7-4d2c-a864-817494a7222b-scripts\") pod \"f04fb303-8db7-4d2c-a864-817494a7222b\" (UID: \"f04fb303-8db7-4d2c-a864-817494a7222b\") " Oct 14 07:21:23 crc kubenswrapper[4870]: I1014 07:21:23.659195 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f04fb303-8db7-4d2c-a864-817494a7222b-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "f04fb303-8db7-4d2c-a864-817494a7222b" (UID: "f04fb303-8db7-4d2c-a864-817494a7222b"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 14 07:21:23 crc kubenswrapper[4870]: I1014 07:21:23.662411 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f04fb303-8db7-4d2c-a864-817494a7222b-scripts" (OuterVolumeSpecName: "scripts") pod "f04fb303-8db7-4d2c-a864-817494a7222b" (UID: "f04fb303-8db7-4d2c-a864-817494a7222b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:21:23 crc kubenswrapper[4870]: I1014 07:21:23.663613 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f04fb303-8db7-4d2c-a864-817494a7222b-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "f04fb303-8db7-4d2c-a864-817494a7222b" (UID: "f04fb303-8db7-4d2c-a864-817494a7222b"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:21:23 crc kubenswrapper[4870]: I1014 07:21:23.663701 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f04fb303-8db7-4d2c-a864-817494a7222b-kube-api-access-2ctzq" (OuterVolumeSpecName: "kube-api-access-2ctzq") pod "f04fb303-8db7-4d2c-a864-817494a7222b" (UID: "f04fb303-8db7-4d2c-a864-817494a7222b"). InnerVolumeSpecName "kube-api-access-2ctzq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:21:23 crc kubenswrapper[4870]: I1014 07:21:23.711368 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f04fb303-8db7-4d2c-a864-817494a7222b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f04fb303-8db7-4d2c-a864-817494a7222b" (UID: "f04fb303-8db7-4d2c-a864-817494a7222b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:21:23 crc kubenswrapper[4870]: I1014 07:21:23.759377 4870 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f04fb303-8db7-4d2c-a864-817494a7222b-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 07:21:23 crc kubenswrapper[4870]: I1014 07:21:23.759419 4870 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f04fb303-8db7-4d2c-a864-817494a7222b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 07:21:23 crc kubenswrapper[4870]: I1014 07:21:23.759449 4870 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f04fb303-8db7-4d2c-a864-817494a7222b-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 14 07:21:23 crc kubenswrapper[4870]: I1014 07:21:23.759461 4870 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f04fb303-8db7-4d2c-a864-817494a7222b-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 14 07:21:23 crc kubenswrapper[4870]: I1014 07:21:23.759474 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2ctzq\" (UniqueName: \"kubernetes.io/projected/f04fb303-8db7-4d2c-a864-817494a7222b-kube-api-access-2ctzq\") on node \"crc\" DevicePath \"\"" Oct 14 07:21:23 crc kubenswrapper[4870]: I1014 07:21:23.763979 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f04fb303-8db7-4d2c-a864-817494a7222b-config-data" (OuterVolumeSpecName: "config-data") pod "f04fb303-8db7-4d2c-a864-817494a7222b" (UID: "f04fb303-8db7-4d2c-a864-817494a7222b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:21:23 crc kubenswrapper[4870]: I1014 07:21:23.860859 4870 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f04fb303-8db7-4d2c-a864-817494a7222b-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 07:21:23 crc kubenswrapper[4870]: I1014 07:21:23.951305 4870 patch_prober.go:28] interesting pod/machine-config-daemon-7tvc7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 07:21:23 crc kubenswrapper[4870]: I1014 07:21:23.951356 4870 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 07:21:24 crc kubenswrapper[4870]: I1014 07:21:24.137257 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5dd746bf8d-mvg6f" event={"ID":"8962df13-02a8-4049-bf25-767413eaa98f","Type":"ContainerDied","Data":"fc215b00e15e2e22b7b6f6158ff8bbac03f014ae65563e668ff453f53f611ded"} Oct 14 07:21:24 crc kubenswrapper[4870]: I1014 07:21:24.137317 4870 scope.go:117] "RemoveContainer" containerID="fd406aca28373d20897ebf5edf751ce84f312fa9b73009dd19544478e754ccde" Oct 14 07:21:24 crc kubenswrapper[4870]: I1014 07:21:24.137408 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5dd746bf8d-mvg6f" Oct 14 07:21:24 crc kubenswrapper[4870]: I1014 07:21:24.142569 4870 generic.go:334] "Generic (PLEG): container finished" podID="f04fb303-8db7-4d2c-a864-817494a7222b" containerID="b8b3775a268dcc4ebc06b8c9a67360595bbc18617ec1fd7cabd8c847229a0730" exitCode=0 Oct 14 07:21:24 crc kubenswrapper[4870]: I1014 07:21:24.142662 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"f04fb303-8db7-4d2c-a864-817494a7222b","Type":"ContainerDied","Data":"b8b3775a268dcc4ebc06b8c9a67360595bbc18617ec1fd7cabd8c847229a0730"} Oct 14 07:21:24 crc kubenswrapper[4870]: I1014 07:21:24.142696 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 14 07:21:24 crc kubenswrapper[4870]: I1014 07:21:24.142720 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"f04fb303-8db7-4d2c-a864-817494a7222b","Type":"ContainerDied","Data":"0793837b9e19c6e7c28ba4e548221f38ff6f35c3ed269be3912feeca82aa06e6"} Oct 14 07:21:24 crc kubenswrapper[4870]: I1014 07:21:24.145019 4870 generic.go:334] "Generic (PLEG): container finished" podID="452a6ce6-d9dc-4455-9304-414701c8895f" containerID="1fde5cdab42d5388f18054e6d6d66a520f1f153b7c41418b4d919ba2fd643643" exitCode=143 Oct 14 07:21:24 crc kubenswrapper[4870]: I1014 07:21:24.145088 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"452a6ce6-d9dc-4455-9304-414701c8895f","Type":"ContainerDied","Data":"1fde5cdab42d5388f18054e6d6d66a520f1f153b7c41418b4d919ba2fd643643"} Oct 14 07:21:24 crc kubenswrapper[4870]: I1014 07:21:24.149743 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4d00ede0-cd53-4791-9325-c712ee891839","Type":"ContainerStarted","Data":"fdda30c816580a958c0638fccaade56d725b6b8fffdcdb85825f921582f25f4e"} Oct 14 07:21:24 crc kubenswrapper[4870]: I1014 07:21:24.149919 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4d00ede0-cd53-4791-9325-c712ee891839" containerName="ceilometer-central-agent" containerID="cri-o://2bcc81a3a9d5c820c0bcd54b459b4f42446c2e0c7a6b7503e11f9c3e6552fadb" gracePeriod=30 Oct 14 07:21:24 crc kubenswrapper[4870]: I1014 07:21:24.149985 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 14 07:21:24 crc kubenswrapper[4870]: I1014 07:21:24.150006 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4d00ede0-cd53-4791-9325-c712ee891839" containerName="sg-core" containerID="cri-o://c07b4185cb742a8d604acc1b5ef508fa09c68ed0229dc00df2a8b8a8be770694" gracePeriod=30 Oct 14 07:21:24 crc kubenswrapper[4870]: I1014 07:21:24.150024 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4d00ede0-cd53-4791-9325-c712ee891839" containerName="ceilometer-notification-agent" containerID="cri-o://ccd5820b235e529b2ee667f2bc0cb4aaa6d87fc0297a583d5753340e0d346237" gracePeriod=30 Oct 14 07:21:24 crc kubenswrapper[4870]: I1014 07:21:24.150573 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4d00ede0-cd53-4791-9325-c712ee891839" containerName="proxy-httpd" containerID="cri-o://fdda30c816580a958c0638fccaade56d725b6b8fffdcdb85825f921582f25f4e" gracePeriod=30 Oct 14 07:21:24 crc kubenswrapper[4870]: I1014 07:21:24.190644 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.509436192 podStartE2EDuration="8.190616167s" podCreationTimestamp="2025-10-14 07:21:16 +0000 UTC" firstStartedPulling="2025-10-14 07:21:17.305992147 +0000 UTC m=+1213.003352518" lastFinishedPulling="2025-10-14 07:21:22.987172122 +0000 UTC m=+1218.684532493" observedRunningTime="2025-10-14 07:21:24.179758476 +0000 UTC m=+1219.877118837" watchObservedRunningTime="2025-10-14 07:21:24.190616167 +0000 UTC m=+1219.887976538" Oct 14 07:21:24 crc kubenswrapper[4870]: I1014 07:21:24.194097 4870 scope.go:117] "RemoveContainer" containerID="b1d406e0278bc21b8489162b2029d7503109f564a8f097477cfabec3a829c8eb" Oct 14 07:21:24 crc kubenswrapper[4870]: I1014 07:21:24.224773 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 14 07:21:24 crc kubenswrapper[4870]: I1014 07:21:24.233433 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 14 07:21:24 crc kubenswrapper[4870]: I1014 07:21:24.244601 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-5dd746bf8d-mvg6f"] Oct 14 07:21:24 crc kubenswrapper[4870]: I1014 07:21:24.275889 4870 scope.go:117] "RemoveContainer" containerID="ed5d51b7faa1c7111cd571af1833b5840ccac99e3be071ef421e46b99f9cb4b7" Oct 14 07:21:24 crc kubenswrapper[4870]: I1014 07:21:24.281138 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-5dd746bf8d-mvg6f"] Oct 14 07:21:24 crc kubenswrapper[4870]: I1014 07:21:24.287972 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Oct 14 07:21:24 crc kubenswrapper[4870]: E1014 07:21:24.288400 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d42ee9f9-390b-437c-a48a-ff7ead370f61" containerName="init" Oct 14 07:21:24 crc kubenswrapper[4870]: I1014 07:21:24.288422 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="d42ee9f9-390b-437c-a48a-ff7ead370f61" containerName="init" Oct 14 07:21:24 crc kubenswrapper[4870]: E1014 07:21:24.288449 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8962df13-02a8-4049-bf25-767413eaa98f" containerName="neutron-api" Oct 14 07:21:24 crc kubenswrapper[4870]: I1014 07:21:24.288456 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="8962df13-02a8-4049-bf25-767413eaa98f" containerName="neutron-api" Oct 14 07:21:24 crc kubenswrapper[4870]: E1014 07:21:24.288469 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d42ee9f9-390b-437c-a48a-ff7ead370f61" containerName="dnsmasq-dns" Oct 14 07:21:24 crc kubenswrapper[4870]: I1014 07:21:24.288475 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="d42ee9f9-390b-437c-a48a-ff7ead370f61" containerName="dnsmasq-dns" Oct 14 07:21:24 crc kubenswrapper[4870]: E1014 07:21:24.288489 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8962df13-02a8-4049-bf25-767413eaa98f" containerName="neutron-httpd" Oct 14 07:21:24 crc kubenswrapper[4870]: I1014 07:21:24.288496 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="8962df13-02a8-4049-bf25-767413eaa98f" containerName="neutron-httpd" Oct 14 07:21:24 crc kubenswrapper[4870]: E1014 07:21:24.288510 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f04fb303-8db7-4d2c-a864-817494a7222b" containerName="cinder-scheduler" Oct 14 07:21:24 crc kubenswrapper[4870]: I1014 07:21:24.288516 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="f04fb303-8db7-4d2c-a864-817494a7222b" containerName="cinder-scheduler" Oct 14 07:21:24 crc kubenswrapper[4870]: E1014 07:21:24.288527 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f04fb303-8db7-4d2c-a864-817494a7222b" containerName="probe" Oct 14 07:21:24 crc kubenswrapper[4870]: I1014 07:21:24.288535 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="f04fb303-8db7-4d2c-a864-817494a7222b" containerName="probe" Oct 14 07:21:24 crc kubenswrapper[4870]: I1014 07:21:24.288700 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="f04fb303-8db7-4d2c-a864-817494a7222b" containerName="cinder-scheduler" Oct 14 07:21:24 crc kubenswrapper[4870]: I1014 07:21:24.288712 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="8962df13-02a8-4049-bf25-767413eaa98f" containerName="neutron-api" Oct 14 07:21:24 crc kubenswrapper[4870]: I1014 07:21:24.288727 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="8962df13-02a8-4049-bf25-767413eaa98f" containerName="neutron-httpd" Oct 14 07:21:24 crc kubenswrapper[4870]: I1014 07:21:24.288738 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="f04fb303-8db7-4d2c-a864-817494a7222b" containerName="probe" Oct 14 07:21:24 crc kubenswrapper[4870]: I1014 07:21:24.288755 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="d42ee9f9-390b-437c-a48a-ff7ead370f61" containerName="dnsmasq-dns" Oct 14 07:21:24 crc kubenswrapper[4870]: I1014 07:21:24.289673 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 14 07:21:24 crc kubenswrapper[4870]: I1014 07:21:24.292382 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Oct 14 07:21:24 crc kubenswrapper[4870]: I1014 07:21:24.307114 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 14 07:21:24 crc kubenswrapper[4870]: I1014 07:21:24.372758 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bb5cv\" (UniqueName: \"kubernetes.io/projected/7bc76b72-8679-4b33-8966-0ee70c1788e5-kube-api-access-bb5cv\") pod \"cinder-scheduler-0\" (UID: \"7bc76b72-8679-4b33-8966-0ee70c1788e5\") " pod="openstack/cinder-scheduler-0" Oct 14 07:21:24 crc kubenswrapper[4870]: I1014 07:21:24.372807 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7bc76b72-8679-4b33-8966-0ee70c1788e5-scripts\") pod \"cinder-scheduler-0\" (UID: \"7bc76b72-8679-4b33-8966-0ee70c1788e5\") " pod="openstack/cinder-scheduler-0" Oct 14 07:21:24 crc kubenswrapper[4870]: I1014 07:21:24.372823 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7bc76b72-8679-4b33-8966-0ee70c1788e5-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"7bc76b72-8679-4b33-8966-0ee70c1788e5\") " pod="openstack/cinder-scheduler-0" Oct 14 07:21:24 crc kubenswrapper[4870]: I1014 07:21:24.372849 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7bc76b72-8679-4b33-8966-0ee70c1788e5-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"7bc76b72-8679-4b33-8966-0ee70c1788e5\") " pod="openstack/cinder-scheduler-0" Oct 14 07:21:24 crc kubenswrapper[4870]: I1014 07:21:24.372917 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7bc76b72-8679-4b33-8966-0ee70c1788e5-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"7bc76b72-8679-4b33-8966-0ee70c1788e5\") " pod="openstack/cinder-scheduler-0" Oct 14 07:21:24 crc kubenswrapper[4870]: I1014 07:21:24.373002 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7bc76b72-8679-4b33-8966-0ee70c1788e5-config-data\") pod \"cinder-scheduler-0\" (UID: \"7bc76b72-8679-4b33-8966-0ee70c1788e5\") " pod="openstack/cinder-scheduler-0" Oct 14 07:21:24 crc kubenswrapper[4870]: I1014 07:21:24.376710 4870 scope.go:117] "RemoveContainer" containerID="b8b3775a268dcc4ebc06b8c9a67360595bbc18617ec1fd7cabd8c847229a0730" Oct 14 07:21:24 crc kubenswrapper[4870]: I1014 07:21:24.396130 4870 scope.go:117] "RemoveContainer" containerID="ed5d51b7faa1c7111cd571af1833b5840ccac99e3be071ef421e46b99f9cb4b7" Oct 14 07:21:24 crc kubenswrapper[4870]: E1014 07:21:24.396598 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ed5d51b7faa1c7111cd571af1833b5840ccac99e3be071ef421e46b99f9cb4b7\": container with ID starting with ed5d51b7faa1c7111cd571af1833b5840ccac99e3be071ef421e46b99f9cb4b7 not found: ID does not exist" containerID="ed5d51b7faa1c7111cd571af1833b5840ccac99e3be071ef421e46b99f9cb4b7" Oct 14 07:21:24 crc kubenswrapper[4870]: I1014 07:21:24.396631 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ed5d51b7faa1c7111cd571af1833b5840ccac99e3be071ef421e46b99f9cb4b7"} err="failed to get container status \"ed5d51b7faa1c7111cd571af1833b5840ccac99e3be071ef421e46b99f9cb4b7\": rpc error: code = NotFound desc = could not find container \"ed5d51b7faa1c7111cd571af1833b5840ccac99e3be071ef421e46b99f9cb4b7\": container with ID starting with ed5d51b7faa1c7111cd571af1833b5840ccac99e3be071ef421e46b99f9cb4b7 not found: ID does not exist" Oct 14 07:21:24 crc kubenswrapper[4870]: I1014 07:21:24.396653 4870 scope.go:117] "RemoveContainer" containerID="b8b3775a268dcc4ebc06b8c9a67360595bbc18617ec1fd7cabd8c847229a0730" Oct 14 07:21:24 crc kubenswrapper[4870]: E1014 07:21:24.396858 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b8b3775a268dcc4ebc06b8c9a67360595bbc18617ec1fd7cabd8c847229a0730\": container with ID starting with b8b3775a268dcc4ebc06b8c9a67360595bbc18617ec1fd7cabd8c847229a0730 not found: ID does not exist" containerID="b8b3775a268dcc4ebc06b8c9a67360595bbc18617ec1fd7cabd8c847229a0730" Oct 14 07:21:24 crc kubenswrapper[4870]: I1014 07:21:24.396878 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b8b3775a268dcc4ebc06b8c9a67360595bbc18617ec1fd7cabd8c847229a0730"} err="failed to get container status \"b8b3775a268dcc4ebc06b8c9a67360595bbc18617ec1fd7cabd8c847229a0730\": rpc error: code = NotFound desc = could not find container \"b8b3775a268dcc4ebc06b8c9a67360595bbc18617ec1fd7cabd8c847229a0730\": container with ID starting with b8b3775a268dcc4ebc06b8c9a67360595bbc18617ec1fd7cabd8c847229a0730 not found: ID does not exist" Oct 14 07:21:24 crc kubenswrapper[4870]: I1014 07:21:24.474743 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7bc76b72-8679-4b33-8966-0ee70c1788e5-config-data\") pod \"cinder-scheduler-0\" (UID: \"7bc76b72-8679-4b33-8966-0ee70c1788e5\") " pod="openstack/cinder-scheduler-0" Oct 14 07:21:24 crc kubenswrapper[4870]: I1014 07:21:24.474870 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bb5cv\" (UniqueName: \"kubernetes.io/projected/7bc76b72-8679-4b33-8966-0ee70c1788e5-kube-api-access-bb5cv\") pod \"cinder-scheduler-0\" (UID: \"7bc76b72-8679-4b33-8966-0ee70c1788e5\") " pod="openstack/cinder-scheduler-0" Oct 14 07:21:24 crc kubenswrapper[4870]: I1014 07:21:24.474899 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7bc76b72-8679-4b33-8966-0ee70c1788e5-scripts\") pod \"cinder-scheduler-0\" (UID: \"7bc76b72-8679-4b33-8966-0ee70c1788e5\") " pod="openstack/cinder-scheduler-0" Oct 14 07:21:24 crc kubenswrapper[4870]: I1014 07:21:24.474914 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7bc76b72-8679-4b33-8966-0ee70c1788e5-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"7bc76b72-8679-4b33-8966-0ee70c1788e5\") " pod="openstack/cinder-scheduler-0" Oct 14 07:21:24 crc kubenswrapper[4870]: I1014 07:21:24.474933 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7bc76b72-8679-4b33-8966-0ee70c1788e5-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"7bc76b72-8679-4b33-8966-0ee70c1788e5\") " pod="openstack/cinder-scheduler-0" Oct 14 07:21:24 crc kubenswrapper[4870]: I1014 07:21:24.475093 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7bc76b72-8679-4b33-8966-0ee70c1788e5-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"7bc76b72-8679-4b33-8966-0ee70c1788e5\") " pod="openstack/cinder-scheduler-0" Oct 14 07:21:24 crc kubenswrapper[4870]: I1014 07:21:24.475646 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7bc76b72-8679-4b33-8966-0ee70c1788e5-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"7bc76b72-8679-4b33-8966-0ee70c1788e5\") " pod="openstack/cinder-scheduler-0" Oct 14 07:21:24 crc kubenswrapper[4870]: I1014 07:21:24.478694 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7bc76b72-8679-4b33-8966-0ee70c1788e5-scripts\") pod \"cinder-scheduler-0\" (UID: \"7bc76b72-8679-4b33-8966-0ee70c1788e5\") " pod="openstack/cinder-scheduler-0" Oct 14 07:21:24 crc kubenswrapper[4870]: I1014 07:21:24.479244 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7bc76b72-8679-4b33-8966-0ee70c1788e5-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"7bc76b72-8679-4b33-8966-0ee70c1788e5\") " pod="openstack/cinder-scheduler-0" Oct 14 07:21:24 crc kubenswrapper[4870]: I1014 07:21:24.479501 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7bc76b72-8679-4b33-8966-0ee70c1788e5-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"7bc76b72-8679-4b33-8966-0ee70c1788e5\") " pod="openstack/cinder-scheduler-0" Oct 14 07:21:24 crc kubenswrapper[4870]: I1014 07:21:24.494890 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bb5cv\" (UniqueName: \"kubernetes.io/projected/7bc76b72-8679-4b33-8966-0ee70c1788e5-kube-api-access-bb5cv\") pod \"cinder-scheduler-0\" (UID: \"7bc76b72-8679-4b33-8966-0ee70c1788e5\") " pod="openstack/cinder-scheduler-0" Oct 14 07:21:24 crc kubenswrapper[4870]: I1014 07:21:24.495103 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7bc76b72-8679-4b33-8966-0ee70c1788e5-config-data\") pod \"cinder-scheduler-0\" (UID: \"7bc76b72-8679-4b33-8966-0ee70c1788e5\") " pod="openstack/cinder-scheduler-0" Oct 14 07:21:24 crc kubenswrapper[4870]: I1014 07:21:24.680951 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 14 07:21:25 crc kubenswrapper[4870]: I1014 07:21:25.054203 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8962df13-02a8-4049-bf25-767413eaa98f" path="/var/lib/kubelet/pods/8962df13-02a8-4049-bf25-767413eaa98f/volumes" Oct 14 07:21:25 crc kubenswrapper[4870]: I1014 07:21:25.055289 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f04fb303-8db7-4d2c-a864-817494a7222b" path="/var/lib/kubelet/pods/f04fb303-8db7-4d2c-a864-817494a7222b/volumes" Oct 14 07:21:25 crc kubenswrapper[4870]: I1014 07:21:25.155251 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 14 07:21:25 crc kubenswrapper[4870]: I1014 07:21:25.173513 4870 generic.go:334] "Generic (PLEG): container finished" podID="4d00ede0-cd53-4791-9325-c712ee891839" containerID="fdda30c816580a958c0638fccaade56d725b6b8fffdcdb85825f921582f25f4e" exitCode=0 Oct 14 07:21:25 crc kubenswrapper[4870]: I1014 07:21:25.173547 4870 generic.go:334] "Generic (PLEG): container finished" podID="4d00ede0-cd53-4791-9325-c712ee891839" containerID="c07b4185cb742a8d604acc1b5ef508fa09c68ed0229dc00df2a8b8a8be770694" exitCode=2 Oct 14 07:21:25 crc kubenswrapper[4870]: I1014 07:21:25.173557 4870 generic.go:334] "Generic (PLEG): container finished" podID="4d00ede0-cd53-4791-9325-c712ee891839" containerID="ccd5820b235e529b2ee667f2bc0cb4aaa6d87fc0297a583d5753340e0d346237" exitCode=0 Oct 14 07:21:25 crc kubenswrapper[4870]: I1014 07:21:25.173582 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4d00ede0-cd53-4791-9325-c712ee891839","Type":"ContainerDied","Data":"fdda30c816580a958c0638fccaade56d725b6b8fffdcdb85825f921582f25f4e"} Oct 14 07:21:25 crc kubenswrapper[4870]: I1014 07:21:25.173641 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4d00ede0-cd53-4791-9325-c712ee891839","Type":"ContainerDied","Data":"c07b4185cb742a8d604acc1b5ef508fa09c68ed0229dc00df2a8b8a8be770694"} Oct 14 07:21:25 crc kubenswrapper[4870]: I1014 07:21:25.173651 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4d00ede0-cd53-4791-9325-c712ee891839","Type":"ContainerDied","Data":"ccd5820b235e529b2ee667f2bc0cb4aaa6d87fc0297a583d5753340e0d346237"} Oct 14 07:21:25 crc kubenswrapper[4870]: W1014 07:21:25.174884 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7bc76b72_8679_4b33_8966_0ee70c1788e5.slice/crio-61320ff1ecf9ca47176c06537cb5c503a9ed81f16e0538a006104df47500851e WatchSource:0}: Error finding container 61320ff1ecf9ca47176c06537cb5c503a9ed81f16e0538a006104df47500851e: Status 404 returned error can't find the container with id 61320ff1ecf9ca47176c06537cb5c503a9ed81f16e0538a006104df47500851e Oct 14 07:21:25 crc kubenswrapper[4870]: I1014 07:21:25.193045 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 14 07:21:25 crc kubenswrapper[4870]: I1014 07:21:25.193447 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="385ca977-7d50-426e-86e9-6a9416ad872c" containerName="glance-log" containerID="cri-o://4bdba3786e90a36656b26936418c10c2f99a9784397e74e7361942e8723a91ce" gracePeriod=30 Oct 14 07:21:25 crc kubenswrapper[4870]: I1014 07:21:25.193585 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="385ca977-7d50-426e-86e9-6a9416ad872c" containerName="glance-httpd" containerID="cri-o://f8730cc9a043783b0ae420ff4c40073296e6d3942a94ac10a57a106576eecfda" gracePeriod=30 Oct 14 07:21:26 crc kubenswrapper[4870]: I1014 07:21:26.198760 4870 generic.go:334] "Generic (PLEG): container finished" podID="385ca977-7d50-426e-86e9-6a9416ad872c" containerID="4bdba3786e90a36656b26936418c10c2f99a9784397e74e7361942e8723a91ce" exitCode=143 Oct 14 07:21:26 crc kubenswrapper[4870]: I1014 07:21:26.199163 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"385ca977-7d50-426e-86e9-6a9416ad872c","Type":"ContainerDied","Data":"4bdba3786e90a36656b26936418c10c2f99a9784397e74e7361942e8723a91ce"} Oct 14 07:21:26 crc kubenswrapper[4870]: I1014 07:21:26.200945 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"7bc76b72-8679-4b33-8966-0ee70c1788e5","Type":"ContainerStarted","Data":"e5c43911d04a168b064b8b753711f3e5ed521dc55a16d5687c49934c8d0eb3a6"} Oct 14 07:21:26 crc kubenswrapper[4870]: I1014 07:21:26.200975 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"7bc76b72-8679-4b33-8966-0ee70c1788e5","Type":"ContainerStarted","Data":"61320ff1ecf9ca47176c06537cb5c503a9ed81f16e0538a006104df47500851e"} Oct 14 07:21:27 crc kubenswrapper[4870]: I1014 07:21:27.211115 4870 generic.go:334] "Generic (PLEG): container finished" podID="452a6ce6-d9dc-4455-9304-414701c8895f" containerID="5db90be32cea50de9fa48a1c7946854f77a0364acaf6f2929af1ec8f226fb41f" exitCode=0 Oct 14 07:21:27 crc kubenswrapper[4870]: I1014 07:21:27.211193 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"452a6ce6-d9dc-4455-9304-414701c8895f","Type":"ContainerDied","Data":"5db90be32cea50de9fa48a1c7946854f77a0364acaf6f2929af1ec8f226fb41f"} Oct 14 07:21:27 crc kubenswrapper[4870]: I1014 07:21:27.214314 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"7bc76b72-8679-4b33-8966-0ee70c1788e5","Type":"ContainerStarted","Data":"a1f9a0d181ab7f3f80bee6c1008e3b393dd082ba31b2614d29e2a4b0d5534a46"} Oct 14 07:21:27 crc kubenswrapper[4870]: I1014 07:21:27.248235 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.248219357 podStartE2EDuration="3.248219357s" podCreationTimestamp="2025-10-14 07:21:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:21:27.247065988 +0000 UTC m=+1222.944426399" watchObservedRunningTime="2025-10-14 07:21:27.248219357 +0000 UTC m=+1222.945579728" Oct 14 07:21:28 crc kubenswrapper[4870]: I1014 07:21:28.231773 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Oct 14 07:21:29 crc kubenswrapper[4870]: I1014 07:21:29.238598 4870 generic.go:334] "Generic (PLEG): container finished" podID="385ca977-7d50-426e-86e9-6a9416ad872c" containerID="f8730cc9a043783b0ae420ff4c40073296e6d3942a94ac10a57a106576eecfda" exitCode=0 Oct 14 07:21:29 crc kubenswrapper[4870]: I1014 07:21:29.238687 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"385ca977-7d50-426e-86e9-6a9416ad872c","Type":"ContainerDied","Data":"f8730cc9a043783b0ae420ff4c40073296e6d3942a94ac10a57a106576eecfda"} Oct 14 07:21:29 crc kubenswrapper[4870]: I1014 07:21:29.681934 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Oct 14 07:21:30 crc kubenswrapper[4870]: I1014 07:21:30.248968 4870 generic.go:334] "Generic (PLEG): container finished" podID="4d00ede0-cd53-4791-9325-c712ee891839" containerID="2bcc81a3a9d5c820c0bcd54b459b4f42446c2e0c7a6b7503e11f9c3e6552fadb" exitCode=0 Oct 14 07:21:30 crc kubenswrapper[4870]: I1014 07:21:30.249248 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4d00ede0-cd53-4791-9325-c712ee891839","Type":"ContainerDied","Data":"2bcc81a3a9d5c820c0bcd54b459b4f42446c2e0c7a6b7503e11f9c3e6552fadb"} Oct 14 07:21:31 crc kubenswrapper[4870]: I1014 07:21:31.890007 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 14 07:21:31 crc kubenswrapper[4870]: I1014 07:21:31.948280 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 14 07:21:32 crc kubenswrapper[4870]: I1014 07:21:32.064622 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/452a6ce6-d9dc-4455-9304-414701c8895f-combined-ca-bundle\") pod \"452a6ce6-d9dc-4455-9304-414701c8895f\" (UID: \"452a6ce6-d9dc-4455-9304-414701c8895f\") " Oct 14 07:21:32 crc kubenswrapper[4870]: I1014 07:21:32.064727 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"385ca977-7d50-426e-86e9-6a9416ad872c\" (UID: \"385ca977-7d50-426e-86e9-6a9416ad872c\") " Oct 14 07:21:32 crc kubenswrapper[4870]: I1014 07:21:32.064780 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/385ca977-7d50-426e-86e9-6a9416ad872c-logs\") pod \"385ca977-7d50-426e-86e9-6a9416ad872c\" (UID: \"385ca977-7d50-426e-86e9-6a9416ad872c\") " Oct 14 07:21:32 crc kubenswrapper[4870]: I1014 07:21:32.064900 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/385ca977-7d50-426e-86e9-6a9416ad872c-config-data\") pod \"385ca977-7d50-426e-86e9-6a9416ad872c\" (UID: \"385ca977-7d50-426e-86e9-6a9416ad872c\") " Oct 14 07:21:32 crc kubenswrapper[4870]: I1014 07:21:32.064936 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-98996\" (UniqueName: \"kubernetes.io/projected/452a6ce6-d9dc-4455-9304-414701c8895f-kube-api-access-98996\") pod \"452a6ce6-d9dc-4455-9304-414701c8895f\" (UID: \"452a6ce6-d9dc-4455-9304-414701c8895f\") " Oct 14 07:21:32 crc kubenswrapper[4870]: I1014 07:21:32.064967 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/452a6ce6-d9dc-4455-9304-414701c8895f-config-data\") pod \"452a6ce6-d9dc-4455-9304-414701c8895f\" (UID: \"452a6ce6-d9dc-4455-9304-414701c8895f\") " Oct 14 07:21:32 crc kubenswrapper[4870]: I1014 07:21:32.065032 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/452a6ce6-d9dc-4455-9304-414701c8895f-scripts\") pod \"452a6ce6-d9dc-4455-9304-414701c8895f\" (UID: \"452a6ce6-d9dc-4455-9304-414701c8895f\") " Oct 14 07:21:32 crc kubenswrapper[4870]: I1014 07:21:32.065102 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/452a6ce6-d9dc-4455-9304-414701c8895f-public-tls-certs\") pod \"452a6ce6-d9dc-4455-9304-414701c8895f\" (UID: \"452a6ce6-d9dc-4455-9304-414701c8895f\") " Oct 14 07:21:32 crc kubenswrapper[4870]: I1014 07:21:32.065154 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/385ca977-7d50-426e-86e9-6a9416ad872c-combined-ca-bundle\") pod \"385ca977-7d50-426e-86e9-6a9416ad872c\" (UID: \"385ca977-7d50-426e-86e9-6a9416ad872c\") " Oct 14 07:21:32 crc kubenswrapper[4870]: I1014 07:21:32.065207 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/452a6ce6-d9dc-4455-9304-414701c8895f-httpd-run\") pod \"452a6ce6-d9dc-4455-9304-414701c8895f\" (UID: \"452a6ce6-d9dc-4455-9304-414701c8895f\") " Oct 14 07:21:32 crc kubenswrapper[4870]: I1014 07:21:32.065370 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w8zs9\" (UniqueName: \"kubernetes.io/projected/385ca977-7d50-426e-86e9-6a9416ad872c-kube-api-access-w8zs9\") pod \"385ca977-7d50-426e-86e9-6a9416ad872c\" (UID: \"385ca977-7d50-426e-86e9-6a9416ad872c\") " Oct 14 07:21:32 crc kubenswrapper[4870]: I1014 07:21:32.065418 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"452a6ce6-d9dc-4455-9304-414701c8895f\" (UID: \"452a6ce6-d9dc-4455-9304-414701c8895f\") " Oct 14 07:21:32 crc kubenswrapper[4870]: I1014 07:21:32.065453 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/385ca977-7d50-426e-86e9-6a9416ad872c-internal-tls-certs\") pod \"385ca977-7d50-426e-86e9-6a9416ad872c\" (UID: \"385ca977-7d50-426e-86e9-6a9416ad872c\") " Oct 14 07:21:32 crc kubenswrapper[4870]: I1014 07:21:32.065478 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/385ca977-7d50-426e-86e9-6a9416ad872c-scripts\") pod \"385ca977-7d50-426e-86e9-6a9416ad872c\" (UID: \"385ca977-7d50-426e-86e9-6a9416ad872c\") " Oct 14 07:21:32 crc kubenswrapper[4870]: I1014 07:21:32.065552 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/452a6ce6-d9dc-4455-9304-414701c8895f-logs\") pod \"452a6ce6-d9dc-4455-9304-414701c8895f\" (UID: \"452a6ce6-d9dc-4455-9304-414701c8895f\") " Oct 14 07:21:32 crc kubenswrapper[4870]: I1014 07:21:32.065578 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/385ca977-7d50-426e-86e9-6a9416ad872c-httpd-run\") pod \"385ca977-7d50-426e-86e9-6a9416ad872c\" (UID: \"385ca977-7d50-426e-86e9-6a9416ad872c\") " Oct 14 07:21:32 crc kubenswrapper[4870]: I1014 07:21:32.067638 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/385ca977-7d50-426e-86e9-6a9416ad872c-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "385ca977-7d50-426e-86e9-6a9416ad872c" (UID: "385ca977-7d50-426e-86e9-6a9416ad872c"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:21:32 crc kubenswrapper[4870]: I1014 07:21:32.071294 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/385ca977-7d50-426e-86e9-6a9416ad872c-logs" (OuterVolumeSpecName: "logs") pod "385ca977-7d50-426e-86e9-6a9416ad872c" (UID: "385ca977-7d50-426e-86e9-6a9416ad872c"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:21:32 crc kubenswrapper[4870]: I1014 07:21:32.072105 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/452a6ce6-d9dc-4455-9304-414701c8895f-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "452a6ce6-d9dc-4455-9304-414701c8895f" (UID: "452a6ce6-d9dc-4455-9304-414701c8895f"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:21:32 crc kubenswrapper[4870]: I1014 07:21:32.076630 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/452a6ce6-d9dc-4455-9304-414701c8895f-kube-api-access-98996" (OuterVolumeSpecName: "kube-api-access-98996") pod "452a6ce6-d9dc-4455-9304-414701c8895f" (UID: "452a6ce6-d9dc-4455-9304-414701c8895f"). InnerVolumeSpecName "kube-api-access-98996". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:21:32 crc kubenswrapper[4870]: I1014 07:21:32.079087 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/385ca977-7d50-426e-86e9-6a9416ad872c-kube-api-access-w8zs9" (OuterVolumeSpecName: "kube-api-access-w8zs9") pod "385ca977-7d50-426e-86e9-6a9416ad872c" (UID: "385ca977-7d50-426e-86e9-6a9416ad872c"). InnerVolumeSpecName "kube-api-access-w8zs9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:21:32 crc kubenswrapper[4870]: I1014 07:21:32.079479 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/452a6ce6-d9dc-4455-9304-414701c8895f-logs" (OuterVolumeSpecName: "logs") pod "452a6ce6-d9dc-4455-9304-414701c8895f" (UID: "452a6ce6-d9dc-4455-9304-414701c8895f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:21:32 crc kubenswrapper[4870]: I1014 07:21:32.087005 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage06-crc" (OuterVolumeSpecName: "glance") pod "385ca977-7d50-426e-86e9-6a9416ad872c" (UID: "385ca977-7d50-426e-86e9-6a9416ad872c"). InnerVolumeSpecName "local-storage06-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 14 07:21:32 crc kubenswrapper[4870]: I1014 07:21:32.095031 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/385ca977-7d50-426e-86e9-6a9416ad872c-scripts" (OuterVolumeSpecName: "scripts") pod "385ca977-7d50-426e-86e9-6a9416ad872c" (UID: "385ca977-7d50-426e-86e9-6a9416ad872c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:21:32 crc kubenswrapper[4870]: I1014 07:21:32.096331 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/452a6ce6-d9dc-4455-9304-414701c8895f-scripts" (OuterVolumeSpecName: "scripts") pod "452a6ce6-d9dc-4455-9304-414701c8895f" (UID: "452a6ce6-d9dc-4455-9304-414701c8895f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:21:32 crc kubenswrapper[4870]: I1014 07:21:32.107760 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "glance") pod "452a6ce6-d9dc-4455-9304-414701c8895f" (UID: "452a6ce6-d9dc-4455-9304-414701c8895f"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 14 07:21:32 crc kubenswrapper[4870]: I1014 07:21:32.171584 4870 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" " Oct 14 07:21:32 crc kubenswrapper[4870]: I1014 07:21:32.171646 4870 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/385ca977-7d50-426e-86e9-6a9416ad872c-logs\") on node \"crc\" DevicePath \"\"" Oct 14 07:21:32 crc kubenswrapper[4870]: I1014 07:21:32.171656 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-98996\" (UniqueName: \"kubernetes.io/projected/452a6ce6-d9dc-4455-9304-414701c8895f-kube-api-access-98996\") on node \"crc\" DevicePath \"\"" Oct 14 07:21:32 crc kubenswrapper[4870]: I1014 07:21:32.171669 4870 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/452a6ce6-d9dc-4455-9304-414701c8895f-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 07:21:32 crc kubenswrapper[4870]: I1014 07:21:32.171677 4870 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/452a6ce6-d9dc-4455-9304-414701c8895f-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 14 07:21:32 crc kubenswrapper[4870]: I1014 07:21:32.171710 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w8zs9\" (UniqueName: \"kubernetes.io/projected/385ca977-7d50-426e-86e9-6a9416ad872c-kube-api-access-w8zs9\") on node \"crc\" DevicePath \"\"" Oct 14 07:21:32 crc kubenswrapper[4870]: I1014 07:21:32.171727 4870 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Oct 14 07:21:32 crc kubenswrapper[4870]: I1014 07:21:32.171736 4870 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/385ca977-7d50-426e-86e9-6a9416ad872c-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 07:21:32 crc kubenswrapper[4870]: I1014 07:21:32.171744 4870 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/452a6ce6-d9dc-4455-9304-414701c8895f-logs\") on node \"crc\" DevicePath \"\"" Oct 14 07:21:32 crc kubenswrapper[4870]: I1014 07:21:32.171752 4870 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/385ca977-7d50-426e-86e9-6a9416ad872c-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 14 07:21:32 crc kubenswrapper[4870]: I1014 07:21:32.174214 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/452a6ce6-d9dc-4455-9304-414701c8895f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "452a6ce6-d9dc-4455-9304-414701c8895f" (UID: "452a6ce6-d9dc-4455-9304-414701c8895f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:21:32 crc kubenswrapper[4870]: I1014 07:21:32.185732 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/385ca977-7d50-426e-86e9-6a9416ad872c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "385ca977-7d50-426e-86e9-6a9416ad872c" (UID: "385ca977-7d50-426e-86e9-6a9416ad872c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:21:32 crc kubenswrapper[4870]: I1014 07:21:32.187178 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/452a6ce6-d9dc-4455-9304-414701c8895f-config-data" (OuterVolumeSpecName: "config-data") pod "452a6ce6-d9dc-4455-9304-414701c8895f" (UID: "452a6ce6-d9dc-4455-9304-414701c8895f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:21:32 crc kubenswrapper[4870]: I1014 07:21:32.241666 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/452a6ce6-d9dc-4455-9304-414701c8895f-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "452a6ce6-d9dc-4455-9304-414701c8895f" (UID: "452a6ce6-d9dc-4455-9304-414701c8895f"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:21:32 crc kubenswrapper[4870]: I1014 07:21:32.248278 4870 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage06-crc" (UniqueName: "kubernetes.io/local-volume/local-storage06-crc") on node "crc" Oct 14 07:21:32 crc kubenswrapper[4870]: I1014 07:21:32.257085 4870 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Oct 14 07:21:32 crc kubenswrapper[4870]: I1014 07:21:32.267835 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/385ca977-7d50-426e-86e9-6a9416ad872c-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "385ca977-7d50-426e-86e9-6a9416ad872c" (UID: "385ca977-7d50-426e-86e9-6a9416ad872c"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:21:32 crc kubenswrapper[4870]: I1014 07:21:32.273820 4870 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/385ca977-7d50-426e-86e9-6a9416ad872c-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 14 07:21:32 crc kubenswrapper[4870]: I1014 07:21:32.273859 4870 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Oct 14 07:21:32 crc kubenswrapper[4870]: I1014 07:21:32.273873 4870 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/452a6ce6-d9dc-4455-9304-414701c8895f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 07:21:32 crc kubenswrapper[4870]: I1014 07:21:32.273885 4870 reconciler_common.go:293] "Volume detached for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" DevicePath \"\"" Oct 14 07:21:32 crc kubenswrapper[4870]: I1014 07:21:32.273895 4870 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/452a6ce6-d9dc-4455-9304-414701c8895f-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 07:21:32 crc kubenswrapper[4870]: I1014 07:21:32.273940 4870 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/452a6ce6-d9dc-4455-9304-414701c8895f-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 14 07:21:32 crc kubenswrapper[4870]: I1014 07:21:32.273953 4870 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/385ca977-7d50-426e-86e9-6a9416ad872c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 07:21:32 crc kubenswrapper[4870]: I1014 07:21:32.287970 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"385ca977-7d50-426e-86e9-6a9416ad872c","Type":"ContainerDied","Data":"a61556670dfd1432f61da5383f63bfcbcca889d5c7b1e8b257bf7a66ee23267e"} Oct 14 07:21:32 crc kubenswrapper[4870]: I1014 07:21:32.288048 4870 scope.go:117] "RemoveContainer" containerID="f8730cc9a043783b0ae420ff4c40073296e6d3942a94ac10a57a106576eecfda" Oct 14 07:21:32 crc kubenswrapper[4870]: I1014 07:21:32.288236 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 14 07:21:32 crc kubenswrapper[4870]: I1014 07:21:32.299062 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 14 07:21:32 crc kubenswrapper[4870]: I1014 07:21:32.299076 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"452a6ce6-d9dc-4455-9304-414701c8895f","Type":"ContainerDied","Data":"32f0b83cd662611a3d7e7503c4a4b8bd18f13afd5fd1f80d21f562d4375742fd"} Oct 14 07:21:32 crc kubenswrapper[4870]: I1014 07:21:32.304826 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4d00ede0-cd53-4791-9325-c712ee891839","Type":"ContainerDied","Data":"18af3f55a518be26512e94d22bb6a5861f9b0a595aab3ee40db0994485db6334"} Oct 14 07:21:32 crc kubenswrapper[4870]: I1014 07:21:32.304944 4870 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="18af3f55a518be26512e94d22bb6a5861f9b0a595aab3ee40db0994485db6334" Oct 14 07:21:32 crc kubenswrapper[4870]: I1014 07:21:32.308215 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/385ca977-7d50-426e-86e9-6a9416ad872c-config-data" (OuterVolumeSpecName: "config-data") pod "385ca977-7d50-426e-86e9-6a9416ad872c" (UID: "385ca977-7d50-426e-86e9-6a9416ad872c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:21:32 crc kubenswrapper[4870]: I1014 07:21:32.309518 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-7jfpt" event={"ID":"bdb3c945-edde-4d43-b3a5-234aaabf00d5","Type":"ContainerStarted","Data":"2841eadea64f22ddd21b4d710878d4871c5c5ceec4c6a266af64c8c08bfe682a"} Oct 14 07:21:32 crc kubenswrapper[4870]: I1014 07:21:32.333568 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-7jfpt" podStartSLOduration=1.978578233 podStartE2EDuration="12.333547436s" podCreationTimestamp="2025-10-14 07:21:20 +0000 UTC" firstStartedPulling="2025-10-14 07:21:21.526550568 +0000 UTC m=+1217.223910939" lastFinishedPulling="2025-10-14 07:21:31.881519771 +0000 UTC m=+1227.578880142" observedRunningTime="2025-10-14 07:21:32.332279954 +0000 UTC m=+1228.029640325" watchObservedRunningTime="2025-10-14 07:21:32.333547436 +0000 UTC m=+1228.030907807" Oct 14 07:21:32 crc kubenswrapper[4870]: I1014 07:21:32.365065 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 14 07:21:32 crc kubenswrapper[4870]: I1014 07:21:32.367694 4870 scope.go:117] "RemoveContainer" containerID="4bdba3786e90a36656b26936418c10c2f99a9784397e74e7361942e8723a91ce" Oct 14 07:21:32 crc kubenswrapper[4870]: I1014 07:21:32.392230 4870 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/385ca977-7d50-426e-86e9-6a9416ad872c-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 07:21:32 crc kubenswrapper[4870]: I1014 07:21:32.393340 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 14 07:21:32 crc kubenswrapper[4870]: I1014 07:21:32.418024 4870 scope.go:117] "RemoveContainer" containerID="5db90be32cea50de9fa48a1c7946854f77a0364acaf6f2929af1ec8f226fb41f" Oct 14 07:21:32 crc kubenswrapper[4870]: I1014 07:21:32.423366 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 14 07:21:32 crc kubenswrapper[4870]: I1014 07:21:32.458591 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 14 07:21:32 crc kubenswrapper[4870]: E1014 07:21:32.459075 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d00ede0-cd53-4791-9325-c712ee891839" containerName="ceilometer-notification-agent" Oct 14 07:21:32 crc kubenswrapper[4870]: I1014 07:21:32.459100 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d00ede0-cd53-4791-9325-c712ee891839" containerName="ceilometer-notification-agent" Oct 14 07:21:32 crc kubenswrapper[4870]: E1014 07:21:32.459115 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="385ca977-7d50-426e-86e9-6a9416ad872c" containerName="glance-log" Oct 14 07:21:32 crc kubenswrapper[4870]: I1014 07:21:32.459125 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="385ca977-7d50-426e-86e9-6a9416ad872c" containerName="glance-log" Oct 14 07:21:32 crc kubenswrapper[4870]: E1014 07:21:32.459134 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="385ca977-7d50-426e-86e9-6a9416ad872c" containerName="glance-httpd" Oct 14 07:21:32 crc kubenswrapper[4870]: I1014 07:21:32.459140 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="385ca977-7d50-426e-86e9-6a9416ad872c" containerName="glance-httpd" Oct 14 07:21:32 crc kubenswrapper[4870]: E1014 07:21:32.459153 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d00ede0-cd53-4791-9325-c712ee891839" containerName="sg-core" Oct 14 07:21:32 crc kubenswrapper[4870]: I1014 07:21:32.459158 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d00ede0-cd53-4791-9325-c712ee891839" containerName="sg-core" Oct 14 07:21:32 crc kubenswrapper[4870]: E1014 07:21:32.459168 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d00ede0-cd53-4791-9325-c712ee891839" containerName="proxy-httpd" Oct 14 07:21:32 crc kubenswrapper[4870]: I1014 07:21:32.459174 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d00ede0-cd53-4791-9325-c712ee891839" containerName="proxy-httpd" Oct 14 07:21:32 crc kubenswrapper[4870]: E1014 07:21:32.459183 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="452a6ce6-d9dc-4455-9304-414701c8895f" containerName="glance-log" Oct 14 07:21:32 crc kubenswrapper[4870]: I1014 07:21:32.459188 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="452a6ce6-d9dc-4455-9304-414701c8895f" containerName="glance-log" Oct 14 07:21:32 crc kubenswrapper[4870]: E1014 07:21:32.459197 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="452a6ce6-d9dc-4455-9304-414701c8895f" containerName="glance-httpd" Oct 14 07:21:32 crc kubenswrapper[4870]: I1014 07:21:32.459203 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="452a6ce6-d9dc-4455-9304-414701c8895f" containerName="glance-httpd" Oct 14 07:21:32 crc kubenswrapper[4870]: E1014 07:21:32.459222 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d00ede0-cd53-4791-9325-c712ee891839" containerName="ceilometer-central-agent" Oct 14 07:21:32 crc kubenswrapper[4870]: I1014 07:21:32.459231 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d00ede0-cd53-4791-9325-c712ee891839" containerName="ceilometer-central-agent" Oct 14 07:21:32 crc kubenswrapper[4870]: I1014 07:21:32.459432 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="4d00ede0-cd53-4791-9325-c712ee891839" containerName="ceilometer-central-agent" Oct 14 07:21:32 crc kubenswrapper[4870]: I1014 07:21:32.459464 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="4d00ede0-cd53-4791-9325-c712ee891839" containerName="proxy-httpd" Oct 14 07:21:32 crc kubenswrapper[4870]: I1014 07:21:32.459477 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="4d00ede0-cd53-4791-9325-c712ee891839" containerName="ceilometer-notification-agent" Oct 14 07:21:32 crc kubenswrapper[4870]: I1014 07:21:32.459501 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="452a6ce6-d9dc-4455-9304-414701c8895f" containerName="glance-log" Oct 14 07:21:32 crc kubenswrapper[4870]: I1014 07:21:32.459508 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="452a6ce6-d9dc-4455-9304-414701c8895f" containerName="glance-httpd" Oct 14 07:21:32 crc kubenswrapper[4870]: I1014 07:21:32.459518 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="4d00ede0-cd53-4791-9325-c712ee891839" containerName="sg-core" Oct 14 07:21:32 crc kubenswrapper[4870]: I1014 07:21:32.459529 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="385ca977-7d50-426e-86e9-6a9416ad872c" containerName="glance-log" Oct 14 07:21:32 crc kubenswrapper[4870]: I1014 07:21:32.459537 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="385ca977-7d50-426e-86e9-6a9416ad872c" containerName="glance-httpd" Oct 14 07:21:32 crc kubenswrapper[4870]: I1014 07:21:32.468569 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 14 07:21:32 crc kubenswrapper[4870]: I1014 07:21:32.471394 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 14 07:21:32 crc kubenswrapper[4870]: I1014 07:21:32.473329 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Oct 14 07:21:32 crc kubenswrapper[4870]: I1014 07:21:32.484828 4870 scope.go:117] "RemoveContainer" containerID="1fde5cdab42d5388f18054e6d6d66a520f1f153b7c41418b4d919ba2fd643643" Oct 14 07:21:32 crc kubenswrapper[4870]: I1014 07:21:32.498663 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 14 07:21:32 crc kubenswrapper[4870]: I1014 07:21:32.499394 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4d00ede0-cd53-4791-9325-c712ee891839-run-httpd\") pod \"4d00ede0-cd53-4791-9325-c712ee891839\" (UID: \"4d00ede0-cd53-4791-9325-c712ee891839\") " Oct 14 07:21:32 crc kubenswrapper[4870]: I1014 07:21:32.499561 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x6hww\" (UniqueName: \"kubernetes.io/projected/4d00ede0-cd53-4791-9325-c712ee891839-kube-api-access-x6hww\") pod \"4d00ede0-cd53-4791-9325-c712ee891839\" (UID: \"4d00ede0-cd53-4791-9325-c712ee891839\") " Oct 14 07:21:32 crc kubenswrapper[4870]: I1014 07:21:32.499739 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4d00ede0-cd53-4791-9325-c712ee891839-log-httpd\") pod \"4d00ede0-cd53-4791-9325-c712ee891839\" (UID: \"4d00ede0-cd53-4791-9325-c712ee891839\") " Oct 14 07:21:32 crc kubenswrapper[4870]: I1014 07:21:32.499896 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d00ede0-cd53-4791-9325-c712ee891839-combined-ca-bundle\") pod \"4d00ede0-cd53-4791-9325-c712ee891839\" (UID: \"4d00ede0-cd53-4791-9325-c712ee891839\") " Oct 14 07:21:32 crc kubenswrapper[4870]: I1014 07:21:32.500151 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4d00ede0-cd53-4791-9325-c712ee891839-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "4d00ede0-cd53-4791-9325-c712ee891839" (UID: "4d00ede0-cd53-4791-9325-c712ee891839"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:21:32 crc kubenswrapper[4870]: I1014 07:21:32.500402 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4d00ede0-cd53-4791-9325-c712ee891839-scripts\") pod \"4d00ede0-cd53-4791-9325-c712ee891839\" (UID: \"4d00ede0-cd53-4791-9325-c712ee891839\") " Oct 14 07:21:32 crc kubenswrapper[4870]: I1014 07:21:32.500628 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4d00ede0-cd53-4791-9325-c712ee891839-sg-core-conf-yaml\") pod \"4d00ede0-cd53-4791-9325-c712ee891839\" (UID: \"4d00ede0-cd53-4791-9325-c712ee891839\") " Oct 14 07:21:32 crc kubenswrapper[4870]: I1014 07:21:32.500660 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4d00ede0-cd53-4791-9325-c712ee891839-config-data\") pod \"4d00ede0-cd53-4791-9325-c712ee891839\" (UID: \"4d00ede0-cd53-4791-9325-c712ee891839\") " Oct 14 07:21:32 crc kubenswrapper[4870]: I1014 07:21:32.500629 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4d00ede0-cd53-4791-9325-c712ee891839-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "4d00ede0-cd53-4791-9325-c712ee891839" (UID: "4d00ede0-cd53-4791-9325-c712ee891839"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:21:32 crc kubenswrapper[4870]: I1014 07:21:32.502201 4870 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4d00ede0-cd53-4791-9325-c712ee891839-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 14 07:21:32 crc kubenswrapper[4870]: I1014 07:21:32.502238 4870 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4d00ede0-cd53-4791-9325-c712ee891839-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 14 07:21:32 crc kubenswrapper[4870]: I1014 07:21:32.505647 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4d00ede0-cd53-4791-9325-c712ee891839-kube-api-access-x6hww" (OuterVolumeSpecName: "kube-api-access-x6hww") pod "4d00ede0-cd53-4791-9325-c712ee891839" (UID: "4d00ede0-cd53-4791-9325-c712ee891839"). InnerVolumeSpecName "kube-api-access-x6hww". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:21:32 crc kubenswrapper[4870]: I1014 07:21:32.519729 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d00ede0-cd53-4791-9325-c712ee891839-scripts" (OuterVolumeSpecName: "scripts") pod "4d00ede0-cd53-4791-9325-c712ee891839" (UID: "4d00ede0-cd53-4791-9325-c712ee891839"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:21:32 crc kubenswrapper[4870]: I1014 07:21:32.546257 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d00ede0-cd53-4791-9325-c712ee891839-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "4d00ede0-cd53-4791-9325-c712ee891839" (UID: "4d00ede0-cd53-4791-9325-c712ee891839"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:21:32 crc kubenswrapper[4870]: I1014 07:21:32.603707 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c712b87a-cf4d-437e-abe0-1a566046ca66-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"c712b87a-cf4d-437e-abe0-1a566046ca66\") " pod="openstack/glance-default-external-api-0" Oct 14 07:21:32 crc kubenswrapper[4870]: I1014 07:21:32.603971 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hzjsv\" (UniqueName: \"kubernetes.io/projected/c712b87a-cf4d-437e-abe0-1a566046ca66-kube-api-access-hzjsv\") pod \"glance-default-external-api-0\" (UID: \"c712b87a-cf4d-437e-abe0-1a566046ca66\") " pod="openstack/glance-default-external-api-0" Oct 14 07:21:32 crc kubenswrapper[4870]: I1014 07:21:32.604138 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c712b87a-cf4d-437e-abe0-1a566046ca66-config-data\") pod \"glance-default-external-api-0\" (UID: \"c712b87a-cf4d-437e-abe0-1a566046ca66\") " pod="openstack/glance-default-external-api-0" Oct 14 07:21:32 crc kubenswrapper[4870]: I1014 07:21:32.604242 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c712b87a-cf4d-437e-abe0-1a566046ca66-scripts\") pod \"glance-default-external-api-0\" (UID: \"c712b87a-cf4d-437e-abe0-1a566046ca66\") " pod="openstack/glance-default-external-api-0" Oct 14 07:21:32 crc kubenswrapper[4870]: I1014 07:21:32.604363 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c712b87a-cf4d-437e-abe0-1a566046ca66-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"c712b87a-cf4d-437e-abe0-1a566046ca66\") " pod="openstack/glance-default-external-api-0" Oct 14 07:21:32 crc kubenswrapper[4870]: I1014 07:21:32.604490 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c712b87a-cf4d-437e-abe0-1a566046ca66-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"c712b87a-cf4d-437e-abe0-1a566046ca66\") " pod="openstack/glance-default-external-api-0" Oct 14 07:21:32 crc kubenswrapper[4870]: I1014 07:21:32.605072 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"c712b87a-cf4d-437e-abe0-1a566046ca66\") " pod="openstack/glance-default-external-api-0" Oct 14 07:21:32 crc kubenswrapper[4870]: I1014 07:21:32.605183 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c712b87a-cf4d-437e-abe0-1a566046ca66-logs\") pod \"glance-default-external-api-0\" (UID: \"c712b87a-cf4d-437e-abe0-1a566046ca66\") " pod="openstack/glance-default-external-api-0" Oct 14 07:21:32 crc kubenswrapper[4870]: I1014 07:21:32.605377 4870 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4d00ede0-cd53-4791-9325-c712ee891839-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 07:21:32 crc kubenswrapper[4870]: I1014 07:21:32.605477 4870 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4d00ede0-cd53-4791-9325-c712ee891839-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 14 07:21:32 crc kubenswrapper[4870]: I1014 07:21:32.605556 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x6hww\" (UniqueName: \"kubernetes.io/projected/4d00ede0-cd53-4791-9325-c712ee891839-kube-api-access-x6hww\") on node \"crc\" DevicePath \"\"" Oct 14 07:21:32 crc kubenswrapper[4870]: I1014 07:21:32.635691 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d00ede0-cd53-4791-9325-c712ee891839-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4d00ede0-cd53-4791-9325-c712ee891839" (UID: "4d00ede0-cd53-4791-9325-c712ee891839"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:21:32 crc kubenswrapper[4870]: I1014 07:21:32.643508 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 14 07:21:32 crc kubenswrapper[4870]: I1014 07:21:32.669524 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 14 07:21:32 crc kubenswrapper[4870]: I1014 07:21:32.691194 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 14 07:21:32 crc kubenswrapper[4870]: I1014 07:21:32.693592 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 14 07:21:32 crc kubenswrapper[4870]: I1014 07:21:32.703311 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 14 07:21:32 crc kubenswrapper[4870]: I1014 07:21:32.708652 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c712b87a-cf4d-437e-abe0-1a566046ca66-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"c712b87a-cf4d-437e-abe0-1a566046ca66\") " pod="openstack/glance-default-external-api-0" Oct 14 07:21:32 crc kubenswrapper[4870]: I1014 07:21:32.708700 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hzjsv\" (UniqueName: \"kubernetes.io/projected/c712b87a-cf4d-437e-abe0-1a566046ca66-kube-api-access-hzjsv\") pod \"glance-default-external-api-0\" (UID: \"c712b87a-cf4d-437e-abe0-1a566046ca66\") " pod="openstack/glance-default-external-api-0" Oct 14 07:21:32 crc kubenswrapper[4870]: I1014 07:21:32.708778 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c712b87a-cf4d-437e-abe0-1a566046ca66-config-data\") pod \"glance-default-external-api-0\" (UID: \"c712b87a-cf4d-437e-abe0-1a566046ca66\") " pod="openstack/glance-default-external-api-0" Oct 14 07:21:32 crc kubenswrapper[4870]: I1014 07:21:32.708811 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c712b87a-cf4d-437e-abe0-1a566046ca66-scripts\") pod \"glance-default-external-api-0\" (UID: \"c712b87a-cf4d-437e-abe0-1a566046ca66\") " pod="openstack/glance-default-external-api-0" Oct 14 07:21:32 crc kubenswrapper[4870]: I1014 07:21:32.708830 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c712b87a-cf4d-437e-abe0-1a566046ca66-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"c712b87a-cf4d-437e-abe0-1a566046ca66\") " pod="openstack/glance-default-external-api-0" Oct 14 07:21:32 crc kubenswrapper[4870]: I1014 07:21:32.708849 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c712b87a-cf4d-437e-abe0-1a566046ca66-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"c712b87a-cf4d-437e-abe0-1a566046ca66\") " pod="openstack/glance-default-external-api-0" Oct 14 07:21:32 crc kubenswrapper[4870]: I1014 07:21:32.708874 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"c712b87a-cf4d-437e-abe0-1a566046ca66\") " pod="openstack/glance-default-external-api-0" Oct 14 07:21:32 crc kubenswrapper[4870]: I1014 07:21:32.708898 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c712b87a-cf4d-437e-abe0-1a566046ca66-logs\") pod \"glance-default-external-api-0\" (UID: \"c712b87a-cf4d-437e-abe0-1a566046ca66\") " pod="openstack/glance-default-external-api-0" Oct 14 07:21:32 crc kubenswrapper[4870]: I1014 07:21:32.708946 4870 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d00ede0-cd53-4791-9325-c712ee891839-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 07:21:32 crc kubenswrapper[4870]: I1014 07:21:32.709338 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c712b87a-cf4d-437e-abe0-1a566046ca66-logs\") pod \"glance-default-external-api-0\" (UID: \"c712b87a-cf4d-437e-abe0-1a566046ca66\") " pod="openstack/glance-default-external-api-0" Oct 14 07:21:32 crc kubenswrapper[4870]: I1014 07:21:32.710323 4870 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"c712b87a-cf4d-437e-abe0-1a566046ca66\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/glance-default-external-api-0" Oct 14 07:21:32 crc kubenswrapper[4870]: I1014 07:21:32.711850 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c712b87a-cf4d-437e-abe0-1a566046ca66-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"c712b87a-cf4d-437e-abe0-1a566046ca66\") " pod="openstack/glance-default-external-api-0" Oct 14 07:21:32 crc kubenswrapper[4870]: I1014 07:21:32.716963 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Oct 14 07:21:32 crc kubenswrapper[4870]: I1014 07:21:32.733544 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c712b87a-cf4d-437e-abe0-1a566046ca66-config-data\") pod \"glance-default-external-api-0\" (UID: \"c712b87a-cf4d-437e-abe0-1a566046ca66\") " pod="openstack/glance-default-external-api-0" Oct 14 07:21:32 crc kubenswrapper[4870]: I1014 07:21:32.733986 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c712b87a-cf4d-437e-abe0-1a566046ca66-scripts\") pod \"glance-default-external-api-0\" (UID: \"c712b87a-cf4d-437e-abe0-1a566046ca66\") " pod="openstack/glance-default-external-api-0" Oct 14 07:21:32 crc kubenswrapper[4870]: I1014 07:21:32.734013 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c712b87a-cf4d-437e-abe0-1a566046ca66-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"c712b87a-cf4d-437e-abe0-1a566046ca66\") " pod="openstack/glance-default-external-api-0" Oct 14 07:21:32 crc kubenswrapper[4870]: I1014 07:21:32.735928 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 14 07:21:32 crc kubenswrapper[4870]: I1014 07:21:32.747841 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c712b87a-cf4d-437e-abe0-1a566046ca66-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"c712b87a-cf4d-437e-abe0-1a566046ca66\") " pod="openstack/glance-default-external-api-0" Oct 14 07:21:32 crc kubenswrapper[4870]: I1014 07:21:32.771742 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"c712b87a-cf4d-437e-abe0-1a566046ca66\") " pod="openstack/glance-default-external-api-0" Oct 14 07:21:32 crc kubenswrapper[4870]: I1014 07:21:32.773282 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hzjsv\" (UniqueName: \"kubernetes.io/projected/c712b87a-cf4d-437e-abe0-1a566046ca66-kube-api-access-hzjsv\") pod \"glance-default-external-api-0\" (UID: \"c712b87a-cf4d-437e-abe0-1a566046ca66\") " pod="openstack/glance-default-external-api-0" Oct 14 07:21:32 crc kubenswrapper[4870]: I1014 07:21:32.781616 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d00ede0-cd53-4791-9325-c712ee891839-config-data" (OuterVolumeSpecName: "config-data") pod "4d00ede0-cd53-4791-9325-c712ee891839" (UID: "4d00ede0-cd53-4791-9325-c712ee891839"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:21:32 crc kubenswrapper[4870]: I1014 07:21:32.804932 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 14 07:21:32 crc kubenswrapper[4870]: I1014 07:21:32.810456 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/56eb286f-8aa2-46aa-a74d-031288a9aad1-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"56eb286f-8aa2-46aa-a74d-031288a9aad1\") " pod="openstack/glance-default-internal-api-0" Oct 14 07:21:32 crc kubenswrapper[4870]: I1014 07:21:32.810519 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"56eb286f-8aa2-46aa-a74d-031288a9aad1\") " pod="openstack/glance-default-internal-api-0" Oct 14 07:21:32 crc kubenswrapper[4870]: I1014 07:21:32.810546 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/56eb286f-8aa2-46aa-a74d-031288a9aad1-scripts\") pod \"glance-default-internal-api-0\" (UID: \"56eb286f-8aa2-46aa-a74d-031288a9aad1\") " pod="openstack/glance-default-internal-api-0" Oct 14 07:21:32 crc kubenswrapper[4870]: I1014 07:21:32.810568 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/56eb286f-8aa2-46aa-a74d-031288a9aad1-logs\") pod \"glance-default-internal-api-0\" (UID: \"56eb286f-8aa2-46aa-a74d-031288a9aad1\") " pod="openstack/glance-default-internal-api-0" Oct 14 07:21:32 crc kubenswrapper[4870]: I1014 07:21:32.810584 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/56eb286f-8aa2-46aa-a74d-031288a9aad1-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"56eb286f-8aa2-46aa-a74d-031288a9aad1\") " pod="openstack/glance-default-internal-api-0" Oct 14 07:21:32 crc kubenswrapper[4870]: I1014 07:21:32.810631 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56eb286f-8aa2-46aa-a74d-031288a9aad1-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"56eb286f-8aa2-46aa-a74d-031288a9aad1\") " pod="openstack/glance-default-internal-api-0" Oct 14 07:21:32 crc kubenswrapper[4870]: I1014 07:21:32.810651 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/56eb286f-8aa2-46aa-a74d-031288a9aad1-config-data\") pod \"glance-default-internal-api-0\" (UID: \"56eb286f-8aa2-46aa-a74d-031288a9aad1\") " pod="openstack/glance-default-internal-api-0" Oct 14 07:21:32 crc kubenswrapper[4870]: I1014 07:21:32.810676 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4w2qr\" (UniqueName: \"kubernetes.io/projected/56eb286f-8aa2-46aa-a74d-031288a9aad1-kube-api-access-4w2qr\") pod \"glance-default-internal-api-0\" (UID: \"56eb286f-8aa2-46aa-a74d-031288a9aad1\") " pod="openstack/glance-default-internal-api-0" Oct 14 07:21:32 crc kubenswrapper[4870]: I1014 07:21:32.810725 4870 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4d00ede0-cd53-4791-9325-c712ee891839-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 07:21:32 crc kubenswrapper[4870]: I1014 07:21:32.911708 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/56eb286f-8aa2-46aa-a74d-031288a9aad1-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"56eb286f-8aa2-46aa-a74d-031288a9aad1\") " pod="openstack/glance-default-internal-api-0" Oct 14 07:21:32 crc kubenswrapper[4870]: I1014 07:21:32.912059 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"56eb286f-8aa2-46aa-a74d-031288a9aad1\") " pod="openstack/glance-default-internal-api-0" Oct 14 07:21:32 crc kubenswrapper[4870]: I1014 07:21:32.912091 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/56eb286f-8aa2-46aa-a74d-031288a9aad1-scripts\") pod \"glance-default-internal-api-0\" (UID: \"56eb286f-8aa2-46aa-a74d-031288a9aad1\") " pod="openstack/glance-default-internal-api-0" Oct 14 07:21:32 crc kubenswrapper[4870]: I1014 07:21:32.912116 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/56eb286f-8aa2-46aa-a74d-031288a9aad1-logs\") pod \"glance-default-internal-api-0\" (UID: \"56eb286f-8aa2-46aa-a74d-031288a9aad1\") " pod="openstack/glance-default-internal-api-0" Oct 14 07:21:32 crc kubenswrapper[4870]: I1014 07:21:32.912144 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/56eb286f-8aa2-46aa-a74d-031288a9aad1-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"56eb286f-8aa2-46aa-a74d-031288a9aad1\") " pod="openstack/glance-default-internal-api-0" Oct 14 07:21:32 crc kubenswrapper[4870]: I1014 07:21:32.912208 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/56eb286f-8aa2-46aa-a74d-031288a9aad1-config-data\") pod \"glance-default-internal-api-0\" (UID: \"56eb286f-8aa2-46aa-a74d-031288a9aad1\") " pod="openstack/glance-default-internal-api-0" Oct 14 07:21:32 crc kubenswrapper[4870]: I1014 07:21:32.912231 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56eb286f-8aa2-46aa-a74d-031288a9aad1-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"56eb286f-8aa2-46aa-a74d-031288a9aad1\") " pod="openstack/glance-default-internal-api-0" Oct 14 07:21:32 crc kubenswrapper[4870]: I1014 07:21:32.912262 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4w2qr\" (UniqueName: \"kubernetes.io/projected/56eb286f-8aa2-46aa-a74d-031288a9aad1-kube-api-access-4w2qr\") pod \"glance-default-internal-api-0\" (UID: \"56eb286f-8aa2-46aa-a74d-031288a9aad1\") " pod="openstack/glance-default-internal-api-0" Oct 14 07:21:32 crc kubenswrapper[4870]: I1014 07:21:32.912344 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/56eb286f-8aa2-46aa-a74d-031288a9aad1-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"56eb286f-8aa2-46aa-a74d-031288a9aad1\") " pod="openstack/glance-default-internal-api-0" Oct 14 07:21:32 crc kubenswrapper[4870]: I1014 07:21:32.912633 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/56eb286f-8aa2-46aa-a74d-031288a9aad1-logs\") pod \"glance-default-internal-api-0\" (UID: \"56eb286f-8aa2-46aa-a74d-031288a9aad1\") " pod="openstack/glance-default-internal-api-0" Oct 14 07:21:32 crc kubenswrapper[4870]: I1014 07:21:32.912944 4870 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"56eb286f-8aa2-46aa-a74d-031288a9aad1\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/glance-default-internal-api-0" Oct 14 07:21:32 crc kubenswrapper[4870]: I1014 07:21:32.917882 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/56eb286f-8aa2-46aa-a74d-031288a9aad1-config-data\") pod \"glance-default-internal-api-0\" (UID: \"56eb286f-8aa2-46aa-a74d-031288a9aad1\") " pod="openstack/glance-default-internal-api-0" Oct 14 07:21:32 crc kubenswrapper[4870]: I1014 07:21:32.923585 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/56eb286f-8aa2-46aa-a74d-031288a9aad1-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"56eb286f-8aa2-46aa-a74d-031288a9aad1\") " pod="openstack/glance-default-internal-api-0" Oct 14 07:21:32 crc kubenswrapper[4870]: I1014 07:21:32.926244 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56eb286f-8aa2-46aa-a74d-031288a9aad1-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"56eb286f-8aa2-46aa-a74d-031288a9aad1\") " pod="openstack/glance-default-internal-api-0" Oct 14 07:21:32 crc kubenswrapper[4870]: I1014 07:21:32.930966 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/56eb286f-8aa2-46aa-a74d-031288a9aad1-scripts\") pod \"glance-default-internal-api-0\" (UID: \"56eb286f-8aa2-46aa-a74d-031288a9aad1\") " pod="openstack/glance-default-internal-api-0" Oct 14 07:21:32 crc kubenswrapper[4870]: I1014 07:21:32.931935 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4w2qr\" (UniqueName: \"kubernetes.io/projected/56eb286f-8aa2-46aa-a74d-031288a9aad1-kube-api-access-4w2qr\") pod \"glance-default-internal-api-0\" (UID: \"56eb286f-8aa2-46aa-a74d-031288a9aad1\") " pod="openstack/glance-default-internal-api-0" Oct 14 07:21:32 crc kubenswrapper[4870]: I1014 07:21:32.967273 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"56eb286f-8aa2-46aa-a74d-031288a9aad1\") " pod="openstack/glance-default-internal-api-0" Oct 14 07:21:33 crc kubenswrapper[4870]: I1014 07:21:33.033948 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 14 07:21:33 crc kubenswrapper[4870]: I1014 07:21:33.047746 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="385ca977-7d50-426e-86e9-6a9416ad872c" path="/var/lib/kubelet/pods/385ca977-7d50-426e-86e9-6a9416ad872c/volumes" Oct 14 07:21:33 crc kubenswrapper[4870]: I1014 07:21:33.048599 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="452a6ce6-d9dc-4455-9304-414701c8895f" path="/var/lib/kubelet/pods/452a6ce6-d9dc-4455-9304-414701c8895f/volumes" Oct 14 07:21:33 crc kubenswrapper[4870]: I1014 07:21:33.321633 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 14 07:21:33 crc kubenswrapper[4870]: I1014 07:21:33.355739 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 14 07:21:33 crc kubenswrapper[4870]: I1014 07:21:33.371760 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 14 07:21:33 crc kubenswrapper[4870]: I1014 07:21:33.378980 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 14 07:21:33 crc kubenswrapper[4870]: I1014 07:21:33.381505 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 14 07:21:33 crc kubenswrapper[4870]: I1014 07:21:33.383551 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 14 07:21:33 crc kubenswrapper[4870]: I1014 07:21:33.383944 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 14 07:21:33 crc kubenswrapper[4870]: I1014 07:21:33.390035 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 14 07:21:33 crc kubenswrapper[4870]: I1014 07:21:33.423451 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 14 07:21:33 crc kubenswrapper[4870]: I1014 07:21:33.437572 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b24fea1-48c7-4100-9811-5739493548d9-config-data\") pod \"ceilometer-0\" (UID: \"5b24fea1-48c7-4100-9811-5739493548d9\") " pod="openstack/ceilometer-0" Oct 14 07:21:33 crc kubenswrapper[4870]: I1014 07:21:33.437702 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-62d7x\" (UniqueName: \"kubernetes.io/projected/5b24fea1-48c7-4100-9811-5739493548d9-kube-api-access-62d7x\") pod \"ceilometer-0\" (UID: \"5b24fea1-48c7-4100-9811-5739493548d9\") " pod="openstack/ceilometer-0" Oct 14 07:21:33 crc kubenswrapper[4870]: I1014 07:21:33.437754 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5b24fea1-48c7-4100-9811-5739493548d9-run-httpd\") pod \"ceilometer-0\" (UID: \"5b24fea1-48c7-4100-9811-5739493548d9\") " pod="openstack/ceilometer-0" Oct 14 07:21:33 crc kubenswrapper[4870]: I1014 07:21:33.437862 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b24fea1-48c7-4100-9811-5739493548d9-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"5b24fea1-48c7-4100-9811-5739493548d9\") " pod="openstack/ceilometer-0" Oct 14 07:21:33 crc kubenswrapper[4870]: I1014 07:21:33.437901 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5b24fea1-48c7-4100-9811-5739493548d9-log-httpd\") pod \"ceilometer-0\" (UID: \"5b24fea1-48c7-4100-9811-5739493548d9\") " pod="openstack/ceilometer-0" Oct 14 07:21:33 crc kubenswrapper[4870]: I1014 07:21:33.437928 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5b24fea1-48c7-4100-9811-5739493548d9-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"5b24fea1-48c7-4100-9811-5739493548d9\") " pod="openstack/ceilometer-0" Oct 14 07:21:33 crc kubenswrapper[4870]: I1014 07:21:33.438002 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5b24fea1-48c7-4100-9811-5739493548d9-scripts\") pod \"ceilometer-0\" (UID: \"5b24fea1-48c7-4100-9811-5739493548d9\") " pod="openstack/ceilometer-0" Oct 14 07:21:33 crc kubenswrapper[4870]: I1014 07:21:33.540026 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b24fea1-48c7-4100-9811-5739493548d9-config-data\") pod \"ceilometer-0\" (UID: \"5b24fea1-48c7-4100-9811-5739493548d9\") " pod="openstack/ceilometer-0" Oct 14 07:21:33 crc kubenswrapper[4870]: I1014 07:21:33.540089 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-62d7x\" (UniqueName: \"kubernetes.io/projected/5b24fea1-48c7-4100-9811-5739493548d9-kube-api-access-62d7x\") pod \"ceilometer-0\" (UID: \"5b24fea1-48c7-4100-9811-5739493548d9\") " pod="openstack/ceilometer-0" Oct 14 07:21:33 crc kubenswrapper[4870]: I1014 07:21:33.540119 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5b24fea1-48c7-4100-9811-5739493548d9-run-httpd\") pod \"ceilometer-0\" (UID: \"5b24fea1-48c7-4100-9811-5739493548d9\") " pod="openstack/ceilometer-0" Oct 14 07:21:33 crc kubenswrapper[4870]: I1014 07:21:33.540178 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b24fea1-48c7-4100-9811-5739493548d9-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"5b24fea1-48c7-4100-9811-5739493548d9\") " pod="openstack/ceilometer-0" Oct 14 07:21:33 crc kubenswrapper[4870]: I1014 07:21:33.540202 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5b24fea1-48c7-4100-9811-5739493548d9-log-httpd\") pod \"ceilometer-0\" (UID: \"5b24fea1-48c7-4100-9811-5739493548d9\") " pod="openstack/ceilometer-0" Oct 14 07:21:33 crc kubenswrapper[4870]: I1014 07:21:33.540222 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5b24fea1-48c7-4100-9811-5739493548d9-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"5b24fea1-48c7-4100-9811-5739493548d9\") " pod="openstack/ceilometer-0" Oct 14 07:21:33 crc kubenswrapper[4870]: I1014 07:21:33.540268 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5b24fea1-48c7-4100-9811-5739493548d9-scripts\") pod \"ceilometer-0\" (UID: \"5b24fea1-48c7-4100-9811-5739493548d9\") " pod="openstack/ceilometer-0" Oct 14 07:21:33 crc kubenswrapper[4870]: I1014 07:21:33.540845 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5b24fea1-48c7-4100-9811-5739493548d9-log-httpd\") pod \"ceilometer-0\" (UID: \"5b24fea1-48c7-4100-9811-5739493548d9\") " pod="openstack/ceilometer-0" Oct 14 07:21:33 crc kubenswrapper[4870]: I1014 07:21:33.541125 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5b24fea1-48c7-4100-9811-5739493548d9-run-httpd\") pod \"ceilometer-0\" (UID: \"5b24fea1-48c7-4100-9811-5739493548d9\") " pod="openstack/ceilometer-0" Oct 14 07:21:33 crc kubenswrapper[4870]: I1014 07:21:33.547673 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b24fea1-48c7-4100-9811-5739493548d9-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"5b24fea1-48c7-4100-9811-5739493548d9\") " pod="openstack/ceilometer-0" Oct 14 07:21:33 crc kubenswrapper[4870]: I1014 07:21:33.548202 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b24fea1-48c7-4100-9811-5739493548d9-config-data\") pod \"ceilometer-0\" (UID: \"5b24fea1-48c7-4100-9811-5739493548d9\") " pod="openstack/ceilometer-0" Oct 14 07:21:33 crc kubenswrapper[4870]: I1014 07:21:33.565190 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5b24fea1-48c7-4100-9811-5739493548d9-scripts\") pod \"ceilometer-0\" (UID: \"5b24fea1-48c7-4100-9811-5739493548d9\") " pod="openstack/ceilometer-0" Oct 14 07:21:33 crc kubenswrapper[4870]: I1014 07:21:33.568136 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-62d7x\" (UniqueName: \"kubernetes.io/projected/5b24fea1-48c7-4100-9811-5739493548d9-kube-api-access-62d7x\") pod \"ceilometer-0\" (UID: \"5b24fea1-48c7-4100-9811-5739493548d9\") " pod="openstack/ceilometer-0" Oct 14 07:21:33 crc kubenswrapper[4870]: I1014 07:21:33.573648 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 14 07:21:33 crc kubenswrapper[4870]: I1014 07:21:33.576604 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5b24fea1-48c7-4100-9811-5739493548d9-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"5b24fea1-48c7-4100-9811-5739493548d9\") " pod="openstack/ceilometer-0" Oct 14 07:21:33 crc kubenswrapper[4870]: W1014 07:21:33.592165 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod56eb286f_8aa2_46aa_a74d_031288a9aad1.slice/crio-d2ecb31883121abad720cfb94674a6f06ef98dccfae53eb34ebb2671e32a13ec WatchSource:0}: Error finding container d2ecb31883121abad720cfb94674a6f06ef98dccfae53eb34ebb2671e32a13ec: Status 404 returned error can't find the container with id d2ecb31883121abad720cfb94674a6f06ef98dccfae53eb34ebb2671e32a13ec Oct 14 07:21:33 crc kubenswrapper[4870]: I1014 07:21:33.698045 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 14 07:21:34 crc kubenswrapper[4870]: I1014 07:21:34.245252 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 14 07:21:34 crc kubenswrapper[4870]: I1014 07:21:34.330476 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"56eb286f-8aa2-46aa-a74d-031288a9aad1","Type":"ContainerStarted","Data":"9fa2f1146313b498f55c9bb3361bca48d735f25641dcd24bdbeac19bc4ae6377"} Oct 14 07:21:34 crc kubenswrapper[4870]: I1014 07:21:34.330524 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"56eb286f-8aa2-46aa-a74d-031288a9aad1","Type":"ContainerStarted","Data":"d2ecb31883121abad720cfb94674a6f06ef98dccfae53eb34ebb2671e32a13ec"} Oct 14 07:21:34 crc kubenswrapper[4870]: I1014 07:21:34.332274 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"c712b87a-cf4d-437e-abe0-1a566046ca66","Type":"ContainerStarted","Data":"6c0de18ab40ba16d1bc75ff9b157bb26252b1a8044809c3ee46ea0a31ed39019"} Oct 14 07:21:34 crc kubenswrapper[4870]: I1014 07:21:34.332300 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"c712b87a-cf4d-437e-abe0-1a566046ca66","Type":"ContainerStarted","Data":"c7369a66d5079dd52c04fcb0951209f2fbdc37e0b77e054874a3ae686da65c90"} Oct 14 07:21:34 crc kubenswrapper[4870]: I1014 07:21:34.334746 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5b24fea1-48c7-4100-9811-5739493548d9","Type":"ContainerStarted","Data":"d51986a7b83298615f06406cd1aa51ad839cd41bd4765cea9283e62969c08b9e"} Oct 14 07:21:34 crc kubenswrapper[4870]: I1014 07:21:34.933960 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Oct 14 07:21:35 crc kubenswrapper[4870]: I1014 07:21:35.061469 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4d00ede0-cd53-4791-9325-c712ee891839" path="/var/lib/kubelet/pods/4d00ede0-cd53-4791-9325-c712ee891839/volumes" Oct 14 07:21:35 crc kubenswrapper[4870]: I1014 07:21:35.342580 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"c712b87a-cf4d-437e-abe0-1a566046ca66","Type":"ContainerStarted","Data":"f744eb9b855e3191279bf5fc01139d0a3359c09a6178c62d0602563d9d13ff25"} Oct 14 07:21:35 crc kubenswrapper[4870]: I1014 07:21:35.343980 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5b24fea1-48c7-4100-9811-5739493548d9","Type":"ContainerStarted","Data":"fd6ab4f83f9cec1b430d4f0a7333b16f2dbbe88473f79bb5bab6c7bdc39b4fa9"} Oct 14 07:21:35 crc kubenswrapper[4870]: I1014 07:21:35.347778 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"56eb286f-8aa2-46aa-a74d-031288a9aad1","Type":"ContainerStarted","Data":"d8fce18d04004b32792441ce25ddcd9d74c06df092b20d09c14fca60675539d4"} Oct 14 07:21:35 crc kubenswrapper[4870]: I1014 07:21:35.376814 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=3.3767887659999998 podStartE2EDuration="3.376788766s" podCreationTimestamp="2025-10-14 07:21:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:21:35.375003144 +0000 UTC m=+1231.072363515" watchObservedRunningTime="2025-10-14 07:21:35.376788766 +0000 UTC m=+1231.074149137" Oct 14 07:21:35 crc kubenswrapper[4870]: I1014 07:21:35.401264 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=3.401239097 podStartE2EDuration="3.401239097s" podCreationTimestamp="2025-10-14 07:21:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:21:35.395779259 +0000 UTC m=+1231.093139630" watchObservedRunningTime="2025-10-14 07:21:35.401239097 +0000 UTC m=+1231.098599468" Oct 14 07:21:36 crc kubenswrapper[4870]: I1014 07:21:36.369066 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5b24fea1-48c7-4100-9811-5739493548d9","Type":"ContainerStarted","Data":"1d2d421419015d3d234317a1289d58133709959153d10dddb4ab2b71247221fe"} Oct 14 07:21:37 crc kubenswrapper[4870]: I1014 07:21:37.379022 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5b24fea1-48c7-4100-9811-5739493548d9","Type":"ContainerStarted","Data":"1ffa055509f25c5728647dcd15f7a11a8d373ff6501e1d0878f249cceb376ac0"} Oct 14 07:21:38 crc kubenswrapper[4870]: I1014 07:21:38.391580 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5b24fea1-48c7-4100-9811-5739493548d9","Type":"ContainerStarted","Data":"af919d31f8bfe4643ff283680c9954ca75a7f7e3d0864f81c9c01c34d175d412"} Oct 14 07:21:38 crc kubenswrapper[4870]: I1014 07:21:38.392143 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 14 07:21:38 crc kubenswrapper[4870]: I1014 07:21:38.419571 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.004213094 podStartE2EDuration="5.41955535s" podCreationTimestamp="2025-10-14 07:21:33 +0000 UTC" firstStartedPulling="2025-10-14 07:21:34.234669716 +0000 UTC m=+1229.932030087" lastFinishedPulling="2025-10-14 07:21:37.650011972 +0000 UTC m=+1233.347372343" observedRunningTime="2025-10-14 07:21:38.418040565 +0000 UTC m=+1234.115400936" watchObservedRunningTime="2025-10-14 07:21:38.41955535 +0000 UTC m=+1234.116915721" Oct 14 07:21:42 crc kubenswrapper[4870]: I1014 07:21:42.382813 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 14 07:21:42 crc kubenswrapper[4870]: I1014 07:21:42.384935 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="5b24fea1-48c7-4100-9811-5739493548d9" containerName="ceilometer-central-agent" containerID="cri-o://fd6ab4f83f9cec1b430d4f0a7333b16f2dbbe88473f79bb5bab6c7bdc39b4fa9" gracePeriod=30 Oct 14 07:21:42 crc kubenswrapper[4870]: I1014 07:21:42.385043 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="5b24fea1-48c7-4100-9811-5739493548d9" containerName="proxy-httpd" containerID="cri-o://af919d31f8bfe4643ff283680c9954ca75a7f7e3d0864f81c9c01c34d175d412" gracePeriod=30 Oct 14 07:21:42 crc kubenswrapper[4870]: I1014 07:21:42.385093 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="5b24fea1-48c7-4100-9811-5739493548d9" containerName="sg-core" containerID="cri-o://1ffa055509f25c5728647dcd15f7a11a8d373ff6501e1d0878f249cceb376ac0" gracePeriod=30 Oct 14 07:21:42 crc kubenswrapper[4870]: I1014 07:21:42.385145 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="5b24fea1-48c7-4100-9811-5739493548d9" containerName="ceilometer-notification-agent" containerID="cri-o://1d2d421419015d3d234317a1289d58133709959153d10dddb4ab2b71247221fe" gracePeriod=30 Oct 14 07:21:42 crc kubenswrapper[4870]: I1014 07:21:42.805944 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 14 07:21:42 crc kubenswrapper[4870]: I1014 07:21:42.805994 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 14 07:21:42 crc kubenswrapper[4870]: I1014 07:21:42.840358 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 14 07:21:42 crc kubenswrapper[4870]: I1014 07:21:42.861920 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 14 07:21:43 crc kubenswrapper[4870]: I1014 07:21:43.054924 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 14 07:21:43 crc kubenswrapper[4870]: I1014 07:21:43.055238 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 14 07:21:43 crc kubenswrapper[4870]: I1014 07:21:43.088754 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 14 07:21:43 crc kubenswrapper[4870]: I1014 07:21:43.114090 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 14 07:21:43 crc kubenswrapper[4870]: I1014 07:21:43.162073 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 14 07:21:43 crc kubenswrapper[4870]: I1014 07:21:43.274067 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b24fea1-48c7-4100-9811-5739493548d9-config-data\") pod \"5b24fea1-48c7-4100-9811-5739493548d9\" (UID: \"5b24fea1-48c7-4100-9811-5739493548d9\") " Oct 14 07:21:43 crc kubenswrapper[4870]: I1014 07:21:43.274216 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5b24fea1-48c7-4100-9811-5739493548d9-run-httpd\") pod \"5b24fea1-48c7-4100-9811-5739493548d9\" (UID: \"5b24fea1-48c7-4100-9811-5739493548d9\") " Oct 14 07:21:43 crc kubenswrapper[4870]: I1014 07:21:43.275001 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5b24fea1-48c7-4100-9811-5739493548d9-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "5b24fea1-48c7-4100-9811-5739493548d9" (UID: "5b24fea1-48c7-4100-9811-5739493548d9"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:21:43 crc kubenswrapper[4870]: I1014 07:21:43.275191 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-62d7x\" (UniqueName: \"kubernetes.io/projected/5b24fea1-48c7-4100-9811-5739493548d9-kube-api-access-62d7x\") pod \"5b24fea1-48c7-4100-9811-5739493548d9\" (UID: \"5b24fea1-48c7-4100-9811-5739493548d9\") " Oct 14 07:21:43 crc kubenswrapper[4870]: I1014 07:21:43.275856 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5b24fea1-48c7-4100-9811-5739493548d9-scripts\") pod \"5b24fea1-48c7-4100-9811-5739493548d9\" (UID: \"5b24fea1-48c7-4100-9811-5739493548d9\") " Oct 14 07:21:43 crc kubenswrapper[4870]: I1014 07:21:43.275979 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5b24fea1-48c7-4100-9811-5739493548d9-log-httpd\") pod \"5b24fea1-48c7-4100-9811-5739493548d9\" (UID: \"5b24fea1-48c7-4100-9811-5739493548d9\") " Oct 14 07:21:43 crc kubenswrapper[4870]: I1014 07:21:43.276065 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b24fea1-48c7-4100-9811-5739493548d9-combined-ca-bundle\") pod \"5b24fea1-48c7-4100-9811-5739493548d9\" (UID: \"5b24fea1-48c7-4100-9811-5739493548d9\") " Oct 14 07:21:43 crc kubenswrapper[4870]: I1014 07:21:43.276114 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5b24fea1-48c7-4100-9811-5739493548d9-sg-core-conf-yaml\") pod \"5b24fea1-48c7-4100-9811-5739493548d9\" (UID: \"5b24fea1-48c7-4100-9811-5739493548d9\") " Oct 14 07:21:43 crc kubenswrapper[4870]: I1014 07:21:43.276477 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5b24fea1-48c7-4100-9811-5739493548d9-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "5b24fea1-48c7-4100-9811-5739493548d9" (UID: "5b24fea1-48c7-4100-9811-5739493548d9"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:21:43 crc kubenswrapper[4870]: I1014 07:21:43.277081 4870 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5b24fea1-48c7-4100-9811-5739493548d9-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 14 07:21:43 crc kubenswrapper[4870]: I1014 07:21:43.277101 4870 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5b24fea1-48c7-4100-9811-5739493548d9-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 14 07:21:43 crc kubenswrapper[4870]: I1014 07:21:43.280328 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b24fea1-48c7-4100-9811-5739493548d9-kube-api-access-62d7x" (OuterVolumeSpecName: "kube-api-access-62d7x") pod "5b24fea1-48c7-4100-9811-5739493548d9" (UID: "5b24fea1-48c7-4100-9811-5739493548d9"). InnerVolumeSpecName "kube-api-access-62d7x". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:21:43 crc kubenswrapper[4870]: I1014 07:21:43.280644 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b24fea1-48c7-4100-9811-5739493548d9-scripts" (OuterVolumeSpecName: "scripts") pod "5b24fea1-48c7-4100-9811-5739493548d9" (UID: "5b24fea1-48c7-4100-9811-5739493548d9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:21:43 crc kubenswrapper[4870]: I1014 07:21:43.310586 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b24fea1-48c7-4100-9811-5739493548d9-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "5b24fea1-48c7-4100-9811-5739493548d9" (UID: "5b24fea1-48c7-4100-9811-5739493548d9"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:21:43 crc kubenswrapper[4870]: I1014 07:21:43.346652 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b24fea1-48c7-4100-9811-5739493548d9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5b24fea1-48c7-4100-9811-5739493548d9" (UID: "5b24fea1-48c7-4100-9811-5739493548d9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:21:43 crc kubenswrapper[4870]: I1014 07:21:43.378426 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-62d7x\" (UniqueName: \"kubernetes.io/projected/5b24fea1-48c7-4100-9811-5739493548d9-kube-api-access-62d7x\") on node \"crc\" DevicePath \"\"" Oct 14 07:21:43 crc kubenswrapper[4870]: I1014 07:21:43.378469 4870 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5b24fea1-48c7-4100-9811-5739493548d9-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 07:21:43 crc kubenswrapper[4870]: I1014 07:21:43.378478 4870 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b24fea1-48c7-4100-9811-5739493548d9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 07:21:43 crc kubenswrapper[4870]: I1014 07:21:43.378487 4870 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5b24fea1-48c7-4100-9811-5739493548d9-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 14 07:21:43 crc kubenswrapper[4870]: I1014 07:21:43.391422 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b24fea1-48c7-4100-9811-5739493548d9-config-data" (OuterVolumeSpecName: "config-data") pod "5b24fea1-48c7-4100-9811-5739493548d9" (UID: "5b24fea1-48c7-4100-9811-5739493548d9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:21:43 crc kubenswrapper[4870]: I1014 07:21:43.446173 4870 generic.go:334] "Generic (PLEG): container finished" podID="5b24fea1-48c7-4100-9811-5739493548d9" containerID="af919d31f8bfe4643ff283680c9954ca75a7f7e3d0864f81c9c01c34d175d412" exitCode=0 Oct 14 07:21:43 crc kubenswrapper[4870]: I1014 07:21:43.446211 4870 generic.go:334] "Generic (PLEG): container finished" podID="5b24fea1-48c7-4100-9811-5739493548d9" containerID="1ffa055509f25c5728647dcd15f7a11a8d373ff6501e1d0878f249cceb376ac0" exitCode=2 Oct 14 07:21:43 crc kubenswrapper[4870]: I1014 07:21:43.446223 4870 generic.go:334] "Generic (PLEG): container finished" podID="5b24fea1-48c7-4100-9811-5739493548d9" containerID="1d2d421419015d3d234317a1289d58133709959153d10dddb4ab2b71247221fe" exitCode=0 Oct 14 07:21:43 crc kubenswrapper[4870]: I1014 07:21:43.446232 4870 generic.go:334] "Generic (PLEG): container finished" podID="5b24fea1-48c7-4100-9811-5739493548d9" containerID="fd6ab4f83f9cec1b430d4f0a7333b16f2dbbe88473f79bb5bab6c7bdc39b4fa9" exitCode=0 Oct 14 07:21:43 crc kubenswrapper[4870]: I1014 07:21:43.446242 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 14 07:21:43 crc kubenswrapper[4870]: I1014 07:21:43.446267 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5b24fea1-48c7-4100-9811-5739493548d9","Type":"ContainerDied","Data":"af919d31f8bfe4643ff283680c9954ca75a7f7e3d0864f81c9c01c34d175d412"} Oct 14 07:21:43 crc kubenswrapper[4870]: I1014 07:21:43.446311 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5b24fea1-48c7-4100-9811-5739493548d9","Type":"ContainerDied","Data":"1ffa055509f25c5728647dcd15f7a11a8d373ff6501e1d0878f249cceb376ac0"} Oct 14 07:21:43 crc kubenswrapper[4870]: I1014 07:21:43.446323 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5b24fea1-48c7-4100-9811-5739493548d9","Type":"ContainerDied","Data":"1d2d421419015d3d234317a1289d58133709959153d10dddb4ab2b71247221fe"} Oct 14 07:21:43 crc kubenswrapper[4870]: I1014 07:21:43.446334 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5b24fea1-48c7-4100-9811-5739493548d9","Type":"ContainerDied","Data":"fd6ab4f83f9cec1b430d4f0a7333b16f2dbbe88473f79bb5bab6c7bdc39b4fa9"} Oct 14 07:21:43 crc kubenswrapper[4870]: I1014 07:21:43.446344 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5b24fea1-48c7-4100-9811-5739493548d9","Type":"ContainerDied","Data":"d51986a7b83298615f06406cd1aa51ad839cd41bd4765cea9283e62969c08b9e"} Oct 14 07:21:43 crc kubenswrapper[4870]: I1014 07:21:43.446357 4870 scope.go:117] "RemoveContainer" containerID="af919d31f8bfe4643ff283680c9954ca75a7f7e3d0864f81c9c01c34d175d412" Oct 14 07:21:43 crc kubenswrapper[4870]: I1014 07:21:43.446981 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 14 07:21:43 crc kubenswrapper[4870]: I1014 07:21:43.447476 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 14 07:21:43 crc kubenswrapper[4870]: I1014 07:21:43.447522 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 14 07:21:43 crc kubenswrapper[4870]: I1014 07:21:43.447536 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 14 07:21:43 crc kubenswrapper[4870]: I1014 07:21:43.483630 4870 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b24fea1-48c7-4100-9811-5739493548d9-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 07:21:43 crc kubenswrapper[4870]: I1014 07:21:43.484339 4870 scope.go:117] "RemoveContainer" containerID="1ffa055509f25c5728647dcd15f7a11a8d373ff6501e1d0878f249cceb376ac0" Oct 14 07:21:43 crc kubenswrapper[4870]: I1014 07:21:43.490871 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 14 07:21:43 crc kubenswrapper[4870]: I1014 07:21:43.507514 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 14 07:21:43 crc kubenswrapper[4870]: I1014 07:21:43.515496 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 14 07:21:43 crc kubenswrapper[4870]: E1014 07:21:43.515934 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b24fea1-48c7-4100-9811-5739493548d9" containerName="proxy-httpd" Oct 14 07:21:43 crc kubenswrapper[4870]: I1014 07:21:43.515952 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b24fea1-48c7-4100-9811-5739493548d9" containerName="proxy-httpd" Oct 14 07:21:43 crc kubenswrapper[4870]: E1014 07:21:43.515981 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b24fea1-48c7-4100-9811-5739493548d9" containerName="sg-core" Oct 14 07:21:43 crc kubenswrapper[4870]: I1014 07:21:43.515988 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b24fea1-48c7-4100-9811-5739493548d9" containerName="sg-core" Oct 14 07:21:43 crc kubenswrapper[4870]: E1014 07:21:43.516008 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b24fea1-48c7-4100-9811-5739493548d9" containerName="ceilometer-notification-agent" Oct 14 07:21:43 crc kubenswrapper[4870]: I1014 07:21:43.516014 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b24fea1-48c7-4100-9811-5739493548d9" containerName="ceilometer-notification-agent" Oct 14 07:21:43 crc kubenswrapper[4870]: E1014 07:21:43.516027 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b24fea1-48c7-4100-9811-5739493548d9" containerName="ceilometer-central-agent" Oct 14 07:21:43 crc kubenswrapper[4870]: I1014 07:21:43.516036 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b24fea1-48c7-4100-9811-5739493548d9" containerName="ceilometer-central-agent" Oct 14 07:21:43 crc kubenswrapper[4870]: I1014 07:21:43.516214 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="5b24fea1-48c7-4100-9811-5739493548d9" containerName="sg-core" Oct 14 07:21:43 crc kubenswrapper[4870]: I1014 07:21:43.516232 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="5b24fea1-48c7-4100-9811-5739493548d9" containerName="proxy-httpd" Oct 14 07:21:43 crc kubenswrapper[4870]: I1014 07:21:43.516243 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="5b24fea1-48c7-4100-9811-5739493548d9" containerName="ceilometer-notification-agent" Oct 14 07:21:43 crc kubenswrapper[4870]: I1014 07:21:43.516258 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="5b24fea1-48c7-4100-9811-5739493548d9" containerName="ceilometer-central-agent" Oct 14 07:21:43 crc kubenswrapper[4870]: I1014 07:21:43.517859 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 14 07:21:43 crc kubenswrapper[4870]: I1014 07:21:43.524005 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 14 07:21:43 crc kubenswrapper[4870]: I1014 07:21:43.527852 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 14 07:21:43 crc kubenswrapper[4870]: I1014 07:21:43.529560 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 14 07:21:43 crc kubenswrapper[4870]: I1014 07:21:43.611993 4870 scope.go:117] "RemoveContainer" containerID="1d2d421419015d3d234317a1289d58133709959153d10dddb4ab2b71247221fe" Oct 14 07:21:43 crc kubenswrapper[4870]: I1014 07:21:43.633924 4870 scope.go:117] "RemoveContainer" containerID="fd6ab4f83f9cec1b430d4f0a7333b16f2dbbe88473f79bb5bab6c7bdc39b4fa9" Oct 14 07:21:43 crc kubenswrapper[4870]: I1014 07:21:43.652316 4870 scope.go:117] "RemoveContainer" containerID="af919d31f8bfe4643ff283680c9954ca75a7f7e3d0864f81c9c01c34d175d412" Oct 14 07:21:43 crc kubenswrapper[4870]: E1014 07:21:43.652661 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"af919d31f8bfe4643ff283680c9954ca75a7f7e3d0864f81c9c01c34d175d412\": container with ID starting with af919d31f8bfe4643ff283680c9954ca75a7f7e3d0864f81c9c01c34d175d412 not found: ID does not exist" containerID="af919d31f8bfe4643ff283680c9954ca75a7f7e3d0864f81c9c01c34d175d412" Oct 14 07:21:43 crc kubenswrapper[4870]: I1014 07:21:43.652694 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"af919d31f8bfe4643ff283680c9954ca75a7f7e3d0864f81c9c01c34d175d412"} err="failed to get container status \"af919d31f8bfe4643ff283680c9954ca75a7f7e3d0864f81c9c01c34d175d412\": rpc error: code = NotFound desc = could not find container \"af919d31f8bfe4643ff283680c9954ca75a7f7e3d0864f81c9c01c34d175d412\": container with ID starting with af919d31f8bfe4643ff283680c9954ca75a7f7e3d0864f81c9c01c34d175d412 not found: ID does not exist" Oct 14 07:21:43 crc kubenswrapper[4870]: I1014 07:21:43.652717 4870 scope.go:117] "RemoveContainer" containerID="1ffa055509f25c5728647dcd15f7a11a8d373ff6501e1d0878f249cceb376ac0" Oct 14 07:21:43 crc kubenswrapper[4870]: E1014 07:21:43.652942 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1ffa055509f25c5728647dcd15f7a11a8d373ff6501e1d0878f249cceb376ac0\": container with ID starting with 1ffa055509f25c5728647dcd15f7a11a8d373ff6501e1d0878f249cceb376ac0 not found: ID does not exist" containerID="1ffa055509f25c5728647dcd15f7a11a8d373ff6501e1d0878f249cceb376ac0" Oct 14 07:21:43 crc kubenswrapper[4870]: I1014 07:21:43.652997 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1ffa055509f25c5728647dcd15f7a11a8d373ff6501e1d0878f249cceb376ac0"} err="failed to get container status \"1ffa055509f25c5728647dcd15f7a11a8d373ff6501e1d0878f249cceb376ac0\": rpc error: code = NotFound desc = could not find container \"1ffa055509f25c5728647dcd15f7a11a8d373ff6501e1d0878f249cceb376ac0\": container with ID starting with 1ffa055509f25c5728647dcd15f7a11a8d373ff6501e1d0878f249cceb376ac0 not found: ID does not exist" Oct 14 07:21:43 crc kubenswrapper[4870]: I1014 07:21:43.653018 4870 scope.go:117] "RemoveContainer" containerID="1d2d421419015d3d234317a1289d58133709959153d10dddb4ab2b71247221fe" Oct 14 07:21:43 crc kubenswrapper[4870]: E1014 07:21:43.653217 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1d2d421419015d3d234317a1289d58133709959153d10dddb4ab2b71247221fe\": container with ID starting with 1d2d421419015d3d234317a1289d58133709959153d10dddb4ab2b71247221fe not found: ID does not exist" containerID="1d2d421419015d3d234317a1289d58133709959153d10dddb4ab2b71247221fe" Oct 14 07:21:43 crc kubenswrapper[4870]: I1014 07:21:43.653237 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1d2d421419015d3d234317a1289d58133709959153d10dddb4ab2b71247221fe"} err="failed to get container status \"1d2d421419015d3d234317a1289d58133709959153d10dddb4ab2b71247221fe\": rpc error: code = NotFound desc = could not find container \"1d2d421419015d3d234317a1289d58133709959153d10dddb4ab2b71247221fe\": container with ID starting with 1d2d421419015d3d234317a1289d58133709959153d10dddb4ab2b71247221fe not found: ID does not exist" Oct 14 07:21:43 crc kubenswrapper[4870]: I1014 07:21:43.653251 4870 scope.go:117] "RemoveContainer" containerID="fd6ab4f83f9cec1b430d4f0a7333b16f2dbbe88473f79bb5bab6c7bdc39b4fa9" Oct 14 07:21:43 crc kubenswrapper[4870]: E1014 07:21:43.653425 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fd6ab4f83f9cec1b430d4f0a7333b16f2dbbe88473f79bb5bab6c7bdc39b4fa9\": container with ID starting with fd6ab4f83f9cec1b430d4f0a7333b16f2dbbe88473f79bb5bab6c7bdc39b4fa9 not found: ID does not exist" containerID="fd6ab4f83f9cec1b430d4f0a7333b16f2dbbe88473f79bb5bab6c7bdc39b4fa9" Oct 14 07:21:43 crc kubenswrapper[4870]: I1014 07:21:43.653483 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fd6ab4f83f9cec1b430d4f0a7333b16f2dbbe88473f79bb5bab6c7bdc39b4fa9"} err="failed to get container status \"fd6ab4f83f9cec1b430d4f0a7333b16f2dbbe88473f79bb5bab6c7bdc39b4fa9\": rpc error: code = NotFound desc = could not find container \"fd6ab4f83f9cec1b430d4f0a7333b16f2dbbe88473f79bb5bab6c7bdc39b4fa9\": container with ID starting with fd6ab4f83f9cec1b430d4f0a7333b16f2dbbe88473f79bb5bab6c7bdc39b4fa9 not found: ID does not exist" Oct 14 07:21:43 crc kubenswrapper[4870]: I1014 07:21:43.653494 4870 scope.go:117] "RemoveContainer" containerID="af919d31f8bfe4643ff283680c9954ca75a7f7e3d0864f81c9c01c34d175d412" Oct 14 07:21:43 crc kubenswrapper[4870]: I1014 07:21:43.653682 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"af919d31f8bfe4643ff283680c9954ca75a7f7e3d0864f81c9c01c34d175d412"} err="failed to get container status \"af919d31f8bfe4643ff283680c9954ca75a7f7e3d0864f81c9c01c34d175d412\": rpc error: code = NotFound desc = could not find container \"af919d31f8bfe4643ff283680c9954ca75a7f7e3d0864f81c9c01c34d175d412\": container with ID starting with af919d31f8bfe4643ff283680c9954ca75a7f7e3d0864f81c9c01c34d175d412 not found: ID does not exist" Oct 14 07:21:43 crc kubenswrapper[4870]: I1014 07:21:43.653710 4870 scope.go:117] "RemoveContainer" containerID="1ffa055509f25c5728647dcd15f7a11a8d373ff6501e1d0878f249cceb376ac0" Oct 14 07:21:43 crc kubenswrapper[4870]: I1014 07:21:43.653912 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1ffa055509f25c5728647dcd15f7a11a8d373ff6501e1d0878f249cceb376ac0"} err="failed to get container status \"1ffa055509f25c5728647dcd15f7a11a8d373ff6501e1d0878f249cceb376ac0\": rpc error: code = NotFound desc = could not find container \"1ffa055509f25c5728647dcd15f7a11a8d373ff6501e1d0878f249cceb376ac0\": container with ID starting with 1ffa055509f25c5728647dcd15f7a11a8d373ff6501e1d0878f249cceb376ac0 not found: ID does not exist" Oct 14 07:21:43 crc kubenswrapper[4870]: I1014 07:21:43.653958 4870 scope.go:117] "RemoveContainer" containerID="1d2d421419015d3d234317a1289d58133709959153d10dddb4ab2b71247221fe" Oct 14 07:21:43 crc kubenswrapper[4870]: I1014 07:21:43.654130 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1d2d421419015d3d234317a1289d58133709959153d10dddb4ab2b71247221fe"} err="failed to get container status \"1d2d421419015d3d234317a1289d58133709959153d10dddb4ab2b71247221fe\": rpc error: code = NotFound desc = could not find container \"1d2d421419015d3d234317a1289d58133709959153d10dddb4ab2b71247221fe\": container with ID starting with 1d2d421419015d3d234317a1289d58133709959153d10dddb4ab2b71247221fe not found: ID does not exist" Oct 14 07:21:43 crc kubenswrapper[4870]: I1014 07:21:43.654175 4870 scope.go:117] "RemoveContainer" containerID="fd6ab4f83f9cec1b430d4f0a7333b16f2dbbe88473f79bb5bab6c7bdc39b4fa9" Oct 14 07:21:43 crc kubenswrapper[4870]: I1014 07:21:43.654343 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fd6ab4f83f9cec1b430d4f0a7333b16f2dbbe88473f79bb5bab6c7bdc39b4fa9"} err="failed to get container status \"fd6ab4f83f9cec1b430d4f0a7333b16f2dbbe88473f79bb5bab6c7bdc39b4fa9\": rpc error: code = NotFound desc = could not find container \"fd6ab4f83f9cec1b430d4f0a7333b16f2dbbe88473f79bb5bab6c7bdc39b4fa9\": container with ID starting with fd6ab4f83f9cec1b430d4f0a7333b16f2dbbe88473f79bb5bab6c7bdc39b4fa9 not found: ID does not exist" Oct 14 07:21:43 crc kubenswrapper[4870]: I1014 07:21:43.654372 4870 scope.go:117] "RemoveContainer" containerID="af919d31f8bfe4643ff283680c9954ca75a7f7e3d0864f81c9c01c34d175d412" Oct 14 07:21:43 crc kubenswrapper[4870]: I1014 07:21:43.654560 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"af919d31f8bfe4643ff283680c9954ca75a7f7e3d0864f81c9c01c34d175d412"} err="failed to get container status \"af919d31f8bfe4643ff283680c9954ca75a7f7e3d0864f81c9c01c34d175d412\": rpc error: code = NotFound desc = could not find container \"af919d31f8bfe4643ff283680c9954ca75a7f7e3d0864f81c9c01c34d175d412\": container with ID starting with af919d31f8bfe4643ff283680c9954ca75a7f7e3d0864f81c9c01c34d175d412 not found: ID does not exist" Oct 14 07:21:43 crc kubenswrapper[4870]: I1014 07:21:43.654578 4870 scope.go:117] "RemoveContainer" containerID="1ffa055509f25c5728647dcd15f7a11a8d373ff6501e1d0878f249cceb376ac0" Oct 14 07:21:43 crc kubenswrapper[4870]: I1014 07:21:43.654756 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1ffa055509f25c5728647dcd15f7a11a8d373ff6501e1d0878f249cceb376ac0"} err="failed to get container status \"1ffa055509f25c5728647dcd15f7a11a8d373ff6501e1d0878f249cceb376ac0\": rpc error: code = NotFound desc = could not find container \"1ffa055509f25c5728647dcd15f7a11a8d373ff6501e1d0878f249cceb376ac0\": container with ID starting with 1ffa055509f25c5728647dcd15f7a11a8d373ff6501e1d0878f249cceb376ac0 not found: ID does not exist" Oct 14 07:21:43 crc kubenswrapper[4870]: I1014 07:21:43.654781 4870 scope.go:117] "RemoveContainer" containerID="1d2d421419015d3d234317a1289d58133709959153d10dddb4ab2b71247221fe" Oct 14 07:21:43 crc kubenswrapper[4870]: I1014 07:21:43.654976 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1d2d421419015d3d234317a1289d58133709959153d10dddb4ab2b71247221fe"} err="failed to get container status \"1d2d421419015d3d234317a1289d58133709959153d10dddb4ab2b71247221fe\": rpc error: code = NotFound desc = could not find container \"1d2d421419015d3d234317a1289d58133709959153d10dddb4ab2b71247221fe\": container with ID starting with 1d2d421419015d3d234317a1289d58133709959153d10dddb4ab2b71247221fe not found: ID does not exist" Oct 14 07:21:43 crc kubenswrapper[4870]: I1014 07:21:43.655020 4870 scope.go:117] "RemoveContainer" containerID="fd6ab4f83f9cec1b430d4f0a7333b16f2dbbe88473f79bb5bab6c7bdc39b4fa9" Oct 14 07:21:43 crc kubenswrapper[4870]: I1014 07:21:43.655191 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fd6ab4f83f9cec1b430d4f0a7333b16f2dbbe88473f79bb5bab6c7bdc39b4fa9"} err="failed to get container status \"fd6ab4f83f9cec1b430d4f0a7333b16f2dbbe88473f79bb5bab6c7bdc39b4fa9\": rpc error: code = NotFound desc = could not find container \"fd6ab4f83f9cec1b430d4f0a7333b16f2dbbe88473f79bb5bab6c7bdc39b4fa9\": container with ID starting with fd6ab4f83f9cec1b430d4f0a7333b16f2dbbe88473f79bb5bab6c7bdc39b4fa9 not found: ID does not exist" Oct 14 07:21:43 crc kubenswrapper[4870]: I1014 07:21:43.655215 4870 scope.go:117] "RemoveContainer" containerID="af919d31f8bfe4643ff283680c9954ca75a7f7e3d0864f81c9c01c34d175d412" Oct 14 07:21:43 crc kubenswrapper[4870]: I1014 07:21:43.655375 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"af919d31f8bfe4643ff283680c9954ca75a7f7e3d0864f81c9c01c34d175d412"} err="failed to get container status \"af919d31f8bfe4643ff283680c9954ca75a7f7e3d0864f81c9c01c34d175d412\": rpc error: code = NotFound desc = could not find container \"af919d31f8bfe4643ff283680c9954ca75a7f7e3d0864f81c9c01c34d175d412\": container with ID starting with af919d31f8bfe4643ff283680c9954ca75a7f7e3d0864f81c9c01c34d175d412 not found: ID does not exist" Oct 14 07:21:43 crc kubenswrapper[4870]: I1014 07:21:43.655397 4870 scope.go:117] "RemoveContainer" containerID="1ffa055509f25c5728647dcd15f7a11a8d373ff6501e1d0878f249cceb376ac0" Oct 14 07:21:43 crc kubenswrapper[4870]: I1014 07:21:43.655678 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1ffa055509f25c5728647dcd15f7a11a8d373ff6501e1d0878f249cceb376ac0"} err="failed to get container status \"1ffa055509f25c5728647dcd15f7a11a8d373ff6501e1d0878f249cceb376ac0\": rpc error: code = NotFound desc = could not find container \"1ffa055509f25c5728647dcd15f7a11a8d373ff6501e1d0878f249cceb376ac0\": container with ID starting with 1ffa055509f25c5728647dcd15f7a11a8d373ff6501e1d0878f249cceb376ac0 not found: ID does not exist" Oct 14 07:21:43 crc kubenswrapper[4870]: I1014 07:21:43.655695 4870 scope.go:117] "RemoveContainer" containerID="1d2d421419015d3d234317a1289d58133709959153d10dddb4ab2b71247221fe" Oct 14 07:21:43 crc kubenswrapper[4870]: I1014 07:21:43.655929 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1d2d421419015d3d234317a1289d58133709959153d10dddb4ab2b71247221fe"} err="failed to get container status \"1d2d421419015d3d234317a1289d58133709959153d10dddb4ab2b71247221fe\": rpc error: code = NotFound desc = could not find container \"1d2d421419015d3d234317a1289d58133709959153d10dddb4ab2b71247221fe\": container with ID starting with 1d2d421419015d3d234317a1289d58133709959153d10dddb4ab2b71247221fe not found: ID does not exist" Oct 14 07:21:43 crc kubenswrapper[4870]: I1014 07:21:43.655972 4870 scope.go:117] "RemoveContainer" containerID="fd6ab4f83f9cec1b430d4f0a7333b16f2dbbe88473f79bb5bab6c7bdc39b4fa9" Oct 14 07:21:43 crc kubenswrapper[4870]: I1014 07:21:43.656154 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fd6ab4f83f9cec1b430d4f0a7333b16f2dbbe88473f79bb5bab6c7bdc39b4fa9"} err="failed to get container status \"fd6ab4f83f9cec1b430d4f0a7333b16f2dbbe88473f79bb5bab6c7bdc39b4fa9\": rpc error: code = NotFound desc = could not find container \"fd6ab4f83f9cec1b430d4f0a7333b16f2dbbe88473f79bb5bab6c7bdc39b4fa9\": container with ID starting with fd6ab4f83f9cec1b430d4f0a7333b16f2dbbe88473f79bb5bab6c7bdc39b4fa9 not found: ID does not exist" Oct 14 07:21:43 crc kubenswrapper[4870]: I1014 07:21:43.687396 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ph2df\" (UniqueName: \"kubernetes.io/projected/2edb41c1-bdff-4251-91e3-8c5d61beccaf-kube-api-access-ph2df\") pod \"ceilometer-0\" (UID: \"2edb41c1-bdff-4251-91e3-8c5d61beccaf\") " pod="openstack/ceilometer-0" Oct 14 07:21:43 crc kubenswrapper[4870]: I1014 07:21:43.687471 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2edb41c1-bdff-4251-91e3-8c5d61beccaf-config-data\") pod \"ceilometer-0\" (UID: \"2edb41c1-bdff-4251-91e3-8c5d61beccaf\") " pod="openstack/ceilometer-0" Oct 14 07:21:43 crc kubenswrapper[4870]: I1014 07:21:43.687702 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2edb41c1-bdff-4251-91e3-8c5d61beccaf-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2edb41c1-bdff-4251-91e3-8c5d61beccaf\") " pod="openstack/ceilometer-0" Oct 14 07:21:43 crc kubenswrapper[4870]: I1014 07:21:43.687776 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2edb41c1-bdff-4251-91e3-8c5d61beccaf-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2edb41c1-bdff-4251-91e3-8c5d61beccaf\") " pod="openstack/ceilometer-0" Oct 14 07:21:43 crc kubenswrapper[4870]: I1014 07:21:43.687863 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2edb41c1-bdff-4251-91e3-8c5d61beccaf-scripts\") pod \"ceilometer-0\" (UID: \"2edb41c1-bdff-4251-91e3-8c5d61beccaf\") " pod="openstack/ceilometer-0" Oct 14 07:21:43 crc kubenswrapper[4870]: I1014 07:21:43.687924 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2edb41c1-bdff-4251-91e3-8c5d61beccaf-run-httpd\") pod \"ceilometer-0\" (UID: \"2edb41c1-bdff-4251-91e3-8c5d61beccaf\") " pod="openstack/ceilometer-0" Oct 14 07:21:43 crc kubenswrapper[4870]: I1014 07:21:43.687975 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2edb41c1-bdff-4251-91e3-8c5d61beccaf-log-httpd\") pod \"ceilometer-0\" (UID: \"2edb41c1-bdff-4251-91e3-8c5d61beccaf\") " pod="openstack/ceilometer-0" Oct 14 07:21:43 crc kubenswrapper[4870]: I1014 07:21:43.789641 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2edb41c1-bdff-4251-91e3-8c5d61beccaf-log-httpd\") pod \"ceilometer-0\" (UID: \"2edb41c1-bdff-4251-91e3-8c5d61beccaf\") " pod="openstack/ceilometer-0" Oct 14 07:21:43 crc kubenswrapper[4870]: I1014 07:21:43.789727 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ph2df\" (UniqueName: \"kubernetes.io/projected/2edb41c1-bdff-4251-91e3-8c5d61beccaf-kube-api-access-ph2df\") pod \"ceilometer-0\" (UID: \"2edb41c1-bdff-4251-91e3-8c5d61beccaf\") " pod="openstack/ceilometer-0" Oct 14 07:21:43 crc kubenswrapper[4870]: I1014 07:21:43.789757 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2edb41c1-bdff-4251-91e3-8c5d61beccaf-config-data\") pod \"ceilometer-0\" (UID: \"2edb41c1-bdff-4251-91e3-8c5d61beccaf\") " pod="openstack/ceilometer-0" Oct 14 07:21:43 crc kubenswrapper[4870]: I1014 07:21:43.789806 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2edb41c1-bdff-4251-91e3-8c5d61beccaf-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2edb41c1-bdff-4251-91e3-8c5d61beccaf\") " pod="openstack/ceilometer-0" Oct 14 07:21:43 crc kubenswrapper[4870]: I1014 07:21:43.789864 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2edb41c1-bdff-4251-91e3-8c5d61beccaf-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2edb41c1-bdff-4251-91e3-8c5d61beccaf\") " pod="openstack/ceilometer-0" Oct 14 07:21:43 crc kubenswrapper[4870]: I1014 07:21:43.789929 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2edb41c1-bdff-4251-91e3-8c5d61beccaf-scripts\") pod \"ceilometer-0\" (UID: \"2edb41c1-bdff-4251-91e3-8c5d61beccaf\") " pod="openstack/ceilometer-0" Oct 14 07:21:43 crc kubenswrapper[4870]: I1014 07:21:43.789965 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2edb41c1-bdff-4251-91e3-8c5d61beccaf-run-httpd\") pod \"ceilometer-0\" (UID: \"2edb41c1-bdff-4251-91e3-8c5d61beccaf\") " pod="openstack/ceilometer-0" Oct 14 07:21:43 crc kubenswrapper[4870]: I1014 07:21:43.790132 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2edb41c1-bdff-4251-91e3-8c5d61beccaf-log-httpd\") pod \"ceilometer-0\" (UID: \"2edb41c1-bdff-4251-91e3-8c5d61beccaf\") " pod="openstack/ceilometer-0" Oct 14 07:21:43 crc kubenswrapper[4870]: I1014 07:21:43.793613 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2edb41c1-bdff-4251-91e3-8c5d61beccaf-run-httpd\") pod \"ceilometer-0\" (UID: \"2edb41c1-bdff-4251-91e3-8c5d61beccaf\") " pod="openstack/ceilometer-0" Oct 14 07:21:43 crc kubenswrapper[4870]: I1014 07:21:43.795903 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2edb41c1-bdff-4251-91e3-8c5d61beccaf-scripts\") pod \"ceilometer-0\" (UID: \"2edb41c1-bdff-4251-91e3-8c5d61beccaf\") " pod="openstack/ceilometer-0" Oct 14 07:21:43 crc kubenswrapper[4870]: I1014 07:21:43.797085 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2edb41c1-bdff-4251-91e3-8c5d61beccaf-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2edb41c1-bdff-4251-91e3-8c5d61beccaf\") " pod="openstack/ceilometer-0" Oct 14 07:21:43 crc kubenswrapper[4870]: I1014 07:21:43.803564 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2edb41c1-bdff-4251-91e3-8c5d61beccaf-config-data\") pod \"ceilometer-0\" (UID: \"2edb41c1-bdff-4251-91e3-8c5d61beccaf\") " pod="openstack/ceilometer-0" Oct 14 07:21:43 crc kubenswrapper[4870]: I1014 07:21:43.804097 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2edb41c1-bdff-4251-91e3-8c5d61beccaf-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2edb41c1-bdff-4251-91e3-8c5d61beccaf\") " pod="openstack/ceilometer-0" Oct 14 07:21:43 crc kubenswrapper[4870]: I1014 07:21:43.809146 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ph2df\" (UniqueName: \"kubernetes.io/projected/2edb41c1-bdff-4251-91e3-8c5d61beccaf-kube-api-access-ph2df\") pod \"ceilometer-0\" (UID: \"2edb41c1-bdff-4251-91e3-8c5d61beccaf\") " pod="openstack/ceilometer-0" Oct 14 07:21:43 crc kubenswrapper[4870]: I1014 07:21:43.928400 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 14 07:21:44 crc kubenswrapper[4870]: I1014 07:21:44.414584 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 14 07:21:44 crc kubenswrapper[4870]: W1014 07:21:44.419428 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2edb41c1_bdff_4251_91e3_8c5d61beccaf.slice/crio-848497eeca968f5f9da4a37de9d21e29add6201e651810d078b383a6f2ef4676 WatchSource:0}: Error finding container 848497eeca968f5f9da4a37de9d21e29add6201e651810d078b383a6f2ef4676: Status 404 returned error can't find the container with id 848497eeca968f5f9da4a37de9d21e29add6201e651810d078b383a6f2ef4676 Oct 14 07:21:44 crc kubenswrapper[4870]: I1014 07:21:44.459818 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2edb41c1-bdff-4251-91e3-8c5d61beccaf","Type":"ContainerStarted","Data":"848497eeca968f5f9da4a37de9d21e29add6201e651810d078b383a6f2ef4676"} Oct 14 07:21:45 crc kubenswrapper[4870]: I1014 07:21:45.048953 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b24fea1-48c7-4100-9811-5739493548d9" path="/var/lib/kubelet/pods/5b24fea1-48c7-4100-9811-5739493548d9/volumes" Oct 14 07:21:45 crc kubenswrapper[4870]: I1014 07:21:45.319028 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 14 07:21:45 crc kubenswrapper[4870]: I1014 07:21:45.365459 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 14 07:21:45 crc kubenswrapper[4870]: I1014 07:21:45.447911 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 14 07:21:45 crc kubenswrapper[4870]: I1014 07:21:45.478793 4870 generic.go:334] "Generic (PLEG): container finished" podID="bdb3c945-edde-4d43-b3a5-234aaabf00d5" containerID="2841eadea64f22ddd21b4d710878d4871c5c5ceec4c6a266af64c8c08bfe682a" exitCode=0 Oct 14 07:21:45 crc kubenswrapper[4870]: I1014 07:21:45.478964 4870 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 14 07:21:45 crc kubenswrapper[4870]: I1014 07:21:45.482460 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-7jfpt" event={"ID":"bdb3c945-edde-4d43-b3a5-234aaabf00d5","Type":"ContainerDied","Data":"2841eadea64f22ddd21b4d710878d4871c5c5ceec4c6a266af64c8c08bfe682a"} Oct 14 07:21:45 crc kubenswrapper[4870]: I1014 07:21:45.491560 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 14 07:21:46 crc kubenswrapper[4870]: I1014 07:21:46.492802 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2edb41c1-bdff-4251-91e3-8c5d61beccaf","Type":"ContainerStarted","Data":"678b866a2d0c6b6a4671816ebea308062057589d8d7a174cab754b0cf7da1ccb"} Oct 14 07:21:46 crc kubenswrapper[4870]: I1014 07:21:46.937705 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-7jfpt" Oct 14 07:21:47 crc kubenswrapper[4870]: I1014 07:21:47.038982 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bdb3c945-edde-4d43-b3a5-234aaabf00d5-scripts\") pod \"bdb3c945-edde-4d43-b3a5-234aaabf00d5\" (UID: \"bdb3c945-edde-4d43-b3a5-234aaabf00d5\") " Oct 14 07:21:47 crc kubenswrapper[4870]: I1014 07:21:47.039056 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-trbf7\" (UniqueName: \"kubernetes.io/projected/bdb3c945-edde-4d43-b3a5-234aaabf00d5-kube-api-access-trbf7\") pod \"bdb3c945-edde-4d43-b3a5-234aaabf00d5\" (UID: \"bdb3c945-edde-4d43-b3a5-234aaabf00d5\") " Oct 14 07:21:47 crc kubenswrapper[4870]: I1014 07:21:47.039124 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bdb3c945-edde-4d43-b3a5-234aaabf00d5-combined-ca-bundle\") pod \"bdb3c945-edde-4d43-b3a5-234aaabf00d5\" (UID: \"bdb3c945-edde-4d43-b3a5-234aaabf00d5\") " Oct 14 07:21:47 crc kubenswrapper[4870]: I1014 07:21:47.039171 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bdb3c945-edde-4d43-b3a5-234aaabf00d5-config-data\") pod \"bdb3c945-edde-4d43-b3a5-234aaabf00d5\" (UID: \"bdb3c945-edde-4d43-b3a5-234aaabf00d5\") " Oct 14 07:21:47 crc kubenswrapper[4870]: I1014 07:21:47.045960 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bdb3c945-edde-4d43-b3a5-234aaabf00d5-kube-api-access-trbf7" (OuterVolumeSpecName: "kube-api-access-trbf7") pod "bdb3c945-edde-4d43-b3a5-234aaabf00d5" (UID: "bdb3c945-edde-4d43-b3a5-234aaabf00d5"). InnerVolumeSpecName "kube-api-access-trbf7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:21:47 crc kubenswrapper[4870]: I1014 07:21:47.050596 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bdb3c945-edde-4d43-b3a5-234aaabf00d5-scripts" (OuterVolumeSpecName: "scripts") pod "bdb3c945-edde-4d43-b3a5-234aaabf00d5" (UID: "bdb3c945-edde-4d43-b3a5-234aaabf00d5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:21:47 crc kubenswrapper[4870]: I1014 07:21:47.077138 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bdb3c945-edde-4d43-b3a5-234aaabf00d5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bdb3c945-edde-4d43-b3a5-234aaabf00d5" (UID: "bdb3c945-edde-4d43-b3a5-234aaabf00d5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:21:47 crc kubenswrapper[4870]: I1014 07:21:47.077569 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bdb3c945-edde-4d43-b3a5-234aaabf00d5-config-data" (OuterVolumeSpecName: "config-data") pod "bdb3c945-edde-4d43-b3a5-234aaabf00d5" (UID: "bdb3c945-edde-4d43-b3a5-234aaabf00d5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:21:47 crc kubenswrapper[4870]: I1014 07:21:47.141805 4870 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bdb3c945-edde-4d43-b3a5-234aaabf00d5-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 07:21:47 crc kubenswrapper[4870]: I1014 07:21:47.141850 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-trbf7\" (UniqueName: \"kubernetes.io/projected/bdb3c945-edde-4d43-b3a5-234aaabf00d5-kube-api-access-trbf7\") on node \"crc\" DevicePath \"\"" Oct 14 07:21:47 crc kubenswrapper[4870]: I1014 07:21:47.141866 4870 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bdb3c945-edde-4d43-b3a5-234aaabf00d5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 07:21:47 crc kubenswrapper[4870]: I1014 07:21:47.141874 4870 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bdb3c945-edde-4d43-b3a5-234aaabf00d5-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 07:21:47 crc kubenswrapper[4870]: I1014 07:21:47.519841 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2edb41c1-bdff-4251-91e3-8c5d61beccaf","Type":"ContainerStarted","Data":"ebd6c5d9bac96edcd5363084598bdcd47a36847351f24ebcb2ffa94c2ba6be41"} Oct 14 07:21:47 crc kubenswrapper[4870]: I1014 07:21:47.528146 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-7jfpt" Oct 14 07:21:47 crc kubenswrapper[4870]: I1014 07:21:47.530615 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-7jfpt" event={"ID":"bdb3c945-edde-4d43-b3a5-234aaabf00d5","Type":"ContainerDied","Data":"3b32a42b8d5e3833de5b1a277eafebef3875d31df1829e6d4edf1bee07df70a2"} Oct 14 07:21:47 crc kubenswrapper[4870]: I1014 07:21:47.530691 4870 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3b32a42b8d5e3833de5b1a277eafebef3875d31df1829e6d4edf1bee07df70a2" Oct 14 07:21:47 crc kubenswrapper[4870]: I1014 07:21:47.634361 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 14 07:21:47 crc kubenswrapper[4870]: E1014 07:21:47.634787 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bdb3c945-edde-4d43-b3a5-234aaabf00d5" containerName="nova-cell0-conductor-db-sync" Oct 14 07:21:47 crc kubenswrapper[4870]: I1014 07:21:47.634810 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="bdb3c945-edde-4d43-b3a5-234aaabf00d5" containerName="nova-cell0-conductor-db-sync" Oct 14 07:21:47 crc kubenswrapper[4870]: I1014 07:21:47.635050 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="bdb3c945-edde-4d43-b3a5-234aaabf00d5" containerName="nova-cell0-conductor-db-sync" Oct 14 07:21:47 crc kubenswrapper[4870]: I1014 07:21:47.635755 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 14 07:21:47 crc kubenswrapper[4870]: I1014 07:21:47.645900 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-bbjpq" Oct 14 07:21:47 crc kubenswrapper[4870]: I1014 07:21:47.646100 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Oct 14 07:21:47 crc kubenswrapper[4870]: I1014 07:21:47.662819 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 14 07:21:47 crc kubenswrapper[4870]: I1014 07:21:47.753961 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bad45b30-5ba2-4062-ab14-04f923679c4e-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"bad45b30-5ba2-4062-ab14-04f923679c4e\") " pod="openstack/nova-cell0-conductor-0" Oct 14 07:21:47 crc kubenswrapper[4870]: I1014 07:21:47.754128 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bad45b30-5ba2-4062-ab14-04f923679c4e-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"bad45b30-5ba2-4062-ab14-04f923679c4e\") " pod="openstack/nova-cell0-conductor-0" Oct 14 07:21:47 crc kubenswrapper[4870]: I1014 07:21:47.754200 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jflv9\" (UniqueName: \"kubernetes.io/projected/bad45b30-5ba2-4062-ab14-04f923679c4e-kube-api-access-jflv9\") pod \"nova-cell0-conductor-0\" (UID: \"bad45b30-5ba2-4062-ab14-04f923679c4e\") " pod="openstack/nova-cell0-conductor-0" Oct 14 07:21:47 crc kubenswrapper[4870]: I1014 07:21:47.856250 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bad45b30-5ba2-4062-ab14-04f923679c4e-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"bad45b30-5ba2-4062-ab14-04f923679c4e\") " pod="openstack/nova-cell0-conductor-0" Oct 14 07:21:47 crc kubenswrapper[4870]: I1014 07:21:47.856305 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bad45b30-5ba2-4062-ab14-04f923679c4e-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"bad45b30-5ba2-4062-ab14-04f923679c4e\") " pod="openstack/nova-cell0-conductor-0" Oct 14 07:21:47 crc kubenswrapper[4870]: I1014 07:21:47.856344 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jflv9\" (UniqueName: \"kubernetes.io/projected/bad45b30-5ba2-4062-ab14-04f923679c4e-kube-api-access-jflv9\") pod \"nova-cell0-conductor-0\" (UID: \"bad45b30-5ba2-4062-ab14-04f923679c4e\") " pod="openstack/nova-cell0-conductor-0" Oct 14 07:21:47 crc kubenswrapper[4870]: I1014 07:21:47.870150 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bad45b30-5ba2-4062-ab14-04f923679c4e-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"bad45b30-5ba2-4062-ab14-04f923679c4e\") " pod="openstack/nova-cell0-conductor-0" Oct 14 07:21:47 crc kubenswrapper[4870]: I1014 07:21:47.873725 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bad45b30-5ba2-4062-ab14-04f923679c4e-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"bad45b30-5ba2-4062-ab14-04f923679c4e\") " pod="openstack/nova-cell0-conductor-0" Oct 14 07:21:47 crc kubenswrapper[4870]: I1014 07:21:47.877745 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jflv9\" (UniqueName: \"kubernetes.io/projected/bad45b30-5ba2-4062-ab14-04f923679c4e-kube-api-access-jflv9\") pod \"nova-cell0-conductor-0\" (UID: \"bad45b30-5ba2-4062-ab14-04f923679c4e\") " pod="openstack/nova-cell0-conductor-0" Oct 14 07:21:47 crc kubenswrapper[4870]: I1014 07:21:47.963095 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 14 07:21:48 crc kubenswrapper[4870]: I1014 07:21:48.528548 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 14 07:21:48 crc kubenswrapper[4870]: W1014 07:21:48.528947 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbad45b30_5ba2_4062_ab14_04f923679c4e.slice/crio-27afa672a6627a62a0ae978b968b82988141797545411ad2982940d134c3309d WatchSource:0}: Error finding container 27afa672a6627a62a0ae978b968b82988141797545411ad2982940d134c3309d: Status 404 returned error can't find the container with id 27afa672a6627a62a0ae978b968b82988141797545411ad2982940d134c3309d Oct 14 07:21:48 crc kubenswrapper[4870]: I1014 07:21:48.551042 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2edb41c1-bdff-4251-91e3-8c5d61beccaf","Type":"ContainerStarted","Data":"4318a0f883a52b92c348661736366f87656f73b3037904058c4301e93f49bf25"} Oct 14 07:21:48 crc kubenswrapper[4870]: I1014 07:21:48.556164 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"bad45b30-5ba2-4062-ab14-04f923679c4e","Type":"ContainerStarted","Data":"27afa672a6627a62a0ae978b968b82988141797545411ad2982940d134c3309d"} Oct 14 07:21:49 crc kubenswrapper[4870]: I1014 07:21:49.572015 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"bad45b30-5ba2-4062-ab14-04f923679c4e","Type":"ContainerStarted","Data":"e6478418e9645b5d3cf9a2a8e1b1a4090b34dcf697735e42cf6fb41d3708bbeb"} Oct 14 07:21:49 crc kubenswrapper[4870]: I1014 07:21:49.572554 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Oct 14 07:21:49 crc kubenswrapper[4870]: I1014 07:21:49.599614 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.599587925 podStartE2EDuration="2.599587925s" podCreationTimestamp="2025-10-14 07:21:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:21:49.590961503 +0000 UTC m=+1245.288321874" watchObservedRunningTime="2025-10-14 07:21:49.599587925 +0000 UTC m=+1245.296948306" Oct 14 07:21:50 crc kubenswrapper[4870]: I1014 07:21:50.584277 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2edb41c1-bdff-4251-91e3-8c5d61beccaf","Type":"ContainerStarted","Data":"130e23ed1e9f1ea0d4e7922866a5deec8f115d0bcc35cbfb3ca734a72f6e3a13"} Oct 14 07:21:50 crc kubenswrapper[4870]: I1014 07:21:50.584592 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 14 07:21:50 crc kubenswrapper[4870]: I1014 07:21:50.615068 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.867155044 podStartE2EDuration="7.615041911s" podCreationTimestamp="2025-10-14 07:21:43 +0000 UTC" firstStartedPulling="2025-10-14 07:21:44.421717671 +0000 UTC m=+1240.119078042" lastFinishedPulling="2025-10-14 07:21:50.169604538 +0000 UTC m=+1245.866964909" observedRunningTime="2025-10-14 07:21:50.610961825 +0000 UTC m=+1246.308322246" watchObservedRunningTime="2025-10-14 07:21:50.615041911 +0000 UTC m=+1246.312402312" Oct 14 07:21:53 crc kubenswrapper[4870]: I1014 07:21:53.951609 4870 patch_prober.go:28] interesting pod/machine-config-daemon-7tvc7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 07:21:53 crc kubenswrapper[4870]: I1014 07:21:53.951972 4870 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 07:21:57 crc kubenswrapper[4870]: I1014 07:21:57.998425 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Oct 14 07:21:58 crc kubenswrapper[4870]: I1014 07:21:58.503034 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-q76l4"] Oct 14 07:21:58 crc kubenswrapper[4870]: I1014 07:21:58.504330 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-q76l4" Oct 14 07:21:58 crc kubenswrapper[4870]: I1014 07:21:58.507017 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Oct 14 07:21:58 crc kubenswrapper[4870]: I1014 07:21:58.508112 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Oct 14 07:21:58 crc kubenswrapper[4870]: I1014 07:21:58.515673 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-q76l4"] Oct 14 07:21:58 crc kubenswrapper[4870]: I1014 07:21:58.646311 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 14 07:21:58 crc kubenswrapper[4870]: I1014 07:21:58.647922 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 14 07:21:58 crc kubenswrapper[4870]: I1014 07:21:58.650026 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 14 07:21:58 crc kubenswrapper[4870]: I1014 07:21:58.683632 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d147118e-e08e-4c64-b157-da3ca3afb6d7-scripts\") pod \"nova-cell0-cell-mapping-q76l4\" (UID: \"d147118e-e08e-4c64-b157-da3ca3afb6d7\") " pod="openstack/nova-cell0-cell-mapping-q76l4" Oct 14 07:21:58 crc kubenswrapper[4870]: I1014 07:21:58.683749 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d147118e-e08e-4c64-b157-da3ca3afb6d7-config-data\") pod \"nova-cell0-cell-mapping-q76l4\" (UID: \"d147118e-e08e-4c64-b157-da3ca3afb6d7\") " pod="openstack/nova-cell0-cell-mapping-q76l4" Oct 14 07:21:58 crc kubenswrapper[4870]: I1014 07:21:58.685650 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d147118e-e08e-4c64-b157-da3ca3afb6d7-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-q76l4\" (UID: \"d147118e-e08e-4c64-b157-da3ca3afb6d7\") " pod="openstack/nova-cell0-cell-mapping-q76l4" Oct 14 07:21:58 crc kubenswrapper[4870]: I1014 07:21:58.685734 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zjn89\" (UniqueName: \"kubernetes.io/projected/d147118e-e08e-4c64-b157-da3ca3afb6d7-kube-api-access-zjn89\") pod \"nova-cell0-cell-mapping-q76l4\" (UID: \"d147118e-e08e-4c64-b157-da3ca3afb6d7\") " pod="openstack/nova-cell0-cell-mapping-q76l4" Oct 14 07:21:58 crc kubenswrapper[4870]: I1014 07:21:58.695768 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 14 07:21:58 crc kubenswrapper[4870]: I1014 07:21:58.748398 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 14 07:21:58 crc kubenswrapper[4870]: I1014 07:21:58.749534 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 14 07:21:58 crc kubenswrapper[4870]: I1014 07:21:58.790019 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d147118e-e08e-4c64-b157-da3ca3afb6d7-config-data\") pod \"nova-cell0-cell-mapping-q76l4\" (UID: \"d147118e-e08e-4c64-b157-da3ca3afb6d7\") " pod="openstack/nova-cell0-cell-mapping-q76l4" Oct 14 07:21:58 crc kubenswrapper[4870]: I1014 07:21:58.790171 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd44bae6-b5d1-4b08-88de-7c5271f54ed3-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"cd44bae6-b5d1-4b08-88de-7c5271f54ed3\") " pod="openstack/nova-api-0" Oct 14 07:21:58 crc kubenswrapper[4870]: I1014 07:21:58.790320 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d147118e-e08e-4c64-b157-da3ca3afb6d7-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-q76l4\" (UID: \"d147118e-e08e-4c64-b157-da3ca3afb6d7\") " pod="openstack/nova-cell0-cell-mapping-q76l4" Oct 14 07:21:58 crc kubenswrapper[4870]: I1014 07:21:58.790344 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd44bae6-b5d1-4b08-88de-7c5271f54ed3-config-data\") pod \"nova-api-0\" (UID: \"cd44bae6-b5d1-4b08-88de-7c5271f54ed3\") " pod="openstack/nova-api-0" Oct 14 07:21:58 crc kubenswrapper[4870]: I1014 07:21:58.790365 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cd44bae6-b5d1-4b08-88de-7c5271f54ed3-logs\") pod \"nova-api-0\" (UID: \"cd44bae6-b5d1-4b08-88de-7c5271f54ed3\") " pod="openstack/nova-api-0" Oct 14 07:21:58 crc kubenswrapper[4870]: I1014 07:21:58.805353 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d147118e-e08e-4c64-b157-da3ca3afb6d7-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-q76l4\" (UID: \"d147118e-e08e-4c64-b157-da3ca3afb6d7\") " pod="openstack/nova-cell0-cell-mapping-q76l4" Oct 14 07:21:58 crc kubenswrapper[4870]: I1014 07:21:58.805781 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 14 07:21:58 crc kubenswrapper[4870]: I1014 07:21:58.790451 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zjn89\" (UniqueName: \"kubernetes.io/projected/d147118e-e08e-4c64-b157-da3ca3afb6d7-kube-api-access-zjn89\") pod \"nova-cell0-cell-mapping-q76l4\" (UID: \"d147118e-e08e-4c64-b157-da3ca3afb6d7\") " pod="openstack/nova-cell0-cell-mapping-q76l4" Oct 14 07:21:58 crc kubenswrapper[4870]: I1014 07:21:58.807983 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b9xkb\" (UniqueName: \"kubernetes.io/projected/cd44bae6-b5d1-4b08-88de-7c5271f54ed3-kube-api-access-b9xkb\") pod \"nova-api-0\" (UID: \"cd44bae6-b5d1-4b08-88de-7c5271f54ed3\") " pod="openstack/nova-api-0" Oct 14 07:21:58 crc kubenswrapper[4870]: I1014 07:21:58.808106 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d147118e-e08e-4c64-b157-da3ca3afb6d7-scripts\") pod \"nova-cell0-cell-mapping-q76l4\" (UID: \"d147118e-e08e-4c64-b157-da3ca3afb6d7\") " pod="openstack/nova-cell0-cell-mapping-q76l4" Oct 14 07:21:58 crc kubenswrapper[4870]: I1014 07:21:58.808459 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d147118e-e08e-4c64-b157-da3ca3afb6d7-config-data\") pod \"nova-cell0-cell-mapping-q76l4\" (UID: \"d147118e-e08e-4c64-b157-da3ca3afb6d7\") " pod="openstack/nova-cell0-cell-mapping-q76l4" Oct 14 07:21:58 crc kubenswrapper[4870]: I1014 07:21:58.821996 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d147118e-e08e-4c64-b157-da3ca3afb6d7-scripts\") pod \"nova-cell0-cell-mapping-q76l4\" (UID: \"d147118e-e08e-4c64-b157-da3ca3afb6d7\") " pod="openstack/nova-cell0-cell-mapping-q76l4" Oct 14 07:21:58 crc kubenswrapper[4870]: I1014 07:21:58.841896 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 14 07:21:58 crc kubenswrapper[4870]: I1014 07:21:58.874858 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zjn89\" (UniqueName: \"kubernetes.io/projected/d147118e-e08e-4c64-b157-da3ca3afb6d7-kube-api-access-zjn89\") pod \"nova-cell0-cell-mapping-q76l4\" (UID: \"d147118e-e08e-4c64-b157-da3ca3afb6d7\") " pod="openstack/nova-cell0-cell-mapping-q76l4" Oct 14 07:21:58 crc kubenswrapper[4870]: I1014 07:21:58.890093 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 14 07:21:58 crc kubenswrapper[4870]: I1014 07:21:58.906419 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 14 07:21:58 crc kubenswrapper[4870]: I1014 07:21:58.915192 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd44bae6-b5d1-4b08-88de-7c5271f54ed3-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"cd44bae6-b5d1-4b08-88de-7c5271f54ed3\") " pod="openstack/nova-api-0" Oct 14 07:21:58 crc kubenswrapper[4870]: I1014 07:21:58.915292 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tqmf6\" (UniqueName: \"kubernetes.io/projected/67e09a03-b028-4444-8aba-63be9c7e1abd-kube-api-access-tqmf6\") pod \"nova-scheduler-0\" (UID: \"67e09a03-b028-4444-8aba-63be9c7e1abd\") " pod="openstack/nova-scheduler-0" Oct 14 07:21:58 crc kubenswrapper[4870]: I1014 07:21:58.915339 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67e09a03-b028-4444-8aba-63be9c7e1abd-config-data\") pod \"nova-scheduler-0\" (UID: \"67e09a03-b028-4444-8aba-63be9c7e1abd\") " pod="openstack/nova-scheduler-0" Oct 14 07:21:58 crc kubenswrapper[4870]: I1014 07:21:58.915372 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd44bae6-b5d1-4b08-88de-7c5271f54ed3-config-data\") pod \"nova-api-0\" (UID: \"cd44bae6-b5d1-4b08-88de-7c5271f54ed3\") " pod="openstack/nova-api-0" Oct 14 07:21:58 crc kubenswrapper[4870]: I1014 07:21:58.915396 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cd44bae6-b5d1-4b08-88de-7c5271f54ed3-logs\") pod \"nova-api-0\" (UID: \"cd44bae6-b5d1-4b08-88de-7c5271f54ed3\") " pod="openstack/nova-api-0" Oct 14 07:21:58 crc kubenswrapper[4870]: I1014 07:21:58.915468 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b9xkb\" (UniqueName: \"kubernetes.io/projected/cd44bae6-b5d1-4b08-88de-7c5271f54ed3-kube-api-access-b9xkb\") pod \"nova-api-0\" (UID: \"cd44bae6-b5d1-4b08-88de-7c5271f54ed3\") " pod="openstack/nova-api-0" Oct 14 07:21:58 crc kubenswrapper[4870]: I1014 07:21:58.915527 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67e09a03-b028-4444-8aba-63be9c7e1abd-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"67e09a03-b028-4444-8aba-63be9c7e1abd\") " pod="openstack/nova-scheduler-0" Oct 14 07:21:58 crc kubenswrapper[4870]: I1014 07:21:58.915869 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 14 07:21:58 crc kubenswrapper[4870]: I1014 07:21:58.918894 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cd44bae6-b5d1-4b08-88de-7c5271f54ed3-logs\") pod \"nova-api-0\" (UID: \"cd44bae6-b5d1-4b08-88de-7c5271f54ed3\") " pod="openstack/nova-api-0" Oct 14 07:21:58 crc kubenswrapper[4870]: I1014 07:21:58.925136 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd44bae6-b5d1-4b08-88de-7c5271f54ed3-config-data\") pod \"nova-api-0\" (UID: \"cd44bae6-b5d1-4b08-88de-7c5271f54ed3\") " pod="openstack/nova-api-0" Oct 14 07:21:58 crc kubenswrapper[4870]: I1014 07:21:58.943960 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 14 07:21:58 crc kubenswrapper[4870]: I1014 07:21:58.955204 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd44bae6-b5d1-4b08-88de-7c5271f54ed3-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"cd44bae6-b5d1-4b08-88de-7c5271f54ed3\") " pod="openstack/nova-api-0" Oct 14 07:21:58 crc kubenswrapper[4870]: I1014 07:21:58.965763 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 14 07:21:58 crc kubenswrapper[4870]: I1014 07:21:58.976810 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 14 07:21:59 crc kubenswrapper[4870]: I1014 07:21:59.007851 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Oct 14 07:21:59 crc kubenswrapper[4870]: I1014 07:21:59.021542 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 14 07:21:59 crc kubenswrapper[4870]: I1014 07:21:59.022612 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67e09a03-b028-4444-8aba-63be9c7e1abd-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"67e09a03-b028-4444-8aba-63be9c7e1abd\") " pod="openstack/nova-scheduler-0" Oct 14 07:21:59 crc kubenswrapper[4870]: I1014 07:21:59.022665 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ncz4r\" (UniqueName: \"kubernetes.io/projected/ed189910-4793-4a5f-9688-d643f5edaf2c-kube-api-access-ncz4r\") pod \"nova-metadata-0\" (UID: \"ed189910-4793-4a5f-9688-d643f5edaf2c\") " pod="openstack/nova-metadata-0" Oct 14 07:21:59 crc kubenswrapper[4870]: I1014 07:21:59.022736 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ed189910-4793-4a5f-9688-d643f5edaf2c-config-data\") pod \"nova-metadata-0\" (UID: \"ed189910-4793-4a5f-9688-d643f5edaf2c\") " pod="openstack/nova-metadata-0" Oct 14 07:21:59 crc kubenswrapper[4870]: I1014 07:21:59.022756 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed189910-4793-4a5f-9688-d643f5edaf2c-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"ed189910-4793-4a5f-9688-d643f5edaf2c\") " pod="openstack/nova-metadata-0" Oct 14 07:21:59 crc kubenswrapper[4870]: I1014 07:21:59.022802 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tqmf6\" (UniqueName: \"kubernetes.io/projected/67e09a03-b028-4444-8aba-63be9c7e1abd-kube-api-access-tqmf6\") pod \"nova-scheduler-0\" (UID: \"67e09a03-b028-4444-8aba-63be9c7e1abd\") " pod="openstack/nova-scheduler-0" Oct 14 07:21:59 crc kubenswrapper[4870]: I1014 07:21:59.022821 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ed189910-4793-4a5f-9688-d643f5edaf2c-logs\") pod \"nova-metadata-0\" (UID: \"ed189910-4793-4a5f-9688-d643f5edaf2c\") " pod="openstack/nova-metadata-0" Oct 14 07:21:59 crc kubenswrapper[4870]: I1014 07:21:59.022848 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67e09a03-b028-4444-8aba-63be9c7e1abd-config-data\") pod \"nova-scheduler-0\" (UID: \"67e09a03-b028-4444-8aba-63be9c7e1abd\") " pod="openstack/nova-scheduler-0" Oct 14 07:21:59 crc kubenswrapper[4870]: I1014 07:21:59.024377 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b9xkb\" (UniqueName: \"kubernetes.io/projected/cd44bae6-b5d1-4b08-88de-7c5271f54ed3-kube-api-access-b9xkb\") pod \"nova-api-0\" (UID: \"cd44bae6-b5d1-4b08-88de-7c5271f54ed3\") " pod="openstack/nova-api-0" Oct 14 07:21:59 crc kubenswrapper[4870]: I1014 07:21:59.043055 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67e09a03-b028-4444-8aba-63be9c7e1abd-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"67e09a03-b028-4444-8aba-63be9c7e1abd\") " pod="openstack/nova-scheduler-0" Oct 14 07:21:59 crc kubenswrapper[4870]: I1014 07:21:59.069011 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67e09a03-b028-4444-8aba-63be9c7e1abd-config-data\") pod \"nova-scheduler-0\" (UID: \"67e09a03-b028-4444-8aba-63be9c7e1abd\") " pod="openstack/nova-scheduler-0" Oct 14 07:21:59 crc kubenswrapper[4870]: I1014 07:21:59.077894 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tqmf6\" (UniqueName: \"kubernetes.io/projected/67e09a03-b028-4444-8aba-63be9c7e1abd-kube-api-access-tqmf6\") pod \"nova-scheduler-0\" (UID: \"67e09a03-b028-4444-8aba-63be9c7e1abd\") " pod="openstack/nova-scheduler-0" Oct 14 07:21:59 crc kubenswrapper[4870]: I1014 07:21:59.124455 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ed189910-4793-4a5f-9688-d643f5edaf2c-logs\") pod \"nova-metadata-0\" (UID: \"ed189910-4793-4a5f-9688-d643f5edaf2c\") " pod="openstack/nova-metadata-0" Oct 14 07:21:59 crc kubenswrapper[4870]: I1014 07:21:59.124510 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hflfq\" (UniqueName: \"kubernetes.io/projected/ed7c5b6f-eb0a-44fc-a6a3-6f43909b6406-kube-api-access-hflfq\") pod \"nova-cell1-novncproxy-0\" (UID: \"ed7c5b6f-eb0a-44fc-a6a3-6f43909b6406\") " pod="openstack/nova-cell1-novncproxy-0" Oct 14 07:21:59 crc kubenswrapper[4870]: I1014 07:21:59.124541 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed7c5b6f-eb0a-44fc-a6a3-6f43909b6406-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"ed7c5b6f-eb0a-44fc-a6a3-6f43909b6406\") " pod="openstack/nova-cell1-novncproxy-0" Oct 14 07:21:59 crc kubenswrapper[4870]: I1014 07:21:59.124657 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ncz4r\" (UniqueName: \"kubernetes.io/projected/ed189910-4793-4a5f-9688-d643f5edaf2c-kube-api-access-ncz4r\") pod \"nova-metadata-0\" (UID: \"ed189910-4793-4a5f-9688-d643f5edaf2c\") " pod="openstack/nova-metadata-0" Oct 14 07:21:59 crc kubenswrapper[4870]: I1014 07:21:59.124706 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ed189910-4793-4a5f-9688-d643f5edaf2c-config-data\") pod \"nova-metadata-0\" (UID: \"ed189910-4793-4a5f-9688-d643f5edaf2c\") " pod="openstack/nova-metadata-0" Oct 14 07:21:59 crc kubenswrapper[4870]: I1014 07:21:59.124730 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed189910-4793-4a5f-9688-d643f5edaf2c-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"ed189910-4793-4a5f-9688-d643f5edaf2c\") " pod="openstack/nova-metadata-0" Oct 14 07:21:59 crc kubenswrapper[4870]: I1014 07:21:59.124796 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ed7c5b6f-eb0a-44fc-a6a3-6f43909b6406-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"ed7c5b6f-eb0a-44fc-a6a3-6f43909b6406\") " pod="openstack/nova-cell1-novncproxy-0" Oct 14 07:21:59 crc kubenswrapper[4870]: I1014 07:21:59.125366 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ed189910-4793-4a5f-9688-d643f5edaf2c-logs\") pod \"nova-metadata-0\" (UID: \"ed189910-4793-4a5f-9688-d643f5edaf2c\") " pod="openstack/nova-metadata-0" Oct 14 07:21:59 crc kubenswrapper[4870]: I1014 07:21:59.136559 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6957666cb7-wgpsx"] Oct 14 07:21:59 crc kubenswrapper[4870]: I1014 07:21:59.138255 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6957666cb7-wgpsx" Oct 14 07:21:59 crc kubenswrapper[4870]: I1014 07:21:59.140726 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-q76l4" Oct 14 07:21:59 crc kubenswrapper[4870]: I1014 07:21:59.149860 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6957666cb7-wgpsx"] Oct 14 07:21:59 crc kubenswrapper[4870]: I1014 07:21:59.155902 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed189910-4793-4a5f-9688-d643f5edaf2c-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"ed189910-4793-4a5f-9688-d643f5edaf2c\") " pod="openstack/nova-metadata-0" Oct 14 07:21:59 crc kubenswrapper[4870]: I1014 07:21:59.156637 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ed189910-4793-4a5f-9688-d643f5edaf2c-config-data\") pod \"nova-metadata-0\" (UID: \"ed189910-4793-4a5f-9688-d643f5edaf2c\") " pod="openstack/nova-metadata-0" Oct 14 07:21:59 crc kubenswrapper[4870]: I1014 07:21:59.176319 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ncz4r\" (UniqueName: \"kubernetes.io/projected/ed189910-4793-4a5f-9688-d643f5edaf2c-kube-api-access-ncz4r\") pod \"nova-metadata-0\" (UID: \"ed189910-4793-4a5f-9688-d643f5edaf2c\") " pod="openstack/nova-metadata-0" Oct 14 07:21:59 crc kubenswrapper[4870]: I1014 07:21:59.227339 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed7c5b6f-eb0a-44fc-a6a3-6f43909b6406-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"ed7c5b6f-eb0a-44fc-a6a3-6f43909b6406\") " pod="openstack/nova-cell1-novncproxy-0" Oct 14 07:21:59 crc kubenswrapper[4870]: I1014 07:21:59.227407 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bqmbb\" (UniqueName: \"kubernetes.io/projected/d5d3e78a-8ed8-40a5-8a64-734c7836155c-kube-api-access-bqmbb\") pod \"dnsmasq-dns-6957666cb7-wgpsx\" (UID: \"d5d3e78a-8ed8-40a5-8a64-734c7836155c\") " pod="openstack/dnsmasq-dns-6957666cb7-wgpsx" Oct 14 07:21:59 crc kubenswrapper[4870]: I1014 07:21:59.227487 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d5d3e78a-8ed8-40a5-8a64-734c7836155c-ovsdbserver-sb\") pod \"dnsmasq-dns-6957666cb7-wgpsx\" (UID: \"d5d3e78a-8ed8-40a5-8a64-734c7836155c\") " pod="openstack/dnsmasq-dns-6957666cb7-wgpsx" Oct 14 07:21:59 crc kubenswrapper[4870]: I1014 07:21:59.227524 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d5d3e78a-8ed8-40a5-8a64-734c7836155c-dns-swift-storage-0\") pod \"dnsmasq-dns-6957666cb7-wgpsx\" (UID: \"d5d3e78a-8ed8-40a5-8a64-734c7836155c\") " pod="openstack/dnsmasq-dns-6957666cb7-wgpsx" Oct 14 07:21:59 crc kubenswrapper[4870]: I1014 07:21:59.227576 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d5d3e78a-8ed8-40a5-8a64-734c7836155c-ovsdbserver-nb\") pod \"dnsmasq-dns-6957666cb7-wgpsx\" (UID: \"d5d3e78a-8ed8-40a5-8a64-734c7836155c\") " pod="openstack/dnsmasq-dns-6957666cb7-wgpsx" Oct 14 07:21:59 crc kubenswrapper[4870]: I1014 07:21:59.227614 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d5d3e78a-8ed8-40a5-8a64-734c7836155c-dns-svc\") pod \"dnsmasq-dns-6957666cb7-wgpsx\" (UID: \"d5d3e78a-8ed8-40a5-8a64-734c7836155c\") " pod="openstack/dnsmasq-dns-6957666cb7-wgpsx" Oct 14 07:21:59 crc kubenswrapper[4870]: I1014 07:21:59.227651 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d5d3e78a-8ed8-40a5-8a64-734c7836155c-config\") pod \"dnsmasq-dns-6957666cb7-wgpsx\" (UID: \"d5d3e78a-8ed8-40a5-8a64-734c7836155c\") " pod="openstack/dnsmasq-dns-6957666cb7-wgpsx" Oct 14 07:21:59 crc kubenswrapper[4870]: I1014 07:21:59.227670 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ed7c5b6f-eb0a-44fc-a6a3-6f43909b6406-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"ed7c5b6f-eb0a-44fc-a6a3-6f43909b6406\") " pod="openstack/nova-cell1-novncproxy-0" Oct 14 07:21:59 crc kubenswrapper[4870]: I1014 07:21:59.227708 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hflfq\" (UniqueName: \"kubernetes.io/projected/ed7c5b6f-eb0a-44fc-a6a3-6f43909b6406-kube-api-access-hflfq\") pod \"nova-cell1-novncproxy-0\" (UID: \"ed7c5b6f-eb0a-44fc-a6a3-6f43909b6406\") " pod="openstack/nova-cell1-novncproxy-0" Oct 14 07:21:59 crc kubenswrapper[4870]: I1014 07:21:59.232027 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed7c5b6f-eb0a-44fc-a6a3-6f43909b6406-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"ed7c5b6f-eb0a-44fc-a6a3-6f43909b6406\") " pod="openstack/nova-cell1-novncproxy-0" Oct 14 07:21:59 crc kubenswrapper[4870]: I1014 07:21:59.233095 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ed7c5b6f-eb0a-44fc-a6a3-6f43909b6406-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"ed7c5b6f-eb0a-44fc-a6a3-6f43909b6406\") " pod="openstack/nova-cell1-novncproxy-0" Oct 14 07:21:59 crc kubenswrapper[4870]: I1014 07:21:59.245294 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hflfq\" (UniqueName: \"kubernetes.io/projected/ed7c5b6f-eb0a-44fc-a6a3-6f43909b6406-kube-api-access-hflfq\") pod \"nova-cell1-novncproxy-0\" (UID: \"ed7c5b6f-eb0a-44fc-a6a3-6f43909b6406\") " pod="openstack/nova-cell1-novncproxy-0" Oct 14 07:21:59 crc kubenswrapper[4870]: I1014 07:21:59.283322 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 14 07:21:59 crc kubenswrapper[4870]: I1014 07:21:59.328874 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d5d3e78a-8ed8-40a5-8a64-734c7836155c-dns-swift-storage-0\") pod \"dnsmasq-dns-6957666cb7-wgpsx\" (UID: \"d5d3e78a-8ed8-40a5-8a64-734c7836155c\") " pod="openstack/dnsmasq-dns-6957666cb7-wgpsx" Oct 14 07:21:59 crc kubenswrapper[4870]: I1014 07:21:59.328942 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d5d3e78a-8ed8-40a5-8a64-734c7836155c-ovsdbserver-nb\") pod \"dnsmasq-dns-6957666cb7-wgpsx\" (UID: \"d5d3e78a-8ed8-40a5-8a64-734c7836155c\") " pod="openstack/dnsmasq-dns-6957666cb7-wgpsx" Oct 14 07:21:59 crc kubenswrapper[4870]: I1014 07:21:59.328979 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d5d3e78a-8ed8-40a5-8a64-734c7836155c-dns-svc\") pod \"dnsmasq-dns-6957666cb7-wgpsx\" (UID: \"d5d3e78a-8ed8-40a5-8a64-734c7836155c\") " pod="openstack/dnsmasq-dns-6957666cb7-wgpsx" Oct 14 07:21:59 crc kubenswrapper[4870]: I1014 07:21:59.329014 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d5d3e78a-8ed8-40a5-8a64-734c7836155c-config\") pod \"dnsmasq-dns-6957666cb7-wgpsx\" (UID: \"d5d3e78a-8ed8-40a5-8a64-734c7836155c\") " pod="openstack/dnsmasq-dns-6957666cb7-wgpsx" Oct 14 07:21:59 crc kubenswrapper[4870]: I1014 07:21:59.329089 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bqmbb\" (UniqueName: \"kubernetes.io/projected/d5d3e78a-8ed8-40a5-8a64-734c7836155c-kube-api-access-bqmbb\") pod \"dnsmasq-dns-6957666cb7-wgpsx\" (UID: \"d5d3e78a-8ed8-40a5-8a64-734c7836155c\") " pod="openstack/dnsmasq-dns-6957666cb7-wgpsx" Oct 14 07:21:59 crc kubenswrapper[4870]: I1014 07:21:59.329110 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d5d3e78a-8ed8-40a5-8a64-734c7836155c-ovsdbserver-sb\") pod \"dnsmasq-dns-6957666cb7-wgpsx\" (UID: \"d5d3e78a-8ed8-40a5-8a64-734c7836155c\") " pod="openstack/dnsmasq-dns-6957666cb7-wgpsx" Oct 14 07:21:59 crc kubenswrapper[4870]: I1014 07:21:59.329979 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d5d3e78a-8ed8-40a5-8a64-734c7836155c-ovsdbserver-sb\") pod \"dnsmasq-dns-6957666cb7-wgpsx\" (UID: \"d5d3e78a-8ed8-40a5-8a64-734c7836155c\") " pod="openstack/dnsmasq-dns-6957666cb7-wgpsx" Oct 14 07:21:59 crc kubenswrapper[4870]: I1014 07:21:59.330662 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d5d3e78a-8ed8-40a5-8a64-734c7836155c-dns-swift-storage-0\") pod \"dnsmasq-dns-6957666cb7-wgpsx\" (UID: \"d5d3e78a-8ed8-40a5-8a64-734c7836155c\") " pod="openstack/dnsmasq-dns-6957666cb7-wgpsx" Oct 14 07:21:59 crc kubenswrapper[4870]: I1014 07:21:59.331359 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d5d3e78a-8ed8-40a5-8a64-734c7836155c-ovsdbserver-nb\") pod \"dnsmasq-dns-6957666cb7-wgpsx\" (UID: \"d5d3e78a-8ed8-40a5-8a64-734c7836155c\") " pod="openstack/dnsmasq-dns-6957666cb7-wgpsx" Oct 14 07:21:59 crc kubenswrapper[4870]: I1014 07:21:59.332074 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d5d3e78a-8ed8-40a5-8a64-734c7836155c-dns-svc\") pod \"dnsmasq-dns-6957666cb7-wgpsx\" (UID: \"d5d3e78a-8ed8-40a5-8a64-734c7836155c\") " pod="openstack/dnsmasq-dns-6957666cb7-wgpsx" Oct 14 07:21:59 crc kubenswrapper[4870]: I1014 07:21:59.332653 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d5d3e78a-8ed8-40a5-8a64-734c7836155c-config\") pod \"dnsmasq-dns-6957666cb7-wgpsx\" (UID: \"d5d3e78a-8ed8-40a5-8a64-734c7836155c\") " pod="openstack/dnsmasq-dns-6957666cb7-wgpsx" Oct 14 07:21:59 crc kubenswrapper[4870]: I1014 07:21:59.359546 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bqmbb\" (UniqueName: \"kubernetes.io/projected/d5d3e78a-8ed8-40a5-8a64-734c7836155c-kube-api-access-bqmbb\") pod \"dnsmasq-dns-6957666cb7-wgpsx\" (UID: \"d5d3e78a-8ed8-40a5-8a64-734c7836155c\") " pod="openstack/dnsmasq-dns-6957666cb7-wgpsx" Oct 14 07:21:59 crc kubenswrapper[4870]: I1014 07:21:59.369655 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 14 07:21:59 crc kubenswrapper[4870]: I1014 07:21:59.448104 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 14 07:21:59 crc kubenswrapper[4870]: I1014 07:21:59.482658 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 14 07:21:59 crc kubenswrapper[4870]: I1014 07:21:59.515781 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6957666cb7-wgpsx" Oct 14 07:21:59 crc kubenswrapper[4870]: I1014 07:21:59.754582 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-q76l4"] Oct 14 07:21:59 crc kubenswrapper[4870]: I1014 07:21:59.825206 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-x2gj8"] Oct 14 07:21:59 crc kubenswrapper[4870]: I1014 07:21:59.832361 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-x2gj8" Oct 14 07:21:59 crc kubenswrapper[4870]: I1014 07:21:59.841472 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Oct 14 07:21:59 crc kubenswrapper[4870]: I1014 07:21:59.841707 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Oct 14 07:21:59 crc kubenswrapper[4870]: I1014 07:21:59.842301 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-x2gj8"] Oct 14 07:21:59 crc kubenswrapper[4870]: I1014 07:21:59.850307 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 14 07:21:59 crc kubenswrapper[4870]: I1014 07:21:59.941954 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 14 07:21:59 crc kubenswrapper[4870]: I1014 07:21:59.945468 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b265c7a-3909-4fd6-9981-c1723fbc1e02-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-x2gj8\" (UID: \"3b265c7a-3909-4fd6-9981-c1723fbc1e02\") " pod="openstack/nova-cell1-conductor-db-sync-x2gj8" Oct 14 07:21:59 crc kubenswrapper[4870]: I1014 07:21:59.945529 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3b265c7a-3909-4fd6-9981-c1723fbc1e02-scripts\") pod \"nova-cell1-conductor-db-sync-x2gj8\" (UID: \"3b265c7a-3909-4fd6-9981-c1723fbc1e02\") " pod="openstack/nova-cell1-conductor-db-sync-x2gj8" Oct 14 07:21:59 crc kubenswrapper[4870]: I1014 07:21:59.945611 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3b265c7a-3909-4fd6-9981-c1723fbc1e02-config-data\") pod \"nova-cell1-conductor-db-sync-x2gj8\" (UID: \"3b265c7a-3909-4fd6-9981-c1723fbc1e02\") " pod="openstack/nova-cell1-conductor-db-sync-x2gj8" Oct 14 07:21:59 crc kubenswrapper[4870]: I1014 07:21:59.945635 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t6pqs\" (UniqueName: \"kubernetes.io/projected/3b265c7a-3909-4fd6-9981-c1723fbc1e02-kube-api-access-t6pqs\") pod \"nova-cell1-conductor-db-sync-x2gj8\" (UID: \"3b265c7a-3909-4fd6-9981-c1723fbc1e02\") " pod="openstack/nova-cell1-conductor-db-sync-x2gj8" Oct 14 07:22:00 crc kubenswrapper[4870]: I1014 07:22:00.046880 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b265c7a-3909-4fd6-9981-c1723fbc1e02-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-x2gj8\" (UID: \"3b265c7a-3909-4fd6-9981-c1723fbc1e02\") " pod="openstack/nova-cell1-conductor-db-sync-x2gj8" Oct 14 07:22:00 crc kubenswrapper[4870]: I1014 07:22:00.047120 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3b265c7a-3909-4fd6-9981-c1723fbc1e02-scripts\") pod \"nova-cell1-conductor-db-sync-x2gj8\" (UID: \"3b265c7a-3909-4fd6-9981-c1723fbc1e02\") " pod="openstack/nova-cell1-conductor-db-sync-x2gj8" Oct 14 07:22:00 crc kubenswrapper[4870]: I1014 07:22:00.047213 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3b265c7a-3909-4fd6-9981-c1723fbc1e02-config-data\") pod \"nova-cell1-conductor-db-sync-x2gj8\" (UID: \"3b265c7a-3909-4fd6-9981-c1723fbc1e02\") " pod="openstack/nova-cell1-conductor-db-sync-x2gj8" Oct 14 07:22:00 crc kubenswrapper[4870]: I1014 07:22:00.047241 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t6pqs\" (UniqueName: \"kubernetes.io/projected/3b265c7a-3909-4fd6-9981-c1723fbc1e02-kube-api-access-t6pqs\") pod \"nova-cell1-conductor-db-sync-x2gj8\" (UID: \"3b265c7a-3909-4fd6-9981-c1723fbc1e02\") " pod="openstack/nova-cell1-conductor-db-sync-x2gj8" Oct 14 07:22:00 crc kubenswrapper[4870]: I1014 07:22:00.055722 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3b265c7a-3909-4fd6-9981-c1723fbc1e02-config-data\") pod \"nova-cell1-conductor-db-sync-x2gj8\" (UID: \"3b265c7a-3909-4fd6-9981-c1723fbc1e02\") " pod="openstack/nova-cell1-conductor-db-sync-x2gj8" Oct 14 07:22:00 crc kubenswrapper[4870]: I1014 07:22:00.056509 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3b265c7a-3909-4fd6-9981-c1723fbc1e02-scripts\") pod \"nova-cell1-conductor-db-sync-x2gj8\" (UID: \"3b265c7a-3909-4fd6-9981-c1723fbc1e02\") " pod="openstack/nova-cell1-conductor-db-sync-x2gj8" Oct 14 07:22:00 crc kubenswrapper[4870]: I1014 07:22:00.063474 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b265c7a-3909-4fd6-9981-c1723fbc1e02-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-x2gj8\" (UID: \"3b265c7a-3909-4fd6-9981-c1723fbc1e02\") " pod="openstack/nova-cell1-conductor-db-sync-x2gj8" Oct 14 07:22:00 crc kubenswrapper[4870]: I1014 07:22:00.065658 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t6pqs\" (UniqueName: \"kubernetes.io/projected/3b265c7a-3909-4fd6-9981-c1723fbc1e02-kube-api-access-t6pqs\") pod \"nova-cell1-conductor-db-sync-x2gj8\" (UID: \"3b265c7a-3909-4fd6-9981-c1723fbc1e02\") " pod="openstack/nova-cell1-conductor-db-sync-x2gj8" Oct 14 07:22:00 crc kubenswrapper[4870]: I1014 07:22:00.102574 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 14 07:22:00 crc kubenswrapper[4870]: I1014 07:22:00.167388 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 14 07:22:00 crc kubenswrapper[4870]: W1014 07:22:00.170963 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poded189910_4793_4a5f_9688_d643f5edaf2c.slice/crio-532f4ff8321952bd5ee06531a7a8ee449c14ed657e7d310a304c8c033c0063c0 WatchSource:0}: Error finding container 532f4ff8321952bd5ee06531a7a8ee449c14ed657e7d310a304c8c033c0063c0: Status 404 returned error can't find the container with id 532f4ff8321952bd5ee06531a7a8ee449c14ed657e7d310a304c8c033c0063c0 Oct 14 07:22:00 crc kubenswrapper[4870]: I1014 07:22:00.306030 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6957666cb7-wgpsx"] Oct 14 07:22:00 crc kubenswrapper[4870]: W1014 07:22:00.309659 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd5d3e78a_8ed8_40a5_8a64_734c7836155c.slice/crio-42174d886426663b8c0ce64845f0962beb3274ed0289eeb2e8d5350a53924767 WatchSource:0}: Error finding container 42174d886426663b8c0ce64845f0962beb3274ed0289eeb2e8d5350a53924767: Status 404 returned error can't find the container with id 42174d886426663b8c0ce64845f0962beb3274ed0289eeb2e8d5350a53924767 Oct 14 07:22:00 crc kubenswrapper[4870]: I1014 07:22:00.323932 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-x2gj8" Oct 14 07:22:00 crc kubenswrapper[4870]: I1014 07:22:00.798517 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"67e09a03-b028-4444-8aba-63be9c7e1abd","Type":"ContainerStarted","Data":"64736b327ed16b55d75fbb107d9283e261cb66851f8f552b53a406fa64433954"} Oct 14 07:22:00 crc kubenswrapper[4870]: I1014 07:22:00.802559 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-q76l4" event={"ID":"d147118e-e08e-4c64-b157-da3ca3afb6d7","Type":"ContainerStarted","Data":"2488802b352a138347373046951250f07e48d44d2dce6d3f42f5ae6763c49ffd"} Oct 14 07:22:00 crc kubenswrapper[4870]: I1014 07:22:00.802593 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-q76l4" event={"ID":"d147118e-e08e-4c64-b157-da3ca3afb6d7","Type":"ContainerStarted","Data":"be7b946d0f4be9d9a635043bfffd44d4df11908f22cc9babed3e86a2dc740e4b"} Oct 14 07:22:00 crc kubenswrapper[4870]: I1014 07:22:00.811010 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-x2gj8"] Oct 14 07:22:00 crc kubenswrapper[4870]: I1014 07:22:00.812745 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"ed7c5b6f-eb0a-44fc-a6a3-6f43909b6406","Type":"ContainerStarted","Data":"6a1a19a25ac1b542a719c1ec08cc06c9b012fd4a776ab1490d6937b80b9cf21f"} Oct 14 07:22:00 crc kubenswrapper[4870]: I1014 07:22:00.816599 4870 generic.go:334] "Generic (PLEG): container finished" podID="d5d3e78a-8ed8-40a5-8a64-734c7836155c" containerID="da1b8f24e5ce4601d5b9ab11fdd2ff2e4c7ba1c1afbb715dc02d5dbe6b6257d2" exitCode=0 Oct 14 07:22:00 crc kubenswrapper[4870]: I1014 07:22:00.817545 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6957666cb7-wgpsx" event={"ID":"d5d3e78a-8ed8-40a5-8a64-734c7836155c","Type":"ContainerDied","Data":"da1b8f24e5ce4601d5b9ab11fdd2ff2e4c7ba1c1afbb715dc02d5dbe6b6257d2"} Oct 14 07:22:00 crc kubenswrapper[4870]: I1014 07:22:00.817581 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6957666cb7-wgpsx" event={"ID":"d5d3e78a-8ed8-40a5-8a64-734c7836155c","Type":"ContainerStarted","Data":"42174d886426663b8c0ce64845f0962beb3274ed0289eeb2e8d5350a53924767"} Oct 14 07:22:00 crc kubenswrapper[4870]: I1014 07:22:00.827282 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ed189910-4793-4a5f-9688-d643f5edaf2c","Type":"ContainerStarted","Data":"532f4ff8321952bd5ee06531a7a8ee449c14ed657e7d310a304c8c033c0063c0"} Oct 14 07:22:00 crc kubenswrapper[4870]: I1014 07:22:00.830260 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"cd44bae6-b5d1-4b08-88de-7c5271f54ed3","Type":"ContainerStarted","Data":"6dee636cd479af70c001abe0df98e5cba74b39c63236703b96a2e4b1ac1f3358"} Oct 14 07:22:00 crc kubenswrapper[4870]: I1014 07:22:00.846924 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-q76l4" podStartSLOduration=2.846903452 podStartE2EDuration="2.846903452s" podCreationTimestamp="2025-10-14 07:21:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:22:00.821394486 +0000 UTC m=+1256.518754867" watchObservedRunningTime="2025-10-14 07:22:00.846903452 +0000 UTC m=+1256.544263823" Oct 14 07:22:01 crc kubenswrapper[4870]: I1014 07:22:01.858867 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6957666cb7-wgpsx" event={"ID":"d5d3e78a-8ed8-40a5-8a64-734c7836155c","Type":"ContainerStarted","Data":"49adfddfc295310c01014e57a90893bd2516bc1e4d16c17bd66c300fee355163"} Oct 14 07:22:01 crc kubenswrapper[4870]: I1014 07:22:01.859265 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6957666cb7-wgpsx" Oct 14 07:22:01 crc kubenswrapper[4870]: I1014 07:22:01.869812 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-x2gj8" event={"ID":"3b265c7a-3909-4fd6-9981-c1723fbc1e02","Type":"ContainerStarted","Data":"ce1d38864bb409e404978e6b3a0f6bd89d3dba826ce7be995022fe4e0ac9b188"} Oct 14 07:22:01 crc kubenswrapper[4870]: I1014 07:22:01.870027 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-x2gj8" event={"ID":"3b265c7a-3909-4fd6-9981-c1723fbc1e02","Type":"ContainerStarted","Data":"5df333dc10c08df3ccbdcd54ec0537a2f6802e8ed904078ac97b1d308eb9d1b7"} Oct 14 07:22:01 crc kubenswrapper[4870]: I1014 07:22:01.893928 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6957666cb7-wgpsx" podStartSLOduration=2.893908636 podStartE2EDuration="2.893908636s" podCreationTimestamp="2025-10-14 07:21:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:22:01.890897576 +0000 UTC m=+1257.588257947" watchObservedRunningTime="2025-10-14 07:22:01.893908636 +0000 UTC m=+1257.591269007" Oct 14 07:22:01 crc kubenswrapper[4870]: I1014 07:22:01.924576 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-x2gj8" podStartSLOduration=2.9245515920000003 podStartE2EDuration="2.924551592s" podCreationTimestamp="2025-10-14 07:21:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:22:01.919431503 +0000 UTC m=+1257.616791884" watchObservedRunningTime="2025-10-14 07:22:01.924551592 +0000 UTC m=+1257.621911953" Oct 14 07:22:02 crc kubenswrapper[4870]: I1014 07:22:02.904421 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 14 07:22:02 crc kubenswrapper[4870]: I1014 07:22:02.913499 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 14 07:22:06 crc kubenswrapper[4870]: I1014 07:22:06.925913 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"cd44bae6-b5d1-4b08-88de-7c5271f54ed3","Type":"ContainerStarted","Data":"d311512ab3d18b4243743bfbfb8a259533302afd01cbbcfda23ec6a678817840"} Oct 14 07:22:06 crc kubenswrapper[4870]: I1014 07:22:06.926520 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"cd44bae6-b5d1-4b08-88de-7c5271f54ed3","Type":"ContainerStarted","Data":"76f2a940e9d8e2f75865166a0366231116cc1a0a333f49e74aef93fb29a8e1a2"} Oct 14 07:22:06 crc kubenswrapper[4870]: I1014 07:22:06.929804 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"67e09a03-b028-4444-8aba-63be9c7e1abd","Type":"ContainerStarted","Data":"231197d5071a9dbc186c3fce6f80efb27d4af4bc8993ba7cab278415accb4013"} Oct 14 07:22:06 crc kubenswrapper[4870]: I1014 07:22:06.931882 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"ed7c5b6f-eb0a-44fc-a6a3-6f43909b6406","Type":"ContainerStarted","Data":"4d1574b6ecc78807ca0e045857f0ad61906f9011616d9b3dbeba62e89d80183d"} Oct 14 07:22:06 crc kubenswrapper[4870]: I1014 07:22:06.931921 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="ed7c5b6f-eb0a-44fc-a6a3-6f43909b6406" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://4d1574b6ecc78807ca0e045857f0ad61906f9011616d9b3dbeba62e89d80183d" gracePeriod=30 Oct 14 07:22:06 crc kubenswrapper[4870]: I1014 07:22:06.933930 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ed189910-4793-4a5f-9688-d643f5edaf2c","Type":"ContainerStarted","Data":"fe11e71b162e5fc6426b83a49c8f249cde31ad6438fbb56de2ebdf3bf7917d83"} Oct 14 07:22:06 crc kubenswrapper[4870]: I1014 07:22:06.933967 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ed189910-4793-4a5f-9688-d643f5edaf2c","Type":"ContainerStarted","Data":"6e42464e530466170317bb05bbfed402ef371c00aa013191553cc02c6b7f14e9"} Oct 14 07:22:06 crc kubenswrapper[4870]: I1014 07:22:06.934129 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="ed189910-4793-4a5f-9688-d643f5edaf2c" containerName="nova-metadata-metadata" containerID="cri-o://fe11e71b162e5fc6426b83a49c8f249cde31ad6438fbb56de2ebdf3bf7917d83" gracePeriod=30 Oct 14 07:22:06 crc kubenswrapper[4870]: I1014 07:22:06.934111 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="ed189910-4793-4a5f-9688-d643f5edaf2c" containerName="nova-metadata-log" containerID="cri-o://6e42464e530466170317bb05bbfed402ef371c00aa013191553cc02c6b7f14e9" gracePeriod=30 Oct 14 07:22:06 crc kubenswrapper[4870]: I1014 07:22:06.983630 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=3.032237884 podStartE2EDuration="8.983606138s" podCreationTimestamp="2025-10-14 07:21:58 +0000 UTC" firstStartedPulling="2025-10-14 07:21:59.860384572 +0000 UTC m=+1255.557744943" lastFinishedPulling="2025-10-14 07:22:05.811752836 +0000 UTC m=+1261.509113197" observedRunningTime="2025-10-14 07:22:06.945676342 +0000 UTC m=+1262.643036723" watchObservedRunningTime="2025-10-14 07:22:06.983606138 +0000 UTC m=+1262.680966509" Oct 14 07:22:06 crc kubenswrapper[4870]: I1014 07:22:06.985016 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=3.343622603 podStartE2EDuration="8.985004271s" podCreationTimestamp="2025-10-14 07:21:58 +0000 UTC" firstStartedPulling="2025-10-14 07:22:00.174073955 +0000 UTC m=+1255.871434326" lastFinishedPulling="2025-10-14 07:22:05.815455623 +0000 UTC m=+1261.512815994" observedRunningTime="2025-10-14 07:22:06.968362502 +0000 UTC m=+1262.665722883" watchObservedRunningTime="2025-10-14 07:22:06.985004271 +0000 UTC m=+1262.682364662" Oct 14 07:22:06 crc kubenswrapper[4870]: I1014 07:22:06.990580 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=3.30798058 podStartE2EDuration="8.990561831s" podCreationTimestamp="2025-10-14 07:21:58 +0000 UTC" firstStartedPulling="2025-10-14 07:22:00.129250817 +0000 UTC m=+1255.826611188" lastFinishedPulling="2025-10-14 07:22:05.811832058 +0000 UTC m=+1261.509192439" observedRunningTime="2025-10-14 07:22:06.98622488 +0000 UTC m=+1262.683585251" watchObservedRunningTime="2025-10-14 07:22:06.990561831 +0000 UTC m=+1262.687922202" Oct 14 07:22:07 crc kubenswrapper[4870]: I1014 07:22:07.009600 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=3.170424266 podStartE2EDuration="9.009576616s" podCreationTimestamp="2025-10-14 07:21:58 +0000 UTC" firstStartedPulling="2025-10-14 07:21:59.972598706 +0000 UTC m=+1255.669959077" lastFinishedPulling="2025-10-14 07:22:05.811751056 +0000 UTC m=+1261.509111427" observedRunningTime="2025-10-14 07:22:07.006103314 +0000 UTC m=+1262.703463685" watchObservedRunningTime="2025-10-14 07:22:07.009576616 +0000 UTC m=+1262.706936997" Oct 14 07:22:07 crc kubenswrapper[4870]: I1014 07:22:07.594675 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 14 07:22:07 crc kubenswrapper[4870]: I1014 07:22:07.699693 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ncz4r\" (UniqueName: \"kubernetes.io/projected/ed189910-4793-4a5f-9688-d643f5edaf2c-kube-api-access-ncz4r\") pod \"ed189910-4793-4a5f-9688-d643f5edaf2c\" (UID: \"ed189910-4793-4a5f-9688-d643f5edaf2c\") " Oct 14 07:22:07 crc kubenswrapper[4870]: I1014 07:22:07.699769 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ed189910-4793-4a5f-9688-d643f5edaf2c-logs\") pod \"ed189910-4793-4a5f-9688-d643f5edaf2c\" (UID: \"ed189910-4793-4a5f-9688-d643f5edaf2c\") " Oct 14 07:22:07 crc kubenswrapper[4870]: I1014 07:22:07.699805 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ed189910-4793-4a5f-9688-d643f5edaf2c-config-data\") pod \"ed189910-4793-4a5f-9688-d643f5edaf2c\" (UID: \"ed189910-4793-4a5f-9688-d643f5edaf2c\") " Oct 14 07:22:07 crc kubenswrapper[4870]: I1014 07:22:07.699940 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed189910-4793-4a5f-9688-d643f5edaf2c-combined-ca-bundle\") pod \"ed189910-4793-4a5f-9688-d643f5edaf2c\" (UID: \"ed189910-4793-4a5f-9688-d643f5edaf2c\") " Oct 14 07:22:07 crc kubenswrapper[4870]: I1014 07:22:07.701706 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ed189910-4793-4a5f-9688-d643f5edaf2c-logs" (OuterVolumeSpecName: "logs") pod "ed189910-4793-4a5f-9688-d643f5edaf2c" (UID: "ed189910-4793-4a5f-9688-d643f5edaf2c"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:22:07 crc kubenswrapper[4870]: I1014 07:22:07.706251 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ed189910-4793-4a5f-9688-d643f5edaf2c-kube-api-access-ncz4r" (OuterVolumeSpecName: "kube-api-access-ncz4r") pod "ed189910-4793-4a5f-9688-d643f5edaf2c" (UID: "ed189910-4793-4a5f-9688-d643f5edaf2c"). InnerVolumeSpecName "kube-api-access-ncz4r". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:22:07 crc kubenswrapper[4870]: I1014 07:22:07.730010 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ed189910-4793-4a5f-9688-d643f5edaf2c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ed189910-4793-4a5f-9688-d643f5edaf2c" (UID: "ed189910-4793-4a5f-9688-d643f5edaf2c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:22:07 crc kubenswrapper[4870]: I1014 07:22:07.732763 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ed189910-4793-4a5f-9688-d643f5edaf2c-config-data" (OuterVolumeSpecName: "config-data") pod "ed189910-4793-4a5f-9688-d643f5edaf2c" (UID: "ed189910-4793-4a5f-9688-d643f5edaf2c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:22:07 crc kubenswrapper[4870]: I1014 07:22:07.802080 4870 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed189910-4793-4a5f-9688-d643f5edaf2c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 07:22:07 crc kubenswrapper[4870]: I1014 07:22:07.802599 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ncz4r\" (UniqueName: \"kubernetes.io/projected/ed189910-4793-4a5f-9688-d643f5edaf2c-kube-api-access-ncz4r\") on node \"crc\" DevicePath \"\"" Oct 14 07:22:07 crc kubenswrapper[4870]: I1014 07:22:07.802615 4870 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ed189910-4793-4a5f-9688-d643f5edaf2c-logs\") on node \"crc\" DevicePath \"\"" Oct 14 07:22:07 crc kubenswrapper[4870]: I1014 07:22:07.802625 4870 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ed189910-4793-4a5f-9688-d643f5edaf2c-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 07:22:07 crc kubenswrapper[4870]: I1014 07:22:07.946326 4870 generic.go:334] "Generic (PLEG): container finished" podID="ed189910-4793-4a5f-9688-d643f5edaf2c" containerID="fe11e71b162e5fc6426b83a49c8f249cde31ad6438fbb56de2ebdf3bf7917d83" exitCode=0 Oct 14 07:22:07 crc kubenswrapper[4870]: I1014 07:22:07.946362 4870 generic.go:334] "Generic (PLEG): container finished" podID="ed189910-4793-4a5f-9688-d643f5edaf2c" containerID="6e42464e530466170317bb05bbfed402ef371c00aa013191553cc02c6b7f14e9" exitCode=143 Oct 14 07:22:07 crc kubenswrapper[4870]: I1014 07:22:07.946525 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 14 07:22:07 crc kubenswrapper[4870]: I1014 07:22:07.946705 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ed189910-4793-4a5f-9688-d643f5edaf2c","Type":"ContainerDied","Data":"fe11e71b162e5fc6426b83a49c8f249cde31ad6438fbb56de2ebdf3bf7917d83"} Oct 14 07:22:07 crc kubenswrapper[4870]: I1014 07:22:07.946739 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ed189910-4793-4a5f-9688-d643f5edaf2c","Type":"ContainerDied","Data":"6e42464e530466170317bb05bbfed402ef371c00aa013191553cc02c6b7f14e9"} Oct 14 07:22:07 crc kubenswrapper[4870]: I1014 07:22:07.946758 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ed189910-4793-4a5f-9688-d643f5edaf2c","Type":"ContainerDied","Data":"532f4ff8321952bd5ee06531a7a8ee449c14ed657e7d310a304c8c033c0063c0"} Oct 14 07:22:07 crc kubenswrapper[4870]: I1014 07:22:07.946779 4870 scope.go:117] "RemoveContainer" containerID="fe11e71b162e5fc6426b83a49c8f249cde31ad6438fbb56de2ebdf3bf7917d83" Oct 14 07:22:07 crc kubenswrapper[4870]: I1014 07:22:07.985210 4870 scope.go:117] "RemoveContainer" containerID="6e42464e530466170317bb05bbfed402ef371c00aa013191553cc02c6b7f14e9" Oct 14 07:22:07 crc kubenswrapper[4870]: I1014 07:22:07.989628 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 14 07:22:08 crc kubenswrapper[4870]: I1014 07:22:08.000463 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 14 07:22:08 crc kubenswrapper[4870]: I1014 07:22:08.016668 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 14 07:22:08 crc kubenswrapper[4870]: E1014 07:22:08.017205 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed189910-4793-4a5f-9688-d643f5edaf2c" containerName="nova-metadata-metadata" Oct 14 07:22:08 crc kubenswrapper[4870]: I1014 07:22:08.017231 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed189910-4793-4a5f-9688-d643f5edaf2c" containerName="nova-metadata-metadata" Oct 14 07:22:08 crc kubenswrapper[4870]: E1014 07:22:08.017272 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed189910-4793-4a5f-9688-d643f5edaf2c" containerName="nova-metadata-log" Oct 14 07:22:08 crc kubenswrapper[4870]: I1014 07:22:08.017281 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed189910-4793-4a5f-9688-d643f5edaf2c" containerName="nova-metadata-log" Oct 14 07:22:08 crc kubenswrapper[4870]: I1014 07:22:08.017520 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="ed189910-4793-4a5f-9688-d643f5edaf2c" containerName="nova-metadata-metadata" Oct 14 07:22:08 crc kubenswrapper[4870]: I1014 07:22:08.017550 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="ed189910-4793-4a5f-9688-d643f5edaf2c" containerName="nova-metadata-log" Oct 14 07:22:08 crc kubenswrapper[4870]: I1014 07:22:08.020010 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 14 07:22:08 crc kubenswrapper[4870]: I1014 07:22:08.020423 4870 scope.go:117] "RemoveContainer" containerID="fe11e71b162e5fc6426b83a49c8f249cde31ad6438fbb56de2ebdf3bf7917d83" Oct 14 07:22:08 crc kubenswrapper[4870]: E1014 07:22:08.020929 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fe11e71b162e5fc6426b83a49c8f249cde31ad6438fbb56de2ebdf3bf7917d83\": container with ID starting with fe11e71b162e5fc6426b83a49c8f249cde31ad6438fbb56de2ebdf3bf7917d83 not found: ID does not exist" containerID="fe11e71b162e5fc6426b83a49c8f249cde31ad6438fbb56de2ebdf3bf7917d83" Oct 14 07:22:08 crc kubenswrapper[4870]: I1014 07:22:08.020961 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fe11e71b162e5fc6426b83a49c8f249cde31ad6438fbb56de2ebdf3bf7917d83"} err="failed to get container status \"fe11e71b162e5fc6426b83a49c8f249cde31ad6438fbb56de2ebdf3bf7917d83\": rpc error: code = NotFound desc = could not find container \"fe11e71b162e5fc6426b83a49c8f249cde31ad6438fbb56de2ebdf3bf7917d83\": container with ID starting with fe11e71b162e5fc6426b83a49c8f249cde31ad6438fbb56de2ebdf3bf7917d83 not found: ID does not exist" Oct 14 07:22:08 crc kubenswrapper[4870]: I1014 07:22:08.020987 4870 scope.go:117] "RemoveContainer" containerID="6e42464e530466170317bb05bbfed402ef371c00aa013191553cc02c6b7f14e9" Oct 14 07:22:08 crc kubenswrapper[4870]: E1014 07:22:08.021293 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6e42464e530466170317bb05bbfed402ef371c00aa013191553cc02c6b7f14e9\": container with ID starting with 6e42464e530466170317bb05bbfed402ef371c00aa013191553cc02c6b7f14e9 not found: ID does not exist" containerID="6e42464e530466170317bb05bbfed402ef371c00aa013191553cc02c6b7f14e9" Oct 14 07:22:08 crc kubenswrapper[4870]: I1014 07:22:08.021340 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6e42464e530466170317bb05bbfed402ef371c00aa013191553cc02c6b7f14e9"} err="failed to get container status \"6e42464e530466170317bb05bbfed402ef371c00aa013191553cc02c6b7f14e9\": rpc error: code = NotFound desc = could not find container \"6e42464e530466170317bb05bbfed402ef371c00aa013191553cc02c6b7f14e9\": container with ID starting with 6e42464e530466170317bb05bbfed402ef371c00aa013191553cc02c6b7f14e9 not found: ID does not exist" Oct 14 07:22:08 crc kubenswrapper[4870]: I1014 07:22:08.021370 4870 scope.go:117] "RemoveContainer" containerID="fe11e71b162e5fc6426b83a49c8f249cde31ad6438fbb56de2ebdf3bf7917d83" Oct 14 07:22:08 crc kubenswrapper[4870]: I1014 07:22:08.021711 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fe11e71b162e5fc6426b83a49c8f249cde31ad6438fbb56de2ebdf3bf7917d83"} err="failed to get container status \"fe11e71b162e5fc6426b83a49c8f249cde31ad6438fbb56de2ebdf3bf7917d83\": rpc error: code = NotFound desc = could not find container \"fe11e71b162e5fc6426b83a49c8f249cde31ad6438fbb56de2ebdf3bf7917d83\": container with ID starting with fe11e71b162e5fc6426b83a49c8f249cde31ad6438fbb56de2ebdf3bf7917d83 not found: ID does not exist" Oct 14 07:22:08 crc kubenswrapper[4870]: I1014 07:22:08.021737 4870 scope.go:117] "RemoveContainer" containerID="6e42464e530466170317bb05bbfed402ef371c00aa013191553cc02c6b7f14e9" Oct 14 07:22:08 crc kubenswrapper[4870]: I1014 07:22:08.021820 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 14 07:22:08 crc kubenswrapper[4870]: I1014 07:22:08.021946 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6e42464e530466170317bb05bbfed402ef371c00aa013191553cc02c6b7f14e9"} err="failed to get container status \"6e42464e530466170317bb05bbfed402ef371c00aa013191553cc02c6b7f14e9\": rpc error: code = NotFound desc = could not find container \"6e42464e530466170317bb05bbfed402ef371c00aa013191553cc02c6b7f14e9\": container with ID starting with 6e42464e530466170317bb05bbfed402ef371c00aa013191553cc02c6b7f14e9 not found: ID does not exist" Oct 14 07:22:08 crc kubenswrapper[4870]: I1014 07:22:08.027081 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Oct 14 07:22:08 crc kubenswrapper[4870]: I1014 07:22:08.041395 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 14 07:22:08 crc kubenswrapper[4870]: I1014 07:22:08.111035 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c70361b-b549-48b9-92f7-2c8a611c30bd-config-data\") pod \"nova-metadata-0\" (UID: \"8c70361b-b549-48b9-92f7-2c8a611c30bd\") " pod="openstack/nova-metadata-0" Oct 14 07:22:08 crc kubenswrapper[4870]: I1014 07:22:08.111130 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6zx9w\" (UniqueName: \"kubernetes.io/projected/8c70361b-b549-48b9-92f7-2c8a611c30bd-kube-api-access-6zx9w\") pod \"nova-metadata-0\" (UID: \"8c70361b-b549-48b9-92f7-2c8a611c30bd\") " pod="openstack/nova-metadata-0" Oct 14 07:22:08 crc kubenswrapper[4870]: I1014 07:22:08.111178 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c70361b-b549-48b9-92f7-2c8a611c30bd-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"8c70361b-b549-48b9-92f7-2c8a611c30bd\") " pod="openstack/nova-metadata-0" Oct 14 07:22:08 crc kubenswrapper[4870]: I1014 07:22:08.111235 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8c70361b-b549-48b9-92f7-2c8a611c30bd-logs\") pod \"nova-metadata-0\" (UID: \"8c70361b-b549-48b9-92f7-2c8a611c30bd\") " pod="openstack/nova-metadata-0" Oct 14 07:22:08 crc kubenswrapper[4870]: I1014 07:22:08.111528 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/8c70361b-b549-48b9-92f7-2c8a611c30bd-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"8c70361b-b549-48b9-92f7-2c8a611c30bd\") " pod="openstack/nova-metadata-0" Oct 14 07:22:08 crc kubenswrapper[4870]: I1014 07:22:08.213887 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/8c70361b-b549-48b9-92f7-2c8a611c30bd-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"8c70361b-b549-48b9-92f7-2c8a611c30bd\") " pod="openstack/nova-metadata-0" Oct 14 07:22:08 crc kubenswrapper[4870]: I1014 07:22:08.214013 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c70361b-b549-48b9-92f7-2c8a611c30bd-config-data\") pod \"nova-metadata-0\" (UID: \"8c70361b-b549-48b9-92f7-2c8a611c30bd\") " pod="openstack/nova-metadata-0" Oct 14 07:22:08 crc kubenswrapper[4870]: I1014 07:22:08.214066 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6zx9w\" (UniqueName: \"kubernetes.io/projected/8c70361b-b549-48b9-92f7-2c8a611c30bd-kube-api-access-6zx9w\") pod \"nova-metadata-0\" (UID: \"8c70361b-b549-48b9-92f7-2c8a611c30bd\") " pod="openstack/nova-metadata-0" Oct 14 07:22:08 crc kubenswrapper[4870]: I1014 07:22:08.214100 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c70361b-b549-48b9-92f7-2c8a611c30bd-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"8c70361b-b549-48b9-92f7-2c8a611c30bd\") " pod="openstack/nova-metadata-0" Oct 14 07:22:08 crc kubenswrapper[4870]: I1014 07:22:08.214142 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8c70361b-b549-48b9-92f7-2c8a611c30bd-logs\") pod \"nova-metadata-0\" (UID: \"8c70361b-b549-48b9-92f7-2c8a611c30bd\") " pod="openstack/nova-metadata-0" Oct 14 07:22:08 crc kubenswrapper[4870]: I1014 07:22:08.214700 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8c70361b-b549-48b9-92f7-2c8a611c30bd-logs\") pod \"nova-metadata-0\" (UID: \"8c70361b-b549-48b9-92f7-2c8a611c30bd\") " pod="openstack/nova-metadata-0" Oct 14 07:22:08 crc kubenswrapper[4870]: I1014 07:22:08.219110 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/8c70361b-b549-48b9-92f7-2c8a611c30bd-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"8c70361b-b549-48b9-92f7-2c8a611c30bd\") " pod="openstack/nova-metadata-0" Oct 14 07:22:08 crc kubenswrapper[4870]: I1014 07:22:08.220844 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c70361b-b549-48b9-92f7-2c8a611c30bd-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"8c70361b-b549-48b9-92f7-2c8a611c30bd\") " pod="openstack/nova-metadata-0" Oct 14 07:22:08 crc kubenswrapper[4870]: I1014 07:22:08.232271 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c70361b-b549-48b9-92f7-2c8a611c30bd-config-data\") pod \"nova-metadata-0\" (UID: \"8c70361b-b549-48b9-92f7-2c8a611c30bd\") " pod="openstack/nova-metadata-0" Oct 14 07:22:08 crc kubenswrapper[4870]: I1014 07:22:08.234137 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6zx9w\" (UniqueName: \"kubernetes.io/projected/8c70361b-b549-48b9-92f7-2c8a611c30bd-kube-api-access-6zx9w\") pod \"nova-metadata-0\" (UID: \"8c70361b-b549-48b9-92f7-2c8a611c30bd\") " pod="openstack/nova-metadata-0" Oct 14 07:22:08 crc kubenswrapper[4870]: I1014 07:22:08.356413 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 14 07:22:08 crc kubenswrapper[4870]: I1014 07:22:08.889182 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 14 07:22:08 crc kubenswrapper[4870]: W1014 07:22:08.896918 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8c70361b_b549_48b9_92f7_2c8a611c30bd.slice/crio-71bfc80fe07fdfc3a8eda33139fb3cdd01f7c27904dceabe861c504ae7714ea8 WatchSource:0}: Error finding container 71bfc80fe07fdfc3a8eda33139fb3cdd01f7c27904dceabe861c504ae7714ea8: Status 404 returned error can't find the container with id 71bfc80fe07fdfc3a8eda33139fb3cdd01f7c27904dceabe861c504ae7714ea8 Oct 14 07:22:08 crc kubenswrapper[4870]: I1014 07:22:08.964639 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"8c70361b-b549-48b9-92f7-2c8a611c30bd","Type":"ContainerStarted","Data":"71bfc80fe07fdfc3a8eda33139fb3cdd01f7c27904dceabe861c504ae7714ea8"} Oct 14 07:22:08 crc kubenswrapper[4870]: I1014 07:22:08.970827 4870 generic.go:334] "Generic (PLEG): container finished" podID="d147118e-e08e-4c64-b157-da3ca3afb6d7" containerID="2488802b352a138347373046951250f07e48d44d2dce6d3f42f5ae6763c49ffd" exitCode=0 Oct 14 07:22:08 crc kubenswrapper[4870]: I1014 07:22:08.970900 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-q76l4" event={"ID":"d147118e-e08e-4c64-b157-da3ca3afb6d7","Type":"ContainerDied","Data":"2488802b352a138347373046951250f07e48d44d2dce6d3f42f5ae6763c49ffd"} Oct 14 07:22:09 crc kubenswrapper[4870]: I1014 07:22:09.062982 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ed189910-4793-4a5f-9688-d643f5edaf2c" path="/var/lib/kubelet/pods/ed189910-4793-4a5f-9688-d643f5edaf2c/volumes" Oct 14 07:22:09 crc kubenswrapper[4870]: I1014 07:22:09.284310 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 14 07:22:09 crc kubenswrapper[4870]: I1014 07:22:09.284787 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 14 07:22:09 crc kubenswrapper[4870]: I1014 07:22:09.371025 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 14 07:22:09 crc kubenswrapper[4870]: I1014 07:22:09.371090 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 14 07:22:09 crc kubenswrapper[4870]: I1014 07:22:09.418534 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 14 07:22:09 crc kubenswrapper[4870]: I1014 07:22:09.484000 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Oct 14 07:22:09 crc kubenswrapper[4870]: I1014 07:22:09.517822 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6957666cb7-wgpsx" Oct 14 07:22:09 crc kubenswrapper[4870]: I1014 07:22:09.581310 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5b7d94f5b5-4qfsd"] Oct 14 07:22:09 crc kubenswrapper[4870]: I1014 07:22:09.581642 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5b7d94f5b5-4qfsd" podUID="0dba2f3b-4ddd-4a1f-8593-1f98dc4479b1" containerName="dnsmasq-dns" containerID="cri-o://9593b500b23a3c6e6593a46342c30a20d936527ddde619eb3f787d9d70f26504" gracePeriod=10 Oct 14 07:22:10 crc kubenswrapper[4870]: I1014 07:22:10.012786 4870 generic.go:334] "Generic (PLEG): container finished" podID="0dba2f3b-4ddd-4a1f-8593-1f98dc4479b1" containerID="9593b500b23a3c6e6593a46342c30a20d936527ddde619eb3f787d9d70f26504" exitCode=0 Oct 14 07:22:10 crc kubenswrapper[4870]: I1014 07:22:10.013063 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b7d94f5b5-4qfsd" event={"ID":"0dba2f3b-4ddd-4a1f-8593-1f98dc4479b1","Type":"ContainerDied","Data":"9593b500b23a3c6e6593a46342c30a20d936527ddde619eb3f787d9d70f26504"} Oct 14 07:22:10 crc kubenswrapper[4870]: I1014 07:22:10.017790 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"8c70361b-b549-48b9-92f7-2c8a611c30bd","Type":"ContainerStarted","Data":"653ad6afcaad03b604bd34bc66a67a40e5bccc9a6b8caee133a16cf728616a1b"} Oct 14 07:22:10 crc kubenswrapper[4870]: I1014 07:22:10.017824 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"8c70361b-b549-48b9-92f7-2c8a611c30bd","Type":"ContainerStarted","Data":"d859d55bc66f6b1bbc5f156fd01df4786718392416a30495c11d5c0a1c4b33bf"} Oct 14 07:22:10 crc kubenswrapper[4870]: I1014 07:22:10.054645 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=3.054621284 podStartE2EDuration="3.054621284s" podCreationTimestamp="2025-10-14 07:22:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:22:10.049966945 +0000 UTC m=+1265.747327316" watchObservedRunningTime="2025-10-14 07:22:10.054621284 +0000 UTC m=+1265.751981655" Oct 14 07:22:10 crc kubenswrapper[4870]: I1014 07:22:10.075646 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 14 07:22:10 crc kubenswrapper[4870]: I1014 07:22:10.227628 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b7d94f5b5-4qfsd" Oct 14 07:22:10 crc kubenswrapper[4870]: I1014 07:22:10.365065 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0dba2f3b-4ddd-4a1f-8593-1f98dc4479b1-dns-svc\") pod \"0dba2f3b-4ddd-4a1f-8593-1f98dc4479b1\" (UID: \"0dba2f3b-4ddd-4a1f-8593-1f98dc4479b1\") " Oct 14 07:22:10 crc kubenswrapper[4870]: I1014 07:22:10.365135 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0dba2f3b-4ddd-4a1f-8593-1f98dc4479b1-dns-swift-storage-0\") pod \"0dba2f3b-4ddd-4a1f-8593-1f98dc4479b1\" (UID: \"0dba2f3b-4ddd-4a1f-8593-1f98dc4479b1\") " Oct 14 07:22:10 crc kubenswrapper[4870]: I1014 07:22:10.365373 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tdcpf\" (UniqueName: \"kubernetes.io/projected/0dba2f3b-4ddd-4a1f-8593-1f98dc4479b1-kube-api-access-tdcpf\") pod \"0dba2f3b-4ddd-4a1f-8593-1f98dc4479b1\" (UID: \"0dba2f3b-4ddd-4a1f-8593-1f98dc4479b1\") " Oct 14 07:22:10 crc kubenswrapper[4870]: I1014 07:22:10.366557 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0dba2f3b-4ddd-4a1f-8593-1f98dc4479b1-ovsdbserver-nb\") pod \"0dba2f3b-4ddd-4a1f-8593-1f98dc4479b1\" (UID: \"0dba2f3b-4ddd-4a1f-8593-1f98dc4479b1\") " Oct 14 07:22:10 crc kubenswrapper[4870]: I1014 07:22:10.366616 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0dba2f3b-4ddd-4a1f-8593-1f98dc4479b1-ovsdbserver-sb\") pod \"0dba2f3b-4ddd-4a1f-8593-1f98dc4479b1\" (UID: \"0dba2f3b-4ddd-4a1f-8593-1f98dc4479b1\") " Oct 14 07:22:10 crc kubenswrapper[4870]: I1014 07:22:10.366660 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0dba2f3b-4ddd-4a1f-8593-1f98dc4479b1-config\") pod \"0dba2f3b-4ddd-4a1f-8593-1f98dc4479b1\" (UID: \"0dba2f3b-4ddd-4a1f-8593-1f98dc4479b1\") " Oct 14 07:22:10 crc kubenswrapper[4870]: I1014 07:22:10.366671 4870 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="cd44bae6-b5d1-4b08-88de-7c5271f54ed3" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.183:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 14 07:22:10 crc kubenswrapper[4870]: I1014 07:22:10.367026 4870 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="cd44bae6-b5d1-4b08-88de-7c5271f54ed3" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.183:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 14 07:22:10 crc kubenswrapper[4870]: I1014 07:22:10.372398 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0dba2f3b-4ddd-4a1f-8593-1f98dc4479b1-kube-api-access-tdcpf" (OuterVolumeSpecName: "kube-api-access-tdcpf") pod "0dba2f3b-4ddd-4a1f-8593-1f98dc4479b1" (UID: "0dba2f3b-4ddd-4a1f-8593-1f98dc4479b1"). InnerVolumeSpecName "kube-api-access-tdcpf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:22:10 crc kubenswrapper[4870]: I1014 07:22:10.421974 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-q76l4" Oct 14 07:22:10 crc kubenswrapper[4870]: I1014 07:22:10.433526 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0dba2f3b-4ddd-4a1f-8593-1f98dc4479b1-config" (OuterVolumeSpecName: "config") pod "0dba2f3b-4ddd-4a1f-8593-1f98dc4479b1" (UID: "0dba2f3b-4ddd-4a1f-8593-1f98dc4479b1"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:22:10 crc kubenswrapper[4870]: I1014 07:22:10.434222 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0dba2f3b-4ddd-4a1f-8593-1f98dc4479b1-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "0dba2f3b-4ddd-4a1f-8593-1f98dc4479b1" (UID: "0dba2f3b-4ddd-4a1f-8593-1f98dc4479b1"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:22:10 crc kubenswrapper[4870]: I1014 07:22:10.434305 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0dba2f3b-4ddd-4a1f-8593-1f98dc4479b1-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "0dba2f3b-4ddd-4a1f-8593-1f98dc4479b1" (UID: "0dba2f3b-4ddd-4a1f-8593-1f98dc4479b1"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:22:10 crc kubenswrapper[4870]: I1014 07:22:10.453236 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0dba2f3b-4ddd-4a1f-8593-1f98dc4479b1-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "0dba2f3b-4ddd-4a1f-8593-1f98dc4479b1" (UID: "0dba2f3b-4ddd-4a1f-8593-1f98dc4479b1"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:22:10 crc kubenswrapper[4870]: I1014 07:22:10.457291 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0dba2f3b-4ddd-4a1f-8593-1f98dc4479b1-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "0dba2f3b-4ddd-4a1f-8593-1f98dc4479b1" (UID: "0dba2f3b-4ddd-4a1f-8593-1f98dc4479b1"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:22:10 crc kubenswrapper[4870]: I1014 07:22:10.469611 4870 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0dba2f3b-4ddd-4a1f-8593-1f98dc4479b1-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 14 07:22:10 crc kubenswrapper[4870]: I1014 07:22:10.469658 4870 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0dba2f3b-4ddd-4a1f-8593-1f98dc4479b1-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 14 07:22:10 crc kubenswrapper[4870]: I1014 07:22:10.469673 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tdcpf\" (UniqueName: \"kubernetes.io/projected/0dba2f3b-4ddd-4a1f-8593-1f98dc4479b1-kube-api-access-tdcpf\") on node \"crc\" DevicePath \"\"" Oct 14 07:22:10 crc kubenswrapper[4870]: I1014 07:22:10.469684 4870 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0dba2f3b-4ddd-4a1f-8593-1f98dc4479b1-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 14 07:22:10 crc kubenswrapper[4870]: I1014 07:22:10.469694 4870 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0dba2f3b-4ddd-4a1f-8593-1f98dc4479b1-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 14 07:22:10 crc kubenswrapper[4870]: I1014 07:22:10.469705 4870 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0dba2f3b-4ddd-4a1f-8593-1f98dc4479b1-config\") on node \"crc\" DevicePath \"\"" Oct 14 07:22:10 crc kubenswrapper[4870]: I1014 07:22:10.571101 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d147118e-e08e-4c64-b157-da3ca3afb6d7-combined-ca-bundle\") pod \"d147118e-e08e-4c64-b157-da3ca3afb6d7\" (UID: \"d147118e-e08e-4c64-b157-da3ca3afb6d7\") " Oct 14 07:22:10 crc kubenswrapper[4870]: I1014 07:22:10.571144 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zjn89\" (UniqueName: \"kubernetes.io/projected/d147118e-e08e-4c64-b157-da3ca3afb6d7-kube-api-access-zjn89\") pod \"d147118e-e08e-4c64-b157-da3ca3afb6d7\" (UID: \"d147118e-e08e-4c64-b157-da3ca3afb6d7\") " Oct 14 07:22:10 crc kubenswrapper[4870]: I1014 07:22:10.571265 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d147118e-e08e-4c64-b157-da3ca3afb6d7-scripts\") pod \"d147118e-e08e-4c64-b157-da3ca3afb6d7\" (UID: \"d147118e-e08e-4c64-b157-da3ca3afb6d7\") " Oct 14 07:22:10 crc kubenswrapper[4870]: I1014 07:22:10.571350 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d147118e-e08e-4c64-b157-da3ca3afb6d7-config-data\") pod \"d147118e-e08e-4c64-b157-da3ca3afb6d7\" (UID: \"d147118e-e08e-4c64-b157-da3ca3afb6d7\") " Oct 14 07:22:10 crc kubenswrapper[4870]: I1014 07:22:10.575041 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d147118e-e08e-4c64-b157-da3ca3afb6d7-scripts" (OuterVolumeSpecName: "scripts") pod "d147118e-e08e-4c64-b157-da3ca3afb6d7" (UID: "d147118e-e08e-4c64-b157-da3ca3afb6d7"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:22:10 crc kubenswrapper[4870]: I1014 07:22:10.575113 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d147118e-e08e-4c64-b157-da3ca3afb6d7-kube-api-access-zjn89" (OuterVolumeSpecName: "kube-api-access-zjn89") pod "d147118e-e08e-4c64-b157-da3ca3afb6d7" (UID: "d147118e-e08e-4c64-b157-da3ca3afb6d7"). InnerVolumeSpecName "kube-api-access-zjn89". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:22:10 crc kubenswrapper[4870]: I1014 07:22:10.600229 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d147118e-e08e-4c64-b157-da3ca3afb6d7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d147118e-e08e-4c64-b157-da3ca3afb6d7" (UID: "d147118e-e08e-4c64-b157-da3ca3afb6d7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:22:10 crc kubenswrapper[4870]: I1014 07:22:10.603568 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d147118e-e08e-4c64-b157-da3ca3afb6d7-config-data" (OuterVolumeSpecName: "config-data") pod "d147118e-e08e-4c64-b157-da3ca3afb6d7" (UID: "d147118e-e08e-4c64-b157-da3ca3afb6d7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:22:10 crc kubenswrapper[4870]: I1014 07:22:10.673055 4870 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d147118e-e08e-4c64-b157-da3ca3afb6d7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 07:22:10 crc kubenswrapper[4870]: I1014 07:22:10.673092 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zjn89\" (UniqueName: \"kubernetes.io/projected/d147118e-e08e-4c64-b157-da3ca3afb6d7-kube-api-access-zjn89\") on node \"crc\" DevicePath \"\"" Oct 14 07:22:10 crc kubenswrapper[4870]: I1014 07:22:10.673104 4870 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d147118e-e08e-4c64-b157-da3ca3afb6d7-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 07:22:10 crc kubenswrapper[4870]: I1014 07:22:10.673115 4870 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d147118e-e08e-4c64-b157-da3ca3afb6d7-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 07:22:11 crc kubenswrapper[4870]: I1014 07:22:11.026556 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-q76l4" Oct 14 07:22:11 crc kubenswrapper[4870]: I1014 07:22:11.026608 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-q76l4" event={"ID":"d147118e-e08e-4c64-b157-da3ca3afb6d7","Type":"ContainerDied","Data":"be7b946d0f4be9d9a635043bfffd44d4df11908f22cc9babed3e86a2dc740e4b"} Oct 14 07:22:11 crc kubenswrapper[4870]: I1014 07:22:11.026754 4870 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="be7b946d0f4be9d9a635043bfffd44d4df11908f22cc9babed3e86a2dc740e4b" Oct 14 07:22:11 crc kubenswrapper[4870]: I1014 07:22:11.028705 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b7d94f5b5-4qfsd" event={"ID":"0dba2f3b-4ddd-4a1f-8593-1f98dc4479b1","Type":"ContainerDied","Data":"458ddb1619a03b4bad71231c7d62bfef1e7348a6b0df30d3b5a0cbd0daefffe5"} Oct 14 07:22:11 crc kubenswrapper[4870]: I1014 07:22:11.028772 4870 scope.go:117] "RemoveContainer" containerID="9593b500b23a3c6e6593a46342c30a20d936527ddde619eb3f787d9d70f26504" Oct 14 07:22:11 crc kubenswrapper[4870]: I1014 07:22:11.028944 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b7d94f5b5-4qfsd" Oct 14 07:22:11 crc kubenswrapper[4870]: I1014 07:22:11.078683 4870 scope.go:117] "RemoveContainer" containerID="083d3c82a5af91bb717224dcca9b7742e957f81b1ee53ab1a10a591aa705f77c" Oct 14 07:22:11 crc kubenswrapper[4870]: I1014 07:22:11.079988 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5b7d94f5b5-4qfsd"] Oct 14 07:22:11 crc kubenswrapper[4870]: I1014 07:22:11.088138 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5b7d94f5b5-4qfsd"] Oct 14 07:22:11 crc kubenswrapper[4870]: I1014 07:22:11.210247 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 14 07:22:11 crc kubenswrapper[4870]: I1014 07:22:11.219317 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 14 07:22:11 crc kubenswrapper[4870]: I1014 07:22:11.219569 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="cd44bae6-b5d1-4b08-88de-7c5271f54ed3" containerName="nova-api-log" containerID="cri-o://76f2a940e9d8e2f75865166a0366231116cc1a0a333f49e74aef93fb29a8e1a2" gracePeriod=30 Oct 14 07:22:11 crc kubenswrapper[4870]: I1014 07:22:11.219724 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="cd44bae6-b5d1-4b08-88de-7c5271f54ed3" containerName="nova-api-api" containerID="cri-o://d311512ab3d18b4243743bfbfb8a259533302afd01cbbcfda23ec6a678817840" gracePeriod=30 Oct 14 07:22:11 crc kubenswrapper[4870]: I1014 07:22:11.285405 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 14 07:22:12 crc kubenswrapper[4870]: I1014 07:22:12.043193 4870 generic.go:334] "Generic (PLEG): container finished" podID="cd44bae6-b5d1-4b08-88de-7c5271f54ed3" containerID="76f2a940e9d8e2f75865166a0366231116cc1a0a333f49e74aef93fb29a8e1a2" exitCode=143 Oct 14 07:22:12 crc kubenswrapper[4870]: I1014 07:22:12.043271 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"cd44bae6-b5d1-4b08-88de-7c5271f54ed3","Type":"ContainerDied","Data":"76f2a940e9d8e2f75865166a0366231116cc1a0a333f49e74aef93fb29a8e1a2"} Oct 14 07:22:12 crc kubenswrapper[4870]: I1014 07:22:12.047266 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="8c70361b-b549-48b9-92f7-2c8a611c30bd" containerName="nova-metadata-log" containerID="cri-o://d859d55bc66f6b1bbc5f156fd01df4786718392416a30495c11d5c0a1c4b33bf" gracePeriod=30 Oct 14 07:22:12 crc kubenswrapper[4870]: I1014 07:22:12.047418 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="8c70361b-b549-48b9-92f7-2c8a611c30bd" containerName="nova-metadata-metadata" containerID="cri-o://653ad6afcaad03b604bd34bc66a67a40e5bccc9a6b8caee133a16cf728616a1b" gracePeriod=30 Oct 14 07:22:12 crc kubenswrapper[4870]: I1014 07:22:12.047487 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="67e09a03-b028-4444-8aba-63be9c7e1abd" containerName="nova-scheduler-scheduler" containerID="cri-o://231197d5071a9dbc186c3fce6f80efb27d4af4bc8993ba7cab278415accb4013" gracePeriod=30 Oct 14 07:22:13 crc kubenswrapper[4870]: I1014 07:22:13.051120 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0dba2f3b-4ddd-4a1f-8593-1f98dc4479b1" path="/var/lib/kubelet/pods/0dba2f3b-4ddd-4a1f-8593-1f98dc4479b1/volumes" Oct 14 07:22:13 crc kubenswrapper[4870]: I1014 07:22:13.065693 4870 generic.go:334] "Generic (PLEG): container finished" podID="67e09a03-b028-4444-8aba-63be9c7e1abd" containerID="231197d5071a9dbc186c3fce6f80efb27d4af4bc8993ba7cab278415accb4013" exitCode=0 Oct 14 07:22:13 crc kubenswrapper[4870]: I1014 07:22:13.065773 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"67e09a03-b028-4444-8aba-63be9c7e1abd","Type":"ContainerDied","Data":"231197d5071a9dbc186c3fce6f80efb27d4af4bc8993ba7cab278415accb4013"} Oct 14 07:22:13 crc kubenswrapper[4870]: I1014 07:22:13.071374 4870 generic.go:334] "Generic (PLEG): container finished" podID="8c70361b-b549-48b9-92f7-2c8a611c30bd" containerID="653ad6afcaad03b604bd34bc66a67a40e5bccc9a6b8caee133a16cf728616a1b" exitCode=0 Oct 14 07:22:13 crc kubenswrapper[4870]: I1014 07:22:13.071410 4870 generic.go:334] "Generic (PLEG): container finished" podID="8c70361b-b549-48b9-92f7-2c8a611c30bd" containerID="d859d55bc66f6b1bbc5f156fd01df4786718392416a30495c11d5c0a1c4b33bf" exitCode=143 Oct 14 07:22:13 crc kubenswrapper[4870]: I1014 07:22:13.071433 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"8c70361b-b549-48b9-92f7-2c8a611c30bd","Type":"ContainerDied","Data":"653ad6afcaad03b604bd34bc66a67a40e5bccc9a6b8caee133a16cf728616a1b"} Oct 14 07:22:13 crc kubenswrapper[4870]: I1014 07:22:13.071503 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"8c70361b-b549-48b9-92f7-2c8a611c30bd","Type":"ContainerDied","Data":"d859d55bc66f6b1bbc5f156fd01df4786718392416a30495c11d5c0a1c4b33bf"} Oct 14 07:22:13 crc kubenswrapper[4870]: I1014 07:22:13.357465 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 14 07:22:13 crc kubenswrapper[4870]: I1014 07:22:13.357524 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 14 07:22:13 crc kubenswrapper[4870]: I1014 07:22:13.375913 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 14 07:22:13 crc kubenswrapper[4870]: I1014 07:22:13.464375 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c70361b-b549-48b9-92f7-2c8a611c30bd-config-data\") pod \"8c70361b-b549-48b9-92f7-2c8a611c30bd\" (UID: \"8c70361b-b549-48b9-92f7-2c8a611c30bd\") " Oct 14 07:22:13 crc kubenswrapper[4870]: I1014 07:22:13.464450 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6zx9w\" (UniqueName: \"kubernetes.io/projected/8c70361b-b549-48b9-92f7-2c8a611c30bd-kube-api-access-6zx9w\") pod \"8c70361b-b549-48b9-92f7-2c8a611c30bd\" (UID: \"8c70361b-b549-48b9-92f7-2c8a611c30bd\") " Oct 14 07:22:13 crc kubenswrapper[4870]: I1014 07:22:13.464682 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/8c70361b-b549-48b9-92f7-2c8a611c30bd-nova-metadata-tls-certs\") pod \"8c70361b-b549-48b9-92f7-2c8a611c30bd\" (UID: \"8c70361b-b549-48b9-92f7-2c8a611c30bd\") " Oct 14 07:22:13 crc kubenswrapper[4870]: I1014 07:22:13.464746 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c70361b-b549-48b9-92f7-2c8a611c30bd-combined-ca-bundle\") pod \"8c70361b-b549-48b9-92f7-2c8a611c30bd\" (UID: \"8c70361b-b549-48b9-92f7-2c8a611c30bd\") " Oct 14 07:22:13 crc kubenswrapper[4870]: I1014 07:22:13.464794 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8c70361b-b549-48b9-92f7-2c8a611c30bd-logs\") pod \"8c70361b-b549-48b9-92f7-2c8a611c30bd\" (UID: \"8c70361b-b549-48b9-92f7-2c8a611c30bd\") " Oct 14 07:22:13 crc kubenswrapper[4870]: I1014 07:22:13.465806 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8c70361b-b549-48b9-92f7-2c8a611c30bd-logs" (OuterVolumeSpecName: "logs") pod "8c70361b-b549-48b9-92f7-2c8a611c30bd" (UID: "8c70361b-b549-48b9-92f7-2c8a611c30bd"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:22:13 crc kubenswrapper[4870]: I1014 07:22:13.503756 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8c70361b-b549-48b9-92f7-2c8a611c30bd-kube-api-access-6zx9w" (OuterVolumeSpecName: "kube-api-access-6zx9w") pod "8c70361b-b549-48b9-92f7-2c8a611c30bd" (UID: "8c70361b-b549-48b9-92f7-2c8a611c30bd"). InnerVolumeSpecName "kube-api-access-6zx9w". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:22:13 crc kubenswrapper[4870]: I1014 07:22:13.513946 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c70361b-b549-48b9-92f7-2c8a611c30bd-config-data" (OuterVolumeSpecName: "config-data") pod "8c70361b-b549-48b9-92f7-2c8a611c30bd" (UID: "8c70361b-b549-48b9-92f7-2c8a611c30bd"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:22:13 crc kubenswrapper[4870]: I1014 07:22:13.534636 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c70361b-b549-48b9-92f7-2c8a611c30bd-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "8c70361b-b549-48b9-92f7-2c8a611c30bd" (UID: "8c70361b-b549-48b9-92f7-2c8a611c30bd"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:22:13 crc kubenswrapper[4870]: I1014 07:22:13.537510 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c70361b-b549-48b9-92f7-2c8a611c30bd-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8c70361b-b549-48b9-92f7-2c8a611c30bd" (UID: "8c70361b-b549-48b9-92f7-2c8a611c30bd"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:22:13 crc kubenswrapper[4870]: I1014 07:22:13.566376 4870 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/8c70361b-b549-48b9-92f7-2c8a611c30bd-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 14 07:22:13 crc kubenswrapper[4870]: I1014 07:22:13.566405 4870 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c70361b-b549-48b9-92f7-2c8a611c30bd-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 07:22:13 crc kubenswrapper[4870]: I1014 07:22:13.566416 4870 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8c70361b-b549-48b9-92f7-2c8a611c30bd-logs\") on node \"crc\" DevicePath \"\"" Oct 14 07:22:13 crc kubenswrapper[4870]: I1014 07:22:13.566424 4870 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c70361b-b549-48b9-92f7-2c8a611c30bd-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 07:22:13 crc kubenswrapper[4870]: I1014 07:22:13.566433 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6zx9w\" (UniqueName: \"kubernetes.io/projected/8c70361b-b549-48b9-92f7-2c8a611c30bd-kube-api-access-6zx9w\") on node \"crc\" DevicePath \"\"" Oct 14 07:22:13 crc kubenswrapper[4870]: I1014 07:22:13.645217 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 14 07:22:13 crc kubenswrapper[4870]: I1014 07:22:13.770367 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67e09a03-b028-4444-8aba-63be9c7e1abd-config-data\") pod \"67e09a03-b028-4444-8aba-63be9c7e1abd\" (UID: \"67e09a03-b028-4444-8aba-63be9c7e1abd\") " Oct 14 07:22:13 crc kubenswrapper[4870]: I1014 07:22:13.770716 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67e09a03-b028-4444-8aba-63be9c7e1abd-combined-ca-bundle\") pod \"67e09a03-b028-4444-8aba-63be9c7e1abd\" (UID: \"67e09a03-b028-4444-8aba-63be9c7e1abd\") " Oct 14 07:22:13 crc kubenswrapper[4870]: I1014 07:22:13.770803 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tqmf6\" (UniqueName: \"kubernetes.io/projected/67e09a03-b028-4444-8aba-63be9c7e1abd-kube-api-access-tqmf6\") pod \"67e09a03-b028-4444-8aba-63be9c7e1abd\" (UID: \"67e09a03-b028-4444-8aba-63be9c7e1abd\") " Oct 14 07:22:13 crc kubenswrapper[4870]: I1014 07:22:13.774736 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/67e09a03-b028-4444-8aba-63be9c7e1abd-kube-api-access-tqmf6" (OuterVolumeSpecName: "kube-api-access-tqmf6") pod "67e09a03-b028-4444-8aba-63be9c7e1abd" (UID: "67e09a03-b028-4444-8aba-63be9c7e1abd"). InnerVolumeSpecName "kube-api-access-tqmf6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:22:13 crc kubenswrapper[4870]: I1014 07:22:13.794295 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/67e09a03-b028-4444-8aba-63be9c7e1abd-config-data" (OuterVolumeSpecName: "config-data") pod "67e09a03-b028-4444-8aba-63be9c7e1abd" (UID: "67e09a03-b028-4444-8aba-63be9c7e1abd"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:22:13 crc kubenswrapper[4870]: I1014 07:22:13.806220 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/67e09a03-b028-4444-8aba-63be9c7e1abd-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "67e09a03-b028-4444-8aba-63be9c7e1abd" (UID: "67e09a03-b028-4444-8aba-63be9c7e1abd"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:22:13 crc kubenswrapper[4870]: I1014 07:22:13.873135 4870 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67e09a03-b028-4444-8aba-63be9c7e1abd-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 07:22:13 crc kubenswrapper[4870]: I1014 07:22:13.873174 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tqmf6\" (UniqueName: \"kubernetes.io/projected/67e09a03-b028-4444-8aba-63be9c7e1abd-kube-api-access-tqmf6\") on node \"crc\" DevicePath \"\"" Oct 14 07:22:13 crc kubenswrapper[4870]: I1014 07:22:13.873190 4870 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67e09a03-b028-4444-8aba-63be9c7e1abd-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 07:22:13 crc kubenswrapper[4870]: I1014 07:22:13.938780 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Oct 14 07:22:14 crc kubenswrapper[4870]: I1014 07:22:14.086580 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"8c70361b-b549-48b9-92f7-2c8a611c30bd","Type":"ContainerDied","Data":"71bfc80fe07fdfc3a8eda33139fb3cdd01f7c27904dceabe861c504ae7714ea8"} Oct 14 07:22:14 crc kubenswrapper[4870]: I1014 07:22:14.086642 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 14 07:22:14 crc kubenswrapper[4870]: I1014 07:22:14.086664 4870 scope.go:117] "RemoveContainer" containerID="653ad6afcaad03b604bd34bc66a67a40e5bccc9a6b8caee133a16cf728616a1b" Oct 14 07:22:14 crc kubenswrapper[4870]: I1014 07:22:14.088256 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"67e09a03-b028-4444-8aba-63be9c7e1abd","Type":"ContainerDied","Data":"64736b327ed16b55d75fbb107d9283e261cb66851f8f552b53a406fa64433954"} Oct 14 07:22:14 crc kubenswrapper[4870]: I1014 07:22:14.088277 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 14 07:22:14 crc kubenswrapper[4870]: I1014 07:22:14.113331 4870 scope.go:117] "RemoveContainer" containerID="d859d55bc66f6b1bbc5f156fd01df4786718392416a30495c11d5c0a1c4b33bf" Oct 14 07:22:14 crc kubenswrapper[4870]: I1014 07:22:14.124062 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 14 07:22:14 crc kubenswrapper[4870]: I1014 07:22:14.133396 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 14 07:22:14 crc kubenswrapper[4870]: I1014 07:22:14.146613 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 14 07:22:14 crc kubenswrapper[4870]: I1014 07:22:14.148327 4870 scope.go:117] "RemoveContainer" containerID="231197d5071a9dbc186c3fce6f80efb27d4af4bc8993ba7cab278415accb4013" Oct 14 07:22:14 crc kubenswrapper[4870]: I1014 07:22:14.155386 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Oct 14 07:22:14 crc kubenswrapper[4870]: I1014 07:22:14.165189 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 14 07:22:14 crc kubenswrapper[4870]: E1014 07:22:14.165785 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0dba2f3b-4ddd-4a1f-8593-1f98dc4479b1" containerName="init" Oct 14 07:22:14 crc kubenswrapper[4870]: I1014 07:22:14.165809 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="0dba2f3b-4ddd-4a1f-8593-1f98dc4479b1" containerName="init" Oct 14 07:22:14 crc kubenswrapper[4870]: E1014 07:22:14.165826 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67e09a03-b028-4444-8aba-63be9c7e1abd" containerName="nova-scheduler-scheduler" Oct 14 07:22:14 crc kubenswrapper[4870]: I1014 07:22:14.165833 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="67e09a03-b028-4444-8aba-63be9c7e1abd" containerName="nova-scheduler-scheduler" Oct 14 07:22:14 crc kubenswrapper[4870]: E1014 07:22:14.165864 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d147118e-e08e-4c64-b157-da3ca3afb6d7" containerName="nova-manage" Oct 14 07:22:14 crc kubenswrapper[4870]: I1014 07:22:14.165871 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="d147118e-e08e-4c64-b157-da3ca3afb6d7" containerName="nova-manage" Oct 14 07:22:14 crc kubenswrapper[4870]: E1014 07:22:14.165890 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0dba2f3b-4ddd-4a1f-8593-1f98dc4479b1" containerName="dnsmasq-dns" Oct 14 07:22:14 crc kubenswrapper[4870]: I1014 07:22:14.165898 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="0dba2f3b-4ddd-4a1f-8593-1f98dc4479b1" containerName="dnsmasq-dns" Oct 14 07:22:14 crc kubenswrapper[4870]: E1014 07:22:14.165921 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c70361b-b549-48b9-92f7-2c8a611c30bd" containerName="nova-metadata-log" Oct 14 07:22:14 crc kubenswrapper[4870]: I1014 07:22:14.165928 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c70361b-b549-48b9-92f7-2c8a611c30bd" containerName="nova-metadata-log" Oct 14 07:22:14 crc kubenswrapper[4870]: E1014 07:22:14.165944 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c70361b-b549-48b9-92f7-2c8a611c30bd" containerName="nova-metadata-metadata" Oct 14 07:22:14 crc kubenswrapper[4870]: I1014 07:22:14.165952 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c70361b-b549-48b9-92f7-2c8a611c30bd" containerName="nova-metadata-metadata" Oct 14 07:22:14 crc kubenswrapper[4870]: I1014 07:22:14.166179 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="0dba2f3b-4ddd-4a1f-8593-1f98dc4479b1" containerName="dnsmasq-dns" Oct 14 07:22:14 crc kubenswrapper[4870]: I1014 07:22:14.166196 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="8c70361b-b549-48b9-92f7-2c8a611c30bd" containerName="nova-metadata-log" Oct 14 07:22:14 crc kubenswrapper[4870]: I1014 07:22:14.166216 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="d147118e-e08e-4c64-b157-da3ca3afb6d7" containerName="nova-manage" Oct 14 07:22:14 crc kubenswrapper[4870]: I1014 07:22:14.166235 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="67e09a03-b028-4444-8aba-63be9c7e1abd" containerName="nova-scheduler-scheduler" Oct 14 07:22:14 crc kubenswrapper[4870]: I1014 07:22:14.166253 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="8c70361b-b549-48b9-92f7-2c8a611c30bd" containerName="nova-metadata-metadata" Oct 14 07:22:14 crc kubenswrapper[4870]: I1014 07:22:14.168937 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 14 07:22:14 crc kubenswrapper[4870]: I1014 07:22:14.172646 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 14 07:22:14 crc kubenswrapper[4870]: I1014 07:22:14.195791 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Oct 14 07:22:14 crc kubenswrapper[4870]: I1014 07:22:14.212319 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 14 07:22:14 crc kubenswrapper[4870]: I1014 07:22:14.214403 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 14 07:22:14 crc kubenswrapper[4870]: I1014 07:22:14.223951 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 14 07:22:14 crc kubenswrapper[4870]: I1014 07:22:14.239522 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 14 07:22:14 crc kubenswrapper[4870]: I1014 07:22:14.303452 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-22wcl\" (UniqueName: \"kubernetes.io/projected/4a361161-c100-49cb-a4ee-19563b724033-kube-api-access-22wcl\") pod \"nova-metadata-0\" (UID: \"4a361161-c100-49cb-a4ee-19563b724033\") " pod="openstack/nova-metadata-0" Oct 14 07:22:14 crc kubenswrapper[4870]: I1014 07:22:14.303568 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/4a361161-c100-49cb-a4ee-19563b724033-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"4a361161-c100-49cb-a4ee-19563b724033\") " pod="openstack/nova-metadata-0" Oct 14 07:22:14 crc kubenswrapper[4870]: I1014 07:22:14.303627 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a361161-c100-49cb-a4ee-19563b724033-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"4a361161-c100-49cb-a4ee-19563b724033\") " pod="openstack/nova-metadata-0" Oct 14 07:22:14 crc kubenswrapper[4870]: I1014 07:22:14.303657 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4a361161-c100-49cb-a4ee-19563b724033-logs\") pod \"nova-metadata-0\" (UID: \"4a361161-c100-49cb-a4ee-19563b724033\") " pod="openstack/nova-metadata-0" Oct 14 07:22:14 crc kubenswrapper[4870]: I1014 07:22:14.303681 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4a361161-c100-49cb-a4ee-19563b724033-config-data\") pod \"nova-metadata-0\" (UID: \"4a361161-c100-49cb-a4ee-19563b724033\") " pod="openstack/nova-metadata-0" Oct 14 07:22:14 crc kubenswrapper[4870]: I1014 07:22:14.322603 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 14 07:22:14 crc kubenswrapper[4870]: I1014 07:22:14.405195 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ffe53321-c66f-4b32-b18d-18bc483f923b-config-data\") pod \"nova-scheduler-0\" (UID: \"ffe53321-c66f-4b32-b18d-18bc483f923b\") " pod="openstack/nova-scheduler-0" Oct 14 07:22:14 crc kubenswrapper[4870]: I1014 07:22:14.405631 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-22wcl\" (UniqueName: \"kubernetes.io/projected/4a361161-c100-49cb-a4ee-19563b724033-kube-api-access-22wcl\") pod \"nova-metadata-0\" (UID: \"4a361161-c100-49cb-a4ee-19563b724033\") " pod="openstack/nova-metadata-0" Oct 14 07:22:14 crc kubenswrapper[4870]: I1014 07:22:14.405664 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ffe53321-c66f-4b32-b18d-18bc483f923b-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"ffe53321-c66f-4b32-b18d-18bc483f923b\") " pod="openstack/nova-scheduler-0" Oct 14 07:22:14 crc kubenswrapper[4870]: I1014 07:22:14.405717 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sf6ps\" (UniqueName: \"kubernetes.io/projected/ffe53321-c66f-4b32-b18d-18bc483f923b-kube-api-access-sf6ps\") pod \"nova-scheduler-0\" (UID: \"ffe53321-c66f-4b32-b18d-18bc483f923b\") " pod="openstack/nova-scheduler-0" Oct 14 07:22:14 crc kubenswrapper[4870]: I1014 07:22:14.405809 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/4a361161-c100-49cb-a4ee-19563b724033-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"4a361161-c100-49cb-a4ee-19563b724033\") " pod="openstack/nova-metadata-0" Oct 14 07:22:14 crc kubenswrapper[4870]: I1014 07:22:14.405863 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a361161-c100-49cb-a4ee-19563b724033-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"4a361161-c100-49cb-a4ee-19563b724033\") " pod="openstack/nova-metadata-0" Oct 14 07:22:14 crc kubenswrapper[4870]: I1014 07:22:14.405900 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4a361161-c100-49cb-a4ee-19563b724033-logs\") pod \"nova-metadata-0\" (UID: \"4a361161-c100-49cb-a4ee-19563b724033\") " pod="openstack/nova-metadata-0" Oct 14 07:22:14 crc kubenswrapper[4870]: I1014 07:22:14.405927 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4a361161-c100-49cb-a4ee-19563b724033-config-data\") pod \"nova-metadata-0\" (UID: \"4a361161-c100-49cb-a4ee-19563b724033\") " pod="openstack/nova-metadata-0" Oct 14 07:22:14 crc kubenswrapper[4870]: I1014 07:22:14.406323 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4a361161-c100-49cb-a4ee-19563b724033-logs\") pod \"nova-metadata-0\" (UID: \"4a361161-c100-49cb-a4ee-19563b724033\") " pod="openstack/nova-metadata-0" Oct 14 07:22:14 crc kubenswrapper[4870]: I1014 07:22:14.411001 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/4a361161-c100-49cb-a4ee-19563b724033-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"4a361161-c100-49cb-a4ee-19563b724033\") " pod="openstack/nova-metadata-0" Oct 14 07:22:14 crc kubenswrapper[4870]: I1014 07:22:14.411465 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a361161-c100-49cb-a4ee-19563b724033-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"4a361161-c100-49cb-a4ee-19563b724033\") " pod="openstack/nova-metadata-0" Oct 14 07:22:14 crc kubenswrapper[4870]: I1014 07:22:14.411904 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4a361161-c100-49cb-a4ee-19563b724033-config-data\") pod \"nova-metadata-0\" (UID: \"4a361161-c100-49cb-a4ee-19563b724033\") " pod="openstack/nova-metadata-0" Oct 14 07:22:14 crc kubenswrapper[4870]: I1014 07:22:14.423089 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-22wcl\" (UniqueName: \"kubernetes.io/projected/4a361161-c100-49cb-a4ee-19563b724033-kube-api-access-22wcl\") pod \"nova-metadata-0\" (UID: \"4a361161-c100-49cb-a4ee-19563b724033\") " pod="openstack/nova-metadata-0" Oct 14 07:22:14 crc kubenswrapper[4870]: I1014 07:22:14.509026 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ffe53321-c66f-4b32-b18d-18bc483f923b-config-data\") pod \"nova-scheduler-0\" (UID: \"ffe53321-c66f-4b32-b18d-18bc483f923b\") " pod="openstack/nova-scheduler-0" Oct 14 07:22:14 crc kubenswrapper[4870]: I1014 07:22:14.509176 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ffe53321-c66f-4b32-b18d-18bc483f923b-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"ffe53321-c66f-4b32-b18d-18bc483f923b\") " pod="openstack/nova-scheduler-0" Oct 14 07:22:14 crc kubenswrapper[4870]: I1014 07:22:14.509260 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sf6ps\" (UniqueName: \"kubernetes.io/projected/ffe53321-c66f-4b32-b18d-18bc483f923b-kube-api-access-sf6ps\") pod \"nova-scheduler-0\" (UID: \"ffe53321-c66f-4b32-b18d-18bc483f923b\") " pod="openstack/nova-scheduler-0" Oct 14 07:22:14 crc kubenswrapper[4870]: I1014 07:22:14.511698 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 14 07:22:14 crc kubenswrapper[4870]: I1014 07:22:14.514296 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ffe53321-c66f-4b32-b18d-18bc483f923b-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"ffe53321-c66f-4b32-b18d-18bc483f923b\") " pod="openstack/nova-scheduler-0" Oct 14 07:22:14 crc kubenswrapper[4870]: I1014 07:22:14.518245 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ffe53321-c66f-4b32-b18d-18bc483f923b-config-data\") pod \"nova-scheduler-0\" (UID: \"ffe53321-c66f-4b32-b18d-18bc483f923b\") " pod="openstack/nova-scheduler-0" Oct 14 07:22:14 crc kubenswrapper[4870]: I1014 07:22:14.526749 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sf6ps\" (UniqueName: \"kubernetes.io/projected/ffe53321-c66f-4b32-b18d-18bc483f923b-kube-api-access-sf6ps\") pod \"nova-scheduler-0\" (UID: \"ffe53321-c66f-4b32-b18d-18bc483f923b\") " pod="openstack/nova-scheduler-0" Oct 14 07:22:14 crc kubenswrapper[4870]: I1014 07:22:14.636214 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 14 07:22:14 crc kubenswrapper[4870]: I1014 07:22:14.956213 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 14 07:22:15 crc kubenswrapper[4870]: I1014 07:22:15.069849 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="67e09a03-b028-4444-8aba-63be9c7e1abd" path="/var/lib/kubelet/pods/67e09a03-b028-4444-8aba-63be9c7e1abd/volumes" Oct 14 07:22:15 crc kubenswrapper[4870]: I1014 07:22:15.071007 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8c70361b-b549-48b9-92f7-2c8a611c30bd" path="/var/lib/kubelet/pods/8c70361b-b549-48b9-92f7-2c8a611c30bd/volumes" Oct 14 07:22:15 crc kubenswrapper[4870]: I1014 07:22:15.100531 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"4a361161-c100-49cb-a4ee-19563b724033","Type":"ContainerStarted","Data":"7ba3a53d0ecb65545d234f202c569d57bc35424c5757f7dcb97fdefa9c3c5700"} Oct 14 07:22:15 crc kubenswrapper[4870]: W1014 07:22:15.106155 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podffe53321_c66f_4b32_b18d_18bc483f923b.slice/crio-5e80cf2ed5aa73531137db469a8d31176086b6df05b2af4503656e893daa0f4a WatchSource:0}: Error finding container 5e80cf2ed5aa73531137db469a8d31176086b6df05b2af4503656e893daa0f4a: Status 404 returned error can't find the container with id 5e80cf2ed5aa73531137db469a8d31176086b6df05b2af4503656e893daa0f4a Oct 14 07:22:15 crc kubenswrapper[4870]: I1014 07:22:15.114237 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 14 07:22:16 crc kubenswrapper[4870]: I1014 07:22:16.127920 4870 generic.go:334] "Generic (PLEG): container finished" podID="cd44bae6-b5d1-4b08-88de-7c5271f54ed3" containerID="d311512ab3d18b4243743bfbfb8a259533302afd01cbbcfda23ec6a678817840" exitCode=0 Oct 14 07:22:16 crc kubenswrapper[4870]: I1014 07:22:16.128157 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"cd44bae6-b5d1-4b08-88de-7c5271f54ed3","Type":"ContainerDied","Data":"d311512ab3d18b4243743bfbfb8a259533302afd01cbbcfda23ec6a678817840"} Oct 14 07:22:16 crc kubenswrapper[4870]: I1014 07:22:16.128919 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"cd44bae6-b5d1-4b08-88de-7c5271f54ed3","Type":"ContainerDied","Data":"6dee636cd479af70c001abe0df98e5cba74b39c63236703b96a2e4b1ac1f3358"} Oct 14 07:22:16 crc kubenswrapper[4870]: I1014 07:22:16.128933 4870 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6dee636cd479af70c001abe0df98e5cba74b39c63236703b96a2e4b1ac1f3358" Oct 14 07:22:16 crc kubenswrapper[4870]: I1014 07:22:16.131503 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"4a361161-c100-49cb-a4ee-19563b724033","Type":"ContainerStarted","Data":"98556816573a1ddbc56f6c6dcdd861e5bfd4e061a6c46d5d705c33318916ee5c"} Oct 14 07:22:16 crc kubenswrapper[4870]: I1014 07:22:16.131565 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"4a361161-c100-49cb-a4ee-19563b724033","Type":"ContainerStarted","Data":"1a08f0b4fd874fefc01d80d5ab91d598f23dd77047d80c172612080dc2670dbc"} Oct 14 07:22:16 crc kubenswrapper[4870]: I1014 07:22:16.133210 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"ffe53321-c66f-4b32-b18d-18bc483f923b","Type":"ContainerStarted","Data":"56e70d8d8996b806615045a193a6672b63fec50f57799575de3be2115113662c"} Oct 14 07:22:16 crc kubenswrapper[4870]: I1014 07:22:16.133242 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"ffe53321-c66f-4b32-b18d-18bc483f923b","Type":"ContainerStarted","Data":"5e80cf2ed5aa73531137db469a8d31176086b6df05b2af4503656e893daa0f4a"} Oct 14 07:22:16 crc kubenswrapper[4870]: I1014 07:22:16.162560 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.162540376 podStartE2EDuration="2.162540376s" podCreationTimestamp="2025-10-14 07:22:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:22:16.153983557 +0000 UTC m=+1271.851343918" watchObservedRunningTime="2025-10-14 07:22:16.162540376 +0000 UTC m=+1271.859900747" Oct 14 07:22:16 crc kubenswrapper[4870]: I1014 07:22:16.188812 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.18878766 podStartE2EDuration="2.18878766s" podCreationTimestamp="2025-10-14 07:22:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:22:16.181844418 +0000 UTC m=+1271.879204809" watchObservedRunningTime="2025-10-14 07:22:16.18878766 +0000 UTC m=+1271.886148031" Oct 14 07:22:16 crc kubenswrapper[4870]: I1014 07:22:16.197595 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 14 07:22:16 crc kubenswrapper[4870]: I1014 07:22:16.350211 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd44bae6-b5d1-4b08-88de-7c5271f54ed3-combined-ca-bundle\") pod \"cd44bae6-b5d1-4b08-88de-7c5271f54ed3\" (UID: \"cd44bae6-b5d1-4b08-88de-7c5271f54ed3\") " Oct 14 07:22:16 crc kubenswrapper[4870]: I1014 07:22:16.350354 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cd44bae6-b5d1-4b08-88de-7c5271f54ed3-logs\") pod \"cd44bae6-b5d1-4b08-88de-7c5271f54ed3\" (UID: \"cd44bae6-b5d1-4b08-88de-7c5271f54ed3\") " Oct 14 07:22:16 crc kubenswrapper[4870]: I1014 07:22:16.350645 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b9xkb\" (UniqueName: \"kubernetes.io/projected/cd44bae6-b5d1-4b08-88de-7c5271f54ed3-kube-api-access-b9xkb\") pod \"cd44bae6-b5d1-4b08-88de-7c5271f54ed3\" (UID: \"cd44bae6-b5d1-4b08-88de-7c5271f54ed3\") " Oct 14 07:22:16 crc kubenswrapper[4870]: I1014 07:22:16.350738 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd44bae6-b5d1-4b08-88de-7c5271f54ed3-config-data\") pod \"cd44bae6-b5d1-4b08-88de-7c5271f54ed3\" (UID: \"cd44bae6-b5d1-4b08-88de-7c5271f54ed3\") " Oct 14 07:22:16 crc kubenswrapper[4870]: I1014 07:22:16.350831 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cd44bae6-b5d1-4b08-88de-7c5271f54ed3-logs" (OuterVolumeSpecName: "logs") pod "cd44bae6-b5d1-4b08-88de-7c5271f54ed3" (UID: "cd44bae6-b5d1-4b08-88de-7c5271f54ed3"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:22:16 crc kubenswrapper[4870]: I1014 07:22:16.351382 4870 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cd44bae6-b5d1-4b08-88de-7c5271f54ed3-logs\") on node \"crc\" DevicePath \"\"" Oct 14 07:22:16 crc kubenswrapper[4870]: I1014 07:22:16.367346 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd44bae6-b5d1-4b08-88de-7c5271f54ed3-kube-api-access-b9xkb" (OuterVolumeSpecName: "kube-api-access-b9xkb") pod "cd44bae6-b5d1-4b08-88de-7c5271f54ed3" (UID: "cd44bae6-b5d1-4b08-88de-7c5271f54ed3"). InnerVolumeSpecName "kube-api-access-b9xkb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:22:16 crc kubenswrapper[4870]: I1014 07:22:16.393258 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cd44bae6-b5d1-4b08-88de-7c5271f54ed3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cd44bae6-b5d1-4b08-88de-7c5271f54ed3" (UID: "cd44bae6-b5d1-4b08-88de-7c5271f54ed3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:22:16 crc kubenswrapper[4870]: I1014 07:22:16.395407 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cd44bae6-b5d1-4b08-88de-7c5271f54ed3-config-data" (OuterVolumeSpecName: "config-data") pod "cd44bae6-b5d1-4b08-88de-7c5271f54ed3" (UID: "cd44bae6-b5d1-4b08-88de-7c5271f54ed3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:22:16 crc kubenswrapper[4870]: I1014 07:22:16.452750 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b9xkb\" (UniqueName: \"kubernetes.io/projected/cd44bae6-b5d1-4b08-88de-7c5271f54ed3-kube-api-access-b9xkb\") on node \"crc\" DevicePath \"\"" Oct 14 07:22:16 crc kubenswrapper[4870]: I1014 07:22:16.452788 4870 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd44bae6-b5d1-4b08-88de-7c5271f54ed3-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 07:22:16 crc kubenswrapper[4870]: I1014 07:22:16.452801 4870 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd44bae6-b5d1-4b08-88de-7c5271f54ed3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 07:22:17 crc kubenswrapper[4870]: I1014 07:22:17.142970 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 14 07:22:17 crc kubenswrapper[4870]: I1014 07:22:17.172005 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 14 07:22:17 crc kubenswrapper[4870]: I1014 07:22:17.183748 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 14 07:22:17 crc kubenswrapper[4870]: I1014 07:22:17.203275 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 14 07:22:17 crc kubenswrapper[4870]: E1014 07:22:17.203960 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd44bae6-b5d1-4b08-88de-7c5271f54ed3" containerName="nova-api-log" Oct 14 07:22:17 crc kubenswrapper[4870]: I1014 07:22:17.203989 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd44bae6-b5d1-4b08-88de-7c5271f54ed3" containerName="nova-api-log" Oct 14 07:22:17 crc kubenswrapper[4870]: E1014 07:22:17.204037 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd44bae6-b5d1-4b08-88de-7c5271f54ed3" containerName="nova-api-api" Oct 14 07:22:17 crc kubenswrapper[4870]: I1014 07:22:17.204053 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd44bae6-b5d1-4b08-88de-7c5271f54ed3" containerName="nova-api-api" Oct 14 07:22:17 crc kubenswrapper[4870]: I1014 07:22:17.204342 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="cd44bae6-b5d1-4b08-88de-7c5271f54ed3" containerName="nova-api-log" Oct 14 07:22:17 crc kubenswrapper[4870]: I1014 07:22:17.204385 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="cd44bae6-b5d1-4b08-88de-7c5271f54ed3" containerName="nova-api-api" Oct 14 07:22:17 crc kubenswrapper[4870]: I1014 07:22:17.205947 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 14 07:22:17 crc kubenswrapper[4870]: I1014 07:22:17.210568 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 14 07:22:17 crc kubenswrapper[4870]: I1014 07:22:17.221125 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 14 07:22:17 crc kubenswrapper[4870]: I1014 07:22:17.377982 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2688cfd1-fbb5-49bb-b13e-a140ef4ed017-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"2688cfd1-fbb5-49bb-b13e-a140ef4ed017\") " pod="openstack/nova-api-0" Oct 14 07:22:17 crc kubenswrapper[4870]: I1014 07:22:17.378088 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2688cfd1-fbb5-49bb-b13e-a140ef4ed017-logs\") pod \"nova-api-0\" (UID: \"2688cfd1-fbb5-49bb-b13e-a140ef4ed017\") " pod="openstack/nova-api-0" Oct 14 07:22:17 crc kubenswrapper[4870]: I1014 07:22:17.378169 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kqtt7\" (UniqueName: \"kubernetes.io/projected/2688cfd1-fbb5-49bb-b13e-a140ef4ed017-kube-api-access-kqtt7\") pod \"nova-api-0\" (UID: \"2688cfd1-fbb5-49bb-b13e-a140ef4ed017\") " pod="openstack/nova-api-0" Oct 14 07:22:17 crc kubenswrapper[4870]: I1014 07:22:17.378248 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2688cfd1-fbb5-49bb-b13e-a140ef4ed017-config-data\") pod \"nova-api-0\" (UID: \"2688cfd1-fbb5-49bb-b13e-a140ef4ed017\") " pod="openstack/nova-api-0" Oct 14 07:22:17 crc kubenswrapper[4870]: I1014 07:22:17.479817 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2688cfd1-fbb5-49bb-b13e-a140ef4ed017-config-data\") pod \"nova-api-0\" (UID: \"2688cfd1-fbb5-49bb-b13e-a140ef4ed017\") " pod="openstack/nova-api-0" Oct 14 07:22:17 crc kubenswrapper[4870]: I1014 07:22:17.480026 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2688cfd1-fbb5-49bb-b13e-a140ef4ed017-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"2688cfd1-fbb5-49bb-b13e-a140ef4ed017\") " pod="openstack/nova-api-0" Oct 14 07:22:17 crc kubenswrapper[4870]: I1014 07:22:17.480056 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2688cfd1-fbb5-49bb-b13e-a140ef4ed017-logs\") pod \"nova-api-0\" (UID: \"2688cfd1-fbb5-49bb-b13e-a140ef4ed017\") " pod="openstack/nova-api-0" Oct 14 07:22:17 crc kubenswrapper[4870]: I1014 07:22:17.480082 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kqtt7\" (UniqueName: \"kubernetes.io/projected/2688cfd1-fbb5-49bb-b13e-a140ef4ed017-kube-api-access-kqtt7\") pod \"nova-api-0\" (UID: \"2688cfd1-fbb5-49bb-b13e-a140ef4ed017\") " pod="openstack/nova-api-0" Oct 14 07:22:17 crc kubenswrapper[4870]: I1014 07:22:17.480714 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2688cfd1-fbb5-49bb-b13e-a140ef4ed017-logs\") pod \"nova-api-0\" (UID: \"2688cfd1-fbb5-49bb-b13e-a140ef4ed017\") " pod="openstack/nova-api-0" Oct 14 07:22:17 crc kubenswrapper[4870]: I1014 07:22:17.488696 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2688cfd1-fbb5-49bb-b13e-a140ef4ed017-config-data\") pod \"nova-api-0\" (UID: \"2688cfd1-fbb5-49bb-b13e-a140ef4ed017\") " pod="openstack/nova-api-0" Oct 14 07:22:17 crc kubenswrapper[4870]: I1014 07:22:17.489478 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2688cfd1-fbb5-49bb-b13e-a140ef4ed017-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"2688cfd1-fbb5-49bb-b13e-a140ef4ed017\") " pod="openstack/nova-api-0" Oct 14 07:22:17 crc kubenswrapper[4870]: I1014 07:22:17.502607 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kqtt7\" (UniqueName: \"kubernetes.io/projected/2688cfd1-fbb5-49bb-b13e-a140ef4ed017-kube-api-access-kqtt7\") pod \"nova-api-0\" (UID: \"2688cfd1-fbb5-49bb-b13e-a140ef4ed017\") " pod="openstack/nova-api-0" Oct 14 07:22:17 crc kubenswrapper[4870]: I1014 07:22:17.535652 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 14 07:22:18 crc kubenswrapper[4870]: I1014 07:22:18.082089 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 14 07:22:18 crc kubenswrapper[4870]: I1014 07:22:18.152546 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2688cfd1-fbb5-49bb-b13e-a140ef4ed017","Type":"ContainerStarted","Data":"1418e6f00a15bb66ee2a7b4cbc6499a1dd4796549c8faac82f925a9d250e35a2"} Oct 14 07:22:18 crc kubenswrapper[4870]: I1014 07:22:18.201891 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 14 07:22:18 crc kubenswrapper[4870]: I1014 07:22:18.202100 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="b2d073e2-e7bb-44c9-b074-c39afeae16b8" containerName="kube-state-metrics" containerID="cri-o://ffd73f77d45e4b6cd97b896c63da4c555a8d71621706a5dd9aa9fade449b5ce5" gracePeriod=30 Oct 14 07:22:18 crc kubenswrapper[4870]: I1014 07:22:18.763154 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 14 07:22:18 crc kubenswrapper[4870]: I1014 07:22:18.905055 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nmlrt\" (UniqueName: \"kubernetes.io/projected/b2d073e2-e7bb-44c9-b074-c39afeae16b8-kube-api-access-nmlrt\") pod \"b2d073e2-e7bb-44c9-b074-c39afeae16b8\" (UID: \"b2d073e2-e7bb-44c9-b074-c39afeae16b8\") " Oct 14 07:22:18 crc kubenswrapper[4870]: I1014 07:22:18.921860 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b2d073e2-e7bb-44c9-b074-c39afeae16b8-kube-api-access-nmlrt" (OuterVolumeSpecName: "kube-api-access-nmlrt") pod "b2d073e2-e7bb-44c9-b074-c39afeae16b8" (UID: "b2d073e2-e7bb-44c9-b074-c39afeae16b8"). InnerVolumeSpecName "kube-api-access-nmlrt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:22:19 crc kubenswrapper[4870]: I1014 07:22:19.007744 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nmlrt\" (UniqueName: \"kubernetes.io/projected/b2d073e2-e7bb-44c9-b074-c39afeae16b8-kube-api-access-nmlrt\") on node \"crc\" DevicePath \"\"" Oct 14 07:22:19 crc kubenswrapper[4870]: I1014 07:22:19.047167 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd44bae6-b5d1-4b08-88de-7c5271f54ed3" path="/var/lib/kubelet/pods/cd44bae6-b5d1-4b08-88de-7c5271f54ed3/volumes" Oct 14 07:22:19 crc kubenswrapper[4870]: I1014 07:22:19.166391 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2688cfd1-fbb5-49bb-b13e-a140ef4ed017","Type":"ContainerStarted","Data":"7a6eec2a77f8c1ad24c5475d626f8ac6b4f062c417e46b252b6ad0d2df13bdf9"} Oct 14 07:22:19 crc kubenswrapper[4870]: I1014 07:22:19.166458 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2688cfd1-fbb5-49bb-b13e-a140ef4ed017","Type":"ContainerStarted","Data":"40a16de7c474abe785156f022176336a66c23c1c1750e3f876fdd77eea54ecfa"} Oct 14 07:22:19 crc kubenswrapper[4870]: I1014 07:22:19.168994 4870 generic.go:334] "Generic (PLEG): container finished" podID="3b265c7a-3909-4fd6-9981-c1723fbc1e02" containerID="ce1d38864bb409e404978e6b3a0f6bd89d3dba826ce7be995022fe4e0ac9b188" exitCode=0 Oct 14 07:22:19 crc kubenswrapper[4870]: I1014 07:22:19.169025 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-x2gj8" event={"ID":"3b265c7a-3909-4fd6-9981-c1723fbc1e02","Type":"ContainerDied","Data":"ce1d38864bb409e404978e6b3a0f6bd89d3dba826ce7be995022fe4e0ac9b188"} Oct 14 07:22:19 crc kubenswrapper[4870]: I1014 07:22:19.171811 4870 generic.go:334] "Generic (PLEG): container finished" podID="b2d073e2-e7bb-44c9-b074-c39afeae16b8" containerID="ffd73f77d45e4b6cd97b896c63da4c555a8d71621706a5dd9aa9fade449b5ce5" exitCode=2 Oct 14 07:22:19 crc kubenswrapper[4870]: I1014 07:22:19.171857 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 14 07:22:19 crc kubenswrapper[4870]: I1014 07:22:19.171893 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"b2d073e2-e7bb-44c9-b074-c39afeae16b8","Type":"ContainerDied","Data":"ffd73f77d45e4b6cd97b896c63da4c555a8d71621706a5dd9aa9fade449b5ce5"} Oct 14 07:22:19 crc kubenswrapper[4870]: I1014 07:22:19.171918 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"b2d073e2-e7bb-44c9-b074-c39afeae16b8","Type":"ContainerDied","Data":"bd238754e5c63ab7ff61dc40da6abf66375324653fb9c0c31ea19bac500142ed"} Oct 14 07:22:19 crc kubenswrapper[4870]: I1014 07:22:19.171970 4870 scope.go:117] "RemoveContainer" containerID="ffd73f77d45e4b6cd97b896c63da4c555a8d71621706a5dd9aa9fade449b5ce5" Oct 14 07:22:19 crc kubenswrapper[4870]: I1014 07:22:19.191113 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.191087179 podStartE2EDuration="2.191087179s" podCreationTimestamp="2025-10-14 07:22:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:22:19.188358685 +0000 UTC m=+1274.885719096" watchObservedRunningTime="2025-10-14 07:22:19.191087179 +0000 UTC m=+1274.888447580" Oct 14 07:22:19 crc kubenswrapper[4870]: I1014 07:22:19.232065 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 14 07:22:19 crc kubenswrapper[4870]: I1014 07:22:19.233306 4870 scope.go:117] "RemoveContainer" containerID="ffd73f77d45e4b6cd97b896c63da4c555a8d71621706a5dd9aa9fade449b5ce5" Oct 14 07:22:19 crc kubenswrapper[4870]: E1014 07:22:19.234651 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ffd73f77d45e4b6cd97b896c63da4c555a8d71621706a5dd9aa9fade449b5ce5\": container with ID starting with ffd73f77d45e4b6cd97b896c63da4c555a8d71621706a5dd9aa9fade449b5ce5 not found: ID does not exist" containerID="ffd73f77d45e4b6cd97b896c63da4c555a8d71621706a5dd9aa9fade449b5ce5" Oct 14 07:22:19 crc kubenswrapper[4870]: I1014 07:22:19.234711 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ffd73f77d45e4b6cd97b896c63da4c555a8d71621706a5dd9aa9fade449b5ce5"} err="failed to get container status \"ffd73f77d45e4b6cd97b896c63da4c555a8d71621706a5dd9aa9fade449b5ce5\": rpc error: code = NotFound desc = could not find container \"ffd73f77d45e4b6cd97b896c63da4c555a8d71621706a5dd9aa9fade449b5ce5\": container with ID starting with ffd73f77d45e4b6cd97b896c63da4c555a8d71621706a5dd9aa9fade449b5ce5 not found: ID does not exist" Oct 14 07:22:19 crc kubenswrapper[4870]: I1014 07:22:19.243657 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 14 07:22:19 crc kubenswrapper[4870]: I1014 07:22:19.270264 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Oct 14 07:22:19 crc kubenswrapper[4870]: E1014 07:22:19.270951 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b2d073e2-e7bb-44c9-b074-c39afeae16b8" containerName="kube-state-metrics" Oct 14 07:22:19 crc kubenswrapper[4870]: I1014 07:22:19.270978 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="b2d073e2-e7bb-44c9-b074-c39afeae16b8" containerName="kube-state-metrics" Oct 14 07:22:19 crc kubenswrapper[4870]: I1014 07:22:19.271235 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="b2d073e2-e7bb-44c9-b074-c39afeae16b8" containerName="kube-state-metrics" Oct 14 07:22:19 crc kubenswrapper[4870]: I1014 07:22:19.272021 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 14 07:22:19 crc kubenswrapper[4870]: I1014 07:22:19.274016 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Oct 14 07:22:19 crc kubenswrapper[4870]: I1014 07:22:19.274796 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Oct 14 07:22:19 crc kubenswrapper[4870]: I1014 07:22:19.282556 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 14 07:22:19 crc kubenswrapper[4870]: I1014 07:22:19.414657 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/c603f7b5-015f-4c2e-9144-b3e8349c3da3-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"c603f7b5-015f-4c2e-9144-b3e8349c3da3\") " pod="openstack/kube-state-metrics-0" Oct 14 07:22:19 crc kubenswrapper[4870]: I1014 07:22:19.414793 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l62g6\" (UniqueName: \"kubernetes.io/projected/c603f7b5-015f-4c2e-9144-b3e8349c3da3-kube-api-access-l62g6\") pod \"kube-state-metrics-0\" (UID: \"c603f7b5-015f-4c2e-9144-b3e8349c3da3\") " pod="openstack/kube-state-metrics-0" Oct 14 07:22:19 crc kubenswrapper[4870]: I1014 07:22:19.414910 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/c603f7b5-015f-4c2e-9144-b3e8349c3da3-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"c603f7b5-015f-4c2e-9144-b3e8349c3da3\") " pod="openstack/kube-state-metrics-0" Oct 14 07:22:19 crc kubenswrapper[4870]: I1014 07:22:19.414959 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c603f7b5-015f-4c2e-9144-b3e8349c3da3-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"c603f7b5-015f-4c2e-9144-b3e8349c3da3\") " pod="openstack/kube-state-metrics-0" Oct 14 07:22:19 crc kubenswrapper[4870]: I1014 07:22:19.512673 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 14 07:22:19 crc kubenswrapper[4870]: I1014 07:22:19.512731 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 14 07:22:19 crc kubenswrapper[4870]: I1014 07:22:19.516513 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l62g6\" (UniqueName: \"kubernetes.io/projected/c603f7b5-015f-4c2e-9144-b3e8349c3da3-kube-api-access-l62g6\") pod \"kube-state-metrics-0\" (UID: \"c603f7b5-015f-4c2e-9144-b3e8349c3da3\") " pod="openstack/kube-state-metrics-0" Oct 14 07:22:19 crc kubenswrapper[4870]: I1014 07:22:19.516671 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/c603f7b5-015f-4c2e-9144-b3e8349c3da3-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"c603f7b5-015f-4c2e-9144-b3e8349c3da3\") " pod="openstack/kube-state-metrics-0" Oct 14 07:22:19 crc kubenswrapper[4870]: I1014 07:22:19.516725 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c603f7b5-015f-4c2e-9144-b3e8349c3da3-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"c603f7b5-015f-4c2e-9144-b3e8349c3da3\") " pod="openstack/kube-state-metrics-0" Oct 14 07:22:19 crc kubenswrapper[4870]: I1014 07:22:19.516924 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/c603f7b5-015f-4c2e-9144-b3e8349c3da3-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"c603f7b5-015f-4c2e-9144-b3e8349c3da3\") " pod="openstack/kube-state-metrics-0" Oct 14 07:22:19 crc kubenswrapper[4870]: I1014 07:22:19.523384 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/c603f7b5-015f-4c2e-9144-b3e8349c3da3-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"c603f7b5-015f-4c2e-9144-b3e8349c3da3\") " pod="openstack/kube-state-metrics-0" Oct 14 07:22:19 crc kubenswrapper[4870]: I1014 07:22:19.536173 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c603f7b5-015f-4c2e-9144-b3e8349c3da3-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"c603f7b5-015f-4c2e-9144-b3e8349c3da3\") " pod="openstack/kube-state-metrics-0" Oct 14 07:22:19 crc kubenswrapper[4870]: I1014 07:22:19.540342 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/c603f7b5-015f-4c2e-9144-b3e8349c3da3-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"c603f7b5-015f-4c2e-9144-b3e8349c3da3\") " pod="openstack/kube-state-metrics-0" Oct 14 07:22:19 crc kubenswrapper[4870]: I1014 07:22:19.554679 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l62g6\" (UniqueName: \"kubernetes.io/projected/c603f7b5-015f-4c2e-9144-b3e8349c3da3-kube-api-access-l62g6\") pod \"kube-state-metrics-0\" (UID: \"c603f7b5-015f-4c2e-9144-b3e8349c3da3\") " pod="openstack/kube-state-metrics-0" Oct 14 07:22:19 crc kubenswrapper[4870]: I1014 07:22:19.590643 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 14 07:22:19 crc kubenswrapper[4870]: I1014 07:22:19.636462 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 14 07:22:20 crc kubenswrapper[4870]: I1014 07:22:20.037531 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 14 07:22:20 crc kubenswrapper[4870]: I1014 07:22:20.060944 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 14 07:22:20 crc kubenswrapper[4870]: I1014 07:22:20.066550 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2edb41c1-bdff-4251-91e3-8c5d61beccaf" containerName="ceilometer-central-agent" containerID="cri-o://678b866a2d0c6b6a4671816ebea308062057589d8d7a174cab754b0cf7da1ccb" gracePeriod=30 Oct 14 07:22:20 crc kubenswrapper[4870]: I1014 07:22:20.066659 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2edb41c1-bdff-4251-91e3-8c5d61beccaf" containerName="ceilometer-notification-agent" containerID="cri-o://ebd6c5d9bac96edcd5363084598bdcd47a36847351f24ebcb2ffa94c2ba6be41" gracePeriod=30 Oct 14 07:22:20 crc kubenswrapper[4870]: I1014 07:22:20.066681 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2edb41c1-bdff-4251-91e3-8c5d61beccaf" containerName="proxy-httpd" containerID="cri-o://130e23ed1e9f1ea0d4e7922866a5deec8f115d0bcc35cbfb3ca734a72f6e3a13" gracePeriod=30 Oct 14 07:22:20 crc kubenswrapper[4870]: I1014 07:22:20.066695 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2edb41c1-bdff-4251-91e3-8c5d61beccaf" containerName="sg-core" containerID="cri-o://4318a0f883a52b92c348661736366f87656f73b3037904058c4301e93f49bf25" gracePeriod=30 Oct 14 07:22:20 crc kubenswrapper[4870]: I1014 07:22:20.186226 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"c603f7b5-015f-4c2e-9144-b3e8349c3da3","Type":"ContainerStarted","Data":"9849fe4942d0101a5e43df2d44e038448ff1198af84d34460865495f68e86877"} Oct 14 07:22:20 crc kubenswrapper[4870]: I1014 07:22:20.488102 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-x2gj8" Oct 14 07:22:20 crc kubenswrapper[4870]: I1014 07:22:20.657145 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t6pqs\" (UniqueName: \"kubernetes.io/projected/3b265c7a-3909-4fd6-9981-c1723fbc1e02-kube-api-access-t6pqs\") pod \"3b265c7a-3909-4fd6-9981-c1723fbc1e02\" (UID: \"3b265c7a-3909-4fd6-9981-c1723fbc1e02\") " Oct 14 07:22:20 crc kubenswrapper[4870]: I1014 07:22:20.657292 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3b265c7a-3909-4fd6-9981-c1723fbc1e02-scripts\") pod \"3b265c7a-3909-4fd6-9981-c1723fbc1e02\" (UID: \"3b265c7a-3909-4fd6-9981-c1723fbc1e02\") " Oct 14 07:22:20 crc kubenswrapper[4870]: I1014 07:22:20.657343 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b265c7a-3909-4fd6-9981-c1723fbc1e02-combined-ca-bundle\") pod \"3b265c7a-3909-4fd6-9981-c1723fbc1e02\" (UID: \"3b265c7a-3909-4fd6-9981-c1723fbc1e02\") " Oct 14 07:22:20 crc kubenswrapper[4870]: I1014 07:22:20.657388 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3b265c7a-3909-4fd6-9981-c1723fbc1e02-config-data\") pod \"3b265c7a-3909-4fd6-9981-c1723fbc1e02\" (UID: \"3b265c7a-3909-4fd6-9981-c1723fbc1e02\") " Oct 14 07:22:20 crc kubenswrapper[4870]: I1014 07:22:20.661799 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3b265c7a-3909-4fd6-9981-c1723fbc1e02-scripts" (OuterVolumeSpecName: "scripts") pod "3b265c7a-3909-4fd6-9981-c1723fbc1e02" (UID: "3b265c7a-3909-4fd6-9981-c1723fbc1e02"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:22:20 crc kubenswrapper[4870]: I1014 07:22:20.664547 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3b265c7a-3909-4fd6-9981-c1723fbc1e02-kube-api-access-t6pqs" (OuterVolumeSpecName: "kube-api-access-t6pqs") pod "3b265c7a-3909-4fd6-9981-c1723fbc1e02" (UID: "3b265c7a-3909-4fd6-9981-c1723fbc1e02"). InnerVolumeSpecName "kube-api-access-t6pqs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:22:20 crc kubenswrapper[4870]: I1014 07:22:20.698575 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3b265c7a-3909-4fd6-9981-c1723fbc1e02-config-data" (OuterVolumeSpecName: "config-data") pod "3b265c7a-3909-4fd6-9981-c1723fbc1e02" (UID: "3b265c7a-3909-4fd6-9981-c1723fbc1e02"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:22:20 crc kubenswrapper[4870]: I1014 07:22:20.704975 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3b265c7a-3909-4fd6-9981-c1723fbc1e02-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3b265c7a-3909-4fd6-9981-c1723fbc1e02" (UID: "3b265c7a-3909-4fd6-9981-c1723fbc1e02"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:22:20 crc kubenswrapper[4870]: I1014 07:22:20.759611 4870 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3b265c7a-3909-4fd6-9981-c1723fbc1e02-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 07:22:20 crc kubenswrapper[4870]: I1014 07:22:20.759646 4870 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b265c7a-3909-4fd6-9981-c1723fbc1e02-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 07:22:20 crc kubenswrapper[4870]: I1014 07:22:20.759658 4870 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3b265c7a-3909-4fd6-9981-c1723fbc1e02-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 07:22:20 crc kubenswrapper[4870]: I1014 07:22:20.759667 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t6pqs\" (UniqueName: \"kubernetes.io/projected/3b265c7a-3909-4fd6-9981-c1723fbc1e02-kube-api-access-t6pqs\") on node \"crc\" DevicePath \"\"" Oct 14 07:22:21 crc kubenswrapper[4870]: I1014 07:22:21.046838 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b2d073e2-e7bb-44c9-b074-c39afeae16b8" path="/var/lib/kubelet/pods/b2d073e2-e7bb-44c9-b074-c39afeae16b8/volumes" Oct 14 07:22:21 crc kubenswrapper[4870]: I1014 07:22:21.200071 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"c603f7b5-015f-4c2e-9144-b3e8349c3da3","Type":"ContainerStarted","Data":"c79902c86d1a79331c9994ac2e40ba61e03d57293be1023f0938cc002e01194a"} Oct 14 07:22:21 crc kubenswrapper[4870]: I1014 07:22:21.200237 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Oct 14 07:22:21 crc kubenswrapper[4870]: I1014 07:22:21.203768 4870 generic.go:334] "Generic (PLEG): container finished" podID="2edb41c1-bdff-4251-91e3-8c5d61beccaf" containerID="130e23ed1e9f1ea0d4e7922866a5deec8f115d0bcc35cbfb3ca734a72f6e3a13" exitCode=0 Oct 14 07:22:21 crc kubenswrapper[4870]: I1014 07:22:21.203806 4870 generic.go:334] "Generic (PLEG): container finished" podID="2edb41c1-bdff-4251-91e3-8c5d61beccaf" containerID="4318a0f883a52b92c348661736366f87656f73b3037904058c4301e93f49bf25" exitCode=2 Oct 14 07:22:21 crc kubenswrapper[4870]: I1014 07:22:21.203821 4870 generic.go:334] "Generic (PLEG): container finished" podID="2edb41c1-bdff-4251-91e3-8c5d61beccaf" containerID="678b866a2d0c6b6a4671816ebea308062057589d8d7a174cab754b0cf7da1ccb" exitCode=0 Oct 14 07:22:21 crc kubenswrapper[4870]: I1014 07:22:21.203832 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2edb41c1-bdff-4251-91e3-8c5d61beccaf","Type":"ContainerDied","Data":"130e23ed1e9f1ea0d4e7922866a5deec8f115d0bcc35cbfb3ca734a72f6e3a13"} Oct 14 07:22:21 crc kubenswrapper[4870]: I1014 07:22:21.203871 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2edb41c1-bdff-4251-91e3-8c5d61beccaf","Type":"ContainerDied","Data":"4318a0f883a52b92c348661736366f87656f73b3037904058c4301e93f49bf25"} Oct 14 07:22:21 crc kubenswrapper[4870]: I1014 07:22:21.203881 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2edb41c1-bdff-4251-91e3-8c5d61beccaf","Type":"ContainerDied","Data":"678b866a2d0c6b6a4671816ebea308062057589d8d7a174cab754b0cf7da1ccb"} Oct 14 07:22:21 crc kubenswrapper[4870]: I1014 07:22:21.206738 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-x2gj8" event={"ID":"3b265c7a-3909-4fd6-9981-c1723fbc1e02","Type":"ContainerDied","Data":"5df333dc10c08df3ccbdcd54ec0537a2f6802e8ed904078ac97b1d308eb9d1b7"} Oct 14 07:22:21 crc kubenswrapper[4870]: I1014 07:22:21.206772 4870 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5df333dc10c08df3ccbdcd54ec0537a2f6802e8ed904078ac97b1d308eb9d1b7" Oct 14 07:22:21 crc kubenswrapper[4870]: I1014 07:22:21.206824 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-x2gj8" Oct 14 07:22:21 crc kubenswrapper[4870]: I1014 07:22:21.227708 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=1.869338329 podStartE2EDuration="2.227684855s" podCreationTimestamp="2025-10-14 07:22:19 +0000 UTC" firstStartedPulling="2025-10-14 07:22:20.054896271 +0000 UTC m=+1275.752256642" lastFinishedPulling="2025-10-14 07:22:20.413242797 +0000 UTC m=+1276.110603168" observedRunningTime="2025-10-14 07:22:21.223246282 +0000 UTC m=+1276.920606693" watchObservedRunningTime="2025-10-14 07:22:21.227684855 +0000 UTC m=+1276.925045226" Oct 14 07:22:21 crc kubenswrapper[4870]: I1014 07:22:21.311673 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 14 07:22:21 crc kubenswrapper[4870]: E1014 07:22:21.312113 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b265c7a-3909-4fd6-9981-c1723fbc1e02" containerName="nova-cell1-conductor-db-sync" Oct 14 07:22:21 crc kubenswrapper[4870]: I1014 07:22:21.312132 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b265c7a-3909-4fd6-9981-c1723fbc1e02" containerName="nova-cell1-conductor-db-sync" Oct 14 07:22:21 crc kubenswrapper[4870]: I1014 07:22:21.312392 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="3b265c7a-3909-4fd6-9981-c1723fbc1e02" containerName="nova-cell1-conductor-db-sync" Oct 14 07:22:21 crc kubenswrapper[4870]: I1014 07:22:21.313069 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 14 07:22:21 crc kubenswrapper[4870]: I1014 07:22:21.320413 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Oct 14 07:22:21 crc kubenswrapper[4870]: I1014 07:22:21.322990 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 14 07:22:21 crc kubenswrapper[4870]: I1014 07:22:21.472380 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/61885759-41a7-4ee8-904a-3f85bb66cd19-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"61885759-41a7-4ee8-904a-3f85bb66cd19\") " pod="openstack/nova-cell1-conductor-0" Oct 14 07:22:21 crc kubenswrapper[4870]: I1014 07:22:21.472736 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/61885759-41a7-4ee8-904a-3f85bb66cd19-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"61885759-41a7-4ee8-904a-3f85bb66cd19\") " pod="openstack/nova-cell1-conductor-0" Oct 14 07:22:21 crc kubenswrapper[4870]: I1014 07:22:21.472774 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bv2wt\" (UniqueName: \"kubernetes.io/projected/61885759-41a7-4ee8-904a-3f85bb66cd19-kube-api-access-bv2wt\") pod \"nova-cell1-conductor-0\" (UID: \"61885759-41a7-4ee8-904a-3f85bb66cd19\") " pod="openstack/nova-cell1-conductor-0" Oct 14 07:22:21 crc kubenswrapper[4870]: I1014 07:22:21.575336 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/61885759-41a7-4ee8-904a-3f85bb66cd19-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"61885759-41a7-4ee8-904a-3f85bb66cd19\") " pod="openstack/nova-cell1-conductor-0" Oct 14 07:22:21 crc kubenswrapper[4870]: I1014 07:22:21.575432 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bv2wt\" (UniqueName: \"kubernetes.io/projected/61885759-41a7-4ee8-904a-3f85bb66cd19-kube-api-access-bv2wt\") pod \"nova-cell1-conductor-0\" (UID: \"61885759-41a7-4ee8-904a-3f85bb66cd19\") " pod="openstack/nova-cell1-conductor-0" Oct 14 07:22:21 crc kubenswrapper[4870]: I1014 07:22:21.575533 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/61885759-41a7-4ee8-904a-3f85bb66cd19-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"61885759-41a7-4ee8-904a-3f85bb66cd19\") " pod="openstack/nova-cell1-conductor-0" Oct 14 07:22:21 crc kubenswrapper[4870]: I1014 07:22:21.580031 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/61885759-41a7-4ee8-904a-3f85bb66cd19-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"61885759-41a7-4ee8-904a-3f85bb66cd19\") " pod="openstack/nova-cell1-conductor-0" Oct 14 07:22:21 crc kubenswrapper[4870]: I1014 07:22:21.584128 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/61885759-41a7-4ee8-904a-3f85bb66cd19-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"61885759-41a7-4ee8-904a-3f85bb66cd19\") " pod="openstack/nova-cell1-conductor-0" Oct 14 07:22:21 crc kubenswrapper[4870]: I1014 07:22:21.607158 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bv2wt\" (UniqueName: \"kubernetes.io/projected/61885759-41a7-4ee8-904a-3f85bb66cd19-kube-api-access-bv2wt\") pod \"nova-cell1-conductor-0\" (UID: \"61885759-41a7-4ee8-904a-3f85bb66cd19\") " pod="openstack/nova-cell1-conductor-0" Oct 14 07:22:21 crc kubenswrapper[4870]: I1014 07:22:21.638115 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 14 07:22:22 crc kubenswrapper[4870]: I1014 07:22:22.229662 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 14 07:22:22 crc kubenswrapper[4870]: I1014 07:22:22.931819 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 14 07:22:23 crc kubenswrapper[4870]: I1014 07:22:23.113592 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2edb41c1-bdff-4251-91e3-8c5d61beccaf-scripts\") pod \"2edb41c1-bdff-4251-91e3-8c5d61beccaf\" (UID: \"2edb41c1-bdff-4251-91e3-8c5d61beccaf\") " Oct 14 07:22:23 crc kubenswrapper[4870]: I1014 07:22:23.113677 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2edb41c1-bdff-4251-91e3-8c5d61beccaf-sg-core-conf-yaml\") pod \"2edb41c1-bdff-4251-91e3-8c5d61beccaf\" (UID: \"2edb41c1-bdff-4251-91e3-8c5d61beccaf\") " Oct 14 07:22:23 crc kubenswrapper[4870]: I1014 07:22:23.114614 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2edb41c1-bdff-4251-91e3-8c5d61beccaf-combined-ca-bundle\") pod \"2edb41c1-bdff-4251-91e3-8c5d61beccaf\" (UID: \"2edb41c1-bdff-4251-91e3-8c5d61beccaf\") " Oct 14 07:22:23 crc kubenswrapper[4870]: I1014 07:22:23.114668 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ph2df\" (UniqueName: \"kubernetes.io/projected/2edb41c1-bdff-4251-91e3-8c5d61beccaf-kube-api-access-ph2df\") pod \"2edb41c1-bdff-4251-91e3-8c5d61beccaf\" (UID: \"2edb41c1-bdff-4251-91e3-8c5d61beccaf\") " Oct 14 07:22:23 crc kubenswrapper[4870]: I1014 07:22:23.114712 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2edb41c1-bdff-4251-91e3-8c5d61beccaf-log-httpd\") pod \"2edb41c1-bdff-4251-91e3-8c5d61beccaf\" (UID: \"2edb41c1-bdff-4251-91e3-8c5d61beccaf\") " Oct 14 07:22:23 crc kubenswrapper[4870]: I1014 07:22:23.114760 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2edb41c1-bdff-4251-91e3-8c5d61beccaf-run-httpd\") pod \"2edb41c1-bdff-4251-91e3-8c5d61beccaf\" (UID: \"2edb41c1-bdff-4251-91e3-8c5d61beccaf\") " Oct 14 07:22:23 crc kubenswrapper[4870]: I1014 07:22:23.114807 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2edb41c1-bdff-4251-91e3-8c5d61beccaf-config-data\") pod \"2edb41c1-bdff-4251-91e3-8c5d61beccaf\" (UID: \"2edb41c1-bdff-4251-91e3-8c5d61beccaf\") " Oct 14 07:22:23 crc kubenswrapper[4870]: I1014 07:22:23.115705 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2edb41c1-bdff-4251-91e3-8c5d61beccaf-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "2edb41c1-bdff-4251-91e3-8c5d61beccaf" (UID: "2edb41c1-bdff-4251-91e3-8c5d61beccaf"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:22:23 crc kubenswrapper[4870]: I1014 07:22:23.116013 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2edb41c1-bdff-4251-91e3-8c5d61beccaf-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "2edb41c1-bdff-4251-91e3-8c5d61beccaf" (UID: "2edb41c1-bdff-4251-91e3-8c5d61beccaf"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:22:23 crc kubenswrapper[4870]: I1014 07:22:23.120034 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2edb41c1-bdff-4251-91e3-8c5d61beccaf-kube-api-access-ph2df" (OuterVolumeSpecName: "kube-api-access-ph2df") pod "2edb41c1-bdff-4251-91e3-8c5d61beccaf" (UID: "2edb41c1-bdff-4251-91e3-8c5d61beccaf"). InnerVolumeSpecName "kube-api-access-ph2df". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:22:23 crc kubenswrapper[4870]: I1014 07:22:23.137900 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2edb41c1-bdff-4251-91e3-8c5d61beccaf-scripts" (OuterVolumeSpecName: "scripts") pod "2edb41c1-bdff-4251-91e3-8c5d61beccaf" (UID: "2edb41c1-bdff-4251-91e3-8c5d61beccaf"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:22:23 crc kubenswrapper[4870]: I1014 07:22:23.156912 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2edb41c1-bdff-4251-91e3-8c5d61beccaf-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "2edb41c1-bdff-4251-91e3-8c5d61beccaf" (UID: "2edb41c1-bdff-4251-91e3-8c5d61beccaf"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:22:23 crc kubenswrapper[4870]: I1014 07:22:23.211917 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2edb41c1-bdff-4251-91e3-8c5d61beccaf-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2edb41c1-bdff-4251-91e3-8c5d61beccaf" (UID: "2edb41c1-bdff-4251-91e3-8c5d61beccaf"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:22:23 crc kubenswrapper[4870]: I1014 07:22:23.217073 4870 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2edb41c1-bdff-4251-91e3-8c5d61beccaf-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 14 07:22:23 crc kubenswrapper[4870]: I1014 07:22:23.217127 4870 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2edb41c1-bdff-4251-91e3-8c5d61beccaf-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 07:22:23 crc kubenswrapper[4870]: I1014 07:22:23.217149 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ph2df\" (UniqueName: \"kubernetes.io/projected/2edb41c1-bdff-4251-91e3-8c5d61beccaf-kube-api-access-ph2df\") on node \"crc\" DevicePath \"\"" Oct 14 07:22:23 crc kubenswrapper[4870]: I1014 07:22:23.217164 4870 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2edb41c1-bdff-4251-91e3-8c5d61beccaf-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 14 07:22:23 crc kubenswrapper[4870]: I1014 07:22:23.217175 4870 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2edb41c1-bdff-4251-91e3-8c5d61beccaf-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 14 07:22:23 crc kubenswrapper[4870]: I1014 07:22:23.217185 4870 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2edb41c1-bdff-4251-91e3-8c5d61beccaf-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 07:22:23 crc kubenswrapper[4870]: I1014 07:22:23.234409 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"61885759-41a7-4ee8-904a-3f85bb66cd19","Type":"ContainerStarted","Data":"925b675acc2471fa8d4cffebdeddb3b79263bf8e5254078a5b2f740ce843cfd5"} Oct 14 07:22:23 crc kubenswrapper[4870]: I1014 07:22:23.235330 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"61885759-41a7-4ee8-904a-3f85bb66cd19","Type":"ContainerStarted","Data":"2c5fb8deafb920aec9e2d2835198a9dfcbe4d42a4bb7c408109aef8440352e4f"} Oct 14 07:22:23 crc kubenswrapper[4870]: I1014 07:22:23.235525 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Oct 14 07:22:23 crc kubenswrapper[4870]: I1014 07:22:23.239534 4870 generic.go:334] "Generic (PLEG): container finished" podID="2edb41c1-bdff-4251-91e3-8c5d61beccaf" containerID="ebd6c5d9bac96edcd5363084598bdcd47a36847351f24ebcb2ffa94c2ba6be41" exitCode=0 Oct 14 07:22:23 crc kubenswrapper[4870]: I1014 07:22:23.239649 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2edb41c1-bdff-4251-91e3-8c5d61beccaf","Type":"ContainerDied","Data":"ebd6c5d9bac96edcd5363084598bdcd47a36847351f24ebcb2ffa94c2ba6be41"} Oct 14 07:22:23 crc kubenswrapper[4870]: I1014 07:22:23.239768 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2edb41c1-bdff-4251-91e3-8c5d61beccaf","Type":"ContainerDied","Data":"848497eeca968f5f9da4a37de9d21e29add6201e651810d078b383a6f2ef4676"} Oct 14 07:22:23 crc kubenswrapper[4870]: I1014 07:22:23.239863 4870 scope.go:117] "RemoveContainer" containerID="130e23ed1e9f1ea0d4e7922866a5deec8f115d0bcc35cbfb3ca734a72f6e3a13" Oct 14 07:22:23 crc kubenswrapper[4870]: I1014 07:22:23.239663 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 14 07:22:23 crc kubenswrapper[4870]: I1014 07:22:23.255371 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2edb41c1-bdff-4251-91e3-8c5d61beccaf-config-data" (OuterVolumeSpecName: "config-data") pod "2edb41c1-bdff-4251-91e3-8c5d61beccaf" (UID: "2edb41c1-bdff-4251-91e3-8c5d61beccaf"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:22:23 crc kubenswrapper[4870]: I1014 07:22:23.261005 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.260988444 podStartE2EDuration="2.260988444s" podCreationTimestamp="2025-10-14 07:22:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:22:23.249962306 +0000 UTC m=+1278.947322677" watchObservedRunningTime="2025-10-14 07:22:23.260988444 +0000 UTC m=+1278.958348815" Oct 14 07:22:23 crc kubenswrapper[4870]: I1014 07:22:23.269355 4870 scope.go:117] "RemoveContainer" containerID="4318a0f883a52b92c348661736366f87656f73b3037904058c4301e93f49bf25" Oct 14 07:22:23 crc kubenswrapper[4870]: I1014 07:22:23.300876 4870 scope.go:117] "RemoveContainer" containerID="ebd6c5d9bac96edcd5363084598bdcd47a36847351f24ebcb2ffa94c2ba6be41" Oct 14 07:22:23 crc kubenswrapper[4870]: I1014 07:22:23.319198 4870 scope.go:117] "RemoveContainer" containerID="678b866a2d0c6b6a4671816ebea308062057589d8d7a174cab754b0cf7da1ccb" Oct 14 07:22:23 crc kubenswrapper[4870]: I1014 07:22:23.327094 4870 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2edb41c1-bdff-4251-91e3-8c5d61beccaf-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 07:22:23 crc kubenswrapper[4870]: I1014 07:22:23.342968 4870 scope.go:117] "RemoveContainer" containerID="130e23ed1e9f1ea0d4e7922866a5deec8f115d0bcc35cbfb3ca734a72f6e3a13" Oct 14 07:22:23 crc kubenswrapper[4870]: E1014 07:22:23.343491 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"130e23ed1e9f1ea0d4e7922866a5deec8f115d0bcc35cbfb3ca734a72f6e3a13\": container with ID starting with 130e23ed1e9f1ea0d4e7922866a5deec8f115d0bcc35cbfb3ca734a72f6e3a13 not found: ID does not exist" containerID="130e23ed1e9f1ea0d4e7922866a5deec8f115d0bcc35cbfb3ca734a72f6e3a13" Oct 14 07:22:23 crc kubenswrapper[4870]: I1014 07:22:23.343520 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"130e23ed1e9f1ea0d4e7922866a5deec8f115d0bcc35cbfb3ca734a72f6e3a13"} err="failed to get container status \"130e23ed1e9f1ea0d4e7922866a5deec8f115d0bcc35cbfb3ca734a72f6e3a13\": rpc error: code = NotFound desc = could not find container \"130e23ed1e9f1ea0d4e7922866a5deec8f115d0bcc35cbfb3ca734a72f6e3a13\": container with ID starting with 130e23ed1e9f1ea0d4e7922866a5deec8f115d0bcc35cbfb3ca734a72f6e3a13 not found: ID does not exist" Oct 14 07:22:23 crc kubenswrapper[4870]: I1014 07:22:23.343541 4870 scope.go:117] "RemoveContainer" containerID="4318a0f883a52b92c348661736366f87656f73b3037904058c4301e93f49bf25" Oct 14 07:22:23 crc kubenswrapper[4870]: E1014 07:22:23.343793 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4318a0f883a52b92c348661736366f87656f73b3037904058c4301e93f49bf25\": container with ID starting with 4318a0f883a52b92c348661736366f87656f73b3037904058c4301e93f49bf25 not found: ID does not exist" containerID="4318a0f883a52b92c348661736366f87656f73b3037904058c4301e93f49bf25" Oct 14 07:22:23 crc kubenswrapper[4870]: I1014 07:22:23.343818 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4318a0f883a52b92c348661736366f87656f73b3037904058c4301e93f49bf25"} err="failed to get container status \"4318a0f883a52b92c348661736366f87656f73b3037904058c4301e93f49bf25\": rpc error: code = NotFound desc = could not find container \"4318a0f883a52b92c348661736366f87656f73b3037904058c4301e93f49bf25\": container with ID starting with 4318a0f883a52b92c348661736366f87656f73b3037904058c4301e93f49bf25 not found: ID does not exist" Oct 14 07:22:23 crc kubenswrapper[4870]: I1014 07:22:23.343832 4870 scope.go:117] "RemoveContainer" containerID="ebd6c5d9bac96edcd5363084598bdcd47a36847351f24ebcb2ffa94c2ba6be41" Oct 14 07:22:23 crc kubenswrapper[4870]: E1014 07:22:23.344133 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ebd6c5d9bac96edcd5363084598bdcd47a36847351f24ebcb2ffa94c2ba6be41\": container with ID starting with ebd6c5d9bac96edcd5363084598bdcd47a36847351f24ebcb2ffa94c2ba6be41 not found: ID does not exist" containerID="ebd6c5d9bac96edcd5363084598bdcd47a36847351f24ebcb2ffa94c2ba6be41" Oct 14 07:22:23 crc kubenswrapper[4870]: I1014 07:22:23.344155 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ebd6c5d9bac96edcd5363084598bdcd47a36847351f24ebcb2ffa94c2ba6be41"} err="failed to get container status \"ebd6c5d9bac96edcd5363084598bdcd47a36847351f24ebcb2ffa94c2ba6be41\": rpc error: code = NotFound desc = could not find container \"ebd6c5d9bac96edcd5363084598bdcd47a36847351f24ebcb2ffa94c2ba6be41\": container with ID starting with ebd6c5d9bac96edcd5363084598bdcd47a36847351f24ebcb2ffa94c2ba6be41 not found: ID does not exist" Oct 14 07:22:23 crc kubenswrapper[4870]: I1014 07:22:23.344201 4870 scope.go:117] "RemoveContainer" containerID="678b866a2d0c6b6a4671816ebea308062057589d8d7a174cab754b0cf7da1ccb" Oct 14 07:22:23 crc kubenswrapper[4870]: E1014 07:22:23.344391 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"678b866a2d0c6b6a4671816ebea308062057589d8d7a174cab754b0cf7da1ccb\": container with ID starting with 678b866a2d0c6b6a4671816ebea308062057589d8d7a174cab754b0cf7da1ccb not found: ID does not exist" containerID="678b866a2d0c6b6a4671816ebea308062057589d8d7a174cab754b0cf7da1ccb" Oct 14 07:22:23 crc kubenswrapper[4870]: I1014 07:22:23.344449 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"678b866a2d0c6b6a4671816ebea308062057589d8d7a174cab754b0cf7da1ccb"} err="failed to get container status \"678b866a2d0c6b6a4671816ebea308062057589d8d7a174cab754b0cf7da1ccb\": rpc error: code = NotFound desc = could not find container \"678b866a2d0c6b6a4671816ebea308062057589d8d7a174cab754b0cf7da1ccb\": container with ID starting with 678b866a2d0c6b6a4671816ebea308062057589d8d7a174cab754b0cf7da1ccb not found: ID does not exist" Oct 14 07:22:23 crc kubenswrapper[4870]: I1014 07:22:23.589725 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 14 07:22:23 crc kubenswrapper[4870]: I1014 07:22:23.610366 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 14 07:22:23 crc kubenswrapper[4870]: I1014 07:22:23.627604 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 14 07:22:23 crc kubenswrapper[4870]: E1014 07:22:23.628807 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2edb41c1-bdff-4251-91e3-8c5d61beccaf" containerName="ceilometer-central-agent" Oct 14 07:22:23 crc kubenswrapper[4870]: I1014 07:22:23.628835 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="2edb41c1-bdff-4251-91e3-8c5d61beccaf" containerName="ceilometer-central-agent" Oct 14 07:22:23 crc kubenswrapper[4870]: E1014 07:22:23.628890 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2edb41c1-bdff-4251-91e3-8c5d61beccaf" containerName="sg-core" Oct 14 07:22:23 crc kubenswrapper[4870]: I1014 07:22:23.628901 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="2edb41c1-bdff-4251-91e3-8c5d61beccaf" containerName="sg-core" Oct 14 07:22:23 crc kubenswrapper[4870]: E1014 07:22:23.628929 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2edb41c1-bdff-4251-91e3-8c5d61beccaf" containerName="ceilometer-notification-agent" Oct 14 07:22:23 crc kubenswrapper[4870]: I1014 07:22:23.628939 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="2edb41c1-bdff-4251-91e3-8c5d61beccaf" containerName="ceilometer-notification-agent" Oct 14 07:22:23 crc kubenswrapper[4870]: E1014 07:22:23.628982 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2edb41c1-bdff-4251-91e3-8c5d61beccaf" containerName="proxy-httpd" Oct 14 07:22:23 crc kubenswrapper[4870]: I1014 07:22:23.628992 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="2edb41c1-bdff-4251-91e3-8c5d61beccaf" containerName="proxy-httpd" Oct 14 07:22:23 crc kubenswrapper[4870]: I1014 07:22:23.629634 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="2edb41c1-bdff-4251-91e3-8c5d61beccaf" containerName="proxy-httpd" Oct 14 07:22:23 crc kubenswrapper[4870]: I1014 07:22:23.629675 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="2edb41c1-bdff-4251-91e3-8c5d61beccaf" containerName="ceilometer-central-agent" Oct 14 07:22:23 crc kubenswrapper[4870]: I1014 07:22:23.629686 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="2edb41c1-bdff-4251-91e3-8c5d61beccaf" containerName="sg-core" Oct 14 07:22:23 crc kubenswrapper[4870]: I1014 07:22:23.629716 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="2edb41c1-bdff-4251-91e3-8c5d61beccaf" containerName="ceilometer-notification-agent" Oct 14 07:22:23 crc kubenswrapper[4870]: I1014 07:22:23.642434 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 14 07:22:23 crc kubenswrapper[4870]: I1014 07:22:23.648643 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Oct 14 07:22:23 crc kubenswrapper[4870]: I1014 07:22:23.648752 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 14 07:22:23 crc kubenswrapper[4870]: I1014 07:22:23.648952 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 14 07:22:23 crc kubenswrapper[4870]: I1014 07:22:23.671725 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 14 07:22:23 crc kubenswrapper[4870]: I1014 07:22:23.746433 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b84cf82f-2718-4712-912e-5cf227d2ac49-config-data\") pod \"ceilometer-0\" (UID: \"b84cf82f-2718-4712-912e-5cf227d2ac49\") " pod="openstack/ceilometer-0" Oct 14 07:22:23 crc kubenswrapper[4870]: I1014 07:22:23.746515 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b84cf82f-2718-4712-912e-5cf227d2ac49-run-httpd\") pod \"ceilometer-0\" (UID: \"b84cf82f-2718-4712-912e-5cf227d2ac49\") " pod="openstack/ceilometer-0" Oct 14 07:22:23 crc kubenswrapper[4870]: I1014 07:22:23.746540 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/b84cf82f-2718-4712-912e-5cf227d2ac49-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"b84cf82f-2718-4712-912e-5cf227d2ac49\") " pod="openstack/ceilometer-0" Oct 14 07:22:23 crc kubenswrapper[4870]: I1014 07:22:23.746972 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b84cf82f-2718-4712-912e-5cf227d2ac49-log-httpd\") pod \"ceilometer-0\" (UID: \"b84cf82f-2718-4712-912e-5cf227d2ac49\") " pod="openstack/ceilometer-0" Oct 14 07:22:23 crc kubenswrapper[4870]: I1014 07:22:23.747084 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b84cf82f-2718-4712-912e-5cf227d2ac49-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b84cf82f-2718-4712-912e-5cf227d2ac49\") " pod="openstack/ceilometer-0" Oct 14 07:22:23 crc kubenswrapper[4870]: I1014 07:22:23.747136 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b84cf82f-2718-4712-912e-5cf227d2ac49-scripts\") pod \"ceilometer-0\" (UID: \"b84cf82f-2718-4712-912e-5cf227d2ac49\") " pod="openstack/ceilometer-0" Oct 14 07:22:23 crc kubenswrapper[4870]: I1014 07:22:23.747164 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b84cf82f-2718-4712-912e-5cf227d2ac49-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b84cf82f-2718-4712-912e-5cf227d2ac49\") " pod="openstack/ceilometer-0" Oct 14 07:22:23 crc kubenswrapper[4870]: I1014 07:22:23.747258 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7dz7f\" (UniqueName: \"kubernetes.io/projected/b84cf82f-2718-4712-912e-5cf227d2ac49-kube-api-access-7dz7f\") pod \"ceilometer-0\" (UID: \"b84cf82f-2718-4712-912e-5cf227d2ac49\") " pod="openstack/ceilometer-0" Oct 14 07:22:23 crc kubenswrapper[4870]: I1014 07:22:23.848920 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b84cf82f-2718-4712-912e-5cf227d2ac49-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b84cf82f-2718-4712-912e-5cf227d2ac49\") " pod="openstack/ceilometer-0" Oct 14 07:22:23 crc kubenswrapper[4870]: I1014 07:22:23.848989 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7dz7f\" (UniqueName: \"kubernetes.io/projected/b84cf82f-2718-4712-912e-5cf227d2ac49-kube-api-access-7dz7f\") pod \"ceilometer-0\" (UID: \"b84cf82f-2718-4712-912e-5cf227d2ac49\") " pod="openstack/ceilometer-0" Oct 14 07:22:23 crc kubenswrapper[4870]: I1014 07:22:23.849039 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b84cf82f-2718-4712-912e-5cf227d2ac49-config-data\") pod \"ceilometer-0\" (UID: \"b84cf82f-2718-4712-912e-5cf227d2ac49\") " pod="openstack/ceilometer-0" Oct 14 07:22:23 crc kubenswrapper[4870]: I1014 07:22:23.849079 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b84cf82f-2718-4712-912e-5cf227d2ac49-run-httpd\") pod \"ceilometer-0\" (UID: \"b84cf82f-2718-4712-912e-5cf227d2ac49\") " pod="openstack/ceilometer-0" Oct 14 07:22:23 crc kubenswrapper[4870]: I1014 07:22:23.849100 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/b84cf82f-2718-4712-912e-5cf227d2ac49-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"b84cf82f-2718-4712-912e-5cf227d2ac49\") " pod="openstack/ceilometer-0" Oct 14 07:22:23 crc kubenswrapper[4870]: I1014 07:22:23.849213 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b84cf82f-2718-4712-912e-5cf227d2ac49-log-httpd\") pod \"ceilometer-0\" (UID: \"b84cf82f-2718-4712-912e-5cf227d2ac49\") " pod="openstack/ceilometer-0" Oct 14 07:22:23 crc kubenswrapper[4870]: I1014 07:22:23.849250 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b84cf82f-2718-4712-912e-5cf227d2ac49-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b84cf82f-2718-4712-912e-5cf227d2ac49\") " pod="openstack/ceilometer-0" Oct 14 07:22:23 crc kubenswrapper[4870]: I1014 07:22:23.849286 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b84cf82f-2718-4712-912e-5cf227d2ac49-scripts\") pod \"ceilometer-0\" (UID: \"b84cf82f-2718-4712-912e-5cf227d2ac49\") " pod="openstack/ceilometer-0" Oct 14 07:22:23 crc kubenswrapper[4870]: I1014 07:22:23.849719 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b84cf82f-2718-4712-912e-5cf227d2ac49-run-httpd\") pod \"ceilometer-0\" (UID: \"b84cf82f-2718-4712-912e-5cf227d2ac49\") " pod="openstack/ceilometer-0" Oct 14 07:22:23 crc kubenswrapper[4870]: I1014 07:22:23.849787 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b84cf82f-2718-4712-912e-5cf227d2ac49-log-httpd\") pod \"ceilometer-0\" (UID: \"b84cf82f-2718-4712-912e-5cf227d2ac49\") " pod="openstack/ceilometer-0" Oct 14 07:22:23 crc kubenswrapper[4870]: I1014 07:22:23.854008 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b84cf82f-2718-4712-912e-5cf227d2ac49-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b84cf82f-2718-4712-912e-5cf227d2ac49\") " pod="openstack/ceilometer-0" Oct 14 07:22:23 crc kubenswrapper[4870]: I1014 07:22:23.855197 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/b84cf82f-2718-4712-912e-5cf227d2ac49-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"b84cf82f-2718-4712-912e-5cf227d2ac49\") " pod="openstack/ceilometer-0" Oct 14 07:22:23 crc kubenswrapper[4870]: I1014 07:22:23.855383 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b84cf82f-2718-4712-912e-5cf227d2ac49-scripts\") pod \"ceilometer-0\" (UID: \"b84cf82f-2718-4712-912e-5cf227d2ac49\") " pod="openstack/ceilometer-0" Oct 14 07:22:23 crc kubenswrapper[4870]: I1014 07:22:23.856611 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b84cf82f-2718-4712-912e-5cf227d2ac49-config-data\") pod \"ceilometer-0\" (UID: \"b84cf82f-2718-4712-912e-5cf227d2ac49\") " pod="openstack/ceilometer-0" Oct 14 07:22:23 crc kubenswrapper[4870]: I1014 07:22:23.857465 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b84cf82f-2718-4712-912e-5cf227d2ac49-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b84cf82f-2718-4712-912e-5cf227d2ac49\") " pod="openstack/ceilometer-0" Oct 14 07:22:23 crc kubenswrapper[4870]: I1014 07:22:23.873048 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7dz7f\" (UniqueName: \"kubernetes.io/projected/b84cf82f-2718-4712-912e-5cf227d2ac49-kube-api-access-7dz7f\") pod \"ceilometer-0\" (UID: \"b84cf82f-2718-4712-912e-5cf227d2ac49\") " pod="openstack/ceilometer-0" Oct 14 07:22:23 crc kubenswrapper[4870]: I1014 07:22:23.951841 4870 patch_prober.go:28] interesting pod/machine-config-daemon-7tvc7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 07:22:23 crc kubenswrapper[4870]: I1014 07:22:23.951894 4870 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 07:22:23 crc kubenswrapper[4870]: I1014 07:22:23.951935 4870 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" Oct 14 07:22:23 crc kubenswrapper[4870]: I1014 07:22:23.952788 4870 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"5389ca0979ade5a4bccec9faa4aa51cda4fcec373b2752e9cf9faa85bf74bfdf"} pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 14 07:22:23 crc kubenswrapper[4870]: I1014 07:22:23.952852 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerName="machine-config-daemon" containerID="cri-o://5389ca0979ade5a4bccec9faa4aa51cda4fcec373b2752e9cf9faa85bf74bfdf" gracePeriod=600 Oct 14 07:22:23 crc kubenswrapper[4870]: I1014 07:22:23.993820 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 14 07:22:24 crc kubenswrapper[4870]: I1014 07:22:24.269880 4870 generic.go:334] "Generic (PLEG): container finished" podID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerID="5389ca0979ade5a4bccec9faa4aa51cda4fcec373b2752e9cf9faa85bf74bfdf" exitCode=0 Oct 14 07:22:24 crc kubenswrapper[4870]: I1014 07:22:24.270005 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" event={"ID":"874a5e45-dffd-4d17-b609-b7d3ed2eab07","Type":"ContainerDied","Data":"5389ca0979ade5a4bccec9faa4aa51cda4fcec373b2752e9cf9faa85bf74bfdf"} Oct 14 07:22:24 crc kubenswrapper[4870]: I1014 07:22:24.271076 4870 scope.go:117] "RemoveContainer" containerID="afa79ed981f5045c93ba7103bf82588637ff210f79db4fa50132dc5fa3e91338" Oct 14 07:22:24 crc kubenswrapper[4870]: I1014 07:22:24.507776 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 14 07:22:24 crc kubenswrapper[4870]: I1014 07:22:24.512145 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 14 07:22:24 crc kubenswrapper[4870]: I1014 07:22:24.512202 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 14 07:22:24 crc kubenswrapper[4870]: W1014 07:22:24.518029 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb84cf82f_2718_4712_912e_5cf227d2ac49.slice/crio-d8be7dcf1e489a881ca8ec158ac6f0d640590fd54799a321c88795ee1085b376 WatchSource:0}: Error finding container d8be7dcf1e489a881ca8ec158ac6f0d640590fd54799a321c88795ee1085b376: Status 404 returned error can't find the container with id d8be7dcf1e489a881ca8ec158ac6f0d640590fd54799a321c88795ee1085b376 Oct 14 07:22:24 crc kubenswrapper[4870]: I1014 07:22:24.636672 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 14 07:22:24 crc kubenswrapper[4870]: I1014 07:22:24.681350 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 14 07:22:25 crc kubenswrapper[4870]: I1014 07:22:25.050559 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2edb41c1-bdff-4251-91e3-8c5d61beccaf" path="/var/lib/kubelet/pods/2edb41c1-bdff-4251-91e3-8c5d61beccaf/volumes" Oct 14 07:22:25 crc kubenswrapper[4870]: I1014 07:22:25.284166 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b84cf82f-2718-4712-912e-5cf227d2ac49","Type":"ContainerStarted","Data":"d8be7dcf1e489a881ca8ec158ac6f0d640590fd54799a321c88795ee1085b376"} Oct 14 07:22:25 crc kubenswrapper[4870]: I1014 07:22:25.289125 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" event={"ID":"874a5e45-dffd-4d17-b609-b7d3ed2eab07","Type":"ContainerStarted","Data":"45dee68d970a494a60daa9fc176a487f297e0f291860479dd1d60c51287b32b3"} Oct 14 07:22:25 crc kubenswrapper[4870]: I1014 07:22:25.326624 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 14 07:22:25 crc kubenswrapper[4870]: I1014 07:22:25.533575 4870 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="4a361161-c100-49cb-a4ee-19563b724033" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.190:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 14 07:22:25 crc kubenswrapper[4870]: I1014 07:22:25.533605 4870 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="4a361161-c100-49cb-a4ee-19563b724033" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.190:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 14 07:22:26 crc kubenswrapper[4870]: I1014 07:22:26.297084 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b84cf82f-2718-4712-912e-5cf227d2ac49","Type":"ContainerStarted","Data":"afdc51519c47922b5bfa269bb1c40631b673cbe7369edef8c1bbd214ecf43404"} Oct 14 07:22:26 crc kubenswrapper[4870]: I1014 07:22:26.297572 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b84cf82f-2718-4712-912e-5cf227d2ac49","Type":"ContainerStarted","Data":"329969d702bc03fa3e7cbff0f2848fe8513a9c5e0faffdf1fe3af56ec9824334"} Oct 14 07:22:27 crc kubenswrapper[4870]: I1014 07:22:27.318208 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b84cf82f-2718-4712-912e-5cf227d2ac49","Type":"ContainerStarted","Data":"e4b4c364010cd10dadc15bc77a031557a06d832f6383878ace726fe6521b31a8"} Oct 14 07:22:27 crc kubenswrapper[4870]: I1014 07:22:27.540860 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 14 07:22:27 crc kubenswrapper[4870]: I1014 07:22:27.540947 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 14 07:22:28 crc kubenswrapper[4870]: I1014 07:22:28.622631 4870 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="2688cfd1-fbb5-49bb-b13e-a140ef4ed017" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.192:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 14 07:22:28 crc kubenswrapper[4870]: I1014 07:22:28.622644 4870 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="2688cfd1-fbb5-49bb-b13e-a140ef4ed017" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.192:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 14 07:22:29 crc kubenswrapper[4870]: I1014 07:22:29.342828 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b84cf82f-2718-4712-912e-5cf227d2ac49","Type":"ContainerStarted","Data":"ef05451915593f58eb36ee5eaabc7426eb6d048cba46b0a2eaabf3ee6921e683"} Oct 14 07:22:29 crc kubenswrapper[4870]: I1014 07:22:29.343390 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 14 07:22:29 crc kubenswrapper[4870]: I1014 07:22:29.387108 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.792617081 podStartE2EDuration="6.387085192s" podCreationTimestamp="2025-10-14 07:22:23 +0000 UTC" firstStartedPulling="2025-10-14 07:22:24.520834143 +0000 UTC m=+1280.218194514" lastFinishedPulling="2025-10-14 07:22:28.115302244 +0000 UTC m=+1283.812662625" observedRunningTime="2025-10-14 07:22:29.379193057 +0000 UTC m=+1285.076553438" watchObservedRunningTime="2025-10-14 07:22:29.387085192 +0000 UTC m=+1285.084445573" Oct 14 07:22:29 crc kubenswrapper[4870]: I1014 07:22:29.646705 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Oct 14 07:22:31 crc kubenswrapper[4870]: I1014 07:22:31.698372 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Oct 14 07:22:34 crc kubenswrapper[4870]: I1014 07:22:34.524818 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 14 07:22:34 crc kubenswrapper[4870]: I1014 07:22:34.537201 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 14 07:22:34 crc kubenswrapper[4870]: I1014 07:22:34.537305 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 14 07:22:35 crc kubenswrapper[4870]: I1014 07:22:35.441605 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 14 07:22:37 crc kubenswrapper[4870]: I1014 07:22:37.459753 4870 generic.go:334] "Generic (PLEG): container finished" podID="ed7c5b6f-eb0a-44fc-a6a3-6f43909b6406" containerID="4d1574b6ecc78807ca0e045857f0ad61906f9011616d9b3dbeba62e89d80183d" exitCode=137 Oct 14 07:22:37 crc kubenswrapper[4870]: I1014 07:22:37.459890 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"ed7c5b6f-eb0a-44fc-a6a3-6f43909b6406","Type":"ContainerDied","Data":"4d1574b6ecc78807ca0e045857f0ad61906f9011616d9b3dbeba62e89d80183d"} Oct 14 07:22:37 crc kubenswrapper[4870]: I1014 07:22:37.461595 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"ed7c5b6f-eb0a-44fc-a6a3-6f43909b6406","Type":"ContainerDied","Data":"6a1a19a25ac1b542a719c1ec08cc06c9b012fd4a776ab1490d6937b80b9cf21f"} Oct 14 07:22:37 crc kubenswrapper[4870]: I1014 07:22:37.461634 4870 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6a1a19a25ac1b542a719c1ec08cc06c9b012fd4a776ab1490d6937b80b9cf21f" Oct 14 07:22:37 crc kubenswrapper[4870]: I1014 07:22:37.471300 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 14 07:22:37 crc kubenswrapper[4870]: I1014 07:22:37.544930 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 14 07:22:37 crc kubenswrapper[4870]: I1014 07:22:37.545017 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 14 07:22:37 crc kubenswrapper[4870]: I1014 07:22:37.545951 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 14 07:22:37 crc kubenswrapper[4870]: I1014 07:22:37.545985 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 14 07:22:37 crc kubenswrapper[4870]: I1014 07:22:37.552628 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 14 07:22:37 crc kubenswrapper[4870]: I1014 07:22:37.558214 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 14 07:22:37 crc kubenswrapper[4870]: I1014 07:22:37.585686 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ed7c5b6f-eb0a-44fc-a6a3-6f43909b6406-config-data\") pod \"ed7c5b6f-eb0a-44fc-a6a3-6f43909b6406\" (UID: \"ed7c5b6f-eb0a-44fc-a6a3-6f43909b6406\") " Oct 14 07:22:37 crc kubenswrapper[4870]: I1014 07:22:37.585838 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed7c5b6f-eb0a-44fc-a6a3-6f43909b6406-combined-ca-bundle\") pod \"ed7c5b6f-eb0a-44fc-a6a3-6f43909b6406\" (UID: \"ed7c5b6f-eb0a-44fc-a6a3-6f43909b6406\") " Oct 14 07:22:37 crc kubenswrapper[4870]: I1014 07:22:37.586058 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hflfq\" (UniqueName: \"kubernetes.io/projected/ed7c5b6f-eb0a-44fc-a6a3-6f43909b6406-kube-api-access-hflfq\") pod \"ed7c5b6f-eb0a-44fc-a6a3-6f43909b6406\" (UID: \"ed7c5b6f-eb0a-44fc-a6a3-6f43909b6406\") " Oct 14 07:22:37 crc kubenswrapper[4870]: I1014 07:22:37.605846 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ed7c5b6f-eb0a-44fc-a6a3-6f43909b6406-kube-api-access-hflfq" (OuterVolumeSpecName: "kube-api-access-hflfq") pod "ed7c5b6f-eb0a-44fc-a6a3-6f43909b6406" (UID: "ed7c5b6f-eb0a-44fc-a6a3-6f43909b6406"). InnerVolumeSpecName "kube-api-access-hflfq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:22:37 crc kubenswrapper[4870]: I1014 07:22:37.624804 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ed7c5b6f-eb0a-44fc-a6a3-6f43909b6406-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ed7c5b6f-eb0a-44fc-a6a3-6f43909b6406" (UID: "ed7c5b6f-eb0a-44fc-a6a3-6f43909b6406"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:22:37 crc kubenswrapper[4870]: I1014 07:22:37.629366 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ed7c5b6f-eb0a-44fc-a6a3-6f43909b6406-config-data" (OuterVolumeSpecName: "config-data") pod "ed7c5b6f-eb0a-44fc-a6a3-6f43909b6406" (UID: "ed7c5b6f-eb0a-44fc-a6a3-6f43909b6406"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:22:37 crc kubenswrapper[4870]: I1014 07:22:37.708044 4870 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ed7c5b6f-eb0a-44fc-a6a3-6f43909b6406-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 07:22:37 crc kubenswrapper[4870]: I1014 07:22:37.708077 4870 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed7c5b6f-eb0a-44fc-a6a3-6f43909b6406-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 07:22:37 crc kubenswrapper[4870]: I1014 07:22:37.708095 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hflfq\" (UniqueName: \"kubernetes.io/projected/ed7c5b6f-eb0a-44fc-a6a3-6f43909b6406-kube-api-access-hflfq\") on node \"crc\" DevicePath \"\"" Oct 14 07:22:37 crc kubenswrapper[4870]: I1014 07:22:37.751166 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7cf4d694c5-wjxfs"] Oct 14 07:22:37 crc kubenswrapper[4870]: E1014 07:22:37.751593 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed7c5b6f-eb0a-44fc-a6a3-6f43909b6406" containerName="nova-cell1-novncproxy-novncproxy" Oct 14 07:22:37 crc kubenswrapper[4870]: I1014 07:22:37.751610 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed7c5b6f-eb0a-44fc-a6a3-6f43909b6406" containerName="nova-cell1-novncproxy-novncproxy" Oct 14 07:22:37 crc kubenswrapper[4870]: I1014 07:22:37.751805 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="ed7c5b6f-eb0a-44fc-a6a3-6f43909b6406" containerName="nova-cell1-novncproxy-novncproxy" Oct 14 07:22:37 crc kubenswrapper[4870]: I1014 07:22:37.752804 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7cf4d694c5-wjxfs" Oct 14 07:22:37 crc kubenswrapper[4870]: I1014 07:22:37.766871 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7cf4d694c5-wjxfs"] Oct 14 07:22:37 crc kubenswrapper[4870]: I1014 07:22:37.912298 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/751daf42-b5c1-4318-a15b-d5911c9d61ea-ovsdbserver-sb\") pod \"dnsmasq-dns-7cf4d694c5-wjxfs\" (UID: \"751daf42-b5c1-4318-a15b-d5911c9d61ea\") " pod="openstack/dnsmasq-dns-7cf4d694c5-wjxfs" Oct 14 07:22:37 crc kubenswrapper[4870]: I1014 07:22:37.912701 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s8f8j\" (UniqueName: \"kubernetes.io/projected/751daf42-b5c1-4318-a15b-d5911c9d61ea-kube-api-access-s8f8j\") pod \"dnsmasq-dns-7cf4d694c5-wjxfs\" (UID: \"751daf42-b5c1-4318-a15b-d5911c9d61ea\") " pod="openstack/dnsmasq-dns-7cf4d694c5-wjxfs" Oct 14 07:22:37 crc kubenswrapper[4870]: I1014 07:22:37.912732 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/751daf42-b5c1-4318-a15b-d5911c9d61ea-ovsdbserver-nb\") pod \"dnsmasq-dns-7cf4d694c5-wjxfs\" (UID: \"751daf42-b5c1-4318-a15b-d5911c9d61ea\") " pod="openstack/dnsmasq-dns-7cf4d694c5-wjxfs" Oct 14 07:22:37 crc kubenswrapper[4870]: I1014 07:22:37.912789 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/751daf42-b5c1-4318-a15b-d5911c9d61ea-config\") pod \"dnsmasq-dns-7cf4d694c5-wjxfs\" (UID: \"751daf42-b5c1-4318-a15b-d5911c9d61ea\") " pod="openstack/dnsmasq-dns-7cf4d694c5-wjxfs" Oct 14 07:22:37 crc kubenswrapper[4870]: I1014 07:22:37.912980 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/751daf42-b5c1-4318-a15b-d5911c9d61ea-dns-svc\") pod \"dnsmasq-dns-7cf4d694c5-wjxfs\" (UID: \"751daf42-b5c1-4318-a15b-d5911c9d61ea\") " pod="openstack/dnsmasq-dns-7cf4d694c5-wjxfs" Oct 14 07:22:37 crc kubenswrapper[4870]: I1014 07:22:37.913150 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/751daf42-b5c1-4318-a15b-d5911c9d61ea-dns-swift-storage-0\") pod \"dnsmasq-dns-7cf4d694c5-wjxfs\" (UID: \"751daf42-b5c1-4318-a15b-d5911c9d61ea\") " pod="openstack/dnsmasq-dns-7cf4d694c5-wjxfs" Oct 14 07:22:38 crc kubenswrapper[4870]: I1014 07:22:38.014715 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/751daf42-b5c1-4318-a15b-d5911c9d61ea-dns-swift-storage-0\") pod \"dnsmasq-dns-7cf4d694c5-wjxfs\" (UID: \"751daf42-b5c1-4318-a15b-d5911c9d61ea\") " pod="openstack/dnsmasq-dns-7cf4d694c5-wjxfs" Oct 14 07:22:38 crc kubenswrapper[4870]: I1014 07:22:38.014815 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/751daf42-b5c1-4318-a15b-d5911c9d61ea-ovsdbserver-sb\") pod \"dnsmasq-dns-7cf4d694c5-wjxfs\" (UID: \"751daf42-b5c1-4318-a15b-d5911c9d61ea\") " pod="openstack/dnsmasq-dns-7cf4d694c5-wjxfs" Oct 14 07:22:38 crc kubenswrapper[4870]: I1014 07:22:38.014850 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s8f8j\" (UniqueName: \"kubernetes.io/projected/751daf42-b5c1-4318-a15b-d5911c9d61ea-kube-api-access-s8f8j\") pod \"dnsmasq-dns-7cf4d694c5-wjxfs\" (UID: \"751daf42-b5c1-4318-a15b-d5911c9d61ea\") " pod="openstack/dnsmasq-dns-7cf4d694c5-wjxfs" Oct 14 07:22:38 crc kubenswrapper[4870]: I1014 07:22:38.014867 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/751daf42-b5c1-4318-a15b-d5911c9d61ea-ovsdbserver-nb\") pod \"dnsmasq-dns-7cf4d694c5-wjxfs\" (UID: \"751daf42-b5c1-4318-a15b-d5911c9d61ea\") " pod="openstack/dnsmasq-dns-7cf4d694c5-wjxfs" Oct 14 07:22:38 crc kubenswrapper[4870]: I1014 07:22:38.014921 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/751daf42-b5c1-4318-a15b-d5911c9d61ea-config\") pod \"dnsmasq-dns-7cf4d694c5-wjxfs\" (UID: \"751daf42-b5c1-4318-a15b-d5911c9d61ea\") " pod="openstack/dnsmasq-dns-7cf4d694c5-wjxfs" Oct 14 07:22:38 crc kubenswrapper[4870]: I1014 07:22:38.014961 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/751daf42-b5c1-4318-a15b-d5911c9d61ea-dns-svc\") pod \"dnsmasq-dns-7cf4d694c5-wjxfs\" (UID: \"751daf42-b5c1-4318-a15b-d5911c9d61ea\") " pod="openstack/dnsmasq-dns-7cf4d694c5-wjxfs" Oct 14 07:22:38 crc kubenswrapper[4870]: I1014 07:22:38.015663 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/751daf42-b5c1-4318-a15b-d5911c9d61ea-dns-swift-storage-0\") pod \"dnsmasq-dns-7cf4d694c5-wjxfs\" (UID: \"751daf42-b5c1-4318-a15b-d5911c9d61ea\") " pod="openstack/dnsmasq-dns-7cf4d694c5-wjxfs" Oct 14 07:22:38 crc kubenswrapper[4870]: I1014 07:22:38.015933 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/751daf42-b5c1-4318-a15b-d5911c9d61ea-ovsdbserver-sb\") pod \"dnsmasq-dns-7cf4d694c5-wjxfs\" (UID: \"751daf42-b5c1-4318-a15b-d5911c9d61ea\") " pod="openstack/dnsmasq-dns-7cf4d694c5-wjxfs" Oct 14 07:22:38 crc kubenswrapper[4870]: I1014 07:22:38.016175 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/751daf42-b5c1-4318-a15b-d5911c9d61ea-dns-svc\") pod \"dnsmasq-dns-7cf4d694c5-wjxfs\" (UID: \"751daf42-b5c1-4318-a15b-d5911c9d61ea\") " pod="openstack/dnsmasq-dns-7cf4d694c5-wjxfs" Oct 14 07:22:38 crc kubenswrapper[4870]: I1014 07:22:38.016361 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/751daf42-b5c1-4318-a15b-d5911c9d61ea-config\") pod \"dnsmasq-dns-7cf4d694c5-wjxfs\" (UID: \"751daf42-b5c1-4318-a15b-d5911c9d61ea\") " pod="openstack/dnsmasq-dns-7cf4d694c5-wjxfs" Oct 14 07:22:38 crc kubenswrapper[4870]: I1014 07:22:38.017848 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/751daf42-b5c1-4318-a15b-d5911c9d61ea-ovsdbserver-nb\") pod \"dnsmasq-dns-7cf4d694c5-wjxfs\" (UID: \"751daf42-b5c1-4318-a15b-d5911c9d61ea\") " pod="openstack/dnsmasq-dns-7cf4d694c5-wjxfs" Oct 14 07:22:38 crc kubenswrapper[4870]: I1014 07:22:38.039646 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s8f8j\" (UniqueName: \"kubernetes.io/projected/751daf42-b5c1-4318-a15b-d5911c9d61ea-kube-api-access-s8f8j\") pod \"dnsmasq-dns-7cf4d694c5-wjxfs\" (UID: \"751daf42-b5c1-4318-a15b-d5911c9d61ea\") " pod="openstack/dnsmasq-dns-7cf4d694c5-wjxfs" Oct 14 07:22:38 crc kubenswrapper[4870]: I1014 07:22:38.081547 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7cf4d694c5-wjxfs" Oct 14 07:22:38 crc kubenswrapper[4870]: I1014 07:22:38.469014 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 14 07:22:38 crc kubenswrapper[4870]: I1014 07:22:38.517023 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 14 07:22:38 crc kubenswrapper[4870]: I1014 07:22:38.527924 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 14 07:22:38 crc kubenswrapper[4870]: I1014 07:22:38.540697 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 14 07:22:38 crc kubenswrapper[4870]: I1014 07:22:38.542487 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 14 07:22:38 crc kubenswrapper[4870]: I1014 07:22:38.548615 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Oct 14 07:22:38 crc kubenswrapper[4870]: I1014 07:22:38.549152 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Oct 14 07:22:38 crc kubenswrapper[4870]: I1014 07:22:38.549327 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Oct 14 07:22:38 crc kubenswrapper[4870]: I1014 07:22:38.550988 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 14 07:22:38 crc kubenswrapper[4870]: I1014 07:22:38.604871 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7cf4d694c5-wjxfs"] Oct 14 07:22:38 crc kubenswrapper[4870]: W1014 07:22:38.609409 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod751daf42_b5c1_4318_a15b_d5911c9d61ea.slice/crio-9f72086a1f5451771a6f36acdca98d32c056502ec4a9d0b1a7f2a1edc601460b WatchSource:0}: Error finding container 9f72086a1f5451771a6f36acdca98d32c056502ec4a9d0b1a7f2a1edc601460b: Status 404 returned error can't find the container with id 9f72086a1f5451771a6f36acdca98d32c056502ec4a9d0b1a7f2a1edc601460b Oct 14 07:22:38 crc kubenswrapper[4870]: I1014 07:22:38.733602 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8abede26-36e8-46ec-a5c4-6cd53319133a-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"8abede26-36e8-46ec-a5c4-6cd53319133a\") " pod="openstack/nova-cell1-novncproxy-0" Oct 14 07:22:38 crc kubenswrapper[4870]: I1014 07:22:38.733722 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/8abede26-36e8-46ec-a5c4-6cd53319133a-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"8abede26-36e8-46ec-a5c4-6cd53319133a\") " pod="openstack/nova-cell1-novncproxy-0" Oct 14 07:22:38 crc kubenswrapper[4870]: I1014 07:22:38.733790 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/8abede26-36e8-46ec-a5c4-6cd53319133a-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"8abede26-36e8-46ec-a5c4-6cd53319133a\") " pod="openstack/nova-cell1-novncproxy-0" Oct 14 07:22:38 crc kubenswrapper[4870]: I1014 07:22:38.733911 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8abede26-36e8-46ec-a5c4-6cd53319133a-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"8abede26-36e8-46ec-a5c4-6cd53319133a\") " pod="openstack/nova-cell1-novncproxy-0" Oct 14 07:22:38 crc kubenswrapper[4870]: I1014 07:22:38.734220 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zpqm8\" (UniqueName: \"kubernetes.io/projected/8abede26-36e8-46ec-a5c4-6cd53319133a-kube-api-access-zpqm8\") pod \"nova-cell1-novncproxy-0\" (UID: \"8abede26-36e8-46ec-a5c4-6cd53319133a\") " pod="openstack/nova-cell1-novncproxy-0" Oct 14 07:22:38 crc kubenswrapper[4870]: I1014 07:22:38.836662 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zpqm8\" (UniqueName: \"kubernetes.io/projected/8abede26-36e8-46ec-a5c4-6cd53319133a-kube-api-access-zpqm8\") pod \"nova-cell1-novncproxy-0\" (UID: \"8abede26-36e8-46ec-a5c4-6cd53319133a\") " pod="openstack/nova-cell1-novncproxy-0" Oct 14 07:22:38 crc kubenswrapper[4870]: I1014 07:22:38.837037 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8abede26-36e8-46ec-a5c4-6cd53319133a-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"8abede26-36e8-46ec-a5c4-6cd53319133a\") " pod="openstack/nova-cell1-novncproxy-0" Oct 14 07:22:38 crc kubenswrapper[4870]: I1014 07:22:38.837102 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/8abede26-36e8-46ec-a5c4-6cd53319133a-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"8abede26-36e8-46ec-a5c4-6cd53319133a\") " pod="openstack/nova-cell1-novncproxy-0" Oct 14 07:22:38 crc kubenswrapper[4870]: I1014 07:22:38.837145 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/8abede26-36e8-46ec-a5c4-6cd53319133a-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"8abede26-36e8-46ec-a5c4-6cd53319133a\") " pod="openstack/nova-cell1-novncproxy-0" Oct 14 07:22:38 crc kubenswrapper[4870]: I1014 07:22:38.837198 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8abede26-36e8-46ec-a5c4-6cd53319133a-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"8abede26-36e8-46ec-a5c4-6cd53319133a\") " pod="openstack/nova-cell1-novncproxy-0" Oct 14 07:22:38 crc kubenswrapper[4870]: I1014 07:22:38.844468 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8abede26-36e8-46ec-a5c4-6cd53319133a-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"8abede26-36e8-46ec-a5c4-6cd53319133a\") " pod="openstack/nova-cell1-novncproxy-0" Oct 14 07:22:38 crc kubenswrapper[4870]: I1014 07:22:38.852396 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/8abede26-36e8-46ec-a5c4-6cd53319133a-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"8abede26-36e8-46ec-a5c4-6cd53319133a\") " pod="openstack/nova-cell1-novncproxy-0" Oct 14 07:22:38 crc kubenswrapper[4870]: I1014 07:22:38.852684 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8abede26-36e8-46ec-a5c4-6cd53319133a-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"8abede26-36e8-46ec-a5c4-6cd53319133a\") " pod="openstack/nova-cell1-novncproxy-0" Oct 14 07:22:38 crc kubenswrapper[4870]: I1014 07:22:38.853111 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/8abede26-36e8-46ec-a5c4-6cd53319133a-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"8abede26-36e8-46ec-a5c4-6cd53319133a\") " pod="openstack/nova-cell1-novncproxy-0" Oct 14 07:22:38 crc kubenswrapper[4870]: I1014 07:22:38.856120 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zpqm8\" (UniqueName: \"kubernetes.io/projected/8abede26-36e8-46ec-a5c4-6cd53319133a-kube-api-access-zpqm8\") pod \"nova-cell1-novncproxy-0\" (UID: \"8abede26-36e8-46ec-a5c4-6cd53319133a\") " pod="openstack/nova-cell1-novncproxy-0" Oct 14 07:22:38 crc kubenswrapper[4870]: I1014 07:22:38.871693 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 14 07:22:39 crc kubenswrapper[4870]: I1014 07:22:39.052254 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ed7c5b6f-eb0a-44fc-a6a3-6f43909b6406" path="/var/lib/kubelet/pods/ed7c5b6f-eb0a-44fc-a6a3-6f43909b6406/volumes" Oct 14 07:22:39 crc kubenswrapper[4870]: I1014 07:22:39.440045 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 14 07:22:39 crc kubenswrapper[4870]: W1014 07:22:39.443693 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8abede26_36e8_46ec_a5c4_6cd53319133a.slice/crio-5a4ce260cbfb825ad1358d2122d60517f96cbcdbd83613ef2c1f69ea0ec0bf41 WatchSource:0}: Error finding container 5a4ce260cbfb825ad1358d2122d60517f96cbcdbd83613ef2c1f69ea0ec0bf41: Status 404 returned error can't find the container with id 5a4ce260cbfb825ad1358d2122d60517f96cbcdbd83613ef2c1f69ea0ec0bf41 Oct 14 07:22:39 crc kubenswrapper[4870]: I1014 07:22:39.479005 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"8abede26-36e8-46ec-a5c4-6cd53319133a","Type":"ContainerStarted","Data":"5a4ce260cbfb825ad1358d2122d60517f96cbcdbd83613ef2c1f69ea0ec0bf41"} Oct 14 07:22:39 crc kubenswrapper[4870]: I1014 07:22:39.481047 4870 generic.go:334] "Generic (PLEG): container finished" podID="751daf42-b5c1-4318-a15b-d5911c9d61ea" containerID="739fd4dfae2ee47a08c26ba0aab10d4ad5944e6d3c334d62f3c9e4aaf95c2394" exitCode=0 Oct 14 07:22:39 crc kubenswrapper[4870]: I1014 07:22:39.482581 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cf4d694c5-wjxfs" event={"ID":"751daf42-b5c1-4318-a15b-d5911c9d61ea","Type":"ContainerDied","Data":"739fd4dfae2ee47a08c26ba0aab10d4ad5944e6d3c334d62f3c9e4aaf95c2394"} Oct 14 07:22:39 crc kubenswrapper[4870]: I1014 07:22:39.482613 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cf4d694c5-wjxfs" event={"ID":"751daf42-b5c1-4318-a15b-d5911c9d61ea","Type":"ContainerStarted","Data":"9f72086a1f5451771a6f36acdca98d32c056502ec4a9d0b1a7f2a1edc601460b"} Oct 14 07:22:39 crc kubenswrapper[4870]: I1014 07:22:39.707075 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 14 07:22:39 crc kubenswrapper[4870]: I1014 07:22:39.707645 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b84cf82f-2718-4712-912e-5cf227d2ac49" containerName="ceilometer-central-agent" containerID="cri-o://329969d702bc03fa3e7cbff0f2848fe8513a9c5e0faffdf1fe3af56ec9824334" gracePeriod=30 Oct 14 07:22:39 crc kubenswrapper[4870]: I1014 07:22:39.707786 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b84cf82f-2718-4712-912e-5cf227d2ac49" containerName="proxy-httpd" containerID="cri-o://ef05451915593f58eb36ee5eaabc7426eb6d048cba46b0a2eaabf3ee6921e683" gracePeriod=30 Oct 14 07:22:39 crc kubenswrapper[4870]: I1014 07:22:39.707821 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b84cf82f-2718-4712-912e-5cf227d2ac49" containerName="sg-core" containerID="cri-o://e4b4c364010cd10dadc15bc77a031557a06d832f6383878ace726fe6521b31a8" gracePeriod=30 Oct 14 07:22:39 crc kubenswrapper[4870]: I1014 07:22:39.707854 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b84cf82f-2718-4712-912e-5cf227d2ac49" containerName="ceilometer-notification-agent" containerID="cri-o://afdc51519c47922b5bfa269bb1c40631b673cbe7369edef8c1bbd214ecf43404" gracePeriod=30 Oct 14 07:22:39 crc kubenswrapper[4870]: I1014 07:22:39.714383 4870 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="b84cf82f-2718-4712-912e-5cf227d2ac49" containerName="proxy-httpd" probeResult="failure" output="Get \"https://10.217.0.195:3000/\": read tcp 10.217.0.2:45742->10.217.0.195:3000: read: connection reset by peer" Oct 14 07:22:40 crc kubenswrapper[4870]: I1014 07:22:40.503969 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"8abede26-36e8-46ec-a5c4-6cd53319133a","Type":"ContainerStarted","Data":"45200242c420e7ff44ca9c10b9cff44cc8bb4c9294423cc3e3e0ad235691f6b4"} Oct 14 07:22:40 crc kubenswrapper[4870]: I1014 07:22:40.507424 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cf4d694c5-wjxfs" event={"ID":"751daf42-b5c1-4318-a15b-d5911c9d61ea","Type":"ContainerStarted","Data":"e7f985e91ecb8ba5c0b4601a782047823bc6426fd3f278547d0ef69704ee2fcc"} Oct 14 07:22:40 crc kubenswrapper[4870]: I1014 07:22:40.507772 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7cf4d694c5-wjxfs" Oct 14 07:22:40 crc kubenswrapper[4870]: I1014 07:22:40.512960 4870 generic.go:334] "Generic (PLEG): container finished" podID="b84cf82f-2718-4712-912e-5cf227d2ac49" containerID="ef05451915593f58eb36ee5eaabc7426eb6d048cba46b0a2eaabf3ee6921e683" exitCode=0 Oct 14 07:22:40 crc kubenswrapper[4870]: I1014 07:22:40.512988 4870 generic.go:334] "Generic (PLEG): container finished" podID="b84cf82f-2718-4712-912e-5cf227d2ac49" containerID="e4b4c364010cd10dadc15bc77a031557a06d832f6383878ace726fe6521b31a8" exitCode=2 Oct 14 07:22:40 crc kubenswrapper[4870]: I1014 07:22:40.512998 4870 generic.go:334] "Generic (PLEG): container finished" podID="b84cf82f-2718-4712-912e-5cf227d2ac49" containerID="329969d702bc03fa3e7cbff0f2848fe8513a9c5e0faffdf1fe3af56ec9824334" exitCode=0 Oct 14 07:22:40 crc kubenswrapper[4870]: I1014 07:22:40.513021 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b84cf82f-2718-4712-912e-5cf227d2ac49","Type":"ContainerDied","Data":"ef05451915593f58eb36ee5eaabc7426eb6d048cba46b0a2eaabf3ee6921e683"} Oct 14 07:22:40 crc kubenswrapper[4870]: I1014 07:22:40.513041 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b84cf82f-2718-4712-912e-5cf227d2ac49","Type":"ContainerDied","Data":"e4b4c364010cd10dadc15bc77a031557a06d832f6383878ace726fe6521b31a8"} Oct 14 07:22:40 crc kubenswrapper[4870]: I1014 07:22:40.513053 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b84cf82f-2718-4712-912e-5cf227d2ac49","Type":"ContainerDied","Data":"329969d702bc03fa3e7cbff0f2848fe8513a9c5e0faffdf1fe3af56ec9824334"} Oct 14 07:22:40 crc kubenswrapper[4870]: I1014 07:22:40.525632 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.5256146360000002 podStartE2EDuration="2.525614636s" podCreationTimestamp="2025-10-14 07:22:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:22:40.520721892 +0000 UTC m=+1296.218082273" watchObservedRunningTime="2025-10-14 07:22:40.525614636 +0000 UTC m=+1296.222975007" Oct 14 07:22:40 crc kubenswrapper[4870]: I1014 07:22:40.553636 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7cf4d694c5-wjxfs" podStartSLOduration=3.553613201 podStartE2EDuration="3.553613201s" podCreationTimestamp="2025-10-14 07:22:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:22:40.543564546 +0000 UTC m=+1296.240924927" watchObservedRunningTime="2025-10-14 07:22:40.553613201 +0000 UTC m=+1296.250973572" Oct 14 07:22:40 crc kubenswrapper[4870]: I1014 07:22:40.583398 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 14 07:22:40 crc kubenswrapper[4870]: I1014 07:22:40.583626 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="2688cfd1-fbb5-49bb-b13e-a140ef4ed017" containerName="nova-api-log" containerID="cri-o://40a16de7c474abe785156f022176336a66c23c1c1750e3f876fdd77eea54ecfa" gracePeriod=30 Oct 14 07:22:40 crc kubenswrapper[4870]: I1014 07:22:40.583737 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="2688cfd1-fbb5-49bb-b13e-a140ef4ed017" containerName="nova-api-api" containerID="cri-o://7a6eec2a77f8c1ad24c5475d626f8ac6b4f062c417e46b252b6ad0d2df13bdf9" gracePeriod=30 Oct 14 07:22:41 crc kubenswrapper[4870]: I1014 07:22:41.522513 4870 generic.go:334] "Generic (PLEG): container finished" podID="2688cfd1-fbb5-49bb-b13e-a140ef4ed017" containerID="40a16de7c474abe785156f022176336a66c23c1c1750e3f876fdd77eea54ecfa" exitCode=143 Oct 14 07:22:41 crc kubenswrapper[4870]: I1014 07:22:41.524221 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2688cfd1-fbb5-49bb-b13e-a140ef4ed017","Type":"ContainerDied","Data":"40a16de7c474abe785156f022176336a66c23c1c1750e3f876fdd77eea54ecfa"} Oct 14 07:22:42 crc kubenswrapper[4870]: I1014 07:22:42.299351 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 14 07:22:42 crc kubenswrapper[4870]: I1014 07:22:42.398835 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b84cf82f-2718-4712-912e-5cf227d2ac49-config-data\") pod \"b84cf82f-2718-4712-912e-5cf227d2ac49\" (UID: \"b84cf82f-2718-4712-912e-5cf227d2ac49\") " Oct 14 07:22:42 crc kubenswrapper[4870]: I1014 07:22:42.398911 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b84cf82f-2718-4712-912e-5cf227d2ac49-scripts\") pod \"b84cf82f-2718-4712-912e-5cf227d2ac49\" (UID: \"b84cf82f-2718-4712-912e-5cf227d2ac49\") " Oct 14 07:22:42 crc kubenswrapper[4870]: I1014 07:22:42.399024 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b84cf82f-2718-4712-912e-5cf227d2ac49-sg-core-conf-yaml\") pod \"b84cf82f-2718-4712-912e-5cf227d2ac49\" (UID: \"b84cf82f-2718-4712-912e-5cf227d2ac49\") " Oct 14 07:22:42 crc kubenswrapper[4870]: I1014 07:22:42.399051 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/b84cf82f-2718-4712-912e-5cf227d2ac49-ceilometer-tls-certs\") pod \"b84cf82f-2718-4712-912e-5cf227d2ac49\" (UID: \"b84cf82f-2718-4712-912e-5cf227d2ac49\") " Oct 14 07:22:42 crc kubenswrapper[4870]: I1014 07:22:42.399076 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b84cf82f-2718-4712-912e-5cf227d2ac49-run-httpd\") pod \"b84cf82f-2718-4712-912e-5cf227d2ac49\" (UID: \"b84cf82f-2718-4712-912e-5cf227d2ac49\") " Oct 14 07:22:42 crc kubenswrapper[4870]: I1014 07:22:42.399100 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b84cf82f-2718-4712-912e-5cf227d2ac49-log-httpd\") pod \"b84cf82f-2718-4712-912e-5cf227d2ac49\" (UID: \"b84cf82f-2718-4712-912e-5cf227d2ac49\") " Oct 14 07:22:42 crc kubenswrapper[4870]: I1014 07:22:42.399135 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b84cf82f-2718-4712-912e-5cf227d2ac49-combined-ca-bundle\") pod \"b84cf82f-2718-4712-912e-5cf227d2ac49\" (UID: \"b84cf82f-2718-4712-912e-5cf227d2ac49\") " Oct 14 07:22:42 crc kubenswrapper[4870]: I1014 07:22:42.399539 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b84cf82f-2718-4712-912e-5cf227d2ac49-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "b84cf82f-2718-4712-912e-5cf227d2ac49" (UID: "b84cf82f-2718-4712-912e-5cf227d2ac49"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:22:42 crc kubenswrapper[4870]: I1014 07:22:42.399165 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7dz7f\" (UniqueName: \"kubernetes.io/projected/b84cf82f-2718-4712-912e-5cf227d2ac49-kube-api-access-7dz7f\") pod \"b84cf82f-2718-4712-912e-5cf227d2ac49\" (UID: \"b84cf82f-2718-4712-912e-5cf227d2ac49\") " Oct 14 07:22:42 crc kubenswrapper[4870]: I1014 07:22:42.399744 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b84cf82f-2718-4712-912e-5cf227d2ac49-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "b84cf82f-2718-4712-912e-5cf227d2ac49" (UID: "b84cf82f-2718-4712-912e-5cf227d2ac49"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:22:42 crc kubenswrapper[4870]: I1014 07:22:42.400272 4870 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b84cf82f-2718-4712-912e-5cf227d2ac49-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 14 07:22:42 crc kubenswrapper[4870]: I1014 07:22:42.400300 4870 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b84cf82f-2718-4712-912e-5cf227d2ac49-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 14 07:22:42 crc kubenswrapper[4870]: I1014 07:22:42.405350 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b84cf82f-2718-4712-912e-5cf227d2ac49-kube-api-access-7dz7f" (OuterVolumeSpecName: "kube-api-access-7dz7f") pod "b84cf82f-2718-4712-912e-5cf227d2ac49" (UID: "b84cf82f-2718-4712-912e-5cf227d2ac49"). InnerVolumeSpecName "kube-api-access-7dz7f". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:22:42 crc kubenswrapper[4870]: I1014 07:22:42.405746 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b84cf82f-2718-4712-912e-5cf227d2ac49-scripts" (OuterVolumeSpecName: "scripts") pod "b84cf82f-2718-4712-912e-5cf227d2ac49" (UID: "b84cf82f-2718-4712-912e-5cf227d2ac49"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:22:42 crc kubenswrapper[4870]: I1014 07:22:42.455743 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b84cf82f-2718-4712-912e-5cf227d2ac49-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "b84cf82f-2718-4712-912e-5cf227d2ac49" (UID: "b84cf82f-2718-4712-912e-5cf227d2ac49"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:22:42 crc kubenswrapper[4870]: I1014 07:22:42.458742 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b84cf82f-2718-4712-912e-5cf227d2ac49-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "b84cf82f-2718-4712-912e-5cf227d2ac49" (UID: "b84cf82f-2718-4712-912e-5cf227d2ac49"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:22:42 crc kubenswrapper[4870]: I1014 07:22:42.496276 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b84cf82f-2718-4712-912e-5cf227d2ac49-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b84cf82f-2718-4712-912e-5cf227d2ac49" (UID: "b84cf82f-2718-4712-912e-5cf227d2ac49"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:22:42 crc kubenswrapper[4870]: I1014 07:22:42.501399 4870 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b84cf82f-2718-4712-912e-5cf227d2ac49-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 07:22:42 crc kubenswrapper[4870]: I1014 07:22:42.501420 4870 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b84cf82f-2718-4712-912e-5cf227d2ac49-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 14 07:22:42 crc kubenswrapper[4870]: I1014 07:22:42.501431 4870 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/b84cf82f-2718-4712-912e-5cf227d2ac49-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 14 07:22:42 crc kubenswrapper[4870]: I1014 07:22:42.501477 4870 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b84cf82f-2718-4712-912e-5cf227d2ac49-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 07:22:42 crc kubenswrapper[4870]: I1014 07:22:42.501488 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7dz7f\" (UniqueName: \"kubernetes.io/projected/b84cf82f-2718-4712-912e-5cf227d2ac49-kube-api-access-7dz7f\") on node \"crc\" DevicePath \"\"" Oct 14 07:22:42 crc kubenswrapper[4870]: I1014 07:22:42.516564 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b84cf82f-2718-4712-912e-5cf227d2ac49-config-data" (OuterVolumeSpecName: "config-data") pod "b84cf82f-2718-4712-912e-5cf227d2ac49" (UID: "b84cf82f-2718-4712-912e-5cf227d2ac49"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:22:42 crc kubenswrapper[4870]: I1014 07:22:42.534620 4870 generic.go:334] "Generic (PLEG): container finished" podID="b84cf82f-2718-4712-912e-5cf227d2ac49" containerID="afdc51519c47922b5bfa269bb1c40631b673cbe7369edef8c1bbd214ecf43404" exitCode=0 Oct 14 07:22:42 crc kubenswrapper[4870]: I1014 07:22:42.534678 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b84cf82f-2718-4712-912e-5cf227d2ac49","Type":"ContainerDied","Data":"afdc51519c47922b5bfa269bb1c40631b673cbe7369edef8c1bbd214ecf43404"} Oct 14 07:22:42 crc kubenswrapper[4870]: I1014 07:22:42.534792 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 14 07:22:42 crc kubenswrapper[4870]: I1014 07:22:42.534809 4870 scope.go:117] "RemoveContainer" containerID="ef05451915593f58eb36ee5eaabc7426eb6d048cba46b0a2eaabf3ee6921e683" Oct 14 07:22:42 crc kubenswrapper[4870]: I1014 07:22:42.535547 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b84cf82f-2718-4712-912e-5cf227d2ac49","Type":"ContainerDied","Data":"d8be7dcf1e489a881ca8ec158ac6f0d640590fd54799a321c88795ee1085b376"} Oct 14 07:22:42 crc kubenswrapper[4870]: I1014 07:22:42.585885 4870 scope.go:117] "RemoveContainer" containerID="e4b4c364010cd10dadc15bc77a031557a06d832f6383878ace726fe6521b31a8" Oct 14 07:22:42 crc kubenswrapper[4870]: I1014 07:22:42.594629 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 14 07:22:42 crc kubenswrapper[4870]: I1014 07:22:42.602174 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 14 07:22:42 crc kubenswrapper[4870]: I1014 07:22:42.603095 4870 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b84cf82f-2718-4712-912e-5cf227d2ac49-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 07:22:42 crc kubenswrapper[4870]: I1014 07:22:42.606167 4870 scope.go:117] "RemoveContainer" containerID="afdc51519c47922b5bfa269bb1c40631b673cbe7369edef8c1bbd214ecf43404" Oct 14 07:22:42 crc kubenswrapper[4870]: I1014 07:22:42.617808 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 14 07:22:42 crc kubenswrapper[4870]: E1014 07:22:42.618143 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b84cf82f-2718-4712-912e-5cf227d2ac49" containerName="ceilometer-notification-agent" Oct 14 07:22:42 crc kubenswrapper[4870]: I1014 07:22:42.618159 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="b84cf82f-2718-4712-912e-5cf227d2ac49" containerName="ceilometer-notification-agent" Oct 14 07:22:42 crc kubenswrapper[4870]: E1014 07:22:42.618177 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b84cf82f-2718-4712-912e-5cf227d2ac49" containerName="sg-core" Oct 14 07:22:42 crc kubenswrapper[4870]: I1014 07:22:42.618184 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="b84cf82f-2718-4712-912e-5cf227d2ac49" containerName="sg-core" Oct 14 07:22:42 crc kubenswrapper[4870]: E1014 07:22:42.618221 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b84cf82f-2718-4712-912e-5cf227d2ac49" containerName="proxy-httpd" Oct 14 07:22:42 crc kubenswrapper[4870]: I1014 07:22:42.618230 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="b84cf82f-2718-4712-912e-5cf227d2ac49" containerName="proxy-httpd" Oct 14 07:22:42 crc kubenswrapper[4870]: E1014 07:22:42.618239 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b84cf82f-2718-4712-912e-5cf227d2ac49" containerName="ceilometer-central-agent" Oct 14 07:22:42 crc kubenswrapper[4870]: I1014 07:22:42.618245 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="b84cf82f-2718-4712-912e-5cf227d2ac49" containerName="ceilometer-central-agent" Oct 14 07:22:42 crc kubenswrapper[4870]: I1014 07:22:42.618403 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="b84cf82f-2718-4712-912e-5cf227d2ac49" containerName="ceilometer-central-agent" Oct 14 07:22:42 crc kubenswrapper[4870]: I1014 07:22:42.618417 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="b84cf82f-2718-4712-912e-5cf227d2ac49" containerName="ceilometer-notification-agent" Oct 14 07:22:42 crc kubenswrapper[4870]: I1014 07:22:42.618449 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="b84cf82f-2718-4712-912e-5cf227d2ac49" containerName="sg-core" Oct 14 07:22:42 crc kubenswrapper[4870]: I1014 07:22:42.618460 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="b84cf82f-2718-4712-912e-5cf227d2ac49" containerName="proxy-httpd" Oct 14 07:22:42 crc kubenswrapper[4870]: I1014 07:22:42.620297 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 14 07:22:42 crc kubenswrapper[4870]: I1014 07:22:42.624707 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 14 07:22:42 crc kubenswrapper[4870]: I1014 07:22:42.624729 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 14 07:22:42 crc kubenswrapper[4870]: I1014 07:22:42.624777 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Oct 14 07:22:42 crc kubenswrapper[4870]: I1014 07:22:42.641057 4870 scope.go:117] "RemoveContainer" containerID="329969d702bc03fa3e7cbff0f2848fe8513a9c5e0faffdf1fe3af56ec9824334" Oct 14 07:22:42 crc kubenswrapper[4870]: I1014 07:22:42.644847 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 14 07:22:42 crc kubenswrapper[4870]: I1014 07:22:42.672471 4870 scope.go:117] "RemoveContainer" containerID="ef05451915593f58eb36ee5eaabc7426eb6d048cba46b0a2eaabf3ee6921e683" Oct 14 07:22:42 crc kubenswrapper[4870]: E1014 07:22:42.672977 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ef05451915593f58eb36ee5eaabc7426eb6d048cba46b0a2eaabf3ee6921e683\": container with ID starting with ef05451915593f58eb36ee5eaabc7426eb6d048cba46b0a2eaabf3ee6921e683 not found: ID does not exist" containerID="ef05451915593f58eb36ee5eaabc7426eb6d048cba46b0a2eaabf3ee6921e683" Oct 14 07:22:42 crc kubenswrapper[4870]: I1014 07:22:42.673013 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ef05451915593f58eb36ee5eaabc7426eb6d048cba46b0a2eaabf3ee6921e683"} err="failed to get container status \"ef05451915593f58eb36ee5eaabc7426eb6d048cba46b0a2eaabf3ee6921e683\": rpc error: code = NotFound desc = could not find container \"ef05451915593f58eb36ee5eaabc7426eb6d048cba46b0a2eaabf3ee6921e683\": container with ID starting with ef05451915593f58eb36ee5eaabc7426eb6d048cba46b0a2eaabf3ee6921e683 not found: ID does not exist" Oct 14 07:22:42 crc kubenswrapper[4870]: I1014 07:22:42.673035 4870 scope.go:117] "RemoveContainer" containerID="e4b4c364010cd10dadc15bc77a031557a06d832f6383878ace726fe6521b31a8" Oct 14 07:22:42 crc kubenswrapper[4870]: E1014 07:22:42.673320 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e4b4c364010cd10dadc15bc77a031557a06d832f6383878ace726fe6521b31a8\": container with ID starting with e4b4c364010cd10dadc15bc77a031557a06d832f6383878ace726fe6521b31a8 not found: ID does not exist" containerID="e4b4c364010cd10dadc15bc77a031557a06d832f6383878ace726fe6521b31a8" Oct 14 07:22:42 crc kubenswrapper[4870]: I1014 07:22:42.673338 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e4b4c364010cd10dadc15bc77a031557a06d832f6383878ace726fe6521b31a8"} err="failed to get container status \"e4b4c364010cd10dadc15bc77a031557a06d832f6383878ace726fe6521b31a8\": rpc error: code = NotFound desc = could not find container \"e4b4c364010cd10dadc15bc77a031557a06d832f6383878ace726fe6521b31a8\": container with ID starting with e4b4c364010cd10dadc15bc77a031557a06d832f6383878ace726fe6521b31a8 not found: ID does not exist" Oct 14 07:22:42 crc kubenswrapper[4870]: I1014 07:22:42.673353 4870 scope.go:117] "RemoveContainer" containerID="afdc51519c47922b5bfa269bb1c40631b673cbe7369edef8c1bbd214ecf43404" Oct 14 07:22:42 crc kubenswrapper[4870]: E1014 07:22:42.673708 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"afdc51519c47922b5bfa269bb1c40631b673cbe7369edef8c1bbd214ecf43404\": container with ID starting with afdc51519c47922b5bfa269bb1c40631b673cbe7369edef8c1bbd214ecf43404 not found: ID does not exist" containerID="afdc51519c47922b5bfa269bb1c40631b673cbe7369edef8c1bbd214ecf43404" Oct 14 07:22:42 crc kubenswrapper[4870]: I1014 07:22:42.673740 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"afdc51519c47922b5bfa269bb1c40631b673cbe7369edef8c1bbd214ecf43404"} err="failed to get container status \"afdc51519c47922b5bfa269bb1c40631b673cbe7369edef8c1bbd214ecf43404\": rpc error: code = NotFound desc = could not find container \"afdc51519c47922b5bfa269bb1c40631b673cbe7369edef8c1bbd214ecf43404\": container with ID starting with afdc51519c47922b5bfa269bb1c40631b673cbe7369edef8c1bbd214ecf43404 not found: ID does not exist" Oct 14 07:22:42 crc kubenswrapper[4870]: I1014 07:22:42.673759 4870 scope.go:117] "RemoveContainer" containerID="329969d702bc03fa3e7cbff0f2848fe8513a9c5e0faffdf1fe3af56ec9824334" Oct 14 07:22:42 crc kubenswrapper[4870]: E1014 07:22:42.674136 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"329969d702bc03fa3e7cbff0f2848fe8513a9c5e0faffdf1fe3af56ec9824334\": container with ID starting with 329969d702bc03fa3e7cbff0f2848fe8513a9c5e0faffdf1fe3af56ec9824334 not found: ID does not exist" containerID="329969d702bc03fa3e7cbff0f2848fe8513a9c5e0faffdf1fe3af56ec9824334" Oct 14 07:22:42 crc kubenswrapper[4870]: I1014 07:22:42.674169 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"329969d702bc03fa3e7cbff0f2848fe8513a9c5e0faffdf1fe3af56ec9824334"} err="failed to get container status \"329969d702bc03fa3e7cbff0f2848fe8513a9c5e0faffdf1fe3af56ec9824334\": rpc error: code = NotFound desc = could not find container \"329969d702bc03fa3e7cbff0f2848fe8513a9c5e0faffdf1fe3af56ec9824334\": container with ID starting with 329969d702bc03fa3e7cbff0f2848fe8513a9c5e0faffdf1fe3af56ec9824334 not found: ID does not exist" Oct 14 07:22:42 crc kubenswrapper[4870]: I1014 07:22:42.704889 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/524b0b2a-77a2-4c05-be39-c8b97666ab11-log-httpd\") pod \"ceilometer-0\" (UID: \"524b0b2a-77a2-4c05-be39-c8b97666ab11\") " pod="openstack/ceilometer-0" Oct 14 07:22:42 crc kubenswrapper[4870]: I1014 07:22:42.704938 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/524b0b2a-77a2-4c05-be39-c8b97666ab11-run-httpd\") pod \"ceilometer-0\" (UID: \"524b0b2a-77a2-4c05-be39-c8b97666ab11\") " pod="openstack/ceilometer-0" Oct 14 07:22:42 crc kubenswrapper[4870]: I1014 07:22:42.704982 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/524b0b2a-77a2-4c05-be39-c8b97666ab11-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"524b0b2a-77a2-4c05-be39-c8b97666ab11\") " pod="openstack/ceilometer-0" Oct 14 07:22:42 crc kubenswrapper[4870]: I1014 07:22:42.705004 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/524b0b2a-77a2-4c05-be39-c8b97666ab11-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"524b0b2a-77a2-4c05-be39-c8b97666ab11\") " pod="openstack/ceilometer-0" Oct 14 07:22:42 crc kubenswrapper[4870]: I1014 07:22:42.705042 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/524b0b2a-77a2-4c05-be39-c8b97666ab11-config-data\") pod \"ceilometer-0\" (UID: \"524b0b2a-77a2-4c05-be39-c8b97666ab11\") " pod="openstack/ceilometer-0" Oct 14 07:22:42 crc kubenswrapper[4870]: I1014 07:22:42.705164 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gr66j\" (UniqueName: \"kubernetes.io/projected/524b0b2a-77a2-4c05-be39-c8b97666ab11-kube-api-access-gr66j\") pod \"ceilometer-0\" (UID: \"524b0b2a-77a2-4c05-be39-c8b97666ab11\") " pod="openstack/ceilometer-0" Oct 14 07:22:42 crc kubenswrapper[4870]: I1014 07:22:42.705209 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/524b0b2a-77a2-4c05-be39-c8b97666ab11-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"524b0b2a-77a2-4c05-be39-c8b97666ab11\") " pod="openstack/ceilometer-0" Oct 14 07:22:42 crc kubenswrapper[4870]: I1014 07:22:42.705243 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/524b0b2a-77a2-4c05-be39-c8b97666ab11-scripts\") pod \"ceilometer-0\" (UID: \"524b0b2a-77a2-4c05-be39-c8b97666ab11\") " pod="openstack/ceilometer-0" Oct 14 07:22:42 crc kubenswrapper[4870]: I1014 07:22:42.807051 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gr66j\" (UniqueName: \"kubernetes.io/projected/524b0b2a-77a2-4c05-be39-c8b97666ab11-kube-api-access-gr66j\") pod \"ceilometer-0\" (UID: \"524b0b2a-77a2-4c05-be39-c8b97666ab11\") " pod="openstack/ceilometer-0" Oct 14 07:22:42 crc kubenswrapper[4870]: I1014 07:22:42.807116 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/524b0b2a-77a2-4c05-be39-c8b97666ab11-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"524b0b2a-77a2-4c05-be39-c8b97666ab11\") " pod="openstack/ceilometer-0" Oct 14 07:22:42 crc kubenswrapper[4870]: I1014 07:22:42.807157 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/524b0b2a-77a2-4c05-be39-c8b97666ab11-scripts\") pod \"ceilometer-0\" (UID: \"524b0b2a-77a2-4c05-be39-c8b97666ab11\") " pod="openstack/ceilometer-0" Oct 14 07:22:42 crc kubenswrapper[4870]: I1014 07:22:42.807208 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/524b0b2a-77a2-4c05-be39-c8b97666ab11-log-httpd\") pod \"ceilometer-0\" (UID: \"524b0b2a-77a2-4c05-be39-c8b97666ab11\") " pod="openstack/ceilometer-0" Oct 14 07:22:42 crc kubenswrapper[4870]: I1014 07:22:42.807236 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/524b0b2a-77a2-4c05-be39-c8b97666ab11-run-httpd\") pod \"ceilometer-0\" (UID: \"524b0b2a-77a2-4c05-be39-c8b97666ab11\") " pod="openstack/ceilometer-0" Oct 14 07:22:42 crc kubenswrapper[4870]: I1014 07:22:42.807269 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/524b0b2a-77a2-4c05-be39-c8b97666ab11-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"524b0b2a-77a2-4c05-be39-c8b97666ab11\") " pod="openstack/ceilometer-0" Oct 14 07:22:42 crc kubenswrapper[4870]: I1014 07:22:42.807289 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/524b0b2a-77a2-4c05-be39-c8b97666ab11-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"524b0b2a-77a2-4c05-be39-c8b97666ab11\") " pod="openstack/ceilometer-0" Oct 14 07:22:42 crc kubenswrapper[4870]: I1014 07:22:42.807321 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/524b0b2a-77a2-4c05-be39-c8b97666ab11-config-data\") pod \"ceilometer-0\" (UID: \"524b0b2a-77a2-4c05-be39-c8b97666ab11\") " pod="openstack/ceilometer-0" Oct 14 07:22:42 crc kubenswrapper[4870]: I1014 07:22:42.808390 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/524b0b2a-77a2-4c05-be39-c8b97666ab11-log-httpd\") pod \"ceilometer-0\" (UID: \"524b0b2a-77a2-4c05-be39-c8b97666ab11\") " pod="openstack/ceilometer-0" Oct 14 07:22:42 crc kubenswrapper[4870]: I1014 07:22:42.808970 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/524b0b2a-77a2-4c05-be39-c8b97666ab11-run-httpd\") pod \"ceilometer-0\" (UID: \"524b0b2a-77a2-4c05-be39-c8b97666ab11\") " pod="openstack/ceilometer-0" Oct 14 07:22:42 crc kubenswrapper[4870]: I1014 07:22:42.811430 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/524b0b2a-77a2-4c05-be39-c8b97666ab11-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"524b0b2a-77a2-4c05-be39-c8b97666ab11\") " pod="openstack/ceilometer-0" Oct 14 07:22:42 crc kubenswrapper[4870]: I1014 07:22:42.811858 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/524b0b2a-77a2-4c05-be39-c8b97666ab11-config-data\") pod \"ceilometer-0\" (UID: \"524b0b2a-77a2-4c05-be39-c8b97666ab11\") " pod="openstack/ceilometer-0" Oct 14 07:22:42 crc kubenswrapper[4870]: I1014 07:22:42.812280 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/524b0b2a-77a2-4c05-be39-c8b97666ab11-scripts\") pod \"ceilometer-0\" (UID: \"524b0b2a-77a2-4c05-be39-c8b97666ab11\") " pod="openstack/ceilometer-0" Oct 14 07:22:42 crc kubenswrapper[4870]: I1014 07:22:42.814899 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/524b0b2a-77a2-4c05-be39-c8b97666ab11-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"524b0b2a-77a2-4c05-be39-c8b97666ab11\") " pod="openstack/ceilometer-0" Oct 14 07:22:42 crc kubenswrapper[4870]: I1014 07:22:42.816558 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/524b0b2a-77a2-4c05-be39-c8b97666ab11-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"524b0b2a-77a2-4c05-be39-c8b97666ab11\") " pod="openstack/ceilometer-0" Oct 14 07:22:42 crc kubenswrapper[4870]: I1014 07:22:42.838537 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gr66j\" (UniqueName: \"kubernetes.io/projected/524b0b2a-77a2-4c05-be39-c8b97666ab11-kube-api-access-gr66j\") pod \"ceilometer-0\" (UID: \"524b0b2a-77a2-4c05-be39-c8b97666ab11\") " pod="openstack/ceilometer-0" Oct 14 07:22:42 crc kubenswrapper[4870]: I1014 07:22:42.942480 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 14 07:22:43 crc kubenswrapper[4870]: I1014 07:22:43.047843 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b84cf82f-2718-4712-912e-5cf227d2ac49" path="/var/lib/kubelet/pods/b84cf82f-2718-4712-912e-5cf227d2ac49/volumes" Oct 14 07:22:43 crc kubenswrapper[4870]: I1014 07:22:43.380151 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 14 07:22:43 crc kubenswrapper[4870]: W1014 07:22:43.395709 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod524b0b2a_77a2_4c05_be39_c8b97666ab11.slice/crio-3e53f14bf4af7bc198470fc2fda53430a9cde12e3adbf27cbd98bbaf80b6393b WatchSource:0}: Error finding container 3e53f14bf4af7bc198470fc2fda53430a9cde12e3adbf27cbd98bbaf80b6393b: Status 404 returned error can't find the container with id 3e53f14bf4af7bc198470fc2fda53430a9cde12e3adbf27cbd98bbaf80b6393b Oct 14 07:22:43 crc kubenswrapper[4870]: I1014 07:22:43.553299 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"524b0b2a-77a2-4c05-be39-c8b97666ab11","Type":"ContainerStarted","Data":"3e53f14bf4af7bc198470fc2fda53430a9cde12e3adbf27cbd98bbaf80b6393b"} Oct 14 07:22:43 crc kubenswrapper[4870]: I1014 07:22:43.871953 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Oct 14 07:22:44 crc kubenswrapper[4870]: I1014 07:22:44.278549 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 14 07:22:44 crc kubenswrapper[4870]: I1014 07:22:44.460824 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kqtt7\" (UniqueName: \"kubernetes.io/projected/2688cfd1-fbb5-49bb-b13e-a140ef4ed017-kube-api-access-kqtt7\") pod \"2688cfd1-fbb5-49bb-b13e-a140ef4ed017\" (UID: \"2688cfd1-fbb5-49bb-b13e-a140ef4ed017\") " Oct 14 07:22:44 crc kubenswrapper[4870]: I1014 07:22:44.460960 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2688cfd1-fbb5-49bb-b13e-a140ef4ed017-combined-ca-bundle\") pod \"2688cfd1-fbb5-49bb-b13e-a140ef4ed017\" (UID: \"2688cfd1-fbb5-49bb-b13e-a140ef4ed017\") " Oct 14 07:22:44 crc kubenswrapper[4870]: I1014 07:22:44.461207 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2688cfd1-fbb5-49bb-b13e-a140ef4ed017-logs\") pod \"2688cfd1-fbb5-49bb-b13e-a140ef4ed017\" (UID: \"2688cfd1-fbb5-49bb-b13e-a140ef4ed017\") " Oct 14 07:22:44 crc kubenswrapper[4870]: I1014 07:22:44.461252 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2688cfd1-fbb5-49bb-b13e-a140ef4ed017-config-data\") pod \"2688cfd1-fbb5-49bb-b13e-a140ef4ed017\" (UID: \"2688cfd1-fbb5-49bb-b13e-a140ef4ed017\") " Oct 14 07:22:44 crc kubenswrapper[4870]: I1014 07:22:44.461864 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2688cfd1-fbb5-49bb-b13e-a140ef4ed017-logs" (OuterVolumeSpecName: "logs") pod "2688cfd1-fbb5-49bb-b13e-a140ef4ed017" (UID: "2688cfd1-fbb5-49bb-b13e-a140ef4ed017"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:22:44 crc kubenswrapper[4870]: I1014 07:22:44.467277 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2688cfd1-fbb5-49bb-b13e-a140ef4ed017-kube-api-access-kqtt7" (OuterVolumeSpecName: "kube-api-access-kqtt7") pod "2688cfd1-fbb5-49bb-b13e-a140ef4ed017" (UID: "2688cfd1-fbb5-49bb-b13e-a140ef4ed017"). InnerVolumeSpecName "kube-api-access-kqtt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:22:44 crc kubenswrapper[4870]: I1014 07:22:44.488206 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2688cfd1-fbb5-49bb-b13e-a140ef4ed017-config-data" (OuterVolumeSpecName: "config-data") pod "2688cfd1-fbb5-49bb-b13e-a140ef4ed017" (UID: "2688cfd1-fbb5-49bb-b13e-a140ef4ed017"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:22:44 crc kubenswrapper[4870]: I1014 07:22:44.508577 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2688cfd1-fbb5-49bb-b13e-a140ef4ed017-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2688cfd1-fbb5-49bb-b13e-a140ef4ed017" (UID: "2688cfd1-fbb5-49bb-b13e-a140ef4ed017"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:22:44 crc kubenswrapper[4870]: I1014 07:22:44.563469 4870 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2688cfd1-fbb5-49bb-b13e-a140ef4ed017-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 07:22:44 crc kubenswrapper[4870]: I1014 07:22:44.563544 4870 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2688cfd1-fbb5-49bb-b13e-a140ef4ed017-logs\") on node \"crc\" DevicePath \"\"" Oct 14 07:22:44 crc kubenswrapper[4870]: I1014 07:22:44.563568 4870 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2688cfd1-fbb5-49bb-b13e-a140ef4ed017-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 07:22:44 crc kubenswrapper[4870]: I1014 07:22:44.563582 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kqtt7\" (UniqueName: \"kubernetes.io/projected/2688cfd1-fbb5-49bb-b13e-a140ef4ed017-kube-api-access-kqtt7\") on node \"crc\" DevicePath \"\"" Oct 14 07:22:44 crc kubenswrapper[4870]: I1014 07:22:44.564896 4870 generic.go:334] "Generic (PLEG): container finished" podID="2688cfd1-fbb5-49bb-b13e-a140ef4ed017" containerID="7a6eec2a77f8c1ad24c5475d626f8ac6b4f062c417e46b252b6ad0d2df13bdf9" exitCode=0 Oct 14 07:22:44 crc kubenswrapper[4870]: I1014 07:22:44.564967 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2688cfd1-fbb5-49bb-b13e-a140ef4ed017","Type":"ContainerDied","Data":"7a6eec2a77f8c1ad24c5475d626f8ac6b4f062c417e46b252b6ad0d2df13bdf9"} Oct 14 07:22:44 crc kubenswrapper[4870]: I1014 07:22:44.564984 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 14 07:22:44 crc kubenswrapper[4870]: I1014 07:22:44.565000 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2688cfd1-fbb5-49bb-b13e-a140ef4ed017","Type":"ContainerDied","Data":"1418e6f00a15bb66ee2a7b4cbc6499a1dd4796549c8faac82f925a9d250e35a2"} Oct 14 07:22:44 crc kubenswrapper[4870]: I1014 07:22:44.565020 4870 scope.go:117] "RemoveContainer" containerID="7a6eec2a77f8c1ad24c5475d626f8ac6b4f062c417e46b252b6ad0d2df13bdf9" Oct 14 07:22:44 crc kubenswrapper[4870]: I1014 07:22:44.566835 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"524b0b2a-77a2-4c05-be39-c8b97666ab11","Type":"ContainerStarted","Data":"f24a75c62a58aefe115cf6dc0ab70e01b811b66e1369401528bc3134ca7d1cc4"} Oct 14 07:22:44 crc kubenswrapper[4870]: I1014 07:22:44.599825 4870 scope.go:117] "RemoveContainer" containerID="40a16de7c474abe785156f022176336a66c23c1c1750e3f876fdd77eea54ecfa" Oct 14 07:22:44 crc kubenswrapper[4870]: I1014 07:22:44.603691 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 14 07:22:44 crc kubenswrapper[4870]: I1014 07:22:44.622464 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 14 07:22:44 crc kubenswrapper[4870]: I1014 07:22:44.631035 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 14 07:22:44 crc kubenswrapper[4870]: E1014 07:22:44.631759 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2688cfd1-fbb5-49bb-b13e-a140ef4ed017" containerName="nova-api-log" Oct 14 07:22:44 crc kubenswrapper[4870]: I1014 07:22:44.631853 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="2688cfd1-fbb5-49bb-b13e-a140ef4ed017" containerName="nova-api-log" Oct 14 07:22:44 crc kubenswrapper[4870]: E1014 07:22:44.631980 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2688cfd1-fbb5-49bb-b13e-a140ef4ed017" containerName="nova-api-api" Oct 14 07:22:44 crc kubenswrapper[4870]: I1014 07:22:44.632067 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="2688cfd1-fbb5-49bb-b13e-a140ef4ed017" containerName="nova-api-api" Oct 14 07:22:44 crc kubenswrapper[4870]: I1014 07:22:44.632555 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="2688cfd1-fbb5-49bb-b13e-a140ef4ed017" containerName="nova-api-api" Oct 14 07:22:44 crc kubenswrapper[4870]: I1014 07:22:44.632676 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="2688cfd1-fbb5-49bb-b13e-a140ef4ed017" containerName="nova-api-log" Oct 14 07:22:44 crc kubenswrapper[4870]: I1014 07:22:44.634515 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 14 07:22:44 crc kubenswrapper[4870]: I1014 07:22:44.637594 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 14 07:22:44 crc kubenswrapper[4870]: I1014 07:22:44.637887 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Oct 14 07:22:44 crc kubenswrapper[4870]: I1014 07:22:44.637896 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Oct 14 07:22:44 crc kubenswrapper[4870]: I1014 07:22:44.638310 4870 scope.go:117] "RemoveContainer" containerID="7a6eec2a77f8c1ad24c5475d626f8ac6b4f062c417e46b252b6ad0d2df13bdf9" Oct 14 07:22:44 crc kubenswrapper[4870]: E1014 07:22:44.645418 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7a6eec2a77f8c1ad24c5475d626f8ac6b4f062c417e46b252b6ad0d2df13bdf9\": container with ID starting with 7a6eec2a77f8c1ad24c5475d626f8ac6b4f062c417e46b252b6ad0d2df13bdf9 not found: ID does not exist" containerID="7a6eec2a77f8c1ad24c5475d626f8ac6b4f062c417e46b252b6ad0d2df13bdf9" Oct 14 07:22:44 crc kubenswrapper[4870]: I1014 07:22:44.645466 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7a6eec2a77f8c1ad24c5475d626f8ac6b4f062c417e46b252b6ad0d2df13bdf9"} err="failed to get container status \"7a6eec2a77f8c1ad24c5475d626f8ac6b4f062c417e46b252b6ad0d2df13bdf9\": rpc error: code = NotFound desc = could not find container \"7a6eec2a77f8c1ad24c5475d626f8ac6b4f062c417e46b252b6ad0d2df13bdf9\": container with ID starting with 7a6eec2a77f8c1ad24c5475d626f8ac6b4f062c417e46b252b6ad0d2df13bdf9 not found: ID does not exist" Oct 14 07:22:44 crc kubenswrapper[4870]: I1014 07:22:44.645491 4870 scope.go:117] "RemoveContainer" containerID="40a16de7c474abe785156f022176336a66c23c1c1750e3f876fdd77eea54ecfa" Oct 14 07:22:44 crc kubenswrapper[4870]: E1014 07:22:44.646016 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"40a16de7c474abe785156f022176336a66c23c1c1750e3f876fdd77eea54ecfa\": container with ID starting with 40a16de7c474abe785156f022176336a66c23c1c1750e3f876fdd77eea54ecfa not found: ID does not exist" containerID="40a16de7c474abe785156f022176336a66c23c1c1750e3f876fdd77eea54ecfa" Oct 14 07:22:44 crc kubenswrapper[4870]: I1014 07:22:44.646061 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"40a16de7c474abe785156f022176336a66c23c1c1750e3f876fdd77eea54ecfa"} err="failed to get container status \"40a16de7c474abe785156f022176336a66c23c1c1750e3f876fdd77eea54ecfa\": rpc error: code = NotFound desc = could not find container \"40a16de7c474abe785156f022176336a66c23c1c1750e3f876fdd77eea54ecfa\": container with ID starting with 40a16de7c474abe785156f022176336a66c23c1c1750e3f876fdd77eea54ecfa not found: ID does not exist" Oct 14 07:22:44 crc kubenswrapper[4870]: I1014 07:22:44.662938 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 14 07:22:44 crc kubenswrapper[4870]: I1014 07:22:44.768083 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1f09945e-c7d1-4bfe-a8d7-7421b1a590ec-config-data\") pod \"nova-api-0\" (UID: \"1f09945e-c7d1-4bfe-a8d7-7421b1a590ec\") " pod="openstack/nova-api-0" Oct 14 07:22:44 crc kubenswrapper[4870]: I1014 07:22:44.768147 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1f09945e-c7d1-4bfe-a8d7-7421b1a590ec-logs\") pod \"nova-api-0\" (UID: \"1f09945e-c7d1-4bfe-a8d7-7421b1a590ec\") " pod="openstack/nova-api-0" Oct 14 07:22:44 crc kubenswrapper[4870]: I1014 07:22:44.768185 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1f09945e-c7d1-4bfe-a8d7-7421b1a590ec-public-tls-certs\") pod \"nova-api-0\" (UID: \"1f09945e-c7d1-4bfe-a8d7-7421b1a590ec\") " pod="openstack/nova-api-0" Oct 14 07:22:44 crc kubenswrapper[4870]: I1014 07:22:44.768327 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k52mj\" (UniqueName: \"kubernetes.io/projected/1f09945e-c7d1-4bfe-a8d7-7421b1a590ec-kube-api-access-k52mj\") pod \"nova-api-0\" (UID: \"1f09945e-c7d1-4bfe-a8d7-7421b1a590ec\") " pod="openstack/nova-api-0" Oct 14 07:22:44 crc kubenswrapper[4870]: I1014 07:22:44.768426 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1f09945e-c7d1-4bfe-a8d7-7421b1a590ec-internal-tls-certs\") pod \"nova-api-0\" (UID: \"1f09945e-c7d1-4bfe-a8d7-7421b1a590ec\") " pod="openstack/nova-api-0" Oct 14 07:22:44 crc kubenswrapper[4870]: I1014 07:22:44.768557 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1f09945e-c7d1-4bfe-a8d7-7421b1a590ec-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"1f09945e-c7d1-4bfe-a8d7-7421b1a590ec\") " pod="openstack/nova-api-0" Oct 14 07:22:44 crc kubenswrapper[4870]: I1014 07:22:44.870369 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1f09945e-c7d1-4bfe-a8d7-7421b1a590ec-config-data\") pod \"nova-api-0\" (UID: \"1f09945e-c7d1-4bfe-a8d7-7421b1a590ec\") " pod="openstack/nova-api-0" Oct 14 07:22:44 crc kubenswrapper[4870]: I1014 07:22:44.870433 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1f09945e-c7d1-4bfe-a8d7-7421b1a590ec-logs\") pod \"nova-api-0\" (UID: \"1f09945e-c7d1-4bfe-a8d7-7421b1a590ec\") " pod="openstack/nova-api-0" Oct 14 07:22:44 crc kubenswrapper[4870]: I1014 07:22:44.870483 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1f09945e-c7d1-4bfe-a8d7-7421b1a590ec-public-tls-certs\") pod \"nova-api-0\" (UID: \"1f09945e-c7d1-4bfe-a8d7-7421b1a590ec\") " pod="openstack/nova-api-0" Oct 14 07:22:44 crc kubenswrapper[4870]: I1014 07:22:44.870537 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k52mj\" (UniqueName: \"kubernetes.io/projected/1f09945e-c7d1-4bfe-a8d7-7421b1a590ec-kube-api-access-k52mj\") pod \"nova-api-0\" (UID: \"1f09945e-c7d1-4bfe-a8d7-7421b1a590ec\") " pod="openstack/nova-api-0" Oct 14 07:22:44 crc kubenswrapper[4870]: I1014 07:22:44.870568 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1f09945e-c7d1-4bfe-a8d7-7421b1a590ec-internal-tls-certs\") pod \"nova-api-0\" (UID: \"1f09945e-c7d1-4bfe-a8d7-7421b1a590ec\") " pod="openstack/nova-api-0" Oct 14 07:22:44 crc kubenswrapper[4870]: I1014 07:22:44.870596 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1f09945e-c7d1-4bfe-a8d7-7421b1a590ec-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"1f09945e-c7d1-4bfe-a8d7-7421b1a590ec\") " pod="openstack/nova-api-0" Oct 14 07:22:44 crc kubenswrapper[4870]: I1014 07:22:44.871962 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1f09945e-c7d1-4bfe-a8d7-7421b1a590ec-logs\") pod \"nova-api-0\" (UID: \"1f09945e-c7d1-4bfe-a8d7-7421b1a590ec\") " pod="openstack/nova-api-0" Oct 14 07:22:44 crc kubenswrapper[4870]: I1014 07:22:44.876313 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1f09945e-c7d1-4bfe-a8d7-7421b1a590ec-internal-tls-certs\") pod \"nova-api-0\" (UID: \"1f09945e-c7d1-4bfe-a8d7-7421b1a590ec\") " pod="openstack/nova-api-0" Oct 14 07:22:44 crc kubenswrapper[4870]: I1014 07:22:44.876654 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1f09945e-c7d1-4bfe-a8d7-7421b1a590ec-public-tls-certs\") pod \"nova-api-0\" (UID: \"1f09945e-c7d1-4bfe-a8d7-7421b1a590ec\") " pod="openstack/nova-api-0" Oct 14 07:22:44 crc kubenswrapper[4870]: I1014 07:22:44.876690 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1f09945e-c7d1-4bfe-a8d7-7421b1a590ec-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"1f09945e-c7d1-4bfe-a8d7-7421b1a590ec\") " pod="openstack/nova-api-0" Oct 14 07:22:44 crc kubenswrapper[4870]: I1014 07:22:44.877111 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1f09945e-c7d1-4bfe-a8d7-7421b1a590ec-config-data\") pod \"nova-api-0\" (UID: \"1f09945e-c7d1-4bfe-a8d7-7421b1a590ec\") " pod="openstack/nova-api-0" Oct 14 07:22:44 crc kubenswrapper[4870]: I1014 07:22:44.894770 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k52mj\" (UniqueName: \"kubernetes.io/projected/1f09945e-c7d1-4bfe-a8d7-7421b1a590ec-kube-api-access-k52mj\") pod \"nova-api-0\" (UID: \"1f09945e-c7d1-4bfe-a8d7-7421b1a590ec\") " pod="openstack/nova-api-0" Oct 14 07:22:44 crc kubenswrapper[4870]: I1014 07:22:44.952308 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 14 07:22:45 crc kubenswrapper[4870]: I1014 07:22:45.072534 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2688cfd1-fbb5-49bb-b13e-a140ef4ed017" path="/var/lib/kubelet/pods/2688cfd1-fbb5-49bb-b13e-a140ef4ed017/volumes" Oct 14 07:22:45 crc kubenswrapper[4870]: I1014 07:22:45.465291 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 14 07:22:45 crc kubenswrapper[4870]: I1014 07:22:45.578847 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"1f09945e-c7d1-4bfe-a8d7-7421b1a590ec","Type":"ContainerStarted","Data":"330bf42bfc588c1c90bed31fab0a50f49168b23128a1e4ed5db6b8467a935dac"} Oct 14 07:22:45 crc kubenswrapper[4870]: I1014 07:22:45.585364 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"524b0b2a-77a2-4c05-be39-c8b97666ab11","Type":"ContainerStarted","Data":"92059a83ebd09e581c802c9d18a6cc53c68a57ad45e7c0b7e480b8f24d3f2bd8"} Oct 14 07:22:46 crc kubenswrapper[4870]: I1014 07:22:46.604820 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"1f09945e-c7d1-4bfe-a8d7-7421b1a590ec","Type":"ContainerStarted","Data":"e1959691705f2a2f3aa974444388035c568669302c7da6551de910de00176330"} Oct 14 07:22:46 crc kubenswrapper[4870]: I1014 07:22:46.605579 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"1f09945e-c7d1-4bfe-a8d7-7421b1a590ec","Type":"ContainerStarted","Data":"ef229458c64f8f5cfd522f061876879d09c15e53eebb20bb3651a9f2a7a32418"} Oct 14 07:22:46 crc kubenswrapper[4870]: I1014 07:22:46.610750 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"524b0b2a-77a2-4c05-be39-c8b97666ab11","Type":"ContainerStarted","Data":"a7f06f2b8b7523a5efb4f46ab915a339acb0f160d1f9753afa2035de3c545c6d"} Oct 14 07:22:46 crc kubenswrapper[4870]: I1014 07:22:46.678479 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.67845653 podStartE2EDuration="2.67845653s" podCreationTimestamp="2025-10-14 07:22:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:22:46.630909449 +0000 UTC m=+1302.328269850" watchObservedRunningTime="2025-10-14 07:22:46.67845653 +0000 UTC m=+1302.375816921" Oct 14 07:22:48 crc kubenswrapper[4870]: I1014 07:22:48.082610 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7cf4d694c5-wjxfs" Oct 14 07:22:48 crc kubenswrapper[4870]: I1014 07:22:48.161023 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6957666cb7-wgpsx"] Oct 14 07:22:48 crc kubenswrapper[4870]: I1014 07:22:48.161366 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6957666cb7-wgpsx" podUID="d5d3e78a-8ed8-40a5-8a64-734c7836155c" containerName="dnsmasq-dns" containerID="cri-o://49adfddfc295310c01014e57a90893bd2516bc1e4d16c17bd66c300fee355163" gracePeriod=10 Oct 14 07:22:48 crc kubenswrapper[4870]: I1014 07:22:48.634704 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"524b0b2a-77a2-4c05-be39-c8b97666ab11","Type":"ContainerStarted","Data":"018378adb14febcf4afcbaac2d9f70a84cc9e7f96675c9b85ee751bb465d9632"} Oct 14 07:22:48 crc kubenswrapper[4870]: I1014 07:22:48.636052 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 14 07:22:48 crc kubenswrapper[4870]: I1014 07:22:48.639345 4870 generic.go:334] "Generic (PLEG): container finished" podID="d5d3e78a-8ed8-40a5-8a64-734c7836155c" containerID="49adfddfc295310c01014e57a90893bd2516bc1e4d16c17bd66c300fee355163" exitCode=0 Oct 14 07:22:48 crc kubenswrapper[4870]: I1014 07:22:48.639381 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6957666cb7-wgpsx" event={"ID":"d5d3e78a-8ed8-40a5-8a64-734c7836155c","Type":"ContainerDied","Data":"49adfddfc295310c01014e57a90893bd2516bc1e4d16c17bd66c300fee355163"} Oct 14 07:22:48 crc kubenswrapper[4870]: I1014 07:22:48.639399 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6957666cb7-wgpsx" event={"ID":"d5d3e78a-8ed8-40a5-8a64-734c7836155c","Type":"ContainerDied","Data":"42174d886426663b8c0ce64845f0962beb3274ed0289eeb2e8d5350a53924767"} Oct 14 07:22:48 crc kubenswrapper[4870]: I1014 07:22:48.639415 4870 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="42174d886426663b8c0ce64845f0962beb3274ed0289eeb2e8d5350a53924767" Oct 14 07:22:48 crc kubenswrapper[4870]: I1014 07:22:48.667844 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.502805714 podStartE2EDuration="6.667820061s" podCreationTimestamp="2025-10-14 07:22:42 +0000 UTC" firstStartedPulling="2025-10-14 07:22:43.399182437 +0000 UTC m=+1299.096542848" lastFinishedPulling="2025-10-14 07:22:47.564196804 +0000 UTC m=+1303.261557195" observedRunningTime="2025-10-14 07:22:48.654595032 +0000 UTC m=+1304.351955413" watchObservedRunningTime="2025-10-14 07:22:48.667820061 +0000 UTC m=+1304.365180432" Oct 14 07:22:48 crc kubenswrapper[4870]: I1014 07:22:48.678545 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6957666cb7-wgpsx" Oct 14 07:22:48 crc kubenswrapper[4870]: I1014 07:22:48.856033 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d5d3e78a-8ed8-40a5-8a64-734c7836155c-config\") pod \"d5d3e78a-8ed8-40a5-8a64-734c7836155c\" (UID: \"d5d3e78a-8ed8-40a5-8a64-734c7836155c\") " Oct 14 07:22:48 crc kubenswrapper[4870]: I1014 07:22:48.856408 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d5d3e78a-8ed8-40a5-8a64-734c7836155c-ovsdbserver-nb\") pod \"d5d3e78a-8ed8-40a5-8a64-734c7836155c\" (UID: \"d5d3e78a-8ed8-40a5-8a64-734c7836155c\") " Oct 14 07:22:48 crc kubenswrapper[4870]: I1014 07:22:48.856525 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d5d3e78a-8ed8-40a5-8a64-734c7836155c-dns-svc\") pod \"d5d3e78a-8ed8-40a5-8a64-734c7836155c\" (UID: \"d5d3e78a-8ed8-40a5-8a64-734c7836155c\") " Oct 14 07:22:48 crc kubenswrapper[4870]: I1014 07:22:48.856600 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bqmbb\" (UniqueName: \"kubernetes.io/projected/d5d3e78a-8ed8-40a5-8a64-734c7836155c-kube-api-access-bqmbb\") pod \"d5d3e78a-8ed8-40a5-8a64-734c7836155c\" (UID: \"d5d3e78a-8ed8-40a5-8a64-734c7836155c\") " Oct 14 07:22:48 crc kubenswrapper[4870]: I1014 07:22:48.856691 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d5d3e78a-8ed8-40a5-8a64-734c7836155c-dns-swift-storage-0\") pod \"d5d3e78a-8ed8-40a5-8a64-734c7836155c\" (UID: \"d5d3e78a-8ed8-40a5-8a64-734c7836155c\") " Oct 14 07:22:48 crc kubenswrapper[4870]: I1014 07:22:48.856713 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d5d3e78a-8ed8-40a5-8a64-734c7836155c-ovsdbserver-sb\") pod \"d5d3e78a-8ed8-40a5-8a64-734c7836155c\" (UID: \"d5d3e78a-8ed8-40a5-8a64-734c7836155c\") " Oct 14 07:22:48 crc kubenswrapper[4870]: I1014 07:22:48.868261 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d5d3e78a-8ed8-40a5-8a64-734c7836155c-kube-api-access-bqmbb" (OuterVolumeSpecName: "kube-api-access-bqmbb") pod "d5d3e78a-8ed8-40a5-8a64-734c7836155c" (UID: "d5d3e78a-8ed8-40a5-8a64-734c7836155c"). InnerVolumeSpecName "kube-api-access-bqmbb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:22:48 crc kubenswrapper[4870]: I1014 07:22:48.872508 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Oct 14 07:22:48 crc kubenswrapper[4870]: I1014 07:22:48.907980 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Oct 14 07:22:48 crc kubenswrapper[4870]: I1014 07:22:48.930205 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d5d3e78a-8ed8-40a5-8a64-734c7836155c-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "d5d3e78a-8ed8-40a5-8a64-734c7836155c" (UID: "d5d3e78a-8ed8-40a5-8a64-734c7836155c"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:22:48 crc kubenswrapper[4870]: I1014 07:22:48.954556 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d5d3e78a-8ed8-40a5-8a64-734c7836155c-config" (OuterVolumeSpecName: "config") pod "d5d3e78a-8ed8-40a5-8a64-734c7836155c" (UID: "d5d3e78a-8ed8-40a5-8a64-734c7836155c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:22:48 crc kubenswrapper[4870]: I1014 07:22:48.955799 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d5d3e78a-8ed8-40a5-8a64-734c7836155c-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "d5d3e78a-8ed8-40a5-8a64-734c7836155c" (UID: "d5d3e78a-8ed8-40a5-8a64-734c7836155c"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:22:48 crc kubenswrapper[4870]: I1014 07:22:48.955833 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d5d3e78a-8ed8-40a5-8a64-734c7836155c-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "d5d3e78a-8ed8-40a5-8a64-734c7836155c" (UID: "d5d3e78a-8ed8-40a5-8a64-734c7836155c"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:22:48 crc kubenswrapper[4870]: I1014 07:22:48.959771 4870 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d5d3e78a-8ed8-40a5-8a64-734c7836155c-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 14 07:22:48 crc kubenswrapper[4870]: I1014 07:22:48.959812 4870 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d5d3e78a-8ed8-40a5-8a64-734c7836155c-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 14 07:22:48 crc kubenswrapper[4870]: I1014 07:22:48.959826 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bqmbb\" (UniqueName: \"kubernetes.io/projected/d5d3e78a-8ed8-40a5-8a64-734c7836155c-kube-api-access-bqmbb\") on node \"crc\" DevicePath \"\"" Oct 14 07:22:48 crc kubenswrapper[4870]: I1014 07:22:48.959842 4870 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d5d3e78a-8ed8-40a5-8a64-734c7836155c-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 14 07:22:48 crc kubenswrapper[4870]: I1014 07:22:48.959857 4870 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d5d3e78a-8ed8-40a5-8a64-734c7836155c-config\") on node \"crc\" DevicePath \"\"" Oct 14 07:22:48 crc kubenswrapper[4870]: I1014 07:22:48.965706 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d5d3e78a-8ed8-40a5-8a64-734c7836155c-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "d5d3e78a-8ed8-40a5-8a64-734c7836155c" (UID: "d5d3e78a-8ed8-40a5-8a64-734c7836155c"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:22:49 crc kubenswrapper[4870]: I1014 07:22:49.066971 4870 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d5d3e78a-8ed8-40a5-8a64-734c7836155c-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 14 07:22:49 crc kubenswrapper[4870]: I1014 07:22:49.672083 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6957666cb7-wgpsx" Oct 14 07:22:49 crc kubenswrapper[4870]: I1014 07:22:49.697179 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Oct 14 07:22:49 crc kubenswrapper[4870]: I1014 07:22:49.737663 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6957666cb7-wgpsx"] Oct 14 07:22:49 crc kubenswrapper[4870]: I1014 07:22:49.765904 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6957666cb7-wgpsx"] Oct 14 07:22:49 crc kubenswrapper[4870]: I1014 07:22:49.901790 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-xtkkt"] Oct 14 07:22:49 crc kubenswrapper[4870]: E1014 07:22:49.902504 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d5d3e78a-8ed8-40a5-8a64-734c7836155c" containerName="dnsmasq-dns" Oct 14 07:22:49 crc kubenswrapper[4870]: I1014 07:22:49.902516 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="d5d3e78a-8ed8-40a5-8a64-734c7836155c" containerName="dnsmasq-dns" Oct 14 07:22:49 crc kubenswrapper[4870]: E1014 07:22:49.902533 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d5d3e78a-8ed8-40a5-8a64-734c7836155c" containerName="init" Oct 14 07:22:49 crc kubenswrapper[4870]: I1014 07:22:49.902538 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="d5d3e78a-8ed8-40a5-8a64-734c7836155c" containerName="init" Oct 14 07:22:49 crc kubenswrapper[4870]: I1014 07:22:49.902720 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="d5d3e78a-8ed8-40a5-8a64-734c7836155c" containerName="dnsmasq-dns" Oct 14 07:22:49 crc kubenswrapper[4870]: I1014 07:22:49.903347 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-xtkkt" Oct 14 07:22:49 crc kubenswrapper[4870]: I1014 07:22:49.907803 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Oct 14 07:22:49 crc kubenswrapper[4870]: I1014 07:22:49.908870 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Oct 14 07:22:49 crc kubenswrapper[4870]: I1014 07:22:49.931356 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-xtkkt"] Oct 14 07:22:49 crc kubenswrapper[4870]: I1014 07:22:49.985161 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ptkg8\" (UniqueName: \"kubernetes.io/projected/1110cc74-c641-4a3a-b46d-ea85668515b0-kube-api-access-ptkg8\") pod \"nova-cell1-cell-mapping-xtkkt\" (UID: \"1110cc74-c641-4a3a-b46d-ea85668515b0\") " pod="openstack/nova-cell1-cell-mapping-xtkkt" Oct 14 07:22:49 crc kubenswrapper[4870]: I1014 07:22:49.985243 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1110cc74-c641-4a3a-b46d-ea85668515b0-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-xtkkt\" (UID: \"1110cc74-c641-4a3a-b46d-ea85668515b0\") " pod="openstack/nova-cell1-cell-mapping-xtkkt" Oct 14 07:22:49 crc kubenswrapper[4870]: I1014 07:22:49.985288 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1110cc74-c641-4a3a-b46d-ea85668515b0-scripts\") pod \"nova-cell1-cell-mapping-xtkkt\" (UID: \"1110cc74-c641-4a3a-b46d-ea85668515b0\") " pod="openstack/nova-cell1-cell-mapping-xtkkt" Oct 14 07:22:49 crc kubenswrapper[4870]: I1014 07:22:49.985372 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1110cc74-c641-4a3a-b46d-ea85668515b0-config-data\") pod \"nova-cell1-cell-mapping-xtkkt\" (UID: \"1110cc74-c641-4a3a-b46d-ea85668515b0\") " pod="openstack/nova-cell1-cell-mapping-xtkkt" Oct 14 07:22:50 crc kubenswrapper[4870]: I1014 07:22:50.087702 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1110cc74-c641-4a3a-b46d-ea85668515b0-scripts\") pod \"nova-cell1-cell-mapping-xtkkt\" (UID: \"1110cc74-c641-4a3a-b46d-ea85668515b0\") " pod="openstack/nova-cell1-cell-mapping-xtkkt" Oct 14 07:22:50 crc kubenswrapper[4870]: I1014 07:22:50.087885 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1110cc74-c641-4a3a-b46d-ea85668515b0-config-data\") pod \"nova-cell1-cell-mapping-xtkkt\" (UID: \"1110cc74-c641-4a3a-b46d-ea85668515b0\") " pod="openstack/nova-cell1-cell-mapping-xtkkt" Oct 14 07:22:50 crc kubenswrapper[4870]: I1014 07:22:50.087943 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ptkg8\" (UniqueName: \"kubernetes.io/projected/1110cc74-c641-4a3a-b46d-ea85668515b0-kube-api-access-ptkg8\") pod \"nova-cell1-cell-mapping-xtkkt\" (UID: \"1110cc74-c641-4a3a-b46d-ea85668515b0\") " pod="openstack/nova-cell1-cell-mapping-xtkkt" Oct 14 07:22:50 crc kubenswrapper[4870]: I1014 07:22:50.087996 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1110cc74-c641-4a3a-b46d-ea85668515b0-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-xtkkt\" (UID: \"1110cc74-c641-4a3a-b46d-ea85668515b0\") " pod="openstack/nova-cell1-cell-mapping-xtkkt" Oct 14 07:22:50 crc kubenswrapper[4870]: I1014 07:22:50.094607 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1110cc74-c641-4a3a-b46d-ea85668515b0-scripts\") pod \"nova-cell1-cell-mapping-xtkkt\" (UID: \"1110cc74-c641-4a3a-b46d-ea85668515b0\") " pod="openstack/nova-cell1-cell-mapping-xtkkt" Oct 14 07:22:50 crc kubenswrapper[4870]: I1014 07:22:50.095008 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1110cc74-c641-4a3a-b46d-ea85668515b0-config-data\") pod \"nova-cell1-cell-mapping-xtkkt\" (UID: \"1110cc74-c641-4a3a-b46d-ea85668515b0\") " pod="openstack/nova-cell1-cell-mapping-xtkkt" Oct 14 07:22:50 crc kubenswrapper[4870]: I1014 07:22:50.103042 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1110cc74-c641-4a3a-b46d-ea85668515b0-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-xtkkt\" (UID: \"1110cc74-c641-4a3a-b46d-ea85668515b0\") " pod="openstack/nova-cell1-cell-mapping-xtkkt" Oct 14 07:22:50 crc kubenswrapper[4870]: I1014 07:22:50.110752 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ptkg8\" (UniqueName: \"kubernetes.io/projected/1110cc74-c641-4a3a-b46d-ea85668515b0-kube-api-access-ptkg8\") pod \"nova-cell1-cell-mapping-xtkkt\" (UID: \"1110cc74-c641-4a3a-b46d-ea85668515b0\") " pod="openstack/nova-cell1-cell-mapping-xtkkt" Oct 14 07:22:50 crc kubenswrapper[4870]: I1014 07:22:50.231791 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-xtkkt" Oct 14 07:22:50 crc kubenswrapper[4870]: I1014 07:22:50.692738 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-xtkkt"] Oct 14 07:22:50 crc kubenswrapper[4870]: W1014 07:22:50.700092 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1110cc74_c641_4a3a_b46d_ea85668515b0.slice/crio-2f75009e64c638fe3a6e314e4996b1df604c42302e40f69fae9215403d7fdbc7 WatchSource:0}: Error finding container 2f75009e64c638fe3a6e314e4996b1df604c42302e40f69fae9215403d7fdbc7: Status 404 returned error can't find the container with id 2f75009e64c638fe3a6e314e4996b1df604c42302e40f69fae9215403d7fdbc7 Oct 14 07:22:51 crc kubenswrapper[4870]: I1014 07:22:51.055927 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d5d3e78a-8ed8-40a5-8a64-734c7836155c" path="/var/lib/kubelet/pods/d5d3e78a-8ed8-40a5-8a64-734c7836155c/volumes" Oct 14 07:22:51 crc kubenswrapper[4870]: I1014 07:22:51.706477 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-xtkkt" event={"ID":"1110cc74-c641-4a3a-b46d-ea85668515b0","Type":"ContainerStarted","Data":"c3d6d43a38bf5b35a30900bbcca66f0dcbacc37ec8b96a0733086e4ae6871167"} Oct 14 07:22:51 crc kubenswrapper[4870]: I1014 07:22:51.706532 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-xtkkt" event={"ID":"1110cc74-c641-4a3a-b46d-ea85668515b0","Type":"ContainerStarted","Data":"2f75009e64c638fe3a6e314e4996b1df604c42302e40f69fae9215403d7fdbc7"} Oct 14 07:22:51 crc kubenswrapper[4870]: I1014 07:22:51.736108 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-xtkkt" podStartSLOduration=2.736089502 podStartE2EDuration="2.736089502s" podCreationTimestamp="2025-10-14 07:22:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:22:51.721198714 +0000 UTC m=+1307.418559135" watchObservedRunningTime="2025-10-14 07:22:51.736089502 +0000 UTC m=+1307.433449883" Oct 14 07:22:54 crc kubenswrapper[4870]: I1014 07:22:54.954257 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 14 07:22:54 crc kubenswrapper[4870]: I1014 07:22:54.954815 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 14 07:22:55 crc kubenswrapper[4870]: I1014 07:22:55.755686 4870 generic.go:334] "Generic (PLEG): container finished" podID="1110cc74-c641-4a3a-b46d-ea85668515b0" containerID="c3d6d43a38bf5b35a30900bbcca66f0dcbacc37ec8b96a0733086e4ae6871167" exitCode=0 Oct 14 07:22:55 crc kubenswrapper[4870]: I1014 07:22:55.756131 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-xtkkt" event={"ID":"1110cc74-c641-4a3a-b46d-ea85668515b0","Type":"ContainerDied","Data":"c3d6d43a38bf5b35a30900bbcca66f0dcbacc37ec8b96a0733086e4ae6871167"} Oct 14 07:22:55 crc kubenswrapper[4870]: I1014 07:22:55.966677 4870 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="1f09945e-c7d1-4bfe-a8d7-7421b1a590ec" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.199:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 14 07:22:55 crc kubenswrapper[4870]: I1014 07:22:55.966728 4870 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="1f09945e-c7d1-4bfe-a8d7-7421b1a590ec" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.199:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 14 07:22:57 crc kubenswrapper[4870]: I1014 07:22:57.207206 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-xtkkt" Oct 14 07:22:57 crc kubenswrapper[4870]: I1014 07:22:57.311008 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ptkg8\" (UniqueName: \"kubernetes.io/projected/1110cc74-c641-4a3a-b46d-ea85668515b0-kube-api-access-ptkg8\") pod \"1110cc74-c641-4a3a-b46d-ea85668515b0\" (UID: \"1110cc74-c641-4a3a-b46d-ea85668515b0\") " Oct 14 07:22:57 crc kubenswrapper[4870]: I1014 07:22:57.311136 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1110cc74-c641-4a3a-b46d-ea85668515b0-combined-ca-bundle\") pod \"1110cc74-c641-4a3a-b46d-ea85668515b0\" (UID: \"1110cc74-c641-4a3a-b46d-ea85668515b0\") " Oct 14 07:22:57 crc kubenswrapper[4870]: I1014 07:22:57.311267 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1110cc74-c641-4a3a-b46d-ea85668515b0-config-data\") pod \"1110cc74-c641-4a3a-b46d-ea85668515b0\" (UID: \"1110cc74-c641-4a3a-b46d-ea85668515b0\") " Oct 14 07:22:57 crc kubenswrapper[4870]: I1014 07:22:57.311352 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1110cc74-c641-4a3a-b46d-ea85668515b0-scripts\") pod \"1110cc74-c641-4a3a-b46d-ea85668515b0\" (UID: \"1110cc74-c641-4a3a-b46d-ea85668515b0\") " Oct 14 07:22:57 crc kubenswrapper[4870]: I1014 07:22:57.317580 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1110cc74-c641-4a3a-b46d-ea85668515b0-kube-api-access-ptkg8" (OuterVolumeSpecName: "kube-api-access-ptkg8") pod "1110cc74-c641-4a3a-b46d-ea85668515b0" (UID: "1110cc74-c641-4a3a-b46d-ea85668515b0"). InnerVolumeSpecName "kube-api-access-ptkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:22:57 crc kubenswrapper[4870]: I1014 07:22:57.319701 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1110cc74-c641-4a3a-b46d-ea85668515b0-scripts" (OuterVolumeSpecName: "scripts") pod "1110cc74-c641-4a3a-b46d-ea85668515b0" (UID: "1110cc74-c641-4a3a-b46d-ea85668515b0"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:22:57 crc kubenswrapper[4870]: I1014 07:22:57.339811 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1110cc74-c641-4a3a-b46d-ea85668515b0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1110cc74-c641-4a3a-b46d-ea85668515b0" (UID: "1110cc74-c641-4a3a-b46d-ea85668515b0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:22:57 crc kubenswrapper[4870]: I1014 07:22:57.344179 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1110cc74-c641-4a3a-b46d-ea85668515b0-config-data" (OuterVolumeSpecName: "config-data") pod "1110cc74-c641-4a3a-b46d-ea85668515b0" (UID: "1110cc74-c641-4a3a-b46d-ea85668515b0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:22:57 crc kubenswrapper[4870]: I1014 07:22:57.413846 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ptkg8\" (UniqueName: \"kubernetes.io/projected/1110cc74-c641-4a3a-b46d-ea85668515b0-kube-api-access-ptkg8\") on node \"crc\" DevicePath \"\"" Oct 14 07:22:57 crc kubenswrapper[4870]: I1014 07:22:57.413888 4870 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1110cc74-c641-4a3a-b46d-ea85668515b0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 07:22:57 crc kubenswrapper[4870]: I1014 07:22:57.413905 4870 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1110cc74-c641-4a3a-b46d-ea85668515b0-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 07:22:57 crc kubenswrapper[4870]: I1014 07:22:57.413918 4870 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1110cc74-c641-4a3a-b46d-ea85668515b0-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 07:22:57 crc kubenswrapper[4870]: I1014 07:22:57.810885 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-xtkkt" event={"ID":"1110cc74-c641-4a3a-b46d-ea85668515b0","Type":"ContainerDied","Data":"2f75009e64c638fe3a6e314e4996b1df604c42302e40f69fae9215403d7fdbc7"} Oct 14 07:22:57 crc kubenswrapper[4870]: I1014 07:22:57.811386 4870 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2f75009e64c638fe3a6e314e4996b1df604c42302e40f69fae9215403d7fdbc7" Oct 14 07:22:57 crc kubenswrapper[4870]: I1014 07:22:57.811618 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-xtkkt" Oct 14 07:22:58 crc kubenswrapper[4870]: I1014 07:22:58.081324 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 14 07:22:58 crc kubenswrapper[4870]: I1014 07:22:58.081789 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="1f09945e-c7d1-4bfe-a8d7-7421b1a590ec" containerName="nova-api-log" containerID="cri-o://ef229458c64f8f5cfd522f061876879d09c15e53eebb20bb3651a9f2a7a32418" gracePeriod=30 Oct 14 07:22:58 crc kubenswrapper[4870]: I1014 07:22:58.081939 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="1f09945e-c7d1-4bfe-a8d7-7421b1a590ec" containerName="nova-api-api" containerID="cri-o://e1959691705f2a2f3aa974444388035c568669302c7da6551de910de00176330" gracePeriod=30 Oct 14 07:22:58 crc kubenswrapper[4870]: I1014 07:22:58.106416 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 14 07:22:58 crc kubenswrapper[4870]: I1014 07:22:58.106739 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="ffe53321-c66f-4b32-b18d-18bc483f923b" containerName="nova-scheduler-scheduler" containerID="cri-o://56e70d8d8996b806615045a193a6672b63fec50f57799575de3be2115113662c" gracePeriod=30 Oct 14 07:22:58 crc kubenswrapper[4870]: I1014 07:22:58.121613 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 14 07:22:58 crc kubenswrapper[4870]: I1014 07:22:58.121898 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="4a361161-c100-49cb-a4ee-19563b724033" containerName="nova-metadata-log" containerID="cri-o://1a08f0b4fd874fefc01d80d5ab91d598f23dd77047d80c172612080dc2670dbc" gracePeriod=30 Oct 14 07:22:58 crc kubenswrapper[4870]: I1014 07:22:58.121999 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="4a361161-c100-49cb-a4ee-19563b724033" containerName="nova-metadata-metadata" containerID="cri-o://98556816573a1ddbc56f6c6dcdd861e5bfd4e061a6c46d5d705c33318916ee5c" gracePeriod=30 Oct 14 07:22:58 crc kubenswrapper[4870]: I1014 07:22:58.824175 4870 generic.go:334] "Generic (PLEG): container finished" podID="1f09945e-c7d1-4bfe-a8d7-7421b1a590ec" containerID="ef229458c64f8f5cfd522f061876879d09c15e53eebb20bb3651a9f2a7a32418" exitCode=143 Oct 14 07:22:58 crc kubenswrapper[4870]: I1014 07:22:58.824472 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"1f09945e-c7d1-4bfe-a8d7-7421b1a590ec","Type":"ContainerDied","Data":"ef229458c64f8f5cfd522f061876879d09c15e53eebb20bb3651a9f2a7a32418"} Oct 14 07:22:58 crc kubenswrapper[4870]: I1014 07:22:58.827616 4870 generic.go:334] "Generic (PLEG): container finished" podID="4a361161-c100-49cb-a4ee-19563b724033" containerID="1a08f0b4fd874fefc01d80d5ab91d598f23dd77047d80c172612080dc2670dbc" exitCode=143 Oct 14 07:22:58 crc kubenswrapper[4870]: I1014 07:22:58.827725 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"4a361161-c100-49cb-a4ee-19563b724033","Type":"ContainerDied","Data":"1a08f0b4fd874fefc01d80d5ab91d598f23dd77047d80c172612080dc2670dbc"} Oct 14 07:22:59 crc kubenswrapper[4870]: E1014 07:22:59.640058 4870 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="56e70d8d8996b806615045a193a6672b63fec50f57799575de3be2115113662c" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 14 07:22:59 crc kubenswrapper[4870]: E1014 07:22:59.641706 4870 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="56e70d8d8996b806615045a193a6672b63fec50f57799575de3be2115113662c" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 14 07:22:59 crc kubenswrapper[4870]: E1014 07:22:59.643143 4870 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="56e70d8d8996b806615045a193a6672b63fec50f57799575de3be2115113662c" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 14 07:22:59 crc kubenswrapper[4870]: E1014 07:22:59.643288 4870 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="ffe53321-c66f-4b32-b18d-18bc483f923b" containerName="nova-scheduler-scheduler" Oct 14 07:23:01 crc kubenswrapper[4870]: I1014 07:23:01.592056 4870 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="4a361161-c100-49cb-a4ee-19563b724033" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.190:8775/\": read tcp 10.217.0.2:55838->10.217.0.190:8775: read: connection reset by peer" Oct 14 07:23:01 crc kubenswrapper[4870]: I1014 07:23:01.592065 4870 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="4a361161-c100-49cb-a4ee-19563b724033" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.190:8775/\": read tcp 10.217.0.2:55822->10.217.0.190:8775: read: connection reset by peer" Oct 14 07:23:01 crc kubenswrapper[4870]: I1014 07:23:01.868987 4870 generic.go:334] "Generic (PLEG): container finished" podID="4a361161-c100-49cb-a4ee-19563b724033" containerID="98556816573a1ddbc56f6c6dcdd861e5bfd4e061a6c46d5d705c33318916ee5c" exitCode=0 Oct 14 07:23:01 crc kubenswrapper[4870]: I1014 07:23:01.869210 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"4a361161-c100-49cb-a4ee-19563b724033","Type":"ContainerDied","Data":"98556816573a1ddbc56f6c6dcdd861e5bfd4e061a6c46d5d705c33318916ee5c"} Oct 14 07:23:01 crc kubenswrapper[4870]: I1014 07:23:01.871257 4870 generic.go:334] "Generic (PLEG): container finished" podID="1f09945e-c7d1-4bfe-a8d7-7421b1a590ec" containerID="e1959691705f2a2f3aa974444388035c568669302c7da6551de910de00176330" exitCode=0 Oct 14 07:23:01 crc kubenswrapper[4870]: I1014 07:23:01.871289 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"1f09945e-c7d1-4bfe-a8d7-7421b1a590ec","Type":"ContainerDied","Data":"e1959691705f2a2f3aa974444388035c568669302c7da6551de910de00176330"} Oct 14 07:23:01 crc kubenswrapper[4870]: I1014 07:23:01.871318 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"1f09945e-c7d1-4bfe-a8d7-7421b1a590ec","Type":"ContainerDied","Data":"330bf42bfc588c1c90bed31fab0a50f49168b23128a1e4ed5db6b8467a935dac"} Oct 14 07:23:01 crc kubenswrapper[4870]: I1014 07:23:01.871330 4870 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="330bf42bfc588c1c90bed31fab0a50f49168b23128a1e4ed5db6b8467a935dac" Oct 14 07:23:01 crc kubenswrapper[4870]: I1014 07:23:01.878601 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 14 07:23:01 crc kubenswrapper[4870]: I1014 07:23:01.999198 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 14 07:23:02 crc kubenswrapper[4870]: I1014 07:23:02.006819 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1f09945e-c7d1-4bfe-a8d7-7421b1a590ec-logs\") pod \"1f09945e-c7d1-4bfe-a8d7-7421b1a590ec\" (UID: \"1f09945e-c7d1-4bfe-a8d7-7421b1a590ec\") " Oct 14 07:23:02 crc kubenswrapper[4870]: I1014 07:23:02.006880 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1f09945e-c7d1-4bfe-a8d7-7421b1a590ec-combined-ca-bundle\") pod \"1f09945e-c7d1-4bfe-a8d7-7421b1a590ec\" (UID: \"1f09945e-c7d1-4bfe-a8d7-7421b1a590ec\") " Oct 14 07:23:02 crc kubenswrapper[4870]: I1014 07:23:02.007013 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1f09945e-c7d1-4bfe-a8d7-7421b1a590ec-config-data\") pod \"1f09945e-c7d1-4bfe-a8d7-7421b1a590ec\" (UID: \"1f09945e-c7d1-4bfe-a8d7-7421b1a590ec\") " Oct 14 07:23:02 crc kubenswrapper[4870]: I1014 07:23:02.007092 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k52mj\" (UniqueName: \"kubernetes.io/projected/1f09945e-c7d1-4bfe-a8d7-7421b1a590ec-kube-api-access-k52mj\") pod \"1f09945e-c7d1-4bfe-a8d7-7421b1a590ec\" (UID: \"1f09945e-c7d1-4bfe-a8d7-7421b1a590ec\") " Oct 14 07:23:02 crc kubenswrapper[4870]: I1014 07:23:02.007130 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1f09945e-c7d1-4bfe-a8d7-7421b1a590ec-public-tls-certs\") pod \"1f09945e-c7d1-4bfe-a8d7-7421b1a590ec\" (UID: \"1f09945e-c7d1-4bfe-a8d7-7421b1a590ec\") " Oct 14 07:23:02 crc kubenswrapper[4870]: I1014 07:23:02.007237 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1f09945e-c7d1-4bfe-a8d7-7421b1a590ec-internal-tls-certs\") pod \"1f09945e-c7d1-4bfe-a8d7-7421b1a590ec\" (UID: \"1f09945e-c7d1-4bfe-a8d7-7421b1a590ec\") " Oct 14 07:23:02 crc kubenswrapper[4870]: I1014 07:23:02.010363 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1f09945e-c7d1-4bfe-a8d7-7421b1a590ec-logs" (OuterVolumeSpecName: "logs") pod "1f09945e-c7d1-4bfe-a8d7-7421b1a590ec" (UID: "1f09945e-c7d1-4bfe-a8d7-7421b1a590ec"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:23:02 crc kubenswrapper[4870]: I1014 07:23:02.013883 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1f09945e-c7d1-4bfe-a8d7-7421b1a590ec-kube-api-access-k52mj" (OuterVolumeSpecName: "kube-api-access-k52mj") pod "1f09945e-c7d1-4bfe-a8d7-7421b1a590ec" (UID: "1f09945e-c7d1-4bfe-a8d7-7421b1a590ec"). InnerVolumeSpecName "kube-api-access-k52mj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:23:02 crc kubenswrapper[4870]: I1014 07:23:02.072262 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1f09945e-c7d1-4bfe-a8d7-7421b1a590ec-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1f09945e-c7d1-4bfe-a8d7-7421b1a590ec" (UID: "1f09945e-c7d1-4bfe-a8d7-7421b1a590ec"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:23:02 crc kubenswrapper[4870]: I1014 07:23:02.076176 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1f09945e-c7d1-4bfe-a8d7-7421b1a590ec-config-data" (OuterVolumeSpecName: "config-data") pod "1f09945e-c7d1-4bfe-a8d7-7421b1a590ec" (UID: "1f09945e-c7d1-4bfe-a8d7-7421b1a590ec"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:23:02 crc kubenswrapper[4870]: I1014 07:23:02.100805 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1f09945e-c7d1-4bfe-a8d7-7421b1a590ec-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "1f09945e-c7d1-4bfe-a8d7-7421b1a590ec" (UID: "1f09945e-c7d1-4bfe-a8d7-7421b1a590ec"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:23:02 crc kubenswrapper[4870]: I1014 07:23:02.108358 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4a361161-c100-49cb-a4ee-19563b724033-config-data\") pod \"4a361161-c100-49cb-a4ee-19563b724033\" (UID: \"4a361161-c100-49cb-a4ee-19563b724033\") " Oct 14 07:23:02 crc kubenswrapper[4870]: I1014 07:23:02.108590 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-22wcl\" (UniqueName: \"kubernetes.io/projected/4a361161-c100-49cb-a4ee-19563b724033-kube-api-access-22wcl\") pod \"4a361161-c100-49cb-a4ee-19563b724033\" (UID: \"4a361161-c100-49cb-a4ee-19563b724033\") " Oct 14 07:23:02 crc kubenswrapper[4870]: I1014 07:23:02.108704 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a361161-c100-49cb-a4ee-19563b724033-combined-ca-bundle\") pod \"4a361161-c100-49cb-a4ee-19563b724033\" (UID: \"4a361161-c100-49cb-a4ee-19563b724033\") " Oct 14 07:23:02 crc kubenswrapper[4870]: I1014 07:23:02.108740 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4a361161-c100-49cb-a4ee-19563b724033-logs\") pod \"4a361161-c100-49cb-a4ee-19563b724033\" (UID: \"4a361161-c100-49cb-a4ee-19563b724033\") " Oct 14 07:23:02 crc kubenswrapper[4870]: I1014 07:23:02.108824 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/4a361161-c100-49cb-a4ee-19563b724033-nova-metadata-tls-certs\") pod \"4a361161-c100-49cb-a4ee-19563b724033\" (UID: \"4a361161-c100-49cb-a4ee-19563b724033\") " Oct 14 07:23:02 crc kubenswrapper[4870]: I1014 07:23:02.109384 4870 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1f09945e-c7d1-4bfe-a8d7-7421b1a590ec-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:02 crc kubenswrapper[4870]: I1014 07:23:02.109405 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k52mj\" (UniqueName: \"kubernetes.io/projected/1f09945e-c7d1-4bfe-a8d7-7421b1a590ec-kube-api-access-k52mj\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:02 crc kubenswrapper[4870]: I1014 07:23:02.109390 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4a361161-c100-49cb-a4ee-19563b724033-logs" (OuterVolumeSpecName: "logs") pod "4a361161-c100-49cb-a4ee-19563b724033" (UID: "4a361161-c100-49cb-a4ee-19563b724033"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:23:02 crc kubenswrapper[4870]: I1014 07:23:02.109416 4870 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1f09945e-c7d1-4bfe-a8d7-7421b1a590ec-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:02 crc kubenswrapper[4870]: I1014 07:23:02.109500 4870 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1f09945e-c7d1-4bfe-a8d7-7421b1a590ec-logs\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:02 crc kubenswrapper[4870]: I1014 07:23:02.109517 4870 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1f09945e-c7d1-4bfe-a8d7-7421b1a590ec-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:02 crc kubenswrapper[4870]: I1014 07:23:02.111030 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1f09945e-c7d1-4bfe-a8d7-7421b1a590ec-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "1f09945e-c7d1-4bfe-a8d7-7421b1a590ec" (UID: "1f09945e-c7d1-4bfe-a8d7-7421b1a590ec"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:23:02 crc kubenswrapper[4870]: I1014 07:23:02.111542 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4a361161-c100-49cb-a4ee-19563b724033-kube-api-access-22wcl" (OuterVolumeSpecName: "kube-api-access-22wcl") pod "4a361161-c100-49cb-a4ee-19563b724033" (UID: "4a361161-c100-49cb-a4ee-19563b724033"). InnerVolumeSpecName "kube-api-access-22wcl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:23:02 crc kubenswrapper[4870]: I1014 07:23:02.133981 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4a361161-c100-49cb-a4ee-19563b724033-config-data" (OuterVolumeSpecName: "config-data") pod "4a361161-c100-49cb-a4ee-19563b724033" (UID: "4a361161-c100-49cb-a4ee-19563b724033"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:23:02 crc kubenswrapper[4870]: I1014 07:23:02.135739 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4a361161-c100-49cb-a4ee-19563b724033-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4a361161-c100-49cb-a4ee-19563b724033" (UID: "4a361161-c100-49cb-a4ee-19563b724033"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:23:02 crc kubenswrapper[4870]: I1014 07:23:02.175714 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4a361161-c100-49cb-a4ee-19563b724033-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "4a361161-c100-49cb-a4ee-19563b724033" (UID: "4a361161-c100-49cb-a4ee-19563b724033"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:23:02 crc kubenswrapper[4870]: I1014 07:23:02.211932 4870 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1f09945e-c7d1-4bfe-a8d7-7421b1a590ec-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:02 crc kubenswrapper[4870]: I1014 07:23:02.211980 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-22wcl\" (UniqueName: \"kubernetes.io/projected/4a361161-c100-49cb-a4ee-19563b724033-kube-api-access-22wcl\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:02 crc kubenswrapper[4870]: I1014 07:23:02.212002 4870 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a361161-c100-49cb-a4ee-19563b724033-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:02 crc kubenswrapper[4870]: I1014 07:23:02.212023 4870 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4a361161-c100-49cb-a4ee-19563b724033-logs\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:02 crc kubenswrapper[4870]: I1014 07:23:02.212076 4870 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/4a361161-c100-49cb-a4ee-19563b724033-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:02 crc kubenswrapper[4870]: I1014 07:23:02.212097 4870 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4a361161-c100-49cb-a4ee-19563b724033-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:03 crc kubenswrapper[4870]: I1014 07:23:02.889126 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 14 07:23:03 crc kubenswrapper[4870]: I1014 07:23:02.889176 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 14 07:23:03 crc kubenswrapper[4870]: I1014 07:23:02.889207 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"4a361161-c100-49cb-a4ee-19563b724033","Type":"ContainerDied","Data":"7ba3a53d0ecb65545d234f202c569d57bc35424c5757f7dcb97fdefa9c3c5700"} Oct 14 07:23:03 crc kubenswrapper[4870]: I1014 07:23:02.889951 4870 scope.go:117] "RemoveContainer" containerID="98556816573a1ddbc56f6c6dcdd861e5bfd4e061a6c46d5d705c33318916ee5c" Oct 14 07:23:03 crc kubenswrapper[4870]: I1014 07:23:02.963455 4870 scope.go:117] "RemoveContainer" containerID="1a08f0b4fd874fefc01d80d5ab91d598f23dd77047d80c172612080dc2670dbc" Oct 14 07:23:03 crc kubenswrapper[4870]: I1014 07:23:02.963485 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 14 07:23:03 crc kubenswrapper[4870]: I1014 07:23:02.991557 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 14 07:23:03 crc kubenswrapper[4870]: I1014 07:23:03.000804 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 14 07:23:03 crc kubenswrapper[4870]: E1014 07:23:03.001330 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f09945e-c7d1-4bfe-a8d7-7421b1a590ec" containerName="nova-api-api" Oct 14 07:23:03 crc kubenswrapper[4870]: I1014 07:23:03.001347 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f09945e-c7d1-4bfe-a8d7-7421b1a590ec" containerName="nova-api-api" Oct 14 07:23:03 crc kubenswrapper[4870]: E1014 07:23:03.001375 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4a361161-c100-49cb-a4ee-19563b724033" containerName="nova-metadata-log" Oct 14 07:23:03 crc kubenswrapper[4870]: I1014 07:23:03.001384 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="4a361161-c100-49cb-a4ee-19563b724033" containerName="nova-metadata-log" Oct 14 07:23:03 crc kubenswrapper[4870]: E1014 07:23:03.001418 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1110cc74-c641-4a3a-b46d-ea85668515b0" containerName="nova-manage" Oct 14 07:23:03 crc kubenswrapper[4870]: I1014 07:23:03.001427 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="1110cc74-c641-4a3a-b46d-ea85668515b0" containerName="nova-manage" Oct 14 07:23:03 crc kubenswrapper[4870]: E1014 07:23:03.001453 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f09945e-c7d1-4bfe-a8d7-7421b1a590ec" containerName="nova-api-log" Oct 14 07:23:03 crc kubenswrapper[4870]: I1014 07:23:03.001461 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f09945e-c7d1-4bfe-a8d7-7421b1a590ec" containerName="nova-api-log" Oct 14 07:23:03 crc kubenswrapper[4870]: E1014 07:23:03.001481 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4a361161-c100-49cb-a4ee-19563b724033" containerName="nova-metadata-metadata" Oct 14 07:23:03 crc kubenswrapper[4870]: I1014 07:23:03.001489 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="4a361161-c100-49cb-a4ee-19563b724033" containerName="nova-metadata-metadata" Oct 14 07:23:03 crc kubenswrapper[4870]: I1014 07:23:03.001714 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="1110cc74-c641-4a3a-b46d-ea85668515b0" containerName="nova-manage" Oct 14 07:23:03 crc kubenswrapper[4870]: I1014 07:23:03.001742 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="1f09945e-c7d1-4bfe-a8d7-7421b1a590ec" containerName="nova-api-api" Oct 14 07:23:03 crc kubenswrapper[4870]: I1014 07:23:03.001761 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="1f09945e-c7d1-4bfe-a8d7-7421b1a590ec" containerName="nova-api-log" Oct 14 07:23:03 crc kubenswrapper[4870]: I1014 07:23:03.001773 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="4a361161-c100-49cb-a4ee-19563b724033" containerName="nova-metadata-log" Oct 14 07:23:03 crc kubenswrapper[4870]: I1014 07:23:03.001785 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="4a361161-c100-49cb-a4ee-19563b724033" containerName="nova-metadata-metadata" Oct 14 07:23:03 crc kubenswrapper[4870]: I1014 07:23:03.003054 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 14 07:23:03 crc kubenswrapper[4870]: I1014 07:23:03.008495 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 14 07:23:03 crc kubenswrapper[4870]: I1014 07:23:03.008902 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Oct 14 07:23:03 crc kubenswrapper[4870]: I1014 07:23:03.009179 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Oct 14 07:23:03 crc kubenswrapper[4870]: I1014 07:23:03.011175 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 14 07:23:03 crc kubenswrapper[4870]: I1014 07:23:03.061061 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1f09945e-c7d1-4bfe-a8d7-7421b1a590ec" path="/var/lib/kubelet/pods/1f09945e-c7d1-4bfe-a8d7-7421b1a590ec/volumes" Oct 14 07:23:03 crc kubenswrapper[4870]: I1014 07:23:03.061791 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 14 07:23:03 crc kubenswrapper[4870]: I1014 07:23:03.077930 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 14 07:23:03 crc kubenswrapper[4870]: I1014 07:23:03.089967 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 14 07:23:03 crc kubenswrapper[4870]: I1014 07:23:03.091562 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 14 07:23:03 crc kubenswrapper[4870]: I1014 07:23:03.096018 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Oct 14 07:23:03 crc kubenswrapper[4870]: I1014 07:23:03.096244 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 14 07:23:03 crc kubenswrapper[4870]: I1014 07:23:03.097642 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 14 07:23:03 crc kubenswrapper[4870]: I1014 07:23:03.133664 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6cpz2\" (UniqueName: \"kubernetes.io/projected/f0b766fd-ccb9-4620-84df-8fe2f194b0da-kube-api-access-6cpz2\") pod \"nova-metadata-0\" (UID: \"f0b766fd-ccb9-4620-84df-8fe2f194b0da\") " pod="openstack/nova-metadata-0" Oct 14 07:23:03 crc kubenswrapper[4870]: I1014 07:23:03.133740 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f0b766fd-ccb9-4620-84df-8fe2f194b0da-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"f0b766fd-ccb9-4620-84df-8fe2f194b0da\") " pod="openstack/nova-metadata-0" Oct 14 07:23:03 crc kubenswrapper[4870]: I1014 07:23:03.133906 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pvhxn\" (UniqueName: \"kubernetes.io/projected/ff28d060-b280-4b89-bea1-d10aa367a3cc-kube-api-access-pvhxn\") pod \"nova-api-0\" (UID: \"ff28d060-b280-4b89-bea1-d10aa367a3cc\") " pod="openstack/nova-api-0" Oct 14 07:23:03 crc kubenswrapper[4870]: I1014 07:23:03.134007 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff28d060-b280-4b89-bea1-d10aa367a3cc-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"ff28d060-b280-4b89-bea1-d10aa367a3cc\") " pod="openstack/nova-api-0" Oct 14 07:23:03 crc kubenswrapper[4870]: I1014 07:23:03.134098 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/f0b766fd-ccb9-4620-84df-8fe2f194b0da-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"f0b766fd-ccb9-4620-84df-8fe2f194b0da\") " pod="openstack/nova-metadata-0" Oct 14 07:23:03 crc kubenswrapper[4870]: I1014 07:23:03.134203 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f0b766fd-ccb9-4620-84df-8fe2f194b0da-logs\") pod \"nova-metadata-0\" (UID: \"f0b766fd-ccb9-4620-84df-8fe2f194b0da\") " pod="openstack/nova-metadata-0" Oct 14 07:23:03 crc kubenswrapper[4870]: I1014 07:23:03.134263 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff28d060-b280-4b89-bea1-d10aa367a3cc-config-data\") pod \"nova-api-0\" (UID: \"ff28d060-b280-4b89-bea1-d10aa367a3cc\") " pod="openstack/nova-api-0" Oct 14 07:23:03 crc kubenswrapper[4870]: I1014 07:23:03.134284 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f0b766fd-ccb9-4620-84df-8fe2f194b0da-config-data\") pod \"nova-metadata-0\" (UID: \"f0b766fd-ccb9-4620-84df-8fe2f194b0da\") " pod="openstack/nova-metadata-0" Oct 14 07:23:03 crc kubenswrapper[4870]: I1014 07:23:03.134399 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ff28d060-b280-4b89-bea1-d10aa367a3cc-public-tls-certs\") pod \"nova-api-0\" (UID: \"ff28d060-b280-4b89-bea1-d10aa367a3cc\") " pod="openstack/nova-api-0" Oct 14 07:23:03 crc kubenswrapper[4870]: I1014 07:23:03.134476 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ff28d060-b280-4b89-bea1-d10aa367a3cc-internal-tls-certs\") pod \"nova-api-0\" (UID: \"ff28d060-b280-4b89-bea1-d10aa367a3cc\") " pod="openstack/nova-api-0" Oct 14 07:23:03 crc kubenswrapper[4870]: I1014 07:23:03.134572 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ff28d060-b280-4b89-bea1-d10aa367a3cc-logs\") pod \"nova-api-0\" (UID: \"ff28d060-b280-4b89-bea1-d10aa367a3cc\") " pod="openstack/nova-api-0" Oct 14 07:23:03 crc kubenswrapper[4870]: I1014 07:23:03.236517 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6cpz2\" (UniqueName: \"kubernetes.io/projected/f0b766fd-ccb9-4620-84df-8fe2f194b0da-kube-api-access-6cpz2\") pod \"nova-metadata-0\" (UID: \"f0b766fd-ccb9-4620-84df-8fe2f194b0da\") " pod="openstack/nova-metadata-0" Oct 14 07:23:03 crc kubenswrapper[4870]: I1014 07:23:03.236654 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f0b766fd-ccb9-4620-84df-8fe2f194b0da-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"f0b766fd-ccb9-4620-84df-8fe2f194b0da\") " pod="openstack/nova-metadata-0" Oct 14 07:23:03 crc kubenswrapper[4870]: I1014 07:23:03.236725 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pvhxn\" (UniqueName: \"kubernetes.io/projected/ff28d060-b280-4b89-bea1-d10aa367a3cc-kube-api-access-pvhxn\") pod \"nova-api-0\" (UID: \"ff28d060-b280-4b89-bea1-d10aa367a3cc\") " pod="openstack/nova-api-0" Oct 14 07:23:03 crc kubenswrapper[4870]: I1014 07:23:03.236781 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff28d060-b280-4b89-bea1-d10aa367a3cc-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"ff28d060-b280-4b89-bea1-d10aa367a3cc\") " pod="openstack/nova-api-0" Oct 14 07:23:03 crc kubenswrapper[4870]: I1014 07:23:03.236851 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/f0b766fd-ccb9-4620-84df-8fe2f194b0da-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"f0b766fd-ccb9-4620-84df-8fe2f194b0da\") " pod="openstack/nova-metadata-0" Oct 14 07:23:03 crc kubenswrapper[4870]: I1014 07:23:03.236897 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f0b766fd-ccb9-4620-84df-8fe2f194b0da-logs\") pod \"nova-metadata-0\" (UID: \"f0b766fd-ccb9-4620-84df-8fe2f194b0da\") " pod="openstack/nova-metadata-0" Oct 14 07:23:03 crc kubenswrapper[4870]: I1014 07:23:03.236945 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff28d060-b280-4b89-bea1-d10aa367a3cc-config-data\") pod \"nova-api-0\" (UID: \"ff28d060-b280-4b89-bea1-d10aa367a3cc\") " pod="openstack/nova-api-0" Oct 14 07:23:03 crc kubenswrapper[4870]: I1014 07:23:03.236977 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f0b766fd-ccb9-4620-84df-8fe2f194b0da-config-data\") pod \"nova-metadata-0\" (UID: \"f0b766fd-ccb9-4620-84df-8fe2f194b0da\") " pod="openstack/nova-metadata-0" Oct 14 07:23:03 crc kubenswrapper[4870]: I1014 07:23:03.237039 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ff28d060-b280-4b89-bea1-d10aa367a3cc-public-tls-certs\") pod \"nova-api-0\" (UID: \"ff28d060-b280-4b89-bea1-d10aa367a3cc\") " pod="openstack/nova-api-0" Oct 14 07:23:03 crc kubenswrapper[4870]: I1014 07:23:03.237090 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ff28d060-b280-4b89-bea1-d10aa367a3cc-internal-tls-certs\") pod \"nova-api-0\" (UID: \"ff28d060-b280-4b89-bea1-d10aa367a3cc\") " pod="openstack/nova-api-0" Oct 14 07:23:03 crc kubenswrapper[4870]: I1014 07:23:03.237123 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ff28d060-b280-4b89-bea1-d10aa367a3cc-logs\") pod \"nova-api-0\" (UID: \"ff28d060-b280-4b89-bea1-d10aa367a3cc\") " pod="openstack/nova-api-0" Oct 14 07:23:03 crc kubenswrapper[4870]: I1014 07:23:03.237730 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ff28d060-b280-4b89-bea1-d10aa367a3cc-logs\") pod \"nova-api-0\" (UID: \"ff28d060-b280-4b89-bea1-d10aa367a3cc\") " pod="openstack/nova-api-0" Oct 14 07:23:03 crc kubenswrapper[4870]: I1014 07:23:03.238214 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f0b766fd-ccb9-4620-84df-8fe2f194b0da-logs\") pod \"nova-metadata-0\" (UID: \"f0b766fd-ccb9-4620-84df-8fe2f194b0da\") " pod="openstack/nova-metadata-0" Oct 14 07:23:03 crc kubenswrapper[4870]: I1014 07:23:03.240576 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f0b766fd-ccb9-4620-84df-8fe2f194b0da-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"f0b766fd-ccb9-4620-84df-8fe2f194b0da\") " pod="openstack/nova-metadata-0" Oct 14 07:23:03 crc kubenswrapper[4870]: I1014 07:23:03.243573 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ff28d060-b280-4b89-bea1-d10aa367a3cc-internal-tls-certs\") pod \"nova-api-0\" (UID: \"ff28d060-b280-4b89-bea1-d10aa367a3cc\") " pod="openstack/nova-api-0" Oct 14 07:23:03 crc kubenswrapper[4870]: I1014 07:23:03.246532 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff28d060-b280-4b89-bea1-d10aa367a3cc-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"ff28d060-b280-4b89-bea1-d10aa367a3cc\") " pod="openstack/nova-api-0" Oct 14 07:23:03 crc kubenswrapper[4870]: I1014 07:23:03.250241 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/f0b766fd-ccb9-4620-84df-8fe2f194b0da-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"f0b766fd-ccb9-4620-84df-8fe2f194b0da\") " pod="openstack/nova-metadata-0" Oct 14 07:23:03 crc kubenswrapper[4870]: I1014 07:23:03.251750 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff28d060-b280-4b89-bea1-d10aa367a3cc-config-data\") pod \"nova-api-0\" (UID: \"ff28d060-b280-4b89-bea1-d10aa367a3cc\") " pod="openstack/nova-api-0" Oct 14 07:23:03 crc kubenswrapper[4870]: I1014 07:23:03.254673 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f0b766fd-ccb9-4620-84df-8fe2f194b0da-config-data\") pod \"nova-metadata-0\" (UID: \"f0b766fd-ccb9-4620-84df-8fe2f194b0da\") " pod="openstack/nova-metadata-0" Oct 14 07:23:03 crc kubenswrapper[4870]: I1014 07:23:03.258686 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ff28d060-b280-4b89-bea1-d10aa367a3cc-public-tls-certs\") pod \"nova-api-0\" (UID: \"ff28d060-b280-4b89-bea1-d10aa367a3cc\") " pod="openstack/nova-api-0" Oct 14 07:23:03 crc kubenswrapper[4870]: I1014 07:23:03.263651 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6cpz2\" (UniqueName: \"kubernetes.io/projected/f0b766fd-ccb9-4620-84df-8fe2f194b0da-kube-api-access-6cpz2\") pod \"nova-metadata-0\" (UID: \"f0b766fd-ccb9-4620-84df-8fe2f194b0da\") " pod="openstack/nova-metadata-0" Oct 14 07:23:03 crc kubenswrapper[4870]: I1014 07:23:03.266756 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pvhxn\" (UniqueName: \"kubernetes.io/projected/ff28d060-b280-4b89-bea1-d10aa367a3cc-kube-api-access-pvhxn\") pod \"nova-api-0\" (UID: \"ff28d060-b280-4b89-bea1-d10aa367a3cc\") " pod="openstack/nova-api-0" Oct 14 07:23:03 crc kubenswrapper[4870]: I1014 07:23:03.331605 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 14 07:23:03 crc kubenswrapper[4870]: I1014 07:23:03.406585 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 14 07:23:03 crc kubenswrapper[4870]: I1014 07:23:03.785474 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 14 07:23:03 crc kubenswrapper[4870]: I1014 07:23:03.870262 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 14 07:23:03 crc kubenswrapper[4870]: I1014 07:23:03.923371 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 14 07:23:03 crc kubenswrapper[4870]: I1014 07:23:03.923432 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"ffe53321-c66f-4b32-b18d-18bc483f923b","Type":"ContainerDied","Data":"56e70d8d8996b806615045a193a6672b63fec50f57799575de3be2115113662c"} Oct 14 07:23:03 crc kubenswrapper[4870]: I1014 07:23:03.923502 4870 generic.go:334] "Generic (PLEG): container finished" podID="ffe53321-c66f-4b32-b18d-18bc483f923b" containerID="56e70d8d8996b806615045a193a6672b63fec50f57799575de3be2115113662c" exitCode=0 Oct 14 07:23:03 crc kubenswrapper[4870]: I1014 07:23:03.924126 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"ffe53321-c66f-4b32-b18d-18bc483f923b","Type":"ContainerDied","Data":"5e80cf2ed5aa73531137db469a8d31176086b6df05b2af4503656e893daa0f4a"} Oct 14 07:23:03 crc kubenswrapper[4870]: I1014 07:23:03.924232 4870 scope.go:117] "RemoveContainer" containerID="56e70d8d8996b806615045a193a6672b63fec50f57799575de3be2115113662c" Oct 14 07:23:03 crc kubenswrapper[4870]: I1014 07:23:03.929195 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ff28d060-b280-4b89-bea1-d10aa367a3cc","Type":"ContainerStarted","Data":"d2c1e6ab1ae6834143ee40554a51da1a2d368bfb5f191a9abce3ca8a5ee4d09a"} Oct 14 07:23:03 crc kubenswrapper[4870]: I1014 07:23:03.938418 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 14 07:23:03 crc kubenswrapper[4870]: I1014 07:23:03.948175 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ffe53321-c66f-4b32-b18d-18bc483f923b-config-data\") pod \"ffe53321-c66f-4b32-b18d-18bc483f923b\" (UID: \"ffe53321-c66f-4b32-b18d-18bc483f923b\") " Oct 14 07:23:03 crc kubenswrapper[4870]: I1014 07:23:03.948286 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ffe53321-c66f-4b32-b18d-18bc483f923b-combined-ca-bundle\") pod \"ffe53321-c66f-4b32-b18d-18bc483f923b\" (UID: \"ffe53321-c66f-4b32-b18d-18bc483f923b\") " Oct 14 07:23:03 crc kubenswrapper[4870]: I1014 07:23:03.948385 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sf6ps\" (UniqueName: \"kubernetes.io/projected/ffe53321-c66f-4b32-b18d-18bc483f923b-kube-api-access-sf6ps\") pod \"ffe53321-c66f-4b32-b18d-18bc483f923b\" (UID: \"ffe53321-c66f-4b32-b18d-18bc483f923b\") " Oct 14 07:23:03 crc kubenswrapper[4870]: I1014 07:23:03.956116 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ffe53321-c66f-4b32-b18d-18bc483f923b-kube-api-access-sf6ps" (OuterVolumeSpecName: "kube-api-access-sf6ps") pod "ffe53321-c66f-4b32-b18d-18bc483f923b" (UID: "ffe53321-c66f-4b32-b18d-18bc483f923b"). InnerVolumeSpecName "kube-api-access-sf6ps". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:23:03 crc kubenswrapper[4870]: I1014 07:23:03.994972 4870 scope.go:117] "RemoveContainer" containerID="56e70d8d8996b806615045a193a6672b63fec50f57799575de3be2115113662c" Oct 14 07:23:03 crc kubenswrapper[4870]: E1014 07:23:03.995584 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"56e70d8d8996b806615045a193a6672b63fec50f57799575de3be2115113662c\": container with ID starting with 56e70d8d8996b806615045a193a6672b63fec50f57799575de3be2115113662c not found: ID does not exist" containerID="56e70d8d8996b806615045a193a6672b63fec50f57799575de3be2115113662c" Oct 14 07:23:03 crc kubenswrapper[4870]: I1014 07:23:03.995661 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"56e70d8d8996b806615045a193a6672b63fec50f57799575de3be2115113662c"} err="failed to get container status \"56e70d8d8996b806615045a193a6672b63fec50f57799575de3be2115113662c\": rpc error: code = NotFound desc = could not find container \"56e70d8d8996b806615045a193a6672b63fec50f57799575de3be2115113662c\": container with ID starting with 56e70d8d8996b806615045a193a6672b63fec50f57799575de3be2115113662c not found: ID does not exist" Oct 14 07:23:04 crc kubenswrapper[4870]: I1014 07:23:04.009143 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ffe53321-c66f-4b32-b18d-18bc483f923b-config-data" (OuterVolumeSpecName: "config-data") pod "ffe53321-c66f-4b32-b18d-18bc483f923b" (UID: "ffe53321-c66f-4b32-b18d-18bc483f923b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:23:04 crc kubenswrapper[4870]: I1014 07:23:04.016042 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ffe53321-c66f-4b32-b18d-18bc483f923b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ffe53321-c66f-4b32-b18d-18bc483f923b" (UID: "ffe53321-c66f-4b32-b18d-18bc483f923b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:23:04 crc kubenswrapper[4870]: I1014 07:23:04.051524 4870 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ffe53321-c66f-4b32-b18d-18bc483f923b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:04 crc kubenswrapper[4870]: I1014 07:23:04.051558 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sf6ps\" (UniqueName: \"kubernetes.io/projected/ffe53321-c66f-4b32-b18d-18bc483f923b-kube-api-access-sf6ps\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:04 crc kubenswrapper[4870]: I1014 07:23:04.051571 4870 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ffe53321-c66f-4b32-b18d-18bc483f923b-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:04 crc kubenswrapper[4870]: I1014 07:23:04.267582 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 14 07:23:04 crc kubenswrapper[4870]: I1014 07:23:04.280333 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Oct 14 07:23:04 crc kubenswrapper[4870]: I1014 07:23:04.292497 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 14 07:23:04 crc kubenswrapper[4870]: E1014 07:23:04.293022 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ffe53321-c66f-4b32-b18d-18bc483f923b" containerName="nova-scheduler-scheduler" Oct 14 07:23:04 crc kubenswrapper[4870]: I1014 07:23:04.293041 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="ffe53321-c66f-4b32-b18d-18bc483f923b" containerName="nova-scheduler-scheduler" Oct 14 07:23:04 crc kubenswrapper[4870]: I1014 07:23:04.293256 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="ffe53321-c66f-4b32-b18d-18bc483f923b" containerName="nova-scheduler-scheduler" Oct 14 07:23:04 crc kubenswrapper[4870]: I1014 07:23:04.294184 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 14 07:23:04 crc kubenswrapper[4870]: I1014 07:23:04.296462 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 14 07:23:04 crc kubenswrapper[4870]: I1014 07:23:04.304242 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 14 07:23:04 crc kubenswrapper[4870]: I1014 07:23:04.357965 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f5b700ed-3919-458f-860e-5630b8e4b800-config-data\") pod \"nova-scheduler-0\" (UID: \"f5b700ed-3919-458f-860e-5630b8e4b800\") " pod="openstack/nova-scheduler-0" Oct 14 07:23:04 crc kubenswrapper[4870]: I1014 07:23:04.358328 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n4c9p\" (UniqueName: \"kubernetes.io/projected/f5b700ed-3919-458f-860e-5630b8e4b800-kube-api-access-n4c9p\") pod \"nova-scheduler-0\" (UID: \"f5b700ed-3919-458f-860e-5630b8e4b800\") " pod="openstack/nova-scheduler-0" Oct 14 07:23:04 crc kubenswrapper[4870]: I1014 07:23:04.358537 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5b700ed-3919-458f-860e-5630b8e4b800-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"f5b700ed-3919-458f-860e-5630b8e4b800\") " pod="openstack/nova-scheduler-0" Oct 14 07:23:04 crc kubenswrapper[4870]: I1014 07:23:04.460333 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f5b700ed-3919-458f-860e-5630b8e4b800-config-data\") pod \"nova-scheduler-0\" (UID: \"f5b700ed-3919-458f-860e-5630b8e4b800\") " pod="openstack/nova-scheduler-0" Oct 14 07:23:04 crc kubenswrapper[4870]: I1014 07:23:04.460713 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n4c9p\" (UniqueName: \"kubernetes.io/projected/f5b700ed-3919-458f-860e-5630b8e4b800-kube-api-access-n4c9p\") pod \"nova-scheduler-0\" (UID: \"f5b700ed-3919-458f-860e-5630b8e4b800\") " pod="openstack/nova-scheduler-0" Oct 14 07:23:04 crc kubenswrapper[4870]: I1014 07:23:04.460754 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5b700ed-3919-458f-860e-5630b8e4b800-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"f5b700ed-3919-458f-860e-5630b8e4b800\") " pod="openstack/nova-scheduler-0" Oct 14 07:23:04 crc kubenswrapper[4870]: I1014 07:23:04.466083 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f5b700ed-3919-458f-860e-5630b8e4b800-config-data\") pod \"nova-scheduler-0\" (UID: \"f5b700ed-3919-458f-860e-5630b8e4b800\") " pod="openstack/nova-scheduler-0" Oct 14 07:23:04 crc kubenswrapper[4870]: I1014 07:23:04.466195 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5b700ed-3919-458f-860e-5630b8e4b800-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"f5b700ed-3919-458f-860e-5630b8e4b800\") " pod="openstack/nova-scheduler-0" Oct 14 07:23:04 crc kubenswrapper[4870]: I1014 07:23:04.491486 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n4c9p\" (UniqueName: \"kubernetes.io/projected/f5b700ed-3919-458f-860e-5630b8e4b800-kube-api-access-n4c9p\") pod \"nova-scheduler-0\" (UID: \"f5b700ed-3919-458f-860e-5630b8e4b800\") " pod="openstack/nova-scheduler-0" Oct 14 07:23:04 crc kubenswrapper[4870]: I1014 07:23:04.625628 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 14 07:23:04 crc kubenswrapper[4870]: I1014 07:23:04.946372 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ff28d060-b280-4b89-bea1-d10aa367a3cc","Type":"ContainerStarted","Data":"9867cb00fe7f9cff6394884f9edfc786d6b848597f09a56abffe01cb5e8ca4cf"} Oct 14 07:23:04 crc kubenswrapper[4870]: I1014 07:23:04.946791 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ff28d060-b280-4b89-bea1-d10aa367a3cc","Type":"ContainerStarted","Data":"6d95049e97ec6fbbfa6d28e3e543ed1f7fc74569f83d0862a392737896e3463d"} Oct 14 07:23:04 crc kubenswrapper[4870]: I1014 07:23:04.949896 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f0b766fd-ccb9-4620-84df-8fe2f194b0da","Type":"ContainerStarted","Data":"65770c73c87b754f9177bcc262af5253b96b0534c636832b413175980c00c785"} Oct 14 07:23:04 crc kubenswrapper[4870]: I1014 07:23:04.949955 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f0b766fd-ccb9-4620-84df-8fe2f194b0da","Type":"ContainerStarted","Data":"6fb3ea4d6cd5802f3695559f939e61f05770876bd3c155dcf6040899ef93330a"} Oct 14 07:23:04 crc kubenswrapper[4870]: I1014 07:23:04.949971 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f0b766fd-ccb9-4620-84df-8fe2f194b0da","Type":"ContainerStarted","Data":"f20638fcf7487657845b9e4bea7d6a1a36ecd836a895de79af97612151ecb066"} Oct 14 07:23:04 crc kubenswrapper[4870]: I1014 07:23:04.967696 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.9676829209999998 podStartE2EDuration="2.967682921s" podCreationTimestamp="2025-10-14 07:23:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:23:04.964073211 +0000 UTC m=+1320.661433582" watchObservedRunningTime="2025-10-14 07:23:04.967682921 +0000 UTC m=+1320.665043292" Oct 14 07:23:04 crc kubenswrapper[4870]: I1014 07:23:04.992026 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.9920064870000003 podStartE2EDuration="2.992006487s" podCreationTimestamp="2025-10-14 07:23:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:23:04.984839378 +0000 UTC m=+1320.682199749" watchObservedRunningTime="2025-10-14 07:23:04.992006487 +0000 UTC m=+1320.689366858" Oct 14 07:23:05 crc kubenswrapper[4870]: I1014 07:23:05.024511 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 14 07:23:05 crc kubenswrapper[4870]: W1014 07:23:05.035759 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf5b700ed_3919_458f_860e_5630b8e4b800.slice/crio-aeed74d1fcfc0d32441c8025c1ed06c4fa2e3b5be6c136e68861143e7f83a891 WatchSource:0}: Error finding container aeed74d1fcfc0d32441c8025c1ed06c4fa2e3b5be6c136e68861143e7f83a891: Status 404 returned error can't find the container with id aeed74d1fcfc0d32441c8025c1ed06c4fa2e3b5be6c136e68861143e7f83a891 Oct 14 07:23:05 crc kubenswrapper[4870]: I1014 07:23:05.046778 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4a361161-c100-49cb-a4ee-19563b724033" path="/var/lib/kubelet/pods/4a361161-c100-49cb-a4ee-19563b724033/volumes" Oct 14 07:23:05 crc kubenswrapper[4870]: I1014 07:23:05.047517 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ffe53321-c66f-4b32-b18d-18bc483f923b" path="/var/lib/kubelet/pods/ffe53321-c66f-4b32-b18d-18bc483f923b/volumes" Oct 14 07:23:05 crc kubenswrapper[4870]: I1014 07:23:05.968369 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"f5b700ed-3919-458f-860e-5630b8e4b800","Type":"ContainerStarted","Data":"19bfdce54348e895d0d26a8651c758e77337522908dc3ad6eb2470e00c72c2cf"} Oct 14 07:23:05 crc kubenswrapper[4870]: I1014 07:23:05.968794 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"f5b700ed-3919-458f-860e-5630b8e4b800","Type":"ContainerStarted","Data":"aeed74d1fcfc0d32441c8025c1ed06c4fa2e3b5be6c136e68861143e7f83a891"} Oct 14 07:23:05 crc kubenswrapper[4870]: I1014 07:23:05.990089 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=1.990068699 podStartE2EDuration="1.990068699s" podCreationTimestamp="2025-10-14 07:23:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 07:23:05.984295535 +0000 UTC m=+1321.681655906" watchObservedRunningTime="2025-10-14 07:23:05.990068699 +0000 UTC m=+1321.687429070" Oct 14 07:23:08 crc kubenswrapper[4870]: I1014 07:23:08.407666 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 14 07:23:08 crc kubenswrapper[4870]: I1014 07:23:08.407744 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 14 07:23:09 crc kubenswrapper[4870]: I1014 07:23:09.625827 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 14 07:23:12 crc kubenswrapper[4870]: I1014 07:23:12.954135 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Oct 14 07:23:13 crc kubenswrapper[4870]: I1014 07:23:13.333026 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 14 07:23:13 crc kubenswrapper[4870]: I1014 07:23:13.333098 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 14 07:23:13 crc kubenswrapper[4870]: I1014 07:23:13.407177 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 14 07:23:13 crc kubenswrapper[4870]: I1014 07:23:13.407253 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 14 07:23:14 crc kubenswrapper[4870]: I1014 07:23:14.351584 4870 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="ff28d060-b280-4b89-bea1-d10aa367a3cc" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.201:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 14 07:23:14 crc kubenswrapper[4870]: I1014 07:23:14.351633 4870 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="ff28d060-b280-4b89-bea1-d10aa367a3cc" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.201:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 14 07:23:14 crc kubenswrapper[4870]: I1014 07:23:14.420727 4870 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="f0b766fd-ccb9-4620-84df-8fe2f194b0da" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.202:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 14 07:23:14 crc kubenswrapper[4870]: I1014 07:23:14.420744 4870 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="f0b766fd-ccb9-4620-84df-8fe2f194b0da" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.202:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 14 07:23:14 crc kubenswrapper[4870]: I1014 07:23:14.626461 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 14 07:23:14 crc kubenswrapper[4870]: I1014 07:23:14.679712 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 14 07:23:15 crc kubenswrapper[4870]: I1014 07:23:15.137588 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 14 07:23:23 crc kubenswrapper[4870]: I1014 07:23:23.346745 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 14 07:23:23 crc kubenswrapper[4870]: I1014 07:23:23.349294 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 14 07:23:23 crc kubenswrapper[4870]: I1014 07:23:23.349361 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 14 07:23:23 crc kubenswrapper[4870]: I1014 07:23:23.359136 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 14 07:23:23 crc kubenswrapper[4870]: I1014 07:23:23.424110 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 14 07:23:23 crc kubenswrapper[4870]: I1014 07:23:23.424967 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 14 07:23:23 crc kubenswrapper[4870]: I1014 07:23:23.435997 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 14 07:23:24 crc kubenswrapper[4870]: I1014 07:23:24.192887 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 14 07:23:24 crc kubenswrapper[4870]: I1014 07:23:24.200495 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 14 07:23:24 crc kubenswrapper[4870]: I1014 07:23:24.205679 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 14 07:23:42 crc kubenswrapper[4870]: I1014 07:23:42.223490 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 14 07:23:42 crc kubenswrapper[4870]: I1014 07:23:42.279048 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Oct 14 07:23:42 crc kubenswrapper[4870]: I1014 07:23:42.279574 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstackclient" podUID="0965b7e6-2aa0-4940-a130-324cfb08de5a" containerName="openstackclient" containerID="cri-o://044ea035cd6094777207ef8f180d32297801ac2dd26be47e987ae4b7318523d6" gracePeriod=2 Oct 14 07:23:42 crc kubenswrapper[4870]: I1014 07:23:42.287570 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Oct 14 07:23:42 crc kubenswrapper[4870]: E1014 07:23:42.357928 4870 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Oct 14 07:23:42 crc kubenswrapper[4870]: E1014 07:23:42.358425 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5212ba20-83c3-41ff-a3d9-5a315f009c1b-config-data podName:5212ba20-83c3-41ff-a3d9-5a315f009c1b nodeName:}" failed. No retries permitted until 2025-10-14 07:23:42.858391194 +0000 UTC m=+1358.555751565 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/5212ba20-83c3-41ff-a3d9-5a315f009c1b-config-data") pod "rabbitmq-cell1-server-0" (UID: "5212ba20-83c3-41ff-a3d9-5a315f009c1b") : configmap "rabbitmq-cell1-config-data" not found Oct 14 07:23:42 crc kubenswrapper[4870]: I1014 07:23:42.444852 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 14 07:23:42 crc kubenswrapper[4870]: I1014 07:23:42.445524 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovsdbserver-nb-0" podUID="09f89569-6fe3-4b3f-9394-a990ad8cde30" containerName="openstack-network-exporter" containerID="cri-o://b03ec54549f3acea91303d6cc01b2627a9ebd5ec3e78dcb4527b100efe1ebd9c" gracePeriod=300 Oct 14 07:23:42 crc kubenswrapper[4870]: I1014 07:23:42.491399 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 14 07:23:42 crc kubenswrapper[4870]: I1014 07:23:42.539699 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement1a80-account-delete-8q7fq"] Oct 14 07:23:42 crc kubenswrapper[4870]: E1014 07:23:42.540269 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0965b7e6-2aa0-4940-a130-324cfb08de5a" containerName="openstackclient" Oct 14 07:23:42 crc kubenswrapper[4870]: I1014 07:23:42.540281 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="0965b7e6-2aa0-4940-a130-324cfb08de5a" containerName="openstackclient" Oct 14 07:23:42 crc kubenswrapper[4870]: I1014 07:23:42.540529 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="0965b7e6-2aa0-4940-a130-324cfb08de5a" containerName="openstackclient" Oct 14 07:23:42 crc kubenswrapper[4870]: I1014 07:23:42.541304 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement1a80-account-delete-8q7fq" Oct 14 07:23:42 crc kubenswrapper[4870]: I1014 07:23:42.558537 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder5c98-account-delete-lhjjg"] Oct 14 07:23:42 crc kubenswrapper[4870]: I1014 07:23:42.559811 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder5c98-account-delete-lhjjg" Oct 14 07:23:42 crc kubenswrapper[4870]: I1014 07:23:42.564807 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j2gjx\" (UniqueName: \"kubernetes.io/projected/de479f5d-22a2-4f7e-8034-cffefc511b53-kube-api-access-j2gjx\") pod \"cinder5c98-account-delete-lhjjg\" (UID: \"de479f5d-22a2-4f7e-8034-cffefc511b53\") " pod="openstack/cinder5c98-account-delete-lhjjg" Oct 14 07:23:42 crc kubenswrapper[4870]: I1014 07:23:42.564869 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mnlql\" (UniqueName: \"kubernetes.io/projected/7cbdbbdd-ea4f-4d14-8f3c-4784bda45874-kube-api-access-mnlql\") pod \"placement1a80-account-delete-8q7fq\" (UID: \"7cbdbbdd-ea4f-4d14-8f3c-4784bda45874\") " pod="openstack/placement1a80-account-delete-8q7fq" Oct 14 07:23:42 crc kubenswrapper[4870]: I1014 07:23:42.594932 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement1a80-account-delete-8q7fq"] Oct 14 07:23:42 crc kubenswrapper[4870]: I1014 07:23:42.604813 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder5c98-account-delete-lhjjg"] Oct 14 07:23:42 crc kubenswrapper[4870]: I1014 07:23:42.629151 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovsdbserver-nb-0" podUID="09f89569-6fe3-4b3f-9394-a990ad8cde30" containerName="ovsdbserver-nb" containerID="cri-o://76d17f5aef5402fa0f6135921f3ba0aeb9f5640152e827b4385348efdbe620d3" gracePeriod=300 Oct 14 07:23:42 crc kubenswrapper[4870]: I1014 07:23:42.649704 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican6547-account-delete-v5crl"] Oct 14 07:23:42 crc kubenswrapper[4870]: I1014 07:23:42.650896 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican6547-account-delete-v5crl" Oct 14 07:23:42 crc kubenswrapper[4870]: I1014 07:23:42.664171 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican6547-account-delete-v5crl"] Oct 14 07:23:42 crc kubenswrapper[4870]: I1014 07:23:42.674614 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j2gjx\" (UniqueName: \"kubernetes.io/projected/de479f5d-22a2-4f7e-8034-cffefc511b53-kube-api-access-j2gjx\") pod \"cinder5c98-account-delete-lhjjg\" (UID: \"de479f5d-22a2-4f7e-8034-cffefc511b53\") " pod="openstack/cinder5c98-account-delete-lhjjg" Oct 14 07:23:42 crc kubenswrapper[4870]: I1014 07:23:42.674692 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mnlql\" (UniqueName: \"kubernetes.io/projected/7cbdbbdd-ea4f-4d14-8f3c-4784bda45874-kube-api-access-mnlql\") pod \"placement1a80-account-delete-8q7fq\" (UID: \"7cbdbbdd-ea4f-4d14-8f3c-4784bda45874\") " pod="openstack/placement1a80-account-delete-8q7fq" Oct 14 07:23:42 crc kubenswrapper[4870]: E1014 07:23:42.676252 4870 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Oct 14 07:23:42 crc kubenswrapper[4870]: E1014 07:23:42.676314 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/da8a9731-f917-462f-9932-b37b6abb9a64-config-data podName:da8a9731-f917-462f-9932-b37b6abb9a64 nodeName:}" failed. No retries permitted until 2025-10-14 07:23:43.176296396 +0000 UTC m=+1358.873656767 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/da8a9731-f917-462f-9932-b37b6abb9a64-config-data") pod "rabbitmq-server-0" (UID: "da8a9731-f917-462f-9932-b37b6abb9a64") : configmap "rabbitmq-config-data" not found Oct 14 07:23:42 crc kubenswrapper[4870]: I1014 07:23:42.711208 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j2gjx\" (UniqueName: \"kubernetes.io/projected/de479f5d-22a2-4f7e-8034-cffefc511b53-kube-api-access-j2gjx\") pod \"cinder5c98-account-delete-lhjjg\" (UID: \"de479f5d-22a2-4f7e-8034-cffefc511b53\") " pod="openstack/cinder5c98-account-delete-lhjjg" Oct 14 07:23:42 crc kubenswrapper[4870]: I1014 07:23:42.732322 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mnlql\" (UniqueName: \"kubernetes.io/projected/7cbdbbdd-ea4f-4d14-8f3c-4784bda45874-kube-api-access-mnlql\") pod \"placement1a80-account-delete-8q7fq\" (UID: \"7cbdbbdd-ea4f-4d14-8f3c-4784bda45874\") " pod="openstack/placement1a80-account-delete-8q7fq" Oct 14 07:23:42 crc kubenswrapper[4870]: I1014 07:23:42.740738 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-northd-0"] Oct 14 07:23:42 crc kubenswrapper[4870]: I1014 07:23:42.741008 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-northd-0" podUID="416d7b39-4541-44a8-a55f-924bc86fee32" containerName="ovn-northd" containerID="cri-o://457905ca1130c3804c068cf6fc44521ad064a3fb34449d666cf858f14e2feace" gracePeriod=30 Oct 14 07:23:42 crc kubenswrapper[4870]: I1014 07:23:42.741668 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-northd-0" podUID="416d7b39-4541-44a8-a55f-924bc86fee32" containerName="openstack-network-exporter" containerID="cri-o://11e57f381235088d13225057331ed61d6e2ab6f98595ff6fe4992063044159e4" gracePeriod=30 Oct 14 07:23:42 crc kubenswrapper[4870]: I1014 07:23:42.772581 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance1aaa-account-delete-nct6c"] Oct 14 07:23:42 crc kubenswrapper[4870]: I1014 07:23:42.774244 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance1aaa-account-delete-nct6c" Oct 14 07:23:42 crc kubenswrapper[4870]: I1014 07:23:42.775793 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z9mzh\" (UniqueName: \"kubernetes.io/projected/2abee394-ffb3-4dba-8d09-19c1cb476595-kube-api-access-z9mzh\") pod \"barbican6547-account-delete-v5crl\" (UID: \"2abee394-ffb3-4dba-8d09-19c1cb476595\") " pod="openstack/barbican6547-account-delete-v5crl" Oct 14 07:23:42 crc kubenswrapper[4870]: I1014 07:23:42.781531 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance1aaa-account-delete-nct6c"] Oct 14 07:23:42 crc kubenswrapper[4870]: I1014 07:23:42.801243 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-nmtf9"] Oct 14 07:23:42 crc kubenswrapper[4870]: I1014 07:23:42.827591 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-nmtf9"] Oct 14 07:23:42 crc kubenswrapper[4870]: I1014 07:23:42.879524 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sklwp\" (UniqueName: \"kubernetes.io/projected/a9d8a25d-2f1c-483d-bf0d-76a2bf30f137-kube-api-access-sklwp\") pod \"glance1aaa-account-delete-nct6c\" (UID: \"a9d8a25d-2f1c-483d-bf0d-76a2bf30f137\") " pod="openstack/glance1aaa-account-delete-nct6c" Oct 14 07:23:42 crc kubenswrapper[4870]: I1014 07:23:42.879665 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z9mzh\" (UniqueName: \"kubernetes.io/projected/2abee394-ffb3-4dba-8d09-19c1cb476595-kube-api-access-z9mzh\") pod \"barbican6547-account-delete-v5crl\" (UID: \"2abee394-ffb3-4dba-8d09-19c1cb476595\") " pod="openstack/barbican6547-account-delete-v5crl" Oct 14 07:23:42 crc kubenswrapper[4870]: E1014 07:23:42.879968 4870 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Oct 14 07:23:42 crc kubenswrapper[4870]: E1014 07:23:42.880021 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5212ba20-83c3-41ff-a3d9-5a315f009c1b-config-data podName:5212ba20-83c3-41ff-a3d9-5a315f009c1b nodeName:}" failed. No retries permitted until 2025-10-14 07:23:43.880007943 +0000 UTC m=+1359.577368314 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/5212ba20-83c3-41ff-a3d9-5a315f009c1b-config-data") pod "rabbitmq-cell1-server-0" (UID: "5212ba20-83c3-41ff-a3d9-5a315f009c1b") : configmap "rabbitmq-cell1-config-data" not found Oct 14 07:23:42 crc kubenswrapper[4870]: I1014 07:23:42.910959 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z9mzh\" (UniqueName: \"kubernetes.io/projected/2abee394-ffb3-4dba-8d09-19c1cb476595-kube-api-access-z9mzh\") pod \"barbican6547-account-delete-v5crl\" (UID: \"2abee394-ffb3-4dba-8d09-19c1cb476595\") " pod="openstack/barbican6547-account-delete-v5crl" Oct 14 07:23:42 crc kubenswrapper[4870]: I1014 07:23:42.926539 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/novacell0ce16-account-delete-bwgsq"] Oct 14 07:23:42 crc kubenswrapper[4870]: I1014 07:23:42.927826 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/novacell0ce16-account-delete-bwgsq" Oct 14 07:23:42 crc kubenswrapper[4870]: I1014 07:23:42.930229 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/novacell0ce16-account-delete-bwgsq"] Oct 14 07:23:42 crc kubenswrapper[4870]: I1014 07:23:42.936650 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement1a80-account-delete-8q7fq" Oct 14 07:23:42 crc kubenswrapper[4870]: I1014 07:23:42.979222 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder5c98-account-delete-lhjjg" Oct 14 07:23:42 crc kubenswrapper[4870]: I1014 07:23:42.980637 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sklwp\" (UniqueName: \"kubernetes.io/projected/a9d8a25d-2f1c-483d-bf0d-76a2bf30f137-kube-api-access-sklwp\") pod \"glance1aaa-account-delete-nct6c\" (UID: \"a9d8a25d-2f1c-483d-bf0d-76a2bf30f137\") " pod="openstack/glance1aaa-account-delete-nct6c" Oct 14 07:23:42 crc kubenswrapper[4870]: I1014 07:23:42.989239 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican6547-account-delete-v5crl" Oct 14 07:23:43 crc kubenswrapper[4870]: I1014 07:23:43.019863 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sklwp\" (UniqueName: \"kubernetes.io/projected/a9d8a25d-2f1c-483d-bf0d-76a2bf30f137-kube-api-access-sklwp\") pod \"glance1aaa-account-delete-nct6c\" (UID: \"a9d8a25d-2f1c-483d-bf0d-76a2bf30f137\") " pod="openstack/glance1aaa-account-delete-nct6c" Oct 14 07:23:43 crc kubenswrapper[4870]: I1014 07:23:43.022233 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-h587c"] Oct 14 07:23:43 crc kubenswrapper[4870]: I1014 07:23:43.082033 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fpns6\" (UniqueName: \"kubernetes.io/projected/a6f02e4f-8dca-4b97-91e9-ee3368e5e44b-kube-api-access-fpns6\") pod \"novacell0ce16-account-delete-bwgsq\" (UID: \"a6f02e4f-8dca-4b97-91e9-ee3368e5e44b\") " pod="openstack/novacell0ce16-account-delete-bwgsq" Oct 14 07:23:43 crc kubenswrapper[4870]: I1014 07:23:43.204909 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9dd93b33-f231-4339-9db0-0507cde1dcf5" path="/var/lib/kubelet/pods/9dd93b33-f231-4339-9db0-0507cde1dcf5/volumes" Oct 14 07:23:43 crc kubenswrapper[4870]: I1014 07:23:43.206222 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-h587c"] Oct 14 07:23:43 crc kubenswrapper[4870]: I1014 07:23:43.206251 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/novaapie66c-account-delete-ftd24"] Oct 14 07:23:43 crc kubenswrapper[4870]: I1014 07:23:43.207942 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fpns6\" (UniqueName: \"kubernetes.io/projected/a6f02e4f-8dca-4b97-91e9-ee3368e5e44b-kube-api-access-fpns6\") pod \"novacell0ce16-account-delete-bwgsq\" (UID: \"a6f02e4f-8dca-4b97-91e9-ee3368e5e44b\") " pod="openstack/novacell0ce16-account-delete-bwgsq" Oct 14 07:23:43 crc kubenswrapper[4870]: E1014 07:23:43.208762 4870 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Oct 14 07:23:43 crc kubenswrapper[4870]: E1014 07:23:43.208819 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/da8a9731-f917-462f-9932-b37b6abb9a64-config-data podName:da8a9731-f917-462f-9932-b37b6abb9a64 nodeName:}" failed. No retries permitted until 2025-10-14 07:23:44.208802207 +0000 UTC m=+1359.906162578 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/da8a9731-f917-462f-9932-b37b6abb9a64-config-data") pod "rabbitmq-server-0" (UID: "da8a9731-f917-462f-9932-b37b6abb9a64") : configmap "rabbitmq-config-data" not found Oct 14 07:23:43 crc kubenswrapper[4870]: I1014 07:23:43.210751 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/novaapie66c-account-delete-ftd24"] Oct 14 07:23:43 crc kubenswrapper[4870]: I1014 07:23:43.210772 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-97xhm"] Oct 14 07:23:43 crc kubenswrapper[4870]: I1014 07:23:43.210786 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-97xhm"] Oct 14 07:23:43 crc kubenswrapper[4870]: I1014 07:23:43.210958 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/novaapie66c-account-delete-ftd24" Oct 14 07:23:43 crc kubenswrapper[4870]: I1014 07:23:43.224549 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-ovs-6482n"] Oct 14 07:23:43 crc kubenswrapper[4870]: I1014 07:23:43.240559 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-rn98c"] Oct 14 07:23:43 crc kubenswrapper[4870]: I1014 07:23:43.254302 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-metrics-9dqlp"] Oct 14 07:23:43 crc kubenswrapper[4870]: I1014 07:23:43.254582 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-controller-metrics-9dqlp" podUID="4c8d0269-7501-4c96-9f9e-30bea4b0fb92" containerName="openstack-network-exporter" containerID="cri-o://f1b865fdd1614804c6e4ae198a96114c844ab93d4e346b066d14705872321268" gracePeriod=30 Oct 14 07:23:43 crc kubenswrapper[4870]: I1014 07:23:43.261567 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fpns6\" (UniqueName: \"kubernetes.io/projected/a6f02e4f-8dca-4b97-91e9-ee3368e5e44b-kube-api-access-fpns6\") pod \"novacell0ce16-account-delete-bwgsq\" (UID: \"a6f02e4f-8dca-4b97-91e9-ee3368e5e44b\") " pod="openstack/novacell0ce16-account-delete-bwgsq" Oct 14 07:23:43 crc kubenswrapper[4870]: I1014 07:23:43.270535 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-v2d79"] Oct 14 07:23:43 crc kubenswrapper[4870]: I1014 07:23:43.281916 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance1aaa-account-delete-nct6c" Oct 14 07:23:43 crc kubenswrapper[4870]: I1014 07:23:43.311654 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-582bv\" (UniqueName: \"kubernetes.io/projected/208754ea-7cc0-4222-a186-9ba01b274add-kube-api-access-582bv\") pod \"novaapie66c-account-delete-ftd24\" (UID: \"208754ea-7cc0-4222-a186-9ba01b274add\") " pod="openstack/novaapie66c-account-delete-ftd24" Oct 14 07:23:43 crc kubenswrapper[4870]: I1014 07:23:43.323962 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/novacell0ce16-account-delete-bwgsq" Oct 14 07:23:43 crc kubenswrapper[4870]: I1014 07:23:43.325131 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-v2d79"] Oct 14 07:23:43 crc kubenswrapper[4870]: I1014 07:23:43.396789 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-lqpbp"] Oct 14 07:23:43 crc kubenswrapper[4870]: I1014 07:23:43.409974 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-ring-rebalance-lqpbp"] Oct 14 07:23:43 crc kubenswrapper[4870]: I1014 07:23:43.418846 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7cf4d694c5-wjxfs"] Oct 14 07:23:43 crc kubenswrapper[4870]: I1014 07:23:43.419079 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7cf4d694c5-wjxfs" podUID="751daf42-b5c1-4318-a15b-d5911c9d61ea" containerName="dnsmasq-dns" containerID="cri-o://e7f985e91ecb8ba5c0b4601a782047823bc6426fd3f278547d0ef69704ee2fcc" gracePeriod=10 Oct 14 07:23:43 crc kubenswrapper[4870]: I1014 07:23:43.419455 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-582bv\" (UniqueName: \"kubernetes.io/projected/208754ea-7cc0-4222-a186-9ba01b274add-kube-api-access-582bv\") pod \"novaapie66c-account-delete-ftd24\" (UID: \"208754ea-7cc0-4222-a186-9ba01b274add\") " pod="openstack/novaapie66c-account-delete-ftd24" Oct 14 07:23:43 crc kubenswrapper[4870]: I1014 07:23:43.445644 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-vtmvg"] Oct 14 07:23:43 crc kubenswrapper[4870]: I1014 07:23:43.454812 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-582bv\" (UniqueName: \"kubernetes.io/projected/208754ea-7cc0-4222-a186-9ba01b274add-kube-api-access-582bv\") pod \"novaapie66c-account-delete-ftd24\" (UID: \"208754ea-7cc0-4222-a186-9ba01b274add\") " pod="openstack/novaapie66c-account-delete-ftd24" Oct 14 07:23:43 crc kubenswrapper[4870]: I1014 07:23:43.460199 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-vtmvg"] Oct 14 07:23:43 crc kubenswrapper[4870]: I1014 07:23:43.552886 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-storage-0"] Oct 14 07:23:43 crc kubenswrapper[4870]: I1014 07:23:43.553361 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="2bd1f41f-3f1a-4ca7-8789-429104ce2120" containerName="account-server" containerID="cri-o://edba2f1cac0476dc717b946f00e8879272b5577916eb07fb08304de266852951" gracePeriod=30 Oct 14 07:23:43 crc kubenswrapper[4870]: I1014 07:23:43.553557 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="2bd1f41f-3f1a-4ca7-8789-429104ce2120" containerName="account-reaper" containerID="cri-o://177c2ea0ed77d9c56284c56d85f024d847d0abd99647a2f3eedfd9a880d3b512" gracePeriod=30 Oct 14 07:23:43 crc kubenswrapper[4870]: I1014 07:23:43.553576 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="2bd1f41f-3f1a-4ca7-8789-429104ce2120" containerName="container-updater" containerID="cri-o://42954c6bfcca40f75f1c6f4bcad9b47bf39a17a66a9a6f9bbf941bbed4d28ebd" gracePeriod=30 Oct 14 07:23:43 crc kubenswrapper[4870]: I1014 07:23:43.553618 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="2bd1f41f-3f1a-4ca7-8789-429104ce2120" containerName="object-updater" containerID="cri-o://ad9dccd6af895adbf1e3c3fcd2eca40527602d0f3bfc6d6d7ea2730f4a267119" gracePeriod=30 Oct 14 07:23:43 crc kubenswrapper[4870]: I1014 07:23:43.553624 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="2bd1f41f-3f1a-4ca7-8789-429104ce2120" containerName="account-auditor" containerID="cri-o://e3fe8e57380100f6a293cf390d246d76bd9db2c91dc07edc65f2b23cfa5d4871" gracePeriod=30 Oct 14 07:23:43 crc kubenswrapper[4870]: I1014 07:23:43.553702 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="2bd1f41f-3f1a-4ca7-8789-429104ce2120" containerName="container-auditor" containerID="cri-o://a2214a07cbc03e9d4c3657396572e211461dec75c50cb838dd53f64e5d4ae7de" gracePeriod=30 Oct 14 07:23:43 crc kubenswrapper[4870]: I1014 07:23:43.553743 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="2bd1f41f-3f1a-4ca7-8789-429104ce2120" containerName="swift-recon-cron" containerID="cri-o://9104317bae5cf155b62976217e7c03cd64c95837b39e557bc2162f991c3e3920" gracePeriod=30 Oct 14 07:23:43 crc kubenswrapper[4870]: I1014 07:23:43.553748 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="2bd1f41f-3f1a-4ca7-8789-429104ce2120" containerName="object-replicator" containerID="cri-o://b51d4fe9abeae800a8b66afb49b1fafee55d1cdb96718d65b7d5b35f6fd7519b" gracePeriod=30 Oct 14 07:23:43 crc kubenswrapper[4870]: I1014 07:23:43.553758 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="2bd1f41f-3f1a-4ca7-8789-429104ce2120" containerName="container-server" containerID="cri-o://642a18f2173ab4124b7442763988bbe3a2c0fdd80e87caad5e48f172fd55efc7" gracePeriod=30 Oct 14 07:23:43 crc kubenswrapper[4870]: I1014 07:23:43.553748 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="2bd1f41f-3f1a-4ca7-8789-429104ce2120" containerName="container-replicator" containerID="cri-o://418f1f077bbfc2949a0abda0f2ddd66da4131527026acec0d19a4c08dec4dc66" gracePeriod=30 Oct 14 07:23:43 crc kubenswrapper[4870]: I1014 07:23:43.553806 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="2bd1f41f-3f1a-4ca7-8789-429104ce2120" containerName="rsync" containerID="cri-o://eaac2b7e339adcaf70b58432d22ea22656cd6aebcd6871bb19adeae4be5f56f0" gracePeriod=30 Oct 14 07:23:43 crc kubenswrapper[4870]: I1014 07:23:43.553821 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="2bd1f41f-3f1a-4ca7-8789-429104ce2120" containerName="object-expirer" containerID="cri-o://7c8fe31733cd77d0b1ee88ab091697a9851a822a29c50d287246ced8eca4d502" gracePeriod=30 Oct 14 07:23:43 crc kubenswrapper[4870]: I1014 07:23:43.553830 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="2bd1f41f-3f1a-4ca7-8789-429104ce2120" containerName="object-auditor" containerID="cri-o://18b08dbc7de7f34cc2bbb65969ca2e9d3e2da4bddab756fdee2d7db4e53e9180" gracePeriod=30 Oct 14 07:23:43 crc kubenswrapper[4870]: I1014 07:23:43.553807 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="2bd1f41f-3f1a-4ca7-8789-429104ce2120" containerName="object-server" containerID="cri-o://5330b69d8d67676f3cc5cd3efe7cfe42a11bfa0006bfe019a9c536b30e4c1f0d" gracePeriod=30 Oct 14 07:23:43 crc kubenswrapper[4870]: I1014 07:23:43.553907 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="2bd1f41f-3f1a-4ca7-8789-429104ce2120" containerName="account-replicator" containerID="cri-o://427d65d1b5cbeeeb2e36fcfaa93892d1227204dbfad3f98d9cee60708c1f3a0d" gracePeriod=30 Oct 14 07:23:43 crc kubenswrapper[4870]: I1014 07:23:43.571205 4870 generic.go:334] "Generic (PLEG): container finished" podID="416d7b39-4541-44a8-a55f-924bc86fee32" containerID="11e57f381235088d13225057331ed61d6e2ab6f98595ff6fe4992063044159e4" exitCode=2 Oct 14 07:23:43 crc kubenswrapper[4870]: I1014 07:23:43.571298 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"416d7b39-4541-44a8-a55f-924bc86fee32","Type":"ContainerDied","Data":"11e57f381235088d13225057331ed61d6e2ab6f98595ff6fe4992063044159e4"} Oct 14 07:23:43 crc kubenswrapper[4870]: I1014 07:23:43.577503 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/novaapie66c-account-delete-ftd24" Oct 14 07:23:43 crc kubenswrapper[4870]: I1014 07:23:43.659760 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_09f89569-6fe3-4b3f-9394-a990ad8cde30/ovsdbserver-nb/0.log" Oct 14 07:23:43 crc kubenswrapper[4870]: I1014 07:23:43.659804 4870 generic.go:334] "Generic (PLEG): container finished" podID="09f89569-6fe3-4b3f-9394-a990ad8cde30" containerID="b03ec54549f3acea91303d6cc01b2627a9ebd5ec3e78dcb4527b100efe1ebd9c" exitCode=2 Oct 14 07:23:43 crc kubenswrapper[4870]: I1014 07:23:43.659820 4870 generic.go:334] "Generic (PLEG): container finished" podID="09f89569-6fe3-4b3f-9394-a990ad8cde30" containerID="76d17f5aef5402fa0f6135921f3ba0aeb9f5640152e827b4385348efdbe620d3" exitCode=143 Oct 14 07:23:43 crc kubenswrapper[4870]: I1014 07:23:43.659845 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"09f89569-6fe3-4b3f-9394-a990ad8cde30","Type":"ContainerDied","Data":"b03ec54549f3acea91303d6cc01b2627a9ebd5ec3e78dcb4527b100efe1ebd9c"} Oct 14 07:23:43 crc kubenswrapper[4870]: I1014 07:23:43.659984 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"09f89569-6fe3-4b3f-9394-a990ad8cde30","Type":"ContainerDied","Data":"76d17f5aef5402fa0f6135921f3ba0aeb9f5640152e827b4385348efdbe620d3"} Oct 14 07:23:43 crc kubenswrapper[4870]: I1014 07:23:43.668724 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-597ff968b-msxbn"] Oct 14 07:23:43 crc kubenswrapper[4870]: I1014 07:23:43.669015 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/placement-597ff968b-msxbn" podUID="2130e3f2-cbea-46e1-b5d3-daefdd972935" containerName="placement-log" containerID="cri-o://f4ef48299206032c6fdb947f48913c54fe5fa0a39a9b5ffe9e3ea9fd8ac49f8e" gracePeriod=30 Oct 14 07:23:43 crc kubenswrapper[4870]: I1014 07:23:43.669457 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/placement-597ff968b-msxbn" podUID="2130e3f2-cbea-46e1-b5d3-daefdd972935" containerName="placement-api" containerID="cri-o://4339244c7626324d778551bfb8f861a2c704b3e5aa7199c34ba643e475017090" gracePeriod=30 Oct 14 07:23:43 crc kubenswrapper[4870]: I1014 07:23:43.812527 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-xtkkt"] Oct 14 07:23:43 crc kubenswrapper[4870]: I1014 07:23:43.868256 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-q76l4"] Oct 14 07:23:43 crc kubenswrapper[4870]: I1014 07:23:43.875250 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_09f89569-6fe3-4b3f-9394-a990ad8cde30/ovsdbserver-nb/0.log" Oct 14 07:23:43 crc kubenswrapper[4870]: I1014 07:23:43.875338 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Oct 14 07:23:43 crc kubenswrapper[4870]: I1014 07:23:43.895565 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-xtkkt"] Oct 14 07:23:43 crc kubenswrapper[4870]: I1014 07:23:43.955585 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-q76l4"] Oct 14 07:23:43 crc kubenswrapper[4870]: I1014 07:23:43.980037 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/09f89569-6fe3-4b3f-9394-a990ad8cde30-metrics-certs-tls-certs\") pod \"09f89569-6fe3-4b3f-9394-a990ad8cde30\" (UID: \"09f89569-6fe3-4b3f-9394-a990ad8cde30\") " Oct 14 07:23:43 crc kubenswrapper[4870]: I1014 07:23:43.980137 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09f89569-6fe3-4b3f-9394-a990ad8cde30-config\") pod \"09f89569-6fe3-4b3f-9394-a990ad8cde30\" (UID: \"09f89569-6fe3-4b3f-9394-a990ad8cde30\") " Oct 14 07:23:43 crc kubenswrapper[4870]: I1014 07:23:43.980176 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8hks2\" (UniqueName: \"kubernetes.io/projected/09f89569-6fe3-4b3f-9394-a990ad8cde30-kube-api-access-8hks2\") pod \"09f89569-6fe3-4b3f-9394-a990ad8cde30\" (UID: \"09f89569-6fe3-4b3f-9394-a990ad8cde30\") " Oct 14 07:23:43 crc kubenswrapper[4870]: I1014 07:23:43.980212 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/09f89569-6fe3-4b3f-9394-a990ad8cde30-ovsdbserver-nb-tls-certs\") pod \"09f89569-6fe3-4b3f-9394-a990ad8cde30\" (UID: \"09f89569-6fe3-4b3f-9394-a990ad8cde30\") " Oct 14 07:23:43 crc kubenswrapper[4870]: I1014 07:23:43.980230 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/09f89569-6fe3-4b3f-9394-a990ad8cde30-combined-ca-bundle\") pod \"09f89569-6fe3-4b3f-9394-a990ad8cde30\" (UID: \"09f89569-6fe3-4b3f-9394-a990ad8cde30\") " Oct 14 07:23:43 crc kubenswrapper[4870]: I1014 07:23:43.980278 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/09f89569-6fe3-4b3f-9394-a990ad8cde30-ovsdb-rundir\") pod \"09f89569-6fe3-4b3f-9394-a990ad8cde30\" (UID: \"09f89569-6fe3-4b3f-9394-a990ad8cde30\") " Oct 14 07:23:43 crc kubenswrapper[4870]: I1014 07:23:43.980346 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/09f89569-6fe3-4b3f-9394-a990ad8cde30-scripts\") pod \"09f89569-6fe3-4b3f-9394-a990ad8cde30\" (UID: \"09f89569-6fe3-4b3f-9394-a990ad8cde30\") " Oct 14 07:23:43 crc kubenswrapper[4870]: I1014 07:23:43.980359 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndbcluster-nb-etc-ovn\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"09f89569-6fe3-4b3f-9394-a990ad8cde30\" (UID: \"09f89569-6fe3-4b3f-9394-a990ad8cde30\") " Oct 14 07:23:43 crc kubenswrapper[4870]: E1014 07:23:43.980801 4870 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Oct 14 07:23:43 crc kubenswrapper[4870]: E1014 07:23:43.980845 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5212ba20-83c3-41ff-a3d9-5a315f009c1b-config-data podName:5212ba20-83c3-41ff-a3d9-5a315f009c1b nodeName:}" failed. No retries permitted until 2025-10-14 07:23:45.980832367 +0000 UTC m=+1361.678192738 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/5212ba20-83c3-41ff-a3d9-5a315f009c1b-config-data") pod "rabbitmq-cell1-server-0" (UID: "5212ba20-83c3-41ff-a3d9-5a315f009c1b") : configmap "rabbitmq-cell1-config-data" not found Oct 14 07:23:43 crc kubenswrapper[4870]: I1014 07:23:43.984188 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09f89569-6fe3-4b3f-9394-a990ad8cde30-scripts" (OuterVolumeSpecName: "scripts") pod "09f89569-6fe3-4b3f-9394-a990ad8cde30" (UID: "09f89569-6fe3-4b3f-9394-a990ad8cde30"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:23:43 crc kubenswrapper[4870]: I1014 07:23:43.984268 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09f89569-6fe3-4b3f-9394-a990ad8cde30-config" (OuterVolumeSpecName: "config") pod "09f89569-6fe3-4b3f-9394-a990ad8cde30" (UID: "09f89569-6fe3-4b3f-9394-a990ad8cde30"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:23:43 crc kubenswrapper[4870]: I1014 07:23:43.985904 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/09f89569-6fe3-4b3f-9394-a990ad8cde30-ovsdb-rundir" (OuterVolumeSpecName: "ovsdb-rundir") pod "09f89569-6fe3-4b3f-9394-a990ad8cde30" (UID: "09f89569-6fe3-4b3f-9394-a990ad8cde30"). InnerVolumeSpecName "ovsdb-rundir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:23:43 crc kubenswrapper[4870]: I1014 07:23:43.989517 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 14 07:23:43 crc kubenswrapper[4870]: I1014 07:23:43.989785 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="7bc76b72-8679-4b33-8966-0ee70c1788e5" containerName="cinder-scheduler" containerID="cri-o://e5c43911d04a168b064b8b753711f3e5ed521dc55a16d5687c49934c8d0eb3a6" gracePeriod=30 Oct 14 07:23:43 crc kubenswrapper[4870]: I1014 07:23:43.990301 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="7bc76b72-8679-4b33-8966-0ee70c1788e5" containerName="probe" containerID="cri-o://a1f9a0d181ab7f3f80bee6c1008e3b393dd082ba31b2614d29e2a4b0d5534a46" gracePeriod=30 Oct 14 07:23:44 crc kubenswrapper[4870]: I1014 07:23:44.035197 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage08-crc" (OuterVolumeSpecName: "ovndbcluster-nb-etc-ovn") pod "09f89569-6fe3-4b3f-9394-a990ad8cde30" (UID: "09f89569-6fe3-4b3f-9394-a990ad8cde30"). InnerVolumeSpecName "local-storage08-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 14 07:23:44 crc kubenswrapper[4870]: I1014 07:23:44.045774 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09f89569-6fe3-4b3f-9394-a990ad8cde30-kube-api-access-8hks2" (OuterVolumeSpecName: "kube-api-access-8hks2") pod "09f89569-6fe3-4b3f-9394-a990ad8cde30" (UID: "09f89569-6fe3-4b3f-9394-a990ad8cde30"). InnerVolumeSpecName "kube-api-access-8hks2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:23:44 crc kubenswrapper[4870]: I1014 07:23:44.082122 4870 reconciler_common.go:293] "Volume detached for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/09f89569-6fe3-4b3f-9394-a990ad8cde30-ovsdb-rundir\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:44 crc kubenswrapper[4870]: I1014 07:23:44.082154 4870 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/09f89569-6fe3-4b3f-9394-a990ad8cde30-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:44 crc kubenswrapper[4870]: I1014 07:23:44.082177 4870 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" " Oct 14 07:23:44 crc kubenswrapper[4870]: I1014 07:23:44.082188 4870 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09f89569-6fe3-4b3f-9394-a990ad8cde30-config\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:44 crc kubenswrapper[4870]: I1014 07:23:44.082197 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8hks2\" (UniqueName: \"kubernetes.io/projected/09f89569-6fe3-4b3f-9394-a990ad8cde30-kube-api-access-8hks2\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:44 crc kubenswrapper[4870]: I1014 07:23:44.087661 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 14 07:23:44 crc kubenswrapper[4870]: I1014 07:23:44.088324 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovsdbserver-sb-0" podUID="36717039-e3e2-4b74-b612-655556122574" containerName="openstack-network-exporter" containerID="cri-o://660968003babd11d1e6d54c59b0e91ea2368857634bd407e25468d0645470948" gracePeriod=300 Oct 14 07:23:44 crc kubenswrapper[4870]: I1014 07:23:44.109274 4870 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage08-crc" (UniqueName: "kubernetes.io/local-volume/local-storage08-crc") on node "crc" Oct 14 07:23:44 crc kubenswrapper[4870]: E1014 07:23:44.173463 4870 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="457905ca1130c3804c068cf6fc44521ad064a3fb34449d666cf858f14e2feace" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Oct 14 07:23:44 crc kubenswrapper[4870]: I1014 07:23:44.194109 4870 reconciler_common.go:293] "Volume detached for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:44 crc kubenswrapper[4870]: I1014 07:23:44.206840 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09f89569-6fe3-4b3f-9394-a990ad8cde30-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "09f89569-6fe3-4b3f-9394-a990ad8cde30" (UID: "09f89569-6fe3-4b3f-9394-a990ad8cde30"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:23:44 crc kubenswrapper[4870]: E1014 07:23:44.246772 4870 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="457905ca1130c3804c068cf6fc44521ad064a3fb34449d666cf858f14e2feace" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Oct 14 07:23:44 crc kubenswrapper[4870]: I1014 07:23:44.278192 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 14 07:23:44 crc kubenswrapper[4870]: E1014 07:23:44.284595 4870 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="457905ca1130c3804c068cf6fc44521ad064a3fb34449d666cf858f14e2feace" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Oct 14 07:23:44 crc kubenswrapper[4870]: E1014 07:23:44.284667 4870 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-northd-0" podUID="416d7b39-4541-44a8-a55f-924bc86fee32" containerName="ovn-northd" Oct 14 07:23:44 crc kubenswrapper[4870]: I1014 07:23:44.301730 4870 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/09f89569-6fe3-4b3f-9394-a990ad8cde30-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:44 crc kubenswrapper[4870]: E1014 07:23:44.302690 4870 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Oct 14 07:23:44 crc kubenswrapper[4870]: E1014 07:23:44.302731 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/da8a9731-f917-462f-9932-b37b6abb9a64-config-data podName:da8a9731-f917-462f-9932-b37b6abb9a64 nodeName:}" failed. No retries permitted until 2025-10-14 07:23:46.302717139 +0000 UTC m=+1362.000077510 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/da8a9731-f917-462f-9932-b37b6abb9a64-config-data") pod "rabbitmq-server-0" (UID: "da8a9731-f917-462f-9932-b37b6abb9a64") : configmap "rabbitmq-config-data" not found Oct 14 07:23:44 crc kubenswrapper[4870]: I1014 07:23:44.343548 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Oct 14 07:23:44 crc kubenswrapper[4870]: I1014 07:23:44.343845 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="51de8090-c48e-498c-99ab-794a7527252c" containerName="cinder-api-log" containerID="cri-o://deba864879defc602edfa6ada18f77ffb9b44bf36a5292d8a32f7922bcd29311" gracePeriod=30 Oct 14 07:23:44 crc kubenswrapper[4870]: I1014 07:23:44.344295 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="51de8090-c48e-498c-99ab-794a7527252c" containerName="cinder-api" containerID="cri-o://571dfedb34f9303684e8f8fe09485d339b6c363eaa75f6950ae538f2526d0129" gracePeriod=30 Oct 14 07:23:44 crc kubenswrapper[4870]: E1014 07:23:44.344894 4870 handlers.go:78] "Exec lifecycle hook for Container in Pod failed" err="command '/usr/share/ovn/scripts/ovn-ctl stop_controller' exited with 137: " execCommand=["/usr/share/ovn/scripts/ovn-ctl","stop_controller"] containerName="ovn-controller" pod="openstack/ovn-controller-rn98c" message=< Oct 14 07:23:44 crc kubenswrapper[4870]: Exiting ovn-controller (1) [ OK ] Oct 14 07:23:44 crc kubenswrapper[4870]: > Oct 14 07:23:44 crc kubenswrapper[4870]: E1014 07:23:44.344917 4870 kuberuntime_container.go:691] "PreStop hook failed" err="command '/usr/share/ovn/scripts/ovn-ctl stop_controller' exited with 137: " pod="openstack/ovn-controller-rn98c" podUID="503a07f8-2c05-458a-80ff-1abfe973dbd5" containerName="ovn-controller" containerID="cri-o://124ba16adf714b9b67c2030611bc8afc10924c09b62003e8aa1a0045f6015014" Oct 14 07:23:44 crc kubenswrapper[4870]: I1014 07:23:44.344941 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-controller-rn98c" podUID="503a07f8-2c05-458a-80ff-1abfe973dbd5" containerName="ovn-controller" containerID="cri-o://124ba16adf714b9b67c2030611bc8afc10924c09b62003e8aa1a0045f6015014" gracePeriod=29 Oct 14 07:23:44 crc kubenswrapper[4870]: I1014 07:23:44.386516 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 14 07:23:44 crc kubenswrapper[4870]: I1014 07:23:44.387078 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="56eb286f-8aa2-46aa-a74d-031288a9aad1" containerName="glance-log" containerID="cri-o://9fa2f1146313b498f55c9bb3361bca48d735f25641dcd24bdbeac19bc4ae6377" gracePeriod=30 Oct 14 07:23:44 crc kubenswrapper[4870]: I1014 07:23:44.387212 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="56eb286f-8aa2-46aa-a74d-031288a9aad1" containerName="glance-httpd" containerID="cri-o://d8fce18d04004b32792441ce25ddcd9d74c06df092b20d09c14fca60675539d4" gracePeriod=30 Oct 14 07:23:44 crc kubenswrapper[4870]: I1014 07:23:44.413522 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 14 07:23:44 crc kubenswrapper[4870]: I1014 07:23:44.413787 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="c712b87a-cf4d-437e-abe0-1a566046ca66" containerName="glance-log" containerID="cri-o://6c0de18ab40ba16d1bc75ff9b157bb26252b1a8044809c3ee46ea0a31ed39019" gracePeriod=30 Oct 14 07:23:44 crc kubenswrapper[4870]: I1014 07:23:44.414250 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="c712b87a-cf4d-437e-abe0-1a566046ca66" containerName="glance-httpd" containerID="cri-o://f744eb9b855e3191279bf5fc01139d0a3359c09a6178c62d0602563d9d13ff25" gracePeriod=30 Oct 14 07:23:44 crc kubenswrapper[4870]: I1014 07:23:44.424215 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-nzvfg"] Oct 14 07:23:44 crc kubenswrapper[4870]: I1014 07:23:44.435568 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement1a80-account-delete-8q7fq"] Oct 14 07:23:44 crc kubenswrapper[4870]: I1014 07:23:44.445677 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-546b769ccc-fdhsk"] Oct 14 07:23:44 crc kubenswrapper[4870]: I1014 07:23:44.445884 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-546b769ccc-fdhsk" podUID="0ba0e2ed-57d0-4535-883e-10b9b985b3d3" containerName="neutron-api" containerID="cri-o://b6b74fcdf90322787a4403ce2a20ab64b512335fa185f5d3caa8a7cd3a4837f4" gracePeriod=30 Oct 14 07:23:44 crc kubenswrapper[4870]: I1014 07:23:44.446266 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-546b769ccc-fdhsk" podUID="0ba0e2ed-57d0-4535-883e-10b9b985b3d3" containerName="neutron-httpd" containerID="cri-o://f84e5ed457b1301ae67b84003d43c8e27a59446623074e29e6e2859280f620fd" gracePeriod=30 Oct 14 07:23:44 crc kubenswrapper[4870]: I1014 07:23:44.453215 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09f89569-6fe3-4b3f-9394-a990ad8cde30-ovsdbserver-nb-tls-certs" (OuterVolumeSpecName: "ovsdbserver-nb-tls-certs") pod "09f89569-6fe3-4b3f-9394-a990ad8cde30" (UID: "09f89569-6fe3-4b3f-9394-a990ad8cde30"). InnerVolumeSpecName "ovsdbserver-nb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:23:44 crc kubenswrapper[4870]: I1014 07:23:44.522246 4870 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/09f89569-6fe3-4b3f-9394-a990ad8cde30-ovsdbserver-nb-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:44 crc kubenswrapper[4870]: I1014 07:23:44.525730 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-nzvfg"] Oct 14 07:23:44 crc kubenswrapper[4870]: I1014 07:23:44.586714 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09f89569-6fe3-4b3f-9394-a990ad8cde30-metrics-certs-tls-certs" (OuterVolumeSpecName: "metrics-certs-tls-certs") pod "09f89569-6fe3-4b3f-9394-a990ad8cde30" (UID: "09f89569-6fe3-4b3f-9394-a990ad8cde30"). InnerVolumeSpecName "metrics-certs-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:23:44 crc kubenswrapper[4870]: I1014 07:23:44.587794 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-vvgl7"] Oct 14 07:23:44 crc kubenswrapper[4870]: I1014 07:23:44.635882 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-1a80-account-create-rcm8s"] Oct 14 07:23:44 crc kubenswrapper[4870]: I1014 07:23:44.638232 4870 reconciler_common.go:293] "Volume detached for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/09f89569-6fe3-4b3f-9394-a990ad8cde30-metrics-certs-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:44 crc kubenswrapper[4870]: I1014 07:23:44.660913 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-1a80-account-create-rcm8s"] Oct 14 07:23:44 crc kubenswrapper[4870]: I1014 07:23:44.683918 4870 generic.go:334] "Generic (PLEG): container finished" podID="51de8090-c48e-498c-99ab-794a7527252c" containerID="deba864879defc602edfa6ada18f77ffb9b44bf36a5292d8a32f7922bcd29311" exitCode=143 Oct 14 07:23:44 crc kubenswrapper[4870]: I1014 07:23:44.683988 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"51de8090-c48e-498c-99ab-794a7527252c","Type":"ContainerDied","Data":"deba864879defc602edfa6ada18f77ffb9b44bf36a5292d8a32f7922bcd29311"} Oct 14 07:23:44 crc kubenswrapper[4870]: I1014 07:23:44.689494 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-9dqlp_4c8d0269-7501-4c96-9f9e-30bea4b0fb92/openstack-network-exporter/0.log" Oct 14 07:23:44 crc kubenswrapper[4870]: I1014 07:23:44.689531 4870 generic.go:334] "Generic (PLEG): container finished" podID="4c8d0269-7501-4c96-9f9e-30bea4b0fb92" containerID="f1b865fdd1614804c6e4ae198a96114c844ab93d4e346b066d14705872321268" exitCode=2 Oct 14 07:23:44 crc kubenswrapper[4870]: I1014 07:23:44.689643 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-9dqlp" event={"ID":"4c8d0269-7501-4c96-9f9e-30bea4b0fb92","Type":"ContainerDied","Data":"f1b865fdd1614804c6e4ae198a96114c844ab93d4e346b066d14705872321268"} Oct 14 07:23:44 crc kubenswrapper[4870]: I1014 07:23:44.689664 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-9dqlp" event={"ID":"4c8d0269-7501-4c96-9f9e-30bea4b0fb92","Type":"ContainerDied","Data":"b0345ab44f21f08a1613972c5c05f4e2a7385879b829d80e6f6d9c546fe68b40"} Oct 14 07:23:44 crc kubenswrapper[4870]: I1014 07:23:44.689676 4870 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b0345ab44f21f08a1613972c5c05f4e2a7385879b829d80e6f6d9c546fe68b40" Oct 14 07:23:44 crc kubenswrapper[4870]: I1014 07:23:44.693962 4870 generic.go:334] "Generic (PLEG): container finished" podID="36717039-e3e2-4b74-b612-655556122574" containerID="660968003babd11d1e6d54c59b0e91ea2368857634bd407e25468d0645470948" exitCode=2 Oct 14 07:23:44 crc kubenswrapper[4870]: I1014 07:23:44.694069 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"36717039-e3e2-4b74-b612-655556122574","Type":"ContainerDied","Data":"660968003babd11d1e6d54c59b0e91ea2368857634bd407e25468d0645470948"} Oct 14 07:23:44 crc kubenswrapper[4870]: I1014 07:23:44.703709 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-vvgl7"] Oct 14 07:23:44 crc kubenswrapper[4870]: I1014 07:23:44.709607 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-5c98-account-create-9bjhh"] Oct 14 07:23:44 crc kubenswrapper[4870]: I1014 07:23:44.722456 4870 generic.go:334] "Generic (PLEG): container finished" podID="2bd1f41f-3f1a-4ca7-8789-429104ce2120" containerID="eaac2b7e339adcaf70b58432d22ea22656cd6aebcd6871bb19adeae4be5f56f0" exitCode=0 Oct 14 07:23:44 crc kubenswrapper[4870]: I1014 07:23:44.722481 4870 generic.go:334] "Generic (PLEG): container finished" podID="2bd1f41f-3f1a-4ca7-8789-429104ce2120" containerID="7c8fe31733cd77d0b1ee88ab091697a9851a822a29c50d287246ced8eca4d502" exitCode=0 Oct 14 07:23:44 crc kubenswrapper[4870]: I1014 07:23:44.722491 4870 generic.go:334] "Generic (PLEG): container finished" podID="2bd1f41f-3f1a-4ca7-8789-429104ce2120" containerID="ad9dccd6af895adbf1e3c3fcd2eca40527602d0f3bfc6d6d7ea2730f4a267119" exitCode=0 Oct 14 07:23:44 crc kubenswrapper[4870]: I1014 07:23:44.722498 4870 generic.go:334] "Generic (PLEG): container finished" podID="2bd1f41f-3f1a-4ca7-8789-429104ce2120" containerID="18b08dbc7de7f34cc2bbb65969ca2e9d3e2da4bddab756fdee2d7db4e53e9180" exitCode=0 Oct 14 07:23:44 crc kubenswrapper[4870]: I1014 07:23:44.722504 4870 generic.go:334] "Generic (PLEG): container finished" podID="2bd1f41f-3f1a-4ca7-8789-429104ce2120" containerID="b51d4fe9abeae800a8b66afb49b1fafee55d1cdb96718d65b7d5b35f6fd7519b" exitCode=0 Oct 14 07:23:44 crc kubenswrapper[4870]: I1014 07:23:44.722511 4870 generic.go:334] "Generic (PLEG): container finished" podID="2bd1f41f-3f1a-4ca7-8789-429104ce2120" containerID="5330b69d8d67676f3cc5cd3efe7cfe42a11bfa0006bfe019a9c536b30e4c1f0d" exitCode=0 Oct 14 07:23:44 crc kubenswrapper[4870]: I1014 07:23:44.722518 4870 generic.go:334] "Generic (PLEG): container finished" podID="2bd1f41f-3f1a-4ca7-8789-429104ce2120" containerID="42954c6bfcca40f75f1c6f4bcad9b47bf39a17a66a9a6f9bbf941bbed4d28ebd" exitCode=0 Oct 14 07:23:44 crc kubenswrapper[4870]: I1014 07:23:44.722524 4870 generic.go:334] "Generic (PLEG): container finished" podID="2bd1f41f-3f1a-4ca7-8789-429104ce2120" containerID="a2214a07cbc03e9d4c3657396572e211461dec75c50cb838dd53f64e5d4ae7de" exitCode=0 Oct 14 07:23:44 crc kubenswrapper[4870]: I1014 07:23:44.722531 4870 generic.go:334] "Generic (PLEG): container finished" podID="2bd1f41f-3f1a-4ca7-8789-429104ce2120" containerID="418f1f077bbfc2949a0abda0f2ddd66da4131527026acec0d19a4c08dec4dc66" exitCode=0 Oct 14 07:23:44 crc kubenswrapper[4870]: I1014 07:23:44.722538 4870 generic.go:334] "Generic (PLEG): container finished" podID="2bd1f41f-3f1a-4ca7-8789-429104ce2120" containerID="642a18f2173ab4124b7442763988bbe3a2c0fdd80e87caad5e48f172fd55efc7" exitCode=0 Oct 14 07:23:44 crc kubenswrapper[4870]: I1014 07:23:44.722545 4870 generic.go:334] "Generic (PLEG): container finished" podID="2bd1f41f-3f1a-4ca7-8789-429104ce2120" containerID="177c2ea0ed77d9c56284c56d85f024d847d0abd99647a2f3eedfd9a880d3b512" exitCode=0 Oct 14 07:23:44 crc kubenswrapper[4870]: I1014 07:23:44.722551 4870 generic.go:334] "Generic (PLEG): container finished" podID="2bd1f41f-3f1a-4ca7-8789-429104ce2120" containerID="e3fe8e57380100f6a293cf390d246d76bd9db2c91dc07edc65f2b23cfa5d4871" exitCode=0 Oct 14 07:23:44 crc kubenswrapper[4870]: I1014 07:23:44.722558 4870 generic.go:334] "Generic (PLEG): container finished" podID="2bd1f41f-3f1a-4ca7-8789-429104ce2120" containerID="427d65d1b5cbeeeb2e36fcfaa93892d1227204dbfad3f98d9cee60708c1f3a0d" exitCode=0 Oct 14 07:23:44 crc kubenswrapper[4870]: I1014 07:23:44.722511 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"2bd1f41f-3f1a-4ca7-8789-429104ce2120","Type":"ContainerDied","Data":"eaac2b7e339adcaf70b58432d22ea22656cd6aebcd6871bb19adeae4be5f56f0"} Oct 14 07:23:44 crc kubenswrapper[4870]: I1014 07:23:44.722630 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"2bd1f41f-3f1a-4ca7-8789-429104ce2120","Type":"ContainerDied","Data":"7c8fe31733cd77d0b1ee88ab091697a9851a822a29c50d287246ced8eca4d502"} Oct 14 07:23:44 crc kubenswrapper[4870]: I1014 07:23:44.722645 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"2bd1f41f-3f1a-4ca7-8789-429104ce2120","Type":"ContainerDied","Data":"ad9dccd6af895adbf1e3c3fcd2eca40527602d0f3bfc6d6d7ea2730f4a267119"} Oct 14 07:23:44 crc kubenswrapper[4870]: I1014 07:23:44.722655 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"2bd1f41f-3f1a-4ca7-8789-429104ce2120","Type":"ContainerDied","Data":"18b08dbc7de7f34cc2bbb65969ca2e9d3e2da4bddab756fdee2d7db4e53e9180"} Oct 14 07:23:44 crc kubenswrapper[4870]: I1014 07:23:44.722666 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"2bd1f41f-3f1a-4ca7-8789-429104ce2120","Type":"ContainerDied","Data":"b51d4fe9abeae800a8b66afb49b1fafee55d1cdb96718d65b7d5b35f6fd7519b"} Oct 14 07:23:44 crc kubenswrapper[4870]: I1014 07:23:44.722676 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"2bd1f41f-3f1a-4ca7-8789-429104ce2120","Type":"ContainerDied","Data":"5330b69d8d67676f3cc5cd3efe7cfe42a11bfa0006bfe019a9c536b30e4c1f0d"} Oct 14 07:23:44 crc kubenswrapper[4870]: I1014 07:23:44.722686 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"2bd1f41f-3f1a-4ca7-8789-429104ce2120","Type":"ContainerDied","Data":"42954c6bfcca40f75f1c6f4bcad9b47bf39a17a66a9a6f9bbf941bbed4d28ebd"} Oct 14 07:23:44 crc kubenswrapper[4870]: I1014 07:23:44.722695 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"2bd1f41f-3f1a-4ca7-8789-429104ce2120","Type":"ContainerDied","Data":"a2214a07cbc03e9d4c3657396572e211461dec75c50cb838dd53f64e5d4ae7de"} Oct 14 07:23:44 crc kubenswrapper[4870]: I1014 07:23:44.722704 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"2bd1f41f-3f1a-4ca7-8789-429104ce2120","Type":"ContainerDied","Data":"418f1f077bbfc2949a0abda0f2ddd66da4131527026acec0d19a4c08dec4dc66"} Oct 14 07:23:44 crc kubenswrapper[4870]: I1014 07:23:44.722713 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"2bd1f41f-3f1a-4ca7-8789-429104ce2120","Type":"ContainerDied","Data":"642a18f2173ab4124b7442763988bbe3a2c0fdd80e87caad5e48f172fd55efc7"} Oct 14 07:23:44 crc kubenswrapper[4870]: I1014 07:23:44.722723 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"2bd1f41f-3f1a-4ca7-8789-429104ce2120","Type":"ContainerDied","Data":"177c2ea0ed77d9c56284c56d85f024d847d0abd99647a2f3eedfd9a880d3b512"} Oct 14 07:23:44 crc kubenswrapper[4870]: I1014 07:23:44.722734 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"2bd1f41f-3f1a-4ca7-8789-429104ce2120","Type":"ContainerDied","Data":"e3fe8e57380100f6a293cf390d246d76bd9db2c91dc07edc65f2b23cfa5d4871"} Oct 14 07:23:44 crc kubenswrapper[4870]: I1014 07:23:44.722744 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"2bd1f41f-3f1a-4ca7-8789-429104ce2120","Type":"ContainerDied","Data":"427d65d1b5cbeeeb2e36fcfaa93892d1227204dbfad3f98d9cee60708c1f3a0d"} Oct 14 07:23:44 crc kubenswrapper[4870]: I1014 07:23:44.724593 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-5c98-account-create-9bjhh"] Oct 14 07:23:44 crc kubenswrapper[4870]: I1014 07:23:44.737380 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_09f89569-6fe3-4b3f-9394-a990ad8cde30/ovsdbserver-nb/0.log" Oct 14 07:23:44 crc kubenswrapper[4870]: I1014 07:23:44.737757 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"09f89569-6fe3-4b3f-9394-a990ad8cde30","Type":"ContainerDied","Data":"bc1f5cf5d7dba356af68fb612233bc6c1440a741d5141cff777e43bb708f2fb5"} Oct 14 07:23:44 crc kubenswrapper[4870]: I1014 07:23:44.737810 4870 scope.go:117] "RemoveContainer" containerID="b03ec54549f3acea91303d6cc01b2627a9ebd5ec3e78dcb4527b100efe1ebd9c" Oct 14 07:23:44 crc kubenswrapper[4870]: I1014 07:23:44.738010 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Oct 14 07:23:44 crc kubenswrapper[4870]: I1014 07:23:44.743941 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder5c98-account-delete-lhjjg"] Oct 14 07:23:44 crc kubenswrapper[4870]: I1014 07:23:44.764084 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-m2g9b"] Oct 14 07:23:44 crc kubenswrapper[4870]: I1014 07:23:44.764172 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-m2g9b"] Oct 14 07:23:44 crc kubenswrapper[4870]: I1014 07:23:44.774914 4870 generic.go:334] "Generic (PLEG): container finished" podID="0965b7e6-2aa0-4940-a130-324cfb08de5a" containerID="044ea035cd6094777207ef8f180d32297801ac2dd26be47e987ae4b7318523d6" exitCode=137 Oct 14 07:23:44 crc kubenswrapper[4870]: I1014 07:23:44.777575 4870 generic.go:334] "Generic (PLEG): container finished" podID="2130e3f2-cbea-46e1-b5d3-daefdd972935" containerID="f4ef48299206032c6fdb947f48913c54fe5fa0a39a9b5ffe9e3ea9fd8ac49f8e" exitCode=143 Oct 14 07:23:44 crc kubenswrapper[4870]: I1014 07:23:44.777629 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-597ff968b-msxbn" event={"ID":"2130e3f2-cbea-46e1-b5d3-daefdd972935","Type":"ContainerDied","Data":"f4ef48299206032c6fdb947f48913c54fe5fa0a39a9b5ffe9e3ea9fd8ac49f8e"} Oct 14 07:23:44 crc kubenswrapper[4870]: I1014 07:23:44.778857 4870 generic.go:334] "Generic (PLEG): container finished" podID="751daf42-b5c1-4318-a15b-d5911c9d61ea" containerID="e7f985e91ecb8ba5c0b4601a782047823bc6426fd3f278547d0ef69704ee2fcc" exitCode=0 Oct 14 07:23:44 crc kubenswrapper[4870]: I1014 07:23:44.778897 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cf4d694c5-wjxfs" event={"ID":"751daf42-b5c1-4318-a15b-d5911c9d61ea","Type":"ContainerDied","Data":"e7f985e91ecb8ba5c0b4601a782047823bc6426fd3f278547d0ef69704ee2fcc"} Oct 14 07:23:44 crc kubenswrapper[4870]: I1014 07:23:44.791727 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="5212ba20-83c3-41ff-a3d9-5a315f009c1b" containerName="rabbitmq" containerID="cri-o://eeb00ec87974e56f742f7ae182732d1dfbbc2c6523987de46327c67f0a729b18" gracePeriod=604800 Oct 14 07:23:44 crc kubenswrapper[4870]: I1014 07:23:44.792231 4870 generic.go:334] "Generic (PLEG): container finished" podID="56eb286f-8aa2-46aa-a74d-031288a9aad1" containerID="9fa2f1146313b498f55c9bb3361bca48d735f25641dcd24bdbeac19bc4ae6377" exitCode=143 Oct 14 07:23:44 crc kubenswrapper[4870]: I1014 07:23:44.792310 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"56eb286f-8aa2-46aa-a74d-031288a9aad1","Type":"ContainerDied","Data":"9fa2f1146313b498f55c9bb3361bca48d735f25641dcd24bdbeac19bc4ae6377"} Oct 14 07:23:44 crc kubenswrapper[4870]: I1014 07:23:44.797422 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovsdbserver-sb-0" podUID="36717039-e3e2-4b74-b612-655556122574" containerName="ovsdbserver-sb" containerID="cri-o://afc051c5e25bf94e2cbc69eabd5bf3dc5066c8c7364b3f767d2bb33960df9191" gracePeriod=300 Oct 14 07:23:44 crc kubenswrapper[4870]: I1014 07:23:44.805979 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-cedd-account-create-gknx9"] Oct 14 07:23:44 crc kubenswrapper[4870]: I1014 07:23:44.811299 4870 generic.go:334] "Generic (PLEG): container finished" podID="503a07f8-2c05-458a-80ff-1abfe973dbd5" containerID="124ba16adf714b9b67c2030611bc8afc10924c09b62003e8aa1a0045f6015014" exitCode=0 Oct 14 07:23:44 crc kubenswrapper[4870]: I1014 07:23:44.811358 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-rn98c" event={"ID":"503a07f8-2c05-458a-80ff-1abfe973dbd5","Type":"ContainerDied","Data":"124ba16adf714b9b67c2030611bc8afc10924c09b62003e8aa1a0045f6015014"} Oct 14 07:23:44 crc kubenswrapper[4870]: I1014 07:23:44.836938 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-wg6ln"] Oct 14 07:23:44 crc kubenswrapper[4870]: I1014 07:23:44.865956 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-cedd-account-create-gknx9"] Oct 14 07:23:44 crc kubenswrapper[4870]: E1014 07:23:44.894908 4870 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of afc051c5e25bf94e2cbc69eabd5bf3dc5066c8c7364b3f767d2bb33960df9191 is running failed: container process not found" containerID="afc051c5e25bf94e2cbc69eabd5bf3dc5066c8c7364b3f767d2bb33960df9191" cmd=["/usr/bin/pidof","ovsdb-server"] Oct 14 07:23:44 crc kubenswrapper[4870]: E1014 07:23:44.908716 4870 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of afc051c5e25bf94e2cbc69eabd5bf3dc5066c8c7364b3f767d2bb33960df9191 is running failed: container process not found" containerID="afc051c5e25bf94e2cbc69eabd5bf3dc5066c8c7364b3f767d2bb33960df9191" cmd=["/usr/bin/pidof","ovsdb-server"] Oct 14 07:23:44 crc kubenswrapper[4870]: E1014 07:23:44.909259 4870 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of afc051c5e25bf94e2cbc69eabd5bf3dc5066c8c7364b3f767d2bb33960df9191 is running failed: container process not found" containerID="afc051c5e25bf94e2cbc69eabd5bf3dc5066c8c7364b3f767d2bb33960df9191" cmd=["/usr/bin/pidof","ovsdb-server"] Oct 14 07:23:44 crc kubenswrapper[4870]: E1014 07:23:44.909327 4870 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of afc051c5e25bf94e2cbc69eabd5bf3dc5066c8c7364b3f767d2bb33960df9191 is running failed: container process not found" probeType="Readiness" pod="openstack/ovsdbserver-sb-0" podUID="36717039-e3e2-4b74-b612-655556122574" containerName="ovsdbserver-sb" Oct 14 07:23:44 crc kubenswrapper[4870]: I1014 07:23:44.914750 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-wg6ln"] Oct 14 07:23:44 crc kubenswrapper[4870]: I1014 07:23:44.929363 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-6547-account-create-vqn99"] Oct 14 07:23:44 crc kubenswrapper[4870]: I1014 07:23:44.941296 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-6547-account-create-vqn99"] Oct 14 07:23:44 crc kubenswrapper[4870]: I1014 07:23:44.948333 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-controller-ovs-6482n" podUID="91a2e1de-104f-4e14-9192-c6587ed13be1" containerName="ovs-vswitchd" containerID="cri-o://72ead03af69bb2d536002fe16f85a88974968529ec05c0e9e78254a3f796b690" gracePeriod=29 Oct 14 07:23:44 crc kubenswrapper[4870]: I1014 07:23:44.959391 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican6547-account-delete-v5crl"] Oct 14 07:23:44 crc kubenswrapper[4870]: I1014 07:23:44.973900 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-f6n86"] Oct 14 07:23:44 crc kubenswrapper[4870]: I1014 07:23:44.986683 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-f6n86"] Oct 14 07:23:44 crc kubenswrapper[4870]: I1014 07:23:44.992314 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance1aaa-account-delete-nct6c"] Oct 14 07:23:44 crc kubenswrapper[4870]: E1014 07:23:44.998940 4870 handlers.go:78] "Exec lifecycle hook for Container in Pod failed" err=< Oct 14 07:23:44 crc kubenswrapper[4870]: command '/usr/local/bin/container-scripts/stop-ovsdb-server.sh' exited with 137: ++ dirname /usr/local/bin/container-scripts/stop-ovsdb-server.sh Oct 14 07:23:44 crc kubenswrapper[4870]: + source /usr/local/bin/container-scripts/functions Oct 14 07:23:44 crc kubenswrapper[4870]: ++ OVNBridge=br-int Oct 14 07:23:44 crc kubenswrapper[4870]: ++ OVNRemote=tcp:localhost:6642 Oct 14 07:23:44 crc kubenswrapper[4870]: ++ OVNEncapType=geneve Oct 14 07:23:44 crc kubenswrapper[4870]: ++ OVNAvailabilityZones= Oct 14 07:23:44 crc kubenswrapper[4870]: ++ EnableChassisAsGateway=true Oct 14 07:23:44 crc kubenswrapper[4870]: ++ PhysicalNetworks= Oct 14 07:23:44 crc kubenswrapper[4870]: ++ OVNHostName= Oct 14 07:23:44 crc kubenswrapper[4870]: ++ DB_FILE=/etc/openvswitch/conf.db Oct 14 07:23:44 crc kubenswrapper[4870]: ++ ovs_dir=/var/lib/openvswitch Oct 14 07:23:44 crc kubenswrapper[4870]: ++ FLOWS_RESTORE_SCRIPT=/var/lib/openvswitch/flows-script Oct 14 07:23:44 crc kubenswrapper[4870]: ++ FLOWS_RESTORE_DIR=/var/lib/openvswitch/saved-flows Oct 14 07:23:44 crc kubenswrapper[4870]: ++ SAFE_TO_STOP_OVSDB_SERVER_SEMAPHORE=/var/lib/openvswitch/is_safe_to_stop_ovsdb_server Oct 14 07:23:44 crc kubenswrapper[4870]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Oct 14 07:23:44 crc kubenswrapper[4870]: + sleep 0.5 Oct 14 07:23:44 crc kubenswrapper[4870]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Oct 14 07:23:44 crc kubenswrapper[4870]: + sleep 0.5 Oct 14 07:23:44 crc kubenswrapper[4870]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Oct 14 07:23:44 crc kubenswrapper[4870]: + sleep 0.5 Oct 14 07:23:44 crc kubenswrapper[4870]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Oct 14 07:23:44 crc kubenswrapper[4870]: + cleanup_ovsdb_server_semaphore Oct 14 07:23:44 crc kubenswrapper[4870]: + rm -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server Oct 14 07:23:44 crc kubenswrapper[4870]: + /usr/share/openvswitch/scripts/ovs-ctl stop --no-ovs-vswitchd Oct 14 07:23:44 crc kubenswrapper[4870]: > execCommand=["/usr/local/bin/container-scripts/stop-ovsdb-server.sh"] containerName="ovsdb-server" pod="openstack/ovn-controller-ovs-6482n" message=< Oct 14 07:23:44 crc kubenswrapper[4870]: Exiting ovsdb-server (5) [ OK ] Oct 14 07:23:44 crc kubenswrapper[4870]: ++ dirname /usr/local/bin/container-scripts/stop-ovsdb-server.sh Oct 14 07:23:44 crc kubenswrapper[4870]: + source /usr/local/bin/container-scripts/functions Oct 14 07:23:44 crc kubenswrapper[4870]: ++ OVNBridge=br-int Oct 14 07:23:44 crc kubenswrapper[4870]: ++ OVNRemote=tcp:localhost:6642 Oct 14 07:23:44 crc kubenswrapper[4870]: ++ OVNEncapType=geneve Oct 14 07:23:44 crc kubenswrapper[4870]: ++ OVNAvailabilityZones= Oct 14 07:23:44 crc kubenswrapper[4870]: ++ EnableChassisAsGateway=true Oct 14 07:23:44 crc kubenswrapper[4870]: ++ PhysicalNetworks= Oct 14 07:23:44 crc kubenswrapper[4870]: ++ OVNHostName= Oct 14 07:23:44 crc kubenswrapper[4870]: ++ DB_FILE=/etc/openvswitch/conf.db Oct 14 07:23:44 crc kubenswrapper[4870]: ++ ovs_dir=/var/lib/openvswitch Oct 14 07:23:44 crc kubenswrapper[4870]: ++ FLOWS_RESTORE_SCRIPT=/var/lib/openvswitch/flows-script Oct 14 07:23:44 crc kubenswrapper[4870]: ++ FLOWS_RESTORE_DIR=/var/lib/openvswitch/saved-flows Oct 14 07:23:44 crc kubenswrapper[4870]: ++ SAFE_TO_STOP_OVSDB_SERVER_SEMAPHORE=/var/lib/openvswitch/is_safe_to_stop_ovsdb_server Oct 14 07:23:44 crc kubenswrapper[4870]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Oct 14 07:23:44 crc kubenswrapper[4870]: + sleep 0.5 Oct 14 07:23:44 crc kubenswrapper[4870]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Oct 14 07:23:44 crc kubenswrapper[4870]: + sleep 0.5 Oct 14 07:23:44 crc kubenswrapper[4870]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Oct 14 07:23:44 crc kubenswrapper[4870]: + sleep 0.5 Oct 14 07:23:44 crc kubenswrapper[4870]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Oct 14 07:23:44 crc kubenswrapper[4870]: + cleanup_ovsdb_server_semaphore Oct 14 07:23:44 crc kubenswrapper[4870]: + rm -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server Oct 14 07:23:44 crc kubenswrapper[4870]: + /usr/share/openvswitch/scripts/ovs-ctl stop --no-ovs-vswitchd Oct 14 07:23:44 crc kubenswrapper[4870]: > Oct 14 07:23:44 crc kubenswrapper[4870]: E1014 07:23:44.999002 4870 kuberuntime_container.go:691] "PreStop hook failed" err=< Oct 14 07:23:44 crc kubenswrapper[4870]: command '/usr/local/bin/container-scripts/stop-ovsdb-server.sh' exited with 137: ++ dirname /usr/local/bin/container-scripts/stop-ovsdb-server.sh Oct 14 07:23:44 crc kubenswrapper[4870]: + source /usr/local/bin/container-scripts/functions Oct 14 07:23:44 crc kubenswrapper[4870]: ++ OVNBridge=br-int Oct 14 07:23:44 crc kubenswrapper[4870]: ++ OVNRemote=tcp:localhost:6642 Oct 14 07:23:44 crc kubenswrapper[4870]: ++ OVNEncapType=geneve Oct 14 07:23:44 crc kubenswrapper[4870]: ++ OVNAvailabilityZones= Oct 14 07:23:44 crc kubenswrapper[4870]: ++ EnableChassisAsGateway=true Oct 14 07:23:44 crc kubenswrapper[4870]: ++ PhysicalNetworks= Oct 14 07:23:44 crc kubenswrapper[4870]: ++ OVNHostName= Oct 14 07:23:44 crc kubenswrapper[4870]: ++ DB_FILE=/etc/openvswitch/conf.db Oct 14 07:23:44 crc kubenswrapper[4870]: ++ ovs_dir=/var/lib/openvswitch Oct 14 07:23:44 crc kubenswrapper[4870]: ++ FLOWS_RESTORE_SCRIPT=/var/lib/openvswitch/flows-script Oct 14 07:23:44 crc kubenswrapper[4870]: ++ FLOWS_RESTORE_DIR=/var/lib/openvswitch/saved-flows Oct 14 07:23:44 crc kubenswrapper[4870]: ++ SAFE_TO_STOP_OVSDB_SERVER_SEMAPHORE=/var/lib/openvswitch/is_safe_to_stop_ovsdb_server Oct 14 07:23:44 crc kubenswrapper[4870]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Oct 14 07:23:44 crc kubenswrapper[4870]: + sleep 0.5 Oct 14 07:23:44 crc kubenswrapper[4870]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Oct 14 07:23:44 crc kubenswrapper[4870]: + sleep 0.5 Oct 14 07:23:44 crc kubenswrapper[4870]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Oct 14 07:23:44 crc kubenswrapper[4870]: + sleep 0.5 Oct 14 07:23:44 crc kubenswrapper[4870]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Oct 14 07:23:44 crc kubenswrapper[4870]: + cleanup_ovsdb_server_semaphore Oct 14 07:23:44 crc kubenswrapper[4870]: + rm -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server Oct 14 07:23:44 crc kubenswrapper[4870]: + /usr/share/openvswitch/scripts/ovs-ctl stop --no-ovs-vswitchd Oct 14 07:23:44 crc kubenswrapper[4870]: > pod="openstack/ovn-controller-ovs-6482n" podUID="91a2e1de-104f-4e14-9192-c6587ed13be1" containerName="ovsdb-server" containerID="cri-o://6f14317c64b4cb248d4254d308d86fadbdb28d2daee87abbf75414c219c2c477" Oct 14 07:23:44 crc kubenswrapper[4870]: I1014 07:23:44.999055 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-controller-ovs-6482n" podUID="91a2e1de-104f-4e14-9192-c6587ed13be1" containerName="ovsdb-server" containerID="cri-o://6f14317c64b4cb248d4254d308d86fadbdb28d2daee87abbf75414c219c2c477" gracePeriod=29 Oct 14 07:23:45 crc kubenswrapper[4870]: I1014 07:23:45.010640 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-1aaa-account-create-6t7t7"] Oct 14 07:23:45 crc kubenswrapper[4870]: I1014 07:23:45.020993 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-1aaa-account-create-6t7t7"] Oct 14 07:23:45 crc kubenswrapper[4870]: I1014 07:23:45.090825 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="000463bc-f3ac-40b7-90c8-32366e327a19" path="/var/lib/kubelet/pods/000463bc-f3ac-40b7-90c8-32366e327a19/volumes" Oct 14 07:23:45 crc kubenswrapper[4870]: I1014 07:23:45.091687 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0c16dc18-0486-420e-86c8-ed2374a9a705" path="/var/lib/kubelet/pods/0c16dc18-0486-420e-86c8-ed2374a9a705/volumes" Oct 14 07:23:45 crc kubenswrapper[4870]: I1014 07:23:45.092224 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1110cc74-c641-4a3a-b46d-ea85668515b0" path="/var/lib/kubelet/pods/1110cc74-c641-4a3a-b46d-ea85668515b0/volumes" Oct 14 07:23:45 crc kubenswrapper[4870]: I1014 07:23:45.094836 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3c2004c3-445d-4495-b24d-862f402ff5a6" path="/var/lib/kubelet/pods/3c2004c3-445d-4495-b24d-862f402ff5a6/volumes" Oct 14 07:23:45 crc kubenswrapper[4870]: I1014 07:23:45.095980 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5056171a-6fc0-4d48-9ff0-e275137253f3" path="/var/lib/kubelet/pods/5056171a-6fc0-4d48-9ff0-e275137253f3/volumes" Oct 14 07:23:45 crc kubenswrapper[4870]: I1014 07:23:45.096521 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="625bb722-0865-47f2-b680-ab642ee6c9e5" path="/var/lib/kubelet/pods/625bb722-0865-47f2-b680-ab642ee6c9e5/volumes" Oct 14 07:23:45 crc kubenswrapper[4870]: I1014 07:23:45.097401 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="67d40a7a-1bb5-457c-9738-b4a05378fee4" path="/var/lib/kubelet/pods/67d40a7a-1bb5-457c-9738-b4a05378fee4/volumes" Oct 14 07:23:45 crc kubenswrapper[4870]: I1014 07:23:45.097981 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="685539fb-44a2-4f75-b418-65383d02152e" path="/var/lib/kubelet/pods/685539fb-44a2-4f75-b418-65383d02152e/volumes" Oct 14 07:23:45 crc kubenswrapper[4870]: I1014 07:23:45.098560 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6e87f821-3bc7-4b88-9c72-ccd80b4b19e9" path="/var/lib/kubelet/pods/6e87f821-3bc7-4b88-9c72-ccd80b4b19e9/volumes" Oct 14 07:23:45 crc kubenswrapper[4870]: I1014 07:23:45.099661 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7170a9ca-4322-4d2c-9550-ee587ec981db" path="/var/lib/kubelet/pods/7170a9ca-4322-4d2c-9550-ee587ec981db/volumes" Oct 14 07:23:45 crc kubenswrapper[4870]: I1014 07:23:45.100323 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="73ffa9f3-6390-4ed0-b0ba-5a3e2a44ae1f" path="/var/lib/kubelet/pods/73ffa9f3-6390-4ed0-b0ba-5a3e2a44ae1f/volumes" Oct 14 07:23:45 crc kubenswrapper[4870]: I1014 07:23:45.100846 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="835bb9d2-3b63-4fa5-bda6-9cc6a9a899a2" path="/var/lib/kubelet/pods/835bb9d2-3b63-4fa5-bda6-9cc6a9a899a2/volumes" Oct 14 07:23:45 crc kubenswrapper[4870]: I1014 07:23:45.102362 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="85aaeb67-e809-416d-b4e0-eb7605cdd307" path="/var/lib/kubelet/pods/85aaeb67-e809-416d-b4e0-eb7605cdd307/volumes" Oct 14 07:23:45 crc kubenswrapper[4870]: I1014 07:23:45.103028 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ae296f6e-96fb-4e1d-bb24-4f4143558ee2" path="/var/lib/kubelet/pods/ae296f6e-96fb-4e1d-bb24-4f4143558ee2/volumes" Oct 14 07:23:45 crc kubenswrapper[4870]: I1014 07:23:45.103622 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd4cf561-b2b4-486b-b317-bd88b71badd9" path="/var/lib/kubelet/pods/cd4cf561-b2b4-486b-b317-bd88b71badd9/volumes" Oct 14 07:23:45 crc kubenswrapper[4870]: I1014 07:23:45.104236 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d147118e-e08e-4c64-b157-da3ca3afb6d7" path="/var/lib/kubelet/pods/d147118e-e08e-4c64-b157-da3ca3afb6d7/volumes" Oct 14 07:23:45 crc kubenswrapper[4870]: I1014 07:23:45.105347 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f5e5bd22-1692-477a-a375-3eecad66c746" path="/var/lib/kubelet/pods/f5e5bd22-1692-477a-a375-3eecad66c746/volumes" Oct 14 07:23:45 crc kubenswrapper[4870]: I1014 07:23:45.107197 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 14 07:23:45 crc kubenswrapper[4870]: I1014 07:23:45.107234 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 14 07:23:45 crc kubenswrapper[4870]: I1014 07:23:45.107248 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 14 07:23:45 crc kubenswrapper[4870]: I1014 07:23:45.108645 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="f5b700ed-3919-458f-860e-5630b8e4b800" containerName="nova-scheduler-scheduler" containerID="cri-o://19bfdce54348e895d0d26a8651c758e77337522908dc3ad6eb2470e00c72c2cf" gracePeriod=30 Oct 14 07:23:45 crc kubenswrapper[4870]: I1014 07:23:45.109698 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="ff28d060-b280-4b89-bea1-d10aa367a3cc" containerName="nova-api-log" containerID="cri-o://6d95049e97ec6fbbfa6d28e3e543ed1f7fc74569f83d0862a392737896e3463d" gracePeriod=30 Oct 14 07:23:45 crc kubenswrapper[4870]: I1014 07:23:45.109891 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="ff28d060-b280-4b89-bea1-d10aa367a3cc" containerName="nova-api-api" containerID="cri-o://9867cb00fe7f9cff6394884f9edfc786d6b848597f09a56abffe01cb5e8ca4cf" gracePeriod=30 Oct 14 07:23:45 crc kubenswrapper[4870]: I1014 07:23:45.122080 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-52fb-account-create-8kmf7"] Oct 14 07:23:45 crc kubenswrapper[4870]: I1014 07:23:45.135573 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 14 07:23:45 crc kubenswrapper[4870]: I1014 07:23:45.136222 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="f0b766fd-ccb9-4620-84df-8fe2f194b0da" containerName="nova-metadata-metadata" containerID="cri-o://65770c73c87b754f9177bcc262af5253b96b0534c636832b413175980c00c785" gracePeriod=30 Oct 14 07:23:45 crc kubenswrapper[4870]: I1014 07:23:45.136378 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="f0b766fd-ccb9-4620-84df-8fe2f194b0da" containerName="nova-metadata-log" containerID="cri-o://6fb3ea4d6cd5802f3695559f939e61f05770876bd3c155dcf6040899ef93330a" gracePeriod=30 Oct 14 07:23:45 crc kubenswrapper[4870]: I1014 07:23:45.153494 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-29752"] Oct 14 07:23:45 crc kubenswrapper[4870]: I1014 07:23:45.172025 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-52fb-account-create-8kmf7"] Oct 14 07:23:45 crc kubenswrapper[4870]: I1014 07:23:45.214773 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/novacell0ce16-account-delete-bwgsq"] Oct 14 07:23:45 crc kubenswrapper[4870]: I1014 07:23:45.234737 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-9dqlp_4c8d0269-7501-4c96-9f9e-30bea4b0fb92/openstack-network-exporter/0.log" Oct 14 07:23:45 crc kubenswrapper[4870]: I1014 07:23:45.234819 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-9dqlp" Oct 14 07:23:45 crc kubenswrapper[4870]: I1014 07:23:45.260539 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7cf4d694c5-wjxfs" Oct 14 07:23:45 crc kubenswrapper[4870]: I1014 07:23:45.270628 4870 scope.go:117] "RemoveContainer" containerID="76d17f5aef5402fa0f6135921f3ba0aeb9f5640152e827b4385348efdbe620d3" Oct 14 07:23:45 crc kubenswrapper[4870]: I1014 07:23:45.294446 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-29752"] Oct 14 07:23:45 crc kubenswrapper[4870]: I1014 07:23:45.311908 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-snx2f"] Oct 14 07:23:45 crc kubenswrapper[4870]: I1014 07:23:45.334899 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-ce16-account-create-nqg26"] Oct 14 07:23:45 crc kubenswrapper[4870]: I1014 07:23:45.337320 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-rn98c" Oct 14 07:23:45 crc kubenswrapper[4870]: I1014 07:23:45.348771 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-keystone-listener-57d8bd87b6-vnpct"] Oct 14 07:23:45 crc kubenswrapper[4870]: I1014 07:23:45.349043 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-keystone-listener-57d8bd87b6-vnpct" podUID="0a2e4c80-d44f-40c0-9c04-74e9ce98ae36" containerName="barbican-keystone-listener-log" containerID="cri-o://bc61afdd96aedc563124f07cd981291ea8c01a0a034c78fba2e57e484564a0ee" gracePeriod=30 Oct 14 07:23:45 crc kubenswrapper[4870]: I1014 07:23:45.349183 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-keystone-listener-57d8bd87b6-vnpct" podUID="0a2e4c80-d44f-40c0-9c04-74e9ce98ae36" containerName="barbican-keystone-listener" containerID="cri-o://ea29b92302ab943ca5bc51b26bb9eba8dd8e4aa271ebf5c213d4ae71f3523daf" gracePeriod=30 Oct 14 07:23:45 crc kubenswrapper[4870]: I1014 07:23:45.360324 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-snx2f"] Oct 14 07:23:45 crc kubenswrapper[4870]: I1014 07:23:45.364630 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/4c8d0269-7501-4c96-9f9e-30bea4b0fb92-metrics-certs-tls-certs\") pod \"4c8d0269-7501-4c96-9f9e-30bea4b0fb92\" (UID: \"4c8d0269-7501-4c96-9f9e-30bea4b0fb92\") " Oct 14 07:23:45 crc kubenswrapper[4870]: I1014 07:23:45.364795 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/751daf42-b5c1-4318-a15b-d5911c9d61ea-ovsdbserver-sb\") pod \"751daf42-b5c1-4318-a15b-d5911c9d61ea\" (UID: \"751daf42-b5c1-4318-a15b-d5911c9d61ea\") " Oct 14 07:23:45 crc kubenswrapper[4870]: I1014 07:23:45.364848 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/4c8d0269-7501-4c96-9f9e-30bea4b0fb92-ovn-rundir\") pod \"4c8d0269-7501-4c96-9f9e-30bea4b0fb92\" (UID: \"4c8d0269-7501-4c96-9f9e-30bea4b0fb92\") " Oct 14 07:23:45 crc kubenswrapper[4870]: I1014 07:23:45.364888 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/4c8d0269-7501-4c96-9f9e-30bea4b0fb92-ovs-rundir\") pod \"4c8d0269-7501-4c96-9f9e-30bea4b0fb92\" (UID: \"4c8d0269-7501-4c96-9f9e-30bea4b0fb92\") " Oct 14 07:23:45 crc kubenswrapper[4870]: I1014 07:23:45.364905 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c8d0269-7501-4c96-9f9e-30bea4b0fb92-combined-ca-bundle\") pod \"4c8d0269-7501-4c96-9f9e-30bea4b0fb92\" (UID: \"4c8d0269-7501-4c96-9f9e-30bea4b0fb92\") " Oct 14 07:23:45 crc kubenswrapper[4870]: I1014 07:23:45.364947 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4c8d0269-7501-4c96-9f9e-30bea4b0fb92-config\") pod \"4c8d0269-7501-4c96-9f9e-30bea4b0fb92\" (UID: \"4c8d0269-7501-4c96-9f9e-30bea4b0fb92\") " Oct 14 07:23:45 crc kubenswrapper[4870]: I1014 07:23:45.365048 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/751daf42-b5c1-4318-a15b-d5911c9d61ea-ovsdbserver-nb\") pod \"751daf42-b5c1-4318-a15b-d5911c9d61ea\" (UID: \"751daf42-b5c1-4318-a15b-d5911c9d61ea\") " Oct 14 07:23:45 crc kubenswrapper[4870]: I1014 07:23:45.365162 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s8f8j\" (UniqueName: \"kubernetes.io/projected/751daf42-b5c1-4318-a15b-d5911c9d61ea-kube-api-access-s8f8j\") pod \"751daf42-b5c1-4318-a15b-d5911c9d61ea\" (UID: \"751daf42-b5c1-4318-a15b-d5911c9d61ea\") " Oct 14 07:23:45 crc kubenswrapper[4870]: I1014 07:23:45.365237 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8t6l9\" (UniqueName: \"kubernetes.io/projected/4c8d0269-7501-4c96-9f9e-30bea4b0fb92-kube-api-access-8t6l9\") pod \"4c8d0269-7501-4c96-9f9e-30bea4b0fb92\" (UID: \"4c8d0269-7501-4c96-9f9e-30bea4b0fb92\") " Oct 14 07:23:45 crc kubenswrapper[4870]: I1014 07:23:45.365271 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/751daf42-b5c1-4318-a15b-d5911c9d61ea-config\") pod \"751daf42-b5c1-4318-a15b-d5911c9d61ea\" (UID: \"751daf42-b5c1-4318-a15b-d5911c9d61ea\") " Oct 14 07:23:45 crc kubenswrapper[4870]: I1014 07:23:45.365293 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/751daf42-b5c1-4318-a15b-d5911c9d61ea-dns-swift-storage-0\") pod \"751daf42-b5c1-4318-a15b-d5911c9d61ea\" (UID: \"751daf42-b5c1-4318-a15b-d5911c9d61ea\") " Oct 14 07:23:45 crc kubenswrapper[4870]: I1014 07:23:45.365321 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/751daf42-b5c1-4318-a15b-d5911c9d61ea-dns-svc\") pod \"751daf42-b5c1-4318-a15b-d5911c9d61ea\" (UID: \"751daf42-b5c1-4318-a15b-d5911c9d61ea\") " Oct 14 07:23:45 crc kubenswrapper[4870]: I1014 07:23:45.367252 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4c8d0269-7501-4c96-9f9e-30bea4b0fb92-config" (OuterVolumeSpecName: "config") pod "4c8d0269-7501-4c96-9f9e-30bea4b0fb92" (UID: "4c8d0269-7501-4c96-9f9e-30bea4b0fb92"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:23:45 crc kubenswrapper[4870]: I1014 07:23:45.370923 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4c8d0269-7501-4c96-9f9e-30bea4b0fb92-ovn-rundir" (OuterVolumeSpecName: "ovn-rundir") pod "4c8d0269-7501-4c96-9f9e-30bea4b0fb92" (UID: "4c8d0269-7501-4c96-9f9e-30bea4b0fb92"). InnerVolumeSpecName "ovn-rundir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 14 07:23:45 crc kubenswrapper[4870]: I1014 07:23:45.375037 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4c8d0269-7501-4c96-9f9e-30bea4b0fb92-ovs-rundir" (OuterVolumeSpecName: "ovs-rundir") pod "4c8d0269-7501-4c96-9f9e-30bea4b0fb92" (UID: "4c8d0269-7501-4c96-9f9e-30bea4b0fb92"). InnerVolumeSpecName "ovs-rundir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 14 07:23:45 crc kubenswrapper[4870]: I1014 07:23:45.375102 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-ce16-account-create-nqg26"] Oct 14 07:23:45 crc kubenswrapper[4870]: I1014 07:23:45.376030 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-e66c-account-create-wks4d"] Oct 14 07:23:45 crc kubenswrapper[4870]: I1014 07:23:45.382542 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-e66c-account-create-wks4d"] Oct 14 07:23:45 crc kubenswrapper[4870]: I1014 07:23:45.389209 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstack-cell1-galera-0" podUID="4d74ef50-1861-4d70-b781-f92d5f4b35ba" containerName="galera" containerID="cri-o://9389d35044ef25b0e127f63d901f00d3d48e58573671b63fa03de1238628ba21" gracePeriod=30 Oct 14 07:23:45 crc kubenswrapper[4870]: I1014 07:23:45.393831 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4c8d0269-7501-4c96-9f9e-30bea4b0fb92-kube-api-access-8t6l9" (OuterVolumeSpecName: "kube-api-access-8t6l9") pod "4c8d0269-7501-4c96-9f9e-30bea4b0fb92" (UID: "4c8d0269-7501-4c96-9f9e-30bea4b0fb92"). InnerVolumeSpecName "kube-api-access-8t6l9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:23:45 crc kubenswrapper[4870]: I1014 07:23:45.397914 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 14 07:23:45 crc kubenswrapper[4870]: I1014 07:23:45.399739 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/751daf42-b5c1-4318-a15b-d5911c9d61ea-kube-api-access-s8f8j" (OuterVolumeSpecName: "kube-api-access-s8f8j") pod "751daf42-b5c1-4318-a15b-d5911c9d61ea" (UID: "751daf42-b5c1-4318-a15b-d5911c9d61ea"). InnerVolumeSpecName "kube-api-access-s8f8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:23:45 crc kubenswrapper[4870]: I1014 07:23:45.408511 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-hjj9z"] Oct 14 07:23:45 crc kubenswrapper[4870]: I1014 07:23:45.425962 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-hjj9z"] Oct 14 07:23:45 crc kubenswrapper[4870]: I1014 07:23:45.431755 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/novaapie66c-account-delete-ftd24"] Oct 14 07:23:45 crc kubenswrapper[4870]: I1014 07:23:45.455540 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-worker-85f775bc47-d95qn"] Oct 14 07:23:45 crc kubenswrapper[4870]: I1014 07:23:45.455884 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-worker-85f775bc47-d95qn" podUID="c5bb3fde-cfaf-4d09-ad31-e23fd830eb91" containerName="barbican-worker-log" containerID="cri-o://a77bf6d6eb57cf53a83d8f0ff81a500083d29df6ab1d56a72558d63fae3b342f" gracePeriod=30 Oct 14 07:23:45 crc kubenswrapper[4870]: I1014 07:23:45.456492 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-worker-85f775bc47-d95qn" podUID="c5bb3fde-cfaf-4d09-ad31-e23fd830eb91" containerName="barbican-worker" containerID="cri-o://94a43ed16da62727b0d42c318ec266880c3131b81bd05fdd1dfddbb2a180533e" gracePeriod=30 Oct 14 07:23:45 crc kubenswrapper[4870]: I1014 07:23:45.482102 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 14 07:23:45 crc kubenswrapper[4870]: I1014 07:23:45.482166 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-79cc757548-zxf4g"] Oct 14 07:23:45 crc kubenswrapper[4870]: I1014 07:23:45.497880 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-79cc757548-zxf4g" podUID="cae5b655-5d7c-4063-90ad-4d9913792d5b" containerName="barbican-api-log" containerID="cri-o://a507ea103a81fde95c96ade7a313c18b4faefbe53e02c99e5cd596e0c8dcb53e" gracePeriod=30 Oct 14 07:23:45 crc kubenswrapper[4870]: I1014 07:23:45.498953 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2jg2f\" (UniqueName: \"kubernetes.io/projected/503a07f8-2c05-458a-80ff-1abfe973dbd5-kube-api-access-2jg2f\") pod \"503a07f8-2c05-458a-80ff-1abfe973dbd5\" (UID: \"503a07f8-2c05-458a-80ff-1abfe973dbd5\") " Oct 14 07:23:45 crc kubenswrapper[4870]: I1014 07:23:45.499020 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/503a07f8-2c05-458a-80ff-1abfe973dbd5-var-log-ovn\") pod \"503a07f8-2c05-458a-80ff-1abfe973dbd5\" (UID: \"503a07f8-2c05-458a-80ff-1abfe973dbd5\") " Oct 14 07:23:45 crc kubenswrapper[4870]: I1014 07:23:45.499101 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/503a07f8-2c05-458a-80ff-1abfe973dbd5-scripts\") pod \"503a07f8-2c05-458a-80ff-1abfe973dbd5\" (UID: \"503a07f8-2c05-458a-80ff-1abfe973dbd5\") " Oct 14 07:23:45 crc kubenswrapper[4870]: I1014 07:23:45.499138 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/503a07f8-2c05-458a-80ff-1abfe973dbd5-ovn-controller-tls-certs\") pod \"503a07f8-2c05-458a-80ff-1abfe973dbd5\" (UID: \"503a07f8-2c05-458a-80ff-1abfe973dbd5\") " Oct 14 07:23:45 crc kubenswrapper[4870]: I1014 07:23:45.499187 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/503a07f8-2c05-458a-80ff-1abfe973dbd5-var-run\") pod \"503a07f8-2c05-458a-80ff-1abfe973dbd5\" (UID: \"503a07f8-2c05-458a-80ff-1abfe973dbd5\") " Oct 14 07:23:45 crc kubenswrapper[4870]: I1014 07:23:45.499373 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/503a07f8-2c05-458a-80ff-1abfe973dbd5-combined-ca-bundle\") pod \"503a07f8-2c05-458a-80ff-1abfe973dbd5\" (UID: \"503a07f8-2c05-458a-80ff-1abfe973dbd5\") " Oct 14 07:23:45 crc kubenswrapper[4870]: I1014 07:23:45.499415 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/503a07f8-2c05-458a-80ff-1abfe973dbd5-var-run-ovn\") pod \"503a07f8-2c05-458a-80ff-1abfe973dbd5\" (UID: \"503a07f8-2c05-458a-80ff-1abfe973dbd5\") " Oct 14 07:23:45 crc kubenswrapper[4870]: I1014 07:23:45.499453 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-79cc757548-zxf4g" podUID="cae5b655-5d7c-4063-90ad-4d9913792d5b" containerName="barbican-api" containerID="cri-o://b97a3d2764961f388fdc1a6459d91ca5f1248920ce92c572b651657c109e7fa4" gracePeriod=30 Oct 14 07:23:45 crc kubenswrapper[4870]: I1014 07:23:45.500907 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/503a07f8-2c05-458a-80ff-1abfe973dbd5-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "503a07f8-2c05-458a-80ff-1abfe973dbd5" (UID: "503a07f8-2c05-458a-80ff-1abfe973dbd5"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 14 07:23:45 crc kubenswrapper[4870]: I1014 07:23:45.501441 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/503a07f8-2c05-458a-80ff-1abfe973dbd5-var-run" (OuterVolumeSpecName: "var-run") pod "503a07f8-2c05-458a-80ff-1abfe973dbd5" (UID: "503a07f8-2c05-458a-80ff-1abfe973dbd5"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 14 07:23:45 crc kubenswrapper[4870]: I1014 07:23:45.502315 4870 reconciler_common.go:293] "Volume detached for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/4c8d0269-7501-4c96-9f9e-30bea4b0fb92-ovn-rundir\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:45 crc kubenswrapper[4870]: I1014 07:23:45.502334 4870 reconciler_common.go:293] "Volume detached for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/4c8d0269-7501-4c96-9f9e-30bea4b0fb92-ovs-rundir\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:45 crc kubenswrapper[4870]: I1014 07:23:45.502351 4870 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/503a07f8-2c05-458a-80ff-1abfe973dbd5-var-log-ovn\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:45 crc kubenswrapper[4870]: I1014 07:23:45.502361 4870 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4c8d0269-7501-4c96-9f9e-30bea4b0fb92-config\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:45 crc kubenswrapper[4870]: I1014 07:23:45.502369 4870 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/503a07f8-2c05-458a-80ff-1abfe973dbd5-var-run\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:45 crc kubenswrapper[4870]: I1014 07:23:45.502382 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s8f8j\" (UniqueName: \"kubernetes.io/projected/751daf42-b5c1-4318-a15b-d5911c9d61ea-kube-api-access-s8f8j\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:45 crc kubenswrapper[4870]: I1014 07:23:45.502393 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8t6l9\" (UniqueName: \"kubernetes.io/projected/4c8d0269-7501-4c96-9f9e-30bea4b0fb92-kube-api-access-8t6l9\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:45 crc kubenswrapper[4870]: I1014 07:23:45.503475 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/503a07f8-2c05-458a-80ff-1abfe973dbd5-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "503a07f8-2c05-458a-80ff-1abfe973dbd5" (UID: "503a07f8-2c05-458a-80ff-1abfe973dbd5"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 14 07:23:45 crc kubenswrapper[4870]: I1014 07:23:45.504645 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/503a07f8-2c05-458a-80ff-1abfe973dbd5-scripts" (OuterVolumeSpecName: "scripts") pod "503a07f8-2c05-458a-80ff-1abfe973dbd5" (UID: "503a07f8-2c05-458a-80ff-1abfe973dbd5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:23:45 crc kubenswrapper[4870]: I1014 07:23:45.537022 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/503a07f8-2c05-458a-80ff-1abfe973dbd5-kube-api-access-2jg2f" (OuterVolumeSpecName: "kube-api-access-2jg2f") pod "503a07f8-2c05-458a-80ff-1abfe973dbd5" (UID: "503a07f8-2c05-458a-80ff-1abfe973dbd5"). InnerVolumeSpecName "kube-api-access-2jg2f". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:23:45 crc kubenswrapper[4870]: I1014 07:23:45.588816 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="da8a9731-f917-462f-9932-b37b6abb9a64" containerName="rabbitmq" containerID="cri-o://bc4b5271d735615160d1289dff31839e1ba9c4495e333ee1f50cab05851fd959" gracePeriod=604800 Oct 14 07:23:45 crc kubenswrapper[4870]: I1014 07:23:45.596742 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_36717039-e3e2-4b74-b612-655556122574/ovsdbserver-sb/0.log" Oct 14 07:23:45 crc kubenswrapper[4870]: I1014 07:23:45.596897 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Oct 14 07:23:45 crc kubenswrapper[4870]: I1014 07:23:45.597107 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 14 07:23:45 crc kubenswrapper[4870]: I1014 07:23:45.612297 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/0965b7e6-2aa0-4940-a130-324cfb08de5a-openstack-config\") pod \"0965b7e6-2aa0-4940-a130-324cfb08de5a\" (UID: \"0965b7e6-2aa0-4940-a130-324cfb08de5a\") " Oct 14 07:23:45 crc kubenswrapper[4870]: I1014 07:23:45.612467 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/0965b7e6-2aa0-4940-a130-324cfb08de5a-openstack-config-secret\") pod \"0965b7e6-2aa0-4940-a130-324cfb08de5a\" (UID: \"0965b7e6-2aa0-4940-a130-324cfb08de5a\") " Oct 14 07:23:45 crc kubenswrapper[4870]: I1014 07:23:45.612585 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6zpqx\" (UniqueName: \"kubernetes.io/projected/0965b7e6-2aa0-4940-a130-324cfb08de5a-kube-api-access-6zpqx\") pod \"0965b7e6-2aa0-4940-a130-324cfb08de5a\" (UID: \"0965b7e6-2aa0-4940-a130-324cfb08de5a\") " Oct 14 07:23:45 crc kubenswrapper[4870]: I1014 07:23:45.612606 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0965b7e6-2aa0-4940-a130-324cfb08de5a-combined-ca-bundle\") pod \"0965b7e6-2aa0-4940-a130-324cfb08de5a\" (UID: \"0965b7e6-2aa0-4940-a130-324cfb08de5a\") " Oct 14 07:23:45 crc kubenswrapper[4870]: I1014 07:23:45.613805 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="8abede26-36e8-46ec-a5c4-6cd53319133a" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://45200242c420e7ff44ca9c10b9cff44cc8bb4c9294423cc3e3e0ad235691f6b4" gracePeriod=30 Oct 14 07:23:45 crc kubenswrapper[4870]: I1014 07:23:45.615734 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2jg2f\" (UniqueName: \"kubernetes.io/projected/503a07f8-2c05-458a-80ff-1abfe973dbd5-kube-api-access-2jg2f\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:45 crc kubenswrapper[4870]: I1014 07:23:45.615764 4870 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/503a07f8-2c05-458a-80ff-1abfe973dbd5-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:45 crc kubenswrapper[4870]: I1014 07:23:45.615774 4870 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/503a07f8-2c05-458a-80ff-1abfe973dbd5-var-run-ovn\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:45 crc kubenswrapper[4870]: I1014 07:23:45.651596 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/751daf42-b5c1-4318-a15b-d5911c9d61ea-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "751daf42-b5c1-4318-a15b-d5911c9d61ea" (UID: "751daf42-b5c1-4318-a15b-d5911c9d61ea"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:23:45 crc kubenswrapper[4870]: I1014 07:23:45.668262 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement1a80-account-delete-8q7fq"] Oct 14 07:23:45 crc kubenswrapper[4870]: I1014 07:23:45.721030 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/36717039-e3e2-4b74-b612-655556122574-ovsdb-rundir\") pod \"36717039-e3e2-4b74-b612-655556122574\" (UID: \"36717039-e3e2-4b74-b612-655556122574\") " Oct 14 07:23:45 crc kubenswrapper[4870]: I1014 07:23:45.721109 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36717039-e3e2-4b74-b612-655556122574-combined-ca-bundle\") pod \"36717039-e3e2-4b74-b612-655556122574\" (UID: \"36717039-e3e2-4b74-b612-655556122574\") " Oct 14 07:23:45 crc kubenswrapper[4870]: I1014 07:23:45.721212 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2gt69\" (UniqueName: \"kubernetes.io/projected/36717039-e3e2-4b74-b612-655556122574-kube-api-access-2gt69\") pod \"36717039-e3e2-4b74-b612-655556122574\" (UID: \"36717039-e3e2-4b74-b612-655556122574\") " Oct 14 07:23:45 crc kubenswrapper[4870]: I1014 07:23:45.721279 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndbcluster-sb-etc-ovn\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"36717039-e3e2-4b74-b612-655556122574\" (UID: \"36717039-e3e2-4b74-b612-655556122574\") " Oct 14 07:23:45 crc kubenswrapper[4870]: I1014 07:23:45.721337 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/36717039-e3e2-4b74-b612-655556122574-metrics-certs-tls-certs\") pod \"36717039-e3e2-4b74-b612-655556122574\" (UID: \"36717039-e3e2-4b74-b612-655556122574\") " Oct 14 07:23:45 crc kubenswrapper[4870]: I1014 07:23:45.721444 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/36717039-e3e2-4b74-b612-655556122574-scripts\") pod \"36717039-e3e2-4b74-b612-655556122574\" (UID: \"36717039-e3e2-4b74-b612-655556122574\") " Oct 14 07:23:45 crc kubenswrapper[4870]: I1014 07:23:45.721481 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/36717039-e3e2-4b74-b612-655556122574-config\") pod \"36717039-e3e2-4b74-b612-655556122574\" (UID: \"36717039-e3e2-4b74-b612-655556122574\") " Oct 14 07:23:45 crc kubenswrapper[4870]: I1014 07:23:45.721658 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/36717039-e3e2-4b74-b612-655556122574-ovsdbserver-sb-tls-certs\") pod \"36717039-e3e2-4b74-b612-655556122574\" (UID: \"36717039-e3e2-4b74-b612-655556122574\") " Oct 14 07:23:45 crc kubenswrapper[4870]: I1014 07:23:45.722238 4870 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/751daf42-b5c1-4318-a15b-d5911c9d61ea-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:45 crc kubenswrapper[4870]: I1014 07:23:45.724931 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/36717039-e3e2-4b74-b612-655556122574-config" (OuterVolumeSpecName: "config") pod "36717039-e3e2-4b74-b612-655556122574" (UID: "36717039-e3e2-4b74-b612-655556122574"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:23:45 crc kubenswrapper[4870]: I1014 07:23:45.725007 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/36717039-e3e2-4b74-b612-655556122574-scripts" (OuterVolumeSpecName: "scripts") pod "36717039-e3e2-4b74-b612-655556122574" (UID: "36717039-e3e2-4b74-b612-655556122574"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:23:45 crc kubenswrapper[4870]: I1014 07:23:45.725339 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/36717039-e3e2-4b74-b612-655556122574-ovsdb-rundir" (OuterVolumeSpecName: "ovsdb-rundir") pod "36717039-e3e2-4b74-b612-655556122574" (UID: "36717039-e3e2-4b74-b612-655556122574"). InnerVolumeSpecName "ovsdb-rundir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:23:45 crc kubenswrapper[4870]: I1014 07:23:45.732146 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 14 07:23:45 crc kubenswrapper[4870]: I1014 07:23:45.732539 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-conductor-0" podUID="61885759-41a7-4ee8-904a-3f85bb66cd19" containerName="nova-cell1-conductor-conductor" containerID="cri-o://925b675acc2471fa8d4cffebdeddb3b79263bf8e5254078a5b2f740ce843cfd5" gracePeriod=30 Oct 14 07:23:45 crc kubenswrapper[4870]: I1014 07:23:45.743557 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-x2gj8"] Oct 14 07:23:45 crc kubenswrapper[4870]: I1014 07:23:45.744942 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0965b7e6-2aa0-4940-a130-324cfb08de5a-kube-api-access-6zpqx" (OuterVolumeSpecName: "kube-api-access-6zpqx") pod "0965b7e6-2aa0-4940-a130-324cfb08de5a" (UID: "0965b7e6-2aa0-4940-a130-324cfb08de5a"). InnerVolumeSpecName "kube-api-access-6zpqx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:23:45 crc kubenswrapper[4870]: I1014 07:23:45.751591 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-x2gj8"] Oct 14 07:23:45 crc kubenswrapper[4870]: I1014 07:23:45.760032 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-7jfpt"] Oct 14 07:23:45 crc kubenswrapper[4870]: I1014 07:23:45.767244 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 14 07:23:45 crc kubenswrapper[4870]: I1014 07:23:45.768093 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c8d0269-7501-4c96-9f9e-30bea4b0fb92-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4c8d0269-7501-4c96-9f9e-30bea4b0fb92" (UID: "4c8d0269-7501-4c96-9f9e-30bea4b0fb92"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:23:45 crc kubenswrapper[4870]: I1014 07:23:45.770609 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell0-conductor-0" podUID="bad45b30-5ba2-4062-ab14-04f923679c4e" containerName="nova-cell0-conductor-conductor" containerID="cri-o://e6478418e9645b5d3cf9a2a8e1b1a4090b34dcf697735e42cf6fb41d3708bbeb" gracePeriod=30 Oct 14 07:23:45 crc kubenswrapper[4870]: I1014 07:23:45.774641 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage11-crc" (OuterVolumeSpecName: "ovndbcluster-sb-etc-ovn") pod "36717039-e3e2-4b74-b612-655556122574" (UID: "36717039-e3e2-4b74-b612-655556122574"). InnerVolumeSpecName "local-storage11-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 14 07:23:45 crc kubenswrapper[4870]: I1014 07:23:45.776960 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/36717039-e3e2-4b74-b612-655556122574-kube-api-access-2gt69" (OuterVolumeSpecName: "kube-api-access-2gt69") pod "36717039-e3e2-4b74-b612-655556122574" (UID: "36717039-e3e2-4b74-b612-655556122574"). InnerVolumeSpecName "kube-api-access-2gt69". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:23:45 crc kubenswrapper[4870]: I1014 07:23:45.777946 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-7jfpt"] Oct 14 07:23:45 crc kubenswrapper[4870]: I1014 07:23:45.802359 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder5c98-account-delete-lhjjg"] Oct 14 07:23:45 crc kubenswrapper[4870]: I1014 07:23:45.823327 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican6547-account-delete-v5crl"] Oct 14 07:23:45 crc kubenswrapper[4870]: I1014 07:23:45.824217 4870 reconciler_common.go:293] "Volume detached for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/36717039-e3e2-4b74-b612-655556122574-ovsdb-rundir\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:45 crc kubenswrapper[4870]: I1014 07:23:45.824251 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2gt69\" (UniqueName: \"kubernetes.io/projected/36717039-e3e2-4b74-b612-655556122574-kube-api-access-2gt69\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:45 crc kubenswrapper[4870]: I1014 07:23:45.824277 4870 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" " Oct 14 07:23:45 crc kubenswrapper[4870]: I1014 07:23:45.824290 4870 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c8d0269-7501-4c96-9f9e-30bea4b0fb92-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:45 crc kubenswrapper[4870]: I1014 07:23:45.824301 4870 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/36717039-e3e2-4b74-b612-655556122574-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:45 crc kubenswrapper[4870]: I1014 07:23:45.824311 4870 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/36717039-e3e2-4b74-b612-655556122574-config\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:45 crc kubenswrapper[4870]: I1014 07:23:45.824325 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6zpqx\" (UniqueName: \"kubernetes.io/projected/0965b7e6-2aa0-4940-a130-324cfb08de5a-kube-api-access-6zpqx\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:45 crc kubenswrapper[4870]: I1014 07:23:45.848706 4870 generic.go:334] "Generic (PLEG): container finished" podID="f0b766fd-ccb9-4620-84df-8fe2f194b0da" containerID="6fb3ea4d6cd5802f3695559f939e61f05770876bd3c155dcf6040899ef93330a" exitCode=143 Oct 14 07:23:45 crc kubenswrapper[4870]: I1014 07:23:45.848769 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f0b766fd-ccb9-4620-84df-8fe2f194b0da","Type":"ContainerDied","Data":"6fb3ea4d6cd5802f3695559f939e61f05770876bd3c155dcf6040899ef93330a"} Oct 14 07:23:45 crc kubenswrapper[4870]: I1014 07:23:45.856171 4870 generic.go:334] "Generic (PLEG): container finished" podID="2bd1f41f-3f1a-4ca7-8789-429104ce2120" containerID="edba2f1cac0476dc717b946f00e8879272b5577916eb07fb08304de266852951" exitCode=0 Oct 14 07:23:45 crc kubenswrapper[4870]: I1014 07:23:45.856216 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"2bd1f41f-3f1a-4ca7-8789-429104ce2120","Type":"ContainerDied","Data":"edba2f1cac0476dc717b946f00e8879272b5577916eb07fb08304de266852951"} Oct 14 07:23:45 crc kubenswrapper[4870]: I1014 07:23:45.865519 4870 generic.go:334] "Generic (PLEG): container finished" podID="c5bb3fde-cfaf-4d09-ad31-e23fd830eb91" containerID="a77bf6d6eb57cf53a83d8f0ff81a500083d29df6ab1d56a72558d63fae3b342f" exitCode=143 Oct 14 07:23:45 crc kubenswrapper[4870]: I1014 07:23:45.865572 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-85f775bc47-d95qn" event={"ID":"c5bb3fde-cfaf-4d09-ad31-e23fd830eb91","Type":"ContainerDied","Data":"a77bf6d6eb57cf53a83d8f0ff81a500083d29df6ab1d56a72558d63fae3b342f"} Oct 14 07:23:45 crc kubenswrapper[4870]: I1014 07:23:45.908034 4870 generic.go:334] "Generic (PLEG): container finished" podID="c712b87a-cf4d-437e-abe0-1a566046ca66" containerID="6c0de18ab40ba16d1bc75ff9b157bb26252b1a8044809c3ee46ea0a31ed39019" exitCode=143 Oct 14 07:23:45 crc kubenswrapper[4870]: I1014 07:23:45.908132 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"c712b87a-cf4d-437e-abe0-1a566046ca66","Type":"ContainerDied","Data":"6c0de18ab40ba16d1bc75ff9b157bb26252b1a8044809c3ee46ea0a31ed39019"} Oct 14 07:23:45 crc kubenswrapper[4870]: I1014 07:23:45.913295 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/novacell0ce16-account-delete-bwgsq"] Oct 14 07:23:45 crc kubenswrapper[4870]: I1014 07:23:45.922895 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-rn98c" event={"ID":"503a07f8-2c05-458a-80ff-1abfe973dbd5","Type":"ContainerDied","Data":"39ecc2d2eb7b47544395be53fdac9994576cc65c656cd6d71aaa3811d2ed728b"} Oct 14 07:23:45 crc kubenswrapper[4870]: I1014 07:23:45.922987 4870 scope.go:117] "RemoveContainer" containerID="124ba16adf714b9b67c2030611bc8afc10924c09b62003e8aa1a0045f6015014" Oct 14 07:23:45 crc kubenswrapper[4870]: I1014 07:23:45.922916 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-rn98c" Oct 14 07:23:45 crc kubenswrapper[4870]: I1014 07:23:45.926830 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/751daf42-b5c1-4318-a15b-d5911c9d61ea-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "751daf42-b5c1-4318-a15b-d5911c9d61ea" (UID: "751daf42-b5c1-4318-a15b-d5911c9d61ea"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:23:45 crc kubenswrapper[4870]: I1014 07:23:45.927141 4870 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/751daf42-b5c1-4318-a15b-d5911c9d61ea-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:45 crc kubenswrapper[4870]: I1014 07:23:45.931519 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance1aaa-account-delete-nct6c"] Oct 14 07:23:45 crc kubenswrapper[4870]: I1014 07:23:45.940367 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder5c98-account-delete-lhjjg" event={"ID":"de479f5d-22a2-4f7e-8034-cffefc511b53","Type":"ContainerStarted","Data":"8b623fe0ae243b386c95275cd06b0fe710ca79ddd29998a1c3cb917add453e2e"} Oct 14 07:23:45 crc kubenswrapper[4870]: I1014 07:23:45.941899 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 14 07:23:45 crc kubenswrapper[4870]: I1014 07:23:45.962309 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/novaapie66c-account-delete-ftd24"] Oct 14 07:23:45 crc kubenswrapper[4870]: I1014 07:23:45.963214 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7cf4d694c5-wjxfs" Oct 14 07:23:45 crc kubenswrapper[4870]: I1014 07:23:45.963287 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cf4d694c5-wjxfs" event={"ID":"751daf42-b5c1-4318-a15b-d5911c9d61ea","Type":"ContainerDied","Data":"9f72086a1f5451771a6f36acdca98d32c056502ec4a9d0b1a7f2a1edc601460b"} Oct 14 07:23:45 crc kubenswrapper[4870]: W1014 07:23:45.971225 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda9d8a25d_2f1c_483d_bf0d_76a2bf30f137.slice/crio-a646110e9a961f59dfe446e486ce2da0f818db4b1ed54f19f0b9124ac62a80df WatchSource:0}: Error finding container a646110e9a961f59dfe446e486ce2da0f818db4b1ed54f19f0b9124ac62a80df: Status 404 returned error can't find the container with id a646110e9a961f59dfe446e486ce2da0f818db4b1ed54f19f0b9124ac62a80df Oct 14 07:23:45 crc kubenswrapper[4870]: I1014 07:23:45.974006 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement1a80-account-delete-8q7fq" event={"ID":"7cbdbbdd-ea4f-4d14-8f3c-4784bda45874","Type":"ContainerStarted","Data":"bca8553e8397dc43e04b16f525a283963e83bdf774088be23914b72c95434ac0"} Oct 14 07:23:45 crc kubenswrapper[4870]: I1014 07:23:45.989623 4870 generic.go:334] "Generic (PLEG): container finished" podID="0ba0e2ed-57d0-4535-883e-10b9b985b3d3" containerID="f84e5ed457b1301ae67b84003d43c8e27a59446623074e29e6e2859280f620fd" exitCode=0 Oct 14 07:23:45 crc kubenswrapper[4870]: I1014 07:23:45.989690 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-546b769ccc-fdhsk" event={"ID":"0ba0e2ed-57d0-4535-883e-10b9b985b3d3","Type":"ContainerDied","Data":"f84e5ed457b1301ae67b84003d43c8e27a59446623074e29e6e2859280f620fd"} Oct 14 07:23:46 crc kubenswrapper[4870]: I1014 07:23:46.001980 4870 generic.go:334] "Generic (PLEG): container finished" podID="0a2e4c80-d44f-40c0-9c04-74e9ce98ae36" containerID="bc61afdd96aedc563124f07cd981291ea8c01a0a034c78fba2e57e484564a0ee" exitCode=143 Oct 14 07:23:46 crc kubenswrapper[4870]: I1014 07:23:46.002076 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-57d8bd87b6-vnpct" event={"ID":"0a2e4c80-d44f-40c0-9c04-74e9ce98ae36","Type":"ContainerDied","Data":"bc61afdd96aedc563124f07cd981291ea8c01a0a034c78fba2e57e484564a0ee"} Oct 14 07:23:46 crc kubenswrapper[4870]: I1014 07:23:46.021534 4870 generic.go:334] "Generic (PLEG): container finished" podID="ff28d060-b280-4b89-bea1-d10aa367a3cc" containerID="6d95049e97ec6fbbfa6d28e3e543ed1f7fc74569f83d0862a392737896e3463d" exitCode=143 Oct 14 07:23:46 crc kubenswrapper[4870]: I1014 07:23:46.021588 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ff28d060-b280-4b89-bea1-d10aa367a3cc","Type":"ContainerDied","Data":"6d95049e97ec6fbbfa6d28e3e543ed1f7fc74569f83d0862a392737896e3463d"} Oct 14 07:23:46 crc kubenswrapper[4870]: E1014 07:23:46.028286 4870 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Oct 14 07:23:46 crc kubenswrapper[4870]: E1014 07:23:46.028334 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5212ba20-83c3-41ff-a3d9-5a315f009c1b-config-data podName:5212ba20-83c3-41ff-a3d9-5a315f009c1b nodeName:}" failed. No retries permitted until 2025-10-14 07:23:50.028321425 +0000 UTC m=+1365.725681796 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/5212ba20-83c3-41ff-a3d9-5a315f009c1b-config-data") pod "rabbitmq-cell1-server-0" (UID: "5212ba20-83c3-41ff-a3d9-5a315f009c1b") : configmap "rabbitmq-cell1-config-data" not found Oct 14 07:23:46 crc kubenswrapper[4870]: I1014 07:23:46.035222 4870 generic.go:334] "Generic (PLEG): container finished" podID="cae5b655-5d7c-4063-90ad-4d9913792d5b" containerID="a507ea103a81fde95c96ade7a313c18b4faefbe53e02c99e5cd596e0c8dcb53e" exitCode=143 Oct 14 07:23:46 crc kubenswrapper[4870]: I1014 07:23:46.037327 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-79cc757548-zxf4g" event={"ID":"cae5b655-5d7c-4063-90ad-4d9913792d5b","Type":"ContainerDied","Data":"a507ea103a81fde95c96ade7a313c18b4faefbe53e02c99e5cd596e0c8dcb53e"} Oct 14 07:23:46 crc kubenswrapper[4870]: I1014 07:23:46.047944 4870 generic.go:334] "Generic (PLEG): container finished" podID="91a2e1de-104f-4e14-9192-c6587ed13be1" containerID="6f14317c64b4cb248d4254d308d86fadbdb28d2daee87abbf75414c219c2c477" exitCode=0 Oct 14 07:23:46 crc kubenswrapper[4870]: I1014 07:23:46.048013 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-6482n" event={"ID":"91a2e1de-104f-4e14-9192-c6587ed13be1","Type":"ContainerDied","Data":"6f14317c64b4cb248d4254d308d86fadbdb28d2daee87abbf75414c219c2c477"} Oct 14 07:23:46 crc kubenswrapper[4870]: I1014 07:23:46.056812 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_36717039-e3e2-4b74-b612-655556122574/ovsdbserver-sb/0.log" Oct 14 07:23:46 crc kubenswrapper[4870]: I1014 07:23:46.056883 4870 generic.go:334] "Generic (PLEG): container finished" podID="36717039-e3e2-4b74-b612-655556122574" containerID="afc051c5e25bf94e2cbc69eabd5bf3dc5066c8c7364b3f767d2bb33960df9191" exitCode=143 Oct 14 07:23:46 crc kubenswrapper[4870]: I1014 07:23:46.057007 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"36717039-e3e2-4b74-b612-655556122574","Type":"ContainerDied","Data":"afc051c5e25bf94e2cbc69eabd5bf3dc5066c8c7364b3f767d2bb33960df9191"} Oct 14 07:23:46 crc kubenswrapper[4870]: I1014 07:23:46.057060 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Oct 14 07:23:46 crc kubenswrapper[4870]: I1014 07:23:46.057558 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"36717039-e3e2-4b74-b612-655556122574","Type":"ContainerDied","Data":"b62fd674ae95524d37f10dfead22b74ed9112cfa1180b36f995c9dd46c127b3e"} Oct 14 07:23:46 crc kubenswrapper[4870]: I1014 07:23:46.064676 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/751daf42-b5c1-4318-a15b-d5911c9d61ea-config" (OuterVolumeSpecName: "config") pod "751daf42-b5c1-4318-a15b-d5911c9d61ea" (UID: "751daf42-b5c1-4318-a15b-d5911c9d61ea"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:23:46 crc kubenswrapper[4870]: I1014 07:23:46.106259 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/503a07f8-2c05-458a-80ff-1abfe973dbd5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "503a07f8-2c05-458a-80ff-1abfe973dbd5" (UID: "503a07f8-2c05-458a-80ff-1abfe973dbd5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:23:46 crc kubenswrapper[4870]: I1014 07:23:46.106775 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/751daf42-b5c1-4318-a15b-d5911c9d61ea-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "751daf42-b5c1-4318-a15b-d5911c9d61ea" (UID: "751daf42-b5c1-4318-a15b-d5911c9d61ea"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:23:46 crc kubenswrapper[4870]: I1014 07:23:46.128855 4870 generic.go:334] "Generic (PLEG): container finished" podID="7bc76b72-8679-4b33-8966-0ee70c1788e5" containerID="a1f9a0d181ab7f3f80bee6c1008e3b393dd082ba31b2614d29e2a4b0d5534a46" exitCode=0 Oct 14 07:23:46 crc kubenswrapper[4870]: I1014 07:23:46.128960 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-9dqlp" Oct 14 07:23:46 crc kubenswrapper[4870]: I1014 07:23:46.129810 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"7bc76b72-8679-4b33-8966-0ee70c1788e5","Type":"ContainerDied","Data":"a1f9a0d181ab7f3f80bee6c1008e3b393dd082ba31b2614d29e2a4b0d5534a46"} Oct 14 07:23:46 crc kubenswrapper[4870]: I1014 07:23:46.130745 4870 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/751daf42-b5c1-4318-a15b-d5911c9d61ea-config\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:46 crc kubenswrapper[4870]: I1014 07:23:46.130759 4870 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/503a07f8-2c05-458a-80ff-1abfe973dbd5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:46 crc kubenswrapper[4870]: I1014 07:23:46.130770 4870 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/751daf42-b5c1-4318-a15b-d5911c9d61ea-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:46 crc kubenswrapper[4870]: I1014 07:23:46.192274 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/751daf42-b5c1-4318-a15b-d5911c9d61ea-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "751daf42-b5c1-4318-a15b-d5911c9d61ea" (UID: "751daf42-b5c1-4318-a15b-d5911c9d61ea"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:23:46 crc kubenswrapper[4870]: I1014 07:23:46.214302 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-proxy-5485b69bc9-ftmmq"] Oct 14 07:23:46 crc kubenswrapper[4870]: I1014 07:23:46.214613 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-proxy-5485b69bc9-ftmmq" podUID="3927fb78-4f30-4371-8fbb-62fd276c0460" containerName="proxy-httpd" containerID="cri-o://58c55f50ceab03915ab9a353ebd2b0d046666be23591bd23a6a60026b468abe2" gracePeriod=30 Oct 14 07:23:46 crc kubenswrapper[4870]: I1014 07:23:46.215027 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-proxy-5485b69bc9-ftmmq" podUID="3927fb78-4f30-4371-8fbb-62fd276c0460" containerName="proxy-server" containerID="cri-o://009e0a5765edbfd4635dd4efcc0b112f3d5e0691352f5b4e2d147aac46326707" gracePeriod=30 Oct 14 07:23:46 crc kubenswrapper[4870]: I1014 07:23:46.218790 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/36717039-e3e2-4b74-b612-655556122574-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "36717039-e3e2-4b74-b612-655556122574" (UID: "36717039-e3e2-4b74-b612-655556122574"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:23:46 crc kubenswrapper[4870]: I1014 07:23:46.229328 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0965b7e6-2aa0-4940-a130-324cfb08de5a-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "0965b7e6-2aa0-4940-a130-324cfb08de5a" (UID: "0965b7e6-2aa0-4940-a130-324cfb08de5a"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:23:46 crc kubenswrapper[4870]: I1014 07:23:46.233016 4870 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/751daf42-b5c1-4318-a15b-d5911c9d61ea-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:46 crc kubenswrapper[4870]: I1014 07:23:46.233067 4870 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36717039-e3e2-4b74-b612-655556122574-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:46 crc kubenswrapper[4870]: I1014 07:23:46.233078 4870 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/0965b7e6-2aa0-4940-a130-324cfb08de5a-openstack-config\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:46 crc kubenswrapper[4870]: I1014 07:23:46.248103 4870 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage11-crc" (UniqueName: "kubernetes.io/local-volume/local-storage11-crc") on node "crc" Oct 14 07:23:46 crc kubenswrapper[4870]: I1014 07:23:46.264446 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0965b7e6-2aa0-4940-a130-324cfb08de5a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0965b7e6-2aa0-4940-a130-324cfb08de5a" (UID: "0965b7e6-2aa0-4940-a130-324cfb08de5a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:23:46 crc kubenswrapper[4870]: I1014 07:23:46.334944 4870 reconciler_common.go:293] "Volume detached for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:46 crc kubenswrapper[4870]: I1014 07:23:46.334985 4870 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0965b7e6-2aa0-4940-a130-324cfb08de5a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:46 crc kubenswrapper[4870]: E1014 07:23:46.335076 4870 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Oct 14 07:23:46 crc kubenswrapper[4870]: E1014 07:23:46.335140 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/da8a9731-f917-462f-9932-b37b6abb9a64-config-data podName:da8a9731-f917-462f-9932-b37b6abb9a64 nodeName:}" failed. No retries permitted until 2025-10-14 07:23:50.335127142 +0000 UTC m=+1366.032487513 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/da8a9731-f917-462f-9932-b37b6abb9a64-config-data") pod "rabbitmq-server-0" (UID: "da8a9731-f917-462f-9932-b37b6abb9a64") : configmap "rabbitmq-config-data" not found Oct 14 07:23:46 crc kubenswrapper[4870]: I1014 07:23:46.366295 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c8d0269-7501-4c96-9f9e-30bea4b0fb92-metrics-certs-tls-certs" (OuterVolumeSpecName: "metrics-certs-tls-certs") pod "4c8d0269-7501-4c96-9f9e-30bea4b0fb92" (UID: "4c8d0269-7501-4c96-9f9e-30bea4b0fb92"). InnerVolumeSpecName "metrics-certs-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:23:46 crc kubenswrapper[4870]: I1014 07:23:46.394889 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/36717039-e3e2-4b74-b612-655556122574-metrics-certs-tls-certs" (OuterVolumeSpecName: "metrics-certs-tls-certs") pod "36717039-e3e2-4b74-b612-655556122574" (UID: "36717039-e3e2-4b74-b612-655556122574"). InnerVolumeSpecName "metrics-certs-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:23:46 crc kubenswrapper[4870]: I1014 07:23:46.437409 4870 reconciler_common.go:293] "Volume detached for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/36717039-e3e2-4b74-b612-655556122574-metrics-certs-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:46 crc kubenswrapper[4870]: I1014 07:23:46.437833 4870 reconciler_common.go:293] "Volume detached for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/4c8d0269-7501-4c96-9f9e-30bea4b0fb92-metrics-certs-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:46 crc kubenswrapper[4870]: I1014 07:23:46.437522 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/36717039-e3e2-4b74-b612-655556122574-ovsdbserver-sb-tls-certs" (OuterVolumeSpecName: "ovsdbserver-sb-tls-certs") pod "36717039-e3e2-4b74-b612-655556122574" (UID: "36717039-e3e2-4b74-b612-655556122574"). InnerVolumeSpecName "ovsdbserver-sb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:23:46 crc kubenswrapper[4870]: I1014 07:23:46.445412 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0965b7e6-2aa0-4940-a130-324cfb08de5a-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "0965b7e6-2aa0-4940-a130-324cfb08de5a" (UID: "0965b7e6-2aa0-4940-a130-324cfb08de5a"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:23:46 crc kubenswrapper[4870]: I1014 07:23:46.453638 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/503a07f8-2c05-458a-80ff-1abfe973dbd5-ovn-controller-tls-certs" (OuterVolumeSpecName: "ovn-controller-tls-certs") pod "503a07f8-2c05-458a-80ff-1abfe973dbd5" (UID: "503a07f8-2c05-458a-80ff-1abfe973dbd5"). InnerVolumeSpecName "ovn-controller-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:23:46 crc kubenswrapper[4870]: I1014 07:23:46.539428 4870 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/36717039-e3e2-4b74-b612-655556122574-ovsdbserver-sb-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:46 crc kubenswrapper[4870]: I1014 07:23:46.539477 4870 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/0965b7e6-2aa0-4940-a130-324cfb08de5a-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:46 crc kubenswrapper[4870]: I1014 07:23:46.539493 4870 reconciler_common.go:293] "Volume detached for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/503a07f8-2c05-458a-80ff-1abfe973dbd5-ovn-controller-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:46 crc kubenswrapper[4870]: I1014 07:23:46.602382 4870 scope.go:117] "RemoveContainer" containerID="044ea035cd6094777207ef8f180d32297801ac2dd26be47e987ae4b7318523d6" Oct 14 07:23:46 crc kubenswrapper[4870]: E1014 07:23:46.638750 4870 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 6f14317c64b4cb248d4254d308d86fadbdb28d2daee87abbf75414c219c2c477 is running failed: container process not found" containerID="6f14317c64b4cb248d4254d308d86fadbdb28d2daee87abbf75414c219c2c477" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 14 07:23:46 crc kubenswrapper[4870]: E1014 07:23:46.639319 4870 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 6f14317c64b4cb248d4254d308d86fadbdb28d2daee87abbf75414c219c2c477 is running failed: container process not found" containerID="6f14317c64b4cb248d4254d308d86fadbdb28d2daee87abbf75414c219c2c477" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 14 07:23:46 crc kubenswrapper[4870]: E1014 07:23:46.639856 4870 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 6f14317c64b4cb248d4254d308d86fadbdb28d2daee87abbf75414c219c2c477 is running failed: container process not found" containerID="6f14317c64b4cb248d4254d308d86fadbdb28d2daee87abbf75414c219c2c477" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 14 07:23:46 crc kubenswrapper[4870]: E1014 07:23:46.639900 4870 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 6f14317c64b4cb248d4254d308d86fadbdb28d2daee87abbf75414c219c2c477 is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-6482n" podUID="91a2e1de-104f-4e14-9192-c6587ed13be1" containerName="ovsdb-server" Oct 14 07:23:46 crc kubenswrapper[4870]: E1014 07:23:46.656832 4870 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="72ead03af69bb2d536002fe16f85a88974968529ec05c0e9e78254a3f796b690" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 14 07:23:46 crc kubenswrapper[4870]: E1014 07:23:46.667575 4870 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="925b675acc2471fa8d4cffebdeddb3b79263bf8e5254078a5b2f740ce843cfd5" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 14 07:23:46 crc kubenswrapper[4870]: E1014 07:23:46.677876 4870 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="925b675acc2471fa8d4cffebdeddb3b79263bf8e5254078a5b2f740ce843cfd5" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 14 07:23:46 crc kubenswrapper[4870]: E1014 07:23:46.678386 4870 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="72ead03af69bb2d536002fe16f85a88974968529ec05c0e9e78254a3f796b690" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 14 07:23:46 crc kubenswrapper[4870]: E1014 07:23:46.685509 4870 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="925b675acc2471fa8d4cffebdeddb3b79263bf8e5254078a5b2f740ce843cfd5" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 14 07:23:46 crc kubenswrapper[4870]: E1014 07:23:46.685586 4870 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-cell1-conductor-0" podUID="61885759-41a7-4ee8-904a-3f85bb66cd19" containerName="nova-cell1-conductor-conductor" Oct 14 07:23:46 crc kubenswrapper[4870]: I1014 07:23:46.685822 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7cf4d694c5-wjxfs"] Oct 14 07:23:46 crc kubenswrapper[4870]: E1014 07:23:46.702388 4870 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="72ead03af69bb2d536002fe16f85a88974968529ec05c0e9e78254a3f796b690" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 14 07:23:46 crc kubenswrapper[4870]: I1014 07:23:46.702507 4870 scope.go:117] "RemoveContainer" containerID="e7f985e91ecb8ba5c0b4601a782047823bc6426fd3f278547d0ef69704ee2fcc" Oct 14 07:23:46 crc kubenswrapper[4870]: E1014 07:23:46.702553 4870 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-6482n" podUID="91a2e1de-104f-4e14-9192-c6587ed13be1" containerName="ovs-vswitchd" Oct 14 07:23:46 crc kubenswrapper[4870]: I1014 07:23:46.731764 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7cf4d694c5-wjxfs"] Oct 14 07:23:46 crc kubenswrapper[4870]: I1014 07:23:46.794330 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-metrics-9dqlp"] Oct 14 07:23:46 crc kubenswrapper[4870]: I1014 07:23:46.812539 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-metrics-9dqlp"] Oct 14 07:23:46 crc kubenswrapper[4870]: I1014 07:23:46.813010 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-rn98c"] Oct 14 07:23:46 crc kubenswrapper[4870]: I1014 07:23:46.817494 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-rn98c"] Oct 14 07:23:46 crc kubenswrapper[4870]: I1014 07:23:46.822801 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 14 07:23:46 crc kubenswrapper[4870]: I1014 07:23:46.827609 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 14 07:23:46 crc kubenswrapper[4870]: I1014 07:23:46.948654 4870 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/neutron-546b769ccc-fdhsk" podUID="0ba0e2ed-57d0-4535-883e-10b9b985b3d3" containerName="neutron-httpd" probeResult="failure" output="Get \"https://10.217.0.158:9696/\": dial tcp 10.217.0.158:9696: connect: connection refused" Oct 14 07:23:47 crc kubenswrapper[4870]: I1014 07:23:47.055894 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0965b7e6-2aa0-4940-a130-324cfb08de5a" path="/var/lib/kubelet/pods/0965b7e6-2aa0-4940-a130-324cfb08de5a/volumes" Oct 14 07:23:47 crc kubenswrapper[4870]: I1014 07:23:47.056583 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="36717039-e3e2-4b74-b612-655556122574" path="/var/lib/kubelet/pods/36717039-e3e2-4b74-b612-655556122574/volumes" Oct 14 07:23:47 crc kubenswrapper[4870]: I1014 07:23:47.057133 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3b265c7a-3909-4fd6-9981-c1723fbc1e02" path="/var/lib/kubelet/pods/3b265c7a-3909-4fd6-9981-c1723fbc1e02/volumes" Oct 14 07:23:47 crc kubenswrapper[4870]: I1014 07:23:47.058770 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ea2c25b-4ae5-473c-9480-ec8444bbee56" path="/var/lib/kubelet/pods/3ea2c25b-4ae5-473c-9480-ec8444bbee56/volumes" Oct 14 07:23:47 crc kubenswrapper[4870]: I1014 07:23:47.061171 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4c8d0269-7501-4c96-9f9e-30bea4b0fb92" path="/var/lib/kubelet/pods/4c8d0269-7501-4c96-9f9e-30bea4b0fb92/volumes" Oct 14 07:23:47 crc kubenswrapper[4870]: I1014 07:23:47.063845 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="503a07f8-2c05-458a-80ff-1abfe973dbd5" path="/var/lib/kubelet/pods/503a07f8-2c05-458a-80ff-1abfe973dbd5/volumes" Oct 14 07:23:47 crc kubenswrapper[4870]: I1014 07:23:47.065141 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="52ebf07b-1f8d-4005-b8bc-14e250c921d4" path="/var/lib/kubelet/pods/52ebf07b-1f8d-4005-b8bc-14e250c921d4/volumes" Oct 14 07:23:47 crc kubenswrapper[4870]: I1014 07:23:47.066494 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="751daf42-b5c1-4318-a15b-d5911c9d61ea" path="/var/lib/kubelet/pods/751daf42-b5c1-4318-a15b-d5911c9d61ea/volumes" Oct 14 07:23:47 crc kubenswrapper[4870]: I1014 07:23:47.068169 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="93414260-afe1-4af0-8673-d51ec4591776" path="/var/lib/kubelet/pods/93414260-afe1-4af0-8673-d51ec4591776/volumes" Oct 14 07:23:47 crc kubenswrapper[4870]: I1014 07:23:47.068868 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="93d44382-d93e-4db6-b267-181f5e02682b" path="/var/lib/kubelet/pods/93d44382-d93e-4db6-b267-181f5e02682b/volumes" Oct 14 07:23:47 crc kubenswrapper[4870]: I1014 07:23:47.069471 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a4ae5421-6680-4e79-ac45-578fdbbc70f2" path="/var/lib/kubelet/pods/a4ae5421-6680-4e79-ac45-578fdbbc70f2/volumes" Oct 14 07:23:47 crc kubenswrapper[4870]: I1014 07:23:47.070100 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bdb3c945-edde-4d43-b3a5-234aaabf00d5" path="/var/lib/kubelet/pods/bdb3c945-edde-4d43-b3a5-234aaabf00d5/volumes" Oct 14 07:23:47 crc kubenswrapper[4870]: I1014 07:23:47.078508 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf415c18-d584-4cf1-ae86-3bebdc74cad1" path="/var/lib/kubelet/pods/bf415c18-d584-4cf1-ae86-3bebdc74cad1/volumes" Oct 14 07:23:47 crc kubenswrapper[4870]: I1014 07:23:47.155710 4870 generic.go:334] "Generic (PLEG): container finished" podID="8abede26-36e8-46ec-a5c4-6cd53319133a" containerID="45200242c420e7ff44ca9c10b9cff44cc8bb4c9294423cc3e3e0ad235691f6b4" exitCode=0 Oct 14 07:23:47 crc kubenswrapper[4870]: I1014 07:23:47.155814 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"8abede26-36e8-46ec-a5c4-6cd53319133a","Type":"ContainerDied","Data":"45200242c420e7ff44ca9c10b9cff44cc8bb4c9294423cc3e3e0ad235691f6b4"} Oct 14 07:23:47 crc kubenswrapper[4870]: I1014 07:23:47.155841 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"8abede26-36e8-46ec-a5c4-6cd53319133a","Type":"ContainerDied","Data":"5a4ce260cbfb825ad1358d2122d60517f96cbcdbd83613ef2c1f69ea0ec0bf41"} Oct 14 07:23:47 crc kubenswrapper[4870]: I1014 07:23:47.155885 4870 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5a4ce260cbfb825ad1358d2122d60517f96cbcdbd83613ef2c1f69ea0ec0bf41" Oct 14 07:23:47 crc kubenswrapper[4870]: I1014 07:23:47.157439 4870 generic.go:334] "Generic (PLEG): container finished" podID="7cbdbbdd-ea4f-4d14-8f3c-4784bda45874" containerID="51cc7a5037c45783ceb5efa6e804746e76684dbe52ba7dd329814c5da8fb28db" exitCode=0 Oct 14 07:23:47 crc kubenswrapper[4870]: I1014 07:23:47.157501 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement1a80-account-delete-8q7fq" event={"ID":"7cbdbbdd-ea4f-4d14-8f3c-4784bda45874","Type":"ContainerDied","Data":"51cc7a5037c45783ceb5efa6e804746e76684dbe52ba7dd329814c5da8fb28db"} Oct 14 07:23:47 crc kubenswrapper[4870]: I1014 07:23:47.191421 4870 generic.go:334] "Generic (PLEG): container finished" podID="2abee394-ffb3-4dba-8d09-19c1cb476595" containerID="769de1c723d5f9ddc1f6d062434e4b09d67bad6377a8cd0e80434875d4f6a655" exitCode=0 Oct 14 07:23:47 crc kubenswrapper[4870]: I1014 07:23:47.192385 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican6547-account-delete-v5crl" event={"ID":"2abee394-ffb3-4dba-8d09-19c1cb476595","Type":"ContainerDied","Data":"769de1c723d5f9ddc1f6d062434e4b09d67bad6377a8cd0e80434875d4f6a655"} Oct 14 07:23:47 crc kubenswrapper[4870]: I1014 07:23:47.192455 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican6547-account-delete-v5crl" event={"ID":"2abee394-ffb3-4dba-8d09-19c1cb476595","Type":"ContainerStarted","Data":"dc92ede877d28eaf350cc8faaf42ce060c07d06e071039d2b9a83dbb16901f1a"} Oct 14 07:23:47 crc kubenswrapper[4870]: I1014 07:23:47.210726 4870 generic.go:334] "Generic (PLEG): container finished" podID="de479f5d-22a2-4f7e-8034-cffefc511b53" containerID="a7dc7bd1a8ffe2b7e76587cb479c6f7ae957cc2c7d3a7fcfe34ba24bbd24c682" exitCode=0 Oct 14 07:23:47 crc kubenswrapper[4870]: I1014 07:23:47.210855 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder5c98-account-delete-lhjjg" event={"ID":"de479f5d-22a2-4f7e-8034-cffefc511b53","Type":"ContainerDied","Data":"a7dc7bd1a8ffe2b7e76587cb479c6f7ae957cc2c7d3a7fcfe34ba24bbd24c682"} Oct 14 07:23:47 crc kubenswrapper[4870]: I1014 07:23:47.232729 4870 generic.go:334] "Generic (PLEG): container finished" podID="4d74ef50-1861-4d70-b781-f92d5f4b35ba" containerID="9389d35044ef25b0e127f63d901f00d3d48e58573671b63fa03de1238628ba21" exitCode=0 Oct 14 07:23:47 crc kubenswrapper[4870]: I1014 07:23:47.232846 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"4d74ef50-1861-4d70-b781-f92d5f4b35ba","Type":"ContainerDied","Data":"9389d35044ef25b0e127f63d901f00d3d48e58573671b63fa03de1238628ba21"} Oct 14 07:23:47 crc kubenswrapper[4870]: I1014 07:23:47.232876 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"4d74ef50-1861-4d70-b781-f92d5f4b35ba","Type":"ContainerDied","Data":"474bb87795c2229a29716f4b23006219a79aa8b329f0418dda86379779caeef2"} Oct 14 07:23:47 crc kubenswrapper[4870]: I1014 07:23:47.232889 4870 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="474bb87795c2229a29716f4b23006219a79aa8b329f0418dda86379779caeef2" Oct 14 07:23:47 crc kubenswrapper[4870]: I1014 07:23:47.272552 4870 generic.go:334] "Generic (PLEG): container finished" podID="a6f02e4f-8dca-4b97-91e9-ee3368e5e44b" containerID="649fea192c7edfde335ef6265aaeb17d682a7f0427cec9179859e9c93e49f399" exitCode=0 Oct 14 07:23:47 crc kubenswrapper[4870]: I1014 07:23:47.272711 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novacell0ce16-account-delete-bwgsq" event={"ID":"a6f02e4f-8dca-4b97-91e9-ee3368e5e44b","Type":"ContainerDied","Data":"649fea192c7edfde335ef6265aaeb17d682a7f0427cec9179859e9c93e49f399"} Oct 14 07:23:47 crc kubenswrapper[4870]: I1014 07:23:47.272743 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novacell0ce16-account-delete-bwgsq" event={"ID":"a6f02e4f-8dca-4b97-91e9-ee3368e5e44b","Type":"ContainerStarted","Data":"1776e2a9afb3ac3237ca85fa6c702e235efdc65ae828037d00005680033973a2"} Oct 14 07:23:47 crc kubenswrapper[4870]: I1014 07:23:47.313045 4870 generic.go:334] "Generic (PLEG): container finished" podID="61885759-41a7-4ee8-904a-3f85bb66cd19" containerID="925b675acc2471fa8d4cffebdeddb3b79263bf8e5254078a5b2f740ce843cfd5" exitCode=0 Oct 14 07:23:47 crc kubenswrapper[4870]: I1014 07:23:47.314267 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 14 07:23:47 crc kubenswrapper[4870]: I1014 07:23:47.314309 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"61885759-41a7-4ee8-904a-3f85bb66cd19","Type":"ContainerDied","Data":"925b675acc2471fa8d4cffebdeddb3b79263bf8e5254078a5b2f740ce843cfd5"} Oct 14 07:23:47 crc kubenswrapper[4870]: I1014 07:23:47.314736 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="524b0b2a-77a2-4c05-be39-c8b97666ab11" containerName="ceilometer-central-agent" containerID="cri-o://f24a75c62a58aefe115cf6dc0ab70e01b811b66e1369401528bc3134ca7d1cc4" gracePeriod=30 Oct 14 07:23:47 crc kubenswrapper[4870]: I1014 07:23:47.315708 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="524b0b2a-77a2-4c05-be39-c8b97666ab11" containerName="sg-core" containerID="cri-o://a7f06f2b8b7523a5efb4f46ab915a339acb0f160d1f9753afa2035de3c545c6d" gracePeriod=30 Oct 14 07:23:47 crc kubenswrapper[4870]: I1014 07:23:47.315788 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="524b0b2a-77a2-4c05-be39-c8b97666ab11" containerName="proxy-httpd" containerID="cri-o://018378adb14febcf4afcbaac2d9f70a84cc9e7f96675c9b85ee751bb465d9632" gracePeriod=30 Oct 14 07:23:47 crc kubenswrapper[4870]: I1014 07:23:47.315874 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="524b0b2a-77a2-4c05-be39-c8b97666ab11" containerName="ceilometer-notification-agent" containerID="cri-o://92059a83ebd09e581c802c9d18a6cc53c68a57ad45e7c0b7e480b8f24d3f2bd8" gracePeriod=30 Oct 14 07:23:47 crc kubenswrapper[4870]: I1014 07:23:47.372038 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 14 07:23:47 crc kubenswrapper[4870]: I1014 07:23:47.372282 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="c603f7b5-015f-4c2e-9144-b3e8349c3da3" containerName="kube-state-metrics" containerID="cri-o://c79902c86d1a79331c9994ac2e40ba61e03d57293be1023f0938cc002e01194a" gracePeriod=30 Oct 14 07:23:47 crc kubenswrapper[4870]: I1014 07:23:47.405627 4870 scope.go:117] "RemoveContainer" containerID="739fd4dfae2ee47a08c26ba0aab10d4ad5944e6d3c334d62f3c9e4aaf95c2394" Oct 14 07:23:47 crc kubenswrapper[4870]: I1014 07:23:47.417099 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Oct 14 07:23:47 crc kubenswrapper[4870]: I1014 07:23:47.429956 4870 generic.go:334] "Generic (PLEG): container finished" podID="3927fb78-4f30-4371-8fbb-62fd276c0460" containerID="009e0a5765edbfd4635dd4efcc0b112f3d5e0691352f5b4e2d147aac46326707" exitCode=0 Oct 14 07:23:47 crc kubenswrapper[4870]: I1014 07:23:47.429983 4870 generic.go:334] "Generic (PLEG): container finished" podID="3927fb78-4f30-4371-8fbb-62fd276c0460" containerID="58c55f50ceab03915ab9a353ebd2b0d046666be23591bd23a6a60026b468abe2" exitCode=0 Oct 14 07:23:47 crc kubenswrapper[4870]: I1014 07:23:47.430030 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-5485b69bc9-ftmmq" event={"ID":"3927fb78-4f30-4371-8fbb-62fd276c0460","Type":"ContainerDied","Data":"009e0a5765edbfd4635dd4efcc0b112f3d5e0691352f5b4e2d147aac46326707"} Oct 14 07:23:47 crc kubenswrapper[4870]: I1014 07:23:47.430055 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-5485b69bc9-ftmmq" event={"ID":"3927fb78-4f30-4371-8fbb-62fd276c0460","Type":"ContainerDied","Data":"58c55f50ceab03915ab9a353ebd2b0d046666be23591bd23a6a60026b468abe2"} Oct 14 07:23:47 crc kubenswrapper[4870]: I1014 07:23:47.430156 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 14 07:23:47 crc kubenswrapper[4870]: I1014 07:23:47.431644 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/memcached-0"] Oct 14 07:23:47 crc kubenswrapper[4870]: I1014 07:23:47.431862 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/memcached-0" podUID="3bd1e649-3d09-4cda-921f-ac9a84a5066e" containerName="memcached" containerID="cri-o://cd6dabf959a5bb827e348f8580f0325c6287615f71bc7279bb2e2744f5cf4f7c" gracePeriod=30 Oct 14 07:23:47 crc kubenswrapper[4870]: I1014 07:23:47.474770 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novaapie66c-account-delete-ftd24" event={"ID":"208754ea-7cc0-4222-a186-9ba01b274add","Type":"ContainerStarted","Data":"719f5c94f8b4f323a4cf25aa23aedc41ecbda25519cbe0736cf69b37d5ab0c75"} Oct 14 07:23:47 crc kubenswrapper[4870]: I1014 07:23:47.474903 4870 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="da8a9731-f917-462f-9932-b37b6abb9a64" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.101:5671: connect: connection refused" Oct 14 07:23:47 crc kubenswrapper[4870]: I1014 07:23:47.488707 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-mwp7r"] Oct 14 07:23:47 crc kubenswrapper[4870]: I1014 07:23:47.492115 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/4d74ef50-1861-4d70-b781-f92d5f4b35ba-secrets\") pod \"4d74ef50-1861-4d70-b781-f92d5f4b35ba\" (UID: \"4d74ef50-1861-4d70-b781-f92d5f4b35ba\") " Oct 14 07:23:47 crc kubenswrapper[4870]: I1014 07:23:47.492159 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/4d74ef50-1861-4d70-b781-f92d5f4b35ba-kolla-config\") pod \"4d74ef50-1861-4d70-b781-f92d5f4b35ba\" (UID: \"4d74ef50-1861-4d70-b781-f92d5f4b35ba\") " Oct 14 07:23:47 crc kubenswrapper[4870]: I1014 07:23:47.492181 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mysql-db\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"4d74ef50-1861-4d70-b781-f92d5f4b35ba\" (UID: \"4d74ef50-1861-4d70-b781-f92d5f4b35ba\") " Oct 14 07:23:47 crc kubenswrapper[4870]: I1014 07:23:47.492228 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/4d74ef50-1861-4d70-b781-f92d5f4b35ba-galera-tls-certs\") pod \"4d74ef50-1861-4d70-b781-f92d5f4b35ba\" (UID: \"4d74ef50-1861-4d70-b781-f92d5f4b35ba\") " Oct 14 07:23:47 crc kubenswrapper[4870]: I1014 07:23:47.492245 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8mr4z\" (UniqueName: \"kubernetes.io/projected/4d74ef50-1861-4d70-b781-f92d5f4b35ba-kube-api-access-8mr4z\") pod \"4d74ef50-1861-4d70-b781-f92d5f4b35ba\" (UID: \"4d74ef50-1861-4d70-b781-f92d5f4b35ba\") " Oct 14 07:23:47 crc kubenswrapper[4870]: I1014 07:23:47.492283 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d74ef50-1861-4d70-b781-f92d5f4b35ba-combined-ca-bundle\") pod \"4d74ef50-1861-4d70-b781-f92d5f4b35ba\" (UID: \"4d74ef50-1861-4d70-b781-f92d5f4b35ba\") " Oct 14 07:23:47 crc kubenswrapper[4870]: I1014 07:23:47.492349 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/8abede26-36e8-46ec-a5c4-6cd53319133a-nova-novncproxy-tls-certs\") pod \"8abede26-36e8-46ec-a5c4-6cd53319133a\" (UID: \"8abede26-36e8-46ec-a5c4-6cd53319133a\") " Oct 14 07:23:47 crc kubenswrapper[4870]: I1014 07:23:47.492368 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zpqm8\" (UniqueName: \"kubernetes.io/projected/8abede26-36e8-46ec-a5c4-6cd53319133a-kube-api-access-zpqm8\") pod \"8abede26-36e8-46ec-a5c4-6cd53319133a\" (UID: \"8abede26-36e8-46ec-a5c4-6cd53319133a\") " Oct 14 07:23:47 crc kubenswrapper[4870]: I1014 07:23:47.492402 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4d74ef50-1861-4d70-b781-f92d5f4b35ba-operator-scripts\") pod \"4d74ef50-1861-4d70-b781-f92d5f4b35ba\" (UID: \"4d74ef50-1861-4d70-b781-f92d5f4b35ba\") " Oct 14 07:23:47 crc kubenswrapper[4870]: I1014 07:23:47.492449 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/4d74ef50-1861-4d70-b781-f92d5f4b35ba-config-data-generated\") pod \"4d74ef50-1861-4d70-b781-f92d5f4b35ba\" (UID: \"4d74ef50-1861-4d70-b781-f92d5f4b35ba\") " Oct 14 07:23:47 crc kubenswrapper[4870]: I1014 07:23:47.492504 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/8abede26-36e8-46ec-a5c4-6cd53319133a-vencrypt-tls-certs\") pod \"8abede26-36e8-46ec-a5c4-6cd53319133a\" (UID: \"8abede26-36e8-46ec-a5c4-6cd53319133a\") " Oct 14 07:23:47 crc kubenswrapper[4870]: I1014 07:23:47.492521 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8abede26-36e8-46ec-a5c4-6cd53319133a-config-data\") pod \"8abede26-36e8-46ec-a5c4-6cd53319133a\" (UID: \"8abede26-36e8-46ec-a5c4-6cd53319133a\") " Oct 14 07:23:47 crc kubenswrapper[4870]: I1014 07:23:47.492563 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/4d74ef50-1861-4d70-b781-f92d5f4b35ba-config-data-default\") pod \"4d74ef50-1861-4d70-b781-f92d5f4b35ba\" (UID: \"4d74ef50-1861-4d70-b781-f92d5f4b35ba\") " Oct 14 07:23:47 crc kubenswrapper[4870]: I1014 07:23:47.492601 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8abede26-36e8-46ec-a5c4-6cd53319133a-combined-ca-bundle\") pod \"8abede26-36e8-46ec-a5c4-6cd53319133a\" (UID: \"8abede26-36e8-46ec-a5c4-6cd53319133a\") " Oct 14 07:23:47 crc kubenswrapper[4870]: I1014 07:23:47.500141 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4d74ef50-1861-4d70-b781-f92d5f4b35ba-kolla-config" (OuterVolumeSpecName: "kolla-config") pod "4d74ef50-1861-4d70-b781-f92d5f4b35ba" (UID: "4d74ef50-1861-4d70-b781-f92d5f4b35ba"). InnerVolumeSpecName "kolla-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:23:47 crc kubenswrapper[4870]: I1014 07:23:47.510655 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4d74ef50-1861-4d70-b781-f92d5f4b35ba-config-data-generated" (OuterVolumeSpecName: "config-data-generated") pod "4d74ef50-1861-4d70-b781-f92d5f4b35ba" (UID: "4d74ef50-1861-4d70-b781-f92d5f4b35ba"). InnerVolumeSpecName "config-data-generated". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:23:47 crc kubenswrapper[4870]: I1014 07:23:47.513955 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4d74ef50-1861-4d70-b781-f92d5f4b35ba-config-data-default" (OuterVolumeSpecName: "config-data-default") pod "4d74ef50-1861-4d70-b781-f92d5f4b35ba" (UID: "4d74ef50-1861-4d70-b781-f92d5f4b35ba"). InnerVolumeSpecName "config-data-default". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:23:47 crc kubenswrapper[4870]: I1014 07:23:47.516511 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4d74ef50-1861-4d70-b781-f92d5f4b35ba-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "4d74ef50-1861-4d70-b781-f92d5f4b35ba" (UID: "4d74ef50-1861-4d70-b781-f92d5f4b35ba"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:23:47 crc kubenswrapper[4870]: I1014 07:23:47.517909 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d74ef50-1861-4d70-b781-f92d5f4b35ba-secrets" (OuterVolumeSpecName: "secrets") pod "4d74ef50-1861-4d70-b781-f92d5f4b35ba" (UID: "4d74ef50-1861-4d70-b781-f92d5f4b35ba"). InnerVolumeSpecName "secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:23:47 crc kubenswrapper[4870]: I1014 07:23:47.539960 4870 generic.go:334] "Generic (PLEG): container finished" podID="a9d8a25d-2f1c-483d-bf0d-76a2bf30f137" containerID="fbcb455b6e940e1124fe304e22169a2100dbb718060f679a722916c473baef26" exitCode=0 Oct 14 07:23:47 crc kubenswrapper[4870]: I1014 07:23:47.540042 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance1aaa-account-delete-nct6c" event={"ID":"a9d8a25d-2f1c-483d-bf0d-76a2bf30f137","Type":"ContainerDied","Data":"fbcb455b6e940e1124fe304e22169a2100dbb718060f679a722916c473baef26"} Oct 14 07:23:47 crc kubenswrapper[4870]: I1014 07:23:47.553688 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-mwp7r"] Oct 14 07:23:47 crc kubenswrapper[4870]: I1014 07:23:47.553761 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-cbhx4"] Oct 14 07:23:47 crc kubenswrapper[4870]: I1014 07:23:47.553815 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance1aaa-account-delete-nct6c" event={"ID":"a9d8a25d-2f1c-483d-bf0d-76a2bf30f137","Type":"ContainerStarted","Data":"a646110e9a961f59dfe446e486ce2da0f818db4b1ed54f19f0b9124ac62a80df"} Oct 14 07:23:47 crc kubenswrapper[4870]: I1014 07:23:47.562790 4870 scope.go:117] "RemoveContainer" containerID="660968003babd11d1e6d54c59b0e91ea2368857634bd407e25468d0645470948" Oct 14 07:23:47 crc kubenswrapper[4870]: I1014 07:23:47.573018 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-cbhx4"] Oct 14 07:23:47 crc kubenswrapper[4870]: I1014 07:23:47.586091 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-865c567797-jljrm"] Oct 14 07:23:47 crc kubenswrapper[4870]: I1014 07:23:47.586312 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/keystone-865c567797-jljrm" podUID="60de9d04-7c24-42a4-9d5e-74b611487b69" containerName="keystone-api" containerID="cri-o://d8615c16655ea7e7d454161d16b61973188b63bffc966ed4b3805db817746641" gracePeriod=30 Oct 14 07:23:47 crc kubenswrapper[4870]: I1014 07:23:47.587322 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8abede26-36e8-46ec-a5c4-6cd53319133a-kube-api-access-zpqm8" (OuterVolumeSpecName: "kube-api-access-zpqm8") pod "8abede26-36e8-46ec-a5c4-6cd53319133a" (UID: "8abede26-36e8-46ec-a5c4-6cd53319133a"). InnerVolumeSpecName "kube-api-access-zpqm8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:23:47 crc kubenswrapper[4870]: I1014 07:23:47.587507 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4d74ef50-1861-4d70-b781-f92d5f4b35ba-kube-api-access-8mr4z" (OuterVolumeSpecName: "kube-api-access-8mr4z") pod "4d74ef50-1861-4d70-b781-f92d5f4b35ba" (UID: "4d74ef50-1861-4d70-b781-f92d5f4b35ba"). InnerVolumeSpecName "kube-api-access-8mr4z". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:23:47 crc kubenswrapper[4870]: I1014 07:23:47.613055 4870 reconciler_common.go:293] "Volume detached for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/4d74ef50-1861-4d70-b781-f92d5f4b35ba-secrets\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:47 crc kubenswrapper[4870]: I1014 07:23:47.613083 4870 reconciler_common.go:293] "Volume detached for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/4d74ef50-1861-4d70-b781-f92d5f4b35ba-kolla-config\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:47 crc kubenswrapper[4870]: I1014 07:23:47.613097 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8mr4z\" (UniqueName: \"kubernetes.io/projected/4d74ef50-1861-4d70-b781-f92d5f4b35ba-kube-api-access-8mr4z\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:47 crc kubenswrapper[4870]: I1014 07:23:47.613108 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zpqm8\" (UniqueName: \"kubernetes.io/projected/8abede26-36e8-46ec-a5c4-6cd53319133a-kube-api-access-zpqm8\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:47 crc kubenswrapper[4870]: I1014 07:23:47.613119 4870 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4d74ef50-1861-4d70-b781-f92d5f4b35ba-operator-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:47 crc kubenswrapper[4870]: I1014 07:23:47.613132 4870 reconciler_common.go:293] "Volume detached for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/4d74ef50-1861-4d70-b781-f92d5f4b35ba-config-data-generated\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:47 crc kubenswrapper[4870]: I1014 07:23:47.613144 4870 reconciler_common.go:293] "Volume detached for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/4d74ef50-1861-4d70-b781-f92d5f4b35ba-config-data-default\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:47 crc kubenswrapper[4870]: I1014 07:23:47.617980 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstack-galera-0"] Oct 14 07:23:47 crc kubenswrapper[4870]: I1014 07:23:47.645370 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-4dkbn"] Oct 14 07:23:47 crc kubenswrapper[4870]: I1014 07:23:47.645820 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage01-crc" (OuterVolumeSpecName: "mysql-db") pod "4d74ef50-1861-4d70-b781-f92d5f4b35ba" (UID: "4d74ef50-1861-4d70-b781-f92d5f4b35ba"). InnerVolumeSpecName "local-storage01-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 14 07:23:47 crc kubenswrapper[4870]: I1014 07:23:47.695205 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-4dkbn"] Oct 14 07:23:47 crc kubenswrapper[4870]: I1014 07:23:47.718192 4870 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" " Oct 14 07:23:47 crc kubenswrapper[4870]: I1014 07:23:47.720196 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-220b-account-create-gthws"] Oct 14 07:23:47 crc kubenswrapper[4870]: I1014 07:23:47.729955 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-220b-account-create-gthws"] Oct 14 07:23:47 crc kubenswrapper[4870]: I1014 07:23:47.798944 4870 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/cinder-api-0" podUID="51de8090-c48e-498c-99ab-794a7527252c" containerName="cinder-api" probeResult="failure" output="Get \"https://10.217.0.173:8776/healthcheck\": read tcp 10.217.0.2:55904->10.217.0.173:8776: read: connection reset by peer" Oct 14 07:23:47 crc kubenswrapper[4870]: I1014 07:23:47.810382 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d74ef50-1861-4d70-b781-f92d5f4b35ba-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4d74ef50-1861-4d70-b781-f92d5f4b35ba" (UID: "4d74ef50-1861-4d70-b781-f92d5f4b35ba"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:23:47 crc kubenswrapper[4870]: I1014 07:23:47.821384 4870 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d74ef50-1861-4d70-b781-f92d5f4b35ba-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:47 crc kubenswrapper[4870]: I1014 07:23:47.827002 4870 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage01-crc" (UniqueName: "kubernetes.io/local-volume/local-storage01-crc") on node "crc" Oct 14 07:23:47 crc kubenswrapper[4870]: I1014 07:23:47.877079 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8abede26-36e8-46ec-a5c4-6cd53319133a-config-data" (OuterVolumeSpecName: "config-data") pod "8abede26-36e8-46ec-a5c4-6cd53319133a" (UID: "8abede26-36e8-46ec-a5c4-6cd53319133a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:23:47 crc kubenswrapper[4870]: I1014 07:23:47.885967 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8abede26-36e8-46ec-a5c4-6cd53319133a-vencrypt-tls-certs" (OuterVolumeSpecName: "vencrypt-tls-certs") pod "8abede26-36e8-46ec-a5c4-6cd53319133a" (UID: "8abede26-36e8-46ec-a5c4-6cd53319133a"). InnerVolumeSpecName "vencrypt-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:23:47 crc kubenswrapper[4870]: I1014 07:23:47.903277 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d74ef50-1861-4d70-b781-f92d5f4b35ba-galera-tls-certs" (OuterVolumeSpecName: "galera-tls-certs") pod "4d74ef50-1861-4d70-b781-f92d5f4b35ba" (UID: "4d74ef50-1861-4d70-b781-f92d5f4b35ba"). InnerVolumeSpecName "galera-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:23:47 crc kubenswrapper[4870]: I1014 07:23:47.904767 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8abede26-36e8-46ec-a5c4-6cd53319133a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8abede26-36e8-46ec-a5c4-6cd53319133a" (UID: "8abede26-36e8-46ec-a5c4-6cd53319133a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:23:47 crc kubenswrapper[4870]: I1014 07:23:47.906137 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8abede26-36e8-46ec-a5c4-6cd53319133a-nova-novncproxy-tls-certs" (OuterVolumeSpecName: "nova-novncproxy-tls-certs") pod "8abede26-36e8-46ec-a5c4-6cd53319133a" (UID: "8abede26-36e8-46ec-a5c4-6cd53319133a"). InnerVolumeSpecName "nova-novncproxy-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:23:47 crc kubenswrapper[4870]: I1014 07:23:47.926796 4870 reconciler_common.go:293] "Volume detached for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:47 crc kubenswrapper[4870]: I1014 07:23:47.926832 4870 reconciler_common.go:293] "Volume detached for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/4d74ef50-1861-4d70-b781-f92d5f4b35ba-galera-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:47 crc kubenswrapper[4870]: I1014 07:23:47.926844 4870 reconciler_common.go:293] "Volume detached for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/8abede26-36e8-46ec-a5c4-6cd53319133a-nova-novncproxy-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:47 crc kubenswrapper[4870]: I1014 07:23:47.926853 4870 reconciler_common.go:293] "Volume detached for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/8abede26-36e8-46ec-a5c4-6cd53319133a-vencrypt-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:47 crc kubenswrapper[4870]: I1014 07:23:47.926862 4870 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8abede26-36e8-46ec-a5c4-6cd53319133a-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:47 crc kubenswrapper[4870]: I1014 07:23:47.926870 4870 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8abede26-36e8-46ec-a5c4-6cd53319133a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:47 crc kubenswrapper[4870]: E1014 07:23:47.969012 4870 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of e6478418e9645b5d3cf9a2a8e1b1a4090b34dcf697735e42cf6fb41d3708bbeb is running failed: container process not found" containerID="e6478418e9645b5d3cf9a2a8e1b1a4090b34dcf697735e42cf6fb41d3708bbeb" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 14 07:23:47 crc kubenswrapper[4870]: E1014 07:23:47.969514 4870 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of e6478418e9645b5d3cf9a2a8e1b1a4090b34dcf697735e42cf6fb41d3708bbeb is running failed: container process not found" containerID="e6478418e9645b5d3cf9a2a8e1b1a4090b34dcf697735e42cf6fb41d3708bbeb" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 14 07:23:47 crc kubenswrapper[4870]: E1014 07:23:47.969833 4870 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of e6478418e9645b5d3cf9a2a8e1b1a4090b34dcf697735e42cf6fb41d3708bbeb is running failed: container process not found" containerID="e6478418e9645b5d3cf9a2a8e1b1a4090b34dcf697735e42cf6fb41d3708bbeb" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 14 07:23:47 crc kubenswrapper[4870]: E1014 07:23:47.969870 4870 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of e6478418e9645b5d3cf9a2a8e1b1a4090b34dcf697735e42cf6fb41d3708bbeb is running failed: container process not found" probeType="Readiness" pod="openstack/nova-cell0-conductor-0" podUID="bad45b30-5ba2-4062-ab14-04f923679c4e" containerName="nova-cell0-conductor-conductor" Oct 14 07:23:48 crc kubenswrapper[4870]: I1014 07:23:48.000496 4870 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="5212ba20-83c3-41ff-a3d9-5a315f009c1b" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.102:5671: connect: connection refused" Oct 14 07:23:48 crc kubenswrapper[4870]: I1014 07:23:48.006315 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstack-galera-0" podUID="3de2167a-7663-4b00-9743-229ff1864cb8" containerName="galera" containerID="cri-o://1b53e989d30152b230c176967064509ca1fafba96b986cc96bc23bdcafabf49f" gracePeriod=30 Oct 14 07:23:48 crc kubenswrapper[4870]: I1014 07:23:48.408790 4870 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="f0b766fd-ccb9-4620-84df-8fe2f194b0da" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.202:8775/\": dial tcp 10.217.0.202:8775: connect: connection refused" Oct 14 07:23:48 crc kubenswrapper[4870]: I1014 07:23:48.413346 4870 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="f0b766fd-ccb9-4620-84df-8fe2f194b0da" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.202:8775/\": dial tcp 10.217.0.202:8775: connect: connection refused" Oct 14 07:23:48 crc kubenswrapper[4870]: I1014 07:23:48.547030 4870 scope.go:117] "RemoveContainer" containerID="afc051c5e25bf94e2cbc69eabd5bf3dc5066c8c7364b3f767d2bb33960df9191" Oct 14 07:23:48 crc kubenswrapper[4870]: I1014 07:23:48.550716 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 14 07:23:48 crc kubenswrapper[4870]: I1014 07:23:48.562213 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement1a80-account-delete-8q7fq" Oct 14 07:23:48 crc kubenswrapper[4870]: I1014 07:23:48.562618 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-5485b69bc9-ftmmq" Oct 14 07:23:48 crc kubenswrapper[4870]: I1014 07:23:48.587386 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican6547-account-delete-v5crl" Oct 14 07:23:48 crc kubenswrapper[4870]: I1014 07:23:48.593619 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/novacell0ce16-account-delete-bwgsq" Oct 14 07:23:48 crc kubenswrapper[4870]: I1014 07:23:48.597063 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-5485b69bc9-ftmmq" event={"ID":"3927fb78-4f30-4371-8fbb-62fd276c0460","Type":"ContainerDied","Data":"428db2cf2ded2f12b61c1ef1b5e8692a727c90d13e0dd04dc1e44a5dff104a1c"} Oct 14 07:23:48 crc kubenswrapper[4870]: I1014 07:23:48.597247 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-5485b69bc9-ftmmq" Oct 14 07:23:48 crc kubenswrapper[4870]: I1014 07:23:48.627810 4870 generic.go:334] "Generic (PLEG): container finished" podID="f0b766fd-ccb9-4620-84df-8fe2f194b0da" containerID="65770c73c87b754f9177bcc262af5253b96b0534c636832b413175980c00c785" exitCode=0 Oct 14 07:23:48 crc kubenswrapper[4870]: I1014 07:23:48.635798 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f0b766fd-ccb9-4620-84df-8fe2f194b0da","Type":"ContainerDied","Data":"65770c73c87b754f9177bcc262af5253b96b0534c636832b413175980c00c785"} Oct 14 07:23:48 crc kubenswrapper[4870]: I1014 07:23:48.641095 4870 scope.go:117] "RemoveContainer" containerID="660968003babd11d1e6d54c59b0e91ea2368857634bd407e25468d0645470948" Oct 14 07:23:48 crc kubenswrapper[4870]: I1014 07:23:48.642274 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9m4qf\" (UniqueName: \"kubernetes.io/projected/3927fb78-4f30-4371-8fbb-62fd276c0460-kube-api-access-9m4qf\") pod \"3927fb78-4f30-4371-8fbb-62fd276c0460\" (UID: \"3927fb78-4f30-4371-8fbb-62fd276c0460\") " Oct 14 07:23:48 crc kubenswrapper[4870]: I1014 07:23:48.642384 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/61885759-41a7-4ee8-904a-3f85bb66cd19-config-data\") pod \"61885759-41a7-4ee8-904a-3f85bb66cd19\" (UID: \"61885759-41a7-4ee8-904a-3f85bb66cd19\") " Oct 14 07:23:48 crc kubenswrapper[4870]: I1014 07:23:48.642474 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/61885759-41a7-4ee8-904a-3f85bb66cd19-combined-ca-bundle\") pod \"61885759-41a7-4ee8-904a-3f85bb66cd19\" (UID: \"61885759-41a7-4ee8-904a-3f85bb66cd19\") " Oct 14 07:23:48 crc kubenswrapper[4870]: I1014 07:23:48.642520 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3927fb78-4f30-4371-8fbb-62fd276c0460-combined-ca-bundle\") pod \"3927fb78-4f30-4371-8fbb-62fd276c0460\" (UID: \"3927fb78-4f30-4371-8fbb-62fd276c0460\") " Oct 14 07:23:48 crc kubenswrapper[4870]: I1014 07:23:48.642580 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnlql\" (UniqueName: \"kubernetes.io/projected/7cbdbbdd-ea4f-4d14-8f3c-4784bda45874-kube-api-access-mnlql\") pod \"7cbdbbdd-ea4f-4d14-8f3c-4784bda45874\" (UID: \"7cbdbbdd-ea4f-4d14-8f3c-4784bda45874\") " Oct 14 07:23:48 crc kubenswrapper[4870]: I1014 07:23:48.642631 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z9mzh\" (UniqueName: \"kubernetes.io/projected/2abee394-ffb3-4dba-8d09-19c1cb476595-kube-api-access-z9mzh\") pod \"2abee394-ffb3-4dba-8d09-19c1cb476595\" (UID: \"2abee394-ffb3-4dba-8d09-19c1cb476595\") " Oct 14 07:23:48 crc kubenswrapper[4870]: I1014 07:23:48.642659 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/3927fb78-4f30-4371-8fbb-62fd276c0460-etc-swift\") pod \"3927fb78-4f30-4371-8fbb-62fd276c0460\" (UID: \"3927fb78-4f30-4371-8fbb-62fd276c0460\") " Oct 14 07:23:48 crc kubenswrapper[4870]: I1014 07:23:48.642717 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bv2wt\" (UniqueName: \"kubernetes.io/projected/61885759-41a7-4ee8-904a-3f85bb66cd19-kube-api-access-bv2wt\") pod \"61885759-41a7-4ee8-904a-3f85bb66cd19\" (UID: \"61885759-41a7-4ee8-904a-3f85bb66cd19\") " Oct 14 07:23:48 crc kubenswrapper[4870]: I1014 07:23:48.642785 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fpns6\" (UniqueName: \"kubernetes.io/projected/a6f02e4f-8dca-4b97-91e9-ee3368e5e44b-kube-api-access-fpns6\") pod \"a6f02e4f-8dca-4b97-91e9-ee3368e5e44b\" (UID: \"a6f02e4f-8dca-4b97-91e9-ee3368e5e44b\") " Oct 14 07:23:48 crc kubenswrapper[4870]: I1014 07:23:48.642817 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3927fb78-4f30-4371-8fbb-62fd276c0460-internal-tls-certs\") pod \"3927fb78-4f30-4371-8fbb-62fd276c0460\" (UID: \"3927fb78-4f30-4371-8fbb-62fd276c0460\") " Oct 14 07:23:48 crc kubenswrapper[4870]: I1014 07:23:48.642868 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3927fb78-4f30-4371-8fbb-62fd276c0460-public-tls-certs\") pod \"3927fb78-4f30-4371-8fbb-62fd276c0460\" (UID: \"3927fb78-4f30-4371-8fbb-62fd276c0460\") " Oct 14 07:23:48 crc kubenswrapper[4870]: I1014 07:23:48.642896 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3927fb78-4f30-4371-8fbb-62fd276c0460-log-httpd\") pod \"3927fb78-4f30-4371-8fbb-62fd276c0460\" (UID: \"3927fb78-4f30-4371-8fbb-62fd276c0460\") " Oct 14 07:23:48 crc kubenswrapper[4870]: I1014 07:23:48.642934 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3927fb78-4f30-4371-8fbb-62fd276c0460-config-data\") pod \"3927fb78-4f30-4371-8fbb-62fd276c0460\" (UID: \"3927fb78-4f30-4371-8fbb-62fd276c0460\") " Oct 14 07:23:48 crc kubenswrapper[4870]: I1014 07:23:48.642973 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3927fb78-4f30-4371-8fbb-62fd276c0460-run-httpd\") pod \"3927fb78-4f30-4371-8fbb-62fd276c0460\" (UID: \"3927fb78-4f30-4371-8fbb-62fd276c0460\") " Oct 14 07:23:48 crc kubenswrapper[4870]: I1014 07:23:48.643901 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3927fb78-4f30-4371-8fbb-62fd276c0460-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "3927fb78-4f30-4371-8fbb-62fd276c0460" (UID: "3927fb78-4f30-4371-8fbb-62fd276c0460"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:23:48 crc kubenswrapper[4870]: E1014 07:23:48.651663 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"660968003babd11d1e6d54c59b0e91ea2368857634bd407e25468d0645470948\": container with ID starting with 660968003babd11d1e6d54c59b0e91ea2368857634bd407e25468d0645470948 not found: ID does not exist" containerID="660968003babd11d1e6d54c59b0e91ea2368857634bd407e25468d0645470948" Oct 14 07:23:48 crc kubenswrapper[4870]: I1014 07:23:48.651704 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"660968003babd11d1e6d54c59b0e91ea2368857634bd407e25468d0645470948"} err="failed to get container status \"660968003babd11d1e6d54c59b0e91ea2368857634bd407e25468d0645470948\": rpc error: code = NotFound desc = could not find container \"660968003babd11d1e6d54c59b0e91ea2368857634bd407e25468d0645470948\": container with ID starting with 660968003babd11d1e6d54c59b0e91ea2368857634bd407e25468d0645470948 not found: ID does not exist" Oct 14 07:23:48 crc kubenswrapper[4870]: I1014 07:23:48.651726 4870 scope.go:117] "RemoveContainer" containerID="afc051c5e25bf94e2cbc69eabd5bf3dc5066c8c7364b3f767d2bb33960df9191" Oct 14 07:23:48 crc kubenswrapper[4870]: E1014 07:23:48.652837 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"afc051c5e25bf94e2cbc69eabd5bf3dc5066c8c7364b3f767d2bb33960df9191\": container with ID starting with afc051c5e25bf94e2cbc69eabd5bf3dc5066c8c7364b3f767d2bb33960df9191 not found: ID does not exist" containerID="afc051c5e25bf94e2cbc69eabd5bf3dc5066c8c7364b3f767d2bb33960df9191" Oct 14 07:23:48 crc kubenswrapper[4870]: I1014 07:23:48.652990 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"afc051c5e25bf94e2cbc69eabd5bf3dc5066c8c7364b3f767d2bb33960df9191"} err="failed to get container status \"afc051c5e25bf94e2cbc69eabd5bf3dc5066c8c7364b3f767d2bb33960df9191\": rpc error: code = NotFound desc = could not find container \"afc051c5e25bf94e2cbc69eabd5bf3dc5066c8c7364b3f767d2bb33960df9191\": container with ID starting with afc051c5e25bf94e2cbc69eabd5bf3dc5066c8c7364b3f767d2bb33960df9191 not found: ID does not exist" Oct 14 07:23:48 crc kubenswrapper[4870]: I1014 07:23:48.653021 4870 scope.go:117] "RemoveContainer" containerID="009e0a5765edbfd4635dd4efcc0b112f3d5e0691352f5b4e2d147aac46326707" Oct 14 07:23:48 crc kubenswrapper[4870]: I1014 07:23:48.657390 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3927fb78-4f30-4371-8fbb-62fd276c0460-kube-api-access-9m4qf" (OuterVolumeSpecName: "kube-api-access-9m4qf") pod "3927fb78-4f30-4371-8fbb-62fd276c0460" (UID: "3927fb78-4f30-4371-8fbb-62fd276c0460"). InnerVolumeSpecName "kube-api-access-9m4qf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:23:48 crc kubenswrapper[4870]: I1014 07:23:48.658522 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"61885759-41a7-4ee8-904a-3f85bb66cd19","Type":"ContainerDied","Data":"2c5fb8deafb920aec9e2d2835198a9dfcbe4d42a4bb7c408109aef8440352e4f"} Oct 14 07:23:48 crc kubenswrapper[4870]: I1014 07:23:48.658863 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 14 07:23:48 crc kubenswrapper[4870]: I1014 07:23:48.659372 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3927fb78-4f30-4371-8fbb-62fd276c0460-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "3927fb78-4f30-4371-8fbb-62fd276c0460" (UID: "3927fb78-4f30-4371-8fbb-62fd276c0460"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:23:48 crc kubenswrapper[4870]: I1014 07:23:48.659917 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder5c98-account-delete-lhjjg" Oct 14 07:23:48 crc kubenswrapper[4870]: I1014 07:23:48.659996 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a6f02e4f-8dca-4b97-91e9-ee3368e5e44b-kube-api-access-fpns6" (OuterVolumeSpecName: "kube-api-access-fpns6") pod "a6f02e4f-8dca-4b97-91e9-ee3368e5e44b" (UID: "a6f02e4f-8dca-4b97-91e9-ee3368e5e44b"). InnerVolumeSpecName "kube-api-access-fpns6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:23:48 crc kubenswrapper[4870]: I1014 07:23:48.660013 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3927fb78-4f30-4371-8fbb-62fd276c0460-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "3927fb78-4f30-4371-8fbb-62fd276c0460" (UID: "3927fb78-4f30-4371-8fbb-62fd276c0460"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:23:48 crc kubenswrapper[4870]: I1014 07:23:48.662001 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement1a80-account-delete-8q7fq" event={"ID":"7cbdbbdd-ea4f-4d14-8f3c-4784bda45874","Type":"ContainerDied","Data":"bca8553e8397dc43e04b16f525a283963e83bdf774088be23914b72c95434ac0"} Oct 14 07:23:48 crc kubenswrapper[4870]: I1014 07:23:48.662380 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement1a80-account-delete-8q7fq" Oct 14 07:23:48 crc kubenswrapper[4870]: I1014 07:23:48.663826 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/61885759-41a7-4ee8-904a-3f85bb66cd19-kube-api-access-bv2wt" (OuterVolumeSpecName: "kube-api-access-bv2wt") pod "61885759-41a7-4ee8-904a-3f85bb66cd19" (UID: "61885759-41a7-4ee8-904a-3f85bb66cd19"). InnerVolumeSpecName "kube-api-access-bv2wt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:23:48 crc kubenswrapper[4870]: I1014 07:23:48.669355 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 14 07:23:48 crc kubenswrapper[4870]: I1014 07:23:48.685353 4870 generic.go:334] "Generic (PLEG): container finished" podID="56eb286f-8aa2-46aa-a74d-031288a9aad1" containerID="d8fce18d04004b32792441ce25ddcd9d74c06df092b20d09c14fca60675539d4" exitCode=0 Oct 14 07:23:48 crc kubenswrapper[4870]: I1014 07:23:48.685666 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"56eb286f-8aa2-46aa-a74d-031288a9aad1","Type":"ContainerDied","Data":"d8fce18d04004b32792441ce25ddcd9d74c06df092b20d09c14fca60675539d4"} Oct 14 07:23:48 crc kubenswrapper[4870]: I1014 07:23:48.694770 4870 scope.go:117] "RemoveContainer" containerID="58c55f50ceab03915ab9a353ebd2b0d046666be23591bd23a6a60026b468abe2" Oct 14 07:23:48 crc kubenswrapper[4870]: I1014 07:23:48.697337 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2abee394-ffb3-4dba-8d09-19c1cb476595-kube-api-access-z9mzh" (OuterVolumeSpecName: "kube-api-access-z9mzh") pod "2abee394-ffb3-4dba-8d09-19c1cb476595" (UID: "2abee394-ffb3-4dba-8d09-19c1cb476595"). InnerVolumeSpecName "kube-api-access-z9mzh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:23:48 crc kubenswrapper[4870]: I1014 07:23:48.698225 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/61885759-41a7-4ee8-904a-3f85bb66cd19-config-data" (OuterVolumeSpecName: "config-data") pod "61885759-41a7-4ee8-904a-3f85bb66cd19" (UID: "61885759-41a7-4ee8-904a-3f85bb66cd19"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:23:48 crc kubenswrapper[4870]: I1014 07:23:48.700250 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance1aaa-account-delete-nct6c" Oct 14 07:23:48 crc kubenswrapper[4870]: I1014 07:23:48.700702 4870 generic.go:334] "Generic (PLEG): container finished" podID="2130e3f2-cbea-46e1-b5d3-daefdd972935" containerID="4339244c7626324d778551bfb8f861a2c704b3e5aa7199c34ba643e475017090" exitCode=0 Oct 14 07:23:48 crc kubenswrapper[4870]: I1014 07:23:48.700765 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-597ff968b-msxbn" event={"ID":"2130e3f2-cbea-46e1-b5d3-daefdd972935","Type":"ContainerDied","Data":"4339244c7626324d778551bfb8f861a2c704b3e5aa7199c34ba643e475017090"} Oct 14 07:23:48 crc kubenswrapper[4870]: I1014 07:23:48.703844 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7cbdbbdd-ea4f-4d14-8f3c-4784bda45874-kube-api-access-mnlql" (OuterVolumeSpecName: "kube-api-access-mnlql") pod "7cbdbbdd-ea4f-4d14-8f3c-4784bda45874" (UID: "7cbdbbdd-ea4f-4d14-8f3c-4784bda45874"). InnerVolumeSpecName "kube-api-access-mnlql". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:23:48 crc kubenswrapper[4870]: I1014 07:23:48.728181 4870 generic.go:334] "Generic (PLEG): container finished" podID="c712b87a-cf4d-437e-abe0-1a566046ca66" containerID="f744eb9b855e3191279bf5fc01139d0a3359c09a6178c62d0602563d9d13ff25" exitCode=0 Oct 14 07:23:48 crc kubenswrapper[4870]: I1014 07:23:48.728256 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"c712b87a-cf4d-437e-abe0-1a566046ca66","Type":"ContainerDied","Data":"f744eb9b855e3191279bf5fc01139d0a3359c09a6178c62d0602563d9d13ff25"} Oct 14 07:23:48 crc kubenswrapper[4870]: I1014 07:23:48.728385 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 14 07:23:48 crc kubenswrapper[4870]: I1014 07:23:48.732998 4870 generic.go:334] "Generic (PLEG): container finished" podID="c603f7b5-015f-4c2e-9144-b3e8349c3da3" containerID="c79902c86d1a79331c9994ac2e40ba61e03d57293be1023f0938cc002e01194a" exitCode=2 Oct 14 07:23:48 crc kubenswrapper[4870]: I1014 07:23:48.733090 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"c603f7b5-015f-4c2e-9144-b3e8349c3da3","Type":"ContainerDied","Data":"c79902c86d1a79331c9994ac2e40ba61e03d57293be1023f0938cc002e01194a"} Oct 14 07:23:48 crc kubenswrapper[4870]: I1014 07:23:48.733135 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"c603f7b5-015f-4c2e-9144-b3e8349c3da3","Type":"ContainerDied","Data":"9849fe4942d0101a5e43df2d44e038448ff1198af84d34460865495f68e86877"} Oct 14 07:23:48 crc kubenswrapper[4870]: I1014 07:23:48.733209 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 14 07:23:48 crc kubenswrapper[4870]: I1014 07:23:48.737309 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican6547-account-delete-v5crl" event={"ID":"2abee394-ffb3-4dba-8d09-19c1cb476595","Type":"ContainerDied","Data":"dc92ede877d28eaf350cc8faaf42ce060c07d06e071039d2b9a83dbb16901f1a"} Oct 14 07:23:48 crc kubenswrapper[4870]: I1014 07:23:48.737384 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican6547-account-delete-v5crl" Oct 14 07:23:48 crc kubenswrapper[4870]: I1014 07:23:48.744335 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/c603f7b5-015f-4c2e-9144-b3e8349c3da3-kube-state-metrics-tls-config\") pod \"c603f7b5-015f-4c2e-9144-b3e8349c3da3\" (UID: \"c603f7b5-015f-4c2e-9144-b3e8349c3da3\") " Oct 14 07:23:48 crc kubenswrapper[4870]: I1014 07:23:48.744472 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l62g6\" (UniqueName: \"kubernetes.io/projected/c603f7b5-015f-4c2e-9144-b3e8349c3da3-kube-api-access-l62g6\") pod \"c603f7b5-015f-4c2e-9144-b3e8349c3da3\" (UID: \"c603f7b5-015f-4c2e-9144-b3e8349c3da3\") " Oct 14 07:23:48 crc kubenswrapper[4870]: I1014 07:23:48.744528 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j2gjx\" (UniqueName: \"kubernetes.io/projected/de479f5d-22a2-4f7e-8034-cffefc511b53-kube-api-access-j2gjx\") pod \"de479f5d-22a2-4f7e-8034-cffefc511b53\" (UID: \"de479f5d-22a2-4f7e-8034-cffefc511b53\") " Oct 14 07:23:48 crc kubenswrapper[4870]: I1014 07:23:48.744598 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c603f7b5-015f-4c2e-9144-b3e8349c3da3-combined-ca-bundle\") pod \"c603f7b5-015f-4c2e-9144-b3e8349c3da3\" (UID: \"c603f7b5-015f-4c2e-9144-b3e8349c3da3\") " Oct 14 07:23:48 crc kubenswrapper[4870]: I1014 07:23:48.744627 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/c603f7b5-015f-4c2e-9144-b3e8349c3da3-kube-state-metrics-tls-certs\") pod \"c603f7b5-015f-4c2e-9144-b3e8349c3da3\" (UID: \"c603f7b5-015f-4c2e-9144-b3e8349c3da3\") " Oct 14 07:23:48 crc kubenswrapper[4870]: I1014 07:23:48.762259 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnlql\" (UniqueName: \"kubernetes.io/projected/7cbdbbdd-ea4f-4d14-8f3c-4784bda45874-kube-api-access-mnlql\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:48 crc kubenswrapper[4870]: I1014 07:23:48.762291 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z9mzh\" (UniqueName: \"kubernetes.io/projected/2abee394-ffb3-4dba-8d09-19c1cb476595-kube-api-access-z9mzh\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:48 crc kubenswrapper[4870]: I1014 07:23:48.762300 4870 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/3927fb78-4f30-4371-8fbb-62fd276c0460-etc-swift\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:48 crc kubenswrapper[4870]: I1014 07:23:48.762312 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bv2wt\" (UniqueName: \"kubernetes.io/projected/61885759-41a7-4ee8-904a-3f85bb66cd19-kube-api-access-bv2wt\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:48 crc kubenswrapper[4870]: I1014 07:23:48.762321 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fpns6\" (UniqueName: \"kubernetes.io/projected/a6f02e4f-8dca-4b97-91e9-ee3368e5e44b-kube-api-access-fpns6\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:48 crc kubenswrapper[4870]: I1014 07:23:48.762328 4870 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3927fb78-4f30-4371-8fbb-62fd276c0460-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:48 crc kubenswrapper[4870]: I1014 07:23:48.762338 4870 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3927fb78-4f30-4371-8fbb-62fd276c0460-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:48 crc kubenswrapper[4870]: I1014 07:23:48.762396 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9m4qf\" (UniqueName: \"kubernetes.io/projected/3927fb78-4f30-4371-8fbb-62fd276c0460-kube-api-access-9m4qf\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:48 crc kubenswrapper[4870]: I1014 07:23:48.762405 4870 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/61885759-41a7-4ee8-904a-3f85bb66cd19-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:48 crc kubenswrapper[4870]: I1014 07:23:48.764080 4870 generic.go:334] "Generic (PLEG): container finished" podID="bad45b30-5ba2-4062-ab14-04f923679c4e" containerID="e6478418e9645b5d3cf9a2a8e1b1a4090b34dcf697735e42cf6fb41d3708bbeb" exitCode=0 Oct 14 07:23:48 crc kubenswrapper[4870]: I1014 07:23:48.764178 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 14 07:23:48 crc kubenswrapper[4870]: I1014 07:23:48.764356 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"bad45b30-5ba2-4062-ab14-04f923679c4e","Type":"ContainerDied","Data":"e6478418e9645b5d3cf9a2a8e1b1a4090b34dcf697735e42cf6fb41d3708bbeb"} Oct 14 07:23:48 crc kubenswrapper[4870]: I1014 07:23:48.764401 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"bad45b30-5ba2-4062-ab14-04f923679c4e","Type":"ContainerDied","Data":"27afa672a6627a62a0ae978b968b82988141797545411ad2982940d134c3309d"} Oct 14 07:23:48 crc kubenswrapper[4870]: I1014 07:23:48.785355 4870 generic.go:334] "Generic (PLEG): container finished" podID="208754ea-7cc0-4222-a186-9ba01b274add" containerID="cd54f5f17ff79e109b4abe40e2d80763b88c28e0edaa5a29d13dfa1146069254" exitCode=0 Oct 14 07:23:48 crc kubenswrapper[4870]: I1014 07:23:48.785482 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novaapie66c-account-delete-ftd24" event={"ID":"208754ea-7cc0-4222-a186-9ba01b274add","Type":"ContainerDied","Data":"cd54f5f17ff79e109b4abe40e2d80763b88c28e0edaa5a29d13dfa1146069254"} Oct 14 07:23:48 crc kubenswrapper[4870]: I1014 07:23:48.805516 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c603f7b5-015f-4c2e-9144-b3e8349c3da3-kube-api-access-l62g6" (OuterVolumeSpecName: "kube-api-access-l62g6") pod "c603f7b5-015f-4c2e-9144-b3e8349c3da3" (UID: "c603f7b5-015f-4c2e-9144-b3e8349c3da3"). InnerVolumeSpecName "kube-api-access-l62g6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:23:48 crc kubenswrapper[4870]: I1014 07:23:48.808428 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/de479f5d-22a2-4f7e-8034-cffefc511b53-kube-api-access-j2gjx" (OuterVolumeSpecName: "kube-api-access-j2gjx") pod "de479f5d-22a2-4f7e-8034-cffefc511b53" (UID: "de479f5d-22a2-4f7e-8034-cffefc511b53"). InnerVolumeSpecName "kube-api-access-j2gjx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:23:48 crc kubenswrapper[4870]: I1014 07:23:48.813779 4870 generic.go:334] "Generic (PLEG): container finished" podID="524b0b2a-77a2-4c05-be39-c8b97666ab11" containerID="018378adb14febcf4afcbaac2d9f70a84cc9e7f96675c9b85ee751bb465d9632" exitCode=0 Oct 14 07:23:48 crc kubenswrapper[4870]: I1014 07:23:48.813806 4870 generic.go:334] "Generic (PLEG): container finished" podID="524b0b2a-77a2-4c05-be39-c8b97666ab11" containerID="a7f06f2b8b7523a5efb4f46ab915a339acb0f160d1f9753afa2035de3c545c6d" exitCode=2 Oct 14 07:23:48 crc kubenswrapper[4870]: I1014 07:23:48.813814 4870 generic.go:334] "Generic (PLEG): container finished" podID="524b0b2a-77a2-4c05-be39-c8b97666ab11" containerID="f24a75c62a58aefe115cf6dc0ab70e01b811b66e1369401528bc3134ca7d1cc4" exitCode=0 Oct 14 07:23:48 crc kubenswrapper[4870]: I1014 07:23:48.813856 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"524b0b2a-77a2-4c05-be39-c8b97666ab11","Type":"ContainerDied","Data":"018378adb14febcf4afcbaac2d9f70a84cc9e7f96675c9b85ee751bb465d9632"} Oct 14 07:23:48 crc kubenswrapper[4870]: I1014 07:23:48.813881 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"524b0b2a-77a2-4c05-be39-c8b97666ab11","Type":"ContainerDied","Data":"a7f06f2b8b7523a5efb4f46ab915a339acb0f160d1f9753afa2035de3c545c6d"} Oct 14 07:23:48 crc kubenswrapper[4870]: I1014 07:23:48.813892 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"524b0b2a-77a2-4c05-be39-c8b97666ab11","Type":"ContainerDied","Data":"f24a75c62a58aefe115cf6dc0ab70e01b811b66e1369401528bc3134ca7d1cc4"} Oct 14 07:23:48 crc kubenswrapper[4870]: I1014 07:23:48.828946 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/61885759-41a7-4ee8-904a-3f85bb66cd19-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "61885759-41a7-4ee8-904a-3f85bb66cd19" (UID: "61885759-41a7-4ee8-904a-3f85bb66cd19"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:23:48 crc kubenswrapper[4870]: I1014 07:23:48.837685 4870 scope.go:117] "RemoveContainer" containerID="925b675acc2471fa8d4cffebdeddb3b79263bf8e5254078a5b2f740ce843cfd5" Oct 14 07:23:48 crc kubenswrapper[4870]: I1014 07:23:48.844964 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance1aaa-account-delete-nct6c" event={"ID":"a9d8a25d-2f1c-483d-bf0d-76a2bf30f137","Type":"ContainerDied","Data":"a646110e9a961f59dfe446e486ce2da0f818db4b1ed54f19f0b9124ac62a80df"} Oct 14 07:23:48 crc kubenswrapper[4870]: I1014 07:23:48.845282 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance1aaa-account-delete-nct6c" Oct 14 07:23:48 crc kubenswrapper[4870]: I1014 07:23:48.863192 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jflv9\" (UniqueName: \"kubernetes.io/projected/bad45b30-5ba2-4062-ab14-04f923679c4e-kube-api-access-jflv9\") pod \"bad45b30-5ba2-4062-ab14-04f923679c4e\" (UID: \"bad45b30-5ba2-4062-ab14-04f923679c4e\") " Oct 14 07:23:48 crc kubenswrapper[4870]: I1014 07:23:48.863439 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sklwp\" (UniqueName: \"kubernetes.io/projected/a9d8a25d-2f1c-483d-bf0d-76a2bf30f137-kube-api-access-sklwp\") pod \"a9d8a25d-2f1c-483d-bf0d-76a2bf30f137\" (UID: \"a9d8a25d-2f1c-483d-bf0d-76a2bf30f137\") " Oct 14 07:23:48 crc kubenswrapper[4870]: I1014 07:23:48.863596 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bad45b30-5ba2-4062-ab14-04f923679c4e-combined-ca-bundle\") pod \"bad45b30-5ba2-4062-ab14-04f923679c4e\" (UID: \"bad45b30-5ba2-4062-ab14-04f923679c4e\") " Oct 14 07:23:48 crc kubenswrapper[4870]: I1014 07:23:48.863751 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bad45b30-5ba2-4062-ab14-04f923679c4e-config-data\") pod \"bad45b30-5ba2-4062-ab14-04f923679c4e\" (UID: \"bad45b30-5ba2-4062-ab14-04f923679c4e\") " Oct 14 07:23:48 crc kubenswrapper[4870]: I1014 07:23:48.864402 4870 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/61885759-41a7-4ee8-904a-3f85bb66cd19-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:48 crc kubenswrapper[4870]: I1014 07:23:48.865075 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l62g6\" (UniqueName: \"kubernetes.io/projected/c603f7b5-015f-4c2e-9144-b3e8349c3da3-kube-api-access-l62g6\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:48 crc kubenswrapper[4870]: I1014 07:23:48.865168 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j2gjx\" (UniqueName: \"kubernetes.io/projected/de479f5d-22a2-4f7e-8034-cffefc511b53-kube-api-access-j2gjx\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:48 crc kubenswrapper[4870]: I1014 07:23:48.864982 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c603f7b5-015f-4c2e-9144-b3e8349c3da3-kube-state-metrics-tls-config" (OuterVolumeSpecName: "kube-state-metrics-tls-config") pod "c603f7b5-015f-4c2e-9144-b3e8349c3da3" (UID: "c603f7b5-015f-4c2e-9144-b3e8349c3da3"). InnerVolumeSpecName "kube-state-metrics-tls-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:23:48 crc kubenswrapper[4870]: I1014 07:23:48.865368 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3927fb78-4f30-4371-8fbb-62fd276c0460-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "3927fb78-4f30-4371-8fbb-62fd276c0460" (UID: "3927fb78-4f30-4371-8fbb-62fd276c0460"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:23:48 crc kubenswrapper[4870]: I1014 07:23:48.868540 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a9d8a25d-2f1c-483d-bf0d-76a2bf30f137-kube-api-access-sklwp" (OuterVolumeSpecName: "kube-api-access-sklwp") pod "a9d8a25d-2f1c-483d-bf0d-76a2bf30f137" (UID: "a9d8a25d-2f1c-483d-bf0d-76a2bf30f137"). InnerVolumeSpecName "kube-api-access-sklwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:23:48 crc kubenswrapper[4870]: I1014 07:23:48.869544 4870 generic.go:334] "Generic (PLEG): container finished" podID="3bd1e649-3d09-4cda-921f-ac9a84a5066e" containerID="cd6dabf959a5bb827e348f8580f0325c6287615f71bc7279bb2e2744f5cf4f7c" exitCode=0 Oct 14 07:23:48 crc kubenswrapper[4870]: I1014 07:23:48.869594 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"3bd1e649-3d09-4cda-921f-ac9a84a5066e","Type":"ContainerDied","Data":"cd6dabf959a5bb827e348f8580f0325c6287615f71bc7279bb2e2744f5cf4f7c"} Oct 14 07:23:48 crc kubenswrapper[4870]: I1014 07:23:48.871320 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bad45b30-5ba2-4062-ab14-04f923679c4e-kube-api-access-jflv9" (OuterVolumeSpecName: "kube-api-access-jflv9") pod "bad45b30-5ba2-4062-ab14-04f923679c4e" (UID: "bad45b30-5ba2-4062-ab14-04f923679c4e"). InnerVolumeSpecName "kube-api-access-jflv9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:23:48 crc kubenswrapper[4870]: E1014 07:23:48.872984 4870 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="1b53e989d30152b230c176967064509ca1fafba96b986cc96bc23bdcafabf49f" cmd=["/bin/bash","/var/lib/operator-scripts/mysql_probe.sh","readiness"] Oct 14 07:23:48 crc kubenswrapper[4870]: I1014 07:23:48.873508 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c603f7b5-015f-4c2e-9144-b3e8349c3da3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c603f7b5-015f-4c2e-9144-b3e8349c3da3" (UID: "c603f7b5-015f-4c2e-9144-b3e8349c3da3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:23:48 crc kubenswrapper[4870]: I1014 07:23:48.875479 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder5c98-account-delete-lhjjg" event={"ID":"de479f5d-22a2-4f7e-8034-cffefc511b53","Type":"ContainerDied","Data":"8b623fe0ae243b386c95275cd06b0fe710ca79ddd29998a1c3cb917add453e2e"} Oct 14 07:23:48 crc kubenswrapper[4870]: I1014 07:23:48.875967 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder5c98-account-delete-lhjjg" Oct 14 07:23:48 crc kubenswrapper[4870]: I1014 07:23:48.876420 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3927fb78-4f30-4371-8fbb-62fd276c0460-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3927fb78-4f30-4371-8fbb-62fd276c0460" (UID: "3927fb78-4f30-4371-8fbb-62fd276c0460"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:23:48 crc kubenswrapper[4870]: I1014 07:23:48.880220 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novacell0ce16-account-delete-bwgsq" event={"ID":"a6f02e4f-8dca-4b97-91e9-ee3368e5e44b","Type":"ContainerDied","Data":"1776e2a9afb3ac3237ca85fa6c702e235efdc65ae828037d00005680033973a2"} Oct 14 07:23:48 crc kubenswrapper[4870]: I1014 07:23:48.880309 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/novacell0ce16-account-delete-bwgsq" Oct 14 07:23:48 crc kubenswrapper[4870]: I1014 07:23:48.887427 4870 generic.go:334] "Generic (PLEG): container finished" podID="51de8090-c48e-498c-99ab-794a7527252c" containerID="571dfedb34f9303684e8f8fe09485d339b6c363eaa75f6950ae538f2526d0129" exitCode=0 Oct 14 07:23:48 crc kubenswrapper[4870]: I1014 07:23:48.887549 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Oct 14 07:23:48 crc kubenswrapper[4870]: I1014 07:23:48.890597 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"51de8090-c48e-498c-99ab-794a7527252c","Type":"ContainerDied","Data":"571dfedb34f9303684e8f8fe09485d339b6c363eaa75f6950ae538f2526d0129"} Oct 14 07:23:48 crc kubenswrapper[4870]: I1014 07:23:48.890635 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"51de8090-c48e-498c-99ab-794a7527252c","Type":"ContainerDied","Data":"c95d4c3fdb901eaa3800b8cbe01a1d7f2794649a3a2e3635950989ee7a523852"} Oct 14 07:23:48 crc kubenswrapper[4870]: I1014 07:23:48.890648 4870 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c95d4c3fdb901eaa3800b8cbe01a1d7f2794649a3a2e3635950989ee7a523852" Oct 14 07:23:48 crc kubenswrapper[4870]: I1014 07:23:48.890692 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 14 07:23:48 crc kubenswrapper[4870]: I1014 07:23:48.891066 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3927fb78-4f30-4371-8fbb-62fd276c0460-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "3927fb78-4f30-4371-8fbb-62fd276c0460" (UID: "3927fb78-4f30-4371-8fbb-62fd276c0460"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:23:48 crc kubenswrapper[4870]: I1014 07:23:48.901624 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3927fb78-4f30-4371-8fbb-62fd276c0460-config-data" (OuterVolumeSpecName: "config-data") pod "3927fb78-4f30-4371-8fbb-62fd276c0460" (UID: "3927fb78-4f30-4371-8fbb-62fd276c0460"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:23:48 crc kubenswrapper[4870]: I1014 07:23:48.901752 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c603f7b5-015f-4c2e-9144-b3e8349c3da3-kube-state-metrics-tls-certs" (OuterVolumeSpecName: "kube-state-metrics-tls-certs") pod "c603f7b5-015f-4c2e-9144-b3e8349c3da3" (UID: "c603f7b5-015f-4c2e-9144-b3e8349c3da3"). InnerVolumeSpecName "kube-state-metrics-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:23:48 crc kubenswrapper[4870]: E1014 07:23:48.905149 4870 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="1b53e989d30152b230c176967064509ca1fafba96b986cc96bc23bdcafabf49f" cmd=["/bin/bash","/var/lib/operator-scripts/mysql_probe.sh","readiness"] Oct 14 07:23:48 crc kubenswrapper[4870]: E1014 07:23:48.906735 4870 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="1b53e989d30152b230c176967064509ca1fafba96b986cc96bc23bdcafabf49f" cmd=["/bin/bash","/var/lib/operator-scripts/mysql_probe.sh","readiness"] Oct 14 07:23:48 crc kubenswrapper[4870]: E1014 07:23:48.906774 4870 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/openstack-galera-0" podUID="3de2167a-7663-4b00-9743-229ff1864cb8" containerName="galera" Oct 14 07:23:48 crc kubenswrapper[4870]: I1014 07:23:48.927610 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bad45b30-5ba2-4062-ab14-04f923679c4e-config-data" (OuterVolumeSpecName: "config-data") pod "bad45b30-5ba2-4062-ab14-04f923679c4e" (UID: "bad45b30-5ba2-4062-ab14-04f923679c4e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:23:48 crc kubenswrapper[4870]: I1014 07:23:48.927719 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bad45b30-5ba2-4062-ab14-04f923679c4e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bad45b30-5ba2-4062-ab14-04f923679c4e" (UID: "bad45b30-5ba2-4062-ab14-04f923679c4e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:23:48 crc kubenswrapper[4870]: I1014 07:23:48.967284 4870 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3927fb78-4f30-4371-8fbb-62fd276c0460-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:48 crc kubenswrapper[4870]: I1014 07:23:48.967325 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jflv9\" (UniqueName: \"kubernetes.io/projected/bad45b30-5ba2-4062-ab14-04f923679c4e-kube-api-access-jflv9\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:48 crc kubenswrapper[4870]: I1014 07:23:48.967342 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sklwp\" (UniqueName: \"kubernetes.io/projected/a9d8a25d-2f1c-483d-bf0d-76a2bf30f137-kube-api-access-sklwp\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:48 crc kubenswrapper[4870]: I1014 07:23:48.967355 4870 reconciler_common.go:293] "Volume detached for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/c603f7b5-015f-4c2e-9144-b3e8349c3da3-kube-state-metrics-tls-config\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:48 crc kubenswrapper[4870]: I1014 07:23:48.967371 4870 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bad45b30-5ba2-4062-ab14-04f923679c4e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:48 crc kubenswrapper[4870]: I1014 07:23:48.967383 4870 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3927fb78-4f30-4371-8fbb-62fd276c0460-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:48 crc kubenswrapper[4870]: I1014 07:23:48.968113 4870 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3927fb78-4f30-4371-8fbb-62fd276c0460-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:48 crc kubenswrapper[4870]: I1014 07:23:48.968140 4870 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3927fb78-4f30-4371-8fbb-62fd276c0460-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:48 crc kubenswrapper[4870]: I1014 07:23:48.968152 4870 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bad45b30-5ba2-4062-ab14-04f923679c4e-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:48 crc kubenswrapper[4870]: I1014 07:23:48.968178 4870 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c603f7b5-015f-4c2e-9144-b3e8349c3da3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:48 crc kubenswrapper[4870]: I1014 07:23:48.968194 4870 reconciler_common.go:293] "Volume detached for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/c603f7b5-015f-4c2e-9144-b3e8349c3da3-kube-state-metrics-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.021728 4870 scope.go:117] "RemoveContainer" containerID="51cc7a5037c45783ceb5efa6e804746e76684dbe52ba7dd329814c5da8fb28db" Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.042376 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 14 07:23:49 crc kubenswrapper[4870]: E1014 07:23:49.141565 4870 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="457905ca1130c3804c068cf6fc44521ad064a3fb34449d666cf858f14e2feace" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Oct 14 07:23:49 crc kubenswrapper[4870]: E1014 07:23:49.144382 4870 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="457905ca1130c3804c068cf6fc44521ad064a3fb34449d666cf858f14e2feace" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Oct 14 07:23:49 crc kubenswrapper[4870]: E1014 07:23:49.145714 4870 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="457905ca1130c3804c068cf6fc44521ad064a3fb34449d666cf858f14e2feace" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Oct 14 07:23:49 crc kubenswrapper[4870]: E1014 07:23:49.145745 4870 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-northd-0" podUID="416d7b39-4541-44a8-a55f-924bc86fee32" containerName="ovn-northd" Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.179801 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/51de8090-c48e-498c-99ab-794a7527252c-config-data-custom\") pod \"51de8090-c48e-498c-99ab-794a7527252c\" (UID: \"51de8090-c48e-498c-99ab-794a7527252c\") " Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.185882 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/51de8090-c48e-498c-99ab-794a7527252c-etc-machine-id\") pod \"51de8090-c48e-498c-99ab-794a7527252c\" (UID: \"51de8090-c48e-498c-99ab-794a7527252c\") " Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.185997 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/51de8090-c48e-498c-99ab-794a7527252c-public-tls-certs\") pod \"51de8090-c48e-498c-99ab-794a7527252c\" (UID: \"51de8090-c48e-498c-99ab-794a7527252c\") " Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.185999 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/51de8090-c48e-498c-99ab-794a7527252c-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "51de8090-c48e-498c-99ab-794a7527252c" (UID: "51de8090-c48e-498c-99ab-794a7527252c"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.186022 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/51de8090-c48e-498c-99ab-794a7527252c-config-data\") pod \"51de8090-c48e-498c-99ab-794a7527252c\" (UID: \"51de8090-c48e-498c-99ab-794a7527252c\") " Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.186058 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4dhzj\" (UniqueName: \"kubernetes.io/projected/51de8090-c48e-498c-99ab-794a7527252c-kube-api-access-4dhzj\") pod \"51de8090-c48e-498c-99ab-794a7527252c\" (UID: \"51de8090-c48e-498c-99ab-794a7527252c\") " Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.186181 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/51de8090-c48e-498c-99ab-794a7527252c-internal-tls-certs\") pod \"51de8090-c48e-498c-99ab-794a7527252c\" (UID: \"51de8090-c48e-498c-99ab-794a7527252c\") " Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.186268 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/51de8090-c48e-498c-99ab-794a7527252c-logs\") pod \"51de8090-c48e-498c-99ab-794a7527252c\" (UID: \"51de8090-c48e-498c-99ab-794a7527252c\") " Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.186298 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/51de8090-c48e-498c-99ab-794a7527252c-scripts\") pod \"51de8090-c48e-498c-99ab-794a7527252c\" (UID: \"51de8090-c48e-498c-99ab-794a7527252c\") " Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.186330 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51de8090-c48e-498c-99ab-794a7527252c-combined-ca-bundle\") pod \"51de8090-c48e-498c-99ab-794a7527252c\" (UID: \"51de8090-c48e-498c-99ab-794a7527252c\") " Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.187040 4870 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/51de8090-c48e-498c-99ab-794a7527252c-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.191685 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/51de8090-c48e-498c-99ab-794a7527252c-logs" (OuterVolumeSpecName: "logs") pod "51de8090-c48e-498c-99ab-794a7527252c" (UID: "51de8090-c48e-498c-99ab-794a7527252c"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.193755 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/51de8090-c48e-498c-99ab-794a7527252c-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "51de8090-c48e-498c-99ab-794a7527252c" (UID: "51de8090-c48e-498c-99ab-794a7527252c"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.217783 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25f5ffd0-3c55-46ee-ade3-ec4cd9471e21" path="/var/lib/kubelet/pods/25f5ffd0-3c55-46ee-ade3-ec4cd9471e21/volumes" Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.218637 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="79dfe32a-c048-41d8-aead-de78ae909cd8" path="/var/lib/kubelet/pods/79dfe32a-c048-41d8-aead-de78ae909cd8/volumes" Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.219282 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d318a57f-fb03-4112-b027-b8eaf943b00d" path="/var/lib/kubelet/pods/d318a57f-fb03-4112-b027-b8eaf943b00d/volumes" Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.219858 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fe66a2e1-364a-4cb9-b857-c208c7d1b108" path="/var/lib/kubelet/pods/fe66a2e1-364a-4cb9-b857-c208c7d1b108/volumes" Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.240672 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/51de8090-c48e-498c-99ab-794a7527252c-scripts" (OuterVolumeSpecName: "scripts") pod "51de8090-c48e-498c-99ab-794a7527252c" (UID: "51de8090-c48e-498c-99ab-794a7527252c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.288802 4870 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/51de8090-c48e-498c-99ab-794a7527252c-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.288831 4870 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/51de8090-c48e-498c-99ab-794a7527252c-logs\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.288841 4870 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/51de8090-c48e-498c-99ab-794a7527252c-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.295030 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/51de8090-c48e-498c-99ab-794a7527252c-kube-api-access-4dhzj" (OuterVolumeSpecName: "kube-api-access-4dhzj") pod "51de8090-c48e-498c-99ab-794a7527252c" (UID: "51de8090-c48e-498c-99ab-794a7527252c"). InnerVolumeSpecName "kube-api-access-4dhzj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.310100 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/51de8090-c48e-498c-99ab-794a7527252c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "51de8090-c48e-498c-99ab-794a7527252c" (UID: "51de8090-c48e-498c-99ab-794a7527252c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.318382 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/51de8090-c48e-498c-99ab-794a7527252c-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "51de8090-c48e-498c-99ab-794a7527252c" (UID: "51de8090-c48e-498c-99ab-794a7527252c"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.332025 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/51de8090-c48e-498c-99ab-794a7527252c-config-data" (OuterVolumeSpecName: "config-data") pod "51de8090-c48e-498c-99ab-794a7527252c" (UID: "51de8090-c48e-498c-99ab-794a7527252c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.332063 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/51de8090-c48e-498c-99ab-794a7527252c-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "51de8090-c48e-498c-99ab-794a7527252c" (UID: "51de8090-c48e-498c-99ab-794a7527252c"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.395481 4870 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/51de8090-c48e-498c-99ab-794a7527252c-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.395509 4870 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51de8090-c48e-498c-99ab-794a7527252c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.395520 4870 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/51de8090-c48e-498c-99ab-794a7527252c-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.395531 4870 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/51de8090-c48e-498c-99ab-794a7527252c-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.395542 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4dhzj\" (UniqueName: \"kubernetes.io/projected/51de8090-c48e-498c-99ab-794a7527252c-kube-api-access-4dhzj\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:49 crc kubenswrapper[4870]: E1014 07:23:49.460744 4870 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda6f02e4f_8dca_4b97_91e9_ee3368e5e44b.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda6f02e4f_8dca_4b97_91e9_ee3368e5e44b.slice/crio-1776e2a9afb3ac3237ca85fa6c702e235efdc65ae828037d00005680033973a2\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8abede26_36e8_46ec_a5c4_6cd53319133a.slice/crio-5a4ce260cbfb825ad1358d2122d60517f96cbcdbd83613ef2c1f69ea0ec0bf41\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8abede26_36e8_46ec_a5c4_6cd53319133a.slice\": RecentStats: unable to find data in memory cache]" Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.535300 4870 scope.go:117] "RemoveContainer" containerID="c79902c86d1a79331c9994ac2e40ba61e03d57293be1023f0938cc002e01194a" Oct 14 07:23:49 crc kubenswrapper[4870]: E1014 07:23:49.627391 4870 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 19bfdce54348e895d0d26a8651c758e77337522908dc3ad6eb2470e00c72c2cf is running failed: container process not found" containerID="19bfdce54348e895d0d26a8651c758e77337522908dc3ad6eb2470e00c72c2cf" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 14 07:23:49 crc kubenswrapper[4870]: E1014 07:23:49.628412 4870 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 19bfdce54348e895d0d26a8651c758e77337522908dc3ad6eb2470e00c72c2cf is running failed: container process not found" containerID="19bfdce54348e895d0d26a8651c758e77337522908dc3ad6eb2470e00c72c2cf" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 14 07:23:49 crc kubenswrapper[4870]: E1014 07:23:49.637581 4870 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 19bfdce54348e895d0d26a8651c758e77337522908dc3ad6eb2470e00c72c2cf is running failed: container process not found" containerID="19bfdce54348e895d0d26a8651c758e77337522908dc3ad6eb2470e00c72c2cf" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 14 07:23:49 crc kubenswrapper[4870]: E1014 07:23:49.637687 4870 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 19bfdce54348e895d0d26a8651c758e77337522908dc3ad6eb2470e00c72c2cf is running failed: container process not found" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="f5b700ed-3919-458f-860e-5630b8e4b800" containerName="nova-scheduler-scheduler" Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.637805 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.656798 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder5c98-account-delete-lhjjg"] Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.662521 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.662820 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder5c98-account-delete-lhjjg"] Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.670677 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.677295 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.685868 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.692908 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.699811 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/56eb286f-8aa2-46aa-a74d-031288a9aad1-scripts\") pod \"56eb286f-8aa2-46aa-a74d-031288a9aad1\" (UID: \"56eb286f-8aa2-46aa-a74d-031288a9aad1\") " Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.699886 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4w2qr\" (UniqueName: \"kubernetes.io/projected/56eb286f-8aa2-46aa-a74d-031288a9aad1-kube-api-access-4w2qr\") pod \"56eb286f-8aa2-46aa-a74d-031288a9aad1\" (UID: \"56eb286f-8aa2-46aa-a74d-031288a9aad1\") " Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.699919 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"56eb286f-8aa2-46aa-a74d-031288a9aad1\" (UID: \"56eb286f-8aa2-46aa-a74d-031288a9aad1\") " Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.699990 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/56eb286f-8aa2-46aa-a74d-031288a9aad1-internal-tls-certs\") pod \"56eb286f-8aa2-46aa-a74d-031288a9aad1\" (UID: \"56eb286f-8aa2-46aa-a74d-031288a9aad1\") " Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.700024 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/56eb286f-8aa2-46aa-a74d-031288a9aad1-config-data\") pod \"56eb286f-8aa2-46aa-a74d-031288a9aad1\" (UID: \"56eb286f-8aa2-46aa-a74d-031288a9aad1\") " Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.700088 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56eb286f-8aa2-46aa-a74d-031288a9aad1-combined-ca-bundle\") pod \"56eb286f-8aa2-46aa-a74d-031288a9aad1\" (UID: \"56eb286f-8aa2-46aa-a74d-031288a9aad1\") " Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.700218 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/56eb286f-8aa2-46aa-a74d-031288a9aad1-logs\") pod \"56eb286f-8aa2-46aa-a74d-031288a9aad1\" (UID: \"56eb286f-8aa2-46aa-a74d-031288a9aad1\") " Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.700271 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/56eb286f-8aa2-46aa-a74d-031288a9aad1-httpd-run\") pod \"56eb286f-8aa2-46aa-a74d-031288a9aad1\" (UID: \"56eb286f-8aa2-46aa-a74d-031288a9aad1\") " Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.701238 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/56eb286f-8aa2-46aa-a74d-031288a9aad1-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "56eb286f-8aa2-46aa-a74d-031288a9aad1" (UID: "56eb286f-8aa2-46aa-a74d-031288a9aad1"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.701492 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/56eb286f-8aa2-46aa-a74d-031288a9aad1-logs" (OuterVolumeSpecName: "logs") pod "56eb286f-8aa2-46aa-a74d-031288a9aad1" (UID: "56eb286f-8aa2-46aa-a74d-031288a9aad1"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.706851 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/56eb286f-8aa2-46aa-a74d-031288a9aad1-scripts" (OuterVolumeSpecName: "scripts") pod "56eb286f-8aa2-46aa-a74d-031288a9aad1" (UID: "56eb286f-8aa2-46aa-a74d-031288a9aad1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.718753 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-597ff968b-msxbn" Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.721310 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/56eb286f-8aa2-46aa-a74d-031288a9aad1-kube-api-access-4w2qr" (OuterVolumeSpecName: "kube-api-access-4w2qr") pod "56eb286f-8aa2-46aa-a74d-031288a9aad1" (UID: "56eb286f-8aa2-46aa-a74d-031288a9aad1"). InnerVolumeSpecName "kube-api-access-4w2qr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.722517 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage06-crc" (OuterVolumeSpecName: "glance") pod "56eb286f-8aa2-46aa-a74d-031288a9aad1" (UID: "56eb286f-8aa2-46aa-a74d-031288a9aad1"). InnerVolumeSpecName "local-storage06-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.722705 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.732368 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/56eb286f-8aa2-46aa-a74d-031288a9aad1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "56eb286f-8aa2-46aa-a74d-031288a9aad1" (UID: "56eb286f-8aa2-46aa-a74d-031288a9aad1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.734500 4870 scope.go:117] "RemoveContainer" containerID="c79902c86d1a79331c9994ac2e40ba61e03d57293be1023f0938cc002e01194a" Oct 14 07:23:49 crc kubenswrapper[4870]: E1014 07:23:49.735228 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c79902c86d1a79331c9994ac2e40ba61e03d57293be1023f0938cc002e01194a\": container with ID starting with c79902c86d1a79331c9994ac2e40ba61e03d57293be1023f0938cc002e01194a not found: ID does not exist" containerID="c79902c86d1a79331c9994ac2e40ba61e03d57293be1023f0938cc002e01194a" Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.735320 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c79902c86d1a79331c9994ac2e40ba61e03d57293be1023f0938cc002e01194a"} err="failed to get container status \"c79902c86d1a79331c9994ac2e40ba61e03d57293be1023f0938cc002e01194a\": rpc error: code = NotFound desc = could not find container \"c79902c86d1a79331c9994ac2e40ba61e03d57293be1023f0938cc002e01194a\": container with ID starting with c79902c86d1a79331c9994ac2e40ba61e03d57293be1023f0938cc002e01194a not found: ID does not exist" Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.735422 4870 scope.go:117] "RemoveContainer" containerID="769de1c723d5f9ddc1f6d062434e4b09d67bad6377a8cd0e80434875d4f6a655" Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.739210 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-79cc757548-zxf4g" Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.748210 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.751082 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.755810 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/novaapie66c-account-delete-ftd24" Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.763792 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/novacell0ce16-account-delete-bwgsq"] Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.770790 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/novacell0ce16-account-delete-bwgsq"] Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.777429 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican6547-account-delete-v5crl"] Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.783621 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican6547-account-delete-v5crl"] Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.787119 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/56eb286f-8aa2-46aa-a74d-031288a9aad1-config-data" (OuterVolumeSpecName: "config-data") pod "56eb286f-8aa2-46aa-a74d-031288a9aad1" (UID: "56eb286f-8aa2-46aa-a74d-031288a9aad1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.790571 4870 scope.go:117] "RemoveContainer" containerID="e6478418e9645b5d3cf9a2a8e1b1a4090b34dcf697735e42cf6fb41d3708bbeb" Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.809653 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6cpz2\" (UniqueName: \"kubernetes.io/projected/f0b766fd-ccb9-4620-84df-8fe2f194b0da-kube-api-access-6cpz2\") pod \"f0b766fd-ccb9-4620-84df-8fe2f194b0da\" (UID: \"f0b766fd-ccb9-4620-84df-8fe2f194b0da\") " Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.809698 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/f0b766fd-ccb9-4620-84df-8fe2f194b0da-nova-metadata-tls-certs\") pod \"f0b766fd-ccb9-4620-84df-8fe2f194b0da\" (UID: \"f0b766fd-ccb9-4620-84df-8fe2f194b0da\") " Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.809725 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/cae5b655-5d7c-4063-90ad-4d9913792d5b-internal-tls-certs\") pod \"cae5b655-5d7c-4063-90ad-4d9913792d5b\" (UID: \"cae5b655-5d7c-4063-90ad-4d9913792d5b\") " Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.809773 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/3bd1e649-3d09-4cda-921f-ac9a84a5066e-kolla-config\") pod \"3bd1e649-3d09-4cda-921f-ac9a84a5066e\" (UID: \"3bd1e649-3d09-4cda-921f-ac9a84a5066e\") " Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.809827 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f0b766fd-ccb9-4620-84df-8fe2f194b0da-combined-ca-bundle\") pod \"f0b766fd-ccb9-4620-84df-8fe2f194b0da\" (UID: \"f0b766fd-ccb9-4620-84df-8fe2f194b0da\") " Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.809854 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cae5b655-5d7c-4063-90ad-4d9913792d5b-logs\") pod \"cae5b655-5d7c-4063-90ad-4d9913792d5b\" (UID: \"cae5b655-5d7c-4063-90ad-4d9913792d5b\") " Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.809873 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2130e3f2-cbea-46e1-b5d3-daefdd972935-logs\") pod \"2130e3f2-cbea-46e1-b5d3-daefdd972935\" (UID: \"2130e3f2-cbea-46e1-b5d3-daefdd972935\") " Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.809904 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2130e3f2-cbea-46e1-b5d3-daefdd972935-combined-ca-bundle\") pod \"2130e3f2-cbea-46e1-b5d3-daefdd972935\" (UID: \"2130e3f2-cbea-46e1-b5d3-daefdd972935\") " Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.809923 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c712b87a-cf4d-437e-abe0-1a566046ca66-combined-ca-bundle\") pod \"c712b87a-cf4d-437e-abe0-1a566046ca66\" (UID: \"c712b87a-cf4d-437e-abe0-1a566046ca66\") " Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.809943 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2130e3f2-cbea-46e1-b5d3-daefdd972935-public-tls-certs\") pod \"2130e3f2-cbea-46e1-b5d3-daefdd972935\" (UID: \"2130e3f2-cbea-46e1-b5d3-daefdd972935\") " Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.809969 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c712b87a-cf4d-437e-abe0-1a566046ca66-public-tls-certs\") pod \"c712b87a-cf4d-437e-abe0-1a566046ca66\" (UID: \"c712b87a-cf4d-437e-abe0-1a566046ca66\") " Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.810123 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3bd1e649-3d09-4cda-921f-ac9a84a5066e-config-data\") pod \"3bd1e649-3d09-4cda-921f-ac9a84a5066e\" (UID: \"3bd1e649-3d09-4cda-921f-ac9a84a5066e\") " Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.810147 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2130e3f2-cbea-46e1-b5d3-daefdd972935-scripts\") pod \"2130e3f2-cbea-46e1-b5d3-daefdd972935\" (UID: \"2130e3f2-cbea-46e1-b5d3-daefdd972935\") " Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.810167 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c712b87a-cf4d-437e-abe0-1a566046ca66-config-data\") pod \"c712b87a-cf4d-437e-abe0-1a566046ca66\" (UID: \"c712b87a-cf4d-437e-abe0-1a566046ca66\") " Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.810188 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"c712b87a-cf4d-437e-abe0-1a566046ca66\" (UID: \"c712b87a-cf4d-437e-abe0-1a566046ca66\") " Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.810215 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bgmmx\" (UniqueName: \"kubernetes.io/projected/3bd1e649-3d09-4cda-921f-ac9a84a5066e-kube-api-access-bgmmx\") pod \"3bd1e649-3d09-4cda-921f-ac9a84a5066e\" (UID: \"3bd1e649-3d09-4cda-921f-ac9a84a5066e\") " Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.810241 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hzjsv\" (UniqueName: \"kubernetes.io/projected/c712b87a-cf4d-437e-abe0-1a566046ca66-kube-api-access-hzjsv\") pod \"c712b87a-cf4d-437e-abe0-1a566046ca66\" (UID: \"c712b87a-cf4d-437e-abe0-1a566046ca66\") " Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.810270 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c712b87a-cf4d-437e-abe0-1a566046ca66-scripts\") pod \"c712b87a-cf4d-437e-abe0-1a566046ca66\" (UID: \"c712b87a-cf4d-437e-abe0-1a566046ca66\") " Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.810299 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3bd1e649-3d09-4cda-921f-ac9a84a5066e-combined-ca-bundle\") pod \"3bd1e649-3d09-4cda-921f-ac9a84a5066e\" (UID: \"3bd1e649-3d09-4cda-921f-ac9a84a5066e\") " Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.810319 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c712b87a-cf4d-437e-abe0-1a566046ca66-httpd-run\") pod \"c712b87a-cf4d-437e-abe0-1a566046ca66\" (UID: \"c712b87a-cf4d-437e-abe0-1a566046ca66\") " Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.810334 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2130e3f2-cbea-46e1-b5d3-daefdd972935-config-data\") pod \"2130e3f2-cbea-46e1-b5d3-daefdd972935\" (UID: \"2130e3f2-cbea-46e1-b5d3-daefdd972935\") " Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.810357 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cae5b655-5d7c-4063-90ad-4d9913792d5b-public-tls-certs\") pod \"cae5b655-5d7c-4063-90ad-4d9913792d5b\" (UID: \"cae5b655-5d7c-4063-90ad-4d9913792d5b\") " Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.810375 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2130e3f2-cbea-46e1-b5d3-daefdd972935-internal-tls-certs\") pod \"2130e3f2-cbea-46e1-b5d3-daefdd972935\" (UID: \"2130e3f2-cbea-46e1-b5d3-daefdd972935\") " Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.810403 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cae5b655-5d7c-4063-90ad-4d9913792d5b-config-data-custom\") pod \"cae5b655-5d7c-4063-90ad-4d9913792d5b\" (UID: \"cae5b655-5d7c-4063-90ad-4d9913792d5b\") " Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.810426 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c712b87a-cf4d-437e-abe0-1a566046ca66-logs\") pod \"c712b87a-cf4d-437e-abe0-1a566046ca66\" (UID: \"c712b87a-cf4d-437e-abe0-1a566046ca66\") " Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.810487 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cae5b655-5d7c-4063-90ad-4d9913792d5b-config-data\") pod \"cae5b655-5d7c-4063-90ad-4d9913792d5b\" (UID: \"cae5b655-5d7c-4063-90ad-4d9913792d5b\") " Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.810532 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgsbd\" (UniqueName: \"kubernetes.io/projected/cae5b655-5d7c-4063-90ad-4d9913792d5b-kube-api-access-zgsbd\") pod \"cae5b655-5d7c-4063-90ad-4d9913792d5b\" (UID: \"cae5b655-5d7c-4063-90ad-4d9913792d5b\") " Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.810563 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8rt82\" (UniqueName: \"kubernetes.io/projected/2130e3f2-cbea-46e1-b5d3-daefdd972935-kube-api-access-8rt82\") pod \"2130e3f2-cbea-46e1-b5d3-daefdd972935\" (UID: \"2130e3f2-cbea-46e1-b5d3-daefdd972935\") " Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.810590 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/3bd1e649-3d09-4cda-921f-ac9a84a5066e-memcached-tls-certs\") pod \"3bd1e649-3d09-4cda-921f-ac9a84a5066e\" (UID: \"3bd1e649-3d09-4cda-921f-ac9a84a5066e\") " Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.810618 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f0b766fd-ccb9-4620-84df-8fe2f194b0da-logs\") pod \"f0b766fd-ccb9-4620-84df-8fe2f194b0da\" (UID: \"f0b766fd-ccb9-4620-84df-8fe2f194b0da\") " Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.810639 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cae5b655-5d7c-4063-90ad-4d9913792d5b-combined-ca-bundle\") pod \"cae5b655-5d7c-4063-90ad-4d9913792d5b\" (UID: \"cae5b655-5d7c-4063-90ad-4d9913792d5b\") " Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.810656 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f0b766fd-ccb9-4620-84df-8fe2f194b0da-config-data\") pod \"f0b766fd-ccb9-4620-84df-8fe2f194b0da\" (UID: \"f0b766fd-ccb9-4620-84df-8fe2f194b0da\") " Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.811078 4870 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/56eb286f-8aa2-46aa-a74d-031288a9aad1-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.811094 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4w2qr\" (UniqueName: \"kubernetes.io/projected/56eb286f-8aa2-46aa-a74d-031288a9aad1-kube-api-access-4w2qr\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.811115 4870 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" " Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.811124 4870 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/56eb286f-8aa2-46aa-a74d-031288a9aad1-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.811134 4870 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56eb286f-8aa2-46aa-a74d-031288a9aad1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.811143 4870 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/56eb286f-8aa2-46aa-a74d-031288a9aad1-logs\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.811151 4870 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/56eb286f-8aa2-46aa-a74d-031288a9aad1-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.812399 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c712b87a-cf4d-437e-abe0-1a566046ca66-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "c712b87a-cf4d-437e-abe0-1a566046ca66" (UID: "c712b87a-cf4d-437e-abe0-1a566046ca66"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.823274 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c712b87a-cf4d-437e-abe0-1a566046ca66-logs" (OuterVolumeSpecName: "logs") pod "c712b87a-cf4d-437e-abe0-1a566046ca66" (UID: "c712b87a-cf4d-437e-abe0-1a566046ca66"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.842666 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3bd1e649-3d09-4cda-921f-ac9a84a5066e-kolla-config" (OuterVolumeSpecName: "kolla-config") pod "3bd1e649-3d09-4cda-921f-ac9a84a5066e" (UID: "3bd1e649-3d09-4cda-921f-ac9a84a5066e"). InnerVolumeSpecName "kolla-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.844884 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f0b766fd-ccb9-4620-84df-8fe2f194b0da-logs" (OuterVolumeSpecName: "logs") pod "f0b766fd-ccb9-4620-84df-8fe2f194b0da" (UID: "f0b766fd-ccb9-4620-84df-8fe2f194b0da"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.847748 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "glance") pod "c712b87a-cf4d-437e-abe0-1a566046ca66" (UID: "c712b87a-cf4d-437e-abe0-1a566046ca66"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.849263 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cae5b655-5d7c-4063-90ad-4d9913792d5b-logs" (OuterVolumeSpecName: "logs") pod "cae5b655-5d7c-4063-90ad-4d9913792d5b" (UID: "cae5b655-5d7c-4063-90ad-4d9913792d5b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.849773 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2130e3f2-cbea-46e1-b5d3-daefdd972935-logs" (OuterVolumeSpecName: "logs") pod "2130e3f2-cbea-46e1-b5d3-daefdd972935" (UID: "2130e3f2-cbea-46e1-b5d3-daefdd972935"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.850930 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cae5b655-5d7c-4063-90ad-4d9913792d5b-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "cae5b655-5d7c-4063-90ad-4d9913792d5b" (UID: "cae5b655-5d7c-4063-90ad-4d9913792d5b"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.851293 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f0b766fd-ccb9-4620-84df-8fe2f194b0da-kube-api-access-6cpz2" (OuterVolumeSpecName: "kube-api-access-6cpz2") pod "f0b766fd-ccb9-4620-84df-8fe2f194b0da" (UID: "f0b766fd-ccb9-4620-84df-8fe2f194b0da"). InnerVolumeSpecName "kube-api-access-6cpz2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.853996 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3bd1e649-3d09-4cda-921f-ac9a84a5066e-config-data" (OuterVolumeSpecName: "config-data") pod "3bd1e649-3d09-4cda-921f-ac9a84a5066e" (UID: "3bd1e649-3d09-4cda-921f-ac9a84a5066e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.875314 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c712b87a-cf4d-437e-abe0-1a566046ca66-kube-api-access-hzjsv" (OuterVolumeSpecName: "kube-api-access-hzjsv") pod "c712b87a-cf4d-437e-abe0-1a566046ca66" (UID: "c712b87a-cf4d-437e-abe0-1a566046ca66"). InnerVolumeSpecName "kube-api-access-hzjsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.879231 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.884629 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3bd1e649-3d09-4cda-921f-ac9a84a5066e-kube-api-access-bgmmx" (OuterVolumeSpecName: "kube-api-access-bgmmx") pod "3bd1e649-3d09-4cda-921f-ac9a84a5066e" (UID: "3bd1e649-3d09-4cda-921f-ac9a84a5066e"). InnerVolumeSpecName "kube-api-access-bgmmx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.888745 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cae5b655-5d7c-4063-90ad-4d9913792d5b-kube-api-access-zgsbd" (OuterVolumeSpecName: "kube-api-access-zgsbd") pod "cae5b655-5d7c-4063-90ad-4d9913792d5b" (UID: "cae5b655-5d7c-4063-90ad-4d9913792d5b"). InnerVolumeSpecName "kube-api-access-zgsbd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.889975 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c712b87a-cf4d-437e-abe0-1a566046ca66-scripts" (OuterVolumeSpecName: "scripts") pod "c712b87a-cf4d-437e-abe0-1a566046ca66" (UID: "c712b87a-cf4d-437e-abe0-1a566046ca66"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.890557 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2130e3f2-cbea-46e1-b5d3-daefdd972935-kube-api-access-8rt82" (OuterVolumeSpecName: "kube-api-access-8rt82") pod "2130e3f2-cbea-46e1-b5d3-daefdd972935" (UID: "2130e3f2-cbea-46e1-b5d3-daefdd972935"). InnerVolumeSpecName "kube-api-access-8rt82". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.905756 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"c712b87a-cf4d-437e-abe0-1a566046ca66","Type":"ContainerDied","Data":"c7369a66d5079dd52c04fcb0951209f2fbdc37e0b77e054874a3ae686da65c90"} Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.905844 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.908947 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.918077 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.923185 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2130e3f2-cbea-46e1-b5d3-daefdd972935-scripts" (OuterVolumeSpecName: "scripts") pod "2130e3f2-cbea-46e1-b5d3-daefdd972935" (UID: "2130e3f2-cbea-46e1-b5d3-daefdd972935"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.924547 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.924654 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f0b766fd-ccb9-4620-84df-8fe2f194b0da","Type":"ContainerDied","Data":"f20638fcf7487657845b9e4bea7d6a1a36ecd836a895de79af97612151ecb066"} Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.924813 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ff28d060-b280-4b89-bea1-d10aa367a3cc-public-tls-certs\") pod \"ff28d060-b280-4b89-bea1-d10aa367a3cc\" (UID: \"ff28d060-b280-4b89-bea1-d10aa367a3cc\") " Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.926081 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pvhxn\" (UniqueName: \"kubernetes.io/projected/ff28d060-b280-4b89-bea1-d10aa367a3cc-kube-api-access-pvhxn\") pod \"ff28d060-b280-4b89-bea1-d10aa367a3cc\" (UID: \"ff28d060-b280-4b89-bea1-d10aa367a3cc\") " Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.926168 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ff28d060-b280-4b89-bea1-d10aa367a3cc-internal-tls-certs\") pod \"ff28d060-b280-4b89-bea1-d10aa367a3cc\" (UID: \"ff28d060-b280-4b89-bea1-d10aa367a3cc\") " Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.926219 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff28d060-b280-4b89-bea1-d10aa367a3cc-config-data\") pod \"ff28d060-b280-4b89-bea1-d10aa367a3cc\" (UID: \"ff28d060-b280-4b89-bea1-d10aa367a3cc\") " Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.926301 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ff28d060-b280-4b89-bea1-d10aa367a3cc-logs\") pod \"ff28d060-b280-4b89-bea1-d10aa367a3cc\" (UID: \"ff28d060-b280-4b89-bea1-d10aa367a3cc\") " Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.926478 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-582bv\" (UniqueName: \"kubernetes.io/projected/208754ea-7cc0-4222-a186-9ba01b274add-kube-api-access-582bv\") pod \"208754ea-7cc0-4222-a186-9ba01b274add\" (UID: \"208754ea-7cc0-4222-a186-9ba01b274add\") " Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.926540 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff28d060-b280-4b89-bea1-d10aa367a3cc-combined-ca-bundle\") pod \"ff28d060-b280-4b89-bea1-d10aa367a3cc\" (UID: \"ff28d060-b280-4b89-bea1-d10aa367a3cc\") " Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.927780 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.929384 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ff28d060-b280-4b89-bea1-d10aa367a3cc-logs" (OuterVolumeSpecName: "logs") pod "ff28d060-b280-4b89-bea1-d10aa367a3cc" (UID: "ff28d060-b280-4b89-bea1-d10aa367a3cc"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.942424 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hzjsv\" (UniqueName: \"kubernetes.io/projected/c712b87a-cf4d-437e-abe0-1a566046ca66-kube-api-access-hzjsv\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.950019 4870 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c712b87a-cf4d-437e-abe0-1a566046ca66-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.947875 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.954749 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.956561 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"524b0b2a-77a2-4c05-be39-c8b97666ab11","Type":"ContainerDied","Data":"92059a83ebd09e581c802c9d18a6cc53c68a57ad45e7c0b7e480b8f24d3f2bd8"} Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.951627 4870 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c712b87a-cf4d-437e-abe0-1a566046ca66-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.957294 4870 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ff28d060-b280-4b89-bea1-d10aa367a3cc-logs\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.957396 4870 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cae5b655-5d7c-4063-90ad-4d9913792d5b-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.957493 4870 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c712b87a-cf4d-437e-abe0-1a566046ca66-logs\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.957551 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgsbd\" (UniqueName: \"kubernetes.io/projected/cae5b655-5d7c-4063-90ad-4d9913792d5b-kube-api-access-zgsbd\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.957623 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8rt82\" (UniqueName: \"kubernetes.io/projected/2130e3f2-cbea-46e1-b5d3-daefdd972935-kube-api-access-8rt82\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.957680 4870 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f0b766fd-ccb9-4620-84df-8fe2f194b0da-logs\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.957734 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6cpz2\" (UniqueName: \"kubernetes.io/projected/f0b766fd-ccb9-4620-84df-8fe2f194b0da-kube-api-access-6cpz2\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.957786 4870 reconciler_common.go:293] "Volume detached for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/3bd1e649-3d09-4cda-921f-ac9a84a5066e-kolla-config\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.957840 4870 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cae5b655-5d7c-4063-90ad-4d9913792d5b-logs\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.957898 4870 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2130e3f2-cbea-46e1-b5d3-daefdd972935-logs\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.957952 4870 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3bd1e649-3d09-4cda-921f-ac9a84a5066e-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.958011 4870 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2130e3f2-cbea-46e1-b5d3-daefdd972935-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.958080 4870 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.958155 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bgmmx\" (UniqueName: \"kubernetes.io/projected/3bd1e649-3d09-4cda-921f-ac9a84a5066e-kube-api-access-bgmmx\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.954324 4870 generic.go:334] "Generic (PLEG): container finished" podID="524b0b2a-77a2-4c05-be39-c8b97666ab11" containerID="92059a83ebd09e581c802c9d18a6cc53c68a57ad45e7c0b7e480b8f24d3f2bd8" exitCode=0 Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.958469 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"524b0b2a-77a2-4c05-be39-c8b97666ab11","Type":"ContainerDied","Data":"3e53f14bf4af7bc198470fc2fda53430a9cde12e3adbf27cbd98bbaf80b6393b"} Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.958518 4870 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3e53f14bf4af7bc198470fc2fda53430a9cde12e3adbf27cbd98bbaf80b6393b" Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.961587 4870 generic.go:334] "Generic (PLEG): container finished" podID="7bc76b72-8679-4b33-8966-0ee70c1788e5" containerID="e5c43911d04a168b064b8b753711f3e5ed521dc55a16d5687c49934c8d0eb3a6" exitCode=0 Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.961628 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"7bc76b72-8679-4b33-8966-0ee70c1788e5","Type":"ContainerDied","Data":"e5c43911d04a168b064b8b753711f3e5ed521dc55a16d5687c49934c8d0eb3a6"} Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.962806 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"3bd1e649-3d09-4cda-921f-ac9a84a5066e","Type":"ContainerDied","Data":"052bde8548ecfddf46ee9484a1311dff89a89809aee66740b22b81c7a218cd25"} Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.962921 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.963668 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/208754ea-7cc0-4222-a186-9ba01b274add-kube-api-access-582bv" (OuterVolumeSpecName: "kube-api-access-582bv") pod "208754ea-7cc0-4222-a186-9ba01b274add" (UID: "208754ea-7cc0-4222-a186-9ba01b274add"). InnerVolumeSpecName "kube-api-access-582bv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.969935 4870 generic.go:334] "Generic (PLEG): container finished" podID="ff28d060-b280-4b89-bea1-d10aa367a3cc" containerID="9867cb00fe7f9cff6394884f9edfc786d6b848597f09a56abffe01cb5e8ca4cf" exitCode=0 Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.970016 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ff28d060-b280-4b89-bea1-d10aa367a3cc","Type":"ContainerDied","Data":"9867cb00fe7f9cff6394884f9edfc786d6b848597f09a56abffe01cb5e8ca4cf"} Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.970079 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ff28d060-b280-4b89-bea1-d10aa367a3cc","Type":"ContainerDied","Data":"d2c1e6ab1ae6834143ee40554a51da1a2d368bfb5f191a9abce3ca8a5ee4d09a"} Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.970145 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.972045 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ff28d060-b280-4b89-bea1-d10aa367a3cc-kube-api-access-pvhxn" (OuterVolumeSpecName: "kube-api-access-pvhxn") pod "ff28d060-b280-4b89-bea1-d10aa367a3cc" (UID: "ff28d060-b280-4b89-bea1-d10aa367a3cc"). InnerVolumeSpecName "kube-api-access-pvhxn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.972607 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novaapie66c-account-delete-ftd24" event={"ID":"208754ea-7cc0-4222-a186-9ba01b274add","Type":"ContainerDied","Data":"719f5c94f8b4f323a4cf25aa23aedc41ecbda25519cbe0736cf69b37d5ab0c75"} Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.972639 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/novaapie66c-account-delete-ftd24" Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.976205 4870 generic.go:334] "Generic (PLEG): container finished" podID="f5b700ed-3919-458f-860e-5630b8e4b800" containerID="19bfdce54348e895d0d26a8651c758e77337522908dc3ad6eb2470e00c72c2cf" exitCode=0 Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.976263 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"f5b700ed-3919-458f-860e-5630b8e4b800","Type":"ContainerDied","Data":"19bfdce54348e895d0d26a8651c758e77337522908dc3ad6eb2470e00c72c2cf"} Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.977369 4870 generic.go:334] "Generic (PLEG): container finished" podID="cae5b655-5d7c-4063-90ad-4d9913792d5b" containerID="b97a3d2764961f388fdc1a6459d91ca5f1248920ce92c572b651657c109e7fa4" exitCode=0 Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.977399 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-79cc757548-zxf4g" event={"ID":"cae5b655-5d7c-4063-90ad-4d9913792d5b","Type":"ContainerDied","Data":"b97a3d2764961f388fdc1a6459d91ca5f1248920ce92c572b651657c109e7fa4"} Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.977415 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-79cc757548-zxf4g" event={"ID":"cae5b655-5d7c-4063-90ad-4d9913792d5b","Type":"ContainerDied","Data":"1348cc0e9dccef391f47727a7b0673d69bdf58a16020dc4ffad5b01888b723b3"} Oct 14 07:23:49 crc kubenswrapper[4870]: I1014 07:23:49.977484 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-79cc757548-zxf4g" Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.035823 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/56eb286f-8aa2-46aa-a74d-031288a9aad1-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "56eb286f-8aa2-46aa-a74d-031288a9aad1" (UID: "56eb286f-8aa2-46aa-a74d-031288a9aad1"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.039245 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement1a80-account-delete-8q7fq"] Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.049437 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-597ff968b-msxbn" Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.049265 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-597ff968b-msxbn" event={"ID":"2130e3f2-cbea-46e1-b5d3-daefdd972935","Type":"ContainerDied","Data":"a137c2c8f449d4fac987a138026d8768098fc2fa5e342a9df21e1f2d3ff32046"} Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.051214 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"56eb286f-8aa2-46aa-a74d-031288a9aad1","Type":"ContainerDied","Data":"d2ecb31883121abad720cfb94674a6f06ef98dccfae53eb34ebb2671e32a13ec"} Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.051663 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.059725 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.076498 4870 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/56eb286f-8aa2-46aa-a74d-031288a9aad1-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:50 crc kubenswrapper[4870]: E1014 07:23:50.077469 4870 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Oct 14 07:23:50 crc kubenswrapper[4870]: E1014 07:23:50.077530 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5212ba20-83c3-41ff-a3d9-5a315f009c1b-config-data podName:5212ba20-83c3-41ff-a3d9-5a315f009c1b nodeName:}" failed. No retries permitted until 2025-10-14 07:23:58.077514985 +0000 UTC m=+1373.774875356 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/5212ba20-83c3-41ff-a3d9-5a315f009c1b-config-data") pod "rabbitmq-cell1-server-0" (UID: "5212ba20-83c3-41ff-a3d9-5a315f009c1b") : configmap "rabbitmq-cell1-config-data" not found Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.078892 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-582bv\" (UniqueName: \"kubernetes.io/projected/208754ea-7cc0-4222-a186-9ba01b274add-kube-api-access-582bv\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.079022 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pvhxn\" (UniqueName: \"kubernetes.io/projected/ff28d060-b280-4b89-bea1-d10aa367a3cc-kube-api-access-pvhxn\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.081499 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement1a80-account-delete-8q7fq"] Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.086863 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f0b766fd-ccb9-4620-84df-8fe2f194b0da-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f0b766fd-ccb9-4620-84df-8fe2f194b0da" (UID: "f0b766fd-ccb9-4620-84df-8fe2f194b0da"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.114845 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-proxy-5485b69bc9-ftmmq"] Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.134364 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-proxy-5485b69bc9-ftmmq"] Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.158168 4870 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage06-crc" (UniqueName: "kubernetes.io/local-volume/local-storage06-crc") on node "crc" Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.164952 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance1aaa-account-delete-nct6c"] Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.170309 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance1aaa-account-delete-nct6c"] Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.178685 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cae5b655-5d7c-4063-90ad-4d9913792d5b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cae5b655-5d7c-4063-90ad-4d9913792d5b" (UID: "cae5b655-5d7c-4063-90ad-4d9913792d5b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.180956 4870 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cae5b655-5d7c-4063-90ad-4d9913792d5b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.180999 4870 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f0b766fd-ccb9-4620-84df-8fe2f194b0da-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.181014 4870 reconciler_common.go:293] "Volume detached for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.198887 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f0b766fd-ccb9-4620-84df-8fe2f194b0da-config-data" (OuterVolumeSpecName: "config-data") pod "f0b766fd-ccb9-4620-84df-8fe2f194b0da" (UID: "f0b766fd-ccb9-4620-84df-8fe2f194b0da"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.246801 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c712b87a-cf4d-437e-abe0-1a566046ca66-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c712b87a-cf4d-437e-abe0-1a566046ca66" (UID: "c712b87a-cf4d-437e-abe0-1a566046ca66"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.283262 4870 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f0b766fd-ccb9-4620-84df-8fe2f194b0da-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.283296 4870 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c712b87a-cf4d-437e-abe0-1a566046ca66-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.301358 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2130e3f2-cbea-46e1-b5d3-daefdd972935-config-data" (OuterVolumeSpecName: "config-data") pod "2130e3f2-cbea-46e1-b5d3-daefdd972935" (UID: "2130e3f2-cbea-46e1-b5d3-daefdd972935"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.306089 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f0b766fd-ccb9-4620-84df-8fe2f194b0da-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "f0b766fd-ccb9-4620-84df-8fe2f194b0da" (UID: "f0b766fd-ccb9-4620-84df-8fe2f194b0da"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.309319 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3bd1e649-3d09-4cda-921f-ac9a84a5066e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3bd1e649-3d09-4cda-921f-ac9a84a5066e" (UID: "3bd1e649-3d09-4cda-921f-ac9a84a5066e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.309626 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cae5b655-5d7c-4063-90ad-4d9913792d5b-config-data" (OuterVolumeSpecName: "config-data") pod "cae5b655-5d7c-4063-90ad-4d9913792d5b" (UID: "cae5b655-5d7c-4063-90ad-4d9913792d5b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.310165 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cae5b655-5d7c-4063-90ad-4d9913792d5b-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "cae5b655-5d7c-4063-90ad-4d9913792d5b" (UID: "cae5b655-5d7c-4063-90ad-4d9913792d5b"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.339837 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c712b87a-cf4d-437e-abe0-1a566046ca66-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "c712b87a-cf4d-437e-abe0-1a566046ca66" (UID: "c712b87a-cf4d-437e-abe0-1a566046ca66"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.349710 4870 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.377246 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ff28d060-b280-4b89-bea1-d10aa367a3cc-config-data" (OuterVolumeSpecName: "config-data") pod "ff28d060-b280-4b89-bea1-d10aa367a3cc" (UID: "ff28d060-b280-4b89-bea1-d10aa367a3cc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.385398 4870 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c712b87a-cf4d-437e-abe0-1a566046ca66-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.385431 4870 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.385453 4870 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff28d060-b280-4b89-bea1-d10aa367a3cc-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.385463 4870 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3bd1e649-3d09-4cda-921f-ac9a84a5066e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.385473 4870 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2130e3f2-cbea-46e1-b5d3-daefdd972935-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.385482 4870 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cae5b655-5d7c-4063-90ad-4d9913792d5b-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.385491 4870 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/f0b766fd-ccb9-4620-84df-8fe2f194b0da-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.385502 4870 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/cae5b655-5d7c-4063-90ad-4d9913792d5b-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:50 crc kubenswrapper[4870]: E1014 07:23:50.385535 4870 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Oct 14 07:23:50 crc kubenswrapper[4870]: E1014 07:23:50.385616 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/da8a9731-f917-462f-9932-b37b6abb9a64-config-data podName:da8a9731-f917-462f-9932-b37b6abb9a64 nodeName:}" failed. No retries permitted until 2025-10-14 07:23:58.385593184 +0000 UTC m=+1374.082953635 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/da8a9731-f917-462f-9932-b37b6abb9a64-config-data") pod "rabbitmq-server-0" (UID: "da8a9731-f917-462f-9932-b37b6abb9a64") : configmap "rabbitmq-config-data" not found Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.390428 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3bd1e649-3d09-4cda-921f-ac9a84a5066e-memcached-tls-certs" (OuterVolumeSpecName: "memcached-tls-certs") pod "3bd1e649-3d09-4cda-921f-ac9a84a5066e" (UID: "3bd1e649-3d09-4cda-921f-ac9a84a5066e"). InnerVolumeSpecName "memcached-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.409295 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.412793 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2130e3f2-cbea-46e1-b5d3-daefdd972935-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2130e3f2-cbea-46e1-b5d3-daefdd972935" (UID: "2130e3f2-cbea-46e1-b5d3-daefdd972935"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.416315 4870 scope.go:117] "RemoveContainer" containerID="e6478418e9645b5d3cf9a2a8e1b1a4090b34dcf697735e42cf6fb41d3708bbeb" Oct 14 07:23:50 crc kubenswrapper[4870]: E1014 07:23:50.419539 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e6478418e9645b5d3cf9a2a8e1b1a4090b34dcf697735e42cf6fb41d3708bbeb\": container with ID starting with e6478418e9645b5d3cf9a2a8e1b1a4090b34dcf697735e42cf6fb41d3708bbeb not found: ID does not exist" containerID="e6478418e9645b5d3cf9a2a8e1b1a4090b34dcf697735e42cf6fb41d3708bbeb" Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.419590 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e6478418e9645b5d3cf9a2a8e1b1a4090b34dcf697735e42cf6fb41d3708bbeb"} err="failed to get container status \"e6478418e9645b5d3cf9a2a8e1b1a4090b34dcf697735e42cf6fb41d3708bbeb\": rpc error: code = NotFound desc = could not find container \"e6478418e9645b5d3cf9a2a8e1b1a4090b34dcf697735e42cf6fb41d3708bbeb\": container with ID starting with e6478418e9645b5d3cf9a2a8e1b1a4090b34dcf697735e42cf6fb41d3708bbeb not found: ID does not exist" Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.419621 4870 scope.go:117] "RemoveContainer" containerID="fbcb455b6e940e1124fe304e22169a2100dbb718060f679a722916c473baef26" Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.427069 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cae5b655-5d7c-4063-90ad-4d9913792d5b-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "cae5b655-5d7c-4063-90ad-4d9913792d5b" (UID: "cae5b655-5d7c-4063-90ad-4d9913792d5b"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.440417 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ff28d060-b280-4b89-bea1-d10aa367a3cc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ff28d060-b280-4b89-bea1-d10aa367a3cc" (UID: "ff28d060-b280-4b89-bea1-d10aa367a3cc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.445713 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.457702 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.469755 4870 scope.go:117] "RemoveContainer" containerID="a7dc7bd1a8ffe2b7e76587cb479c6f7ae957cc2c7d3a7fcfe34ba24bbd24c682" Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.478059 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.486604 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7bc76b72-8679-4b33-8966-0ee70c1788e5-config-data\") pod \"7bc76b72-8679-4b33-8966-0ee70c1788e5\" (UID: \"7bc76b72-8679-4b33-8966-0ee70c1788e5\") " Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.486648 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f5b700ed-3919-458f-860e-5630b8e4b800-config-data\") pod \"f5b700ed-3919-458f-860e-5630b8e4b800\" (UID: \"f5b700ed-3919-458f-860e-5630b8e4b800\") " Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.486675 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7bc76b72-8679-4b33-8966-0ee70c1788e5-etc-machine-id\") pod \"7bc76b72-8679-4b33-8966-0ee70c1788e5\" (UID: \"7bc76b72-8679-4b33-8966-0ee70c1788e5\") " Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.487907 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.489609 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7bc76b72-8679-4b33-8966-0ee70c1788e5-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "7bc76b72-8679-4b33-8966-0ee70c1788e5" (UID: "7bc76b72-8679-4b33-8966-0ee70c1788e5"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.489700 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/524b0b2a-77a2-4c05-be39-c8b97666ab11-log-httpd\") pod \"524b0b2a-77a2-4c05-be39-c8b97666ab11\" (UID: \"524b0b2a-77a2-4c05-be39-c8b97666ab11\") " Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.489776 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bb5cv\" (UniqueName: \"kubernetes.io/projected/7bc76b72-8679-4b33-8966-0ee70c1788e5-kube-api-access-bb5cv\") pod \"7bc76b72-8679-4b33-8966-0ee70c1788e5\" (UID: \"7bc76b72-8679-4b33-8966-0ee70c1788e5\") " Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.489810 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n4c9p\" (UniqueName: \"kubernetes.io/projected/f5b700ed-3919-458f-860e-5630b8e4b800-kube-api-access-n4c9p\") pod \"f5b700ed-3919-458f-860e-5630b8e4b800\" (UID: \"f5b700ed-3919-458f-860e-5630b8e4b800\") " Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.489903 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7bc76b72-8679-4b33-8966-0ee70c1788e5-combined-ca-bundle\") pod \"7bc76b72-8679-4b33-8966-0ee70c1788e5\" (UID: \"7bc76b72-8679-4b33-8966-0ee70c1788e5\") " Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.489994 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7bc76b72-8679-4b33-8966-0ee70c1788e5-scripts\") pod \"7bc76b72-8679-4b33-8966-0ee70c1788e5\" (UID: \"7bc76b72-8679-4b33-8966-0ee70c1788e5\") " Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.490034 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/524b0b2a-77a2-4c05-be39-c8b97666ab11-combined-ca-bundle\") pod \"524b0b2a-77a2-4c05-be39-c8b97666ab11\" (UID: \"524b0b2a-77a2-4c05-be39-c8b97666ab11\") " Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.490419 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/524b0b2a-77a2-4c05-be39-c8b97666ab11-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "524b0b2a-77a2-4c05-be39-c8b97666ab11" (UID: "524b0b2a-77a2-4c05-be39-c8b97666ab11"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.491833 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/524b0b2a-77a2-4c05-be39-c8b97666ab11-sg-core-conf-yaml\") pod \"524b0b2a-77a2-4c05-be39-c8b97666ab11\" (UID: \"524b0b2a-77a2-4c05-be39-c8b97666ab11\") " Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.491887 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5b700ed-3919-458f-860e-5630b8e4b800-combined-ca-bundle\") pod \"f5b700ed-3919-458f-860e-5630b8e4b800\" (UID: \"f5b700ed-3919-458f-860e-5630b8e4b800\") " Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.491940 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/524b0b2a-77a2-4c05-be39-c8b97666ab11-config-data\") pod \"524b0b2a-77a2-4c05-be39-c8b97666ab11\" (UID: \"524b0b2a-77a2-4c05-be39-c8b97666ab11\") " Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.491965 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7bc76b72-8679-4b33-8966-0ee70c1788e5-config-data-custom\") pod \"7bc76b72-8679-4b33-8966-0ee70c1788e5\" (UID: \"7bc76b72-8679-4b33-8966-0ee70c1788e5\") " Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.491990 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/524b0b2a-77a2-4c05-be39-c8b97666ab11-run-httpd\") pod \"524b0b2a-77a2-4c05-be39-c8b97666ab11\" (UID: \"524b0b2a-77a2-4c05-be39-c8b97666ab11\") " Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.492010 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gr66j\" (UniqueName: \"kubernetes.io/projected/524b0b2a-77a2-4c05-be39-c8b97666ab11-kube-api-access-gr66j\") pod \"524b0b2a-77a2-4c05-be39-c8b97666ab11\" (UID: \"524b0b2a-77a2-4c05-be39-c8b97666ab11\") " Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.492070 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/524b0b2a-77a2-4c05-be39-c8b97666ab11-ceilometer-tls-certs\") pod \"524b0b2a-77a2-4c05-be39-c8b97666ab11\" (UID: \"524b0b2a-77a2-4c05-be39-c8b97666ab11\") " Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.493365 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c712b87a-cf4d-437e-abe0-1a566046ca66-config-data" (OuterVolumeSpecName: "config-data") pod "c712b87a-cf4d-437e-abe0-1a566046ca66" (UID: "c712b87a-cf4d-437e-abe0-1a566046ca66"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.498374 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/524b0b2a-77a2-4c05-be39-c8b97666ab11-scripts\") pod \"524b0b2a-77a2-4c05-be39-c8b97666ab11\" (UID: \"524b0b2a-77a2-4c05-be39-c8b97666ab11\") " Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.498481 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/524b0b2a-77a2-4c05-be39-c8b97666ab11-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "524b0b2a-77a2-4c05-be39-c8b97666ab11" (UID: "524b0b2a-77a2-4c05-be39-c8b97666ab11"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.499076 4870 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/524b0b2a-77a2-4c05-be39-c8b97666ab11-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.499096 4870 reconciler_common.go:293] "Volume detached for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/3bd1e649-3d09-4cda-921f-ac9a84a5066e-memcached-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.499109 4870 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff28d060-b280-4b89-bea1-d10aa367a3cc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.499122 4870 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7bc76b72-8679-4b33-8966-0ee70c1788e5-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.499132 4870 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/524b0b2a-77a2-4c05-be39-c8b97666ab11-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.499141 4870 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2130e3f2-cbea-46e1-b5d3-daefdd972935-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.499150 4870 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c712b87a-cf4d-437e-abe0-1a566046ca66-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.499162 4870 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cae5b655-5d7c-4063-90ad-4d9913792d5b-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.499842 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7bc76b72-8679-4b33-8966-0ee70c1788e5-scripts" (OuterVolumeSpecName: "scripts") pod "7bc76b72-8679-4b33-8966-0ee70c1788e5" (UID: "7bc76b72-8679-4b33-8966-0ee70c1788e5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.507231 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7bc76b72-8679-4b33-8966-0ee70c1788e5-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "7bc76b72-8679-4b33-8966-0ee70c1788e5" (UID: "7bc76b72-8679-4b33-8966-0ee70c1788e5"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.507681 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2130e3f2-cbea-46e1-b5d3-daefdd972935-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "2130e3f2-cbea-46e1-b5d3-daefdd972935" (UID: "2130e3f2-cbea-46e1-b5d3-daefdd972935"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.517940 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f5b700ed-3919-458f-860e-5630b8e4b800-kube-api-access-n4c9p" (OuterVolumeSpecName: "kube-api-access-n4c9p") pod "f5b700ed-3919-458f-860e-5630b8e4b800" (UID: "f5b700ed-3919-458f-860e-5630b8e4b800"). InnerVolumeSpecName "kube-api-access-n4c9p". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.518214 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/524b0b2a-77a2-4c05-be39-c8b97666ab11-scripts" (OuterVolumeSpecName: "scripts") pod "524b0b2a-77a2-4c05-be39-c8b97666ab11" (UID: "524b0b2a-77a2-4c05-be39-c8b97666ab11"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.520946 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bc76b72-8679-4b33-8966-0ee70c1788e5-kube-api-access-bb5cv" (OuterVolumeSpecName: "kube-api-access-bb5cv") pod "7bc76b72-8679-4b33-8966-0ee70c1788e5" (UID: "7bc76b72-8679-4b33-8966-0ee70c1788e5"). InnerVolumeSpecName "kube-api-access-bb5cv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.529083 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/novaapie66c-account-delete-ftd24"] Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.534753 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/524b0b2a-77a2-4c05-be39-c8b97666ab11-kube-api-access-gr66j" (OuterVolumeSpecName: "kube-api-access-gr66j") pod "524b0b2a-77a2-4c05-be39-c8b97666ab11" (UID: "524b0b2a-77a2-4c05-be39-c8b97666ab11"). InnerVolumeSpecName "kube-api-access-gr66j". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.536644 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.567062 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/novaapie66c-account-delete-ftd24"] Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.600641 4870 scope.go:117] "RemoveContainer" containerID="649fea192c7edfde335ef6265aaeb17d682a7f0427cec9179859e9c93e49f399" Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.601621 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.606682 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ff28d060-b280-4b89-bea1-d10aa367a3cc-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "ff28d060-b280-4b89-bea1-d10aa367a3cc" (UID: "ff28d060-b280-4b89-bea1-d10aa367a3cc"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.609197 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.609322 4870 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2130e3f2-cbea-46e1-b5d3-daefdd972935-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.609350 4870 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7bc76b72-8679-4b33-8966-0ee70c1788e5-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.609365 4870 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7bc76b72-8679-4b33-8966-0ee70c1788e5-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.609382 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gr66j\" (UniqueName: \"kubernetes.io/projected/524b0b2a-77a2-4c05-be39-c8b97666ab11-kube-api-access-gr66j\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.609395 4870 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/524b0b2a-77a2-4c05-be39-c8b97666ab11-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.609408 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bb5cv\" (UniqueName: \"kubernetes.io/projected/7bc76b72-8679-4b33-8966-0ee70c1788e5-kube-api-access-bb5cv\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.609421 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n4c9p\" (UniqueName: \"kubernetes.io/projected/f5b700ed-3919-458f-860e-5630b8e4b800-kube-api-access-n4c9p\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.625321 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.630981 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ff28d060-b280-4b89-bea1-d10aa367a3cc-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "ff28d060-b280-4b89-bea1-d10aa367a3cc" (UID: "ff28d060-b280-4b89-bea1-d10aa367a3cc"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.631895 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.632205 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f5b700ed-3919-458f-860e-5630b8e4b800-config-data" (OuterVolumeSpecName: "config-data") pod "f5b700ed-3919-458f-860e-5630b8e4b800" (UID: "f5b700ed-3919-458f-860e-5630b8e4b800"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.632723 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f5b700ed-3919-458f-860e-5630b8e4b800-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f5b700ed-3919-458f-860e-5630b8e4b800" (UID: "f5b700ed-3919-458f-860e-5630b8e4b800"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.638546 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2130e3f2-cbea-46e1-b5d3-daefdd972935-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "2130e3f2-cbea-46e1-b5d3-daefdd972935" (UID: "2130e3f2-cbea-46e1-b5d3-daefdd972935"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.660654 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7bc76b72-8679-4b33-8966-0ee70c1788e5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7bc76b72-8679-4b33-8966-0ee70c1788e5" (UID: "7bc76b72-8679-4b33-8966-0ee70c1788e5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.669992 4870 scope.go:117] "RemoveContainer" containerID="f744eb9b855e3191279bf5fc01139d0a3359c09a6178c62d0602563d9d13ff25" Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.672345 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/524b0b2a-77a2-4c05-be39-c8b97666ab11-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "524b0b2a-77a2-4c05-be39-c8b97666ab11" (UID: "524b0b2a-77a2-4c05-be39-c8b97666ab11"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.672746 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/524b0b2a-77a2-4c05-be39-c8b97666ab11-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "524b0b2a-77a2-4c05-be39-c8b97666ab11" (UID: "524b0b2a-77a2-4c05-be39-c8b97666ab11"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.674676 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/524b0b2a-77a2-4c05-be39-c8b97666ab11-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "524b0b2a-77a2-4c05-be39-c8b97666ab11" (UID: "524b0b2a-77a2-4c05-be39-c8b97666ab11"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.676552 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.699339 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.699375 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7bc76b72-8679-4b33-8966-0ee70c1788e5-config-data" (OuterVolumeSpecName: "config-data") pod "7bc76b72-8679-4b33-8966-0ee70c1788e5" (UID: "7bc76b72-8679-4b33-8966-0ee70c1788e5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.710875 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/3de2167a-7663-4b00-9743-229ff1864cb8-secrets\") pod \"3de2167a-7663-4b00-9743-229ff1864cb8\" (UID: \"3de2167a-7663-4b00-9743-229ff1864cb8\") " Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.710935 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/3de2167a-7663-4b00-9743-229ff1864cb8-galera-tls-certs\") pod \"3de2167a-7663-4b00-9743-229ff1864cb8\" (UID: \"3de2167a-7663-4b00-9743-229ff1864cb8\") " Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.710981 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bz55d\" (UniqueName: \"kubernetes.io/projected/3de2167a-7663-4b00-9743-229ff1864cb8-kube-api-access-bz55d\") pod \"3de2167a-7663-4b00-9743-229ff1864cb8\" (UID: \"3de2167a-7663-4b00-9743-229ff1864cb8\") " Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.711012 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3de2167a-7663-4b00-9743-229ff1864cb8-combined-ca-bundle\") pod \"3de2167a-7663-4b00-9743-229ff1864cb8\" (UID: \"3de2167a-7663-4b00-9743-229ff1864cb8\") " Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.711055 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/3de2167a-7663-4b00-9743-229ff1864cb8-config-data-generated\") pod \"3de2167a-7663-4b00-9743-229ff1864cb8\" (UID: \"3de2167a-7663-4b00-9743-229ff1864cb8\") " Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.711097 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mysql-db\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"3de2167a-7663-4b00-9743-229ff1864cb8\" (UID: \"3de2167a-7663-4b00-9743-229ff1864cb8\") " Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.711114 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/3de2167a-7663-4b00-9743-229ff1864cb8-config-data-default\") pod \"3de2167a-7663-4b00-9743-229ff1864cb8\" (UID: \"3de2167a-7663-4b00-9743-229ff1864cb8\") " Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.711130 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3de2167a-7663-4b00-9743-229ff1864cb8-operator-scripts\") pod \"3de2167a-7663-4b00-9743-229ff1864cb8\" (UID: \"3de2167a-7663-4b00-9743-229ff1864cb8\") " Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.711177 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/3de2167a-7663-4b00-9743-229ff1864cb8-kolla-config\") pod \"3de2167a-7663-4b00-9743-229ff1864cb8\" (UID: \"3de2167a-7663-4b00-9743-229ff1864cb8\") " Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.711439 4870 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7bc76b72-8679-4b33-8966-0ee70c1788e5-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.711516 4870 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f5b700ed-3919-458f-860e-5630b8e4b800-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.711526 4870 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ff28d060-b280-4b89-bea1-d10aa367a3cc-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.711536 4870 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7bc76b72-8679-4b33-8966-0ee70c1788e5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.711544 4870 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ff28d060-b280-4b89-bea1-d10aa367a3cc-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.711554 4870 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/524b0b2a-77a2-4c05-be39-c8b97666ab11-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.711565 4870 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/524b0b2a-77a2-4c05-be39-c8b97666ab11-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.711573 4870 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5b700ed-3919-458f-860e-5630b8e4b800-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.711582 4870 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2130e3f2-cbea-46e1-b5d3-daefdd972935-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.711591 4870 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/524b0b2a-77a2-4c05-be39-c8b97666ab11-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.713714 4870 scope.go:117] "RemoveContainer" containerID="6c0de18ab40ba16d1bc75ff9b157bb26252b1a8044809c3ee46ea0a31ed39019" Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.714242 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3de2167a-7663-4b00-9743-229ff1864cb8-config-data-generated" (OuterVolumeSpecName: "config-data-generated") pod "3de2167a-7663-4b00-9743-229ff1864cb8" (UID: "3de2167a-7663-4b00-9743-229ff1864cb8"). InnerVolumeSpecName "config-data-generated". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.714421 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/memcached-0"] Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.714825 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3de2167a-7663-4b00-9743-229ff1864cb8-config-data-default" (OuterVolumeSpecName: "config-data-default") pod "3de2167a-7663-4b00-9743-229ff1864cb8" (UID: "3de2167a-7663-4b00-9743-229ff1864cb8"). InnerVolumeSpecName "config-data-default". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.715259 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3de2167a-7663-4b00-9743-229ff1864cb8-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "3de2167a-7663-4b00-9743-229ff1864cb8" (UID: "3de2167a-7663-4b00-9743-229ff1864cb8"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.715760 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3de2167a-7663-4b00-9743-229ff1864cb8-kolla-config" (OuterVolumeSpecName: "kolla-config") pod "3de2167a-7663-4b00-9743-229ff1864cb8" (UID: "3de2167a-7663-4b00-9743-229ff1864cb8"). InnerVolumeSpecName "kolla-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.719647 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3de2167a-7663-4b00-9743-229ff1864cb8-secrets" (OuterVolumeSpecName: "secrets") pod "3de2167a-7663-4b00-9743-229ff1864cb8" (UID: "3de2167a-7663-4b00-9743-229ff1864cb8"). InnerVolumeSpecName "secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.719744 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3de2167a-7663-4b00-9743-229ff1864cb8-kube-api-access-bz55d" (OuterVolumeSpecName: "kube-api-access-bz55d") pod "3de2167a-7663-4b00-9743-229ff1864cb8" (UID: "3de2167a-7663-4b00-9743-229ff1864cb8"). InnerVolumeSpecName "kube-api-access-bz55d". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.725047 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/memcached-0"] Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.732917 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage03-crc" (OuterVolumeSpecName: "mysql-db") pod "3de2167a-7663-4b00-9743-229ff1864cb8" (UID: "3de2167a-7663-4b00-9743-229ff1864cb8"). InnerVolumeSpecName "local-storage03-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.735412 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-79cc757548-zxf4g"] Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.741898 4870 scope.go:117] "RemoveContainer" containerID="65770c73c87b754f9177bcc262af5253b96b0534c636832b413175980c00c785" Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.744926 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-79cc757548-zxf4g"] Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.755151 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-597ff968b-msxbn"] Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.760609 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-597ff968b-msxbn"] Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.769330 4870 scope.go:117] "RemoveContainer" containerID="6fb3ea4d6cd5802f3695559f939e61f05770876bd3c155dcf6040899ef93330a" Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.799719 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3de2167a-7663-4b00-9743-229ff1864cb8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3de2167a-7663-4b00-9743-229ff1864cb8" (UID: "3de2167a-7663-4b00-9743-229ff1864cb8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.812288 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3de2167a-7663-4b00-9743-229ff1864cb8-galera-tls-certs" (OuterVolumeSpecName: "galera-tls-certs") pod "3de2167a-7663-4b00-9743-229ff1864cb8" (UID: "3de2167a-7663-4b00-9743-229ff1864cb8"). InnerVolumeSpecName "galera-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.812850 4870 reconciler_common.go:293] "Volume detached for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/3de2167a-7663-4b00-9743-229ff1864cb8-kolla-config\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.813788 4870 reconciler_common.go:293] "Volume detached for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/3de2167a-7663-4b00-9743-229ff1864cb8-secrets\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.813848 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bz55d\" (UniqueName: \"kubernetes.io/projected/3de2167a-7663-4b00-9743-229ff1864cb8-kube-api-access-bz55d\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.813917 4870 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3de2167a-7663-4b00-9743-229ff1864cb8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.813980 4870 reconciler_common.go:293] "Volume detached for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/3de2167a-7663-4b00-9743-229ff1864cb8-config-data-generated\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.814052 4870 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" " Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.814113 4870 reconciler_common.go:293] "Volume detached for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/3de2167a-7663-4b00-9743-229ff1864cb8-config-data-default\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.814175 4870 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3de2167a-7663-4b00-9743-229ff1864cb8-operator-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.818669 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/524b0b2a-77a2-4c05-be39-c8b97666ab11-config-data" (OuterVolumeSpecName: "config-data") pod "524b0b2a-77a2-4c05-be39-c8b97666ab11" (UID: "524b0b2a-77a2-4c05-be39-c8b97666ab11"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.836690 4870 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage03-crc" (UniqueName: "kubernetes.io/local-volume/local-storage03-crc") on node "crc" Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.919800 4870 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/524b0b2a-77a2-4c05-be39-c8b97666ab11-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.919845 4870 reconciler_common.go:293] "Volume detached for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.919862 4870 reconciler_common.go:293] "Volume detached for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/3de2167a-7663-4b00-9743-229ff1864cb8-galera-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.986799 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 14 07:23:50 crc kubenswrapper[4870]: I1014 07:23:50.992384 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 14 07:23:51 crc kubenswrapper[4870]: I1014 07:23:51.020277 4870 scope.go:117] "RemoveContainer" containerID="cd6dabf959a5bb827e348f8580f0325c6287615f71bc7279bb2e2744f5cf4f7c" Oct 14 07:23:51 crc kubenswrapper[4870]: I1014 07:23:51.043616 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="208754ea-7cc0-4222-a186-9ba01b274add" path="/var/lib/kubelet/pods/208754ea-7cc0-4222-a186-9ba01b274add/volumes" Oct 14 07:23:51 crc kubenswrapper[4870]: I1014 07:23:51.044175 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2130e3f2-cbea-46e1-b5d3-daefdd972935" path="/var/lib/kubelet/pods/2130e3f2-cbea-46e1-b5d3-daefdd972935/volumes" Oct 14 07:23:51 crc kubenswrapper[4870]: I1014 07:23:51.044935 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2abee394-ffb3-4dba-8d09-19c1cb476595" path="/var/lib/kubelet/pods/2abee394-ffb3-4dba-8d09-19c1cb476595/volumes" Oct 14 07:23:51 crc kubenswrapper[4870]: I1014 07:23:51.045940 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3927fb78-4f30-4371-8fbb-62fd276c0460" path="/var/lib/kubelet/pods/3927fb78-4f30-4371-8fbb-62fd276c0460/volumes" Oct 14 07:23:51 crc kubenswrapper[4870]: I1014 07:23:51.046733 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3bd1e649-3d09-4cda-921f-ac9a84a5066e" path="/var/lib/kubelet/pods/3bd1e649-3d09-4cda-921f-ac9a84a5066e/volumes" Oct 14 07:23:51 crc kubenswrapper[4870]: I1014 07:23:51.047301 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4d74ef50-1861-4d70-b781-f92d5f4b35ba" path="/var/lib/kubelet/pods/4d74ef50-1861-4d70-b781-f92d5f4b35ba/volumes" Oct 14 07:23:51 crc kubenswrapper[4870]: I1014 07:23:51.048326 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="51de8090-c48e-498c-99ab-794a7527252c" path="/var/lib/kubelet/pods/51de8090-c48e-498c-99ab-794a7527252c/volumes" Oct 14 07:23:51 crc kubenswrapper[4870]: I1014 07:23:51.049655 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="56eb286f-8aa2-46aa-a74d-031288a9aad1" path="/var/lib/kubelet/pods/56eb286f-8aa2-46aa-a74d-031288a9aad1/volumes" Oct 14 07:23:51 crc kubenswrapper[4870]: I1014 07:23:51.052732 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="61885759-41a7-4ee8-904a-3f85bb66cd19" path="/var/lib/kubelet/pods/61885759-41a7-4ee8-904a-3f85bb66cd19/volumes" Oct 14 07:23:51 crc kubenswrapper[4870]: I1014 07:23:51.053278 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7cbdbbdd-ea4f-4d14-8f3c-4784bda45874" path="/var/lib/kubelet/pods/7cbdbbdd-ea4f-4d14-8f3c-4784bda45874/volumes" Oct 14 07:23:51 crc kubenswrapper[4870]: I1014 07:23:51.056324 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8abede26-36e8-46ec-a5c4-6cd53319133a" path="/var/lib/kubelet/pods/8abede26-36e8-46ec-a5c4-6cd53319133a/volumes" Oct 14 07:23:51 crc kubenswrapper[4870]: I1014 07:23:51.056773 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a6f02e4f-8dca-4b97-91e9-ee3368e5e44b" path="/var/lib/kubelet/pods/a6f02e4f-8dca-4b97-91e9-ee3368e5e44b/volumes" Oct 14 07:23:51 crc kubenswrapper[4870]: I1014 07:23:51.057263 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a9d8a25d-2f1c-483d-bf0d-76a2bf30f137" path="/var/lib/kubelet/pods/a9d8a25d-2f1c-483d-bf0d-76a2bf30f137/volumes" Oct 14 07:23:51 crc kubenswrapper[4870]: I1014 07:23:51.058253 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bad45b30-5ba2-4062-ab14-04f923679c4e" path="/var/lib/kubelet/pods/bad45b30-5ba2-4062-ab14-04f923679c4e/volumes" Oct 14 07:23:51 crc kubenswrapper[4870]: I1014 07:23:51.062045 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c603f7b5-015f-4c2e-9144-b3e8349c3da3" path="/var/lib/kubelet/pods/c603f7b5-015f-4c2e-9144-b3e8349c3da3/volumes" Oct 14 07:23:51 crc kubenswrapper[4870]: I1014 07:23:51.062763 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c712b87a-cf4d-437e-abe0-1a566046ca66" path="/var/lib/kubelet/pods/c712b87a-cf4d-437e-abe0-1a566046ca66/volumes" Oct 14 07:23:51 crc kubenswrapper[4870]: I1014 07:23:51.063299 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cae5b655-5d7c-4063-90ad-4d9913792d5b" path="/var/lib/kubelet/pods/cae5b655-5d7c-4063-90ad-4d9913792d5b/volumes" Oct 14 07:23:51 crc kubenswrapper[4870]: I1014 07:23:51.064163 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="de479f5d-22a2-4f7e-8034-cffefc511b53" path="/var/lib/kubelet/pods/de479f5d-22a2-4f7e-8034-cffefc511b53/volumes" Oct 14 07:23:51 crc kubenswrapper[4870]: I1014 07:23:51.064674 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f0b766fd-ccb9-4620-84df-8fe2f194b0da" path="/var/lib/kubelet/pods/f0b766fd-ccb9-4620-84df-8fe2f194b0da/volumes" Oct 14 07:23:51 crc kubenswrapper[4870]: I1014 07:23:51.065240 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ff28d060-b280-4b89-bea1-d10aa367a3cc" path="/var/lib/kubelet/pods/ff28d060-b280-4b89-bea1-d10aa367a3cc/volumes" Oct 14 07:23:51 crc kubenswrapper[4870]: I1014 07:23:51.084807 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"f5b700ed-3919-458f-860e-5630b8e4b800","Type":"ContainerDied","Data":"aeed74d1fcfc0d32441c8025c1ed06c4fa2e3b5be6c136e68861143e7f83a891"} Oct 14 07:23:51 crc kubenswrapper[4870]: I1014 07:23:51.084916 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 14 07:23:51 crc kubenswrapper[4870]: I1014 07:23:51.095590 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_416d7b39-4541-44a8-a55f-924bc86fee32/ovn-northd/0.log" Oct 14 07:23:51 crc kubenswrapper[4870]: I1014 07:23:51.095632 4870 generic.go:334] "Generic (PLEG): container finished" podID="416d7b39-4541-44a8-a55f-924bc86fee32" containerID="457905ca1130c3804c068cf6fc44521ad064a3fb34449d666cf858f14e2feace" exitCode=139 Oct 14 07:23:51 crc kubenswrapper[4870]: I1014 07:23:51.095687 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"416d7b39-4541-44a8-a55f-924bc86fee32","Type":"ContainerDied","Data":"457905ca1130c3804c068cf6fc44521ad064a3fb34449d666cf858f14e2feace"} Oct 14 07:23:51 crc kubenswrapper[4870]: I1014 07:23:51.095713 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"416d7b39-4541-44a8-a55f-924bc86fee32","Type":"ContainerDied","Data":"ae426546eb62a779272b920709b17c66107242f70b6d596ad360a33101291fe0"} Oct 14 07:23:51 crc kubenswrapper[4870]: I1014 07:23:51.095724 4870 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ae426546eb62a779272b920709b17c66107242f70b6d596ad360a33101291fe0" Oct 14 07:23:51 crc kubenswrapper[4870]: I1014 07:23:51.098636 4870 generic.go:334] "Generic (PLEG): container finished" podID="5212ba20-83c3-41ff-a3d9-5a315f009c1b" containerID="eeb00ec87974e56f742f7ae182732d1dfbbc2c6523987de46327c67f0a729b18" exitCode=0 Oct 14 07:23:51 crc kubenswrapper[4870]: I1014 07:23:51.098688 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"5212ba20-83c3-41ff-a3d9-5a315f009c1b","Type":"ContainerDied","Data":"eeb00ec87974e56f742f7ae182732d1dfbbc2c6523987de46327c67f0a729b18"} Oct 14 07:23:51 crc kubenswrapper[4870]: I1014 07:23:51.109676 4870 generic.go:334] "Generic (PLEG): container finished" podID="3de2167a-7663-4b00-9743-229ff1864cb8" containerID="1b53e989d30152b230c176967064509ca1fafba96b986cc96bc23bdcafabf49f" exitCode=0 Oct 14 07:23:51 crc kubenswrapper[4870]: I1014 07:23:51.109744 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"3de2167a-7663-4b00-9743-229ff1864cb8","Type":"ContainerDied","Data":"1b53e989d30152b230c176967064509ca1fafba96b986cc96bc23bdcafabf49f"} Oct 14 07:23:51 crc kubenswrapper[4870]: I1014 07:23:51.109803 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"3de2167a-7663-4b00-9743-229ff1864cb8","Type":"ContainerDied","Data":"e83ba8ed593070f7993afb4580d91a1da037ccd817bcfc2a6029ab4e2b0edf86"} Oct 14 07:23:51 crc kubenswrapper[4870]: I1014 07:23:51.109972 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Oct 14 07:23:51 crc kubenswrapper[4870]: I1014 07:23:51.114218 4870 generic.go:334] "Generic (PLEG): container finished" podID="60de9d04-7c24-42a4-9d5e-74b611487b69" containerID="d8615c16655ea7e7d454161d16b61973188b63bffc966ed4b3805db817746641" exitCode=0 Oct 14 07:23:51 crc kubenswrapper[4870]: I1014 07:23:51.114261 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-865c567797-jljrm" event={"ID":"60de9d04-7c24-42a4-9d5e-74b611487b69","Type":"ContainerDied","Data":"d8615c16655ea7e7d454161d16b61973188b63bffc966ed4b3805db817746641"} Oct 14 07:23:51 crc kubenswrapper[4870]: I1014 07:23:51.125149 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 14 07:23:51 crc kubenswrapper[4870]: I1014 07:23:51.125839 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 14 07:23:51 crc kubenswrapper[4870]: I1014 07:23:51.126041 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"7bc76b72-8679-4b33-8966-0ee70c1788e5","Type":"ContainerDied","Data":"61320ff1ecf9ca47176c06537cb5c503a9ed81f16e0538a006104df47500851e"} Oct 14 07:23:51 crc kubenswrapper[4870]: I1014 07:23:51.126667 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_416d7b39-4541-44a8-a55f-924bc86fee32/ovn-northd/0.log" Oct 14 07:23:51 crc kubenswrapper[4870]: I1014 07:23:51.126740 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Oct 14 07:23:51 crc kubenswrapper[4870]: I1014 07:23:51.140039 4870 scope.go:117] "RemoveContainer" containerID="9867cb00fe7f9cff6394884f9edfc786d6b848597f09a56abffe01cb5e8ca4cf" Oct 14 07:23:51 crc kubenswrapper[4870]: I1014 07:23:51.152080 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 14 07:23:51 crc kubenswrapper[4870]: I1014 07:23:51.159230 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 14 07:23:51 crc kubenswrapper[4870]: I1014 07:23:51.219084 4870 scope.go:117] "RemoveContainer" containerID="6d95049e97ec6fbbfa6d28e3e543ed1f7fc74569f83d0862a392737896e3463d" Oct 14 07:23:51 crc kubenswrapper[4870]: I1014 07:23:51.226317 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d28jg\" (UniqueName: \"kubernetes.io/projected/416d7b39-4541-44a8-a55f-924bc86fee32-kube-api-access-d28jg\") pod \"416d7b39-4541-44a8-a55f-924bc86fee32\" (UID: \"416d7b39-4541-44a8-a55f-924bc86fee32\") " Oct 14 07:23:51 crc kubenswrapper[4870]: I1014 07:23:51.226361 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/416d7b39-4541-44a8-a55f-924bc86fee32-ovn-northd-tls-certs\") pod \"416d7b39-4541-44a8-a55f-924bc86fee32\" (UID: \"416d7b39-4541-44a8-a55f-924bc86fee32\") " Oct 14 07:23:51 crc kubenswrapper[4870]: I1014 07:23:51.226400 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/416d7b39-4541-44a8-a55f-924bc86fee32-metrics-certs-tls-certs\") pod \"416d7b39-4541-44a8-a55f-924bc86fee32\" (UID: \"416d7b39-4541-44a8-a55f-924bc86fee32\") " Oct 14 07:23:51 crc kubenswrapper[4870]: I1014 07:23:51.226440 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/416d7b39-4541-44a8-a55f-924bc86fee32-config\") pod \"416d7b39-4541-44a8-a55f-924bc86fee32\" (UID: \"416d7b39-4541-44a8-a55f-924bc86fee32\") " Oct 14 07:23:51 crc kubenswrapper[4870]: I1014 07:23:51.226512 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/416d7b39-4541-44a8-a55f-924bc86fee32-combined-ca-bundle\") pod \"416d7b39-4541-44a8-a55f-924bc86fee32\" (UID: \"416d7b39-4541-44a8-a55f-924bc86fee32\") " Oct 14 07:23:51 crc kubenswrapper[4870]: I1014 07:23:51.226583 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/416d7b39-4541-44a8-a55f-924bc86fee32-scripts\") pod \"416d7b39-4541-44a8-a55f-924bc86fee32\" (UID: \"416d7b39-4541-44a8-a55f-924bc86fee32\") " Oct 14 07:23:51 crc kubenswrapper[4870]: I1014 07:23:51.226681 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/416d7b39-4541-44a8-a55f-924bc86fee32-ovn-rundir\") pod \"416d7b39-4541-44a8-a55f-924bc86fee32\" (UID: \"416d7b39-4541-44a8-a55f-924bc86fee32\") " Oct 14 07:23:51 crc kubenswrapper[4870]: I1014 07:23:51.227509 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/416d7b39-4541-44a8-a55f-924bc86fee32-config" (OuterVolumeSpecName: "config") pod "416d7b39-4541-44a8-a55f-924bc86fee32" (UID: "416d7b39-4541-44a8-a55f-924bc86fee32"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:23:51 crc kubenswrapper[4870]: I1014 07:23:51.227592 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/416d7b39-4541-44a8-a55f-924bc86fee32-ovn-rundir" (OuterVolumeSpecName: "ovn-rundir") pod "416d7b39-4541-44a8-a55f-924bc86fee32" (UID: "416d7b39-4541-44a8-a55f-924bc86fee32"). InnerVolumeSpecName "ovn-rundir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:23:51 crc kubenswrapper[4870]: I1014 07:23:51.228019 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/416d7b39-4541-44a8-a55f-924bc86fee32-scripts" (OuterVolumeSpecName: "scripts") pod "416d7b39-4541-44a8-a55f-924bc86fee32" (UID: "416d7b39-4541-44a8-a55f-924bc86fee32"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:23:51 crc kubenswrapper[4870]: I1014 07:23:51.231857 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/416d7b39-4541-44a8-a55f-924bc86fee32-kube-api-access-d28jg" (OuterVolumeSpecName: "kube-api-access-d28jg") pod "416d7b39-4541-44a8-a55f-924bc86fee32" (UID: "416d7b39-4541-44a8-a55f-924bc86fee32"). InnerVolumeSpecName "kube-api-access-d28jg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:23:51 crc kubenswrapper[4870]: I1014 07:23:51.241175 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 14 07:23:51 crc kubenswrapper[4870]: I1014 07:23:51.254804 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Oct 14 07:23:51 crc kubenswrapper[4870]: I1014 07:23:51.260644 4870 scope.go:117] "RemoveContainer" containerID="9867cb00fe7f9cff6394884f9edfc786d6b848597f09a56abffe01cb5e8ca4cf" Oct 14 07:23:51 crc kubenswrapper[4870]: E1014 07:23:51.261163 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9867cb00fe7f9cff6394884f9edfc786d6b848597f09a56abffe01cb5e8ca4cf\": container with ID starting with 9867cb00fe7f9cff6394884f9edfc786d6b848597f09a56abffe01cb5e8ca4cf not found: ID does not exist" containerID="9867cb00fe7f9cff6394884f9edfc786d6b848597f09a56abffe01cb5e8ca4cf" Oct 14 07:23:51 crc kubenswrapper[4870]: I1014 07:23:51.261318 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9867cb00fe7f9cff6394884f9edfc786d6b848597f09a56abffe01cb5e8ca4cf"} err="failed to get container status \"9867cb00fe7f9cff6394884f9edfc786d6b848597f09a56abffe01cb5e8ca4cf\": rpc error: code = NotFound desc = could not find container \"9867cb00fe7f9cff6394884f9edfc786d6b848597f09a56abffe01cb5e8ca4cf\": container with ID starting with 9867cb00fe7f9cff6394884f9edfc786d6b848597f09a56abffe01cb5e8ca4cf not found: ID does not exist" Oct 14 07:23:51 crc kubenswrapper[4870]: I1014 07:23:51.261367 4870 scope.go:117] "RemoveContainer" containerID="6d95049e97ec6fbbfa6d28e3e543ed1f7fc74569f83d0862a392737896e3463d" Oct 14 07:23:51 crc kubenswrapper[4870]: E1014 07:23:51.261739 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6d95049e97ec6fbbfa6d28e3e543ed1f7fc74569f83d0862a392737896e3463d\": container with ID starting with 6d95049e97ec6fbbfa6d28e3e543ed1f7fc74569f83d0862a392737896e3463d not found: ID does not exist" containerID="6d95049e97ec6fbbfa6d28e3e543ed1f7fc74569f83d0862a392737896e3463d" Oct 14 07:23:51 crc kubenswrapper[4870]: I1014 07:23:51.261779 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6d95049e97ec6fbbfa6d28e3e543ed1f7fc74569f83d0862a392737896e3463d"} err="failed to get container status \"6d95049e97ec6fbbfa6d28e3e543ed1f7fc74569f83d0862a392737896e3463d\": rpc error: code = NotFound desc = could not find container \"6d95049e97ec6fbbfa6d28e3e543ed1f7fc74569f83d0862a392737896e3463d\": container with ID starting with 6d95049e97ec6fbbfa6d28e3e543ed1f7fc74569f83d0862a392737896e3463d not found: ID does not exist" Oct 14 07:23:51 crc kubenswrapper[4870]: I1014 07:23:51.261798 4870 scope.go:117] "RemoveContainer" containerID="cd54f5f17ff79e109b4abe40e2d80763b88c28e0edaa5a29d13dfa1146069254" Oct 14 07:23:51 crc kubenswrapper[4870]: I1014 07:23:51.265583 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 14 07:23:51 crc kubenswrapper[4870]: I1014 07:23:51.271614 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 14 07:23:51 crc kubenswrapper[4870]: I1014 07:23:51.273624 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/416d7b39-4541-44a8-a55f-924bc86fee32-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "416d7b39-4541-44a8-a55f-924bc86fee32" (UID: "416d7b39-4541-44a8-a55f-924bc86fee32"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:23:51 crc kubenswrapper[4870]: I1014 07:23:51.277053 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstack-galera-0"] Oct 14 07:23:51 crc kubenswrapper[4870]: I1014 07:23:51.278681 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-865c567797-jljrm" Oct 14 07:23:51 crc kubenswrapper[4870]: I1014 07:23:51.282254 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstack-galera-0"] Oct 14 07:23:51 crc kubenswrapper[4870]: I1014 07:23:51.292711 4870 scope.go:117] "RemoveContainer" containerID="b97a3d2764961f388fdc1a6459d91ca5f1248920ce92c572b651657c109e7fa4" Oct 14 07:23:51 crc kubenswrapper[4870]: I1014 07:23:51.331345 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/60de9d04-7c24-42a4-9d5e-74b611487b69-fernet-keys\") pod \"60de9d04-7c24-42a4-9d5e-74b611487b69\" (UID: \"60de9d04-7c24-42a4-9d5e-74b611487b69\") " Oct 14 07:23:51 crc kubenswrapper[4870]: I1014 07:23:51.331386 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60de9d04-7c24-42a4-9d5e-74b611487b69-combined-ca-bundle\") pod \"60de9d04-7c24-42a4-9d5e-74b611487b69\" (UID: \"60de9d04-7c24-42a4-9d5e-74b611487b69\") " Oct 14 07:23:51 crc kubenswrapper[4870]: I1014 07:23:51.331433 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vhds6\" (UniqueName: \"kubernetes.io/projected/60de9d04-7c24-42a4-9d5e-74b611487b69-kube-api-access-vhds6\") pod \"60de9d04-7c24-42a4-9d5e-74b611487b69\" (UID: \"60de9d04-7c24-42a4-9d5e-74b611487b69\") " Oct 14 07:23:51 crc kubenswrapper[4870]: I1014 07:23:51.331483 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/60de9d04-7c24-42a4-9d5e-74b611487b69-credential-keys\") pod \"60de9d04-7c24-42a4-9d5e-74b611487b69\" (UID: \"60de9d04-7c24-42a4-9d5e-74b611487b69\") " Oct 14 07:23:51 crc kubenswrapper[4870]: I1014 07:23:51.331524 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/60de9d04-7c24-42a4-9d5e-74b611487b69-internal-tls-certs\") pod \"60de9d04-7c24-42a4-9d5e-74b611487b69\" (UID: \"60de9d04-7c24-42a4-9d5e-74b611487b69\") " Oct 14 07:23:51 crc kubenswrapper[4870]: I1014 07:23:51.331563 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/60de9d04-7c24-42a4-9d5e-74b611487b69-scripts\") pod \"60de9d04-7c24-42a4-9d5e-74b611487b69\" (UID: \"60de9d04-7c24-42a4-9d5e-74b611487b69\") " Oct 14 07:23:51 crc kubenswrapper[4870]: I1014 07:23:51.332196 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/60de9d04-7c24-42a4-9d5e-74b611487b69-config-data\") pod \"60de9d04-7c24-42a4-9d5e-74b611487b69\" (UID: \"60de9d04-7c24-42a4-9d5e-74b611487b69\") " Oct 14 07:23:51 crc kubenswrapper[4870]: I1014 07:23:51.332219 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/60de9d04-7c24-42a4-9d5e-74b611487b69-public-tls-certs\") pod \"60de9d04-7c24-42a4-9d5e-74b611487b69\" (UID: \"60de9d04-7c24-42a4-9d5e-74b611487b69\") " Oct 14 07:23:51 crc kubenswrapper[4870]: I1014 07:23:51.332810 4870 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/416d7b39-4541-44a8-a55f-924bc86fee32-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:51 crc kubenswrapper[4870]: I1014 07:23:51.332825 4870 reconciler_common.go:293] "Volume detached for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/416d7b39-4541-44a8-a55f-924bc86fee32-ovn-rundir\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:51 crc kubenswrapper[4870]: I1014 07:23:51.332836 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d28jg\" (UniqueName: \"kubernetes.io/projected/416d7b39-4541-44a8-a55f-924bc86fee32-kube-api-access-d28jg\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:51 crc kubenswrapper[4870]: I1014 07:23:51.332846 4870 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/416d7b39-4541-44a8-a55f-924bc86fee32-config\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:51 crc kubenswrapper[4870]: I1014 07:23:51.332854 4870 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/416d7b39-4541-44a8-a55f-924bc86fee32-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:51 crc kubenswrapper[4870]: I1014 07:23:51.336791 4870 scope.go:117] "RemoveContainer" containerID="a507ea103a81fde95c96ade7a313c18b4faefbe53e02c99e5cd596e0c8dcb53e" Oct 14 07:23:51 crc kubenswrapper[4870]: I1014 07:23:51.338928 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/416d7b39-4541-44a8-a55f-924bc86fee32-ovn-northd-tls-certs" (OuterVolumeSpecName: "ovn-northd-tls-certs") pod "416d7b39-4541-44a8-a55f-924bc86fee32" (UID: "416d7b39-4541-44a8-a55f-924bc86fee32"). InnerVolumeSpecName "ovn-northd-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:23:51 crc kubenswrapper[4870]: I1014 07:23:51.347098 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/60de9d04-7c24-42a4-9d5e-74b611487b69-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "60de9d04-7c24-42a4-9d5e-74b611487b69" (UID: "60de9d04-7c24-42a4-9d5e-74b611487b69"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:23:51 crc kubenswrapper[4870]: I1014 07:23:51.349232 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/60de9d04-7c24-42a4-9d5e-74b611487b69-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "60de9d04-7c24-42a4-9d5e-74b611487b69" (UID: "60de9d04-7c24-42a4-9d5e-74b611487b69"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:23:51 crc kubenswrapper[4870]: I1014 07:23:51.351680 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/60de9d04-7c24-42a4-9d5e-74b611487b69-kube-api-access-vhds6" (OuterVolumeSpecName: "kube-api-access-vhds6") pod "60de9d04-7c24-42a4-9d5e-74b611487b69" (UID: "60de9d04-7c24-42a4-9d5e-74b611487b69"). InnerVolumeSpecName "kube-api-access-vhds6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:23:51 crc kubenswrapper[4870]: I1014 07:23:51.351794 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/60de9d04-7c24-42a4-9d5e-74b611487b69-scripts" (OuterVolumeSpecName: "scripts") pod "60de9d04-7c24-42a4-9d5e-74b611487b69" (UID: "60de9d04-7c24-42a4-9d5e-74b611487b69"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:23:51 crc kubenswrapper[4870]: I1014 07:23:51.368030 4870 scope.go:117] "RemoveContainer" containerID="b97a3d2764961f388fdc1a6459d91ca5f1248920ce92c572b651657c109e7fa4" Oct 14 07:23:51 crc kubenswrapper[4870]: I1014 07:23:51.370733 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/60de9d04-7c24-42a4-9d5e-74b611487b69-config-data" (OuterVolumeSpecName: "config-data") pod "60de9d04-7c24-42a4-9d5e-74b611487b69" (UID: "60de9d04-7c24-42a4-9d5e-74b611487b69"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:23:51 crc kubenswrapper[4870]: E1014 07:23:51.372159 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b97a3d2764961f388fdc1a6459d91ca5f1248920ce92c572b651657c109e7fa4\": container with ID starting with b97a3d2764961f388fdc1a6459d91ca5f1248920ce92c572b651657c109e7fa4 not found: ID does not exist" containerID="b97a3d2764961f388fdc1a6459d91ca5f1248920ce92c572b651657c109e7fa4" Oct 14 07:23:51 crc kubenswrapper[4870]: I1014 07:23:51.372258 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b97a3d2764961f388fdc1a6459d91ca5f1248920ce92c572b651657c109e7fa4"} err="failed to get container status \"b97a3d2764961f388fdc1a6459d91ca5f1248920ce92c572b651657c109e7fa4\": rpc error: code = NotFound desc = could not find container \"b97a3d2764961f388fdc1a6459d91ca5f1248920ce92c572b651657c109e7fa4\": container with ID starting with b97a3d2764961f388fdc1a6459d91ca5f1248920ce92c572b651657c109e7fa4 not found: ID does not exist" Oct 14 07:23:51 crc kubenswrapper[4870]: I1014 07:23:51.372360 4870 scope.go:117] "RemoveContainer" containerID="a507ea103a81fde95c96ade7a313c18b4faefbe53e02c99e5cd596e0c8dcb53e" Oct 14 07:23:51 crc kubenswrapper[4870]: E1014 07:23:51.375466 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a507ea103a81fde95c96ade7a313c18b4faefbe53e02c99e5cd596e0c8dcb53e\": container with ID starting with a507ea103a81fde95c96ade7a313c18b4faefbe53e02c99e5cd596e0c8dcb53e not found: ID does not exist" containerID="a507ea103a81fde95c96ade7a313c18b4faefbe53e02c99e5cd596e0c8dcb53e" Oct 14 07:23:51 crc kubenswrapper[4870]: I1014 07:23:51.375519 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a507ea103a81fde95c96ade7a313c18b4faefbe53e02c99e5cd596e0c8dcb53e"} err="failed to get container status \"a507ea103a81fde95c96ade7a313c18b4faefbe53e02c99e5cd596e0c8dcb53e\": rpc error: code = NotFound desc = could not find container \"a507ea103a81fde95c96ade7a313c18b4faefbe53e02c99e5cd596e0c8dcb53e\": container with ID starting with a507ea103a81fde95c96ade7a313c18b4faefbe53e02c99e5cd596e0c8dcb53e not found: ID does not exist" Oct 14 07:23:51 crc kubenswrapper[4870]: I1014 07:23:51.375558 4870 scope.go:117] "RemoveContainer" containerID="4339244c7626324d778551bfb8f861a2c704b3e5aa7199c34ba643e475017090" Oct 14 07:23:51 crc kubenswrapper[4870]: I1014 07:23:51.396660 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/60de9d04-7c24-42a4-9d5e-74b611487b69-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "60de9d04-7c24-42a4-9d5e-74b611487b69" (UID: "60de9d04-7c24-42a4-9d5e-74b611487b69"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:23:51 crc kubenswrapper[4870]: I1014 07:23:51.397807 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/60de9d04-7c24-42a4-9d5e-74b611487b69-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "60de9d04-7c24-42a4-9d5e-74b611487b69" (UID: "60de9d04-7c24-42a4-9d5e-74b611487b69"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:23:51 crc kubenswrapper[4870]: I1014 07:23:51.399761 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/416d7b39-4541-44a8-a55f-924bc86fee32-metrics-certs-tls-certs" (OuterVolumeSpecName: "metrics-certs-tls-certs") pod "416d7b39-4541-44a8-a55f-924bc86fee32" (UID: "416d7b39-4541-44a8-a55f-924bc86fee32"). InnerVolumeSpecName "metrics-certs-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:23:51 crc kubenswrapper[4870]: I1014 07:23:51.414949 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 14 07:23:51 crc kubenswrapper[4870]: I1014 07:23:51.420861 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/60de9d04-7c24-42a4-9d5e-74b611487b69-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "60de9d04-7c24-42a4-9d5e-74b611487b69" (UID: "60de9d04-7c24-42a4-9d5e-74b611487b69"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:23:51 crc kubenswrapper[4870]: I1014 07:23:51.421748 4870 scope.go:117] "RemoveContainer" containerID="f4ef48299206032c6fdb947f48913c54fe5fa0a39a9b5ffe9e3ea9fd8ac49f8e" Oct 14 07:23:51 crc kubenswrapper[4870]: I1014 07:23:51.434070 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/5212ba20-83c3-41ff-a3d9-5a315f009c1b-rabbitmq-erlang-cookie\") pod \"5212ba20-83c3-41ff-a3d9-5a315f009c1b\" (UID: \"5212ba20-83c3-41ff-a3d9-5a315f009c1b\") " Oct 14 07:23:51 crc kubenswrapper[4870]: I1014 07:23:51.434112 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/5212ba20-83c3-41ff-a3d9-5a315f009c1b-pod-info\") pod \"5212ba20-83c3-41ff-a3d9-5a315f009c1b\" (UID: \"5212ba20-83c3-41ff-a3d9-5a315f009c1b\") " Oct 14 07:23:51 crc kubenswrapper[4870]: I1014 07:23:51.434188 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/5212ba20-83c3-41ff-a3d9-5a315f009c1b-rabbitmq-plugins\") pod \"5212ba20-83c3-41ff-a3d9-5a315f009c1b\" (UID: \"5212ba20-83c3-41ff-a3d9-5a315f009c1b\") " Oct 14 07:23:51 crc kubenswrapper[4870]: I1014 07:23:51.434209 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/5212ba20-83c3-41ff-a3d9-5a315f009c1b-plugins-conf\") pod \"5212ba20-83c3-41ff-a3d9-5a315f009c1b\" (UID: \"5212ba20-83c3-41ff-a3d9-5a315f009c1b\") " Oct 14 07:23:51 crc kubenswrapper[4870]: I1014 07:23:51.434269 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5212ba20-83c3-41ff-a3d9-5a315f009c1b-config-data\") pod \"5212ba20-83c3-41ff-a3d9-5a315f009c1b\" (UID: \"5212ba20-83c3-41ff-a3d9-5a315f009c1b\") " Oct 14 07:23:51 crc kubenswrapper[4870]: I1014 07:23:51.434286 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/5212ba20-83c3-41ff-a3d9-5a315f009c1b-rabbitmq-tls\") pod \"5212ba20-83c3-41ff-a3d9-5a315f009c1b\" (UID: \"5212ba20-83c3-41ff-a3d9-5a315f009c1b\") " Oct 14 07:23:51 crc kubenswrapper[4870]: I1014 07:23:51.434316 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/5212ba20-83c3-41ff-a3d9-5a315f009c1b-server-conf\") pod \"5212ba20-83c3-41ff-a3d9-5a315f009c1b\" (UID: \"5212ba20-83c3-41ff-a3d9-5a315f009c1b\") " Oct 14 07:23:51 crc kubenswrapper[4870]: I1014 07:23:51.434340 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"5212ba20-83c3-41ff-a3d9-5a315f009c1b\" (UID: \"5212ba20-83c3-41ff-a3d9-5a315f009c1b\") " Oct 14 07:23:51 crc kubenswrapper[4870]: I1014 07:23:51.434362 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/5212ba20-83c3-41ff-a3d9-5a315f009c1b-rabbitmq-confd\") pod \"5212ba20-83c3-41ff-a3d9-5a315f009c1b\" (UID: \"5212ba20-83c3-41ff-a3d9-5a315f009c1b\") " Oct 14 07:23:51 crc kubenswrapper[4870]: I1014 07:23:51.434380 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/5212ba20-83c3-41ff-a3d9-5a315f009c1b-erlang-cookie-secret\") pod \"5212ba20-83c3-41ff-a3d9-5a315f009c1b\" (UID: \"5212ba20-83c3-41ff-a3d9-5a315f009c1b\") " Oct 14 07:23:51 crc kubenswrapper[4870]: I1014 07:23:51.434403 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4r8n\" (UniqueName: \"kubernetes.io/projected/5212ba20-83c3-41ff-a3d9-5a315f009c1b-kube-api-access-d4r8n\") pod \"5212ba20-83c3-41ff-a3d9-5a315f009c1b\" (UID: \"5212ba20-83c3-41ff-a3d9-5a315f009c1b\") " Oct 14 07:23:51 crc kubenswrapper[4870]: I1014 07:23:51.434674 4870 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/60de9d04-7c24-42a4-9d5e-74b611487b69-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:51 crc kubenswrapper[4870]: I1014 07:23:51.434686 4870 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/60de9d04-7c24-42a4-9d5e-74b611487b69-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:51 crc kubenswrapper[4870]: I1014 07:23:51.434696 4870 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/60de9d04-7c24-42a4-9d5e-74b611487b69-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:51 crc kubenswrapper[4870]: I1014 07:23:51.434705 4870 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60de9d04-7c24-42a4-9d5e-74b611487b69-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:51 crc kubenswrapper[4870]: I1014 07:23:51.434713 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vhds6\" (UniqueName: \"kubernetes.io/projected/60de9d04-7c24-42a4-9d5e-74b611487b69-kube-api-access-vhds6\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:51 crc kubenswrapper[4870]: I1014 07:23:51.434722 4870 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/60de9d04-7c24-42a4-9d5e-74b611487b69-credential-keys\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:51 crc kubenswrapper[4870]: I1014 07:23:51.434743 4870 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/60de9d04-7c24-42a4-9d5e-74b611487b69-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:51 crc kubenswrapper[4870]: I1014 07:23:51.434752 4870 reconciler_common.go:293] "Volume detached for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/416d7b39-4541-44a8-a55f-924bc86fee32-ovn-northd-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:51 crc kubenswrapper[4870]: I1014 07:23:51.434759 4870 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/60de9d04-7c24-42a4-9d5e-74b611487b69-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:51 crc kubenswrapper[4870]: I1014 07:23:51.434767 4870 reconciler_common.go:293] "Volume detached for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/416d7b39-4541-44a8-a55f-924bc86fee32-metrics-certs-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:51 crc kubenswrapper[4870]: I1014 07:23:51.436003 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5212ba20-83c3-41ff-a3d9-5a315f009c1b-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "5212ba20-83c3-41ff-a3d9-5a315f009c1b" (UID: "5212ba20-83c3-41ff-a3d9-5a315f009c1b"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:23:51 crc kubenswrapper[4870]: I1014 07:23:51.436029 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5212ba20-83c3-41ff-a3d9-5a315f009c1b-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "5212ba20-83c3-41ff-a3d9-5a315f009c1b" (UID: "5212ba20-83c3-41ff-a3d9-5a315f009c1b"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:23:51 crc kubenswrapper[4870]: I1014 07:23:51.436139 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5212ba20-83c3-41ff-a3d9-5a315f009c1b-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "5212ba20-83c3-41ff-a3d9-5a315f009c1b" (UID: "5212ba20-83c3-41ff-a3d9-5a315f009c1b"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:23:51 crc kubenswrapper[4870]: I1014 07:23:51.440020 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage04-crc" (OuterVolumeSpecName: "persistence") pod "5212ba20-83c3-41ff-a3d9-5a315f009c1b" (UID: "5212ba20-83c3-41ff-a3d9-5a315f009c1b"). InnerVolumeSpecName "local-storage04-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 14 07:23:51 crc kubenswrapper[4870]: I1014 07:23:51.440781 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/5212ba20-83c3-41ff-a3d9-5a315f009c1b-pod-info" (OuterVolumeSpecName: "pod-info") pod "5212ba20-83c3-41ff-a3d9-5a315f009c1b" (UID: "5212ba20-83c3-41ff-a3d9-5a315f009c1b"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Oct 14 07:23:51 crc kubenswrapper[4870]: I1014 07:23:51.441424 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5212ba20-83c3-41ff-a3d9-5a315f009c1b-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "5212ba20-83c3-41ff-a3d9-5a315f009c1b" (UID: "5212ba20-83c3-41ff-a3d9-5a315f009c1b"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:23:51 crc kubenswrapper[4870]: I1014 07:23:51.442524 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5212ba20-83c3-41ff-a3d9-5a315f009c1b-kube-api-access-d4r8n" (OuterVolumeSpecName: "kube-api-access-d4r8n") pod "5212ba20-83c3-41ff-a3d9-5a315f009c1b" (UID: "5212ba20-83c3-41ff-a3d9-5a315f009c1b"). InnerVolumeSpecName "kube-api-access-d4r8n". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:23:51 crc kubenswrapper[4870]: I1014 07:23:51.445504 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5212ba20-83c3-41ff-a3d9-5a315f009c1b-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "5212ba20-83c3-41ff-a3d9-5a315f009c1b" (UID: "5212ba20-83c3-41ff-a3d9-5a315f009c1b"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:23:51 crc kubenswrapper[4870]: I1014 07:23:51.453322 4870 scope.go:117] "RemoveContainer" containerID="d8fce18d04004b32792441ce25ddcd9d74c06df092b20d09c14fca60675539d4" Oct 14 07:23:51 crc kubenswrapper[4870]: I1014 07:23:51.454814 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5212ba20-83c3-41ff-a3d9-5a315f009c1b-config-data" (OuterVolumeSpecName: "config-data") pod "5212ba20-83c3-41ff-a3d9-5a315f009c1b" (UID: "5212ba20-83c3-41ff-a3d9-5a315f009c1b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:23:51 crc kubenswrapper[4870]: I1014 07:23:51.474178 4870 scope.go:117] "RemoveContainer" containerID="9fa2f1146313b498f55c9bb3361bca48d735f25641dcd24bdbeac19bc4ae6377" Oct 14 07:23:51 crc kubenswrapper[4870]: I1014 07:23:51.479265 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5212ba20-83c3-41ff-a3d9-5a315f009c1b-server-conf" (OuterVolumeSpecName: "server-conf") pod "5212ba20-83c3-41ff-a3d9-5a315f009c1b" (UID: "5212ba20-83c3-41ff-a3d9-5a315f009c1b"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:23:51 crc kubenswrapper[4870]: I1014 07:23:51.496779 4870 scope.go:117] "RemoveContainer" containerID="19bfdce54348e895d0d26a8651c758e77337522908dc3ad6eb2470e00c72c2cf" Oct 14 07:23:51 crc kubenswrapper[4870]: I1014 07:23:51.514149 4870 scope.go:117] "RemoveContainer" containerID="1b53e989d30152b230c176967064509ca1fafba96b986cc96bc23bdcafabf49f" Oct 14 07:23:51 crc kubenswrapper[4870]: I1014 07:23:51.521665 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5212ba20-83c3-41ff-a3d9-5a315f009c1b-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "5212ba20-83c3-41ff-a3d9-5a315f009c1b" (UID: "5212ba20-83c3-41ff-a3d9-5a315f009c1b"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:23:51 crc kubenswrapper[4870]: I1014 07:23:51.536222 4870 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/5212ba20-83c3-41ff-a3d9-5a315f009c1b-server-conf\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:51 crc kubenswrapper[4870]: I1014 07:23:51.536279 4870 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" " Oct 14 07:23:51 crc kubenswrapper[4870]: I1014 07:23:51.536291 4870 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/5212ba20-83c3-41ff-a3d9-5a315f009c1b-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:51 crc kubenswrapper[4870]: I1014 07:23:51.536302 4870 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/5212ba20-83c3-41ff-a3d9-5a315f009c1b-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:51 crc kubenswrapper[4870]: I1014 07:23:51.536314 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4r8n\" (UniqueName: \"kubernetes.io/projected/5212ba20-83c3-41ff-a3d9-5a315f009c1b-kube-api-access-d4r8n\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:51 crc kubenswrapper[4870]: I1014 07:23:51.536322 4870 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/5212ba20-83c3-41ff-a3d9-5a315f009c1b-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:51 crc kubenswrapper[4870]: I1014 07:23:51.536331 4870 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/5212ba20-83c3-41ff-a3d9-5a315f009c1b-pod-info\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:51 crc kubenswrapper[4870]: I1014 07:23:51.536339 4870 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/5212ba20-83c3-41ff-a3d9-5a315f009c1b-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:51 crc kubenswrapper[4870]: I1014 07:23:51.536347 4870 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/5212ba20-83c3-41ff-a3d9-5a315f009c1b-plugins-conf\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:51 crc kubenswrapper[4870]: I1014 07:23:51.536354 4870 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5212ba20-83c3-41ff-a3d9-5a315f009c1b-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:51 crc kubenswrapper[4870]: I1014 07:23:51.536362 4870 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/5212ba20-83c3-41ff-a3d9-5a315f009c1b-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:51 crc kubenswrapper[4870]: I1014 07:23:51.547602 4870 scope.go:117] "RemoveContainer" containerID="eb1654e601e030081e70bbe69e7b024b64dd889e97e317b911a3c3468c7fcfb8" Oct 14 07:23:51 crc kubenswrapper[4870]: I1014 07:23:51.552730 4870 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage04-crc" (UniqueName: "kubernetes.io/local-volume/local-storage04-crc") on node "crc" Oct 14 07:23:51 crc kubenswrapper[4870]: I1014 07:23:51.574726 4870 scope.go:117] "RemoveContainer" containerID="1b53e989d30152b230c176967064509ca1fafba96b986cc96bc23bdcafabf49f" Oct 14 07:23:51 crc kubenswrapper[4870]: E1014 07:23:51.575239 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1b53e989d30152b230c176967064509ca1fafba96b986cc96bc23bdcafabf49f\": container with ID starting with 1b53e989d30152b230c176967064509ca1fafba96b986cc96bc23bdcafabf49f not found: ID does not exist" containerID="1b53e989d30152b230c176967064509ca1fafba96b986cc96bc23bdcafabf49f" Oct 14 07:23:51 crc kubenswrapper[4870]: I1014 07:23:51.575274 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1b53e989d30152b230c176967064509ca1fafba96b986cc96bc23bdcafabf49f"} err="failed to get container status \"1b53e989d30152b230c176967064509ca1fafba96b986cc96bc23bdcafabf49f\": rpc error: code = NotFound desc = could not find container \"1b53e989d30152b230c176967064509ca1fafba96b986cc96bc23bdcafabf49f\": container with ID starting with 1b53e989d30152b230c176967064509ca1fafba96b986cc96bc23bdcafabf49f not found: ID does not exist" Oct 14 07:23:51 crc kubenswrapper[4870]: I1014 07:23:51.575301 4870 scope.go:117] "RemoveContainer" containerID="eb1654e601e030081e70bbe69e7b024b64dd889e97e317b911a3c3468c7fcfb8" Oct 14 07:23:51 crc kubenswrapper[4870]: E1014 07:23:51.575578 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eb1654e601e030081e70bbe69e7b024b64dd889e97e317b911a3c3468c7fcfb8\": container with ID starting with eb1654e601e030081e70bbe69e7b024b64dd889e97e317b911a3c3468c7fcfb8 not found: ID does not exist" containerID="eb1654e601e030081e70bbe69e7b024b64dd889e97e317b911a3c3468c7fcfb8" Oct 14 07:23:51 crc kubenswrapper[4870]: I1014 07:23:51.575609 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eb1654e601e030081e70bbe69e7b024b64dd889e97e317b911a3c3468c7fcfb8"} err="failed to get container status \"eb1654e601e030081e70bbe69e7b024b64dd889e97e317b911a3c3468c7fcfb8\": rpc error: code = NotFound desc = could not find container \"eb1654e601e030081e70bbe69e7b024b64dd889e97e317b911a3c3468c7fcfb8\": container with ID starting with eb1654e601e030081e70bbe69e7b024b64dd889e97e317b911a3c3468c7fcfb8 not found: ID does not exist" Oct 14 07:23:51 crc kubenswrapper[4870]: I1014 07:23:51.575626 4870 scope.go:117] "RemoveContainer" containerID="a1f9a0d181ab7f3f80bee6c1008e3b393dd082ba31b2614d29e2a4b0d5534a46" Oct 14 07:23:51 crc kubenswrapper[4870]: I1014 07:23:51.609297 4870 scope.go:117] "RemoveContainer" containerID="e5c43911d04a168b064b8b753711f3e5ed521dc55a16d5687c49934c8d0eb3a6" Oct 14 07:23:51 crc kubenswrapper[4870]: I1014 07:23:51.637793 4870 reconciler_common.go:293] "Volume detached for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:51 crc kubenswrapper[4870]: E1014 07:23:51.638274 4870 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 6f14317c64b4cb248d4254d308d86fadbdb28d2daee87abbf75414c219c2c477 is running failed: container process not found" containerID="6f14317c64b4cb248d4254d308d86fadbdb28d2daee87abbf75414c219c2c477" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 14 07:23:51 crc kubenswrapper[4870]: E1014 07:23:51.638589 4870 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="72ead03af69bb2d536002fe16f85a88974968529ec05c0e9e78254a3f796b690" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 14 07:23:51 crc kubenswrapper[4870]: E1014 07:23:51.638700 4870 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 6f14317c64b4cb248d4254d308d86fadbdb28d2daee87abbf75414c219c2c477 is running failed: container process not found" containerID="6f14317c64b4cb248d4254d308d86fadbdb28d2daee87abbf75414c219c2c477" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 14 07:23:51 crc kubenswrapper[4870]: E1014 07:23:51.639094 4870 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 6f14317c64b4cb248d4254d308d86fadbdb28d2daee87abbf75414c219c2c477 is running failed: container process not found" containerID="6f14317c64b4cb248d4254d308d86fadbdb28d2daee87abbf75414c219c2c477" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 14 07:23:51 crc kubenswrapper[4870]: E1014 07:23:51.639133 4870 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 6f14317c64b4cb248d4254d308d86fadbdb28d2daee87abbf75414c219c2c477 is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-6482n" podUID="91a2e1de-104f-4e14-9192-c6587ed13be1" containerName="ovsdb-server" Oct 14 07:23:51 crc kubenswrapper[4870]: E1014 07:23:51.639826 4870 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="72ead03af69bb2d536002fe16f85a88974968529ec05c0e9e78254a3f796b690" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 14 07:23:51 crc kubenswrapper[4870]: E1014 07:23:51.642866 4870 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="72ead03af69bb2d536002fe16f85a88974968529ec05c0e9e78254a3f796b690" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 14 07:23:51 crc kubenswrapper[4870]: E1014 07:23:51.643016 4870 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-6482n" podUID="91a2e1de-104f-4e14-9192-c6587ed13be1" containerName="ovs-vswitchd" Oct 14 07:23:52 crc kubenswrapper[4870]: I1014 07:23:52.146320 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-865c567797-jljrm" event={"ID":"60de9d04-7c24-42a4-9d5e-74b611487b69","Type":"ContainerDied","Data":"bb99a031932012bc36d4d56aba1c27bfd8482f6a5153da2e1843ed879054a61a"} Oct 14 07:23:52 crc kubenswrapper[4870]: I1014 07:23:52.146380 4870 scope.go:117] "RemoveContainer" containerID="d8615c16655ea7e7d454161d16b61973188b63bffc966ed4b3805db817746641" Oct 14 07:23:52 crc kubenswrapper[4870]: I1014 07:23:52.146334 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-865c567797-jljrm" Oct 14 07:23:52 crc kubenswrapper[4870]: I1014 07:23:52.161195 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 14 07:23:52 crc kubenswrapper[4870]: I1014 07:23:52.161640 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"5212ba20-83c3-41ff-a3d9-5a315f009c1b","Type":"ContainerDied","Data":"d98e1ffb219c955289fa2faf4924c6941b1140c0dffec8bed2ba720a0fc61177"} Oct 14 07:23:52 crc kubenswrapper[4870]: I1014 07:23:52.179006 4870 scope.go:117] "RemoveContainer" containerID="eeb00ec87974e56f742f7ae182732d1dfbbc2c6523987de46327c67f0a729b18" Oct 14 07:23:52 crc kubenswrapper[4870]: I1014 07:23:52.194896 4870 generic.go:334] "Generic (PLEG): container finished" podID="da8a9731-f917-462f-9932-b37b6abb9a64" containerID="bc4b5271d735615160d1289dff31839e1ba9c4495e333ee1f50cab05851fd959" exitCode=0 Oct 14 07:23:52 crc kubenswrapper[4870]: I1014 07:23:52.194987 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Oct 14 07:23:52 crc kubenswrapper[4870]: I1014 07:23:52.196522 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"da8a9731-f917-462f-9932-b37b6abb9a64","Type":"ContainerDied","Data":"bc4b5271d735615160d1289dff31839e1ba9c4495e333ee1f50cab05851fd959"} Oct 14 07:23:52 crc kubenswrapper[4870]: I1014 07:23:52.198150 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-865c567797-jljrm"] Oct 14 07:23:52 crc kubenswrapper[4870]: I1014 07:23:52.207195 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-865c567797-jljrm"] Oct 14 07:23:52 crc kubenswrapper[4870]: I1014 07:23:52.225287 4870 scope.go:117] "RemoveContainer" containerID="80776853a799d9822a89d9d86e1010ddb56ba32acd904f622c45923cef6744ac" Oct 14 07:23:52 crc kubenswrapper[4870]: I1014 07:23:52.228209 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 14 07:23:52 crc kubenswrapper[4870]: I1014 07:23:52.236176 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 14 07:23:52 crc kubenswrapper[4870]: I1014 07:23:52.243181 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-northd-0"] Oct 14 07:23:52 crc kubenswrapper[4870]: I1014 07:23:52.250425 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-northd-0"] Oct 14 07:23:52 crc kubenswrapper[4870]: I1014 07:23:52.326224 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 14 07:23:52 crc kubenswrapper[4870]: I1014 07:23:52.353360 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nl2bs\" (UniqueName: \"kubernetes.io/projected/da8a9731-f917-462f-9932-b37b6abb9a64-kube-api-access-nl2bs\") pod \"da8a9731-f917-462f-9932-b37b6abb9a64\" (UID: \"da8a9731-f917-462f-9932-b37b6abb9a64\") " Oct 14 07:23:52 crc kubenswrapper[4870]: I1014 07:23:52.353403 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/da8a9731-f917-462f-9932-b37b6abb9a64-rabbitmq-erlang-cookie\") pod \"da8a9731-f917-462f-9932-b37b6abb9a64\" (UID: \"da8a9731-f917-462f-9932-b37b6abb9a64\") " Oct 14 07:23:52 crc kubenswrapper[4870]: I1014 07:23:52.353425 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/da8a9731-f917-462f-9932-b37b6abb9a64-rabbitmq-confd\") pod \"da8a9731-f917-462f-9932-b37b6abb9a64\" (UID: \"da8a9731-f917-462f-9932-b37b6abb9a64\") " Oct 14 07:23:52 crc kubenswrapper[4870]: I1014 07:23:52.353462 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/da8a9731-f917-462f-9932-b37b6abb9a64-erlang-cookie-secret\") pod \"da8a9731-f917-462f-9932-b37b6abb9a64\" (UID: \"da8a9731-f917-462f-9932-b37b6abb9a64\") " Oct 14 07:23:52 crc kubenswrapper[4870]: I1014 07:23:52.353486 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/da8a9731-f917-462f-9932-b37b6abb9a64-plugins-conf\") pod \"da8a9731-f917-462f-9932-b37b6abb9a64\" (UID: \"da8a9731-f917-462f-9932-b37b6abb9a64\") " Oct 14 07:23:52 crc kubenswrapper[4870]: I1014 07:23:52.353504 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/da8a9731-f917-462f-9932-b37b6abb9a64-rabbitmq-tls\") pod \"da8a9731-f917-462f-9932-b37b6abb9a64\" (UID: \"da8a9731-f917-462f-9932-b37b6abb9a64\") " Oct 14 07:23:52 crc kubenswrapper[4870]: I1014 07:23:52.353537 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/da8a9731-f917-462f-9932-b37b6abb9a64-config-data\") pod \"da8a9731-f917-462f-9932-b37b6abb9a64\" (UID: \"da8a9731-f917-462f-9932-b37b6abb9a64\") " Oct 14 07:23:52 crc kubenswrapper[4870]: I1014 07:23:52.353560 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/da8a9731-f917-462f-9932-b37b6abb9a64-server-conf\") pod \"da8a9731-f917-462f-9932-b37b6abb9a64\" (UID: \"da8a9731-f917-462f-9932-b37b6abb9a64\") " Oct 14 07:23:52 crc kubenswrapper[4870]: I1014 07:23:52.353577 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/da8a9731-f917-462f-9932-b37b6abb9a64-pod-info\") pod \"da8a9731-f917-462f-9932-b37b6abb9a64\" (UID: \"da8a9731-f917-462f-9932-b37b6abb9a64\") " Oct 14 07:23:52 crc kubenswrapper[4870]: I1014 07:23:52.353597 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/da8a9731-f917-462f-9932-b37b6abb9a64-rabbitmq-plugins\") pod \"da8a9731-f917-462f-9932-b37b6abb9a64\" (UID: \"da8a9731-f917-462f-9932-b37b6abb9a64\") " Oct 14 07:23:52 crc kubenswrapper[4870]: I1014 07:23:52.353614 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"da8a9731-f917-462f-9932-b37b6abb9a64\" (UID: \"da8a9731-f917-462f-9932-b37b6abb9a64\") " Oct 14 07:23:52 crc kubenswrapper[4870]: I1014 07:23:52.358509 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/da8a9731-f917-462f-9932-b37b6abb9a64-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "da8a9731-f917-462f-9932-b37b6abb9a64" (UID: "da8a9731-f917-462f-9932-b37b6abb9a64"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:23:52 crc kubenswrapper[4870]: I1014 07:23:52.359925 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/da8a9731-f917-462f-9932-b37b6abb9a64-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "da8a9731-f917-462f-9932-b37b6abb9a64" (UID: "da8a9731-f917-462f-9932-b37b6abb9a64"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:23:52 crc kubenswrapper[4870]: I1014 07:23:52.361900 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/da8a9731-f917-462f-9932-b37b6abb9a64-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "da8a9731-f917-462f-9932-b37b6abb9a64" (UID: "da8a9731-f917-462f-9932-b37b6abb9a64"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:23:52 crc kubenswrapper[4870]: I1014 07:23:52.362379 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/da8a9731-f917-462f-9932-b37b6abb9a64-kube-api-access-nl2bs" (OuterVolumeSpecName: "kube-api-access-nl2bs") pod "da8a9731-f917-462f-9932-b37b6abb9a64" (UID: "da8a9731-f917-462f-9932-b37b6abb9a64"). InnerVolumeSpecName "kube-api-access-nl2bs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:23:52 crc kubenswrapper[4870]: I1014 07:23:52.363984 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/da8a9731-f917-462f-9932-b37b6abb9a64-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "da8a9731-f917-462f-9932-b37b6abb9a64" (UID: "da8a9731-f917-462f-9932-b37b6abb9a64"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:23:52 crc kubenswrapper[4870]: I1014 07:23:52.364555 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage02-crc" (OuterVolumeSpecName: "persistence") pod "da8a9731-f917-462f-9932-b37b6abb9a64" (UID: "da8a9731-f917-462f-9932-b37b6abb9a64"). InnerVolumeSpecName "local-storage02-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 14 07:23:52 crc kubenswrapper[4870]: I1014 07:23:52.365439 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/da8a9731-f917-462f-9932-b37b6abb9a64-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "da8a9731-f917-462f-9932-b37b6abb9a64" (UID: "da8a9731-f917-462f-9932-b37b6abb9a64"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:23:52 crc kubenswrapper[4870]: I1014 07:23:52.378464 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/da8a9731-f917-462f-9932-b37b6abb9a64-pod-info" (OuterVolumeSpecName: "pod-info") pod "da8a9731-f917-462f-9932-b37b6abb9a64" (UID: "da8a9731-f917-462f-9932-b37b6abb9a64"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Oct 14 07:23:52 crc kubenswrapper[4870]: I1014 07:23:52.382614 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/da8a9731-f917-462f-9932-b37b6abb9a64-config-data" (OuterVolumeSpecName: "config-data") pod "da8a9731-f917-462f-9932-b37b6abb9a64" (UID: "da8a9731-f917-462f-9932-b37b6abb9a64"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:23:52 crc kubenswrapper[4870]: I1014 07:23:52.428936 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/da8a9731-f917-462f-9932-b37b6abb9a64-server-conf" (OuterVolumeSpecName: "server-conf") pod "da8a9731-f917-462f-9932-b37b6abb9a64" (UID: "da8a9731-f917-462f-9932-b37b6abb9a64"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:23:52 crc kubenswrapper[4870]: I1014 07:23:52.455249 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nl2bs\" (UniqueName: \"kubernetes.io/projected/da8a9731-f917-462f-9932-b37b6abb9a64-kube-api-access-nl2bs\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:52 crc kubenswrapper[4870]: I1014 07:23:52.455280 4870 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/da8a9731-f917-462f-9932-b37b6abb9a64-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:52 crc kubenswrapper[4870]: I1014 07:23:52.455291 4870 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/da8a9731-f917-462f-9932-b37b6abb9a64-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:52 crc kubenswrapper[4870]: I1014 07:23:52.455301 4870 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/da8a9731-f917-462f-9932-b37b6abb9a64-plugins-conf\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:52 crc kubenswrapper[4870]: I1014 07:23:52.455311 4870 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/da8a9731-f917-462f-9932-b37b6abb9a64-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:52 crc kubenswrapper[4870]: I1014 07:23:52.455319 4870 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/da8a9731-f917-462f-9932-b37b6abb9a64-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:52 crc kubenswrapper[4870]: I1014 07:23:52.455327 4870 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/da8a9731-f917-462f-9932-b37b6abb9a64-server-conf\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:52 crc kubenswrapper[4870]: I1014 07:23:52.455335 4870 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/da8a9731-f917-462f-9932-b37b6abb9a64-pod-info\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:52 crc kubenswrapper[4870]: I1014 07:23:52.455343 4870 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/da8a9731-f917-462f-9932-b37b6abb9a64-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:52 crc kubenswrapper[4870]: I1014 07:23:52.455373 4870 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" " Oct 14 07:23:52 crc kubenswrapper[4870]: I1014 07:23:52.469514 4870 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage02-crc" (UniqueName: "kubernetes.io/local-volume/local-storage02-crc") on node "crc" Oct 14 07:23:52 crc kubenswrapper[4870]: I1014 07:23:52.487142 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/da8a9731-f917-462f-9932-b37b6abb9a64-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "da8a9731-f917-462f-9932-b37b6abb9a64" (UID: "da8a9731-f917-462f-9932-b37b6abb9a64"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:23:52 crc kubenswrapper[4870]: I1014 07:23:52.558310 4870 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/da8a9731-f917-462f-9932-b37b6abb9a64-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:52 crc kubenswrapper[4870]: I1014 07:23:52.558345 4870 reconciler_common.go:293] "Volume detached for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:52 crc kubenswrapper[4870]: I1014 07:23:52.851161 4870 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/swift-proxy-5485b69bc9-ftmmq" podUID="3927fb78-4f30-4371-8fbb-62fd276c0460" containerName="proxy-httpd" probeResult="failure" output="Get \"https://10.217.0.162:8080/healthcheck\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Oct 14 07:23:52 crc kubenswrapper[4870]: I1014 07:23:52.851319 4870 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/swift-proxy-5485b69bc9-ftmmq" podUID="3927fb78-4f30-4371-8fbb-62fd276c0460" containerName="proxy-server" probeResult="failure" output="Get \"https://10.217.0.162:8080/healthcheck\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Oct 14 07:23:53 crc kubenswrapper[4870]: I1014 07:23:53.043984 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3de2167a-7663-4b00-9743-229ff1864cb8" path="/var/lib/kubelet/pods/3de2167a-7663-4b00-9743-229ff1864cb8/volumes" Oct 14 07:23:53 crc kubenswrapper[4870]: I1014 07:23:53.045431 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="416d7b39-4541-44a8-a55f-924bc86fee32" path="/var/lib/kubelet/pods/416d7b39-4541-44a8-a55f-924bc86fee32/volumes" Oct 14 07:23:53 crc kubenswrapper[4870]: I1014 07:23:53.047850 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5212ba20-83c3-41ff-a3d9-5a315f009c1b" path="/var/lib/kubelet/pods/5212ba20-83c3-41ff-a3d9-5a315f009c1b/volumes" Oct 14 07:23:53 crc kubenswrapper[4870]: I1014 07:23:53.049123 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="524b0b2a-77a2-4c05-be39-c8b97666ab11" path="/var/lib/kubelet/pods/524b0b2a-77a2-4c05-be39-c8b97666ab11/volumes" Oct 14 07:23:53 crc kubenswrapper[4870]: I1014 07:23:53.050659 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="60de9d04-7c24-42a4-9d5e-74b611487b69" path="/var/lib/kubelet/pods/60de9d04-7c24-42a4-9d5e-74b611487b69/volumes" Oct 14 07:23:53 crc kubenswrapper[4870]: I1014 07:23:53.052400 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bc76b72-8679-4b33-8966-0ee70c1788e5" path="/var/lib/kubelet/pods/7bc76b72-8679-4b33-8966-0ee70c1788e5/volumes" Oct 14 07:23:53 crc kubenswrapper[4870]: I1014 07:23:53.053860 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f5b700ed-3919-458f-860e-5630b8e4b800" path="/var/lib/kubelet/pods/f5b700ed-3919-458f-860e-5630b8e4b800/volumes" Oct 14 07:23:53 crc kubenswrapper[4870]: I1014 07:23:53.208861 4870 generic.go:334] "Generic (PLEG): container finished" podID="c5bb3fde-cfaf-4d09-ad31-e23fd830eb91" containerID="94a43ed16da62727b0d42c318ec266880c3131b81bd05fdd1dfddbb2a180533e" exitCode=0 Oct 14 07:23:53 crc kubenswrapper[4870]: I1014 07:23:53.208911 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-85f775bc47-d95qn" event={"ID":"c5bb3fde-cfaf-4d09-ad31-e23fd830eb91","Type":"ContainerDied","Data":"94a43ed16da62727b0d42c318ec266880c3131b81bd05fdd1dfddbb2a180533e"} Oct 14 07:23:53 crc kubenswrapper[4870]: I1014 07:23:53.211138 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 14 07:23:53 crc kubenswrapper[4870]: I1014 07:23:53.211289 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"da8a9731-f917-462f-9932-b37b6abb9a64","Type":"ContainerDied","Data":"b2190e6313a8dbb2a9d5a3f14b18f0e56fba0be535cabc899a22fee9bf2d18bf"} Oct 14 07:23:53 crc kubenswrapper[4870]: I1014 07:23:53.211324 4870 scope.go:117] "RemoveContainer" containerID="bc4b5271d735615160d1289dff31839e1ba9c4495e333ee1f50cab05851fd959" Oct 14 07:23:53 crc kubenswrapper[4870]: I1014 07:23:53.219595 4870 generic.go:334] "Generic (PLEG): container finished" podID="0a2e4c80-d44f-40c0-9c04-74e9ce98ae36" containerID="ea29b92302ab943ca5bc51b26bb9eba8dd8e4aa271ebf5c213d4ae71f3523daf" exitCode=0 Oct 14 07:23:53 crc kubenswrapper[4870]: I1014 07:23:53.219683 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-57d8bd87b6-vnpct" event={"ID":"0a2e4c80-d44f-40c0-9c04-74e9ce98ae36","Type":"ContainerDied","Data":"ea29b92302ab943ca5bc51b26bb9eba8dd8e4aa271ebf5c213d4ae71f3523daf"} Oct 14 07:23:53 crc kubenswrapper[4870]: I1014 07:23:53.239684 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 14 07:23:53 crc kubenswrapper[4870]: I1014 07:23:53.246389 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 14 07:23:53 crc kubenswrapper[4870]: I1014 07:23:53.255718 4870 scope.go:117] "RemoveContainer" containerID="fdc546ad7b75936b8391d1c25cb200fd0e53c0d3629443c510ae84c5a357cc6e" Oct 14 07:23:53 crc kubenswrapper[4870]: I1014 07:23:53.620919 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-85f775bc47-d95qn" Oct 14 07:23:53 crc kubenswrapper[4870]: I1014 07:23:53.670300 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-57d8bd87b6-vnpct" Oct 14 07:23:53 crc kubenswrapper[4870]: I1014 07:23:53.784877 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a2e4c80-d44f-40c0-9c04-74e9ce98ae36-combined-ca-bundle\") pod \"0a2e4c80-d44f-40c0-9c04-74e9ce98ae36\" (UID: \"0a2e4c80-d44f-40c0-9c04-74e9ce98ae36\") " Oct 14 07:23:53 crc kubenswrapper[4870]: I1014 07:23:53.785034 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c5bb3fde-cfaf-4d09-ad31-e23fd830eb91-config-data\") pod \"c5bb3fde-cfaf-4d09-ad31-e23fd830eb91\" (UID: \"c5bb3fde-cfaf-4d09-ad31-e23fd830eb91\") " Oct 14 07:23:53 crc kubenswrapper[4870]: I1014 07:23:53.785067 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a2e4c80-d44f-40c0-9c04-74e9ce98ae36-config-data\") pod \"0a2e4c80-d44f-40c0-9c04-74e9ce98ae36\" (UID: \"0a2e4c80-d44f-40c0-9c04-74e9ce98ae36\") " Oct 14 07:23:53 crc kubenswrapper[4870]: I1014 07:23:53.785104 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5bb3fde-cfaf-4d09-ad31-e23fd830eb91-combined-ca-bundle\") pod \"c5bb3fde-cfaf-4d09-ad31-e23fd830eb91\" (UID: \"c5bb3fde-cfaf-4d09-ad31-e23fd830eb91\") " Oct 14 07:23:53 crc kubenswrapper[4870]: I1014 07:23:53.785207 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0a2e4c80-d44f-40c0-9c04-74e9ce98ae36-config-data-custom\") pod \"0a2e4c80-d44f-40c0-9c04-74e9ce98ae36\" (UID: \"0a2e4c80-d44f-40c0-9c04-74e9ce98ae36\") " Oct 14 07:23:53 crc kubenswrapper[4870]: I1014 07:23:53.785946 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zxtlp\" (UniqueName: \"kubernetes.io/projected/c5bb3fde-cfaf-4d09-ad31-e23fd830eb91-kube-api-access-zxtlp\") pod \"c5bb3fde-cfaf-4d09-ad31-e23fd830eb91\" (UID: \"c5bb3fde-cfaf-4d09-ad31-e23fd830eb91\") " Oct 14 07:23:53 crc kubenswrapper[4870]: I1014 07:23:53.786013 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c5bb3fde-cfaf-4d09-ad31-e23fd830eb91-logs\") pod \"c5bb3fde-cfaf-4d09-ad31-e23fd830eb91\" (UID: \"c5bb3fde-cfaf-4d09-ad31-e23fd830eb91\") " Oct 14 07:23:53 crc kubenswrapper[4870]: I1014 07:23:53.786064 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c5bb3fde-cfaf-4d09-ad31-e23fd830eb91-config-data-custom\") pod \"c5bb3fde-cfaf-4d09-ad31-e23fd830eb91\" (UID: \"c5bb3fde-cfaf-4d09-ad31-e23fd830eb91\") " Oct 14 07:23:53 crc kubenswrapper[4870]: I1014 07:23:53.786099 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hbl6f\" (UniqueName: \"kubernetes.io/projected/0a2e4c80-d44f-40c0-9c04-74e9ce98ae36-kube-api-access-hbl6f\") pod \"0a2e4c80-d44f-40c0-9c04-74e9ce98ae36\" (UID: \"0a2e4c80-d44f-40c0-9c04-74e9ce98ae36\") " Oct 14 07:23:53 crc kubenswrapper[4870]: I1014 07:23:53.786258 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0a2e4c80-d44f-40c0-9c04-74e9ce98ae36-logs\") pod \"0a2e4c80-d44f-40c0-9c04-74e9ce98ae36\" (UID: \"0a2e4c80-d44f-40c0-9c04-74e9ce98ae36\") " Oct 14 07:23:53 crc kubenswrapper[4870]: I1014 07:23:53.786500 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c5bb3fde-cfaf-4d09-ad31-e23fd830eb91-logs" (OuterVolumeSpecName: "logs") pod "c5bb3fde-cfaf-4d09-ad31-e23fd830eb91" (UID: "c5bb3fde-cfaf-4d09-ad31-e23fd830eb91"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:23:53 crc kubenswrapper[4870]: I1014 07:23:53.787314 4870 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c5bb3fde-cfaf-4d09-ad31-e23fd830eb91-logs\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:53 crc kubenswrapper[4870]: I1014 07:23:53.788520 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0a2e4c80-d44f-40c0-9c04-74e9ce98ae36-logs" (OuterVolumeSpecName: "logs") pod "0a2e4c80-d44f-40c0-9c04-74e9ce98ae36" (UID: "0a2e4c80-d44f-40c0-9c04-74e9ce98ae36"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:23:53 crc kubenswrapper[4870]: I1014 07:23:53.789347 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a2e4c80-d44f-40c0-9c04-74e9ce98ae36-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "0a2e4c80-d44f-40c0-9c04-74e9ce98ae36" (UID: "0a2e4c80-d44f-40c0-9c04-74e9ce98ae36"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:23:53 crc kubenswrapper[4870]: I1014 07:23:53.792306 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0a2e4c80-d44f-40c0-9c04-74e9ce98ae36-kube-api-access-hbl6f" (OuterVolumeSpecName: "kube-api-access-hbl6f") pod "0a2e4c80-d44f-40c0-9c04-74e9ce98ae36" (UID: "0a2e4c80-d44f-40c0-9c04-74e9ce98ae36"). InnerVolumeSpecName "kube-api-access-hbl6f". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:23:53 crc kubenswrapper[4870]: I1014 07:23:53.792528 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c5bb3fde-cfaf-4d09-ad31-e23fd830eb91-kube-api-access-zxtlp" (OuterVolumeSpecName: "kube-api-access-zxtlp") pod "c5bb3fde-cfaf-4d09-ad31-e23fd830eb91" (UID: "c5bb3fde-cfaf-4d09-ad31-e23fd830eb91"). InnerVolumeSpecName "kube-api-access-zxtlp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:23:53 crc kubenswrapper[4870]: I1014 07:23:53.793869 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c5bb3fde-cfaf-4d09-ad31-e23fd830eb91-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "c5bb3fde-cfaf-4d09-ad31-e23fd830eb91" (UID: "c5bb3fde-cfaf-4d09-ad31-e23fd830eb91"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:23:53 crc kubenswrapper[4870]: I1014 07:23:53.805508 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a2e4c80-d44f-40c0-9c04-74e9ce98ae36-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0a2e4c80-d44f-40c0-9c04-74e9ce98ae36" (UID: "0a2e4c80-d44f-40c0-9c04-74e9ce98ae36"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:23:53 crc kubenswrapper[4870]: I1014 07:23:53.825094 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c5bb3fde-cfaf-4d09-ad31-e23fd830eb91-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c5bb3fde-cfaf-4d09-ad31-e23fd830eb91" (UID: "c5bb3fde-cfaf-4d09-ad31-e23fd830eb91"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:23:53 crc kubenswrapper[4870]: I1014 07:23:53.828664 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a2e4c80-d44f-40c0-9c04-74e9ce98ae36-config-data" (OuterVolumeSpecName: "config-data") pod "0a2e4c80-d44f-40c0-9c04-74e9ce98ae36" (UID: "0a2e4c80-d44f-40c0-9c04-74e9ce98ae36"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:23:53 crc kubenswrapper[4870]: I1014 07:23:53.843675 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c5bb3fde-cfaf-4d09-ad31-e23fd830eb91-config-data" (OuterVolumeSpecName: "config-data") pod "c5bb3fde-cfaf-4d09-ad31-e23fd830eb91" (UID: "c5bb3fde-cfaf-4d09-ad31-e23fd830eb91"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:23:53 crc kubenswrapper[4870]: I1014 07:23:53.888786 4870 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a2e4c80-d44f-40c0-9c04-74e9ce98ae36-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:53 crc kubenswrapper[4870]: I1014 07:23:53.888822 4870 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c5bb3fde-cfaf-4d09-ad31-e23fd830eb91-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:53 crc kubenswrapper[4870]: I1014 07:23:53.888831 4870 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a2e4c80-d44f-40c0-9c04-74e9ce98ae36-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:53 crc kubenswrapper[4870]: I1014 07:23:53.888840 4870 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5bb3fde-cfaf-4d09-ad31-e23fd830eb91-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:53 crc kubenswrapper[4870]: I1014 07:23:53.888849 4870 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0a2e4c80-d44f-40c0-9c04-74e9ce98ae36-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:53 crc kubenswrapper[4870]: I1014 07:23:53.888858 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zxtlp\" (UniqueName: \"kubernetes.io/projected/c5bb3fde-cfaf-4d09-ad31-e23fd830eb91-kube-api-access-zxtlp\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:53 crc kubenswrapper[4870]: I1014 07:23:53.888868 4870 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c5bb3fde-cfaf-4d09-ad31-e23fd830eb91-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:53 crc kubenswrapper[4870]: I1014 07:23:53.888876 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hbl6f\" (UniqueName: \"kubernetes.io/projected/0a2e4c80-d44f-40c0-9c04-74e9ce98ae36-kube-api-access-hbl6f\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:53 crc kubenswrapper[4870]: I1014 07:23:53.888884 4870 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0a2e4c80-d44f-40c0-9c04-74e9ce98ae36-logs\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:54 crc kubenswrapper[4870]: I1014 07:23:54.233772 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-57d8bd87b6-vnpct" event={"ID":"0a2e4c80-d44f-40c0-9c04-74e9ce98ae36","Type":"ContainerDied","Data":"8fe0611de8a26b46ab89d8cd485aaf180738d54c520c20fdca54ea25c93e6929"} Oct 14 07:23:54 crc kubenswrapper[4870]: I1014 07:23:54.233830 4870 scope.go:117] "RemoveContainer" containerID="ea29b92302ab943ca5bc51b26bb9eba8dd8e4aa271ebf5c213d4ae71f3523daf" Oct 14 07:23:54 crc kubenswrapper[4870]: I1014 07:23:54.233970 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-57d8bd87b6-vnpct" Oct 14 07:23:54 crc kubenswrapper[4870]: I1014 07:23:54.239005 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-85f775bc47-d95qn" event={"ID":"c5bb3fde-cfaf-4d09-ad31-e23fd830eb91","Type":"ContainerDied","Data":"f01c5670c5028a8cacb48b675dfcae6eaa4e1828d52432656e0cc9a9b15bd160"} Oct 14 07:23:54 crc kubenswrapper[4870]: I1014 07:23:54.239179 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-85f775bc47-d95qn" Oct 14 07:23:54 crc kubenswrapper[4870]: I1014 07:23:54.272179 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-keystone-listener-57d8bd87b6-vnpct"] Oct 14 07:23:54 crc kubenswrapper[4870]: I1014 07:23:54.277736 4870 scope.go:117] "RemoveContainer" containerID="bc61afdd96aedc563124f07cd981291ea8c01a0a034c78fba2e57e484564a0ee" Oct 14 07:23:54 crc kubenswrapper[4870]: I1014 07:23:54.281205 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-keystone-listener-57d8bd87b6-vnpct"] Oct 14 07:23:54 crc kubenswrapper[4870]: I1014 07:23:54.298905 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-worker-85f775bc47-d95qn"] Oct 14 07:23:54 crc kubenswrapper[4870]: I1014 07:23:54.306355 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-worker-85f775bc47-d95qn"] Oct 14 07:23:54 crc kubenswrapper[4870]: I1014 07:23:54.314189 4870 scope.go:117] "RemoveContainer" containerID="94a43ed16da62727b0d42c318ec266880c3131b81bd05fdd1dfddbb2a180533e" Oct 14 07:23:54 crc kubenswrapper[4870]: I1014 07:23:54.337617 4870 scope.go:117] "RemoveContainer" containerID="a77bf6d6eb57cf53a83d8f0ff81a500083d29df6ab1d56a72558d63fae3b342f" Oct 14 07:23:55 crc kubenswrapper[4870]: I1014 07:23:55.060920 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0a2e4c80-d44f-40c0-9c04-74e9ce98ae36" path="/var/lib/kubelet/pods/0a2e4c80-d44f-40c0-9c04-74e9ce98ae36/volumes" Oct 14 07:23:55 crc kubenswrapper[4870]: I1014 07:23:55.062266 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c5bb3fde-cfaf-4d09-ad31-e23fd830eb91" path="/var/lib/kubelet/pods/c5bb3fde-cfaf-4d09-ad31-e23fd830eb91/volumes" Oct 14 07:23:55 crc kubenswrapper[4870]: I1014 07:23:55.063071 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="da8a9731-f917-462f-9932-b37b6abb9a64" path="/var/lib/kubelet/pods/da8a9731-f917-462f-9932-b37b6abb9a64/volumes" Oct 14 07:23:56 crc kubenswrapper[4870]: I1014 07:23:56.272421 4870 generic.go:334] "Generic (PLEG): container finished" podID="0ba0e2ed-57d0-4535-883e-10b9b985b3d3" containerID="b6b74fcdf90322787a4403ce2a20ab64b512335fa185f5d3caa8a7cd3a4837f4" exitCode=0 Oct 14 07:23:56 crc kubenswrapper[4870]: I1014 07:23:56.272649 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-546b769ccc-fdhsk" event={"ID":"0ba0e2ed-57d0-4535-883e-10b9b985b3d3","Type":"ContainerDied","Data":"b6b74fcdf90322787a4403ce2a20ab64b512335fa185f5d3caa8a7cd3a4837f4"} Oct 14 07:23:56 crc kubenswrapper[4870]: I1014 07:23:56.325312 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-546b769ccc-fdhsk" Oct 14 07:23:56 crc kubenswrapper[4870]: I1014 07:23:56.425423 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-szj76\" (UniqueName: \"kubernetes.io/projected/0ba0e2ed-57d0-4535-883e-10b9b985b3d3-kube-api-access-szj76\") pod \"0ba0e2ed-57d0-4535-883e-10b9b985b3d3\" (UID: \"0ba0e2ed-57d0-4535-883e-10b9b985b3d3\") " Oct 14 07:23:56 crc kubenswrapper[4870]: I1014 07:23:56.425803 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/0ba0e2ed-57d0-4535-883e-10b9b985b3d3-config\") pod \"0ba0e2ed-57d0-4535-883e-10b9b985b3d3\" (UID: \"0ba0e2ed-57d0-4535-883e-10b9b985b3d3\") " Oct 14 07:23:56 crc kubenswrapper[4870]: I1014 07:23:56.426484 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ba0e2ed-57d0-4535-883e-10b9b985b3d3-combined-ca-bundle\") pod \"0ba0e2ed-57d0-4535-883e-10b9b985b3d3\" (UID: \"0ba0e2ed-57d0-4535-883e-10b9b985b3d3\") " Oct 14 07:23:56 crc kubenswrapper[4870]: I1014 07:23:56.426620 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/0ba0e2ed-57d0-4535-883e-10b9b985b3d3-ovndb-tls-certs\") pod \"0ba0e2ed-57d0-4535-883e-10b9b985b3d3\" (UID: \"0ba0e2ed-57d0-4535-883e-10b9b985b3d3\") " Oct 14 07:23:56 crc kubenswrapper[4870]: I1014 07:23:56.426827 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0ba0e2ed-57d0-4535-883e-10b9b985b3d3-internal-tls-certs\") pod \"0ba0e2ed-57d0-4535-883e-10b9b985b3d3\" (UID: \"0ba0e2ed-57d0-4535-883e-10b9b985b3d3\") " Oct 14 07:23:56 crc kubenswrapper[4870]: I1014 07:23:56.427018 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0ba0e2ed-57d0-4535-883e-10b9b985b3d3-public-tls-certs\") pod \"0ba0e2ed-57d0-4535-883e-10b9b985b3d3\" (UID: \"0ba0e2ed-57d0-4535-883e-10b9b985b3d3\") " Oct 14 07:23:56 crc kubenswrapper[4870]: I1014 07:23:56.427205 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/0ba0e2ed-57d0-4535-883e-10b9b985b3d3-httpd-config\") pod \"0ba0e2ed-57d0-4535-883e-10b9b985b3d3\" (UID: \"0ba0e2ed-57d0-4535-883e-10b9b985b3d3\") " Oct 14 07:23:56 crc kubenswrapper[4870]: I1014 07:23:56.443782 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0ba0e2ed-57d0-4535-883e-10b9b985b3d3-kube-api-access-szj76" (OuterVolumeSpecName: "kube-api-access-szj76") pod "0ba0e2ed-57d0-4535-883e-10b9b985b3d3" (UID: "0ba0e2ed-57d0-4535-883e-10b9b985b3d3"). InnerVolumeSpecName "kube-api-access-szj76". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:23:56 crc kubenswrapper[4870]: I1014 07:23:56.446913 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0ba0e2ed-57d0-4535-883e-10b9b985b3d3-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "0ba0e2ed-57d0-4535-883e-10b9b985b3d3" (UID: "0ba0e2ed-57d0-4535-883e-10b9b985b3d3"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:23:56 crc kubenswrapper[4870]: I1014 07:23:56.470853 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0ba0e2ed-57d0-4535-883e-10b9b985b3d3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0ba0e2ed-57d0-4535-883e-10b9b985b3d3" (UID: "0ba0e2ed-57d0-4535-883e-10b9b985b3d3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:23:56 crc kubenswrapper[4870]: I1014 07:23:56.474257 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0ba0e2ed-57d0-4535-883e-10b9b985b3d3-config" (OuterVolumeSpecName: "config") pod "0ba0e2ed-57d0-4535-883e-10b9b985b3d3" (UID: "0ba0e2ed-57d0-4535-883e-10b9b985b3d3"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:23:56 crc kubenswrapper[4870]: I1014 07:23:56.477273 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0ba0e2ed-57d0-4535-883e-10b9b985b3d3-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "0ba0e2ed-57d0-4535-883e-10b9b985b3d3" (UID: "0ba0e2ed-57d0-4535-883e-10b9b985b3d3"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:23:56 crc kubenswrapper[4870]: I1014 07:23:56.486160 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0ba0e2ed-57d0-4535-883e-10b9b985b3d3-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "0ba0e2ed-57d0-4535-883e-10b9b985b3d3" (UID: "0ba0e2ed-57d0-4535-883e-10b9b985b3d3"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:23:56 crc kubenswrapper[4870]: I1014 07:23:56.506275 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0ba0e2ed-57d0-4535-883e-10b9b985b3d3-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "0ba0e2ed-57d0-4535-883e-10b9b985b3d3" (UID: "0ba0e2ed-57d0-4535-883e-10b9b985b3d3"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:23:56 crc kubenswrapper[4870]: I1014 07:23:56.528919 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-szj76\" (UniqueName: \"kubernetes.io/projected/0ba0e2ed-57d0-4535-883e-10b9b985b3d3-kube-api-access-szj76\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:56 crc kubenswrapper[4870]: I1014 07:23:56.528954 4870 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/0ba0e2ed-57d0-4535-883e-10b9b985b3d3-config\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:56 crc kubenswrapper[4870]: I1014 07:23:56.528968 4870 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ba0e2ed-57d0-4535-883e-10b9b985b3d3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:56 crc kubenswrapper[4870]: I1014 07:23:56.528978 4870 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/0ba0e2ed-57d0-4535-883e-10b9b985b3d3-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:56 crc kubenswrapper[4870]: I1014 07:23:56.528987 4870 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0ba0e2ed-57d0-4535-883e-10b9b985b3d3-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:56 crc kubenswrapper[4870]: I1014 07:23:56.528999 4870 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0ba0e2ed-57d0-4535-883e-10b9b985b3d3-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:56 crc kubenswrapper[4870]: I1014 07:23:56.529009 4870 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/0ba0e2ed-57d0-4535-883e-10b9b985b3d3-httpd-config\") on node \"crc\" DevicePath \"\"" Oct 14 07:23:56 crc kubenswrapper[4870]: E1014 07:23:56.637123 4870 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 6f14317c64b4cb248d4254d308d86fadbdb28d2daee87abbf75414c219c2c477 is running failed: container process not found" containerID="6f14317c64b4cb248d4254d308d86fadbdb28d2daee87abbf75414c219c2c477" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 14 07:23:56 crc kubenswrapper[4870]: E1014 07:23:56.637597 4870 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 6f14317c64b4cb248d4254d308d86fadbdb28d2daee87abbf75414c219c2c477 is running failed: container process not found" containerID="6f14317c64b4cb248d4254d308d86fadbdb28d2daee87abbf75414c219c2c477" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 14 07:23:56 crc kubenswrapper[4870]: E1014 07:23:56.637898 4870 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 6f14317c64b4cb248d4254d308d86fadbdb28d2daee87abbf75414c219c2c477 is running failed: container process not found" containerID="6f14317c64b4cb248d4254d308d86fadbdb28d2daee87abbf75414c219c2c477" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 14 07:23:56 crc kubenswrapper[4870]: E1014 07:23:56.637984 4870 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 6f14317c64b4cb248d4254d308d86fadbdb28d2daee87abbf75414c219c2c477 is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-6482n" podUID="91a2e1de-104f-4e14-9192-c6587ed13be1" containerName="ovsdb-server" Oct 14 07:23:56 crc kubenswrapper[4870]: E1014 07:23:56.638933 4870 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="72ead03af69bb2d536002fe16f85a88974968529ec05c0e9e78254a3f796b690" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 14 07:23:56 crc kubenswrapper[4870]: E1014 07:23:56.641016 4870 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="72ead03af69bb2d536002fe16f85a88974968529ec05c0e9e78254a3f796b690" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 14 07:23:56 crc kubenswrapper[4870]: E1014 07:23:56.642264 4870 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="72ead03af69bb2d536002fe16f85a88974968529ec05c0e9e78254a3f796b690" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 14 07:23:56 crc kubenswrapper[4870]: E1014 07:23:56.642294 4870 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-6482n" podUID="91a2e1de-104f-4e14-9192-c6587ed13be1" containerName="ovs-vswitchd" Oct 14 07:23:57 crc kubenswrapper[4870]: I1014 07:23:57.285824 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-546b769ccc-fdhsk" event={"ID":"0ba0e2ed-57d0-4535-883e-10b9b985b3d3","Type":"ContainerDied","Data":"1c553e152dd7faad019d3848c24b46d39b8e66a6593592e2f0e2fb4653ee3b84"} Oct 14 07:23:57 crc kubenswrapper[4870]: I1014 07:23:57.285881 4870 scope.go:117] "RemoveContainer" containerID="f84e5ed457b1301ae67b84003d43c8e27a59446623074e29e6e2859280f620fd" Oct 14 07:23:57 crc kubenswrapper[4870]: I1014 07:23:57.286001 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-546b769ccc-fdhsk" Oct 14 07:23:57 crc kubenswrapper[4870]: I1014 07:23:57.312197 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-546b769ccc-fdhsk"] Oct 14 07:23:57 crc kubenswrapper[4870]: I1014 07:23:57.316847 4870 scope.go:117] "RemoveContainer" containerID="b6b74fcdf90322787a4403ce2a20ab64b512335fa185f5d3caa8a7cd3a4837f4" Oct 14 07:23:57 crc kubenswrapper[4870]: I1014 07:23:57.320078 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-546b769ccc-fdhsk"] Oct 14 07:23:59 crc kubenswrapper[4870]: I1014 07:23:59.049010 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0ba0e2ed-57d0-4535-883e-10b9b985b3d3" path="/var/lib/kubelet/pods/0ba0e2ed-57d0-4535-883e-10b9b985b3d3/volumes" Oct 14 07:24:01 crc kubenswrapper[4870]: E1014 07:24:01.637057 4870 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 6f14317c64b4cb248d4254d308d86fadbdb28d2daee87abbf75414c219c2c477 is running failed: container process not found" containerID="6f14317c64b4cb248d4254d308d86fadbdb28d2daee87abbf75414c219c2c477" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 14 07:24:01 crc kubenswrapper[4870]: E1014 07:24:01.637767 4870 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 6f14317c64b4cb248d4254d308d86fadbdb28d2daee87abbf75414c219c2c477 is running failed: container process not found" containerID="6f14317c64b4cb248d4254d308d86fadbdb28d2daee87abbf75414c219c2c477" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 14 07:24:01 crc kubenswrapper[4870]: E1014 07:24:01.638107 4870 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 6f14317c64b4cb248d4254d308d86fadbdb28d2daee87abbf75414c219c2c477 is running failed: container process not found" containerID="6f14317c64b4cb248d4254d308d86fadbdb28d2daee87abbf75414c219c2c477" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 14 07:24:01 crc kubenswrapper[4870]: E1014 07:24:01.638209 4870 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 6f14317c64b4cb248d4254d308d86fadbdb28d2daee87abbf75414c219c2c477 is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-6482n" podUID="91a2e1de-104f-4e14-9192-c6587ed13be1" containerName="ovsdb-server" Oct 14 07:24:01 crc kubenswrapper[4870]: E1014 07:24:01.638484 4870 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="72ead03af69bb2d536002fe16f85a88974968529ec05c0e9e78254a3f796b690" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 14 07:24:01 crc kubenswrapper[4870]: E1014 07:24:01.640238 4870 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="72ead03af69bb2d536002fe16f85a88974968529ec05c0e9e78254a3f796b690" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 14 07:24:01 crc kubenswrapper[4870]: E1014 07:24:01.641699 4870 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="72ead03af69bb2d536002fe16f85a88974968529ec05c0e9e78254a3f796b690" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 14 07:24:01 crc kubenswrapper[4870]: E1014 07:24:01.641790 4870 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-6482n" podUID="91a2e1de-104f-4e14-9192-c6587ed13be1" containerName="ovs-vswitchd" Oct 14 07:24:06 crc kubenswrapper[4870]: E1014 07:24:06.638040 4870 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 6f14317c64b4cb248d4254d308d86fadbdb28d2daee87abbf75414c219c2c477 is running failed: container process not found" containerID="6f14317c64b4cb248d4254d308d86fadbdb28d2daee87abbf75414c219c2c477" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 14 07:24:06 crc kubenswrapper[4870]: E1014 07:24:06.638875 4870 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 6f14317c64b4cb248d4254d308d86fadbdb28d2daee87abbf75414c219c2c477 is running failed: container process not found" containerID="6f14317c64b4cb248d4254d308d86fadbdb28d2daee87abbf75414c219c2c477" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 14 07:24:06 crc kubenswrapper[4870]: E1014 07:24:06.639330 4870 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 6f14317c64b4cb248d4254d308d86fadbdb28d2daee87abbf75414c219c2c477 is running failed: container process not found" containerID="6f14317c64b4cb248d4254d308d86fadbdb28d2daee87abbf75414c219c2c477" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 14 07:24:06 crc kubenswrapper[4870]: E1014 07:24:06.639431 4870 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 6f14317c64b4cb248d4254d308d86fadbdb28d2daee87abbf75414c219c2c477 is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-6482n" podUID="91a2e1de-104f-4e14-9192-c6587ed13be1" containerName="ovsdb-server" Oct 14 07:24:06 crc kubenswrapper[4870]: E1014 07:24:06.639932 4870 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="72ead03af69bb2d536002fe16f85a88974968529ec05c0e9e78254a3f796b690" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 14 07:24:06 crc kubenswrapper[4870]: E1014 07:24:06.643230 4870 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="72ead03af69bb2d536002fe16f85a88974968529ec05c0e9e78254a3f796b690" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 14 07:24:06 crc kubenswrapper[4870]: E1014 07:24:06.645134 4870 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="72ead03af69bb2d536002fe16f85a88974968529ec05c0e9e78254a3f796b690" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 14 07:24:06 crc kubenswrapper[4870]: E1014 07:24:06.645196 4870 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-6482n" podUID="91a2e1de-104f-4e14-9192-c6587ed13be1" containerName="ovs-vswitchd" Oct 14 07:24:11 crc kubenswrapper[4870]: E1014 07:24:11.637904 4870 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 6f14317c64b4cb248d4254d308d86fadbdb28d2daee87abbf75414c219c2c477 is running failed: container process not found" containerID="6f14317c64b4cb248d4254d308d86fadbdb28d2daee87abbf75414c219c2c477" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 14 07:24:11 crc kubenswrapper[4870]: E1014 07:24:11.639061 4870 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 6f14317c64b4cb248d4254d308d86fadbdb28d2daee87abbf75414c219c2c477 is running failed: container process not found" containerID="6f14317c64b4cb248d4254d308d86fadbdb28d2daee87abbf75414c219c2c477" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 14 07:24:11 crc kubenswrapper[4870]: E1014 07:24:11.639547 4870 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 6f14317c64b4cb248d4254d308d86fadbdb28d2daee87abbf75414c219c2c477 is running failed: container process not found" containerID="6f14317c64b4cb248d4254d308d86fadbdb28d2daee87abbf75414c219c2c477" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 14 07:24:11 crc kubenswrapper[4870]: E1014 07:24:11.639622 4870 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 6f14317c64b4cb248d4254d308d86fadbdb28d2daee87abbf75414c219c2c477 is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-6482n" podUID="91a2e1de-104f-4e14-9192-c6587ed13be1" containerName="ovsdb-server" Oct 14 07:24:11 crc kubenswrapper[4870]: E1014 07:24:11.640114 4870 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="72ead03af69bb2d536002fe16f85a88974968529ec05c0e9e78254a3f796b690" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 14 07:24:11 crc kubenswrapper[4870]: E1014 07:24:11.642408 4870 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="72ead03af69bb2d536002fe16f85a88974968529ec05c0e9e78254a3f796b690" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 14 07:24:11 crc kubenswrapper[4870]: E1014 07:24:11.643992 4870 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="72ead03af69bb2d536002fe16f85a88974968529ec05c0e9e78254a3f796b690" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 14 07:24:11 crc kubenswrapper[4870]: E1014 07:24:11.644031 4870 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-6482n" podUID="91a2e1de-104f-4e14-9192-c6587ed13be1" containerName="ovs-vswitchd" Oct 14 07:24:14 crc kubenswrapper[4870]: I1014 07:24:14.145196 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Oct 14 07:24:14 crc kubenswrapper[4870]: I1014 07:24:14.173793 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/2bd1f41f-3f1a-4ca7-8789-429104ce2120-lock\") pod \"2bd1f41f-3f1a-4ca7-8789-429104ce2120\" (UID: \"2bd1f41f-3f1a-4ca7-8789-429104ce2120\") " Oct 14 07:24:14 crc kubenswrapper[4870]: I1014 07:24:14.173865 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swift\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"2bd1f41f-3f1a-4ca7-8789-429104ce2120\" (UID: \"2bd1f41f-3f1a-4ca7-8789-429104ce2120\") " Oct 14 07:24:14 crc kubenswrapper[4870]: I1014 07:24:14.174013 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/2bd1f41f-3f1a-4ca7-8789-429104ce2120-cache\") pod \"2bd1f41f-3f1a-4ca7-8789-429104ce2120\" (UID: \"2bd1f41f-3f1a-4ca7-8789-429104ce2120\") " Oct 14 07:24:14 crc kubenswrapper[4870]: I1014 07:24:14.174081 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gsdsq\" (UniqueName: \"kubernetes.io/projected/2bd1f41f-3f1a-4ca7-8789-429104ce2120-kube-api-access-gsdsq\") pod \"2bd1f41f-3f1a-4ca7-8789-429104ce2120\" (UID: \"2bd1f41f-3f1a-4ca7-8789-429104ce2120\") " Oct 14 07:24:14 crc kubenswrapper[4870]: I1014 07:24:14.174156 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/2bd1f41f-3f1a-4ca7-8789-429104ce2120-etc-swift\") pod \"2bd1f41f-3f1a-4ca7-8789-429104ce2120\" (UID: \"2bd1f41f-3f1a-4ca7-8789-429104ce2120\") " Oct 14 07:24:14 crc kubenswrapper[4870]: I1014 07:24:14.174610 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2bd1f41f-3f1a-4ca7-8789-429104ce2120-cache" (OuterVolumeSpecName: "cache") pod "2bd1f41f-3f1a-4ca7-8789-429104ce2120" (UID: "2bd1f41f-3f1a-4ca7-8789-429104ce2120"). InnerVolumeSpecName "cache". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:24:14 crc kubenswrapper[4870]: I1014 07:24:14.174613 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2bd1f41f-3f1a-4ca7-8789-429104ce2120-lock" (OuterVolumeSpecName: "lock") pod "2bd1f41f-3f1a-4ca7-8789-429104ce2120" (UID: "2bd1f41f-3f1a-4ca7-8789-429104ce2120"). InnerVolumeSpecName "lock". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:24:14 crc kubenswrapper[4870]: I1014 07:24:14.175579 4870 reconciler_common.go:293] "Volume detached for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/2bd1f41f-3f1a-4ca7-8789-429104ce2120-lock\") on node \"crc\" DevicePath \"\"" Oct 14 07:24:14 crc kubenswrapper[4870]: I1014 07:24:14.175614 4870 reconciler_common.go:293] "Volume detached for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/2bd1f41f-3f1a-4ca7-8789-429104ce2120-cache\") on node \"crc\" DevicePath \"\"" Oct 14 07:24:14 crc kubenswrapper[4870]: I1014 07:24:14.181028 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2bd1f41f-3f1a-4ca7-8789-429104ce2120-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "2bd1f41f-3f1a-4ca7-8789-429104ce2120" (UID: "2bd1f41f-3f1a-4ca7-8789-429104ce2120"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:24:14 crc kubenswrapper[4870]: I1014 07:24:14.181887 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage10-crc" (OuterVolumeSpecName: "swift") pod "2bd1f41f-3f1a-4ca7-8789-429104ce2120" (UID: "2bd1f41f-3f1a-4ca7-8789-429104ce2120"). InnerVolumeSpecName "local-storage10-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 14 07:24:14 crc kubenswrapper[4870]: I1014 07:24:14.182942 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2bd1f41f-3f1a-4ca7-8789-429104ce2120-kube-api-access-gsdsq" (OuterVolumeSpecName: "kube-api-access-gsdsq") pod "2bd1f41f-3f1a-4ca7-8789-429104ce2120" (UID: "2bd1f41f-3f1a-4ca7-8789-429104ce2120"). InnerVolumeSpecName "kube-api-access-gsdsq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:24:14 crc kubenswrapper[4870]: I1014 07:24:14.277358 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gsdsq\" (UniqueName: \"kubernetes.io/projected/2bd1f41f-3f1a-4ca7-8789-429104ce2120-kube-api-access-gsdsq\") on node \"crc\" DevicePath \"\"" Oct 14 07:24:14 crc kubenswrapper[4870]: I1014 07:24:14.277416 4870 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/2bd1f41f-3f1a-4ca7-8789-429104ce2120-etc-swift\") on node \"crc\" DevicePath \"\"" Oct 14 07:24:14 crc kubenswrapper[4870]: I1014 07:24:14.277495 4870 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" " Oct 14 07:24:14 crc kubenswrapper[4870]: I1014 07:24:14.297075 4870 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage10-crc" (UniqueName: "kubernetes.io/local-volume/local-storage10-crc") on node "crc" Oct 14 07:24:14 crc kubenswrapper[4870]: I1014 07:24:14.381749 4870 reconciler_common.go:293] "Volume detached for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" DevicePath \"\"" Oct 14 07:24:14 crc kubenswrapper[4870]: I1014 07:24:14.501241 4870 generic.go:334] "Generic (PLEG): container finished" podID="2bd1f41f-3f1a-4ca7-8789-429104ce2120" containerID="9104317bae5cf155b62976217e7c03cd64c95837b39e557bc2162f991c3e3920" exitCode=137 Oct 14 07:24:14 crc kubenswrapper[4870]: I1014 07:24:14.501342 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"2bd1f41f-3f1a-4ca7-8789-429104ce2120","Type":"ContainerDied","Data":"9104317bae5cf155b62976217e7c03cd64c95837b39e557bc2162f991c3e3920"} Oct 14 07:24:14 crc kubenswrapper[4870]: I1014 07:24:14.501388 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"2bd1f41f-3f1a-4ca7-8789-429104ce2120","Type":"ContainerDied","Data":"b3bced4776c0abee54d19d97669ffda4f01fe4106bd9452fcfe3165b4e59267e"} Oct 14 07:24:14 crc kubenswrapper[4870]: I1014 07:24:14.501418 4870 scope.go:117] "RemoveContainer" containerID="9104317bae5cf155b62976217e7c03cd64c95837b39e557bc2162f991c3e3920" Oct 14 07:24:14 crc kubenswrapper[4870]: I1014 07:24:14.501753 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Oct 14 07:24:14 crc kubenswrapper[4870]: I1014 07:24:14.507928 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-6482n_91a2e1de-104f-4e14-9192-c6587ed13be1/ovs-vswitchd/0.log" Oct 14 07:24:14 crc kubenswrapper[4870]: I1014 07:24:14.508924 4870 generic.go:334] "Generic (PLEG): container finished" podID="91a2e1de-104f-4e14-9192-c6587ed13be1" containerID="72ead03af69bb2d536002fe16f85a88974968529ec05c0e9e78254a3f796b690" exitCode=137 Oct 14 07:24:14 crc kubenswrapper[4870]: I1014 07:24:14.508980 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-6482n" event={"ID":"91a2e1de-104f-4e14-9192-c6587ed13be1","Type":"ContainerDied","Data":"72ead03af69bb2d536002fe16f85a88974968529ec05c0e9e78254a3f796b690"} Oct 14 07:24:14 crc kubenswrapper[4870]: I1014 07:24:14.543819 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-6482n_91a2e1de-104f-4e14-9192-c6587ed13be1/ovs-vswitchd/0.log" Oct 14 07:24:14 crc kubenswrapper[4870]: I1014 07:24:14.546931 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-6482n" Oct 14 07:24:14 crc kubenswrapper[4870]: I1014 07:24:14.553237 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-storage-0"] Oct 14 07:24:14 crc kubenswrapper[4870]: I1014 07:24:14.555761 4870 scope.go:117] "RemoveContainer" containerID="eaac2b7e339adcaf70b58432d22ea22656cd6aebcd6871bb19adeae4be5f56f0" Oct 14 07:24:14 crc kubenswrapper[4870]: I1014 07:24:14.567591 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-storage-0"] Oct 14 07:24:14 crc kubenswrapper[4870]: I1014 07:24:14.583635 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/91a2e1de-104f-4e14-9192-c6587ed13be1-var-lib\") pod \"91a2e1de-104f-4e14-9192-c6587ed13be1\" (UID: \"91a2e1de-104f-4e14-9192-c6587ed13be1\") " Oct 14 07:24:14 crc kubenswrapper[4870]: I1014 07:24:14.583767 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/91a2e1de-104f-4e14-9192-c6587ed13be1-scripts\") pod \"91a2e1de-104f-4e14-9192-c6587ed13be1\" (UID: \"91a2e1de-104f-4e14-9192-c6587ed13be1\") " Oct 14 07:24:14 crc kubenswrapper[4870]: I1014 07:24:14.583789 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/91a2e1de-104f-4e14-9192-c6587ed13be1-etc-ovs\") pod \"91a2e1de-104f-4e14-9192-c6587ed13be1\" (UID: \"91a2e1de-104f-4e14-9192-c6587ed13be1\") " Oct 14 07:24:14 crc kubenswrapper[4870]: I1014 07:24:14.583811 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ldpld\" (UniqueName: \"kubernetes.io/projected/91a2e1de-104f-4e14-9192-c6587ed13be1-kube-api-access-ldpld\") pod \"91a2e1de-104f-4e14-9192-c6587ed13be1\" (UID: \"91a2e1de-104f-4e14-9192-c6587ed13be1\") " Oct 14 07:24:14 crc kubenswrapper[4870]: I1014 07:24:14.583840 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/91a2e1de-104f-4e14-9192-c6587ed13be1-var-lib" (OuterVolumeSpecName: "var-lib") pod "91a2e1de-104f-4e14-9192-c6587ed13be1" (UID: "91a2e1de-104f-4e14-9192-c6587ed13be1"). InnerVolumeSpecName "var-lib". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 14 07:24:14 crc kubenswrapper[4870]: I1014 07:24:14.583858 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/91a2e1de-104f-4e14-9192-c6587ed13be1-var-run\") pod \"91a2e1de-104f-4e14-9192-c6587ed13be1\" (UID: \"91a2e1de-104f-4e14-9192-c6587ed13be1\") " Oct 14 07:24:14 crc kubenswrapper[4870]: I1014 07:24:14.583901 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/91a2e1de-104f-4e14-9192-c6587ed13be1-var-run" (OuterVolumeSpecName: "var-run") pod "91a2e1de-104f-4e14-9192-c6587ed13be1" (UID: "91a2e1de-104f-4e14-9192-c6587ed13be1"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 14 07:24:14 crc kubenswrapper[4870]: I1014 07:24:14.583955 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/91a2e1de-104f-4e14-9192-c6587ed13be1-var-log\") pod \"91a2e1de-104f-4e14-9192-c6587ed13be1\" (UID: \"91a2e1de-104f-4e14-9192-c6587ed13be1\") " Oct 14 07:24:14 crc kubenswrapper[4870]: I1014 07:24:14.584655 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/91a2e1de-104f-4e14-9192-c6587ed13be1-etc-ovs" (OuterVolumeSpecName: "etc-ovs") pod "91a2e1de-104f-4e14-9192-c6587ed13be1" (UID: "91a2e1de-104f-4e14-9192-c6587ed13be1"). InnerVolumeSpecName "etc-ovs". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 14 07:24:14 crc kubenswrapper[4870]: I1014 07:24:14.584698 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/91a2e1de-104f-4e14-9192-c6587ed13be1-var-log" (OuterVolumeSpecName: "var-log") pod "91a2e1de-104f-4e14-9192-c6587ed13be1" (UID: "91a2e1de-104f-4e14-9192-c6587ed13be1"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 14 07:24:14 crc kubenswrapper[4870]: I1014 07:24:14.584759 4870 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/91a2e1de-104f-4e14-9192-c6587ed13be1-var-run\") on node \"crc\" DevicePath \"\"" Oct 14 07:24:14 crc kubenswrapper[4870]: I1014 07:24:14.584781 4870 scope.go:117] "RemoveContainer" containerID="7c8fe31733cd77d0b1ee88ab091697a9851a822a29c50d287246ced8eca4d502" Oct 14 07:24:14 crc kubenswrapper[4870]: I1014 07:24:14.584786 4870 reconciler_common.go:293] "Volume detached for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/91a2e1de-104f-4e14-9192-c6587ed13be1-var-lib\") on node \"crc\" DevicePath \"\"" Oct 14 07:24:14 crc kubenswrapper[4870]: I1014 07:24:14.584915 4870 reconciler_common.go:293] "Volume detached for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/91a2e1de-104f-4e14-9192-c6587ed13be1-etc-ovs\") on node \"crc\" DevicePath \"\"" Oct 14 07:24:14 crc kubenswrapper[4870]: I1014 07:24:14.585070 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/91a2e1de-104f-4e14-9192-c6587ed13be1-scripts" (OuterVolumeSpecName: "scripts") pod "91a2e1de-104f-4e14-9192-c6587ed13be1" (UID: "91a2e1de-104f-4e14-9192-c6587ed13be1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:24:14 crc kubenswrapper[4870]: I1014 07:24:14.587980 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/91a2e1de-104f-4e14-9192-c6587ed13be1-kube-api-access-ldpld" (OuterVolumeSpecName: "kube-api-access-ldpld") pod "91a2e1de-104f-4e14-9192-c6587ed13be1" (UID: "91a2e1de-104f-4e14-9192-c6587ed13be1"). InnerVolumeSpecName "kube-api-access-ldpld". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:24:14 crc kubenswrapper[4870]: I1014 07:24:14.613134 4870 scope.go:117] "RemoveContainer" containerID="ad9dccd6af895adbf1e3c3fcd2eca40527602d0f3bfc6d6d7ea2730f4a267119" Oct 14 07:24:14 crc kubenswrapper[4870]: I1014 07:24:14.636503 4870 scope.go:117] "RemoveContainer" containerID="18b08dbc7de7f34cc2bbb65969ca2e9d3e2da4bddab756fdee2d7db4e53e9180" Oct 14 07:24:14 crc kubenswrapper[4870]: I1014 07:24:14.659426 4870 scope.go:117] "RemoveContainer" containerID="b51d4fe9abeae800a8b66afb49b1fafee55d1cdb96718d65b7d5b35f6fd7519b" Oct 14 07:24:14 crc kubenswrapper[4870]: I1014 07:24:14.685545 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ldpld\" (UniqueName: \"kubernetes.io/projected/91a2e1de-104f-4e14-9192-c6587ed13be1-kube-api-access-ldpld\") on node \"crc\" DevicePath \"\"" Oct 14 07:24:14 crc kubenswrapper[4870]: I1014 07:24:14.685579 4870 scope.go:117] "RemoveContainer" containerID="5330b69d8d67676f3cc5cd3efe7cfe42a11bfa0006bfe019a9c536b30e4c1f0d" Oct 14 07:24:14 crc kubenswrapper[4870]: I1014 07:24:14.685595 4870 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/91a2e1de-104f-4e14-9192-c6587ed13be1-var-log\") on node \"crc\" DevicePath \"\"" Oct 14 07:24:14 crc kubenswrapper[4870]: I1014 07:24:14.685768 4870 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/91a2e1de-104f-4e14-9192-c6587ed13be1-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 07:24:14 crc kubenswrapper[4870]: I1014 07:24:14.735292 4870 scope.go:117] "RemoveContainer" containerID="42954c6bfcca40f75f1c6f4bcad9b47bf39a17a66a9a6f9bbf941bbed4d28ebd" Oct 14 07:24:14 crc kubenswrapper[4870]: I1014 07:24:14.767207 4870 scope.go:117] "RemoveContainer" containerID="a2214a07cbc03e9d4c3657396572e211461dec75c50cb838dd53f64e5d4ae7de" Oct 14 07:24:14 crc kubenswrapper[4870]: I1014 07:24:14.795730 4870 scope.go:117] "RemoveContainer" containerID="418f1f077bbfc2949a0abda0f2ddd66da4131527026acec0d19a4c08dec4dc66" Oct 14 07:24:14 crc kubenswrapper[4870]: I1014 07:24:14.830604 4870 scope.go:117] "RemoveContainer" containerID="642a18f2173ab4124b7442763988bbe3a2c0fdd80e87caad5e48f172fd55efc7" Oct 14 07:24:14 crc kubenswrapper[4870]: I1014 07:24:14.863525 4870 scope.go:117] "RemoveContainer" containerID="177c2ea0ed77d9c56284c56d85f024d847d0abd99647a2f3eedfd9a880d3b512" Oct 14 07:24:14 crc kubenswrapper[4870]: I1014 07:24:14.892792 4870 scope.go:117] "RemoveContainer" containerID="e3fe8e57380100f6a293cf390d246d76bd9db2c91dc07edc65f2b23cfa5d4871" Oct 14 07:24:14 crc kubenswrapper[4870]: I1014 07:24:14.923199 4870 scope.go:117] "RemoveContainer" containerID="427d65d1b5cbeeeb2e36fcfaa93892d1227204dbfad3f98d9cee60708c1f3a0d" Oct 14 07:24:14 crc kubenswrapper[4870]: I1014 07:24:14.956231 4870 scope.go:117] "RemoveContainer" containerID="edba2f1cac0476dc717b946f00e8879272b5577916eb07fb08304de266852951" Oct 14 07:24:14 crc kubenswrapper[4870]: I1014 07:24:14.982703 4870 scope.go:117] "RemoveContainer" containerID="9104317bae5cf155b62976217e7c03cd64c95837b39e557bc2162f991c3e3920" Oct 14 07:24:14 crc kubenswrapper[4870]: E1014 07:24:14.983310 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9104317bae5cf155b62976217e7c03cd64c95837b39e557bc2162f991c3e3920\": container with ID starting with 9104317bae5cf155b62976217e7c03cd64c95837b39e557bc2162f991c3e3920 not found: ID does not exist" containerID="9104317bae5cf155b62976217e7c03cd64c95837b39e557bc2162f991c3e3920" Oct 14 07:24:14 crc kubenswrapper[4870]: I1014 07:24:14.983547 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9104317bae5cf155b62976217e7c03cd64c95837b39e557bc2162f991c3e3920"} err="failed to get container status \"9104317bae5cf155b62976217e7c03cd64c95837b39e557bc2162f991c3e3920\": rpc error: code = NotFound desc = could not find container \"9104317bae5cf155b62976217e7c03cd64c95837b39e557bc2162f991c3e3920\": container with ID starting with 9104317bae5cf155b62976217e7c03cd64c95837b39e557bc2162f991c3e3920 not found: ID does not exist" Oct 14 07:24:14 crc kubenswrapper[4870]: I1014 07:24:14.983737 4870 scope.go:117] "RemoveContainer" containerID="eaac2b7e339adcaf70b58432d22ea22656cd6aebcd6871bb19adeae4be5f56f0" Oct 14 07:24:14 crc kubenswrapper[4870]: E1014 07:24:14.984414 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eaac2b7e339adcaf70b58432d22ea22656cd6aebcd6871bb19adeae4be5f56f0\": container with ID starting with eaac2b7e339adcaf70b58432d22ea22656cd6aebcd6871bb19adeae4be5f56f0 not found: ID does not exist" containerID="eaac2b7e339adcaf70b58432d22ea22656cd6aebcd6871bb19adeae4be5f56f0" Oct 14 07:24:14 crc kubenswrapper[4870]: I1014 07:24:14.984525 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eaac2b7e339adcaf70b58432d22ea22656cd6aebcd6871bb19adeae4be5f56f0"} err="failed to get container status \"eaac2b7e339adcaf70b58432d22ea22656cd6aebcd6871bb19adeae4be5f56f0\": rpc error: code = NotFound desc = could not find container \"eaac2b7e339adcaf70b58432d22ea22656cd6aebcd6871bb19adeae4be5f56f0\": container with ID starting with eaac2b7e339adcaf70b58432d22ea22656cd6aebcd6871bb19adeae4be5f56f0 not found: ID does not exist" Oct 14 07:24:14 crc kubenswrapper[4870]: I1014 07:24:14.984575 4870 scope.go:117] "RemoveContainer" containerID="7c8fe31733cd77d0b1ee88ab091697a9851a822a29c50d287246ced8eca4d502" Oct 14 07:24:14 crc kubenswrapper[4870]: E1014 07:24:14.984971 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7c8fe31733cd77d0b1ee88ab091697a9851a822a29c50d287246ced8eca4d502\": container with ID starting with 7c8fe31733cd77d0b1ee88ab091697a9851a822a29c50d287246ced8eca4d502 not found: ID does not exist" containerID="7c8fe31733cd77d0b1ee88ab091697a9851a822a29c50d287246ced8eca4d502" Oct 14 07:24:14 crc kubenswrapper[4870]: I1014 07:24:14.985028 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7c8fe31733cd77d0b1ee88ab091697a9851a822a29c50d287246ced8eca4d502"} err="failed to get container status \"7c8fe31733cd77d0b1ee88ab091697a9851a822a29c50d287246ced8eca4d502\": rpc error: code = NotFound desc = could not find container \"7c8fe31733cd77d0b1ee88ab091697a9851a822a29c50d287246ced8eca4d502\": container with ID starting with 7c8fe31733cd77d0b1ee88ab091697a9851a822a29c50d287246ced8eca4d502 not found: ID does not exist" Oct 14 07:24:14 crc kubenswrapper[4870]: I1014 07:24:14.985065 4870 scope.go:117] "RemoveContainer" containerID="ad9dccd6af895adbf1e3c3fcd2eca40527602d0f3bfc6d6d7ea2730f4a267119" Oct 14 07:24:14 crc kubenswrapper[4870]: E1014 07:24:14.985894 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ad9dccd6af895adbf1e3c3fcd2eca40527602d0f3bfc6d6d7ea2730f4a267119\": container with ID starting with ad9dccd6af895adbf1e3c3fcd2eca40527602d0f3bfc6d6d7ea2730f4a267119 not found: ID does not exist" containerID="ad9dccd6af895adbf1e3c3fcd2eca40527602d0f3bfc6d6d7ea2730f4a267119" Oct 14 07:24:14 crc kubenswrapper[4870]: I1014 07:24:14.985938 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ad9dccd6af895adbf1e3c3fcd2eca40527602d0f3bfc6d6d7ea2730f4a267119"} err="failed to get container status \"ad9dccd6af895adbf1e3c3fcd2eca40527602d0f3bfc6d6d7ea2730f4a267119\": rpc error: code = NotFound desc = could not find container \"ad9dccd6af895adbf1e3c3fcd2eca40527602d0f3bfc6d6d7ea2730f4a267119\": container with ID starting with ad9dccd6af895adbf1e3c3fcd2eca40527602d0f3bfc6d6d7ea2730f4a267119 not found: ID does not exist" Oct 14 07:24:14 crc kubenswrapper[4870]: I1014 07:24:14.985965 4870 scope.go:117] "RemoveContainer" containerID="18b08dbc7de7f34cc2bbb65969ca2e9d3e2da4bddab756fdee2d7db4e53e9180" Oct 14 07:24:14 crc kubenswrapper[4870]: E1014 07:24:14.986390 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"18b08dbc7de7f34cc2bbb65969ca2e9d3e2da4bddab756fdee2d7db4e53e9180\": container with ID starting with 18b08dbc7de7f34cc2bbb65969ca2e9d3e2da4bddab756fdee2d7db4e53e9180 not found: ID does not exist" containerID="18b08dbc7de7f34cc2bbb65969ca2e9d3e2da4bddab756fdee2d7db4e53e9180" Oct 14 07:24:14 crc kubenswrapper[4870]: I1014 07:24:14.986523 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"18b08dbc7de7f34cc2bbb65969ca2e9d3e2da4bddab756fdee2d7db4e53e9180"} err="failed to get container status \"18b08dbc7de7f34cc2bbb65969ca2e9d3e2da4bddab756fdee2d7db4e53e9180\": rpc error: code = NotFound desc = could not find container \"18b08dbc7de7f34cc2bbb65969ca2e9d3e2da4bddab756fdee2d7db4e53e9180\": container with ID starting with 18b08dbc7de7f34cc2bbb65969ca2e9d3e2da4bddab756fdee2d7db4e53e9180 not found: ID does not exist" Oct 14 07:24:14 crc kubenswrapper[4870]: I1014 07:24:14.986573 4870 scope.go:117] "RemoveContainer" containerID="b51d4fe9abeae800a8b66afb49b1fafee55d1cdb96718d65b7d5b35f6fd7519b" Oct 14 07:24:14 crc kubenswrapper[4870]: E1014 07:24:14.986955 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b51d4fe9abeae800a8b66afb49b1fafee55d1cdb96718d65b7d5b35f6fd7519b\": container with ID starting with b51d4fe9abeae800a8b66afb49b1fafee55d1cdb96718d65b7d5b35f6fd7519b not found: ID does not exist" containerID="b51d4fe9abeae800a8b66afb49b1fafee55d1cdb96718d65b7d5b35f6fd7519b" Oct 14 07:24:14 crc kubenswrapper[4870]: I1014 07:24:14.987001 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b51d4fe9abeae800a8b66afb49b1fafee55d1cdb96718d65b7d5b35f6fd7519b"} err="failed to get container status \"b51d4fe9abeae800a8b66afb49b1fafee55d1cdb96718d65b7d5b35f6fd7519b\": rpc error: code = NotFound desc = could not find container \"b51d4fe9abeae800a8b66afb49b1fafee55d1cdb96718d65b7d5b35f6fd7519b\": container with ID starting with b51d4fe9abeae800a8b66afb49b1fafee55d1cdb96718d65b7d5b35f6fd7519b not found: ID does not exist" Oct 14 07:24:14 crc kubenswrapper[4870]: I1014 07:24:14.987039 4870 scope.go:117] "RemoveContainer" containerID="5330b69d8d67676f3cc5cd3efe7cfe42a11bfa0006bfe019a9c536b30e4c1f0d" Oct 14 07:24:14 crc kubenswrapper[4870]: E1014 07:24:14.987419 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5330b69d8d67676f3cc5cd3efe7cfe42a11bfa0006bfe019a9c536b30e4c1f0d\": container with ID starting with 5330b69d8d67676f3cc5cd3efe7cfe42a11bfa0006bfe019a9c536b30e4c1f0d not found: ID does not exist" containerID="5330b69d8d67676f3cc5cd3efe7cfe42a11bfa0006bfe019a9c536b30e4c1f0d" Oct 14 07:24:14 crc kubenswrapper[4870]: I1014 07:24:14.987503 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5330b69d8d67676f3cc5cd3efe7cfe42a11bfa0006bfe019a9c536b30e4c1f0d"} err="failed to get container status \"5330b69d8d67676f3cc5cd3efe7cfe42a11bfa0006bfe019a9c536b30e4c1f0d\": rpc error: code = NotFound desc = could not find container \"5330b69d8d67676f3cc5cd3efe7cfe42a11bfa0006bfe019a9c536b30e4c1f0d\": container with ID starting with 5330b69d8d67676f3cc5cd3efe7cfe42a11bfa0006bfe019a9c536b30e4c1f0d not found: ID does not exist" Oct 14 07:24:14 crc kubenswrapper[4870]: I1014 07:24:14.987534 4870 scope.go:117] "RemoveContainer" containerID="42954c6bfcca40f75f1c6f4bcad9b47bf39a17a66a9a6f9bbf941bbed4d28ebd" Oct 14 07:24:14 crc kubenswrapper[4870]: E1014 07:24:14.987917 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"42954c6bfcca40f75f1c6f4bcad9b47bf39a17a66a9a6f9bbf941bbed4d28ebd\": container with ID starting with 42954c6bfcca40f75f1c6f4bcad9b47bf39a17a66a9a6f9bbf941bbed4d28ebd not found: ID does not exist" containerID="42954c6bfcca40f75f1c6f4bcad9b47bf39a17a66a9a6f9bbf941bbed4d28ebd" Oct 14 07:24:14 crc kubenswrapper[4870]: I1014 07:24:14.987965 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"42954c6bfcca40f75f1c6f4bcad9b47bf39a17a66a9a6f9bbf941bbed4d28ebd"} err="failed to get container status \"42954c6bfcca40f75f1c6f4bcad9b47bf39a17a66a9a6f9bbf941bbed4d28ebd\": rpc error: code = NotFound desc = could not find container \"42954c6bfcca40f75f1c6f4bcad9b47bf39a17a66a9a6f9bbf941bbed4d28ebd\": container with ID starting with 42954c6bfcca40f75f1c6f4bcad9b47bf39a17a66a9a6f9bbf941bbed4d28ebd not found: ID does not exist" Oct 14 07:24:14 crc kubenswrapper[4870]: I1014 07:24:14.988054 4870 scope.go:117] "RemoveContainer" containerID="a2214a07cbc03e9d4c3657396572e211461dec75c50cb838dd53f64e5d4ae7de" Oct 14 07:24:14 crc kubenswrapper[4870]: E1014 07:24:14.988511 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a2214a07cbc03e9d4c3657396572e211461dec75c50cb838dd53f64e5d4ae7de\": container with ID starting with a2214a07cbc03e9d4c3657396572e211461dec75c50cb838dd53f64e5d4ae7de not found: ID does not exist" containerID="a2214a07cbc03e9d4c3657396572e211461dec75c50cb838dd53f64e5d4ae7de" Oct 14 07:24:14 crc kubenswrapper[4870]: I1014 07:24:14.988563 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a2214a07cbc03e9d4c3657396572e211461dec75c50cb838dd53f64e5d4ae7de"} err="failed to get container status \"a2214a07cbc03e9d4c3657396572e211461dec75c50cb838dd53f64e5d4ae7de\": rpc error: code = NotFound desc = could not find container \"a2214a07cbc03e9d4c3657396572e211461dec75c50cb838dd53f64e5d4ae7de\": container with ID starting with a2214a07cbc03e9d4c3657396572e211461dec75c50cb838dd53f64e5d4ae7de not found: ID does not exist" Oct 14 07:24:14 crc kubenswrapper[4870]: I1014 07:24:14.988593 4870 scope.go:117] "RemoveContainer" containerID="418f1f077bbfc2949a0abda0f2ddd66da4131527026acec0d19a4c08dec4dc66" Oct 14 07:24:14 crc kubenswrapper[4870]: E1014 07:24:14.989140 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"418f1f077bbfc2949a0abda0f2ddd66da4131527026acec0d19a4c08dec4dc66\": container with ID starting with 418f1f077bbfc2949a0abda0f2ddd66da4131527026acec0d19a4c08dec4dc66 not found: ID does not exist" containerID="418f1f077bbfc2949a0abda0f2ddd66da4131527026acec0d19a4c08dec4dc66" Oct 14 07:24:14 crc kubenswrapper[4870]: I1014 07:24:14.989191 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"418f1f077bbfc2949a0abda0f2ddd66da4131527026acec0d19a4c08dec4dc66"} err="failed to get container status \"418f1f077bbfc2949a0abda0f2ddd66da4131527026acec0d19a4c08dec4dc66\": rpc error: code = NotFound desc = could not find container \"418f1f077bbfc2949a0abda0f2ddd66da4131527026acec0d19a4c08dec4dc66\": container with ID starting with 418f1f077bbfc2949a0abda0f2ddd66da4131527026acec0d19a4c08dec4dc66 not found: ID does not exist" Oct 14 07:24:14 crc kubenswrapper[4870]: I1014 07:24:14.989227 4870 scope.go:117] "RemoveContainer" containerID="642a18f2173ab4124b7442763988bbe3a2c0fdd80e87caad5e48f172fd55efc7" Oct 14 07:24:14 crc kubenswrapper[4870]: E1014 07:24:14.989637 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"642a18f2173ab4124b7442763988bbe3a2c0fdd80e87caad5e48f172fd55efc7\": container with ID starting with 642a18f2173ab4124b7442763988bbe3a2c0fdd80e87caad5e48f172fd55efc7 not found: ID does not exist" containerID="642a18f2173ab4124b7442763988bbe3a2c0fdd80e87caad5e48f172fd55efc7" Oct 14 07:24:14 crc kubenswrapper[4870]: I1014 07:24:14.989688 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"642a18f2173ab4124b7442763988bbe3a2c0fdd80e87caad5e48f172fd55efc7"} err="failed to get container status \"642a18f2173ab4124b7442763988bbe3a2c0fdd80e87caad5e48f172fd55efc7\": rpc error: code = NotFound desc = could not find container \"642a18f2173ab4124b7442763988bbe3a2c0fdd80e87caad5e48f172fd55efc7\": container with ID starting with 642a18f2173ab4124b7442763988bbe3a2c0fdd80e87caad5e48f172fd55efc7 not found: ID does not exist" Oct 14 07:24:14 crc kubenswrapper[4870]: I1014 07:24:14.989721 4870 scope.go:117] "RemoveContainer" containerID="177c2ea0ed77d9c56284c56d85f024d847d0abd99647a2f3eedfd9a880d3b512" Oct 14 07:24:14 crc kubenswrapper[4870]: E1014 07:24:14.990097 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"177c2ea0ed77d9c56284c56d85f024d847d0abd99647a2f3eedfd9a880d3b512\": container with ID starting with 177c2ea0ed77d9c56284c56d85f024d847d0abd99647a2f3eedfd9a880d3b512 not found: ID does not exist" containerID="177c2ea0ed77d9c56284c56d85f024d847d0abd99647a2f3eedfd9a880d3b512" Oct 14 07:24:14 crc kubenswrapper[4870]: I1014 07:24:14.990144 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"177c2ea0ed77d9c56284c56d85f024d847d0abd99647a2f3eedfd9a880d3b512"} err="failed to get container status \"177c2ea0ed77d9c56284c56d85f024d847d0abd99647a2f3eedfd9a880d3b512\": rpc error: code = NotFound desc = could not find container \"177c2ea0ed77d9c56284c56d85f024d847d0abd99647a2f3eedfd9a880d3b512\": container with ID starting with 177c2ea0ed77d9c56284c56d85f024d847d0abd99647a2f3eedfd9a880d3b512 not found: ID does not exist" Oct 14 07:24:14 crc kubenswrapper[4870]: I1014 07:24:14.990173 4870 scope.go:117] "RemoveContainer" containerID="e3fe8e57380100f6a293cf390d246d76bd9db2c91dc07edc65f2b23cfa5d4871" Oct 14 07:24:14 crc kubenswrapper[4870]: E1014 07:24:14.990786 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e3fe8e57380100f6a293cf390d246d76bd9db2c91dc07edc65f2b23cfa5d4871\": container with ID starting with e3fe8e57380100f6a293cf390d246d76bd9db2c91dc07edc65f2b23cfa5d4871 not found: ID does not exist" containerID="e3fe8e57380100f6a293cf390d246d76bd9db2c91dc07edc65f2b23cfa5d4871" Oct 14 07:24:14 crc kubenswrapper[4870]: I1014 07:24:14.990833 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e3fe8e57380100f6a293cf390d246d76bd9db2c91dc07edc65f2b23cfa5d4871"} err="failed to get container status \"e3fe8e57380100f6a293cf390d246d76bd9db2c91dc07edc65f2b23cfa5d4871\": rpc error: code = NotFound desc = could not find container \"e3fe8e57380100f6a293cf390d246d76bd9db2c91dc07edc65f2b23cfa5d4871\": container with ID starting with e3fe8e57380100f6a293cf390d246d76bd9db2c91dc07edc65f2b23cfa5d4871 not found: ID does not exist" Oct 14 07:24:14 crc kubenswrapper[4870]: I1014 07:24:14.990863 4870 scope.go:117] "RemoveContainer" containerID="427d65d1b5cbeeeb2e36fcfaa93892d1227204dbfad3f98d9cee60708c1f3a0d" Oct 14 07:24:14 crc kubenswrapper[4870]: E1014 07:24:14.991208 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"427d65d1b5cbeeeb2e36fcfaa93892d1227204dbfad3f98d9cee60708c1f3a0d\": container with ID starting with 427d65d1b5cbeeeb2e36fcfaa93892d1227204dbfad3f98d9cee60708c1f3a0d not found: ID does not exist" containerID="427d65d1b5cbeeeb2e36fcfaa93892d1227204dbfad3f98d9cee60708c1f3a0d" Oct 14 07:24:14 crc kubenswrapper[4870]: I1014 07:24:14.991251 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"427d65d1b5cbeeeb2e36fcfaa93892d1227204dbfad3f98d9cee60708c1f3a0d"} err="failed to get container status \"427d65d1b5cbeeeb2e36fcfaa93892d1227204dbfad3f98d9cee60708c1f3a0d\": rpc error: code = NotFound desc = could not find container \"427d65d1b5cbeeeb2e36fcfaa93892d1227204dbfad3f98d9cee60708c1f3a0d\": container with ID starting with 427d65d1b5cbeeeb2e36fcfaa93892d1227204dbfad3f98d9cee60708c1f3a0d not found: ID does not exist" Oct 14 07:24:14 crc kubenswrapper[4870]: I1014 07:24:14.991280 4870 scope.go:117] "RemoveContainer" containerID="edba2f1cac0476dc717b946f00e8879272b5577916eb07fb08304de266852951" Oct 14 07:24:14 crc kubenswrapper[4870]: E1014 07:24:14.991862 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"edba2f1cac0476dc717b946f00e8879272b5577916eb07fb08304de266852951\": container with ID starting with edba2f1cac0476dc717b946f00e8879272b5577916eb07fb08304de266852951 not found: ID does not exist" containerID="edba2f1cac0476dc717b946f00e8879272b5577916eb07fb08304de266852951" Oct 14 07:24:14 crc kubenswrapper[4870]: I1014 07:24:14.991910 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"edba2f1cac0476dc717b946f00e8879272b5577916eb07fb08304de266852951"} err="failed to get container status \"edba2f1cac0476dc717b946f00e8879272b5577916eb07fb08304de266852951\": rpc error: code = NotFound desc = could not find container \"edba2f1cac0476dc717b946f00e8879272b5577916eb07fb08304de266852951\": container with ID starting with edba2f1cac0476dc717b946f00e8879272b5577916eb07fb08304de266852951 not found: ID does not exist" Oct 14 07:24:15 crc kubenswrapper[4870]: I1014 07:24:15.053321 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2bd1f41f-3f1a-4ca7-8789-429104ce2120" path="/var/lib/kubelet/pods/2bd1f41f-3f1a-4ca7-8789-429104ce2120/volumes" Oct 14 07:24:15 crc kubenswrapper[4870]: I1014 07:24:15.217904 4870 pod_container_manager_linux.go:210] "Failed to delete cgroup paths" cgroupName=["kubepods","besteffort","pod09f89569-6fe3-4b3f-9394-a990ad8cde30"] err="unable to destroy cgroup paths for cgroup [kubepods besteffort pod09f89569-6fe3-4b3f-9394-a990ad8cde30] : Timed out while waiting for systemd to remove kubepods-besteffort-pod09f89569_6fe3_4b3f_9394_a990ad8cde30.slice" Oct 14 07:24:15 crc kubenswrapper[4870]: E1014 07:24:15.218782 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to delete cgroup paths for [kubepods besteffort pod09f89569-6fe3-4b3f-9394-a990ad8cde30] : unable to destroy cgroup paths for cgroup [kubepods besteffort pod09f89569-6fe3-4b3f-9394-a990ad8cde30] : Timed out while waiting for systemd to remove kubepods-besteffort-pod09f89569_6fe3_4b3f_9394_a990ad8cde30.slice" pod="openstack/ovsdbserver-nb-0" podUID="09f89569-6fe3-4b3f-9394-a990ad8cde30" Oct 14 07:24:15 crc kubenswrapper[4870]: I1014 07:24:15.525262 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-6482n_91a2e1de-104f-4e14-9192-c6587ed13be1/ovs-vswitchd/0.log" Oct 14 07:24:15 crc kubenswrapper[4870]: I1014 07:24:15.526940 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Oct 14 07:24:15 crc kubenswrapper[4870]: I1014 07:24:15.526984 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-6482n" event={"ID":"91a2e1de-104f-4e14-9192-c6587ed13be1","Type":"ContainerDied","Data":"ca705e1a5041f27d32a37ea1dfaf6496c090f4c3a3d80f271e9fc688a3c8e78f"} Oct 14 07:24:15 crc kubenswrapper[4870]: I1014 07:24:15.527032 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-6482n" Oct 14 07:24:15 crc kubenswrapper[4870]: I1014 07:24:15.527062 4870 scope.go:117] "RemoveContainer" containerID="72ead03af69bb2d536002fe16f85a88974968529ec05c0e9e78254a3f796b690" Oct 14 07:24:15 crc kubenswrapper[4870]: I1014 07:24:15.573164 4870 scope.go:117] "RemoveContainer" containerID="6f14317c64b4cb248d4254d308d86fadbdb28d2daee87abbf75414c219c2c477" Oct 14 07:24:15 crc kubenswrapper[4870]: I1014 07:24:15.596060 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-ovs-6482n"] Oct 14 07:24:15 crc kubenswrapper[4870]: I1014 07:24:15.605917 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-ovs-6482n"] Oct 14 07:24:15 crc kubenswrapper[4870]: I1014 07:24:15.622504 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 14 07:24:15 crc kubenswrapper[4870]: I1014 07:24:15.623523 4870 scope.go:117] "RemoveContainer" containerID="31bee80630d9bb4bd377e738c7f57f0caeb9aafbcc1210717378d78bab78ce72" Oct 14 07:24:15 crc kubenswrapper[4870]: I1014 07:24:15.629165 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 14 07:24:17 crc kubenswrapper[4870]: I1014 07:24:17.051285 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09f89569-6fe3-4b3f-9394-a990ad8cde30" path="/var/lib/kubelet/pods/09f89569-6fe3-4b3f-9394-a990ad8cde30/volumes" Oct 14 07:24:17 crc kubenswrapper[4870]: I1014 07:24:17.052551 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="91a2e1de-104f-4e14-9192-c6587ed13be1" path="/var/lib/kubelet/pods/91a2e1de-104f-4e14-9192-c6587ed13be1/volumes" Oct 14 07:24:25 crc kubenswrapper[4870]: I1014 07:24:25.467528 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-qm9qp"] Oct 14 07:24:25 crc kubenswrapper[4870]: E1014 07:24:25.468592 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5212ba20-83c3-41ff-a3d9-5a315f009c1b" containerName="setup-container" Oct 14 07:24:25 crc kubenswrapper[4870]: I1014 07:24:25.468617 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="5212ba20-83c3-41ff-a3d9-5a315f009c1b" containerName="setup-container" Oct 14 07:24:25 crc kubenswrapper[4870]: E1014 07:24:25.468641 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="36717039-e3e2-4b74-b612-655556122574" containerName="ovsdbserver-sb" Oct 14 07:24:25 crc kubenswrapper[4870]: I1014 07:24:25.468651 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="36717039-e3e2-4b74-b612-655556122574" containerName="ovsdbserver-sb" Oct 14 07:24:25 crc kubenswrapper[4870]: E1014 07:24:25.468669 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09f89569-6fe3-4b3f-9394-a990ad8cde30" containerName="openstack-network-exporter" Oct 14 07:24:25 crc kubenswrapper[4870]: I1014 07:24:25.468680 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="09f89569-6fe3-4b3f-9394-a990ad8cde30" containerName="openstack-network-exporter" Oct 14 07:24:25 crc kubenswrapper[4870]: E1014 07:24:25.468693 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d74ef50-1861-4d70-b781-f92d5f4b35ba" containerName="galera" Oct 14 07:24:25 crc kubenswrapper[4870]: I1014 07:24:25.468703 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d74ef50-1861-4d70-b781-f92d5f4b35ba" containerName="galera" Oct 14 07:24:25 crc kubenswrapper[4870]: E1014 07:24:25.468726 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c5bb3fde-cfaf-4d09-ad31-e23fd830eb91" containerName="barbican-worker-log" Oct 14 07:24:25 crc kubenswrapper[4870]: I1014 07:24:25.468737 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="c5bb3fde-cfaf-4d09-ad31-e23fd830eb91" containerName="barbican-worker-log" Oct 14 07:24:25 crc kubenswrapper[4870]: E1014 07:24:25.468753 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="416d7b39-4541-44a8-a55f-924bc86fee32" containerName="ovn-northd" Oct 14 07:24:25 crc kubenswrapper[4870]: I1014 07:24:25.468762 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="416d7b39-4541-44a8-a55f-924bc86fee32" containerName="ovn-northd" Oct 14 07:24:25 crc kubenswrapper[4870]: E1014 07:24:25.468779 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09f89569-6fe3-4b3f-9394-a990ad8cde30" containerName="ovsdbserver-nb" Oct 14 07:24:25 crc kubenswrapper[4870]: I1014 07:24:25.468788 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="09f89569-6fe3-4b3f-9394-a990ad8cde30" containerName="ovsdbserver-nb" Oct 14 07:24:25 crc kubenswrapper[4870]: E1014 07:24:25.468802 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a2e4c80-d44f-40c0-9c04-74e9ce98ae36" containerName="barbican-keystone-listener-log" Oct 14 07:24:25 crc kubenswrapper[4870]: I1014 07:24:25.468812 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a2e4c80-d44f-40c0-9c04-74e9ce98ae36" containerName="barbican-keystone-listener-log" Oct 14 07:24:25 crc kubenswrapper[4870]: E1014 07:24:25.468832 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c5bb3fde-cfaf-4d09-ad31-e23fd830eb91" containerName="barbican-worker" Oct 14 07:24:25 crc kubenswrapper[4870]: I1014 07:24:25.468843 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="c5bb3fde-cfaf-4d09-ad31-e23fd830eb91" containerName="barbican-worker" Oct 14 07:24:25 crc kubenswrapper[4870]: E1014 07:24:25.468864 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="524b0b2a-77a2-4c05-be39-c8b97666ab11" containerName="ceilometer-central-agent" Oct 14 07:24:25 crc kubenswrapper[4870]: I1014 07:24:25.468875 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="524b0b2a-77a2-4c05-be39-c8b97666ab11" containerName="ceilometer-central-agent" Oct 14 07:24:25 crc kubenswrapper[4870]: E1014 07:24:25.468887 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="416d7b39-4541-44a8-a55f-924bc86fee32" containerName="openstack-network-exporter" Oct 14 07:24:25 crc kubenswrapper[4870]: I1014 07:24:25.468896 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="416d7b39-4541-44a8-a55f-924bc86fee32" containerName="openstack-network-exporter" Oct 14 07:24:25 crc kubenswrapper[4870]: E1014 07:24:25.468911 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="524b0b2a-77a2-4c05-be39-c8b97666ab11" containerName="sg-core" Oct 14 07:24:25 crc kubenswrapper[4870]: I1014 07:24:25.468920 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="524b0b2a-77a2-4c05-be39-c8b97666ab11" containerName="sg-core" Oct 14 07:24:25 crc kubenswrapper[4870]: E1014 07:24:25.468932 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2bd1f41f-3f1a-4ca7-8789-429104ce2120" containerName="account-replicator" Oct 14 07:24:25 crc kubenswrapper[4870]: I1014 07:24:25.468943 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="2bd1f41f-3f1a-4ca7-8789-429104ce2120" containerName="account-replicator" Oct 14 07:24:25 crc kubenswrapper[4870]: E1014 07:24:25.468955 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="524b0b2a-77a2-4c05-be39-c8b97666ab11" containerName="proxy-httpd" Oct 14 07:24:25 crc kubenswrapper[4870]: I1014 07:24:25.468965 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="524b0b2a-77a2-4c05-be39-c8b97666ab11" containerName="proxy-httpd" Oct 14 07:24:25 crc kubenswrapper[4870]: E1014 07:24:25.468980 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5b700ed-3919-458f-860e-5630b8e4b800" containerName="nova-scheduler-scheduler" Oct 14 07:24:25 crc kubenswrapper[4870]: I1014 07:24:25.468991 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5b700ed-3919-458f-860e-5630b8e4b800" containerName="nova-scheduler-scheduler" Oct 14 07:24:25 crc kubenswrapper[4870]: E1014 07:24:25.469008 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2bd1f41f-3f1a-4ca7-8789-429104ce2120" containerName="container-replicator" Oct 14 07:24:25 crc kubenswrapper[4870]: I1014 07:24:25.469018 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="2bd1f41f-3f1a-4ca7-8789-429104ce2120" containerName="container-replicator" Oct 14 07:24:25 crc kubenswrapper[4870]: E1014 07:24:25.469038 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="60de9d04-7c24-42a4-9d5e-74b611487b69" containerName="keystone-api" Oct 14 07:24:25 crc kubenswrapper[4870]: I1014 07:24:25.469048 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="60de9d04-7c24-42a4-9d5e-74b611487b69" containerName="keystone-api" Oct 14 07:24:25 crc kubenswrapper[4870]: E1014 07:24:25.469060 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ba0e2ed-57d0-4535-883e-10b9b985b3d3" containerName="neutron-api" Oct 14 07:24:25 crc kubenswrapper[4870]: I1014 07:24:25.469070 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ba0e2ed-57d0-4535-883e-10b9b985b3d3" containerName="neutron-api" Oct 14 07:24:25 crc kubenswrapper[4870]: E1014 07:24:25.469086 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="751daf42-b5c1-4318-a15b-d5911c9d61ea" containerName="dnsmasq-dns" Oct 14 07:24:25 crc kubenswrapper[4870]: I1014 07:24:25.469097 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="751daf42-b5c1-4318-a15b-d5911c9d61ea" containerName="dnsmasq-dns" Oct 14 07:24:25 crc kubenswrapper[4870]: E1014 07:24:25.469113 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91a2e1de-104f-4e14-9192-c6587ed13be1" containerName="ovsdb-server" Oct 14 07:24:25 crc kubenswrapper[4870]: I1014 07:24:25.469123 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="91a2e1de-104f-4e14-9192-c6587ed13be1" containerName="ovsdb-server" Oct 14 07:24:25 crc kubenswrapper[4870]: E1014 07:24:25.469139 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cae5b655-5d7c-4063-90ad-4d9913792d5b" containerName="barbican-api" Oct 14 07:24:25 crc kubenswrapper[4870]: I1014 07:24:25.469149 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="cae5b655-5d7c-4063-90ad-4d9913792d5b" containerName="barbican-api" Oct 14 07:24:25 crc kubenswrapper[4870]: E1014 07:24:25.469165 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c712b87a-cf4d-437e-abe0-1a566046ca66" containerName="glance-log" Oct 14 07:24:25 crc kubenswrapper[4870]: I1014 07:24:25.469176 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="c712b87a-cf4d-437e-abe0-1a566046ca66" containerName="glance-log" Oct 14 07:24:25 crc kubenswrapper[4870]: E1014 07:24:25.469192 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d74ef50-1861-4d70-b781-f92d5f4b35ba" containerName="mysql-bootstrap" Oct 14 07:24:25 crc kubenswrapper[4870]: I1014 07:24:25.469202 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d74ef50-1861-4d70-b781-f92d5f4b35ba" containerName="mysql-bootstrap" Oct 14 07:24:25 crc kubenswrapper[4870]: E1014 07:24:25.469217 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ba0e2ed-57d0-4535-883e-10b9b985b3d3" containerName="neutron-httpd" Oct 14 07:24:25 crc kubenswrapper[4870]: I1014 07:24:25.469230 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ba0e2ed-57d0-4535-883e-10b9b985b3d3" containerName="neutron-httpd" Oct 14 07:24:25 crc kubenswrapper[4870]: E1014 07:24:25.469241 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2bd1f41f-3f1a-4ca7-8789-429104ce2120" containerName="account-auditor" Oct 14 07:24:25 crc kubenswrapper[4870]: I1014 07:24:25.469252 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="2bd1f41f-3f1a-4ca7-8789-429104ce2120" containerName="account-auditor" Oct 14 07:24:25 crc kubenswrapper[4870]: E1014 07:24:25.469269 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2bd1f41f-3f1a-4ca7-8789-429104ce2120" containerName="object-replicator" Oct 14 07:24:25 crc kubenswrapper[4870]: I1014 07:24:25.469279 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="2bd1f41f-3f1a-4ca7-8789-429104ce2120" containerName="object-replicator" Oct 14 07:24:25 crc kubenswrapper[4870]: E1014 07:24:25.469298 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7bc76b72-8679-4b33-8966-0ee70c1788e5" containerName="cinder-scheduler" Oct 14 07:24:25 crc kubenswrapper[4870]: I1014 07:24:25.469308 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="7bc76b72-8679-4b33-8966-0ee70c1788e5" containerName="cinder-scheduler" Oct 14 07:24:25 crc kubenswrapper[4870]: E1014 07:24:25.469327 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2bd1f41f-3f1a-4ca7-8789-429104ce2120" containerName="container-updater" Oct 14 07:24:25 crc kubenswrapper[4870]: I1014 07:24:25.469336 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="2bd1f41f-3f1a-4ca7-8789-429104ce2120" containerName="container-updater" Oct 14 07:24:25 crc kubenswrapper[4870]: E1014 07:24:25.469349 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3927fb78-4f30-4371-8fbb-62fd276c0460" containerName="proxy-httpd" Oct 14 07:24:25 crc kubenswrapper[4870]: I1014 07:24:25.469359 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="3927fb78-4f30-4371-8fbb-62fd276c0460" containerName="proxy-httpd" Oct 14 07:24:25 crc kubenswrapper[4870]: E1014 07:24:25.469371 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7cbdbbdd-ea4f-4d14-8f3c-4784bda45874" containerName="mariadb-account-delete" Oct 14 07:24:25 crc kubenswrapper[4870]: I1014 07:24:25.469379 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="7cbdbbdd-ea4f-4d14-8f3c-4784bda45874" containerName="mariadb-account-delete" Oct 14 07:24:25 crc kubenswrapper[4870]: E1014 07:24:25.469390 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2bd1f41f-3f1a-4ca7-8789-429104ce2120" containerName="object-expirer" Oct 14 07:24:25 crc kubenswrapper[4870]: I1014 07:24:25.469398 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="2bd1f41f-3f1a-4ca7-8789-429104ce2120" containerName="object-expirer" Oct 14 07:24:25 crc kubenswrapper[4870]: E1014 07:24:25.469409 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="208754ea-7cc0-4222-a186-9ba01b274add" containerName="mariadb-account-delete" Oct 14 07:24:25 crc kubenswrapper[4870]: I1014 07:24:25.469417 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="208754ea-7cc0-4222-a186-9ba01b274add" containerName="mariadb-account-delete" Oct 14 07:24:25 crc kubenswrapper[4870]: E1014 07:24:25.469429 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2bd1f41f-3f1a-4ca7-8789-429104ce2120" containerName="rsync" Oct 14 07:24:25 crc kubenswrapper[4870]: I1014 07:24:25.469460 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="2bd1f41f-3f1a-4ca7-8789-429104ce2120" containerName="rsync" Oct 14 07:24:25 crc kubenswrapper[4870]: E1014 07:24:25.469471 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2130e3f2-cbea-46e1-b5d3-daefdd972935" containerName="placement-log" Oct 14 07:24:25 crc kubenswrapper[4870]: I1014 07:24:25.469482 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="2130e3f2-cbea-46e1-b5d3-daefdd972935" containerName="placement-log" Oct 14 07:24:25 crc kubenswrapper[4870]: E1014 07:24:25.469496 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91a2e1de-104f-4e14-9192-c6587ed13be1" containerName="ovs-vswitchd" Oct 14 07:24:25 crc kubenswrapper[4870]: I1014 07:24:25.469506 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="91a2e1de-104f-4e14-9192-c6587ed13be1" containerName="ovs-vswitchd" Oct 14 07:24:25 crc kubenswrapper[4870]: E1014 07:24:25.469526 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2bd1f41f-3f1a-4ca7-8789-429104ce2120" containerName="container-server" Oct 14 07:24:25 crc kubenswrapper[4870]: I1014 07:24:25.469534 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="2bd1f41f-3f1a-4ca7-8789-429104ce2120" containerName="container-server" Oct 14 07:24:25 crc kubenswrapper[4870]: E1014 07:24:25.469543 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c712b87a-cf4d-437e-abe0-1a566046ca66" containerName="glance-httpd" Oct 14 07:24:25 crc kubenswrapper[4870]: I1014 07:24:25.469551 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="c712b87a-cf4d-437e-abe0-1a566046ca66" containerName="glance-httpd" Oct 14 07:24:25 crc kubenswrapper[4870]: E1014 07:24:25.469564 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="36717039-e3e2-4b74-b612-655556122574" containerName="openstack-network-exporter" Oct 14 07:24:25 crc kubenswrapper[4870]: I1014 07:24:25.469572 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="36717039-e3e2-4b74-b612-655556122574" containerName="openstack-network-exporter" Oct 14 07:24:25 crc kubenswrapper[4870]: E1014 07:24:25.469587 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="503a07f8-2c05-458a-80ff-1abfe973dbd5" containerName="ovn-controller" Oct 14 07:24:25 crc kubenswrapper[4870]: I1014 07:24:25.469595 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="503a07f8-2c05-458a-80ff-1abfe973dbd5" containerName="ovn-controller" Oct 14 07:24:25 crc kubenswrapper[4870]: E1014 07:24:25.469605 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="51de8090-c48e-498c-99ab-794a7527252c" containerName="cinder-api" Oct 14 07:24:25 crc kubenswrapper[4870]: I1014 07:24:25.469613 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="51de8090-c48e-498c-99ab-794a7527252c" containerName="cinder-api" Oct 14 07:24:25 crc kubenswrapper[4870]: E1014 07:24:25.469627 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7bc76b72-8679-4b33-8966-0ee70c1788e5" containerName="probe" Oct 14 07:24:25 crc kubenswrapper[4870]: I1014 07:24:25.469635 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="7bc76b72-8679-4b33-8966-0ee70c1788e5" containerName="probe" Oct 14 07:24:25 crc kubenswrapper[4870]: E1014 07:24:25.469650 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2bd1f41f-3f1a-4ca7-8789-429104ce2120" containerName="swift-recon-cron" Oct 14 07:24:25 crc kubenswrapper[4870]: I1014 07:24:25.469658 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="2bd1f41f-3f1a-4ca7-8789-429104ce2120" containerName="swift-recon-cron" Oct 14 07:24:25 crc kubenswrapper[4870]: E1014 07:24:25.469697 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91a2e1de-104f-4e14-9192-c6587ed13be1" containerName="ovsdb-server-init" Oct 14 07:24:25 crc kubenswrapper[4870]: I1014 07:24:25.469705 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="91a2e1de-104f-4e14-9192-c6587ed13be1" containerName="ovsdb-server-init" Oct 14 07:24:25 crc kubenswrapper[4870]: E1014 07:24:25.469715 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2bd1f41f-3f1a-4ca7-8789-429104ce2120" containerName="object-auditor" Oct 14 07:24:25 crc kubenswrapper[4870]: I1014 07:24:25.469722 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="2bd1f41f-3f1a-4ca7-8789-429104ce2120" containerName="object-auditor" Oct 14 07:24:25 crc kubenswrapper[4870]: E1014 07:24:25.469734 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2bd1f41f-3f1a-4ca7-8789-429104ce2120" containerName="object-updater" Oct 14 07:24:25 crc kubenswrapper[4870]: I1014 07:24:25.469741 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="2bd1f41f-3f1a-4ca7-8789-429104ce2120" containerName="object-updater" Oct 14 07:24:25 crc kubenswrapper[4870]: E1014 07:24:25.469752 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8abede26-36e8-46ec-a5c4-6cd53319133a" containerName="nova-cell1-novncproxy-novncproxy" Oct 14 07:24:25 crc kubenswrapper[4870]: I1014 07:24:25.469762 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="8abede26-36e8-46ec-a5c4-6cd53319133a" containerName="nova-cell1-novncproxy-novncproxy" Oct 14 07:24:25 crc kubenswrapper[4870]: E1014 07:24:25.469774 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a6f02e4f-8dca-4b97-91e9-ee3368e5e44b" containerName="mariadb-account-delete" Oct 14 07:24:25 crc kubenswrapper[4870]: I1014 07:24:25.469782 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="a6f02e4f-8dca-4b97-91e9-ee3368e5e44b" containerName="mariadb-account-delete" Oct 14 07:24:25 crc kubenswrapper[4870]: E1014 07:24:25.469793 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3de2167a-7663-4b00-9743-229ff1864cb8" containerName="mysql-bootstrap" Oct 14 07:24:25 crc kubenswrapper[4870]: I1014 07:24:25.469801 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="3de2167a-7663-4b00-9743-229ff1864cb8" containerName="mysql-bootstrap" Oct 14 07:24:25 crc kubenswrapper[4870]: E1014 07:24:25.469810 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61885759-41a7-4ee8-904a-3f85bb66cd19" containerName="nova-cell1-conductor-conductor" Oct 14 07:24:25 crc kubenswrapper[4870]: I1014 07:24:25.469818 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="61885759-41a7-4ee8-904a-3f85bb66cd19" containerName="nova-cell1-conductor-conductor" Oct 14 07:24:25 crc kubenswrapper[4870]: E1014 07:24:25.469833 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="524b0b2a-77a2-4c05-be39-c8b97666ab11" containerName="ceilometer-notification-agent" Oct 14 07:24:25 crc kubenswrapper[4870]: I1014 07:24:25.469841 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="524b0b2a-77a2-4c05-be39-c8b97666ab11" containerName="ceilometer-notification-agent" Oct 14 07:24:25 crc kubenswrapper[4870]: E1014 07:24:25.469850 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de479f5d-22a2-4f7e-8034-cffefc511b53" containerName="mariadb-account-delete" Oct 14 07:24:25 crc kubenswrapper[4870]: I1014 07:24:25.469858 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="de479f5d-22a2-4f7e-8034-cffefc511b53" containerName="mariadb-account-delete" Oct 14 07:24:25 crc kubenswrapper[4870]: E1014 07:24:25.469873 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da8a9731-f917-462f-9932-b37b6abb9a64" containerName="setup-container" Oct 14 07:24:25 crc kubenswrapper[4870]: I1014 07:24:25.469881 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="da8a9731-f917-462f-9932-b37b6abb9a64" containerName="setup-container" Oct 14 07:24:25 crc kubenswrapper[4870]: E1014 07:24:25.469893 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a2e4c80-d44f-40c0-9c04-74e9ce98ae36" containerName="barbican-keystone-listener" Oct 14 07:24:25 crc kubenswrapper[4870]: I1014 07:24:25.469903 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a2e4c80-d44f-40c0-9c04-74e9ce98ae36" containerName="barbican-keystone-listener" Oct 14 07:24:25 crc kubenswrapper[4870]: E1014 07:24:25.469918 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3de2167a-7663-4b00-9743-229ff1864cb8" containerName="galera" Oct 14 07:24:25 crc kubenswrapper[4870]: I1014 07:24:25.469927 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="3de2167a-7663-4b00-9743-229ff1864cb8" containerName="galera" Oct 14 07:24:25 crc kubenswrapper[4870]: E1014 07:24:25.469943 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da8a9731-f917-462f-9932-b37b6abb9a64" containerName="rabbitmq" Oct 14 07:24:25 crc kubenswrapper[4870]: I1014 07:24:25.469951 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="da8a9731-f917-462f-9932-b37b6abb9a64" containerName="rabbitmq" Oct 14 07:24:25 crc kubenswrapper[4870]: E1014 07:24:25.469964 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5212ba20-83c3-41ff-a3d9-5a315f009c1b" containerName="rabbitmq" Oct 14 07:24:25 crc kubenswrapper[4870]: I1014 07:24:25.469972 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="5212ba20-83c3-41ff-a3d9-5a315f009c1b" containerName="rabbitmq" Oct 14 07:24:25 crc kubenswrapper[4870]: E1014 07:24:25.469984 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bad45b30-5ba2-4062-ab14-04f923679c4e" containerName="nova-cell0-conductor-conductor" Oct 14 07:24:25 crc kubenswrapper[4870]: I1014 07:24:25.469992 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="bad45b30-5ba2-4062-ab14-04f923679c4e" containerName="nova-cell0-conductor-conductor" Oct 14 07:24:25 crc kubenswrapper[4870]: E1014 07:24:25.470002 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c603f7b5-015f-4c2e-9144-b3e8349c3da3" containerName="kube-state-metrics" Oct 14 07:24:25 crc kubenswrapper[4870]: I1014 07:24:25.470010 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="c603f7b5-015f-4c2e-9144-b3e8349c3da3" containerName="kube-state-metrics" Oct 14 07:24:25 crc kubenswrapper[4870]: E1014 07:24:25.470020 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3bd1e649-3d09-4cda-921f-ac9a84a5066e" containerName="memcached" Oct 14 07:24:25 crc kubenswrapper[4870]: I1014 07:24:25.470028 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="3bd1e649-3d09-4cda-921f-ac9a84a5066e" containerName="memcached" Oct 14 07:24:25 crc kubenswrapper[4870]: E1014 07:24:25.470041 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2130e3f2-cbea-46e1-b5d3-daefdd972935" containerName="placement-api" Oct 14 07:24:25 crc kubenswrapper[4870]: I1014 07:24:25.470049 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="2130e3f2-cbea-46e1-b5d3-daefdd972935" containerName="placement-api" Oct 14 07:24:25 crc kubenswrapper[4870]: E1014 07:24:25.470063 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="751daf42-b5c1-4318-a15b-d5911c9d61ea" containerName="init" Oct 14 07:24:25 crc kubenswrapper[4870]: I1014 07:24:25.470071 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="751daf42-b5c1-4318-a15b-d5911c9d61ea" containerName="init" Oct 14 07:24:25 crc kubenswrapper[4870]: E1014 07:24:25.470082 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2bd1f41f-3f1a-4ca7-8789-429104ce2120" containerName="account-server" Oct 14 07:24:25 crc kubenswrapper[4870]: I1014 07:24:25.470090 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="2bd1f41f-3f1a-4ca7-8789-429104ce2120" containerName="account-server" Oct 14 07:24:25 crc kubenswrapper[4870]: E1014 07:24:25.470100 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56eb286f-8aa2-46aa-a74d-031288a9aad1" containerName="glance-log" Oct 14 07:24:25 crc kubenswrapper[4870]: I1014 07:24:25.470107 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="56eb286f-8aa2-46aa-a74d-031288a9aad1" containerName="glance-log" Oct 14 07:24:25 crc kubenswrapper[4870]: E1014 07:24:25.470117 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff28d060-b280-4b89-bea1-d10aa367a3cc" containerName="nova-api-log" Oct 14 07:24:25 crc kubenswrapper[4870]: I1014 07:24:25.470125 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff28d060-b280-4b89-bea1-d10aa367a3cc" containerName="nova-api-log" Oct 14 07:24:25 crc kubenswrapper[4870]: E1014 07:24:25.470138 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c8d0269-7501-4c96-9f9e-30bea4b0fb92" containerName="openstack-network-exporter" Oct 14 07:24:25 crc kubenswrapper[4870]: I1014 07:24:25.470146 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c8d0269-7501-4c96-9f9e-30bea4b0fb92" containerName="openstack-network-exporter" Oct 14 07:24:25 crc kubenswrapper[4870]: E1014 07:24:25.470159 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2bd1f41f-3f1a-4ca7-8789-429104ce2120" containerName="container-auditor" Oct 14 07:24:25 crc kubenswrapper[4870]: I1014 07:24:25.470169 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="2bd1f41f-3f1a-4ca7-8789-429104ce2120" containerName="container-auditor" Oct 14 07:24:25 crc kubenswrapper[4870]: E1014 07:24:25.470180 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2bd1f41f-3f1a-4ca7-8789-429104ce2120" containerName="object-server" Oct 14 07:24:25 crc kubenswrapper[4870]: I1014 07:24:25.470187 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="2bd1f41f-3f1a-4ca7-8789-429104ce2120" containerName="object-server" Oct 14 07:24:25 crc kubenswrapper[4870]: E1014 07:24:25.470198 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cae5b655-5d7c-4063-90ad-4d9913792d5b" containerName="barbican-api-log" Oct 14 07:24:25 crc kubenswrapper[4870]: I1014 07:24:25.470205 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="cae5b655-5d7c-4063-90ad-4d9913792d5b" containerName="barbican-api-log" Oct 14 07:24:25 crc kubenswrapper[4870]: E1014 07:24:25.470219 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9d8a25d-2f1c-483d-bf0d-76a2bf30f137" containerName="mariadb-account-delete" Oct 14 07:24:25 crc kubenswrapper[4870]: I1014 07:24:25.470227 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9d8a25d-2f1c-483d-bf0d-76a2bf30f137" containerName="mariadb-account-delete" Oct 14 07:24:25 crc kubenswrapper[4870]: E1014 07:24:25.470240 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2bd1f41f-3f1a-4ca7-8789-429104ce2120" containerName="account-reaper" Oct 14 07:24:25 crc kubenswrapper[4870]: I1014 07:24:25.470247 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="2bd1f41f-3f1a-4ca7-8789-429104ce2120" containerName="account-reaper" Oct 14 07:24:25 crc kubenswrapper[4870]: E1014 07:24:25.470261 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3927fb78-4f30-4371-8fbb-62fd276c0460" containerName="proxy-server" Oct 14 07:24:25 crc kubenswrapper[4870]: I1014 07:24:25.470268 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="3927fb78-4f30-4371-8fbb-62fd276c0460" containerName="proxy-server" Oct 14 07:24:25 crc kubenswrapper[4870]: E1014 07:24:25.470281 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56eb286f-8aa2-46aa-a74d-031288a9aad1" containerName="glance-httpd" Oct 14 07:24:25 crc kubenswrapper[4870]: I1014 07:24:25.470289 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="56eb286f-8aa2-46aa-a74d-031288a9aad1" containerName="glance-httpd" Oct 14 07:24:25 crc kubenswrapper[4870]: E1014 07:24:25.470311 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="51de8090-c48e-498c-99ab-794a7527252c" containerName="cinder-api-log" Oct 14 07:24:25 crc kubenswrapper[4870]: I1014 07:24:25.470319 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="51de8090-c48e-498c-99ab-794a7527252c" containerName="cinder-api-log" Oct 14 07:24:25 crc kubenswrapper[4870]: E1014 07:24:25.470329 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f0b766fd-ccb9-4620-84df-8fe2f194b0da" containerName="nova-metadata-log" Oct 14 07:24:25 crc kubenswrapper[4870]: I1014 07:24:25.470337 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="f0b766fd-ccb9-4620-84df-8fe2f194b0da" containerName="nova-metadata-log" Oct 14 07:24:25 crc kubenswrapper[4870]: E1014 07:24:25.470347 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff28d060-b280-4b89-bea1-d10aa367a3cc" containerName="nova-api-api" Oct 14 07:24:25 crc kubenswrapper[4870]: I1014 07:24:25.470354 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff28d060-b280-4b89-bea1-d10aa367a3cc" containerName="nova-api-api" Oct 14 07:24:25 crc kubenswrapper[4870]: E1014 07:24:25.470362 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f0b766fd-ccb9-4620-84df-8fe2f194b0da" containerName="nova-metadata-metadata" Oct 14 07:24:25 crc kubenswrapper[4870]: I1014 07:24:25.470369 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="f0b766fd-ccb9-4620-84df-8fe2f194b0da" containerName="nova-metadata-metadata" Oct 14 07:24:25 crc kubenswrapper[4870]: E1014 07:24:25.470383 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2abee394-ffb3-4dba-8d09-19c1cb476595" containerName="mariadb-account-delete" Oct 14 07:24:25 crc kubenswrapper[4870]: I1014 07:24:25.470391 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="2abee394-ffb3-4dba-8d09-19c1cb476595" containerName="mariadb-account-delete" Oct 14 07:24:25 crc kubenswrapper[4870]: I1014 07:24:25.470612 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="0a2e4c80-d44f-40c0-9c04-74e9ce98ae36" containerName="barbican-keystone-listener" Oct 14 07:24:25 crc kubenswrapper[4870]: I1014 07:24:25.470634 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="4c8d0269-7501-4c96-9f9e-30bea4b0fb92" containerName="openstack-network-exporter" Oct 14 07:24:25 crc kubenswrapper[4870]: I1014 07:24:25.470649 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="2bd1f41f-3f1a-4ca7-8789-429104ce2120" containerName="swift-recon-cron" Oct 14 07:24:25 crc kubenswrapper[4870]: I1014 07:24:25.470661 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="3de2167a-7663-4b00-9743-229ff1864cb8" containerName="galera" Oct 14 07:24:25 crc kubenswrapper[4870]: I1014 07:24:25.470674 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="3bd1e649-3d09-4cda-921f-ac9a84a5066e" containerName="memcached" Oct 14 07:24:25 crc kubenswrapper[4870]: I1014 07:24:25.470688 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="2bd1f41f-3f1a-4ca7-8789-429104ce2120" containerName="container-auditor" Oct 14 07:24:25 crc kubenswrapper[4870]: I1014 07:24:25.470700 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="416d7b39-4541-44a8-a55f-924bc86fee32" containerName="ovn-northd" Oct 14 07:24:25 crc kubenswrapper[4870]: I1014 07:24:25.470712 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="3927fb78-4f30-4371-8fbb-62fd276c0460" containerName="proxy-server" Oct 14 07:24:25 crc kubenswrapper[4870]: I1014 07:24:25.470720 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="7bc76b72-8679-4b33-8966-0ee70c1788e5" containerName="cinder-scheduler" Oct 14 07:24:25 crc kubenswrapper[4870]: I1014 07:24:25.470735 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="4d74ef50-1861-4d70-b781-f92d5f4b35ba" containerName="galera" Oct 14 07:24:25 crc kubenswrapper[4870]: I1014 07:24:25.470742 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="2bd1f41f-3f1a-4ca7-8789-429104ce2120" containerName="rsync" Oct 14 07:24:25 crc kubenswrapper[4870]: I1014 07:24:25.470755 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="208754ea-7cc0-4222-a186-9ba01b274add" containerName="mariadb-account-delete" Oct 14 07:24:25 crc kubenswrapper[4870]: I1014 07:24:25.470771 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="524b0b2a-77a2-4c05-be39-c8b97666ab11" containerName="ceilometer-notification-agent" Oct 14 07:24:25 crc kubenswrapper[4870]: I1014 07:24:25.470784 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="2bd1f41f-3f1a-4ca7-8789-429104ce2120" containerName="object-server" Oct 14 07:24:25 crc kubenswrapper[4870]: I1014 07:24:25.470795 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="ff28d060-b280-4b89-bea1-d10aa367a3cc" containerName="nova-api-api" Oct 14 07:24:25 crc kubenswrapper[4870]: I1014 07:24:25.470815 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="36717039-e3e2-4b74-b612-655556122574" containerName="openstack-network-exporter" Oct 14 07:24:25 crc kubenswrapper[4870]: I1014 07:24:25.470827 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="cae5b655-5d7c-4063-90ad-4d9913792d5b" containerName="barbican-api" Oct 14 07:24:25 crc kubenswrapper[4870]: I1014 07:24:25.470840 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="0ba0e2ed-57d0-4535-883e-10b9b985b3d3" containerName="neutron-api" Oct 14 07:24:25 crc kubenswrapper[4870]: I1014 07:24:25.470854 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="2bd1f41f-3f1a-4ca7-8789-429104ce2120" containerName="account-auditor" Oct 14 07:24:25 crc kubenswrapper[4870]: I1014 07:24:25.470866 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="2bd1f41f-3f1a-4ca7-8789-429104ce2120" containerName="object-replicator" Oct 14 07:24:25 crc kubenswrapper[4870]: I1014 07:24:25.470880 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="524b0b2a-77a2-4c05-be39-c8b97666ab11" containerName="sg-core" Oct 14 07:24:25 crc kubenswrapper[4870]: I1014 07:24:25.470890 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="2bd1f41f-3f1a-4ca7-8789-429104ce2120" containerName="object-expirer" Oct 14 07:24:25 crc kubenswrapper[4870]: I1014 07:24:25.470903 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="51de8090-c48e-498c-99ab-794a7527252c" containerName="cinder-api" Oct 14 07:24:25 crc kubenswrapper[4870]: I1014 07:24:25.470914 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="7bc76b72-8679-4b33-8966-0ee70c1788e5" containerName="probe" Oct 14 07:24:25 crc kubenswrapper[4870]: I1014 07:24:25.470922 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="2bd1f41f-3f1a-4ca7-8789-429104ce2120" containerName="container-replicator" Oct 14 07:24:25 crc kubenswrapper[4870]: I1014 07:24:25.470930 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="2abee394-ffb3-4dba-8d09-19c1cb476595" containerName="mariadb-account-delete" Oct 14 07:24:25 crc kubenswrapper[4870]: I1014 07:24:25.470940 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="f5b700ed-3919-458f-860e-5630b8e4b800" containerName="nova-scheduler-scheduler" Oct 14 07:24:25 crc kubenswrapper[4870]: I1014 07:24:25.470949 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="bad45b30-5ba2-4062-ab14-04f923679c4e" containerName="nova-cell0-conductor-conductor" Oct 14 07:24:25 crc kubenswrapper[4870]: I1014 07:24:25.470961 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="c5bb3fde-cfaf-4d09-ad31-e23fd830eb91" containerName="barbican-worker" Oct 14 07:24:25 crc kubenswrapper[4870]: I1014 07:24:25.470970 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="2bd1f41f-3f1a-4ca7-8789-429104ce2120" containerName="account-replicator" Oct 14 07:24:25 crc kubenswrapper[4870]: I1014 07:24:25.470981 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="c5bb3fde-cfaf-4d09-ad31-e23fd830eb91" containerName="barbican-worker-log" Oct 14 07:24:25 crc kubenswrapper[4870]: I1014 07:24:25.470990 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="36717039-e3e2-4b74-b612-655556122574" containerName="ovsdbserver-sb" Oct 14 07:24:25 crc kubenswrapper[4870]: I1014 07:24:25.470999 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="2130e3f2-cbea-46e1-b5d3-daefdd972935" containerName="placement-api" Oct 14 07:24:25 crc kubenswrapper[4870]: I1014 07:24:25.471006 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="c712b87a-cf4d-437e-abe0-1a566046ca66" containerName="glance-log" Oct 14 07:24:25 crc kubenswrapper[4870]: I1014 07:24:25.471019 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="ff28d060-b280-4b89-bea1-d10aa367a3cc" containerName="nova-api-log" Oct 14 07:24:25 crc kubenswrapper[4870]: I1014 07:24:25.471029 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="09f89569-6fe3-4b3f-9394-a990ad8cde30" containerName="openstack-network-exporter" Oct 14 07:24:25 crc kubenswrapper[4870]: I1014 07:24:25.471042 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="56eb286f-8aa2-46aa-a74d-031288a9aad1" containerName="glance-httpd" Oct 14 07:24:25 crc kubenswrapper[4870]: I1014 07:24:25.471051 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="503a07f8-2c05-458a-80ff-1abfe973dbd5" containerName="ovn-controller" Oct 14 07:24:25 crc kubenswrapper[4870]: I1014 07:24:25.471061 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="91a2e1de-104f-4e14-9192-c6587ed13be1" containerName="ovsdb-server" Oct 14 07:24:25 crc kubenswrapper[4870]: I1014 07:24:25.471070 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="cae5b655-5d7c-4063-90ad-4d9913792d5b" containerName="barbican-api-log" Oct 14 07:24:25 crc kubenswrapper[4870]: I1014 07:24:25.471079 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="de479f5d-22a2-4f7e-8034-cffefc511b53" containerName="mariadb-account-delete" Oct 14 07:24:25 crc kubenswrapper[4870]: I1014 07:24:25.471088 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="5212ba20-83c3-41ff-a3d9-5a315f009c1b" containerName="rabbitmq" Oct 14 07:24:25 crc kubenswrapper[4870]: I1014 07:24:25.471099 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="2bd1f41f-3f1a-4ca7-8789-429104ce2120" containerName="object-updater" Oct 14 07:24:25 crc kubenswrapper[4870]: I1014 07:24:25.471108 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="60de9d04-7c24-42a4-9d5e-74b611487b69" containerName="keystone-api" Oct 14 07:24:25 crc kubenswrapper[4870]: I1014 07:24:25.471117 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="8abede26-36e8-46ec-a5c4-6cd53319133a" containerName="nova-cell1-novncproxy-novncproxy" Oct 14 07:24:25 crc kubenswrapper[4870]: I1014 07:24:25.471128 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="524b0b2a-77a2-4c05-be39-c8b97666ab11" containerName="ceilometer-central-agent" Oct 14 07:24:25 crc kubenswrapper[4870]: I1014 07:24:25.471141 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="a9d8a25d-2f1c-483d-bf0d-76a2bf30f137" containerName="mariadb-account-delete" Oct 14 07:24:25 crc kubenswrapper[4870]: I1014 07:24:25.471148 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="2bd1f41f-3f1a-4ca7-8789-429104ce2120" containerName="account-reaper" Oct 14 07:24:25 crc kubenswrapper[4870]: I1014 07:24:25.471156 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="2bd1f41f-3f1a-4ca7-8789-429104ce2120" containerName="container-server" Oct 14 07:24:25 crc kubenswrapper[4870]: I1014 07:24:25.471166 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="51de8090-c48e-498c-99ab-794a7527252c" containerName="cinder-api-log" Oct 14 07:24:25 crc kubenswrapper[4870]: I1014 07:24:25.471175 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="416d7b39-4541-44a8-a55f-924bc86fee32" containerName="openstack-network-exporter" Oct 14 07:24:25 crc kubenswrapper[4870]: I1014 07:24:25.471186 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="3927fb78-4f30-4371-8fbb-62fd276c0460" containerName="proxy-httpd" Oct 14 07:24:25 crc kubenswrapper[4870]: I1014 07:24:25.471199 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="2bd1f41f-3f1a-4ca7-8789-429104ce2120" containerName="container-updater" Oct 14 07:24:25 crc kubenswrapper[4870]: I1014 07:24:25.471209 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="0a2e4c80-d44f-40c0-9c04-74e9ce98ae36" containerName="barbican-keystone-listener-log" Oct 14 07:24:25 crc kubenswrapper[4870]: I1014 07:24:25.471222 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="91a2e1de-104f-4e14-9192-c6587ed13be1" containerName="ovs-vswitchd" Oct 14 07:24:25 crc kubenswrapper[4870]: I1014 07:24:25.471231 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="a6f02e4f-8dca-4b97-91e9-ee3368e5e44b" containerName="mariadb-account-delete" Oct 14 07:24:25 crc kubenswrapper[4870]: I1014 07:24:25.471242 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="f0b766fd-ccb9-4620-84df-8fe2f194b0da" containerName="nova-metadata-metadata" Oct 14 07:24:25 crc kubenswrapper[4870]: I1014 07:24:25.471250 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="f0b766fd-ccb9-4620-84df-8fe2f194b0da" containerName="nova-metadata-log" Oct 14 07:24:25 crc kubenswrapper[4870]: I1014 07:24:25.471261 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="2130e3f2-cbea-46e1-b5d3-daefdd972935" containerName="placement-log" Oct 14 07:24:25 crc kubenswrapper[4870]: I1014 07:24:25.471273 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="524b0b2a-77a2-4c05-be39-c8b97666ab11" containerName="proxy-httpd" Oct 14 07:24:25 crc kubenswrapper[4870]: I1014 07:24:25.471281 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="2bd1f41f-3f1a-4ca7-8789-429104ce2120" containerName="account-server" Oct 14 07:24:25 crc kubenswrapper[4870]: I1014 07:24:25.471292 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="0ba0e2ed-57d0-4535-883e-10b9b985b3d3" containerName="neutron-httpd" Oct 14 07:24:25 crc kubenswrapper[4870]: I1014 07:24:25.471304 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="751daf42-b5c1-4318-a15b-d5911c9d61ea" containerName="dnsmasq-dns" Oct 14 07:24:25 crc kubenswrapper[4870]: I1014 07:24:25.471317 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="7cbdbbdd-ea4f-4d14-8f3c-4784bda45874" containerName="mariadb-account-delete" Oct 14 07:24:25 crc kubenswrapper[4870]: I1014 07:24:25.471326 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="09f89569-6fe3-4b3f-9394-a990ad8cde30" containerName="ovsdbserver-nb" Oct 14 07:24:25 crc kubenswrapper[4870]: I1014 07:24:25.471339 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="da8a9731-f917-462f-9932-b37b6abb9a64" containerName="rabbitmq" Oct 14 07:24:25 crc kubenswrapper[4870]: I1014 07:24:25.471349 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="56eb286f-8aa2-46aa-a74d-031288a9aad1" containerName="glance-log" Oct 14 07:24:25 crc kubenswrapper[4870]: I1014 07:24:25.471358 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="2bd1f41f-3f1a-4ca7-8789-429104ce2120" containerName="object-auditor" Oct 14 07:24:25 crc kubenswrapper[4870]: I1014 07:24:25.471371 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="c712b87a-cf4d-437e-abe0-1a566046ca66" containerName="glance-httpd" Oct 14 07:24:25 crc kubenswrapper[4870]: I1014 07:24:25.471379 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="61885759-41a7-4ee8-904a-3f85bb66cd19" containerName="nova-cell1-conductor-conductor" Oct 14 07:24:25 crc kubenswrapper[4870]: I1014 07:24:25.471392 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="c603f7b5-015f-4c2e-9144-b3e8349c3da3" containerName="kube-state-metrics" Oct 14 07:24:25 crc kubenswrapper[4870]: I1014 07:24:25.472688 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qm9qp" Oct 14 07:24:25 crc kubenswrapper[4870]: I1014 07:24:25.495563 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-qm9qp"] Oct 14 07:24:25 crc kubenswrapper[4870]: I1014 07:24:25.575684 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e32ab0b1-c28f-4d01-9b07-e1f0963a53cc-catalog-content\") pod \"redhat-operators-qm9qp\" (UID: \"e32ab0b1-c28f-4d01-9b07-e1f0963a53cc\") " pod="openshift-marketplace/redhat-operators-qm9qp" Oct 14 07:24:25 crc kubenswrapper[4870]: I1014 07:24:25.576181 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e32ab0b1-c28f-4d01-9b07-e1f0963a53cc-utilities\") pod \"redhat-operators-qm9qp\" (UID: \"e32ab0b1-c28f-4d01-9b07-e1f0963a53cc\") " pod="openshift-marketplace/redhat-operators-qm9qp" Oct 14 07:24:25 crc kubenswrapper[4870]: I1014 07:24:25.576262 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xnv98\" (UniqueName: \"kubernetes.io/projected/e32ab0b1-c28f-4d01-9b07-e1f0963a53cc-kube-api-access-xnv98\") pod \"redhat-operators-qm9qp\" (UID: \"e32ab0b1-c28f-4d01-9b07-e1f0963a53cc\") " pod="openshift-marketplace/redhat-operators-qm9qp" Oct 14 07:24:25 crc kubenswrapper[4870]: I1014 07:24:25.677376 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e32ab0b1-c28f-4d01-9b07-e1f0963a53cc-utilities\") pod \"redhat-operators-qm9qp\" (UID: \"e32ab0b1-c28f-4d01-9b07-e1f0963a53cc\") " pod="openshift-marketplace/redhat-operators-qm9qp" Oct 14 07:24:25 crc kubenswrapper[4870]: I1014 07:24:25.677762 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xnv98\" (UniqueName: \"kubernetes.io/projected/e32ab0b1-c28f-4d01-9b07-e1f0963a53cc-kube-api-access-xnv98\") pod \"redhat-operators-qm9qp\" (UID: \"e32ab0b1-c28f-4d01-9b07-e1f0963a53cc\") " pod="openshift-marketplace/redhat-operators-qm9qp" Oct 14 07:24:25 crc kubenswrapper[4870]: I1014 07:24:25.677887 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e32ab0b1-c28f-4d01-9b07-e1f0963a53cc-catalog-content\") pod \"redhat-operators-qm9qp\" (UID: \"e32ab0b1-c28f-4d01-9b07-e1f0963a53cc\") " pod="openshift-marketplace/redhat-operators-qm9qp" Oct 14 07:24:25 crc kubenswrapper[4870]: I1014 07:24:25.678050 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e32ab0b1-c28f-4d01-9b07-e1f0963a53cc-utilities\") pod \"redhat-operators-qm9qp\" (UID: \"e32ab0b1-c28f-4d01-9b07-e1f0963a53cc\") " pod="openshift-marketplace/redhat-operators-qm9qp" Oct 14 07:24:25 crc kubenswrapper[4870]: I1014 07:24:25.678277 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e32ab0b1-c28f-4d01-9b07-e1f0963a53cc-catalog-content\") pod \"redhat-operators-qm9qp\" (UID: \"e32ab0b1-c28f-4d01-9b07-e1f0963a53cc\") " pod="openshift-marketplace/redhat-operators-qm9qp" Oct 14 07:24:25 crc kubenswrapper[4870]: I1014 07:24:25.717192 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xnv98\" (UniqueName: \"kubernetes.io/projected/e32ab0b1-c28f-4d01-9b07-e1f0963a53cc-kube-api-access-xnv98\") pod \"redhat-operators-qm9qp\" (UID: \"e32ab0b1-c28f-4d01-9b07-e1f0963a53cc\") " pod="openshift-marketplace/redhat-operators-qm9qp" Oct 14 07:24:25 crc kubenswrapper[4870]: I1014 07:24:25.862156 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qm9qp" Oct 14 07:24:26 crc kubenswrapper[4870]: I1014 07:24:26.348264 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-qm9qp"] Oct 14 07:24:26 crc kubenswrapper[4870]: I1014 07:24:26.677612 4870 generic.go:334] "Generic (PLEG): container finished" podID="e32ab0b1-c28f-4d01-9b07-e1f0963a53cc" containerID="fd5a77e4d86864dcda09e067e4784774d48cffc18bb30899ef6435fdec875298" exitCode=0 Oct 14 07:24:26 crc kubenswrapper[4870]: I1014 07:24:26.677670 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qm9qp" event={"ID":"e32ab0b1-c28f-4d01-9b07-e1f0963a53cc","Type":"ContainerDied","Data":"fd5a77e4d86864dcda09e067e4784774d48cffc18bb30899ef6435fdec875298"} Oct 14 07:24:26 crc kubenswrapper[4870]: I1014 07:24:26.677730 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qm9qp" event={"ID":"e32ab0b1-c28f-4d01-9b07-e1f0963a53cc","Type":"ContainerStarted","Data":"cc720749673811cc2d8f9e0efff3c1a7a51ae049ef20299c01c9b94276211966"} Oct 14 07:24:26 crc kubenswrapper[4870]: I1014 07:24:26.679610 4870 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 14 07:24:27 crc kubenswrapper[4870]: I1014 07:24:27.690378 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qm9qp" event={"ID":"e32ab0b1-c28f-4d01-9b07-e1f0963a53cc","Type":"ContainerStarted","Data":"e18bf216d97e6578d131eabde3066737b5f66b7a5951c8380dde012e398782e7"} Oct 14 07:24:30 crc kubenswrapper[4870]: I1014 07:24:30.732618 4870 generic.go:334] "Generic (PLEG): container finished" podID="e32ab0b1-c28f-4d01-9b07-e1f0963a53cc" containerID="e18bf216d97e6578d131eabde3066737b5f66b7a5951c8380dde012e398782e7" exitCode=0 Oct 14 07:24:30 crc kubenswrapper[4870]: I1014 07:24:30.732786 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qm9qp" event={"ID":"e32ab0b1-c28f-4d01-9b07-e1f0963a53cc","Type":"ContainerDied","Data":"e18bf216d97e6578d131eabde3066737b5f66b7a5951c8380dde012e398782e7"} Oct 14 07:24:31 crc kubenswrapper[4870]: I1014 07:24:31.760916 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qm9qp" event={"ID":"e32ab0b1-c28f-4d01-9b07-e1f0963a53cc","Type":"ContainerStarted","Data":"ef038290196d8c055db8c2c98a122600289cfe5d79287793bc3f754001d94c19"} Oct 14 07:24:31 crc kubenswrapper[4870]: I1014 07:24:31.812229 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-qm9qp" podStartSLOduration=2.304241889 podStartE2EDuration="6.812191725s" podCreationTimestamp="2025-10-14 07:24:25 +0000 UTC" firstStartedPulling="2025-10-14 07:24:26.679393271 +0000 UTC m=+1402.376753642" lastFinishedPulling="2025-10-14 07:24:31.187343077 +0000 UTC m=+1406.884703478" observedRunningTime="2025-10-14 07:24:31.79303079 +0000 UTC m=+1407.490391161" watchObservedRunningTime="2025-10-14 07:24:31.812191725 +0000 UTC m=+1407.509552126" Oct 14 07:24:35 crc kubenswrapper[4870]: I1014 07:24:35.862495 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-qm9qp" Oct 14 07:24:35 crc kubenswrapper[4870]: I1014 07:24:35.863513 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-qm9qp" Oct 14 07:24:36 crc kubenswrapper[4870]: I1014 07:24:36.045821 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-7pfsx"] Oct 14 07:24:36 crc kubenswrapper[4870]: I1014 07:24:36.047198 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7pfsx" Oct 14 07:24:36 crc kubenswrapper[4870]: I1014 07:24:36.065903 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-7pfsx"] Oct 14 07:24:36 crc kubenswrapper[4870]: I1014 07:24:36.101983 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c0a12284-835c-4d00-9d95-095f282df8a2-utilities\") pod \"community-operators-7pfsx\" (UID: \"c0a12284-835c-4d00-9d95-095f282df8a2\") " pod="openshift-marketplace/community-operators-7pfsx" Oct 14 07:24:36 crc kubenswrapper[4870]: I1014 07:24:36.102609 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c0a12284-835c-4d00-9d95-095f282df8a2-catalog-content\") pod \"community-operators-7pfsx\" (UID: \"c0a12284-835c-4d00-9d95-095f282df8a2\") " pod="openshift-marketplace/community-operators-7pfsx" Oct 14 07:24:36 crc kubenswrapper[4870]: I1014 07:24:36.102930 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b28lb\" (UniqueName: \"kubernetes.io/projected/c0a12284-835c-4d00-9d95-095f282df8a2-kube-api-access-b28lb\") pod \"community-operators-7pfsx\" (UID: \"c0a12284-835c-4d00-9d95-095f282df8a2\") " pod="openshift-marketplace/community-operators-7pfsx" Oct 14 07:24:36 crc kubenswrapper[4870]: I1014 07:24:36.205644 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c0a12284-835c-4d00-9d95-095f282df8a2-catalog-content\") pod \"community-operators-7pfsx\" (UID: \"c0a12284-835c-4d00-9d95-095f282df8a2\") " pod="openshift-marketplace/community-operators-7pfsx" Oct 14 07:24:36 crc kubenswrapper[4870]: I1014 07:24:36.205714 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b28lb\" (UniqueName: \"kubernetes.io/projected/c0a12284-835c-4d00-9d95-095f282df8a2-kube-api-access-b28lb\") pod \"community-operators-7pfsx\" (UID: \"c0a12284-835c-4d00-9d95-095f282df8a2\") " pod="openshift-marketplace/community-operators-7pfsx" Oct 14 07:24:36 crc kubenswrapper[4870]: I1014 07:24:36.205818 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c0a12284-835c-4d00-9d95-095f282df8a2-utilities\") pod \"community-operators-7pfsx\" (UID: \"c0a12284-835c-4d00-9d95-095f282df8a2\") " pod="openshift-marketplace/community-operators-7pfsx" Oct 14 07:24:36 crc kubenswrapper[4870]: I1014 07:24:36.206342 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c0a12284-835c-4d00-9d95-095f282df8a2-catalog-content\") pod \"community-operators-7pfsx\" (UID: \"c0a12284-835c-4d00-9d95-095f282df8a2\") " pod="openshift-marketplace/community-operators-7pfsx" Oct 14 07:24:36 crc kubenswrapper[4870]: I1014 07:24:36.206416 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c0a12284-835c-4d00-9d95-095f282df8a2-utilities\") pod \"community-operators-7pfsx\" (UID: \"c0a12284-835c-4d00-9d95-095f282df8a2\") " pod="openshift-marketplace/community-operators-7pfsx" Oct 14 07:24:36 crc kubenswrapper[4870]: I1014 07:24:36.228108 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b28lb\" (UniqueName: \"kubernetes.io/projected/c0a12284-835c-4d00-9d95-095f282df8a2-kube-api-access-b28lb\") pod \"community-operators-7pfsx\" (UID: \"c0a12284-835c-4d00-9d95-095f282df8a2\") " pod="openshift-marketplace/community-operators-7pfsx" Oct 14 07:24:36 crc kubenswrapper[4870]: I1014 07:24:36.423520 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7pfsx" Oct 14 07:24:36 crc kubenswrapper[4870]: I1014 07:24:36.743996 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-7pfsx"] Oct 14 07:24:36 crc kubenswrapper[4870]: I1014 07:24:36.842385 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7pfsx" event={"ID":"c0a12284-835c-4d00-9d95-095f282df8a2","Type":"ContainerStarted","Data":"1d75776e40b84764d43f6200436743fbfc7843daf8b9013489cc52d11b199576"} Oct 14 07:24:36 crc kubenswrapper[4870]: I1014 07:24:36.961615 4870 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-qm9qp" podUID="e32ab0b1-c28f-4d01-9b07-e1f0963a53cc" containerName="registry-server" probeResult="failure" output=< Oct 14 07:24:36 crc kubenswrapper[4870]: timeout: failed to connect service ":50051" within 1s Oct 14 07:24:36 crc kubenswrapper[4870]: > Oct 14 07:24:37 crc kubenswrapper[4870]: I1014 07:24:37.855870 4870 generic.go:334] "Generic (PLEG): container finished" podID="c0a12284-835c-4d00-9d95-095f282df8a2" containerID="bb3faed94690768d73d92e92692fa460871e516e331988f9227a928626397a7d" exitCode=0 Oct 14 07:24:37 crc kubenswrapper[4870]: I1014 07:24:37.855915 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7pfsx" event={"ID":"c0a12284-835c-4d00-9d95-095f282df8a2","Type":"ContainerDied","Data":"bb3faed94690768d73d92e92692fa460871e516e331988f9227a928626397a7d"} Oct 14 07:24:38 crc kubenswrapper[4870]: I1014 07:24:38.869144 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7pfsx" event={"ID":"c0a12284-835c-4d00-9d95-095f282df8a2","Type":"ContainerStarted","Data":"d7f9b6fe350ea98afbe743938ca3bf64228279b4b5c0024ddf8a9a9b72a6154a"} Oct 14 07:24:39 crc kubenswrapper[4870]: I1014 07:24:39.885422 4870 generic.go:334] "Generic (PLEG): container finished" podID="c0a12284-835c-4d00-9d95-095f282df8a2" containerID="d7f9b6fe350ea98afbe743938ca3bf64228279b4b5c0024ddf8a9a9b72a6154a" exitCode=0 Oct 14 07:24:39 crc kubenswrapper[4870]: I1014 07:24:39.885515 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7pfsx" event={"ID":"c0a12284-835c-4d00-9d95-095f282df8a2","Type":"ContainerDied","Data":"d7f9b6fe350ea98afbe743938ca3bf64228279b4b5c0024ddf8a9a9b72a6154a"} Oct 14 07:24:40 crc kubenswrapper[4870]: I1014 07:24:40.902775 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7pfsx" event={"ID":"c0a12284-835c-4d00-9d95-095f282df8a2","Type":"ContainerStarted","Data":"9787cb76b4aa69a95f7f0c9b633f09d3bae38901849e4e4c9600e17d8c7f71d2"} Oct 14 07:24:40 crc kubenswrapper[4870]: I1014 07:24:40.939242 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-7pfsx" podStartSLOduration=2.237105145 podStartE2EDuration="4.939217066s" podCreationTimestamp="2025-10-14 07:24:36 +0000 UTC" firstStartedPulling="2025-10-14 07:24:37.857962351 +0000 UTC m=+1413.555322732" lastFinishedPulling="2025-10-14 07:24:40.560074272 +0000 UTC m=+1416.257434653" observedRunningTime="2025-10-14 07:24:40.932913363 +0000 UTC m=+1416.630273774" watchObservedRunningTime="2025-10-14 07:24:40.939217066 +0000 UTC m=+1416.636577467" Oct 14 07:24:45 crc kubenswrapper[4870]: I1014 07:24:45.945822 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-qm9qp" Oct 14 07:24:46 crc kubenswrapper[4870]: I1014 07:24:46.009546 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-qm9qp" Oct 14 07:24:46 crc kubenswrapper[4870]: I1014 07:24:46.200378 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-qm9qp"] Oct 14 07:24:46 crc kubenswrapper[4870]: I1014 07:24:46.424901 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-7pfsx" Oct 14 07:24:46 crc kubenswrapper[4870]: I1014 07:24:46.424995 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-7pfsx" Oct 14 07:24:46 crc kubenswrapper[4870]: I1014 07:24:46.493313 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-7pfsx" Oct 14 07:24:46 crc kubenswrapper[4870]: I1014 07:24:46.979318 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-qm9qp" podUID="e32ab0b1-c28f-4d01-9b07-e1f0963a53cc" containerName="registry-server" containerID="cri-o://ef038290196d8c055db8c2c98a122600289cfe5d79287793bc3f754001d94c19" gracePeriod=2 Oct 14 07:24:47 crc kubenswrapper[4870]: I1014 07:24:47.058899 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-7pfsx" Oct 14 07:24:47 crc kubenswrapper[4870]: I1014 07:24:47.556546 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qm9qp" Oct 14 07:24:47 crc kubenswrapper[4870]: I1014 07:24:47.647839 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xnv98\" (UniqueName: \"kubernetes.io/projected/e32ab0b1-c28f-4d01-9b07-e1f0963a53cc-kube-api-access-xnv98\") pod \"e32ab0b1-c28f-4d01-9b07-e1f0963a53cc\" (UID: \"e32ab0b1-c28f-4d01-9b07-e1f0963a53cc\") " Oct 14 07:24:47 crc kubenswrapper[4870]: I1014 07:24:47.647921 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e32ab0b1-c28f-4d01-9b07-e1f0963a53cc-catalog-content\") pod \"e32ab0b1-c28f-4d01-9b07-e1f0963a53cc\" (UID: \"e32ab0b1-c28f-4d01-9b07-e1f0963a53cc\") " Oct 14 07:24:47 crc kubenswrapper[4870]: I1014 07:24:47.647995 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e32ab0b1-c28f-4d01-9b07-e1f0963a53cc-utilities\") pod \"e32ab0b1-c28f-4d01-9b07-e1f0963a53cc\" (UID: \"e32ab0b1-c28f-4d01-9b07-e1f0963a53cc\") " Oct 14 07:24:47 crc kubenswrapper[4870]: I1014 07:24:47.648874 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e32ab0b1-c28f-4d01-9b07-e1f0963a53cc-utilities" (OuterVolumeSpecName: "utilities") pod "e32ab0b1-c28f-4d01-9b07-e1f0963a53cc" (UID: "e32ab0b1-c28f-4d01-9b07-e1f0963a53cc"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:24:47 crc kubenswrapper[4870]: I1014 07:24:47.652963 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e32ab0b1-c28f-4d01-9b07-e1f0963a53cc-kube-api-access-xnv98" (OuterVolumeSpecName: "kube-api-access-xnv98") pod "e32ab0b1-c28f-4d01-9b07-e1f0963a53cc" (UID: "e32ab0b1-c28f-4d01-9b07-e1f0963a53cc"). InnerVolumeSpecName "kube-api-access-xnv98". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:24:47 crc kubenswrapper[4870]: I1014 07:24:47.746308 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e32ab0b1-c28f-4d01-9b07-e1f0963a53cc-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e32ab0b1-c28f-4d01-9b07-e1f0963a53cc" (UID: "e32ab0b1-c28f-4d01-9b07-e1f0963a53cc"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:24:47 crc kubenswrapper[4870]: I1014 07:24:47.749650 4870 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e32ab0b1-c28f-4d01-9b07-e1f0963a53cc-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 07:24:47 crc kubenswrapper[4870]: I1014 07:24:47.749710 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xnv98\" (UniqueName: \"kubernetes.io/projected/e32ab0b1-c28f-4d01-9b07-e1f0963a53cc-kube-api-access-xnv98\") on node \"crc\" DevicePath \"\"" Oct 14 07:24:47 crc kubenswrapper[4870]: I1014 07:24:47.749733 4870 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e32ab0b1-c28f-4d01-9b07-e1f0963a53cc-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 07:24:47 crc kubenswrapper[4870]: I1014 07:24:47.997035 4870 generic.go:334] "Generic (PLEG): container finished" podID="e32ab0b1-c28f-4d01-9b07-e1f0963a53cc" containerID="ef038290196d8c055db8c2c98a122600289cfe5d79287793bc3f754001d94c19" exitCode=0 Oct 14 07:24:47 crc kubenswrapper[4870]: I1014 07:24:47.997170 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qm9qp" Oct 14 07:24:47 crc kubenswrapper[4870]: I1014 07:24:47.997194 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qm9qp" event={"ID":"e32ab0b1-c28f-4d01-9b07-e1f0963a53cc","Type":"ContainerDied","Data":"ef038290196d8c055db8c2c98a122600289cfe5d79287793bc3f754001d94c19"} Oct 14 07:24:47 crc kubenswrapper[4870]: I1014 07:24:47.997284 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qm9qp" event={"ID":"e32ab0b1-c28f-4d01-9b07-e1f0963a53cc","Type":"ContainerDied","Data":"cc720749673811cc2d8f9e0efff3c1a7a51ae049ef20299c01c9b94276211966"} Oct 14 07:24:47 crc kubenswrapper[4870]: I1014 07:24:47.997332 4870 scope.go:117] "RemoveContainer" containerID="ef038290196d8c055db8c2c98a122600289cfe5d79287793bc3f754001d94c19" Oct 14 07:24:48 crc kubenswrapper[4870]: I1014 07:24:48.034489 4870 scope.go:117] "RemoveContainer" containerID="e18bf216d97e6578d131eabde3066737b5f66b7a5951c8380dde012e398782e7" Oct 14 07:24:48 crc kubenswrapper[4870]: I1014 07:24:48.053140 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-qm9qp"] Oct 14 07:24:48 crc kubenswrapper[4870]: I1014 07:24:48.062470 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-qm9qp"] Oct 14 07:24:48 crc kubenswrapper[4870]: I1014 07:24:48.091110 4870 scope.go:117] "RemoveContainer" containerID="fd5a77e4d86864dcda09e067e4784774d48cffc18bb30899ef6435fdec875298" Oct 14 07:24:48 crc kubenswrapper[4870]: I1014 07:24:48.122197 4870 scope.go:117] "RemoveContainer" containerID="ef038290196d8c055db8c2c98a122600289cfe5d79287793bc3f754001d94c19" Oct 14 07:24:48 crc kubenswrapper[4870]: E1014 07:24:48.123134 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ef038290196d8c055db8c2c98a122600289cfe5d79287793bc3f754001d94c19\": container with ID starting with ef038290196d8c055db8c2c98a122600289cfe5d79287793bc3f754001d94c19 not found: ID does not exist" containerID="ef038290196d8c055db8c2c98a122600289cfe5d79287793bc3f754001d94c19" Oct 14 07:24:48 crc kubenswrapper[4870]: I1014 07:24:48.123191 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ef038290196d8c055db8c2c98a122600289cfe5d79287793bc3f754001d94c19"} err="failed to get container status \"ef038290196d8c055db8c2c98a122600289cfe5d79287793bc3f754001d94c19\": rpc error: code = NotFound desc = could not find container \"ef038290196d8c055db8c2c98a122600289cfe5d79287793bc3f754001d94c19\": container with ID starting with ef038290196d8c055db8c2c98a122600289cfe5d79287793bc3f754001d94c19 not found: ID does not exist" Oct 14 07:24:48 crc kubenswrapper[4870]: I1014 07:24:48.123227 4870 scope.go:117] "RemoveContainer" containerID="e18bf216d97e6578d131eabde3066737b5f66b7a5951c8380dde012e398782e7" Oct 14 07:24:48 crc kubenswrapper[4870]: E1014 07:24:48.123981 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e18bf216d97e6578d131eabde3066737b5f66b7a5951c8380dde012e398782e7\": container with ID starting with e18bf216d97e6578d131eabde3066737b5f66b7a5951c8380dde012e398782e7 not found: ID does not exist" containerID="e18bf216d97e6578d131eabde3066737b5f66b7a5951c8380dde012e398782e7" Oct 14 07:24:48 crc kubenswrapper[4870]: I1014 07:24:48.124048 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e18bf216d97e6578d131eabde3066737b5f66b7a5951c8380dde012e398782e7"} err="failed to get container status \"e18bf216d97e6578d131eabde3066737b5f66b7a5951c8380dde012e398782e7\": rpc error: code = NotFound desc = could not find container \"e18bf216d97e6578d131eabde3066737b5f66b7a5951c8380dde012e398782e7\": container with ID starting with e18bf216d97e6578d131eabde3066737b5f66b7a5951c8380dde012e398782e7 not found: ID does not exist" Oct 14 07:24:48 crc kubenswrapper[4870]: I1014 07:24:48.124092 4870 scope.go:117] "RemoveContainer" containerID="fd5a77e4d86864dcda09e067e4784774d48cffc18bb30899ef6435fdec875298" Oct 14 07:24:48 crc kubenswrapper[4870]: E1014 07:24:48.124951 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fd5a77e4d86864dcda09e067e4784774d48cffc18bb30899ef6435fdec875298\": container with ID starting with fd5a77e4d86864dcda09e067e4784774d48cffc18bb30899ef6435fdec875298 not found: ID does not exist" containerID="fd5a77e4d86864dcda09e067e4784774d48cffc18bb30899ef6435fdec875298" Oct 14 07:24:48 crc kubenswrapper[4870]: I1014 07:24:48.125000 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fd5a77e4d86864dcda09e067e4784774d48cffc18bb30899ef6435fdec875298"} err="failed to get container status \"fd5a77e4d86864dcda09e067e4784774d48cffc18bb30899ef6435fdec875298\": rpc error: code = NotFound desc = could not find container \"fd5a77e4d86864dcda09e067e4784774d48cffc18bb30899ef6435fdec875298\": container with ID starting with fd5a77e4d86864dcda09e067e4784774d48cffc18bb30899ef6435fdec875298 not found: ID does not exist" Oct 14 07:24:48 crc kubenswrapper[4870]: I1014 07:24:48.793296 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-7pfsx"] Oct 14 07:24:49 crc kubenswrapper[4870]: I1014 07:24:49.030791 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-7pfsx" podUID="c0a12284-835c-4d00-9d95-095f282df8a2" containerName="registry-server" containerID="cri-o://9787cb76b4aa69a95f7f0c9b633f09d3bae38901849e4e4c9600e17d8c7f71d2" gracePeriod=2 Oct 14 07:24:49 crc kubenswrapper[4870]: I1014 07:24:49.050806 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e32ab0b1-c28f-4d01-9b07-e1f0963a53cc" path="/var/lib/kubelet/pods/e32ab0b1-c28f-4d01-9b07-e1f0963a53cc/volumes" Oct 14 07:24:49 crc kubenswrapper[4870]: I1014 07:24:49.538127 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7pfsx" Oct 14 07:24:49 crc kubenswrapper[4870]: I1014 07:24:49.683140 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b28lb\" (UniqueName: \"kubernetes.io/projected/c0a12284-835c-4d00-9d95-095f282df8a2-kube-api-access-b28lb\") pod \"c0a12284-835c-4d00-9d95-095f282df8a2\" (UID: \"c0a12284-835c-4d00-9d95-095f282df8a2\") " Oct 14 07:24:49 crc kubenswrapper[4870]: I1014 07:24:49.683289 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c0a12284-835c-4d00-9d95-095f282df8a2-utilities\") pod \"c0a12284-835c-4d00-9d95-095f282df8a2\" (UID: \"c0a12284-835c-4d00-9d95-095f282df8a2\") " Oct 14 07:24:49 crc kubenswrapper[4870]: I1014 07:24:49.683342 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c0a12284-835c-4d00-9d95-095f282df8a2-catalog-content\") pod \"c0a12284-835c-4d00-9d95-095f282df8a2\" (UID: \"c0a12284-835c-4d00-9d95-095f282df8a2\") " Oct 14 07:24:49 crc kubenswrapper[4870]: I1014 07:24:49.685408 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c0a12284-835c-4d00-9d95-095f282df8a2-utilities" (OuterVolumeSpecName: "utilities") pod "c0a12284-835c-4d00-9d95-095f282df8a2" (UID: "c0a12284-835c-4d00-9d95-095f282df8a2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:24:49 crc kubenswrapper[4870]: I1014 07:24:49.691995 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c0a12284-835c-4d00-9d95-095f282df8a2-kube-api-access-b28lb" (OuterVolumeSpecName: "kube-api-access-b28lb") pod "c0a12284-835c-4d00-9d95-095f282df8a2" (UID: "c0a12284-835c-4d00-9d95-095f282df8a2"). InnerVolumeSpecName "kube-api-access-b28lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:24:49 crc kubenswrapper[4870]: I1014 07:24:49.786515 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b28lb\" (UniqueName: \"kubernetes.io/projected/c0a12284-835c-4d00-9d95-095f282df8a2-kube-api-access-b28lb\") on node \"crc\" DevicePath \"\"" Oct 14 07:24:49 crc kubenswrapper[4870]: I1014 07:24:49.786562 4870 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c0a12284-835c-4d00-9d95-095f282df8a2-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 07:24:49 crc kubenswrapper[4870]: I1014 07:24:49.817194 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c0a12284-835c-4d00-9d95-095f282df8a2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c0a12284-835c-4d00-9d95-095f282df8a2" (UID: "c0a12284-835c-4d00-9d95-095f282df8a2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:24:49 crc kubenswrapper[4870]: I1014 07:24:49.889025 4870 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c0a12284-835c-4d00-9d95-095f282df8a2-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 07:24:50 crc kubenswrapper[4870]: I1014 07:24:50.048498 4870 generic.go:334] "Generic (PLEG): container finished" podID="c0a12284-835c-4d00-9d95-095f282df8a2" containerID="9787cb76b4aa69a95f7f0c9b633f09d3bae38901849e4e4c9600e17d8c7f71d2" exitCode=0 Oct 14 07:24:50 crc kubenswrapper[4870]: I1014 07:24:50.048586 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7pfsx" event={"ID":"c0a12284-835c-4d00-9d95-095f282df8a2","Type":"ContainerDied","Data":"9787cb76b4aa69a95f7f0c9b633f09d3bae38901849e4e4c9600e17d8c7f71d2"} Oct 14 07:24:50 crc kubenswrapper[4870]: I1014 07:24:50.048629 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7pfsx" Oct 14 07:24:50 crc kubenswrapper[4870]: I1014 07:24:50.048656 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7pfsx" event={"ID":"c0a12284-835c-4d00-9d95-095f282df8a2","Type":"ContainerDied","Data":"1d75776e40b84764d43f6200436743fbfc7843daf8b9013489cc52d11b199576"} Oct 14 07:24:50 crc kubenswrapper[4870]: I1014 07:24:50.048693 4870 scope.go:117] "RemoveContainer" containerID="9787cb76b4aa69a95f7f0c9b633f09d3bae38901849e4e4c9600e17d8c7f71d2" Oct 14 07:24:50 crc kubenswrapper[4870]: I1014 07:24:50.107458 4870 scope.go:117] "RemoveContainer" containerID="d7f9b6fe350ea98afbe743938ca3bf64228279b4b5c0024ddf8a9a9b72a6154a" Oct 14 07:24:50 crc kubenswrapper[4870]: I1014 07:24:50.122774 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-7pfsx"] Oct 14 07:24:50 crc kubenswrapper[4870]: I1014 07:24:50.123496 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-7pfsx"] Oct 14 07:24:50 crc kubenswrapper[4870]: I1014 07:24:50.160311 4870 scope.go:117] "RemoveContainer" containerID="bb3faed94690768d73d92e92692fa460871e516e331988f9227a928626397a7d" Oct 14 07:24:50 crc kubenswrapper[4870]: I1014 07:24:50.195905 4870 scope.go:117] "RemoveContainer" containerID="9787cb76b4aa69a95f7f0c9b633f09d3bae38901849e4e4c9600e17d8c7f71d2" Oct 14 07:24:50 crc kubenswrapper[4870]: E1014 07:24:50.196647 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9787cb76b4aa69a95f7f0c9b633f09d3bae38901849e4e4c9600e17d8c7f71d2\": container with ID starting with 9787cb76b4aa69a95f7f0c9b633f09d3bae38901849e4e4c9600e17d8c7f71d2 not found: ID does not exist" containerID="9787cb76b4aa69a95f7f0c9b633f09d3bae38901849e4e4c9600e17d8c7f71d2" Oct 14 07:24:50 crc kubenswrapper[4870]: I1014 07:24:50.196702 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9787cb76b4aa69a95f7f0c9b633f09d3bae38901849e4e4c9600e17d8c7f71d2"} err="failed to get container status \"9787cb76b4aa69a95f7f0c9b633f09d3bae38901849e4e4c9600e17d8c7f71d2\": rpc error: code = NotFound desc = could not find container \"9787cb76b4aa69a95f7f0c9b633f09d3bae38901849e4e4c9600e17d8c7f71d2\": container with ID starting with 9787cb76b4aa69a95f7f0c9b633f09d3bae38901849e4e4c9600e17d8c7f71d2 not found: ID does not exist" Oct 14 07:24:50 crc kubenswrapper[4870]: I1014 07:24:50.196736 4870 scope.go:117] "RemoveContainer" containerID="d7f9b6fe350ea98afbe743938ca3bf64228279b4b5c0024ddf8a9a9b72a6154a" Oct 14 07:24:50 crc kubenswrapper[4870]: E1014 07:24:50.197229 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d7f9b6fe350ea98afbe743938ca3bf64228279b4b5c0024ddf8a9a9b72a6154a\": container with ID starting with d7f9b6fe350ea98afbe743938ca3bf64228279b4b5c0024ddf8a9a9b72a6154a not found: ID does not exist" containerID="d7f9b6fe350ea98afbe743938ca3bf64228279b4b5c0024ddf8a9a9b72a6154a" Oct 14 07:24:50 crc kubenswrapper[4870]: I1014 07:24:50.197282 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d7f9b6fe350ea98afbe743938ca3bf64228279b4b5c0024ddf8a9a9b72a6154a"} err="failed to get container status \"d7f9b6fe350ea98afbe743938ca3bf64228279b4b5c0024ddf8a9a9b72a6154a\": rpc error: code = NotFound desc = could not find container \"d7f9b6fe350ea98afbe743938ca3bf64228279b4b5c0024ddf8a9a9b72a6154a\": container with ID starting with d7f9b6fe350ea98afbe743938ca3bf64228279b4b5c0024ddf8a9a9b72a6154a not found: ID does not exist" Oct 14 07:24:50 crc kubenswrapper[4870]: I1014 07:24:50.197320 4870 scope.go:117] "RemoveContainer" containerID="bb3faed94690768d73d92e92692fa460871e516e331988f9227a928626397a7d" Oct 14 07:24:50 crc kubenswrapper[4870]: E1014 07:24:50.197719 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bb3faed94690768d73d92e92692fa460871e516e331988f9227a928626397a7d\": container with ID starting with bb3faed94690768d73d92e92692fa460871e516e331988f9227a928626397a7d not found: ID does not exist" containerID="bb3faed94690768d73d92e92692fa460871e516e331988f9227a928626397a7d" Oct 14 07:24:50 crc kubenswrapper[4870]: I1014 07:24:50.197750 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bb3faed94690768d73d92e92692fa460871e516e331988f9227a928626397a7d"} err="failed to get container status \"bb3faed94690768d73d92e92692fa460871e516e331988f9227a928626397a7d\": rpc error: code = NotFound desc = could not find container \"bb3faed94690768d73d92e92692fa460871e516e331988f9227a928626397a7d\": container with ID starting with bb3faed94690768d73d92e92692fa460871e516e331988f9227a928626397a7d not found: ID does not exist" Oct 14 07:24:51 crc kubenswrapper[4870]: I1014 07:24:51.053880 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c0a12284-835c-4d00-9d95-095f282df8a2" path="/var/lib/kubelet/pods/c0a12284-835c-4d00-9d95-095f282df8a2/volumes" Oct 14 07:24:53 crc kubenswrapper[4870]: I1014 07:24:53.950741 4870 patch_prober.go:28] interesting pod/machine-config-daemon-7tvc7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 07:24:53 crc kubenswrapper[4870]: I1014 07:24:53.951285 4870 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 07:25:09 crc kubenswrapper[4870]: I1014 07:25:09.200089 4870 scope.go:117] "RemoveContainer" containerID="9389d35044ef25b0e127f63d901f00d3d48e58573671b63fa03de1238628ba21" Oct 14 07:25:09 crc kubenswrapper[4870]: I1014 07:25:09.261861 4870 scope.go:117] "RemoveContainer" containerID="13f6a7296cd23cf0aa5a1d5c03164c9801caa1ced55914f7f42966f05f4d9aa7" Oct 14 07:25:11 crc kubenswrapper[4870]: I1014 07:25:11.520407 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-sz58s"] Oct 14 07:25:11 crc kubenswrapper[4870]: E1014 07:25:11.521601 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c0a12284-835c-4d00-9d95-095f282df8a2" containerName="registry-server" Oct 14 07:25:11 crc kubenswrapper[4870]: I1014 07:25:11.521622 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="c0a12284-835c-4d00-9d95-095f282df8a2" containerName="registry-server" Oct 14 07:25:11 crc kubenswrapper[4870]: E1014 07:25:11.521656 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e32ab0b1-c28f-4d01-9b07-e1f0963a53cc" containerName="extract-content" Oct 14 07:25:11 crc kubenswrapper[4870]: I1014 07:25:11.521663 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="e32ab0b1-c28f-4d01-9b07-e1f0963a53cc" containerName="extract-content" Oct 14 07:25:11 crc kubenswrapper[4870]: E1014 07:25:11.521676 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c0a12284-835c-4d00-9d95-095f282df8a2" containerName="extract-utilities" Oct 14 07:25:11 crc kubenswrapper[4870]: I1014 07:25:11.521685 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="c0a12284-835c-4d00-9d95-095f282df8a2" containerName="extract-utilities" Oct 14 07:25:11 crc kubenswrapper[4870]: E1014 07:25:11.521705 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e32ab0b1-c28f-4d01-9b07-e1f0963a53cc" containerName="extract-utilities" Oct 14 07:25:11 crc kubenswrapper[4870]: I1014 07:25:11.521712 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="e32ab0b1-c28f-4d01-9b07-e1f0963a53cc" containerName="extract-utilities" Oct 14 07:25:11 crc kubenswrapper[4870]: E1014 07:25:11.521721 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e32ab0b1-c28f-4d01-9b07-e1f0963a53cc" containerName="registry-server" Oct 14 07:25:11 crc kubenswrapper[4870]: I1014 07:25:11.521728 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="e32ab0b1-c28f-4d01-9b07-e1f0963a53cc" containerName="registry-server" Oct 14 07:25:11 crc kubenswrapper[4870]: E1014 07:25:11.521738 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c0a12284-835c-4d00-9d95-095f282df8a2" containerName="extract-content" Oct 14 07:25:11 crc kubenswrapper[4870]: I1014 07:25:11.521744 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="c0a12284-835c-4d00-9d95-095f282df8a2" containerName="extract-content" Oct 14 07:25:11 crc kubenswrapper[4870]: I1014 07:25:11.521893 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="c0a12284-835c-4d00-9d95-095f282df8a2" containerName="registry-server" Oct 14 07:25:11 crc kubenswrapper[4870]: I1014 07:25:11.521917 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="e32ab0b1-c28f-4d01-9b07-e1f0963a53cc" containerName="registry-server" Oct 14 07:25:11 crc kubenswrapper[4870]: I1014 07:25:11.523137 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-sz58s" Oct 14 07:25:11 crc kubenswrapper[4870]: I1014 07:25:11.542597 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-sz58s"] Oct 14 07:25:11 crc kubenswrapper[4870]: I1014 07:25:11.661368 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7382409f-f523-4017-9bc9-80661712ff92-catalog-content\") pod \"certified-operators-sz58s\" (UID: \"7382409f-f523-4017-9bc9-80661712ff92\") " pod="openshift-marketplace/certified-operators-sz58s" Oct 14 07:25:11 crc kubenswrapper[4870]: I1014 07:25:11.661450 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dk5xg\" (UniqueName: \"kubernetes.io/projected/7382409f-f523-4017-9bc9-80661712ff92-kube-api-access-dk5xg\") pod \"certified-operators-sz58s\" (UID: \"7382409f-f523-4017-9bc9-80661712ff92\") " pod="openshift-marketplace/certified-operators-sz58s" Oct 14 07:25:11 crc kubenswrapper[4870]: I1014 07:25:11.661479 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7382409f-f523-4017-9bc9-80661712ff92-utilities\") pod \"certified-operators-sz58s\" (UID: \"7382409f-f523-4017-9bc9-80661712ff92\") " pod="openshift-marketplace/certified-operators-sz58s" Oct 14 07:25:11 crc kubenswrapper[4870]: I1014 07:25:11.763043 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7382409f-f523-4017-9bc9-80661712ff92-catalog-content\") pod \"certified-operators-sz58s\" (UID: \"7382409f-f523-4017-9bc9-80661712ff92\") " pod="openshift-marketplace/certified-operators-sz58s" Oct 14 07:25:11 crc kubenswrapper[4870]: I1014 07:25:11.763130 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dk5xg\" (UniqueName: \"kubernetes.io/projected/7382409f-f523-4017-9bc9-80661712ff92-kube-api-access-dk5xg\") pod \"certified-operators-sz58s\" (UID: \"7382409f-f523-4017-9bc9-80661712ff92\") " pod="openshift-marketplace/certified-operators-sz58s" Oct 14 07:25:11 crc kubenswrapper[4870]: I1014 07:25:11.763170 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7382409f-f523-4017-9bc9-80661712ff92-utilities\") pod \"certified-operators-sz58s\" (UID: \"7382409f-f523-4017-9bc9-80661712ff92\") " pod="openshift-marketplace/certified-operators-sz58s" Oct 14 07:25:11 crc kubenswrapper[4870]: I1014 07:25:11.763848 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7382409f-f523-4017-9bc9-80661712ff92-utilities\") pod \"certified-operators-sz58s\" (UID: \"7382409f-f523-4017-9bc9-80661712ff92\") " pod="openshift-marketplace/certified-operators-sz58s" Oct 14 07:25:11 crc kubenswrapper[4870]: I1014 07:25:11.764281 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7382409f-f523-4017-9bc9-80661712ff92-catalog-content\") pod \"certified-operators-sz58s\" (UID: \"7382409f-f523-4017-9bc9-80661712ff92\") " pod="openshift-marketplace/certified-operators-sz58s" Oct 14 07:25:11 crc kubenswrapper[4870]: I1014 07:25:11.789627 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dk5xg\" (UniqueName: \"kubernetes.io/projected/7382409f-f523-4017-9bc9-80661712ff92-kube-api-access-dk5xg\") pod \"certified-operators-sz58s\" (UID: \"7382409f-f523-4017-9bc9-80661712ff92\") " pod="openshift-marketplace/certified-operators-sz58s" Oct 14 07:25:11 crc kubenswrapper[4870]: I1014 07:25:11.847581 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-sz58s" Oct 14 07:25:12 crc kubenswrapper[4870]: I1014 07:25:12.325514 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-sz58s"] Oct 14 07:25:13 crc kubenswrapper[4870]: I1014 07:25:13.308754 4870 generic.go:334] "Generic (PLEG): container finished" podID="7382409f-f523-4017-9bc9-80661712ff92" containerID="3921b56a57b3359e4823aa8eb9791d4575ec9186fbd0cf145841ed8fb99886a9" exitCode=0 Oct 14 07:25:13 crc kubenswrapper[4870]: I1014 07:25:13.308826 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sz58s" event={"ID":"7382409f-f523-4017-9bc9-80661712ff92","Type":"ContainerDied","Data":"3921b56a57b3359e4823aa8eb9791d4575ec9186fbd0cf145841ed8fb99886a9"} Oct 14 07:25:13 crc kubenswrapper[4870]: I1014 07:25:13.308867 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sz58s" event={"ID":"7382409f-f523-4017-9bc9-80661712ff92","Type":"ContainerStarted","Data":"7cc1b268d4726c556d4b28735ddc4c3e665ac53c87cee8e15dda56d36cd48934"} Oct 14 07:25:14 crc kubenswrapper[4870]: I1014 07:25:14.320375 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sz58s" event={"ID":"7382409f-f523-4017-9bc9-80661712ff92","Type":"ContainerStarted","Data":"2f87670aabbfa190bb739856b391e28f08c9199aff4167eb658103bc10e66d85"} Oct 14 07:25:15 crc kubenswrapper[4870]: I1014 07:25:15.334010 4870 generic.go:334] "Generic (PLEG): container finished" podID="7382409f-f523-4017-9bc9-80661712ff92" containerID="2f87670aabbfa190bb739856b391e28f08c9199aff4167eb658103bc10e66d85" exitCode=0 Oct 14 07:25:15 crc kubenswrapper[4870]: I1014 07:25:15.334220 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sz58s" event={"ID":"7382409f-f523-4017-9bc9-80661712ff92","Type":"ContainerDied","Data":"2f87670aabbfa190bb739856b391e28f08c9199aff4167eb658103bc10e66d85"} Oct 14 07:25:16 crc kubenswrapper[4870]: I1014 07:25:16.346958 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sz58s" event={"ID":"7382409f-f523-4017-9bc9-80661712ff92","Type":"ContainerStarted","Data":"ce827dbd893c32c3b59311cb8d54cbff6a44d5d05b84f0777104d4792cd1a60e"} Oct 14 07:25:16 crc kubenswrapper[4870]: I1014 07:25:16.373452 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-sz58s" podStartSLOduration=2.85883885 podStartE2EDuration="5.373420529s" podCreationTimestamp="2025-10-14 07:25:11 +0000 UTC" firstStartedPulling="2025-10-14 07:25:13.314007325 +0000 UTC m=+1449.011367726" lastFinishedPulling="2025-10-14 07:25:15.828588994 +0000 UTC m=+1451.525949405" observedRunningTime="2025-10-14 07:25:16.368333316 +0000 UTC m=+1452.065693727" watchObservedRunningTime="2025-10-14 07:25:16.373420529 +0000 UTC m=+1452.070780900" Oct 14 07:25:16 crc kubenswrapper[4870]: I1014 07:25:16.902818 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-lpr6m"] Oct 14 07:25:16 crc kubenswrapper[4870]: I1014 07:25:16.906330 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lpr6m" Oct 14 07:25:16 crc kubenswrapper[4870]: I1014 07:25:16.923088 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-lpr6m"] Oct 14 07:25:17 crc kubenswrapper[4870]: I1014 07:25:17.045797 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ad74a576-9ce6-403d-a4d8-2e7f8ac20b0b-utilities\") pod \"redhat-marketplace-lpr6m\" (UID: \"ad74a576-9ce6-403d-a4d8-2e7f8ac20b0b\") " pod="openshift-marketplace/redhat-marketplace-lpr6m" Oct 14 07:25:17 crc kubenswrapper[4870]: I1014 07:25:17.045904 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-djxkj\" (UniqueName: \"kubernetes.io/projected/ad74a576-9ce6-403d-a4d8-2e7f8ac20b0b-kube-api-access-djxkj\") pod \"redhat-marketplace-lpr6m\" (UID: \"ad74a576-9ce6-403d-a4d8-2e7f8ac20b0b\") " pod="openshift-marketplace/redhat-marketplace-lpr6m" Oct 14 07:25:17 crc kubenswrapper[4870]: I1014 07:25:17.046145 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ad74a576-9ce6-403d-a4d8-2e7f8ac20b0b-catalog-content\") pod \"redhat-marketplace-lpr6m\" (UID: \"ad74a576-9ce6-403d-a4d8-2e7f8ac20b0b\") " pod="openshift-marketplace/redhat-marketplace-lpr6m" Oct 14 07:25:17 crc kubenswrapper[4870]: I1014 07:25:17.147803 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-djxkj\" (UniqueName: \"kubernetes.io/projected/ad74a576-9ce6-403d-a4d8-2e7f8ac20b0b-kube-api-access-djxkj\") pod \"redhat-marketplace-lpr6m\" (UID: \"ad74a576-9ce6-403d-a4d8-2e7f8ac20b0b\") " pod="openshift-marketplace/redhat-marketplace-lpr6m" Oct 14 07:25:17 crc kubenswrapper[4870]: I1014 07:25:17.148160 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ad74a576-9ce6-403d-a4d8-2e7f8ac20b0b-catalog-content\") pod \"redhat-marketplace-lpr6m\" (UID: \"ad74a576-9ce6-403d-a4d8-2e7f8ac20b0b\") " pod="openshift-marketplace/redhat-marketplace-lpr6m" Oct 14 07:25:17 crc kubenswrapper[4870]: I1014 07:25:17.148224 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ad74a576-9ce6-403d-a4d8-2e7f8ac20b0b-utilities\") pod \"redhat-marketplace-lpr6m\" (UID: \"ad74a576-9ce6-403d-a4d8-2e7f8ac20b0b\") " pod="openshift-marketplace/redhat-marketplace-lpr6m" Oct 14 07:25:17 crc kubenswrapper[4870]: I1014 07:25:17.148834 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ad74a576-9ce6-403d-a4d8-2e7f8ac20b0b-catalog-content\") pod \"redhat-marketplace-lpr6m\" (UID: \"ad74a576-9ce6-403d-a4d8-2e7f8ac20b0b\") " pod="openshift-marketplace/redhat-marketplace-lpr6m" Oct 14 07:25:17 crc kubenswrapper[4870]: I1014 07:25:17.148864 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ad74a576-9ce6-403d-a4d8-2e7f8ac20b0b-utilities\") pod \"redhat-marketplace-lpr6m\" (UID: \"ad74a576-9ce6-403d-a4d8-2e7f8ac20b0b\") " pod="openshift-marketplace/redhat-marketplace-lpr6m" Oct 14 07:25:17 crc kubenswrapper[4870]: I1014 07:25:17.179793 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-djxkj\" (UniqueName: \"kubernetes.io/projected/ad74a576-9ce6-403d-a4d8-2e7f8ac20b0b-kube-api-access-djxkj\") pod \"redhat-marketplace-lpr6m\" (UID: \"ad74a576-9ce6-403d-a4d8-2e7f8ac20b0b\") " pod="openshift-marketplace/redhat-marketplace-lpr6m" Oct 14 07:25:17 crc kubenswrapper[4870]: I1014 07:25:17.237390 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lpr6m" Oct 14 07:25:17 crc kubenswrapper[4870]: I1014 07:25:17.474408 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-lpr6m"] Oct 14 07:25:17 crc kubenswrapper[4870]: W1014 07:25:17.478557 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podad74a576_9ce6_403d_a4d8_2e7f8ac20b0b.slice/crio-7d688518699f261bb22516adecde11fd417722a9704fe3d0be8843b29df2bed0 WatchSource:0}: Error finding container 7d688518699f261bb22516adecde11fd417722a9704fe3d0be8843b29df2bed0: Status 404 returned error can't find the container with id 7d688518699f261bb22516adecde11fd417722a9704fe3d0be8843b29df2bed0 Oct 14 07:25:18 crc kubenswrapper[4870]: I1014 07:25:18.370095 4870 generic.go:334] "Generic (PLEG): container finished" podID="ad74a576-9ce6-403d-a4d8-2e7f8ac20b0b" containerID="4588b56d1c4995d8d075c8938ddacc433cfc995a2bcddf3d79c948d63d80c1e3" exitCode=0 Oct 14 07:25:18 crc kubenswrapper[4870]: I1014 07:25:18.370150 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lpr6m" event={"ID":"ad74a576-9ce6-403d-a4d8-2e7f8ac20b0b","Type":"ContainerDied","Data":"4588b56d1c4995d8d075c8938ddacc433cfc995a2bcddf3d79c948d63d80c1e3"} Oct 14 07:25:18 crc kubenswrapper[4870]: I1014 07:25:18.370181 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lpr6m" event={"ID":"ad74a576-9ce6-403d-a4d8-2e7f8ac20b0b","Type":"ContainerStarted","Data":"7d688518699f261bb22516adecde11fd417722a9704fe3d0be8843b29df2bed0"} Oct 14 07:25:19 crc kubenswrapper[4870]: I1014 07:25:19.383798 4870 generic.go:334] "Generic (PLEG): container finished" podID="ad74a576-9ce6-403d-a4d8-2e7f8ac20b0b" containerID="7988e63e745a0d766b56efa7cb852ecba17cff4a17a090e64a4059425421ae27" exitCode=0 Oct 14 07:25:19 crc kubenswrapper[4870]: I1014 07:25:19.383943 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lpr6m" event={"ID":"ad74a576-9ce6-403d-a4d8-2e7f8ac20b0b","Type":"ContainerDied","Data":"7988e63e745a0d766b56efa7cb852ecba17cff4a17a090e64a4059425421ae27"} Oct 14 07:25:20 crc kubenswrapper[4870]: I1014 07:25:20.396611 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lpr6m" event={"ID":"ad74a576-9ce6-403d-a4d8-2e7f8ac20b0b","Type":"ContainerStarted","Data":"24d4db48788a056f8e4a7b15c54344fdd02bb5b13e9b2a42f392f36c6cc7ff6e"} Oct 14 07:25:20 crc kubenswrapper[4870]: I1014 07:25:20.426595 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-lpr6m" podStartSLOduration=2.975331518 podStartE2EDuration="4.426567996s" podCreationTimestamp="2025-10-14 07:25:16 +0000 UTC" firstStartedPulling="2025-10-14 07:25:18.371921981 +0000 UTC m=+1454.069282362" lastFinishedPulling="2025-10-14 07:25:19.823158459 +0000 UTC m=+1455.520518840" observedRunningTime="2025-10-14 07:25:20.415282442 +0000 UTC m=+1456.112642833" watchObservedRunningTime="2025-10-14 07:25:20.426567996 +0000 UTC m=+1456.123928367" Oct 14 07:25:21 crc kubenswrapper[4870]: I1014 07:25:21.848580 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-sz58s" Oct 14 07:25:21 crc kubenswrapper[4870]: I1014 07:25:21.849040 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-sz58s" Oct 14 07:25:21 crc kubenswrapper[4870]: I1014 07:25:21.912885 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-sz58s" Oct 14 07:25:22 crc kubenswrapper[4870]: I1014 07:25:22.496410 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-sz58s" Oct 14 07:25:23 crc kubenswrapper[4870]: I1014 07:25:23.084374 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-sz58s"] Oct 14 07:25:23 crc kubenswrapper[4870]: I1014 07:25:23.951386 4870 patch_prober.go:28] interesting pod/machine-config-daemon-7tvc7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 07:25:23 crc kubenswrapper[4870]: I1014 07:25:23.951515 4870 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 07:25:24 crc kubenswrapper[4870]: I1014 07:25:24.440628 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-sz58s" podUID="7382409f-f523-4017-9bc9-80661712ff92" containerName="registry-server" containerID="cri-o://ce827dbd893c32c3b59311cb8d54cbff6a44d5d05b84f0777104d4792cd1a60e" gracePeriod=2 Oct 14 07:25:24 crc kubenswrapper[4870]: I1014 07:25:24.866765 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-sz58s" Oct 14 07:25:24 crc kubenswrapper[4870]: I1014 07:25:24.982829 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7382409f-f523-4017-9bc9-80661712ff92-catalog-content\") pod \"7382409f-f523-4017-9bc9-80661712ff92\" (UID: \"7382409f-f523-4017-9bc9-80661712ff92\") " Oct 14 07:25:24 crc kubenswrapper[4870]: I1014 07:25:24.982962 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dk5xg\" (UniqueName: \"kubernetes.io/projected/7382409f-f523-4017-9bc9-80661712ff92-kube-api-access-dk5xg\") pod \"7382409f-f523-4017-9bc9-80661712ff92\" (UID: \"7382409f-f523-4017-9bc9-80661712ff92\") " Oct 14 07:25:24 crc kubenswrapper[4870]: I1014 07:25:24.983000 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7382409f-f523-4017-9bc9-80661712ff92-utilities\") pod \"7382409f-f523-4017-9bc9-80661712ff92\" (UID: \"7382409f-f523-4017-9bc9-80661712ff92\") " Oct 14 07:25:24 crc kubenswrapper[4870]: I1014 07:25:24.984160 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7382409f-f523-4017-9bc9-80661712ff92-utilities" (OuterVolumeSpecName: "utilities") pod "7382409f-f523-4017-9bc9-80661712ff92" (UID: "7382409f-f523-4017-9bc9-80661712ff92"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:25:24 crc kubenswrapper[4870]: I1014 07:25:24.989233 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7382409f-f523-4017-9bc9-80661712ff92-kube-api-access-dk5xg" (OuterVolumeSpecName: "kube-api-access-dk5xg") pod "7382409f-f523-4017-9bc9-80661712ff92" (UID: "7382409f-f523-4017-9bc9-80661712ff92"). InnerVolumeSpecName "kube-api-access-dk5xg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:25:25 crc kubenswrapper[4870]: I1014 07:25:25.085818 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dk5xg\" (UniqueName: \"kubernetes.io/projected/7382409f-f523-4017-9bc9-80661712ff92-kube-api-access-dk5xg\") on node \"crc\" DevicePath \"\"" Oct 14 07:25:25 crc kubenswrapper[4870]: I1014 07:25:25.085940 4870 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7382409f-f523-4017-9bc9-80661712ff92-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 07:25:25 crc kubenswrapper[4870]: I1014 07:25:25.088998 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7382409f-f523-4017-9bc9-80661712ff92-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7382409f-f523-4017-9bc9-80661712ff92" (UID: "7382409f-f523-4017-9bc9-80661712ff92"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:25:25 crc kubenswrapper[4870]: I1014 07:25:25.187365 4870 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7382409f-f523-4017-9bc9-80661712ff92-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 07:25:25 crc kubenswrapper[4870]: I1014 07:25:25.461057 4870 generic.go:334] "Generic (PLEG): container finished" podID="7382409f-f523-4017-9bc9-80661712ff92" containerID="ce827dbd893c32c3b59311cb8d54cbff6a44d5d05b84f0777104d4792cd1a60e" exitCode=0 Oct 14 07:25:25 crc kubenswrapper[4870]: I1014 07:25:25.461120 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sz58s" event={"ID":"7382409f-f523-4017-9bc9-80661712ff92","Type":"ContainerDied","Data":"ce827dbd893c32c3b59311cb8d54cbff6a44d5d05b84f0777104d4792cd1a60e"} Oct 14 07:25:25 crc kubenswrapper[4870]: I1014 07:25:25.461167 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-sz58s" Oct 14 07:25:25 crc kubenswrapper[4870]: I1014 07:25:25.461186 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sz58s" event={"ID":"7382409f-f523-4017-9bc9-80661712ff92","Type":"ContainerDied","Data":"7cc1b268d4726c556d4b28735ddc4c3e665ac53c87cee8e15dda56d36cd48934"} Oct 14 07:25:25 crc kubenswrapper[4870]: I1014 07:25:25.461239 4870 scope.go:117] "RemoveContainer" containerID="ce827dbd893c32c3b59311cb8d54cbff6a44d5d05b84f0777104d4792cd1a60e" Oct 14 07:25:25 crc kubenswrapper[4870]: I1014 07:25:25.519054 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-sz58s"] Oct 14 07:25:25 crc kubenswrapper[4870]: I1014 07:25:25.519657 4870 scope.go:117] "RemoveContainer" containerID="2f87670aabbfa190bb739856b391e28f08c9199aff4167eb658103bc10e66d85" Oct 14 07:25:25 crc kubenswrapper[4870]: I1014 07:25:25.525502 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-sz58s"] Oct 14 07:25:25 crc kubenswrapper[4870]: I1014 07:25:25.550168 4870 scope.go:117] "RemoveContainer" containerID="3921b56a57b3359e4823aa8eb9791d4575ec9186fbd0cf145841ed8fb99886a9" Oct 14 07:25:25 crc kubenswrapper[4870]: I1014 07:25:25.598992 4870 scope.go:117] "RemoveContainer" containerID="ce827dbd893c32c3b59311cb8d54cbff6a44d5d05b84f0777104d4792cd1a60e" Oct 14 07:25:25 crc kubenswrapper[4870]: E1014 07:25:25.599559 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ce827dbd893c32c3b59311cb8d54cbff6a44d5d05b84f0777104d4792cd1a60e\": container with ID starting with ce827dbd893c32c3b59311cb8d54cbff6a44d5d05b84f0777104d4792cd1a60e not found: ID does not exist" containerID="ce827dbd893c32c3b59311cb8d54cbff6a44d5d05b84f0777104d4792cd1a60e" Oct 14 07:25:25 crc kubenswrapper[4870]: I1014 07:25:25.599605 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ce827dbd893c32c3b59311cb8d54cbff6a44d5d05b84f0777104d4792cd1a60e"} err="failed to get container status \"ce827dbd893c32c3b59311cb8d54cbff6a44d5d05b84f0777104d4792cd1a60e\": rpc error: code = NotFound desc = could not find container \"ce827dbd893c32c3b59311cb8d54cbff6a44d5d05b84f0777104d4792cd1a60e\": container with ID starting with ce827dbd893c32c3b59311cb8d54cbff6a44d5d05b84f0777104d4792cd1a60e not found: ID does not exist" Oct 14 07:25:25 crc kubenswrapper[4870]: I1014 07:25:25.599640 4870 scope.go:117] "RemoveContainer" containerID="2f87670aabbfa190bb739856b391e28f08c9199aff4167eb658103bc10e66d85" Oct 14 07:25:25 crc kubenswrapper[4870]: E1014 07:25:25.600208 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2f87670aabbfa190bb739856b391e28f08c9199aff4167eb658103bc10e66d85\": container with ID starting with 2f87670aabbfa190bb739856b391e28f08c9199aff4167eb658103bc10e66d85 not found: ID does not exist" containerID="2f87670aabbfa190bb739856b391e28f08c9199aff4167eb658103bc10e66d85" Oct 14 07:25:25 crc kubenswrapper[4870]: I1014 07:25:25.600296 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2f87670aabbfa190bb739856b391e28f08c9199aff4167eb658103bc10e66d85"} err="failed to get container status \"2f87670aabbfa190bb739856b391e28f08c9199aff4167eb658103bc10e66d85\": rpc error: code = NotFound desc = could not find container \"2f87670aabbfa190bb739856b391e28f08c9199aff4167eb658103bc10e66d85\": container with ID starting with 2f87670aabbfa190bb739856b391e28f08c9199aff4167eb658103bc10e66d85 not found: ID does not exist" Oct 14 07:25:25 crc kubenswrapper[4870]: I1014 07:25:25.600357 4870 scope.go:117] "RemoveContainer" containerID="3921b56a57b3359e4823aa8eb9791d4575ec9186fbd0cf145841ed8fb99886a9" Oct 14 07:25:25 crc kubenswrapper[4870]: E1014 07:25:25.600865 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3921b56a57b3359e4823aa8eb9791d4575ec9186fbd0cf145841ed8fb99886a9\": container with ID starting with 3921b56a57b3359e4823aa8eb9791d4575ec9186fbd0cf145841ed8fb99886a9 not found: ID does not exist" containerID="3921b56a57b3359e4823aa8eb9791d4575ec9186fbd0cf145841ed8fb99886a9" Oct 14 07:25:25 crc kubenswrapper[4870]: I1014 07:25:25.600911 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3921b56a57b3359e4823aa8eb9791d4575ec9186fbd0cf145841ed8fb99886a9"} err="failed to get container status \"3921b56a57b3359e4823aa8eb9791d4575ec9186fbd0cf145841ed8fb99886a9\": rpc error: code = NotFound desc = could not find container \"3921b56a57b3359e4823aa8eb9791d4575ec9186fbd0cf145841ed8fb99886a9\": container with ID starting with 3921b56a57b3359e4823aa8eb9791d4575ec9186fbd0cf145841ed8fb99886a9 not found: ID does not exist" Oct 14 07:25:27 crc kubenswrapper[4870]: I1014 07:25:27.056837 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7382409f-f523-4017-9bc9-80661712ff92" path="/var/lib/kubelet/pods/7382409f-f523-4017-9bc9-80661712ff92/volumes" Oct 14 07:25:27 crc kubenswrapper[4870]: I1014 07:25:27.238070 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-lpr6m" Oct 14 07:25:27 crc kubenswrapper[4870]: I1014 07:25:27.238170 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-lpr6m" Oct 14 07:25:27 crc kubenswrapper[4870]: I1014 07:25:27.307581 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-lpr6m" Oct 14 07:25:27 crc kubenswrapper[4870]: I1014 07:25:27.559160 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-lpr6m" Oct 14 07:25:28 crc kubenswrapper[4870]: I1014 07:25:28.485315 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-lpr6m"] Oct 14 07:25:29 crc kubenswrapper[4870]: I1014 07:25:29.537083 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-lpr6m" podUID="ad74a576-9ce6-403d-a4d8-2e7f8ac20b0b" containerName="registry-server" containerID="cri-o://24d4db48788a056f8e4a7b15c54344fdd02bb5b13e9b2a42f392f36c6cc7ff6e" gracePeriod=2 Oct 14 07:25:30 crc kubenswrapper[4870]: I1014 07:25:30.014833 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lpr6m" Oct 14 07:25:30 crc kubenswrapper[4870]: I1014 07:25:30.065303 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ad74a576-9ce6-403d-a4d8-2e7f8ac20b0b-utilities\") pod \"ad74a576-9ce6-403d-a4d8-2e7f8ac20b0b\" (UID: \"ad74a576-9ce6-403d-a4d8-2e7f8ac20b0b\") " Oct 14 07:25:30 crc kubenswrapper[4870]: I1014 07:25:30.065356 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-djxkj\" (UniqueName: \"kubernetes.io/projected/ad74a576-9ce6-403d-a4d8-2e7f8ac20b0b-kube-api-access-djxkj\") pod \"ad74a576-9ce6-403d-a4d8-2e7f8ac20b0b\" (UID: \"ad74a576-9ce6-403d-a4d8-2e7f8ac20b0b\") " Oct 14 07:25:30 crc kubenswrapper[4870]: I1014 07:25:30.065492 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ad74a576-9ce6-403d-a4d8-2e7f8ac20b0b-catalog-content\") pod \"ad74a576-9ce6-403d-a4d8-2e7f8ac20b0b\" (UID: \"ad74a576-9ce6-403d-a4d8-2e7f8ac20b0b\") " Oct 14 07:25:30 crc kubenswrapper[4870]: I1014 07:25:30.067934 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ad74a576-9ce6-403d-a4d8-2e7f8ac20b0b-utilities" (OuterVolumeSpecName: "utilities") pod "ad74a576-9ce6-403d-a4d8-2e7f8ac20b0b" (UID: "ad74a576-9ce6-403d-a4d8-2e7f8ac20b0b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:25:30 crc kubenswrapper[4870]: I1014 07:25:30.073334 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ad74a576-9ce6-403d-a4d8-2e7f8ac20b0b-kube-api-access-djxkj" (OuterVolumeSpecName: "kube-api-access-djxkj") pod "ad74a576-9ce6-403d-a4d8-2e7f8ac20b0b" (UID: "ad74a576-9ce6-403d-a4d8-2e7f8ac20b0b"). InnerVolumeSpecName "kube-api-access-djxkj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:25:30 crc kubenswrapper[4870]: I1014 07:25:30.083755 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ad74a576-9ce6-403d-a4d8-2e7f8ac20b0b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ad74a576-9ce6-403d-a4d8-2e7f8ac20b0b" (UID: "ad74a576-9ce6-403d-a4d8-2e7f8ac20b0b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:25:30 crc kubenswrapper[4870]: I1014 07:25:30.166831 4870 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ad74a576-9ce6-403d-a4d8-2e7f8ac20b0b-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 07:25:30 crc kubenswrapper[4870]: I1014 07:25:30.166862 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-djxkj\" (UniqueName: \"kubernetes.io/projected/ad74a576-9ce6-403d-a4d8-2e7f8ac20b0b-kube-api-access-djxkj\") on node \"crc\" DevicePath \"\"" Oct 14 07:25:30 crc kubenswrapper[4870]: I1014 07:25:30.166871 4870 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ad74a576-9ce6-403d-a4d8-2e7f8ac20b0b-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 07:25:30 crc kubenswrapper[4870]: I1014 07:25:30.551023 4870 generic.go:334] "Generic (PLEG): container finished" podID="ad74a576-9ce6-403d-a4d8-2e7f8ac20b0b" containerID="24d4db48788a056f8e4a7b15c54344fdd02bb5b13e9b2a42f392f36c6cc7ff6e" exitCode=0 Oct 14 07:25:30 crc kubenswrapper[4870]: I1014 07:25:30.551108 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lpr6m" event={"ID":"ad74a576-9ce6-403d-a4d8-2e7f8ac20b0b","Type":"ContainerDied","Data":"24d4db48788a056f8e4a7b15c54344fdd02bb5b13e9b2a42f392f36c6cc7ff6e"} Oct 14 07:25:30 crc kubenswrapper[4870]: I1014 07:25:30.552262 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lpr6m" event={"ID":"ad74a576-9ce6-403d-a4d8-2e7f8ac20b0b","Type":"ContainerDied","Data":"7d688518699f261bb22516adecde11fd417722a9704fe3d0be8843b29df2bed0"} Oct 14 07:25:30 crc kubenswrapper[4870]: I1014 07:25:30.551132 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lpr6m" Oct 14 07:25:30 crc kubenswrapper[4870]: I1014 07:25:30.552299 4870 scope.go:117] "RemoveContainer" containerID="24d4db48788a056f8e4a7b15c54344fdd02bb5b13e9b2a42f392f36c6cc7ff6e" Oct 14 07:25:30 crc kubenswrapper[4870]: I1014 07:25:30.584838 4870 scope.go:117] "RemoveContainer" containerID="7988e63e745a0d766b56efa7cb852ecba17cff4a17a090e64a4059425421ae27" Oct 14 07:25:30 crc kubenswrapper[4870]: I1014 07:25:30.608787 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-lpr6m"] Oct 14 07:25:30 crc kubenswrapper[4870]: I1014 07:25:30.615178 4870 scope.go:117] "RemoveContainer" containerID="4588b56d1c4995d8d075c8938ddacc433cfc995a2bcddf3d79c948d63d80c1e3" Oct 14 07:25:30 crc kubenswrapper[4870]: I1014 07:25:30.619484 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-lpr6m"] Oct 14 07:25:30 crc kubenswrapper[4870]: I1014 07:25:30.645569 4870 scope.go:117] "RemoveContainer" containerID="24d4db48788a056f8e4a7b15c54344fdd02bb5b13e9b2a42f392f36c6cc7ff6e" Oct 14 07:25:30 crc kubenswrapper[4870]: E1014 07:25:30.646066 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"24d4db48788a056f8e4a7b15c54344fdd02bb5b13e9b2a42f392f36c6cc7ff6e\": container with ID starting with 24d4db48788a056f8e4a7b15c54344fdd02bb5b13e9b2a42f392f36c6cc7ff6e not found: ID does not exist" containerID="24d4db48788a056f8e4a7b15c54344fdd02bb5b13e9b2a42f392f36c6cc7ff6e" Oct 14 07:25:30 crc kubenswrapper[4870]: I1014 07:25:30.646127 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"24d4db48788a056f8e4a7b15c54344fdd02bb5b13e9b2a42f392f36c6cc7ff6e"} err="failed to get container status \"24d4db48788a056f8e4a7b15c54344fdd02bb5b13e9b2a42f392f36c6cc7ff6e\": rpc error: code = NotFound desc = could not find container \"24d4db48788a056f8e4a7b15c54344fdd02bb5b13e9b2a42f392f36c6cc7ff6e\": container with ID starting with 24d4db48788a056f8e4a7b15c54344fdd02bb5b13e9b2a42f392f36c6cc7ff6e not found: ID does not exist" Oct 14 07:25:30 crc kubenswrapper[4870]: I1014 07:25:30.646161 4870 scope.go:117] "RemoveContainer" containerID="7988e63e745a0d766b56efa7cb852ecba17cff4a17a090e64a4059425421ae27" Oct 14 07:25:30 crc kubenswrapper[4870]: E1014 07:25:30.646567 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7988e63e745a0d766b56efa7cb852ecba17cff4a17a090e64a4059425421ae27\": container with ID starting with 7988e63e745a0d766b56efa7cb852ecba17cff4a17a090e64a4059425421ae27 not found: ID does not exist" containerID="7988e63e745a0d766b56efa7cb852ecba17cff4a17a090e64a4059425421ae27" Oct 14 07:25:30 crc kubenswrapper[4870]: I1014 07:25:30.646696 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7988e63e745a0d766b56efa7cb852ecba17cff4a17a090e64a4059425421ae27"} err="failed to get container status \"7988e63e745a0d766b56efa7cb852ecba17cff4a17a090e64a4059425421ae27\": rpc error: code = NotFound desc = could not find container \"7988e63e745a0d766b56efa7cb852ecba17cff4a17a090e64a4059425421ae27\": container with ID starting with 7988e63e745a0d766b56efa7cb852ecba17cff4a17a090e64a4059425421ae27 not found: ID does not exist" Oct 14 07:25:30 crc kubenswrapper[4870]: I1014 07:25:30.646798 4870 scope.go:117] "RemoveContainer" containerID="4588b56d1c4995d8d075c8938ddacc433cfc995a2bcddf3d79c948d63d80c1e3" Oct 14 07:25:30 crc kubenswrapper[4870]: E1014 07:25:30.647274 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4588b56d1c4995d8d075c8938ddacc433cfc995a2bcddf3d79c948d63d80c1e3\": container with ID starting with 4588b56d1c4995d8d075c8938ddacc433cfc995a2bcddf3d79c948d63d80c1e3 not found: ID does not exist" containerID="4588b56d1c4995d8d075c8938ddacc433cfc995a2bcddf3d79c948d63d80c1e3" Oct 14 07:25:30 crc kubenswrapper[4870]: I1014 07:25:30.647314 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4588b56d1c4995d8d075c8938ddacc433cfc995a2bcddf3d79c948d63d80c1e3"} err="failed to get container status \"4588b56d1c4995d8d075c8938ddacc433cfc995a2bcddf3d79c948d63d80c1e3\": rpc error: code = NotFound desc = could not find container \"4588b56d1c4995d8d075c8938ddacc433cfc995a2bcddf3d79c948d63d80c1e3\": container with ID starting with 4588b56d1c4995d8d075c8938ddacc433cfc995a2bcddf3d79c948d63d80c1e3 not found: ID does not exist" Oct 14 07:25:31 crc kubenswrapper[4870]: I1014 07:25:31.049149 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ad74a576-9ce6-403d-a4d8-2e7f8ac20b0b" path="/var/lib/kubelet/pods/ad74a576-9ce6-403d-a4d8-2e7f8ac20b0b/volumes" Oct 14 07:25:53 crc kubenswrapper[4870]: I1014 07:25:53.951376 4870 patch_prober.go:28] interesting pod/machine-config-daemon-7tvc7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 07:25:53 crc kubenswrapper[4870]: I1014 07:25:53.952164 4870 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 07:25:53 crc kubenswrapper[4870]: I1014 07:25:53.952238 4870 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" Oct 14 07:25:53 crc kubenswrapper[4870]: I1014 07:25:53.953171 4870 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"45dee68d970a494a60daa9fc176a487f297e0f291860479dd1d60c51287b32b3"} pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 14 07:25:53 crc kubenswrapper[4870]: I1014 07:25:53.953275 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerName="machine-config-daemon" containerID="cri-o://45dee68d970a494a60daa9fc176a487f297e0f291860479dd1d60c51287b32b3" gracePeriod=600 Oct 14 07:25:54 crc kubenswrapper[4870]: I1014 07:25:54.831306 4870 generic.go:334] "Generic (PLEG): container finished" podID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerID="45dee68d970a494a60daa9fc176a487f297e0f291860479dd1d60c51287b32b3" exitCode=0 Oct 14 07:25:54 crc kubenswrapper[4870]: I1014 07:25:54.831434 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" event={"ID":"874a5e45-dffd-4d17-b609-b7d3ed2eab07","Type":"ContainerDied","Data":"45dee68d970a494a60daa9fc176a487f297e0f291860479dd1d60c51287b32b3"} Oct 14 07:25:54 crc kubenswrapper[4870]: I1014 07:25:54.831849 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" event={"ID":"874a5e45-dffd-4d17-b609-b7d3ed2eab07","Type":"ContainerStarted","Data":"d754019673e24a68a5e4e484dd11d6018830e5970b17109dcba3bd998a84db74"} Oct 14 07:25:54 crc kubenswrapper[4870]: I1014 07:25:54.831876 4870 scope.go:117] "RemoveContainer" containerID="5389ca0979ade5a4bccec9faa4aa51cda4fcec373b2752e9cf9faa85bf74bfdf" Oct 14 07:26:09 crc kubenswrapper[4870]: I1014 07:26:09.440238 4870 scope.go:117] "RemoveContainer" containerID="f712d6e12449176eab14639b7624cd8368d3985bad7d851e1a3f05c35886b81a" Oct 14 07:26:09 crc kubenswrapper[4870]: I1014 07:26:09.481651 4870 scope.go:117] "RemoveContainer" containerID="f1b865fdd1614804c6e4ae198a96114c844ab93d4e346b066d14705872321268" Oct 14 07:26:09 crc kubenswrapper[4870]: I1014 07:26:09.547372 4870 scope.go:117] "RemoveContainer" containerID="4ffe098a8e8e4c9ba69a5357443ce3b51e9f90cd094ae3dbb2018a4ad085b163" Oct 14 07:26:09 crc kubenswrapper[4870]: I1014 07:26:09.602307 4870 scope.go:117] "RemoveContainer" containerID="d4bce05dddb71b43960de2db8b34835c15e33a352960018f56e19358841c968e" Oct 14 07:26:09 crc kubenswrapper[4870]: I1014 07:26:09.632940 4870 scope.go:117] "RemoveContainer" containerID="cea067cfa733de231a12525a23273b289acdfd787a45cbbf5c5d50bd4fd9998e" Oct 14 07:26:09 crc kubenswrapper[4870]: I1014 07:26:09.665244 4870 scope.go:117] "RemoveContainer" containerID="9e34fe756d7db40d6ed7af9609859cb38ea6bbd896e9819ab8153c17bb627435" Oct 14 07:26:09 crc kubenswrapper[4870]: I1014 07:26:09.719327 4870 scope.go:117] "RemoveContainer" containerID="3e6bcfe5922f64d7458739850f7cc5cf193c595b1c6d384c0bbf95f39c7959bc" Oct 14 07:26:09 crc kubenswrapper[4870]: I1014 07:26:09.751544 4870 scope.go:117] "RemoveContainer" containerID="1d7dbc12c4434da4d09988b87f042f614c4f32064d8308852b271f35c34e44d1" Oct 14 07:26:09 crc kubenswrapper[4870]: I1014 07:26:09.779328 4870 scope.go:117] "RemoveContainer" containerID="b1e3d02e0b9f1f551752996b576a2919f1585fce83c7b33a6bc6758f9bf28cc1" Oct 14 07:26:09 crc kubenswrapper[4870]: I1014 07:26:09.805891 4870 scope.go:117] "RemoveContainer" containerID="1b44e2314629f4fd5b7db5c31f15920edf5685a4b13b8bd037c33027e60486b4" Oct 14 07:26:09 crc kubenswrapper[4870]: I1014 07:26:09.833375 4870 scope.go:117] "RemoveContainer" containerID="365d48ec2f5364c81df1e76cb37b8b718e730cf770131932caead0d2a5c0f467" Oct 14 07:26:09 crc kubenswrapper[4870]: I1014 07:26:09.862305 4870 scope.go:117] "RemoveContainer" containerID="6f74f7987dbeaadf7d1897858c5cff4d3ddcf7dce7dbc12f36e6c1f68826ee1d" Oct 14 07:26:09 crc kubenswrapper[4870]: I1014 07:26:09.915810 4870 scope.go:117] "RemoveContainer" containerID="a83b32049ab0d1907715a8d4bb87cfcec31636757720f25ebb523546eaa5e523" Oct 14 07:26:09 crc kubenswrapper[4870]: I1014 07:26:09.965678 4870 scope.go:117] "RemoveContainer" containerID="2c964df0978df33d463f733f9d5d9c9b3b2b0107067baa2d37c1979f7552ff66" Oct 14 07:26:09 crc kubenswrapper[4870]: I1014 07:26:09.992729 4870 scope.go:117] "RemoveContainer" containerID="130ba2c834779a188207288977bbddd0143d006f944bd94f346093f2993c5ff4" Oct 14 07:26:10 crc kubenswrapper[4870]: I1014 07:26:10.026333 4870 scope.go:117] "RemoveContainer" containerID="eba786db77e4785126a35b98eed013099cfce39a7e939fc837fb1037d741e55c" Oct 14 07:26:10 crc kubenswrapper[4870]: I1014 07:26:10.045995 4870 scope.go:117] "RemoveContainer" containerID="457905ca1130c3804c068cf6fc44521ad064a3fb34449d666cf858f14e2feace" Oct 14 07:26:10 crc kubenswrapper[4870]: I1014 07:26:10.073204 4870 scope.go:117] "RemoveContainer" containerID="53081407fd75b324098b0ed4b06bd603bd305531ac1290feb40806718fcc0af4" Oct 14 07:26:10 crc kubenswrapper[4870]: I1014 07:26:10.100281 4870 scope.go:117] "RemoveContainer" containerID="11e57f381235088d13225057331ed61d6e2ab6f98595ff6fe4992063044159e4" Oct 14 07:26:10 crc kubenswrapper[4870]: I1014 07:26:10.129470 4870 scope.go:117] "RemoveContainer" containerID="c5e07d714b9d6ff354b5240abbc6e77cc9c929d1f020b6844549c47e3eded85e" Oct 14 07:27:10 crc kubenswrapper[4870]: I1014 07:27:10.530358 4870 scope.go:117] "RemoveContainer" containerID="00b9207820bcb2e8026c0e8f4ff4fff1bbbb4e4d5e1181da574ef68c63fba739" Oct 14 07:27:10 crc kubenswrapper[4870]: I1014 07:27:10.559248 4870 scope.go:117] "RemoveContainer" containerID="a82c41053072124e1b79993189c4bbf5862bb71cd34f3bf9755173e9a3637eb1" Oct 14 07:27:10 crc kubenswrapper[4870]: I1014 07:27:10.618912 4870 scope.go:117] "RemoveContainer" containerID="8f57f5ca62644a1c87bd38a50bcf2f2378a79660bfe53b23e9830b3861c114f8" Oct 14 07:27:10 crc kubenswrapper[4870]: I1014 07:27:10.654188 4870 scope.go:117] "RemoveContainer" containerID="250338fb343035b6cec2629aa8ef485b90fd7a39a4592f8aae30cde99b224717" Oct 14 07:27:10 crc kubenswrapper[4870]: I1014 07:27:10.676814 4870 scope.go:117] "RemoveContainer" containerID="28a8e8e086b6d6c6e4aa03f701ff3d36e5350745361841619c40f9371ec5f443" Oct 14 07:27:10 crc kubenswrapper[4870]: I1014 07:27:10.725212 4870 scope.go:117] "RemoveContainer" containerID="474c728d027f33d4775c88671e52d61b3ef2e1359014dd049a62df531455d2e4" Oct 14 07:27:10 crc kubenswrapper[4870]: I1014 07:27:10.747357 4870 scope.go:117] "RemoveContainer" containerID="4a86c8a92708d5909a487ab3cdcb494f586b0ebda954a88a7d9c31a60d00a222" Oct 14 07:27:10 crc kubenswrapper[4870]: I1014 07:27:10.775837 4870 scope.go:117] "RemoveContainer" containerID="77a9d26b6f11833d6d27a01cdd11e90dfaece26453d11c1389ea5966482e89ea" Oct 14 07:27:10 crc kubenswrapper[4870]: I1014 07:27:10.797390 4870 scope.go:117] "RemoveContainer" containerID="afdfea867b44e3792f20e1c1ecf3ae23b03aa598321ac8c6d8a8225512ee0c5a" Oct 14 07:27:10 crc kubenswrapper[4870]: I1014 07:27:10.830057 4870 scope.go:117] "RemoveContainer" containerID="35a065d725ec0b4b5f843ba5a5779e23bb304a93a7feb8c329fffbfe341e4930" Oct 14 07:27:10 crc kubenswrapper[4870]: I1014 07:27:10.867212 4870 scope.go:117] "RemoveContainer" containerID="87f2dc9b9b112ef49726f92a6ed22f035a2c6c59983215e902ee16f2b25877d9" Oct 14 07:28:11 crc kubenswrapper[4870]: I1014 07:28:11.087578 4870 scope.go:117] "RemoveContainer" containerID="8e85e09a8e31af355d315694c672c0246f87999b118cfc6dc5c46c1e00ad0b77" Oct 14 07:28:11 crc kubenswrapper[4870]: I1014 07:28:11.118558 4870 scope.go:117] "RemoveContainer" containerID="3027ab58c61f91be86ef6147ca252c055e21923601e19c2de2f6b7121ee15ada" Oct 14 07:28:11 crc kubenswrapper[4870]: I1014 07:28:11.142577 4870 scope.go:117] "RemoveContainer" containerID="49adfddfc295310c01014e57a90893bd2516bc1e4d16c17bd66c300fee355163" Oct 14 07:28:11 crc kubenswrapper[4870]: I1014 07:28:11.159146 4870 scope.go:117] "RemoveContainer" containerID="2488802b352a138347373046951250f07e48d44d2dce6d3f42f5ae6763c49ffd" Oct 14 07:28:11 crc kubenswrapper[4870]: I1014 07:28:11.188036 4870 scope.go:117] "RemoveContainer" containerID="da1b8f24e5ce4601d5b9ab11fdd2ff2e4c7ba1c1afbb715dc02d5dbe6b6257d2" Oct 14 07:28:11 crc kubenswrapper[4870]: I1014 07:28:11.202173 4870 scope.go:117] "RemoveContainer" containerID="571dfedb34f9303684e8f8fe09485d339b6c363eaa75f6950ae538f2526d0129" Oct 14 07:28:11 crc kubenswrapper[4870]: I1014 07:28:11.230425 4870 scope.go:117] "RemoveContainer" containerID="76f2a940e9d8e2f75865166a0366231116cc1a0a333f49e74aef93fb29a8e1a2" Oct 14 07:28:11 crc kubenswrapper[4870]: I1014 07:28:11.245245 4870 scope.go:117] "RemoveContainer" containerID="fdda30c816580a958c0638fccaade56d725b6b8fffdcdb85825f921582f25f4e" Oct 14 07:28:11 crc kubenswrapper[4870]: I1014 07:28:11.262102 4870 scope.go:117] "RemoveContainer" containerID="deba864879defc602edfa6ada18f77ffb9b44bf36a5292d8a32f7922bcd29311" Oct 14 07:28:11 crc kubenswrapper[4870]: I1014 07:28:11.282193 4870 scope.go:117] "RemoveContainer" containerID="1de36f0eb4ed54f334543bf349d349ea3bac6126555500cdd2af3c26b73dffe7" Oct 14 07:28:11 crc kubenswrapper[4870]: I1014 07:28:11.300307 4870 scope.go:117] "RemoveContainer" containerID="d311512ab3d18b4243743bfbfb8a259533302afd01cbbcfda23ec6a678817840" Oct 14 07:28:11 crc kubenswrapper[4870]: I1014 07:28:11.349607 4870 scope.go:117] "RemoveContainer" containerID="4d1574b6ecc78807ca0e045857f0ad61906f9011616d9b3dbeba62e89d80183d" Oct 14 07:28:11 crc kubenswrapper[4870]: I1014 07:28:11.368668 4870 scope.go:117] "RemoveContainer" containerID="2841eadea64f22ddd21b4d710878d4871c5c5ceec4c6a266af64c8c08bfe682a" Oct 14 07:28:11 crc kubenswrapper[4870]: I1014 07:28:11.403936 4870 scope.go:117] "RemoveContainer" containerID="c07b4185cb742a8d604acc1b5ef508fa09c68ed0229dc00df2a8b8a8be770694" Oct 14 07:28:11 crc kubenswrapper[4870]: I1014 07:28:11.423981 4870 scope.go:117] "RemoveContainer" containerID="2bcc81a3a9d5c820c0bcd54b459b4f42446c2e0c7a6b7503e11f9c3e6552fadb" Oct 14 07:28:11 crc kubenswrapper[4870]: I1014 07:28:11.452891 4870 scope.go:117] "RemoveContainer" containerID="ccd5820b235e529b2ee667f2bc0cb4aaa6d87fc0297a583d5753340e0d346237" Oct 14 07:28:11 crc kubenswrapper[4870]: I1014 07:28:11.476994 4870 scope.go:117] "RemoveContainer" containerID="ce1d38864bb409e404978e6b3a0f6bd89d3dba826ce7be995022fe4e0ac9b188" Oct 14 07:28:23 crc kubenswrapper[4870]: I1014 07:28:23.951662 4870 patch_prober.go:28] interesting pod/machine-config-daemon-7tvc7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 07:28:23 crc kubenswrapper[4870]: I1014 07:28:23.953831 4870 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 07:28:53 crc kubenswrapper[4870]: I1014 07:28:53.951298 4870 patch_prober.go:28] interesting pod/machine-config-daemon-7tvc7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 07:28:53 crc kubenswrapper[4870]: I1014 07:28:53.953816 4870 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 07:29:11 crc kubenswrapper[4870]: I1014 07:29:11.696562 4870 scope.go:117] "RemoveContainer" containerID="f24a75c62a58aefe115cf6dc0ab70e01b811b66e1369401528bc3134ca7d1cc4" Oct 14 07:29:11 crc kubenswrapper[4870]: I1014 07:29:11.726925 4870 scope.go:117] "RemoveContainer" containerID="e1959691705f2a2f3aa974444388035c568669302c7da6551de910de00176330" Oct 14 07:29:11 crc kubenswrapper[4870]: I1014 07:29:11.754391 4870 scope.go:117] "RemoveContainer" containerID="018378adb14febcf4afcbaac2d9f70a84cc9e7f96675c9b85ee751bb465d9632" Oct 14 07:29:11 crc kubenswrapper[4870]: I1014 07:29:11.778943 4870 scope.go:117] "RemoveContainer" containerID="c3d6d43a38bf5b35a30900bbcca66f0dcbacc37ec8b96a0733086e4ae6871167" Oct 14 07:29:11 crc kubenswrapper[4870]: I1014 07:29:11.833926 4870 scope.go:117] "RemoveContainer" containerID="a7f06f2b8b7523a5efb4f46ab915a339acb0f160d1f9753afa2035de3c545c6d" Oct 14 07:29:11 crc kubenswrapper[4870]: I1014 07:29:11.852813 4870 scope.go:117] "RemoveContainer" containerID="45200242c420e7ff44ca9c10b9cff44cc8bb4c9294423cc3e3e0ad235691f6b4" Oct 14 07:29:11 crc kubenswrapper[4870]: I1014 07:29:11.877021 4870 scope.go:117] "RemoveContainer" containerID="92059a83ebd09e581c802c9d18a6cc53c68a57ad45e7c0b7e480b8f24d3f2bd8" Oct 14 07:29:11 crc kubenswrapper[4870]: I1014 07:29:11.900795 4870 scope.go:117] "RemoveContainer" containerID="ef229458c64f8f5cfd522f061876879d09c15e53eebb20bb3651a9f2a7a32418" Oct 14 07:29:23 crc kubenswrapper[4870]: I1014 07:29:23.951176 4870 patch_prober.go:28] interesting pod/machine-config-daemon-7tvc7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 07:29:23 crc kubenswrapper[4870]: I1014 07:29:23.952562 4870 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 07:29:23 crc kubenswrapper[4870]: I1014 07:29:23.952680 4870 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" Oct 14 07:29:23 crc kubenswrapper[4870]: I1014 07:29:23.954174 4870 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d754019673e24a68a5e4e484dd11d6018830e5970b17109dcba3bd998a84db74"} pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 14 07:29:23 crc kubenswrapper[4870]: I1014 07:29:23.954321 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerName="machine-config-daemon" containerID="cri-o://d754019673e24a68a5e4e484dd11d6018830e5970b17109dcba3bd998a84db74" gracePeriod=600 Oct 14 07:29:24 crc kubenswrapper[4870]: E1014 07:29:24.090727 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 07:29:24 crc kubenswrapper[4870]: I1014 07:29:24.101959 4870 generic.go:334] "Generic (PLEG): container finished" podID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerID="d754019673e24a68a5e4e484dd11d6018830e5970b17109dcba3bd998a84db74" exitCode=0 Oct 14 07:29:24 crc kubenswrapper[4870]: I1014 07:29:24.102046 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" event={"ID":"874a5e45-dffd-4d17-b609-b7d3ed2eab07","Type":"ContainerDied","Data":"d754019673e24a68a5e4e484dd11d6018830e5970b17109dcba3bd998a84db74"} Oct 14 07:29:24 crc kubenswrapper[4870]: I1014 07:29:24.102096 4870 scope.go:117] "RemoveContainer" containerID="45dee68d970a494a60daa9fc176a487f297e0f291860479dd1d60c51287b32b3" Oct 14 07:29:24 crc kubenswrapper[4870]: I1014 07:29:24.102995 4870 scope.go:117] "RemoveContainer" containerID="d754019673e24a68a5e4e484dd11d6018830e5970b17109dcba3bd998a84db74" Oct 14 07:29:24 crc kubenswrapper[4870]: E1014 07:29:24.103401 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 07:29:25 crc kubenswrapper[4870]: I1014 07:29:25.117565 4870 scope.go:117] "RemoveContainer" containerID="d754019673e24a68a5e4e484dd11d6018830e5970b17109dcba3bd998a84db74" Oct 14 07:29:25 crc kubenswrapper[4870]: E1014 07:29:25.118008 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 07:29:38 crc kubenswrapper[4870]: I1014 07:29:38.034291 4870 scope.go:117] "RemoveContainer" containerID="d754019673e24a68a5e4e484dd11d6018830e5970b17109dcba3bd998a84db74" Oct 14 07:29:38 crc kubenswrapper[4870]: E1014 07:29:38.035042 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 07:29:52 crc kubenswrapper[4870]: I1014 07:29:52.035265 4870 scope.go:117] "RemoveContainer" containerID="d754019673e24a68a5e4e484dd11d6018830e5970b17109dcba3bd998a84db74" Oct 14 07:29:52 crc kubenswrapper[4870]: E1014 07:29:52.036258 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 07:30:00 crc kubenswrapper[4870]: I1014 07:30:00.163719 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29340450-gx8ch"] Oct 14 07:30:00 crc kubenswrapper[4870]: E1014 07:30:00.167601 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad74a576-9ce6-403d-a4d8-2e7f8ac20b0b" containerName="registry-server" Oct 14 07:30:00 crc kubenswrapper[4870]: I1014 07:30:00.167751 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad74a576-9ce6-403d-a4d8-2e7f8ac20b0b" containerName="registry-server" Oct 14 07:30:00 crc kubenswrapper[4870]: E1014 07:30:00.167868 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7382409f-f523-4017-9bc9-80661712ff92" containerName="registry-server" Oct 14 07:30:00 crc kubenswrapper[4870]: I1014 07:30:00.167972 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="7382409f-f523-4017-9bc9-80661712ff92" containerName="registry-server" Oct 14 07:30:00 crc kubenswrapper[4870]: E1014 07:30:00.168084 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7382409f-f523-4017-9bc9-80661712ff92" containerName="extract-content" Oct 14 07:30:00 crc kubenswrapper[4870]: I1014 07:30:00.168218 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="7382409f-f523-4017-9bc9-80661712ff92" containerName="extract-content" Oct 14 07:30:00 crc kubenswrapper[4870]: E1014 07:30:00.168476 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad74a576-9ce6-403d-a4d8-2e7f8ac20b0b" containerName="extract-utilities" Oct 14 07:30:00 crc kubenswrapper[4870]: I1014 07:30:00.168574 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad74a576-9ce6-403d-a4d8-2e7f8ac20b0b" containerName="extract-utilities" Oct 14 07:30:00 crc kubenswrapper[4870]: E1014 07:30:00.168674 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7382409f-f523-4017-9bc9-80661712ff92" containerName="extract-utilities" Oct 14 07:30:00 crc kubenswrapper[4870]: I1014 07:30:00.168748 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="7382409f-f523-4017-9bc9-80661712ff92" containerName="extract-utilities" Oct 14 07:30:00 crc kubenswrapper[4870]: E1014 07:30:00.168829 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad74a576-9ce6-403d-a4d8-2e7f8ac20b0b" containerName="extract-content" Oct 14 07:30:00 crc kubenswrapper[4870]: I1014 07:30:00.168901 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad74a576-9ce6-403d-a4d8-2e7f8ac20b0b" containerName="extract-content" Oct 14 07:30:00 crc kubenswrapper[4870]: I1014 07:30:00.169175 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="ad74a576-9ce6-403d-a4d8-2e7f8ac20b0b" containerName="registry-server" Oct 14 07:30:00 crc kubenswrapper[4870]: I1014 07:30:00.169296 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="7382409f-f523-4017-9bc9-80661712ff92" containerName="registry-server" Oct 14 07:30:00 crc kubenswrapper[4870]: I1014 07:30:00.170147 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29340450-gx8ch" Oct 14 07:30:00 crc kubenswrapper[4870]: I1014 07:30:00.173175 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 14 07:30:00 crc kubenswrapper[4870]: I1014 07:30:00.173469 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 14 07:30:00 crc kubenswrapper[4870]: I1014 07:30:00.187380 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29340450-gx8ch"] Oct 14 07:30:00 crc kubenswrapper[4870]: I1014 07:30:00.330822 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/767a62c8-6040-4095-8a79-dbaa9401940d-secret-volume\") pod \"collect-profiles-29340450-gx8ch\" (UID: \"767a62c8-6040-4095-8a79-dbaa9401940d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340450-gx8ch" Oct 14 07:30:00 crc kubenswrapper[4870]: I1014 07:30:00.330878 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8b5fm\" (UniqueName: \"kubernetes.io/projected/767a62c8-6040-4095-8a79-dbaa9401940d-kube-api-access-8b5fm\") pod \"collect-profiles-29340450-gx8ch\" (UID: \"767a62c8-6040-4095-8a79-dbaa9401940d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340450-gx8ch" Oct 14 07:30:00 crc kubenswrapper[4870]: I1014 07:30:00.331339 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/767a62c8-6040-4095-8a79-dbaa9401940d-config-volume\") pod \"collect-profiles-29340450-gx8ch\" (UID: \"767a62c8-6040-4095-8a79-dbaa9401940d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340450-gx8ch" Oct 14 07:30:00 crc kubenswrapper[4870]: I1014 07:30:00.432759 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/767a62c8-6040-4095-8a79-dbaa9401940d-config-volume\") pod \"collect-profiles-29340450-gx8ch\" (UID: \"767a62c8-6040-4095-8a79-dbaa9401940d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340450-gx8ch" Oct 14 07:30:00 crc kubenswrapper[4870]: I1014 07:30:00.432886 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/767a62c8-6040-4095-8a79-dbaa9401940d-secret-volume\") pod \"collect-profiles-29340450-gx8ch\" (UID: \"767a62c8-6040-4095-8a79-dbaa9401940d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340450-gx8ch" Oct 14 07:30:00 crc kubenswrapper[4870]: I1014 07:30:00.432961 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8b5fm\" (UniqueName: \"kubernetes.io/projected/767a62c8-6040-4095-8a79-dbaa9401940d-kube-api-access-8b5fm\") pod \"collect-profiles-29340450-gx8ch\" (UID: \"767a62c8-6040-4095-8a79-dbaa9401940d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340450-gx8ch" Oct 14 07:30:00 crc kubenswrapper[4870]: I1014 07:30:00.434223 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/767a62c8-6040-4095-8a79-dbaa9401940d-config-volume\") pod \"collect-profiles-29340450-gx8ch\" (UID: \"767a62c8-6040-4095-8a79-dbaa9401940d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340450-gx8ch" Oct 14 07:30:00 crc kubenswrapper[4870]: I1014 07:30:00.443110 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/767a62c8-6040-4095-8a79-dbaa9401940d-secret-volume\") pod \"collect-profiles-29340450-gx8ch\" (UID: \"767a62c8-6040-4095-8a79-dbaa9401940d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340450-gx8ch" Oct 14 07:30:00 crc kubenswrapper[4870]: I1014 07:30:00.454552 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8b5fm\" (UniqueName: \"kubernetes.io/projected/767a62c8-6040-4095-8a79-dbaa9401940d-kube-api-access-8b5fm\") pod \"collect-profiles-29340450-gx8ch\" (UID: \"767a62c8-6040-4095-8a79-dbaa9401940d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340450-gx8ch" Oct 14 07:30:00 crc kubenswrapper[4870]: I1014 07:30:00.508035 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29340450-gx8ch" Oct 14 07:30:00 crc kubenswrapper[4870]: I1014 07:30:00.810866 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29340450-gx8ch"] Oct 14 07:30:01 crc kubenswrapper[4870]: I1014 07:30:01.515950 4870 generic.go:334] "Generic (PLEG): container finished" podID="767a62c8-6040-4095-8a79-dbaa9401940d" containerID="4b8393d462d3cba7296c43eb753cc6a1d10b653d0f087f8b55fd1cf7c9573ba5" exitCode=0 Oct 14 07:30:01 crc kubenswrapper[4870]: I1014 07:30:01.516035 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29340450-gx8ch" event={"ID":"767a62c8-6040-4095-8a79-dbaa9401940d","Type":"ContainerDied","Data":"4b8393d462d3cba7296c43eb753cc6a1d10b653d0f087f8b55fd1cf7c9573ba5"} Oct 14 07:30:01 crc kubenswrapper[4870]: I1014 07:30:01.516354 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29340450-gx8ch" event={"ID":"767a62c8-6040-4095-8a79-dbaa9401940d","Type":"ContainerStarted","Data":"2313122f735c8de63228d91f84b574ad37820796c72e22c568dd7bdd3f7942aa"} Oct 14 07:30:02 crc kubenswrapper[4870]: I1014 07:30:02.890168 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29340450-gx8ch" Oct 14 07:30:02 crc kubenswrapper[4870]: I1014 07:30:02.976157 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/767a62c8-6040-4095-8a79-dbaa9401940d-secret-volume\") pod \"767a62c8-6040-4095-8a79-dbaa9401940d\" (UID: \"767a62c8-6040-4095-8a79-dbaa9401940d\") " Oct 14 07:30:02 crc kubenswrapper[4870]: I1014 07:30:02.983148 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/767a62c8-6040-4095-8a79-dbaa9401940d-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "767a62c8-6040-4095-8a79-dbaa9401940d" (UID: "767a62c8-6040-4095-8a79-dbaa9401940d"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:30:03 crc kubenswrapper[4870]: I1014 07:30:03.078496 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/767a62c8-6040-4095-8a79-dbaa9401940d-config-volume\") pod \"767a62c8-6040-4095-8a79-dbaa9401940d\" (UID: \"767a62c8-6040-4095-8a79-dbaa9401940d\") " Oct 14 07:30:03 crc kubenswrapper[4870]: I1014 07:30:03.078596 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8b5fm\" (UniqueName: \"kubernetes.io/projected/767a62c8-6040-4095-8a79-dbaa9401940d-kube-api-access-8b5fm\") pod \"767a62c8-6040-4095-8a79-dbaa9401940d\" (UID: \"767a62c8-6040-4095-8a79-dbaa9401940d\") " Oct 14 07:30:03 crc kubenswrapper[4870]: I1014 07:30:03.079009 4870 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/767a62c8-6040-4095-8a79-dbaa9401940d-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 14 07:30:03 crc kubenswrapper[4870]: I1014 07:30:03.080128 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/767a62c8-6040-4095-8a79-dbaa9401940d-config-volume" (OuterVolumeSpecName: "config-volume") pod "767a62c8-6040-4095-8a79-dbaa9401940d" (UID: "767a62c8-6040-4095-8a79-dbaa9401940d"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:30:03 crc kubenswrapper[4870]: I1014 07:30:03.085668 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/767a62c8-6040-4095-8a79-dbaa9401940d-kube-api-access-8b5fm" (OuterVolumeSpecName: "kube-api-access-8b5fm") pod "767a62c8-6040-4095-8a79-dbaa9401940d" (UID: "767a62c8-6040-4095-8a79-dbaa9401940d"). InnerVolumeSpecName "kube-api-access-8b5fm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:30:03 crc kubenswrapper[4870]: I1014 07:30:03.180957 4870 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/767a62c8-6040-4095-8a79-dbaa9401940d-config-volume\") on node \"crc\" DevicePath \"\"" Oct 14 07:30:03 crc kubenswrapper[4870]: I1014 07:30:03.181026 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8b5fm\" (UniqueName: \"kubernetes.io/projected/767a62c8-6040-4095-8a79-dbaa9401940d-kube-api-access-8b5fm\") on node \"crc\" DevicePath \"\"" Oct 14 07:30:03 crc kubenswrapper[4870]: I1014 07:30:03.540977 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29340450-gx8ch" event={"ID":"767a62c8-6040-4095-8a79-dbaa9401940d","Type":"ContainerDied","Data":"2313122f735c8de63228d91f84b574ad37820796c72e22c568dd7bdd3f7942aa"} Oct 14 07:30:03 crc kubenswrapper[4870]: I1014 07:30:03.541029 4870 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2313122f735c8de63228d91f84b574ad37820796c72e22c568dd7bdd3f7942aa" Oct 14 07:30:03 crc kubenswrapper[4870]: I1014 07:30:03.541051 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29340450-gx8ch" Oct 14 07:30:05 crc kubenswrapper[4870]: I1014 07:30:05.043716 4870 scope.go:117] "RemoveContainer" containerID="d754019673e24a68a5e4e484dd11d6018830e5970b17109dcba3bd998a84db74" Oct 14 07:30:05 crc kubenswrapper[4870]: E1014 07:30:05.044116 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 07:30:18 crc kubenswrapper[4870]: I1014 07:30:18.033762 4870 scope.go:117] "RemoveContainer" containerID="d754019673e24a68a5e4e484dd11d6018830e5970b17109dcba3bd998a84db74" Oct 14 07:30:18 crc kubenswrapper[4870]: E1014 07:30:18.034367 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 07:30:33 crc kubenswrapper[4870]: I1014 07:30:33.033638 4870 scope.go:117] "RemoveContainer" containerID="d754019673e24a68a5e4e484dd11d6018830e5970b17109dcba3bd998a84db74" Oct 14 07:30:33 crc kubenswrapper[4870]: E1014 07:30:33.034366 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 07:30:48 crc kubenswrapper[4870]: I1014 07:30:48.034801 4870 scope.go:117] "RemoveContainer" containerID="d754019673e24a68a5e4e484dd11d6018830e5970b17109dcba3bd998a84db74" Oct 14 07:30:48 crc kubenswrapper[4870]: E1014 07:30:48.035835 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 07:31:02 crc kubenswrapper[4870]: I1014 07:31:02.035041 4870 scope.go:117] "RemoveContainer" containerID="d754019673e24a68a5e4e484dd11d6018830e5970b17109dcba3bd998a84db74" Oct 14 07:31:02 crc kubenswrapper[4870]: E1014 07:31:02.036680 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 07:31:17 crc kubenswrapper[4870]: I1014 07:31:17.035513 4870 scope.go:117] "RemoveContainer" containerID="d754019673e24a68a5e4e484dd11d6018830e5970b17109dcba3bd998a84db74" Oct 14 07:31:17 crc kubenswrapper[4870]: E1014 07:31:17.036582 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 07:31:29 crc kubenswrapper[4870]: I1014 07:31:29.034225 4870 scope.go:117] "RemoveContainer" containerID="d754019673e24a68a5e4e484dd11d6018830e5970b17109dcba3bd998a84db74" Oct 14 07:31:29 crc kubenswrapper[4870]: E1014 07:31:29.035419 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 07:31:41 crc kubenswrapper[4870]: I1014 07:31:41.035124 4870 scope.go:117] "RemoveContainer" containerID="d754019673e24a68a5e4e484dd11d6018830e5970b17109dcba3bd998a84db74" Oct 14 07:31:41 crc kubenswrapper[4870]: E1014 07:31:41.036267 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 07:31:55 crc kubenswrapper[4870]: I1014 07:31:55.041068 4870 scope.go:117] "RemoveContainer" containerID="d754019673e24a68a5e4e484dd11d6018830e5970b17109dcba3bd998a84db74" Oct 14 07:31:55 crc kubenswrapper[4870]: E1014 07:31:55.042015 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 07:32:07 crc kubenswrapper[4870]: I1014 07:32:07.034407 4870 scope.go:117] "RemoveContainer" containerID="d754019673e24a68a5e4e484dd11d6018830e5970b17109dcba3bd998a84db74" Oct 14 07:32:07 crc kubenswrapper[4870]: E1014 07:32:07.035340 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 07:32:21 crc kubenswrapper[4870]: I1014 07:32:21.034795 4870 scope.go:117] "RemoveContainer" containerID="d754019673e24a68a5e4e484dd11d6018830e5970b17109dcba3bd998a84db74" Oct 14 07:32:21 crc kubenswrapper[4870]: E1014 07:32:21.037150 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 07:32:32 crc kubenswrapper[4870]: I1014 07:32:32.034768 4870 scope.go:117] "RemoveContainer" containerID="d754019673e24a68a5e4e484dd11d6018830e5970b17109dcba3bd998a84db74" Oct 14 07:32:32 crc kubenswrapper[4870]: E1014 07:32:32.035844 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 07:32:44 crc kubenswrapper[4870]: I1014 07:32:44.034642 4870 scope.go:117] "RemoveContainer" containerID="d754019673e24a68a5e4e484dd11d6018830e5970b17109dcba3bd998a84db74" Oct 14 07:32:44 crc kubenswrapper[4870]: E1014 07:32:44.035676 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 07:32:55 crc kubenswrapper[4870]: I1014 07:32:55.040169 4870 scope.go:117] "RemoveContainer" containerID="d754019673e24a68a5e4e484dd11d6018830e5970b17109dcba3bd998a84db74" Oct 14 07:32:55 crc kubenswrapper[4870]: E1014 07:32:55.041414 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 07:33:08 crc kubenswrapper[4870]: I1014 07:33:08.034380 4870 scope.go:117] "RemoveContainer" containerID="d754019673e24a68a5e4e484dd11d6018830e5970b17109dcba3bd998a84db74" Oct 14 07:33:08 crc kubenswrapper[4870]: E1014 07:33:08.035355 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 07:33:20 crc kubenswrapper[4870]: I1014 07:33:20.033662 4870 scope.go:117] "RemoveContainer" containerID="d754019673e24a68a5e4e484dd11d6018830e5970b17109dcba3bd998a84db74" Oct 14 07:33:20 crc kubenswrapper[4870]: E1014 07:33:20.034654 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 07:33:35 crc kubenswrapper[4870]: I1014 07:33:35.043423 4870 scope.go:117] "RemoveContainer" containerID="d754019673e24a68a5e4e484dd11d6018830e5970b17109dcba3bd998a84db74" Oct 14 07:33:35 crc kubenswrapper[4870]: E1014 07:33:35.045079 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 07:33:49 crc kubenswrapper[4870]: I1014 07:33:49.034089 4870 scope.go:117] "RemoveContainer" containerID="d754019673e24a68a5e4e484dd11d6018830e5970b17109dcba3bd998a84db74" Oct 14 07:33:49 crc kubenswrapper[4870]: E1014 07:33:49.035885 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 07:34:03 crc kubenswrapper[4870]: I1014 07:34:03.034854 4870 scope.go:117] "RemoveContainer" containerID="d754019673e24a68a5e4e484dd11d6018830e5970b17109dcba3bd998a84db74" Oct 14 07:34:03 crc kubenswrapper[4870]: E1014 07:34:03.036080 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 07:34:18 crc kubenswrapper[4870]: I1014 07:34:18.034467 4870 scope.go:117] "RemoveContainer" containerID="d754019673e24a68a5e4e484dd11d6018830e5970b17109dcba3bd998a84db74" Oct 14 07:34:18 crc kubenswrapper[4870]: E1014 07:34:18.035232 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 07:34:32 crc kubenswrapper[4870]: I1014 07:34:32.034646 4870 scope.go:117] "RemoveContainer" containerID="d754019673e24a68a5e4e484dd11d6018830e5970b17109dcba3bd998a84db74" Oct 14 07:34:33 crc kubenswrapper[4870]: I1014 07:34:33.150536 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" event={"ID":"874a5e45-dffd-4d17-b609-b7d3ed2eab07","Type":"ContainerStarted","Data":"13bbb5c4f6c1ff5e8c8987841a34cf3435bc534c235a0bc4bb5b02514ede6e5e"} Oct 14 07:34:45 crc kubenswrapper[4870]: I1014 07:34:45.709209 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-9qgf9"] Oct 14 07:34:45 crc kubenswrapper[4870]: E1014 07:34:45.711534 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="767a62c8-6040-4095-8a79-dbaa9401940d" containerName="collect-profiles" Oct 14 07:34:45 crc kubenswrapper[4870]: I1014 07:34:45.711574 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="767a62c8-6040-4095-8a79-dbaa9401940d" containerName="collect-profiles" Oct 14 07:34:45 crc kubenswrapper[4870]: I1014 07:34:45.712029 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="767a62c8-6040-4095-8a79-dbaa9401940d" containerName="collect-profiles" Oct 14 07:34:45 crc kubenswrapper[4870]: I1014 07:34:45.714562 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9qgf9" Oct 14 07:34:45 crc kubenswrapper[4870]: I1014 07:34:45.720576 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-9qgf9"] Oct 14 07:34:45 crc kubenswrapper[4870]: I1014 07:34:45.867309 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2baef9db-d79c-4f55-9c3f-cd4ac8488a8d-utilities\") pod \"community-operators-9qgf9\" (UID: \"2baef9db-d79c-4f55-9c3f-cd4ac8488a8d\") " pod="openshift-marketplace/community-operators-9qgf9" Oct 14 07:34:45 crc kubenswrapper[4870]: I1014 07:34:45.867365 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dqzcn\" (UniqueName: \"kubernetes.io/projected/2baef9db-d79c-4f55-9c3f-cd4ac8488a8d-kube-api-access-dqzcn\") pod \"community-operators-9qgf9\" (UID: \"2baef9db-d79c-4f55-9c3f-cd4ac8488a8d\") " pod="openshift-marketplace/community-operators-9qgf9" Oct 14 07:34:45 crc kubenswrapper[4870]: I1014 07:34:45.867689 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2baef9db-d79c-4f55-9c3f-cd4ac8488a8d-catalog-content\") pod \"community-operators-9qgf9\" (UID: \"2baef9db-d79c-4f55-9c3f-cd4ac8488a8d\") " pod="openshift-marketplace/community-operators-9qgf9" Oct 14 07:34:45 crc kubenswrapper[4870]: I1014 07:34:45.913115 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-g5cbp"] Oct 14 07:34:45 crc kubenswrapper[4870]: I1014 07:34:45.916125 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-g5cbp" Oct 14 07:34:45 crc kubenswrapper[4870]: I1014 07:34:45.937631 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-g5cbp"] Oct 14 07:34:45 crc kubenswrapper[4870]: I1014 07:34:45.968708 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2baef9db-d79c-4f55-9c3f-cd4ac8488a8d-catalog-content\") pod \"community-operators-9qgf9\" (UID: \"2baef9db-d79c-4f55-9c3f-cd4ac8488a8d\") " pod="openshift-marketplace/community-operators-9qgf9" Oct 14 07:34:45 crc kubenswrapper[4870]: I1014 07:34:45.968772 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2baef9db-d79c-4f55-9c3f-cd4ac8488a8d-utilities\") pod \"community-operators-9qgf9\" (UID: \"2baef9db-d79c-4f55-9c3f-cd4ac8488a8d\") " pod="openshift-marketplace/community-operators-9qgf9" Oct 14 07:34:45 crc kubenswrapper[4870]: I1014 07:34:45.968832 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dqzcn\" (UniqueName: \"kubernetes.io/projected/2baef9db-d79c-4f55-9c3f-cd4ac8488a8d-kube-api-access-dqzcn\") pod \"community-operators-9qgf9\" (UID: \"2baef9db-d79c-4f55-9c3f-cd4ac8488a8d\") " pod="openshift-marketplace/community-operators-9qgf9" Oct 14 07:34:45 crc kubenswrapper[4870]: I1014 07:34:45.969420 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2baef9db-d79c-4f55-9c3f-cd4ac8488a8d-catalog-content\") pod \"community-operators-9qgf9\" (UID: \"2baef9db-d79c-4f55-9c3f-cd4ac8488a8d\") " pod="openshift-marketplace/community-operators-9qgf9" Oct 14 07:34:45 crc kubenswrapper[4870]: I1014 07:34:45.969462 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2baef9db-d79c-4f55-9c3f-cd4ac8488a8d-utilities\") pod \"community-operators-9qgf9\" (UID: \"2baef9db-d79c-4f55-9c3f-cd4ac8488a8d\") " pod="openshift-marketplace/community-operators-9qgf9" Oct 14 07:34:45 crc kubenswrapper[4870]: I1014 07:34:45.994211 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dqzcn\" (UniqueName: \"kubernetes.io/projected/2baef9db-d79c-4f55-9c3f-cd4ac8488a8d-kube-api-access-dqzcn\") pod \"community-operators-9qgf9\" (UID: \"2baef9db-d79c-4f55-9c3f-cd4ac8488a8d\") " pod="openshift-marketplace/community-operators-9qgf9" Oct 14 07:34:46 crc kubenswrapper[4870]: I1014 07:34:46.071086 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nw9sm\" (UniqueName: \"kubernetes.io/projected/83e0f943-ad16-4731-b196-13cb0165e023-kube-api-access-nw9sm\") pod \"redhat-operators-g5cbp\" (UID: \"83e0f943-ad16-4731-b196-13cb0165e023\") " pod="openshift-marketplace/redhat-operators-g5cbp" Oct 14 07:34:46 crc kubenswrapper[4870]: I1014 07:34:46.071962 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/83e0f943-ad16-4731-b196-13cb0165e023-utilities\") pod \"redhat-operators-g5cbp\" (UID: \"83e0f943-ad16-4731-b196-13cb0165e023\") " pod="openshift-marketplace/redhat-operators-g5cbp" Oct 14 07:34:46 crc kubenswrapper[4870]: I1014 07:34:46.072118 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/83e0f943-ad16-4731-b196-13cb0165e023-catalog-content\") pod \"redhat-operators-g5cbp\" (UID: \"83e0f943-ad16-4731-b196-13cb0165e023\") " pod="openshift-marketplace/redhat-operators-g5cbp" Oct 14 07:34:46 crc kubenswrapper[4870]: I1014 07:34:46.080396 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9qgf9" Oct 14 07:34:46 crc kubenswrapper[4870]: I1014 07:34:46.175059 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/83e0f943-ad16-4731-b196-13cb0165e023-utilities\") pod \"redhat-operators-g5cbp\" (UID: \"83e0f943-ad16-4731-b196-13cb0165e023\") " pod="openshift-marketplace/redhat-operators-g5cbp" Oct 14 07:34:46 crc kubenswrapper[4870]: I1014 07:34:46.175216 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/83e0f943-ad16-4731-b196-13cb0165e023-catalog-content\") pod \"redhat-operators-g5cbp\" (UID: \"83e0f943-ad16-4731-b196-13cb0165e023\") " pod="openshift-marketplace/redhat-operators-g5cbp" Oct 14 07:34:46 crc kubenswrapper[4870]: I1014 07:34:46.175364 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nw9sm\" (UniqueName: \"kubernetes.io/projected/83e0f943-ad16-4731-b196-13cb0165e023-kube-api-access-nw9sm\") pod \"redhat-operators-g5cbp\" (UID: \"83e0f943-ad16-4731-b196-13cb0165e023\") " pod="openshift-marketplace/redhat-operators-g5cbp" Oct 14 07:34:46 crc kubenswrapper[4870]: I1014 07:34:46.177009 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/83e0f943-ad16-4731-b196-13cb0165e023-catalog-content\") pod \"redhat-operators-g5cbp\" (UID: \"83e0f943-ad16-4731-b196-13cb0165e023\") " pod="openshift-marketplace/redhat-operators-g5cbp" Oct 14 07:34:46 crc kubenswrapper[4870]: I1014 07:34:46.176752 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/83e0f943-ad16-4731-b196-13cb0165e023-utilities\") pod \"redhat-operators-g5cbp\" (UID: \"83e0f943-ad16-4731-b196-13cb0165e023\") " pod="openshift-marketplace/redhat-operators-g5cbp" Oct 14 07:34:46 crc kubenswrapper[4870]: I1014 07:34:46.220169 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nw9sm\" (UniqueName: \"kubernetes.io/projected/83e0f943-ad16-4731-b196-13cb0165e023-kube-api-access-nw9sm\") pod \"redhat-operators-g5cbp\" (UID: \"83e0f943-ad16-4731-b196-13cb0165e023\") " pod="openshift-marketplace/redhat-operators-g5cbp" Oct 14 07:34:46 crc kubenswrapper[4870]: I1014 07:34:46.249357 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-g5cbp" Oct 14 07:34:46 crc kubenswrapper[4870]: I1014 07:34:46.593536 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-9qgf9"] Oct 14 07:34:46 crc kubenswrapper[4870]: W1014 07:34:46.596182 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2baef9db_d79c_4f55_9c3f_cd4ac8488a8d.slice/crio-91caa1e999fcc488856c96556dc72c49dde7dde6a08713c0c4bbc551486f7575 WatchSource:0}: Error finding container 91caa1e999fcc488856c96556dc72c49dde7dde6a08713c0c4bbc551486f7575: Status 404 returned error can't find the container with id 91caa1e999fcc488856c96556dc72c49dde7dde6a08713c0c4bbc551486f7575 Oct 14 07:34:46 crc kubenswrapper[4870]: I1014 07:34:46.712522 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-g5cbp"] Oct 14 07:34:46 crc kubenswrapper[4870]: W1014 07:34:46.713247 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod83e0f943_ad16_4731_b196_13cb0165e023.slice/crio-d0d1fce021486be3e4c371a7b53be1c864ae697d50108b8730cdd09fe3b01011 WatchSource:0}: Error finding container d0d1fce021486be3e4c371a7b53be1c864ae697d50108b8730cdd09fe3b01011: Status 404 returned error can't find the container with id d0d1fce021486be3e4c371a7b53be1c864ae697d50108b8730cdd09fe3b01011 Oct 14 07:34:47 crc kubenswrapper[4870]: I1014 07:34:47.286727 4870 generic.go:334] "Generic (PLEG): container finished" podID="83e0f943-ad16-4731-b196-13cb0165e023" containerID="8e153ddd7940c00b231b0dbc24e6cf87df0731938b00091048ac7eda34d6dbb8" exitCode=0 Oct 14 07:34:47 crc kubenswrapper[4870]: I1014 07:34:47.286861 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g5cbp" event={"ID":"83e0f943-ad16-4731-b196-13cb0165e023","Type":"ContainerDied","Data":"8e153ddd7940c00b231b0dbc24e6cf87df0731938b00091048ac7eda34d6dbb8"} Oct 14 07:34:47 crc kubenswrapper[4870]: I1014 07:34:47.286921 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g5cbp" event={"ID":"83e0f943-ad16-4731-b196-13cb0165e023","Type":"ContainerStarted","Data":"d0d1fce021486be3e4c371a7b53be1c864ae697d50108b8730cdd09fe3b01011"} Oct 14 07:34:47 crc kubenswrapper[4870]: I1014 07:34:47.289326 4870 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 14 07:34:47 crc kubenswrapper[4870]: I1014 07:34:47.291093 4870 generic.go:334] "Generic (PLEG): container finished" podID="2baef9db-d79c-4f55-9c3f-cd4ac8488a8d" containerID="663cb46f4f51bf6fb5ceb2a4b0027b914a3dba9aa5486f8dadfab93dd3a1e061" exitCode=0 Oct 14 07:34:47 crc kubenswrapper[4870]: I1014 07:34:47.293261 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9qgf9" event={"ID":"2baef9db-d79c-4f55-9c3f-cd4ac8488a8d","Type":"ContainerDied","Data":"663cb46f4f51bf6fb5ceb2a4b0027b914a3dba9aa5486f8dadfab93dd3a1e061"} Oct 14 07:34:47 crc kubenswrapper[4870]: I1014 07:34:47.293302 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9qgf9" event={"ID":"2baef9db-d79c-4f55-9c3f-cd4ac8488a8d","Type":"ContainerStarted","Data":"91caa1e999fcc488856c96556dc72c49dde7dde6a08713c0c4bbc551486f7575"} Oct 14 07:34:49 crc kubenswrapper[4870]: I1014 07:34:49.310335 4870 generic.go:334] "Generic (PLEG): container finished" podID="83e0f943-ad16-4731-b196-13cb0165e023" containerID="52766fc1cb66a0d68626fa1c09fe6470678cb714084f5e01526187120054c9e4" exitCode=0 Oct 14 07:34:49 crc kubenswrapper[4870]: I1014 07:34:49.310488 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g5cbp" event={"ID":"83e0f943-ad16-4731-b196-13cb0165e023","Type":"ContainerDied","Data":"52766fc1cb66a0d68626fa1c09fe6470678cb714084f5e01526187120054c9e4"} Oct 14 07:34:51 crc kubenswrapper[4870]: I1014 07:34:51.328767 4870 generic.go:334] "Generic (PLEG): container finished" podID="2baef9db-d79c-4f55-9c3f-cd4ac8488a8d" containerID="767a55328722b63ea6d75648d06c370895e1fe0a5429ed7a61f10f13379f3d6c" exitCode=0 Oct 14 07:34:51 crc kubenswrapper[4870]: I1014 07:34:51.328857 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9qgf9" event={"ID":"2baef9db-d79c-4f55-9c3f-cd4ac8488a8d","Type":"ContainerDied","Data":"767a55328722b63ea6d75648d06c370895e1fe0a5429ed7a61f10f13379f3d6c"} Oct 14 07:34:52 crc kubenswrapper[4870]: I1014 07:34:52.339630 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g5cbp" event={"ID":"83e0f943-ad16-4731-b196-13cb0165e023","Type":"ContainerStarted","Data":"fbd9a982702292c4ca7010b6c409fc0a8e6cd189e8fbc4a0ed10a0a0fb8f9dca"} Oct 14 07:34:52 crc kubenswrapper[4870]: I1014 07:34:52.369408 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-g5cbp" podStartSLOduration=3.579117131 podStartE2EDuration="7.36936793s" podCreationTimestamp="2025-10-14 07:34:45 +0000 UTC" firstStartedPulling="2025-10-14 07:34:47.288629437 +0000 UTC m=+2022.985989838" lastFinishedPulling="2025-10-14 07:34:51.078880226 +0000 UTC m=+2026.776240637" observedRunningTime="2025-10-14 07:34:52.367007542 +0000 UTC m=+2028.064367913" watchObservedRunningTime="2025-10-14 07:34:52.36936793 +0000 UTC m=+2028.066728341" Oct 14 07:34:53 crc kubenswrapper[4870]: I1014 07:34:53.351011 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9qgf9" event={"ID":"2baef9db-d79c-4f55-9c3f-cd4ac8488a8d","Type":"ContainerStarted","Data":"877dd9831079fffa4178ebbc2445a85a114bf775ab1effc8694fc1d707ff00a2"} Oct 14 07:34:53 crc kubenswrapper[4870]: I1014 07:34:53.383058 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-9qgf9" podStartSLOduration=3.515625305 podStartE2EDuration="8.383028345s" podCreationTimestamp="2025-10-14 07:34:45 +0000 UTC" firstStartedPulling="2025-10-14 07:34:47.294152264 +0000 UTC m=+2022.991512665" lastFinishedPulling="2025-10-14 07:34:52.161555304 +0000 UTC m=+2027.858915705" observedRunningTime="2025-10-14 07:34:53.374804792 +0000 UTC m=+2029.072165163" watchObservedRunningTime="2025-10-14 07:34:53.383028345 +0000 UTC m=+2029.080388716" Oct 14 07:34:56 crc kubenswrapper[4870]: I1014 07:34:56.080640 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-9qgf9" Oct 14 07:34:56 crc kubenswrapper[4870]: I1014 07:34:56.080950 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-9qgf9" Oct 14 07:34:56 crc kubenswrapper[4870]: I1014 07:34:56.136244 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-9qgf9" Oct 14 07:34:56 crc kubenswrapper[4870]: I1014 07:34:56.251026 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-g5cbp" Oct 14 07:34:56 crc kubenswrapper[4870]: I1014 07:34:56.252027 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-g5cbp" Oct 14 07:34:57 crc kubenswrapper[4870]: I1014 07:34:57.302187 4870 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-g5cbp" podUID="83e0f943-ad16-4731-b196-13cb0165e023" containerName="registry-server" probeResult="failure" output=< Oct 14 07:34:57 crc kubenswrapper[4870]: timeout: failed to connect service ":50051" within 1s Oct 14 07:34:57 crc kubenswrapper[4870]: > Oct 14 07:35:06 crc kubenswrapper[4870]: I1014 07:35:06.174365 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-9qgf9" Oct 14 07:35:06 crc kubenswrapper[4870]: I1014 07:35:06.290164 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-9qgf9"] Oct 14 07:35:06 crc kubenswrapper[4870]: I1014 07:35:06.321969 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-g5cbp" Oct 14 07:35:06 crc kubenswrapper[4870]: I1014 07:35:06.349445 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-gx6t2"] Oct 14 07:35:06 crc kubenswrapper[4870]: I1014 07:35:06.349731 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-gx6t2" podUID="ad17975e-43e4-4fda-86a2-eb50ba104055" containerName="registry-server" containerID="cri-o://39672b19a3ed09209a2959f962d538ebd415f15ec16ce8b7054ec9bafcafe900" gracePeriod=2 Oct 14 07:35:06 crc kubenswrapper[4870]: I1014 07:35:06.384769 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-g5cbp" Oct 14 07:35:06 crc kubenswrapper[4870]: I1014 07:35:06.773375 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gx6t2" Oct 14 07:35:06 crc kubenswrapper[4870]: I1014 07:35:06.951509 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ad17975e-43e4-4fda-86a2-eb50ba104055-catalog-content\") pod \"ad17975e-43e4-4fda-86a2-eb50ba104055\" (UID: \"ad17975e-43e4-4fda-86a2-eb50ba104055\") " Oct 14 07:35:06 crc kubenswrapper[4870]: I1014 07:35:06.951708 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ad17975e-43e4-4fda-86a2-eb50ba104055-utilities\") pod \"ad17975e-43e4-4fda-86a2-eb50ba104055\" (UID: \"ad17975e-43e4-4fda-86a2-eb50ba104055\") " Oct 14 07:35:06 crc kubenswrapper[4870]: I1014 07:35:06.951799 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q4v8w\" (UniqueName: \"kubernetes.io/projected/ad17975e-43e4-4fda-86a2-eb50ba104055-kube-api-access-q4v8w\") pod \"ad17975e-43e4-4fda-86a2-eb50ba104055\" (UID: \"ad17975e-43e4-4fda-86a2-eb50ba104055\") " Oct 14 07:35:06 crc kubenswrapper[4870]: I1014 07:35:06.952703 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ad17975e-43e4-4fda-86a2-eb50ba104055-utilities" (OuterVolumeSpecName: "utilities") pod "ad17975e-43e4-4fda-86a2-eb50ba104055" (UID: "ad17975e-43e4-4fda-86a2-eb50ba104055"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:35:06 crc kubenswrapper[4870]: I1014 07:35:06.964725 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ad17975e-43e4-4fda-86a2-eb50ba104055-kube-api-access-q4v8w" (OuterVolumeSpecName: "kube-api-access-q4v8w") pod "ad17975e-43e4-4fda-86a2-eb50ba104055" (UID: "ad17975e-43e4-4fda-86a2-eb50ba104055"). InnerVolumeSpecName "kube-api-access-q4v8w". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:35:07 crc kubenswrapper[4870]: I1014 07:35:07.001311 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ad17975e-43e4-4fda-86a2-eb50ba104055-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ad17975e-43e4-4fda-86a2-eb50ba104055" (UID: "ad17975e-43e4-4fda-86a2-eb50ba104055"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:35:07 crc kubenswrapper[4870]: I1014 07:35:07.053376 4870 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ad17975e-43e4-4fda-86a2-eb50ba104055-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 07:35:07 crc kubenswrapper[4870]: I1014 07:35:07.053429 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q4v8w\" (UniqueName: \"kubernetes.io/projected/ad17975e-43e4-4fda-86a2-eb50ba104055-kube-api-access-q4v8w\") on node \"crc\" DevicePath \"\"" Oct 14 07:35:07 crc kubenswrapper[4870]: I1014 07:35:07.053447 4870 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ad17975e-43e4-4fda-86a2-eb50ba104055-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 07:35:07 crc kubenswrapper[4870]: I1014 07:35:07.482916 4870 generic.go:334] "Generic (PLEG): container finished" podID="ad17975e-43e4-4fda-86a2-eb50ba104055" containerID="39672b19a3ed09209a2959f962d538ebd415f15ec16ce8b7054ec9bafcafe900" exitCode=0 Oct 14 07:35:07 crc kubenswrapper[4870]: I1014 07:35:07.482997 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gx6t2" event={"ID":"ad17975e-43e4-4fda-86a2-eb50ba104055","Type":"ContainerDied","Data":"39672b19a3ed09209a2959f962d538ebd415f15ec16ce8b7054ec9bafcafe900"} Oct 14 07:35:07 crc kubenswrapper[4870]: I1014 07:35:07.483380 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gx6t2" event={"ID":"ad17975e-43e4-4fda-86a2-eb50ba104055","Type":"ContainerDied","Data":"e1cc2c228224d281e5dcd32cf752f38f21d8b80f4e483ed92cc34b09f507653e"} Oct 14 07:35:07 crc kubenswrapper[4870]: I1014 07:35:07.483082 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gx6t2" Oct 14 07:35:07 crc kubenswrapper[4870]: I1014 07:35:07.483417 4870 scope.go:117] "RemoveContainer" containerID="39672b19a3ed09209a2959f962d538ebd415f15ec16ce8b7054ec9bafcafe900" Oct 14 07:35:07 crc kubenswrapper[4870]: I1014 07:35:07.518385 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-gx6t2"] Oct 14 07:35:07 crc kubenswrapper[4870]: I1014 07:35:07.525745 4870 scope.go:117] "RemoveContainer" containerID="e02875bb6687d2c29029f9c42337dfd4c0b09c990b10d8862351990437a0a8c5" Oct 14 07:35:07 crc kubenswrapper[4870]: I1014 07:35:07.527828 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-gx6t2"] Oct 14 07:35:07 crc kubenswrapper[4870]: I1014 07:35:07.578600 4870 scope.go:117] "RemoveContainer" containerID="dbeb0e2d82b30a81e19d4708dd5f09496b0f79175aa2a1e82c9c488114369d44" Oct 14 07:35:07 crc kubenswrapper[4870]: I1014 07:35:07.601710 4870 scope.go:117] "RemoveContainer" containerID="39672b19a3ed09209a2959f962d538ebd415f15ec16ce8b7054ec9bafcafe900" Oct 14 07:35:07 crc kubenswrapper[4870]: E1014 07:35:07.602369 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"39672b19a3ed09209a2959f962d538ebd415f15ec16ce8b7054ec9bafcafe900\": container with ID starting with 39672b19a3ed09209a2959f962d538ebd415f15ec16ce8b7054ec9bafcafe900 not found: ID does not exist" containerID="39672b19a3ed09209a2959f962d538ebd415f15ec16ce8b7054ec9bafcafe900" Oct 14 07:35:07 crc kubenswrapper[4870]: I1014 07:35:07.602497 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"39672b19a3ed09209a2959f962d538ebd415f15ec16ce8b7054ec9bafcafe900"} err="failed to get container status \"39672b19a3ed09209a2959f962d538ebd415f15ec16ce8b7054ec9bafcafe900\": rpc error: code = NotFound desc = could not find container \"39672b19a3ed09209a2959f962d538ebd415f15ec16ce8b7054ec9bafcafe900\": container with ID starting with 39672b19a3ed09209a2959f962d538ebd415f15ec16ce8b7054ec9bafcafe900 not found: ID does not exist" Oct 14 07:35:07 crc kubenswrapper[4870]: I1014 07:35:07.602561 4870 scope.go:117] "RemoveContainer" containerID="e02875bb6687d2c29029f9c42337dfd4c0b09c990b10d8862351990437a0a8c5" Oct 14 07:35:07 crc kubenswrapper[4870]: E1014 07:35:07.603105 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e02875bb6687d2c29029f9c42337dfd4c0b09c990b10d8862351990437a0a8c5\": container with ID starting with e02875bb6687d2c29029f9c42337dfd4c0b09c990b10d8862351990437a0a8c5 not found: ID does not exist" containerID="e02875bb6687d2c29029f9c42337dfd4c0b09c990b10d8862351990437a0a8c5" Oct 14 07:35:07 crc kubenswrapper[4870]: I1014 07:35:07.603170 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e02875bb6687d2c29029f9c42337dfd4c0b09c990b10d8862351990437a0a8c5"} err="failed to get container status \"e02875bb6687d2c29029f9c42337dfd4c0b09c990b10d8862351990437a0a8c5\": rpc error: code = NotFound desc = could not find container \"e02875bb6687d2c29029f9c42337dfd4c0b09c990b10d8862351990437a0a8c5\": container with ID starting with e02875bb6687d2c29029f9c42337dfd4c0b09c990b10d8862351990437a0a8c5 not found: ID does not exist" Oct 14 07:35:07 crc kubenswrapper[4870]: I1014 07:35:07.603202 4870 scope.go:117] "RemoveContainer" containerID="dbeb0e2d82b30a81e19d4708dd5f09496b0f79175aa2a1e82c9c488114369d44" Oct 14 07:35:07 crc kubenswrapper[4870]: E1014 07:35:07.603666 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dbeb0e2d82b30a81e19d4708dd5f09496b0f79175aa2a1e82c9c488114369d44\": container with ID starting with dbeb0e2d82b30a81e19d4708dd5f09496b0f79175aa2a1e82c9c488114369d44 not found: ID does not exist" containerID="dbeb0e2d82b30a81e19d4708dd5f09496b0f79175aa2a1e82c9c488114369d44" Oct 14 07:35:07 crc kubenswrapper[4870]: I1014 07:35:07.603720 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dbeb0e2d82b30a81e19d4708dd5f09496b0f79175aa2a1e82c9c488114369d44"} err="failed to get container status \"dbeb0e2d82b30a81e19d4708dd5f09496b0f79175aa2a1e82c9c488114369d44\": rpc error: code = NotFound desc = could not find container \"dbeb0e2d82b30a81e19d4708dd5f09496b0f79175aa2a1e82c9c488114369d44\": container with ID starting with dbeb0e2d82b30a81e19d4708dd5f09496b0f79175aa2a1e82c9c488114369d44 not found: ID does not exist" Oct 14 07:35:08 crc kubenswrapper[4870]: I1014 07:35:08.651635 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-g5cbp"] Oct 14 07:35:08 crc kubenswrapper[4870]: I1014 07:35:08.652310 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-g5cbp" podUID="83e0f943-ad16-4731-b196-13cb0165e023" containerName="registry-server" containerID="cri-o://fbd9a982702292c4ca7010b6c409fc0a8e6cd189e8fbc4a0ed10a0a0fb8f9dca" gracePeriod=2 Oct 14 07:35:09 crc kubenswrapper[4870]: I1014 07:35:09.044681 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ad17975e-43e4-4fda-86a2-eb50ba104055" path="/var/lib/kubelet/pods/ad17975e-43e4-4fda-86a2-eb50ba104055/volumes" Oct 14 07:35:09 crc kubenswrapper[4870]: I1014 07:35:09.085851 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-g5cbp" Oct 14 07:35:09 crc kubenswrapper[4870]: I1014 07:35:09.191263 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/83e0f943-ad16-4731-b196-13cb0165e023-utilities\") pod \"83e0f943-ad16-4731-b196-13cb0165e023\" (UID: \"83e0f943-ad16-4731-b196-13cb0165e023\") " Oct 14 07:35:09 crc kubenswrapper[4870]: I1014 07:35:09.191359 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nw9sm\" (UniqueName: \"kubernetes.io/projected/83e0f943-ad16-4731-b196-13cb0165e023-kube-api-access-nw9sm\") pod \"83e0f943-ad16-4731-b196-13cb0165e023\" (UID: \"83e0f943-ad16-4731-b196-13cb0165e023\") " Oct 14 07:35:09 crc kubenswrapper[4870]: I1014 07:35:09.191536 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/83e0f943-ad16-4731-b196-13cb0165e023-catalog-content\") pod \"83e0f943-ad16-4731-b196-13cb0165e023\" (UID: \"83e0f943-ad16-4731-b196-13cb0165e023\") " Oct 14 07:35:09 crc kubenswrapper[4870]: I1014 07:35:09.192033 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/83e0f943-ad16-4731-b196-13cb0165e023-utilities" (OuterVolumeSpecName: "utilities") pod "83e0f943-ad16-4731-b196-13cb0165e023" (UID: "83e0f943-ad16-4731-b196-13cb0165e023"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:35:09 crc kubenswrapper[4870]: I1014 07:35:09.193147 4870 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/83e0f943-ad16-4731-b196-13cb0165e023-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 07:35:09 crc kubenswrapper[4870]: I1014 07:35:09.196585 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/83e0f943-ad16-4731-b196-13cb0165e023-kube-api-access-nw9sm" (OuterVolumeSpecName: "kube-api-access-nw9sm") pod "83e0f943-ad16-4731-b196-13cb0165e023" (UID: "83e0f943-ad16-4731-b196-13cb0165e023"). InnerVolumeSpecName "kube-api-access-nw9sm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:35:09 crc kubenswrapper[4870]: I1014 07:35:09.272728 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/83e0f943-ad16-4731-b196-13cb0165e023-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "83e0f943-ad16-4731-b196-13cb0165e023" (UID: "83e0f943-ad16-4731-b196-13cb0165e023"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:35:09 crc kubenswrapper[4870]: I1014 07:35:09.294919 4870 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/83e0f943-ad16-4731-b196-13cb0165e023-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 07:35:09 crc kubenswrapper[4870]: I1014 07:35:09.294955 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nw9sm\" (UniqueName: \"kubernetes.io/projected/83e0f943-ad16-4731-b196-13cb0165e023-kube-api-access-nw9sm\") on node \"crc\" DevicePath \"\"" Oct 14 07:35:09 crc kubenswrapper[4870]: I1014 07:35:09.519114 4870 generic.go:334] "Generic (PLEG): container finished" podID="83e0f943-ad16-4731-b196-13cb0165e023" containerID="fbd9a982702292c4ca7010b6c409fc0a8e6cd189e8fbc4a0ed10a0a0fb8f9dca" exitCode=0 Oct 14 07:35:09 crc kubenswrapper[4870]: I1014 07:35:09.519172 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g5cbp" event={"ID":"83e0f943-ad16-4731-b196-13cb0165e023","Type":"ContainerDied","Data":"fbd9a982702292c4ca7010b6c409fc0a8e6cd189e8fbc4a0ed10a0a0fb8f9dca"} Oct 14 07:35:09 crc kubenswrapper[4870]: I1014 07:35:09.519207 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g5cbp" event={"ID":"83e0f943-ad16-4731-b196-13cb0165e023","Type":"ContainerDied","Data":"d0d1fce021486be3e4c371a7b53be1c864ae697d50108b8730cdd09fe3b01011"} Oct 14 07:35:09 crc kubenswrapper[4870]: I1014 07:35:09.519228 4870 scope.go:117] "RemoveContainer" containerID="fbd9a982702292c4ca7010b6c409fc0a8e6cd189e8fbc4a0ed10a0a0fb8f9dca" Oct 14 07:35:09 crc kubenswrapper[4870]: I1014 07:35:09.519392 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-g5cbp" Oct 14 07:35:09 crc kubenswrapper[4870]: I1014 07:35:09.537667 4870 scope.go:117] "RemoveContainer" containerID="52766fc1cb66a0d68626fa1c09fe6470678cb714084f5e01526187120054c9e4" Oct 14 07:35:09 crc kubenswrapper[4870]: I1014 07:35:09.554269 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-g5cbp"] Oct 14 07:35:09 crc kubenswrapper[4870]: I1014 07:35:09.564004 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-g5cbp"] Oct 14 07:35:09 crc kubenswrapper[4870]: I1014 07:35:09.564587 4870 scope.go:117] "RemoveContainer" containerID="8e153ddd7940c00b231b0dbc24e6cf87df0731938b00091048ac7eda34d6dbb8" Oct 14 07:35:09 crc kubenswrapper[4870]: I1014 07:35:09.593602 4870 scope.go:117] "RemoveContainer" containerID="fbd9a982702292c4ca7010b6c409fc0a8e6cd189e8fbc4a0ed10a0a0fb8f9dca" Oct 14 07:35:09 crc kubenswrapper[4870]: E1014 07:35:09.595019 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fbd9a982702292c4ca7010b6c409fc0a8e6cd189e8fbc4a0ed10a0a0fb8f9dca\": container with ID starting with fbd9a982702292c4ca7010b6c409fc0a8e6cd189e8fbc4a0ed10a0a0fb8f9dca not found: ID does not exist" containerID="fbd9a982702292c4ca7010b6c409fc0a8e6cd189e8fbc4a0ed10a0a0fb8f9dca" Oct 14 07:35:09 crc kubenswrapper[4870]: I1014 07:35:09.595069 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fbd9a982702292c4ca7010b6c409fc0a8e6cd189e8fbc4a0ed10a0a0fb8f9dca"} err="failed to get container status \"fbd9a982702292c4ca7010b6c409fc0a8e6cd189e8fbc4a0ed10a0a0fb8f9dca\": rpc error: code = NotFound desc = could not find container \"fbd9a982702292c4ca7010b6c409fc0a8e6cd189e8fbc4a0ed10a0a0fb8f9dca\": container with ID starting with fbd9a982702292c4ca7010b6c409fc0a8e6cd189e8fbc4a0ed10a0a0fb8f9dca not found: ID does not exist" Oct 14 07:35:09 crc kubenswrapper[4870]: I1014 07:35:09.595102 4870 scope.go:117] "RemoveContainer" containerID="52766fc1cb66a0d68626fa1c09fe6470678cb714084f5e01526187120054c9e4" Oct 14 07:35:09 crc kubenswrapper[4870]: E1014 07:35:09.595502 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"52766fc1cb66a0d68626fa1c09fe6470678cb714084f5e01526187120054c9e4\": container with ID starting with 52766fc1cb66a0d68626fa1c09fe6470678cb714084f5e01526187120054c9e4 not found: ID does not exist" containerID="52766fc1cb66a0d68626fa1c09fe6470678cb714084f5e01526187120054c9e4" Oct 14 07:35:09 crc kubenswrapper[4870]: I1014 07:35:09.595539 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"52766fc1cb66a0d68626fa1c09fe6470678cb714084f5e01526187120054c9e4"} err="failed to get container status \"52766fc1cb66a0d68626fa1c09fe6470678cb714084f5e01526187120054c9e4\": rpc error: code = NotFound desc = could not find container \"52766fc1cb66a0d68626fa1c09fe6470678cb714084f5e01526187120054c9e4\": container with ID starting with 52766fc1cb66a0d68626fa1c09fe6470678cb714084f5e01526187120054c9e4 not found: ID does not exist" Oct 14 07:35:09 crc kubenswrapper[4870]: I1014 07:35:09.595568 4870 scope.go:117] "RemoveContainer" containerID="8e153ddd7940c00b231b0dbc24e6cf87df0731938b00091048ac7eda34d6dbb8" Oct 14 07:35:09 crc kubenswrapper[4870]: E1014 07:35:09.595876 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8e153ddd7940c00b231b0dbc24e6cf87df0731938b00091048ac7eda34d6dbb8\": container with ID starting with 8e153ddd7940c00b231b0dbc24e6cf87df0731938b00091048ac7eda34d6dbb8 not found: ID does not exist" containerID="8e153ddd7940c00b231b0dbc24e6cf87df0731938b00091048ac7eda34d6dbb8" Oct 14 07:35:09 crc kubenswrapper[4870]: I1014 07:35:09.595899 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8e153ddd7940c00b231b0dbc24e6cf87df0731938b00091048ac7eda34d6dbb8"} err="failed to get container status \"8e153ddd7940c00b231b0dbc24e6cf87df0731938b00091048ac7eda34d6dbb8\": rpc error: code = NotFound desc = could not find container \"8e153ddd7940c00b231b0dbc24e6cf87df0731938b00091048ac7eda34d6dbb8\": container with ID starting with 8e153ddd7940c00b231b0dbc24e6cf87df0731938b00091048ac7eda34d6dbb8 not found: ID does not exist" Oct 14 07:35:11 crc kubenswrapper[4870]: I1014 07:35:11.051306 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="83e0f943-ad16-4731-b196-13cb0165e023" path="/var/lib/kubelet/pods/83e0f943-ad16-4731-b196-13cb0165e023/volumes" Oct 14 07:35:43 crc kubenswrapper[4870]: I1014 07:35:43.295193 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-8qwbr"] Oct 14 07:35:43 crc kubenswrapper[4870]: E1014 07:35:43.297408 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83e0f943-ad16-4731-b196-13cb0165e023" containerName="registry-server" Oct 14 07:35:43 crc kubenswrapper[4870]: I1014 07:35:43.297467 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="83e0f943-ad16-4731-b196-13cb0165e023" containerName="registry-server" Oct 14 07:35:43 crc kubenswrapper[4870]: E1014 07:35:43.297487 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad17975e-43e4-4fda-86a2-eb50ba104055" containerName="registry-server" Oct 14 07:35:43 crc kubenswrapper[4870]: I1014 07:35:43.297497 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad17975e-43e4-4fda-86a2-eb50ba104055" containerName="registry-server" Oct 14 07:35:43 crc kubenswrapper[4870]: E1014 07:35:43.297522 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad17975e-43e4-4fda-86a2-eb50ba104055" containerName="extract-content" Oct 14 07:35:43 crc kubenswrapper[4870]: I1014 07:35:43.297530 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad17975e-43e4-4fda-86a2-eb50ba104055" containerName="extract-content" Oct 14 07:35:43 crc kubenswrapper[4870]: E1014 07:35:43.297549 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83e0f943-ad16-4731-b196-13cb0165e023" containerName="extract-utilities" Oct 14 07:35:43 crc kubenswrapper[4870]: I1014 07:35:43.297558 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="83e0f943-ad16-4731-b196-13cb0165e023" containerName="extract-utilities" Oct 14 07:35:43 crc kubenswrapper[4870]: E1014 07:35:43.297572 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83e0f943-ad16-4731-b196-13cb0165e023" containerName="extract-content" Oct 14 07:35:43 crc kubenswrapper[4870]: I1014 07:35:43.297580 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="83e0f943-ad16-4731-b196-13cb0165e023" containerName="extract-content" Oct 14 07:35:43 crc kubenswrapper[4870]: E1014 07:35:43.297597 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad17975e-43e4-4fda-86a2-eb50ba104055" containerName="extract-utilities" Oct 14 07:35:43 crc kubenswrapper[4870]: I1014 07:35:43.297605 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad17975e-43e4-4fda-86a2-eb50ba104055" containerName="extract-utilities" Oct 14 07:35:43 crc kubenswrapper[4870]: I1014 07:35:43.297793 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="ad17975e-43e4-4fda-86a2-eb50ba104055" containerName="registry-server" Oct 14 07:35:43 crc kubenswrapper[4870]: I1014 07:35:43.297829 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="83e0f943-ad16-4731-b196-13cb0165e023" containerName="registry-server" Oct 14 07:35:43 crc kubenswrapper[4870]: I1014 07:35:43.299136 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8qwbr" Oct 14 07:35:43 crc kubenswrapper[4870]: I1014 07:35:43.318079 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-8qwbr"] Oct 14 07:35:43 crc kubenswrapper[4870]: I1014 07:35:43.490489 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-78pnt\" (UniqueName: \"kubernetes.io/projected/06528a74-01d4-4143-9dfe-f0c09b3449ea-kube-api-access-78pnt\") pod \"certified-operators-8qwbr\" (UID: \"06528a74-01d4-4143-9dfe-f0c09b3449ea\") " pod="openshift-marketplace/certified-operators-8qwbr" Oct 14 07:35:43 crc kubenswrapper[4870]: I1014 07:35:43.490559 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/06528a74-01d4-4143-9dfe-f0c09b3449ea-utilities\") pod \"certified-operators-8qwbr\" (UID: \"06528a74-01d4-4143-9dfe-f0c09b3449ea\") " pod="openshift-marketplace/certified-operators-8qwbr" Oct 14 07:35:43 crc kubenswrapper[4870]: I1014 07:35:43.490667 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/06528a74-01d4-4143-9dfe-f0c09b3449ea-catalog-content\") pod \"certified-operators-8qwbr\" (UID: \"06528a74-01d4-4143-9dfe-f0c09b3449ea\") " pod="openshift-marketplace/certified-operators-8qwbr" Oct 14 07:35:43 crc kubenswrapper[4870]: I1014 07:35:43.592621 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/06528a74-01d4-4143-9dfe-f0c09b3449ea-catalog-content\") pod \"certified-operators-8qwbr\" (UID: \"06528a74-01d4-4143-9dfe-f0c09b3449ea\") " pod="openshift-marketplace/certified-operators-8qwbr" Oct 14 07:35:43 crc kubenswrapper[4870]: I1014 07:35:43.592864 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-78pnt\" (UniqueName: \"kubernetes.io/projected/06528a74-01d4-4143-9dfe-f0c09b3449ea-kube-api-access-78pnt\") pod \"certified-operators-8qwbr\" (UID: \"06528a74-01d4-4143-9dfe-f0c09b3449ea\") " pod="openshift-marketplace/certified-operators-8qwbr" Oct 14 07:35:43 crc kubenswrapper[4870]: I1014 07:35:43.592922 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/06528a74-01d4-4143-9dfe-f0c09b3449ea-utilities\") pod \"certified-operators-8qwbr\" (UID: \"06528a74-01d4-4143-9dfe-f0c09b3449ea\") " pod="openshift-marketplace/certified-operators-8qwbr" Oct 14 07:35:43 crc kubenswrapper[4870]: I1014 07:35:43.593393 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/06528a74-01d4-4143-9dfe-f0c09b3449ea-utilities\") pod \"certified-operators-8qwbr\" (UID: \"06528a74-01d4-4143-9dfe-f0c09b3449ea\") " pod="openshift-marketplace/certified-operators-8qwbr" Oct 14 07:35:43 crc kubenswrapper[4870]: I1014 07:35:43.593393 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/06528a74-01d4-4143-9dfe-f0c09b3449ea-catalog-content\") pod \"certified-operators-8qwbr\" (UID: \"06528a74-01d4-4143-9dfe-f0c09b3449ea\") " pod="openshift-marketplace/certified-operators-8qwbr" Oct 14 07:35:43 crc kubenswrapper[4870]: I1014 07:35:43.613128 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-78pnt\" (UniqueName: \"kubernetes.io/projected/06528a74-01d4-4143-9dfe-f0c09b3449ea-kube-api-access-78pnt\") pod \"certified-operators-8qwbr\" (UID: \"06528a74-01d4-4143-9dfe-f0c09b3449ea\") " pod="openshift-marketplace/certified-operators-8qwbr" Oct 14 07:35:43 crc kubenswrapper[4870]: I1014 07:35:43.673618 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8qwbr" Oct 14 07:35:44 crc kubenswrapper[4870]: I1014 07:35:44.112537 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-8qwbr"] Oct 14 07:35:44 crc kubenswrapper[4870]: I1014 07:35:44.858227 4870 generic.go:334] "Generic (PLEG): container finished" podID="06528a74-01d4-4143-9dfe-f0c09b3449ea" containerID="341ba3c8e0f9fba7b0a148b39c01a9723ba5e49803038611aed6694de0d3a05a" exitCode=0 Oct 14 07:35:44 crc kubenswrapper[4870]: I1014 07:35:44.858300 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8qwbr" event={"ID":"06528a74-01d4-4143-9dfe-f0c09b3449ea","Type":"ContainerDied","Data":"341ba3c8e0f9fba7b0a148b39c01a9723ba5e49803038611aed6694de0d3a05a"} Oct 14 07:35:44 crc kubenswrapper[4870]: I1014 07:35:44.858635 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8qwbr" event={"ID":"06528a74-01d4-4143-9dfe-f0c09b3449ea","Type":"ContainerStarted","Data":"ea2cff4eafde82db52e9363a46756b4c305b3df58d984dd1bb586c7171a7318d"} Oct 14 07:35:46 crc kubenswrapper[4870]: I1014 07:35:46.644251 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-rv6z7"] Oct 14 07:35:46 crc kubenswrapper[4870]: I1014 07:35:46.646098 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rv6z7" Oct 14 07:35:46 crc kubenswrapper[4870]: I1014 07:35:46.660170 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-rv6z7"] Oct 14 07:35:46 crc kubenswrapper[4870]: I1014 07:35:46.743997 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/15fa7584-619a-405d-b126-6a8c285e3ead-catalog-content\") pod \"redhat-marketplace-rv6z7\" (UID: \"15fa7584-619a-405d-b126-6a8c285e3ead\") " pod="openshift-marketplace/redhat-marketplace-rv6z7" Oct 14 07:35:46 crc kubenswrapper[4870]: I1014 07:35:46.744062 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mnsrt\" (UniqueName: \"kubernetes.io/projected/15fa7584-619a-405d-b126-6a8c285e3ead-kube-api-access-mnsrt\") pod \"redhat-marketplace-rv6z7\" (UID: \"15fa7584-619a-405d-b126-6a8c285e3ead\") " pod="openshift-marketplace/redhat-marketplace-rv6z7" Oct 14 07:35:46 crc kubenswrapper[4870]: I1014 07:35:46.744172 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/15fa7584-619a-405d-b126-6a8c285e3ead-utilities\") pod \"redhat-marketplace-rv6z7\" (UID: \"15fa7584-619a-405d-b126-6a8c285e3ead\") " pod="openshift-marketplace/redhat-marketplace-rv6z7" Oct 14 07:35:46 crc kubenswrapper[4870]: I1014 07:35:46.845167 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/15fa7584-619a-405d-b126-6a8c285e3ead-catalog-content\") pod \"redhat-marketplace-rv6z7\" (UID: \"15fa7584-619a-405d-b126-6a8c285e3ead\") " pod="openshift-marketplace/redhat-marketplace-rv6z7" Oct 14 07:35:46 crc kubenswrapper[4870]: I1014 07:35:46.845224 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mnsrt\" (UniqueName: \"kubernetes.io/projected/15fa7584-619a-405d-b126-6a8c285e3ead-kube-api-access-mnsrt\") pod \"redhat-marketplace-rv6z7\" (UID: \"15fa7584-619a-405d-b126-6a8c285e3ead\") " pod="openshift-marketplace/redhat-marketplace-rv6z7" Oct 14 07:35:46 crc kubenswrapper[4870]: I1014 07:35:46.845287 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/15fa7584-619a-405d-b126-6a8c285e3ead-utilities\") pod \"redhat-marketplace-rv6z7\" (UID: \"15fa7584-619a-405d-b126-6a8c285e3ead\") " pod="openshift-marketplace/redhat-marketplace-rv6z7" Oct 14 07:35:46 crc kubenswrapper[4870]: I1014 07:35:46.845717 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/15fa7584-619a-405d-b126-6a8c285e3ead-catalog-content\") pod \"redhat-marketplace-rv6z7\" (UID: \"15fa7584-619a-405d-b126-6a8c285e3ead\") " pod="openshift-marketplace/redhat-marketplace-rv6z7" Oct 14 07:35:46 crc kubenswrapper[4870]: I1014 07:35:46.845748 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/15fa7584-619a-405d-b126-6a8c285e3ead-utilities\") pod \"redhat-marketplace-rv6z7\" (UID: \"15fa7584-619a-405d-b126-6a8c285e3ead\") " pod="openshift-marketplace/redhat-marketplace-rv6z7" Oct 14 07:35:46 crc kubenswrapper[4870]: I1014 07:35:46.864104 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mnsrt\" (UniqueName: \"kubernetes.io/projected/15fa7584-619a-405d-b126-6a8c285e3ead-kube-api-access-mnsrt\") pod \"redhat-marketplace-rv6z7\" (UID: \"15fa7584-619a-405d-b126-6a8c285e3ead\") " pod="openshift-marketplace/redhat-marketplace-rv6z7" Oct 14 07:35:46 crc kubenswrapper[4870]: I1014 07:35:46.876409 4870 generic.go:334] "Generic (PLEG): container finished" podID="06528a74-01d4-4143-9dfe-f0c09b3449ea" containerID="6b029a3750fb2d20b48f224430e20f51f58c044e8d6cd093d497e3b867ef06aa" exitCode=0 Oct 14 07:35:46 crc kubenswrapper[4870]: I1014 07:35:46.876460 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8qwbr" event={"ID":"06528a74-01d4-4143-9dfe-f0c09b3449ea","Type":"ContainerDied","Data":"6b029a3750fb2d20b48f224430e20f51f58c044e8d6cd093d497e3b867ef06aa"} Oct 14 07:35:46 crc kubenswrapper[4870]: I1014 07:35:46.962051 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rv6z7" Oct 14 07:35:47 crc kubenswrapper[4870]: I1014 07:35:47.388680 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-rv6z7"] Oct 14 07:35:47 crc kubenswrapper[4870]: W1014 07:35:47.396699 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod15fa7584_619a_405d_b126_6a8c285e3ead.slice/crio-df8244ccc326fcb06eb7f53603511d7ba29038bde456ff0bde8e5e184df63962 WatchSource:0}: Error finding container df8244ccc326fcb06eb7f53603511d7ba29038bde456ff0bde8e5e184df63962: Status 404 returned error can't find the container with id df8244ccc326fcb06eb7f53603511d7ba29038bde456ff0bde8e5e184df63962 Oct 14 07:35:47 crc kubenswrapper[4870]: I1014 07:35:47.886666 4870 generic.go:334] "Generic (PLEG): container finished" podID="15fa7584-619a-405d-b126-6a8c285e3ead" containerID="8b001833edb1d99817a52d548986c862d77e45d5bf0e958a451270f8fdc3d948" exitCode=0 Oct 14 07:35:47 crc kubenswrapper[4870]: I1014 07:35:47.886827 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rv6z7" event={"ID":"15fa7584-619a-405d-b126-6a8c285e3ead","Type":"ContainerDied","Data":"8b001833edb1d99817a52d548986c862d77e45d5bf0e958a451270f8fdc3d948"} Oct 14 07:35:47 crc kubenswrapper[4870]: I1014 07:35:47.887003 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rv6z7" event={"ID":"15fa7584-619a-405d-b126-6a8c285e3ead","Type":"ContainerStarted","Data":"df8244ccc326fcb06eb7f53603511d7ba29038bde456ff0bde8e5e184df63962"} Oct 14 07:35:47 crc kubenswrapper[4870]: I1014 07:35:47.891190 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8qwbr" event={"ID":"06528a74-01d4-4143-9dfe-f0c09b3449ea","Type":"ContainerStarted","Data":"87a2010cebc886d5aea82ae9f58d6b28c69f02cb4d9d520f0849e84ab8179e72"} Oct 14 07:35:47 crc kubenswrapper[4870]: I1014 07:35:47.934782 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-8qwbr" podStartSLOduration=2.428577393 podStartE2EDuration="4.934752797s" podCreationTimestamp="2025-10-14 07:35:43 +0000 UTC" firstStartedPulling="2025-10-14 07:35:44.860504561 +0000 UTC m=+2080.557864972" lastFinishedPulling="2025-10-14 07:35:47.366679965 +0000 UTC m=+2083.064040376" observedRunningTime="2025-10-14 07:35:47.927128379 +0000 UTC m=+2083.624488800" watchObservedRunningTime="2025-10-14 07:35:47.934752797 +0000 UTC m=+2083.632113208" Oct 14 07:35:48 crc kubenswrapper[4870]: I1014 07:35:48.903068 4870 generic.go:334] "Generic (PLEG): container finished" podID="15fa7584-619a-405d-b126-6a8c285e3ead" containerID="8f0f32e08863195cf74eb7db5d66a026e87fb0c5f782dedf7f1f0c088ecc59dd" exitCode=0 Oct 14 07:35:48 crc kubenswrapper[4870]: I1014 07:35:48.903166 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rv6z7" event={"ID":"15fa7584-619a-405d-b126-6a8c285e3ead","Type":"ContainerDied","Data":"8f0f32e08863195cf74eb7db5d66a026e87fb0c5f782dedf7f1f0c088ecc59dd"} Oct 14 07:35:49 crc kubenswrapper[4870]: I1014 07:35:49.916622 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rv6z7" event={"ID":"15fa7584-619a-405d-b126-6a8c285e3ead","Type":"ContainerStarted","Data":"31bf812c440f6a59e098f62599f2c91a8f82d0cb0d62b25d1a4e066c5a9de817"} Oct 14 07:35:49 crc kubenswrapper[4870]: I1014 07:35:49.945335 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-rv6z7" podStartSLOduration=2.548627791 podStartE2EDuration="3.945307014s" podCreationTimestamp="2025-10-14 07:35:46 +0000 UTC" firstStartedPulling="2025-10-14 07:35:47.888849065 +0000 UTC m=+2083.586209436" lastFinishedPulling="2025-10-14 07:35:49.285528278 +0000 UTC m=+2084.982888659" observedRunningTime="2025-10-14 07:35:49.938197999 +0000 UTC m=+2085.635558380" watchObservedRunningTime="2025-10-14 07:35:49.945307014 +0000 UTC m=+2085.642667425" Oct 14 07:35:53 crc kubenswrapper[4870]: I1014 07:35:53.673870 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-8qwbr" Oct 14 07:35:53 crc kubenswrapper[4870]: I1014 07:35:53.674238 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-8qwbr" Oct 14 07:35:53 crc kubenswrapper[4870]: I1014 07:35:53.718430 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-8qwbr" Oct 14 07:35:54 crc kubenswrapper[4870]: I1014 07:35:54.011951 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-8qwbr" Oct 14 07:35:54 crc kubenswrapper[4870]: I1014 07:35:54.077403 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-8qwbr"] Oct 14 07:35:55 crc kubenswrapper[4870]: I1014 07:35:55.980172 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-8qwbr" podUID="06528a74-01d4-4143-9dfe-f0c09b3449ea" containerName="registry-server" containerID="cri-o://87a2010cebc886d5aea82ae9f58d6b28c69f02cb4d9d520f0849e84ab8179e72" gracePeriod=2 Oct 14 07:35:56 crc kubenswrapper[4870]: I1014 07:35:56.439937 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8qwbr" Oct 14 07:35:56 crc kubenswrapper[4870]: I1014 07:35:56.614631 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/06528a74-01d4-4143-9dfe-f0c09b3449ea-catalog-content\") pod \"06528a74-01d4-4143-9dfe-f0c09b3449ea\" (UID: \"06528a74-01d4-4143-9dfe-f0c09b3449ea\") " Oct 14 07:35:56 crc kubenswrapper[4870]: I1014 07:35:56.614749 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-78pnt\" (UniqueName: \"kubernetes.io/projected/06528a74-01d4-4143-9dfe-f0c09b3449ea-kube-api-access-78pnt\") pod \"06528a74-01d4-4143-9dfe-f0c09b3449ea\" (UID: \"06528a74-01d4-4143-9dfe-f0c09b3449ea\") " Oct 14 07:35:56 crc kubenswrapper[4870]: I1014 07:35:56.614969 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/06528a74-01d4-4143-9dfe-f0c09b3449ea-utilities\") pod \"06528a74-01d4-4143-9dfe-f0c09b3449ea\" (UID: \"06528a74-01d4-4143-9dfe-f0c09b3449ea\") " Oct 14 07:35:56 crc kubenswrapper[4870]: I1014 07:35:56.616643 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/06528a74-01d4-4143-9dfe-f0c09b3449ea-utilities" (OuterVolumeSpecName: "utilities") pod "06528a74-01d4-4143-9dfe-f0c09b3449ea" (UID: "06528a74-01d4-4143-9dfe-f0c09b3449ea"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:35:56 crc kubenswrapper[4870]: I1014 07:35:56.623409 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/06528a74-01d4-4143-9dfe-f0c09b3449ea-kube-api-access-78pnt" (OuterVolumeSpecName: "kube-api-access-78pnt") pod "06528a74-01d4-4143-9dfe-f0c09b3449ea" (UID: "06528a74-01d4-4143-9dfe-f0c09b3449ea"). InnerVolumeSpecName "kube-api-access-78pnt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:35:56 crc kubenswrapper[4870]: I1014 07:35:56.675348 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/06528a74-01d4-4143-9dfe-f0c09b3449ea-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "06528a74-01d4-4143-9dfe-f0c09b3449ea" (UID: "06528a74-01d4-4143-9dfe-f0c09b3449ea"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:35:56 crc kubenswrapper[4870]: I1014 07:35:56.717122 4870 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/06528a74-01d4-4143-9dfe-f0c09b3449ea-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 07:35:56 crc kubenswrapper[4870]: I1014 07:35:56.717211 4870 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/06528a74-01d4-4143-9dfe-f0c09b3449ea-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 07:35:56 crc kubenswrapper[4870]: I1014 07:35:56.717233 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-78pnt\" (UniqueName: \"kubernetes.io/projected/06528a74-01d4-4143-9dfe-f0c09b3449ea-kube-api-access-78pnt\") on node \"crc\" DevicePath \"\"" Oct 14 07:35:56 crc kubenswrapper[4870]: I1014 07:35:56.962362 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-rv6z7" Oct 14 07:35:56 crc kubenswrapper[4870]: I1014 07:35:56.962483 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-rv6z7" Oct 14 07:35:56 crc kubenswrapper[4870]: I1014 07:35:56.996208 4870 generic.go:334] "Generic (PLEG): container finished" podID="06528a74-01d4-4143-9dfe-f0c09b3449ea" containerID="87a2010cebc886d5aea82ae9f58d6b28c69f02cb4d9d520f0849e84ab8179e72" exitCode=0 Oct 14 07:35:56 crc kubenswrapper[4870]: I1014 07:35:56.996310 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8qwbr" Oct 14 07:35:56 crc kubenswrapper[4870]: I1014 07:35:56.996309 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8qwbr" event={"ID":"06528a74-01d4-4143-9dfe-f0c09b3449ea","Type":"ContainerDied","Data":"87a2010cebc886d5aea82ae9f58d6b28c69f02cb4d9d520f0849e84ab8179e72"} Oct 14 07:35:56 crc kubenswrapper[4870]: I1014 07:35:56.996405 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8qwbr" event={"ID":"06528a74-01d4-4143-9dfe-f0c09b3449ea","Type":"ContainerDied","Data":"ea2cff4eafde82db52e9363a46756b4c305b3df58d984dd1bb586c7171a7318d"} Oct 14 07:35:56 crc kubenswrapper[4870]: I1014 07:35:56.996465 4870 scope.go:117] "RemoveContainer" containerID="87a2010cebc886d5aea82ae9f58d6b28c69f02cb4d9d520f0849e84ab8179e72" Oct 14 07:35:57 crc kubenswrapper[4870]: I1014 07:35:57.037561 4870 scope.go:117] "RemoveContainer" containerID="6b029a3750fb2d20b48f224430e20f51f58c044e8d6cd093d497e3b867ef06aa" Oct 14 07:35:57 crc kubenswrapper[4870]: I1014 07:35:57.080521 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-rv6z7" Oct 14 07:35:57 crc kubenswrapper[4870]: I1014 07:35:57.080566 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-8qwbr"] Oct 14 07:35:57 crc kubenswrapper[4870]: I1014 07:35:57.080584 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-8qwbr"] Oct 14 07:35:57 crc kubenswrapper[4870]: I1014 07:35:57.095897 4870 scope.go:117] "RemoveContainer" containerID="341ba3c8e0f9fba7b0a148b39c01a9723ba5e49803038611aed6694de0d3a05a" Oct 14 07:35:57 crc kubenswrapper[4870]: I1014 07:35:57.112223 4870 scope.go:117] "RemoveContainer" containerID="87a2010cebc886d5aea82ae9f58d6b28c69f02cb4d9d520f0849e84ab8179e72" Oct 14 07:35:57 crc kubenswrapper[4870]: E1014 07:35:57.116573 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"87a2010cebc886d5aea82ae9f58d6b28c69f02cb4d9d520f0849e84ab8179e72\": container with ID starting with 87a2010cebc886d5aea82ae9f58d6b28c69f02cb4d9d520f0849e84ab8179e72 not found: ID does not exist" containerID="87a2010cebc886d5aea82ae9f58d6b28c69f02cb4d9d520f0849e84ab8179e72" Oct 14 07:35:57 crc kubenswrapper[4870]: I1014 07:35:57.116627 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"87a2010cebc886d5aea82ae9f58d6b28c69f02cb4d9d520f0849e84ab8179e72"} err="failed to get container status \"87a2010cebc886d5aea82ae9f58d6b28c69f02cb4d9d520f0849e84ab8179e72\": rpc error: code = NotFound desc = could not find container \"87a2010cebc886d5aea82ae9f58d6b28c69f02cb4d9d520f0849e84ab8179e72\": container with ID starting with 87a2010cebc886d5aea82ae9f58d6b28c69f02cb4d9d520f0849e84ab8179e72 not found: ID does not exist" Oct 14 07:35:57 crc kubenswrapper[4870]: I1014 07:35:57.116659 4870 scope.go:117] "RemoveContainer" containerID="6b029a3750fb2d20b48f224430e20f51f58c044e8d6cd093d497e3b867ef06aa" Oct 14 07:35:57 crc kubenswrapper[4870]: E1014 07:35:57.123562 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6b029a3750fb2d20b48f224430e20f51f58c044e8d6cd093d497e3b867ef06aa\": container with ID starting with 6b029a3750fb2d20b48f224430e20f51f58c044e8d6cd093d497e3b867ef06aa not found: ID does not exist" containerID="6b029a3750fb2d20b48f224430e20f51f58c044e8d6cd093d497e3b867ef06aa" Oct 14 07:35:57 crc kubenswrapper[4870]: I1014 07:35:57.123628 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6b029a3750fb2d20b48f224430e20f51f58c044e8d6cd093d497e3b867ef06aa"} err="failed to get container status \"6b029a3750fb2d20b48f224430e20f51f58c044e8d6cd093d497e3b867ef06aa\": rpc error: code = NotFound desc = could not find container \"6b029a3750fb2d20b48f224430e20f51f58c044e8d6cd093d497e3b867ef06aa\": container with ID starting with 6b029a3750fb2d20b48f224430e20f51f58c044e8d6cd093d497e3b867ef06aa not found: ID does not exist" Oct 14 07:35:57 crc kubenswrapper[4870]: I1014 07:35:57.123656 4870 scope.go:117] "RemoveContainer" containerID="341ba3c8e0f9fba7b0a148b39c01a9723ba5e49803038611aed6694de0d3a05a" Oct 14 07:35:57 crc kubenswrapper[4870]: E1014 07:35:57.125830 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"341ba3c8e0f9fba7b0a148b39c01a9723ba5e49803038611aed6694de0d3a05a\": container with ID starting with 341ba3c8e0f9fba7b0a148b39c01a9723ba5e49803038611aed6694de0d3a05a not found: ID does not exist" containerID="341ba3c8e0f9fba7b0a148b39c01a9723ba5e49803038611aed6694de0d3a05a" Oct 14 07:35:57 crc kubenswrapper[4870]: I1014 07:35:57.125854 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"341ba3c8e0f9fba7b0a148b39c01a9723ba5e49803038611aed6694de0d3a05a"} err="failed to get container status \"341ba3c8e0f9fba7b0a148b39c01a9723ba5e49803038611aed6694de0d3a05a\": rpc error: code = NotFound desc = could not find container \"341ba3c8e0f9fba7b0a148b39c01a9723ba5e49803038611aed6694de0d3a05a\": container with ID starting with 341ba3c8e0f9fba7b0a148b39c01a9723ba5e49803038611aed6694de0d3a05a not found: ID does not exist" Oct 14 07:35:57 crc kubenswrapper[4870]: I1014 07:35:57.138126 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-rv6z7" Oct 14 07:35:59 crc kubenswrapper[4870]: I1014 07:35:59.047760 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="06528a74-01d4-4143-9dfe-f0c09b3449ea" path="/var/lib/kubelet/pods/06528a74-01d4-4143-9dfe-f0c09b3449ea/volumes" Oct 14 07:35:59 crc kubenswrapper[4870]: I1014 07:35:59.361104 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-rv6z7"] Oct 14 07:35:59 crc kubenswrapper[4870]: I1014 07:35:59.361388 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-rv6z7" podUID="15fa7584-619a-405d-b126-6a8c285e3ead" containerName="registry-server" containerID="cri-o://31bf812c440f6a59e098f62599f2c91a8f82d0cb0d62b25d1a4e066c5a9de817" gracePeriod=2 Oct 14 07:35:59 crc kubenswrapper[4870]: I1014 07:35:59.809950 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rv6z7" Oct 14 07:35:59 crc kubenswrapper[4870]: I1014 07:35:59.977837 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/15fa7584-619a-405d-b126-6a8c285e3ead-utilities\") pod \"15fa7584-619a-405d-b126-6a8c285e3ead\" (UID: \"15fa7584-619a-405d-b126-6a8c285e3ead\") " Oct 14 07:35:59 crc kubenswrapper[4870]: I1014 07:35:59.978003 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/15fa7584-619a-405d-b126-6a8c285e3ead-catalog-content\") pod \"15fa7584-619a-405d-b126-6a8c285e3ead\" (UID: \"15fa7584-619a-405d-b126-6a8c285e3ead\") " Oct 14 07:35:59 crc kubenswrapper[4870]: I1014 07:35:59.978073 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnsrt\" (UniqueName: \"kubernetes.io/projected/15fa7584-619a-405d-b126-6a8c285e3ead-kube-api-access-mnsrt\") pod \"15fa7584-619a-405d-b126-6a8c285e3ead\" (UID: \"15fa7584-619a-405d-b126-6a8c285e3ead\") " Oct 14 07:35:59 crc kubenswrapper[4870]: I1014 07:35:59.979598 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/15fa7584-619a-405d-b126-6a8c285e3ead-utilities" (OuterVolumeSpecName: "utilities") pod "15fa7584-619a-405d-b126-6a8c285e3ead" (UID: "15fa7584-619a-405d-b126-6a8c285e3ead"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:35:59 crc kubenswrapper[4870]: I1014 07:35:59.985326 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/15fa7584-619a-405d-b126-6a8c285e3ead-kube-api-access-mnsrt" (OuterVolumeSpecName: "kube-api-access-mnsrt") pod "15fa7584-619a-405d-b126-6a8c285e3ead" (UID: "15fa7584-619a-405d-b126-6a8c285e3ead"). InnerVolumeSpecName "kube-api-access-mnsrt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:35:59 crc kubenswrapper[4870]: I1014 07:35:59.996562 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/15fa7584-619a-405d-b126-6a8c285e3ead-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "15fa7584-619a-405d-b126-6a8c285e3ead" (UID: "15fa7584-619a-405d-b126-6a8c285e3ead"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:36:00 crc kubenswrapper[4870]: I1014 07:36:00.047787 4870 generic.go:334] "Generic (PLEG): container finished" podID="15fa7584-619a-405d-b126-6a8c285e3ead" containerID="31bf812c440f6a59e098f62599f2c91a8f82d0cb0d62b25d1a4e066c5a9de817" exitCode=0 Oct 14 07:36:00 crc kubenswrapper[4870]: I1014 07:36:00.047884 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rv6z7" event={"ID":"15fa7584-619a-405d-b126-6a8c285e3ead","Type":"ContainerDied","Data":"31bf812c440f6a59e098f62599f2c91a8f82d0cb0d62b25d1a4e066c5a9de817"} Oct 14 07:36:00 crc kubenswrapper[4870]: I1014 07:36:00.047942 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rv6z7" event={"ID":"15fa7584-619a-405d-b126-6a8c285e3ead","Type":"ContainerDied","Data":"df8244ccc326fcb06eb7f53603511d7ba29038bde456ff0bde8e5e184df63962"} Oct 14 07:36:00 crc kubenswrapper[4870]: I1014 07:36:00.048001 4870 scope.go:117] "RemoveContainer" containerID="31bf812c440f6a59e098f62599f2c91a8f82d0cb0d62b25d1a4e066c5a9de817" Oct 14 07:36:00 crc kubenswrapper[4870]: I1014 07:36:00.048339 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rv6z7" Oct 14 07:36:00 crc kubenswrapper[4870]: I1014 07:36:00.066620 4870 scope.go:117] "RemoveContainer" containerID="8f0f32e08863195cf74eb7db5d66a026e87fb0c5f782dedf7f1f0c088ecc59dd" Oct 14 07:36:00 crc kubenswrapper[4870]: I1014 07:36:00.082160 4870 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/15fa7584-619a-405d-b126-6a8c285e3ead-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 07:36:00 crc kubenswrapper[4870]: I1014 07:36:00.082207 4870 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/15fa7584-619a-405d-b126-6a8c285e3ead-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 07:36:00 crc kubenswrapper[4870]: I1014 07:36:00.082224 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnsrt\" (UniqueName: \"kubernetes.io/projected/15fa7584-619a-405d-b126-6a8c285e3ead-kube-api-access-mnsrt\") on node \"crc\" DevicePath \"\"" Oct 14 07:36:00 crc kubenswrapper[4870]: I1014 07:36:00.096870 4870 scope.go:117] "RemoveContainer" containerID="8b001833edb1d99817a52d548986c862d77e45d5bf0e958a451270f8fdc3d948" Oct 14 07:36:00 crc kubenswrapper[4870]: I1014 07:36:00.099120 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-rv6z7"] Oct 14 07:36:00 crc kubenswrapper[4870]: I1014 07:36:00.104467 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-rv6z7"] Oct 14 07:36:00 crc kubenswrapper[4870]: I1014 07:36:00.118632 4870 scope.go:117] "RemoveContainer" containerID="31bf812c440f6a59e098f62599f2c91a8f82d0cb0d62b25d1a4e066c5a9de817" Oct 14 07:36:00 crc kubenswrapper[4870]: E1014 07:36:00.119123 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"31bf812c440f6a59e098f62599f2c91a8f82d0cb0d62b25d1a4e066c5a9de817\": container with ID starting with 31bf812c440f6a59e098f62599f2c91a8f82d0cb0d62b25d1a4e066c5a9de817 not found: ID does not exist" containerID="31bf812c440f6a59e098f62599f2c91a8f82d0cb0d62b25d1a4e066c5a9de817" Oct 14 07:36:00 crc kubenswrapper[4870]: I1014 07:36:00.119167 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"31bf812c440f6a59e098f62599f2c91a8f82d0cb0d62b25d1a4e066c5a9de817"} err="failed to get container status \"31bf812c440f6a59e098f62599f2c91a8f82d0cb0d62b25d1a4e066c5a9de817\": rpc error: code = NotFound desc = could not find container \"31bf812c440f6a59e098f62599f2c91a8f82d0cb0d62b25d1a4e066c5a9de817\": container with ID starting with 31bf812c440f6a59e098f62599f2c91a8f82d0cb0d62b25d1a4e066c5a9de817 not found: ID does not exist" Oct 14 07:36:00 crc kubenswrapper[4870]: I1014 07:36:00.119197 4870 scope.go:117] "RemoveContainer" containerID="8f0f32e08863195cf74eb7db5d66a026e87fb0c5f782dedf7f1f0c088ecc59dd" Oct 14 07:36:00 crc kubenswrapper[4870]: E1014 07:36:00.119653 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8f0f32e08863195cf74eb7db5d66a026e87fb0c5f782dedf7f1f0c088ecc59dd\": container with ID starting with 8f0f32e08863195cf74eb7db5d66a026e87fb0c5f782dedf7f1f0c088ecc59dd not found: ID does not exist" containerID="8f0f32e08863195cf74eb7db5d66a026e87fb0c5f782dedf7f1f0c088ecc59dd" Oct 14 07:36:00 crc kubenswrapper[4870]: I1014 07:36:00.119678 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8f0f32e08863195cf74eb7db5d66a026e87fb0c5f782dedf7f1f0c088ecc59dd"} err="failed to get container status \"8f0f32e08863195cf74eb7db5d66a026e87fb0c5f782dedf7f1f0c088ecc59dd\": rpc error: code = NotFound desc = could not find container \"8f0f32e08863195cf74eb7db5d66a026e87fb0c5f782dedf7f1f0c088ecc59dd\": container with ID starting with 8f0f32e08863195cf74eb7db5d66a026e87fb0c5f782dedf7f1f0c088ecc59dd not found: ID does not exist" Oct 14 07:36:00 crc kubenswrapper[4870]: I1014 07:36:00.119693 4870 scope.go:117] "RemoveContainer" containerID="8b001833edb1d99817a52d548986c862d77e45d5bf0e958a451270f8fdc3d948" Oct 14 07:36:00 crc kubenswrapper[4870]: E1014 07:36:00.120060 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8b001833edb1d99817a52d548986c862d77e45d5bf0e958a451270f8fdc3d948\": container with ID starting with 8b001833edb1d99817a52d548986c862d77e45d5bf0e958a451270f8fdc3d948 not found: ID does not exist" containerID="8b001833edb1d99817a52d548986c862d77e45d5bf0e958a451270f8fdc3d948" Oct 14 07:36:00 crc kubenswrapper[4870]: I1014 07:36:00.120084 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8b001833edb1d99817a52d548986c862d77e45d5bf0e958a451270f8fdc3d948"} err="failed to get container status \"8b001833edb1d99817a52d548986c862d77e45d5bf0e958a451270f8fdc3d948\": rpc error: code = NotFound desc = could not find container \"8b001833edb1d99817a52d548986c862d77e45d5bf0e958a451270f8fdc3d948\": container with ID starting with 8b001833edb1d99817a52d548986c862d77e45d5bf0e958a451270f8fdc3d948 not found: ID does not exist" Oct 14 07:36:01 crc kubenswrapper[4870]: I1014 07:36:01.056985 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="15fa7584-619a-405d-b126-6a8c285e3ead" path="/var/lib/kubelet/pods/15fa7584-619a-405d-b126-6a8c285e3ead/volumes" Oct 14 07:36:53 crc kubenswrapper[4870]: I1014 07:36:53.951686 4870 patch_prober.go:28] interesting pod/machine-config-daemon-7tvc7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 07:36:53 crc kubenswrapper[4870]: I1014 07:36:53.952422 4870 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 07:37:23 crc kubenswrapper[4870]: I1014 07:37:23.951138 4870 patch_prober.go:28] interesting pod/machine-config-daemon-7tvc7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 07:37:23 crc kubenswrapper[4870]: I1014 07:37:23.951991 4870 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 07:37:53 crc kubenswrapper[4870]: I1014 07:37:53.951412 4870 patch_prober.go:28] interesting pod/machine-config-daemon-7tvc7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 07:37:53 crc kubenswrapper[4870]: I1014 07:37:53.952153 4870 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 07:37:53 crc kubenswrapper[4870]: I1014 07:37:53.952237 4870 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" Oct 14 07:37:53 crc kubenswrapper[4870]: I1014 07:37:53.953169 4870 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"13bbb5c4f6c1ff5e8c8987841a34cf3435bc534c235a0bc4bb5b02514ede6e5e"} pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 14 07:37:53 crc kubenswrapper[4870]: I1014 07:37:53.953271 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerName="machine-config-daemon" containerID="cri-o://13bbb5c4f6c1ff5e8c8987841a34cf3435bc534c235a0bc4bb5b02514ede6e5e" gracePeriod=600 Oct 14 07:37:54 crc kubenswrapper[4870]: I1014 07:37:54.175829 4870 generic.go:334] "Generic (PLEG): container finished" podID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerID="13bbb5c4f6c1ff5e8c8987841a34cf3435bc534c235a0bc4bb5b02514ede6e5e" exitCode=0 Oct 14 07:37:54 crc kubenswrapper[4870]: I1014 07:37:54.175921 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" event={"ID":"874a5e45-dffd-4d17-b609-b7d3ed2eab07","Type":"ContainerDied","Data":"13bbb5c4f6c1ff5e8c8987841a34cf3435bc534c235a0bc4bb5b02514ede6e5e"} Oct 14 07:37:54 crc kubenswrapper[4870]: I1014 07:37:54.176311 4870 scope.go:117] "RemoveContainer" containerID="d754019673e24a68a5e4e484dd11d6018830e5970b17109dcba3bd998a84db74" Oct 14 07:37:55 crc kubenswrapper[4870]: I1014 07:37:55.190526 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" event={"ID":"874a5e45-dffd-4d17-b609-b7d3ed2eab07","Type":"ContainerStarted","Data":"0c23f76241656fa35961005192bf783fc346c05081f076194d712c820a96b05d"} Oct 14 07:40:23 crc kubenswrapper[4870]: I1014 07:40:23.950833 4870 patch_prober.go:28] interesting pod/machine-config-daemon-7tvc7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 07:40:23 crc kubenswrapper[4870]: I1014 07:40:23.951494 4870 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 07:40:53 crc kubenswrapper[4870]: I1014 07:40:53.950978 4870 patch_prober.go:28] interesting pod/machine-config-daemon-7tvc7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 07:40:53 crc kubenswrapper[4870]: I1014 07:40:53.951609 4870 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 07:41:23 crc kubenswrapper[4870]: I1014 07:41:23.951274 4870 patch_prober.go:28] interesting pod/machine-config-daemon-7tvc7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 07:41:23 crc kubenswrapper[4870]: I1014 07:41:23.951897 4870 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 07:41:23 crc kubenswrapper[4870]: I1014 07:41:23.951947 4870 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" Oct 14 07:41:23 crc kubenswrapper[4870]: I1014 07:41:23.952603 4870 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"0c23f76241656fa35961005192bf783fc346c05081f076194d712c820a96b05d"} pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 14 07:41:23 crc kubenswrapper[4870]: I1014 07:41:23.952666 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerName="machine-config-daemon" containerID="cri-o://0c23f76241656fa35961005192bf783fc346c05081f076194d712c820a96b05d" gracePeriod=600 Oct 14 07:41:24 crc kubenswrapper[4870]: E1014 07:41:24.080139 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 07:41:24 crc kubenswrapper[4870]: I1014 07:41:24.107608 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" event={"ID":"874a5e45-dffd-4d17-b609-b7d3ed2eab07","Type":"ContainerDied","Data":"0c23f76241656fa35961005192bf783fc346c05081f076194d712c820a96b05d"} Oct 14 07:41:24 crc kubenswrapper[4870]: I1014 07:41:24.107684 4870 scope.go:117] "RemoveContainer" containerID="13bbb5c4f6c1ff5e8c8987841a34cf3435bc534c235a0bc4bb5b02514ede6e5e" Oct 14 07:41:24 crc kubenswrapper[4870]: I1014 07:41:24.107718 4870 generic.go:334] "Generic (PLEG): container finished" podID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerID="0c23f76241656fa35961005192bf783fc346c05081f076194d712c820a96b05d" exitCode=0 Oct 14 07:41:24 crc kubenswrapper[4870]: I1014 07:41:24.108428 4870 scope.go:117] "RemoveContainer" containerID="0c23f76241656fa35961005192bf783fc346c05081f076194d712c820a96b05d" Oct 14 07:41:24 crc kubenswrapper[4870]: E1014 07:41:24.108791 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 07:41:39 crc kubenswrapper[4870]: I1014 07:41:39.034620 4870 scope.go:117] "RemoveContainer" containerID="0c23f76241656fa35961005192bf783fc346c05081f076194d712c820a96b05d" Oct 14 07:41:39 crc kubenswrapper[4870]: E1014 07:41:39.035598 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 07:41:53 crc kubenswrapper[4870]: I1014 07:41:53.034982 4870 scope.go:117] "RemoveContainer" containerID="0c23f76241656fa35961005192bf783fc346c05081f076194d712c820a96b05d" Oct 14 07:41:53 crc kubenswrapper[4870]: E1014 07:41:53.036104 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 07:42:08 crc kubenswrapper[4870]: I1014 07:42:08.034259 4870 scope.go:117] "RemoveContainer" containerID="0c23f76241656fa35961005192bf783fc346c05081f076194d712c820a96b05d" Oct 14 07:42:08 crc kubenswrapper[4870]: E1014 07:42:08.035301 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 07:42:21 crc kubenswrapper[4870]: I1014 07:42:21.034668 4870 scope.go:117] "RemoveContainer" containerID="0c23f76241656fa35961005192bf783fc346c05081f076194d712c820a96b05d" Oct 14 07:42:21 crc kubenswrapper[4870]: E1014 07:42:21.035655 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 07:42:36 crc kubenswrapper[4870]: I1014 07:42:36.034668 4870 scope.go:117] "RemoveContainer" containerID="0c23f76241656fa35961005192bf783fc346c05081f076194d712c820a96b05d" Oct 14 07:42:36 crc kubenswrapper[4870]: E1014 07:42:36.035389 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 07:42:50 crc kubenswrapper[4870]: I1014 07:42:50.034754 4870 scope.go:117] "RemoveContainer" containerID="0c23f76241656fa35961005192bf783fc346c05081f076194d712c820a96b05d" Oct 14 07:42:50 crc kubenswrapper[4870]: E1014 07:42:50.037649 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 07:43:02 crc kubenswrapper[4870]: I1014 07:43:02.035825 4870 scope.go:117] "RemoveContainer" containerID="0c23f76241656fa35961005192bf783fc346c05081f076194d712c820a96b05d" Oct 14 07:43:02 crc kubenswrapper[4870]: E1014 07:43:02.037171 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 07:43:14 crc kubenswrapper[4870]: I1014 07:43:14.035014 4870 scope.go:117] "RemoveContainer" containerID="0c23f76241656fa35961005192bf783fc346c05081f076194d712c820a96b05d" Oct 14 07:43:14 crc kubenswrapper[4870]: E1014 07:43:14.036294 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 07:43:26 crc kubenswrapper[4870]: I1014 07:43:26.034385 4870 scope.go:117] "RemoveContainer" containerID="0c23f76241656fa35961005192bf783fc346c05081f076194d712c820a96b05d" Oct 14 07:43:26 crc kubenswrapper[4870]: E1014 07:43:26.035556 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 07:43:39 crc kubenswrapper[4870]: I1014 07:43:39.033370 4870 scope.go:117] "RemoveContainer" containerID="0c23f76241656fa35961005192bf783fc346c05081f076194d712c820a96b05d" Oct 14 07:43:39 crc kubenswrapper[4870]: E1014 07:43:39.034423 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 07:43:54 crc kubenswrapper[4870]: I1014 07:43:54.034251 4870 scope.go:117] "RemoveContainer" containerID="0c23f76241656fa35961005192bf783fc346c05081f076194d712c820a96b05d" Oct 14 07:43:54 crc kubenswrapper[4870]: E1014 07:43:54.035109 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 07:44:09 crc kubenswrapper[4870]: I1014 07:44:09.033915 4870 scope.go:117] "RemoveContainer" containerID="0c23f76241656fa35961005192bf783fc346c05081f076194d712c820a96b05d" Oct 14 07:44:09 crc kubenswrapper[4870]: E1014 07:44:09.034965 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 07:44:23 crc kubenswrapper[4870]: I1014 07:44:23.034728 4870 scope.go:117] "RemoveContainer" containerID="0c23f76241656fa35961005192bf783fc346c05081f076194d712c820a96b05d" Oct 14 07:44:23 crc kubenswrapper[4870]: E1014 07:44:23.036027 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 07:44:34 crc kubenswrapper[4870]: I1014 07:44:34.034210 4870 scope.go:117] "RemoveContainer" containerID="0c23f76241656fa35961005192bf783fc346c05081f076194d712c820a96b05d" Oct 14 07:44:34 crc kubenswrapper[4870]: E1014 07:44:34.035288 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 07:44:46 crc kubenswrapper[4870]: I1014 07:44:46.034640 4870 scope.go:117] "RemoveContainer" containerID="0c23f76241656fa35961005192bf783fc346c05081f076194d712c820a96b05d" Oct 14 07:44:46 crc kubenswrapper[4870]: E1014 07:44:46.035706 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 07:44:57 crc kubenswrapper[4870]: I1014 07:44:57.034378 4870 scope.go:117] "RemoveContainer" containerID="0c23f76241656fa35961005192bf783fc346c05081f076194d712c820a96b05d" Oct 14 07:44:57 crc kubenswrapper[4870]: E1014 07:44:57.035688 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 07:45:00 crc kubenswrapper[4870]: I1014 07:45:00.168933 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29340465-nfz9t"] Oct 14 07:45:00 crc kubenswrapper[4870]: E1014 07:45:00.170059 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="06528a74-01d4-4143-9dfe-f0c09b3449ea" containerName="extract-utilities" Oct 14 07:45:00 crc kubenswrapper[4870]: I1014 07:45:00.170079 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="06528a74-01d4-4143-9dfe-f0c09b3449ea" containerName="extract-utilities" Oct 14 07:45:00 crc kubenswrapper[4870]: E1014 07:45:00.170104 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15fa7584-619a-405d-b126-6a8c285e3ead" containerName="extract-content" Oct 14 07:45:00 crc kubenswrapper[4870]: I1014 07:45:00.170115 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="15fa7584-619a-405d-b126-6a8c285e3ead" containerName="extract-content" Oct 14 07:45:00 crc kubenswrapper[4870]: E1014 07:45:00.170130 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15fa7584-619a-405d-b126-6a8c285e3ead" containerName="extract-utilities" Oct 14 07:45:00 crc kubenswrapper[4870]: I1014 07:45:00.170141 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="15fa7584-619a-405d-b126-6a8c285e3ead" containerName="extract-utilities" Oct 14 07:45:00 crc kubenswrapper[4870]: E1014 07:45:00.170160 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="06528a74-01d4-4143-9dfe-f0c09b3449ea" containerName="registry-server" Oct 14 07:45:00 crc kubenswrapper[4870]: I1014 07:45:00.170170 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="06528a74-01d4-4143-9dfe-f0c09b3449ea" containerName="registry-server" Oct 14 07:45:00 crc kubenswrapper[4870]: E1014 07:45:00.170218 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="06528a74-01d4-4143-9dfe-f0c09b3449ea" containerName="extract-content" Oct 14 07:45:00 crc kubenswrapper[4870]: I1014 07:45:00.170228 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="06528a74-01d4-4143-9dfe-f0c09b3449ea" containerName="extract-content" Oct 14 07:45:00 crc kubenswrapper[4870]: E1014 07:45:00.170257 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15fa7584-619a-405d-b126-6a8c285e3ead" containerName="registry-server" Oct 14 07:45:00 crc kubenswrapper[4870]: I1014 07:45:00.170267 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="15fa7584-619a-405d-b126-6a8c285e3ead" containerName="registry-server" Oct 14 07:45:00 crc kubenswrapper[4870]: I1014 07:45:00.170524 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="06528a74-01d4-4143-9dfe-f0c09b3449ea" containerName="registry-server" Oct 14 07:45:00 crc kubenswrapper[4870]: I1014 07:45:00.170556 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="15fa7584-619a-405d-b126-6a8c285e3ead" containerName="registry-server" Oct 14 07:45:00 crc kubenswrapper[4870]: I1014 07:45:00.171304 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29340465-nfz9t" Oct 14 07:45:00 crc kubenswrapper[4870]: I1014 07:45:00.173063 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mxspq\" (UniqueName: \"kubernetes.io/projected/610412a5-52e1-4aa6-9d19-8f8a232eec01-kube-api-access-mxspq\") pod \"collect-profiles-29340465-nfz9t\" (UID: \"610412a5-52e1-4aa6-9d19-8f8a232eec01\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340465-nfz9t" Oct 14 07:45:00 crc kubenswrapper[4870]: I1014 07:45:00.173192 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/610412a5-52e1-4aa6-9d19-8f8a232eec01-config-volume\") pod \"collect-profiles-29340465-nfz9t\" (UID: \"610412a5-52e1-4aa6-9d19-8f8a232eec01\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340465-nfz9t" Oct 14 07:45:00 crc kubenswrapper[4870]: I1014 07:45:00.173281 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/610412a5-52e1-4aa6-9d19-8f8a232eec01-secret-volume\") pod \"collect-profiles-29340465-nfz9t\" (UID: \"610412a5-52e1-4aa6-9d19-8f8a232eec01\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340465-nfz9t" Oct 14 07:45:00 crc kubenswrapper[4870]: I1014 07:45:00.174335 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 14 07:45:00 crc kubenswrapper[4870]: I1014 07:45:00.175267 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 14 07:45:00 crc kubenswrapper[4870]: I1014 07:45:00.192107 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29340465-nfz9t"] Oct 14 07:45:00 crc kubenswrapper[4870]: I1014 07:45:00.274261 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/610412a5-52e1-4aa6-9d19-8f8a232eec01-config-volume\") pod \"collect-profiles-29340465-nfz9t\" (UID: \"610412a5-52e1-4aa6-9d19-8f8a232eec01\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340465-nfz9t" Oct 14 07:45:00 crc kubenswrapper[4870]: I1014 07:45:00.274325 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/610412a5-52e1-4aa6-9d19-8f8a232eec01-secret-volume\") pod \"collect-profiles-29340465-nfz9t\" (UID: \"610412a5-52e1-4aa6-9d19-8f8a232eec01\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340465-nfz9t" Oct 14 07:45:00 crc kubenswrapper[4870]: I1014 07:45:00.274388 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mxspq\" (UniqueName: \"kubernetes.io/projected/610412a5-52e1-4aa6-9d19-8f8a232eec01-kube-api-access-mxspq\") pod \"collect-profiles-29340465-nfz9t\" (UID: \"610412a5-52e1-4aa6-9d19-8f8a232eec01\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340465-nfz9t" Oct 14 07:45:00 crc kubenswrapper[4870]: I1014 07:45:00.275580 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/610412a5-52e1-4aa6-9d19-8f8a232eec01-config-volume\") pod \"collect-profiles-29340465-nfz9t\" (UID: \"610412a5-52e1-4aa6-9d19-8f8a232eec01\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340465-nfz9t" Oct 14 07:45:00 crc kubenswrapper[4870]: I1014 07:45:00.281904 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/610412a5-52e1-4aa6-9d19-8f8a232eec01-secret-volume\") pod \"collect-profiles-29340465-nfz9t\" (UID: \"610412a5-52e1-4aa6-9d19-8f8a232eec01\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340465-nfz9t" Oct 14 07:45:00 crc kubenswrapper[4870]: I1014 07:45:00.305009 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mxspq\" (UniqueName: \"kubernetes.io/projected/610412a5-52e1-4aa6-9d19-8f8a232eec01-kube-api-access-mxspq\") pod \"collect-profiles-29340465-nfz9t\" (UID: \"610412a5-52e1-4aa6-9d19-8f8a232eec01\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340465-nfz9t" Oct 14 07:45:00 crc kubenswrapper[4870]: I1014 07:45:00.494426 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29340465-nfz9t" Oct 14 07:45:01 crc kubenswrapper[4870]: I1014 07:45:01.071189 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29340465-nfz9t"] Oct 14 07:45:01 crc kubenswrapper[4870]: I1014 07:45:01.110459 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29340465-nfz9t" event={"ID":"610412a5-52e1-4aa6-9d19-8f8a232eec01","Type":"ContainerStarted","Data":"fe484b85396e71337d718fd469d1716aa2d85c3ac310cc6e48b1c7e739f8e640"} Oct 14 07:45:02 crc kubenswrapper[4870]: I1014 07:45:02.126912 4870 generic.go:334] "Generic (PLEG): container finished" podID="610412a5-52e1-4aa6-9d19-8f8a232eec01" containerID="2002983608d0a8250ea40b794c7803f7a6d8c52e7b23c7f7a9870044d70831c2" exitCode=0 Oct 14 07:45:02 crc kubenswrapper[4870]: I1014 07:45:02.127042 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29340465-nfz9t" event={"ID":"610412a5-52e1-4aa6-9d19-8f8a232eec01","Type":"ContainerDied","Data":"2002983608d0a8250ea40b794c7803f7a6d8c52e7b23c7f7a9870044d70831c2"} Oct 14 07:45:03 crc kubenswrapper[4870]: I1014 07:45:03.537775 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29340465-nfz9t" Oct 14 07:45:03 crc kubenswrapper[4870]: I1014 07:45:03.633476 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mxspq\" (UniqueName: \"kubernetes.io/projected/610412a5-52e1-4aa6-9d19-8f8a232eec01-kube-api-access-mxspq\") pod \"610412a5-52e1-4aa6-9d19-8f8a232eec01\" (UID: \"610412a5-52e1-4aa6-9d19-8f8a232eec01\") " Oct 14 07:45:03 crc kubenswrapper[4870]: I1014 07:45:03.633575 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/610412a5-52e1-4aa6-9d19-8f8a232eec01-secret-volume\") pod \"610412a5-52e1-4aa6-9d19-8f8a232eec01\" (UID: \"610412a5-52e1-4aa6-9d19-8f8a232eec01\") " Oct 14 07:45:03 crc kubenswrapper[4870]: I1014 07:45:03.633709 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/610412a5-52e1-4aa6-9d19-8f8a232eec01-config-volume\") pod \"610412a5-52e1-4aa6-9d19-8f8a232eec01\" (UID: \"610412a5-52e1-4aa6-9d19-8f8a232eec01\") " Oct 14 07:45:03 crc kubenswrapper[4870]: I1014 07:45:03.634730 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/610412a5-52e1-4aa6-9d19-8f8a232eec01-config-volume" (OuterVolumeSpecName: "config-volume") pod "610412a5-52e1-4aa6-9d19-8f8a232eec01" (UID: "610412a5-52e1-4aa6-9d19-8f8a232eec01"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 07:45:03 crc kubenswrapper[4870]: I1014 07:45:03.635398 4870 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/610412a5-52e1-4aa6-9d19-8f8a232eec01-config-volume\") on node \"crc\" DevicePath \"\"" Oct 14 07:45:03 crc kubenswrapper[4870]: I1014 07:45:03.641975 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/610412a5-52e1-4aa6-9d19-8f8a232eec01-kube-api-access-mxspq" (OuterVolumeSpecName: "kube-api-access-mxspq") pod "610412a5-52e1-4aa6-9d19-8f8a232eec01" (UID: "610412a5-52e1-4aa6-9d19-8f8a232eec01"). InnerVolumeSpecName "kube-api-access-mxspq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:45:03 crc kubenswrapper[4870]: I1014 07:45:03.642838 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/610412a5-52e1-4aa6-9d19-8f8a232eec01-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "610412a5-52e1-4aa6-9d19-8f8a232eec01" (UID: "610412a5-52e1-4aa6-9d19-8f8a232eec01"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 07:45:03 crc kubenswrapper[4870]: I1014 07:45:03.736940 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mxspq\" (UniqueName: \"kubernetes.io/projected/610412a5-52e1-4aa6-9d19-8f8a232eec01-kube-api-access-mxspq\") on node \"crc\" DevicePath \"\"" Oct 14 07:45:03 crc kubenswrapper[4870]: I1014 07:45:03.736991 4870 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/610412a5-52e1-4aa6-9d19-8f8a232eec01-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 14 07:45:04 crc kubenswrapper[4870]: I1014 07:45:04.153228 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29340465-nfz9t" event={"ID":"610412a5-52e1-4aa6-9d19-8f8a232eec01","Type":"ContainerDied","Data":"fe484b85396e71337d718fd469d1716aa2d85c3ac310cc6e48b1c7e739f8e640"} Oct 14 07:45:04 crc kubenswrapper[4870]: I1014 07:45:04.153293 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29340465-nfz9t" Oct 14 07:45:04 crc kubenswrapper[4870]: I1014 07:45:04.153323 4870 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fe484b85396e71337d718fd469d1716aa2d85c3ac310cc6e48b1c7e739f8e640" Oct 14 07:45:04 crc kubenswrapper[4870]: I1014 07:45:04.645995 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29340420-t6cgj"] Oct 14 07:45:04 crc kubenswrapper[4870]: I1014 07:45:04.657997 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29340420-t6cgj"] Oct 14 07:45:05 crc kubenswrapper[4870]: I1014 07:45:05.051267 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="737f304d-12ac-4d4b-97fb-e99506dd08a8" path="/var/lib/kubelet/pods/737f304d-12ac-4d4b-97fb-e99506dd08a8/volumes" Oct 14 07:45:12 crc kubenswrapper[4870]: I1014 07:45:12.034850 4870 scope.go:117] "RemoveContainer" containerID="0c23f76241656fa35961005192bf783fc346c05081f076194d712c820a96b05d" Oct 14 07:45:12 crc kubenswrapper[4870]: E1014 07:45:12.036167 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 07:45:12 crc kubenswrapper[4870]: I1014 07:45:12.354301 4870 scope.go:117] "RemoveContainer" containerID="54dde5dec3d53d9948e6ae0d3f700606e677c020c9d7e4bd7179a25688038920" Oct 14 07:45:27 crc kubenswrapper[4870]: I1014 07:45:27.034300 4870 scope.go:117] "RemoveContainer" containerID="0c23f76241656fa35961005192bf783fc346c05081f076194d712c820a96b05d" Oct 14 07:45:27 crc kubenswrapper[4870]: E1014 07:45:27.035864 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 07:45:32 crc kubenswrapper[4870]: I1014 07:45:32.822636 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-mpxng"] Oct 14 07:45:32 crc kubenswrapper[4870]: E1014 07:45:32.823774 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="610412a5-52e1-4aa6-9d19-8f8a232eec01" containerName="collect-profiles" Oct 14 07:45:32 crc kubenswrapper[4870]: I1014 07:45:32.823796 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="610412a5-52e1-4aa6-9d19-8f8a232eec01" containerName="collect-profiles" Oct 14 07:45:32 crc kubenswrapper[4870]: I1014 07:45:32.824043 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="610412a5-52e1-4aa6-9d19-8f8a232eec01" containerName="collect-profiles" Oct 14 07:45:32 crc kubenswrapper[4870]: I1014 07:45:32.825837 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mpxng" Oct 14 07:45:32 crc kubenswrapper[4870]: I1014 07:45:32.844821 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-mpxng"] Oct 14 07:45:32 crc kubenswrapper[4870]: I1014 07:45:32.928199 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e7619066-6d81-4d1d-aa48-b6f77bc25e11-catalog-content\") pod \"community-operators-mpxng\" (UID: \"e7619066-6d81-4d1d-aa48-b6f77bc25e11\") " pod="openshift-marketplace/community-operators-mpxng" Oct 14 07:45:32 crc kubenswrapper[4870]: I1014 07:45:32.928269 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e7619066-6d81-4d1d-aa48-b6f77bc25e11-utilities\") pod \"community-operators-mpxng\" (UID: \"e7619066-6d81-4d1d-aa48-b6f77bc25e11\") " pod="openshift-marketplace/community-operators-mpxng" Oct 14 07:45:32 crc kubenswrapper[4870]: I1014 07:45:32.928963 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-db279\" (UniqueName: \"kubernetes.io/projected/e7619066-6d81-4d1d-aa48-b6f77bc25e11-kube-api-access-db279\") pod \"community-operators-mpxng\" (UID: \"e7619066-6d81-4d1d-aa48-b6f77bc25e11\") " pod="openshift-marketplace/community-operators-mpxng" Oct 14 07:45:33 crc kubenswrapper[4870]: I1014 07:45:33.030799 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-db279\" (UniqueName: \"kubernetes.io/projected/e7619066-6d81-4d1d-aa48-b6f77bc25e11-kube-api-access-db279\") pod \"community-operators-mpxng\" (UID: \"e7619066-6d81-4d1d-aa48-b6f77bc25e11\") " pod="openshift-marketplace/community-operators-mpxng" Oct 14 07:45:33 crc kubenswrapper[4870]: I1014 07:45:33.030884 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e7619066-6d81-4d1d-aa48-b6f77bc25e11-catalog-content\") pod \"community-operators-mpxng\" (UID: \"e7619066-6d81-4d1d-aa48-b6f77bc25e11\") " pod="openshift-marketplace/community-operators-mpxng" Oct 14 07:45:33 crc kubenswrapper[4870]: I1014 07:45:33.030949 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e7619066-6d81-4d1d-aa48-b6f77bc25e11-utilities\") pod \"community-operators-mpxng\" (UID: \"e7619066-6d81-4d1d-aa48-b6f77bc25e11\") " pod="openshift-marketplace/community-operators-mpxng" Oct 14 07:45:33 crc kubenswrapper[4870]: I1014 07:45:33.032178 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e7619066-6d81-4d1d-aa48-b6f77bc25e11-catalog-content\") pod \"community-operators-mpxng\" (UID: \"e7619066-6d81-4d1d-aa48-b6f77bc25e11\") " pod="openshift-marketplace/community-operators-mpxng" Oct 14 07:45:33 crc kubenswrapper[4870]: I1014 07:45:33.032990 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e7619066-6d81-4d1d-aa48-b6f77bc25e11-utilities\") pod \"community-operators-mpxng\" (UID: \"e7619066-6d81-4d1d-aa48-b6f77bc25e11\") " pod="openshift-marketplace/community-operators-mpxng" Oct 14 07:45:33 crc kubenswrapper[4870]: I1014 07:45:33.060369 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-db279\" (UniqueName: \"kubernetes.io/projected/e7619066-6d81-4d1d-aa48-b6f77bc25e11-kube-api-access-db279\") pod \"community-operators-mpxng\" (UID: \"e7619066-6d81-4d1d-aa48-b6f77bc25e11\") " pod="openshift-marketplace/community-operators-mpxng" Oct 14 07:45:33 crc kubenswrapper[4870]: I1014 07:45:33.156356 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mpxng" Oct 14 07:45:33 crc kubenswrapper[4870]: I1014 07:45:33.700140 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-mpxng"] Oct 14 07:45:33 crc kubenswrapper[4870]: W1014 07:45:33.710265 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode7619066_6d81_4d1d_aa48_b6f77bc25e11.slice/crio-dc9f499d3830061c2af11739df825839a15d395926aa87a95b44a5a304071423 WatchSource:0}: Error finding container dc9f499d3830061c2af11739df825839a15d395926aa87a95b44a5a304071423: Status 404 returned error can't find the container with id dc9f499d3830061c2af11739df825839a15d395926aa87a95b44a5a304071423 Oct 14 07:45:34 crc kubenswrapper[4870]: I1014 07:45:34.446674 4870 generic.go:334] "Generic (PLEG): container finished" podID="e7619066-6d81-4d1d-aa48-b6f77bc25e11" containerID="8c3e16637c4cc2db7020a00d886a072773475e59c1b840e6dceb38e1de444882" exitCode=0 Oct 14 07:45:34 crc kubenswrapper[4870]: I1014 07:45:34.446752 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mpxng" event={"ID":"e7619066-6d81-4d1d-aa48-b6f77bc25e11","Type":"ContainerDied","Data":"8c3e16637c4cc2db7020a00d886a072773475e59c1b840e6dceb38e1de444882"} Oct 14 07:45:34 crc kubenswrapper[4870]: I1014 07:45:34.446838 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mpxng" event={"ID":"e7619066-6d81-4d1d-aa48-b6f77bc25e11","Type":"ContainerStarted","Data":"dc9f499d3830061c2af11739df825839a15d395926aa87a95b44a5a304071423"} Oct 14 07:45:34 crc kubenswrapper[4870]: I1014 07:45:34.449724 4870 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 14 07:45:35 crc kubenswrapper[4870]: I1014 07:45:35.459649 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mpxng" event={"ID":"e7619066-6d81-4d1d-aa48-b6f77bc25e11","Type":"ContainerStarted","Data":"56b34e64d7405d765a8665c5ff6a8c2a3c5e05cd481b2d6a0979621edf6db12d"} Oct 14 07:45:36 crc kubenswrapper[4870]: I1014 07:45:36.472606 4870 generic.go:334] "Generic (PLEG): container finished" podID="e7619066-6d81-4d1d-aa48-b6f77bc25e11" containerID="56b34e64d7405d765a8665c5ff6a8c2a3c5e05cd481b2d6a0979621edf6db12d" exitCode=0 Oct 14 07:45:36 crc kubenswrapper[4870]: I1014 07:45:36.472711 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mpxng" event={"ID":"e7619066-6d81-4d1d-aa48-b6f77bc25e11","Type":"ContainerDied","Data":"56b34e64d7405d765a8665c5ff6a8c2a3c5e05cd481b2d6a0979621edf6db12d"} Oct 14 07:45:37 crc kubenswrapper[4870]: I1014 07:45:37.486296 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mpxng" event={"ID":"e7619066-6d81-4d1d-aa48-b6f77bc25e11","Type":"ContainerStarted","Data":"7c32126ba910fe9a042cefff1eb6cedfee2e498837206d255f4fe3f929968972"} Oct 14 07:45:37 crc kubenswrapper[4870]: I1014 07:45:37.524832 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-mpxng" podStartSLOduration=2.885797813 podStartE2EDuration="5.524804853s" podCreationTimestamp="2025-10-14 07:45:32 +0000 UTC" firstStartedPulling="2025-10-14 07:45:34.449175033 +0000 UTC m=+2670.146535444" lastFinishedPulling="2025-10-14 07:45:37.088182113 +0000 UTC m=+2672.785542484" observedRunningTime="2025-10-14 07:45:37.519380489 +0000 UTC m=+2673.216740900" watchObservedRunningTime="2025-10-14 07:45:37.524804853 +0000 UTC m=+2673.222165264" Oct 14 07:45:42 crc kubenswrapper[4870]: I1014 07:45:42.034469 4870 scope.go:117] "RemoveContainer" containerID="0c23f76241656fa35961005192bf783fc346c05081f076194d712c820a96b05d" Oct 14 07:45:42 crc kubenswrapper[4870]: E1014 07:45:42.035168 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 07:45:43 crc kubenswrapper[4870]: I1014 07:45:43.156960 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-mpxng" Oct 14 07:45:43 crc kubenswrapper[4870]: I1014 07:45:43.157021 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-mpxng" Oct 14 07:45:43 crc kubenswrapper[4870]: I1014 07:45:43.237731 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-mpxng" Oct 14 07:45:43 crc kubenswrapper[4870]: I1014 07:45:43.616509 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-mpxng" Oct 14 07:45:43 crc kubenswrapper[4870]: I1014 07:45:43.685418 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-mpxng"] Oct 14 07:45:45 crc kubenswrapper[4870]: I1014 07:45:45.561272 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-mpxng" podUID="e7619066-6d81-4d1d-aa48-b6f77bc25e11" containerName="registry-server" containerID="cri-o://7c32126ba910fe9a042cefff1eb6cedfee2e498837206d255f4fe3f929968972" gracePeriod=2 Oct 14 07:45:46 crc kubenswrapper[4870]: I1014 07:45:46.110877 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mpxng" Oct 14 07:45:46 crc kubenswrapper[4870]: I1014 07:45:46.245204 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e7619066-6d81-4d1d-aa48-b6f77bc25e11-utilities\") pod \"e7619066-6d81-4d1d-aa48-b6f77bc25e11\" (UID: \"e7619066-6d81-4d1d-aa48-b6f77bc25e11\") " Oct 14 07:45:46 crc kubenswrapper[4870]: I1014 07:45:46.245283 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-db279\" (UniqueName: \"kubernetes.io/projected/e7619066-6d81-4d1d-aa48-b6f77bc25e11-kube-api-access-db279\") pod \"e7619066-6d81-4d1d-aa48-b6f77bc25e11\" (UID: \"e7619066-6d81-4d1d-aa48-b6f77bc25e11\") " Oct 14 07:45:46 crc kubenswrapper[4870]: I1014 07:45:46.245402 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e7619066-6d81-4d1d-aa48-b6f77bc25e11-catalog-content\") pod \"e7619066-6d81-4d1d-aa48-b6f77bc25e11\" (UID: \"e7619066-6d81-4d1d-aa48-b6f77bc25e11\") " Oct 14 07:45:46 crc kubenswrapper[4870]: I1014 07:45:46.248154 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e7619066-6d81-4d1d-aa48-b6f77bc25e11-utilities" (OuterVolumeSpecName: "utilities") pod "e7619066-6d81-4d1d-aa48-b6f77bc25e11" (UID: "e7619066-6d81-4d1d-aa48-b6f77bc25e11"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:45:46 crc kubenswrapper[4870]: I1014 07:45:46.256130 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7619066-6d81-4d1d-aa48-b6f77bc25e11-kube-api-access-db279" (OuterVolumeSpecName: "kube-api-access-db279") pod "e7619066-6d81-4d1d-aa48-b6f77bc25e11" (UID: "e7619066-6d81-4d1d-aa48-b6f77bc25e11"). InnerVolumeSpecName "kube-api-access-db279". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:45:46 crc kubenswrapper[4870]: I1014 07:45:46.314700 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e7619066-6d81-4d1d-aa48-b6f77bc25e11-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e7619066-6d81-4d1d-aa48-b6f77bc25e11" (UID: "e7619066-6d81-4d1d-aa48-b6f77bc25e11"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:45:46 crc kubenswrapper[4870]: I1014 07:45:46.347509 4870 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e7619066-6d81-4d1d-aa48-b6f77bc25e11-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 07:45:46 crc kubenswrapper[4870]: I1014 07:45:46.347561 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-db279\" (UniqueName: \"kubernetes.io/projected/e7619066-6d81-4d1d-aa48-b6f77bc25e11-kube-api-access-db279\") on node \"crc\" DevicePath \"\"" Oct 14 07:45:46 crc kubenswrapper[4870]: I1014 07:45:46.347581 4870 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e7619066-6d81-4d1d-aa48-b6f77bc25e11-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 07:45:46 crc kubenswrapper[4870]: I1014 07:45:46.575256 4870 generic.go:334] "Generic (PLEG): container finished" podID="e7619066-6d81-4d1d-aa48-b6f77bc25e11" containerID="7c32126ba910fe9a042cefff1eb6cedfee2e498837206d255f4fe3f929968972" exitCode=0 Oct 14 07:45:46 crc kubenswrapper[4870]: I1014 07:45:46.575321 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mpxng" event={"ID":"e7619066-6d81-4d1d-aa48-b6f77bc25e11","Type":"ContainerDied","Data":"7c32126ba910fe9a042cefff1eb6cedfee2e498837206d255f4fe3f929968972"} Oct 14 07:45:46 crc kubenswrapper[4870]: I1014 07:45:46.575377 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mpxng" event={"ID":"e7619066-6d81-4d1d-aa48-b6f77bc25e11","Type":"ContainerDied","Data":"dc9f499d3830061c2af11739df825839a15d395926aa87a95b44a5a304071423"} Oct 14 07:45:46 crc kubenswrapper[4870]: I1014 07:45:46.575386 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mpxng" Oct 14 07:45:46 crc kubenswrapper[4870]: I1014 07:45:46.575422 4870 scope.go:117] "RemoveContainer" containerID="7c32126ba910fe9a042cefff1eb6cedfee2e498837206d255f4fe3f929968972" Oct 14 07:45:46 crc kubenswrapper[4870]: I1014 07:45:46.612158 4870 scope.go:117] "RemoveContainer" containerID="56b34e64d7405d765a8665c5ff6a8c2a3c5e05cd481b2d6a0979621edf6db12d" Oct 14 07:45:46 crc kubenswrapper[4870]: I1014 07:45:46.629251 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-mpxng"] Oct 14 07:45:46 crc kubenswrapper[4870]: I1014 07:45:46.638698 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-mpxng"] Oct 14 07:45:46 crc kubenswrapper[4870]: I1014 07:45:46.652970 4870 scope.go:117] "RemoveContainer" containerID="8c3e16637c4cc2db7020a00d886a072773475e59c1b840e6dceb38e1de444882" Oct 14 07:45:46 crc kubenswrapper[4870]: I1014 07:45:46.680988 4870 scope.go:117] "RemoveContainer" containerID="7c32126ba910fe9a042cefff1eb6cedfee2e498837206d255f4fe3f929968972" Oct 14 07:45:46 crc kubenswrapper[4870]: E1014 07:45:46.681557 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7c32126ba910fe9a042cefff1eb6cedfee2e498837206d255f4fe3f929968972\": container with ID starting with 7c32126ba910fe9a042cefff1eb6cedfee2e498837206d255f4fe3f929968972 not found: ID does not exist" containerID="7c32126ba910fe9a042cefff1eb6cedfee2e498837206d255f4fe3f929968972" Oct 14 07:45:46 crc kubenswrapper[4870]: I1014 07:45:46.681625 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7c32126ba910fe9a042cefff1eb6cedfee2e498837206d255f4fe3f929968972"} err="failed to get container status \"7c32126ba910fe9a042cefff1eb6cedfee2e498837206d255f4fe3f929968972\": rpc error: code = NotFound desc = could not find container \"7c32126ba910fe9a042cefff1eb6cedfee2e498837206d255f4fe3f929968972\": container with ID starting with 7c32126ba910fe9a042cefff1eb6cedfee2e498837206d255f4fe3f929968972 not found: ID does not exist" Oct 14 07:45:46 crc kubenswrapper[4870]: I1014 07:45:46.681666 4870 scope.go:117] "RemoveContainer" containerID="56b34e64d7405d765a8665c5ff6a8c2a3c5e05cd481b2d6a0979621edf6db12d" Oct 14 07:45:46 crc kubenswrapper[4870]: E1014 07:45:46.682156 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"56b34e64d7405d765a8665c5ff6a8c2a3c5e05cd481b2d6a0979621edf6db12d\": container with ID starting with 56b34e64d7405d765a8665c5ff6a8c2a3c5e05cd481b2d6a0979621edf6db12d not found: ID does not exist" containerID="56b34e64d7405d765a8665c5ff6a8c2a3c5e05cd481b2d6a0979621edf6db12d" Oct 14 07:45:46 crc kubenswrapper[4870]: I1014 07:45:46.682209 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"56b34e64d7405d765a8665c5ff6a8c2a3c5e05cd481b2d6a0979621edf6db12d"} err="failed to get container status \"56b34e64d7405d765a8665c5ff6a8c2a3c5e05cd481b2d6a0979621edf6db12d\": rpc error: code = NotFound desc = could not find container \"56b34e64d7405d765a8665c5ff6a8c2a3c5e05cd481b2d6a0979621edf6db12d\": container with ID starting with 56b34e64d7405d765a8665c5ff6a8c2a3c5e05cd481b2d6a0979621edf6db12d not found: ID does not exist" Oct 14 07:45:46 crc kubenswrapper[4870]: I1014 07:45:46.682253 4870 scope.go:117] "RemoveContainer" containerID="8c3e16637c4cc2db7020a00d886a072773475e59c1b840e6dceb38e1de444882" Oct 14 07:45:46 crc kubenswrapper[4870]: E1014 07:45:46.682658 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8c3e16637c4cc2db7020a00d886a072773475e59c1b840e6dceb38e1de444882\": container with ID starting with 8c3e16637c4cc2db7020a00d886a072773475e59c1b840e6dceb38e1de444882 not found: ID does not exist" containerID="8c3e16637c4cc2db7020a00d886a072773475e59c1b840e6dceb38e1de444882" Oct 14 07:45:46 crc kubenswrapper[4870]: I1014 07:45:46.682690 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8c3e16637c4cc2db7020a00d886a072773475e59c1b840e6dceb38e1de444882"} err="failed to get container status \"8c3e16637c4cc2db7020a00d886a072773475e59c1b840e6dceb38e1de444882\": rpc error: code = NotFound desc = could not find container \"8c3e16637c4cc2db7020a00d886a072773475e59c1b840e6dceb38e1de444882\": container with ID starting with 8c3e16637c4cc2db7020a00d886a072773475e59c1b840e6dceb38e1de444882 not found: ID does not exist" Oct 14 07:45:47 crc kubenswrapper[4870]: I1014 07:45:47.051820 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7619066-6d81-4d1d-aa48-b6f77bc25e11" path="/var/lib/kubelet/pods/e7619066-6d81-4d1d-aa48-b6f77bc25e11/volumes" Oct 14 07:45:52 crc kubenswrapper[4870]: I1014 07:45:52.622345 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-dnxcs"] Oct 14 07:45:52 crc kubenswrapper[4870]: E1014 07:45:52.624377 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7619066-6d81-4d1d-aa48-b6f77bc25e11" containerName="extract-content" Oct 14 07:45:52 crc kubenswrapper[4870]: I1014 07:45:52.624480 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7619066-6d81-4d1d-aa48-b6f77bc25e11" containerName="extract-content" Oct 14 07:45:52 crc kubenswrapper[4870]: E1014 07:45:52.624564 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7619066-6d81-4d1d-aa48-b6f77bc25e11" containerName="extract-utilities" Oct 14 07:45:52 crc kubenswrapper[4870]: I1014 07:45:52.624627 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7619066-6d81-4d1d-aa48-b6f77bc25e11" containerName="extract-utilities" Oct 14 07:45:52 crc kubenswrapper[4870]: E1014 07:45:52.624715 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7619066-6d81-4d1d-aa48-b6f77bc25e11" containerName="registry-server" Oct 14 07:45:52 crc kubenswrapper[4870]: I1014 07:45:52.624790 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7619066-6d81-4d1d-aa48-b6f77bc25e11" containerName="registry-server" Oct 14 07:45:52 crc kubenswrapper[4870]: I1014 07:45:52.625020 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="e7619066-6d81-4d1d-aa48-b6f77bc25e11" containerName="registry-server" Oct 14 07:45:52 crc kubenswrapper[4870]: I1014 07:45:52.626399 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dnxcs" Oct 14 07:45:52 crc kubenswrapper[4870]: I1014 07:45:52.634708 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-dnxcs"] Oct 14 07:45:52 crc kubenswrapper[4870]: I1014 07:45:52.662955 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rsnfc\" (UniqueName: \"kubernetes.io/projected/c25b883e-0b06-4477-abf7-5f404393ce2e-kube-api-access-rsnfc\") pod \"redhat-operators-dnxcs\" (UID: \"c25b883e-0b06-4477-abf7-5f404393ce2e\") " pod="openshift-marketplace/redhat-operators-dnxcs" Oct 14 07:45:52 crc kubenswrapper[4870]: I1014 07:45:52.663044 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c25b883e-0b06-4477-abf7-5f404393ce2e-utilities\") pod \"redhat-operators-dnxcs\" (UID: \"c25b883e-0b06-4477-abf7-5f404393ce2e\") " pod="openshift-marketplace/redhat-operators-dnxcs" Oct 14 07:45:52 crc kubenswrapper[4870]: I1014 07:45:52.663091 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c25b883e-0b06-4477-abf7-5f404393ce2e-catalog-content\") pod \"redhat-operators-dnxcs\" (UID: \"c25b883e-0b06-4477-abf7-5f404393ce2e\") " pod="openshift-marketplace/redhat-operators-dnxcs" Oct 14 07:45:52 crc kubenswrapper[4870]: I1014 07:45:52.764590 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rsnfc\" (UniqueName: \"kubernetes.io/projected/c25b883e-0b06-4477-abf7-5f404393ce2e-kube-api-access-rsnfc\") pod \"redhat-operators-dnxcs\" (UID: \"c25b883e-0b06-4477-abf7-5f404393ce2e\") " pod="openshift-marketplace/redhat-operators-dnxcs" Oct 14 07:45:52 crc kubenswrapper[4870]: I1014 07:45:52.765170 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c25b883e-0b06-4477-abf7-5f404393ce2e-utilities\") pod \"redhat-operators-dnxcs\" (UID: \"c25b883e-0b06-4477-abf7-5f404393ce2e\") " pod="openshift-marketplace/redhat-operators-dnxcs" Oct 14 07:45:52 crc kubenswrapper[4870]: I1014 07:45:52.765874 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c25b883e-0b06-4477-abf7-5f404393ce2e-utilities\") pod \"redhat-operators-dnxcs\" (UID: \"c25b883e-0b06-4477-abf7-5f404393ce2e\") " pod="openshift-marketplace/redhat-operators-dnxcs" Oct 14 07:45:52 crc kubenswrapper[4870]: I1014 07:45:52.766054 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c25b883e-0b06-4477-abf7-5f404393ce2e-catalog-content\") pod \"redhat-operators-dnxcs\" (UID: \"c25b883e-0b06-4477-abf7-5f404393ce2e\") " pod="openshift-marketplace/redhat-operators-dnxcs" Oct 14 07:45:52 crc kubenswrapper[4870]: I1014 07:45:52.766494 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c25b883e-0b06-4477-abf7-5f404393ce2e-catalog-content\") pod \"redhat-operators-dnxcs\" (UID: \"c25b883e-0b06-4477-abf7-5f404393ce2e\") " pod="openshift-marketplace/redhat-operators-dnxcs" Oct 14 07:45:52 crc kubenswrapper[4870]: I1014 07:45:52.785270 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rsnfc\" (UniqueName: \"kubernetes.io/projected/c25b883e-0b06-4477-abf7-5f404393ce2e-kube-api-access-rsnfc\") pod \"redhat-operators-dnxcs\" (UID: \"c25b883e-0b06-4477-abf7-5f404393ce2e\") " pod="openshift-marketplace/redhat-operators-dnxcs" Oct 14 07:45:52 crc kubenswrapper[4870]: I1014 07:45:52.955361 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dnxcs" Oct 14 07:45:53 crc kubenswrapper[4870]: I1014 07:45:53.034266 4870 scope.go:117] "RemoveContainer" containerID="0c23f76241656fa35961005192bf783fc346c05081f076194d712c820a96b05d" Oct 14 07:45:53 crc kubenswrapper[4870]: E1014 07:45:53.034778 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 07:45:53 crc kubenswrapper[4870]: I1014 07:45:53.409909 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-dnxcs"] Oct 14 07:45:53 crc kubenswrapper[4870]: I1014 07:45:53.677884 4870 generic.go:334] "Generic (PLEG): container finished" podID="c25b883e-0b06-4477-abf7-5f404393ce2e" containerID="cacdd034e541074d7a171a0cf4015790649ff7ea6aa5171022c37201daa54e9f" exitCode=0 Oct 14 07:45:53 crc kubenswrapper[4870]: I1014 07:45:53.677996 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dnxcs" event={"ID":"c25b883e-0b06-4477-abf7-5f404393ce2e","Type":"ContainerDied","Data":"cacdd034e541074d7a171a0cf4015790649ff7ea6aa5171022c37201daa54e9f"} Oct 14 07:45:53 crc kubenswrapper[4870]: I1014 07:45:53.678169 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dnxcs" event={"ID":"c25b883e-0b06-4477-abf7-5f404393ce2e","Type":"ContainerStarted","Data":"a8ca3f83976b040fedff5ca737809aa049458a83ca3e04efdd5287ebe4d95da1"} Oct 14 07:45:55 crc kubenswrapper[4870]: I1014 07:45:55.719024 4870 generic.go:334] "Generic (PLEG): container finished" podID="c25b883e-0b06-4477-abf7-5f404393ce2e" containerID="79b2866f1d829b5ca2e61c60f2dfba9ceeaae79a530fe400145b2c56a768be9e" exitCode=0 Oct 14 07:45:55 crc kubenswrapper[4870]: I1014 07:45:55.719109 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dnxcs" event={"ID":"c25b883e-0b06-4477-abf7-5f404393ce2e","Type":"ContainerDied","Data":"79b2866f1d829b5ca2e61c60f2dfba9ceeaae79a530fe400145b2c56a768be9e"} Oct 14 07:45:56 crc kubenswrapper[4870]: I1014 07:45:56.747424 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dnxcs" event={"ID":"c25b883e-0b06-4477-abf7-5f404393ce2e","Type":"ContainerStarted","Data":"00c3fd16946e0933cd0d9db945c6b57f1f2eb2b2fdbab5368e82ec0c761f9793"} Oct 14 07:45:56 crc kubenswrapper[4870]: I1014 07:45:56.777359 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-dnxcs" podStartSLOduration=2.230760545 podStartE2EDuration="4.777336048s" podCreationTimestamp="2025-10-14 07:45:52 +0000 UTC" firstStartedPulling="2025-10-14 07:45:53.679506819 +0000 UTC m=+2689.376867230" lastFinishedPulling="2025-10-14 07:45:56.226082322 +0000 UTC m=+2691.923442733" observedRunningTime="2025-10-14 07:45:56.774108798 +0000 UTC m=+2692.471469249" watchObservedRunningTime="2025-10-14 07:45:56.777336048 +0000 UTC m=+2692.474696429" Oct 14 07:46:02 crc kubenswrapper[4870]: I1014 07:46:02.956682 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-dnxcs" Oct 14 07:46:02 crc kubenswrapper[4870]: I1014 07:46:02.958637 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-dnxcs" Oct 14 07:46:03 crc kubenswrapper[4870]: I1014 07:46:03.050751 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-dnxcs" Oct 14 07:46:03 crc kubenswrapper[4870]: I1014 07:46:03.882813 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-dnxcs" Oct 14 07:46:04 crc kubenswrapper[4870]: I1014 07:46:04.020919 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-dnxcs"] Oct 14 07:46:05 crc kubenswrapper[4870]: I1014 07:46:05.828916 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-dnxcs" podUID="c25b883e-0b06-4477-abf7-5f404393ce2e" containerName="registry-server" containerID="cri-o://00c3fd16946e0933cd0d9db945c6b57f1f2eb2b2fdbab5368e82ec0c761f9793" gracePeriod=2 Oct 14 07:46:06 crc kubenswrapper[4870]: I1014 07:46:06.833006 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dnxcs" Oct 14 07:46:06 crc kubenswrapper[4870]: I1014 07:46:06.838428 4870 generic.go:334] "Generic (PLEG): container finished" podID="c25b883e-0b06-4477-abf7-5f404393ce2e" containerID="00c3fd16946e0933cd0d9db945c6b57f1f2eb2b2fdbab5368e82ec0c761f9793" exitCode=0 Oct 14 07:46:06 crc kubenswrapper[4870]: I1014 07:46:06.838480 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dnxcs" event={"ID":"c25b883e-0b06-4477-abf7-5f404393ce2e","Type":"ContainerDied","Data":"00c3fd16946e0933cd0d9db945c6b57f1f2eb2b2fdbab5368e82ec0c761f9793"} Oct 14 07:46:06 crc kubenswrapper[4870]: I1014 07:46:06.838502 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dnxcs" event={"ID":"c25b883e-0b06-4477-abf7-5f404393ce2e","Type":"ContainerDied","Data":"a8ca3f83976b040fedff5ca737809aa049458a83ca3e04efdd5287ebe4d95da1"} Oct 14 07:46:06 crc kubenswrapper[4870]: I1014 07:46:06.838520 4870 scope.go:117] "RemoveContainer" containerID="00c3fd16946e0933cd0d9db945c6b57f1f2eb2b2fdbab5368e82ec0c761f9793" Oct 14 07:46:06 crc kubenswrapper[4870]: I1014 07:46:06.838566 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dnxcs" Oct 14 07:46:06 crc kubenswrapper[4870]: I1014 07:46:06.882387 4870 scope.go:117] "RemoveContainer" containerID="79b2866f1d829b5ca2e61c60f2dfba9ceeaae79a530fe400145b2c56a768be9e" Oct 14 07:46:06 crc kubenswrapper[4870]: I1014 07:46:06.918026 4870 scope.go:117] "RemoveContainer" containerID="cacdd034e541074d7a171a0cf4015790649ff7ea6aa5171022c37201daa54e9f" Oct 14 07:46:06 crc kubenswrapper[4870]: I1014 07:46:06.945039 4870 scope.go:117] "RemoveContainer" containerID="00c3fd16946e0933cd0d9db945c6b57f1f2eb2b2fdbab5368e82ec0c761f9793" Oct 14 07:46:06 crc kubenswrapper[4870]: E1014 07:46:06.945727 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"00c3fd16946e0933cd0d9db945c6b57f1f2eb2b2fdbab5368e82ec0c761f9793\": container with ID starting with 00c3fd16946e0933cd0d9db945c6b57f1f2eb2b2fdbab5368e82ec0c761f9793 not found: ID does not exist" containerID="00c3fd16946e0933cd0d9db945c6b57f1f2eb2b2fdbab5368e82ec0c761f9793" Oct 14 07:46:06 crc kubenswrapper[4870]: I1014 07:46:06.945789 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"00c3fd16946e0933cd0d9db945c6b57f1f2eb2b2fdbab5368e82ec0c761f9793"} err="failed to get container status \"00c3fd16946e0933cd0d9db945c6b57f1f2eb2b2fdbab5368e82ec0c761f9793\": rpc error: code = NotFound desc = could not find container \"00c3fd16946e0933cd0d9db945c6b57f1f2eb2b2fdbab5368e82ec0c761f9793\": container with ID starting with 00c3fd16946e0933cd0d9db945c6b57f1f2eb2b2fdbab5368e82ec0c761f9793 not found: ID does not exist" Oct 14 07:46:06 crc kubenswrapper[4870]: I1014 07:46:06.945827 4870 scope.go:117] "RemoveContainer" containerID="79b2866f1d829b5ca2e61c60f2dfba9ceeaae79a530fe400145b2c56a768be9e" Oct 14 07:46:06 crc kubenswrapper[4870]: E1014 07:46:06.946158 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"79b2866f1d829b5ca2e61c60f2dfba9ceeaae79a530fe400145b2c56a768be9e\": container with ID starting with 79b2866f1d829b5ca2e61c60f2dfba9ceeaae79a530fe400145b2c56a768be9e not found: ID does not exist" containerID="79b2866f1d829b5ca2e61c60f2dfba9ceeaae79a530fe400145b2c56a768be9e" Oct 14 07:46:06 crc kubenswrapper[4870]: I1014 07:46:06.946205 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"79b2866f1d829b5ca2e61c60f2dfba9ceeaae79a530fe400145b2c56a768be9e"} err="failed to get container status \"79b2866f1d829b5ca2e61c60f2dfba9ceeaae79a530fe400145b2c56a768be9e\": rpc error: code = NotFound desc = could not find container \"79b2866f1d829b5ca2e61c60f2dfba9ceeaae79a530fe400145b2c56a768be9e\": container with ID starting with 79b2866f1d829b5ca2e61c60f2dfba9ceeaae79a530fe400145b2c56a768be9e not found: ID does not exist" Oct 14 07:46:06 crc kubenswrapper[4870]: I1014 07:46:06.946237 4870 scope.go:117] "RemoveContainer" containerID="cacdd034e541074d7a171a0cf4015790649ff7ea6aa5171022c37201daa54e9f" Oct 14 07:46:06 crc kubenswrapper[4870]: E1014 07:46:06.946495 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cacdd034e541074d7a171a0cf4015790649ff7ea6aa5171022c37201daa54e9f\": container with ID starting with cacdd034e541074d7a171a0cf4015790649ff7ea6aa5171022c37201daa54e9f not found: ID does not exist" containerID="cacdd034e541074d7a171a0cf4015790649ff7ea6aa5171022c37201daa54e9f" Oct 14 07:46:06 crc kubenswrapper[4870]: I1014 07:46:06.946525 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cacdd034e541074d7a171a0cf4015790649ff7ea6aa5171022c37201daa54e9f"} err="failed to get container status \"cacdd034e541074d7a171a0cf4015790649ff7ea6aa5171022c37201daa54e9f\": rpc error: code = NotFound desc = could not find container \"cacdd034e541074d7a171a0cf4015790649ff7ea6aa5171022c37201daa54e9f\": container with ID starting with cacdd034e541074d7a171a0cf4015790649ff7ea6aa5171022c37201daa54e9f not found: ID does not exist" Oct 14 07:46:07 crc kubenswrapper[4870]: I1014 07:46:07.017393 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c25b883e-0b06-4477-abf7-5f404393ce2e-utilities\") pod \"c25b883e-0b06-4477-abf7-5f404393ce2e\" (UID: \"c25b883e-0b06-4477-abf7-5f404393ce2e\") " Oct 14 07:46:07 crc kubenswrapper[4870]: I1014 07:46:07.017537 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rsnfc\" (UniqueName: \"kubernetes.io/projected/c25b883e-0b06-4477-abf7-5f404393ce2e-kube-api-access-rsnfc\") pod \"c25b883e-0b06-4477-abf7-5f404393ce2e\" (UID: \"c25b883e-0b06-4477-abf7-5f404393ce2e\") " Oct 14 07:46:07 crc kubenswrapper[4870]: I1014 07:46:07.017615 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c25b883e-0b06-4477-abf7-5f404393ce2e-catalog-content\") pod \"c25b883e-0b06-4477-abf7-5f404393ce2e\" (UID: \"c25b883e-0b06-4477-abf7-5f404393ce2e\") " Oct 14 07:46:07 crc kubenswrapper[4870]: I1014 07:46:07.019503 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c25b883e-0b06-4477-abf7-5f404393ce2e-utilities" (OuterVolumeSpecName: "utilities") pod "c25b883e-0b06-4477-abf7-5f404393ce2e" (UID: "c25b883e-0b06-4477-abf7-5f404393ce2e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:46:07 crc kubenswrapper[4870]: I1014 07:46:07.029321 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c25b883e-0b06-4477-abf7-5f404393ce2e-kube-api-access-rsnfc" (OuterVolumeSpecName: "kube-api-access-rsnfc") pod "c25b883e-0b06-4477-abf7-5f404393ce2e" (UID: "c25b883e-0b06-4477-abf7-5f404393ce2e"). InnerVolumeSpecName "kube-api-access-rsnfc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:46:07 crc kubenswrapper[4870]: I1014 07:46:07.034562 4870 scope.go:117] "RemoveContainer" containerID="0c23f76241656fa35961005192bf783fc346c05081f076194d712c820a96b05d" Oct 14 07:46:07 crc kubenswrapper[4870]: E1014 07:46:07.034825 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 07:46:07 crc kubenswrapper[4870]: I1014 07:46:07.126923 4870 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c25b883e-0b06-4477-abf7-5f404393ce2e-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 07:46:07 crc kubenswrapper[4870]: I1014 07:46:07.126987 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rsnfc\" (UniqueName: \"kubernetes.io/projected/c25b883e-0b06-4477-abf7-5f404393ce2e-kube-api-access-rsnfc\") on node \"crc\" DevicePath \"\"" Oct 14 07:46:07 crc kubenswrapper[4870]: I1014 07:46:07.194089 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c25b883e-0b06-4477-abf7-5f404393ce2e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c25b883e-0b06-4477-abf7-5f404393ce2e" (UID: "c25b883e-0b06-4477-abf7-5f404393ce2e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:46:07 crc kubenswrapper[4870]: I1014 07:46:07.228858 4870 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c25b883e-0b06-4477-abf7-5f404393ce2e-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 07:46:07 crc kubenswrapper[4870]: I1014 07:46:07.483290 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-dnxcs"] Oct 14 07:46:07 crc kubenswrapper[4870]: I1014 07:46:07.491101 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-dnxcs"] Oct 14 07:46:09 crc kubenswrapper[4870]: I1014 07:46:09.047348 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c25b883e-0b06-4477-abf7-5f404393ce2e" path="/var/lib/kubelet/pods/c25b883e-0b06-4477-abf7-5f404393ce2e/volumes" Oct 14 07:46:12 crc kubenswrapper[4870]: I1014 07:46:12.443638 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-9jz9m"] Oct 14 07:46:12 crc kubenswrapper[4870]: E1014 07:46:12.449659 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c25b883e-0b06-4477-abf7-5f404393ce2e" containerName="registry-server" Oct 14 07:46:12 crc kubenswrapper[4870]: I1014 07:46:12.449928 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="c25b883e-0b06-4477-abf7-5f404393ce2e" containerName="registry-server" Oct 14 07:46:12 crc kubenswrapper[4870]: E1014 07:46:12.450335 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c25b883e-0b06-4477-abf7-5f404393ce2e" containerName="extract-content" Oct 14 07:46:12 crc kubenswrapper[4870]: I1014 07:46:12.450544 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="c25b883e-0b06-4477-abf7-5f404393ce2e" containerName="extract-content" Oct 14 07:46:12 crc kubenswrapper[4870]: E1014 07:46:12.450761 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c25b883e-0b06-4477-abf7-5f404393ce2e" containerName="extract-utilities" Oct 14 07:46:12 crc kubenswrapper[4870]: I1014 07:46:12.450926 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="c25b883e-0b06-4477-abf7-5f404393ce2e" containerName="extract-utilities" Oct 14 07:46:12 crc kubenswrapper[4870]: I1014 07:46:12.452017 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="c25b883e-0b06-4477-abf7-5f404393ce2e" containerName="registry-server" Oct 14 07:46:12 crc kubenswrapper[4870]: I1014 07:46:12.458837 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9jz9m" Oct 14 07:46:12 crc kubenswrapper[4870]: I1014 07:46:12.471708 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-9jz9m"] Oct 14 07:46:12 crc kubenswrapper[4870]: I1014 07:46:12.625520 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/68c9a2ec-af40-419e-8c62-7e3b6a6e2a2f-catalog-content\") pod \"redhat-marketplace-9jz9m\" (UID: \"68c9a2ec-af40-419e-8c62-7e3b6a6e2a2f\") " pod="openshift-marketplace/redhat-marketplace-9jz9m" Oct 14 07:46:12 crc kubenswrapper[4870]: I1014 07:46:12.625720 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/68c9a2ec-af40-419e-8c62-7e3b6a6e2a2f-utilities\") pod \"redhat-marketplace-9jz9m\" (UID: \"68c9a2ec-af40-419e-8c62-7e3b6a6e2a2f\") " pod="openshift-marketplace/redhat-marketplace-9jz9m" Oct 14 07:46:12 crc kubenswrapper[4870]: I1014 07:46:12.625786 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6jg8p\" (UniqueName: \"kubernetes.io/projected/68c9a2ec-af40-419e-8c62-7e3b6a6e2a2f-kube-api-access-6jg8p\") pod \"redhat-marketplace-9jz9m\" (UID: \"68c9a2ec-af40-419e-8c62-7e3b6a6e2a2f\") " pod="openshift-marketplace/redhat-marketplace-9jz9m" Oct 14 07:46:12 crc kubenswrapper[4870]: I1014 07:46:12.727405 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/68c9a2ec-af40-419e-8c62-7e3b6a6e2a2f-catalog-content\") pod \"redhat-marketplace-9jz9m\" (UID: \"68c9a2ec-af40-419e-8c62-7e3b6a6e2a2f\") " pod="openshift-marketplace/redhat-marketplace-9jz9m" Oct 14 07:46:12 crc kubenswrapper[4870]: I1014 07:46:12.727549 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/68c9a2ec-af40-419e-8c62-7e3b6a6e2a2f-utilities\") pod \"redhat-marketplace-9jz9m\" (UID: \"68c9a2ec-af40-419e-8c62-7e3b6a6e2a2f\") " pod="openshift-marketplace/redhat-marketplace-9jz9m" Oct 14 07:46:12 crc kubenswrapper[4870]: I1014 07:46:12.727983 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/68c9a2ec-af40-419e-8c62-7e3b6a6e2a2f-catalog-content\") pod \"redhat-marketplace-9jz9m\" (UID: \"68c9a2ec-af40-419e-8c62-7e3b6a6e2a2f\") " pod="openshift-marketplace/redhat-marketplace-9jz9m" Oct 14 07:46:12 crc kubenswrapper[4870]: I1014 07:46:12.728072 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/68c9a2ec-af40-419e-8c62-7e3b6a6e2a2f-utilities\") pod \"redhat-marketplace-9jz9m\" (UID: \"68c9a2ec-af40-419e-8c62-7e3b6a6e2a2f\") " pod="openshift-marketplace/redhat-marketplace-9jz9m" Oct 14 07:46:12 crc kubenswrapper[4870]: I1014 07:46:12.727581 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6jg8p\" (UniqueName: \"kubernetes.io/projected/68c9a2ec-af40-419e-8c62-7e3b6a6e2a2f-kube-api-access-6jg8p\") pod \"redhat-marketplace-9jz9m\" (UID: \"68c9a2ec-af40-419e-8c62-7e3b6a6e2a2f\") " pod="openshift-marketplace/redhat-marketplace-9jz9m" Oct 14 07:46:12 crc kubenswrapper[4870]: I1014 07:46:12.761816 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6jg8p\" (UniqueName: \"kubernetes.io/projected/68c9a2ec-af40-419e-8c62-7e3b6a6e2a2f-kube-api-access-6jg8p\") pod \"redhat-marketplace-9jz9m\" (UID: \"68c9a2ec-af40-419e-8c62-7e3b6a6e2a2f\") " pod="openshift-marketplace/redhat-marketplace-9jz9m" Oct 14 07:46:12 crc kubenswrapper[4870]: I1014 07:46:12.797107 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9jz9m" Oct 14 07:46:13 crc kubenswrapper[4870]: I1014 07:46:13.315426 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-9jz9m"] Oct 14 07:46:13 crc kubenswrapper[4870]: W1014 07:46:13.324606 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod68c9a2ec_af40_419e_8c62_7e3b6a6e2a2f.slice/crio-e3fb314e14eefd42729c7f316d80f3f8a573c2541cb5a2aa418f0ff656c944bd WatchSource:0}: Error finding container e3fb314e14eefd42729c7f316d80f3f8a573c2541cb5a2aa418f0ff656c944bd: Status 404 returned error can't find the container with id e3fb314e14eefd42729c7f316d80f3f8a573c2541cb5a2aa418f0ff656c944bd Oct 14 07:46:13 crc kubenswrapper[4870]: I1014 07:46:13.920868 4870 generic.go:334] "Generic (PLEG): container finished" podID="68c9a2ec-af40-419e-8c62-7e3b6a6e2a2f" containerID="5df19ebfbfe10106a45f4842930d4b9faa38625b109042f1d58472e79d4e6889" exitCode=0 Oct 14 07:46:13 crc kubenswrapper[4870]: I1014 07:46:13.921001 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9jz9m" event={"ID":"68c9a2ec-af40-419e-8c62-7e3b6a6e2a2f","Type":"ContainerDied","Data":"5df19ebfbfe10106a45f4842930d4b9faa38625b109042f1d58472e79d4e6889"} Oct 14 07:46:13 crc kubenswrapper[4870]: I1014 07:46:13.921335 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9jz9m" event={"ID":"68c9a2ec-af40-419e-8c62-7e3b6a6e2a2f","Type":"ContainerStarted","Data":"e3fb314e14eefd42729c7f316d80f3f8a573c2541cb5a2aa418f0ff656c944bd"} Oct 14 07:46:15 crc kubenswrapper[4870]: I1014 07:46:15.959939 4870 generic.go:334] "Generic (PLEG): container finished" podID="68c9a2ec-af40-419e-8c62-7e3b6a6e2a2f" containerID="792670f5c339de01437b6f02ec4a9986b6a1597fb291df0d169c9809228a8cac" exitCode=0 Oct 14 07:46:15 crc kubenswrapper[4870]: I1014 07:46:15.960176 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9jz9m" event={"ID":"68c9a2ec-af40-419e-8c62-7e3b6a6e2a2f","Type":"ContainerDied","Data":"792670f5c339de01437b6f02ec4a9986b6a1597fb291df0d169c9809228a8cac"} Oct 14 07:46:16 crc kubenswrapper[4870]: I1014 07:46:16.974115 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9jz9m" event={"ID":"68c9a2ec-af40-419e-8c62-7e3b6a6e2a2f","Type":"ContainerStarted","Data":"a783e70ae8d191fd34f19b5098be6bc9569b5209e60e812a31ce554e45d1fa8c"} Oct 14 07:46:17 crc kubenswrapper[4870]: I1014 07:46:17.009524 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-9jz9m" podStartSLOduration=2.3270363720000002 podStartE2EDuration="5.009501537s" podCreationTimestamp="2025-10-14 07:46:12 +0000 UTC" firstStartedPulling="2025-10-14 07:46:13.925845358 +0000 UTC m=+2709.623205769" lastFinishedPulling="2025-10-14 07:46:16.608310563 +0000 UTC m=+2712.305670934" observedRunningTime="2025-10-14 07:46:17.00396088 +0000 UTC m=+2712.701321251" watchObservedRunningTime="2025-10-14 07:46:17.009501537 +0000 UTC m=+2712.706861918" Oct 14 07:46:18 crc kubenswrapper[4870]: I1014 07:46:18.034198 4870 scope.go:117] "RemoveContainer" containerID="0c23f76241656fa35961005192bf783fc346c05081f076194d712c820a96b05d" Oct 14 07:46:18 crc kubenswrapper[4870]: E1014 07:46:18.034766 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 07:46:22 crc kubenswrapper[4870]: I1014 07:46:22.797255 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-9jz9m" Oct 14 07:46:22 crc kubenswrapper[4870]: I1014 07:46:22.797767 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-9jz9m" Oct 14 07:46:22 crc kubenswrapper[4870]: I1014 07:46:22.861948 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-9jz9m" Oct 14 07:46:23 crc kubenswrapper[4870]: I1014 07:46:23.088942 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-9jz9m" Oct 14 07:46:23 crc kubenswrapper[4870]: I1014 07:46:23.819858 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-9jz9m"] Oct 14 07:46:25 crc kubenswrapper[4870]: I1014 07:46:25.040250 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-9jz9m" podUID="68c9a2ec-af40-419e-8c62-7e3b6a6e2a2f" containerName="registry-server" containerID="cri-o://a783e70ae8d191fd34f19b5098be6bc9569b5209e60e812a31ce554e45d1fa8c" gracePeriod=2 Oct 14 07:46:25 crc kubenswrapper[4870]: I1014 07:46:25.466004 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9jz9m" Oct 14 07:46:25 crc kubenswrapper[4870]: I1014 07:46:25.639901 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6jg8p\" (UniqueName: \"kubernetes.io/projected/68c9a2ec-af40-419e-8c62-7e3b6a6e2a2f-kube-api-access-6jg8p\") pod \"68c9a2ec-af40-419e-8c62-7e3b6a6e2a2f\" (UID: \"68c9a2ec-af40-419e-8c62-7e3b6a6e2a2f\") " Oct 14 07:46:25 crc kubenswrapper[4870]: I1014 07:46:25.640507 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/68c9a2ec-af40-419e-8c62-7e3b6a6e2a2f-catalog-content\") pod \"68c9a2ec-af40-419e-8c62-7e3b6a6e2a2f\" (UID: \"68c9a2ec-af40-419e-8c62-7e3b6a6e2a2f\") " Oct 14 07:46:25 crc kubenswrapper[4870]: I1014 07:46:25.640542 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/68c9a2ec-af40-419e-8c62-7e3b6a6e2a2f-utilities\") pod \"68c9a2ec-af40-419e-8c62-7e3b6a6e2a2f\" (UID: \"68c9a2ec-af40-419e-8c62-7e3b6a6e2a2f\") " Oct 14 07:46:25 crc kubenswrapper[4870]: I1014 07:46:25.641911 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/68c9a2ec-af40-419e-8c62-7e3b6a6e2a2f-utilities" (OuterVolumeSpecName: "utilities") pod "68c9a2ec-af40-419e-8c62-7e3b6a6e2a2f" (UID: "68c9a2ec-af40-419e-8c62-7e3b6a6e2a2f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:46:25 crc kubenswrapper[4870]: I1014 07:46:25.645676 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/68c9a2ec-af40-419e-8c62-7e3b6a6e2a2f-kube-api-access-6jg8p" (OuterVolumeSpecName: "kube-api-access-6jg8p") pod "68c9a2ec-af40-419e-8c62-7e3b6a6e2a2f" (UID: "68c9a2ec-af40-419e-8c62-7e3b6a6e2a2f"). InnerVolumeSpecName "kube-api-access-6jg8p". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:46:25 crc kubenswrapper[4870]: I1014 07:46:25.659697 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/68c9a2ec-af40-419e-8c62-7e3b6a6e2a2f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "68c9a2ec-af40-419e-8c62-7e3b6a6e2a2f" (UID: "68c9a2ec-af40-419e-8c62-7e3b6a6e2a2f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:46:25 crc kubenswrapper[4870]: I1014 07:46:25.742164 4870 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/68c9a2ec-af40-419e-8c62-7e3b6a6e2a2f-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 07:46:25 crc kubenswrapper[4870]: I1014 07:46:25.742221 4870 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/68c9a2ec-af40-419e-8c62-7e3b6a6e2a2f-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 07:46:25 crc kubenswrapper[4870]: I1014 07:46:25.742243 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6jg8p\" (UniqueName: \"kubernetes.io/projected/68c9a2ec-af40-419e-8c62-7e3b6a6e2a2f-kube-api-access-6jg8p\") on node \"crc\" DevicePath \"\"" Oct 14 07:46:26 crc kubenswrapper[4870]: I1014 07:46:26.048068 4870 generic.go:334] "Generic (PLEG): container finished" podID="68c9a2ec-af40-419e-8c62-7e3b6a6e2a2f" containerID="a783e70ae8d191fd34f19b5098be6bc9569b5209e60e812a31ce554e45d1fa8c" exitCode=0 Oct 14 07:46:26 crc kubenswrapper[4870]: I1014 07:46:26.048138 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9jz9m" event={"ID":"68c9a2ec-af40-419e-8c62-7e3b6a6e2a2f","Type":"ContainerDied","Data":"a783e70ae8d191fd34f19b5098be6bc9569b5209e60e812a31ce554e45d1fa8c"} Oct 14 07:46:26 crc kubenswrapper[4870]: I1014 07:46:26.048159 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9jz9m" Oct 14 07:46:26 crc kubenswrapper[4870]: I1014 07:46:26.048187 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9jz9m" event={"ID":"68c9a2ec-af40-419e-8c62-7e3b6a6e2a2f","Type":"ContainerDied","Data":"e3fb314e14eefd42729c7f316d80f3f8a573c2541cb5a2aa418f0ff656c944bd"} Oct 14 07:46:26 crc kubenswrapper[4870]: I1014 07:46:26.048216 4870 scope.go:117] "RemoveContainer" containerID="a783e70ae8d191fd34f19b5098be6bc9569b5209e60e812a31ce554e45d1fa8c" Oct 14 07:46:26 crc kubenswrapper[4870]: I1014 07:46:26.067792 4870 scope.go:117] "RemoveContainer" containerID="792670f5c339de01437b6f02ec4a9986b6a1597fb291df0d169c9809228a8cac" Oct 14 07:46:26 crc kubenswrapper[4870]: I1014 07:46:26.090551 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-9jz9m"] Oct 14 07:46:26 crc kubenswrapper[4870]: I1014 07:46:26.096993 4870 scope.go:117] "RemoveContainer" containerID="5df19ebfbfe10106a45f4842930d4b9faa38625b109042f1d58472e79d4e6889" Oct 14 07:46:26 crc kubenswrapper[4870]: I1014 07:46:26.100668 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-9jz9m"] Oct 14 07:46:26 crc kubenswrapper[4870]: I1014 07:46:26.129470 4870 scope.go:117] "RemoveContainer" containerID="a783e70ae8d191fd34f19b5098be6bc9569b5209e60e812a31ce554e45d1fa8c" Oct 14 07:46:26 crc kubenswrapper[4870]: E1014 07:46:26.129859 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a783e70ae8d191fd34f19b5098be6bc9569b5209e60e812a31ce554e45d1fa8c\": container with ID starting with a783e70ae8d191fd34f19b5098be6bc9569b5209e60e812a31ce554e45d1fa8c not found: ID does not exist" containerID="a783e70ae8d191fd34f19b5098be6bc9569b5209e60e812a31ce554e45d1fa8c" Oct 14 07:46:26 crc kubenswrapper[4870]: I1014 07:46:26.129897 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a783e70ae8d191fd34f19b5098be6bc9569b5209e60e812a31ce554e45d1fa8c"} err="failed to get container status \"a783e70ae8d191fd34f19b5098be6bc9569b5209e60e812a31ce554e45d1fa8c\": rpc error: code = NotFound desc = could not find container \"a783e70ae8d191fd34f19b5098be6bc9569b5209e60e812a31ce554e45d1fa8c\": container with ID starting with a783e70ae8d191fd34f19b5098be6bc9569b5209e60e812a31ce554e45d1fa8c not found: ID does not exist" Oct 14 07:46:26 crc kubenswrapper[4870]: I1014 07:46:26.129922 4870 scope.go:117] "RemoveContainer" containerID="792670f5c339de01437b6f02ec4a9986b6a1597fb291df0d169c9809228a8cac" Oct 14 07:46:26 crc kubenswrapper[4870]: E1014 07:46:26.130350 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"792670f5c339de01437b6f02ec4a9986b6a1597fb291df0d169c9809228a8cac\": container with ID starting with 792670f5c339de01437b6f02ec4a9986b6a1597fb291df0d169c9809228a8cac not found: ID does not exist" containerID="792670f5c339de01437b6f02ec4a9986b6a1597fb291df0d169c9809228a8cac" Oct 14 07:46:26 crc kubenswrapper[4870]: I1014 07:46:26.130378 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"792670f5c339de01437b6f02ec4a9986b6a1597fb291df0d169c9809228a8cac"} err="failed to get container status \"792670f5c339de01437b6f02ec4a9986b6a1597fb291df0d169c9809228a8cac\": rpc error: code = NotFound desc = could not find container \"792670f5c339de01437b6f02ec4a9986b6a1597fb291df0d169c9809228a8cac\": container with ID starting with 792670f5c339de01437b6f02ec4a9986b6a1597fb291df0d169c9809228a8cac not found: ID does not exist" Oct 14 07:46:26 crc kubenswrapper[4870]: I1014 07:46:26.130396 4870 scope.go:117] "RemoveContainer" containerID="5df19ebfbfe10106a45f4842930d4b9faa38625b109042f1d58472e79d4e6889" Oct 14 07:46:26 crc kubenswrapper[4870]: E1014 07:46:26.130660 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5df19ebfbfe10106a45f4842930d4b9faa38625b109042f1d58472e79d4e6889\": container with ID starting with 5df19ebfbfe10106a45f4842930d4b9faa38625b109042f1d58472e79d4e6889 not found: ID does not exist" containerID="5df19ebfbfe10106a45f4842930d4b9faa38625b109042f1d58472e79d4e6889" Oct 14 07:46:26 crc kubenswrapper[4870]: I1014 07:46:26.130688 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5df19ebfbfe10106a45f4842930d4b9faa38625b109042f1d58472e79d4e6889"} err="failed to get container status \"5df19ebfbfe10106a45f4842930d4b9faa38625b109042f1d58472e79d4e6889\": rpc error: code = NotFound desc = could not find container \"5df19ebfbfe10106a45f4842930d4b9faa38625b109042f1d58472e79d4e6889\": container with ID starting with 5df19ebfbfe10106a45f4842930d4b9faa38625b109042f1d58472e79d4e6889 not found: ID does not exist" Oct 14 07:46:27 crc kubenswrapper[4870]: I1014 07:46:27.049804 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="68c9a2ec-af40-419e-8c62-7e3b6a6e2a2f" path="/var/lib/kubelet/pods/68c9a2ec-af40-419e-8c62-7e3b6a6e2a2f/volumes" Oct 14 07:46:29 crc kubenswrapper[4870]: I1014 07:46:29.034722 4870 scope.go:117] "RemoveContainer" containerID="0c23f76241656fa35961005192bf783fc346c05081f076194d712c820a96b05d" Oct 14 07:46:30 crc kubenswrapper[4870]: I1014 07:46:30.090163 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" event={"ID":"874a5e45-dffd-4d17-b609-b7d3ed2eab07","Type":"ContainerStarted","Data":"46c3508808d7a164a3a6d8bec958ad1e42310dfe2ae689cbdcdff84796e75681"} Oct 14 07:47:11 crc kubenswrapper[4870]: I1014 07:47:11.058314 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-g9pdl"] Oct 14 07:47:11 crc kubenswrapper[4870]: E1014 07:47:11.059131 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68c9a2ec-af40-419e-8c62-7e3b6a6e2a2f" containerName="extract-utilities" Oct 14 07:47:11 crc kubenswrapper[4870]: I1014 07:47:11.059146 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="68c9a2ec-af40-419e-8c62-7e3b6a6e2a2f" containerName="extract-utilities" Oct 14 07:47:11 crc kubenswrapper[4870]: E1014 07:47:11.059167 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68c9a2ec-af40-419e-8c62-7e3b6a6e2a2f" containerName="extract-content" Oct 14 07:47:11 crc kubenswrapper[4870]: I1014 07:47:11.059176 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="68c9a2ec-af40-419e-8c62-7e3b6a6e2a2f" containerName="extract-content" Oct 14 07:47:11 crc kubenswrapper[4870]: E1014 07:47:11.059192 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68c9a2ec-af40-419e-8c62-7e3b6a6e2a2f" containerName="registry-server" Oct 14 07:47:11 crc kubenswrapper[4870]: I1014 07:47:11.059199 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="68c9a2ec-af40-419e-8c62-7e3b6a6e2a2f" containerName="registry-server" Oct 14 07:47:11 crc kubenswrapper[4870]: I1014 07:47:11.059363 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="68c9a2ec-af40-419e-8c62-7e3b6a6e2a2f" containerName="registry-server" Oct 14 07:47:11 crc kubenswrapper[4870]: I1014 07:47:11.060315 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-g9pdl" Oct 14 07:47:11 crc kubenswrapper[4870]: I1014 07:47:11.071016 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-g9pdl"] Oct 14 07:47:11 crc kubenswrapper[4870]: I1014 07:47:11.232978 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ff880f6e-cf37-4ed8-8812-1384da2c9522-catalog-content\") pod \"certified-operators-g9pdl\" (UID: \"ff880f6e-cf37-4ed8-8812-1384da2c9522\") " pod="openshift-marketplace/certified-operators-g9pdl" Oct 14 07:47:11 crc kubenswrapper[4870]: I1014 07:47:11.233068 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ff880f6e-cf37-4ed8-8812-1384da2c9522-utilities\") pod \"certified-operators-g9pdl\" (UID: \"ff880f6e-cf37-4ed8-8812-1384da2c9522\") " pod="openshift-marketplace/certified-operators-g9pdl" Oct 14 07:47:11 crc kubenswrapper[4870]: I1014 07:47:11.233088 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bctpw\" (UniqueName: \"kubernetes.io/projected/ff880f6e-cf37-4ed8-8812-1384da2c9522-kube-api-access-bctpw\") pod \"certified-operators-g9pdl\" (UID: \"ff880f6e-cf37-4ed8-8812-1384da2c9522\") " pod="openshift-marketplace/certified-operators-g9pdl" Oct 14 07:47:11 crc kubenswrapper[4870]: I1014 07:47:11.334677 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ff880f6e-cf37-4ed8-8812-1384da2c9522-utilities\") pod \"certified-operators-g9pdl\" (UID: \"ff880f6e-cf37-4ed8-8812-1384da2c9522\") " pod="openshift-marketplace/certified-operators-g9pdl" Oct 14 07:47:11 crc kubenswrapper[4870]: I1014 07:47:11.334732 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bctpw\" (UniqueName: \"kubernetes.io/projected/ff880f6e-cf37-4ed8-8812-1384da2c9522-kube-api-access-bctpw\") pod \"certified-operators-g9pdl\" (UID: \"ff880f6e-cf37-4ed8-8812-1384da2c9522\") " pod="openshift-marketplace/certified-operators-g9pdl" Oct 14 07:47:11 crc kubenswrapper[4870]: I1014 07:47:11.334872 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ff880f6e-cf37-4ed8-8812-1384da2c9522-catalog-content\") pod \"certified-operators-g9pdl\" (UID: \"ff880f6e-cf37-4ed8-8812-1384da2c9522\") " pod="openshift-marketplace/certified-operators-g9pdl" Oct 14 07:47:11 crc kubenswrapper[4870]: I1014 07:47:11.335396 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ff880f6e-cf37-4ed8-8812-1384da2c9522-catalog-content\") pod \"certified-operators-g9pdl\" (UID: \"ff880f6e-cf37-4ed8-8812-1384da2c9522\") " pod="openshift-marketplace/certified-operators-g9pdl" Oct 14 07:47:11 crc kubenswrapper[4870]: I1014 07:47:11.335585 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ff880f6e-cf37-4ed8-8812-1384da2c9522-utilities\") pod \"certified-operators-g9pdl\" (UID: \"ff880f6e-cf37-4ed8-8812-1384da2c9522\") " pod="openshift-marketplace/certified-operators-g9pdl" Oct 14 07:47:11 crc kubenswrapper[4870]: I1014 07:47:11.354104 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bctpw\" (UniqueName: \"kubernetes.io/projected/ff880f6e-cf37-4ed8-8812-1384da2c9522-kube-api-access-bctpw\") pod \"certified-operators-g9pdl\" (UID: \"ff880f6e-cf37-4ed8-8812-1384da2c9522\") " pod="openshift-marketplace/certified-operators-g9pdl" Oct 14 07:47:11 crc kubenswrapper[4870]: I1014 07:47:11.435914 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-g9pdl" Oct 14 07:47:11 crc kubenswrapper[4870]: I1014 07:47:11.906237 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-g9pdl"] Oct 14 07:47:12 crc kubenswrapper[4870]: I1014 07:47:12.490407 4870 generic.go:334] "Generic (PLEG): container finished" podID="ff880f6e-cf37-4ed8-8812-1384da2c9522" containerID="54a0724c067db3a68785a1ce2241644586ff1f033541e917470c1822922e1e0a" exitCode=0 Oct 14 07:47:12 crc kubenswrapper[4870]: I1014 07:47:12.490555 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-g9pdl" event={"ID":"ff880f6e-cf37-4ed8-8812-1384da2c9522","Type":"ContainerDied","Data":"54a0724c067db3a68785a1ce2241644586ff1f033541e917470c1822922e1e0a"} Oct 14 07:47:12 crc kubenswrapper[4870]: I1014 07:47:12.490955 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-g9pdl" event={"ID":"ff880f6e-cf37-4ed8-8812-1384da2c9522","Type":"ContainerStarted","Data":"d841a5910e817fa16e5a90a72168aaad6841fec0b62ed5a99cc2f12c8181a22e"} Oct 14 07:47:13 crc kubenswrapper[4870]: I1014 07:47:13.507825 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-g9pdl" event={"ID":"ff880f6e-cf37-4ed8-8812-1384da2c9522","Type":"ContainerStarted","Data":"6792b0d2412a49ee4df0c3c0ba659c79f9577b64fd5b2769bc2dd7c09de34d2f"} Oct 14 07:47:14 crc kubenswrapper[4870]: I1014 07:47:14.523799 4870 generic.go:334] "Generic (PLEG): container finished" podID="ff880f6e-cf37-4ed8-8812-1384da2c9522" containerID="6792b0d2412a49ee4df0c3c0ba659c79f9577b64fd5b2769bc2dd7c09de34d2f" exitCode=0 Oct 14 07:47:14 crc kubenswrapper[4870]: I1014 07:47:14.523855 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-g9pdl" event={"ID":"ff880f6e-cf37-4ed8-8812-1384da2c9522","Type":"ContainerDied","Data":"6792b0d2412a49ee4df0c3c0ba659c79f9577b64fd5b2769bc2dd7c09de34d2f"} Oct 14 07:47:15 crc kubenswrapper[4870]: I1014 07:47:15.535985 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-g9pdl" event={"ID":"ff880f6e-cf37-4ed8-8812-1384da2c9522","Type":"ContainerStarted","Data":"f94c38a539afadf020792dea9200c6c9ff813d3da76948e60766f5611ee93658"} Oct 14 07:47:15 crc kubenswrapper[4870]: I1014 07:47:15.572996 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-g9pdl" podStartSLOduration=1.9744492660000001 podStartE2EDuration="4.572976123s" podCreationTimestamp="2025-10-14 07:47:11 +0000 UTC" firstStartedPulling="2025-10-14 07:47:12.492641317 +0000 UTC m=+2768.190001718" lastFinishedPulling="2025-10-14 07:47:15.091168164 +0000 UTC m=+2770.788528575" observedRunningTime="2025-10-14 07:47:15.568798539 +0000 UTC m=+2771.266158960" watchObservedRunningTime="2025-10-14 07:47:15.572976123 +0000 UTC m=+2771.270336494" Oct 14 07:47:21 crc kubenswrapper[4870]: I1014 07:47:21.436670 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-g9pdl" Oct 14 07:47:21 crc kubenswrapper[4870]: I1014 07:47:21.437324 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-g9pdl" Oct 14 07:47:21 crc kubenswrapper[4870]: I1014 07:47:21.508382 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-g9pdl" Oct 14 07:47:21 crc kubenswrapper[4870]: I1014 07:47:21.670144 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-g9pdl" Oct 14 07:47:21 crc kubenswrapper[4870]: I1014 07:47:21.751578 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-g9pdl"] Oct 14 07:47:23 crc kubenswrapper[4870]: I1014 07:47:23.615280 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-g9pdl" podUID="ff880f6e-cf37-4ed8-8812-1384da2c9522" containerName="registry-server" containerID="cri-o://f94c38a539afadf020792dea9200c6c9ff813d3da76948e60766f5611ee93658" gracePeriod=2 Oct 14 07:47:24 crc kubenswrapper[4870]: I1014 07:47:24.098778 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-g9pdl" Oct 14 07:47:24 crc kubenswrapper[4870]: I1014 07:47:24.281000 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bctpw\" (UniqueName: \"kubernetes.io/projected/ff880f6e-cf37-4ed8-8812-1384da2c9522-kube-api-access-bctpw\") pod \"ff880f6e-cf37-4ed8-8812-1384da2c9522\" (UID: \"ff880f6e-cf37-4ed8-8812-1384da2c9522\") " Oct 14 07:47:24 crc kubenswrapper[4870]: I1014 07:47:24.281196 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ff880f6e-cf37-4ed8-8812-1384da2c9522-utilities\") pod \"ff880f6e-cf37-4ed8-8812-1384da2c9522\" (UID: \"ff880f6e-cf37-4ed8-8812-1384da2c9522\") " Oct 14 07:47:24 crc kubenswrapper[4870]: I1014 07:47:24.281309 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ff880f6e-cf37-4ed8-8812-1384da2c9522-catalog-content\") pod \"ff880f6e-cf37-4ed8-8812-1384da2c9522\" (UID: \"ff880f6e-cf37-4ed8-8812-1384da2c9522\") " Oct 14 07:47:24 crc kubenswrapper[4870]: I1014 07:47:24.283257 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ff880f6e-cf37-4ed8-8812-1384da2c9522-utilities" (OuterVolumeSpecName: "utilities") pod "ff880f6e-cf37-4ed8-8812-1384da2c9522" (UID: "ff880f6e-cf37-4ed8-8812-1384da2c9522"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:47:24 crc kubenswrapper[4870]: I1014 07:47:24.290595 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ff880f6e-cf37-4ed8-8812-1384da2c9522-kube-api-access-bctpw" (OuterVolumeSpecName: "kube-api-access-bctpw") pod "ff880f6e-cf37-4ed8-8812-1384da2c9522" (UID: "ff880f6e-cf37-4ed8-8812-1384da2c9522"). InnerVolumeSpecName "kube-api-access-bctpw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:47:24 crc kubenswrapper[4870]: I1014 07:47:24.357858 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ff880f6e-cf37-4ed8-8812-1384da2c9522-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ff880f6e-cf37-4ed8-8812-1384da2c9522" (UID: "ff880f6e-cf37-4ed8-8812-1384da2c9522"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:47:24 crc kubenswrapper[4870]: I1014 07:47:24.383005 4870 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ff880f6e-cf37-4ed8-8812-1384da2c9522-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 07:47:24 crc kubenswrapper[4870]: I1014 07:47:24.383053 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bctpw\" (UniqueName: \"kubernetes.io/projected/ff880f6e-cf37-4ed8-8812-1384da2c9522-kube-api-access-bctpw\") on node \"crc\" DevicePath \"\"" Oct 14 07:47:24 crc kubenswrapper[4870]: I1014 07:47:24.383075 4870 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ff880f6e-cf37-4ed8-8812-1384da2c9522-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 07:47:24 crc kubenswrapper[4870]: I1014 07:47:24.633420 4870 generic.go:334] "Generic (PLEG): container finished" podID="ff880f6e-cf37-4ed8-8812-1384da2c9522" containerID="f94c38a539afadf020792dea9200c6c9ff813d3da76948e60766f5611ee93658" exitCode=0 Oct 14 07:47:24 crc kubenswrapper[4870]: I1014 07:47:24.633556 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-g9pdl" Oct 14 07:47:24 crc kubenswrapper[4870]: I1014 07:47:24.633604 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-g9pdl" event={"ID":"ff880f6e-cf37-4ed8-8812-1384da2c9522","Type":"ContainerDied","Data":"f94c38a539afadf020792dea9200c6c9ff813d3da76948e60766f5611ee93658"} Oct 14 07:47:24 crc kubenswrapper[4870]: I1014 07:47:24.635582 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-g9pdl" event={"ID":"ff880f6e-cf37-4ed8-8812-1384da2c9522","Type":"ContainerDied","Data":"d841a5910e817fa16e5a90a72168aaad6841fec0b62ed5a99cc2f12c8181a22e"} Oct 14 07:47:24 crc kubenswrapper[4870]: I1014 07:47:24.635635 4870 scope.go:117] "RemoveContainer" containerID="f94c38a539afadf020792dea9200c6c9ff813d3da76948e60766f5611ee93658" Oct 14 07:47:24 crc kubenswrapper[4870]: I1014 07:47:24.678800 4870 scope.go:117] "RemoveContainer" containerID="6792b0d2412a49ee4df0c3c0ba659c79f9577b64fd5b2769bc2dd7c09de34d2f" Oct 14 07:47:24 crc kubenswrapper[4870]: I1014 07:47:24.698306 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-g9pdl"] Oct 14 07:47:24 crc kubenswrapper[4870]: I1014 07:47:24.709481 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-g9pdl"] Oct 14 07:47:24 crc kubenswrapper[4870]: I1014 07:47:24.716804 4870 scope.go:117] "RemoveContainer" containerID="54a0724c067db3a68785a1ce2241644586ff1f033541e917470c1822922e1e0a" Oct 14 07:47:24 crc kubenswrapper[4870]: I1014 07:47:24.754496 4870 scope.go:117] "RemoveContainer" containerID="f94c38a539afadf020792dea9200c6c9ff813d3da76948e60766f5611ee93658" Oct 14 07:47:24 crc kubenswrapper[4870]: E1014 07:47:24.755203 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f94c38a539afadf020792dea9200c6c9ff813d3da76948e60766f5611ee93658\": container with ID starting with f94c38a539afadf020792dea9200c6c9ff813d3da76948e60766f5611ee93658 not found: ID does not exist" containerID="f94c38a539afadf020792dea9200c6c9ff813d3da76948e60766f5611ee93658" Oct 14 07:47:24 crc kubenswrapper[4870]: I1014 07:47:24.755265 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f94c38a539afadf020792dea9200c6c9ff813d3da76948e60766f5611ee93658"} err="failed to get container status \"f94c38a539afadf020792dea9200c6c9ff813d3da76948e60766f5611ee93658\": rpc error: code = NotFound desc = could not find container \"f94c38a539afadf020792dea9200c6c9ff813d3da76948e60766f5611ee93658\": container with ID starting with f94c38a539afadf020792dea9200c6c9ff813d3da76948e60766f5611ee93658 not found: ID does not exist" Oct 14 07:47:24 crc kubenswrapper[4870]: I1014 07:47:24.755307 4870 scope.go:117] "RemoveContainer" containerID="6792b0d2412a49ee4df0c3c0ba659c79f9577b64fd5b2769bc2dd7c09de34d2f" Oct 14 07:47:24 crc kubenswrapper[4870]: E1014 07:47:24.755963 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6792b0d2412a49ee4df0c3c0ba659c79f9577b64fd5b2769bc2dd7c09de34d2f\": container with ID starting with 6792b0d2412a49ee4df0c3c0ba659c79f9577b64fd5b2769bc2dd7c09de34d2f not found: ID does not exist" containerID="6792b0d2412a49ee4df0c3c0ba659c79f9577b64fd5b2769bc2dd7c09de34d2f" Oct 14 07:47:24 crc kubenswrapper[4870]: I1014 07:47:24.756043 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6792b0d2412a49ee4df0c3c0ba659c79f9577b64fd5b2769bc2dd7c09de34d2f"} err="failed to get container status \"6792b0d2412a49ee4df0c3c0ba659c79f9577b64fd5b2769bc2dd7c09de34d2f\": rpc error: code = NotFound desc = could not find container \"6792b0d2412a49ee4df0c3c0ba659c79f9577b64fd5b2769bc2dd7c09de34d2f\": container with ID starting with 6792b0d2412a49ee4df0c3c0ba659c79f9577b64fd5b2769bc2dd7c09de34d2f not found: ID does not exist" Oct 14 07:47:24 crc kubenswrapper[4870]: I1014 07:47:24.756088 4870 scope.go:117] "RemoveContainer" containerID="54a0724c067db3a68785a1ce2241644586ff1f033541e917470c1822922e1e0a" Oct 14 07:47:24 crc kubenswrapper[4870]: E1014 07:47:24.756640 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"54a0724c067db3a68785a1ce2241644586ff1f033541e917470c1822922e1e0a\": container with ID starting with 54a0724c067db3a68785a1ce2241644586ff1f033541e917470c1822922e1e0a not found: ID does not exist" containerID="54a0724c067db3a68785a1ce2241644586ff1f033541e917470c1822922e1e0a" Oct 14 07:47:24 crc kubenswrapper[4870]: I1014 07:47:24.756689 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"54a0724c067db3a68785a1ce2241644586ff1f033541e917470c1822922e1e0a"} err="failed to get container status \"54a0724c067db3a68785a1ce2241644586ff1f033541e917470c1822922e1e0a\": rpc error: code = NotFound desc = could not find container \"54a0724c067db3a68785a1ce2241644586ff1f033541e917470c1822922e1e0a\": container with ID starting with 54a0724c067db3a68785a1ce2241644586ff1f033541e917470c1822922e1e0a not found: ID does not exist" Oct 14 07:47:25 crc kubenswrapper[4870]: I1014 07:47:25.046721 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ff880f6e-cf37-4ed8-8812-1384da2c9522" path="/var/lib/kubelet/pods/ff880f6e-cf37-4ed8-8812-1384da2c9522/volumes" Oct 14 07:48:53 crc kubenswrapper[4870]: I1014 07:48:53.951080 4870 patch_prober.go:28] interesting pod/machine-config-daemon-7tvc7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 07:48:53 crc kubenswrapper[4870]: I1014 07:48:53.951756 4870 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 07:49:23 crc kubenswrapper[4870]: I1014 07:49:23.951706 4870 patch_prober.go:28] interesting pod/machine-config-daemon-7tvc7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 07:49:23 crc kubenswrapper[4870]: I1014 07:49:23.952294 4870 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 07:49:53 crc kubenswrapper[4870]: I1014 07:49:53.951675 4870 patch_prober.go:28] interesting pod/machine-config-daemon-7tvc7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 07:49:53 crc kubenswrapper[4870]: I1014 07:49:53.952332 4870 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 07:49:53 crc kubenswrapper[4870]: I1014 07:49:53.952396 4870 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" Oct 14 07:49:53 crc kubenswrapper[4870]: I1014 07:49:53.953409 4870 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"46c3508808d7a164a3a6d8bec958ad1e42310dfe2ae689cbdcdff84796e75681"} pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 14 07:49:53 crc kubenswrapper[4870]: I1014 07:49:53.953537 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerName="machine-config-daemon" containerID="cri-o://46c3508808d7a164a3a6d8bec958ad1e42310dfe2ae689cbdcdff84796e75681" gracePeriod=600 Oct 14 07:49:54 crc kubenswrapper[4870]: I1014 07:49:54.121052 4870 generic.go:334] "Generic (PLEG): container finished" podID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerID="46c3508808d7a164a3a6d8bec958ad1e42310dfe2ae689cbdcdff84796e75681" exitCode=0 Oct 14 07:49:54 crc kubenswrapper[4870]: I1014 07:49:54.121401 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" event={"ID":"874a5e45-dffd-4d17-b609-b7d3ed2eab07","Type":"ContainerDied","Data":"46c3508808d7a164a3a6d8bec958ad1e42310dfe2ae689cbdcdff84796e75681"} Oct 14 07:49:54 crc kubenswrapper[4870]: I1014 07:49:54.121576 4870 scope.go:117] "RemoveContainer" containerID="0c23f76241656fa35961005192bf783fc346c05081f076194d712c820a96b05d" Oct 14 07:49:55 crc kubenswrapper[4870]: I1014 07:49:55.136658 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" event={"ID":"874a5e45-dffd-4d17-b609-b7d3ed2eab07","Type":"ContainerStarted","Data":"d9859555f9dac6de3438df8cb32e7aedced80b131b2e63a298d227e4c67cd16e"} Oct 14 07:52:23 crc kubenswrapper[4870]: I1014 07:52:23.950690 4870 patch_prober.go:28] interesting pod/machine-config-daemon-7tvc7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 07:52:23 crc kubenswrapper[4870]: I1014 07:52:23.951386 4870 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 07:52:53 crc kubenswrapper[4870]: I1014 07:52:53.951069 4870 patch_prober.go:28] interesting pod/machine-config-daemon-7tvc7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 07:52:53 crc kubenswrapper[4870]: I1014 07:52:53.951833 4870 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 07:53:23 crc kubenswrapper[4870]: I1014 07:53:23.951330 4870 patch_prober.go:28] interesting pod/machine-config-daemon-7tvc7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 07:53:23 crc kubenswrapper[4870]: I1014 07:53:23.951972 4870 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 07:53:23 crc kubenswrapper[4870]: I1014 07:53:23.952026 4870 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" Oct 14 07:53:23 crc kubenswrapper[4870]: I1014 07:53:23.952704 4870 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d9859555f9dac6de3438df8cb32e7aedced80b131b2e63a298d227e4c67cd16e"} pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 14 07:53:23 crc kubenswrapper[4870]: I1014 07:53:23.952773 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerName="machine-config-daemon" containerID="cri-o://d9859555f9dac6de3438df8cb32e7aedced80b131b2e63a298d227e4c67cd16e" gracePeriod=600 Oct 14 07:53:24 crc kubenswrapper[4870]: E1014 07:53:24.096484 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 07:53:24 crc kubenswrapper[4870]: I1014 07:53:24.172011 4870 generic.go:334] "Generic (PLEG): container finished" podID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerID="d9859555f9dac6de3438df8cb32e7aedced80b131b2e63a298d227e4c67cd16e" exitCode=0 Oct 14 07:53:24 crc kubenswrapper[4870]: I1014 07:53:24.172089 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" event={"ID":"874a5e45-dffd-4d17-b609-b7d3ed2eab07","Type":"ContainerDied","Data":"d9859555f9dac6de3438df8cb32e7aedced80b131b2e63a298d227e4c67cd16e"} Oct 14 07:53:24 crc kubenswrapper[4870]: I1014 07:53:24.172806 4870 scope.go:117] "RemoveContainer" containerID="46c3508808d7a164a3a6d8bec958ad1e42310dfe2ae689cbdcdff84796e75681" Oct 14 07:53:24 crc kubenswrapper[4870]: I1014 07:53:24.173627 4870 scope.go:117] "RemoveContainer" containerID="d9859555f9dac6de3438df8cb32e7aedced80b131b2e63a298d227e4c67cd16e" Oct 14 07:53:24 crc kubenswrapper[4870]: E1014 07:53:24.174551 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 07:53:38 crc kubenswrapper[4870]: I1014 07:53:38.034377 4870 scope.go:117] "RemoveContainer" containerID="d9859555f9dac6de3438df8cb32e7aedced80b131b2e63a298d227e4c67cd16e" Oct 14 07:53:38 crc kubenswrapper[4870]: E1014 07:53:38.035817 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 07:53:51 crc kubenswrapper[4870]: I1014 07:53:51.033729 4870 scope.go:117] "RemoveContainer" containerID="d9859555f9dac6de3438df8cb32e7aedced80b131b2e63a298d227e4c67cd16e" Oct 14 07:53:51 crc kubenswrapper[4870]: E1014 07:53:51.034896 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 07:54:03 crc kubenswrapper[4870]: I1014 07:54:03.034623 4870 scope.go:117] "RemoveContainer" containerID="d9859555f9dac6de3438df8cb32e7aedced80b131b2e63a298d227e4c67cd16e" Oct 14 07:54:03 crc kubenswrapper[4870]: E1014 07:54:03.035586 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 07:54:15 crc kubenswrapper[4870]: I1014 07:54:15.042842 4870 scope.go:117] "RemoveContainer" containerID="d9859555f9dac6de3438df8cb32e7aedced80b131b2e63a298d227e4c67cd16e" Oct 14 07:54:15 crc kubenswrapper[4870]: E1014 07:54:15.043753 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 07:54:28 crc kubenswrapper[4870]: I1014 07:54:28.034791 4870 scope.go:117] "RemoveContainer" containerID="d9859555f9dac6de3438df8cb32e7aedced80b131b2e63a298d227e4c67cd16e" Oct 14 07:54:28 crc kubenswrapper[4870]: E1014 07:54:28.036230 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 07:54:42 crc kubenswrapper[4870]: I1014 07:54:42.035026 4870 scope.go:117] "RemoveContainer" containerID="d9859555f9dac6de3438df8cb32e7aedced80b131b2e63a298d227e4c67cd16e" Oct 14 07:54:42 crc kubenswrapper[4870]: E1014 07:54:42.036009 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 07:54:57 crc kubenswrapper[4870]: I1014 07:54:57.034579 4870 scope.go:117] "RemoveContainer" containerID="d9859555f9dac6de3438df8cb32e7aedced80b131b2e63a298d227e4c67cd16e" Oct 14 07:54:57 crc kubenswrapper[4870]: E1014 07:54:57.035522 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 07:55:11 crc kubenswrapper[4870]: I1014 07:55:11.035121 4870 scope.go:117] "RemoveContainer" containerID="d9859555f9dac6de3438df8cb32e7aedced80b131b2e63a298d227e4c67cd16e" Oct 14 07:55:11 crc kubenswrapper[4870]: E1014 07:55:11.036829 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 07:55:22 crc kubenswrapper[4870]: I1014 07:55:22.035224 4870 scope.go:117] "RemoveContainer" containerID="d9859555f9dac6de3438df8cb32e7aedced80b131b2e63a298d227e4c67cd16e" Oct 14 07:55:22 crc kubenswrapper[4870]: E1014 07:55:22.036526 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 07:55:33 crc kubenswrapper[4870]: I1014 07:55:33.033695 4870 scope.go:117] "RemoveContainer" containerID="d9859555f9dac6de3438df8cb32e7aedced80b131b2e63a298d227e4c67cd16e" Oct 14 07:55:33 crc kubenswrapper[4870]: E1014 07:55:33.034933 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 07:55:44 crc kubenswrapper[4870]: I1014 07:55:44.035164 4870 scope.go:117] "RemoveContainer" containerID="d9859555f9dac6de3438df8cb32e7aedced80b131b2e63a298d227e4c67cd16e" Oct 14 07:55:44 crc kubenswrapper[4870]: E1014 07:55:44.036226 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 07:55:47 crc kubenswrapper[4870]: I1014 07:55:47.382016 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-b8j7v"] Oct 14 07:55:47 crc kubenswrapper[4870]: E1014 07:55:47.382648 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff880f6e-cf37-4ed8-8812-1384da2c9522" containerName="registry-server" Oct 14 07:55:47 crc kubenswrapper[4870]: I1014 07:55:47.382661 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff880f6e-cf37-4ed8-8812-1384da2c9522" containerName="registry-server" Oct 14 07:55:47 crc kubenswrapper[4870]: E1014 07:55:47.382698 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff880f6e-cf37-4ed8-8812-1384da2c9522" containerName="extract-content" Oct 14 07:55:47 crc kubenswrapper[4870]: I1014 07:55:47.382705 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff880f6e-cf37-4ed8-8812-1384da2c9522" containerName="extract-content" Oct 14 07:55:47 crc kubenswrapper[4870]: E1014 07:55:47.382715 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff880f6e-cf37-4ed8-8812-1384da2c9522" containerName="extract-utilities" Oct 14 07:55:47 crc kubenswrapper[4870]: I1014 07:55:47.382722 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff880f6e-cf37-4ed8-8812-1384da2c9522" containerName="extract-utilities" Oct 14 07:55:47 crc kubenswrapper[4870]: I1014 07:55:47.382873 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="ff880f6e-cf37-4ed8-8812-1384da2c9522" containerName="registry-server" Oct 14 07:55:47 crc kubenswrapper[4870]: I1014 07:55:47.383998 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-b8j7v" Oct 14 07:55:47 crc kubenswrapper[4870]: I1014 07:55:47.392565 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-b8j7v"] Oct 14 07:55:47 crc kubenswrapper[4870]: I1014 07:55:47.545078 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d98071c7-9eda-459d-a24b-7225c18a3d7b-catalog-content\") pod \"community-operators-b8j7v\" (UID: \"d98071c7-9eda-459d-a24b-7225c18a3d7b\") " pod="openshift-marketplace/community-operators-b8j7v" Oct 14 07:55:47 crc kubenswrapper[4870]: I1014 07:55:47.545174 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d98071c7-9eda-459d-a24b-7225c18a3d7b-utilities\") pod \"community-operators-b8j7v\" (UID: \"d98071c7-9eda-459d-a24b-7225c18a3d7b\") " pod="openshift-marketplace/community-operators-b8j7v" Oct 14 07:55:47 crc kubenswrapper[4870]: I1014 07:55:47.545215 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qqc8b\" (UniqueName: \"kubernetes.io/projected/d98071c7-9eda-459d-a24b-7225c18a3d7b-kube-api-access-qqc8b\") pod \"community-operators-b8j7v\" (UID: \"d98071c7-9eda-459d-a24b-7225c18a3d7b\") " pod="openshift-marketplace/community-operators-b8j7v" Oct 14 07:55:47 crc kubenswrapper[4870]: I1014 07:55:47.646267 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d98071c7-9eda-459d-a24b-7225c18a3d7b-catalog-content\") pod \"community-operators-b8j7v\" (UID: \"d98071c7-9eda-459d-a24b-7225c18a3d7b\") " pod="openshift-marketplace/community-operators-b8j7v" Oct 14 07:55:47 crc kubenswrapper[4870]: I1014 07:55:47.646387 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d98071c7-9eda-459d-a24b-7225c18a3d7b-utilities\") pod \"community-operators-b8j7v\" (UID: \"d98071c7-9eda-459d-a24b-7225c18a3d7b\") " pod="openshift-marketplace/community-operators-b8j7v" Oct 14 07:55:47 crc kubenswrapper[4870]: I1014 07:55:47.646425 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qqc8b\" (UniqueName: \"kubernetes.io/projected/d98071c7-9eda-459d-a24b-7225c18a3d7b-kube-api-access-qqc8b\") pod \"community-operators-b8j7v\" (UID: \"d98071c7-9eda-459d-a24b-7225c18a3d7b\") " pod="openshift-marketplace/community-operators-b8j7v" Oct 14 07:55:47 crc kubenswrapper[4870]: I1014 07:55:47.646806 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d98071c7-9eda-459d-a24b-7225c18a3d7b-catalog-content\") pod \"community-operators-b8j7v\" (UID: \"d98071c7-9eda-459d-a24b-7225c18a3d7b\") " pod="openshift-marketplace/community-operators-b8j7v" Oct 14 07:55:47 crc kubenswrapper[4870]: I1014 07:55:47.647051 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d98071c7-9eda-459d-a24b-7225c18a3d7b-utilities\") pod \"community-operators-b8j7v\" (UID: \"d98071c7-9eda-459d-a24b-7225c18a3d7b\") " pod="openshift-marketplace/community-operators-b8j7v" Oct 14 07:55:47 crc kubenswrapper[4870]: I1014 07:55:47.669193 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qqc8b\" (UniqueName: \"kubernetes.io/projected/d98071c7-9eda-459d-a24b-7225c18a3d7b-kube-api-access-qqc8b\") pod \"community-operators-b8j7v\" (UID: \"d98071c7-9eda-459d-a24b-7225c18a3d7b\") " pod="openshift-marketplace/community-operators-b8j7v" Oct 14 07:55:47 crc kubenswrapper[4870]: I1014 07:55:47.719586 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-b8j7v" Oct 14 07:55:47 crc kubenswrapper[4870]: I1014 07:55:47.983534 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-b8j7v"] Oct 14 07:55:48 crc kubenswrapper[4870]: I1014 07:55:48.636039 4870 generic.go:334] "Generic (PLEG): container finished" podID="d98071c7-9eda-459d-a24b-7225c18a3d7b" containerID="41dbad60d81c981324a1b530af3b2e360de4131c56e643abb7c5771ce24be865" exitCode=0 Oct 14 07:55:48 crc kubenswrapper[4870]: I1014 07:55:48.636114 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-b8j7v" event={"ID":"d98071c7-9eda-459d-a24b-7225c18a3d7b","Type":"ContainerDied","Data":"41dbad60d81c981324a1b530af3b2e360de4131c56e643abb7c5771ce24be865"} Oct 14 07:55:48 crc kubenswrapper[4870]: I1014 07:55:48.636184 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-b8j7v" event={"ID":"d98071c7-9eda-459d-a24b-7225c18a3d7b","Type":"ContainerStarted","Data":"b8f1c0df88a7d2af7eee769fb6903607cff70aef6d3288cdad2ea4ad870eab34"} Oct 14 07:55:48 crc kubenswrapper[4870]: I1014 07:55:48.639053 4870 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 14 07:55:50 crc kubenswrapper[4870]: I1014 07:55:50.657072 4870 generic.go:334] "Generic (PLEG): container finished" podID="d98071c7-9eda-459d-a24b-7225c18a3d7b" containerID="1c891de60b2aba4ce6f5897069bce41433f7783849c8b3ca836920b395978f29" exitCode=0 Oct 14 07:55:50 crc kubenswrapper[4870]: I1014 07:55:50.657195 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-b8j7v" event={"ID":"d98071c7-9eda-459d-a24b-7225c18a3d7b","Type":"ContainerDied","Data":"1c891de60b2aba4ce6f5897069bce41433f7783849c8b3ca836920b395978f29"} Oct 14 07:55:51 crc kubenswrapper[4870]: I1014 07:55:51.669324 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-b8j7v" event={"ID":"d98071c7-9eda-459d-a24b-7225c18a3d7b","Type":"ContainerStarted","Data":"a6a6134016ce9828e6239b3f5b548db9078a39db1cec35155e8e2c9918b4206b"} Oct 14 07:55:51 crc kubenswrapper[4870]: I1014 07:55:51.706869 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-b8j7v" podStartSLOduration=2.231270419 podStartE2EDuration="4.706836992s" podCreationTimestamp="2025-10-14 07:55:47 +0000 UTC" firstStartedPulling="2025-10-14 07:55:48.638612488 +0000 UTC m=+3284.335972899" lastFinishedPulling="2025-10-14 07:55:51.114179061 +0000 UTC m=+3286.811539472" observedRunningTime="2025-10-14 07:55:51.698978872 +0000 UTC m=+3287.396339313" watchObservedRunningTime="2025-10-14 07:55:51.706836992 +0000 UTC m=+3287.404197403" Oct 14 07:55:57 crc kubenswrapper[4870]: I1014 07:55:57.720269 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-b8j7v" Oct 14 07:55:57 crc kubenswrapper[4870]: I1014 07:55:57.720814 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-b8j7v" Oct 14 07:55:57 crc kubenswrapper[4870]: I1014 07:55:57.784249 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-b8j7v" Oct 14 07:55:58 crc kubenswrapper[4870]: I1014 07:55:58.782031 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-b8j7v" Oct 14 07:55:58 crc kubenswrapper[4870]: I1014 07:55:58.852951 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-b8j7v"] Oct 14 07:55:59 crc kubenswrapper[4870]: I1014 07:55:59.034511 4870 scope.go:117] "RemoveContainer" containerID="d9859555f9dac6de3438df8cb32e7aedced80b131b2e63a298d227e4c67cd16e" Oct 14 07:55:59 crc kubenswrapper[4870]: E1014 07:55:59.034840 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 07:56:00 crc kubenswrapper[4870]: I1014 07:56:00.744542 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-b8j7v" podUID="d98071c7-9eda-459d-a24b-7225c18a3d7b" containerName="registry-server" containerID="cri-o://a6a6134016ce9828e6239b3f5b548db9078a39db1cec35155e8e2c9918b4206b" gracePeriod=2 Oct 14 07:56:01 crc kubenswrapper[4870]: I1014 07:56:01.284504 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-b8j7v" Oct 14 07:56:01 crc kubenswrapper[4870]: I1014 07:56:01.466734 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qqc8b\" (UniqueName: \"kubernetes.io/projected/d98071c7-9eda-459d-a24b-7225c18a3d7b-kube-api-access-qqc8b\") pod \"d98071c7-9eda-459d-a24b-7225c18a3d7b\" (UID: \"d98071c7-9eda-459d-a24b-7225c18a3d7b\") " Oct 14 07:56:01 crc kubenswrapper[4870]: I1014 07:56:01.466807 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d98071c7-9eda-459d-a24b-7225c18a3d7b-utilities\") pod \"d98071c7-9eda-459d-a24b-7225c18a3d7b\" (UID: \"d98071c7-9eda-459d-a24b-7225c18a3d7b\") " Oct 14 07:56:01 crc kubenswrapper[4870]: I1014 07:56:01.466975 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d98071c7-9eda-459d-a24b-7225c18a3d7b-catalog-content\") pod \"d98071c7-9eda-459d-a24b-7225c18a3d7b\" (UID: \"d98071c7-9eda-459d-a24b-7225c18a3d7b\") " Oct 14 07:56:01 crc kubenswrapper[4870]: I1014 07:56:01.468184 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d98071c7-9eda-459d-a24b-7225c18a3d7b-utilities" (OuterVolumeSpecName: "utilities") pod "d98071c7-9eda-459d-a24b-7225c18a3d7b" (UID: "d98071c7-9eda-459d-a24b-7225c18a3d7b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:56:01 crc kubenswrapper[4870]: I1014 07:56:01.477105 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d98071c7-9eda-459d-a24b-7225c18a3d7b-kube-api-access-qqc8b" (OuterVolumeSpecName: "kube-api-access-qqc8b") pod "d98071c7-9eda-459d-a24b-7225c18a3d7b" (UID: "d98071c7-9eda-459d-a24b-7225c18a3d7b"). InnerVolumeSpecName "kube-api-access-qqc8b". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:56:01 crc kubenswrapper[4870]: I1014 07:56:01.516823 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d98071c7-9eda-459d-a24b-7225c18a3d7b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d98071c7-9eda-459d-a24b-7225c18a3d7b" (UID: "d98071c7-9eda-459d-a24b-7225c18a3d7b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:56:01 crc kubenswrapper[4870]: I1014 07:56:01.568592 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qqc8b\" (UniqueName: \"kubernetes.io/projected/d98071c7-9eda-459d-a24b-7225c18a3d7b-kube-api-access-qqc8b\") on node \"crc\" DevicePath \"\"" Oct 14 07:56:01 crc kubenswrapper[4870]: I1014 07:56:01.568624 4870 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d98071c7-9eda-459d-a24b-7225c18a3d7b-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 07:56:01 crc kubenswrapper[4870]: I1014 07:56:01.568634 4870 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d98071c7-9eda-459d-a24b-7225c18a3d7b-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 07:56:01 crc kubenswrapper[4870]: I1014 07:56:01.754576 4870 generic.go:334] "Generic (PLEG): container finished" podID="d98071c7-9eda-459d-a24b-7225c18a3d7b" containerID="a6a6134016ce9828e6239b3f5b548db9078a39db1cec35155e8e2c9918b4206b" exitCode=0 Oct 14 07:56:01 crc kubenswrapper[4870]: I1014 07:56:01.754650 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-b8j7v" Oct 14 07:56:01 crc kubenswrapper[4870]: I1014 07:56:01.754659 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-b8j7v" event={"ID":"d98071c7-9eda-459d-a24b-7225c18a3d7b","Type":"ContainerDied","Data":"a6a6134016ce9828e6239b3f5b548db9078a39db1cec35155e8e2c9918b4206b"} Oct 14 07:56:01 crc kubenswrapper[4870]: I1014 07:56:01.754714 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-b8j7v" event={"ID":"d98071c7-9eda-459d-a24b-7225c18a3d7b","Type":"ContainerDied","Data":"b8f1c0df88a7d2af7eee769fb6903607cff70aef6d3288cdad2ea4ad870eab34"} Oct 14 07:56:01 crc kubenswrapper[4870]: I1014 07:56:01.754753 4870 scope.go:117] "RemoveContainer" containerID="a6a6134016ce9828e6239b3f5b548db9078a39db1cec35155e8e2c9918b4206b" Oct 14 07:56:01 crc kubenswrapper[4870]: I1014 07:56:01.780669 4870 scope.go:117] "RemoveContainer" containerID="1c891de60b2aba4ce6f5897069bce41433f7783849c8b3ca836920b395978f29" Oct 14 07:56:01 crc kubenswrapper[4870]: I1014 07:56:01.827174 4870 scope.go:117] "RemoveContainer" containerID="41dbad60d81c981324a1b530af3b2e360de4131c56e643abb7c5771ce24be865" Oct 14 07:56:01 crc kubenswrapper[4870]: I1014 07:56:01.834691 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-b8j7v"] Oct 14 07:56:01 crc kubenswrapper[4870]: I1014 07:56:01.844291 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-b8j7v"] Oct 14 07:56:01 crc kubenswrapper[4870]: I1014 07:56:01.855882 4870 scope.go:117] "RemoveContainer" containerID="a6a6134016ce9828e6239b3f5b548db9078a39db1cec35155e8e2c9918b4206b" Oct 14 07:56:01 crc kubenswrapper[4870]: E1014 07:56:01.856683 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a6a6134016ce9828e6239b3f5b548db9078a39db1cec35155e8e2c9918b4206b\": container with ID starting with a6a6134016ce9828e6239b3f5b548db9078a39db1cec35155e8e2c9918b4206b not found: ID does not exist" containerID="a6a6134016ce9828e6239b3f5b548db9078a39db1cec35155e8e2c9918b4206b" Oct 14 07:56:01 crc kubenswrapper[4870]: I1014 07:56:01.856802 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a6a6134016ce9828e6239b3f5b548db9078a39db1cec35155e8e2c9918b4206b"} err="failed to get container status \"a6a6134016ce9828e6239b3f5b548db9078a39db1cec35155e8e2c9918b4206b\": rpc error: code = NotFound desc = could not find container \"a6a6134016ce9828e6239b3f5b548db9078a39db1cec35155e8e2c9918b4206b\": container with ID starting with a6a6134016ce9828e6239b3f5b548db9078a39db1cec35155e8e2c9918b4206b not found: ID does not exist" Oct 14 07:56:01 crc kubenswrapper[4870]: I1014 07:56:01.856844 4870 scope.go:117] "RemoveContainer" containerID="1c891de60b2aba4ce6f5897069bce41433f7783849c8b3ca836920b395978f29" Oct 14 07:56:01 crc kubenswrapper[4870]: E1014 07:56:01.857221 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1c891de60b2aba4ce6f5897069bce41433f7783849c8b3ca836920b395978f29\": container with ID starting with 1c891de60b2aba4ce6f5897069bce41433f7783849c8b3ca836920b395978f29 not found: ID does not exist" containerID="1c891de60b2aba4ce6f5897069bce41433f7783849c8b3ca836920b395978f29" Oct 14 07:56:01 crc kubenswrapper[4870]: I1014 07:56:01.857270 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1c891de60b2aba4ce6f5897069bce41433f7783849c8b3ca836920b395978f29"} err="failed to get container status \"1c891de60b2aba4ce6f5897069bce41433f7783849c8b3ca836920b395978f29\": rpc error: code = NotFound desc = could not find container \"1c891de60b2aba4ce6f5897069bce41433f7783849c8b3ca836920b395978f29\": container with ID starting with 1c891de60b2aba4ce6f5897069bce41433f7783849c8b3ca836920b395978f29 not found: ID does not exist" Oct 14 07:56:01 crc kubenswrapper[4870]: I1014 07:56:01.857306 4870 scope.go:117] "RemoveContainer" containerID="41dbad60d81c981324a1b530af3b2e360de4131c56e643abb7c5771ce24be865" Oct 14 07:56:01 crc kubenswrapper[4870]: E1014 07:56:01.857936 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"41dbad60d81c981324a1b530af3b2e360de4131c56e643abb7c5771ce24be865\": container with ID starting with 41dbad60d81c981324a1b530af3b2e360de4131c56e643abb7c5771ce24be865 not found: ID does not exist" containerID="41dbad60d81c981324a1b530af3b2e360de4131c56e643abb7c5771ce24be865" Oct 14 07:56:01 crc kubenswrapper[4870]: I1014 07:56:01.858023 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"41dbad60d81c981324a1b530af3b2e360de4131c56e643abb7c5771ce24be865"} err="failed to get container status \"41dbad60d81c981324a1b530af3b2e360de4131c56e643abb7c5771ce24be865\": rpc error: code = NotFound desc = could not find container \"41dbad60d81c981324a1b530af3b2e360de4131c56e643abb7c5771ce24be865\": container with ID starting with 41dbad60d81c981324a1b530af3b2e360de4131c56e643abb7c5771ce24be865 not found: ID does not exist" Oct 14 07:56:03 crc kubenswrapper[4870]: I1014 07:56:03.050435 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d98071c7-9eda-459d-a24b-7225c18a3d7b" path="/var/lib/kubelet/pods/d98071c7-9eda-459d-a24b-7225c18a3d7b/volumes" Oct 14 07:56:12 crc kubenswrapper[4870]: I1014 07:56:12.035183 4870 scope.go:117] "RemoveContainer" containerID="d9859555f9dac6de3438df8cb32e7aedced80b131b2e63a298d227e4c67cd16e" Oct 14 07:56:12 crc kubenswrapper[4870]: E1014 07:56:12.036156 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 07:56:26 crc kubenswrapper[4870]: I1014 07:56:26.033522 4870 scope.go:117] "RemoveContainer" containerID="d9859555f9dac6de3438df8cb32e7aedced80b131b2e63a298d227e4c67cd16e" Oct 14 07:56:26 crc kubenswrapper[4870]: E1014 07:56:26.034247 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 07:56:31 crc kubenswrapper[4870]: I1014 07:56:31.480959 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-dvns9"] Oct 14 07:56:31 crc kubenswrapper[4870]: E1014 07:56:31.481867 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d98071c7-9eda-459d-a24b-7225c18a3d7b" containerName="extract-utilities" Oct 14 07:56:31 crc kubenswrapper[4870]: I1014 07:56:31.481883 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="d98071c7-9eda-459d-a24b-7225c18a3d7b" containerName="extract-utilities" Oct 14 07:56:31 crc kubenswrapper[4870]: E1014 07:56:31.481911 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d98071c7-9eda-459d-a24b-7225c18a3d7b" containerName="extract-content" Oct 14 07:56:31 crc kubenswrapper[4870]: I1014 07:56:31.481919 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="d98071c7-9eda-459d-a24b-7225c18a3d7b" containerName="extract-content" Oct 14 07:56:31 crc kubenswrapper[4870]: E1014 07:56:31.481929 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d98071c7-9eda-459d-a24b-7225c18a3d7b" containerName="registry-server" Oct 14 07:56:31 crc kubenswrapper[4870]: I1014 07:56:31.481937 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="d98071c7-9eda-459d-a24b-7225c18a3d7b" containerName="registry-server" Oct 14 07:56:31 crc kubenswrapper[4870]: I1014 07:56:31.482128 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="d98071c7-9eda-459d-a24b-7225c18a3d7b" containerName="registry-server" Oct 14 07:56:31 crc kubenswrapper[4870]: I1014 07:56:31.483320 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dvns9" Oct 14 07:56:31 crc kubenswrapper[4870]: I1014 07:56:31.503222 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-dvns9"] Oct 14 07:56:31 crc kubenswrapper[4870]: I1014 07:56:31.638791 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9zdxq\" (UniqueName: \"kubernetes.io/projected/d736983d-775c-48e3-bf18-72c745b03ef3-kube-api-access-9zdxq\") pod \"redhat-marketplace-dvns9\" (UID: \"d736983d-775c-48e3-bf18-72c745b03ef3\") " pod="openshift-marketplace/redhat-marketplace-dvns9" Oct 14 07:56:31 crc kubenswrapper[4870]: I1014 07:56:31.638838 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d736983d-775c-48e3-bf18-72c745b03ef3-utilities\") pod \"redhat-marketplace-dvns9\" (UID: \"d736983d-775c-48e3-bf18-72c745b03ef3\") " pod="openshift-marketplace/redhat-marketplace-dvns9" Oct 14 07:56:31 crc kubenswrapper[4870]: I1014 07:56:31.638861 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d736983d-775c-48e3-bf18-72c745b03ef3-catalog-content\") pod \"redhat-marketplace-dvns9\" (UID: \"d736983d-775c-48e3-bf18-72c745b03ef3\") " pod="openshift-marketplace/redhat-marketplace-dvns9" Oct 14 07:56:31 crc kubenswrapper[4870]: I1014 07:56:31.740741 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9zdxq\" (UniqueName: \"kubernetes.io/projected/d736983d-775c-48e3-bf18-72c745b03ef3-kube-api-access-9zdxq\") pod \"redhat-marketplace-dvns9\" (UID: \"d736983d-775c-48e3-bf18-72c745b03ef3\") " pod="openshift-marketplace/redhat-marketplace-dvns9" Oct 14 07:56:31 crc kubenswrapper[4870]: I1014 07:56:31.740823 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d736983d-775c-48e3-bf18-72c745b03ef3-utilities\") pod \"redhat-marketplace-dvns9\" (UID: \"d736983d-775c-48e3-bf18-72c745b03ef3\") " pod="openshift-marketplace/redhat-marketplace-dvns9" Oct 14 07:56:31 crc kubenswrapper[4870]: I1014 07:56:31.740859 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d736983d-775c-48e3-bf18-72c745b03ef3-catalog-content\") pod \"redhat-marketplace-dvns9\" (UID: \"d736983d-775c-48e3-bf18-72c745b03ef3\") " pod="openshift-marketplace/redhat-marketplace-dvns9" Oct 14 07:56:31 crc kubenswrapper[4870]: I1014 07:56:31.741392 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d736983d-775c-48e3-bf18-72c745b03ef3-utilities\") pod \"redhat-marketplace-dvns9\" (UID: \"d736983d-775c-48e3-bf18-72c745b03ef3\") " pod="openshift-marketplace/redhat-marketplace-dvns9" Oct 14 07:56:31 crc kubenswrapper[4870]: I1014 07:56:31.741487 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d736983d-775c-48e3-bf18-72c745b03ef3-catalog-content\") pod \"redhat-marketplace-dvns9\" (UID: \"d736983d-775c-48e3-bf18-72c745b03ef3\") " pod="openshift-marketplace/redhat-marketplace-dvns9" Oct 14 07:56:31 crc kubenswrapper[4870]: I1014 07:56:31.766313 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9zdxq\" (UniqueName: \"kubernetes.io/projected/d736983d-775c-48e3-bf18-72c745b03ef3-kube-api-access-9zdxq\") pod \"redhat-marketplace-dvns9\" (UID: \"d736983d-775c-48e3-bf18-72c745b03ef3\") " pod="openshift-marketplace/redhat-marketplace-dvns9" Oct 14 07:56:31 crc kubenswrapper[4870]: I1014 07:56:31.815432 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dvns9" Oct 14 07:56:32 crc kubenswrapper[4870]: I1014 07:56:32.236243 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-dvns9"] Oct 14 07:56:33 crc kubenswrapper[4870]: I1014 07:56:33.034276 4870 generic.go:334] "Generic (PLEG): container finished" podID="d736983d-775c-48e3-bf18-72c745b03ef3" containerID="94dbef6e405c6719aade2f4ec442a23a30c543791828e9cbcc52bbc10ea0cc28" exitCode=0 Oct 14 07:56:33 crc kubenswrapper[4870]: I1014 07:56:33.064287 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dvns9" event={"ID":"d736983d-775c-48e3-bf18-72c745b03ef3","Type":"ContainerDied","Data":"94dbef6e405c6719aade2f4ec442a23a30c543791828e9cbcc52bbc10ea0cc28"} Oct 14 07:56:33 crc kubenswrapper[4870]: I1014 07:56:33.064326 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dvns9" event={"ID":"d736983d-775c-48e3-bf18-72c745b03ef3","Type":"ContainerStarted","Data":"fee43b7af0560a2dd3d9b33aa227d71cea105ed36ed2682cfff4dcd82dd2e61c"} Oct 14 07:56:34 crc kubenswrapper[4870]: I1014 07:56:34.045670 4870 generic.go:334] "Generic (PLEG): container finished" podID="d736983d-775c-48e3-bf18-72c745b03ef3" containerID="e7b624ba83d9ba28cf31308a9f456301ac7c2a1b1714fc1a89b072c533e7af37" exitCode=0 Oct 14 07:56:34 crc kubenswrapper[4870]: I1014 07:56:34.045712 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dvns9" event={"ID":"d736983d-775c-48e3-bf18-72c745b03ef3","Type":"ContainerDied","Data":"e7b624ba83d9ba28cf31308a9f456301ac7c2a1b1714fc1a89b072c533e7af37"} Oct 14 07:56:35 crc kubenswrapper[4870]: I1014 07:56:35.055178 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dvns9" event={"ID":"d736983d-775c-48e3-bf18-72c745b03ef3","Type":"ContainerStarted","Data":"0f407f04ad87c68e619b0bd7fe463baa9c7f555d99c7695817f6e336c08ef203"} Oct 14 07:56:35 crc kubenswrapper[4870]: I1014 07:56:35.085226 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-dvns9" podStartSLOduration=2.699702693 podStartE2EDuration="4.085201655s" podCreationTimestamp="2025-10-14 07:56:31 +0000 UTC" firstStartedPulling="2025-10-14 07:56:33.037891489 +0000 UTC m=+3328.735251860" lastFinishedPulling="2025-10-14 07:56:34.423390411 +0000 UTC m=+3330.120750822" observedRunningTime="2025-10-14 07:56:35.080033087 +0000 UTC m=+3330.777393478" watchObservedRunningTime="2025-10-14 07:56:35.085201655 +0000 UTC m=+3330.782562046" Oct 14 07:56:40 crc kubenswrapper[4870]: I1014 07:56:40.034553 4870 scope.go:117] "RemoveContainer" containerID="d9859555f9dac6de3438df8cb32e7aedced80b131b2e63a298d227e4c67cd16e" Oct 14 07:56:40 crc kubenswrapper[4870]: E1014 07:56:40.035531 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 07:56:41 crc kubenswrapper[4870]: I1014 07:56:41.815720 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-dvns9" Oct 14 07:56:41 crc kubenswrapper[4870]: I1014 07:56:41.815814 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-dvns9" Oct 14 07:56:41 crc kubenswrapper[4870]: I1014 07:56:41.878955 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-dvns9" Oct 14 07:56:42 crc kubenswrapper[4870]: I1014 07:56:42.173107 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-dvns9" Oct 14 07:56:42 crc kubenswrapper[4870]: I1014 07:56:42.238544 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-dvns9"] Oct 14 07:56:44 crc kubenswrapper[4870]: I1014 07:56:44.147262 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-dvns9" podUID="d736983d-775c-48e3-bf18-72c745b03ef3" containerName="registry-server" containerID="cri-o://0f407f04ad87c68e619b0bd7fe463baa9c7f555d99c7695817f6e336c08ef203" gracePeriod=2 Oct 14 07:56:44 crc kubenswrapper[4870]: I1014 07:56:44.633773 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dvns9" Oct 14 07:56:44 crc kubenswrapper[4870]: I1014 07:56:44.762862 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d736983d-775c-48e3-bf18-72c745b03ef3-utilities\") pod \"d736983d-775c-48e3-bf18-72c745b03ef3\" (UID: \"d736983d-775c-48e3-bf18-72c745b03ef3\") " Oct 14 07:56:44 crc kubenswrapper[4870]: I1014 07:56:44.762973 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d736983d-775c-48e3-bf18-72c745b03ef3-catalog-content\") pod \"d736983d-775c-48e3-bf18-72c745b03ef3\" (UID: \"d736983d-775c-48e3-bf18-72c745b03ef3\") " Oct 14 07:56:44 crc kubenswrapper[4870]: I1014 07:56:44.763045 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9zdxq\" (UniqueName: \"kubernetes.io/projected/d736983d-775c-48e3-bf18-72c745b03ef3-kube-api-access-9zdxq\") pod \"d736983d-775c-48e3-bf18-72c745b03ef3\" (UID: \"d736983d-775c-48e3-bf18-72c745b03ef3\") " Oct 14 07:56:44 crc kubenswrapper[4870]: I1014 07:56:44.764280 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d736983d-775c-48e3-bf18-72c745b03ef3-utilities" (OuterVolumeSpecName: "utilities") pod "d736983d-775c-48e3-bf18-72c745b03ef3" (UID: "d736983d-775c-48e3-bf18-72c745b03ef3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:56:44 crc kubenswrapper[4870]: I1014 07:56:44.770952 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d736983d-775c-48e3-bf18-72c745b03ef3-kube-api-access-9zdxq" (OuterVolumeSpecName: "kube-api-access-9zdxq") pod "d736983d-775c-48e3-bf18-72c745b03ef3" (UID: "d736983d-775c-48e3-bf18-72c745b03ef3"). InnerVolumeSpecName "kube-api-access-9zdxq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:56:44 crc kubenswrapper[4870]: I1014 07:56:44.790017 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d736983d-775c-48e3-bf18-72c745b03ef3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d736983d-775c-48e3-bf18-72c745b03ef3" (UID: "d736983d-775c-48e3-bf18-72c745b03ef3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:56:44 crc kubenswrapper[4870]: I1014 07:56:44.865164 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9zdxq\" (UniqueName: \"kubernetes.io/projected/d736983d-775c-48e3-bf18-72c745b03ef3-kube-api-access-9zdxq\") on node \"crc\" DevicePath \"\"" Oct 14 07:56:44 crc kubenswrapper[4870]: I1014 07:56:44.865217 4870 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d736983d-775c-48e3-bf18-72c745b03ef3-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 07:56:44 crc kubenswrapper[4870]: I1014 07:56:44.865240 4870 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d736983d-775c-48e3-bf18-72c745b03ef3-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 07:56:45 crc kubenswrapper[4870]: I1014 07:56:45.162570 4870 generic.go:334] "Generic (PLEG): container finished" podID="d736983d-775c-48e3-bf18-72c745b03ef3" containerID="0f407f04ad87c68e619b0bd7fe463baa9c7f555d99c7695817f6e336c08ef203" exitCode=0 Oct 14 07:56:45 crc kubenswrapper[4870]: I1014 07:56:45.162637 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dvns9" event={"ID":"d736983d-775c-48e3-bf18-72c745b03ef3","Type":"ContainerDied","Data":"0f407f04ad87c68e619b0bd7fe463baa9c7f555d99c7695817f6e336c08ef203"} Oct 14 07:56:45 crc kubenswrapper[4870]: I1014 07:56:45.162663 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dvns9" Oct 14 07:56:45 crc kubenswrapper[4870]: I1014 07:56:45.162689 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dvns9" event={"ID":"d736983d-775c-48e3-bf18-72c745b03ef3","Type":"ContainerDied","Data":"fee43b7af0560a2dd3d9b33aa227d71cea105ed36ed2682cfff4dcd82dd2e61c"} Oct 14 07:56:45 crc kubenswrapper[4870]: I1014 07:56:45.162717 4870 scope.go:117] "RemoveContainer" containerID="0f407f04ad87c68e619b0bd7fe463baa9c7f555d99c7695817f6e336c08ef203" Oct 14 07:56:45 crc kubenswrapper[4870]: I1014 07:56:45.200394 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-dvns9"] Oct 14 07:56:45 crc kubenswrapper[4870]: I1014 07:56:45.210915 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-dvns9"] Oct 14 07:56:45 crc kubenswrapper[4870]: I1014 07:56:45.211849 4870 scope.go:117] "RemoveContainer" containerID="e7b624ba83d9ba28cf31308a9f456301ac7c2a1b1714fc1a89b072c533e7af37" Oct 14 07:56:45 crc kubenswrapper[4870]: I1014 07:56:45.238385 4870 scope.go:117] "RemoveContainer" containerID="94dbef6e405c6719aade2f4ec442a23a30c543791828e9cbcc52bbc10ea0cc28" Oct 14 07:56:45 crc kubenswrapper[4870]: I1014 07:56:45.281752 4870 scope.go:117] "RemoveContainer" containerID="0f407f04ad87c68e619b0bd7fe463baa9c7f555d99c7695817f6e336c08ef203" Oct 14 07:56:45 crc kubenswrapper[4870]: E1014 07:56:45.282682 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0f407f04ad87c68e619b0bd7fe463baa9c7f555d99c7695817f6e336c08ef203\": container with ID starting with 0f407f04ad87c68e619b0bd7fe463baa9c7f555d99c7695817f6e336c08ef203 not found: ID does not exist" containerID="0f407f04ad87c68e619b0bd7fe463baa9c7f555d99c7695817f6e336c08ef203" Oct 14 07:56:45 crc kubenswrapper[4870]: I1014 07:56:45.282746 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0f407f04ad87c68e619b0bd7fe463baa9c7f555d99c7695817f6e336c08ef203"} err="failed to get container status \"0f407f04ad87c68e619b0bd7fe463baa9c7f555d99c7695817f6e336c08ef203\": rpc error: code = NotFound desc = could not find container \"0f407f04ad87c68e619b0bd7fe463baa9c7f555d99c7695817f6e336c08ef203\": container with ID starting with 0f407f04ad87c68e619b0bd7fe463baa9c7f555d99c7695817f6e336c08ef203 not found: ID does not exist" Oct 14 07:56:45 crc kubenswrapper[4870]: I1014 07:56:45.282788 4870 scope.go:117] "RemoveContainer" containerID="e7b624ba83d9ba28cf31308a9f456301ac7c2a1b1714fc1a89b072c533e7af37" Oct 14 07:56:45 crc kubenswrapper[4870]: E1014 07:56:45.283576 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e7b624ba83d9ba28cf31308a9f456301ac7c2a1b1714fc1a89b072c533e7af37\": container with ID starting with e7b624ba83d9ba28cf31308a9f456301ac7c2a1b1714fc1a89b072c533e7af37 not found: ID does not exist" containerID="e7b624ba83d9ba28cf31308a9f456301ac7c2a1b1714fc1a89b072c533e7af37" Oct 14 07:56:45 crc kubenswrapper[4870]: I1014 07:56:45.283680 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e7b624ba83d9ba28cf31308a9f456301ac7c2a1b1714fc1a89b072c533e7af37"} err="failed to get container status \"e7b624ba83d9ba28cf31308a9f456301ac7c2a1b1714fc1a89b072c533e7af37\": rpc error: code = NotFound desc = could not find container \"e7b624ba83d9ba28cf31308a9f456301ac7c2a1b1714fc1a89b072c533e7af37\": container with ID starting with e7b624ba83d9ba28cf31308a9f456301ac7c2a1b1714fc1a89b072c533e7af37 not found: ID does not exist" Oct 14 07:56:45 crc kubenswrapper[4870]: I1014 07:56:45.283741 4870 scope.go:117] "RemoveContainer" containerID="94dbef6e405c6719aade2f4ec442a23a30c543791828e9cbcc52bbc10ea0cc28" Oct 14 07:56:45 crc kubenswrapper[4870]: E1014 07:56:45.284409 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"94dbef6e405c6719aade2f4ec442a23a30c543791828e9cbcc52bbc10ea0cc28\": container with ID starting with 94dbef6e405c6719aade2f4ec442a23a30c543791828e9cbcc52bbc10ea0cc28 not found: ID does not exist" containerID="94dbef6e405c6719aade2f4ec442a23a30c543791828e9cbcc52bbc10ea0cc28" Oct 14 07:56:45 crc kubenswrapper[4870]: I1014 07:56:45.284485 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"94dbef6e405c6719aade2f4ec442a23a30c543791828e9cbcc52bbc10ea0cc28"} err="failed to get container status \"94dbef6e405c6719aade2f4ec442a23a30c543791828e9cbcc52bbc10ea0cc28\": rpc error: code = NotFound desc = could not find container \"94dbef6e405c6719aade2f4ec442a23a30c543791828e9cbcc52bbc10ea0cc28\": container with ID starting with 94dbef6e405c6719aade2f4ec442a23a30c543791828e9cbcc52bbc10ea0cc28 not found: ID does not exist" Oct 14 07:56:47 crc kubenswrapper[4870]: I1014 07:56:47.048759 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d736983d-775c-48e3-bf18-72c745b03ef3" path="/var/lib/kubelet/pods/d736983d-775c-48e3-bf18-72c745b03ef3/volumes" Oct 14 07:56:52 crc kubenswrapper[4870]: I1014 07:56:52.034329 4870 scope.go:117] "RemoveContainer" containerID="d9859555f9dac6de3438df8cb32e7aedced80b131b2e63a298d227e4c67cd16e" Oct 14 07:56:52 crc kubenswrapper[4870]: E1014 07:56:52.035117 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 07:57:05 crc kubenswrapper[4870]: I1014 07:57:05.060722 4870 scope.go:117] "RemoveContainer" containerID="d9859555f9dac6de3438df8cb32e7aedced80b131b2e63a298d227e4c67cd16e" Oct 14 07:57:05 crc kubenswrapper[4870]: E1014 07:57:05.062071 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 07:57:16 crc kubenswrapper[4870]: I1014 07:57:16.034121 4870 scope.go:117] "RemoveContainer" containerID="d9859555f9dac6de3438df8cb32e7aedced80b131b2e63a298d227e4c67cd16e" Oct 14 07:57:16 crc kubenswrapper[4870]: E1014 07:57:16.035231 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 07:57:29 crc kubenswrapper[4870]: I1014 07:57:29.033638 4870 scope.go:117] "RemoveContainer" containerID="d9859555f9dac6de3438df8cb32e7aedced80b131b2e63a298d227e4c67cd16e" Oct 14 07:57:29 crc kubenswrapper[4870]: E1014 07:57:29.034766 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 07:57:41 crc kubenswrapper[4870]: I1014 07:57:41.036258 4870 scope.go:117] "RemoveContainer" containerID="d9859555f9dac6de3438df8cb32e7aedced80b131b2e63a298d227e4c67cd16e" Oct 14 07:57:41 crc kubenswrapper[4870]: E1014 07:57:41.037074 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 07:57:52 crc kubenswrapper[4870]: I1014 07:57:52.034695 4870 scope.go:117] "RemoveContainer" containerID="d9859555f9dac6de3438df8cb32e7aedced80b131b2e63a298d227e4c67cd16e" Oct 14 07:57:52 crc kubenswrapper[4870]: E1014 07:57:52.035615 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 07:58:05 crc kubenswrapper[4870]: I1014 07:58:05.043987 4870 scope.go:117] "RemoveContainer" containerID="d9859555f9dac6de3438df8cb32e7aedced80b131b2e63a298d227e4c67cd16e" Oct 14 07:58:05 crc kubenswrapper[4870]: E1014 07:58:05.046971 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 07:58:12 crc kubenswrapper[4870]: I1014 07:58:12.010414 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-68hnm"] Oct 14 07:58:12 crc kubenswrapper[4870]: E1014 07:58:12.013800 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d736983d-775c-48e3-bf18-72c745b03ef3" containerName="registry-server" Oct 14 07:58:12 crc kubenswrapper[4870]: I1014 07:58:12.014037 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="d736983d-775c-48e3-bf18-72c745b03ef3" containerName="registry-server" Oct 14 07:58:12 crc kubenswrapper[4870]: E1014 07:58:12.014230 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d736983d-775c-48e3-bf18-72c745b03ef3" containerName="extract-utilities" Oct 14 07:58:12 crc kubenswrapper[4870]: I1014 07:58:12.014405 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="d736983d-775c-48e3-bf18-72c745b03ef3" containerName="extract-utilities" Oct 14 07:58:12 crc kubenswrapper[4870]: E1014 07:58:12.014678 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d736983d-775c-48e3-bf18-72c745b03ef3" containerName="extract-content" Oct 14 07:58:12 crc kubenswrapper[4870]: I1014 07:58:12.014903 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="d736983d-775c-48e3-bf18-72c745b03ef3" containerName="extract-content" Oct 14 07:58:12 crc kubenswrapper[4870]: I1014 07:58:12.015714 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="d736983d-775c-48e3-bf18-72c745b03ef3" containerName="registry-server" Oct 14 07:58:12 crc kubenswrapper[4870]: I1014 07:58:12.018407 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-68hnm" Oct 14 07:58:12 crc kubenswrapper[4870]: I1014 07:58:12.023413 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-68hnm"] Oct 14 07:58:12 crc kubenswrapper[4870]: I1014 07:58:12.137859 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mf2kz\" (UniqueName: \"kubernetes.io/projected/56990b82-7491-4f38-b9c8-ecc3fc0db284-kube-api-access-mf2kz\") pod \"certified-operators-68hnm\" (UID: \"56990b82-7491-4f38-b9c8-ecc3fc0db284\") " pod="openshift-marketplace/certified-operators-68hnm" Oct 14 07:58:12 crc kubenswrapper[4870]: I1014 07:58:12.137903 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/56990b82-7491-4f38-b9c8-ecc3fc0db284-catalog-content\") pod \"certified-operators-68hnm\" (UID: \"56990b82-7491-4f38-b9c8-ecc3fc0db284\") " pod="openshift-marketplace/certified-operators-68hnm" Oct 14 07:58:12 crc kubenswrapper[4870]: I1014 07:58:12.138476 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/56990b82-7491-4f38-b9c8-ecc3fc0db284-utilities\") pod \"certified-operators-68hnm\" (UID: \"56990b82-7491-4f38-b9c8-ecc3fc0db284\") " pod="openshift-marketplace/certified-operators-68hnm" Oct 14 07:58:12 crc kubenswrapper[4870]: I1014 07:58:12.240296 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/56990b82-7491-4f38-b9c8-ecc3fc0db284-utilities\") pod \"certified-operators-68hnm\" (UID: \"56990b82-7491-4f38-b9c8-ecc3fc0db284\") " pod="openshift-marketplace/certified-operators-68hnm" Oct 14 07:58:12 crc kubenswrapper[4870]: I1014 07:58:12.240389 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mf2kz\" (UniqueName: \"kubernetes.io/projected/56990b82-7491-4f38-b9c8-ecc3fc0db284-kube-api-access-mf2kz\") pod \"certified-operators-68hnm\" (UID: \"56990b82-7491-4f38-b9c8-ecc3fc0db284\") " pod="openshift-marketplace/certified-operators-68hnm" Oct 14 07:58:12 crc kubenswrapper[4870]: I1014 07:58:12.240440 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/56990b82-7491-4f38-b9c8-ecc3fc0db284-catalog-content\") pod \"certified-operators-68hnm\" (UID: \"56990b82-7491-4f38-b9c8-ecc3fc0db284\") " pod="openshift-marketplace/certified-operators-68hnm" Oct 14 07:58:12 crc kubenswrapper[4870]: I1014 07:58:12.241265 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/56990b82-7491-4f38-b9c8-ecc3fc0db284-catalog-content\") pod \"certified-operators-68hnm\" (UID: \"56990b82-7491-4f38-b9c8-ecc3fc0db284\") " pod="openshift-marketplace/certified-operators-68hnm" Oct 14 07:58:12 crc kubenswrapper[4870]: I1014 07:58:12.241262 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/56990b82-7491-4f38-b9c8-ecc3fc0db284-utilities\") pod \"certified-operators-68hnm\" (UID: \"56990b82-7491-4f38-b9c8-ecc3fc0db284\") " pod="openshift-marketplace/certified-operators-68hnm" Oct 14 07:58:12 crc kubenswrapper[4870]: I1014 07:58:12.260619 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mf2kz\" (UniqueName: \"kubernetes.io/projected/56990b82-7491-4f38-b9c8-ecc3fc0db284-kube-api-access-mf2kz\") pod \"certified-operators-68hnm\" (UID: \"56990b82-7491-4f38-b9c8-ecc3fc0db284\") " pod="openshift-marketplace/certified-operators-68hnm" Oct 14 07:58:12 crc kubenswrapper[4870]: I1014 07:58:12.343621 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-68hnm" Oct 14 07:58:12 crc kubenswrapper[4870]: I1014 07:58:12.793165 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-68hnm"] Oct 14 07:58:12 crc kubenswrapper[4870]: I1014 07:58:12.931358 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-68hnm" event={"ID":"56990b82-7491-4f38-b9c8-ecc3fc0db284","Type":"ContainerStarted","Data":"3686e1df7cefb04cd9a7034b5e8548f0674950af5f95c54f35c2f5c182e3a6eb"} Oct 14 07:58:13 crc kubenswrapper[4870]: I1014 07:58:13.945670 4870 generic.go:334] "Generic (PLEG): container finished" podID="56990b82-7491-4f38-b9c8-ecc3fc0db284" containerID="e35d5aace8a24b5f3e25945f36433526aec4e96b4502a759c6ce91707e3a2d47" exitCode=0 Oct 14 07:58:13 crc kubenswrapper[4870]: I1014 07:58:13.945755 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-68hnm" event={"ID":"56990b82-7491-4f38-b9c8-ecc3fc0db284","Type":"ContainerDied","Data":"e35d5aace8a24b5f3e25945f36433526aec4e96b4502a759c6ce91707e3a2d47"} Oct 14 07:58:14 crc kubenswrapper[4870]: I1014 07:58:14.955407 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-68hnm" event={"ID":"56990b82-7491-4f38-b9c8-ecc3fc0db284","Type":"ContainerStarted","Data":"bb7106f5e5627b59abfbb1799fe2b06a55dd92f63ceb98026c15886edc73b769"} Oct 14 07:58:15 crc kubenswrapper[4870]: I1014 07:58:15.967077 4870 generic.go:334] "Generic (PLEG): container finished" podID="56990b82-7491-4f38-b9c8-ecc3fc0db284" containerID="bb7106f5e5627b59abfbb1799fe2b06a55dd92f63ceb98026c15886edc73b769" exitCode=0 Oct 14 07:58:15 crc kubenswrapper[4870]: I1014 07:58:15.967187 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-68hnm" event={"ID":"56990b82-7491-4f38-b9c8-ecc3fc0db284","Type":"ContainerDied","Data":"bb7106f5e5627b59abfbb1799fe2b06a55dd92f63ceb98026c15886edc73b769"} Oct 14 07:58:16 crc kubenswrapper[4870]: I1014 07:58:16.978236 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-68hnm" event={"ID":"56990b82-7491-4f38-b9c8-ecc3fc0db284","Type":"ContainerStarted","Data":"e79e17d967d7faea31b37beff44ff3111eb91b2d1a6fd1b11250e09807a9b7d3"} Oct 14 07:58:16 crc kubenswrapper[4870]: I1014 07:58:16.997191 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-68hnm" podStartSLOduration=3.42584819 podStartE2EDuration="5.997171393s" podCreationTimestamp="2025-10-14 07:58:11 +0000 UTC" firstStartedPulling="2025-10-14 07:58:13.947985597 +0000 UTC m=+3429.645345998" lastFinishedPulling="2025-10-14 07:58:16.51930883 +0000 UTC m=+3432.216669201" observedRunningTime="2025-10-14 07:58:16.994336403 +0000 UTC m=+3432.691696784" watchObservedRunningTime="2025-10-14 07:58:16.997171393 +0000 UTC m=+3432.694531774" Oct 14 07:58:20 crc kubenswrapper[4870]: I1014 07:58:20.034516 4870 scope.go:117] "RemoveContainer" containerID="d9859555f9dac6de3438df8cb32e7aedced80b131b2e63a298d227e4c67cd16e" Oct 14 07:58:20 crc kubenswrapper[4870]: E1014 07:58:20.035394 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 07:58:22 crc kubenswrapper[4870]: I1014 07:58:22.344823 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-68hnm" Oct 14 07:58:22 crc kubenswrapper[4870]: I1014 07:58:22.344924 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-68hnm" Oct 14 07:58:22 crc kubenswrapper[4870]: I1014 07:58:22.425496 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-68hnm" Oct 14 07:58:23 crc kubenswrapper[4870]: I1014 07:58:23.116380 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-68hnm" Oct 14 07:58:23 crc kubenswrapper[4870]: I1014 07:58:23.185365 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-68hnm"] Oct 14 07:58:25 crc kubenswrapper[4870]: I1014 07:58:25.059951 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-68hnm" podUID="56990b82-7491-4f38-b9c8-ecc3fc0db284" containerName="registry-server" containerID="cri-o://e79e17d967d7faea31b37beff44ff3111eb91b2d1a6fd1b11250e09807a9b7d3" gracePeriod=2 Oct 14 07:58:25 crc kubenswrapper[4870]: E1014 07:58:25.225323 4870 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod56990b82_7491_4f38_b9c8_ecc3fc0db284.slice/crio-e79e17d967d7faea31b37beff44ff3111eb91b2d1a6fd1b11250e09807a9b7d3.scope\": RecentStats: unable to find data in memory cache]" Oct 14 07:58:26 crc kubenswrapper[4870]: I1014 07:58:26.068390 4870 generic.go:334] "Generic (PLEG): container finished" podID="56990b82-7491-4f38-b9c8-ecc3fc0db284" containerID="e79e17d967d7faea31b37beff44ff3111eb91b2d1a6fd1b11250e09807a9b7d3" exitCode=0 Oct 14 07:58:26 crc kubenswrapper[4870]: I1014 07:58:26.068429 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-68hnm" event={"ID":"56990b82-7491-4f38-b9c8-ecc3fc0db284","Type":"ContainerDied","Data":"e79e17d967d7faea31b37beff44ff3111eb91b2d1a6fd1b11250e09807a9b7d3"} Oct 14 07:58:26 crc kubenswrapper[4870]: I1014 07:58:26.068470 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-68hnm" event={"ID":"56990b82-7491-4f38-b9c8-ecc3fc0db284","Type":"ContainerDied","Data":"3686e1df7cefb04cd9a7034b5e8548f0674950af5f95c54f35c2f5c182e3a6eb"} Oct 14 07:58:26 crc kubenswrapper[4870]: I1014 07:58:26.068484 4870 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3686e1df7cefb04cd9a7034b5e8548f0674950af5f95c54f35c2f5c182e3a6eb" Oct 14 07:58:26 crc kubenswrapper[4870]: I1014 07:58:26.074078 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-68hnm" Oct 14 07:58:26 crc kubenswrapper[4870]: I1014 07:58:26.090109 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mf2kz\" (UniqueName: \"kubernetes.io/projected/56990b82-7491-4f38-b9c8-ecc3fc0db284-kube-api-access-mf2kz\") pod \"56990b82-7491-4f38-b9c8-ecc3fc0db284\" (UID: \"56990b82-7491-4f38-b9c8-ecc3fc0db284\") " Oct 14 07:58:26 crc kubenswrapper[4870]: I1014 07:58:26.090203 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/56990b82-7491-4f38-b9c8-ecc3fc0db284-catalog-content\") pod \"56990b82-7491-4f38-b9c8-ecc3fc0db284\" (UID: \"56990b82-7491-4f38-b9c8-ecc3fc0db284\") " Oct 14 07:58:26 crc kubenswrapper[4870]: I1014 07:58:26.090256 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/56990b82-7491-4f38-b9c8-ecc3fc0db284-utilities\") pod \"56990b82-7491-4f38-b9c8-ecc3fc0db284\" (UID: \"56990b82-7491-4f38-b9c8-ecc3fc0db284\") " Oct 14 07:58:26 crc kubenswrapper[4870]: I1014 07:58:26.091751 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/56990b82-7491-4f38-b9c8-ecc3fc0db284-utilities" (OuterVolumeSpecName: "utilities") pod "56990b82-7491-4f38-b9c8-ecc3fc0db284" (UID: "56990b82-7491-4f38-b9c8-ecc3fc0db284"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:58:26 crc kubenswrapper[4870]: I1014 07:58:26.106235 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/56990b82-7491-4f38-b9c8-ecc3fc0db284-kube-api-access-mf2kz" (OuterVolumeSpecName: "kube-api-access-mf2kz") pod "56990b82-7491-4f38-b9c8-ecc3fc0db284" (UID: "56990b82-7491-4f38-b9c8-ecc3fc0db284"). InnerVolumeSpecName "kube-api-access-mf2kz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 07:58:26 crc kubenswrapper[4870]: I1014 07:58:26.151183 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/56990b82-7491-4f38-b9c8-ecc3fc0db284-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "56990b82-7491-4f38-b9c8-ecc3fc0db284" (UID: "56990b82-7491-4f38-b9c8-ecc3fc0db284"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 07:58:26 crc kubenswrapper[4870]: I1014 07:58:26.190896 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mf2kz\" (UniqueName: \"kubernetes.io/projected/56990b82-7491-4f38-b9c8-ecc3fc0db284-kube-api-access-mf2kz\") on node \"crc\" DevicePath \"\"" Oct 14 07:58:26 crc kubenswrapper[4870]: I1014 07:58:26.190918 4870 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/56990b82-7491-4f38-b9c8-ecc3fc0db284-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 07:58:26 crc kubenswrapper[4870]: I1014 07:58:26.190927 4870 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/56990b82-7491-4f38-b9c8-ecc3fc0db284-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 07:58:27 crc kubenswrapper[4870]: I1014 07:58:27.076564 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-68hnm" Oct 14 07:58:27 crc kubenswrapper[4870]: I1014 07:58:27.127913 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-68hnm"] Oct 14 07:58:27 crc kubenswrapper[4870]: I1014 07:58:27.135684 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-68hnm"] Oct 14 07:58:29 crc kubenswrapper[4870]: I1014 07:58:29.055035 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="56990b82-7491-4f38-b9c8-ecc3fc0db284" path="/var/lib/kubelet/pods/56990b82-7491-4f38-b9c8-ecc3fc0db284/volumes" Oct 14 07:58:31 crc kubenswrapper[4870]: I1014 07:58:31.034417 4870 scope.go:117] "RemoveContainer" containerID="d9859555f9dac6de3438df8cb32e7aedced80b131b2e63a298d227e4c67cd16e" Oct 14 07:58:32 crc kubenswrapper[4870]: I1014 07:58:32.126226 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" event={"ID":"874a5e45-dffd-4d17-b609-b7d3ed2eab07","Type":"ContainerStarted","Data":"a73a10b6ed5339f2f4f7b402b25425ae167a29a2c96f75901310db50025f68fd"} Oct 14 08:00:00 crc kubenswrapper[4870]: I1014 08:00:00.217041 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29340480-hp2hq"] Oct 14 08:00:00 crc kubenswrapper[4870]: E1014 08:00:00.218143 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56990b82-7491-4f38-b9c8-ecc3fc0db284" containerName="registry-server" Oct 14 08:00:00 crc kubenswrapper[4870]: I1014 08:00:00.218155 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="56990b82-7491-4f38-b9c8-ecc3fc0db284" containerName="registry-server" Oct 14 08:00:00 crc kubenswrapper[4870]: E1014 08:00:00.218181 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56990b82-7491-4f38-b9c8-ecc3fc0db284" containerName="extract-utilities" Oct 14 08:00:00 crc kubenswrapper[4870]: I1014 08:00:00.218188 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="56990b82-7491-4f38-b9c8-ecc3fc0db284" containerName="extract-utilities" Oct 14 08:00:00 crc kubenswrapper[4870]: E1014 08:00:00.218198 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56990b82-7491-4f38-b9c8-ecc3fc0db284" containerName="extract-content" Oct 14 08:00:00 crc kubenswrapper[4870]: I1014 08:00:00.218206 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="56990b82-7491-4f38-b9c8-ecc3fc0db284" containerName="extract-content" Oct 14 08:00:00 crc kubenswrapper[4870]: I1014 08:00:00.218355 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="56990b82-7491-4f38-b9c8-ecc3fc0db284" containerName="registry-server" Oct 14 08:00:00 crc kubenswrapper[4870]: I1014 08:00:00.218850 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29340480-hp2hq" Oct 14 08:00:00 crc kubenswrapper[4870]: I1014 08:00:00.220708 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 14 08:00:00 crc kubenswrapper[4870]: I1014 08:00:00.222270 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 14 08:00:00 crc kubenswrapper[4870]: I1014 08:00:00.232726 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29340480-hp2hq"] Oct 14 08:00:00 crc kubenswrapper[4870]: I1014 08:00:00.301534 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sn5dz\" (UniqueName: \"kubernetes.io/projected/db935c6f-1060-4dfa-a650-d0e894ad8a61-kube-api-access-sn5dz\") pod \"collect-profiles-29340480-hp2hq\" (UID: \"db935c6f-1060-4dfa-a650-d0e894ad8a61\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340480-hp2hq" Oct 14 08:00:00 crc kubenswrapper[4870]: I1014 08:00:00.301586 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/db935c6f-1060-4dfa-a650-d0e894ad8a61-secret-volume\") pod \"collect-profiles-29340480-hp2hq\" (UID: \"db935c6f-1060-4dfa-a650-d0e894ad8a61\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340480-hp2hq" Oct 14 08:00:00 crc kubenswrapper[4870]: I1014 08:00:00.301673 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/db935c6f-1060-4dfa-a650-d0e894ad8a61-config-volume\") pod \"collect-profiles-29340480-hp2hq\" (UID: \"db935c6f-1060-4dfa-a650-d0e894ad8a61\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340480-hp2hq" Oct 14 08:00:00 crc kubenswrapper[4870]: I1014 08:00:00.403332 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/db935c6f-1060-4dfa-a650-d0e894ad8a61-config-volume\") pod \"collect-profiles-29340480-hp2hq\" (UID: \"db935c6f-1060-4dfa-a650-d0e894ad8a61\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340480-hp2hq" Oct 14 08:00:00 crc kubenswrapper[4870]: I1014 08:00:00.403475 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sn5dz\" (UniqueName: \"kubernetes.io/projected/db935c6f-1060-4dfa-a650-d0e894ad8a61-kube-api-access-sn5dz\") pod \"collect-profiles-29340480-hp2hq\" (UID: \"db935c6f-1060-4dfa-a650-d0e894ad8a61\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340480-hp2hq" Oct 14 08:00:00 crc kubenswrapper[4870]: I1014 08:00:00.403537 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/db935c6f-1060-4dfa-a650-d0e894ad8a61-secret-volume\") pod \"collect-profiles-29340480-hp2hq\" (UID: \"db935c6f-1060-4dfa-a650-d0e894ad8a61\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340480-hp2hq" Oct 14 08:00:00 crc kubenswrapper[4870]: I1014 08:00:00.405100 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/db935c6f-1060-4dfa-a650-d0e894ad8a61-config-volume\") pod \"collect-profiles-29340480-hp2hq\" (UID: \"db935c6f-1060-4dfa-a650-d0e894ad8a61\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340480-hp2hq" Oct 14 08:00:00 crc kubenswrapper[4870]: I1014 08:00:00.413911 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/db935c6f-1060-4dfa-a650-d0e894ad8a61-secret-volume\") pod \"collect-profiles-29340480-hp2hq\" (UID: \"db935c6f-1060-4dfa-a650-d0e894ad8a61\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340480-hp2hq" Oct 14 08:00:00 crc kubenswrapper[4870]: I1014 08:00:00.424113 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sn5dz\" (UniqueName: \"kubernetes.io/projected/db935c6f-1060-4dfa-a650-d0e894ad8a61-kube-api-access-sn5dz\") pod \"collect-profiles-29340480-hp2hq\" (UID: \"db935c6f-1060-4dfa-a650-d0e894ad8a61\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340480-hp2hq" Oct 14 08:00:00 crc kubenswrapper[4870]: I1014 08:00:00.554982 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29340480-hp2hq" Oct 14 08:00:01 crc kubenswrapper[4870]: I1014 08:00:01.010355 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29340480-hp2hq"] Oct 14 08:00:02 crc kubenswrapper[4870]: I1014 08:00:02.013521 4870 generic.go:334] "Generic (PLEG): container finished" podID="db935c6f-1060-4dfa-a650-d0e894ad8a61" containerID="747359111df3b191bff560e8b74280264f374cc938e54fbee17b306a8cb4bb9e" exitCode=0 Oct 14 08:00:02 crc kubenswrapper[4870]: I1014 08:00:02.013660 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29340480-hp2hq" event={"ID":"db935c6f-1060-4dfa-a650-d0e894ad8a61","Type":"ContainerDied","Data":"747359111df3b191bff560e8b74280264f374cc938e54fbee17b306a8cb4bb9e"} Oct 14 08:00:02 crc kubenswrapper[4870]: I1014 08:00:02.014049 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29340480-hp2hq" event={"ID":"db935c6f-1060-4dfa-a650-d0e894ad8a61","Type":"ContainerStarted","Data":"37e8c2bebab9c5322a3382f28963583b6871f7c74fd35747e93f64db4d06e533"} Oct 14 08:00:03 crc kubenswrapper[4870]: I1014 08:00:03.431776 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29340480-hp2hq" Oct 14 08:00:03 crc kubenswrapper[4870]: I1014 08:00:03.553666 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/db935c6f-1060-4dfa-a650-d0e894ad8a61-secret-volume\") pod \"db935c6f-1060-4dfa-a650-d0e894ad8a61\" (UID: \"db935c6f-1060-4dfa-a650-d0e894ad8a61\") " Oct 14 08:00:03 crc kubenswrapper[4870]: I1014 08:00:03.553743 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sn5dz\" (UniqueName: \"kubernetes.io/projected/db935c6f-1060-4dfa-a650-d0e894ad8a61-kube-api-access-sn5dz\") pod \"db935c6f-1060-4dfa-a650-d0e894ad8a61\" (UID: \"db935c6f-1060-4dfa-a650-d0e894ad8a61\") " Oct 14 08:00:03 crc kubenswrapper[4870]: I1014 08:00:03.553778 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/db935c6f-1060-4dfa-a650-d0e894ad8a61-config-volume\") pod \"db935c6f-1060-4dfa-a650-d0e894ad8a61\" (UID: \"db935c6f-1060-4dfa-a650-d0e894ad8a61\") " Oct 14 08:00:03 crc kubenswrapper[4870]: I1014 08:00:03.554763 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/db935c6f-1060-4dfa-a650-d0e894ad8a61-config-volume" (OuterVolumeSpecName: "config-volume") pod "db935c6f-1060-4dfa-a650-d0e894ad8a61" (UID: "db935c6f-1060-4dfa-a650-d0e894ad8a61"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:00:03 crc kubenswrapper[4870]: I1014 08:00:03.563763 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/db935c6f-1060-4dfa-a650-d0e894ad8a61-kube-api-access-sn5dz" (OuterVolumeSpecName: "kube-api-access-sn5dz") pod "db935c6f-1060-4dfa-a650-d0e894ad8a61" (UID: "db935c6f-1060-4dfa-a650-d0e894ad8a61"). InnerVolumeSpecName "kube-api-access-sn5dz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:00:03 crc kubenswrapper[4870]: I1014 08:00:03.564196 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/db935c6f-1060-4dfa-a650-d0e894ad8a61-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "db935c6f-1060-4dfa-a650-d0e894ad8a61" (UID: "db935c6f-1060-4dfa-a650-d0e894ad8a61"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:00:03 crc kubenswrapper[4870]: I1014 08:00:03.655429 4870 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/db935c6f-1060-4dfa-a650-d0e894ad8a61-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 14 08:00:03 crc kubenswrapper[4870]: I1014 08:00:03.655494 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sn5dz\" (UniqueName: \"kubernetes.io/projected/db935c6f-1060-4dfa-a650-d0e894ad8a61-kube-api-access-sn5dz\") on node \"crc\" DevicePath \"\"" Oct 14 08:00:03 crc kubenswrapper[4870]: I1014 08:00:03.655513 4870 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/db935c6f-1060-4dfa-a650-d0e894ad8a61-config-volume\") on node \"crc\" DevicePath \"\"" Oct 14 08:00:04 crc kubenswrapper[4870]: I1014 08:00:04.038160 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29340480-hp2hq" event={"ID":"db935c6f-1060-4dfa-a650-d0e894ad8a61","Type":"ContainerDied","Data":"37e8c2bebab9c5322a3382f28963583b6871f7c74fd35747e93f64db4d06e533"} Oct 14 08:00:04 crc kubenswrapper[4870]: I1014 08:00:04.038404 4870 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="37e8c2bebab9c5322a3382f28963583b6871f7c74fd35747e93f64db4d06e533" Oct 14 08:00:04 crc kubenswrapper[4870]: I1014 08:00:04.038257 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29340480-hp2hq" Oct 14 08:00:04 crc kubenswrapper[4870]: I1014 08:00:04.529535 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29340435-g6qxb"] Oct 14 08:00:04 crc kubenswrapper[4870]: I1014 08:00:04.539690 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29340435-g6qxb"] Oct 14 08:00:05 crc kubenswrapper[4870]: I1014 08:00:05.049426 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="89fe6660-6f16-4ddb-9b5c-1582231f5f30" path="/var/lib/kubelet/pods/89fe6660-6f16-4ddb-9b5c-1582231f5f30/volumes" Oct 14 08:00:12 crc kubenswrapper[4870]: I1014 08:00:12.874418 4870 scope.go:117] "RemoveContainer" containerID="2d804bb157221ca38727691cccd928f8917bce656da1d8d959887b79b3c20d34" Oct 14 08:00:53 crc kubenswrapper[4870]: I1014 08:00:53.951153 4870 patch_prober.go:28] interesting pod/machine-config-daemon-7tvc7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 08:00:53 crc kubenswrapper[4870]: I1014 08:00:53.952020 4870 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 08:01:23 crc kubenswrapper[4870]: I1014 08:01:23.951160 4870 patch_prober.go:28] interesting pod/machine-config-daemon-7tvc7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 08:01:23 crc kubenswrapper[4870]: I1014 08:01:23.952074 4870 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 08:01:53 crc kubenswrapper[4870]: I1014 08:01:53.951528 4870 patch_prober.go:28] interesting pod/machine-config-daemon-7tvc7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 08:01:53 crc kubenswrapper[4870]: I1014 08:01:53.952307 4870 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 08:01:53 crc kubenswrapper[4870]: I1014 08:01:53.952386 4870 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" Oct 14 08:01:53 crc kubenswrapper[4870]: I1014 08:01:53.953215 4870 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"a73a10b6ed5339f2f4f7b402b25425ae167a29a2c96f75901310db50025f68fd"} pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 14 08:01:53 crc kubenswrapper[4870]: I1014 08:01:53.953311 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerName="machine-config-daemon" containerID="cri-o://a73a10b6ed5339f2f4f7b402b25425ae167a29a2c96f75901310db50025f68fd" gracePeriod=600 Oct 14 08:01:54 crc kubenswrapper[4870]: I1014 08:01:54.127969 4870 generic.go:334] "Generic (PLEG): container finished" podID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerID="a73a10b6ed5339f2f4f7b402b25425ae167a29a2c96f75901310db50025f68fd" exitCode=0 Oct 14 08:01:54 crc kubenswrapper[4870]: I1014 08:01:54.128100 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" event={"ID":"874a5e45-dffd-4d17-b609-b7d3ed2eab07","Type":"ContainerDied","Data":"a73a10b6ed5339f2f4f7b402b25425ae167a29a2c96f75901310db50025f68fd"} Oct 14 08:01:54 crc kubenswrapper[4870]: I1014 08:01:54.128493 4870 scope.go:117] "RemoveContainer" containerID="d9859555f9dac6de3438df8cb32e7aedced80b131b2e63a298d227e4c67cd16e" Oct 14 08:01:55 crc kubenswrapper[4870]: I1014 08:01:55.142618 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" event={"ID":"874a5e45-dffd-4d17-b609-b7d3ed2eab07","Type":"ContainerStarted","Data":"dca50334a8cd2d1438da606d2f43eb3a283c73f0855c6f8bb11c380e552e3268"} Oct 14 08:03:54 crc kubenswrapper[4870]: I1014 08:03:54.547982 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-gqf44"] Oct 14 08:03:54 crc kubenswrapper[4870]: E1014 08:03:54.549156 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db935c6f-1060-4dfa-a650-d0e894ad8a61" containerName="collect-profiles" Oct 14 08:03:54 crc kubenswrapper[4870]: I1014 08:03:54.549176 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="db935c6f-1060-4dfa-a650-d0e894ad8a61" containerName="collect-profiles" Oct 14 08:03:54 crc kubenswrapper[4870]: I1014 08:03:54.549393 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="db935c6f-1060-4dfa-a650-d0e894ad8a61" containerName="collect-profiles" Oct 14 08:03:54 crc kubenswrapper[4870]: I1014 08:03:54.550861 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gqf44" Oct 14 08:03:54 crc kubenswrapper[4870]: I1014 08:03:54.567179 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-gqf44"] Oct 14 08:03:54 crc kubenswrapper[4870]: I1014 08:03:54.674645 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7c686764-9b09-4725-9d67-55e22adec2c4-utilities\") pod \"redhat-operators-gqf44\" (UID: \"7c686764-9b09-4725-9d67-55e22adec2c4\") " pod="openshift-marketplace/redhat-operators-gqf44" Oct 14 08:03:54 crc kubenswrapper[4870]: I1014 08:03:54.674918 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7c686764-9b09-4725-9d67-55e22adec2c4-catalog-content\") pod \"redhat-operators-gqf44\" (UID: \"7c686764-9b09-4725-9d67-55e22adec2c4\") " pod="openshift-marketplace/redhat-operators-gqf44" Oct 14 08:03:54 crc kubenswrapper[4870]: I1014 08:03:54.675349 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kxfgj\" (UniqueName: \"kubernetes.io/projected/7c686764-9b09-4725-9d67-55e22adec2c4-kube-api-access-kxfgj\") pod \"redhat-operators-gqf44\" (UID: \"7c686764-9b09-4725-9d67-55e22adec2c4\") " pod="openshift-marketplace/redhat-operators-gqf44" Oct 14 08:03:54 crc kubenswrapper[4870]: I1014 08:03:54.778520 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kxfgj\" (UniqueName: \"kubernetes.io/projected/7c686764-9b09-4725-9d67-55e22adec2c4-kube-api-access-kxfgj\") pod \"redhat-operators-gqf44\" (UID: \"7c686764-9b09-4725-9d67-55e22adec2c4\") " pod="openshift-marketplace/redhat-operators-gqf44" Oct 14 08:03:54 crc kubenswrapper[4870]: I1014 08:03:54.778697 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7c686764-9b09-4725-9d67-55e22adec2c4-utilities\") pod \"redhat-operators-gqf44\" (UID: \"7c686764-9b09-4725-9d67-55e22adec2c4\") " pod="openshift-marketplace/redhat-operators-gqf44" Oct 14 08:03:54 crc kubenswrapper[4870]: I1014 08:03:54.778889 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7c686764-9b09-4725-9d67-55e22adec2c4-catalog-content\") pod \"redhat-operators-gqf44\" (UID: \"7c686764-9b09-4725-9d67-55e22adec2c4\") " pod="openshift-marketplace/redhat-operators-gqf44" Oct 14 08:03:54 crc kubenswrapper[4870]: I1014 08:03:54.779689 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7c686764-9b09-4725-9d67-55e22adec2c4-utilities\") pod \"redhat-operators-gqf44\" (UID: \"7c686764-9b09-4725-9d67-55e22adec2c4\") " pod="openshift-marketplace/redhat-operators-gqf44" Oct 14 08:03:54 crc kubenswrapper[4870]: I1014 08:03:54.779922 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7c686764-9b09-4725-9d67-55e22adec2c4-catalog-content\") pod \"redhat-operators-gqf44\" (UID: \"7c686764-9b09-4725-9d67-55e22adec2c4\") " pod="openshift-marketplace/redhat-operators-gqf44" Oct 14 08:03:54 crc kubenswrapper[4870]: I1014 08:03:54.812021 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kxfgj\" (UniqueName: \"kubernetes.io/projected/7c686764-9b09-4725-9d67-55e22adec2c4-kube-api-access-kxfgj\") pod \"redhat-operators-gqf44\" (UID: \"7c686764-9b09-4725-9d67-55e22adec2c4\") " pod="openshift-marketplace/redhat-operators-gqf44" Oct 14 08:03:54 crc kubenswrapper[4870]: I1014 08:03:54.874743 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gqf44" Oct 14 08:03:55 crc kubenswrapper[4870]: I1014 08:03:55.383331 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-gqf44"] Oct 14 08:03:56 crc kubenswrapper[4870]: I1014 08:03:56.385226 4870 generic.go:334] "Generic (PLEG): container finished" podID="7c686764-9b09-4725-9d67-55e22adec2c4" containerID="c92c36f1a614b15ac097143fa90a88aec88420b482eb09b3550c8f04a5fb53ce" exitCode=0 Oct 14 08:03:56 crc kubenswrapper[4870]: I1014 08:03:56.385335 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gqf44" event={"ID":"7c686764-9b09-4725-9d67-55e22adec2c4","Type":"ContainerDied","Data":"c92c36f1a614b15ac097143fa90a88aec88420b482eb09b3550c8f04a5fb53ce"} Oct 14 08:03:56 crc kubenswrapper[4870]: I1014 08:03:56.385714 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gqf44" event={"ID":"7c686764-9b09-4725-9d67-55e22adec2c4","Type":"ContainerStarted","Data":"3835259acc5dde7e5e7533250f8b4122aee82d7152b3654e09487cfce8e09654"} Oct 14 08:03:56 crc kubenswrapper[4870]: I1014 08:03:56.390300 4870 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 14 08:03:57 crc kubenswrapper[4870]: I1014 08:03:57.400338 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gqf44" event={"ID":"7c686764-9b09-4725-9d67-55e22adec2c4","Type":"ContainerStarted","Data":"947b7bf8645365613f89de5d5027694e3d1f8ec0457a3548d9efa408d5c82c01"} Oct 14 08:03:58 crc kubenswrapper[4870]: I1014 08:03:58.412630 4870 generic.go:334] "Generic (PLEG): container finished" podID="7c686764-9b09-4725-9d67-55e22adec2c4" containerID="947b7bf8645365613f89de5d5027694e3d1f8ec0457a3548d9efa408d5c82c01" exitCode=0 Oct 14 08:03:58 crc kubenswrapper[4870]: I1014 08:03:58.412686 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gqf44" event={"ID":"7c686764-9b09-4725-9d67-55e22adec2c4","Type":"ContainerDied","Data":"947b7bf8645365613f89de5d5027694e3d1f8ec0457a3548d9efa408d5c82c01"} Oct 14 08:03:59 crc kubenswrapper[4870]: I1014 08:03:59.426530 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gqf44" event={"ID":"7c686764-9b09-4725-9d67-55e22adec2c4","Type":"ContainerStarted","Data":"9551f936a3b74bb0566c78242f37181088fae400bdba1beeeb5980aaa423a9cb"} Oct 14 08:03:59 crc kubenswrapper[4870]: I1014 08:03:59.456851 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-gqf44" podStartSLOduration=3.006667765 podStartE2EDuration="5.456823402s" podCreationTimestamp="2025-10-14 08:03:54 +0000 UTC" firstStartedPulling="2025-10-14 08:03:56.389719298 +0000 UTC m=+3772.087079699" lastFinishedPulling="2025-10-14 08:03:58.839874935 +0000 UTC m=+3774.537235336" observedRunningTime="2025-10-14 08:03:59.451400749 +0000 UTC m=+3775.148761160" watchObservedRunningTime="2025-10-14 08:03:59.456823402 +0000 UTC m=+3775.154183803" Oct 14 08:04:04 crc kubenswrapper[4870]: I1014 08:04:04.875557 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-gqf44" Oct 14 08:04:04 crc kubenswrapper[4870]: I1014 08:04:04.876417 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-gqf44" Oct 14 08:04:04 crc kubenswrapper[4870]: I1014 08:04:04.934647 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-gqf44" Oct 14 08:04:05 crc kubenswrapper[4870]: I1014 08:04:05.567687 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-gqf44" Oct 14 08:04:06 crc kubenswrapper[4870]: I1014 08:04:06.127826 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-gqf44"] Oct 14 08:04:07 crc kubenswrapper[4870]: I1014 08:04:07.520265 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-gqf44" podUID="7c686764-9b09-4725-9d67-55e22adec2c4" containerName="registry-server" containerID="cri-o://9551f936a3b74bb0566c78242f37181088fae400bdba1beeeb5980aaa423a9cb" gracePeriod=2 Oct 14 08:04:09 crc kubenswrapper[4870]: I1014 08:04:09.113004 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gqf44" Oct 14 08:04:09 crc kubenswrapper[4870]: I1014 08:04:09.242615 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kxfgj\" (UniqueName: \"kubernetes.io/projected/7c686764-9b09-4725-9d67-55e22adec2c4-kube-api-access-kxfgj\") pod \"7c686764-9b09-4725-9d67-55e22adec2c4\" (UID: \"7c686764-9b09-4725-9d67-55e22adec2c4\") " Oct 14 08:04:09 crc kubenswrapper[4870]: I1014 08:04:09.242691 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7c686764-9b09-4725-9d67-55e22adec2c4-utilities\") pod \"7c686764-9b09-4725-9d67-55e22adec2c4\" (UID: \"7c686764-9b09-4725-9d67-55e22adec2c4\") " Oct 14 08:04:09 crc kubenswrapper[4870]: I1014 08:04:09.242799 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7c686764-9b09-4725-9d67-55e22adec2c4-catalog-content\") pod \"7c686764-9b09-4725-9d67-55e22adec2c4\" (UID: \"7c686764-9b09-4725-9d67-55e22adec2c4\") " Oct 14 08:04:09 crc kubenswrapper[4870]: I1014 08:04:09.243813 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7c686764-9b09-4725-9d67-55e22adec2c4-utilities" (OuterVolumeSpecName: "utilities") pod "7c686764-9b09-4725-9d67-55e22adec2c4" (UID: "7c686764-9b09-4725-9d67-55e22adec2c4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:04:09 crc kubenswrapper[4870]: I1014 08:04:09.251358 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7c686764-9b09-4725-9d67-55e22adec2c4-kube-api-access-kxfgj" (OuterVolumeSpecName: "kube-api-access-kxfgj") pod "7c686764-9b09-4725-9d67-55e22adec2c4" (UID: "7c686764-9b09-4725-9d67-55e22adec2c4"). InnerVolumeSpecName "kube-api-access-kxfgj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:04:09 crc kubenswrapper[4870]: I1014 08:04:09.337660 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7c686764-9b09-4725-9d67-55e22adec2c4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7c686764-9b09-4725-9d67-55e22adec2c4" (UID: "7c686764-9b09-4725-9d67-55e22adec2c4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:04:09 crc kubenswrapper[4870]: I1014 08:04:09.344076 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kxfgj\" (UniqueName: \"kubernetes.io/projected/7c686764-9b09-4725-9d67-55e22adec2c4-kube-api-access-kxfgj\") on node \"crc\" DevicePath \"\"" Oct 14 08:04:09 crc kubenswrapper[4870]: I1014 08:04:09.344096 4870 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7c686764-9b09-4725-9d67-55e22adec2c4-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 08:04:09 crc kubenswrapper[4870]: I1014 08:04:09.344107 4870 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7c686764-9b09-4725-9d67-55e22adec2c4-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 08:04:09 crc kubenswrapper[4870]: I1014 08:04:09.543688 4870 generic.go:334] "Generic (PLEG): container finished" podID="7c686764-9b09-4725-9d67-55e22adec2c4" containerID="9551f936a3b74bb0566c78242f37181088fae400bdba1beeeb5980aaa423a9cb" exitCode=0 Oct 14 08:04:09 crc kubenswrapper[4870]: I1014 08:04:09.543767 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gqf44" event={"ID":"7c686764-9b09-4725-9d67-55e22adec2c4","Type":"ContainerDied","Data":"9551f936a3b74bb0566c78242f37181088fae400bdba1beeeb5980aaa423a9cb"} Oct 14 08:04:09 crc kubenswrapper[4870]: I1014 08:04:09.543837 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gqf44" event={"ID":"7c686764-9b09-4725-9d67-55e22adec2c4","Type":"ContainerDied","Data":"3835259acc5dde7e5e7533250f8b4122aee82d7152b3654e09487cfce8e09654"} Oct 14 08:04:09 crc kubenswrapper[4870]: I1014 08:04:09.543870 4870 scope.go:117] "RemoveContainer" containerID="9551f936a3b74bb0566c78242f37181088fae400bdba1beeeb5980aaa423a9cb" Oct 14 08:04:09 crc kubenswrapper[4870]: I1014 08:04:09.543788 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gqf44" Oct 14 08:04:09 crc kubenswrapper[4870]: I1014 08:04:09.570277 4870 scope.go:117] "RemoveContainer" containerID="947b7bf8645365613f89de5d5027694e3d1f8ec0457a3548d9efa408d5c82c01" Oct 14 08:04:09 crc kubenswrapper[4870]: I1014 08:04:09.597028 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-gqf44"] Oct 14 08:04:09 crc kubenswrapper[4870]: I1014 08:04:09.604117 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-gqf44"] Oct 14 08:04:09 crc kubenswrapper[4870]: I1014 08:04:09.608499 4870 scope.go:117] "RemoveContainer" containerID="c92c36f1a614b15ac097143fa90a88aec88420b482eb09b3550c8f04a5fb53ce" Oct 14 08:04:09 crc kubenswrapper[4870]: I1014 08:04:09.622268 4870 scope.go:117] "RemoveContainer" containerID="9551f936a3b74bb0566c78242f37181088fae400bdba1beeeb5980aaa423a9cb" Oct 14 08:04:09 crc kubenswrapper[4870]: E1014 08:04:09.622561 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9551f936a3b74bb0566c78242f37181088fae400bdba1beeeb5980aaa423a9cb\": container with ID starting with 9551f936a3b74bb0566c78242f37181088fae400bdba1beeeb5980aaa423a9cb not found: ID does not exist" containerID="9551f936a3b74bb0566c78242f37181088fae400bdba1beeeb5980aaa423a9cb" Oct 14 08:04:09 crc kubenswrapper[4870]: I1014 08:04:09.622610 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9551f936a3b74bb0566c78242f37181088fae400bdba1beeeb5980aaa423a9cb"} err="failed to get container status \"9551f936a3b74bb0566c78242f37181088fae400bdba1beeeb5980aaa423a9cb\": rpc error: code = NotFound desc = could not find container \"9551f936a3b74bb0566c78242f37181088fae400bdba1beeeb5980aaa423a9cb\": container with ID starting with 9551f936a3b74bb0566c78242f37181088fae400bdba1beeeb5980aaa423a9cb not found: ID does not exist" Oct 14 08:04:09 crc kubenswrapper[4870]: I1014 08:04:09.622638 4870 scope.go:117] "RemoveContainer" containerID="947b7bf8645365613f89de5d5027694e3d1f8ec0457a3548d9efa408d5c82c01" Oct 14 08:04:09 crc kubenswrapper[4870]: E1014 08:04:09.622833 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"947b7bf8645365613f89de5d5027694e3d1f8ec0457a3548d9efa408d5c82c01\": container with ID starting with 947b7bf8645365613f89de5d5027694e3d1f8ec0457a3548d9efa408d5c82c01 not found: ID does not exist" containerID="947b7bf8645365613f89de5d5027694e3d1f8ec0457a3548d9efa408d5c82c01" Oct 14 08:04:09 crc kubenswrapper[4870]: I1014 08:04:09.622859 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"947b7bf8645365613f89de5d5027694e3d1f8ec0457a3548d9efa408d5c82c01"} err="failed to get container status \"947b7bf8645365613f89de5d5027694e3d1f8ec0457a3548d9efa408d5c82c01\": rpc error: code = NotFound desc = could not find container \"947b7bf8645365613f89de5d5027694e3d1f8ec0457a3548d9efa408d5c82c01\": container with ID starting with 947b7bf8645365613f89de5d5027694e3d1f8ec0457a3548d9efa408d5c82c01 not found: ID does not exist" Oct 14 08:04:09 crc kubenswrapper[4870]: I1014 08:04:09.622881 4870 scope.go:117] "RemoveContainer" containerID="c92c36f1a614b15ac097143fa90a88aec88420b482eb09b3550c8f04a5fb53ce" Oct 14 08:04:09 crc kubenswrapper[4870]: E1014 08:04:09.623188 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c92c36f1a614b15ac097143fa90a88aec88420b482eb09b3550c8f04a5fb53ce\": container with ID starting with c92c36f1a614b15ac097143fa90a88aec88420b482eb09b3550c8f04a5fb53ce not found: ID does not exist" containerID="c92c36f1a614b15ac097143fa90a88aec88420b482eb09b3550c8f04a5fb53ce" Oct 14 08:04:09 crc kubenswrapper[4870]: I1014 08:04:09.623215 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c92c36f1a614b15ac097143fa90a88aec88420b482eb09b3550c8f04a5fb53ce"} err="failed to get container status \"c92c36f1a614b15ac097143fa90a88aec88420b482eb09b3550c8f04a5fb53ce\": rpc error: code = NotFound desc = could not find container \"c92c36f1a614b15ac097143fa90a88aec88420b482eb09b3550c8f04a5fb53ce\": container with ID starting with c92c36f1a614b15ac097143fa90a88aec88420b482eb09b3550c8f04a5fb53ce not found: ID does not exist" Oct 14 08:04:11 crc kubenswrapper[4870]: I1014 08:04:11.046815 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7c686764-9b09-4725-9d67-55e22adec2c4" path="/var/lib/kubelet/pods/7c686764-9b09-4725-9d67-55e22adec2c4/volumes" Oct 14 08:04:13 crc kubenswrapper[4870]: I1014 08:04:13.021235 4870 scope.go:117] "RemoveContainer" containerID="e35d5aace8a24b5f3e25945f36433526aec4e96b4502a759c6ce91707e3a2d47" Oct 14 08:04:23 crc kubenswrapper[4870]: I1014 08:04:23.951622 4870 patch_prober.go:28] interesting pod/machine-config-daemon-7tvc7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 08:04:23 crc kubenswrapper[4870]: I1014 08:04:23.952393 4870 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 08:04:53 crc kubenswrapper[4870]: I1014 08:04:53.951368 4870 patch_prober.go:28] interesting pod/machine-config-daemon-7tvc7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 08:04:53 crc kubenswrapper[4870]: I1014 08:04:53.952494 4870 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 08:05:13 crc kubenswrapper[4870]: I1014 08:05:13.104068 4870 scope.go:117] "RemoveContainer" containerID="e79e17d967d7faea31b37beff44ff3111eb91b2d1a6fd1b11250e09807a9b7d3" Oct 14 08:05:13 crc kubenswrapper[4870]: I1014 08:05:13.144148 4870 scope.go:117] "RemoveContainer" containerID="bb7106f5e5627b59abfbb1799fe2b06a55dd92f63ceb98026c15886edc73b769" Oct 14 08:05:23 crc kubenswrapper[4870]: I1014 08:05:23.951384 4870 patch_prober.go:28] interesting pod/machine-config-daemon-7tvc7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 08:05:23 crc kubenswrapper[4870]: I1014 08:05:23.952129 4870 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 08:05:23 crc kubenswrapper[4870]: I1014 08:05:23.952209 4870 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" Oct 14 08:05:23 crc kubenswrapper[4870]: I1014 08:05:23.953194 4870 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"dca50334a8cd2d1438da606d2f43eb3a283c73f0855c6f8bb11c380e552e3268"} pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 14 08:05:23 crc kubenswrapper[4870]: I1014 08:05:23.953301 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerName="machine-config-daemon" containerID="cri-o://dca50334a8cd2d1438da606d2f43eb3a283c73f0855c6f8bb11c380e552e3268" gracePeriod=600 Oct 14 08:05:24 crc kubenswrapper[4870]: E1014 08:05:24.094631 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 08:05:24 crc kubenswrapper[4870]: I1014 08:05:24.281531 4870 generic.go:334] "Generic (PLEG): container finished" podID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerID="dca50334a8cd2d1438da606d2f43eb3a283c73f0855c6f8bb11c380e552e3268" exitCode=0 Oct 14 08:05:24 crc kubenswrapper[4870]: I1014 08:05:24.281609 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" event={"ID":"874a5e45-dffd-4d17-b609-b7d3ed2eab07","Type":"ContainerDied","Data":"dca50334a8cd2d1438da606d2f43eb3a283c73f0855c6f8bb11c380e552e3268"} Oct 14 08:05:24 crc kubenswrapper[4870]: I1014 08:05:24.281670 4870 scope.go:117] "RemoveContainer" containerID="a73a10b6ed5339f2f4f7b402b25425ae167a29a2c96f75901310db50025f68fd" Oct 14 08:05:24 crc kubenswrapper[4870]: I1014 08:05:24.282572 4870 scope.go:117] "RemoveContainer" containerID="dca50334a8cd2d1438da606d2f43eb3a283c73f0855c6f8bb11c380e552e3268" Oct 14 08:05:24 crc kubenswrapper[4870]: E1014 08:05:24.283040 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 08:05:35 crc kubenswrapper[4870]: I1014 08:05:35.042601 4870 scope.go:117] "RemoveContainer" containerID="dca50334a8cd2d1438da606d2f43eb3a283c73f0855c6f8bb11c380e552e3268" Oct 14 08:05:35 crc kubenswrapper[4870]: E1014 08:05:35.044400 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 08:05:46 crc kubenswrapper[4870]: I1014 08:05:46.034497 4870 scope.go:117] "RemoveContainer" containerID="dca50334a8cd2d1438da606d2f43eb3a283c73f0855c6f8bb11c380e552e3268" Oct 14 08:05:46 crc kubenswrapper[4870]: E1014 08:05:46.035860 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 08:05:58 crc kubenswrapper[4870]: I1014 08:05:58.373277 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-66rz4"] Oct 14 08:05:58 crc kubenswrapper[4870]: E1014 08:05:58.374750 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c686764-9b09-4725-9d67-55e22adec2c4" containerName="extract-utilities" Oct 14 08:05:58 crc kubenswrapper[4870]: I1014 08:05:58.374774 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c686764-9b09-4725-9d67-55e22adec2c4" containerName="extract-utilities" Oct 14 08:05:58 crc kubenswrapper[4870]: E1014 08:05:58.374795 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c686764-9b09-4725-9d67-55e22adec2c4" containerName="extract-content" Oct 14 08:05:58 crc kubenswrapper[4870]: I1014 08:05:58.374810 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c686764-9b09-4725-9d67-55e22adec2c4" containerName="extract-content" Oct 14 08:05:58 crc kubenswrapper[4870]: E1014 08:05:58.374865 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c686764-9b09-4725-9d67-55e22adec2c4" containerName="registry-server" Oct 14 08:05:58 crc kubenswrapper[4870]: I1014 08:05:58.374878 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c686764-9b09-4725-9d67-55e22adec2c4" containerName="registry-server" Oct 14 08:05:58 crc kubenswrapper[4870]: I1014 08:05:58.375144 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="7c686764-9b09-4725-9d67-55e22adec2c4" containerName="registry-server" Oct 14 08:05:58 crc kubenswrapper[4870]: I1014 08:05:58.377082 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-66rz4" Oct 14 08:05:58 crc kubenswrapper[4870]: I1014 08:05:58.404527 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-66rz4"] Oct 14 08:05:58 crc kubenswrapper[4870]: I1014 08:05:58.500964 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-blwc6\" (UniqueName: \"kubernetes.io/projected/15d300b8-ae83-4e0f-9d34-522aa236767c-kube-api-access-blwc6\") pod \"community-operators-66rz4\" (UID: \"15d300b8-ae83-4e0f-9d34-522aa236767c\") " pod="openshift-marketplace/community-operators-66rz4" Oct 14 08:05:58 crc kubenswrapper[4870]: I1014 08:05:58.501099 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/15d300b8-ae83-4e0f-9d34-522aa236767c-catalog-content\") pod \"community-operators-66rz4\" (UID: \"15d300b8-ae83-4e0f-9d34-522aa236767c\") " pod="openshift-marketplace/community-operators-66rz4" Oct 14 08:05:58 crc kubenswrapper[4870]: I1014 08:05:58.501174 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/15d300b8-ae83-4e0f-9d34-522aa236767c-utilities\") pod \"community-operators-66rz4\" (UID: \"15d300b8-ae83-4e0f-9d34-522aa236767c\") " pod="openshift-marketplace/community-operators-66rz4" Oct 14 08:05:58 crc kubenswrapper[4870]: I1014 08:05:58.602767 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-blwc6\" (UniqueName: \"kubernetes.io/projected/15d300b8-ae83-4e0f-9d34-522aa236767c-kube-api-access-blwc6\") pod \"community-operators-66rz4\" (UID: \"15d300b8-ae83-4e0f-9d34-522aa236767c\") " pod="openshift-marketplace/community-operators-66rz4" Oct 14 08:05:58 crc kubenswrapper[4870]: I1014 08:05:58.602891 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/15d300b8-ae83-4e0f-9d34-522aa236767c-catalog-content\") pod \"community-operators-66rz4\" (UID: \"15d300b8-ae83-4e0f-9d34-522aa236767c\") " pod="openshift-marketplace/community-operators-66rz4" Oct 14 08:05:58 crc kubenswrapper[4870]: I1014 08:05:58.602981 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/15d300b8-ae83-4e0f-9d34-522aa236767c-utilities\") pod \"community-operators-66rz4\" (UID: \"15d300b8-ae83-4e0f-9d34-522aa236767c\") " pod="openshift-marketplace/community-operators-66rz4" Oct 14 08:05:58 crc kubenswrapper[4870]: I1014 08:05:58.604050 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/15d300b8-ae83-4e0f-9d34-522aa236767c-utilities\") pod \"community-operators-66rz4\" (UID: \"15d300b8-ae83-4e0f-9d34-522aa236767c\") " pod="openshift-marketplace/community-operators-66rz4" Oct 14 08:05:58 crc kubenswrapper[4870]: I1014 08:05:58.604109 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/15d300b8-ae83-4e0f-9d34-522aa236767c-catalog-content\") pod \"community-operators-66rz4\" (UID: \"15d300b8-ae83-4e0f-9d34-522aa236767c\") " pod="openshift-marketplace/community-operators-66rz4" Oct 14 08:05:58 crc kubenswrapper[4870]: I1014 08:05:58.635122 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-blwc6\" (UniqueName: \"kubernetes.io/projected/15d300b8-ae83-4e0f-9d34-522aa236767c-kube-api-access-blwc6\") pod \"community-operators-66rz4\" (UID: \"15d300b8-ae83-4e0f-9d34-522aa236767c\") " pod="openshift-marketplace/community-operators-66rz4" Oct 14 08:05:58 crc kubenswrapper[4870]: I1014 08:05:58.711176 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-66rz4" Oct 14 08:05:59 crc kubenswrapper[4870]: I1014 08:05:59.033946 4870 scope.go:117] "RemoveContainer" containerID="dca50334a8cd2d1438da606d2f43eb3a283c73f0855c6f8bb11c380e552e3268" Oct 14 08:05:59 crc kubenswrapper[4870]: E1014 08:05:59.034157 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 08:05:59 crc kubenswrapper[4870]: I1014 08:05:59.232962 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-66rz4"] Oct 14 08:05:59 crc kubenswrapper[4870]: I1014 08:05:59.652583 4870 generic.go:334] "Generic (PLEG): container finished" podID="15d300b8-ae83-4e0f-9d34-522aa236767c" containerID="3412f9244dc86698c6b0fe7ecca18d4ad2ddd99ccd43fe3bf0454f3cd4ad005a" exitCode=0 Oct 14 08:05:59 crc kubenswrapper[4870]: I1014 08:05:59.652644 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-66rz4" event={"ID":"15d300b8-ae83-4e0f-9d34-522aa236767c","Type":"ContainerDied","Data":"3412f9244dc86698c6b0fe7ecca18d4ad2ddd99ccd43fe3bf0454f3cd4ad005a"} Oct 14 08:05:59 crc kubenswrapper[4870]: I1014 08:05:59.652685 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-66rz4" event={"ID":"15d300b8-ae83-4e0f-9d34-522aa236767c","Type":"ContainerStarted","Data":"f024e809ab0976dee18fb335d1ff96861bd9681d9910875933845e768b6c07db"} Oct 14 08:06:00 crc kubenswrapper[4870]: I1014 08:06:00.674793 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-66rz4" event={"ID":"15d300b8-ae83-4e0f-9d34-522aa236767c","Type":"ContainerStarted","Data":"873f33befd4a1a13388586cc506681a747bdff9eb180bf3dc7b3577e7a82dc8f"} Oct 14 08:06:01 crc kubenswrapper[4870]: I1014 08:06:01.688714 4870 generic.go:334] "Generic (PLEG): container finished" podID="15d300b8-ae83-4e0f-9d34-522aa236767c" containerID="873f33befd4a1a13388586cc506681a747bdff9eb180bf3dc7b3577e7a82dc8f" exitCode=0 Oct 14 08:06:01 crc kubenswrapper[4870]: I1014 08:06:01.688964 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-66rz4" event={"ID":"15d300b8-ae83-4e0f-9d34-522aa236767c","Type":"ContainerDied","Data":"873f33befd4a1a13388586cc506681a747bdff9eb180bf3dc7b3577e7a82dc8f"} Oct 14 08:06:02 crc kubenswrapper[4870]: I1014 08:06:02.703295 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-66rz4" event={"ID":"15d300b8-ae83-4e0f-9d34-522aa236767c","Type":"ContainerStarted","Data":"2d6f958e94344ab7ef46f6f788efa3b33704623581a7c5626a589fa53193e93c"} Oct 14 08:06:02 crc kubenswrapper[4870]: I1014 08:06:02.750489 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-66rz4" podStartSLOduration=2.244223324 podStartE2EDuration="4.750393511s" podCreationTimestamp="2025-10-14 08:05:58 +0000 UTC" firstStartedPulling="2025-10-14 08:05:59.654470642 +0000 UTC m=+3895.351831053" lastFinishedPulling="2025-10-14 08:06:02.160640849 +0000 UTC m=+3897.858001240" observedRunningTime="2025-10-14 08:06:02.730671498 +0000 UTC m=+3898.428031909" watchObservedRunningTime="2025-10-14 08:06:02.750393511 +0000 UTC m=+3898.447753972" Oct 14 08:06:08 crc kubenswrapper[4870]: I1014 08:06:08.712100 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-66rz4" Oct 14 08:06:08 crc kubenswrapper[4870]: I1014 08:06:08.712517 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-66rz4" Oct 14 08:06:08 crc kubenswrapper[4870]: I1014 08:06:08.792476 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-66rz4" Oct 14 08:06:09 crc kubenswrapper[4870]: I1014 08:06:09.868057 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-66rz4" Oct 14 08:06:09 crc kubenswrapper[4870]: I1014 08:06:09.940572 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-66rz4"] Oct 14 08:06:11 crc kubenswrapper[4870]: I1014 08:06:11.802492 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-66rz4" podUID="15d300b8-ae83-4e0f-9d34-522aa236767c" containerName="registry-server" containerID="cri-o://2d6f958e94344ab7ef46f6f788efa3b33704623581a7c5626a589fa53193e93c" gracePeriod=2 Oct 14 08:06:12 crc kubenswrapper[4870]: I1014 08:06:12.296633 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-66rz4" Oct 14 08:06:12 crc kubenswrapper[4870]: I1014 08:06:12.445573 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-blwc6\" (UniqueName: \"kubernetes.io/projected/15d300b8-ae83-4e0f-9d34-522aa236767c-kube-api-access-blwc6\") pod \"15d300b8-ae83-4e0f-9d34-522aa236767c\" (UID: \"15d300b8-ae83-4e0f-9d34-522aa236767c\") " Oct 14 08:06:12 crc kubenswrapper[4870]: I1014 08:06:12.445821 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/15d300b8-ae83-4e0f-9d34-522aa236767c-catalog-content\") pod \"15d300b8-ae83-4e0f-9d34-522aa236767c\" (UID: \"15d300b8-ae83-4e0f-9d34-522aa236767c\") " Oct 14 08:06:12 crc kubenswrapper[4870]: I1014 08:06:12.445886 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/15d300b8-ae83-4e0f-9d34-522aa236767c-utilities\") pod \"15d300b8-ae83-4e0f-9d34-522aa236767c\" (UID: \"15d300b8-ae83-4e0f-9d34-522aa236767c\") " Oct 14 08:06:12 crc kubenswrapper[4870]: I1014 08:06:12.447268 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/15d300b8-ae83-4e0f-9d34-522aa236767c-utilities" (OuterVolumeSpecName: "utilities") pod "15d300b8-ae83-4e0f-9d34-522aa236767c" (UID: "15d300b8-ae83-4e0f-9d34-522aa236767c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:06:12 crc kubenswrapper[4870]: I1014 08:06:12.468839 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/15d300b8-ae83-4e0f-9d34-522aa236767c-kube-api-access-blwc6" (OuterVolumeSpecName: "kube-api-access-blwc6") pod "15d300b8-ae83-4e0f-9d34-522aa236767c" (UID: "15d300b8-ae83-4e0f-9d34-522aa236767c"). InnerVolumeSpecName "kube-api-access-blwc6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:06:12 crc kubenswrapper[4870]: I1014 08:06:12.503086 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/15d300b8-ae83-4e0f-9d34-522aa236767c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "15d300b8-ae83-4e0f-9d34-522aa236767c" (UID: "15d300b8-ae83-4e0f-9d34-522aa236767c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:06:12 crc kubenswrapper[4870]: I1014 08:06:12.548142 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-blwc6\" (UniqueName: \"kubernetes.io/projected/15d300b8-ae83-4e0f-9d34-522aa236767c-kube-api-access-blwc6\") on node \"crc\" DevicePath \"\"" Oct 14 08:06:12 crc kubenswrapper[4870]: I1014 08:06:12.548204 4870 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/15d300b8-ae83-4e0f-9d34-522aa236767c-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 08:06:12 crc kubenswrapper[4870]: I1014 08:06:12.548223 4870 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/15d300b8-ae83-4e0f-9d34-522aa236767c-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 08:06:12 crc kubenswrapper[4870]: I1014 08:06:12.814696 4870 generic.go:334] "Generic (PLEG): container finished" podID="15d300b8-ae83-4e0f-9d34-522aa236767c" containerID="2d6f958e94344ab7ef46f6f788efa3b33704623581a7c5626a589fa53193e93c" exitCode=0 Oct 14 08:06:12 crc kubenswrapper[4870]: I1014 08:06:12.814765 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-66rz4" event={"ID":"15d300b8-ae83-4e0f-9d34-522aa236767c","Type":"ContainerDied","Data":"2d6f958e94344ab7ef46f6f788efa3b33704623581a7c5626a589fa53193e93c"} Oct 14 08:06:12 crc kubenswrapper[4870]: I1014 08:06:12.814803 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-66rz4" event={"ID":"15d300b8-ae83-4e0f-9d34-522aa236767c","Type":"ContainerDied","Data":"f024e809ab0976dee18fb335d1ff96861bd9681d9910875933845e768b6c07db"} Oct 14 08:06:12 crc kubenswrapper[4870]: I1014 08:06:12.814831 4870 scope.go:117] "RemoveContainer" containerID="2d6f958e94344ab7ef46f6f788efa3b33704623581a7c5626a589fa53193e93c" Oct 14 08:06:12 crc kubenswrapper[4870]: I1014 08:06:12.814861 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-66rz4" Oct 14 08:06:12 crc kubenswrapper[4870]: I1014 08:06:12.840976 4870 scope.go:117] "RemoveContainer" containerID="873f33befd4a1a13388586cc506681a747bdff9eb180bf3dc7b3577e7a82dc8f" Oct 14 08:06:12 crc kubenswrapper[4870]: I1014 08:06:12.874062 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-66rz4"] Oct 14 08:06:12 crc kubenswrapper[4870]: I1014 08:06:12.883482 4870 scope.go:117] "RemoveContainer" containerID="3412f9244dc86698c6b0fe7ecca18d4ad2ddd99ccd43fe3bf0454f3cd4ad005a" Oct 14 08:06:12 crc kubenswrapper[4870]: I1014 08:06:12.887768 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-66rz4"] Oct 14 08:06:12 crc kubenswrapper[4870]: I1014 08:06:12.909469 4870 scope.go:117] "RemoveContainer" containerID="2d6f958e94344ab7ef46f6f788efa3b33704623581a7c5626a589fa53193e93c" Oct 14 08:06:12 crc kubenswrapper[4870]: E1014 08:06:12.909952 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2d6f958e94344ab7ef46f6f788efa3b33704623581a7c5626a589fa53193e93c\": container with ID starting with 2d6f958e94344ab7ef46f6f788efa3b33704623581a7c5626a589fa53193e93c not found: ID does not exist" containerID="2d6f958e94344ab7ef46f6f788efa3b33704623581a7c5626a589fa53193e93c" Oct 14 08:06:12 crc kubenswrapper[4870]: I1014 08:06:12.909986 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2d6f958e94344ab7ef46f6f788efa3b33704623581a7c5626a589fa53193e93c"} err="failed to get container status \"2d6f958e94344ab7ef46f6f788efa3b33704623581a7c5626a589fa53193e93c\": rpc error: code = NotFound desc = could not find container \"2d6f958e94344ab7ef46f6f788efa3b33704623581a7c5626a589fa53193e93c\": container with ID starting with 2d6f958e94344ab7ef46f6f788efa3b33704623581a7c5626a589fa53193e93c not found: ID does not exist" Oct 14 08:06:12 crc kubenswrapper[4870]: I1014 08:06:12.910016 4870 scope.go:117] "RemoveContainer" containerID="873f33befd4a1a13388586cc506681a747bdff9eb180bf3dc7b3577e7a82dc8f" Oct 14 08:06:12 crc kubenswrapper[4870]: E1014 08:06:12.910285 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"873f33befd4a1a13388586cc506681a747bdff9eb180bf3dc7b3577e7a82dc8f\": container with ID starting with 873f33befd4a1a13388586cc506681a747bdff9eb180bf3dc7b3577e7a82dc8f not found: ID does not exist" containerID="873f33befd4a1a13388586cc506681a747bdff9eb180bf3dc7b3577e7a82dc8f" Oct 14 08:06:12 crc kubenswrapper[4870]: I1014 08:06:12.910323 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"873f33befd4a1a13388586cc506681a747bdff9eb180bf3dc7b3577e7a82dc8f"} err="failed to get container status \"873f33befd4a1a13388586cc506681a747bdff9eb180bf3dc7b3577e7a82dc8f\": rpc error: code = NotFound desc = could not find container \"873f33befd4a1a13388586cc506681a747bdff9eb180bf3dc7b3577e7a82dc8f\": container with ID starting with 873f33befd4a1a13388586cc506681a747bdff9eb180bf3dc7b3577e7a82dc8f not found: ID does not exist" Oct 14 08:06:12 crc kubenswrapper[4870]: I1014 08:06:12.910348 4870 scope.go:117] "RemoveContainer" containerID="3412f9244dc86698c6b0fe7ecca18d4ad2ddd99ccd43fe3bf0454f3cd4ad005a" Oct 14 08:06:12 crc kubenswrapper[4870]: E1014 08:06:12.910608 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3412f9244dc86698c6b0fe7ecca18d4ad2ddd99ccd43fe3bf0454f3cd4ad005a\": container with ID starting with 3412f9244dc86698c6b0fe7ecca18d4ad2ddd99ccd43fe3bf0454f3cd4ad005a not found: ID does not exist" containerID="3412f9244dc86698c6b0fe7ecca18d4ad2ddd99ccd43fe3bf0454f3cd4ad005a" Oct 14 08:06:12 crc kubenswrapper[4870]: I1014 08:06:12.910637 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3412f9244dc86698c6b0fe7ecca18d4ad2ddd99ccd43fe3bf0454f3cd4ad005a"} err="failed to get container status \"3412f9244dc86698c6b0fe7ecca18d4ad2ddd99ccd43fe3bf0454f3cd4ad005a\": rpc error: code = NotFound desc = could not find container \"3412f9244dc86698c6b0fe7ecca18d4ad2ddd99ccd43fe3bf0454f3cd4ad005a\": container with ID starting with 3412f9244dc86698c6b0fe7ecca18d4ad2ddd99ccd43fe3bf0454f3cd4ad005a not found: ID does not exist" Oct 14 08:06:13 crc kubenswrapper[4870]: I1014 08:06:13.044277 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="15d300b8-ae83-4e0f-9d34-522aa236767c" path="/var/lib/kubelet/pods/15d300b8-ae83-4e0f-9d34-522aa236767c/volumes" Oct 14 08:06:14 crc kubenswrapper[4870]: I1014 08:06:14.034204 4870 scope.go:117] "RemoveContainer" containerID="dca50334a8cd2d1438da606d2f43eb3a283c73f0855c6f8bb11c380e552e3268" Oct 14 08:06:14 crc kubenswrapper[4870]: E1014 08:06:14.034754 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 08:06:26 crc kubenswrapper[4870]: I1014 08:06:26.034409 4870 scope.go:117] "RemoveContainer" containerID="dca50334a8cd2d1438da606d2f43eb3a283c73f0855c6f8bb11c380e552e3268" Oct 14 08:06:26 crc kubenswrapper[4870]: E1014 08:06:26.035464 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 08:06:38 crc kubenswrapper[4870]: I1014 08:06:38.034826 4870 scope.go:117] "RemoveContainer" containerID="dca50334a8cd2d1438da606d2f43eb3a283c73f0855c6f8bb11c380e552e3268" Oct 14 08:06:38 crc kubenswrapper[4870]: E1014 08:06:38.036212 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 08:06:49 crc kubenswrapper[4870]: I1014 08:06:49.033941 4870 scope.go:117] "RemoveContainer" containerID="dca50334a8cd2d1438da606d2f43eb3a283c73f0855c6f8bb11c380e552e3268" Oct 14 08:06:49 crc kubenswrapper[4870]: E1014 08:06:49.035260 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 08:07:01 crc kubenswrapper[4870]: I1014 08:07:01.034878 4870 scope.go:117] "RemoveContainer" containerID="dca50334a8cd2d1438da606d2f43eb3a283c73f0855c6f8bb11c380e552e3268" Oct 14 08:07:01 crc kubenswrapper[4870]: E1014 08:07:01.036002 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 08:07:07 crc kubenswrapper[4870]: I1014 08:07:07.731230 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-mz89p"] Oct 14 08:07:07 crc kubenswrapper[4870]: E1014 08:07:07.733601 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15d300b8-ae83-4e0f-9d34-522aa236767c" containerName="extract-content" Oct 14 08:07:07 crc kubenswrapper[4870]: I1014 08:07:07.733694 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="15d300b8-ae83-4e0f-9d34-522aa236767c" containerName="extract-content" Oct 14 08:07:07 crc kubenswrapper[4870]: E1014 08:07:07.733799 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15d300b8-ae83-4e0f-9d34-522aa236767c" containerName="extract-utilities" Oct 14 08:07:07 crc kubenswrapper[4870]: I1014 08:07:07.733889 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="15d300b8-ae83-4e0f-9d34-522aa236767c" containerName="extract-utilities" Oct 14 08:07:07 crc kubenswrapper[4870]: E1014 08:07:07.733980 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15d300b8-ae83-4e0f-9d34-522aa236767c" containerName="registry-server" Oct 14 08:07:07 crc kubenswrapper[4870]: I1014 08:07:07.734060 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="15d300b8-ae83-4e0f-9d34-522aa236767c" containerName="registry-server" Oct 14 08:07:07 crc kubenswrapper[4870]: I1014 08:07:07.734287 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="15d300b8-ae83-4e0f-9d34-522aa236767c" containerName="registry-server" Oct 14 08:07:07 crc kubenswrapper[4870]: I1014 08:07:07.735465 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mz89p" Oct 14 08:07:07 crc kubenswrapper[4870]: I1014 08:07:07.752753 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-mz89p"] Oct 14 08:07:07 crc kubenswrapper[4870]: I1014 08:07:07.888766 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bqc6m\" (UniqueName: \"kubernetes.io/projected/33599891-137f-4153-ac3c-045703b2bc3e-kube-api-access-bqc6m\") pod \"redhat-marketplace-mz89p\" (UID: \"33599891-137f-4153-ac3c-045703b2bc3e\") " pod="openshift-marketplace/redhat-marketplace-mz89p" Oct 14 08:07:07 crc kubenswrapper[4870]: I1014 08:07:07.889074 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/33599891-137f-4153-ac3c-045703b2bc3e-utilities\") pod \"redhat-marketplace-mz89p\" (UID: \"33599891-137f-4153-ac3c-045703b2bc3e\") " pod="openshift-marketplace/redhat-marketplace-mz89p" Oct 14 08:07:07 crc kubenswrapper[4870]: I1014 08:07:07.889174 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/33599891-137f-4153-ac3c-045703b2bc3e-catalog-content\") pod \"redhat-marketplace-mz89p\" (UID: \"33599891-137f-4153-ac3c-045703b2bc3e\") " pod="openshift-marketplace/redhat-marketplace-mz89p" Oct 14 08:07:07 crc kubenswrapper[4870]: I1014 08:07:07.991090 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bqc6m\" (UniqueName: \"kubernetes.io/projected/33599891-137f-4153-ac3c-045703b2bc3e-kube-api-access-bqc6m\") pod \"redhat-marketplace-mz89p\" (UID: \"33599891-137f-4153-ac3c-045703b2bc3e\") " pod="openshift-marketplace/redhat-marketplace-mz89p" Oct 14 08:07:07 crc kubenswrapper[4870]: I1014 08:07:07.991166 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/33599891-137f-4153-ac3c-045703b2bc3e-utilities\") pod \"redhat-marketplace-mz89p\" (UID: \"33599891-137f-4153-ac3c-045703b2bc3e\") " pod="openshift-marketplace/redhat-marketplace-mz89p" Oct 14 08:07:07 crc kubenswrapper[4870]: I1014 08:07:07.991215 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/33599891-137f-4153-ac3c-045703b2bc3e-catalog-content\") pod \"redhat-marketplace-mz89p\" (UID: \"33599891-137f-4153-ac3c-045703b2bc3e\") " pod="openshift-marketplace/redhat-marketplace-mz89p" Oct 14 08:07:07 crc kubenswrapper[4870]: I1014 08:07:07.991639 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/33599891-137f-4153-ac3c-045703b2bc3e-utilities\") pod \"redhat-marketplace-mz89p\" (UID: \"33599891-137f-4153-ac3c-045703b2bc3e\") " pod="openshift-marketplace/redhat-marketplace-mz89p" Oct 14 08:07:07 crc kubenswrapper[4870]: I1014 08:07:07.991790 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/33599891-137f-4153-ac3c-045703b2bc3e-catalog-content\") pod \"redhat-marketplace-mz89p\" (UID: \"33599891-137f-4153-ac3c-045703b2bc3e\") " pod="openshift-marketplace/redhat-marketplace-mz89p" Oct 14 08:07:08 crc kubenswrapper[4870]: I1014 08:07:08.016719 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bqc6m\" (UniqueName: \"kubernetes.io/projected/33599891-137f-4153-ac3c-045703b2bc3e-kube-api-access-bqc6m\") pod \"redhat-marketplace-mz89p\" (UID: \"33599891-137f-4153-ac3c-045703b2bc3e\") " pod="openshift-marketplace/redhat-marketplace-mz89p" Oct 14 08:07:08 crc kubenswrapper[4870]: I1014 08:07:08.074716 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mz89p" Oct 14 08:07:08 crc kubenswrapper[4870]: I1014 08:07:08.652961 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-mz89p"] Oct 14 08:07:09 crc kubenswrapper[4870]: I1014 08:07:09.351530 4870 generic.go:334] "Generic (PLEG): container finished" podID="33599891-137f-4153-ac3c-045703b2bc3e" containerID="90fac99ebbea77c29addfb827e3df999b9c8eb5856b8c1025a57613212b020a7" exitCode=0 Oct 14 08:07:09 crc kubenswrapper[4870]: I1014 08:07:09.351689 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mz89p" event={"ID":"33599891-137f-4153-ac3c-045703b2bc3e","Type":"ContainerDied","Data":"90fac99ebbea77c29addfb827e3df999b9c8eb5856b8c1025a57613212b020a7"} Oct 14 08:07:09 crc kubenswrapper[4870]: I1014 08:07:09.352065 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mz89p" event={"ID":"33599891-137f-4153-ac3c-045703b2bc3e","Type":"ContainerStarted","Data":"524d7c56d99ac757b9abb6f8bbb5305f08009fc64b9aa6510fdb09c2eb115873"} Oct 14 08:07:10 crc kubenswrapper[4870]: I1014 08:07:10.363637 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mz89p" event={"ID":"33599891-137f-4153-ac3c-045703b2bc3e","Type":"ContainerStarted","Data":"29c69e682c87abf624627e544a3fd7f5e6da27f915777d7f0cea69796167ebcf"} Oct 14 08:07:11 crc kubenswrapper[4870]: I1014 08:07:11.374216 4870 generic.go:334] "Generic (PLEG): container finished" podID="33599891-137f-4153-ac3c-045703b2bc3e" containerID="29c69e682c87abf624627e544a3fd7f5e6da27f915777d7f0cea69796167ebcf" exitCode=0 Oct 14 08:07:11 crc kubenswrapper[4870]: I1014 08:07:11.374270 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mz89p" event={"ID":"33599891-137f-4153-ac3c-045703b2bc3e","Type":"ContainerDied","Data":"29c69e682c87abf624627e544a3fd7f5e6da27f915777d7f0cea69796167ebcf"} Oct 14 08:07:12 crc kubenswrapper[4870]: I1014 08:07:12.393657 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mz89p" event={"ID":"33599891-137f-4153-ac3c-045703b2bc3e","Type":"ContainerStarted","Data":"b884ac300b62bd2d61f63c5a261faf2bebdb196bb7014aafafc0a1bd8f20fe47"} Oct 14 08:07:12 crc kubenswrapper[4870]: I1014 08:07:12.432518 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-mz89p" podStartSLOduration=2.648861155 podStartE2EDuration="5.432489157s" podCreationTimestamp="2025-10-14 08:07:07 +0000 UTC" firstStartedPulling="2025-10-14 08:07:09.354274788 +0000 UTC m=+3965.051635189" lastFinishedPulling="2025-10-14 08:07:12.13790279 +0000 UTC m=+3967.835263191" observedRunningTime="2025-10-14 08:07:12.422112317 +0000 UTC m=+3968.119472728" watchObservedRunningTime="2025-10-14 08:07:12.432489157 +0000 UTC m=+3968.129849568" Oct 14 08:07:14 crc kubenswrapper[4870]: I1014 08:07:14.033878 4870 scope.go:117] "RemoveContainer" containerID="dca50334a8cd2d1438da606d2f43eb3a283c73f0855c6f8bb11c380e552e3268" Oct 14 08:07:14 crc kubenswrapper[4870]: E1014 08:07:14.034736 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 08:07:18 crc kubenswrapper[4870]: I1014 08:07:18.075750 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-mz89p" Oct 14 08:07:18 crc kubenswrapper[4870]: I1014 08:07:18.076128 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-mz89p" Oct 14 08:07:18 crc kubenswrapper[4870]: I1014 08:07:18.134767 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-mz89p" Oct 14 08:07:18 crc kubenswrapper[4870]: I1014 08:07:18.546218 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-mz89p" Oct 14 08:07:18 crc kubenswrapper[4870]: I1014 08:07:18.621785 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-mz89p"] Oct 14 08:07:20 crc kubenswrapper[4870]: I1014 08:07:20.486327 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-mz89p" podUID="33599891-137f-4153-ac3c-045703b2bc3e" containerName="registry-server" containerID="cri-o://b884ac300b62bd2d61f63c5a261faf2bebdb196bb7014aafafc0a1bd8f20fe47" gracePeriod=2 Oct 14 08:07:20 crc kubenswrapper[4870]: I1014 08:07:20.947952 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mz89p" Oct 14 08:07:21 crc kubenswrapper[4870]: I1014 08:07:21.039335 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/33599891-137f-4153-ac3c-045703b2bc3e-utilities\") pod \"33599891-137f-4153-ac3c-045703b2bc3e\" (UID: \"33599891-137f-4153-ac3c-045703b2bc3e\") " Oct 14 08:07:21 crc kubenswrapper[4870]: I1014 08:07:21.039581 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bqc6m\" (UniqueName: \"kubernetes.io/projected/33599891-137f-4153-ac3c-045703b2bc3e-kube-api-access-bqc6m\") pod \"33599891-137f-4153-ac3c-045703b2bc3e\" (UID: \"33599891-137f-4153-ac3c-045703b2bc3e\") " Oct 14 08:07:21 crc kubenswrapper[4870]: I1014 08:07:21.039762 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/33599891-137f-4153-ac3c-045703b2bc3e-catalog-content\") pod \"33599891-137f-4153-ac3c-045703b2bc3e\" (UID: \"33599891-137f-4153-ac3c-045703b2bc3e\") " Oct 14 08:07:21 crc kubenswrapper[4870]: I1014 08:07:21.046934 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/33599891-137f-4153-ac3c-045703b2bc3e-utilities" (OuterVolumeSpecName: "utilities") pod "33599891-137f-4153-ac3c-045703b2bc3e" (UID: "33599891-137f-4153-ac3c-045703b2bc3e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:07:21 crc kubenswrapper[4870]: I1014 08:07:21.054865 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/33599891-137f-4153-ac3c-045703b2bc3e-kube-api-access-bqc6m" (OuterVolumeSpecName: "kube-api-access-bqc6m") pod "33599891-137f-4153-ac3c-045703b2bc3e" (UID: "33599891-137f-4153-ac3c-045703b2bc3e"). InnerVolumeSpecName "kube-api-access-bqc6m". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:07:21 crc kubenswrapper[4870]: I1014 08:07:21.058574 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/33599891-137f-4153-ac3c-045703b2bc3e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "33599891-137f-4153-ac3c-045703b2bc3e" (UID: "33599891-137f-4153-ac3c-045703b2bc3e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:07:21 crc kubenswrapper[4870]: I1014 08:07:21.141973 4870 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/33599891-137f-4153-ac3c-045703b2bc3e-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 08:07:21 crc kubenswrapper[4870]: I1014 08:07:21.142020 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bqc6m\" (UniqueName: \"kubernetes.io/projected/33599891-137f-4153-ac3c-045703b2bc3e-kube-api-access-bqc6m\") on node \"crc\" DevicePath \"\"" Oct 14 08:07:21 crc kubenswrapper[4870]: I1014 08:07:21.142034 4870 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/33599891-137f-4153-ac3c-045703b2bc3e-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 08:07:21 crc kubenswrapper[4870]: I1014 08:07:21.504344 4870 generic.go:334] "Generic (PLEG): container finished" podID="33599891-137f-4153-ac3c-045703b2bc3e" containerID="b884ac300b62bd2d61f63c5a261faf2bebdb196bb7014aafafc0a1bd8f20fe47" exitCode=0 Oct 14 08:07:21 crc kubenswrapper[4870]: I1014 08:07:21.504433 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mz89p" Oct 14 08:07:21 crc kubenswrapper[4870]: I1014 08:07:21.504499 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mz89p" event={"ID":"33599891-137f-4153-ac3c-045703b2bc3e","Type":"ContainerDied","Data":"b884ac300b62bd2d61f63c5a261faf2bebdb196bb7014aafafc0a1bd8f20fe47"} Oct 14 08:07:21 crc kubenswrapper[4870]: I1014 08:07:21.504583 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mz89p" event={"ID":"33599891-137f-4153-ac3c-045703b2bc3e","Type":"ContainerDied","Data":"524d7c56d99ac757b9abb6f8bbb5305f08009fc64b9aa6510fdb09c2eb115873"} Oct 14 08:07:21 crc kubenswrapper[4870]: I1014 08:07:21.504636 4870 scope.go:117] "RemoveContainer" containerID="b884ac300b62bd2d61f63c5a261faf2bebdb196bb7014aafafc0a1bd8f20fe47" Oct 14 08:07:21 crc kubenswrapper[4870]: I1014 08:07:21.533845 4870 scope.go:117] "RemoveContainer" containerID="29c69e682c87abf624627e544a3fd7f5e6da27f915777d7f0cea69796167ebcf" Oct 14 08:07:21 crc kubenswrapper[4870]: I1014 08:07:21.564088 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-mz89p"] Oct 14 08:07:21 crc kubenswrapper[4870]: I1014 08:07:21.570035 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-mz89p"] Oct 14 08:07:21 crc kubenswrapper[4870]: I1014 08:07:21.588648 4870 scope.go:117] "RemoveContainer" containerID="90fac99ebbea77c29addfb827e3df999b9c8eb5856b8c1025a57613212b020a7" Oct 14 08:07:21 crc kubenswrapper[4870]: I1014 08:07:21.607451 4870 scope.go:117] "RemoveContainer" containerID="b884ac300b62bd2d61f63c5a261faf2bebdb196bb7014aafafc0a1bd8f20fe47" Oct 14 08:07:21 crc kubenswrapper[4870]: E1014 08:07:21.607790 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b884ac300b62bd2d61f63c5a261faf2bebdb196bb7014aafafc0a1bd8f20fe47\": container with ID starting with b884ac300b62bd2d61f63c5a261faf2bebdb196bb7014aafafc0a1bd8f20fe47 not found: ID does not exist" containerID="b884ac300b62bd2d61f63c5a261faf2bebdb196bb7014aafafc0a1bd8f20fe47" Oct 14 08:07:21 crc kubenswrapper[4870]: I1014 08:07:21.607821 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b884ac300b62bd2d61f63c5a261faf2bebdb196bb7014aafafc0a1bd8f20fe47"} err="failed to get container status \"b884ac300b62bd2d61f63c5a261faf2bebdb196bb7014aafafc0a1bd8f20fe47\": rpc error: code = NotFound desc = could not find container \"b884ac300b62bd2d61f63c5a261faf2bebdb196bb7014aafafc0a1bd8f20fe47\": container with ID starting with b884ac300b62bd2d61f63c5a261faf2bebdb196bb7014aafafc0a1bd8f20fe47 not found: ID does not exist" Oct 14 08:07:21 crc kubenswrapper[4870]: I1014 08:07:21.607843 4870 scope.go:117] "RemoveContainer" containerID="29c69e682c87abf624627e544a3fd7f5e6da27f915777d7f0cea69796167ebcf" Oct 14 08:07:21 crc kubenswrapper[4870]: E1014 08:07:21.608383 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"29c69e682c87abf624627e544a3fd7f5e6da27f915777d7f0cea69796167ebcf\": container with ID starting with 29c69e682c87abf624627e544a3fd7f5e6da27f915777d7f0cea69796167ebcf not found: ID does not exist" containerID="29c69e682c87abf624627e544a3fd7f5e6da27f915777d7f0cea69796167ebcf" Oct 14 08:07:21 crc kubenswrapper[4870]: I1014 08:07:21.608428 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"29c69e682c87abf624627e544a3fd7f5e6da27f915777d7f0cea69796167ebcf"} err="failed to get container status \"29c69e682c87abf624627e544a3fd7f5e6da27f915777d7f0cea69796167ebcf\": rpc error: code = NotFound desc = could not find container \"29c69e682c87abf624627e544a3fd7f5e6da27f915777d7f0cea69796167ebcf\": container with ID starting with 29c69e682c87abf624627e544a3fd7f5e6da27f915777d7f0cea69796167ebcf not found: ID does not exist" Oct 14 08:07:21 crc kubenswrapper[4870]: I1014 08:07:21.608502 4870 scope.go:117] "RemoveContainer" containerID="90fac99ebbea77c29addfb827e3df999b9c8eb5856b8c1025a57613212b020a7" Oct 14 08:07:21 crc kubenswrapper[4870]: E1014 08:07:21.608811 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"90fac99ebbea77c29addfb827e3df999b9c8eb5856b8c1025a57613212b020a7\": container with ID starting with 90fac99ebbea77c29addfb827e3df999b9c8eb5856b8c1025a57613212b020a7 not found: ID does not exist" containerID="90fac99ebbea77c29addfb827e3df999b9c8eb5856b8c1025a57613212b020a7" Oct 14 08:07:21 crc kubenswrapper[4870]: I1014 08:07:21.608841 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"90fac99ebbea77c29addfb827e3df999b9c8eb5856b8c1025a57613212b020a7"} err="failed to get container status \"90fac99ebbea77c29addfb827e3df999b9c8eb5856b8c1025a57613212b020a7\": rpc error: code = NotFound desc = could not find container \"90fac99ebbea77c29addfb827e3df999b9c8eb5856b8c1025a57613212b020a7\": container with ID starting with 90fac99ebbea77c29addfb827e3df999b9c8eb5856b8c1025a57613212b020a7 not found: ID does not exist" Oct 14 08:07:23 crc kubenswrapper[4870]: I1014 08:07:23.050039 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="33599891-137f-4153-ac3c-045703b2bc3e" path="/var/lib/kubelet/pods/33599891-137f-4153-ac3c-045703b2bc3e/volumes" Oct 14 08:07:25 crc kubenswrapper[4870]: I1014 08:07:25.043833 4870 scope.go:117] "RemoveContainer" containerID="dca50334a8cd2d1438da606d2f43eb3a283c73f0855c6f8bb11c380e552e3268" Oct 14 08:07:25 crc kubenswrapper[4870]: E1014 08:07:25.044225 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 08:07:37 crc kubenswrapper[4870]: I1014 08:07:37.034237 4870 scope.go:117] "RemoveContainer" containerID="dca50334a8cd2d1438da606d2f43eb3a283c73f0855c6f8bb11c380e552e3268" Oct 14 08:07:37 crc kubenswrapper[4870]: E1014 08:07:37.035796 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 08:07:51 crc kubenswrapper[4870]: I1014 08:07:51.034402 4870 scope.go:117] "RemoveContainer" containerID="dca50334a8cd2d1438da606d2f43eb3a283c73f0855c6f8bb11c380e552e3268" Oct 14 08:07:51 crc kubenswrapper[4870]: E1014 08:07:51.035486 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 08:08:02 crc kubenswrapper[4870]: I1014 08:08:02.033760 4870 scope.go:117] "RemoveContainer" containerID="dca50334a8cd2d1438da606d2f43eb3a283c73f0855c6f8bb11c380e552e3268" Oct 14 08:08:02 crc kubenswrapper[4870]: E1014 08:08:02.034896 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 08:08:15 crc kubenswrapper[4870]: I1014 08:08:15.045513 4870 scope.go:117] "RemoveContainer" containerID="dca50334a8cd2d1438da606d2f43eb3a283c73f0855c6f8bb11c380e552e3268" Oct 14 08:08:15 crc kubenswrapper[4870]: E1014 08:08:15.046571 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 08:08:28 crc kubenswrapper[4870]: I1014 08:08:28.034951 4870 scope.go:117] "RemoveContainer" containerID="dca50334a8cd2d1438da606d2f43eb3a283c73f0855c6f8bb11c380e552e3268" Oct 14 08:08:28 crc kubenswrapper[4870]: E1014 08:08:28.036083 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 08:08:41 crc kubenswrapper[4870]: I1014 08:08:41.034192 4870 scope.go:117] "RemoveContainer" containerID="dca50334a8cd2d1438da606d2f43eb3a283c73f0855c6f8bb11c380e552e3268" Oct 14 08:08:41 crc kubenswrapper[4870]: E1014 08:08:41.035068 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 08:08:48 crc kubenswrapper[4870]: I1014 08:08:48.030048 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-ljzfk"] Oct 14 08:08:48 crc kubenswrapper[4870]: E1014 08:08:48.031085 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33599891-137f-4153-ac3c-045703b2bc3e" containerName="extract-content" Oct 14 08:08:48 crc kubenswrapper[4870]: I1014 08:08:48.031116 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="33599891-137f-4153-ac3c-045703b2bc3e" containerName="extract-content" Oct 14 08:08:48 crc kubenswrapper[4870]: E1014 08:08:48.031144 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33599891-137f-4153-ac3c-045703b2bc3e" containerName="registry-server" Oct 14 08:08:48 crc kubenswrapper[4870]: I1014 08:08:48.031161 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="33599891-137f-4153-ac3c-045703b2bc3e" containerName="registry-server" Oct 14 08:08:48 crc kubenswrapper[4870]: E1014 08:08:48.031203 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33599891-137f-4153-ac3c-045703b2bc3e" containerName="extract-utilities" Oct 14 08:08:48 crc kubenswrapper[4870]: I1014 08:08:48.031221 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="33599891-137f-4153-ac3c-045703b2bc3e" containerName="extract-utilities" Oct 14 08:08:48 crc kubenswrapper[4870]: I1014 08:08:48.031598 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="33599891-137f-4153-ac3c-045703b2bc3e" containerName="registry-server" Oct 14 08:08:48 crc kubenswrapper[4870]: I1014 08:08:48.034584 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ljzfk" Oct 14 08:08:48 crc kubenswrapper[4870]: I1014 08:08:48.040320 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-ljzfk"] Oct 14 08:08:48 crc kubenswrapper[4870]: I1014 08:08:48.108179 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e52e7047-2b96-43a8-a320-5a0768e75333-utilities\") pod \"certified-operators-ljzfk\" (UID: \"e52e7047-2b96-43a8-a320-5a0768e75333\") " pod="openshift-marketplace/certified-operators-ljzfk" Oct 14 08:08:48 crc kubenswrapper[4870]: I1014 08:08:48.108650 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e52e7047-2b96-43a8-a320-5a0768e75333-catalog-content\") pod \"certified-operators-ljzfk\" (UID: \"e52e7047-2b96-43a8-a320-5a0768e75333\") " pod="openshift-marketplace/certified-operators-ljzfk" Oct 14 08:08:48 crc kubenswrapper[4870]: I1014 08:08:48.108716 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-72tz6\" (UniqueName: \"kubernetes.io/projected/e52e7047-2b96-43a8-a320-5a0768e75333-kube-api-access-72tz6\") pod \"certified-operators-ljzfk\" (UID: \"e52e7047-2b96-43a8-a320-5a0768e75333\") " pod="openshift-marketplace/certified-operators-ljzfk" Oct 14 08:08:48 crc kubenswrapper[4870]: I1014 08:08:48.210059 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e52e7047-2b96-43a8-a320-5a0768e75333-utilities\") pod \"certified-operators-ljzfk\" (UID: \"e52e7047-2b96-43a8-a320-5a0768e75333\") " pod="openshift-marketplace/certified-operators-ljzfk" Oct 14 08:08:48 crc kubenswrapper[4870]: I1014 08:08:48.210170 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e52e7047-2b96-43a8-a320-5a0768e75333-catalog-content\") pod \"certified-operators-ljzfk\" (UID: \"e52e7047-2b96-43a8-a320-5a0768e75333\") " pod="openshift-marketplace/certified-operators-ljzfk" Oct 14 08:08:48 crc kubenswrapper[4870]: I1014 08:08:48.210265 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-72tz6\" (UniqueName: \"kubernetes.io/projected/e52e7047-2b96-43a8-a320-5a0768e75333-kube-api-access-72tz6\") pod \"certified-operators-ljzfk\" (UID: \"e52e7047-2b96-43a8-a320-5a0768e75333\") " pod="openshift-marketplace/certified-operators-ljzfk" Oct 14 08:08:48 crc kubenswrapper[4870]: I1014 08:08:48.210898 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e52e7047-2b96-43a8-a320-5a0768e75333-utilities\") pod \"certified-operators-ljzfk\" (UID: \"e52e7047-2b96-43a8-a320-5a0768e75333\") " pod="openshift-marketplace/certified-operators-ljzfk" Oct 14 08:08:48 crc kubenswrapper[4870]: I1014 08:08:48.210910 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e52e7047-2b96-43a8-a320-5a0768e75333-catalog-content\") pod \"certified-operators-ljzfk\" (UID: \"e52e7047-2b96-43a8-a320-5a0768e75333\") " pod="openshift-marketplace/certified-operators-ljzfk" Oct 14 08:08:48 crc kubenswrapper[4870]: I1014 08:08:48.252265 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-72tz6\" (UniqueName: \"kubernetes.io/projected/e52e7047-2b96-43a8-a320-5a0768e75333-kube-api-access-72tz6\") pod \"certified-operators-ljzfk\" (UID: \"e52e7047-2b96-43a8-a320-5a0768e75333\") " pod="openshift-marketplace/certified-operators-ljzfk" Oct 14 08:08:48 crc kubenswrapper[4870]: I1014 08:08:48.391794 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ljzfk" Oct 14 08:08:48 crc kubenswrapper[4870]: I1014 08:08:48.698930 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-ljzfk"] Oct 14 08:08:49 crc kubenswrapper[4870]: I1014 08:08:49.331919 4870 generic.go:334] "Generic (PLEG): container finished" podID="e52e7047-2b96-43a8-a320-5a0768e75333" containerID="d3410ea217979507c91d9f028f9ab8ff63f748dfa1205610aae1040430049cc6" exitCode=0 Oct 14 08:08:49 crc kubenswrapper[4870]: I1014 08:08:49.331987 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ljzfk" event={"ID":"e52e7047-2b96-43a8-a320-5a0768e75333","Type":"ContainerDied","Data":"d3410ea217979507c91d9f028f9ab8ff63f748dfa1205610aae1040430049cc6"} Oct 14 08:08:49 crc kubenswrapper[4870]: I1014 08:08:49.332311 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ljzfk" event={"ID":"e52e7047-2b96-43a8-a320-5a0768e75333","Type":"ContainerStarted","Data":"354c5f837949fae4d23b398cd5486d1f77623a10278c77fd2fff90e006c9b545"} Oct 14 08:08:52 crc kubenswrapper[4870]: I1014 08:08:52.365593 4870 generic.go:334] "Generic (PLEG): container finished" podID="e52e7047-2b96-43a8-a320-5a0768e75333" containerID="f07bcd67d2f61869e40f116f0307b300af64a206d272afbad628a55973239814" exitCode=0 Oct 14 08:08:52 crc kubenswrapper[4870]: I1014 08:08:52.365678 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ljzfk" event={"ID":"e52e7047-2b96-43a8-a320-5a0768e75333","Type":"ContainerDied","Data":"f07bcd67d2f61869e40f116f0307b300af64a206d272afbad628a55973239814"} Oct 14 08:08:54 crc kubenswrapper[4870]: I1014 08:08:54.034790 4870 scope.go:117] "RemoveContainer" containerID="dca50334a8cd2d1438da606d2f43eb3a283c73f0855c6f8bb11c380e552e3268" Oct 14 08:08:54 crc kubenswrapper[4870]: E1014 08:08:54.035372 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 08:08:54 crc kubenswrapper[4870]: I1014 08:08:54.391814 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ljzfk" event={"ID":"e52e7047-2b96-43a8-a320-5a0768e75333","Type":"ContainerStarted","Data":"5514ea447bf31937e03d3ef02baebc0bace6c0ff924a4153c4c450f01dbfd87e"} Oct 14 08:08:54 crc kubenswrapper[4870]: I1014 08:08:54.419220 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-ljzfk" podStartSLOduration=3.503316484 podStartE2EDuration="7.419182255s" podCreationTimestamp="2025-10-14 08:08:47 +0000 UTC" firstStartedPulling="2025-10-14 08:08:49.33543157 +0000 UTC m=+4065.032791941" lastFinishedPulling="2025-10-14 08:08:53.251297301 +0000 UTC m=+4068.948657712" observedRunningTime="2025-10-14 08:08:54.415404004 +0000 UTC m=+4070.112764405" watchObservedRunningTime="2025-10-14 08:08:54.419182255 +0000 UTC m=+4070.116542676" Oct 14 08:08:58 crc kubenswrapper[4870]: I1014 08:08:58.392632 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-ljzfk" Oct 14 08:08:58 crc kubenswrapper[4870]: I1014 08:08:58.393890 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-ljzfk" Oct 14 08:08:58 crc kubenswrapper[4870]: I1014 08:08:58.466699 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-ljzfk" Oct 14 08:08:58 crc kubenswrapper[4870]: I1014 08:08:58.531275 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-ljzfk" Oct 14 08:08:58 crc kubenswrapper[4870]: I1014 08:08:58.714573 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-ljzfk"] Oct 14 08:09:00 crc kubenswrapper[4870]: I1014 08:09:00.443869 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-ljzfk" podUID="e52e7047-2b96-43a8-a320-5a0768e75333" containerName="registry-server" containerID="cri-o://5514ea447bf31937e03d3ef02baebc0bace6c0ff924a4153c4c450f01dbfd87e" gracePeriod=2 Oct 14 08:09:01 crc kubenswrapper[4870]: I1014 08:09:01.006967 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ljzfk" Oct 14 08:09:01 crc kubenswrapper[4870]: I1014 08:09:01.196041 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e52e7047-2b96-43a8-a320-5a0768e75333-catalog-content\") pod \"e52e7047-2b96-43a8-a320-5a0768e75333\" (UID: \"e52e7047-2b96-43a8-a320-5a0768e75333\") " Oct 14 08:09:01 crc kubenswrapper[4870]: I1014 08:09:01.196131 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-72tz6\" (UniqueName: \"kubernetes.io/projected/e52e7047-2b96-43a8-a320-5a0768e75333-kube-api-access-72tz6\") pod \"e52e7047-2b96-43a8-a320-5a0768e75333\" (UID: \"e52e7047-2b96-43a8-a320-5a0768e75333\") " Oct 14 08:09:01 crc kubenswrapper[4870]: I1014 08:09:01.196343 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e52e7047-2b96-43a8-a320-5a0768e75333-utilities\") pod \"e52e7047-2b96-43a8-a320-5a0768e75333\" (UID: \"e52e7047-2b96-43a8-a320-5a0768e75333\") " Oct 14 08:09:01 crc kubenswrapper[4870]: I1014 08:09:01.198034 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e52e7047-2b96-43a8-a320-5a0768e75333-utilities" (OuterVolumeSpecName: "utilities") pod "e52e7047-2b96-43a8-a320-5a0768e75333" (UID: "e52e7047-2b96-43a8-a320-5a0768e75333"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:09:01 crc kubenswrapper[4870]: I1014 08:09:01.208748 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e52e7047-2b96-43a8-a320-5a0768e75333-kube-api-access-72tz6" (OuterVolumeSpecName: "kube-api-access-72tz6") pod "e52e7047-2b96-43a8-a320-5a0768e75333" (UID: "e52e7047-2b96-43a8-a320-5a0768e75333"). InnerVolumeSpecName "kube-api-access-72tz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:09:01 crc kubenswrapper[4870]: I1014 08:09:01.271641 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e52e7047-2b96-43a8-a320-5a0768e75333-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e52e7047-2b96-43a8-a320-5a0768e75333" (UID: "e52e7047-2b96-43a8-a320-5a0768e75333"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:09:01 crc kubenswrapper[4870]: I1014 08:09:01.298438 4870 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e52e7047-2b96-43a8-a320-5a0768e75333-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 08:09:01 crc kubenswrapper[4870]: I1014 08:09:01.298486 4870 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e52e7047-2b96-43a8-a320-5a0768e75333-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 08:09:01 crc kubenswrapper[4870]: I1014 08:09:01.298497 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-72tz6\" (UniqueName: \"kubernetes.io/projected/e52e7047-2b96-43a8-a320-5a0768e75333-kube-api-access-72tz6\") on node \"crc\" DevicePath \"\"" Oct 14 08:09:01 crc kubenswrapper[4870]: I1014 08:09:01.457353 4870 generic.go:334] "Generic (PLEG): container finished" podID="e52e7047-2b96-43a8-a320-5a0768e75333" containerID="5514ea447bf31937e03d3ef02baebc0bace6c0ff924a4153c4c450f01dbfd87e" exitCode=0 Oct 14 08:09:01 crc kubenswrapper[4870]: I1014 08:09:01.457480 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ljzfk" Oct 14 08:09:01 crc kubenswrapper[4870]: I1014 08:09:01.457491 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ljzfk" event={"ID":"e52e7047-2b96-43a8-a320-5a0768e75333","Type":"ContainerDied","Data":"5514ea447bf31937e03d3ef02baebc0bace6c0ff924a4153c4c450f01dbfd87e"} Oct 14 08:09:01 crc kubenswrapper[4870]: I1014 08:09:01.457962 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ljzfk" event={"ID":"e52e7047-2b96-43a8-a320-5a0768e75333","Type":"ContainerDied","Data":"354c5f837949fae4d23b398cd5486d1f77623a10278c77fd2fff90e006c9b545"} Oct 14 08:09:01 crc kubenswrapper[4870]: I1014 08:09:01.457998 4870 scope.go:117] "RemoveContainer" containerID="5514ea447bf31937e03d3ef02baebc0bace6c0ff924a4153c4c450f01dbfd87e" Oct 14 08:09:01 crc kubenswrapper[4870]: I1014 08:09:01.499069 4870 scope.go:117] "RemoveContainer" containerID="f07bcd67d2f61869e40f116f0307b300af64a206d272afbad628a55973239814" Oct 14 08:09:01 crc kubenswrapper[4870]: I1014 08:09:01.516754 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-ljzfk"] Oct 14 08:09:01 crc kubenswrapper[4870]: I1014 08:09:01.527017 4870 scope.go:117] "RemoveContainer" containerID="d3410ea217979507c91d9f028f9ab8ff63f748dfa1205610aae1040430049cc6" Oct 14 08:09:01 crc kubenswrapper[4870]: I1014 08:09:01.528404 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-ljzfk"] Oct 14 08:09:01 crc kubenswrapper[4870]: I1014 08:09:01.562558 4870 scope.go:117] "RemoveContainer" containerID="5514ea447bf31937e03d3ef02baebc0bace6c0ff924a4153c4c450f01dbfd87e" Oct 14 08:09:01 crc kubenswrapper[4870]: E1014 08:09:01.563139 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5514ea447bf31937e03d3ef02baebc0bace6c0ff924a4153c4c450f01dbfd87e\": container with ID starting with 5514ea447bf31937e03d3ef02baebc0bace6c0ff924a4153c4c450f01dbfd87e not found: ID does not exist" containerID="5514ea447bf31937e03d3ef02baebc0bace6c0ff924a4153c4c450f01dbfd87e" Oct 14 08:09:01 crc kubenswrapper[4870]: I1014 08:09:01.563197 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5514ea447bf31937e03d3ef02baebc0bace6c0ff924a4153c4c450f01dbfd87e"} err="failed to get container status \"5514ea447bf31937e03d3ef02baebc0bace6c0ff924a4153c4c450f01dbfd87e\": rpc error: code = NotFound desc = could not find container \"5514ea447bf31937e03d3ef02baebc0bace6c0ff924a4153c4c450f01dbfd87e\": container with ID starting with 5514ea447bf31937e03d3ef02baebc0bace6c0ff924a4153c4c450f01dbfd87e not found: ID does not exist" Oct 14 08:09:01 crc kubenswrapper[4870]: I1014 08:09:01.563235 4870 scope.go:117] "RemoveContainer" containerID="f07bcd67d2f61869e40f116f0307b300af64a206d272afbad628a55973239814" Oct 14 08:09:01 crc kubenswrapper[4870]: E1014 08:09:01.563911 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f07bcd67d2f61869e40f116f0307b300af64a206d272afbad628a55973239814\": container with ID starting with f07bcd67d2f61869e40f116f0307b300af64a206d272afbad628a55973239814 not found: ID does not exist" containerID="f07bcd67d2f61869e40f116f0307b300af64a206d272afbad628a55973239814" Oct 14 08:09:01 crc kubenswrapper[4870]: I1014 08:09:01.563998 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f07bcd67d2f61869e40f116f0307b300af64a206d272afbad628a55973239814"} err="failed to get container status \"f07bcd67d2f61869e40f116f0307b300af64a206d272afbad628a55973239814\": rpc error: code = NotFound desc = could not find container \"f07bcd67d2f61869e40f116f0307b300af64a206d272afbad628a55973239814\": container with ID starting with f07bcd67d2f61869e40f116f0307b300af64a206d272afbad628a55973239814 not found: ID does not exist" Oct 14 08:09:01 crc kubenswrapper[4870]: I1014 08:09:01.564035 4870 scope.go:117] "RemoveContainer" containerID="d3410ea217979507c91d9f028f9ab8ff63f748dfa1205610aae1040430049cc6" Oct 14 08:09:01 crc kubenswrapper[4870]: E1014 08:09:01.564514 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d3410ea217979507c91d9f028f9ab8ff63f748dfa1205610aae1040430049cc6\": container with ID starting with d3410ea217979507c91d9f028f9ab8ff63f748dfa1205610aae1040430049cc6 not found: ID does not exist" containerID="d3410ea217979507c91d9f028f9ab8ff63f748dfa1205610aae1040430049cc6" Oct 14 08:09:01 crc kubenswrapper[4870]: I1014 08:09:01.564550 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d3410ea217979507c91d9f028f9ab8ff63f748dfa1205610aae1040430049cc6"} err="failed to get container status \"d3410ea217979507c91d9f028f9ab8ff63f748dfa1205610aae1040430049cc6\": rpc error: code = NotFound desc = could not find container \"d3410ea217979507c91d9f028f9ab8ff63f748dfa1205610aae1040430049cc6\": container with ID starting with d3410ea217979507c91d9f028f9ab8ff63f748dfa1205610aae1040430049cc6 not found: ID does not exist" Oct 14 08:09:03 crc kubenswrapper[4870]: I1014 08:09:03.052281 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e52e7047-2b96-43a8-a320-5a0768e75333" path="/var/lib/kubelet/pods/e52e7047-2b96-43a8-a320-5a0768e75333/volumes" Oct 14 08:09:06 crc kubenswrapper[4870]: I1014 08:09:06.034977 4870 scope.go:117] "RemoveContainer" containerID="dca50334a8cd2d1438da606d2f43eb3a283c73f0855c6f8bb11c380e552e3268" Oct 14 08:09:06 crc kubenswrapper[4870]: E1014 08:09:06.036049 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 08:09:17 crc kubenswrapper[4870]: I1014 08:09:17.034050 4870 scope.go:117] "RemoveContainer" containerID="dca50334a8cd2d1438da606d2f43eb3a283c73f0855c6f8bb11c380e552e3268" Oct 14 08:09:17 crc kubenswrapper[4870]: E1014 08:09:17.035198 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 08:09:28 crc kubenswrapper[4870]: I1014 08:09:28.034517 4870 scope.go:117] "RemoveContainer" containerID="dca50334a8cd2d1438da606d2f43eb3a283c73f0855c6f8bb11c380e552e3268" Oct 14 08:09:28 crc kubenswrapper[4870]: E1014 08:09:28.035504 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 08:09:42 crc kubenswrapper[4870]: I1014 08:09:42.033904 4870 scope.go:117] "RemoveContainer" containerID="dca50334a8cd2d1438da606d2f43eb3a283c73f0855c6f8bb11c380e552e3268" Oct 14 08:09:42 crc kubenswrapper[4870]: E1014 08:09:42.036823 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 08:09:56 crc kubenswrapper[4870]: I1014 08:09:56.034473 4870 scope.go:117] "RemoveContainer" containerID="dca50334a8cd2d1438da606d2f43eb3a283c73f0855c6f8bb11c380e552e3268" Oct 14 08:09:56 crc kubenswrapper[4870]: E1014 08:09:56.035515 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 08:10:08 crc kubenswrapper[4870]: I1014 08:10:08.034805 4870 scope.go:117] "RemoveContainer" containerID="dca50334a8cd2d1438da606d2f43eb3a283c73f0855c6f8bb11c380e552e3268" Oct 14 08:10:08 crc kubenswrapper[4870]: E1014 08:10:08.035984 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 08:10:20 crc kubenswrapper[4870]: I1014 08:10:20.034856 4870 scope.go:117] "RemoveContainer" containerID="dca50334a8cd2d1438da606d2f43eb3a283c73f0855c6f8bb11c380e552e3268" Oct 14 08:10:20 crc kubenswrapper[4870]: E1014 08:10:20.036016 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 08:10:34 crc kubenswrapper[4870]: I1014 08:10:34.033876 4870 scope.go:117] "RemoveContainer" containerID="dca50334a8cd2d1438da606d2f43eb3a283c73f0855c6f8bb11c380e552e3268" Oct 14 08:10:34 crc kubenswrapper[4870]: I1014 08:10:34.437700 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" event={"ID":"874a5e45-dffd-4d17-b609-b7d3ed2eab07","Type":"ContainerStarted","Data":"52e4290890c8147a0b5888fc1327a813d1c142f382c705fe446f84040afd3bef"} Oct 14 08:12:53 crc kubenswrapper[4870]: I1014 08:12:53.951611 4870 patch_prober.go:28] interesting pod/machine-config-daemon-7tvc7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 08:12:53 crc kubenswrapper[4870]: I1014 08:12:53.952206 4870 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 08:13:23 crc kubenswrapper[4870]: I1014 08:13:23.951591 4870 patch_prober.go:28] interesting pod/machine-config-daemon-7tvc7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 08:13:23 crc kubenswrapper[4870]: I1014 08:13:23.952581 4870 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 08:13:53 crc kubenswrapper[4870]: I1014 08:13:53.950709 4870 patch_prober.go:28] interesting pod/machine-config-daemon-7tvc7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 08:13:53 crc kubenswrapper[4870]: I1014 08:13:53.951498 4870 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 08:13:53 crc kubenswrapper[4870]: I1014 08:13:53.951590 4870 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" Oct 14 08:13:53 crc kubenswrapper[4870]: I1014 08:13:53.952405 4870 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"52e4290890c8147a0b5888fc1327a813d1c142f382c705fe446f84040afd3bef"} pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 14 08:13:53 crc kubenswrapper[4870]: I1014 08:13:53.952557 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerName="machine-config-daemon" containerID="cri-o://52e4290890c8147a0b5888fc1327a813d1c142f382c705fe446f84040afd3bef" gracePeriod=600 Oct 14 08:13:54 crc kubenswrapper[4870]: I1014 08:13:54.321975 4870 generic.go:334] "Generic (PLEG): container finished" podID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerID="52e4290890c8147a0b5888fc1327a813d1c142f382c705fe446f84040afd3bef" exitCode=0 Oct 14 08:13:54 crc kubenswrapper[4870]: I1014 08:13:54.322514 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" event={"ID":"874a5e45-dffd-4d17-b609-b7d3ed2eab07","Type":"ContainerDied","Data":"52e4290890c8147a0b5888fc1327a813d1c142f382c705fe446f84040afd3bef"} Oct 14 08:13:54 crc kubenswrapper[4870]: I1014 08:13:54.322556 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" event={"ID":"874a5e45-dffd-4d17-b609-b7d3ed2eab07","Type":"ContainerStarted","Data":"cb5962db9597015a8eaf724ffe6622e32b2c77623d375ac7952b2a86f641e3ec"} Oct 14 08:13:54 crc kubenswrapper[4870]: I1014 08:13:54.322584 4870 scope.go:117] "RemoveContainer" containerID="dca50334a8cd2d1438da606d2f43eb3a283c73f0855c6f8bb11c380e552e3268" Oct 14 08:15:00 crc kubenswrapper[4870]: I1014 08:15:00.168051 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29340495-ldzr5"] Oct 14 08:15:00 crc kubenswrapper[4870]: E1014 08:15:00.169317 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e52e7047-2b96-43a8-a320-5a0768e75333" containerName="extract-content" Oct 14 08:15:00 crc kubenswrapper[4870]: I1014 08:15:00.169344 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="e52e7047-2b96-43a8-a320-5a0768e75333" containerName="extract-content" Oct 14 08:15:00 crc kubenswrapper[4870]: E1014 08:15:00.169376 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e52e7047-2b96-43a8-a320-5a0768e75333" containerName="registry-server" Oct 14 08:15:00 crc kubenswrapper[4870]: I1014 08:15:00.169394 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="e52e7047-2b96-43a8-a320-5a0768e75333" containerName="registry-server" Oct 14 08:15:00 crc kubenswrapper[4870]: E1014 08:15:00.169488 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e52e7047-2b96-43a8-a320-5a0768e75333" containerName="extract-utilities" Oct 14 08:15:00 crc kubenswrapper[4870]: I1014 08:15:00.169508 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="e52e7047-2b96-43a8-a320-5a0768e75333" containerName="extract-utilities" Oct 14 08:15:00 crc kubenswrapper[4870]: I1014 08:15:00.169827 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="e52e7047-2b96-43a8-a320-5a0768e75333" containerName="registry-server" Oct 14 08:15:00 crc kubenswrapper[4870]: I1014 08:15:00.170628 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29340495-ldzr5" Oct 14 08:15:00 crc kubenswrapper[4870]: I1014 08:15:00.179137 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29340495-ldzr5"] Oct 14 08:15:00 crc kubenswrapper[4870]: I1014 08:15:00.199615 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 14 08:15:00 crc kubenswrapper[4870]: I1014 08:15:00.199685 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 14 08:15:00 crc kubenswrapper[4870]: I1014 08:15:00.306546 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/26e2587a-73e2-48c4-9330-09d968b5a9c4-secret-volume\") pod \"collect-profiles-29340495-ldzr5\" (UID: \"26e2587a-73e2-48c4-9330-09d968b5a9c4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340495-ldzr5" Oct 14 08:15:00 crc kubenswrapper[4870]: I1014 08:15:00.307135 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rtwrc\" (UniqueName: \"kubernetes.io/projected/26e2587a-73e2-48c4-9330-09d968b5a9c4-kube-api-access-rtwrc\") pod \"collect-profiles-29340495-ldzr5\" (UID: \"26e2587a-73e2-48c4-9330-09d968b5a9c4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340495-ldzr5" Oct 14 08:15:00 crc kubenswrapper[4870]: I1014 08:15:00.307928 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/26e2587a-73e2-48c4-9330-09d968b5a9c4-config-volume\") pod \"collect-profiles-29340495-ldzr5\" (UID: \"26e2587a-73e2-48c4-9330-09d968b5a9c4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340495-ldzr5" Oct 14 08:15:00 crc kubenswrapper[4870]: I1014 08:15:00.409714 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/26e2587a-73e2-48c4-9330-09d968b5a9c4-secret-volume\") pod \"collect-profiles-29340495-ldzr5\" (UID: \"26e2587a-73e2-48c4-9330-09d968b5a9c4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340495-ldzr5" Oct 14 08:15:00 crc kubenswrapper[4870]: I1014 08:15:00.409802 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rtwrc\" (UniqueName: \"kubernetes.io/projected/26e2587a-73e2-48c4-9330-09d968b5a9c4-kube-api-access-rtwrc\") pod \"collect-profiles-29340495-ldzr5\" (UID: \"26e2587a-73e2-48c4-9330-09d968b5a9c4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340495-ldzr5" Oct 14 08:15:00 crc kubenswrapper[4870]: I1014 08:15:00.409852 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/26e2587a-73e2-48c4-9330-09d968b5a9c4-config-volume\") pod \"collect-profiles-29340495-ldzr5\" (UID: \"26e2587a-73e2-48c4-9330-09d968b5a9c4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340495-ldzr5" Oct 14 08:15:00 crc kubenswrapper[4870]: I1014 08:15:00.411079 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/26e2587a-73e2-48c4-9330-09d968b5a9c4-config-volume\") pod \"collect-profiles-29340495-ldzr5\" (UID: \"26e2587a-73e2-48c4-9330-09d968b5a9c4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340495-ldzr5" Oct 14 08:15:00 crc kubenswrapper[4870]: I1014 08:15:00.421226 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/26e2587a-73e2-48c4-9330-09d968b5a9c4-secret-volume\") pod \"collect-profiles-29340495-ldzr5\" (UID: \"26e2587a-73e2-48c4-9330-09d968b5a9c4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340495-ldzr5" Oct 14 08:15:00 crc kubenswrapper[4870]: I1014 08:15:00.432373 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rtwrc\" (UniqueName: \"kubernetes.io/projected/26e2587a-73e2-48c4-9330-09d968b5a9c4-kube-api-access-rtwrc\") pod \"collect-profiles-29340495-ldzr5\" (UID: \"26e2587a-73e2-48c4-9330-09d968b5a9c4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340495-ldzr5" Oct 14 08:15:00 crc kubenswrapper[4870]: I1014 08:15:00.530559 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29340495-ldzr5" Oct 14 08:15:01 crc kubenswrapper[4870]: I1014 08:15:01.061746 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29340495-ldzr5"] Oct 14 08:15:02 crc kubenswrapper[4870]: I1014 08:15:02.021587 4870 generic.go:334] "Generic (PLEG): container finished" podID="26e2587a-73e2-48c4-9330-09d968b5a9c4" containerID="81cd4308735a167d73e072e83aa982ef464c8eb82a7354700badc7d2297f1f09" exitCode=0 Oct 14 08:15:02 crc kubenswrapper[4870]: I1014 08:15:02.021674 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29340495-ldzr5" event={"ID":"26e2587a-73e2-48c4-9330-09d968b5a9c4","Type":"ContainerDied","Data":"81cd4308735a167d73e072e83aa982ef464c8eb82a7354700badc7d2297f1f09"} Oct 14 08:15:02 crc kubenswrapper[4870]: I1014 08:15:02.023724 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29340495-ldzr5" event={"ID":"26e2587a-73e2-48c4-9330-09d968b5a9c4","Type":"ContainerStarted","Data":"c1bfd2893c20a74715bdbf894975019925cf1e9a7696e46a9afc2beda987a6c9"} Oct 14 08:15:03 crc kubenswrapper[4870]: I1014 08:15:03.434985 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29340495-ldzr5" Oct 14 08:15:03 crc kubenswrapper[4870]: I1014 08:15:03.565491 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rtwrc\" (UniqueName: \"kubernetes.io/projected/26e2587a-73e2-48c4-9330-09d968b5a9c4-kube-api-access-rtwrc\") pod \"26e2587a-73e2-48c4-9330-09d968b5a9c4\" (UID: \"26e2587a-73e2-48c4-9330-09d968b5a9c4\") " Oct 14 08:15:03 crc kubenswrapper[4870]: I1014 08:15:03.565945 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/26e2587a-73e2-48c4-9330-09d968b5a9c4-secret-volume\") pod \"26e2587a-73e2-48c4-9330-09d968b5a9c4\" (UID: \"26e2587a-73e2-48c4-9330-09d968b5a9c4\") " Oct 14 08:15:03 crc kubenswrapper[4870]: I1014 08:15:03.566210 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/26e2587a-73e2-48c4-9330-09d968b5a9c4-config-volume\") pod \"26e2587a-73e2-48c4-9330-09d968b5a9c4\" (UID: \"26e2587a-73e2-48c4-9330-09d968b5a9c4\") " Oct 14 08:15:03 crc kubenswrapper[4870]: I1014 08:15:03.566957 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/26e2587a-73e2-48c4-9330-09d968b5a9c4-config-volume" (OuterVolumeSpecName: "config-volume") pod "26e2587a-73e2-48c4-9330-09d968b5a9c4" (UID: "26e2587a-73e2-48c4-9330-09d968b5a9c4"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:15:03 crc kubenswrapper[4870]: I1014 08:15:03.567316 4870 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/26e2587a-73e2-48c4-9330-09d968b5a9c4-config-volume\") on node \"crc\" DevicePath \"\"" Oct 14 08:15:03 crc kubenswrapper[4870]: I1014 08:15:03.572482 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/26e2587a-73e2-48c4-9330-09d968b5a9c4-kube-api-access-rtwrc" (OuterVolumeSpecName: "kube-api-access-rtwrc") pod "26e2587a-73e2-48c4-9330-09d968b5a9c4" (UID: "26e2587a-73e2-48c4-9330-09d968b5a9c4"). InnerVolumeSpecName "kube-api-access-rtwrc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:15:03 crc kubenswrapper[4870]: I1014 08:15:03.572772 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/26e2587a-73e2-48c4-9330-09d968b5a9c4-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "26e2587a-73e2-48c4-9330-09d968b5a9c4" (UID: "26e2587a-73e2-48c4-9330-09d968b5a9c4"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:15:03 crc kubenswrapper[4870]: I1014 08:15:03.669584 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rtwrc\" (UniqueName: \"kubernetes.io/projected/26e2587a-73e2-48c4-9330-09d968b5a9c4-kube-api-access-rtwrc\") on node \"crc\" DevicePath \"\"" Oct 14 08:15:03 crc kubenswrapper[4870]: I1014 08:15:03.669638 4870 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/26e2587a-73e2-48c4-9330-09d968b5a9c4-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 14 08:15:04 crc kubenswrapper[4870]: I1014 08:15:04.046349 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29340495-ldzr5" event={"ID":"26e2587a-73e2-48c4-9330-09d968b5a9c4","Type":"ContainerDied","Data":"c1bfd2893c20a74715bdbf894975019925cf1e9a7696e46a9afc2beda987a6c9"} Oct 14 08:15:04 crc kubenswrapper[4870]: I1014 08:15:04.046423 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29340495-ldzr5" Oct 14 08:15:04 crc kubenswrapper[4870]: I1014 08:15:04.046494 4870 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c1bfd2893c20a74715bdbf894975019925cf1e9a7696e46a9afc2beda987a6c9" Oct 14 08:15:04 crc kubenswrapper[4870]: I1014 08:15:04.537629 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29340450-gx8ch"] Oct 14 08:15:04 crc kubenswrapper[4870]: I1014 08:15:04.553585 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29340450-gx8ch"] Oct 14 08:15:05 crc kubenswrapper[4870]: I1014 08:15:05.054703 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="767a62c8-6040-4095-8a79-dbaa9401940d" path="/var/lib/kubelet/pods/767a62c8-6040-4095-8a79-dbaa9401940d/volumes" Oct 14 08:15:13 crc kubenswrapper[4870]: I1014 08:15:13.529365 4870 scope.go:117] "RemoveContainer" containerID="4b8393d462d3cba7296c43eb753cc6a1d10b653d0f087f8b55fd1cf7c9573ba5" Oct 14 08:16:23 crc kubenswrapper[4870]: I1014 08:16:23.951479 4870 patch_prober.go:28] interesting pod/machine-config-daemon-7tvc7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 08:16:23 crc kubenswrapper[4870]: I1014 08:16:23.952366 4870 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 08:16:48 crc kubenswrapper[4870]: I1014 08:16:48.251496 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-68cp5"] Oct 14 08:16:48 crc kubenswrapper[4870]: E1014 08:16:48.256332 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26e2587a-73e2-48c4-9330-09d968b5a9c4" containerName="collect-profiles" Oct 14 08:16:48 crc kubenswrapper[4870]: I1014 08:16:48.256510 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="26e2587a-73e2-48c4-9330-09d968b5a9c4" containerName="collect-profiles" Oct 14 08:16:48 crc kubenswrapper[4870]: I1014 08:16:48.256890 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="26e2587a-73e2-48c4-9330-09d968b5a9c4" containerName="collect-profiles" Oct 14 08:16:48 crc kubenswrapper[4870]: I1014 08:16:48.258402 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-68cp5" Oct 14 08:16:48 crc kubenswrapper[4870]: I1014 08:16:48.267532 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-68cp5"] Oct 14 08:16:48 crc kubenswrapper[4870]: I1014 08:16:48.393184 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rptfm\" (UniqueName: \"kubernetes.io/projected/1a841b84-a6d4-4c9e-8a08-e4ec07ecb09b-kube-api-access-rptfm\") pod \"community-operators-68cp5\" (UID: \"1a841b84-a6d4-4c9e-8a08-e4ec07ecb09b\") " pod="openshift-marketplace/community-operators-68cp5" Oct 14 08:16:48 crc kubenswrapper[4870]: I1014 08:16:48.393635 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1a841b84-a6d4-4c9e-8a08-e4ec07ecb09b-utilities\") pod \"community-operators-68cp5\" (UID: \"1a841b84-a6d4-4c9e-8a08-e4ec07ecb09b\") " pod="openshift-marketplace/community-operators-68cp5" Oct 14 08:16:48 crc kubenswrapper[4870]: I1014 08:16:48.393717 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1a841b84-a6d4-4c9e-8a08-e4ec07ecb09b-catalog-content\") pod \"community-operators-68cp5\" (UID: \"1a841b84-a6d4-4c9e-8a08-e4ec07ecb09b\") " pod="openshift-marketplace/community-operators-68cp5" Oct 14 08:16:48 crc kubenswrapper[4870]: I1014 08:16:48.495252 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1a841b84-a6d4-4c9e-8a08-e4ec07ecb09b-catalog-content\") pod \"community-operators-68cp5\" (UID: \"1a841b84-a6d4-4c9e-8a08-e4ec07ecb09b\") " pod="openshift-marketplace/community-operators-68cp5" Oct 14 08:16:48 crc kubenswrapper[4870]: I1014 08:16:48.495517 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rptfm\" (UniqueName: \"kubernetes.io/projected/1a841b84-a6d4-4c9e-8a08-e4ec07ecb09b-kube-api-access-rptfm\") pod \"community-operators-68cp5\" (UID: \"1a841b84-a6d4-4c9e-8a08-e4ec07ecb09b\") " pod="openshift-marketplace/community-operators-68cp5" Oct 14 08:16:48 crc kubenswrapper[4870]: I1014 08:16:48.495579 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1a841b84-a6d4-4c9e-8a08-e4ec07ecb09b-utilities\") pod \"community-operators-68cp5\" (UID: \"1a841b84-a6d4-4c9e-8a08-e4ec07ecb09b\") " pod="openshift-marketplace/community-operators-68cp5" Oct 14 08:16:48 crc kubenswrapper[4870]: I1014 08:16:48.496312 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1a841b84-a6d4-4c9e-8a08-e4ec07ecb09b-catalog-content\") pod \"community-operators-68cp5\" (UID: \"1a841b84-a6d4-4c9e-8a08-e4ec07ecb09b\") " pod="openshift-marketplace/community-operators-68cp5" Oct 14 08:16:48 crc kubenswrapper[4870]: I1014 08:16:48.496571 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1a841b84-a6d4-4c9e-8a08-e4ec07ecb09b-utilities\") pod \"community-operators-68cp5\" (UID: \"1a841b84-a6d4-4c9e-8a08-e4ec07ecb09b\") " pod="openshift-marketplace/community-operators-68cp5" Oct 14 08:16:48 crc kubenswrapper[4870]: I1014 08:16:48.530873 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rptfm\" (UniqueName: \"kubernetes.io/projected/1a841b84-a6d4-4c9e-8a08-e4ec07ecb09b-kube-api-access-rptfm\") pod \"community-operators-68cp5\" (UID: \"1a841b84-a6d4-4c9e-8a08-e4ec07ecb09b\") " pod="openshift-marketplace/community-operators-68cp5" Oct 14 08:16:48 crc kubenswrapper[4870]: I1014 08:16:48.601904 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-68cp5" Oct 14 08:16:49 crc kubenswrapper[4870]: I1014 08:16:49.126308 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-68cp5"] Oct 14 08:16:50 crc kubenswrapper[4870]: I1014 08:16:50.109753 4870 generic.go:334] "Generic (PLEG): container finished" podID="1a841b84-a6d4-4c9e-8a08-e4ec07ecb09b" containerID="46094e6d2286eb2e15c27c779888d02623b0d8079a74d9212af712abfef02823" exitCode=0 Oct 14 08:16:50 crc kubenswrapper[4870]: I1014 08:16:50.109961 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-68cp5" event={"ID":"1a841b84-a6d4-4c9e-8a08-e4ec07ecb09b","Type":"ContainerDied","Data":"46094e6d2286eb2e15c27c779888d02623b0d8079a74d9212af712abfef02823"} Oct 14 08:16:50 crc kubenswrapper[4870]: I1014 08:16:50.110307 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-68cp5" event={"ID":"1a841b84-a6d4-4c9e-8a08-e4ec07ecb09b","Type":"ContainerStarted","Data":"c74f18bcfc06d332a603f13a141f205b952baa1761652f99ccc6f1de9a6f0e5e"} Oct 14 08:16:50 crc kubenswrapper[4870]: I1014 08:16:50.113310 4870 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 14 08:16:51 crc kubenswrapper[4870]: I1014 08:16:51.118710 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-68cp5" event={"ID":"1a841b84-a6d4-4c9e-8a08-e4ec07ecb09b","Type":"ContainerStarted","Data":"7e962ee111cd201712f6a241c9d5526345083b6ef821c2ab8bad06e3138f395a"} Oct 14 08:16:52 crc kubenswrapper[4870]: I1014 08:16:52.131240 4870 generic.go:334] "Generic (PLEG): container finished" podID="1a841b84-a6d4-4c9e-8a08-e4ec07ecb09b" containerID="7e962ee111cd201712f6a241c9d5526345083b6ef821c2ab8bad06e3138f395a" exitCode=0 Oct 14 08:16:52 crc kubenswrapper[4870]: I1014 08:16:52.131297 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-68cp5" event={"ID":"1a841b84-a6d4-4c9e-8a08-e4ec07ecb09b","Type":"ContainerDied","Data":"7e962ee111cd201712f6a241c9d5526345083b6ef821c2ab8bad06e3138f395a"} Oct 14 08:16:53 crc kubenswrapper[4870]: I1014 08:16:53.143309 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-68cp5" event={"ID":"1a841b84-a6d4-4c9e-8a08-e4ec07ecb09b","Type":"ContainerStarted","Data":"358bd4222a8d72e347c4946cae08b7195365a8edb98a86d0939421f6ddd501e1"} Oct 14 08:16:53 crc kubenswrapper[4870]: I1014 08:16:53.951533 4870 patch_prober.go:28] interesting pod/machine-config-daemon-7tvc7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 08:16:53 crc kubenswrapper[4870]: I1014 08:16:53.951635 4870 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 08:16:58 crc kubenswrapper[4870]: I1014 08:16:58.603225 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-68cp5" Oct 14 08:16:58 crc kubenswrapper[4870]: I1014 08:16:58.604360 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-68cp5" Oct 14 08:16:58 crc kubenswrapper[4870]: I1014 08:16:58.686193 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-68cp5" Oct 14 08:16:58 crc kubenswrapper[4870]: I1014 08:16:58.713399 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-68cp5" podStartSLOduration=8.161834165 podStartE2EDuration="10.713368456s" podCreationTimestamp="2025-10-14 08:16:48 +0000 UTC" firstStartedPulling="2025-10-14 08:16:50.112683433 +0000 UTC m=+4545.810043834" lastFinishedPulling="2025-10-14 08:16:52.664217714 +0000 UTC m=+4548.361578125" observedRunningTime="2025-10-14 08:16:53.178485753 +0000 UTC m=+4548.875846164" watchObservedRunningTime="2025-10-14 08:16:58.713368456 +0000 UTC m=+4554.410728867" Oct 14 08:16:59 crc kubenswrapper[4870]: I1014 08:16:59.304618 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-68cp5" Oct 14 08:16:59 crc kubenswrapper[4870]: I1014 08:16:59.358963 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-68cp5"] Oct 14 08:17:01 crc kubenswrapper[4870]: I1014 08:17:01.266058 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-68cp5" podUID="1a841b84-a6d4-4c9e-8a08-e4ec07ecb09b" containerName="registry-server" containerID="cri-o://358bd4222a8d72e347c4946cae08b7195365a8edb98a86d0939421f6ddd501e1" gracePeriod=2 Oct 14 08:17:01 crc kubenswrapper[4870]: I1014 08:17:01.834417 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-68cp5" Oct 14 08:17:01 crc kubenswrapper[4870]: I1014 08:17:01.907015 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1a841b84-a6d4-4c9e-8a08-e4ec07ecb09b-catalog-content\") pod \"1a841b84-a6d4-4c9e-8a08-e4ec07ecb09b\" (UID: \"1a841b84-a6d4-4c9e-8a08-e4ec07ecb09b\") " Oct 14 08:17:01 crc kubenswrapper[4870]: I1014 08:17:01.907097 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rptfm\" (UniqueName: \"kubernetes.io/projected/1a841b84-a6d4-4c9e-8a08-e4ec07ecb09b-kube-api-access-rptfm\") pod \"1a841b84-a6d4-4c9e-8a08-e4ec07ecb09b\" (UID: \"1a841b84-a6d4-4c9e-8a08-e4ec07ecb09b\") " Oct 14 08:17:01 crc kubenswrapper[4870]: I1014 08:17:01.907167 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1a841b84-a6d4-4c9e-8a08-e4ec07ecb09b-utilities\") pod \"1a841b84-a6d4-4c9e-8a08-e4ec07ecb09b\" (UID: \"1a841b84-a6d4-4c9e-8a08-e4ec07ecb09b\") " Oct 14 08:17:01 crc kubenswrapper[4870]: I1014 08:17:01.908351 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1a841b84-a6d4-4c9e-8a08-e4ec07ecb09b-utilities" (OuterVolumeSpecName: "utilities") pod "1a841b84-a6d4-4c9e-8a08-e4ec07ecb09b" (UID: "1a841b84-a6d4-4c9e-8a08-e4ec07ecb09b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:17:01 crc kubenswrapper[4870]: I1014 08:17:01.916016 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1a841b84-a6d4-4c9e-8a08-e4ec07ecb09b-kube-api-access-rptfm" (OuterVolumeSpecName: "kube-api-access-rptfm") pod "1a841b84-a6d4-4c9e-8a08-e4ec07ecb09b" (UID: "1a841b84-a6d4-4c9e-8a08-e4ec07ecb09b"). InnerVolumeSpecName "kube-api-access-rptfm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:17:01 crc kubenswrapper[4870]: I1014 08:17:01.954273 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1a841b84-a6d4-4c9e-8a08-e4ec07ecb09b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1a841b84-a6d4-4c9e-8a08-e4ec07ecb09b" (UID: "1a841b84-a6d4-4c9e-8a08-e4ec07ecb09b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:17:02 crc kubenswrapper[4870]: I1014 08:17:02.009276 4870 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1a841b84-a6d4-4c9e-8a08-e4ec07ecb09b-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 08:17:02 crc kubenswrapper[4870]: I1014 08:17:02.009338 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rptfm\" (UniqueName: \"kubernetes.io/projected/1a841b84-a6d4-4c9e-8a08-e4ec07ecb09b-kube-api-access-rptfm\") on node \"crc\" DevicePath \"\"" Oct 14 08:17:02 crc kubenswrapper[4870]: I1014 08:17:02.009360 4870 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1a841b84-a6d4-4c9e-8a08-e4ec07ecb09b-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 08:17:02 crc kubenswrapper[4870]: I1014 08:17:02.283408 4870 generic.go:334] "Generic (PLEG): container finished" podID="1a841b84-a6d4-4c9e-8a08-e4ec07ecb09b" containerID="358bd4222a8d72e347c4946cae08b7195365a8edb98a86d0939421f6ddd501e1" exitCode=0 Oct 14 08:17:02 crc kubenswrapper[4870]: I1014 08:17:02.283548 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-68cp5" Oct 14 08:17:02 crc kubenswrapper[4870]: I1014 08:17:02.283533 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-68cp5" event={"ID":"1a841b84-a6d4-4c9e-8a08-e4ec07ecb09b","Type":"ContainerDied","Data":"358bd4222a8d72e347c4946cae08b7195365a8edb98a86d0939421f6ddd501e1"} Oct 14 08:17:02 crc kubenswrapper[4870]: I1014 08:17:02.285355 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-68cp5" event={"ID":"1a841b84-a6d4-4c9e-8a08-e4ec07ecb09b","Type":"ContainerDied","Data":"c74f18bcfc06d332a603f13a141f205b952baa1761652f99ccc6f1de9a6f0e5e"} Oct 14 08:17:02 crc kubenswrapper[4870]: I1014 08:17:02.285401 4870 scope.go:117] "RemoveContainer" containerID="358bd4222a8d72e347c4946cae08b7195365a8edb98a86d0939421f6ddd501e1" Oct 14 08:17:02 crc kubenswrapper[4870]: I1014 08:17:02.313391 4870 scope.go:117] "RemoveContainer" containerID="7e962ee111cd201712f6a241c9d5526345083b6ef821c2ab8bad06e3138f395a" Oct 14 08:17:02 crc kubenswrapper[4870]: I1014 08:17:02.344644 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-68cp5"] Oct 14 08:17:02 crc kubenswrapper[4870]: I1014 08:17:02.344799 4870 scope.go:117] "RemoveContainer" containerID="46094e6d2286eb2e15c27c779888d02623b0d8079a74d9212af712abfef02823" Oct 14 08:17:02 crc kubenswrapper[4870]: I1014 08:17:02.354394 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-68cp5"] Oct 14 08:17:02 crc kubenswrapper[4870]: I1014 08:17:02.368225 4870 scope.go:117] "RemoveContainer" containerID="358bd4222a8d72e347c4946cae08b7195365a8edb98a86d0939421f6ddd501e1" Oct 14 08:17:02 crc kubenswrapper[4870]: E1014 08:17:02.368685 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"358bd4222a8d72e347c4946cae08b7195365a8edb98a86d0939421f6ddd501e1\": container with ID starting with 358bd4222a8d72e347c4946cae08b7195365a8edb98a86d0939421f6ddd501e1 not found: ID does not exist" containerID="358bd4222a8d72e347c4946cae08b7195365a8edb98a86d0939421f6ddd501e1" Oct 14 08:17:02 crc kubenswrapper[4870]: I1014 08:17:02.368751 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"358bd4222a8d72e347c4946cae08b7195365a8edb98a86d0939421f6ddd501e1"} err="failed to get container status \"358bd4222a8d72e347c4946cae08b7195365a8edb98a86d0939421f6ddd501e1\": rpc error: code = NotFound desc = could not find container \"358bd4222a8d72e347c4946cae08b7195365a8edb98a86d0939421f6ddd501e1\": container with ID starting with 358bd4222a8d72e347c4946cae08b7195365a8edb98a86d0939421f6ddd501e1 not found: ID does not exist" Oct 14 08:17:02 crc kubenswrapper[4870]: I1014 08:17:02.368794 4870 scope.go:117] "RemoveContainer" containerID="7e962ee111cd201712f6a241c9d5526345083b6ef821c2ab8bad06e3138f395a" Oct 14 08:17:02 crc kubenswrapper[4870]: E1014 08:17:02.371024 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7e962ee111cd201712f6a241c9d5526345083b6ef821c2ab8bad06e3138f395a\": container with ID starting with 7e962ee111cd201712f6a241c9d5526345083b6ef821c2ab8bad06e3138f395a not found: ID does not exist" containerID="7e962ee111cd201712f6a241c9d5526345083b6ef821c2ab8bad06e3138f395a" Oct 14 08:17:02 crc kubenswrapper[4870]: I1014 08:17:02.371054 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7e962ee111cd201712f6a241c9d5526345083b6ef821c2ab8bad06e3138f395a"} err="failed to get container status \"7e962ee111cd201712f6a241c9d5526345083b6ef821c2ab8bad06e3138f395a\": rpc error: code = NotFound desc = could not find container \"7e962ee111cd201712f6a241c9d5526345083b6ef821c2ab8bad06e3138f395a\": container with ID starting with 7e962ee111cd201712f6a241c9d5526345083b6ef821c2ab8bad06e3138f395a not found: ID does not exist" Oct 14 08:17:02 crc kubenswrapper[4870]: I1014 08:17:02.371077 4870 scope.go:117] "RemoveContainer" containerID="46094e6d2286eb2e15c27c779888d02623b0d8079a74d9212af712abfef02823" Oct 14 08:17:02 crc kubenswrapper[4870]: E1014 08:17:02.371639 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"46094e6d2286eb2e15c27c779888d02623b0d8079a74d9212af712abfef02823\": container with ID starting with 46094e6d2286eb2e15c27c779888d02623b0d8079a74d9212af712abfef02823 not found: ID does not exist" containerID="46094e6d2286eb2e15c27c779888d02623b0d8079a74d9212af712abfef02823" Oct 14 08:17:02 crc kubenswrapper[4870]: I1014 08:17:02.371702 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"46094e6d2286eb2e15c27c779888d02623b0d8079a74d9212af712abfef02823"} err="failed to get container status \"46094e6d2286eb2e15c27c779888d02623b0d8079a74d9212af712abfef02823\": rpc error: code = NotFound desc = could not find container \"46094e6d2286eb2e15c27c779888d02623b0d8079a74d9212af712abfef02823\": container with ID starting with 46094e6d2286eb2e15c27c779888d02623b0d8079a74d9212af712abfef02823 not found: ID does not exist" Oct 14 08:17:03 crc kubenswrapper[4870]: I1014 08:17:03.047128 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1a841b84-a6d4-4c9e-8a08-e4ec07ecb09b" path="/var/lib/kubelet/pods/1a841b84-a6d4-4c9e-8a08-e4ec07ecb09b/volumes" Oct 14 08:17:23 crc kubenswrapper[4870]: I1014 08:17:23.951210 4870 patch_prober.go:28] interesting pod/machine-config-daemon-7tvc7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 08:17:23 crc kubenswrapper[4870]: I1014 08:17:23.952116 4870 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 08:17:23 crc kubenswrapper[4870]: I1014 08:17:23.952197 4870 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" Oct 14 08:17:23 crc kubenswrapper[4870]: I1014 08:17:23.953146 4870 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"cb5962db9597015a8eaf724ffe6622e32b2c77623d375ac7952b2a86f641e3ec"} pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 14 08:17:23 crc kubenswrapper[4870]: I1014 08:17:23.953268 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerName="machine-config-daemon" containerID="cri-o://cb5962db9597015a8eaf724ffe6622e32b2c77623d375ac7952b2a86f641e3ec" gracePeriod=600 Oct 14 08:17:24 crc kubenswrapper[4870]: E1014 08:17:24.090284 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 08:17:24 crc kubenswrapper[4870]: I1014 08:17:24.519512 4870 generic.go:334] "Generic (PLEG): container finished" podID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerID="cb5962db9597015a8eaf724ffe6622e32b2c77623d375ac7952b2a86f641e3ec" exitCode=0 Oct 14 08:17:24 crc kubenswrapper[4870]: I1014 08:17:24.520124 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" event={"ID":"874a5e45-dffd-4d17-b609-b7d3ed2eab07","Type":"ContainerDied","Data":"cb5962db9597015a8eaf724ffe6622e32b2c77623d375ac7952b2a86f641e3ec"} Oct 14 08:17:24 crc kubenswrapper[4870]: I1014 08:17:24.520207 4870 scope.go:117] "RemoveContainer" containerID="52e4290890c8147a0b5888fc1327a813d1c142f382c705fe446f84040afd3bef" Oct 14 08:17:24 crc kubenswrapper[4870]: I1014 08:17:24.521054 4870 scope.go:117] "RemoveContainer" containerID="cb5962db9597015a8eaf724ffe6622e32b2c77623d375ac7952b2a86f641e3ec" Oct 14 08:17:24 crc kubenswrapper[4870]: E1014 08:17:24.521588 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 08:17:37 crc kubenswrapper[4870]: I1014 08:17:37.034327 4870 scope.go:117] "RemoveContainer" containerID="cb5962db9597015a8eaf724ffe6622e32b2c77623d375ac7952b2a86f641e3ec" Oct 14 08:17:37 crc kubenswrapper[4870]: E1014 08:17:37.038124 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 08:17:51 crc kubenswrapper[4870]: I1014 08:17:51.035234 4870 scope.go:117] "RemoveContainer" containerID="cb5962db9597015a8eaf724ffe6622e32b2c77623d375ac7952b2a86f641e3ec" Oct 14 08:17:51 crc kubenswrapper[4870]: E1014 08:17:51.036676 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 08:18:02 crc kubenswrapper[4870]: I1014 08:18:02.034394 4870 scope.go:117] "RemoveContainer" containerID="cb5962db9597015a8eaf724ffe6622e32b2c77623d375ac7952b2a86f641e3ec" Oct 14 08:18:02 crc kubenswrapper[4870]: E1014 08:18:02.035270 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 08:18:02 crc kubenswrapper[4870]: I1014 08:18:02.341524 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-krvcx"] Oct 14 08:18:02 crc kubenswrapper[4870]: E1014 08:18:02.341972 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a841b84-a6d4-4c9e-8a08-e4ec07ecb09b" containerName="registry-server" Oct 14 08:18:02 crc kubenswrapper[4870]: I1014 08:18:02.341993 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a841b84-a6d4-4c9e-8a08-e4ec07ecb09b" containerName="registry-server" Oct 14 08:18:02 crc kubenswrapper[4870]: E1014 08:18:02.342064 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a841b84-a6d4-4c9e-8a08-e4ec07ecb09b" containerName="extract-utilities" Oct 14 08:18:02 crc kubenswrapper[4870]: I1014 08:18:02.342081 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a841b84-a6d4-4c9e-8a08-e4ec07ecb09b" containerName="extract-utilities" Oct 14 08:18:02 crc kubenswrapper[4870]: E1014 08:18:02.342120 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a841b84-a6d4-4c9e-8a08-e4ec07ecb09b" containerName="extract-content" Oct 14 08:18:02 crc kubenswrapper[4870]: I1014 08:18:02.342136 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a841b84-a6d4-4c9e-8a08-e4ec07ecb09b" containerName="extract-content" Oct 14 08:18:02 crc kubenswrapper[4870]: I1014 08:18:02.342528 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="1a841b84-a6d4-4c9e-8a08-e4ec07ecb09b" containerName="registry-server" Oct 14 08:18:02 crc kubenswrapper[4870]: I1014 08:18:02.344400 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-krvcx" Oct 14 08:18:02 crc kubenswrapper[4870]: I1014 08:18:02.376860 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-krvcx"] Oct 14 08:18:02 crc kubenswrapper[4870]: I1014 08:18:02.468729 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2fbdb997-d4d3-4737-9858-ed6a861e475b-utilities\") pod \"redhat-marketplace-krvcx\" (UID: \"2fbdb997-d4d3-4737-9858-ed6a861e475b\") " pod="openshift-marketplace/redhat-marketplace-krvcx" Oct 14 08:18:02 crc kubenswrapper[4870]: I1014 08:18:02.468820 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2fbdb997-d4d3-4737-9858-ed6a861e475b-catalog-content\") pod \"redhat-marketplace-krvcx\" (UID: \"2fbdb997-d4d3-4737-9858-ed6a861e475b\") " pod="openshift-marketplace/redhat-marketplace-krvcx" Oct 14 08:18:02 crc kubenswrapper[4870]: I1014 08:18:02.468863 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xhdjv\" (UniqueName: \"kubernetes.io/projected/2fbdb997-d4d3-4737-9858-ed6a861e475b-kube-api-access-xhdjv\") pod \"redhat-marketplace-krvcx\" (UID: \"2fbdb997-d4d3-4737-9858-ed6a861e475b\") " pod="openshift-marketplace/redhat-marketplace-krvcx" Oct 14 08:18:02 crc kubenswrapper[4870]: I1014 08:18:02.570324 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2fbdb997-d4d3-4737-9858-ed6a861e475b-utilities\") pod \"redhat-marketplace-krvcx\" (UID: \"2fbdb997-d4d3-4737-9858-ed6a861e475b\") " pod="openshift-marketplace/redhat-marketplace-krvcx" Oct 14 08:18:02 crc kubenswrapper[4870]: I1014 08:18:02.570412 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2fbdb997-d4d3-4737-9858-ed6a861e475b-catalog-content\") pod \"redhat-marketplace-krvcx\" (UID: \"2fbdb997-d4d3-4737-9858-ed6a861e475b\") " pod="openshift-marketplace/redhat-marketplace-krvcx" Oct 14 08:18:02 crc kubenswrapper[4870]: I1014 08:18:02.570469 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xhdjv\" (UniqueName: \"kubernetes.io/projected/2fbdb997-d4d3-4737-9858-ed6a861e475b-kube-api-access-xhdjv\") pod \"redhat-marketplace-krvcx\" (UID: \"2fbdb997-d4d3-4737-9858-ed6a861e475b\") " pod="openshift-marketplace/redhat-marketplace-krvcx" Oct 14 08:18:02 crc kubenswrapper[4870]: I1014 08:18:02.571294 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2fbdb997-d4d3-4737-9858-ed6a861e475b-utilities\") pod \"redhat-marketplace-krvcx\" (UID: \"2fbdb997-d4d3-4737-9858-ed6a861e475b\") " pod="openshift-marketplace/redhat-marketplace-krvcx" Oct 14 08:18:02 crc kubenswrapper[4870]: I1014 08:18:02.571897 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2fbdb997-d4d3-4737-9858-ed6a861e475b-catalog-content\") pod \"redhat-marketplace-krvcx\" (UID: \"2fbdb997-d4d3-4737-9858-ed6a861e475b\") " pod="openshift-marketplace/redhat-marketplace-krvcx" Oct 14 08:18:02 crc kubenswrapper[4870]: I1014 08:18:02.591627 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xhdjv\" (UniqueName: \"kubernetes.io/projected/2fbdb997-d4d3-4737-9858-ed6a861e475b-kube-api-access-xhdjv\") pod \"redhat-marketplace-krvcx\" (UID: \"2fbdb997-d4d3-4737-9858-ed6a861e475b\") " pod="openshift-marketplace/redhat-marketplace-krvcx" Oct 14 08:18:02 crc kubenswrapper[4870]: I1014 08:18:02.707229 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-krvcx" Oct 14 08:18:03 crc kubenswrapper[4870]: I1014 08:18:03.166604 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-krvcx"] Oct 14 08:18:03 crc kubenswrapper[4870]: I1014 08:18:03.907511 4870 generic.go:334] "Generic (PLEG): container finished" podID="2fbdb997-d4d3-4737-9858-ed6a861e475b" containerID="00594f9c6ade109cf799a6fc59d7b23f5d8ea17c190d4bd32cc227656b9a1996" exitCode=0 Oct 14 08:18:03 crc kubenswrapper[4870]: I1014 08:18:03.907672 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-krvcx" event={"ID":"2fbdb997-d4d3-4737-9858-ed6a861e475b","Type":"ContainerDied","Data":"00594f9c6ade109cf799a6fc59d7b23f5d8ea17c190d4bd32cc227656b9a1996"} Oct 14 08:18:03 crc kubenswrapper[4870]: I1014 08:18:03.907852 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-krvcx" event={"ID":"2fbdb997-d4d3-4737-9858-ed6a861e475b","Type":"ContainerStarted","Data":"ebd35694a43350d1d8f4ab8f85ff1e0172a812a0a766f5aeb7163cd547789862"} Oct 14 08:18:04 crc kubenswrapper[4870]: I1014 08:18:04.920518 4870 generic.go:334] "Generic (PLEG): container finished" podID="2fbdb997-d4d3-4737-9858-ed6a861e475b" containerID="ca7c8eef02d1cc1e73f9bbcd3ba296bd53653e321e2c2de5b64a5baca4ecc9c6" exitCode=0 Oct 14 08:18:04 crc kubenswrapper[4870]: I1014 08:18:04.920632 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-krvcx" event={"ID":"2fbdb997-d4d3-4737-9858-ed6a861e475b","Type":"ContainerDied","Data":"ca7c8eef02d1cc1e73f9bbcd3ba296bd53653e321e2c2de5b64a5baca4ecc9c6"} Oct 14 08:18:05 crc kubenswrapper[4870]: I1014 08:18:05.934535 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-krvcx" event={"ID":"2fbdb997-d4d3-4737-9858-ed6a861e475b","Type":"ContainerStarted","Data":"da89f60aa80a2d52aec7b9ac240a98a6e1802b21cc0c8fbdc90eb1f6b0fce417"} Oct 14 08:18:05 crc kubenswrapper[4870]: I1014 08:18:05.960386 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-krvcx" podStartSLOduration=2.516730198 podStartE2EDuration="3.96035113s" podCreationTimestamp="2025-10-14 08:18:02 +0000 UTC" firstStartedPulling="2025-10-14 08:18:03.91049803 +0000 UTC m=+4619.607858431" lastFinishedPulling="2025-10-14 08:18:05.354118952 +0000 UTC m=+4621.051479363" observedRunningTime="2025-10-14 08:18:05.952214273 +0000 UTC m=+4621.649574684" watchObservedRunningTime="2025-10-14 08:18:05.96035113 +0000 UTC m=+4621.657711541" Oct 14 08:18:12 crc kubenswrapper[4870]: I1014 08:18:12.707840 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-krvcx" Oct 14 08:18:12 crc kubenswrapper[4870]: I1014 08:18:12.708957 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-krvcx" Oct 14 08:18:12 crc kubenswrapper[4870]: I1014 08:18:12.782574 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-krvcx" Oct 14 08:18:13 crc kubenswrapper[4870]: I1014 08:18:13.091634 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-krvcx" Oct 14 08:18:13 crc kubenswrapper[4870]: I1014 08:18:13.160187 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-krvcx"] Oct 14 08:18:15 crc kubenswrapper[4870]: I1014 08:18:15.043049 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-krvcx" podUID="2fbdb997-d4d3-4737-9858-ed6a861e475b" containerName="registry-server" containerID="cri-o://da89f60aa80a2d52aec7b9ac240a98a6e1802b21cc0c8fbdc90eb1f6b0fce417" gracePeriod=2 Oct 14 08:18:15 crc kubenswrapper[4870]: I1014 08:18:15.493938 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-krvcx" Oct 14 08:18:15 crc kubenswrapper[4870]: I1014 08:18:15.601284 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2fbdb997-d4d3-4737-9858-ed6a861e475b-catalog-content\") pod \"2fbdb997-d4d3-4737-9858-ed6a861e475b\" (UID: \"2fbdb997-d4d3-4737-9858-ed6a861e475b\") " Oct 14 08:18:15 crc kubenswrapper[4870]: I1014 08:18:15.601436 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xhdjv\" (UniqueName: \"kubernetes.io/projected/2fbdb997-d4d3-4737-9858-ed6a861e475b-kube-api-access-xhdjv\") pod \"2fbdb997-d4d3-4737-9858-ed6a861e475b\" (UID: \"2fbdb997-d4d3-4737-9858-ed6a861e475b\") " Oct 14 08:18:15 crc kubenswrapper[4870]: I1014 08:18:15.601515 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2fbdb997-d4d3-4737-9858-ed6a861e475b-utilities\") pod \"2fbdb997-d4d3-4737-9858-ed6a861e475b\" (UID: \"2fbdb997-d4d3-4737-9858-ed6a861e475b\") " Oct 14 08:18:15 crc kubenswrapper[4870]: I1014 08:18:15.603276 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2fbdb997-d4d3-4737-9858-ed6a861e475b-utilities" (OuterVolumeSpecName: "utilities") pod "2fbdb997-d4d3-4737-9858-ed6a861e475b" (UID: "2fbdb997-d4d3-4737-9858-ed6a861e475b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:18:15 crc kubenswrapper[4870]: I1014 08:18:15.612614 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2fbdb997-d4d3-4737-9858-ed6a861e475b-kube-api-access-xhdjv" (OuterVolumeSpecName: "kube-api-access-xhdjv") pod "2fbdb997-d4d3-4737-9858-ed6a861e475b" (UID: "2fbdb997-d4d3-4737-9858-ed6a861e475b"). InnerVolumeSpecName "kube-api-access-xhdjv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:18:15 crc kubenswrapper[4870]: I1014 08:18:15.629920 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2fbdb997-d4d3-4737-9858-ed6a861e475b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2fbdb997-d4d3-4737-9858-ed6a861e475b" (UID: "2fbdb997-d4d3-4737-9858-ed6a861e475b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:18:15 crc kubenswrapper[4870]: I1014 08:18:15.703690 4870 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2fbdb997-d4d3-4737-9858-ed6a861e475b-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 08:18:15 crc kubenswrapper[4870]: I1014 08:18:15.703793 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xhdjv\" (UniqueName: \"kubernetes.io/projected/2fbdb997-d4d3-4737-9858-ed6a861e475b-kube-api-access-xhdjv\") on node \"crc\" DevicePath \"\"" Oct 14 08:18:15 crc kubenswrapper[4870]: I1014 08:18:15.703819 4870 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2fbdb997-d4d3-4737-9858-ed6a861e475b-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 08:18:16 crc kubenswrapper[4870]: I1014 08:18:16.034495 4870 scope.go:117] "RemoveContainer" containerID="cb5962db9597015a8eaf724ffe6622e32b2c77623d375ac7952b2a86f641e3ec" Oct 14 08:18:16 crc kubenswrapper[4870]: E1014 08:18:16.035194 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 08:18:16 crc kubenswrapper[4870]: I1014 08:18:16.052707 4870 generic.go:334] "Generic (PLEG): container finished" podID="2fbdb997-d4d3-4737-9858-ed6a861e475b" containerID="da89f60aa80a2d52aec7b9ac240a98a6e1802b21cc0c8fbdc90eb1f6b0fce417" exitCode=0 Oct 14 08:18:16 crc kubenswrapper[4870]: I1014 08:18:16.052779 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-krvcx" event={"ID":"2fbdb997-d4d3-4737-9858-ed6a861e475b","Type":"ContainerDied","Data":"da89f60aa80a2d52aec7b9ac240a98a6e1802b21cc0c8fbdc90eb1f6b0fce417"} Oct 14 08:18:16 crc kubenswrapper[4870]: I1014 08:18:16.052828 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-krvcx" event={"ID":"2fbdb997-d4d3-4737-9858-ed6a861e475b","Type":"ContainerDied","Data":"ebd35694a43350d1d8f4ab8f85ff1e0172a812a0a766f5aeb7163cd547789862"} Oct 14 08:18:16 crc kubenswrapper[4870]: I1014 08:18:16.052866 4870 scope.go:117] "RemoveContainer" containerID="da89f60aa80a2d52aec7b9ac240a98a6e1802b21cc0c8fbdc90eb1f6b0fce417" Oct 14 08:18:16 crc kubenswrapper[4870]: I1014 08:18:16.053120 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-krvcx" Oct 14 08:18:16 crc kubenswrapper[4870]: I1014 08:18:16.091686 4870 scope.go:117] "RemoveContainer" containerID="ca7c8eef02d1cc1e73f9bbcd3ba296bd53653e321e2c2de5b64a5baca4ecc9c6" Oct 14 08:18:16 crc kubenswrapper[4870]: I1014 08:18:16.116354 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-krvcx"] Oct 14 08:18:16 crc kubenswrapper[4870]: I1014 08:18:16.136809 4870 scope.go:117] "RemoveContainer" containerID="00594f9c6ade109cf799a6fc59d7b23f5d8ea17c190d4bd32cc227656b9a1996" Oct 14 08:18:16 crc kubenswrapper[4870]: I1014 08:18:16.137132 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-krvcx"] Oct 14 08:18:16 crc kubenswrapper[4870]: I1014 08:18:16.172178 4870 scope.go:117] "RemoveContainer" containerID="da89f60aa80a2d52aec7b9ac240a98a6e1802b21cc0c8fbdc90eb1f6b0fce417" Oct 14 08:18:16 crc kubenswrapper[4870]: E1014 08:18:16.172989 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"da89f60aa80a2d52aec7b9ac240a98a6e1802b21cc0c8fbdc90eb1f6b0fce417\": container with ID starting with da89f60aa80a2d52aec7b9ac240a98a6e1802b21cc0c8fbdc90eb1f6b0fce417 not found: ID does not exist" containerID="da89f60aa80a2d52aec7b9ac240a98a6e1802b21cc0c8fbdc90eb1f6b0fce417" Oct 14 08:18:16 crc kubenswrapper[4870]: I1014 08:18:16.173134 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"da89f60aa80a2d52aec7b9ac240a98a6e1802b21cc0c8fbdc90eb1f6b0fce417"} err="failed to get container status \"da89f60aa80a2d52aec7b9ac240a98a6e1802b21cc0c8fbdc90eb1f6b0fce417\": rpc error: code = NotFound desc = could not find container \"da89f60aa80a2d52aec7b9ac240a98a6e1802b21cc0c8fbdc90eb1f6b0fce417\": container with ID starting with da89f60aa80a2d52aec7b9ac240a98a6e1802b21cc0c8fbdc90eb1f6b0fce417 not found: ID does not exist" Oct 14 08:18:16 crc kubenswrapper[4870]: I1014 08:18:16.173241 4870 scope.go:117] "RemoveContainer" containerID="ca7c8eef02d1cc1e73f9bbcd3ba296bd53653e321e2c2de5b64a5baca4ecc9c6" Oct 14 08:18:16 crc kubenswrapper[4870]: E1014 08:18:16.174065 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ca7c8eef02d1cc1e73f9bbcd3ba296bd53653e321e2c2de5b64a5baca4ecc9c6\": container with ID starting with ca7c8eef02d1cc1e73f9bbcd3ba296bd53653e321e2c2de5b64a5baca4ecc9c6 not found: ID does not exist" containerID="ca7c8eef02d1cc1e73f9bbcd3ba296bd53653e321e2c2de5b64a5baca4ecc9c6" Oct 14 08:18:16 crc kubenswrapper[4870]: I1014 08:18:16.174123 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ca7c8eef02d1cc1e73f9bbcd3ba296bd53653e321e2c2de5b64a5baca4ecc9c6"} err="failed to get container status \"ca7c8eef02d1cc1e73f9bbcd3ba296bd53653e321e2c2de5b64a5baca4ecc9c6\": rpc error: code = NotFound desc = could not find container \"ca7c8eef02d1cc1e73f9bbcd3ba296bd53653e321e2c2de5b64a5baca4ecc9c6\": container with ID starting with ca7c8eef02d1cc1e73f9bbcd3ba296bd53653e321e2c2de5b64a5baca4ecc9c6 not found: ID does not exist" Oct 14 08:18:16 crc kubenswrapper[4870]: I1014 08:18:16.174160 4870 scope.go:117] "RemoveContainer" containerID="00594f9c6ade109cf799a6fc59d7b23f5d8ea17c190d4bd32cc227656b9a1996" Oct 14 08:18:16 crc kubenswrapper[4870]: E1014 08:18:16.174639 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"00594f9c6ade109cf799a6fc59d7b23f5d8ea17c190d4bd32cc227656b9a1996\": container with ID starting with 00594f9c6ade109cf799a6fc59d7b23f5d8ea17c190d4bd32cc227656b9a1996 not found: ID does not exist" containerID="00594f9c6ade109cf799a6fc59d7b23f5d8ea17c190d4bd32cc227656b9a1996" Oct 14 08:18:16 crc kubenswrapper[4870]: I1014 08:18:16.174694 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"00594f9c6ade109cf799a6fc59d7b23f5d8ea17c190d4bd32cc227656b9a1996"} err="failed to get container status \"00594f9c6ade109cf799a6fc59d7b23f5d8ea17c190d4bd32cc227656b9a1996\": rpc error: code = NotFound desc = could not find container \"00594f9c6ade109cf799a6fc59d7b23f5d8ea17c190d4bd32cc227656b9a1996\": container with ID starting with 00594f9c6ade109cf799a6fc59d7b23f5d8ea17c190d4bd32cc227656b9a1996 not found: ID does not exist" Oct 14 08:18:17 crc kubenswrapper[4870]: I1014 08:18:17.049574 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2fbdb997-d4d3-4737-9858-ed6a861e475b" path="/var/lib/kubelet/pods/2fbdb997-d4d3-4737-9858-ed6a861e475b/volumes" Oct 14 08:18:31 crc kubenswrapper[4870]: I1014 08:18:31.034261 4870 scope.go:117] "RemoveContainer" containerID="cb5962db9597015a8eaf724ffe6622e32b2c77623d375ac7952b2a86f641e3ec" Oct 14 08:18:31 crc kubenswrapper[4870]: E1014 08:18:31.035266 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 08:18:45 crc kubenswrapper[4870]: I1014 08:18:45.042578 4870 scope.go:117] "RemoveContainer" containerID="cb5962db9597015a8eaf724ffe6622e32b2c77623d375ac7952b2a86f641e3ec" Oct 14 08:18:45 crc kubenswrapper[4870]: E1014 08:18:45.043882 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 08:18:56 crc kubenswrapper[4870]: I1014 08:18:56.034596 4870 scope.go:117] "RemoveContainer" containerID="cb5962db9597015a8eaf724ffe6622e32b2c77623d375ac7952b2a86f641e3ec" Oct 14 08:18:56 crc kubenswrapper[4870]: E1014 08:18:56.035509 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 08:18:58 crc kubenswrapper[4870]: I1014 08:18:58.837692 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-j7s7w"] Oct 14 08:18:58 crc kubenswrapper[4870]: E1014 08:18:58.840359 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2fbdb997-d4d3-4737-9858-ed6a861e475b" containerName="extract-content" Oct 14 08:18:58 crc kubenswrapper[4870]: I1014 08:18:58.840572 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="2fbdb997-d4d3-4737-9858-ed6a861e475b" containerName="extract-content" Oct 14 08:18:58 crc kubenswrapper[4870]: E1014 08:18:58.840718 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2fbdb997-d4d3-4737-9858-ed6a861e475b" containerName="registry-server" Oct 14 08:18:58 crc kubenswrapper[4870]: I1014 08:18:58.840844 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="2fbdb997-d4d3-4737-9858-ed6a861e475b" containerName="registry-server" Oct 14 08:18:58 crc kubenswrapper[4870]: E1014 08:18:58.840989 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2fbdb997-d4d3-4737-9858-ed6a861e475b" containerName="extract-utilities" Oct 14 08:18:58 crc kubenswrapper[4870]: I1014 08:18:58.841123 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="2fbdb997-d4d3-4737-9858-ed6a861e475b" containerName="extract-utilities" Oct 14 08:18:58 crc kubenswrapper[4870]: I1014 08:18:58.841574 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="2fbdb997-d4d3-4737-9858-ed6a861e475b" containerName="registry-server" Oct 14 08:18:58 crc kubenswrapper[4870]: I1014 08:18:58.843795 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-j7s7w" Oct 14 08:18:58 crc kubenswrapper[4870]: I1014 08:18:58.855298 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-j7s7w"] Oct 14 08:18:58 crc kubenswrapper[4870]: I1014 08:18:58.892546 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/332ebf1d-e2e7-4408-8d79-b26307dbe30e-catalog-content\") pod \"certified-operators-j7s7w\" (UID: \"332ebf1d-e2e7-4408-8d79-b26307dbe30e\") " pod="openshift-marketplace/certified-operators-j7s7w" Oct 14 08:18:58 crc kubenswrapper[4870]: I1014 08:18:58.892614 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tgwjr\" (UniqueName: \"kubernetes.io/projected/332ebf1d-e2e7-4408-8d79-b26307dbe30e-kube-api-access-tgwjr\") pod \"certified-operators-j7s7w\" (UID: \"332ebf1d-e2e7-4408-8d79-b26307dbe30e\") " pod="openshift-marketplace/certified-operators-j7s7w" Oct 14 08:18:58 crc kubenswrapper[4870]: I1014 08:18:58.892939 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/332ebf1d-e2e7-4408-8d79-b26307dbe30e-utilities\") pod \"certified-operators-j7s7w\" (UID: \"332ebf1d-e2e7-4408-8d79-b26307dbe30e\") " pod="openshift-marketplace/certified-operators-j7s7w" Oct 14 08:18:58 crc kubenswrapper[4870]: I1014 08:18:58.994330 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/332ebf1d-e2e7-4408-8d79-b26307dbe30e-utilities\") pod \"certified-operators-j7s7w\" (UID: \"332ebf1d-e2e7-4408-8d79-b26307dbe30e\") " pod="openshift-marketplace/certified-operators-j7s7w" Oct 14 08:18:58 crc kubenswrapper[4870]: I1014 08:18:58.994463 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/332ebf1d-e2e7-4408-8d79-b26307dbe30e-catalog-content\") pod \"certified-operators-j7s7w\" (UID: \"332ebf1d-e2e7-4408-8d79-b26307dbe30e\") " pod="openshift-marketplace/certified-operators-j7s7w" Oct 14 08:18:58 crc kubenswrapper[4870]: I1014 08:18:58.994503 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tgwjr\" (UniqueName: \"kubernetes.io/projected/332ebf1d-e2e7-4408-8d79-b26307dbe30e-kube-api-access-tgwjr\") pod \"certified-operators-j7s7w\" (UID: \"332ebf1d-e2e7-4408-8d79-b26307dbe30e\") " pod="openshift-marketplace/certified-operators-j7s7w" Oct 14 08:18:58 crc kubenswrapper[4870]: I1014 08:18:58.995172 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/332ebf1d-e2e7-4408-8d79-b26307dbe30e-utilities\") pod \"certified-operators-j7s7w\" (UID: \"332ebf1d-e2e7-4408-8d79-b26307dbe30e\") " pod="openshift-marketplace/certified-operators-j7s7w" Oct 14 08:18:58 crc kubenswrapper[4870]: I1014 08:18:58.995306 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/332ebf1d-e2e7-4408-8d79-b26307dbe30e-catalog-content\") pod \"certified-operators-j7s7w\" (UID: \"332ebf1d-e2e7-4408-8d79-b26307dbe30e\") " pod="openshift-marketplace/certified-operators-j7s7w" Oct 14 08:18:59 crc kubenswrapper[4870]: I1014 08:18:59.033050 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tgwjr\" (UniqueName: \"kubernetes.io/projected/332ebf1d-e2e7-4408-8d79-b26307dbe30e-kube-api-access-tgwjr\") pod \"certified-operators-j7s7w\" (UID: \"332ebf1d-e2e7-4408-8d79-b26307dbe30e\") " pod="openshift-marketplace/certified-operators-j7s7w" Oct 14 08:18:59 crc kubenswrapper[4870]: I1014 08:18:59.169534 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-j7s7w" Oct 14 08:18:59 crc kubenswrapper[4870]: I1014 08:18:59.482659 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-j7s7w"] Oct 14 08:19:00 crc kubenswrapper[4870]: I1014 08:19:00.476728 4870 generic.go:334] "Generic (PLEG): container finished" podID="332ebf1d-e2e7-4408-8d79-b26307dbe30e" containerID="ce85b78bc7dfc7446151e0aabf9e118a3c5703e6582226a0a233717f0a1c852f" exitCode=0 Oct 14 08:19:00 crc kubenswrapper[4870]: I1014 08:19:00.476795 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-j7s7w" event={"ID":"332ebf1d-e2e7-4408-8d79-b26307dbe30e","Type":"ContainerDied","Data":"ce85b78bc7dfc7446151e0aabf9e118a3c5703e6582226a0a233717f0a1c852f"} Oct 14 08:19:00 crc kubenswrapper[4870]: I1014 08:19:00.477208 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-j7s7w" event={"ID":"332ebf1d-e2e7-4408-8d79-b26307dbe30e","Type":"ContainerStarted","Data":"b2847f5dcb70d2b6308fe9235ddec619fa1d66f18dff997d7b1a28824e5ca66a"} Oct 14 08:19:02 crc kubenswrapper[4870]: I1014 08:19:02.502119 4870 generic.go:334] "Generic (PLEG): container finished" podID="332ebf1d-e2e7-4408-8d79-b26307dbe30e" containerID="8a51e8c34ee8df99a33d185f824ef56e00688b34ada912c9acf9f4e3f3f25c14" exitCode=0 Oct 14 08:19:02 crc kubenswrapper[4870]: I1014 08:19:02.502350 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-j7s7w" event={"ID":"332ebf1d-e2e7-4408-8d79-b26307dbe30e","Type":"ContainerDied","Data":"8a51e8c34ee8df99a33d185f824ef56e00688b34ada912c9acf9f4e3f3f25c14"} Oct 14 08:19:03 crc kubenswrapper[4870]: I1014 08:19:03.515524 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-j7s7w" event={"ID":"332ebf1d-e2e7-4408-8d79-b26307dbe30e","Type":"ContainerStarted","Data":"1713eba955d49b8056743d9a328a7062eabf3db04610b8ab8a8e347e9952a71c"} Oct 14 08:19:03 crc kubenswrapper[4870]: I1014 08:19:03.552561 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-j7s7w" podStartSLOduration=3.049595132 podStartE2EDuration="5.552535862s" podCreationTimestamp="2025-10-14 08:18:58 +0000 UTC" firstStartedPulling="2025-10-14 08:19:00.47882193 +0000 UTC m=+4676.176182331" lastFinishedPulling="2025-10-14 08:19:02.98176266 +0000 UTC m=+4678.679123061" observedRunningTime="2025-10-14 08:19:03.541649201 +0000 UTC m=+4679.239009612" watchObservedRunningTime="2025-10-14 08:19:03.552535862 +0000 UTC m=+4679.249896273" Oct 14 08:19:03 crc kubenswrapper[4870]: I1014 08:19:03.828214 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-ck6rx"] Oct 14 08:19:03 crc kubenswrapper[4870]: I1014 08:19:03.830166 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ck6rx" Oct 14 08:19:03 crc kubenswrapper[4870]: I1014 08:19:03.850872 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-ck6rx"] Oct 14 08:19:03 crc kubenswrapper[4870]: I1014 08:19:03.877625 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p87pp\" (UniqueName: \"kubernetes.io/projected/8c58cbac-14a6-482d-8a7a-28e7fcfab25f-kube-api-access-p87pp\") pod \"redhat-operators-ck6rx\" (UID: \"8c58cbac-14a6-482d-8a7a-28e7fcfab25f\") " pod="openshift-marketplace/redhat-operators-ck6rx" Oct 14 08:19:03 crc kubenswrapper[4870]: I1014 08:19:03.877766 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8c58cbac-14a6-482d-8a7a-28e7fcfab25f-catalog-content\") pod \"redhat-operators-ck6rx\" (UID: \"8c58cbac-14a6-482d-8a7a-28e7fcfab25f\") " pod="openshift-marketplace/redhat-operators-ck6rx" Oct 14 08:19:03 crc kubenswrapper[4870]: I1014 08:19:03.877851 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8c58cbac-14a6-482d-8a7a-28e7fcfab25f-utilities\") pod \"redhat-operators-ck6rx\" (UID: \"8c58cbac-14a6-482d-8a7a-28e7fcfab25f\") " pod="openshift-marketplace/redhat-operators-ck6rx" Oct 14 08:19:03 crc kubenswrapper[4870]: I1014 08:19:03.979696 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p87pp\" (UniqueName: \"kubernetes.io/projected/8c58cbac-14a6-482d-8a7a-28e7fcfab25f-kube-api-access-p87pp\") pod \"redhat-operators-ck6rx\" (UID: \"8c58cbac-14a6-482d-8a7a-28e7fcfab25f\") " pod="openshift-marketplace/redhat-operators-ck6rx" Oct 14 08:19:03 crc kubenswrapper[4870]: I1014 08:19:03.979876 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8c58cbac-14a6-482d-8a7a-28e7fcfab25f-catalog-content\") pod \"redhat-operators-ck6rx\" (UID: \"8c58cbac-14a6-482d-8a7a-28e7fcfab25f\") " pod="openshift-marketplace/redhat-operators-ck6rx" Oct 14 08:19:03 crc kubenswrapper[4870]: I1014 08:19:03.980051 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8c58cbac-14a6-482d-8a7a-28e7fcfab25f-utilities\") pod \"redhat-operators-ck6rx\" (UID: \"8c58cbac-14a6-482d-8a7a-28e7fcfab25f\") " pod="openshift-marketplace/redhat-operators-ck6rx" Oct 14 08:19:03 crc kubenswrapper[4870]: I1014 08:19:03.980409 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8c58cbac-14a6-482d-8a7a-28e7fcfab25f-catalog-content\") pod \"redhat-operators-ck6rx\" (UID: \"8c58cbac-14a6-482d-8a7a-28e7fcfab25f\") " pod="openshift-marketplace/redhat-operators-ck6rx" Oct 14 08:19:03 crc kubenswrapper[4870]: I1014 08:19:03.981015 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8c58cbac-14a6-482d-8a7a-28e7fcfab25f-utilities\") pod \"redhat-operators-ck6rx\" (UID: \"8c58cbac-14a6-482d-8a7a-28e7fcfab25f\") " pod="openshift-marketplace/redhat-operators-ck6rx" Oct 14 08:19:04 crc kubenswrapper[4870]: I1014 08:19:04.002627 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p87pp\" (UniqueName: \"kubernetes.io/projected/8c58cbac-14a6-482d-8a7a-28e7fcfab25f-kube-api-access-p87pp\") pod \"redhat-operators-ck6rx\" (UID: \"8c58cbac-14a6-482d-8a7a-28e7fcfab25f\") " pod="openshift-marketplace/redhat-operators-ck6rx" Oct 14 08:19:04 crc kubenswrapper[4870]: I1014 08:19:04.154780 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ck6rx" Oct 14 08:19:04 crc kubenswrapper[4870]: I1014 08:19:04.630381 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-ck6rx"] Oct 14 08:19:04 crc kubenswrapper[4870]: W1014 08:19:04.641100 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8c58cbac_14a6_482d_8a7a_28e7fcfab25f.slice/crio-ff11ea74d118dfd65ac37b96127578085875eb811ff13a3909bcba445a37fb3b WatchSource:0}: Error finding container ff11ea74d118dfd65ac37b96127578085875eb811ff13a3909bcba445a37fb3b: Status 404 returned error can't find the container with id ff11ea74d118dfd65ac37b96127578085875eb811ff13a3909bcba445a37fb3b Oct 14 08:19:05 crc kubenswrapper[4870]: I1014 08:19:05.534269 4870 generic.go:334] "Generic (PLEG): container finished" podID="8c58cbac-14a6-482d-8a7a-28e7fcfab25f" containerID="85cae0e16eafdc3fc2939a91b2b8ae7ccf9f90a0b23ae587261b40830e93b54a" exitCode=0 Oct 14 08:19:05 crc kubenswrapper[4870]: I1014 08:19:05.534309 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ck6rx" event={"ID":"8c58cbac-14a6-482d-8a7a-28e7fcfab25f","Type":"ContainerDied","Data":"85cae0e16eafdc3fc2939a91b2b8ae7ccf9f90a0b23ae587261b40830e93b54a"} Oct 14 08:19:05 crc kubenswrapper[4870]: I1014 08:19:05.534338 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ck6rx" event={"ID":"8c58cbac-14a6-482d-8a7a-28e7fcfab25f","Type":"ContainerStarted","Data":"ff11ea74d118dfd65ac37b96127578085875eb811ff13a3909bcba445a37fb3b"} Oct 14 08:19:06 crc kubenswrapper[4870]: I1014 08:19:06.546909 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ck6rx" event={"ID":"8c58cbac-14a6-482d-8a7a-28e7fcfab25f","Type":"ContainerStarted","Data":"d2637a4924c98d7a6b4dc5488a3f7e8b021018cb947cb0f83ef5cb08e571a74b"} Oct 14 08:19:07 crc kubenswrapper[4870]: I1014 08:19:07.582066 4870 generic.go:334] "Generic (PLEG): container finished" podID="8c58cbac-14a6-482d-8a7a-28e7fcfab25f" containerID="d2637a4924c98d7a6b4dc5488a3f7e8b021018cb947cb0f83ef5cb08e571a74b" exitCode=0 Oct 14 08:19:07 crc kubenswrapper[4870]: I1014 08:19:07.582197 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ck6rx" event={"ID":"8c58cbac-14a6-482d-8a7a-28e7fcfab25f","Type":"ContainerDied","Data":"d2637a4924c98d7a6b4dc5488a3f7e8b021018cb947cb0f83ef5cb08e571a74b"} Oct 14 08:19:08 crc kubenswrapper[4870]: I1014 08:19:08.593787 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ck6rx" event={"ID":"8c58cbac-14a6-482d-8a7a-28e7fcfab25f","Type":"ContainerStarted","Data":"9c6c71451171c06ec1eed187668f6626c739771ae259f280cc2a7e72e7239a5f"} Oct 14 08:19:08 crc kubenswrapper[4870]: I1014 08:19:08.623304 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-ck6rx" podStartSLOduration=3.133462181 podStartE2EDuration="5.623265699s" podCreationTimestamp="2025-10-14 08:19:03 +0000 UTC" firstStartedPulling="2025-10-14 08:19:05.537425597 +0000 UTC m=+4681.234785968" lastFinishedPulling="2025-10-14 08:19:08.027229075 +0000 UTC m=+4683.724589486" observedRunningTime="2025-10-14 08:19:08.613877433 +0000 UTC m=+4684.311237864" watchObservedRunningTime="2025-10-14 08:19:08.623265699 +0000 UTC m=+4684.320626080" Oct 14 08:19:09 crc kubenswrapper[4870]: I1014 08:19:09.034486 4870 scope.go:117] "RemoveContainer" containerID="cb5962db9597015a8eaf724ffe6622e32b2c77623d375ac7952b2a86f641e3ec" Oct 14 08:19:09 crc kubenswrapper[4870]: E1014 08:19:09.034777 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 08:19:09 crc kubenswrapper[4870]: I1014 08:19:09.170386 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-j7s7w" Oct 14 08:19:09 crc kubenswrapper[4870]: I1014 08:19:09.170494 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-j7s7w" Oct 14 08:19:09 crc kubenswrapper[4870]: I1014 08:19:09.247323 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-j7s7w" Oct 14 08:19:09 crc kubenswrapper[4870]: I1014 08:19:09.688107 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-j7s7w" Oct 14 08:19:11 crc kubenswrapper[4870]: I1014 08:19:11.822867 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-j7s7w"] Oct 14 08:19:11 crc kubenswrapper[4870]: I1014 08:19:11.825398 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-j7s7w" podUID="332ebf1d-e2e7-4408-8d79-b26307dbe30e" containerName="registry-server" containerID="cri-o://1713eba955d49b8056743d9a328a7062eabf3db04610b8ab8a8e347e9952a71c" gracePeriod=2 Oct 14 08:19:13 crc kubenswrapper[4870]: I1014 08:19:13.232675 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-j7s7w" Oct 14 08:19:13 crc kubenswrapper[4870]: I1014 08:19:13.314103 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/332ebf1d-e2e7-4408-8d79-b26307dbe30e-utilities\") pod \"332ebf1d-e2e7-4408-8d79-b26307dbe30e\" (UID: \"332ebf1d-e2e7-4408-8d79-b26307dbe30e\") " Oct 14 08:19:13 crc kubenswrapper[4870]: I1014 08:19:13.314249 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tgwjr\" (UniqueName: \"kubernetes.io/projected/332ebf1d-e2e7-4408-8d79-b26307dbe30e-kube-api-access-tgwjr\") pod \"332ebf1d-e2e7-4408-8d79-b26307dbe30e\" (UID: \"332ebf1d-e2e7-4408-8d79-b26307dbe30e\") " Oct 14 08:19:13 crc kubenswrapper[4870]: I1014 08:19:13.314381 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/332ebf1d-e2e7-4408-8d79-b26307dbe30e-catalog-content\") pod \"332ebf1d-e2e7-4408-8d79-b26307dbe30e\" (UID: \"332ebf1d-e2e7-4408-8d79-b26307dbe30e\") " Oct 14 08:19:13 crc kubenswrapper[4870]: I1014 08:19:13.315648 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/332ebf1d-e2e7-4408-8d79-b26307dbe30e-utilities" (OuterVolumeSpecName: "utilities") pod "332ebf1d-e2e7-4408-8d79-b26307dbe30e" (UID: "332ebf1d-e2e7-4408-8d79-b26307dbe30e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:19:13 crc kubenswrapper[4870]: I1014 08:19:13.316995 4870 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/332ebf1d-e2e7-4408-8d79-b26307dbe30e-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 08:19:13 crc kubenswrapper[4870]: I1014 08:19:13.322634 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/332ebf1d-e2e7-4408-8d79-b26307dbe30e-kube-api-access-tgwjr" (OuterVolumeSpecName: "kube-api-access-tgwjr") pod "332ebf1d-e2e7-4408-8d79-b26307dbe30e" (UID: "332ebf1d-e2e7-4408-8d79-b26307dbe30e"). InnerVolumeSpecName "kube-api-access-tgwjr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:19:13 crc kubenswrapper[4870]: I1014 08:19:13.375545 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/332ebf1d-e2e7-4408-8d79-b26307dbe30e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "332ebf1d-e2e7-4408-8d79-b26307dbe30e" (UID: "332ebf1d-e2e7-4408-8d79-b26307dbe30e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:19:13 crc kubenswrapper[4870]: I1014 08:19:13.418637 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tgwjr\" (UniqueName: \"kubernetes.io/projected/332ebf1d-e2e7-4408-8d79-b26307dbe30e-kube-api-access-tgwjr\") on node \"crc\" DevicePath \"\"" Oct 14 08:19:13 crc kubenswrapper[4870]: I1014 08:19:13.418673 4870 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/332ebf1d-e2e7-4408-8d79-b26307dbe30e-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 08:19:13 crc kubenswrapper[4870]: I1014 08:19:13.641258 4870 generic.go:334] "Generic (PLEG): container finished" podID="332ebf1d-e2e7-4408-8d79-b26307dbe30e" containerID="1713eba955d49b8056743d9a328a7062eabf3db04610b8ab8a8e347e9952a71c" exitCode=0 Oct 14 08:19:13 crc kubenswrapper[4870]: I1014 08:19:13.641324 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-j7s7w" event={"ID":"332ebf1d-e2e7-4408-8d79-b26307dbe30e","Type":"ContainerDied","Data":"1713eba955d49b8056743d9a328a7062eabf3db04610b8ab8a8e347e9952a71c"} Oct 14 08:19:13 crc kubenswrapper[4870]: I1014 08:19:13.641367 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-j7s7w" event={"ID":"332ebf1d-e2e7-4408-8d79-b26307dbe30e","Type":"ContainerDied","Data":"b2847f5dcb70d2b6308fe9235ddec619fa1d66f18dff997d7b1a28824e5ca66a"} Oct 14 08:19:13 crc kubenswrapper[4870]: I1014 08:19:13.641398 4870 scope.go:117] "RemoveContainer" containerID="1713eba955d49b8056743d9a328a7062eabf3db04610b8ab8a8e347e9952a71c" Oct 14 08:19:13 crc kubenswrapper[4870]: I1014 08:19:13.641422 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-j7s7w" Oct 14 08:19:13 crc kubenswrapper[4870]: I1014 08:19:13.671213 4870 scope.go:117] "RemoveContainer" containerID="8a51e8c34ee8df99a33d185f824ef56e00688b34ada912c9acf9f4e3f3f25c14" Oct 14 08:19:13 crc kubenswrapper[4870]: I1014 08:19:13.708573 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-j7s7w"] Oct 14 08:19:13 crc kubenswrapper[4870]: I1014 08:19:13.719557 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-j7s7w"] Oct 14 08:19:13 crc kubenswrapper[4870]: I1014 08:19:13.720728 4870 scope.go:117] "RemoveContainer" containerID="ce85b78bc7dfc7446151e0aabf9e118a3c5703e6582226a0a233717f0a1c852f" Oct 14 08:19:13 crc kubenswrapper[4870]: I1014 08:19:13.754638 4870 scope.go:117] "RemoveContainer" containerID="1713eba955d49b8056743d9a328a7062eabf3db04610b8ab8a8e347e9952a71c" Oct 14 08:19:13 crc kubenswrapper[4870]: E1014 08:19:13.755507 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1713eba955d49b8056743d9a328a7062eabf3db04610b8ab8a8e347e9952a71c\": container with ID starting with 1713eba955d49b8056743d9a328a7062eabf3db04610b8ab8a8e347e9952a71c not found: ID does not exist" containerID="1713eba955d49b8056743d9a328a7062eabf3db04610b8ab8a8e347e9952a71c" Oct 14 08:19:13 crc kubenswrapper[4870]: I1014 08:19:13.755554 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1713eba955d49b8056743d9a328a7062eabf3db04610b8ab8a8e347e9952a71c"} err="failed to get container status \"1713eba955d49b8056743d9a328a7062eabf3db04610b8ab8a8e347e9952a71c\": rpc error: code = NotFound desc = could not find container \"1713eba955d49b8056743d9a328a7062eabf3db04610b8ab8a8e347e9952a71c\": container with ID starting with 1713eba955d49b8056743d9a328a7062eabf3db04610b8ab8a8e347e9952a71c not found: ID does not exist" Oct 14 08:19:13 crc kubenswrapper[4870]: I1014 08:19:13.755582 4870 scope.go:117] "RemoveContainer" containerID="8a51e8c34ee8df99a33d185f824ef56e00688b34ada912c9acf9f4e3f3f25c14" Oct 14 08:19:13 crc kubenswrapper[4870]: E1014 08:19:13.756203 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8a51e8c34ee8df99a33d185f824ef56e00688b34ada912c9acf9f4e3f3f25c14\": container with ID starting with 8a51e8c34ee8df99a33d185f824ef56e00688b34ada912c9acf9f4e3f3f25c14 not found: ID does not exist" containerID="8a51e8c34ee8df99a33d185f824ef56e00688b34ada912c9acf9f4e3f3f25c14" Oct 14 08:19:13 crc kubenswrapper[4870]: I1014 08:19:13.756267 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8a51e8c34ee8df99a33d185f824ef56e00688b34ada912c9acf9f4e3f3f25c14"} err="failed to get container status \"8a51e8c34ee8df99a33d185f824ef56e00688b34ada912c9acf9f4e3f3f25c14\": rpc error: code = NotFound desc = could not find container \"8a51e8c34ee8df99a33d185f824ef56e00688b34ada912c9acf9f4e3f3f25c14\": container with ID starting with 8a51e8c34ee8df99a33d185f824ef56e00688b34ada912c9acf9f4e3f3f25c14 not found: ID does not exist" Oct 14 08:19:13 crc kubenswrapper[4870]: I1014 08:19:13.756309 4870 scope.go:117] "RemoveContainer" containerID="ce85b78bc7dfc7446151e0aabf9e118a3c5703e6582226a0a233717f0a1c852f" Oct 14 08:19:13 crc kubenswrapper[4870]: E1014 08:19:13.756946 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ce85b78bc7dfc7446151e0aabf9e118a3c5703e6582226a0a233717f0a1c852f\": container with ID starting with ce85b78bc7dfc7446151e0aabf9e118a3c5703e6582226a0a233717f0a1c852f not found: ID does not exist" containerID="ce85b78bc7dfc7446151e0aabf9e118a3c5703e6582226a0a233717f0a1c852f" Oct 14 08:19:13 crc kubenswrapper[4870]: I1014 08:19:13.756973 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ce85b78bc7dfc7446151e0aabf9e118a3c5703e6582226a0a233717f0a1c852f"} err="failed to get container status \"ce85b78bc7dfc7446151e0aabf9e118a3c5703e6582226a0a233717f0a1c852f\": rpc error: code = NotFound desc = could not find container \"ce85b78bc7dfc7446151e0aabf9e118a3c5703e6582226a0a233717f0a1c852f\": container with ID starting with ce85b78bc7dfc7446151e0aabf9e118a3c5703e6582226a0a233717f0a1c852f not found: ID does not exist" Oct 14 08:19:14 crc kubenswrapper[4870]: I1014 08:19:14.155025 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-ck6rx" Oct 14 08:19:14 crc kubenswrapper[4870]: I1014 08:19:14.155533 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-ck6rx" Oct 14 08:19:14 crc kubenswrapper[4870]: I1014 08:19:14.306090 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-ck6rx" Oct 14 08:19:14 crc kubenswrapper[4870]: I1014 08:19:14.726126 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-ck6rx" Oct 14 08:19:15 crc kubenswrapper[4870]: I1014 08:19:15.047366 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="332ebf1d-e2e7-4408-8d79-b26307dbe30e" path="/var/lib/kubelet/pods/332ebf1d-e2e7-4408-8d79-b26307dbe30e/volumes" Oct 14 08:19:16 crc kubenswrapper[4870]: I1014 08:19:16.617325 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-ck6rx"] Oct 14 08:19:16 crc kubenswrapper[4870]: I1014 08:19:16.685102 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-ck6rx" podUID="8c58cbac-14a6-482d-8a7a-28e7fcfab25f" containerName="registry-server" containerID="cri-o://9c6c71451171c06ec1eed187668f6626c739771ae259f280cc2a7e72e7239a5f" gracePeriod=2 Oct 14 08:19:17 crc kubenswrapper[4870]: I1014 08:19:17.193257 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ck6rx" Oct 14 08:19:17 crc kubenswrapper[4870]: I1014 08:19:17.285184 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8c58cbac-14a6-482d-8a7a-28e7fcfab25f-utilities\") pod \"8c58cbac-14a6-482d-8a7a-28e7fcfab25f\" (UID: \"8c58cbac-14a6-482d-8a7a-28e7fcfab25f\") " Oct 14 08:19:17 crc kubenswrapper[4870]: I1014 08:19:17.285270 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p87pp\" (UniqueName: \"kubernetes.io/projected/8c58cbac-14a6-482d-8a7a-28e7fcfab25f-kube-api-access-p87pp\") pod \"8c58cbac-14a6-482d-8a7a-28e7fcfab25f\" (UID: \"8c58cbac-14a6-482d-8a7a-28e7fcfab25f\") " Oct 14 08:19:17 crc kubenswrapper[4870]: I1014 08:19:17.285383 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8c58cbac-14a6-482d-8a7a-28e7fcfab25f-catalog-content\") pod \"8c58cbac-14a6-482d-8a7a-28e7fcfab25f\" (UID: \"8c58cbac-14a6-482d-8a7a-28e7fcfab25f\") " Oct 14 08:19:17 crc kubenswrapper[4870]: I1014 08:19:17.286483 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8c58cbac-14a6-482d-8a7a-28e7fcfab25f-utilities" (OuterVolumeSpecName: "utilities") pod "8c58cbac-14a6-482d-8a7a-28e7fcfab25f" (UID: "8c58cbac-14a6-482d-8a7a-28e7fcfab25f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:19:17 crc kubenswrapper[4870]: I1014 08:19:17.293689 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8c58cbac-14a6-482d-8a7a-28e7fcfab25f-kube-api-access-p87pp" (OuterVolumeSpecName: "kube-api-access-p87pp") pod "8c58cbac-14a6-482d-8a7a-28e7fcfab25f" (UID: "8c58cbac-14a6-482d-8a7a-28e7fcfab25f"). InnerVolumeSpecName "kube-api-access-p87pp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:19:17 crc kubenswrapper[4870]: I1014 08:19:17.387201 4870 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8c58cbac-14a6-482d-8a7a-28e7fcfab25f-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 08:19:17 crc kubenswrapper[4870]: I1014 08:19:17.387277 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p87pp\" (UniqueName: \"kubernetes.io/projected/8c58cbac-14a6-482d-8a7a-28e7fcfab25f-kube-api-access-p87pp\") on node \"crc\" DevicePath \"\"" Oct 14 08:19:17 crc kubenswrapper[4870]: I1014 08:19:17.699996 4870 generic.go:334] "Generic (PLEG): container finished" podID="8c58cbac-14a6-482d-8a7a-28e7fcfab25f" containerID="9c6c71451171c06ec1eed187668f6626c739771ae259f280cc2a7e72e7239a5f" exitCode=0 Oct 14 08:19:17 crc kubenswrapper[4870]: I1014 08:19:17.700129 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ck6rx" Oct 14 08:19:17 crc kubenswrapper[4870]: I1014 08:19:17.700193 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ck6rx" event={"ID":"8c58cbac-14a6-482d-8a7a-28e7fcfab25f","Type":"ContainerDied","Data":"9c6c71451171c06ec1eed187668f6626c739771ae259f280cc2a7e72e7239a5f"} Oct 14 08:19:17 crc kubenswrapper[4870]: I1014 08:19:17.703026 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ck6rx" event={"ID":"8c58cbac-14a6-482d-8a7a-28e7fcfab25f","Type":"ContainerDied","Data":"ff11ea74d118dfd65ac37b96127578085875eb811ff13a3909bcba445a37fb3b"} Oct 14 08:19:17 crc kubenswrapper[4870]: I1014 08:19:17.703070 4870 scope.go:117] "RemoveContainer" containerID="9c6c71451171c06ec1eed187668f6626c739771ae259f280cc2a7e72e7239a5f" Oct 14 08:19:17 crc kubenswrapper[4870]: I1014 08:19:17.729880 4870 scope.go:117] "RemoveContainer" containerID="d2637a4924c98d7a6b4dc5488a3f7e8b021018cb947cb0f83ef5cb08e571a74b" Oct 14 08:19:17 crc kubenswrapper[4870]: I1014 08:19:17.754607 4870 scope.go:117] "RemoveContainer" containerID="85cae0e16eafdc3fc2939a91b2b8ae7ccf9f90a0b23ae587261b40830e93b54a" Oct 14 08:19:17 crc kubenswrapper[4870]: I1014 08:19:17.790712 4870 scope.go:117] "RemoveContainer" containerID="9c6c71451171c06ec1eed187668f6626c739771ae259f280cc2a7e72e7239a5f" Oct 14 08:19:17 crc kubenswrapper[4870]: E1014 08:19:17.791682 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9c6c71451171c06ec1eed187668f6626c739771ae259f280cc2a7e72e7239a5f\": container with ID starting with 9c6c71451171c06ec1eed187668f6626c739771ae259f280cc2a7e72e7239a5f not found: ID does not exist" containerID="9c6c71451171c06ec1eed187668f6626c739771ae259f280cc2a7e72e7239a5f" Oct 14 08:19:17 crc kubenswrapper[4870]: I1014 08:19:17.791744 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9c6c71451171c06ec1eed187668f6626c739771ae259f280cc2a7e72e7239a5f"} err="failed to get container status \"9c6c71451171c06ec1eed187668f6626c739771ae259f280cc2a7e72e7239a5f\": rpc error: code = NotFound desc = could not find container \"9c6c71451171c06ec1eed187668f6626c739771ae259f280cc2a7e72e7239a5f\": container with ID starting with 9c6c71451171c06ec1eed187668f6626c739771ae259f280cc2a7e72e7239a5f not found: ID does not exist" Oct 14 08:19:17 crc kubenswrapper[4870]: I1014 08:19:17.791781 4870 scope.go:117] "RemoveContainer" containerID="d2637a4924c98d7a6b4dc5488a3f7e8b021018cb947cb0f83ef5cb08e571a74b" Oct 14 08:19:17 crc kubenswrapper[4870]: E1014 08:19:17.792343 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d2637a4924c98d7a6b4dc5488a3f7e8b021018cb947cb0f83ef5cb08e571a74b\": container with ID starting with d2637a4924c98d7a6b4dc5488a3f7e8b021018cb947cb0f83ef5cb08e571a74b not found: ID does not exist" containerID="d2637a4924c98d7a6b4dc5488a3f7e8b021018cb947cb0f83ef5cb08e571a74b" Oct 14 08:19:17 crc kubenswrapper[4870]: I1014 08:19:17.792600 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d2637a4924c98d7a6b4dc5488a3f7e8b021018cb947cb0f83ef5cb08e571a74b"} err="failed to get container status \"d2637a4924c98d7a6b4dc5488a3f7e8b021018cb947cb0f83ef5cb08e571a74b\": rpc error: code = NotFound desc = could not find container \"d2637a4924c98d7a6b4dc5488a3f7e8b021018cb947cb0f83ef5cb08e571a74b\": container with ID starting with d2637a4924c98d7a6b4dc5488a3f7e8b021018cb947cb0f83ef5cb08e571a74b not found: ID does not exist" Oct 14 08:19:17 crc kubenswrapper[4870]: I1014 08:19:17.792782 4870 scope.go:117] "RemoveContainer" containerID="85cae0e16eafdc3fc2939a91b2b8ae7ccf9f90a0b23ae587261b40830e93b54a" Oct 14 08:19:17 crc kubenswrapper[4870]: E1014 08:19:17.793575 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"85cae0e16eafdc3fc2939a91b2b8ae7ccf9f90a0b23ae587261b40830e93b54a\": container with ID starting with 85cae0e16eafdc3fc2939a91b2b8ae7ccf9f90a0b23ae587261b40830e93b54a not found: ID does not exist" containerID="85cae0e16eafdc3fc2939a91b2b8ae7ccf9f90a0b23ae587261b40830e93b54a" Oct 14 08:19:17 crc kubenswrapper[4870]: I1014 08:19:17.793602 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"85cae0e16eafdc3fc2939a91b2b8ae7ccf9f90a0b23ae587261b40830e93b54a"} err="failed to get container status \"85cae0e16eafdc3fc2939a91b2b8ae7ccf9f90a0b23ae587261b40830e93b54a\": rpc error: code = NotFound desc = could not find container \"85cae0e16eafdc3fc2939a91b2b8ae7ccf9f90a0b23ae587261b40830e93b54a\": container with ID starting with 85cae0e16eafdc3fc2939a91b2b8ae7ccf9f90a0b23ae587261b40830e93b54a not found: ID does not exist" Oct 14 08:19:18 crc kubenswrapper[4870]: I1014 08:19:18.276919 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8c58cbac-14a6-482d-8a7a-28e7fcfab25f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8c58cbac-14a6-482d-8a7a-28e7fcfab25f" (UID: "8c58cbac-14a6-482d-8a7a-28e7fcfab25f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:19:18 crc kubenswrapper[4870]: I1014 08:19:18.298840 4870 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8c58cbac-14a6-482d-8a7a-28e7fcfab25f-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 08:19:18 crc kubenswrapper[4870]: I1014 08:19:18.363388 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-ck6rx"] Oct 14 08:19:18 crc kubenswrapper[4870]: I1014 08:19:18.372522 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-ck6rx"] Oct 14 08:19:19 crc kubenswrapper[4870]: I1014 08:19:19.049754 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8c58cbac-14a6-482d-8a7a-28e7fcfab25f" path="/var/lib/kubelet/pods/8c58cbac-14a6-482d-8a7a-28e7fcfab25f/volumes" Oct 14 08:19:22 crc kubenswrapper[4870]: I1014 08:19:22.034068 4870 scope.go:117] "RemoveContainer" containerID="cb5962db9597015a8eaf724ffe6622e32b2c77623d375ac7952b2a86f641e3ec" Oct 14 08:19:22 crc kubenswrapper[4870]: E1014 08:19:22.034906 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 08:19:36 crc kubenswrapper[4870]: I1014 08:19:36.034048 4870 scope.go:117] "RemoveContainer" containerID="cb5962db9597015a8eaf724ffe6622e32b2c77623d375ac7952b2a86f641e3ec" Oct 14 08:19:36 crc kubenswrapper[4870]: E1014 08:19:36.035368 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 08:19:50 crc kubenswrapper[4870]: I1014 08:19:50.034652 4870 scope.go:117] "RemoveContainer" containerID="cb5962db9597015a8eaf724ffe6622e32b2c77623d375ac7952b2a86f641e3ec" Oct 14 08:19:50 crc kubenswrapper[4870]: E1014 08:19:50.035705 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 08:20:04 crc kubenswrapper[4870]: I1014 08:20:04.034699 4870 scope.go:117] "RemoveContainer" containerID="cb5962db9597015a8eaf724ffe6622e32b2c77623d375ac7952b2a86f641e3ec" Oct 14 08:20:04 crc kubenswrapper[4870]: E1014 08:20:04.036011 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 08:20:16 crc kubenswrapper[4870]: I1014 08:20:16.034371 4870 scope.go:117] "RemoveContainer" containerID="cb5962db9597015a8eaf724ffe6622e32b2c77623d375ac7952b2a86f641e3ec" Oct 14 08:20:16 crc kubenswrapper[4870]: E1014 08:20:16.035177 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 08:20:30 crc kubenswrapper[4870]: I1014 08:20:30.034721 4870 scope.go:117] "RemoveContainer" containerID="cb5962db9597015a8eaf724ffe6622e32b2c77623d375ac7952b2a86f641e3ec" Oct 14 08:20:30 crc kubenswrapper[4870]: E1014 08:20:30.035847 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 08:20:42 crc kubenswrapper[4870]: I1014 08:20:42.034704 4870 scope.go:117] "RemoveContainer" containerID="cb5962db9597015a8eaf724ffe6622e32b2c77623d375ac7952b2a86f641e3ec" Oct 14 08:20:42 crc kubenswrapper[4870]: E1014 08:20:42.036183 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 08:20:54 crc kubenswrapper[4870]: I1014 08:20:54.034492 4870 scope.go:117] "RemoveContainer" containerID="cb5962db9597015a8eaf724ffe6622e32b2c77623d375ac7952b2a86f641e3ec" Oct 14 08:20:54 crc kubenswrapper[4870]: E1014 08:20:54.035786 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 08:21:05 crc kubenswrapper[4870]: I1014 08:21:05.047410 4870 scope.go:117] "RemoveContainer" containerID="cb5962db9597015a8eaf724ffe6622e32b2c77623d375ac7952b2a86f641e3ec" Oct 14 08:21:05 crc kubenswrapper[4870]: E1014 08:21:05.050719 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 08:21:19 crc kubenswrapper[4870]: I1014 08:21:19.034120 4870 scope.go:117] "RemoveContainer" containerID="cb5962db9597015a8eaf724ffe6622e32b2c77623d375ac7952b2a86f641e3ec" Oct 14 08:21:19 crc kubenswrapper[4870]: E1014 08:21:19.035434 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 08:21:31 crc kubenswrapper[4870]: I1014 08:21:31.034323 4870 scope.go:117] "RemoveContainer" containerID="cb5962db9597015a8eaf724ffe6622e32b2c77623d375ac7952b2a86f641e3ec" Oct 14 08:21:31 crc kubenswrapper[4870]: E1014 08:21:31.035595 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 08:21:42 crc kubenswrapper[4870]: I1014 08:21:42.035110 4870 scope.go:117] "RemoveContainer" containerID="cb5962db9597015a8eaf724ffe6622e32b2c77623d375ac7952b2a86f641e3ec" Oct 14 08:21:42 crc kubenswrapper[4870]: E1014 08:21:42.036551 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 08:21:57 crc kubenswrapper[4870]: I1014 08:21:57.034934 4870 scope.go:117] "RemoveContainer" containerID="cb5962db9597015a8eaf724ffe6622e32b2c77623d375ac7952b2a86f641e3ec" Oct 14 08:21:57 crc kubenswrapper[4870]: E1014 08:21:57.035855 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 08:22:09 crc kubenswrapper[4870]: I1014 08:22:09.034643 4870 scope.go:117] "RemoveContainer" containerID="cb5962db9597015a8eaf724ffe6622e32b2c77623d375ac7952b2a86f641e3ec" Oct 14 08:22:09 crc kubenswrapper[4870]: E1014 08:22:09.036239 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 08:22:20 crc kubenswrapper[4870]: I1014 08:22:20.033993 4870 scope.go:117] "RemoveContainer" containerID="cb5962db9597015a8eaf724ffe6622e32b2c77623d375ac7952b2a86f641e3ec" Oct 14 08:22:20 crc kubenswrapper[4870]: E1014 08:22:20.035110 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 08:22:34 crc kubenswrapper[4870]: I1014 08:22:34.034386 4870 scope.go:117] "RemoveContainer" containerID="cb5962db9597015a8eaf724ffe6622e32b2c77623d375ac7952b2a86f641e3ec" Oct 14 08:22:34 crc kubenswrapper[4870]: I1014 08:22:34.658164 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" event={"ID":"874a5e45-dffd-4d17-b609-b7d3ed2eab07","Type":"ContainerStarted","Data":"de10f364f1612f1011866f6d30e25476889918a312e7f5d85482df8b3358654e"} Oct 14 08:24:53 crc kubenswrapper[4870]: I1014 08:24:53.951767 4870 patch_prober.go:28] interesting pod/machine-config-daemon-7tvc7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 08:24:53 crc kubenswrapper[4870]: I1014 08:24:53.952664 4870 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 08:25:23 crc kubenswrapper[4870]: I1014 08:25:23.952274 4870 patch_prober.go:28] interesting pod/machine-config-daemon-7tvc7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 08:25:23 crc kubenswrapper[4870]: I1014 08:25:23.953326 4870 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 08:25:53 crc kubenswrapper[4870]: I1014 08:25:53.951744 4870 patch_prober.go:28] interesting pod/machine-config-daemon-7tvc7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 08:25:53 crc kubenswrapper[4870]: I1014 08:25:53.954701 4870 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 08:25:53 crc kubenswrapper[4870]: I1014 08:25:53.954933 4870 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" Oct 14 08:25:53 crc kubenswrapper[4870]: I1014 08:25:53.956332 4870 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"de10f364f1612f1011866f6d30e25476889918a312e7f5d85482df8b3358654e"} pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 14 08:25:53 crc kubenswrapper[4870]: I1014 08:25:53.956655 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerName="machine-config-daemon" containerID="cri-o://de10f364f1612f1011866f6d30e25476889918a312e7f5d85482df8b3358654e" gracePeriod=600 Oct 14 08:25:54 crc kubenswrapper[4870]: I1014 08:25:54.586489 4870 generic.go:334] "Generic (PLEG): container finished" podID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerID="de10f364f1612f1011866f6d30e25476889918a312e7f5d85482df8b3358654e" exitCode=0 Oct 14 08:25:54 crc kubenswrapper[4870]: I1014 08:25:54.586618 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" event={"ID":"874a5e45-dffd-4d17-b609-b7d3ed2eab07","Type":"ContainerDied","Data":"de10f364f1612f1011866f6d30e25476889918a312e7f5d85482df8b3358654e"} Oct 14 08:25:54 crc kubenswrapper[4870]: I1014 08:25:54.586979 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" event={"ID":"874a5e45-dffd-4d17-b609-b7d3ed2eab07","Type":"ContainerStarted","Data":"52468dffc62b6337feac3ee24b71903855af2b6b6e95f02951551f88ff46993a"} Oct 14 08:25:54 crc kubenswrapper[4870]: I1014 08:25:54.587036 4870 scope.go:117] "RemoveContainer" containerID="cb5962db9597015a8eaf724ffe6622e32b2c77623d375ac7952b2a86f641e3ec" Oct 14 08:28:23 crc kubenswrapper[4870]: I1014 08:28:23.951614 4870 patch_prober.go:28] interesting pod/machine-config-daemon-7tvc7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 08:28:23 crc kubenswrapper[4870]: I1014 08:28:23.952310 4870 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 08:28:53 crc kubenswrapper[4870]: I1014 08:28:53.951040 4870 patch_prober.go:28] interesting pod/machine-config-daemon-7tvc7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 08:28:53 crc kubenswrapper[4870]: I1014 08:28:53.951672 4870 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 08:29:15 crc kubenswrapper[4870]: I1014 08:29:15.161805 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-2vsst"] Oct 14 08:29:15 crc kubenswrapper[4870]: E1014 08:29:15.163158 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c58cbac-14a6-482d-8a7a-28e7fcfab25f" containerName="extract-content" Oct 14 08:29:15 crc kubenswrapper[4870]: I1014 08:29:15.163194 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c58cbac-14a6-482d-8a7a-28e7fcfab25f" containerName="extract-content" Oct 14 08:29:15 crc kubenswrapper[4870]: E1014 08:29:15.163240 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="332ebf1d-e2e7-4408-8d79-b26307dbe30e" containerName="registry-server" Oct 14 08:29:15 crc kubenswrapper[4870]: I1014 08:29:15.163256 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="332ebf1d-e2e7-4408-8d79-b26307dbe30e" containerName="registry-server" Oct 14 08:29:15 crc kubenswrapper[4870]: E1014 08:29:15.163296 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c58cbac-14a6-482d-8a7a-28e7fcfab25f" containerName="extract-utilities" Oct 14 08:29:15 crc kubenswrapper[4870]: I1014 08:29:15.163315 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c58cbac-14a6-482d-8a7a-28e7fcfab25f" containerName="extract-utilities" Oct 14 08:29:15 crc kubenswrapper[4870]: E1014 08:29:15.163350 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c58cbac-14a6-482d-8a7a-28e7fcfab25f" containerName="registry-server" Oct 14 08:29:15 crc kubenswrapper[4870]: I1014 08:29:15.163366 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c58cbac-14a6-482d-8a7a-28e7fcfab25f" containerName="registry-server" Oct 14 08:29:15 crc kubenswrapper[4870]: E1014 08:29:15.163403 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="332ebf1d-e2e7-4408-8d79-b26307dbe30e" containerName="extract-utilities" Oct 14 08:29:15 crc kubenswrapper[4870]: I1014 08:29:15.163423 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="332ebf1d-e2e7-4408-8d79-b26307dbe30e" containerName="extract-utilities" Oct 14 08:29:15 crc kubenswrapper[4870]: E1014 08:29:15.163491 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="332ebf1d-e2e7-4408-8d79-b26307dbe30e" containerName="extract-content" Oct 14 08:29:15 crc kubenswrapper[4870]: I1014 08:29:15.163513 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="332ebf1d-e2e7-4408-8d79-b26307dbe30e" containerName="extract-content" Oct 14 08:29:15 crc kubenswrapper[4870]: I1014 08:29:15.163897 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="8c58cbac-14a6-482d-8a7a-28e7fcfab25f" containerName="registry-server" Oct 14 08:29:15 crc kubenswrapper[4870]: I1014 08:29:15.163942 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="332ebf1d-e2e7-4408-8d79-b26307dbe30e" containerName="registry-server" Oct 14 08:29:15 crc kubenswrapper[4870]: I1014 08:29:15.166027 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2vsst" Oct 14 08:29:15 crc kubenswrapper[4870]: I1014 08:29:15.199855 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-2vsst"] Oct 14 08:29:15 crc kubenswrapper[4870]: I1014 08:29:15.234472 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4ff196f5-1fe7-45d8-a641-2b3cd1fc0a87-utilities\") pod \"redhat-operators-2vsst\" (UID: \"4ff196f5-1fe7-45d8-a641-2b3cd1fc0a87\") " pod="openshift-marketplace/redhat-operators-2vsst" Oct 14 08:29:15 crc kubenswrapper[4870]: I1014 08:29:15.234559 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4ff196f5-1fe7-45d8-a641-2b3cd1fc0a87-catalog-content\") pod \"redhat-operators-2vsst\" (UID: \"4ff196f5-1fe7-45d8-a641-2b3cd1fc0a87\") " pod="openshift-marketplace/redhat-operators-2vsst" Oct 14 08:29:15 crc kubenswrapper[4870]: I1014 08:29:15.234618 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pccbk\" (UniqueName: \"kubernetes.io/projected/4ff196f5-1fe7-45d8-a641-2b3cd1fc0a87-kube-api-access-pccbk\") pod \"redhat-operators-2vsst\" (UID: \"4ff196f5-1fe7-45d8-a641-2b3cd1fc0a87\") " pod="openshift-marketplace/redhat-operators-2vsst" Oct 14 08:29:15 crc kubenswrapper[4870]: I1014 08:29:15.336468 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4ff196f5-1fe7-45d8-a641-2b3cd1fc0a87-utilities\") pod \"redhat-operators-2vsst\" (UID: \"4ff196f5-1fe7-45d8-a641-2b3cd1fc0a87\") " pod="openshift-marketplace/redhat-operators-2vsst" Oct 14 08:29:15 crc kubenswrapper[4870]: I1014 08:29:15.336558 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4ff196f5-1fe7-45d8-a641-2b3cd1fc0a87-catalog-content\") pod \"redhat-operators-2vsst\" (UID: \"4ff196f5-1fe7-45d8-a641-2b3cd1fc0a87\") " pod="openshift-marketplace/redhat-operators-2vsst" Oct 14 08:29:15 crc kubenswrapper[4870]: I1014 08:29:15.336620 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pccbk\" (UniqueName: \"kubernetes.io/projected/4ff196f5-1fe7-45d8-a641-2b3cd1fc0a87-kube-api-access-pccbk\") pod \"redhat-operators-2vsst\" (UID: \"4ff196f5-1fe7-45d8-a641-2b3cd1fc0a87\") " pod="openshift-marketplace/redhat-operators-2vsst" Oct 14 08:29:15 crc kubenswrapper[4870]: I1014 08:29:15.337493 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4ff196f5-1fe7-45d8-a641-2b3cd1fc0a87-utilities\") pod \"redhat-operators-2vsst\" (UID: \"4ff196f5-1fe7-45d8-a641-2b3cd1fc0a87\") " pod="openshift-marketplace/redhat-operators-2vsst" Oct 14 08:29:15 crc kubenswrapper[4870]: I1014 08:29:15.337757 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4ff196f5-1fe7-45d8-a641-2b3cd1fc0a87-catalog-content\") pod \"redhat-operators-2vsst\" (UID: \"4ff196f5-1fe7-45d8-a641-2b3cd1fc0a87\") " pod="openshift-marketplace/redhat-operators-2vsst" Oct 14 08:29:15 crc kubenswrapper[4870]: I1014 08:29:15.373704 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pccbk\" (UniqueName: \"kubernetes.io/projected/4ff196f5-1fe7-45d8-a641-2b3cd1fc0a87-kube-api-access-pccbk\") pod \"redhat-operators-2vsst\" (UID: \"4ff196f5-1fe7-45d8-a641-2b3cd1fc0a87\") " pod="openshift-marketplace/redhat-operators-2vsst" Oct 14 08:29:15 crc kubenswrapper[4870]: I1014 08:29:15.538634 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2vsst" Oct 14 08:29:15 crc kubenswrapper[4870]: I1014 08:29:15.958092 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-2vsst"] Oct 14 08:29:16 crc kubenswrapper[4870]: I1014 08:29:16.493826 4870 generic.go:334] "Generic (PLEG): container finished" podID="4ff196f5-1fe7-45d8-a641-2b3cd1fc0a87" containerID="2607bf42dac74d5ce477df127b1f191a43c3c15a2a65aa4bd07866486665f742" exitCode=0 Oct 14 08:29:16 crc kubenswrapper[4870]: I1014 08:29:16.493891 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2vsst" event={"ID":"4ff196f5-1fe7-45d8-a641-2b3cd1fc0a87","Type":"ContainerDied","Data":"2607bf42dac74d5ce477df127b1f191a43c3c15a2a65aa4bd07866486665f742"} Oct 14 08:29:16 crc kubenswrapper[4870]: I1014 08:29:16.494179 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2vsst" event={"ID":"4ff196f5-1fe7-45d8-a641-2b3cd1fc0a87","Type":"ContainerStarted","Data":"fb934b4078c7dbe36b8302b3e3b08b8e38fbccdf425aae8c75e0f931b301b945"} Oct 14 08:29:16 crc kubenswrapper[4870]: I1014 08:29:16.495339 4870 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 14 08:29:17 crc kubenswrapper[4870]: I1014 08:29:17.512070 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2vsst" event={"ID":"4ff196f5-1fe7-45d8-a641-2b3cd1fc0a87","Type":"ContainerStarted","Data":"5162ca6a20ac2f711c0b93e99649ee830e8cc26361336a0d2bbe14cb5a2e2b9b"} Oct 14 08:29:18 crc kubenswrapper[4870]: I1014 08:29:18.521711 4870 generic.go:334] "Generic (PLEG): container finished" podID="4ff196f5-1fe7-45d8-a641-2b3cd1fc0a87" containerID="5162ca6a20ac2f711c0b93e99649ee830e8cc26361336a0d2bbe14cb5a2e2b9b" exitCode=0 Oct 14 08:29:18 crc kubenswrapper[4870]: I1014 08:29:18.521820 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2vsst" event={"ID":"4ff196f5-1fe7-45d8-a641-2b3cd1fc0a87","Type":"ContainerDied","Data":"5162ca6a20ac2f711c0b93e99649ee830e8cc26361336a0d2bbe14cb5a2e2b9b"} Oct 14 08:29:19 crc kubenswrapper[4870]: I1014 08:29:19.534532 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2vsst" event={"ID":"4ff196f5-1fe7-45d8-a641-2b3cd1fc0a87","Type":"ContainerStarted","Data":"09e8844aadda252f7141047e9d9900f014a54494be684f6d9283bab54eaf33e5"} Oct 14 08:29:19 crc kubenswrapper[4870]: I1014 08:29:19.568101 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-2vsst" podStartSLOduration=2.150415046 podStartE2EDuration="4.568077423s" podCreationTimestamp="2025-10-14 08:29:15 +0000 UTC" firstStartedPulling="2025-10-14 08:29:16.495120868 +0000 UTC m=+5292.192481239" lastFinishedPulling="2025-10-14 08:29:18.912783205 +0000 UTC m=+5294.610143616" observedRunningTime="2025-10-14 08:29:19.563539493 +0000 UTC m=+5295.260899894" watchObservedRunningTime="2025-10-14 08:29:19.568077423 +0000 UTC m=+5295.265437824" Oct 14 08:29:23 crc kubenswrapper[4870]: I1014 08:29:23.951072 4870 patch_prober.go:28] interesting pod/machine-config-daemon-7tvc7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 08:29:23 crc kubenswrapper[4870]: I1014 08:29:23.951388 4870 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 08:29:23 crc kubenswrapper[4870]: I1014 08:29:23.951426 4870 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" Oct 14 08:29:23 crc kubenswrapper[4870]: I1014 08:29:23.951872 4870 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"52468dffc62b6337feac3ee24b71903855af2b6b6e95f02951551f88ff46993a"} pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 14 08:29:23 crc kubenswrapper[4870]: I1014 08:29:23.951921 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerName="machine-config-daemon" containerID="cri-o://52468dffc62b6337feac3ee24b71903855af2b6b6e95f02951551f88ff46993a" gracePeriod=600 Oct 14 08:29:24 crc kubenswrapper[4870]: E1014 08:29:24.106551 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 08:29:24 crc kubenswrapper[4870]: I1014 08:29:24.581716 4870 generic.go:334] "Generic (PLEG): container finished" podID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerID="52468dffc62b6337feac3ee24b71903855af2b6b6e95f02951551f88ff46993a" exitCode=0 Oct 14 08:29:24 crc kubenswrapper[4870]: I1014 08:29:24.581789 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" event={"ID":"874a5e45-dffd-4d17-b609-b7d3ed2eab07","Type":"ContainerDied","Data":"52468dffc62b6337feac3ee24b71903855af2b6b6e95f02951551f88ff46993a"} Oct 14 08:29:24 crc kubenswrapper[4870]: I1014 08:29:24.581890 4870 scope.go:117] "RemoveContainer" containerID="de10f364f1612f1011866f6d30e25476889918a312e7f5d85482df8b3358654e" Oct 14 08:29:24 crc kubenswrapper[4870]: I1014 08:29:24.582893 4870 scope.go:117] "RemoveContainer" containerID="52468dffc62b6337feac3ee24b71903855af2b6b6e95f02951551f88ff46993a" Oct 14 08:29:24 crc kubenswrapper[4870]: E1014 08:29:24.587673 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 08:29:25 crc kubenswrapper[4870]: I1014 08:29:25.539395 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-2vsst" Oct 14 08:29:25 crc kubenswrapper[4870]: I1014 08:29:25.541966 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-2vsst" Oct 14 08:29:25 crc kubenswrapper[4870]: I1014 08:29:25.617588 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-2vsst" Oct 14 08:29:26 crc kubenswrapper[4870]: I1014 08:29:26.680335 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-2vsst" Oct 14 08:29:26 crc kubenswrapper[4870]: I1014 08:29:26.754871 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-2vsst"] Oct 14 08:29:28 crc kubenswrapper[4870]: I1014 08:29:28.645536 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-2vsst" podUID="4ff196f5-1fe7-45d8-a641-2b3cd1fc0a87" containerName="registry-server" containerID="cri-o://09e8844aadda252f7141047e9d9900f014a54494be684f6d9283bab54eaf33e5" gracePeriod=2 Oct 14 08:29:29 crc kubenswrapper[4870]: I1014 08:29:29.211668 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2vsst" Oct 14 08:29:29 crc kubenswrapper[4870]: I1014 08:29:29.279089 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4ff196f5-1fe7-45d8-a641-2b3cd1fc0a87-utilities\") pod \"4ff196f5-1fe7-45d8-a641-2b3cd1fc0a87\" (UID: \"4ff196f5-1fe7-45d8-a641-2b3cd1fc0a87\") " Oct 14 08:29:29 crc kubenswrapper[4870]: I1014 08:29:29.279189 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4ff196f5-1fe7-45d8-a641-2b3cd1fc0a87-catalog-content\") pod \"4ff196f5-1fe7-45d8-a641-2b3cd1fc0a87\" (UID: \"4ff196f5-1fe7-45d8-a641-2b3cd1fc0a87\") " Oct 14 08:29:29 crc kubenswrapper[4870]: I1014 08:29:29.279412 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pccbk\" (UniqueName: \"kubernetes.io/projected/4ff196f5-1fe7-45d8-a641-2b3cd1fc0a87-kube-api-access-pccbk\") pod \"4ff196f5-1fe7-45d8-a641-2b3cd1fc0a87\" (UID: \"4ff196f5-1fe7-45d8-a641-2b3cd1fc0a87\") " Oct 14 08:29:29 crc kubenswrapper[4870]: I1014 08:29:29.280721 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4ff196f5-1fe7-45d8-a641-2b3cd1fc0a87-utilities" (OuterVolumeSpecName: "utilities") pod "4ff196f5-1fe7-45d8-a641-2b3cd1fc0a87" (UID: "4ff196f5-1fe7-45d8-a641-2b3cd1fc0a87"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:29:29 crc kubenswrapper[4870]: I1014 08:29:29.280891 4870 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4ff196f5-1fe7-45d8-a641-2b3cd1fc0a87-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 08:29:29 crc kubenswrapper[4870]: I1014 08:29:29.285019 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4ff196f5-1fe7-45d8-a641-2b3cd1fc0a87-kube-api-access-pccbk" (OuterVolumeSpecName: "kube-api-access-pccbk") pod "4ff196f5-1fe7-45d8-a641-2b3cd1fc0a87" (UID: "4ff196f5-1fe7-45d8-a641-2b3cd1fc0a87"). InnerVolumeSpecName "kube-api-access-pccbk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:29:29 crc kubenswrapper[4870]: I1014 08:29:29.362334 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4ff196f5-1fe7-45d8-a641-2b3cd1fc0a87-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4ff196f5-1fe7-45d8-a641-2b3cd1fc0a87" (UID: "4ff196f5-1fe7-45d8-a641-2b3cd1fc0a87"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:29:29 crc kubenswrapper[4870]: I1014 08:29:29.381891 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pccbk\" (UniqueName: \"kubernetes.io/projected/4ff196f5-1fe7-45d8-a641-2b3cd1fc0a87-kube-api-access-pccbk\") on node \"crc\" DevicePath \"\"" Oct 14 08:29:29 crc kubenswrapper[4870]: I1014 08:29:29.381926 4870 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4ff196f5-1fe7-45d8-a641-2b3cd1fc0a87-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 08:29:29 crc kubenswrapper[4870]: I1014 08:29:29.657778 4870 generic.go:334] "Generic (PLEG): container finished" podID="4ff196f5-1fe7-45d8-a641-2b3cd1fc0a87" containerID="09e8844aadda252f7141047e9d9900f014a54494be684f6d9283bab54eaf33e5" exitCode=0 Oct 14 08:29:29 crc kubenswrapper[4870]: I1014 08:29:29.657840 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2vsst" event={"ID":"4ff196f5-1fe7-45d8-a641-2b3cd1fc0a87","Type":"ContainerDied","Data":"09e8844aadda252f7141047e9d9900f014a54494be684f6d9283bab54eaf33e5"} Oct 14 08:29:29 crc kubenswrapper[4870]: I1014 08:29:29.657889 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2vsst" event={"ID":"4ff196f5-1fe7-45d8-a641-2b3cd1fc0a87","Type":"ContainerDied","Data":"fb934b4078c7dbe36b8302b3e3b08b8e38fbccdf425aae8c75e0f931b301b945"} Oct 14 08:29:29 crc kubenswrapper[4870]: I1014 08:29:29.657919 4870 scope.go:117] "RemoveContainer" containerID="09e8844aadda252f7141047e9d9900f014a54494be684f6d9283bab54eaf33e5" Oct 14 08:29:29 crc kubenswrapper[4870]: I1014 08:29:29.657944 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2vsst" Oct 14 08:29:29 crc kubenswrapper[4870]: I1014 08:29:29.699653 4870 scope.go:117] "RemoveContainer" containerID="5162ca6a20ac2f711c0b93e99649ee830e8cc26361336a0d2bbe14cb5a2e2b9b" Oct 14 08:29:29 crc kubenswrapper[4870]: I1014 08:29:29.714611 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-2vsst"] Oct 14 08:29:29 crc kubenswrapper[4870]: I1014 08:29:29.728247 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-2vsst"] Oct 14 08:29:29 crc kubenswrapper[4870]: I1014 08:29:29.746570 4870 scope.go:117] "RemoveContainer" containerID="2607bf42dac74d5ce477df127b1f191a43c3c15a2a65aa4bd07866486665f742" Oct 14 08:29:29 crc kubenswrapper[4870]: I1014 08:29:29.767013 4870 scope.go:117] "RemoveContainer" containerID="09e8844aadda252f7141047e9d9900f014a54494be684f6d9283bab54eaf33e5" Oct 14 08:29:29 crc kubenswrapper[4870]: E1014 08:29:29.767522 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"09e8844aadda252f7141047e9d9900f014a54494be684f6d9283bab54eaf33e5\": container with ID starting with 09e8844aadda252f7141047e9d9900f014a54494be684f6d9283bab54eaf33e5 not found: ID does not exist" containerID="09e8844aadda252f7141047e9d9900f014a54494be684f6d9283bab54eaf33e5" Oct 14 08:29:29 crc kubenswrapper[4870]: I1014 08:29:29.767593 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"09e8844aadda252f7141047e9d9900f014a54494be684f6d9283bab54eaf33e5"} err="failed to get container status \"09e8844aadda252f7141047e9d9900f014a54494be684f6d9283bab54eaf33e5\": rpc error: code = NotFound desc = could not find container \"09e8844aadda252f7141047e9d9900f014a54494be684f6d9283bab54eaf33e5\": container with ID starting with 09e8844aadda252f7141047e9d9900f014a54494be684f6d9283bab54eaf33e5 not found: ID does not exist" Oct 14 08:29:29 crc kubenswrapper[4870]: I1014 08:29:29.767643 4870 scope.go:117] "RemoveContainer" containerID="5162ca6a20ac2f711c0b93e99649ee830e8cc26361336a0d2bbe14cb5a2e2b9b" Oct 14 08:29:29 crc kubenswrapper[4870]: E1014 08:29:29.768052 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5162ca6a20ac2f711c0b93e99649ee830e8cc26361336a0d2bbe14cb5a2e2b9b\": container with ID starting with 5162ca6a20ac2f711c0b93e99649ee830e8cc26361336a0d2bbe14cb5a2e2b9b not found: ID does not exist" containerID="5162ca6a20ac2f711c0b93e99649ee830e8cc26361336a0d2bbe14cb5a2e2b9b" Oct 14 08:29:29 crc kubenswrapper[4870]: I1014 08:29:29.768124 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5162ca6a20ac2f711c0b93e99649ee830e8cc26361336a0d2bbe14cb5a2e2b9b"} err="failed to get container status \"5162ca6a20ac2f711c0b93e99649ee830e8cc26361336a0d2bbe14cb5a2e2b9b\": rpc error: code = NotFound desc = could not find container \"5162ca6a20ac2f711c0b93e99649ee830e8cc26361336a0d2bbe14cb5a2e2b9b\": container with ID starting with 5162ca6a20ac2f711c0b93e99649ee830e8cc26361336a0d2bbe14cb5a2e2b9b not found: ID does not exist" Oct 14 08:29:29 crc kubenswrapper[4870]: I1014 08:29:29.768266 4870 scope.go:117] "RemoveContainer" containerID="2607bf42dac74d5ce477df127b1f191a43c3c15a2a65aa4bd07866486665f742" Oct 14 08:29:29 crc kubenswrapper[4870]: E1014 08:29:29.768723 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2607bf42dac74d5ce477df127b1f191a43c3c15a2a65aa4bd07866486665f742\": container with ID starting with 2607bf42dac74d5ce477df127b1f191a43c3c15a2a65aa4bd07866486665f742 not found: ID does not exist" containerID="2607bf42dac74d5ce477df127b1f191a43c3c15a2a65aa4bd07866486665f742" Oct 14 08:29:29 crc kubenswrapper[4870]: I1014 08:29:29.768797 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2607bf42dac74d5ce477df127b1f191a43c3c15a2a65aa4bd07866486665f742"} err="failed to get container status \"2607bf42dac74d5ce477df127b1f191a43c3c15a2a65aa4bd07866486665f742\": rpc error: code = NotFound desc = could not find container \"2607bf42dac74d5ce477df127b1f191a43c3c15a2a65aa4bd07866486665f742\": container with ID starting with 2607bf42dac74d5ce477df127b1f191a43c3c15a2a65aa4bd07866486665f742 not found: ID does not exist" Oct 14 08:29:31 crc kubenswrapper[4870]: I1014 08:29:31.050021 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4ff196f5-1fe7-45d8-a641-2b3cd1fc0a87" path="/var/lib/kubelet/pods/4ff196f5-1fe7-45d8-a641-2b3cd1fc0a87/volumes" Oct 14 08:29:37 crc kubenswrapper[4870]: I1014 08:29:37.034902 4870 scope.go:117] "RemoveContainer" containerID="52468dffc62b6337feac3ee24b71903855af2b6b6e95f02951551f88ff46993a" Oct 14 08:29:37 crc kubenswrapper[4870]: E1014 08:29:37.035943 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 08:29:49 crc kubenswrapper[4870]: I1014 08:29:49.034997 4870 scope.go:117] "RemoveContainer" containerID="52468dffc62b6337feac3ee24b71903855af2b6b6e95f02951551f88ff46993a" Oct 14 08:29:49 crc kubenswrapper[4870]: E1014 08:29:49.036243 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 08:29:52 crc kubenswrapper[4870]: I1014 08:29:52.305575 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-m5tf6"] Oct 14 08:29:52 crc kubenswrapper[4870]: E1014 08:29:52.306596 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ff196f5-1fe7-45d8-a641-2b3cd1fc0a87" containerName="extract-content" Oct 14 08:29:52 crc kubenswrapper[4870]: I1014 08:29:52.306631 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ff196f5-1fe7-45d8-a641-2b3cd1fc0a87" containerName="extract-content" Oct 14 08:29:52 crc kubenswrapper[4870]: E1014 08:29:52.306679 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ff196f5-1fe7-45d8-a641-2b3cd1fc0a87" containerName="extract-utilities" Oct 14 08:29:52 crc kubenswrapper[4870]: I1014 08:29:52.306695 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ff196f5-1fe7-45d8-a641-2b3cd1fc0a87" containerName="extract-utilities" Oct 14 08:29:52 crc kubenswrapper[4870]: E1014 08:29:52.306717 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ff196f5-1fe7-45d8-a641-2b3cd1fc0a87" containerName="registry-server" Oct 14 08:29:52 crc kubenswrapper[4870]: I1014 08:29:52.306732 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ff196f5-1fe7-45d8-a641-2b3cd1fc0a87" containerName="registry-server" Oct 14 08:29:52 crc kubenswrapper[4870]: I1014 08:29:52.307120 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="4ff196f5-1fe7-45d8-a641-2b3cd1fc0a87" containerName="registry-server" Oct 14 08:29:52 crc kubenswrapper[4870]: I1014 08:29:52.309857 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-m5tf6" Oct 14 08:29:52 crc kubenswrapper[4870]: I1014 08:29:52.329862 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-m5tf6"] Oct 14 08:29:52 crc kubenswrapper[4870]: I1014 08:29:52.365889 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pr9wb\" (UniqueName: \"kubernetes.io/projected/6f0539ab-53f0-46bd-8ed2-13d507fae9ea-kube-api-access-pr9wb\") pod \"certified-operators-m5tf6\" (UID: \"6f0539ab-53f0-46bd-8ed2-13d507fae9ea\") " pod="openshift-marketplace/certified-operators-m5tf6" Oct 14 08:29:52 crc kubenswrapper[4870]: I1014 08:29:52.366171 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6f0539ab-53f0-46bd-8ed2-13d507fae9ea-utilities\") pod \"certified-operators-m5tf6\" (UID: \"6f0539ab-53f0-46bd-8ed2-13d507fae9ea\") " pod="openshift-marketplace/certified-operators-m5tf6" Oct 14 08:29:52 crc kubenswrapper[4870]: I1014 08:29:52.366238 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6f0539ab-53f0-46bd-8ed2-13d507fae9ea-catalog-content\") pod \"certified-operators-m5tf6\" (UID: \"6f0539ab-53f0-46bd-8ed2-13d507fae9ea\") " pod="openshift-marketplace/certified-operators-m5tf6" Oct 14 08:29:52 crc kubenswrapper[4870]: I1014 08:29:52.468129 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6f0539ab-53f0-46bd-8ed2-13d507fae9ea-utilities\") pod \"certified-operators-m5tf6\" (UID: \"6f0539ab-53f0-46bd-8ed2-13d507fae9ea\") " pod="openshift-marketplace/certified-operators-m5tf6" Oct 14 08:29:52 crc kubenswrapper[4870]: I1014 08:29:52.468223 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6f0539ab-53f0-46bd-8ed2-13d507fae9ea-catalog-content\") pod \"certified-operators-m5tf6\" (UID: \"6f0539ab-53f0-46bd-8ed2-13d507fae9ea\") " pod="openshift-marketplace/certified-operators-m5tf6" Oct 14 08:29:52 crc kubenswrapper[4870]: I1014 08:29:52.468346 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pr9wb\" (UniqueName: \"kubernetes.io/projected/6f0539ab-53f0-46bd-8ed2-13d507fae9ea-kube-api-access-pr9wb\") pod \"certified-operators-m5tf6\" (UID: \"6f0539ab-53f0-46bd-8ed2-13d507fae9ea\") " pod="openshift-marketplace/certified-operators-m5tf6" Oct 14 08:29:52 crc kubenswrapper[4870]: I1014 08:29:52.468740 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6f0539ab-53f0-46bd-8ed2-13d507fae9ea-utilities\") pod \"certified-operators-m5tf6\" (UID: \"6f0539ab-53f0-46bd-8ed2-13d507fae9ea\") " pod="openshift-marketplace/certified-operators-m5tf6" Oct 14 08:29:52 crc kubenswrapper[4870]: I1014 08:29:52.469177 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6f0539ab-53f0-46bd-8ed2-13d507fae9ea-catalog-content\") pod \"certified-operators-m5tf6\" (UID: \"6f0539ab-53f0-46bd-8ed2-13d507fae9ea\") " pod="openshift-marketplace/certified-operators-m5tf6" Oct 14 08:29:52 crc kubenswrapper[4870]: I1014 08:29:52.498576 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pr9wb\" (UniqueName: \"kubernetes.io/projected/6f0539ab-53f0-46bd-8ed2-13d507fae9ea-kube-api-access-pr9wb\") pod \"certified-operators-m5tf6\" (UID: \"6f0539ab-53f0-46bd-8ed2-13d507fae9ea\") " pod="openshift-marketplace/certified-operators-m5tf6" Oct 14 08:29:52 crc kubenswrapper[4870]: I1014 08:29:52.664427 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-m5tf6" Oct 14 08:29:52 crc kubenswrapper[4870]: I1014 08:29:52.940870 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-m5tf6"] Oct 14 08:29:53 crc kubenswrapper[4870]: I1014 08:29:53.895272 4870 generic.go:334] "Generic (PLEG): container finished" podID="6f0539ab-53f0-46bd-8ed2-13d507fae9ea" containerID="4337638da6a6a2004924f63309d0f3793b461a0b65c8b915c0de386a11c32e27" exitCode=0 Oct 14 08:29:53 crc kubenswrapper[4870]: I1014 08:29:53.895338 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-m5tf6" event={"ID":"6f0539ab-53f0-46bd-8ed2-13d507fae9ea","Type":"ContainerDied","Data":"4337638da6a6a2004924f63309d0f3793b461a0b65c8b915c0de386a11c32e27"} Oct 14 08:29:53 crc kubenswrapper[4870]: I1014 08:29:53.895812 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-m5tf6" event={"ID":"6f0539ab-53f0-46bd-8ed2-13d507fae9ea","Type":"ContainerStarted","Data":"8bdb1e5e95abf144ab8f5ceeef1982fcfa4af94ee1f70281ec3f169bccd20ce1"} Oct 14 08:29:54 crc kubenswrapper[4870]: I1014 08:29:54.908380 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-m5tf6" event={"ID":"6f0539ab-53f0-46bd-8ed2-13d507fae9ea","Type":"ContainerStarted","Data":"bb6a95c31fb302e27993e69e40d11404874c55d6d894ceddf63f2740d22ed150"} Oct 14 08:29:55 crc kubenswrapper[4870]: I1014 08:29:55.921829 4870 generic.go:334] "Generic (PLEG): container finished" podID="6f0539ab-53f0-46bd-8ed2-13d507fae9ea" containerID="bb6a95c31fb302e27993e69e40d11404874c55d6d894ceddf63f2740d22ed150" exitCode=0 Oct 14 08:29:55 crc kubenswrapper[4870]: I1014 08:29:55.921896 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-m5tf6" event={"ID":"6f0539ab-53f0-46bd-8ed2-13d507fae9ea","Type":"ContainerDied","Data":"bb6a95c31fb302e27993e69e40d11404874c55d6d894ceddf63f2740d22ed150"} Oct 14 08:29:56 crc kubenswrapper[4870]: I1014 08:29:56.933487 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-m5tf6" event={"ID":"6f0539ab-53f0-46bd-8ed2-13d507fae9ea","Type":"ContainerStarted","Data":"4e3418a010441a7a6e393f727cc896cc10236490b882265e11d2f327dbb3efbf"} Oct 14 08:29:56 crc kubenswrapper[4870]: I1014 08:29:56.955405 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-m5tf6" podStartSLOduration=2.352701145 podStartE2EDuration="4.955385928s" podCreationTimestamp="2025-10-14 08:29:52 +0000 UTC" firstStartedPulling="2025-10-14 08:29:53.897305024 +0000 UTC m=+5329.594665425" lastFinishedPulling="2025-10-14 08:29:56.499989797 +0000 UTC m=+5332.197350208" observedRunningTime="2025-10-14 08:29:56.953673646 +0000 UTC m=+5332.651034017" watchObservedRunningTime="2025-10-14 08:29:56.955385928 +0000 UTC m=+5332.652746299" Oct 14 08:30:00 crc kubenswrapper[4870]: I1014 08:30:00.150105 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29340510-vb6cv"] Oct 14 08:30:00 crc kubenswrapper[4870]: I1014 08:30:00.151170 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29340510-vb6cv" Oct 14 08:30:00 crc kubenswrapper[4870]: I1014 08:30:00.154758 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 14 08:30:00 crc kubenswrapper[4870]: I1014 08:30:00.156469 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 14 08:30:00 crc kubenswrapper[4870]: I1014 08:30:00.174994 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29340510-vb6cv"] Oct 14 08:30:00 crc kubenswrapper[4870]: I1014 08:30:00.203124 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6702b606-3463-40c8-934c-baaaf9735e01-config-volume\") pod \"collect-profiles-29340510-vb6cv\" (UID: \"6702b606-3463-40c8-934c-baaaf9735e01\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340510-vb6cv" Oct 14 08:30:00 crc kubenswrapper[4870]: I1014 08:30:00.203195 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n9fmr\" (UniqueName: \"kubernetes.io/projected/6702b606-3463-40c8-934c-baaaf9735e01-kube-api-access-n9fmr\") pod \"collect-profiles-29340510-vb6cv\" (UID: \"6702b606-3463-40c8-934c-baaaf9735e01\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340510-vb6cv" Oct 14 08:30:00 crc kubenswrapper[4870]: I1014 08:30:00.203218 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6702b606-3463-40c8-934c-baaaf9735e01-secret-volume\") pod \"collect-profiles-29340510-vb6cv\" (UID: \"6702b606-3463-40c8-934c-baaaf9735e01\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340510-vb6cv" Oct 14 08:30:00 crc kubenswrapper[4870]: I1014 08:30:00.304309 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6702b606-3463-40c8-934c-baaaf9735e01-config-volume\") pod \"collect-profiles-29340510-vb6cv\" (UID: \"6702b606-3463-40c8-934c-baaaf9735e01\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340510-vb6cv" Oct 14 08:30:00 crc kubenswrapper[4870]: I1014 08:30:00.304376 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n9fmr\" (UniqueName: \"kubernetes.io/projected/6702b606-3463-40c8-934c-baaaf9735e01-kube-api-access-n9fmr\") pod \"collect-profiles-29340510-vb6cv\" (UID: \"6702b606-3463-40c8-934c-baaaf9735e01\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340510-vb6cv" Oct 14 08:30:00 crc kubenswrapper[4870]: I1014 08:30:00.304411 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6702b606-3463-40c8-934c-baaaf9735e01-secret-volume\") pod \"collect-profiles-29340510-vb6cv\" (UID: \"6702b606-3463-40c8-934c-baaaf9735e01\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340510-vb6cv" Oct 14 08:30:00 crc kubenswrapper[4870]: I1014 08:30:00.305633 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6702b606-3463-40c8-934c-baaaf9735e01-config-volume\") pod \"collect-profiles-29340510-vb6cv\" (UID: \"6702b606-3463-40c8-934c-baaaf9735e01\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340510-vb6cv" Oct 14 08:30:00 crc kubenswrapper[4870]: I1014 08:30:00.318518 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6702b606-3463-40c8-934c-baaaf9735e01-secret-volume\") pod \"collect-profiles-29340510-vb6cv\" (UID: \"6702b606-3463-40c8-934c-baaaf9735e01\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340510-vb6cv" Oct 14 08:30:00 crc kubenswrapper[4870]: I1014 08:30:00.327795 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n9fmr\" (UniqueName: \"kubernetes.io/projected/6702b606-3463-40c8-934c-baaaf9735e01-kube-api-access-n9fmr\") pod \"collect-profiles-29340510-vb6cv\" (UID: \"6702b606-3463-40c8-934c-baaaf9735e01\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340510-vb6cv" Oct 14 08:30:00 crc kubenswrapper[4870]: I1014 08:30:00.476107 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29340510-vb6cv" Oct 14 08:30:00 crc kubenswrapper[4870]: I1014 08:30:00.794254 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29340510-vb6cv"] Oct 14 08:30:00 crc kubenswrapper[4870]: W1014 08:30:00.794930 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6702b606_3463_40c8_934c_baaaf9735e01.slice/crio-c925968d48228d6d1f12b4f927102af2d41b5369e8a15078659f99d260cdda54 WatchSource:0}: Error finding container c925968d48228d6d1f12b4f927102af2d41b5369e8a15078659f99d260cdda54: Status 404 returned error can't find the container with id c925968d48228d6d1f12b4f927102af2d41b5369e8a15078659f99d260cdda54 Oct 14 08:30:00 crc kubenswrapper[4870]: I1014 08:30:00.970516 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29340510-vb6cv" event={"ID":"6702b606-3463-40c8-934c-baaaf9735e01","Type":"ContainerStarted","Data":"50ccc31eebb6c05607156d85a7862b4203834eae708689e65c496ea8a5560964"} Oct 14 08:30:00 crc kubenswrapper[4870]: I1014 08:30:00.970881 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29340510-vb6cv" event={"ID":"6702b606-3463-40c8-934c-baaaf9735e01","Type":"ContainerStarted","Data":"c925968d48228d6d1f12b4f927102af2d41b5369e8a15078659f99d260cdda54"} Oct 14 08:30:00 crc kubenswrapper[4870]: I1014 08:30:00.990867 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29340510-vb6cv" podStartSLOduration=0.990843112 podStartE2EDuration="990.843112ms" podCreationTimestamp="2025-10-14 08:30:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 08:30:00.989232803 +0000 UTC m=+5336.686593184" watchObservedRunningTime="2025-10-14 08:30:00.990843112 +0000 UTC m=+5336.688203503" Oct 14 08:30:01 crc kubenswrapper[4870]: I1014 08:30:01.982882 4870 generic.go:334] "Generic (PLEG): container finished" podID="6702b606-3463-40c8-934c-baaaf9735e01" containerID="50ccc31eebb6c05607156d85a7862b4203834eae708689e65c496ea8a5560964" exitCode=0 Oct 14 08:30:01 crc kubenswrapper[4870]: I1014 08:30:01.982954 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29340510-vb6cv" event={"ID":"6702b606-3463-40c8-934c-baaaf9735e01","Type":"ContainerDied","Data":"50ccc31eebb6c05607156d85a7862b4203834eae708689e65c496ea8a5560964"} Oct 14 08:30:02 crc kubenswrapper[4870]: I1014 08:30:02.034401 4870 scope.go:117] "RemoveContainer" containerID="52468dffc62b6337feac3ee24b71903855af2b6b6e95f02951551f88ff46993a" Oct 14 08:30:02 crc kubenswrapper[4870]: E1014 08:30:02.035035 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 08:30:02 crc kubenswrapper[4870]: I1014 08:30:02.665129 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-m5tf6" Oct 14 08:30:02 crc kubenswrapper[4870]: I1014 08:30:02.665718 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-m5tf6" Oct 14 08:30:02 crc kubenswrapper[4870]: I1014 08:30:02.731619 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-m5tf6" Oct 14 08:30:03 crc kubenswrapper[4870]: I1014 08:30:03.047058 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-m5tf6" Oct 14 08:30:03 crc kubenswrapper[4870]: I1014 08:30:03.112590 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-m5tf6"] Oct 14 08:30:03 crc kubenswrapper[4870]: I1014 08:30:03.388970 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29340510-vb6cv" Oct 14 08:30:03 crc kubenswrapper[4870]: I1014 08:30:03.472552 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6702b606-3463-40c8-934c-baaaf9735e01-secret-volume\") pod \"6702b606-3463-40c8-934c-baaaf9735e01\" (UID: \"6702b606-3463-40c8-934c-baaaf9735e01\") " Oct 14 08:30:03 crc kubenswrapper[4870]: I1014 08:30:03.472621 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n9fmr\" (UniqueName: \"kubernetes.io/projected/6702b606-3463-40c8-934c-baaaf9735e01-kube-api-access-n9fmr\") pod \"6702b606-3463-40c8-934c-baaaf9735e01\" (UID: \"6702b606-3463-40c8-934c-baaaf9735e01\") " Oct 14 08:30:03 crc kubenswrapper[4870]: I1014 08:30:03.472752 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6702b606-3463-40c8-934c-baaaf9735e01-config-volume\") pod \"6702b606-3463-40c8-934c-baaaf9735e01\" (UID: \"6702b606-3463-40c8-934c-baaaf9735e01\") " Oct 14 08:30:03 crc kubenswrapper[4870]: I1014 08:30:03.473263 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6702b606-3463-40c8-934c-baaaf9735e01-config-volume" (OuterVolumeSpecName: "config-volume") pod "6702b606-3463-40c8-934c-baaaf9735e01" (UID: "6702b606-3463-40c8-934c-baaaf9735e01"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:30:03 crc kubenswrapper[4870]: I1014 08:30:03.477296 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6702b606-3463-40c8-934c-baaaf9735e01-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "6702b606-3463-40c8-934c-baaaf9735e01" (UID: "6702b606-3463-40c8-934c-baaaf9735e01"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:30:03 crc kubenswrapper[4870]: I1014 08:30:03.477860 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6702b606-3463-40c8-934c-baaaf9735e01-kube-api-access-n9fmr" (OuterVolumeSpecName: "kube-api-access-n9fmr") pod "6702b606-3463-40c8-934c-baaaf9735e01" (UID: "6702b606-3463-40c8-934c-baaaf9735e01"). InnerVolumeSpecName "kube-api-access-n9fmr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:30:03 crc kubenswrapper[4870]: I1014 08:30:03.574757 4870 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6702b606-3463-40c8-934c-baaaf9735e01-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 14 08:30:03 crc kubenswrapper[4870]: I1014 08:30:03.575162 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n9fmr\" (UniqueName: \"kubernetes.io/projected/6702b606-3463-40c8-934c-baaaf9735e01-kube-api-access-n9fmr\") on node \"crc\" DevicePath \"\"" Oct 14 08:30:03 crc kubenswrapper[4870]: I1014 08:30:03.575187 4870 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6702b606-3463-40c8-934c-baaaf9735e01-config-volume\") on node \"crc\" DevicePath \"\"" Oct 14 08:30:04 crc kubenswrapper[4870]: I1014 08:30:04.007075 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29340510-vb6cv" Oct 14 08:30:04 crc kubenswrapper[4870]: I1014 08:30:04.007062 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29340510-vb6cv" event={"ID":"6702b606-3463-40c8-934c-baaaf9735e01","Type":"ContainerDied","Data":"c925968d48228d6d1f12b4f927102af2d41b5369e8a15078659f99d260cdda54"} Oct 14 08:30:04 crc kubenswrapper[4870]: I1014 08:30:04.007181 4870 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c925968d48228d6d1f12b4f927102af2d41b5369e8a15078659f99d260cdda54" Oct 14 08:30:04 crc kubenswrapper[4870]: I1014 08:30:04.483502 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29340465-nfz9t"] Oct 14 08:30:04 crc kubenswrapper[4870]: I1014 08:30:04.494675 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29340465-nfz9t"] Oct 14 08:30:05 crc kubenswrapper[4870]: I1014 08:30:05.015702 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-m5tf6" podUID="6f0539ab-53f0-46bd-8ed2-13d507fae9ea" containerName="registry-server" containerID="cri-o://4e3418a010441a7a6e393f727cc896cc10236490b882265e11d2f327dbb3efbf" gracePeriod=2 Oct 14 08:30:05 crc kubenswrapper[4870]: I1014 08:30:05.065014 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="610412a5-52e1-4aa6-9d19-8f8a232eec01" path="/var/lib/kubelet/pods/610412a5-52e1-4aa6-9d19-8f8a232eec01/volumes" Oct 14 08:30:05 crc kubenswrapper[4870]: I1014 08:30:05.484153 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-m5tf6" Oct 14 08:30:05 crc kubenswrapper[4870]: I1014 08:30:05.604341 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6f0539ab-53f0-46bd-8ed2-13d507fae9ea-utilities\") pod \"6f0539ab-53f0-46bd-8ed2-13d507fae9ea\" (UID: \"6f0539ab-53f0-46bd-8ed2-13d507fae9ea\") " Oct 14 08:30:05 crc kubenswrapper[4870]: I1014 08:30:05.604389 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6f0539ab-53f0-46bd-8ed2-13d507fae9ea-catalog-content\") pod \"6f0539ab-53f0-46bd-8ed2-13d507fae9ea\" (UID: \"6f0539ab-53f0-46bd-8ed2-13d507fae9ea\") " Oct 14 08:30:05 crc kubenswrapper[4870]: I1014 08:30:05.604965 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pr9wb\" (UniqueName: \"kubernetes.io/projected/6f0539ab-53f0-46bd-8ed2-13d507fae9ea-kube-api-access-pr9wb\") pod \"6f0539ab-53f0-46bd-8ed2-13d507fae9ea\" (UID: \"6f0539ab-53f0-46bd-8ed2-13d507fae9ea\") " Oct 14 08:30:05 crc kubenswrapper[4870]: I1014 08:30:05.606704 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6f0539ab-53f0-46bd-8ed2-13d507fae9ea-utilities" (OuterVolumeSpecName: "utilities") pod "6f0539ab-53f0-46bd-8ed2-13d507fae9ea" (UID: "6f0539ab-53f0-46bd-8ed2-13d507fae9ea"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:30:05 crc kubenswrapper[4870]: I1014 08:30:05.611246 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6f0539ab-53f0-46bd-8ed2-13d507fae9ea-kube-api-access-pr9wb" (OuterVolumeSpecName: "kube-api-access-pr9wb") pod "6f0539ab-53f0-46bd-8ed2-13d507fae9ea" (UID: "6f0539ab-53f0-46bd-8ed2-13d507fae9ea"). InnerVolumeSpecName "kube-api-access-pr9wb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:30:05 crc kubenswrapper[4870]: I1014 08:30:05.686715 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6f0539ab-53f0-46bd-8ed2-13d507fae9ea-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6f0539ab-53f0-46bd-8ed2-13d507fae9ea" (UID: "6f0539ab-53f0-46bd-8ed2-13d507fae9ea"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:30:05 crc kubenswrapper[4870]: I1014 08:30:05.707329 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pr9wb\" (UniqueName: \"kubernetes.io/projected/6f0539ab-53f0-46bd-8ed2-13d507fae9ea-kube-api-access-pr9wb\") on node \"crc\" DevicePath \"\"" Oct 14 08:30:05 crc kubenswrapper[4870]: I1014 08:30:05.707376 4870 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6f0539ab-53f0-46bd-8ed2-13d507fae9ea-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 08:30:05 crc kubenswrapper[4870]: I1014 08:30:05.707397 4870 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6f0539ab-53f0-46bd-8ed2-13d507fae9ea-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 08:30:06 crc kubenswrapper[4870]: I1014 08:30:06.028481 4870 generic.go:334] "Generic (PLEG): container finished" podID="6f0539ab-53f0-46bd-8ed2-13d507fae9ea" containerID="4e3418a010441a7a6e393f727cc896cc10236490b882265e11d2f327dbb3efbf" exitCode=0 Oct 14 08:30:06 crc kubenswrapper[4870]: I1014 08:30:06.028552 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-m5tf6" Oct 14 08:30:06 crc kubenswrapper[4870]: I1014 08:30:06.028562 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-m5tf6" event={"ID":"6f0539ab-53f0-46bd-8ed2-13d507fae9ea","Type":"ContainerDied","Data":"4e3418a010441a7a6e393f727cc896cc10236490b882265e11d2f327dbb3efbf"} Oct 14 08:30:06 crc kubenswrapper[4870]: I1014 08:30:06.028709 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-m5tf6" event={"ID":"6f0539ab-53f0-46bd-8ed2-13d507fae9ea","Type":"ContainerDied","Data":"8bdb1e5e95abf144ab8f5ceeef1982fcfa4af94ee1f70281ec3f169bccd20ce1"} Oct 14 08:30:06 crc kubenswrapper[4870]: I1014 08:30:06.028748 4870 scope.go:117] "RemoveContainer" containerID="4e3418a010441a7a6e393f727cc896cc10236490b882265e11d2f327dbb3efbf" Oct 14 08:30:06 crc kubenswrapper[4870]: I1014 08:30:06.078822 4870 scope.go:117] "RemoveContainer" containerID="bb6a95c31fb302e27993e69e40d11404874c55d6d894ceddf63f2740d22ed150" Oct 14 08:30:06 crc kubenswrapper[4870]: I1014 08:30:06.088783 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-m5tf6"] Oct 14 08:30:06 crc kubenswrapper[4870]: I1014 08:30:06.099763 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-m5tf6"] Oct 14 08:30:06 crc kubenswrapper[4870]: I1014 08:30:06.116328 4870 scope.go:117] "RemoveContainer" containerID="4337638da6a6a2004924f63309d0f3793b461a0b65c8b915c0de386a11c32e27" Oct 14 08:30:06 crc kubenswrapper[4870]: I1014 08:30:06.167500 4870 scope.go:117] "RemoveContainer" containerID="4e3418a010441a7a6e393f727cc896cc10236490b882265e11d2f327dbb3efbf" Oct 14 08:30:06 crc kubenswrapper[4870]: E1014 08:30:06.168223 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4e3418a010441a7a6e393f727cc896cc10236490b882265e11d2f327dbb3efbf\": container with ID starting with 4e3418a010441a7a6e393f727cc896cc10236490b882265e11d2f327dbb3efbf not found: ID does not exist" containerID="4e3418a010441a7a6e393f727cc896cc10236490b882265e11d2f327dbb3efbf" Oct 14 08:30:06 crc kubenswrapper[4870]: I1014 08:30:06.168283 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4e3418a010441a7a6e393f727cc896cc10236490b882265e11d2f327dbb3efbf"} err="failed to get container status \"4e3418a010441a7a6e393f727cc896cc10236490b882265e11d2f327dbb3efbf\": rpc error: code = NotFound desc = could not find container \"4e3418a010441a7a6e393f727cc896cc10236490b882265e11d2f327dbb3efbf\": container with ID starting with 4e3418a010441a7a6e393f727cc896cc10236490b882265e11d2f327dbb3efbf not found: ID does not exist" Oct 14 08:30:06 crc kubenswrapper[4870]: I1014 08:30:06.168325 4870 scope.go:117] "RemoveContainer" containerID="bb6a95c31fb302e27993e69e40d11404874c55d6d894ceddf63f2740d22ed150" Oct 14 08:30:06 crc kubenswrapper[4870]: E1014 08:30:06.168959 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bb6a95c31fb302e27993e69e40d11404874c55d6d894ceddf63f2740d22ed150\": container with ID starting with bb6a95c31fb302e27993e69e40d11404874c55d6d894ceddf63f2740d22ed150 not found: ID does not exist" containerID="bb6a95c31fb302e27993e69e40d11404874c55d6d894ceddf63f2740d22ed150" Oct 14 08:30:06 crc kubenswrapper[4870]: I1014 08:30:06.169053 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bb6a95c31fb302e27993e69e40d11404874c55d6d894ceddf63f2740d22ed150"} err="failed to get container status \"bb6a95c31fb302e27993e69e40d11404874c55d6d894ceddf63f2740d22ed150\": rpc error: code = NotFound desc = could not find container \"bb6a95c31fb302e27993e69e40d11404874c55d6d894ceddf63f2740d22ed150\": container with ID starting with bb6a95c31fb302e27993e69e40d11404874c55d6d894ceddf63f2740d22ed150 not found: ID does not exist" Oct 14 08:30:06 crc kubenswrapper[4870]: I1014 08:30:06.169112 4870 scope.go:117] "RemoveContainer" containerID="4337638da6a6a2004924f63309d0f3793b461a0b65c8b915c0de386a11c32e27" Oct 14 08:30:06 crc kubenswrapper[4870]: E1014 08:30:06.169633 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4337638da6a6a2004924f63309d0f3793b461a0b65c8b915c0de386a11c32e27\": container with ID starting with 4337638da6a6a2004924f63309d0f3793b461a0b65c8b915c0de386a11c32e27 not found: ID does not exist" containerID="4337638da6a6a2004924f63309d0f3793b461a0b65c8b915c0de386a11c32e27" Oct 14 08:30:06 crc kubenswrapper[4870]: I1014 08:30:06.169702 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4337638da6a6a2004924f63309d0f3793b461a0b65c8b915c0de386a11c32e27"} err="failed to get container status \"4337638da6a6a2004924f63309d0f3793b461a0b65c8b915c0de386a11c32e27\": rpc error: code = NotFound desc = could not find container \"4337638da6a6a2004924f63309d0f3793b461a0b65c8b915c0de386a11c32e27\": container with ID starting with 4337638da6a6a2004924f63309d0f3793b461a0b65c8b915c0de386a11c32e27 not found: ID does not exist" Oct 14 08:30:07 crc kubenswrapper[4870]: I1014 08:30:07.056561 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6f0539ab-53f0-46bd-8ed2-13d507fae9ea" path="/var/lib/kubelet/pods/6f0539ab-53f0-46bd-8ed2-13d507fae9ea/volumes" Oct 14 08:30:13 crc kubenswrapper[4870]: I1014 08:30:13.989297 4870 scope.go:117] "RemoveContainer" containerID="2002983608d0a8250ea40b794c7803f7a6d8c52e7b23c7f7a9870044d70831c2" Oct 14 08:30:14 crc kubenswrapper[4870]: I1014 08:30:14.034177 4870 scope.go:117] "RemoveContainer" containerID="52468dffc62b6337feac3ee24b71903855af2b6b6e95f02951551f88ff46993a" Oct 14 08:30:14 crc kubenswrapper[4870]: E1014 08:30:14.034714 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 08:30:28 crc kubenswrapper[4870]: I1014 08:30:28.034160 4870 scope.go:117] "RemoveContainer" containerID="52468dffc62b6337feac3ee24b71903855af2b6b6e95f02951551f88ff46993a" Oct 14 08:30:28 crc kubenswrapper[4870]: E1014 08:30:28.035379 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 08:30:34 crc kubenswrapper[4870]: I1014 08:30:34.289087 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-9zljk"] Oct 14 08:30:34 crc kubenswrapper[4870]: E1014 08:30:34.290539 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6702b606-3463-40c8-934c-baaaf9735e01" containerName="collect-profiles" Oct 14 08:30:34 crc kubenswrapper[4870]: I1014 08:30:34.290573 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="6702b606-3463-40c8-934c-baaaf9735e01" containerName="collect-profiles" Oct 14 08:30:34 crc kubenswrapper[4870]: E1014 08:30:34.290612 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6f0539ab-53f0-46bd-8ed2-13d507fae9ea" containerName="registry-server" Oct 14 08:30:34 crc kubenswrapper[4870]: I1014 08:30:34.290630 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="6f0539ab-53f0-46bd-8ed2-13d507fae9ea" containerName="registry-server" Oct 14 08:30:34 crc kubenswrapper[4870]: E1014 08:30:34.290655 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6f0539ab-53f0-46bd-8ed2-13d507fae9ea" containerName="extract-content" Oct 14 08:30:34 crc kubenswrapper[4870]: I1014 08:30:34.290674 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="6f0539ab-53f0-46bd-8ed2-13d507fae9ea" containerName="extract-content" Oct 14 08:30:34 crc kubenswrapper[4870]: E1014 08:30:34.290714 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6f0539ab-53f0-46bd-8ed2-13d507fae9ea" containerName="extract-utilities" Oct 14 08:30:34 crc kubenswrapper[4870]: I1014 08:30:34.290733 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="6f0539ab-53f0-46bd-8ed2-13d507fae9ea" containerName="extract-utilities" Oct 14 08:30:34 crc kubenswrapper[4870]: I1014 08:30:34.291087 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="6f0539ab-53f0-46bd-8ed2-13d507fae9ea" containerName="registry-server" Oct 14 08:30:34 crc kubenswrapper[4870]: I1014 08:30:34.291161 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="6702b606-3463-40c8-934c-baaaf9735e01" containerName="collect-profiles" Oct 14 08:30:34 crc kubenswrapper[4870]: I1014 08:30:34.293620 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9zljk" Oct 14 08:30:34 crc kubenswrapper[4870]: I1014 08:30:34.298472 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-9zljk"] Oct 14 08:30:34 crc kubenswrapper[4870]: I1014 08:30:34.401347 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5dt8f\" (UniqueName: \"kubernetes.io/projected/1e1d704d-45a3-4f38-8394-221d1da341b0-kube-api-access-5dt8f\") pod \"community-operators-9zljk\" (UID: \"1e1d704d-45a3-4f38-8394-221d1da341b0\") " pod="openshift-marketplace/community-operators-9zljk" Oct 14 08:30:34 crc kubenswrapper[4870]: I1014 08:30:34.401476 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1e1d704d-45a3-4f38-8394-221d1da341b0-catalog-content\") pod \"community-operators-9zljk\" (UID: \"1e1d704d-45a3-4f38-8394-221d1da341b0\") " pod="openshift-marketplace/community-operators-9zljk" Oct 14 08:30:34 crc kubenswrapper[4870]: I1014 08:30:34.401649 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1e1d704d-45a3-4f38-8394-221d1da341b0-utilities\") pod \"community-operators-9zljk\" (UID: \"1e1d704d-45a3-4f38-8394-221d1da341b0\") " pod="openshift-marketplace/community-operators-9zljk" Oct 14 08:30:34 crc kubenswrapper[4870]: I1014 08:30:34.502687 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1e1d704d-45a3-4f38-8394-221d1da341b0-catalog-content\") pod \"community-operators-9zljk\" (UID: \"1e1d704d-45a3-4f38-8394-221d1da341b0\") " pod="openshift-marketplace/community-operators-9zljk" Oct 14 08:30:34 crc kubenswrapper[4870]: I1014 08:30:34.503107 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1e1d704d-45a3-4f38-8394-221d1da341b0-catalog-content\") pod \"community-operators-9zljk\" (UID: \"1e1d704d-45a3-4f38-8394-221d1da341b0\") " pod="openshift-marketplace/community-operators-9zljk" Oct 14 08:30:34 crc kubenswrapper[4870]: I1014 08:30:34.503238 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1e1d704d-45a3-4f38-8394-221d1da341b0-utilities\") pod \"community-operators-9zljk\" (UID: \"1e1d704d-45a3-4f38-8394-221d1da341b0\") " pod="openshift-marketplace/community-operators-9zljk" Oct 14 08:30:34 crc kubenswrapper[4870]: I1014 08:30:34.503502 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1e1d704d-45a3-4f38-8394-221d1da341b0-utilities\") pod \"community-operators-9zljk\" (UID: \"1e1d704d-45a3-4f38-8394-221d1da341b0\") " pod="openshift-marketplace/community-operators-9zljk" Oct 14 08:30:34 crc kubenswrapper[4870]: I1014 08:30:34.503903 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5dt8f\" (UniqueName: \"kubernetes.io/projected/1e1d704d-45a3-4f38-8394-221d1da341b0-kube-api-access-5dt8f\") pod \"community-operators-9zljk\" (UID: \"1e1d704d-45a3-4f38-8394-221d1da341b0\") " pod="openshift-marketplace/community-operators-9zljk" Oct 14 08:30:34 crc kubenswrapper[4870]: I1014 08:30:34.527938 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5dt8f\" (UniqueName: \"kubernetes.io/projected/1e1d704d-45a3-4f38-8394-221d1da341b0-kube-api-access-5dt8f\") pod \"community-operators-9zljk\" (UID: \"1e1d704d-45a3-4f38-8394-221d1da341b0\") " pod="openshift-marketplace/community-operators-9zljk" Oct 14 08:30:34 crc kubenswrapper[4870]: I1014 08:30:34.616738 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9zljk" Oct 14 08:30:35 crc kubenswrapper[4870]: I1014 08:30:35.193388 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-9zljk"] Oct 14 08:30:35 crc kubenswrapper[4870]: W1014 08:30:35.212505 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1e1d704d_45a3_4f38_8394_221d1da341b0.slice/crio-e4c16f8aa00666f2e5fe9c0765a4e9338a6f8fdc4d031cd6d53de0696417db80 WatchSource:0}: Error finding container e4c16f8aa00666f2e5fe9c0765a4e9338a6f8fdc4d031cd6d53de0696417db80: Status 404 returned error can't find the container with id e4c16f8aa00666f2e5fe9c0765a4e9338a6f8fdc4d031cd6d53de0696417db80 Oct 14 08:30:35 crc kubenswrapper[4870]: I1014 08:30:35.276724 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-xkxvz"] Oct 14 08:30:35 crc kubenswrapper[4870]: I1014 08:30:35.278496 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xkxvz" Oct 14 08:30:35 crc kubenswrapper[4870]: I1014 08:30:35.299240 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-xkxvz"] Oct 14 08:30:35 crc kubenswrapper[4870]: I1014 08:30:35.314965 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2fb1e2a7-e0cb-4a76-aa5d-5e1a79944c54-catalog-content\") pod \"redhat-marketplace-xkxvz\" (UID: \"2fb1e2a7-e0cb-4a76-aa5d-5e1a79944c54\") " pod="openshift-marketplace/redhat-marketplace-xkxvz" Oct 14 08:30:35 crc kubenswrapper[4870]: I1014 08:30:35.315044 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9wbbx\" (UniqueName: \"kubernetes.io/projected/2fb1e2a7-e0cb-4a76-aa5d-5e1a79944c54-kube-api-access-9wbbx\") pod \"redhat-marketplace-xkxvz\" (UID: \"2fb1e2a7-e0cb-4a76-aa5d-5e1a79944c54\") " pod="openshift-marketplace/redhat-marketplace-xkxvz" Oct 14 08:30:35 crc kubenswrapper[4870]: I1014 08:30:35.315298 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2fb1e2a7-e0cb-4a76-aa5d-5e1a79944c54-utilities\") pod \"redhat-marketplace-xkxvz\" (UID: \"2fb1e2a7-e0cb-4a76-aa5d-5e1a79944c54\") " pod="openshift-marketplace/redhat-marketplace-xkxvz" Oct 14 08:30:35 crc kubenswrapper[4870]: I1014 08:30:35.325046 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9zljk" event={"ID":"1e1d704d-45a3-4f38-8394-221d1da341b0","Type":"ContainerStarted","Data":"e4c16f8aa00666f2e5fe9c0765a4e9338a6f8fdc4d031cd6d53de0696417db80"} Oct 14 08:30:35 crc kubenswrapper[4870]: I1014 08:30:35.416403 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2fb1e2a7-e0cb-4a76-aa5d-5e1a79944c54-utilities\") pod \"redhat-marketplace-xkxvz\" (UID: \"2fb1e2a7-e0cb-4a76-aa5d-5e1a79944c54\") " pod="openshift-marketplace/redhat-marketplace-xkxvz" Oct 14 08:30:35 crc kubenswrapper[4870]: I1014 08:30:35.416500 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2fb1e2a7-e0cb-4a76-aa5d-5e1a79944c54-catalog-content\") pod \"redhat-marketplace-xkxvz\" (UID: \"2fb1e2a7-e0cb-4a76-aa5d-5e1a79944c54\") " pod="openshift-marketplace/redhat-marketplace-xkxvz" Oct 14 08:30:35 crc kubenswrapper[4870]: I1014 08:30:35.416539 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9wbbx\" (UniqueName: \"kubernetes.io/projected/2fb1e2a7-e0cb-4a76-aa5d-5e1a79944c54-kube-api-access-9wbbx\") pod \"redhat-marketplace-xkxvz\" (UID: \"2fb1e2a7-e0cb-4a76-aa5d-5e1a79944c54\") " pod="openshift-marketplace/redhat-marketplace-xkxvz" Oct 14 08:30:35 crc kubenswrapper[4870]: I1014 08:30:35.417180 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2fb1e2a7-e0cb-4a76-aa5d-5e1a79944c54-catalog-content\") pod \"redhat-marketplace-xkxvz\" (UID: \"2fb1e2a7-e0cb-4a76-aa5d-5e1a79944c54\") " pod="openshift-marketplace/redhat-marketplace-xkxvz" Oct 14 08:30:35 crc kubenswrapper[4870]: I1014 08:30:35.417476 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2fb1e2a7-e0cb-4a76-aa5d-5e1a79944c54-utilities\") pod \"redhat-marketplace-xkxvz\" (UID: \"2fb1e2a7-e0cb-4a76-aa5d-5e1a79944c54\") " pod="openshift-marketplace/redhat-marketplace-xkxvz" Oct 14 08:30:35 crc kubenswrapper[4870]: I1014 08:30:35.437670 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9wbbx\" (UniqueName: \"kubernetes.io/projected/2fb1e2a7-e0cb-4a76-aa5d-5e1a79944c54-kube-api-access-9wbbx\") pod \"redhat-marketplace-xkxvz\" (UID: \"2fb1e2a7-e0cb-4a76-aa5d-5e1a79944c54\") " pod="openshift-marketplace/redhat-marketplace-xkxvz" Oct 14 08:30:35 crc kubenswrapper[4870]: I1014 08:30:35.707194 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xkxvz" Oct 14 08:30:36 crc kubenswrapper[4870]: I1014 08:30:36.147806 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-xkxvz"] Oct 14 08:30:36 crc kubenswrapper[4870]: I1014 08:30:36.335991 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xkxvz" event={"ID":"2fb1e2a7-e0cb-4a76-aa5d-5e1a79944c54","Type":"ContainerStarted","Data":"37dda114ea0cc3e17a3a29fe4d9661b20fb473a863d1320c286fccfbb0f10001"} Oct 14 08:30:36 crc kubenswrapper[4870]: I1014 08:30:36.336058 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xkxvz" event={"ID":"2fb1e2a7-e0cb-4a76-aa5d-5e1a79944c54","Type":"ContainerStarted","Data":"6f11862947ed03b2ccc5424d10fd76fbb40de650a66c7ad7fb17c375a7c81925"} Oct 14 08:30:36 crc kubenswrapper[4870]: I1014 08:30:36.342376 4870 generic.go:334] "Generic (PLEG): container finished" podID="1e1d704d-45a3-4f38-8394-221d1da341b0" containerID="00ba238e8f65c09e86a8e4678315ca65a95aa86295de526ff09e0b7782d8bdcc" exitCode=0 Oct 14 08:30:36 crc kubenswrapper[4870]: I1014 08:30:36.342465 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9zljk" event={"ID":"1e1d704d-45a3-4f38-8394-221d1da341b0","Type":"ContainerDied","Data":"00ba238e8f65c09e86a8e4678315ca65a95aa86295de526ff09e0b7782d8bdcc"} Oct 14 08:30:37 crc kubenswrapper[4870]: I1014 08:30:37.352355 4870 generic.go:334] "Generic (PLEG): container finished" podID="2fb1e2a7-e0cb-4a76-aa5d-5e1a79944c54" containerID="37dda114ea0cc3e17a3a29fe4d9661b20fb473a863d1320c286fccfbb0f10001" exitCode=0 Oct 14 08:30:37 crc kubenswrapper[4870]: I1014 08:30:37.352513 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xkxvz" event={"ID":"2fb1e2a7-e0cb-4a76-aa5d-5e1a79944c54","Type":"ContainerDied","Data":"37dda114ea0cc3e17a3a29fe4d9661b20fb473a863d1320c286fccfbb0f10001"} Oct 14 08:30:38 crc kubenswrapper[4870]: I1014 08:30:38.363095 4870 generic.go:334] "Generic (PLEG): container finished" podID="1e1d704d-45a3-4f38-8394-221d1da341b0" containerID="98abf6ba36a988fabb6bad05ee94283e05e41b290ae3b5a3708aefeeb100958b" exitCode=0 Oct 14 08:30:38 crc kubenswrapper[4870]: I1014 08:30:38.363457 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9zljk" event={"ID":"1e1d704d-45a3-4f38-8394-221d1da341b0","Type":"ContainerDied","Data":"98abf6ba36a988fabb6bad05ee94283e05e41b290ae3b5a3708aefeeb100958b"} Oct 14 08:30:39 crc kubenswrapper[4870]: I1014 08:30:39.374924 4870 generic.go:334] "Generic (PLEG): container finished" podID="2fb1e2a7-e0cb-4a76-aa5d-5e1a79944c54" containerID="3984eb62f7fa11ff26b783ab2a9fc3030613f899ff48c9fcb4c13cf3ded0b916" exitCode=0 Oct 14 08:30:39 crc kubenswrapper[4870]: I1014 08:30:39.374961 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xkxvz" event={"ID":"2fb1e2a7-e0cb-4a76-aa5d-5e1a79944c54","Type":"ContainerDied","Data":"3984eb62f7fa11ff26b783ab2a9fc3030613f899ff48c9fcb4c13cf3ded0b916"} Oct 14 08:30:39 crc kubenswrapper[4870]: I1014 08:30:39.378897 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9zljk" event={"ID":"1e1d704d-45a3-4f38-8394-221d1da341b0","Type":"ContainerStarted","Data":"41d719ba40aa737920ec7cb0c8051443a2f2a5c78c156f8e24666090156cff5a"} Oct 14 08:30:39 crc kubenswrapper[4870]: I1014 08:30:39.420640 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-9zljk" podStartSLOduration=2.891104667 podStartE2EDuration="5.420619877s" podCreationTimestamp="2025-10-14 08:30:34 +0000 UTC" firstStartedPulling="2025-10-14 08:30:36.345382197 +0000 UTC m=+5372.042742568" lastFinishedPulling="2025-10-14 08:30:38.874897387 +0000 UTC m=+5374.572257778" observedRunningTime="2025-10-14 08:30:39.417938802 +0000 UTC m=+5375.115299183" watchObservedRunningTime="2025-10-14 08:30:39.420619877 +0000 UTC m=+5375.117980258" Oct 14 08:30:40 crc kubenswrapper[4870]: I1014 08:30:40.390708 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xkxvz" event={"ID":"2fb1e2a7-e0cb-4a76-aa5d-5e1a79944c54","Type":"ContainerStarted","Data":"2adbbafa9659ef18cdc6e3ea241673f5dfcf720233cded625edd1cd4dea062fb"} Oct 14 08:30:40 crc kubenswrapper[4870]: I1014 08:30:40.415061 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-xkxvz" podStartSLOduration=3.004475569 podStartE2EDuration="5.415042354s" podCreationTimestamp="2025-10-14 08:30:35 +0000 UTC" firstStartedPulling="2025-10-14 08:30:37.354140392 +0000 UTC m=+5373.051500763" lastFinishedPulling="2025-10-14 08:30:39.764707187 +0000 UTC m=+5375.462067548" observedRunningTime="2025-10-14 08:30:40.412500412 +0000 UTC m=+5376.109860793" watchObservedRunningTime="2025-10-14 08:30:40.415042354 +0000 UTC m=+5376.112402725" Oct 14 08:30:42 crc kubenswrapper[4870]: I1014 08:30:42.034664 4870 scope.go:117] "RemoveContainer" containerID="52468dffc62b6337feac3ee24b71903855af2b6b6e95f02951551f88ff46993a" Oct 14 08:30:42 crc kubenswrapper[4870]: E1014 08:30:42.035047 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 08:30:44 crc kubenswrapper[4870]: I1014 08:30:44.617605 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-9zljk" Oct 14 08:30:44 crc kubenswrapper[4870]: I1014 08:30:44.618026 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-9zljk" Oct 14 08:30:44 crc kubenswrapper[4870]: I1014 08:30:44.693725 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-9zljk" Oct 14 08:30:45 crc kubenswrapper[4870]: I1014 08:30:45.523812 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-9zljk" Oct 14 08:30:45 crc kubenswrapper[4870]: I1014 08:30:45.584294 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-9zljk"] Oct 14 08:30:45 crc kubenswrapper[4870]: I1014 08:30:45.708414 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-xkxvz" Oct 14 08:30:45 crc kubenswrapper[4870]: I1014 08:30:45.708519 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-xkxvz" Oct 14 08:30:45 crc kubenswrapper[4870]: I1014 08:30:45.783682 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-xkxvz" Oct 14 08:30:46 crc kubenswrapper[4870]: I1014 08:30:46.531107 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-xkxvz" Oct 14 08:30:47 crc kubenswrapper[4870]: I1014 08:30:47.468624 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-9zljk" podUID="1e1d704d-45a3-4f38-8394-221d1da341b0" containerName="registry-server" containerID="cri-o://41d719ba40aa737920ec7cb0c8051443a2f2a5c78c156f8e24666090156cff5a" gracePeriod=2 Oct 14 08:30:47 crc kubenswrapper[4870]: I1014 08:30:47.880743 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-xkxvz"] Oct 14 08:30:48 crc kubenswrapper[4870]: I1014 08:30:48.003798 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9zljk" Oct 14 08:30:48 crc kubenswrapper[4870]: I1014 08:30:48.117194 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1e1d704d-45a3-4f38-8394-221d1da341b0-catalog-content\") pod \"1e1d704d-45a3-4f38-8394-221d1da341b0\" (UID: \"1e1d704d-45a3-4f38-8394-221d1da341b0\") " Oct 14 08:30:48 crc kubenswrapper[4870]: I1014 08:30:48.117349 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1e1d704d-45a3-4f38-8394-221d1da341b0-utilities\") pod \"1e1d704d-45a3-4f38-8394-221d1da341b0\" (UID: \"1e1d704d-45a3-4f38-8394-221d1da341b0\") " Oct 14 08:30:48 crc kubenswrapper[4870]: I1014 08:30:48.117880 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5dt8f\" (UniqueName: \"kubernetes.io/projected/1e1d704d-45a3-4f38-8394-221d1da341b0-kube-api-access-5dt8f\") pod \"1e1d704d-45a3-4f38-8394-221d1da341b0\" (UID: \"1e1d704d-45a3-4f38-8394-221d1da341b0\") " Oct 14 08:30:48 crc kubenswrapper[4870]: I1014 08:30:48.118961 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1e1d704d-45a3-4f38-8394-221d1da341b0-utilities" (OuterVolumeSpecName: "utilities") pod "1e1d704d-45a3-4f38-8394-221d1da341b0" (UID: "1e1d704d-45a3-4f38-8394-221d1da341b0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:30:48 crc kubenswrapper[4870]: I1014 08:30:48.124035 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1e1d704d-45a3-4f38-8394-221d1da341b0-kube-api-access-5dt8f" (OuterVolumeSpecName: "kube-api-access-5dt8f") pod "1e1d704d-45a3-4f38-8394-221d1da341b0" (UID: "1e1d704d-45a3-4f38-8394-221d1da341b0"). InnerVolumeSpecName "kube-api-access-5dt8f". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:30:48 crc kubenswrapper[4870]: I1014 08:30:48.220469 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5dt8f\" (UniqueName: \"kubernetes.io/projected/1e1d704d-45a3-4f38-8394-221d1da341b0-kube-api-access-5dt8f\") on node \"crc\" DevicePath \"\"" Oct 14 08:30:48 crc kubenswrapper[4870]: I1014 08:30:48.220544 4870 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1e1d704d-45a3-4f38-8394-221d1da341b0-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 08:30:48 crc kubenswrapper[4870]: I1014 08:30:48.298561 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1e1d704d-45a3-4f38-8394-221d1da341b0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1e1d704d-45a3-4f38-8394-221d1da341b0" (UID: "1e1d704d-45a3-4f38-8394-221d1da341b0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:30:48 crc kubenswrapper[4870]: I1014 08:30:48.322105 4870 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1e1d704d-45a3-4f38-8394-221d1da341b0-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 08:30:48 crc kubenswrapper[4870]: I1014 08:30:48.482363 4870 generic.go:334] "Generic (PLEG): container finished" podID="1e1d704d-45a3-4f38-8394-221d1da341b0" containerID="41d719ba40aa737920ec7cb0c8051443a2f2a5c78c156f8e24666090156cff5a" exitCode=0 Oct 14 08:30:48 crc kubenswrapper[4870]: I1014 08:30:48.482488 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9zljk" event={"ID":"1e1d704d-45a3-4f38-8394-221d1da341b0","Type":"ContainerDied","Data":"41d719ba40aa737920ec7cb0c8051443a2f2a5c78c156f8e24666090156cff5a"} Oct 14 08:30:48 crc kubenswrapper[4870]: I1014 08:30:48.482560 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9zljk" event={"ID":"1e1d704d-45a3-4f38-8394-221d1da341b0","Type":"ContainerDied","Data":"e4c16f8aa00666f2e5fe9c0765a4e9338a6f8fdc4d031cd6d53de0696417db80"} Oct 14 08:30:48 crc kubenswrapper[4870]: I1014 08:30:48.482617 4870 scope.go:117] "RemoveContainer" containerID="41d719ba40aa737920ec7cb0c8051443a2f2a5c78c156f8e24666090156cff5a" Oct 14 08:30:48 crc kubenswrapper[4870]: I1014 08:30:48.482769 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-xkxvz" podUID="2fb1e2a7-e0cb-4a76-aa5d-5e1a79944c54" containerName="registry-server" containerID="cri-o://2adbbafa9659ef18cdc6e3ea241673f5dfcf720233cded625edd1cd4dea062fb" gracePeriod=2 Oct 14 08:30:48 crc kubenswrapper[4870]: I1014 08:30:48.483649 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9zljk" Oct 14 08:30:48 crc kubenswrapper[4870]: I1014 08:30:48.511604 4870 scope.go:117] "RemoveContainer" containerID="98abf6ba36a988fabb6bad05ee94283e05e41b290ae3b5a3708aefeeb100958b" Oct 14 08:30:48 crc kubenswrapper[4870]: I1014 08:30:48.546505 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-9zljk"] Oct 14 08:30:48 crc kubenswrapper[4870]: I1014 08:30:48.557192 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-9zljk"] Oct 14 08:30:48 crc kubenswrapper[4870]: I1014 08:30:48.587766 4870 scope.go:117] "RemoveContainer" containerID="00ba238e8f65c09e86a8e4678315ca65a95aa86295de526ff09e0b7782d8bdcc" Oct 14 08:30:48 crc kubenswrapper[4870]: I1014 08:30:48.657970 4870 scope.go:117] "RemoveContainer" containerID="41d719ba40aa737920ec7cb0c8051443a2f2a5c78c156f8e24666090156cff5a" Oct 14 08:30:48 crc kubenswrapper[4870]: E1014 08:30:48.658778 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"41d719ba40aa737920ec7cb0c8051443a2f2a5c78c156f8e24666090156cff5a\": container with ID starting with 41d719ba40aa737920ec7cb0c8051443a2f2a5c78c156f8e24666090156cff5a not found: ID does not exist" containerID="41d719ba40aa737920ec7cb0c8051443a2f2a5c78c156f8e24666090156cff5a" Oct 14 08:30:48 crc kubenswrapper[4870]: I1014 08:30:48.658850 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"41d719ba40aa737920ec7cb0c8051443a2f2a5c78c156f8e24666090156cff5a"} err="failed to get container status \"41d719ba40aa737920ec7cb0c8051443a2f2a5c78c156f8e24666090156cff5a\": rpc error: code = NotFound desc = could not find container \"41d719ba40aa737920ec7cb0c8051443a2f2a5c78c156f8e24666090156cff5a\": container with ID starting with 41d719ba40aa737920ec7cb0c8051443a2f2a5c78c156f8e24666090156cff5a not found: ID does not exist" Oct 14 08:30:48 crc kubenswrapper[4870]: I1014 08:30:48.658902 4870 scope.go:117] "RemoveContainer" containerID="98abf6ba36a988fabb6bad05ee94283e05e41b290ae3b5a3708aefeeb100958b" Oct 14 08:30:48 crc kubenswrapper[4870]: E1014 08:30:48.659566 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"98abf6ba36a988fabb6bad05ee94283e05e41b290ae3b5a3708aefeeb100958b\": container with ID starting with 98abf6ba36a988fabb6bad05ee94283e05e41b290ae3b5a3708aefeeb100958b not found: ID does not exist" containerID="98abf6ba36a988fabb6bad05ee94283e05e41b290ae3b5a3708aefeeb100958b" Oct 14 08:30:48 crc kubenswrapper[4870]: I1014 08:30:48.659623 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"98abf6ba36a988fabb6bad05ee94283e05e41b290ae3b5a3708aefeeb100958b"} err="failed to get container status \"98abf6ba36a988fabb6bad05ee94283e05e41b290ae3b5a3708aefeeb100958b\": rpc error: code = NotFound desc = could not find container \"98abf6ba36a988fabb6bad05ee94283e05e41b290ae3b5a3708aefeeb100958b\": container with ID starting with 98abf6ba36a988fabb6bad05ee94283e05e41b290ae3b5a3708aefeeb100958b not found: ID does not exist" Oct 14 08:30:48 crc kubenswrapper[4870]: I1014 08:30:48.659655 4870 scope.go:117] "RemoveContainer" containerID="00ba238e8f65c09e86a8e4678315ca65a95aa86295de526ff09e0b7782d8bdcc" Oct 14 08:30:48 crc kubenswrapper[4870]: E1014 08:30:48.660209 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"00ba238e8f65c09e86a8e4678315ca65a95aa86295de526ff09e0b7782d8bdcc\": container with ID starting with 00ba238e8f65c09e86a8e4678315ca65a95aa86295de526ff09e0b7782d8bdcc not found: ID does not exist" containerID="00ba238e8f65c09e86a8e4678315ca65a95aa86295de526ff09e0b7782d8bdcc" Oct 14 08:30:48 crc kubenswrapper[4870]: I1014 08:30:48.660256 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"00ba238e8f65c09e86a8e4678315ca65a95aa86295de526ff09e0b7782d8bdcc"} err="failed to get container status \"00ba238e8f65c09e86a8e4678315ca65a95aa86295de526ff09e0b7782d8bdcc\": rpc error: code = NotFound desc = could not find container \"00ba238e8f65c09e86a8e4678315ca65a95aa86295de526ff09e0b7782d8bdcc\": container with ID starting with 00ba238e8f65c09e86a8e4678315ca65a95aa86295de526ff09e0b7782d8bdcc not found: ID does not exist" Oct 14 08:30:48 crc kubenswrapper[4870]: I1014 08:30:48.959855 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xkxvz" Oct 14 08:30:49 crc kubenswrapper[4870]: I1014 08:30:49.040766 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9wbbx\" (UniqueName: \"kubernetes.io/projected/2fb1e2a7-e0cb-4a76-aa5d-5e1a79944c54-kube-api-access-9wbbx\") pod \"2fb1e2a7-e0cb-4a76-aa5d-5e1a79944c54\" (UID: \"2fb1e2a7-e0cb-4a76-aa5d-5e1a79944c54\") " Oct 14 08:30:49 crc kubenswrapper[4870]: I1014 08:30:49.040889 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2fb1e2a7-e0cb-4a76-aa5d-5e1a79944c54-utilities\") pod \"2fb1e2a7-e0cb-4a76-aa5d-5e1a79944c54\" (UID: \"2fb1e2a7-e0cb-4a76-aa5d-5e1a79944c54\") " Oct 14 08:30:49 crc kubenswrapper[4870]: I1014 08:30:49.040924 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2fb1e2a7-e0cb-4a76-aa5d-5e1a79944c54-catalog-content\") pod \"2fb1e2a7-e0cb-4a76-aa5d-5e1a79944c54\" (UID: \"2fb1e2a7-e0cb-4a76-aa5d-5e1a79944c54\") " Oct 14 08:30:49 crc kubenswrapper[4870]: I1014 08:30:49.042054 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2fb1e2a7-e0cb-4a76-aa5d-5e1a79944c54-utilities" (OuterVolumeSpecName: "utilities") pod "2fb1e2a7-e0cb-4a76-aa5d-5e1a79944c54" (UID: "2fb1e2a7-e0cb-4a76-aa5d-5e1a79944c54"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:30:49 crc kubenswrapper[4870]: I1014 08:30:49.045657 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1e1d704d-45a3-4f38-8394-221d1da341b0" path="/var/lib/kubelet/pods/1e1d704d-45a3-4f38-8394-221d1da341b0/volumes" Oct 14 08:30:49 crc kubenswrapper[4870]: I1014 08:30:49.046700 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2fb1e2a7-e0cb-4a76-aa5d-5e1a79944c54-kube-api-access-9wbbx" (OuterVolumeSpecName: "kube-api-access-9wbbx") pod "2fb1e2a7-e0cb-4a76-aa5d-5e1a79944c54" (UID: "2fb1e2a7-e0cb-4a76-aa5d-5e1a79944c54"). InnerVolumeSpecName "kube-api-access-9wbbx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:30:49 crc kubenswrapper[4870]: I1014 08:30:49.060717 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2fb1e2a7-e0cb-4a76-aa5d-5e1a79944c54-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2fb1e2a7-e0cb-4a76-aa5d-5e1a79944c54" (UID: "2fb1e2a7-e0cb-4a76-aa5d-5e1a79944c54"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:30:49 crc kubenswrapper[4870]: I1014 08:30:49.142899 4870 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2fb1e2a7-e0cb-4a76-aa5d-5e1a79944c54-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 08:30:49 crc kubenswrapper[4870]: I1014 08:30:49.142943 4870 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2fb1e2a7-e0cb-4a76-aa5d-5e1a79944c54-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 08:30:49 crc kubenswrapper[4870]: I1014 08:30:49.142964 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9wbbx\" (UniqueName: \"kubernetes.io/projected/2fb1e2a7-e0cb-4a76-aa5d-5e1a79944c54-kube-api-access-9wbbx\") on node \"crc\" DevicePath \"\"" Oct 14 08:30:49 crc kubenswrapper[4870]: I1014 08:30:49.496009 4870 generic.go:334] "Generic (PLEG): container finished" podID="2fb1e2a7-e0cb-4a76-aa5d-5e1a79944c54" containerID="2adbbafa9659ef18cdc6e3ea241673f5dfcf720233cded625edd1cd4dea062fb" exitCode=0 Oct 14 08:30:49 crc kubenswrapper[4870]: I1014 08:30:49.496138 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xkxvz" Oct 14 08:30:49 crc kubenswrapper[4870]: I1014 08:30:49.496166 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xkxvz" event={"ID":"2fb1e2a7-e0cb-4a76-aa5d-5e1a79944c54","Type":"ContainerDied","Data":"2adbbafa9659ef18cdc6e3ea241673f5dfcf720233cded625edd1cd4dea062fb"} Oct 14 08:30:49 crc kubenswrapper[4870]: I1014 08:30:49.497333 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xkxvz" event={"ID":"2fb1e2a7-e0cb-4a76-aa5d-5e1a79944c54","Type":"ContainerDied","Data":"6f11862947ed03b2ccc5424d10fd76fbb40de650a66c7ad7fb17c375a7c81925"} Oct 14 08:30:49 crc kubenswrapper[4870]: I1014 08:30:49.497367 4870 scope.go:117] "RemoveContainer" containerID="2adbbafa9659ef18cdc6e3ea241673f5dfcf720233cded625edd1cd4dea062fb" Oct 14 08:30:49 crc kubenswrapper[4870]: I1014 08:30:49.554353 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-xkxvz"] Oct 14 08:30:49 crc kubenswrapper[4870]: I1014 08:30:49.558582 4870 scope.go:117] "RemoveContainer" containerID="3984eb62f7fa11ff26b783ab2a9fc3030613f899ff48c9fcb4c13cf3ded0b916" Oct 14 08:30:49 crc kubenswrapper[4870]: I1014 08:30:49.562193 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-xkxvz"] Oct 14 08:30:49 crc kubenswrapper[4870]: I1014 08:30:49.584063 4870 scope.go:117] "RemoveContainer" containerID="37dda114ea0cc3e17a3a29fe4d9661b20fb473a863d1320c286fccfbb0f10001" Oct 14 08:30:49 crc kubenswrapper[4870]: I1014 08:30:49.605062 4870 scope.go:117] "RemoveContainer" containerID="2adbbafa9659ef18cdc6e3ea241673f5dfcf720233cded625edd1cd4dea062fb" Oct 14 08:30:49 crc kubenswrapper[4870]: E1014 08:30:49.605955 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2adbbafa9659ef18cdc6e3ea241673f5dfcf720233cded625edd1cd4dea062fb\": container with ID starting with 2adbbafa9659ef18cdc6e3ea241673f5dfcf720233cded625edd1cd4dea062fb not found: ID does not exist" containerID="2adbbafa9659ef18cdc6e3ea241673f5dfcf720233cded625edd1cd4dea062fb" Oct 14 08:30:49 crc kubenswrapper[4870]: I1014 08:30:49.606017 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2adbbafa9659ef18cdc6e3ea241673f5dfcf720233cded625edd1cd4dea062fb"} err="failed to get container status \"2adbbafa9659ef18cdc6e3ea241673f5dfcf720233cded625edd1cd4dea062fb\": rpc error: code = NotFound desc = could not find container \"2adbbafa9659ef18cdc6e3ea241673f5dfcf720233cded625edd1cd4dea062fb\": container with ID starting with 2adbbafa9659ef18cdc6e3ea241673f5dfcf720233cded625edd1cd4dea062fb not found: ID does not exist" Oct 14 08:30:49 crc kubenswrapper[4870]: I1014 08:30:49.606054 4870 scope.go:117] "RemoveContainer" containerID="3984eb62f7fa11ff26b783ab2a9fc3030613f899ff48c9fcb4c13cf3ded0b916" Oct 14 08:30:49 crc kubenswrapper[4870]: E1014 08:30:49.606541 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3984eb62f7fa11ff26b783ab2a9fc3030613f899ff48c9fcb4c13cf3ded0b916\": container with ID starting with 3984eb62f7fa11ff26b783ab2a9fc3030613f899ff48c9fcb4c13cf3ded0b916 not found: ID does not exist" containerID="3984eb62f7fa11ff26b783ab2a9fc3030613f899ff48c9fcb4c13cf3ded0b916" Oct 14 08:30:49 crc kubenswrapper[4870]: I1014 08:30:49.606568 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3984eb62f7fa11ff26b783ab2a9fc3030613f899ff48c9fcb4c13cf3ded0b916"} err="failed to get container status \"3984eb62f7fa11ff26b783ab2a9fc3030613f899ff48c9fcb4c13cf3ded0b916\": rpc error: code = NotFound desc = could not find container \"3984eb62f7fa11ff26b783ab2a9fc3030613f899ff48c9fcb4c13cf3ded0b916\": container with ID starting with 3984eb62f7fa11ff26b783ab2a9fc3030613f899ff48c9fcb4c13cf3ded0b916 not found: ID does not exist" Oct 14 08:30:49 crc kubenswrapper[4870]: I1014 08:30:49.606586 4870 scope.go:117] "RemoveContainer" containerID="37dda114ea0cc3e17a3a29fe4d9661b20fb473a863d1320c286fccfbb0f10001" Oct 14 08:30:49 crc kubenswrapper[4870]: E1014 08:30:49.606859 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"37dda114ea0cc3e17a3a29fe4d9661b20fb473a863d1320c286fccfbb0f10001\": container with ID starting with 37dda114ea0cc3e17a3a29fe4d9661b20fb473a863d1320c286fccfbb0f10001 not found: ID does not exist" containerID="37dda114ea0cc3e17a3a29fe4d9661b20fb473a863d1320c286fccfbb0f10001" Oct 14 08:30:49 crc kubenswrapper[4870]: I1014 08:30:49.606885 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"37dda114ea0cc3e17a3a29fe4d9661b20fb473a863d1320c286fccfbb0f10001"} err="failed to get container status \"37dda114ea0cc3e17a3a29fe4d9661b20fb473a863d1320c286fccfbb0f10001\": rpc error: code = NotFound desc = could not find container \"37dda114ea0cc3e17a3a29fe4d9661b20fb473a863d1320c286fccfbb0f10001\": container with ID starting with 37dda114ea0cc3e17a3a29fe4d9661b20fb473a863d1320c286fccfbb0f10001 not found: ID does not exist" Oct 14 08:30:51 crc kubenswrapper[4870]: I1014 08:30:51.053326 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2fb1e2a7-e0cb-4a76-aa5d-5e1a79944c54" path="/var/lib/kubelet/pods/2fb1e2a7-e0cb-4a76-aa5d-5e1a79944c54/volumes" Oct 14 08:30:55 crc kubenswrapper[4870]: I1014 08:30:55.043570 4870 scope.go:117] "RemoveContainer" containerID="52468dffc62b6337feac3ee24b71903855af2b6b6e95f02951551f88ff46993a" Oct 14 08:30:55 crc kubenswrapper[4870]: E1014 08:30:55.044505 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 08:31:09 crc kubenswrapper[4870]: I1014 08:31:09.034028 4870 scope.go:117] "RemoveContainer" containerID="52468dffc62b6337feac3ee24b71903855af2b6b6e95f02951551f88ff46993a" Oct 14 08:31:09 crc kubenswrapper[4870]: E1014 08:31:09.035214 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 08:31:22 crc kubenswrapper[4870]: I1014 08:31:22.034103 4870 scope.go:117] "RemoveContainer" containerID="52468dffc62b6337feac3ee24b71903855af2b6b6e95f02951551f88ff46993a" Oct 14 08:31:22 crc kubenswrapper[4870]: E1014 08:31:22.035128 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 08:31:36 crc kubenswrapper[4870]: I1014 08:31:36.035391 4870 scope.go:117] "RemoveContainer" containerID="52468dffc62b6337feac3ee24b71903855af2b6b6e95f02951551f88ff46993a" Oct 14 08:31:36 crc kubenswrapper[4870]: E1014 08:31:36.036811 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 08:31:50 crc kubenswrapper[4870]: I1014 08:31:50.034054 4870 scope.go:117] "RemoveContainer" containerID="52468dffc62b6337feac3ee24b71903855af2b6b6e95f02951551f88ff46993a" Oct 14 08:31:50 crc kubenswrapper[4870]: E1014 08:31:50.035206 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 08:32:05 crc kubenswrapper[4870]: I1014 08:32:05.046366 4870 scope.go:117] "RemoveContainer" containerID="52468dffc62b6337feac3ee24b71903855af2b6b6e95f02951551f88ff46993a" Oct 14 08:32:05 crc kubenswrapper[4870]: E1014 08:32:05.047980 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 08:32:19 crc kubenswrapper[4870]: I1014 08:32:19.034355 4870 scope.go:117] "RemoveContainer" containerID="52468dffc62b6337feac3ee24b71903855af2b6b6e95f02951551f88ff46993a" Oct 14 08:32:19 crc kubenswrapper[4870]: E1014 08:32:19.036190 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 08:32:33 crc kubenswrapper[4870]: I1014 08:32:33.035138 4870 scope.go:117] "RemoveContainer" containerID="52468dffc62b6337feac3ee24b71903855af2b6b6e95f02951551f88ff46993a" Oct 14 08:32:33 crc kubenswrapper[4870]: E1014 08:32:33.036131 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 08:32:44 crc kubenswrapper[4870]: I1014 08:32:44.034381 4870 scope.go:117] "RemoveContainer" containerID="52468dffc62b6337feac3ee24b71903855af2b6b6e95f02951551f88ff46993a" Oct 14 08:32:44 crc kubenswrapper[4870]: E1014 08:32:44.035529 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 08:32:58 crc kubenswrapper[4870]: I1014 08:32:58.034014 4870 scope.go:117] "RemoveContainer" containerID="52468dffc62b6337feac3ee24b71903855af2b6b6e95f02951551f88ff46993a" Oct 14 08:32:58 crc kubenswrapper[4870]: E1014 08:32:58.035118 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 08:33:11 crc kubenswrapper[4870]: I1014 08:33:11.034519 4870 scope.go:117] "RemoveContainer" containerID="52468dffc62b6337feac3ee24b71903855af2b6b6e95f02951551f88ff46993a" Oct 14 08:33:11 crc kubenswrapper[4870]: E1014 08:33:11.036193 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 08:33:24 crc kubenswrapper[4870]: I1014 08:33:24.034551 4870 scope.go:117] "RemoveContainer" containerID="52468dffc62b6337feac3ee24b71903855af2b6b6e95f02951551f88ff46993a" Oct 14 08:33:24 crc kubenswrapper[4870]: E1014 08:33:24.035830 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 08:33:39 crc kubenswrapper[4870]: I1014 08:33:39.034594 4870 scope.go:117] "RemoveContainer" containerID="52468dffc62b6337feac3ee24b71903855af2b6b6e95f02951551f88ff46993a" Oct 14 08:33:39 crc kubenswrapper[4870]: E1014 08:33:39.035706 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 08:33:54 crc kubenswrapper[4870]: I1014 08:33:54.034465 4870 scope.go:117] "RemoveContainer" containerID="52468dffc62b6337feac3ee24b71903855af2b6b6e95f02951551f88ff46993a" Oct 14 08:33:54 crc kubenswrapper[4870]: E1014 08:33:54.035123 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 08:34:08 crc kubenswrapper[4870]: I1014 08:34:08.055114 4870 scope.go:117] "RemoveContainer" containerID="52468dffc62b6337feac3ee24b71903855af2b6b6e95f02951551f88ff46993a" Oct 14 08:34:08 crc kubenswrapper[4870]: E1014 08:34:08.056217 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 08:34:19 crc kubenswrapper[4870]: I1014 08:34:19.034233 4870 scope.go:117] "RemoveContainer" containerID="52468dffc62b6337feac3ee24b71903855af2b6b6e95f02951551f88ff46993a" Oct 14 08:34:19 crc kubenswrapper[4870]: E1014 08:34:19.035568 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 08:34:33 crc kubenswrapper[4870]: I1014 08:34:33.034013 4870 scope.go:117] "RemoveContainer" containerID="52468dffc62b6337feac3ee24b71903855af2b6b6e95f02951551f88ff46993a" Oct 14 08:34:33 crc kubenswrapper[4870]: I1014 08:34:33.660049 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" event={"ID":"874a5e45-dffd-4d17-b609-b7d3ed2eab07","Type":"ContainerStarted","Data":"6f0398b7e7959f65985e6ee24c4434fc79307d6cce2fb874cfc4c5f50459984b"} Oct 14 08:35:51 crc kubenswrapper[4870]: I1014 08:35:51.816705 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["crc-storage/crc-storage-crc-d2ngg"] Oct 14 08:35:51 crc kubenswrapper[4870]: I1014 08:35:51.827811 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["crc-storage/crc-storage-crc-d2ngg"] Oct 14 08:35:51 crc kubenswrapper[4870]: I1014 08:35:51.982918 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["crc-storage/crc-storage-crc-dwrfd"] Oct 14 08:35:51 crc kubenswrapper[4870]: E1014 08:35:51.983184 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e1d704d-45a3-4f38-8394-221d1da341b0" containerName="extract-content" Oct 14 08:35:51 crc kubenswrapper[4870]: I1014 08:35:51.983199 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e1d704d-45a3-4f38-8394-221d1da341b0" containerName="extract-content" Oct 14 08:35:51 crc kubenswrapper[4870]: E1014 08:35:51.983210 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2fb1e2a7-e0cb-4a76-aa5d-5e1a79944c54" containerName="extract-content" Oct 14 08:35:51 crc kubenswrapper[4870]: I1014 08:35:51.983217 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="2fb1e2a7-e0cb-4a76-aa5d-5e1a79944c54" containerName="extract-content" Oct 14 08:35:51 crc kubenswrapper[4870]: E1014 08:35:51.983228 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2fb1e2a7-e0cb-4a76-aa5d-5e1a79944c54" containerName="registry-server" Oct 14 08:35:51 crc kubenswrapper[4870]: I1014 08:35:51.983235 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="2fb1e2a7-e0cb-4a76-aa5d-5e1a79944c54" containerName="registry-server" Oct 14 08:35:51 crc kubenswrapper[4870]: E1014 08:35:51.983249 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e1d704d-45a3-4f38-8394-221d1da341b0" containerName="extract-utilities" Oct 14 08:35:51 crc kubenswrapper[4870]: I1014 08:35:51.983255 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e1d704d-45a3-4f38-8394-221d1da341b0" containerName="extract-utilities" Oct 14 08:35:51 crc kubenswrapper[4870]: E1014 08:35:51.983262 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e1d704d-45a3-4f38-8394-221d1da341b0" containerName="registry-server" Oct 14 08:35:51 crc kubenswrapper[4870]: I1014 08:35:51.983267 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e1d704d-45a3-4f38-8394-221d1da341b0" containerName="registry-server" Oct 14 08:35:51 crc kubenswrapper[4870]: E1014 08:35:51.983280 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2fb1e2a7-e0cb-4a76-aa5d-5e1a79944c54" containerName="extract-utilities" Oct 14 08:35:51 crc kubenswrapper[4870]: I1014 08:35:51.983286 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="2fb1e2a7-e0cb-4a76-aa5d-5e1a79944c54" containerName="extract-utilities" Oct 14 08:35:51 crc kubenswrapper[4870]: I1014 08:35:51.983418 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="1e1d704d-45a3-4f38-8394-221d1da341b0" containerName="registry-server" Oct 14 08:35:51 crc kubenswrapper[4870]: I1014 08:35:51.983436 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="2fb1e2a7-e0cb-4a76-aa5d-5e1a79944c54" containerName="registry-server" Oct 14 08:35:51 crc kubenswrapper[4870]: I1014 08:35:51.983875 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-dwrfd" Oct 14 08:35:51 crc kubenswrapper[4870]: I1014 08:35:51.986949 4870 reflector.go:368] Caches populated for *v1.Secret from object-"crc-storage"/"crc-storage-dockercfg-pm8px" Oct 14 08:35:51 crc kubenswrapper[4870]: I1014 08:35:51.986948 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"kube-root-ca.crt" Oct 14 08:35:51 crc kubenswrapper[4870]: I1014 08:35:51.987546 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"crc-storage" Oct 14 08:35:51 crc kubenswrapper[4870]: I1014 08:35:51.988244 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"openshift-service-ca.crt" Oct 14 08:35:52 crc kubenswrapper[4870]: I1014 08:35:52.003329 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-dwrfd"] Oct 14 08:35:52 crc kubenswrapper[4870]: I1014 08:35:52.123846 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-58hx5\" (UniqueName: \"kubernetes.io/projected/b2d12dd0-18d6-4507-855e-639b7ef2fc2b-kube-api-access-58hx5\") pod \"crc-storage-crc-dwrfd\" (UID: \"b2d12dd0-18d6-4507-855e-639b7ef2fc2b\") " pod="crc-storage/crc-storage-crc-dwrfd" Oct 14 08:35:52 crc kubenswrapper[4870]: I1014 08:35:52.124070 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/b2d12dd0-18d6-4507-855e-639b7ef2fc2b-crc-storage\") pod \"crc-storage-crc-dwrfd\" (UID: \"b2d12dd0-18d6-4507-855e-639b7ef2fc2b\") " pod="crc-storage/crc-storage-crc-dwrfd" Oct 14 08:35:52 crc kubenswrapper[4870]: I1014 08:35:52.124159 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/b2d12dd0-18d6-4507-855e-639b7ef2fc2b-node-mnt\") pod \"crc-storage-crc-dwrfd\" (UID: \"b2d12dd0-18d6-4507-855e-639b7ef2fc2b\") " pod="crc-storage/crc-storage-crc-dwrfd" Oct 14 08:35:52 crc kubenswrapper[4870]: I1014 08:35:52.226086 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-58hx5\" (UniqueName: \"kubernetes.io/projected/b2d12dd0-18d6-4507-855e-639b7ef2fc2b-kube-api-access-58hx5\") pod \"crc-storage-crc-dwrfd\" (UID: \"b2d12dd0-18d6-4507-855e-639b7ef2fc2b\") " pod="crc-storage/crc-storage-crc-dwrfd" Oct 14 08:35:52 crc kubenswrapper[4870]: I1014 08:35:52.226199 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/b2d12dd0-18d6-4507-855e-639b7ef2fc2b-crc-storage\") pod \"crc-storage-crc-dwrfd\" (UID: \"b2d12dd0-18d6-4507-855e-639b7ef2fc2b\") " pod="crc-storage/crc-storage-crc-dwrfd" Oct 14 08:35:52 crc kubenswrapper[4870]: I1014 08:35:52.226259 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/b2d12dd0-18d6-4507-855e-639b7ef2fc2b-node-mnt\") pod \"crc-storage-crc-dwrfd\" (UID: \"b2d12dd0-18d6-4507-855e-639b7ef2fc2b\") " pod="crc-storage/crc-storage-crc-dwrfd" Oct 14 08:35:52 crc kubenswrapper[4870]: I1014 08:35:52.226822 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/b2d12dd0-18d6-4507-855e-639b7ef2fc2b-node-mnt\") pod \"crc-storage-crc-dwrfd\" (UID: \"b2d12dd0-18d6-4507-855e-639b7ef2fc2b\") " pod="crc-storage/crc-storage-crc-dwrfd" Oct 14 08:35:52 crc kubenswrapper[4870]: I1014 08:35:52.227670 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/b2d12dd0-18d6-4507-855e-639b7ef2fc2b-crc-storage\") pod \"crc-storage-crc-dwrfd\" (UID: \"b2d12dd0-18d6-4507-855e-639b7ef2fc2b\") " pod="crc-storage/crc-storage-crc-dwrfd" Oct 14 08:35:52 crc kubenswrapper[4870]: I1014 08:35:52.251742 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-58hx5\" (UniqueName: \"kubernetes.io/projected/b2d12dd0-18d6-4507-855e-639b7ef2fc2b-kube-api-access-58hx5\") pod \"crc-storage-crc-dwrfd\" (UID: \"b2d12dd0-18d6-4507-855e-639b7ef2fc2b\") " pod="crc-storage/crc-storage-crc-dwrfd" Oct 14 08:35:52 crc kubenswrapper[4870]: I1014 08:35:52.311296 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-dwrfd" Oct 14 08:35:52 crc kubenswrapper[4870]: I1014 08:35:52.820434 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-dwrfd"] Oct 14 08:35:52 crc kubenswrapper[4870]: I1014 08:35:52.829117 4870 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 14 08:35:53 crc kubenswrapper[4870]: I1014 08:35:53.051243 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="436866a2-a717-4431-aefe-56a515024a72" path="/var/lib/kubelet/pods/436866a2-a717-4431-aefe-56a515024a72/volumes" Oct 14 08:35:53 crc kubenswrapper[4870]: I1014 08:35:53.404205 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-dwrfd" event={"ID":"b2d12dd0-18d6-4507-855e-639b7ef2fc2b","Type":"ContainerStarted","Data":"8036524002f5b294344ba9c85a76e9e5914f6ac971151b715c58b4f8c7f0025b"} Oct 14 08:35:54 crc kubenswrapper[4870]: I1014 08:35:54.417730 4870 generic.go:334] "Generic (PLEG): container finished" podID="b2d12dd0-18d6-4507-855e-639b7ef2fc2b" containerID="38057b445ce0f34a4b8f85cd8b85ce6aebbbdae879eb23826923ab6094445017" exitCode=0 Oct 14 08:35:54 crc kubenswrapper[4870]: I1014 08:35:54.417816 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-dwrfd" event={"ID":"b2d12dd0-18d6-4507-855e-639b7ef2fc2b","Type":"ContainerDied","Data":"38057b445ce0f34a4b8f85cd8b85ce6aebbbdae879eb23826923ab6094445017"} Oct 14 08:35:55 crc kubenswrapper[4870]: I1014 08:35:55.769928 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-dwrfd" Oct 14 08:35:55 crc kubenswrapper[4870]: I1014 08:35:55.884116 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/b2d12dd0-18d6-4507-855e-639b7ef2fc2b-crc-storage\") pod \"b2d12dd0-18d6-4507-855e-639b7ef2fc2b\" (UID: \"b2d12dd0-18d6-4507-855e-639b7ef2fc2b\") " Oct 14 08:35:55 crc kubenswrapper[4870]: I1014 08:35:55.884266 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/b2d12dd0-18d6-4507-855e-639b7ef2fc2b-node-mnt\") pod \"b2d12dd0-18d6-4507-855e-639b7ef2fc2b\" (UID: \"b2d12dd0-18d6-4507-855e-639b7ef2fc2b\") " Oct 14 08:35:55 crc kubenswrapper[4870]: I1014 08:35:55.884578 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b2d12dd0-18d6-4507-855e-639b7ef2fc2b-node-mnt" (OuterVolumeSpecName: "node-mnt") pod "b2d12dd0-18d6-4507-855e-639b7ef2fc2b" (UID: "b2d12dd0-18d6-4507-855e-639b7ef2fc2b"). InnerVolumeSpecName "node-mnt". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 14 08:35:55 crc kubenswrapper[4870]: I1014 08:35:55.885588 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-58hx5\" (UniqueName: \"kubernetes.io/projected/b2d12dd0-18d6-4507-855e-639b7ef2fc2b-kube-api-access-58hx5\") pod \"b2d12dd0-18d6-4507-855e-639b7ef2fc2b\" (UID: \"b2d12dd0-18d6-4507-855e-639b7ef2fc2b\") " Oct 14 08:35:55 crc kubenswrapper[4870]: I1014 08:35:55.886345 4870 reconciler_common.go:293] "Volume detached for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/b2d12dd0-18d6-4507-855e-639b7ef2fc2b-node-mnt\") on node \"crc\" DevicePath \"\"" Oct 14 08:35:55 crc kubenswrapper[4870]: I1014 08:35:55.894314 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b2d12dd0-18d6-4507-855e-639b7ef2fc2b-kube-api-access-58hx5" (OuterVolumeSpecName: "kube-api-access-58hx5") pod "b2d12dd0-18d6-4507-855e-639b7ef2fc2b" (UID: "b2d12dd0-18d6-4507-855e-639b7ef2fc2b"). InnerVolumeSpecName "kube-api-access-58hx5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:35:55 crc kubenswrapper[4870]: I1014 08:35:55.911570 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b2d12dd0-18d6-4507-855e-639b7ef2fc2b-crc-storage" (OuterVolumeSpecName: "crc-storage") pod "b2d12dd0-18d6-4507-855e-639b7ef2fc2b" (UID: "b2d12dd0-18d6-4507-855e-639b7ef2fc2b"). InnerVolumeSpecName "crc-storage". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:35:55 crc kubenswrapper[4870]: I1014 08:35:55.987777 4870 reconciler_common.go:293] "Volume detached for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/b2d12dd0-18d6-4507-855e-639b7ef2fc2b-crc-storage\") on node \"crc\" DevicePath \"\"" Oct 14 08:35:55 crc kubenswrapper[4870]: I1014 08:35:55.987815 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-58hx5\" (UniqueName: \"kubernetes.io/projected/b2d12dd0-18d6-4507-855e-639b7ef2fc2b-kube-api-access-58hx5\") on node \"crc\" DevicePath \"\"" Oct 14 08:35:56 crc kubenswrapper[4870]: I1014 08:35:56.441218 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-dwrfd" event={"ID":"b2d12dd0-18d6-4507-855e-639b7ef2fc2b","Type":"ContainerDied","Data":"8036524002f5b294344ba9c85a76e9e5914f6ac971151b715c58b4f8c7f0025b"} Oct 14 08:35:56 crc kubenswrapper[4870]: I1014 08:35:56.441283 4870 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8036524002f5b294344ba9c85a76e9e5914f6ac971151b715c58b4f8c7f0025b" Oct 14 08:35:56 crc kubenswrapper[4870]: I1014 08:35:56.441350 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-dwrfd" Oct 14 08:35:58 crc kubenswrapper[4870]: I1014 08:35:58.159007 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["crc-storage/crc-storage-crc-dwrfd"] Oct 14 08:35:58 crc kubenswrapper[4870]: I1014 08:35:58.166240 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["crc-storage/crc-storage-crc-dwrfd"] Oct 14 08:35:58 crc kubenswrapper[4870]: I1014 08:35:58.279040 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["crc-storage/crc-storage-crc-cqtdk"] Oct 14 08:35:58 crc kubenswrapper[4870]: E1014 08:35:58.279734 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b2d12dd0-18d6-4507-855e-639b7ef2fc2b" containerName="storage" Oct 14 08:35:58 crc kubenswrapper[4870]: I1014 08:35:58.279855 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="b2d12dd0-18d6-4507-855e-639b7ef2fc2b" containerName="storage" Oct 14 08:35:58 crc kubenswrapper[4870]: I1014 08:35:58.280180 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="b2d12dd0-18d6-4507-855e-639b7ef2fc2b" containerName="storage" Oct 14 08:35:58 crc kubenswrapper[4870]: I1014 08:35:58.281143 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-cqtdk" Oct 14 08:35:58 crc kubenswrapper[4870]: I1014 08:35:58.283066 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"openshift-service-ca.crt" Oct 14 08:35:58 crc kubenswrapper[4870]: I1014 08:35:58.283387 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"kube-root-ca.crt" Oct 14 08:35:58 crc kubenswrapper[4870]: I1014 08:35:58.283404 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"crc-storage" Oct 14 08:35:58 crc kubenswrapper[4870]: I1014 08:35:58.283966 4870 reflector.go:368] Caches populated for *v1.Secret from object-"crc-storage"/"crc-storage-dockercfg-pm8px" Oct 14 08:35:58 crc kubenswrapper[4870]: I1014 08:35:58.289203 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-cqtdk"] Oct 14 08:35:58 crc kubenswrapper[4870]: I1014 08:35:58.429496 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/9d64ed29-6d15-448b-a2d2-c2e7c412906e-node-mnt\") pod \"crc-storage-crc-cqtdk\" (UID: \"9d64ed29-6d15-448b-a2d2-c2e7c412906e\") " pod="crc-storage/crc-storage-crc-cqtdk" Oct 14 08:35:58 crc kubenswrapper[4870]: I1014 08:35:58.429854 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/9d64ed29-6d15-448b-a2d2-c2e7c412906e-crc-storage\") pod \"crc-storage-crc-cqtdk\" (UID: \"9d64ed29-6d15-448b-a2d2-c2e7c412906e\") " pod="crc-storage/crc-storage-crc-cqtdk" Oct 14 08:35:58 crc kubenswrapper[4870]: I1014 08:35:58.430308 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2jjw2\" (UniqueName: \"kubernetes.io/projected/9d64ed29-6d15-448b-a2d2-c2e7c412906e-kube-api-access-2jjw2\") pod \"crc-storage-crc-cqtdk\" (UID: \"9d64ed29-6d15-448b-a2d2-c2e7c412906e\") " pod="crc-storage/crc-storage-crc-cqtdk" Oct 14 08:35:58 crc kubenswrapper[4870]: I1014 08:35:58.531929 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/9d64ed29-6d15-448b-a2d2-c2e7c412906e-crc-storage\") pod \"crc-storage-crc-cqtdk\" (UID: \"9d64ed29-6d15-448b-a2d2-c2e7c412906e\") " pod="crc-storage/crc-storage-crc-cqtdk" Oct 14 08:35:58 crc kubenswrapper[4870]: I1014 08:35:58.532048 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2jjw2\" (UniqueName: \"kubernetes.io/projected/9d64ed29-6d15-448b-a2d2-c2e7c412906e-kube-api-access-2jjw2\") pod \"crc-storage-crc-cqtdk\" (UID: \"9d64ed29-6d15-448b-a2d2-c2e7c412906e\") " pod="crc-storage/crc-storage-crc-cqtdk" Oct 14 08:35:58 crc kubenswrapper[4870]: I1014 08:35:58.532089 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/9d64ed29-6d15-448b-a2d2-c2e7c412906e-node-mnt\") pod \"crc-storage-crc-cqtdk\" (UID: \"9d64ed29-6d15-448b-a2d2-c2e7c412906e\") " pod="crc-storage/crc-storage-crc-cqtdk" Oct 14 08:35:58 crc kubenswrapper[4870]: I1014 08:35:58.532409 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/9d64ed29-6d15-448b-a2d2-c2e7c412906e-node-mnt\") pod \"crc-storage-crc-cqtdk\" (UID: \"9d64ed29-6d15-448b-a2d2-c2e7c412906e\") " pod="crc-storage/crc-storage-crc-cqtdk" Oct 14 08:35:58 crc kubenswrapper[4870]: I1014 08:35:58.533115 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/9d64ed29-6d15-448b-a2d2-c2e7c412906e-crc-storage\") pod \"crc-storage-crc-cqtdk\" (UID: \"9d64ed29-6d15-448b-a2d2-c2e7c412906e\") " pod="crc-storage/crc-storage-crc-cqtdk" Oct 14 08:35:58 crc kubenswrapper[4870]: I1014 08:35:58.552217 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2jjw2\" (UniqueName: \"kubernetes.io/projected/9d64ed29-6d15-448b-a2d2-c2e7c412906e-kube-api-access-2jjw2\") pod \"crc-storage-crc-cqtdk\" (UID: \"9d64ed29-6d15-448b-a2d2-c2e7c412906e\") " pod="crc-storage/crc-storage-crc-cqtdk" Oct 14 08:35:58 crc kubenswrapper[4870]: I1014 08:35:58.607457 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-cqtdk" Oct 14 08:35:59 crc kubenswrapper[4870]: I1014 08:35:59.049632 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b2d12dd0-18d6-4507-855e-639b7ef2fc2b" path="/var/lib/kubelet/pods/b2d12dd0-18d6-4507-855e-639b7ef2fc2b/volumes" Oct 14 08:35:59 crc kubenswrapper[4870]: I1014 08:35:59.149799 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-cqtdk"] Oct 14 08:35:59 crc kubenswrapper[4870]: I1014 08:35:59.471338 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-cqtdk" event={"ID":"9d64ed29-6d15-448b-a2d2-c2e7c412906e","Type":"ContainerStarted","Data":"35e3317c677f9822e4ae4cb60c9ed89c3b02af9f3335e772b85a5d23cc6a0764"} Oct 14 08:36:00 crc kubenswrapper[4870]: I1014 08:36:00.483806 4870 generic.go:334] "Generic (PLEG): container finished" podID="9d64ed29-6d15-448b-a2d2-c2e7c412906e" containerID="1c0b44d2200a388aa7749a6f2e4c863ecf5342dfec028ed3e3f8ce16e644e54b" exitCode=0 Oct 14 08:36:00 crc kubenswrapper[4870]: I1014 08:36:00.483881 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-cqtdk" event={"ID":"9d64ed29-6d15-448b-a2d2-c2e7c412906e","Type":"ContainerDied","Data":"1c0b44d2200a388aa7749a6f2e4c863ecf5342dfec028ed3e3f8ce16e644e54b"} Oct 14 08:36:01 crc kubenswrapper[4870]: I1014 08:36:01.862401 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-cqtdk" Oct 14 08:36:01 crc kubenswrapper[4870]: I1014 08:36:01.986583 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/9d64ed29-6d15-448b-a2d2-c2e7c412906e-crc-storage\") pod \"9d64ed29-6d15-448b-a2d2-c2e7c412906e\" (UID: \"9d64ed29-6d15-448b-a2d2-c2e7c412906e\") " Oct 14 08:36:01 crc kubenswrapper[4870]: I1014 08:36:01.986729 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/9d64ed29-6d15-448b-a2d2-c2e7c412906e-node-mnt\") pod \"9d64ed29-6d15-448b-a2d2-c2e7c412906e\" (UID: \"9d64ed29-6d15-448b-a2d2-c2e7c412906e\") " Oct 14 08:36:01 crc kubenswrapper[4870]: I1014 08:36:01.986781 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2jjw2\" (UniqueName: \"kubernetes.io/projected/9d64ed29-6d15-448b-a2d2-c2e7c412906e-kube-api-access-2jjw2\") pod \"9d64ed29-6d15-448b-a2d2-c2e7c412906e\" (UID: \"9d64ed29-6d15-448b-a2d2-c2e7c412906e\") " Oct 14 08:36:01 crc kubenswrapper[4870]: I1014 08:36:01.987160 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9d64ed29-6d15-448b-a2d2-c2e7c412906e-node-mnt" (OuterVolumeSpecName: "node-mnt") pod "9d64ed29-6d15-448b-a2d2-c2e7c412906e" (UID: "9d64ed29-6d15-448b-a2d2-c2e7c412906e"). InnerVolumeSpecName "node-mnt". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 14 08:36:01 crc kubenswrapper[4870]: I1014 08:36:01.995470 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d64ed29-6d15-448b-a2d2-c2e7c412906e-kube-api-access-2jjw2" (OuterVolumeSpecName: "kube-api-access-2jjw2") pod "9d64ed29-6d15-448b-a2d2-c2e7c412906e" (UID: "9d64ed29-6d15-448b-a2d2-c2e7c412906e"). InnerVolumeSpecName "kube-api-access-2jjw2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:36:02 crc kubenswrapper[4870]: I1014 08:36:02.010095 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d64ed29-6d15-448b-a2d2-c2e7c412906e-crc-storage" (OuterVolumeSpecName: "crc-storage") pod "9d64ed29-6d15-448b-a2d2-c2e7c412906e" (UID: "9d64ed29-6d15-448b-a2d2-c2e7c412906e"). InnerVolumeSpecName "crc-storage". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:36:02 crc kubenswrapper[4870]: I1014 08:36:02.088888 4870 reconciler_common.go:293] "Volume detached for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/9d64ed29-6d15-448b-a2d2-c2e7c412906e-node-mnt\") on node \"crc\" DevicePath \"\"" Oct 14 08:36:02 crc kubenswrapper[4870]: I1014 08:36:02.088933 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2jjw2\" (UniqueName: \"kubernetes.io/projected/9d64ed29-6d15-448b-a2d2-c2e7c412906e-kube-api-access-2jjw2\") on node \"crc\" DevicePath \"\"" Oct 14 08:36:02 crc kubenswrapper[4870]: I1014 08:36:02.088950 4870 reconciler_common.go:293] "Volume detached for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/9d64ed29-6d15-448b-a2d2-c2e7c412906e-crc-storage\") on node \"crc\" DevicePath \"\"" Oct 14 08:36:02 crc kubenswrapper[4870]: I1014 08:36:02.504068 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-cqtdk" event={"ID":"9d64ed29-6d15-448b-a2d2-c2e7c412906e","Type":"ContainerDied","Data":"35e3317c677f9822e4ae4cb60c9ed89c3b02af9f3335e772b85a5d23cc6a0764"} Oct 14 08:36:02 crc kubenswrapper[4870]: I1014 08:36:02.504120 4870 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="35e3317c677f9822e4ae4cb60c9ed89c3b02af9f3335e772b85a5d23cc6a0764" Oct 14 08:36:02 crc kubenswrapper[4870]: I1014 08:36:02.504135 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-cqtdk" Oct 14 08:36:14 crc kubenswrapper[4870]: I1014 08:36:14.244023 4870 scope.go:117] "RemoveContainer" containerID="2cbfbf3bceef05399e1e3f329b2560558a7255dc007de0432c52c0cece04e00f" Oct 14 08:36:53 crc kubenswrapper[4870]: I1014 08:36:53.951584 4870 patch_prober.go:28] interesting pod/machine-config-daemon-7tvc7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 08:36:53 crc kubenswrapper[4870]: I1014 08:36:53.952254 4870 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 08:37:23 crc kubenswrapper[4870]: I1014 08:37:23.951065 4870 patch_prober.go:28] interesting pod/machine-config-daemon-7tvc7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 08:37:23 crc kubenswrapper[4870]: I1014 08:37:23.952702 4870 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 08:37:53 crc kubenswrapper[4870]: I1014 08:37:53.951206 4870 patch_prober.go:28] interesting pod/machine-config-daemon-7tvc7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 08:37:53 crc kubenswrapper[4870]: I1014 08:37:53.952175 4870 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 08:37:53 crc kubenswrapper[4870]: I1014 08:37:53.952268 4870 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" Oct 14 08:37:53 crc kubenswrapper[4870]: I1014 08:37:53.953125 4870 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"6f0398b7e7959f65985e6ee24c4434fc79307d6cce2fb874cfc4c5f50459984b"} pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 14 08:37:53 crc kubenswrapper[4870]: I1014 08:37:53.953223 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerName="machine-config-daemon" containerID="cri-o://6f0398b7e7959f65985e6ee24c4434fc79307d6cce2fb874cfc4c5f50459984b" gracePeriod=600 Oct 14 08:37:54 crc kubenswrapper[4870]: I1014 08:37:54.510833 4870 generic.go:334] "Generic (PLEG): container finished" podID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerID="6f0398b7e7959f65985e6ee24c4434fc79307d6cce2fb874cfc4c5f50459984b" exitCode=0 Oct 14 08:37:54 crc kubenswrapper[4870]: I1014 08:37:54.510890 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" event={"ID":"874a5e45-dffd-4d17-b609-b7d3ed2eab07","Type":"ContainerDied","Data":"6f0398b7e7959f65985e6ee24c4434fc79307d6cce2fb874cfc4c5f50459984b"} Oct 14 08:37:54 crc kubenswrapper[4870]: I1014 08:37:54.511222 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" event={"ID":"874a5e45-dffd-4d17-b609-b7d3ed2eab07","Type":"ContainerStarted","Data":"f9f1ebc5d4fe74cf43793adf15e4a8f7b2c05c9c750d6e220918c13c542fb541"} Oct 14 08:37:54 crc kubenswrapper[4870]: I1014 08:37:54.511259 4870 scope.go:117] "RemoveContainer" containerID="52468dffc62b6337feac3ee24b71903855af2b6b6e95f02951551f88ff46993a" Oct 14 08:38:02 crc kubenswrapper[4870]: I1014 08:38:02.613661 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7bd8799f99-g72nj"] Oct 14 08:38:02 crc kubenswrapper[4870]: E1014 08:38:02.614506 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d64ed29-6d15-448b-a2d2-c2e7c412906e" containerName="storage" Oct 14 08:38:02 crc kubenswrapper[4870]: I1014 08:38:02.614523 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d64ed29-6d15-448b-a2d2-c2e7c412906e" containerName="storage" Oct 14 08:38:02 crc kubenswrapper[4870]: I1014 08:38:02.614709 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="9d64ed29-6d15-448b-a2d2-c2e7c412906e" containerName="storage" Oct 14 08:38:02 crc kubenswrapper[4870]: I1014 08:38:02.615770 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7bd8799f99-g72nj" Oct 14 08:38:02 crc kubenswrapper[4870]: I1014 08:38:02.619682 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Oct 14 08:38:02 crc kubenswrapper[4870]: I1014 08:38:02.619797 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Oct 14 08:38:02 crc kubenswrapper[4870]: I1014 08:38:02.619842 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Oct 14 08:38:02 crc kubenswrapper[4870]: I1014 08:38:02.620431 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-qrkg7" Oct 14 08:38:02 crc kubenswrapper[4870]: I1014 08:38:02.619950 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Oct 14 08:38:02 crc kubenswrapper[4870]: I1014 08:38:02.626477 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7bd8799f99-g72nj"] Oct 14 08:38:02 crc kubenswrapper[4870]: I1014 08:38:02.778905 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p9vcb\" (UniqueName: \"kubernetes.io/projected/7482b8c0-86a7-4e3e-b05f-bb343129b9fb-kube-api-access-p9vcb\") pod \"dnsmasq-dns-7bd8799f99-g72nj\" (UID: \"7482b8c0-86a7-4e3e-b05f-bb343129b9fb\") " pod="openstack/dnsmasq-dns-7bd8799f99-g72nj" Oct 14 08:38:02 crc kubenswrapper[4870]: I1014 08:38:02.778954 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7482b8c0-86a7-4e3e-b05f-bb343129b9fb-dns-svc\") pod \"dnsmasq-dns-7bd8799f99-g72nj\" (UID: \"7482b8c0-86a7-4e3e-b05f-bb343129b9fb\") " pod="openstack/dnsmasq-dns-7bd8799f99-g72nj" Oct 14 08:38:02 crc kubenswrapper[4870]: I1014 08:38:02.779033 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7482b8c0-86a7-4e3e-b05f-bb343129b9fb-config\") pod \"dnsmasq-dns-7bd8799f99-g72nj\" (UID: \"7482b8c0-86a7-4e3e-b05f-bb343129b9fb\") " pod="openstack/dnsmasq-dns-7bd8799f99-g72nj" Oct 14 08:38:02 crc kubenswrapper[4870]: I1014 08:38:02.875166 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-54bcdcd99f-wq6qt"] Oct 14 08:38:02 crc kubenswrapper[4870]: I1014 08:38:02.876548 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-54bcdcd99f-wq6qt" Oct 14 08:38:02 crc kubenswrapper[4870]: I1014 08:38:02.879903 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7482b8c0-86a7-4e3e-b05f-bb343129b9fb-config\") pod \"dnsmasq-dns-7bd8799f99-g72nj\" (UID: \"7482b8c0-86a7-4e3e-b05f-bb343129b9fb\") " pod="openstack/dnsmasq-dns-7bd8799f99-g72nj" Oct 14 08:38:02 crc kubenswrapper[4870]: I1014 08:38:02.880008 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p9vcb\" (UniqueName: \"kubernetes.io/projected/7482b8c0-86a7-4e3e-b05f-bb343129b9fb-kube-api-access-p9vcb\") pod \"dnsmasq-dns-7bd8799f99-g72nj\" (UID: \"7482b8c0-86a7-4e3e-b05f-bb343129b9fb\") " pod="openstack/dnsmasq-dns-7bd8799f99-g72nj" Oct 14 08:38:02 crc kubenswrapper[4870]: I1014 08:38:02.880041 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7482b8c0-86a7-4e3e-b05f-bb343129b9fb-dns-svc\") pod \"dnsmasq-dns-7bd8799f99-g72nj\" (UID: \"7482b8c0-86a7-4e3e-b05f-bb343129b9fb\") " pod="openstack/dnsmasq-dns-7bd8799f99-g72nj" Oct 14 08:38:02 crc kubenswrapper[4870]: I1014 08:38:02.880802 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7482b8c0-86a7-4e3e-b05f-bb343129b9fb-config\") pod \"dnsmasq-dns-7bd8799f99-g72nj\" (UID: \"7482b8c0-86a7-4e3e-b05f-bb343129b9fb\") " pod="openstack/dnsmasq-dns-7bd8799f99-g72nj" Oct 14 08:38:02 crc kubenswrapper[4870]: I1014 08:38:02.880926 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7482b8c0-86a7-4e3e-b05f-bb343129b9fb-dns-svc\") pod \"dnsmasq-dns-7bd8799f99-g72nj\" (UID: \"7482b8c0-86a7-4e3e-b05f-bb343129b9fb\") " pod="openstack/dnsmasq-dns-7bd8799f99-g72nj" Oct 14 08:38:02 crc kubenswrapper[4870]: I1014 08:38:02.894492 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-54bcdcd99f-wq6qt"] Oct 14 08:38:02 crc kubenswrapper[4870]: I1014 08:38:02.902024 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p9vcb\" (UniqueName: \"kubernetes.io/projected/7482b8c0-86a7-4e3e-b05f-bb343129b9fb-kube-api-access-p9vcb\") pod \"dnsmasq-dns-7bd8799f99-g72nj\" (UID: \"7482b8c0-86a7-4e3e-b05f-bb343129b9fb\") " pod="openstack/dnsmasq-dns-7bd8799f99-g72nj" Oct 14 08:38:02 crc kubenswrapper[4870]: I1014 08:38:02.940627 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7bd8799f99-g72nj" Oct 14 08:38:02 crc kubenswrapper[4870]: I1014 08:38:02.981126 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/64d3a546-2e9d-4c15-bcb0-53bbdf6070b3-config\") pod \"dnsmasq-dns-54bcdcd99f-wq6qt\" (UID: \"64d3a546-2e9d-4c15-bcb0-53bbdf6070b3\") " pod="openstack/dnsmasq-dns-54bcdcd99f-wq6qt" Oct 14 08:38:02 crc kubenswrapper[4870]: I1014 08:38:02.981362 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cgd65\" (UniqueName: \"kubernetes.io/projected/64d3a546-2e9d-4c15-bcb0-53bbdf6070b3-kube-api-access-cgd65\") pod \"dnsmasq-dns-54bcdcd99f-wq6qt\" (UID: \"64d3a546-2e9d-4c15-bcb0-53bbdf6070b3\") " pod="openstack/dnsmasq-dns-54bcdcd99f-wq6qt" Oct 14 08:38:02 crc kubenswrapper[4870]: I1014 08:38:02.981501 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/64d3a546-2e9d-4c15-bcb0-53bbdf6070b3-dns-svc\") pod \"dnsmasq-dns-54bcdcd99f-wq6qt\" (UID: \"64d3a546-2e9d-4c15-bcb0-53bbdf6070b3\") " pod="openstack/dnsmasq-dns-54bcdcd99f-wq6qt" Oct 14 08:38:03 crc kubenswrapper[4870]: I1014 08:38:03.082652 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/64d3a546-2e9d-4c15-bcb0-53bbdf6070b3-dns-svc\") pod \"dnsmasq-dns-54bcdcd99f-wq6qt\" (UID: \"64d3a546-2e9d-4c15-bcb0-53bbdf6070b3\") " pod="openstack/dnsmasq-dns-54bcdcd99f-wq6qt" Oct 14 08:38:03 crc kubenswrapper[4870]: I1014 08:38:03.082868 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/64d3a546-2e9d-4c15-bcb0-53bbdf6070b3-config\") pod \"dnsmasq-dns-54bcdcd99f-wq6qt\" (UID: \"64d3a546-2e9d-4c15-bcb0-53bbdf6070b3\") " pod="openstack/dnsmasq-dns-54bcdcd99f-wq6qt" Oct 14 08:38:03 crc kubenswrapper[4870]: I1014 08:38:03.082927 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cgd65\" (UniqueName: \"kubernetes.io/projected/64d3a546-2e9d-4c15-bcb0-53bbdf6070b3-kube-api-access-cgd65\") pod \"dnsmasq-dns-54bcdcd99f-wq6qt\" (UID: \"64d3a546-2e9d-4c15-bcb0-53bbdf6070b3\") " pod="openstack/dnsmasq-dns-54bcdcd99f-wq6qt" Oct 14 08:38:03 crc kubenswrapper[4870]: I1014 08:38:03.084000 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/64d3a546-2e9d-4c15-bcb0-53bbdf6070b3-dns-svc\") pod \"dnsmasq-dns-54bcdcd99f-wq6qt\" (UID: \"64d3a546-2e9d-4c15-bcb0-53bbdf6070b3\") " pod="openstack/dnsmasq-dns-54bcdcd99f-wq6qt" Oct 14 08:38:03 crc kubenswrapper[4870]: I1014 08:38:03.084196 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/64d3a546-2e9d-4c15-bcb0-53bbdf6070b3-config\") pod \"dnsmasq-dns-54bcdcd99f-wq6qt\" (UID: \"64d3a546-2e9d-4c15-bcb0-53bbdf6070b3\") " pod="openstack/dnsmasq-dns-54bcdcd99f-wq6qt" Oct 14 08:38:03 crc kubenswrapper[4870]: I1014 08:38:03.102919 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cgd65\" (UniqueName: \"kubernetes.io/projected/64d3a546-2e9d-4c15-bcb0-53bbdf6070b3-kube-api-access-cgd65\") pod \"dnsmasq-dns-54bcdcd99f-wq6qt\" (UID: \"64d3a546-2e9d-4c15-bcb0-53bbdf6070b3\") " pod="openstack/dnsmasq-dns-54bcdcd99f-wq6qt" Oct 14 08:38:03 crc kubenswrapper[4870]: I1014 08:38:03.193117 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-54bcdcd99f-wq6qt" Oct 14 08:38:03 crc kubenswrapper[4870]: I1014 08:38:03.777234 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Oct 14 08:38:03 crc kubenswrapper[4870]: I1014 08:38:03.779684 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 14 08:38:03 crc kubenswrapper[4870]: I1014 08:38:03.785010 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Oct 14 08:38:03 crc kubenswrapper[4870]: I1014 08:38:03.785024 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Oct 14 08:38:03 crc kubenswrapper[4870]: I1014 08:38:03.785279 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-7fhn4" Oct 14 08:38:03 crc kubenswrapper[4870]: I1014 08:38:03.785419 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Oct 14 08:38:03 crc kubenswrapper[4870]: I1014 08:38:03.785703 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Oct 14 08:38:03 crc kubenswrapper[4870]: I1014 08:38:03.789545 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 14 08:38:03 crc kubenswrapper[4870]: I1014 08:38:03.893648 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t4vf8\" (UniqueName: \"kubernetes.io/projected/8e2ef84d-2d3f-4a34-8dec-27e08689779e-kube-api-access-t4vf8\") pod \"rabbitmq-server-0\" (UID: \"8e2ef84d-2d3f-4a34-8dec-27e08689779e\") " pod="openstack/rabbitmq-server-0" Oct 14 08:38:03 crc kubenswrapper[4870]: I1014 08:38:03.893735 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/8e2ef84d-2d3f-4a34-8dec-27e08689779e-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"8e2ef84d-2d3f-4a34-8dec-27e08689779e\") " pod="openstack/rabbitmq-server-0" Oct 14 08:38:03 crc kubenswrapper[4870]: I1014 08:38:03.893858 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/8e2ef84d-2d3f-4a34-8dec-27e08689779e-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"8e2ef84d-2d3f-4a34-8dec-27e08689779e\") " pod="openstack/rabbitmq-server-0" Oct 14 08:38:03 crc kubenswrapper[4870]: I1014 08:38:03.893953 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/8e2ef84d-2d3f-4a34-8dec-27e08689779e-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"8e2ef84d-2d3f-4a34-8dec-27e08689779e\") " pod="openstack/rabbitmq-server-0" Oct 14 08:38:03 crc kubenswrapper[4870]: I1014 08:38:03.894015 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/8e2ef84d-2d3f-4a34-8dec-27e08689779e-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"8e2ef84d-2d3f-4a34-8dec-27e08689779e\") " pod="openstack/rabbitmq-server-0" Oct 14 08:38:03 crc kubenswrapper[4870]: I1014 08:38:03.894068 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/8e2ef84d-2d3f-4a34-8dec-27e08689779e-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"8e2ef84d-2d3f-4a34-8dec-27e08689779e\") " pod="openstack/rabbitmq-server-0" Oct 14 08:38:03 crc kubenswrapper[4870]: I1014 08:38:03.894161 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-dc261995-757b-493e-a935-9844b1fe3864\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-dc261995-757b-493e-a935-9844b1fe3864\") pod \"rabbitmq-server-0\" (UID: \"8e2ef84d-2d3f-4a34-8dec-27e08689779e\") " pod="openstack/rabbitmq-server-0" Oct 14 08:38:03 crc kubenswrapper[4870]: I1014 08:38:03.894284 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/8e2ef84d-2d3f-4a34-8dec-27e08689779e-server-conf\") pod \"rabbitmq-server-0\" (UID: \"8e2ef84d-2d3f-4a34-8dec-27e08689779e\") " pod="openstack/rabbitmq-server-0" Oct 14 08:38:03 crc kubenswrapper[4870]: I1014 08:38:03.894358 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/8e2ef84d-2d3f-4a34-8dec-27e08689779e-pod-info\") pod \"rabbitmq-server-0\" (UID: \"8e2ef84d-2d3f-4a34-8dec-27e08689779e\") " pod="openstack/rabbitmq-server-0" Oct 14 08:38:03 crc kubenswrapper[4870]: I1014 08:38:03.995863 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/8e2ef84d-2d3f-4a34-8dec-27e08689779e-server-conf\") pod \"rabbitmq-server-0\" (UID: \"8e2ef84d-2d3f-4a34-8dec-27e08689779e\") " pod="openstack/rabbitmq-server-0" Oct 14 08:38:03 crc kubenswrapper[4870]: I1014 08:38:03.995935 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/8e2ef84d-2d3f-4a34-8dec-27e08689779e-pod-info\") pod \"rabbitmq-server-0\" (UID: \"8e2ef84d-2d3f-4a34-8dec-27e08689779e\") " pod="openstack/rabbitmq-server-0" Oct 14 08:38:03 crc kubenswrapper[4870]: I1014 08:38:03.995997 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t4vf8\" (UniqueName: \"kubernetes.io/projected/8e2ef84d-2d3f-4a34-8dec-27e08689779e-kube-api-access-t4vf8\") pod \"rabbitmq-server-0\" (UID: \"8e2ef84d-2d3f-4a34-8dec-27e08689779e\") " pod="openstack/rabbitmq-server-0" Oct 14 08:38:03 crc kubenswrapper[4870]: I1014 08:38:03.996052 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/8e2ef84d-2d3f-4a34-8dec-27e08689779e-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"8e2ef84d-2d3f-4a34-8dec-27e08689779e\") " pod="openstack/rabbitmq-server-0" Oct 14 08:38:03 crc kubenswrapper[4870]: I1014 08:38:03.996078 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/8e2ef84d-2d3f-4a34-8dec-27e08689779e-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"8e2ef84d-2d3f-4a34-8dec-27e08689779e\") " pod="openstack/rabbitmq-server-0" Oct 14 08:38:03 crc kubenswrapper[4870]: I1014 08:38:03.996114 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/8e2ef84d-2d3f-4a34-8dec-27e08689779e-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"8e2ef84d-2d3f-4a34-8dec-27e08689779e\") " pod="openstack/rabbitmq-server-0" Oct 14 08:38:03 crc kubenswrapper[4870]: I1014 08:38:03.996142 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/8e2ef84d-2d3f-4a34-8dec-27e08689779e-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"8e2ef84d-2d3f-4a34-8dec-27e08689779e\") " pod="openstack/rabbitmq-server-0" Oct 14 08:38:03 crc kubenswrapper[4870]: I1014 08:38:03.996176 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/8e2ef84d-2d3f-4a34-8dec-27e08689779e-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"8e2ef84d-2d3f-4a34-8dec-27e08689779e\") " pod="openstack/rabbitmq-server-0" Oct 14 08:38:03 crc kubenswrapper[4870]: I1014 08:38:03.996206 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-dc261995-757b-493e-a935-9844b1fe3864\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-dc261995-757b-493e-a935-9844b1fe3864\") pod \"rabbitmq-server-0\" (UID: \"8e2ef84d-2d3f-4a34-8dec-27e08689779e\") " pod="openstack/rabbitmq-server-0" Oct 14 08:38:03 crc kubenswrapper[4870]: I1014 08:38:03.996832 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/8e2ef84d-2d3f-4a34-8dec-27e08689779e-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"8e2ef84d-2d3f-4a34-8dec-27e08689779e\") " pod="openstack/rabbitmq-server-0" Oct 14 08:38:03 crc kubenswrapper[4870]: I1014 08:38:03.997115 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/8e2ef84d-2d3f-4a34-8dec-27e08689779e-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"8e2ef84d-2d3f-4a34-8dec-27e08689779e\") " pod="openstack/rabbitmq-server-0" Oct 14 08:38:03 crc kubenswrapper[4870]: I1014 08:38:03.997229 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/8e2ef84d-2d3f-4a34-8dec-27e08689779e-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"8e2ef84d-2d3f-4a34-8dec-27e08689779e\") " pod="openstack/rabbitmq-server-0" Oct 14 08:38:03 crc kubenswrapper[4870]: I1014 08:38:03.997598 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/8e2ef84d-2d3f-4a34-8dec-27e08689779e-server-conf\") pod \"rabbitmq-server-0\" (UID: \"8e2ef84d-2d3f-4a34-8dec-27e08689779e\") " pod="openstack/rabbitmq-server-0" Oct 14 08:38:04 crc kubenswrapper[4870]: I1014 08:38:03.999964 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/8e2ef84d-2d3f-4a34-8dec-27e08689779e-pod-info\") pod \"rabbitmq-server-0\" (UID: \"8e2ef84d-2d3f-4a34-8dec-27e08689779e\") " pod="openstack/rabbitmq-server-0" Oct 14 08:38:04 crc kubenswrapper[4870]: I1014 08:38:04.000206 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/8e2ef84d-2d3f-4a34-8dec-27e08689779e-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"8e2ef84d-2d3f-4a34-8dec-27e08689779e\") " pod="openstack/rabbitmq-server-0" Oct 14 08:38:04 crc kubenswrapper[4870]: I1014 08:38:04.001472 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/8e2ef84d-2d3f-4a34-8dec-27e08689779e-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"8e2ef84d-2d3f-4a34-8dec-27e08689779e\") " pod="openstack/rabbitmq-server-0" Oct 14 08:38:04 crc kubenswrapper[4870]: I1014 08:38:04.013008 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 14 08:38:04 crc kubenswrapper[4870]: I1014 08:38:04.013254 4870 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 14 08:38:04 crc kubenswrapper[4870]: I1014 08:38:04.019581 4870 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-dc261995-757b-493e-a935-9844b1fe3864\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-dc261995-757b-493e-a935-9844b1fe3864\") pod \"rabbitmq-server-0\" (UID: \"8e2ef84d-2d3f-4a34-8dec-27e08689779e\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/e4cd1d45c92fe27d183871dbe3d8ad739ccac095190c427f162da6f6556256e5/globalmount\"" pod="openstack/rabbitmq-server-0" Oct 14 08:38:04 crc kubenswrapper[4870]: I1014 08:38:04.014552 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:38:04 crc kubenswrapper[4870]: I1014 08:38:04.017419 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t4vf8\" (UniqueName: \"kubernetes.io/projected/8e2ef84d-2d3f-4a34-8dec-27e08689779e-kube-api-access-t4vf8\") pod \"rabbitmq-server-0\" (UID: \"8e2ef84d-2d3f-4a34-8dec-27e08689779e\") " pod="openstack/rabbitmq-server-0" Oct 14 08:38:04 crc kubenswrapper[4870]: I1014 08:38:04.022734 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-kpv9q" Oct 14 08:38:04 crc kubenswrapper[4870]: I1014 08:38:04.023178 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Oct 14 08:38:04 crc kubenswrapper[4870]: I1014 08:38:04.023407 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Oct 14 08:38:04 crc kubenswrapper[4870]: I1014 08:38:04.023429 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Oct 14 08:38:04 crc kubenswrapper[4870]: I1014 08:38:04.023634 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Oct 14 08:38:04 crc kubenswrapper[4870]: I1014 08:38:04.049540 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 14 08:38:04 crc kubenswrapper[4870]: I1014 08:38:04.058898 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7bd8799f99-g72nj"] Oct 14 08:38:04 crc kubenswrapper[4870]: I1014 08:38:04.099856 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-dc261995-757b-493e-a935-9844b1fe3864\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-dc261995-757b-493e-a935-9844b1fe3864\") pod \"rabbitmq-server-0\" (UID: \"8e2ef84d-2d3f-4a34-8dec-27e08689779e\") " pod="openstack/rabbitmq-server-0" Oct 14 08:38:04 crc kubenswrapper[4870]: I1014 08:38:04.101287 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-7f3d7a0e-92c6-421f-ac96-6a093ce92216\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7f3d7a0e-92c6-421f-ac96-6a093ce92216\") pod \"rabbitmq-cell1-server-0\" (UID: \"98a7db06-05d2-4c54-9321-11ad232caeda\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:38:04 crc kubenswrapper[4870]: I1014 08:38:04.101352 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/98a7db06-05d2-4c54-9321-11ad232caeda-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"98a7db06-05d2-4c54-9321-11ad232caeda\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:38:04 crc kubenswrapper[4870]: I1014 08:38:04.101429 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/98a7db06-05d2-4c54-9321-11ad232caeda-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"98a7db06-05d2-4c54-9321-11ad232caeda\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:38:04 crc kubenswrapper[4870]: I1014 08:38:04.101556 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/98a7db06-05d2-4c54-9321-11ad232caeda-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"98a7db06-05d2-4c54-9321-11ad232caeda\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:38:04 crc kubenswrapper[4870]: I1014 08:38:04.101615 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/98a7db06-05d2-4c54-9321-11ad232caeda-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"98a7db06-05d2-4c54-9321-11ad232caeda\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:38:04 crc kubenswrapper[4870]: I1014 08:38:04.101635 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/98a7db06-05d2-4c54-9321-11ad232caeda-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"98a7db06-05d2-4c54-9321-11ad232caeda\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:38:04 crc kubenswrapper[4870]: I1014 08:38:04.101662 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/98a7db06-05d2-4c54-9321-11ad232caeda-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"98a7db06-05d2-4c54-9321-11ad232caeda\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:38:04 crc kubenswrapper[4870]: I1014 08:38:04.101682 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-whccn\" (UniqueName: \"kubernetes.io/projected/98a7db06-05d2-4c54-9321-11ad232caeda-kube-api-access-whccn\") pod \"rabbitmq-cell1-server-0\" (UID: \"98a7db06-05d2-4c54-9321-11ad232caeda\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:38:04 crc kubenswrapper[4870]: I1014 08:38:04.101700 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/98a7db06-05d2-4c54-9321-11ad232caeda-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"98a7db06-05d2-4c54-9321-11ad232caeda\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:38:04 crc kubenswrapper[4870]: I1014 08:38:04.136699 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-54bcdcd99f-wq6qt"] Oct 14 08:38:04 crc kubenswrapper[4870]: I1014 08:38:04.203069 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/98a7db06-05d2-4c54-9321-11ad232caeda-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"98a7db06-05d2-4c54-9321-11ad232caeda\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:38:04 crc kubenswrapper[4870]: I1014 08:38:04.203112 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/98a7db06-05d2-4c54-9321-11ad232caeda-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"98a7db06-05d2-4c54-9321-11ad232caeda\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:38:04 crc kubenswrapper[4870]: I1014 08:38:04.203149 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/98a7db06-05d2-4c54-9321-11ad232caeda-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"98a7db06-05d2-4c54-9321-11ad232caeda\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:38:04 crc kubenswrapper[4870]: I1014 08:38:04.203171 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-whccn\" (UniqueName: \"kubernetes.io/projected/98a7db06-05d2-4c54-9321-11ad232caeda-kube-api-access-whccn\") pod \"rabbitmq-cell1-server-0\" (UID: \"98a7db06-05d2-4c54-9321-11ad232caeda\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:38:04 crc kubenswrapper[4870]: I1014 08:38:04.203191 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/98a7db06-05d2-4c54-9321-11ad232caeda-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"98a7db06-05d2-4c54-9321-11ad232caeda\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:38:04 crc kubenswrapper[4870]: I1014 08:38:04.203244 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-7f3d7a0e-92c6-421f-ac96-6a093ce92216\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7f3d7a0e-92c6-421f-ac96-6a093ce92216\") pod \"rabbitmq-cell1-server-0\" (UID: \"98a7db06-05d2-4c54-9321-11ad232caeda\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:38:04 crc kubenswrapper[4870]: I1014 08:38:04.203266 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/98a7db06-05d2-4c54-9321-11ad232caeda-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"98a7db06-05d2-4c54-9321-11ad232caeda\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:38:04 crc kubenswrapper[4870]: I1014 08:38:04.203322 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/98a7db06-05d2-4c54-9321-11ad232caeda-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"98a7db06-05d2-4c54-9321-11ad232caeda\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:38:04 crc kubenswrapper[4870]: I1014 08:38:04.203347 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/98a7db06-05d2-4c54-9321-11ad232caeda-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"98a7db06-05d2-4c54-9321-11ad232caeda\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:38:04 crc kubenswrapper[4870]: I1014 08:38:04.204055 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/98a7db06-05d2-4c54-9321-11ad232caeda-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"98a7db06-05d2-4c54-9321-11ad232caeda\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:38:04 crc kubenswrapper[4870]: I1014 08:38:04.204324 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/98a7db06-05d2-4c54-9321-11ad232caeda-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"98a7db06-05d2-4c54-9321-11ad232caeda\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:38:04 crc kubenswrapper[4870]: I1014 08:38:04.204520 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/98a7db06-05d2-4c54-9321-11ad232caeda-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"98a7db06-05d2-4c54-9321-11ad232caeda\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:38:04 crc kubenswrapper[4870]: I1014 08:38:04.204618 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/98a7db06-05d2-4c54-9321-11ad232caeda-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"98a7db06-05d2-4c54-9321-11ad232caeda\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:38:04 crc kubenswrapper[4870]: I1014 08:38:04.206843 4870 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 14 08:38:04 crc kubenswrapper[4870]: I1014 08:38:04.206877 4870 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-7f3d7a0e-92c6-421f-ac96-6a093ce92216\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7f3d7a0e-92c6-421f-ac96-6a093ce92216\") pod \"rabbitmq-cell1-server-0\" (UID: \"98a7db06-05d2-4c54-9321-11ad232caeda\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/5ef42abd3faf11736bf20237aa539234ecec053653af230c7b5d392cae64bad5/globalmount\"" pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:38:04 crc kubenswrapper[4870]: I1014 08:38:04.208917 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/98a7db06-05d2-4c54-9321-11ad232caeda-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"98a7db06-05d2-4c54-9321-11ad232caeda\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:38:04 crc kubenswrapper[4870]: I1014 08:38:04.210504 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/98a7db06-05d2-4c54-9321-11ad232caeda-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"98a7db06-05d2-4c54-9321-11ad232caeda\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:38:04 crc kubenswrapper[4870]: I1014 08:38:04.211615 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/98a7db06-05d2-4c54-9321-11ad232caeda-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"98a7db06-05d2-4c54-9321-11ad232caeda\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:38:04 crc kubenswrapper[4870]: I1014 08:38:04.227016 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-whccn\" (UniqueName: \"kubernetes.io/projected/98a7db06-05d2-4c54-9321-11ad232caeda-kube-api-access-whccn\") pod \"rabbitmq-cell1-server-0\" (UID: \"98a7db06-05d2-4c54-9321-11ad232caeda\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:38:04 crc kubenswrapper[4870]: I1014 08:38:04.236493 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-7f3d7a0e-92c6-421f-ac96-6a093ce92216\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7f3d7a0e-92c6-421f-ac96-6a093ce92216\") pod \"rabbitmq-cell1-server-0\" (UID: \"98a7db06-05d2-4c54-9321-11ad232caeda\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:38:04 crc kubenswrapper[4870]: I1014 08:38:04.401421 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 14 08:38:04 crc kubenswrapper[4870]: I1014 08:38:04.448650 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:38:04 crc kubenswrapper[4870]: I1014 08:38:04.607491 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-54bcdcd99f-wq6qt" event={"ID":"64d3a546-2e9d-4c15-bcb0-53bbdf6070b3","Type":"ContainerStarted","Data":"783c8c366bd8f5e01238808715936a1503599cedb5a2ded203425eda63c5808a"} Oct 14 08:38:04 crc kubenswrapper[4870]: I1014 08:38:04.610412 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7bd8799f99-g72nj" event={"ID":"7482b8c0-86a7-4e3e-b05f-bb343129b9fb","Type":"ContainerStarted","Data":"694d032b9c8da5a064f94a58658638e7030bba153b0cbbbdf71e601ebcf930dc"} Oct 14 08:38:04 crc kubenswrapper[4870]: I1014 08:38:04.907670 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 14 08:38:04 crc kubenswrapper[4870]: I1014 08:38:04.919015 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 14 08:38:04 crc kubenswrapper[4870]: W1014 08:38:04.938757 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8e2ef84d_2d3f_4a34_8dec_27e08689779e.slice/crio-d316b97d9c5edf108324aa5c4099c35b60e984ff57c60aae837af076cc2d9662 WatchSource:0}: Error finding container d316b97d9c5edf108324aa5c4099c35b60e984ff57c60aae837af076cc2d9662: Status 404 returned error can't find the container with id d316b97d9c5edf108324aa5c4099c35b60e984ff57c60aae837af076cc2d9662 Oct 14 08:38:05 crc kubenswrapper[4870]: I1014 08:38:05.134946 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Oct 14 08:38:05 crc kubenswrapper[4870]: I1014 08:38:05.136416 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Oct 14 08:38:05 crc kubenswrapper[4870]: I1014 08:38:05.140595 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-x7hnp" Oct 14 08:38:05 crc kubenswrapper[4870]: I1014 08:38:05.140660 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Oct 14 08:38:05 crc kubenswrapper[4870]: I1014 08:38:05.143842 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Oct 14 08:38:05 crc kubenswrapper[4870]: I1014 08:38:05.144835 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Oct 14 08:38:05 crc kubenswrapper[4870]: I1014 08:38:05.149803 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Oct 14 08:38:05 crc kubenswrapper[4870]: I1014 08:38:05.150965 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Oct 14 08:38:05 crc kubenswrapper[4870]: I1014 08:38:05.151458 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Oct 14 08:38:05 crc kubenswrapper[4870]: I1014 08:38:05.215900 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xhwbk\" (UniqueName: \"kubernetes.io/projected/05c1ec22-e0d6-45c4-85fc-d586c046b3b1-kube-api-access-xhwbk\") pod \"openstack-galera-0\" (UID: \"05c1ec22-e0d6-45c4-85fc-d586c046b3b1\") " pod="openstack/openstack-galera-0" Oct 14 08:38:05 crc kubenswrapper[4870]: I1014 08:38:05.215957 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-acda8db4-57c4-42f7-beb2-dc1d70fe6445\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-acda8db4-57c4-42f7-beb2-dc1d70fe6445\") pod \"openstack-galera-0\" (UID: \"05c1ec22-e0d6-45c4-85fc-d586c046b3b1\") " pod="openstack/openstack-galera-0" Oct 14 08:38:05 crc kubenswrapper[4870]: I1014 08:38:05.215991 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/05c1ec22-e0d6-45c4-85fc-d586c046b3b1-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"05c1ec22-e0d6-45c4-85fc-d586c046b3b1\") " pod="openstack/openstack-galera-0" Oct 14 08:38:05 crc kubenswrapper[4870]: I1014 08:38:05.216041 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/05c1ec22-e0d6-45c4-85fc-d586c046b3b1-config-data-default\") pod \"openstack-galera-0\" (UID: \"05c1ec22-e0d6-45c4-85fc-d586c046b3b1\") " pod="openstack/openstack-galera-0" Oct 14 08:38:05 crc kubenswrapper[4870]: I1014 08:38:05.216072 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/05c1ec22-e0d6-45c4-85fc-d586c046b3b1-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"05c1ec22-e0d6-45c4-85fc-d586c046b3b1\") " pod="openstack/openstack-galera-0" Oct 14 08:38:05 crc kubenswrapper[4870]: I1014 08:38:05.216093 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/05c1ec22-e0d6-45c4-85fc-d586c046b3b1-secrets\") pod \"openstack-galera-0\" (UID: \"05c1ec22-e0d6-45c4-85fc-d586c046b3b1\") " pod="openstack/openstack-galera-0" Oct 14 08:38:05 crc kubenswrapper[4870]: I1014 08:38:05.216109 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/05c1ec22-e0d6-45c4-85fc-d586c046b3b1-kolla-config\") pod \"openstack-galera-0\" (UID: \"05c1ec22-e0d6-45c4-85fc-d586c046b3b1\") " pod="openstack/openstack-galera-0" Oct 14 08:38:05 crc kubenswrapper[4870]: I1014 08:38:05.216124 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/05c1ec22-e0d6-45c4-85fc-d586c046b3b1-config-data-generated\") pod \"openstack-galera-0\" (UID: \"05c1ec22-e0d6-45c4-85fc-d586c046b3b1\") " pod="openstack/openstack-galera-0" Oct 14 08:38:05 crc kubenswrapper[4870]: I1014 08:38:05.216167 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/05c1ec22-e0d6-45c4-85fc-d586c046b3b1-operator-scripts\") pod \"openstack-galera-0\" (UID: \"05c1ec22-e0d6-45c4-85fc-d586c046b3b1\") " pod="openstack/openstack-galera-0" Oct 14 08:38:05 crc kubenswrapper[4870]: I1014 08:38:05.317401 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/05c1ec22-e0d6-45c4-85fc-d586c046b3b1-config-data-default\") pod \"openstack-galera-0\" (UID: \"05c1ec22-e0d6-45c4-85fc-d586c046b3b1\") " pod="openstack/openstack-galera-0" Oct 14 08:38:05 crc kubenswrapper[4870]: I1014 08:38:05.317486 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/05c1ec22-e0d6-45c4-85fc-d586c046b3b1-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"05c1ec22-e0d6-45c4-85fc-d586c046b3b1\") " pod="openstack/openstack-galera-0" Oct 14 08:38:05 crc kubenswrapper[4870]: I1014 08:38:05.317544 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/05c1ec22-e0d6-45c4-85fc-d586c046b3b1-secrets\") pod \"openstack-galera-0\" (UID: \"05c1ec22-e0d6-45c4-85fc-d586c046b3b1\") " pod="openstack/openstack-galera-0" Oct 14 08:38:05 crc kubenswrapper[4870]: I1014 08:38:05.317573 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/05c1ec22-e0d6-45c4-85fc-d586c046b3b1-kolla-config\") pod \"openstack-galera-0\" (UID: \"05c1ec22-e0d6-45c4-85fc-d586c046b3b1\") " pod="openstack/openstack-galera-0" Oct 14 08:38:05 crc kubenswrapper[4870]: I1014 08:38:05.317604 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/05c1ec22-e0d6-45c4-85fc-d586c046b3b1-config-data-generated\") pod \"openstack-galera-0\" (UID: \"05c1ec22-e0d6-45c4-85fc-d586c046b3b1\") " pod="openstack/openstack-galera-0" Oct 14 08:38:05 crc kubenswrapper[4870]: I1014 08:38:05.317684 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/05c1ec22-e0d6-45c4-85fc-d586c046b3b1-operator-scripts\") pod \"openstack-galera-0\" (UID: \"05c1ec22-e0d6-45c4-85fc-d586c046b3b1\") " pod="openstack/openstack-galera-0" Oct 14 08:38:05 crc kubenswrapper[4870]: I1014 08:38:05.317717 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xhwbk\" (UniqueName: \"kubernetes.io/projected/05c1ec22-e0d6-45c4-85fc-d586c046b3b1-kube-api-access-xhwbk\") pod \"openstack-galera-0\" (UID: \"05c1ec22-e0d6-45c4-85fc-d586c046b3b1\") " pod="openstack/openstack-galera-0" Oct 14 08:38:05 crc kubenswrapper[4870]: I1014 08:38:05.317749 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-acda8db4-57c4-42f7-beb2-dc1d70fe6445\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-acda8db4-57c4-42f7-beb2-dc1d70fe6445\") pod \"openstack-galera-0\" (UID: \"05c1ec22-e0d6-45c4-85fc-d586c046b3b1\") " pod="openstack/openstack-galera-0" Oct 14 08:38:05 crc kubenswrapper[4870]: I1014 08:38:05.317786 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/05c1ec22-e0d6-45c4-85fc-d586c046b3b1-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"05c1ec22-e0d6-45c4-85fc-d586c046b3b1\") " pod="openstack/openstack-galera-0" Oct 14 08:38:05 crc kubenswrapper[4870]: I1014 08:38:05.320520 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/05c1ec22-e0d6-45c4-85fc-d586c046b3b1-config-data-default\") pod \"openstack-galera-0\" (UID: \"05c1ec22-e0d6-45c4-85fc-d586c046b3b1\") " pod="openstack/openstack-galera-0" Oct 14 08:38:05 crc kubenswrapper[4870]: I1014 08:38:05.321019 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/05c1ec22-e0d6-45c4-85fc-d586c046b3b1-config-data-generated\") pod \"openstack-galera-0\" (UID: \"05c1ec22-e0d6-45c4-85fc-d586c046b3b1\") " pod="openstack/openstack-galera-0" Oct 14 08:38:05 crc kubenswrapper[4870]: I1014 08:38:05.323786 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/05c1ec22-e0d6-45c4-85fc-d586c046b3b1-kolla-config\") pod \"openstack-galera-0\" (UID: \"05c1ec22-e0d6-45c4-85fc-d586c046b3b1\") " pod="openstack/openstack-galera-0" Oct 14 08:38:05 crc kubenswrapper[4870]: I1014 08:38:05.324950 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/05c1ec22-e0d6-45c4-85fc-d586c046b3b1-operator-scripts\") pod \"openstack-galera-0\" (UID: \"05c1ec22-e0d6-45c4-85fc-d586c046b3b1\") " pod="openstack/openstack-galera-0" Oct 14 08:38:05 crc kubenswrapper[4870]: I1014 08:38:05.326077 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/05c1ec22-e0d6-45c4-85fc-d586c046b3b1-secrets\") pod \"openstack-galera-0\" (UID: \"05c1ec22-e0d6-45c4-85fc-d586c046b3b1\") " pod="openstack/openstack-galera-0" Oct 14 08:38:05 crc kubenswrapper[4870]: I1014 08:38:05.326376 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/05c1ec22-e0d6-45c4-85fc-d586c046b3b1-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"05c1ec22-e0d6-45c4-85fc-d586c046b3b1\") " pod="openstack/openstack-galera-0" Oct 14 08:38:05 crc kubenswrapper[4870]: I1014 08:38:05.327906 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/05c1ec22-e0d6-45c4-85fc-d586c046b3b1-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"05c1ec22-e0d6-45c4-85fc-d586c046b3b1\") " pod="openstack/openstack-galera-0" Oct 14 08:38:05 crc kubenswrapper[4870]: I1014 08:38:05.332320 4870 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 14 08:38:05 crc kubenswrapper[4870]: I1014 08:38:05.332350 4870 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-acda8db4-57c4-42f7-beb2-dc1d70fe6445\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-acda8db4-57c4-42f7-beb2-dc1d70fe6445\") pod \"openstack-galera-0\" (UID: \"05c1ec22-e0d6-45c4-85fc-d586c046b3b1\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/799a1b60640a114e37dc4ffdc48648f645cbdebf384c338106e4cf2c332e1cba/globalmount\"" pod="openstack/openstack-galera-0" Oct 14 08:38:05 crc kubenswrapper[4870]: I1014 08:38:05.339546 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xhwbk\" (UniqueName: \"kubernetes.io/projected/05c1ec22-e0d6-45c4-85fc-d586c046b3b1-kube-api-access-xhwbk\") pod \"openstack-galera-0\" (UID: \"05c1ec22-e0d6-45c4-85fc-d586c046b3b1\") " pod="openstack/openstack-galera-0" Oct 14 08:38:05 crc kubenswrapper[4870]: I1014 08:38:05.385749 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-acda8db4-57c4-42f7-beb2-dc1d70fe6445\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-acda8db4-57c4-42f7-beb2-dc1d70fe6445\") pod \"openstack-galera-0\" (UID: \"05c1ec22-e0d6-45c4-85fc-d586c046b3b1\") " pod="openstack/openstack-galera-0" Oct 14 08:38:05 crc kubenswrapper[4870]: I1014 08:38:05.463015 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Oct 14 08:38:05 crc kubenswrapper[4870]: I1014 08:38:05.605505 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Oct 14 08:38:05 crc kubenswrapper[4870]: I1014 08:38:05.606639 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Oct 14 08:38:05 crc kubenswrapper[4870]: I1014 08:38:05.613046 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Oct 14 08:38:05 crc kubenswrapper[4870]: I1014 08:38:05.613381 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-vnjb2" Oct 14 08:38:05 crc kubenswrapper[4870]: I1014 08:38:05.636761 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Oct 14 08:38:05 crc kubenswrapper[4870]: I1014 08:38:05.695071 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"8e2ef84d-2d3f-4a34-8dec-27e08689779e","Type":"ContainerStarted","Data":"d316b97d9c5edf108324aa5c4099c35b60e984ff57c60aae837af076cc2d9662"} Oct 14 08:38:05 crc kubenswrapper[4870]: I1014 08:38:05.696066 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"98a7db06-05d2-4c54-9321-11ad232caeda","Type":"ContainerStarted","Data":"798a8d68fd0b760faf65377048147fa0370de599771dcda098a7cb2ee75b96bf"} Oct 14 08:38:05 crc kubenswrapper[4870]: I1014 08:38:05.726811 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3d2c8717-a111-4e18-99e4-46be8b4b5673-config-data\") pod \"memcached-0\" (UID: \"3d2c8717-a111-4e18-99e4-46be8b4b5673\") " pod="openstack/memcached-0" Oct 14 08:38:05 crc kubenswrapper[4870]: I1014 08:38:05.726882 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/3d2c8717-a111-4e18-99e4-46be8b4b5673-kolla-config\") pod \"memcached-0\" (UID: \"3d2c8717-a111-4e18-99e4-46be8b4b5673\") " pod="openstack/memcached-0" Oct 14 08:38:05 crc kubenswrapper[4870]: I1014 08:38:05.726908 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9gw9d\" (UniqueName: \"kubernetes.io/projected/3d2c8717-a111-4e18-99e4-46be8b4b5673-kube-api-access-9gw9d\") pod \"memcached-0\" (UID: \"3d2c8717-a111-4e18-99e4-46be8b4b5673\") " pod="openstack/memcached-0" Oct 14 08:38:05 crc kubenswrapper[4870]: I1014 08:38:05.827955 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3d2c8717-a111-4e18-99e4-46be8b4b5673-config-data\") pod \"memcached-0\" (UID: \"3d2c8717-a111-4e18-99e4-46be8b4b5673\") " pod="openstack/memcached-0" Oct 14 08:38:05 crc kubenswrapper[4870]: I1014 08:38:05.828019 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/3d2c8717-a111-4e18-99e4-46be8b4b5673-kolla-config\") pod \"memcached-0\" (UID: \"3d2c8717-a111-4e18-99e4-46be8b4b5673\") " pod="openstack/memcached-0" Oct 14 08:38:05 crc kubenswrapper[4870]: I1014 08:38:05.828098 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9gw9d\" (UniqueName: \"kubernetes.io/projected/3d2c8717-a111-4e18-99e4-46be8b4b5673-kube-api-access-9gw9d\") pod \"memcached-0\" (UID: \"3d2c8717-a111-4e18-99e4-46be8b4b5673\") " pod="openstack/memcached-0" Oct 14 08:38:05 crc kubenswrapper[4870]: I1014 08:38:05.828898 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3d2c8717-a111-4e18-99e4-46be8b4b5673-config-data\") pod \"memcached-0\" (UID: \"3d2c8717-a111-4e18-99e4-46be8b4b5673\") " pod="openstack/memcached-0" Oct 14 08:38:05 crc kubenswrapper[4870]: I1014 08:38:05.829483 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/3d2c8717-a111-4e18-99e4-46be8b4b5673-kolla-config\") pod \"memcached-0\" (UID: \"3d2c8717-a111-4e18-99e4-46be8b4b5673\") " pod="openstack/memcached-0" Oct 14 08:38:05 crc kubenswrapper[4870]: I1014 08:38:05.864489 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9gw9d\" (UniqueName: \"kubernetes.io/projected/3d2c8717-a111-4e18-99e4-46be8b4b5673-kube-api-access-9gw9d\") pod \"memcached-0\" (UID: \"3d2c8717-a111-4e18-99e4-46be8b4b5673\") " pod="openstack/memcached-0" Oct 14 08:38:05 crc kubenswrapper[4870]: I1014 08:38:05.954883 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Oct 14 08:38:06 crc kubenswrapper[4870]: I1014 08:38:06.050876 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Oct 14 08:38:06 crc kubenswrapper[4870]: W1014 08:38:06.487659 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod05c1ec22_e0d6_45c4_85fc_d586c046b3b1.slice/crio-63b98e7b00a51fea5aeae28d388a366ee793215524b2a52e215a9019e92c9484 WatchSource:0}: Error finding container 63b98e7b00a51fea5aeae28d388a366ee793215524b2a52e215a9019e92c9484: Status 404 returned error can't find the container with id 63b98e7b00a51fea5aeae28d388a366ee793215524b2a52e215a9019e92c9484 Oct 14 08:38:06 crc kubenswrapper[4870]: I1014 08:38:06.704506 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"05c1ec22-e0d6-45c4-85fc-d586c046b3b1","Type":"ContainerStarted","Data":"63b98e7b00a51fea5aeae28d388a366ee793215524b2a52e215a9019e92c9484"} Oct 14 08:38:06 crc kubenswrapper[4870]: I1014 08:38:06.817368 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 14 08:38:06 crc kubenswrapper[4870]: I1014 08:38:06.820237 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Oct 14 08:38:06 crc kubenswrapper[4870]: I1014 08:38:06.823672 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-rds2l" Oct 14 08:38:06 crc kubenswrapper[4870]: I1014 08:38:06.823906 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Oct 14 08:38:06 crc kubenswrapper[4870]: I1014 08:38:06.824124 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Oct 14 08:38:06 crc kubenswrapper[4870]: I1014 08:38:06.824508 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Oct 14 08:38:06 crc kubenswrapper[4870]: I1014 08:38:06.829251 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 14 08:38:06 crc kubenswrapper[4870]: I1014 08:38:06.845414 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8cd09b24-b663-4aea-9ef4-36cf69ea689c-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"8cd09b24-b663-4aea-9ef4-36cf69ea689c\") " pod="openstack/openstack-cell1-galera-0" Oct 14 08:38:06 crc kubenswrapper[4870]: I1014 08:38:06.845477 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/8cd09b24-b663-4aea-9ef4-36cf69ea689c-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"8cd09b24-b663-4aea-9ef4-36cf69ea689c\") " pod="openstack/openstack-cell1-galera-0" Oct 14 08:38:06 crc kubenswrapper[4870]: I1014 08:38:06.845504 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/8cd09b24-b663-4aea-9ef4-36cf69ea689c-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"8cd09b24-b663-4aea-9ef4-36cf69ea689c\") " pod="openstack/openstack-cell1-galera-0" Oct 14 08:38:06 crc kubenswrapper[4870]: I1014 08:38:06.845577 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/8cd09b24-b663-4aea-9ef4-36cf69ea689c-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"8cd09b24-b663-4aea-9ef4-36cf69ea689c\") " pod="openstack/openstack-cell1-galera-0" Oct 14 08:38:06 crc kubenswrapper[4870]: I1014 08:38:06.845638 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/8cd09b24-b663-4aea-9ef4-36cf69ea689c-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"8cd09b24-b663-4aea-9ef4-36cf69ea689c\") " pod="openstack/openstack-cell1-galera-0" Oct 14 08:38:06 crc kubenswrapper[4870]: I1014 08:38:06.845714 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8cd09b24-b663-4aea-9ef4-36cf69ea689c-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"8cd09b24-b663-4aea-9ef4-36cf69ea689c\") " pod="openstack/openstack-cell1-galera-0" Oct 14 08:38:06 crc kubenswrapper[4870]: I1014 08:38:06.845768 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-4e8f4f4d-760a-4b05-b3d6-c6bf414a367b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-4e8f4f4d-760a-4b05-b3d6-c6bf414a367b\") pod \"openstack-cell1-galera-0\" (UID: \"8cd09b24-b663-4aea-9ef4-36cf69ea689c\") " pod="openstack/openstack-cell1-galera-0" Oct 14 08:38:06 crc kubenswrapper[4870]: I1014 08:38:06.845990 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/8cd09b24-b663-4aea-9ef4-36cf69ea689c-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"8cd09b24-b663-4aea-9ef4-36cf69ea689c\") " pod="openstack/openstack-cell1-galera-0" Oct 14 08:38:06 crc kubenswrapper[4870]: I1014 08:38:06.846068 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2xst5\" (UniqueName: \"kubernetes.io/projected/8cd09b24-b663-4aea-9ef4-36cf69ea689c-kube-api-access-2xst5\") pod \"openstack-cell1-galera-0\" (UID: \"8cd09b24-b663-4aea-9ef4-36cf69ea689c\") " pod="openstack/openstack-cell1-galera-0" Oct 14 08:38:06 crc kubenswrapper[4870]: I1014 08:38:06.873905 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Oct 14 08:38:06 crc kubenswrapper[4870]: W1014 08:38:06.881879 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3d2c8717_a111_4e18_99e4_46be8b4b5673.slice/crio-2920b358aabc34b772dd46555c85bc571c4e94db7673521ccb82adbb884e4c79 WatchSource:0}: Error finding container 2920b358aabc34b772dd46555c85bc571c4e94db7673521ccb82adbb884e4c79: Status 404 returned error can't find the container with id 2920b358aabc34b772dd46555c85bc571c4e94db7673521ccb82adbb884e4c79 Oct 14 08:38:06 crc kubenswrapper[4870]: I1014 08:38:06.947034 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8cd09b24-b663-4aea-9ef4-36cf69ea689c-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"8cd09b24-b663-4aea-9ef4-36cf69ea689c\") " pod="openstack/openstack-cell1-galera-0" Oct 14 08:38:06 crc kubenswrapper[4870]: I1014 08:38:06.947076 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/8cd09b24-b663-4aea-9ef4-36cf69ea689c-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"8cd09b24-b663-4aea-9ef4-36cf69ea689c\") " pod="openstack/openstack-cell1-galera-0" Oct 14 08:38:06 crc kubenswrapper[4870]: I1014 08:38:06.947096 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/8cd09b24-b663-4aea-9ef4-36cf69ea689c-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"8cd09b24-b663-4aea-9ef4-36cf69ea689c\") " pod="openstack/openstack-cell1-galera-0" Oct 14 08:38:06 crc kubenswrapper[4870]: I1014 08:38:06.947128 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/8cd09b24-b663-4aea-9ef4-36cf69ea689c-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"8cd09b24-b663-4aea-9ef4-36cf69ea689c\") " pod="openstack/openstack-cell1-galera-0" Oct 14 08:38:06 crc kubenswrapper[4870]: I1014 08:38:06.947174 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/8cd09b24-b663-4aea-9ef4-36cf69ea689c-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"8cd09b24-b663-4aea-9ef4-36cf69ea689c\") " pod="openstack/openstack-cell1-galera-0" Oct 14 08:38:06 crc kubenswrapper[4870]: I1014 08:38:06.947198 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8cd09b24-b663-4aea-9ef4-36cf69ea689c-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"8cd09b24-b663-4aea-9ef4-36cf69ea689c\") " pod="openstack/openstack-cell1-galera-0" Oct 14 08:38:06 crc kubenswrapper[4870]: I1014 08:38:06.947223 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-4e8f4f4d-760a-4b05-b3d6-c6bf414a367b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-4e8f4f4d-760a-4b05-b3d6-c6bf414a367b\") pod \"openstack-cell1-galera-0\" (UID: \"8cd09b24-b663-4aea-9ef4-36cf69ea689c\") " pod="openstack/openstack-cell1-galera-0" Oct 14 08:38:06 crc kubenswrapper[4870]: I1014 08:38:06.947253 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/8cd09b24-b663-4aea-9ef4-36cf69ea689c-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"8cd09b24-b663-4aea-9ef4-36cf69ea689c\") " pod="openstack/openstack-cell1-galera-0" Oct 14 08:38:06 crc kubenswrapper[4870]: I1014 08:38:06.947312 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2xst5\" (UniqueName: \"kubernetes.io/projected/8cd09b24-b663-4aea-9ef4-36cf69ea689c-kube-api-access-2xst5\") pod \"openstack-cell1-galera-0\" (UID: \"8cd09b24-b663-4aea-9ef4-36cf69ea689c\") " pod="openstack/openstack-cell1-galera-0" Oct 14 08:38:06 crc kubenswrapper[4870]: I1014 08:38:06.948612 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/8cd09b24-b663-4aea-9ef4-36cf69ea689c-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"8cd09b24-b663-4aea-9ef4-36cf69ea689c\") " pod="openstack/openstack-cell1-galera-0" Oct 14 08:38:06 crc kubenswrapper[4870]: I1014 08:38:06.949784 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/8cd09b24-b663-4aea-9ef4-36cf69ea689c-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"8cd09b24-b663-4aea-9ef4-36cf69ea689c\") " pod="openstack/openstack-cell1-galera-0" Oct 14 08:38:06 crc kubenswrapper[4870]: I1014 08:38:06.949950 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/8cd09b24-b663-4aea-9ef4-36cf69ea689c-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"8cd09b24-b663-4aea-9ef4-36cf69ea689c\") " pod="openstack/openstack-cell1-galera-0" Oct 14 08:38:06 crc kubenswrapper[4870]: I1014 08:38:06.950868 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8cd09b24-b663-4aea-9ef4-36cf69ea689c-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"8cd09b24-b663-4aea-9ef4-36cf69ea689c\") " pod="openstack/openstack-cell1-galera-0" Oct 14 08:38:06 crc kubenswrapper[4870]: I1014 08:38:06.953823 4870 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 14 08:38:06 crc kubenswrapper[4870]: I1014 08:38:06.953871 4870 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-4e8f4f4d-760a-4b05-b3d6-c6bf414a367b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-4e8f4f4d-760a-4b05-b3d6-c6bf414a367b\") pod \"openstack-cell1-galera-0\" (UID: \"8cd09b24-b663-4aea-9ef4-36cf69ea689c\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/0c256c24d59676db64f1a0c5a7465dc4d0f695a8698201426148018abfe581a5/globalmount\"" pod="openstack/openstack-cell1-galera-0" Oct 14 08:38:06 crc kubenswrapper[4870]: I1014 08:38:06.954652 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/8cd09b24-b663-4aea-9ef4-36cf69ea689c-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"8cd09b24-b663-4aea-9ef4-36cf69ea689c\") " pod="openstack/openstack-cell1-galera-0" Oct 14 08:38:06 crc kubenswrapper[4870]: I1014 08:38:06.962273 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8cd09b24-b663-4aea-9ef4-36cf69ea689c-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"8cd09b24-b663-4aea-9ef4-36cf69ea689c\") " pod="openstack/openstack-cell1-galera-0" Oct 14 08:38:06 crc kubenswrapper[4870]: I1014 08:38:06.962464 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/8cd09b24-b663-4aea-9ef4-36cf69ea689c-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"8cd09b24-b663-4aea-9ef4-36cf69ea689c\") " pod="openstack/openstack-cell1-galera-0" Oct 14 08:38:06 crc kubenswrapper[4870]: I1014 08:38:06.966401 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2xst5\" (UniqueName: \"kubernetes.io/projected/8cd09b24-b663-4aea-9ef4-36cf69ea689c-kube-api-access-2xst5\") pod \"openstack-cell1-galera-0\" (UID: \"8cd09b24-b663-4aea-9ef4-36cf69ea689c\") " pod="openstack/openstack-cell1-galera-0" Oct 14 08:38:06 crc kubenswrapper[4870]: I1014 08:38:06.989504 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-4e8f4f4d-760a-4b05-b3d6-c6bf414a367b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-4e8f4f4d-760a-4b05-b3d6-c6bf414a367b\") pod \"openstack-cell1-galera-0\" (UID: \"8cd09b24-b663-4aea-9ef4-36cf69ea689c\") " pod="openstack/openstack-cell1-galera-0" Oct 14 08:38:07 crc kubenswrapper[4870]: I1014 08:38:07.146753 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Oct 14 08:38:07 crc kubenswrapper[4870]: I1014 08:38:07.712306 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"3d2c8717-a111-4e18-99e4-46be8b4b5673","Type":"ContainerStarted","Data":"2920b358aabc34b772dd46555c85bc571c4e94db7673521ccb82adbb884e4c79"} Oct 14 08:38:09 crc kubenswrapper[4870]: I1014 08:38:09.829913 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 14 08:38:09 crc kubenswrapper[4870]: W1014 08:38:09.832987 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8cd09b24_b663_4aea_9ef4_36cf69ea689c.slice/crio-1aba3391786b79759fb8786d9570d27e0de65c48bc6f6b0511c4d278bd8d1348 WatchSource:0}: Error finding container 1aba3391786b79759fb8786d9570d27e0de65c48bc6f6b0511c4d278bd8d1348: Status 404 returned error can't find the container with id 1aba3391786b79759fb8786d9570d27e0de65c48bc6f6b0511c4d278bd8d1348 Oct 14 08:38:10 crc kubenswrapper[4870]: I1014 08:38:10.736057 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"8cd09b24-b663-4aea-9ef4-36cf69ea689c","Type":"ContainerStarted","Data":"1aba3391786b79759fb8786d9570d27e0de65c48bc6f6b0511c4d278bd8d1348"} Oct 14 08:38:18 crc kubenswrapper[4870]: I1014 08:38:18.805881 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"3d2c8717-a111-4e18-99e4-46be8b4b5673","Type":"ContainerStarted","Data":"d52ceedbaf1e285b8cb1797c85ee9df7303897ddcda80b230c4cbef66e04930b"} Oct 14 08:38:18 crc kubenswrapper[4870]: I1014 08:38:18.806541 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Oct 14 08:38:18 crc kubenswrapper[4870]: I1014 08:38:18.807875 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"05c1ec22-e0d6-45c4-85fc-d586c046b3b1","Type":"ContainerStarted","Data":"184801cb0d8a46769cb779045ff86b19164e767957a8212c27720516c747c965"} Oct 14 08:38:18 crc kubenswrapper[4870]: I1014 08:38:18.809931 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"8cd09b24-b663-4aea-9ef4-36cf69ea689c","Type":"ContainerStarted","Data":"2aee4f4cdee5cfc8fd541fd787d19de83f87d13156d000c7fadc744c2f79617d"} Oct 14 08:38:18 crc kubenswrapper[4870]: I1014 08:38:18.811988 4870 generic.go:334] "Generic (PLEG): container finished" podID="64d3a546-2e9d-4c15-bcb0-53bbdf6070b3" containerID="6439581ad9335c27c204b5ea46672e6a7ce578a8085a52123db3911f1dc5daaa" exitCode=0 Oct 14 08:38:18 crc kubenswrapper[4870]: I1014 08:38:18.812035 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-54bcdcd99f-wq6qt" event={"ID":"64d3a546-2e9d-4c15-bcb0-53bbdf6070b3","Type":"ContainerDied","Data":"6439581ad9335c27c204b5ea46672e6a7ce578a8085a52123db3911f1dc5daaa"} Oct 14 08:38:18 crc kubenswrapper[4870]: I1014 08:38:18.834745 4870 generic.go:334] "Generic (PLEG): container finished" podID="7482b8c0-86a7-4e3e-b05f-bb343129b9fb" containerID="8a92769c5ae97057f8efd3a4937d6ec40917a765f436ba65b2645bdc43b902f2" exitCode=0 Oct 14 08:38:18 crc kubenswrapper[4870]: I1014 08:38:18.834819 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7bd8799f99-g72nj" event={"ID":"7482b8c0-86a7-4e3e-b05f-bb343129b9fb","Type":"ContainerDied","Data":"8a92769c5ae97057f8efd3a4937d6ec40917a765f436ba65b2645bdc43b902f2"} Oct 14 08:38:18 crc kubenswrapper[4870]: I1014 08:38:18.841167 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=2.7849554100000002 podStartE2EDuration="13.84114845s" podCreationTimestamp="2025-10-14 08:38:05 +0000 UTC" firstStartedPulling="2025-10-14 08:38:06.884075472 +0000 UTC m=+5822.581435843" lastFinishedPulling="2025-10-14 08:38:17.940268472 +0000 UTC m=+5833.637628883" observedRunningTime="2025-10-14 08:38:18.832081669 +0000 UTC m=+5834.529442080" watchObservedRunningTime="2025-10-14 08:38:18.84114845 +0000 UTC m=+5834.538508831" Oct 14 08:38:19 crc kubenswrapper[4870]: I1014 08:38:19.847950 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7bd8799f99-g72nj" event={"ID":"7482b8c0-86a7-4e3e-b05f-bb343129b9fb","Type":"ContainerStarted","Data":"f648a4af2ebece0df50ab0d1a338f4bef248c64644715b5d99c805e3197170b2"} Oct 14 08:38:19 crc kubenswrapper[4870]: I1014 08:38:19.848335 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7bd8799f99-g72nj" Oct 14 08:38:19 crc kubenswrapper[4870]: I1014 08:38:19.850242 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"98a7db06-05d2-4c54-9321-11ad232caeda","Type":"ContainerStarted","Data":"3d14338acf97116dab677fd1d9d7d8f678ba6f131bc38563b2fb8fdddb7bc72e"} Oct 14 08:38:19 crc kubenswrapper[4870]: I1014 08:38:19.854958 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"8e2ef84d-2d3f-4a34-8dec-27e08689779e","Type":"ContainerStarted","Data":"b64502dd493bf8edf885099edf35cb1448b0e5102aa49008be1a2e1934486b35"} Oct 14 08:38:19 crc kubenswrapper[4870]: I1014 08:38:19.858685 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-54bcdcd99f-wq6qt" event={"ID":"64d3a546-2e9d-4c15-bcb0-53bbdf6070b3","Type":"ContainerStarted","Data":"d0df804fea6242a55b2c5b2eab0bbb6e8f48d9c5a84e84bdb0ea37cf2f734564"} Oct 14 08:38:19 crc kubenswrapper[4870]: I1014 08:38:19.859099 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-54bcdcd99f-wq6qt" Oct 14 08:38:19 crc kubenswrapper[4870]: I1014 08:38:19.872086 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7bd8799f99-g72nj" podStartSLOduration=3.9470243419999997 podStartE2EDuration="17.872062046s" podCreationTimestamp="2025-10-14 08:38:02 +0000 UTC" firstStartedPulling="2025-10-14 08:38:04.07567735 +0000 UTC m=+5819.773037711" lastFinishedPulling="2025-10-14 08:38:18.000715044 +0000 UTC m=+5833.698075415" observedRunningTime="2025-10-14 08:38:19.868620552 +0000 UTC m=+5835.565980923" watchObservedRunningTime="2025-10-14 08:38:19.872062046 +0000 UTC m=+5835.569422417" Oct 14 08:38:19 crc kubenswrapper[4870]: I1014 08:38:19.906888 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-54bcdcd99f-wq6qt" podStartSLOduration=4.026902617 podStartE2EDuration="17.906867644s" podCreationTimestamp="2025-10-14 08:38:02 +0000 UTC" firstStartedPulling="2025-10-14 08:38:04.120545692 +0000 UTC m=+5819.817906063" lastFinishedPulling="2025-10-14 08:38:18.000510709 +0000 UTC m=+5833.697871090" observedRunningTime="2025-10-14 08:38:19.89808364 +0000 UTC m=+5835.595444011" watchObservedRunningTime="2025-10-14 08:38:19.906867644 +0000 UTC m=+5835.604228015" Oct 14 08:38:21 crc kubenswrapper[4870]: I1014 08:38:21.881664 4870 generic.go:334] "Generic (PLEG): container finished" podID="05c1ec22-e0d6-45c4-85fc-d586c046b3b1" containerID="184801cb0d8a46769cb779045ff86b19164e767957a8212c27720516c747c965" exitCode=0 Oct 14 08:38:21 crc kubenswrapper[4870]: I1014 08:38:21.881741 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"05c1ec22-e0d6-45c4-85fc-d586c046b3b1","Type":"ContainerDied","Data":"184801cb0d8a46769cb779045ff86b19164e767957a8212c27720516c747c965"} Oct 14 08:38:21 crc kubenswrapper[4870]: I1014 08:38:21.886561 4870 generic.go:334] "Generic (PLEG): container finished" podID="8cd09b24-b663-4aea-9ef4-36cf69ea689c" containerID="2aee4f4cdee5cfc8fd541fd787d19de83f87d13156d000c7fadc744c2f79617d" exitCode=0 Oct 14 08:38:21 crc kubenswrapper[4870]: I1014 08:38:21.886654 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"8cd09b24-b663-4aea-9ef4-36cf69ea689c","Type":"ContainerDied","Data":"2aee4f4cdee5cfc8fd541fd787d19de83f87d13156d000c7fadc744c2f79617d"} Oct 14 08:38:22 crc kubenswrapper[4870]: I1014 08:38:22.896692 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"8cd09b24-b663-4aea-9ef4-36cf69ea689c","Type":"ContainerStarted","Data":"19489406c34bacefa145c418d28adb46a95659733b4d8459b0d73ac3884d5c6c"} Oct 14 08:38:22 crc kubenswrapper[4870]: I1014 08:38:22.899429 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"05c1ec22-e0d6-45c4-85fc-d586c046b3b1","Type":"ContainerStarted","Data":"db696365a775763817b148e747140fe4abe3b002e2b625ca87e703053fdda9a1"} Oct 14 08:38:22 crc kubenswrapper[4870]: I1014 08:38:22.924813 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=9.758093686 podStartE2EDuration="17.924795129s" podCreationTimestamp="2025-10-14 08:38:05 +0000 UTC" firstStartedPulling="2025-10-14 08:38:09.83683096 +0000 UTC m=+5825.534191331" lastFinishedPulling="2025-10-14 08:38:18.003532393 +0000 UTC m=+5833.700892774" observedRunningTime="2025-10-14 08:38:22.923262432 +0000 UTC m=+5838.620622833" watchObservedRunningTime="2025-10-14 08:38:22.924795129 +0000 UTC m=+5838.622155500" Oct 14 08:38:25 crc kubenswrapper[4870]: I1014 08:38:25.464006 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Oct 14 08:38:25 crc kubenswrapper[4870]: I1014 08:38:25.464336 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Oct 14 08:38:25 crc kubenswrapper[4870]: I1014 08:38:25.956182 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Oct 14 08:38:25 crc kubenswrapper[4870]: I1014 08:38:25.977842 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=10.4773601 podStartE2EDuration="21.97781942s" podCreationTimestamp="2025-10-14 08:38:04 +0000 UTC" firstStartedPulling="2025-10-14 08:38:06.500029429 +0000 UTC m=+5822.197389800" lastFinishedPulling="2025-10-14 08:38:18.000488749 +0000 UTC m=+5833.697849120" observedRunningTime="2025-10-14 08:38:22.960478778 +0000 UTC m=+5838.657839189" watchObservedRunningTime="2025-10-14 08:38:25.97781942 +0000 UTC m=+5841.675179801" Oct 14 08:38:27 crc kubenswrapper[4870]: I1014 08:38:27.147665 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Oct 14 08:38:27 crc kubenswrapper[4870]: I1014 08:38:27.147843 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Oct 14 08:38:27 crc kubenswrapper[4870]: I1014 08:38:27.221649 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Oct 14 08:38:27 crc kubenswrapper[4870]: I1014 08:38:27.942923 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7bd8799f99-g72nj" Oct 14 08:38:28 crc kubenswrapper[4870]: I1014 08:38:28.030806 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Oct 14 08:38:28 crc kubenswrapper[4870]: I1014 08:38:28.195856 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-54bcdcd99f-wq6qt" Oct 14 08:38:28 crc kubenswrapper[4870]: I1014 08:38:28.256545 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7bd8799f99-g72nj"] Oct 14 08:38:28 crc kubenswrapper[4870]: I1014 08:38:28.956272 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7bd8799f99-g72nj" podUID="7482b8c0-86a7-4e3e-b05f-bb343129b9fb" containerName="dnsmasq-dns" containerID="cri-o://f648a4af2ebece0df50ab0d1a338f4bef248c64644715b5d99c805e3197170b2" gracePeriod=10 Oct 14 08:38:29 crc kubenswrapper[4870]: I1014 08:38:29.510797 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7bd8799f99-g72nj" Oct 14 08:38:29 crc kubenswrapper[4870]: I1014 08:38:29.552654 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Oct 14 08:38:29 crc kubenswrapper[4870]: I1014 08:38:29.617392 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Oct 14 08:38:29 crc kubenswrapper[4870]: I1014 08:38:29.648837 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7482b8c0-86a7-4e3e-b05f-bb343129b9fb-config\") pod \"7482b8c0-86a7-4e3e-b05f-bb343129b9fb\" (UID: \"7482b8c0-86a7-4e3e-b05f-bb343129b9fb\") " Oct 14 08:38:29 crc kubenswrapper[4870]: I1014 08:38:29.649007 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7482b8c0-86a7-4e3e-b05f-bb343129b9fb-dns-svc\") pod \"7482b8c0-86a7-4e3e-b05f-bb343129b9fb\" (UID: \"7482b8c0-86a7-4e3e-b05f-bb343129b9fb\") " Oct 14 08:38:29 crc kubenswrapper[4870]: I1014 08:38:29.649076 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p9vcb\" (UniqueName: \"kubernetes.io/projected/7482b8c0-86a7-4e3e-b05f-bb343129b9fb-kube-api-access-p9vcb\") pod \"7482b8c0-86a7-4e3e-b05f-bb343129b9fb\" (UID: \"7482b8c0-86a7-4e3e-b05f-bb343129b9fb\") " Oct 14 08:38:29 crc kubenswrapper[4870]: I1014 08:38:29.655372 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7482b8c0-86a7-4e3e-b05f-bb343129b9fb-kube-api-access-p9vcb" (OuterVolumeSpecName: "kube-api-access-p9vcb") pod "7482b8c0-86a7-4e3e-b05f-bb343129b9fb" (UID: "7482b8c0-86a7-4e3e-b05f-bb343129b9fb"). InnerVolumeSpecName "kube-api-access-p9vcb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:38:29 crc kubenswrapper[4870]: I1014 08:38:29.697039 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7482b8c0-86a7-4e3e-b05f-bb343129b9fb-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "7482b8c0-86a7-4e3e-b05f-bb343129b9fb" (UID: "7482b8c0-86a7-4e3e-b05f-bb343129b9fb"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:38:29 crc kubenswrapper[4870]: I1014 08:38:29.700497 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7482b8c0-86a7-4e3e-b05f-bb343129b9fb-config" (OuterVolumeSpecName: "config") pod "7482b8c0-86a7-4e3e-b05f-bb343129b9fb" (UID: "7482b8c0-86a7-4e3e-b05f-bb343129b9fb"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:38:29 crc kubenswrapper[4870]: I1014 08:38:29.751136 4870 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7482b8c0-86a7-4e3e-b05f-bb343129b9fb-config\") on node \"crc\" DevicePath \"\"" Oct 14 08:38:29 crc kubenswrapper[4870]: I1014 08:38:29.751178 4870 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7482b8c0-86a7-4e3e-b05f-bb343129b9fb-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 14 08:38:29 crc kubenswrapper[4870]: I1014 08:38:29.751193 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p9vcb\" (UniqueName: \"kubernetes.io/projected/7482b8c0-86a7-4e3e-b05f-bb343129b9fb-kube-api-access-p9vcb\") on node \"crc\" DevicePath \"\"" Oct 14 08:38:29 crc kubenswrapper[4870]: I1014 08:38:29.963729 4870 generic.go:334] "Generic (PLEG): container finished" podID="7482b8c0-86a7-4e3e-b05f-bb343129b9fb" containerID="f648a4af2ebece0df50ab0d1a338f4bef248c64644715b5d99c805e3197170b2" exitCode=0 Oct 14 08:38:29 crc kubenswrapper[4870]: I1014 08:38:29.963789 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7bd8799f99-g72nj" Oct 14 08:38:29 crc kubenswrapper[4870]: I1014 08:38:29.963811 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7bd8799f99-g72nj" event={"ID":"7482b8c0-86a7-4e3e-b05f-bb343129b9fb","Type":"ContainerDied","Data":"f648a4af2ebece0df50ab0d1a338f4bef248c64644715b5d99c805e3197170b2"} Oct 14 08:38:29 crc kubenswrapper[4870]: I1014 08:38:29.964255 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7bd8799f99-g72nj" event={"ID":"7482b8c0-86a7-4e3e-b05f-bb343129b9fb","Type":"ContainerDied","Data":"694d032b9c8da5a064f94a58658638e7030bba153b0cbbbdf71e601ebcf930dc"} Oct 14 08:38:29 crc kubenswrapper[4870]: I1014 08:38:29.964282 4870 scope.go:117] "RemoveContainer" containerID="f648a4af2ebece0df50ab0d1a338f4bef248c64644715b5d99c805e3197170b2" Oct 14 08:38:29 crc kubenswrapper[4870]: I1014 08:38:29.995657 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7bd8799f99-g72nj"] Oct 14 08:38:30 crc kubenswrapper[4870]: I1014 08:38:30.003879 4870 scope.go:117] "RemoveContainer" containerID="8a92769c5ae97057f8efd3a4937d6ec40917a765f436ba65b2645bdc43b902f2" Oct 14 08:38:30 crc kubenswrapper[4870]: I1014 08:38:30.004957 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7bd8799f99-g72nj"] Oct 14 08:38:30 crc kubenswrapper[4870]: I1014 08:38:30.025864 4870 scope.go:117] "RemoveContainer" containerID="f648a4af2ebece0df50ab0d1a338f4bef248c64644715b5d99c805e3197170b2" Oct 14 08:38:30 crc kubenswrapper[4870]: E1014 08:38:30.026329 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f648a4af2ebece0df50ab0d1a338f4bef248c64644715b5d99c805e3197170b2\": container with ID starting with f648a4af2ebece0df50ab0d1a338f4bef248c64644715b5d99c805e3197170b2 not found: ID does not exist" containerID="f648a4af2ebece0df50ab0d1a338f4bef248c64644715b5d99c805e3197170b2" Oct 14 08:38:30 crc kubenswrapper[4870]: I1014 08:38:30.026357 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f648a4af2ebece0df50ab0d1a338f4bef248c64644715b5d99c805e3197170b2"} err="failed to get container status \"f648a4af2ebece0df50ab0d1a338f4bef248c64644715b5d99c805e3197170b2\": rpc error: code = NotFound desc = could not find container \"f648a4af2ebece0df50ab0d1a338f4bef248c64644715b5d99c805e3197170b2\": container with ID starting with f648a4af2ebece0df50ab0d1a338f4bef248c64644715b5d99c805e3197170b2 not found: ID does not exist" Oct 14 08:38:30 crc kubenswrapper[4870]: I1014 08:38:30.026377 4870 scope.go:117] "RemoveContainer" containerID="8a92769c5ae97057f8efd3a4937d6ec40917a765f436ba65b2645bdc43b902f2" Oct 14 08:38:30 crc kubenswrapper[4870]: E1014 08:38:30.026898 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8a92769c5ae97057f8efd3a4937d6ec40917a765f436ba65b2645bdc43b902f2\": container with ID starting with 8a92769c5ae97057f8efd3a4937d6ec40917a765f436ba65b2645bdc43b902f2 not found: ID does not exist" containerID="8a92769c5ae97057f8efd3a4937d6ec40917a765f436ba65b2645bdc43b902f2" Oct 14 08:38:30 crc kubenswrapper[4870]: I1014 08:38:30.026940 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8a92769c5ae97057f8efd3a4937d6ec40917a765f436ba65b2645bdc43b902f2"} err="failed to get container status \"8a92769c5ae97057f8efd3a4937d6ec40917a765f436ba65b2645bdc43b902f2\": rpc error: code = NotFound desc = could not find container \"8a92769c5ae97057f8efd3a4937d6ec40917a765f436ba65b2645bdc43b902f2\": container with ID starting with 8a92769c5ae97057f8efd3a4937d6ec40917a765f436ba65b2645bdc43b902f2 not found: ID does not exist" Oct 14 08:38:31 crc kubenswrapper[4870]: I1014 08:38:31.051236 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7482b8c0-86a7-4e3e-b05f-bb343129b9fb" path="/var/lib/kubelet/pods/7482b8c0-86a7-4e3e-b05f-bb343129b9fb/volumes" Oct 14 08:38:53 crc kubenswrapper[4870]: I1014 08:38:53.203922 4870 generic.go:334] "Generic (PLEG): container finished" podID="98a7db06-05d2-4c54-9321-11ad232caeda" containerID="3d14338acf97116dab677fd1d9d7d8f678ba6f131bc38563b2fb8fdddb7bc72e" exitCode=0 Oct 14 08:38:53 crc kubenswrapper[4870]: I1014 08:38:53.204014 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"98a7db06-05d2-4c54-9321-11ad232caeda","Type":"ContainerDied","Data":"3d14338acf97116dab677fd1d9d7d8f678ba6f131bc38563b2fb8fdddb7bc72e"} Oct 14 08:38:53 crc kubenswrapper[4870]: I1014 08:38:53.207428 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"8e2ef84d-2d3f-4a34-8dec-27e08689779e","Type":"ContainerDied","Data":"b64502dd493bf8edf885099edf35cb1448b0e5102aa49008be1a2e1934486b35"} Oct 14 08:38:53 crc kubenswrapper[4870]: I1014 08:38:53.207424 4870 generic.go:334] "Generic (PLEG): container finished" podID="8e2ef84d-2d3f-4a34-8dec-27e08689779e" containerID="b64502dd493bf8edf885099edf35cb1448b0e5102aa49008be1a2e1934486b35" exitCode=0 Oct 14 08:38:54 crc kubenswrapper[4870]: I1014 08:38:54.218964 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"8e2ef84d-2d3f-4a34-8dec-27e08689779e","Type":"ContainerStarted","Data":"5369a8d28111984dafc203f7d7c08881e6a18d5758459a04f6ea6b816684a89f"} Oct 14 08:38:54 crc kubenswrapper[4870]: I1014 08:38:54.219531 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Oct 14 08:38:54 crc kubenswrapper[4870]: I1014 08:38:54.221574 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"98a7db06-05d2-4c54-9321-11ad232caeda","Type":"ContainerStarted","Data":"f01402d691527160c02dbc3bd9a412ac37940e802dbce96abd9ae8d86c17e1a2"} Oct 14 08:38:54 crc kubenswrapper[4870]: I1014 08:38:54.222006 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:38:54 crc kubenswrapper[4870]: I1014 08:38:54.270877 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=39.23424946 podStartE2EDuration="52.270853809s" podCreationTimestamp="2025-10-14 08:38:02 +0000 UTC" firstStartedPulling="2025-10-14 08:38:04.9355817 +0000 UTC m=+5820.632942081" lastFinishedPulling="2025-10-14 08:38:17.972186059 +0000 UTC m=+5833.669546430" observedRunningTime="2025-10-14 08:38:54.268273446 +0000 UTC m=+5869.965633827" watchObservedRunningTime="2025-10-14 08:38:54.270853809 +0000 UTC m=+5869.968214190" Oct 14 08:38:54 crc kubenswrapper[4870]: I1014 08:38:54.275946 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=39.284654687 podStartE2EDuration="52.275928002s" podCreationTimestamp="2025-10-14 08:38:02 +0000 UTC" firstStartedPulling="2025-10-14 08:38:04.948959826 +0000 UTC m=+5820.646320197" lastFinishedPulling="2025-10-14 08:38:17.940233141 +0000 UTC m=+5833.637593512" observedRunningTime="2025-10-14 08:38:54.249218112 +0000 UTC m=+5869.946578493" watchObservedRunningTime="2025-10-14 08:38:54.275928002 +0000 UTC m=+5869.973288383" Oct 14 08:39:04 crc kubenswrapper[4870]: I1014 08:39:04.406269 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Oct 14 08:39:04 crc kubenswrapper[4870]: I1014 08:39:04.452496 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:39:11 crc kubenswrapper[4870]: I1014 08:39:11.116966 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7ccc84877c-2ngf9"] Oct 14 08:39:11 crc kubenswrapper[4870]: E1014 08:39:11.118021 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7482b8c0-86a7-4e3e-b05f-bb343129b9fb" containerName="init" Oct 14 08:39:11 crc kubenswrapper[4870]: I1014 08:39:11.118043 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="7482b8c0-86a7-4e3e-b05f-bb343129b9fb" containerName="init" Oct 14 08:39:11 crc kubenswrapper[4870]: E1014 08:39:11.118059 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7482b8c0-86a7-4e3e-b05f-bb343129b9fb" containerName="dnsmasq-dns" Oct 14 08:39:11 crc kubenswrapper[4870]: I1014 08:39:11.118067 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="7482b8c0-86a7-4e3e-b05f-bb343129b9fb" containerName="dnsmasq-dns" Oct 14 08:39:11 crc kubenswrapper[4870]: I1014 08:39:11.118255 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="7482b8c0-86a7-4e3e-b05f-bb343129b9fb" containerName="dnsmasq-dns" Oct 14 08:39:11 crc kubenswrapper[4870]: I1014 08:39:11.119275 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7ccc84877c-2ngf9" Oct 14 08:39:11 crc kubenswrapper[4870]: I1014 08:39:11.123888 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7ccc84877c-2ngf9"] Oct 14 08:39:11 crc kubenswrapper[4870]: I1014 08:39:11.278177 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rj2jz\" (UniqueName: \"kubernetes.io/projected/920f17fd-fc08-4a5e-9e7e-b641cc11528a-kube-api-access-rj2jz\") pod \"dnsmasq-dns-7ccc84877c-2ngf9\" (UID: \"920f17fd-fc08-4a5e-9e7e-b641cc11528a\") " pod="openstack/dnsmasq-dns-7ccc84877c-2ngf9" Oct 14 08:39:11 crc kubenswrapper[4870]: I1014 08:39:11.278872 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/920f17fd-fc08-4a5e-9e7e-b641cc11528a-config\") pod \"dnsmasq-dns-7ccc84877c-2ngf9\" (UID: \"920f17fd-fc08-4a5e-9e7e-b641cc11528a\") " pod="openstack/dnsmasq-dns-7ccc84877c-2ngf9" Oct 14 08:39:11 crc kubenswrapper[4870]: I1014 08:39:11.278996 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/920f17fd-fc08-4a5e-9e7e-b641cc11528a-dns-svc\") pod \"dnsmasq-dns-7ccc84877c-2ngf9\" (UID: \"920f17fd-fc08-4a5e-9e7e-b641cc11528a\") " pod="openstack/dnsmasq-dns-7ccc84877c-2ngf9" Oct 14 08:39:11 crc kubenswrapper[4870]: I1014 08:39:11.380820 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rj2jz\" (UniqueName: \"kubernetes.io/projected/920f17fd-fc08-4a5e-9e7e-b641cc11528a-kube-api-access-rj2jz\") pod \"dnsmasq-dns-7ccc84877c-2ngf9\" (UID: \"920f17fd-fc08-4a5e-9e7e-b641cc11528a\") " pod="openstack/dnsmasq-dns-7ccc84877c-2ngf9" Oct 14 08:39:11 crc kubenswrapper[4870]: I1014 08:39:11.380943 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/920f17fd-fc08-4a5e-9e7e-b641cc11528a-config\") pod \"dnsmasq-dns-7ccc84877c-2ngf9\" (UID: \"920f17fd-fc08-4a5e-9e7e-b641cc11528a\") " pod="openstack/dnsmasq-dns-7ccc84877c-2ngf9" Oct 14 08:39:11 crc kubenswrapper[4870]: I1014 08:39:11.380991 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/920f17fd-fc08-4a5e-9e7e-b641cc11528a-dns-svc\") pod \"dnsmasq-dns-7ccc84877c-2ngf9\" (UID: \"920f17fd-fc08-4a5e-9e7e-b641cc11528a\") " pod="openstack/dnsmasq-dns-7ccc84877c-2ngf9" Oct 14 08:39:11 crc kubenswrapper[4870]: I1014 08:39:11.381973 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/920f17fd-fc08-4a5e-9e7e-b641cc11528a-dns-svc\") pod \"dnsmasq-dns-7ccc84877c-2ngf9\" (UID: \"920f17fd-fc08-4a5e-9e7e-b641cc11528a\") " pod="openstack/dnsmasq-dns-7ccc84877c-2ngf9" Oct 14 08:39:11 crc kubenswrapper[4870]: I1014 08:39:11.382948 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/920f17fd-fc08-4a5e-9e7e-b641cc11528a-config\") pod \"dnsmasq-dns-7ccc84877c-2ngf9\" (UID: \"920f17fd-fc08-4a5e-9e7e-b641cc11528a\") " pod="openstack/dnsmasq-dns-7ccc84877c-2ngf9" Oct 14 08:39:11 crc kubenswrapper[4870]: I1014 08:39:11.411112 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rj2jz\" (UniqueName: \"kubernetes.io/projected/920f17fd-fc08-4a5e-9e7e-b641cc11528a-kube-api-access-rj2jz\") pod \"dnsmasq-dns-7ccc84877c-2ngf9\" (UID: \"920f17fd-fc08-4a5e-9e7e-b641cc11528a\") " pod="openstack/dnsmasq-dns-7ccc84877c-2ngf9" Oct 14 08:39:11 crc kubenswrapper[4870]: I1014 08:39:11.439142 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7ccc84877c-2ngf9" Oct 14 08:39:11 crc kubenswrapper[4870]: I1014 08:39:11.884820 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 14 08:39:11 crc kubenswrapper[4870]: I1014 08:39:11.923876 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7ccc84877c-2ngf9"] Oct 14 08:39:12 crc kubenswrapper[4870]: I1014 08:39:12.382803 4870 generic.go:334] "Generic (PLEG): container finished" podID="920f17fd-fc08-4a5e-9e7e-b641cc11528a" containerID="24710d7457adf6e2a83521401b65bb4a8a96f3a42e3ac761d2a2eb94b1acb0a2" exitCode=0 Oct 14 08:39:12 crc kubenswrapper[4870]: I1014 08:39:12.382990 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7ccc84877c-2ngf9" event={"ID":"920f17fd-fc08-4a5e-9e7e-b641cc11528a","Type":"ContainerDied","Data":"24710d7457adf6e2a83521401b65bb4a8a96f3a42e3ac761d2a2eb94b1acb0a2"} Oct 14 08:39:12 crc kubenswrapper[4870]: I1014 08:39:12.383110 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7ccc84877c-2ngf9" event={"ID":"920f17fd-fc08-4a5e-9e7e-b641cc11528a","Type":"ContainerStarted","Data":"83a45ecc32ec34f4626f3e65f0dbe7d5ac105a88a6aa1f4dd3d3c217b46b0ea0"} Oct 14 08:39:12 crc kubenswrapper[4870]: I1014 08:39:12.696121 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 14 08:39:13 crc kubenswrapper[4870]: I1014 08:39:13.394105 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7ccc84877c-2ngf9" event={"ID":"920f17fd-fc08-4a5e-9e7e-b641cc11528a","Type":"ContainerStarted","Data":"d1f99638f9a135316382888657472849639a8a7ae1413b7da513e34c1354d997"} Oct 14 08:39:13 crc kubenswrapper[4870]: I1014 08:39:13.394664 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7ccc84877c-2ngf9" Oct 14 08:39:13 crc kubenswrapper[4870]: I1014 08:39:13.569314 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="8e2ef84d-2d3f-4a34-8dec-27e08689779e" containerName="rabbitmq" containerID="cri-o://5369a8d28111984dafc203f7d7c08881e6a18d5758459a04f6ea6b816684a89f" gracePeriod=604799 Oct 14 08:39:14 crc kubenswrapper[4870]: I1014 08:39:14.397795 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="98a7db06-05d2-4c54-9321-11ad232caeda" containerName="rabbitmq" containerID="cri-o://f01402d691527160c02dbc3bd9a412ac37940e802dbce96abd9ae8d86c17e1a2" gracePeriod=604799 Oct 14 08:39:14 crc kubenswrapper[4870]: I1014 08:39:14.403141 4870 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="8e2ef84d-2d3f-4a34-8dec-27e08689779e" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.246:5672: connect: connection refused" Oct 14 08:39:14 crc kubenswrapper[4870]: I1014 08:39:14.449804 4870 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="98a7db06-05d2-4c54-9321-11ad232caeda" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.247:5672: connect: connection refused" Oct 14 08:39:20 crc kubenswrapper[4870]: I1014 08:39:20.267842 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 14 08:39:20 crc kubenswrapper[4870]: I1014 08:39:20.310742 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7ccc84877c-2ngf9" podStartSLOduration=9.310715541 podStartE2EDuration="9.310715541s" podCreationTimestamp="2025-10-14 08:39:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 08:39:13.417174283 +0000 UTC m=+5889.114534654" watchObservedRunningTime="2025-10-14 08:39:20.310715541 +0000 UTC m=+5896.008075922" Oct 14 08:39:20 crc kubenswrapper[4870]: I1014 08:39:20.332870 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/8e2ef84d-2d3f-4a34-8dec-27e08689779e-pod-info\") pod \"8e2ef84d-2d3f-4a34-8dec-27e08689779e\" (UID: \"8e2ef84d-2d3f-4a34-8dec-27e08689779e\") " Oct 14 08:39:20 crc kubenswrapper[4870]: I1014 08:39:20.333329 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/8e2ef84d-2d3f-4a34-8dec-27e08689779e-server-conf\") pod \"8e2ef84d-2d3f-4a34-8dec-27e08689779e\" (UID: \"8e2ef84d-2d3f-4a34-8dec-27e08689779e\") " Oct 14 08:39:20 crc kubenswrapper[4870]: I1014 08:39:20.333392 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/8e2ef84d-2d3f-4a34-8dec-27e08689779e-erlang-cookie-secret\") pod \"8e2ef84d-2d3f-4a34-8dec-27e08689779e\" (UID: \"8e2ef84d-2d3f-4a34-8dec-27e08689779e\") " Oct 14 08:39:20 crc kubenswrapper[4870]: I1014 08:39:20.333510 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t4vf8\" (UniqueName: \"kubernetes.io/projected/8e2ef84d-2d3f-4a34-8dec-27e08689779e-kube-api-access-t4vf8\") pod \"8e2ef84d-2d3f-4a34-8dec-27e08689779e\" (UID: \"8e2ef84d-2d3f-4a34-8dec-27e08689779e\") " Oct 14 08:39:20 crc kubenswrapper[4870]: I1014 08:39:20.333559 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/8e2ef84d-2d3f-4a34-8dec-27e08689779e-plugins-conf\") pod \"8e2ef84d-2d3f-4a34-8dec-27e08689779e\" (UID: \"8e2ef84d-2d3f-4a34-8dec-27e08689779e\") " Oct 14 08:39:20 crc kubenswrapper[4870]: I1014 08:39:20.333716 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-dc261995-757b-493e-a935-9844b1fe3864\") pod \"8e2ef84d-2d3f-4a34-8dec-27e08689779e\" (UID: \"8e2ef84d-2d3f-4a34-8dec-27e08689779e\") " Oct 14 08:39:20 crc kubenswrapper[4870]: I1014 08:39:20.333785 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/8e2ef84d-2d3f-4a34-8dec-27e08689779e-rabbitmq-erlang-cookie\") pod \"8e2ef84d-2d3f-4a34-8dec-27e08689779e\" (UID: \"8e2ef84d-2d3f-4a34-8dec-27e08689779e\") " Oct 14 08:39:20 crc kubenswrapper[4870]: I1014 08:39:20.333824 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/8e2ef84d-2d3f-4a34-8dec-27e08689779e-rabbitmq-confd\") pod \"8e2ef84d-2d3f-4a34-8dec-27e08689779e\" (UID: \"8e2ef84d-2d3f-4a34-8dec-27e08689779e\") " Oct 14 08:39:20 crc kubenswrapper[4870]: I1014 08:39:20.333899 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/8e2ef84d-2d3f-4a34-8dec-27e08689779e-rabbitmq-plugins\") pod \"8e2ef84d-2d3f-4a34-8dec-27e08689779e\" (UID: \"8e2ef84d-2d3f-4a34-8dec-27e08689779e\") " Oct 14 08:39:20 crc kubenswrapper[4870]: I1014 08:39:20.335586 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8e2ef84d-2d3f-4a34-8dec-27e08689779e-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "8e2ef84d-2d3f-4a34-8dec-27e08689779e" (UID: "8e2ef84d-2d3f-4a34-8dec-27e08689779e"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:39:20 crc kubenswrapper[4870]: I1014 08:39:20.339393 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8e2ef84d-2d3f-4a34-8dec-27e08689779e-kube-api-access-t4vf8" (OuterVolumeSpecName: "kube-api-access-t4vf8") pod "8e2ef84d-2d3f-4a34-8dec-27e08689779e" (UID: "8e2ef84d-2d3f-4a34-8dec-27e08689779e"). InnerVolumeSpecName "kube-api-access-t4vf8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:39:20 crc kubenswrapper[4870]: I1014 08:39:20.341959 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8e2ef84d-2d3f-4a34-8dec-27e08689779e-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "8e2ef84d-2d3f-4a34-8dec-27e08689779e" (UID: "8e2ef84d-2d3f-4a34-8dec-27e08689779e"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:39:20 crc kubenswrapper[4870]: I1014 08:39:20.342874 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8e2ef84d-2d3f-4a34-8dec-27e08689779e-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "8e2ef84d-2d3f-4a34-8dec-27e08689779e" (UID: "8e2ef84d-2d3f-4a34-8dec-27e08689779e"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:39:20 crc kubenswrapper[4870]: I1014 08:39:20.343035 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8e2ef84d-2d3f-4a34-8dec-27e08689779e-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "8e2ef84d-2d3f-4a34-8dec-27e08689779e" (UID: "8e2ef84d-2d3f-4a34-8dec-27e08689779e"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:39:20 crc kubenswrapper[4870]: I1014 08:39:20.344593 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/8e2ef84d-2d3f-4a34-8dec-27e08689779e-pod-info" (OuterVolumeSpecName: "pod-info") pod "8e2ef84d-2d3f-4a34-8dec-27e08689779e" (UID: "8e2ef84d-2d3f-4a34-8dec-27e08689779e"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Oct 14 08:39:20 crc kubenswrapper[4870]: I1014 08:39:20.361366 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-dc261995-757b-493e-a935-9844b1fe3864" (OuterVolumeSpecName: "persistence") pod "8e2ef84d-2d3f-4a34-8dec-27e08689779e" (UID: "8e2ef84d-2d3f-4a34-8dec-27e08689779e"). InnerVolumeSpecName "pvc-dc261995-757b-493e-a935-9844b1fe3864". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 14 08:39:20 crc kubenswrapper[4870]: I1014 08:39:20.401175 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8e2ef84d-2d3f-4a34-8dec-27e08689779e-server-conf" (OuterVolumeSpecName: "server-conf") pod "8e2ef84d-2d3f-4a34-8dec-27e08689779e" (UID: "8e2ef84d-2d3f-4a34-8dec-27e08689779e"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:39:20 crc kubenswrapper[4870]: I1014 08:39:20.436283 4870 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/8e2ef84d-2d3f-4a34-8dec-27e08689779e-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Oct 14 08:39:20 crc kubenswrapper[4870]: I1014 08:39:20.436355 4870 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/8e2ef84d-2d3f-4a34-8dec-27e08689779e-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Oct 14 08:39:20 crc kubenswrapper[4870]: I1014 08:39:20.436404 4870 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/8e2ef84d-2d3f-4a34-8dec-27e08689779e-pod-info\") on node \"crc\" DevicePath \"\"" Oct 14 08:39:20 crc kubenswrapper[4870]: I1014 08:39:20.436428 4870 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/8e2ef84d-2d3f-4a34-8dec-27e08689779e-server-conf\") on node \"crc\" DevicePath \"\"" Oct 14 08:39:20 crc kubenswrapper[4870]: I1014 08:39:20.436463 4870 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/8e2ef84d-2d3f-4a34-8dec-27e08689779e-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Oct 14 08:39:20 crc kubenswrapper[4870]: I1014 08:39:20.436481 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t4vf8\" (UniqueName: \"kubernetes.io/projected/8e2ef84d-2d3f-4a34-8dec-27e08689779e-kube-api-access-t4vf8\") on node \"crc\" DevicePath \"\"" Oct 14 08:39:20 crc kubenswrapper[4870]: I1014 08:39:20.436495 4870 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/8e2ef84d-2d3f-4a34-8dec-27e08689779e-plugins-conf\") on node \"crc\" DevicePath \"\"" Oct 14 08:39:20 crc kubenswrapper[4870]: I1014 08:39:20.436546 4870 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-dc261995-757b-493e-a935-9844b1fe3864\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-dc261995-757b-493e-a935-9844b1fe3864\") on node \"crc\" " Oct 14 08:39:20 crc kubenswrapper[4870]: I1014 08:39:20.437216 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8e2ef84d-2d3f-4a34-8dec-27e08689779e-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "8e2ef84d-2d3f-4a34-8dec-27e08689779e" (UID: "8e2ef84d-2d3f-4a34-8dec-27e08689779e"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:39:20 crc kubenswrapper[4870]: I1014 08:39:20.459530 4870 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Oct 14 08:39:20 crc kubenswrapper[4870]: I1014 08:39:20.459954 4870 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-dc261995-757b-493e-a935-9844b1fe3864" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-dc261995-757b-493e-a935-9844b1fe3864") on node "crc" Oct 14 08:39:20 crc kubenswrapper[4870]: I1014 08:39:20.467329 4870 generic.go:334] "Generic (PLEG): container finished" podID="8e2ef84d-2d3f-4a34-8dec-27e08689779e" containerID="5369a8d28111984dafc203f7d7c08881e6a18d5758459a04f6ea6b816684a89f" exitCode=0 Oct 14 08:39:20 crc kubenswrapper[4870]: I1014 08:39:20.467573 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"8e2ef84d-2d3f-4a34-8dec-27e08689779e","Type":"ContainerDied","Data":"5369a8d28111984dafc203f7d7c08881e6a18d5758459a04f6ea6b816684a89f"} Oct 14 08:39:20 crc kubenswrapper[4870]: I1014 08:39:20.467698 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"8e2ef84d-2d3f-4a34-8dec-27e08689779e","Type":"ContainerDied","Data":"d316b97d9c5edf108324aa5c4099c35b60e984ff57c60aae837af076cc2d9662"} Oct 14 08:39:20 crc kubenswrapper[4870]: I1014 08:39:20.467842 4870 scope.go:117] "RemoveContainer" containerID="5369a8d28111984dafc203f7d7c08881e6a18d5758459a04f6ea6b816684a89f" Oct 14 08:39:20 crc kubenswrapper[4870]: I1014 08:39:20.468093 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 14 08:39:20 crc kubenswrapper[4870]: I1014 08:39:20.504101 4870 scope.go:117] "RemoveContainer" containerID="b64502dd493bf8edf885099edf35cb1448b0e5102aa49008be1a2e1934486b35" Oct 14 08:39:20 crc kubenswrapper[4870]: I1014 08:39:20.521745 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 14 08:39:20 crc kubenswrapper[4870]: I1014 08:39:20.532826 4870 scope.go:117] "RemoveContainer" containerID="5369a8d28111984dafc203f7d7c08881e6a18d5758459a04f6ea6b816684a89f" Oct 14 08:39:20 crc kubenswrapper[4870]: E1014 08:39:20.535929 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5369a8d28111984dafc203f7d7c08881e6a18d5758459a04f6ea6b816684a89f\": container with ID starting with 5369a8d28111984dafc203f7d7c08881e6a18d5758459a04f6ea6b816684a89f not found: ID does not exist" containerID="5369a8d28111984dafc203f7d7c08881e6a18d5758459a04f6ea6b816684a89f" Oct 14 08:39:20 crc kubenswrapper[4870]: I1014 08:39:20.535998 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5369a8d28111984dafc203f7d7c08881e6a18d5758459a04f6ea6b816684a89f"} err="failed to get container status \"5369a8d28111984dafc203f7d7c08881e6a18d5758459a04f6ea6b816684a89f\": rpc error: code = NotFound desc = could not find container \"5369a8d28111984dafc203f7d7c08881e6a18d5758459a04f6ea6b816684a89f\": container with ID starting with 5369a8d28111984dafc203f7d7c08881e6a18d5758459a04f6ea6b816684a89f not found: ID does not exist" Oct 14 08:39:20 crc kubenswrapper[4870]: I1014 08:39:20.536049 4870 scope.go:117] "RemoveContainer" containerID="b64502dd493bf8edf885099edf35cb1448b0e5102aa49008be1a2e1934486b35" Oct 14 08:39:20 crc kubenswrapper[4870]: E1014 08:39:20.537897 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b64502dd493bf8edf885099edf35cb1448b0e5102aa49008be1a2e1934486b35\": container with ID starting with b64502dd493bf8edf885099edf35cb1448b0e5102aa49008be1a2e1934486b35 not found: ID does not exist" containerID="b64502dd493bf8edf885099edf35cb1448b0e5102aa49008be1a2e1934486b35" Oct 14 08:39:20 crc kubenswrapper[4870]: I1014 08:39:20.537926 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b64502dd493bf8edf885099edf35cb1448b0e5102aa49008be1a2e1934486b35"} err="failed to get container status \"b64502dd493bf8edf885099edf35cb1448b0e5102aa49008be1a2e1934486b35\": rpc error: code = NotFound desc = could not find container \"b64502dd493bf8edf885099edf35cb1448b0e5102aa49008be1a2e1934486b35\": container with ID starting with b64502dd493bf8edf885099edf35cb1448b0e5102aa49008be1a2e1934486b35 not found: ID does not exist" Oct 14 08:39:20 crc kubenswrapper[4870]: I1014 08:39:20.538039 4870 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/8e2ef84d-2d3f-4a34-8dec-27e08689779e-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Oct 14 08:39:20 crc kubenswrapper[4870]: I1014 08:39:20.538071 4870 reconciler_common.go:293] "Volume detached for volume \"pvc-dc261995-757b-493e-a935-9844b1fe3864\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-dc261995-757b-493e-a935-9844b1fe3864\") on node \"crc\" DevicePath \"\"" Oct 14 08:39:20 crc kubenswrapper[4870]: I1014 08:39:20.548569 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 14 08:39:20 crc kubenswrapper[4870]: I1014 08:39:20.562295 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Oct 14 08:39:20 crc kubenswrapper[4870]: E1014 08:39:20.562777 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e2ef84d-2d3f-4a34-8dec-27e08689779e" containerName="setup-container" Oct 14 08:39:20 crc kubenswrapper[4870]: I1014 08:39:20.562816 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e2ef84d-2d3f-4a34-8dec-27e08689779e" containerName="setup-container" Oct 14 08:39:20 crc kubenswrapper[4870]: E1014 08:39:20.562857 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e2ef84d-2d3f-4a34-8dec-27e08689779e" containerName="rabbitmq" Oct 14 08:39:20 crc kubenswrapper[4870]: I1014 08:39:20.562864 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e2ef84d-2d3f-4a34-8dec-27e08689779e" containerName="rabbitmq" Oct 14 08:39:20 crc kubenswrapper[4870]: I1014 08:39:20.563021 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="8e2ef84d-2d3f-4a34-8dec-27e08689779e" containerName="rabbitmq" Oct 14 08:39:20 crc kubenswrapper[4870]: I1014 08:39:20.563888 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 14 08:39:20 crc kubenswrapper[4870]: I1014 08:39:20.566653 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Oct 14 08:39:20 crc kubenswrapper[4870]: I1014 08:39:20.566966 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Oct 14 08:39:20 crc kubenswrapper[4870]: I1014 08:39:20.567307 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Oct 14 08:39:20 crc kubenswrapper[4870]: I1014 08:39:20.567729 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-7fhn4" Oct 14 08:39:20 crc kubenswrapper[4870]: I1014 08:39:20.569523 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Oct 14 08:39:20 crc kubenswrapper[4870]: I1014 08:39:20.576700 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 14 08:39:20 crc kubenswrapper[4870]: I1014 08:39:20.639199 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/a40d8a1a-605f-4db1-bf5f-4e4785efaa4e-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"a40d8a1a-605f-4db1-bf5f-4e4785efaa4e\") " pod="openstack/rabbitmq-server-0" Oct 14 08:39:20 crc kubenswrapper[4870]: I1014 08:39:20.639244 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/a40d8a1a-605f-4db1-bf5f-4e4785efaa4e-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"a40d8a1a-605f-4db1-bf5f-4e4785efaa4e\") " pod="openstack/rabbitmq-server-0" Oct 14 08:39:20 crc kubenswrapper[4870]: I1014 08:39:20.639274 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/a40d8a1a-605f-4db1-bf5f-4e4785efaa4e-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"a40d8a1a-605f-4db1-bf5f-4e4785efaa4e\") " pod="openstack/rabbitmq-server-0" Oct 14 08:39:20 crc kubenswrapper[4870]: I1014 08:39:20.639290 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/a40d8a1a-605f-4db1-bf5f-4e4785efaa4e-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"a40d8a1a-605f-4db1-bf5f-4e4785efaa4e\") " pod="openstack/rabbitmq-server-0" Oct 14 08:39:20 crc kubenswrapper[4870]: I1014 08:39:20.639329 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-trvsp\" (UniqueName: \"kubernetes.io/projected/a40d8a1a-605f-4db1-bf5f-4e4785efaa4e-kube-api-access-trvsp\") pod \"rabbitmq-server-0\" (UID: \"a40d8a1a-605f-4db1-bf5f-4e4785efaa4e\") " pod="openstack/rabbitmq-server-0" Oct 14 08:39:20 crc kubenswrapper[4870]: I1014 08:39:20.639351 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/a40d8a1a-605f-4db1-bf5f-4e4785efaa4e-pod-info\") pod \"rabbitmq-server-0\" (UID: \"a40d8a1a-605f-4db1-bf5f-4e4785efaa4e\") " pod="openstack/rabbitmq-server-0" Oct 14 08:39:20 crc kubenswrapper[4870]: I1014 08:39:20.639378 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/a40d8a1a-605f-4db1-bf5f-4e4785efaa4e-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"a40d8a1a-605f-4db1-bf5f-4e4785efaa4e\") " pod="openstack/rabbitmq-server-0" Oct 14 08:39:20 crc kubenswrapper[4870]: I1014 08:39:20.639393 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/a40d8a1a-605f-4db1-bf5f-4e4785efaa4e-server-conf\") pod \"rabbitmq-server-0\" (UID: \"a40d8a1a-605f-4db1-bf5f-4e4785efaa4e\") " pod="openstack/rabbitmq-server-0" Oct 14 08:39:20 crc kubenswrapper[4870]: I1014 08:39:20.639462 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-dc261995-757b-493e-a935-9844b1fe3864\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-dc261995-757b-493e-a935-9844b1fe3864\") pod \"rabbitmq-server-0\" (UID: \"a40d8a1a-605f-4db1-bf5f-4e4785efaa4e\") " pod="openstack/rabbitmq-server-0" Oct 14 08:39:20 crc kubenswrapper[4870]: I1014 08:39:20.741194 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/a40d8a1a-605f-4db1-bf5f-4e4785efaa4e-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"a40d8a1a-605f-4db1-bf5f-4e4785efaa4e\") " pod="openstack/rabbitmq-server-0" Oct 14 08:39:20 crc kubenswrapper[4870]: I1014 08:39:20.741258 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/a40d8a1a-605f-4db1-bf5f-4e4785efaa4e-server-conf\") pod \"rabbitmq-server-0\" (UID: \"a40d8a1a-605f-4db1-bf5f-4e4785efaa4e\") " pod="openstack/rabbitmq-server-0" Oct 14 08:39:20 crc kubenswrapper[4870]: I1014 08:39:20.741358 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-dc261995-757b-493e-a935-9844b1fe3864\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-dc261995-757b-493e-a935-9844b1fe3864\") pod \"rabbitmq-server-0\" (UID: \"a40d8a1a-605f-4db1-bf5f-4e4785efaa4e\") " pod="openstack/rabbitmq-server-0" Oct 14 08:39:20 crc kubenswrapper[4870]: I1014 08:39:20.741412 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/a40d8a1a-605f-4db1-bf5f-4e4785efaa4e-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"a40d8a1a-605f-4db1-bf5f-4e4785efaa4e\") " pod="openstack/rabbitmq-server-0" Oct 14 08:39:20 crc kubenswrapper[4870]: I1014 08:39:20.741478 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/a40d8a1a-605f-4db1-bf5f-4e4785efaa4e-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"a40d8a1a-605f-4db1-bf5f-4e4785efaa4e\") " pod="openstack/rabbitmq-server-0" Oct 14 08:39:20 crc kubenswrapper[4870]: I1014 08:39:20.741531 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/a40d8a1a-605f-4db1-bf5f-4e4785efaa4e-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"a40d8a1a-605f-4db1-bf5f-4e4785efaa4e\") " pod="openstack/rabbitmq-server-0" Oct 14 08:39:20 crc kubenswrapper[4870]: I1014 08:39:20.741566 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/a40d8a1a-605f-4db1-bf5f-4e4785efaa4e-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"a40d8a1a-605f-4db1-bf5f-4e4785efaa4e\") " pod="openstack/rabbitmq-server-0" Oct 14 08:39:20 crc kubenswrapper[4870]: I1014 08:39:20.741642 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-trvsp\" (UniqueName: \"kubernetes.io/projected/a40d8a1a-605f-4db1-bf5f-4e4785efaa4e-kube-api-access-trvsp\") pod \"rabbitmq-server-0\" (UID: \"a40d8a1a-605f-4db1-bf5f-4e4785efaa4e\") " pod="openstack/rabbitmq-server-0" Oct 14 08:39:20 crc kubenswrapper[4870]: I1014 08:39:20.741679 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/a40d8a1a-605f-4db1-bf5f-4e4785efaa4e-pod-info\") pod \"rabbitmq-server-0\" (UID: \"a40d8a1a-605f-4db1-bf5f-4e4785efaa4e\") " pod="openstack/rabbitmq-server-0" Oct 14 08:39:20 crc kubenswrapper[4870]: I1014 08:39:20.742631 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/a40d8a1a-605f-4db1-bf5f-4e4785efaa4e-server-conf\") pod \"rabbitmq-server-0\" (UID: \"a40d8a1a-605f-4db1-bf5f-4e4785efaa4e\") " pod="openstack/rabbitmq-server-0" Oct 14 08:39:20 crc kubenswrapper[4870]: I1014 08:39:20.742882 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/a40d8a1a-605f-4db1-bf5f-4e4785efaa4e-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"a40d8a1a-605f-4db1-bf5f-4e4785efaa4e\") " pod="openstack/rabbitmq-server-0" Oct 14 08:39:20 crc kubenswrapper[4870]: I1014 08:39:20.743705 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/a40d8a1a-605f-4db1-bf5f-4e4785efaa4e-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"a40d8a1a-605f-4db1-bf5f-4e4785efaa4e\") " pod="openstack/rabbitmq-server-0" Oct 14 08:39:20 crc kubenswrapper[4870]: I1014 08:39:20.743938 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/a40d8a1a-605f-4db1-bf5f-4e4785efaa4e-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"a40d8a1a-605f-4db1-bf5f-4e4785efaa4e\") " pod="openstack/rabbitmq-server-0" Oct 14 08:39:20 crc kubenswrapper[4870]: I1014 08:39:20.744820 4870 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 14 08:39:20 crc kubenswrapper[4870]: I1014 08:39:20.744845 4870 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-dc261995-757b-493e-a935-9844b1fe3864\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-dc261995-757b-493e-a935-9844b1fe3864\") pod \"rabbitmq-server-0\" (UID: \"a40d8a1a-605f-4db1-bf5f-4e4785efaa4e\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/e4cd1d45c92fe27d183871dbe3d8ad739ccac095190c427f162da6f6556256e5/globalmount\"" pod="openstack/rabbitmq-server-0" Oct 14 08:39:20 crc kubenswrapper[4870]: I1014 08:39:20.746607 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/a40d8a1a-605f-4db1-bf5f-4e4785efaa4e-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"a40d8a1a-605f-4db1-bf5f-4e4785efaa4e\") " pod="openstack/rabbitmq-server-0" Oct 14 08:39:20 crc kubenswrapper[4870]: I1014 08:39:20.748605 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/a40d8a1a-605f-4db1-bf5f-4e4785efaa4e-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"a40d8a1a-605f-4db1-bf5f-4e4785efaa4e\") " pod="openstack/rabbitmq-server-0" Oct 14 08:39:20 crc kubenswrapper[4870]: I1014 08:39:20.749641 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/a40d8a1a-605f-4db1-bf5f-4e4785efaa4e-pod-info\") pod \"rabbitmq-server-0\" (UID: \"a40d8a1a-605f-4db1-bf5f-4e4785efaa4e\") " pod="openstack/rabbitmq-server-0" Oct 14 08:39:20 crc kubenswrapper[4870]: I1014 08:39:20.759228 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-trvsp\" (UniqueName: \"kubernetes.io/projected/a40d8a1a-605f-4db1-bf5f-4e4785efaa4e-kube-api-access-trvsp\") pod \"rabbitmq-server-0\" (UID: \"a40d8a1a-605f-4db1-bf5f-4e4785efaa4e\") " pod="openstack/rabbitmq-server-0" Oct 14 08:39:20 crc kubenswrapper[4870]: I1014 08:39:20.772267 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-dc261995-757b-493e-a935-9844b1fe3864\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-dc261995-757b-493e-a935-9844b1fe3864\") pod \"rabbitmq-server-0\" (UID: \"a40d8a1a-605f-4db1-bf5f-4e4785efaa4e\") " pod="openstack/rabbitmq-server-0" Oct 14 08:39:20 crc kubenswrapper[4870]: I1014 08:39:20.930013 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 14 08:39:21 crc kubenswrapper[4870]: I1014 08:39:21.034210 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:39:21 crc kubenswrapper[4870]: I1014 08:39:21.069169 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8e2ef84d-2d3f-4a34-8dec-27e08689779e" path="/var/lib/kubelet/pods/8e2ef84d-2d3f-4a34-8dec-27e08689779e/volumes" Oct 14 08:39:21 crc kubenswrapper[4870]: I1014 08:39:21.147394 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/98a7db06-05d2-4c54-9321-11ad232caeda-erlang-cookie-secret\") pod \"98a7db06-05d2-4c54-9321-11ad232caeda\" (UID: \"98a7db06-05d2-4c54-9321-11ad232caeda\") " Oct 14 08:39:21 crc kubenswrapper[4870]: I1014 08:39:21.147465 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/98a7db06-05d2-4c54-9321-11ad232caeda-rabbitmq-confd\") pod \"98a7db06-05d2-4c54-9321-11ad232caeda\" (UID: \"98a7db06-05d2-4c54-9321-11ad232caeda\") " Oct 14 08:39:21 crc kubenswrapper[4870]: I1014 08:39:21.147502 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-whccn\" (UniqueName: \"kubernetes.io/projected/98a7db06-05d2-4c54-9321-11ad232caeda-kube-api-access-whccn\") pod \"98a7db06-05d2-4c54-9321-11ad232caeda\" (UID: \"98a7db06-05d2-4c54-9321-11ad232caeda\") " Oct 14 08:39:21 crc kubenswrapper[4870]: I1014 08:39:21.147548 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/98a7db06-05d2-4c54-9321-11ad232caeda-server-conf\") pod \"98a7db06-05d2-4c54-9321-11ad232caeda\" (UID: \"98a7db06-05d2-4c54-9321-11ad232caeda\") " Oct 14 08:39:21 crc kubenswrapper[4870]: I1014 08:39:21.147578 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/98a7db06-05d2-4c54-9321-11ad232caeda-rabbitmq-plugins\") pod \"98a7db06-05d2-4c54-9321-11ad232caeda\" (UID: \"98a7db06-05d2-4c54-9321-11ad232caeda\") " Oct 14 08:39:21 crc kubenswrapper[4870]: I1014 08:39:21.147596 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/98a7db06-05d2-4c54-9321-11ad232caeda-rabbitmq-erlang-cookie\") pod \"98a7db06-05d2-4c54-9321-11ad232caeda\" (UID: \"98a7db06-05d2-4c54-9321-11ad232caeda\") " Oct 14 08:39:21 crc kubenswrapper[4870]: I1014 08:39:21.147735 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7f3d7a0e-92c6-421f-ac96-6a093ce92216\") pod \"98a7db06-05d2-4c54-9321-11ad232caeda\" (UID: \"98a7db06-05d2-4c54-9321-11ad232caeda\") " Oct 14 08:39:21 crc kubenswrapper[4870]: I1014 08:39:21.147790 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/98a7db06-05d2-4c54-9321-11ad232caeda-plugins-conf\") pod \"98a7db06-05d2-4c54-9321-11ad232caeda\" (UID: \"98a7db06-05d2-4c54-9321-11ad232caeda\") " Oct 14 08:39:21 crc kubenswrapper[4870]: I1014 08:39:21.147831 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/98a7db06-05d2-4c54-9321-11ad232caeda-pod-info\") pod \"98a7db06-05d2-4c54-9321-11ad232caeda\" (UID: \"98a7db06-05d2-4c54-9321-11ad232caeda\") " Oct 14 08:39:21 crc kubenswrapper[4870]: I1014 08:39:21.148651 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/98a7db06-05d2-4c54-9321-11ad232caeda-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "98a7db06-05d2-4c54-9321-11ad232caeda" (UID: "98a7db06-05d2-4c54-9321-11ad232caeda"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:39:21 crc kubenswrapper[4870]: I1014 08:39:21.148636 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/98a7db06-05d2-4c54-9321-11ad232caeda-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "98a7db06-05d2-4c54-9321-11ad232caeda" (UID: "98a7db06-05d2-4c54-9321-11ad232caeda"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:39:21 crc kubenswrapper[4870]: I1014 08:39:21.149390 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/98a7db06-05d2-4c54-9321-11ad232caeda-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "98a7db06-05d2-4c54-9321-11ad232caeda" (UID: "98a7db06-05d2-4c54-9321-11ad232caeda"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:39:21 crc kubenswrapper[4870]: I1014 08:39:21.155122 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/98a7db06-05d2-4c54-9321-11ad232caeda-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "98a7db06-05d2-4c54-9321-11ad232caeda" (UID: "98a7db06-05d2-4c54-9321-11ad232caeda"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:39:21 crc kubenswrapper[4870]: I1014 08:39:21.155249 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/98a7db06-05d2-4c54-9321-11ad232caeda-pod-info" (OuterVolumeSpecName: "pod-info") pod "98a7db06-05d2-4c54-9321-11ad232caeda" (UID: "98a7db06-05d2-4c54-9321-11ad232caeda"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Oct 14 08:39:21 crc kubenswrapper[4870]: I1014 08:39:21.162513 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7f3d7a0e-92c6-421f-ac96-6a093ce92216" (OuterVolumeSpecName: "persistence") pod "98a7db06-05d2-4c54-9321-11ad232caeda" (UID: "98a7db06-05d2-4c54-9321-11ad232caeda"). InnerVolumeSpecName "pvc-7f3d7a0e-92c6-421f-ac96-6a093ce92216". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 14 08:39:21 crc kubenswrapper[4870]: I1014 08:39:21.162748 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/98a7db06-05d2-4c54-9321-11ad232caeda-kube-api-access-whccn" (OuterVolumeSpecName: "kube-api-access-whccn") pod "98a7db06-05d2-4c54-9321-11ad232caeda" (UID: "98a7db06-05d2-4c54-9321-11ad232caeda"). InnerVolumeSpecName "kube-api-access-whccn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:39:21 crc kubenswrapper[4870]: I1014 08:39:21.171092 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/98a7db06-05d2-4c54-9321-11ad232caeda-server-conf" (OuterVolumeSpecName: "server-conf") pod "98a7db06-05d2-4c54-9321-11ad232caeda" (UID: "98a7db06-05d2-4c54-9321-11ad232caeda"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:39:21 crc kubenswrapper[4870]: I1014 08:39:21.241940 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/98a7db06-05d2-4c54-9321-11ad232caeda-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "98a7db06-05d2-4c54-9321-11ad232caeda" (UID: "98a7db06-05d2-4c54-9321-11ad232caeda"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:39:21 crc kubenswrapper[4870]: I1014 08:39:21.249162 4870 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/98a7db06-05d2-4c54-9321-11ad232caeda-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Oct 14 08:39:21 crc kubenswrapper[4870]: I1014 08:39:21.249190 4870 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/98a7db06-05d2-4c54-9321-11ad232caeda-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Oct 14 08:39:21 crc kubenswrapper[4870]: I1014 08:39:21.249206 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-whccn\" (UniqueName: \"kubernetes.io/projected/98a7db06-05d2-4c54-9321-11ad232caeda-kube-api-access-whccn\") on node \"crc\" DevicePath \"\"" Oct 14 08:39:21 crc kubenswrapper[4870]: I1014 08:39:21.249219 4870 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/98a7db06-05d2-4c54-9321-11ad232caeda-server-conf\") on node \"crc\" DevicePath \"\"" Oct 14 08:39:21 crc kubenswrapper[4870]: I1014 08:39:21.249231 4870 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/98a7db06-05d2-4c54-9321-11ad232caeda-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Oct 14 08:39:21 crc kubenswrapper[4870]: I1014 08:39:21.249244 4870 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/98a7db06-05d2-4c54-9321-11ad232caeda-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Oct 14 08:39:21 crc kubenswrapper[4870]: I1014 08:39:21.249280 4870 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-7f3d7a0e-92c6-421f-ac96-6a093ce92216\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7f3d7a0e-92c6-421f-ac96-6a093ce92216\") on node \"crc\" " Oct 14 08:39:21 crc kubenswrapper[4870]: I1014 08:39:21.249356 4870 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/98a7db06-05d2-4c54-9321-11ad232caeda-plugins-conf\") on node \"crc\" DevicePath \"\"" Oct 14 08:39:21 crc kubenswrapper[4870]: I1014 08:39:21.249373 4870 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/98a7db06-05d2-4c54-9321-11ad232caeda-pod-info\") on node \"crc\" DevicePath \"\"" Oct 14 08:39:21 crc kubenswrapper[4870]: I1014 08:39:21.264180 4870 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Oct 14 08:39:21 crc kubenswrapper[4870]: I1014 08:39:21.264355 4870 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-7f3d7a0e-92c6-421f-ac96-6a093ce92216" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7f3d7a0e-92c6-421f-ac96-6a093ce92216") on node "crc" Oct 14 08:39:21 crc kubenswrapper[4870]: I1014 08:39:21.351206 4870 reconciler_common.go:293] "Volume detached for volume \"pvc-7f3d7a0e-92c6-421f-ac96-6a093ce92216\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7f3d7a0e-92c6-421f-ac96-6a093ce92216\") on node \"crc\" DevicePath \"\"" Oct 14 08:39:21 crc kubenswrapper[4870]: I1014 08:39:21.405752 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 14 08:39:21 crc kubenswrapper[4870]: W1014 08:39:21.421420 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda40d8a1a_605f_4db1_bf5f_4e4785efaa4e.slice/crio-1e67981cee3b5ba70848594a68df8cb990ad547e8789716f6c31fdd07e719b65 WatchSource:0}: Error finding container 1e67981cee3b5ba70848594a68df8cb990ad547e8789716f6c31fdd07e719b65: Status 404 returned error can't find the container with id 1e67981cee3b5ba70848594a68df8cb990ad547e8789716f6c31fdd07e719b65 Oct 14 08:39:21 crc kubenswrapper[4870]: I1014 08:39:21.442091 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7ccc84877c-2ngf9" Oct 14 08:39:21 crc kubenswrapper[4870]: I1014 08:39:21.481067 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"a40d8a1a-605f-4db1-bf5f-4e4785efaa4e","Type":"ContainerStarted","Data":"1e67981cee3b5ba70848594a68df8cb990ad547e8789716f6c31fdd07e719b65"} Oct 14 08:39:21 crc kubenswrapper[4870]: I1014 08:39:21.489944 4870 generic.go:334] "Generic (PLEG): container finished" podID="98a7db06-05d2-4c54-9321-11ad232caeda" containerID="f01402d691527160c02dbc3bd9a412ac37940e802dbce96abd9ae8d86c17e1a2" exitCode=0 Oct 14 08:39:21 crc kubenswrapper[4870]: I1014 08:39:21.490136 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:39:21 crc kubenswrapper[4870]: I1014 08:39:21.490901 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"98a7db06-05d2-4c54-9321-11ad232caeda","Type":"ContainerDied","Data":"f01402d691527160c02dbc3bd9a412ac37940e802dbce96abd9ae8d86c17e1a2"} Oct 14 08:39:21 crc kubenswrapper[4870]: I1014 08:39:21.490969 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"98a7db06-05d2-4c54-9321-11ad232caeda","Type":"ContainerDied","Data":"798a8d68fd0b760faf65377048147fa0370de599771dcda098a7cb2ee75b96bf"} Oct 14 08:39:21 crc kubenswrapper[4870]: I1014 08:39:21.491003 4870 scope.go:117] "RemoveContainer" containerID="f01402d691527160c02dbc3bd9a412ac37940e802dbce96abd9ae8d86c17e1a2" Oct 14 08:39:21 crc kubenswrapper[4870]: I1014 08:39:21.503619 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-54bcdcd99f-wq6qt"] Oct 14 08:39:21 crc kubenswrapper[4870]: I1014 08:39:21.503989 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-54bcdcd99f-wq6qt" podUID="64d3a546-2e9d-4c15-bcb0-53bbdf6070b3" containerName="dnsmasq-dns" containerID="cri-o://d0df804fea6242a55b2c5b2eab0bbb6e8f48d9c5a84e84bdb0ea37cf2f734564" gracePeriod=10 Oct 14 08:39:21 crc kubenswrapper[4870]: I1014 08:39:21.561915 4870 scope.go:117] "RemoveContainer" containerID="3d14338acf97116dab677fd1d9d7d8f678ba6f131bc38563b2fb8fdddb7bc72e" Oct 14 08:39:21 crc kubenswrapper[4870]: I1014 08:39:21.562121 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 14 08:39:21 crc kubenswrapper[4870]: I1014 08:39:21.569157 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 14 08:39:21 crc kubenswrapper[4870]: I1014 08:39:21.587785 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 14 08:39:21 crc kubenswrapper[4870]: E1014 08:39:21.588138 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="98a7db06-05d2-4c54-9321-11ad232caeda" containerName="setup-container" Oct 14 08:39:21 crc kubenswrapper[4870]: I1014 08:39:21.588152 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="98a7db06-05d2-4c54-9321-11ad232caeda" containerName="setup-container" Oct 14 08:39:21 crc kubenswrapper[4870]: E1014 08:39:21.588174 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="98a7db06-05d2-4c54-9321-11ad232caeda" containerName="rabbitmq" Oct 14 08:39:21 crc kubenswrapper[4870]: I1014 08:39:21.588181 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="98a7db06-05d2-4c54-9321-11ad232caeda" containerName="rabbitmq" Oct 14 08:39:21 crc kubenswrapper[4870]: I1014 08:39:21.588365 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="98a7db06-05d2-4c54-9321-11ad232caeda" containerName="rabbitmq" Oct 14 08:39:21 crc kubenswrapper[4870]: I1014 08:39:21.589539 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:39:21 crc kubenswrapper[4870]: I1014 08:39:21.596919 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Oct 14 08:39:21 crc kubenswrapper[4870]: I1014 08:39:21.597100 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Oct 14 08:39:21 crc kubenswrapper[4870]: I1014 08:39:21.597207 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-kpv9q" Oct 14 08:39:21 crc kubenswrapper[4870]: I1014 08:39:21.597638 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Oct 14 08:39:21 crc kubenswrapper[4870]: I1014 08:39:21.597887 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Oct 14 08:39:21 crc kubenswrapper[4870]: I1014 08:39:21.605350 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 14 08:39:21 crc kubenswrapper[4870]: I1014 08:39:21.607020 4870 scope.go:117] "RemoveContainer" containerID="f01402d691527160c02dbc3bd9a412ac37940e802dbce96abd9ae8d86c17e1a2" Oct 14 08:39:21 crc kubenswrapper[4870]: E1014 08:39:21.607524 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f01402d691527160c02dbc3bd9a412ac37940e802dbce96abd9ae8d86c17e1a2\": container with ID starting with f01402d691527160c02dbc3bd9a412ac37940e802dbce96abd9ae8d86c17e1a2 not found: ID does not exist" containerID="f01402d691527160c02dbc3bd9a412ac37940e802dbce96abd9ae8d86c17e1a2" Oct 14 08:39:21 crc kubenswrapper[4870]: I1014 08:39:21.607547 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f01402d691527160c02dbc3bd9a412ac37940e802dbce96abd9ae8d86c17e1a2"} err="failed to get container status \"f01402d691527160c02dbc3bd9a412ac37940e802dbce96abd9ae8d86c17e1a2\": rpc error: code = NotFound desc = could not find container \"f01402d691527160c02dbc3bd9a412ac37940e802dbce96abd9ae8d86c17e1a2\": container with ID starting with f01402d691527160c02dbc3bd9a412ac37940e802dbce96abd9ae8d86c17e1a2 not found: ID does not exist" Oct 14 08:39:21 crc kubenswrapper[4870]: I1014 08:39:21.607565 4870 scope.go:117] "RemoveContainer" containerID="3d14338acf97116dab677fd1d9d7d8f678ba6f131bc38563b2fb8fdddb7bc72e" Oct 14 08:39:21 crc kubenswrapper[4870]: E1014 08:39:21.607887 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3d14338acf97116dab677fd1d9d7d8f678ba6f131bc38563b2fb8fdddb7bc72e\": container with ID starting with 3d14338acf97116dab677fd1d9d7d8f678ba6f131bc38563b2fb8fdddb7bc72e not found: ID does not exist" containerID="3d14338acf97116dab677fd1d9d7d8f678ba6f131bc38563b2fb8fdddb7bc72e" Oct 14 08:39:21 crc kubenswrapper[4870]: I1014 08:39:21.607913 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3d14338acf97116dab677fd1d9d7d8f678ba6f131bc38563b2fb8fdddb7bc72e"} err="failed to get container status \"3d14338acf97116dab677fd1d9d7d8f678ba6f131bc38563b2fb8fdddb7bc72e\": rpc error: code = NotFound desc = could not find container \"3d14338acf97116dab677fd1d9d7d8f678ba6f131bc38563b2fb8fdddb7bc72e\": container with ID starting with 3d14338acf97116dab677fd1d9d7d8f678ba6f131bc38563b2fb8fdddb7bc72e not found: ID does not exist" Oct 14 08:39:21 crc kubenswrapper[4870]: I1014 08:39:21.658786 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/e07b718e-6d3b-491e-9229-d30e5228635c-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"e07b718e-6d3b-491e-9229-d30e5228635c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:39:21 crc kubenswrapper[4870]: I1014 08:39:21.658859 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/e07b718e-6d3b-491e-9229-d30e5228635c-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"e07b718e-6d3b-491e-9229-d30e5228635c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:39:21 crc kubenswrapper[4870]: I1014 08:39:21.658887 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/e07b718e-6d3b-491e-9229-d30e5228635c-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"e07b718e-6d3b-491e-9229-d30e5228635c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:39:21 crc kubenswrapper[4870]: I1014 08:39:21.658912 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/e07b718e-6d3b-491e-9229-d30e5228635c-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"e07b718e-6d3b-491e-9229-d30e5228635c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:39:21 crc kubenswrapper[4870]: I1014 08:39:21.658936 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/e07b718e-6d3b-491e-9229-d30e5228635c-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"e07b718e-6d3b-491e-9229-d30e5228635c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:39:21 crc kubenswrapper[4870]: I1014 08:39:21.658954 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-78khs\" (UniqueName: \"kubernetes.io/projected/e07b718e-6d3b-491e-9229-d30e5228635c-kube-api-access-78khs\") pod \"rabbitmq-cell1-server-0\" (UID: \"e07b718e-6d3b-491e-9229-d30e5228635c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:39:21 crc kubenswrapper[4870]: I1014 08:39:21.658974 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/e07b718e-6d3b-491e-9229-d30e5228635c-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"e07b718e-6d3b-491e-9229-d30e5228635c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:39:21 crc kubenswrapper[4870]: I1014 08:39:21.658995 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/e07b718e-6d3b-491e-9229-d30e5228635c-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"e07b718e-6d3b-491e-9229-d30e5228635c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:39:21 crc kubenswrapper[4870]: I1014 08:39:21.659022 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-7f3d7a0e-92c6-421f-ac96-6a093ce92216\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7f3d7a0e-92c6-421f-ac96-6a093ce92216\") pod \"rabbitmq-cell1-server-0\" (UID: \"e07b718e-6d3b-491e-9229-d30e5228635c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:39:21 crc kubenswrapper[4870]: I1014 08:39:21.760999 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/e07b718e-6d3b-491e-9229-d30e5228635c-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"e07b718e-6d3b-491e-9229-d30e5228635c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:39:21 crc kubenswrapper[4870]: I1014 08:39:21.761133 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/e07b718e-6d3b-491e-9229-d30e5228635c-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"e07b718e-6d3b-491e-9229-d30e5228635c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:39:21 crc kubenswrapper[4870]: I1014 08:39:21.761188 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/e07b718e-6d3b-491e-9229-d30e5228635c-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"e07b718e-6d3b-491e-9229-d30e5228635c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:39:21 crc kubenswrapper[4870]: I1014 08:39:21.761235 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/e07b718e-6d3b-491e-9229-d30e5228635c-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"e07b718e-6d3b-491e-9229-d30e5228635c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:39:21 crc kubenswrapper[4870]: I1014 08:39:21.761275 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/e07b718e-6d3b-491e-9229-d30e5228635c-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"e07b718e-6d3b-491e-9229-d30e5228635c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:39:21 crc kubenswrapper[4870]: I1014 08:39:21.761307 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-78khs\" (UniqueName: \"kubernetes.io/projected/e07b718e-6d3b-491e-9229-d30e5228635c-kube-api-access-78khs\") pod \"rabbitmq-cell1-server-0\" (UID: \"e07b718e-6d3b-491e-9229-d30e5228635c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:39:21 crc kubenswrapper[4870]: I1014 08:39:21.761746 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/e07b718e-6d3b-491e-9229-d30e5228635c-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"e07b718e-6d3b-491e-9229-d30e5228635c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:39:21 crc kubenswrapper[4870]: I1014 08:39:21.761809 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/e07b718e-6d3b-491e-9229-d30e5228635c-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"e07b718e-6d3b-491e-9229-d30e5228635c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:39:21 crc kubenswrapper[4870]: I1014 08:39:21.761851 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/e07b718e-6d3b-491e-9229-d30e5228635c-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"e07b718e-6d3b-491e-9229-d30e5228635c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:39:21 crc kubenswrapper[4870]: I1014 08:39:21.761904 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-7f3d7a0e-92c6-421f-ac96-6a093ce92216\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7f3d7a0e-92c6-421f-ac96-6a093ce92216\") pod \"rabbitmq-cell1-server-0\" (UID: \"e07b718e-6d3b-491e-9229-d30e5228635c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:39:21 crc kubenswrapper[4870]: I1014 08:39:21.762170 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/e07b718e-6d3b-491e-9229-d30e5228635c-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"e07b718e-6d3b-491e-9229-d30e5228635c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:39:21 crc kubenswrapper[4870]: I1014 08:39:21.762177 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/e07b718e-6d3b-491e-9229-d30e5228635c-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"e07b718e-6d3b-491e-9229-d30e5228635c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:39:21 crc kubenswrapper[4870]: I1014 08:39:21.763157 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/e07b718e-6d3b-491e-9229-d30e5228635c-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"e07b718e-6d3b-491e-9229-d30e5228635c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:39:21 crc kubenswrapper[4870]: I1014 08:39:21.764570 4870 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 14 08:39:21 crc kubenswrapper[4870]: I1014 08:39:21.764593 4870 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-7f3d7a0e-92c6-421f-ac96-6a093ce92216\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7f3d7a0e-92c6-421f-ac96-6a093ce92216\") pod \"rabbitmq-cell1-server-0\" (UID: \"e07b718e-6d3b-491e-9229-d30e5228635c\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/5ef42abd3faf11736bf20237aa539234ecec053653af230c7b5d392cae64bad5/globalmount\"" pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:39:21 crc kubenswrapper[4870]: I1014 08:39:21.765732 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/e07b718e-6d3b-491e-9229-d30e5228635c-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"e07b718e-6d3b-491e-9229-d30e5228635c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:39:21 crc kubenswrapper[4870]: I1014 08:39:21.767164 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/e07b718e-6d3b-491e-9229-d30e5228635c-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"e07b718e-6d3b-491e-9229-d30e5228635c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:39:21 crc kubenswrapper[4870]: I1014 08:39:21.768106 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/e07b718e-6d3b-491e-9229-d30e5228635c-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"e07b718e-6d3b-491e-9229-d30e5228635c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:39:21 crc kubenswrapper[4870]: I1014 08:39:21.778495 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-78khs\" (UniqueName: \"kubernetes.io/projected/e07b718e-6d3b-491e-9229-d30e5228635c-kube-api-access-78khs\") pod \"rabbitmq-cell1-server-0\" (UID: \"e07b718e-6d3b-491e-9229-d30e5228635c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:39:21 crc kubenswrapper[4870]: I1014 08:39:21.811649 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-7f3d7a0e-92c6-421f-ac96-6a093ce92216\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7f3d7a0e-92c6-421f-ac96-6a093ce92216\") pod \"rabbitmq-cell1-server-0\" (UID: \"e07b718e-6d3b-491e-9229-d30e5228635c\") " pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:39:21 crc kubenswrapper[4870]: I1014 08:39:21.917133 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:39:22 crc kubenswrapper[4870]: I1014 08:39:22.029854 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-54bcdcd99f-wq6qt" Oct 14 08:39:22 crc kubenswrapper[4870]: I1014 08:39:22.067180 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/64d3a546-2e9d-4c15-bcb0-53bbdf6070b3-dns-svc\") pod \"64d3a546-2e9d-4c15-bcb0-53bbdf6070b3\" (UID: \"64d3a546-2e9d-4c15-bcb0-53bbdf6070b3\") " Oct 14 08:39:22 crc kubenswrapper[4870]: I1014 08:39:22.067269 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cgd65\" (UniqueName: \"kubernetes.io/projected/64d3a546-2e9d-4c15-bcb0-53bbdf6070b3-kube-api-access-cgd65\") pod \"64d3a546-2e9d-4c15-bcb0-53bbdf6070b3\" (UID: \"64d3a546-2e9d-4c15-bcb0-53bbdf6070b3\") " Oct 14 08:39:22 crc kubenswrapper[4870]: I1014 08:39:22.067413 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/64d3a546-2e9d-4c15-bcb0-53bbdf6070b3-config\") pod \"64d3a546-2e9d-4c15-bcb0-53bbdf6070b3\" (UID: \"64d3a546-2e9d-4c15-bcb0-53bbdf6070b3\") " Oct 14 08:39:22 crc kubenswrapper[4870]: I1014 08:39:22.078480 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/64d3a546-2e9d-4c15-bcb0-53bbdf6070b3-kube-api-access-cgd65" (OuterVolumeSpecName: "kube-api-access-cgd65") pod "64d3a546-2e9d-4c15-bcb0-53bbdf6070b3" (UID: "64d3a546-2e9d-4c15-bcb0-53bbdf6070b3"). InnerVolumeSpecName "kube-api-access-cgd65". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:39:22 crc kubenswrapper[4870]: I1014 08:39:22.124478 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/64d3a546-2e9d-4c15-bcb0-53bbdf6070b3-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "64d3a546-2e9d-4c15-bcb0-53bbdf6070b3" (UID: "64d3a546-2e9d-4c15-bcb0-53bbdf6070b3"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:39:22 crc kubenswrapper[4870]: I1014 08:39:22.127717 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/64d3a546-2e9d-4c15-bcb0-53bbdf6070b3-config" (OuterVolumeSpecName: "config") pod "64d3a546-2e9d-4c15-bcb0-53bbdf6070b3" (UID: "64d3a546-2e9d-4c15-bcb0-53bbdf6070b3"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:39:22 crc kubenswrapper[4870]: I1014 08:39:22.168432 4870 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/64d3a546-2e9d-4c15-bcb0-53bbdf6070b3-config\") on node \"crc\" DevicePath \"\"" Oct 14 08:39:22 crc kubenswrapper[4870]: I1014 08:39:22.168473 4870 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/64d3a546-2e9d-4c15-bcb0-53bbdf6070b3-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 14 08:39:22 crc kubenswrapper[4870]: I1014 08:39:22.168484 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cgd65\" (UniqueName: \"kubernetes.io/projected/64d3a546-2e9d-4c15-bcb0-53bbdf6070b3-kube-api-access-cgd65\") on node \"crc\" DevicePath \"\"" Oct 14 08:39:22 crc kubenswrapper[4870]: I1014 08:39:22.406797 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 14 08:39:22 crc kubenswrapper[4870]: I1014 08:39:22.519518 4870 generic.go:334] "Generic (PLEG): container finished" podID="64d3a546-2e9d-4c15-bcb0-53bbdf6070b3" containerID="d0df804fea6242a55b2c5b2eab0bbb6e8f48d9c5a84e84bdb0ea37cf2f734564" exitCode=0 Oct 14 08:39:22 crc kubenswrapper[4870]: I1014 08:39:22.519604 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-54bcdcd99f-wq6qt" event={"ID":"64d3a546-2e9d-4c15-bcb0-53bbdf6070b3","Type":"ContainerDied","Data":"d0df804fea6242a55b2c5b2eab0bbb6e8f48d9c5a84e84bdb0ea37cf2f734564"} Oct 14 08:39:22 crc kubenswrapper[4870]: I1014 08:39:22.519677 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-54bcdcd99f-wq6qt" Oct 14 08:39:22 crc kubenswrapper[4870]: I1014 08:39:22.519910 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-54bcdcd99f-wq6qt" event={"ID":"64d3a546-2e9d-4c15-bcb0-53bbdf6070b3","Type":"ContainerDied","Data":"783c8c366bd8f5e01238808715936a1503599cedb5a2ded203425eda63c5808a"} Oct 14 08:39:22 crc kubenswrapper[4870]: I1014 08:39:22.519932 4870 scope.go:117] "RemoveContainer" containerID="d0df804fea6242a55b2c5b2eab0bbb6e8f48d9c5a84e84bdb0ea37cf2f734564" Oct 14 08:39:22 crc kubenswrapper[4870]: I1014 08:39:22.520841 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"e07b718e-6d3b-491e-9229-d30e5228635c","Type":"ContainerStarted","Data":"6e9126cfa6352b7538c827ed0a1eaf4fe2ae6df3b6d35d6b69553a324a6d1af5"} Oct 14 08:39:22 crc kubenswrapper[4870]: I1014 08:39:22.538862 4870 scope.go:117] "RemoveContainer" containerID="6439581ad9335c27c204b5ea46672e6a7ce578a8085a52123db3911f1dc5daaa" Oct 14 08:39:22 crc kubenswrapper[4870]: I1014 08:39:22.557475 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-54bcdcd99f-wq6qt"] Oct 14 08:39:22 crc kubenswrapper[4870]: I1014 08:39:22.568235 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-54bcdcd99f-wq6qt"] Oct 14 08:39:22 crc kubenswrapper[4870]: I1014 08:39:22.575586 4870 scope.go:117] "RemoveContainer" containerID="d0df804fea6242a55b2c5b2eab0bbb6e8f48d9c5a84e84bdb0ea37cf2f734564" Oct 14 08:39:22 crc kubenswrapper[4870]: E1014 08:39:22.575967 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d0df804fea6242a55b2c5b2eab0bbb6e8f48d9c5a84e84bdb0ea37cf2f734564\": container with ID starting with d0df804fea6242a55b2c5b2eab0bbb6e8f48d9c5a84e84bdb0ea37cf2f734564 not found: ID does not exist" containerID="d0df804fea6242a55b2c5b2eab0bbb6e8f48d9c5a84e84bdb0ea37cf2f734564" Oct 14 08:39:22 crc kubenswrapper[4870]: I1014 08:39:22.576002 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d0df804fea6242a55b2c5b2eab0bbb6e8f48d9c5a84e84bdb0ea37cf2f734564"} err="failed to get container status \"d0df804fea6242a55b2c5b2eab0bbb6e8f48d9c5a84e84bdb0ea37cf2f734564\": rpc error: code = NotFound desc = could not find container \"d0df804fea6242a55b2c5b2eab0bbb6e8f48d9c5a84e84bdb0ea37cf2f734564\": container with ID starting with d0df804fea6242a55b2c5b2eab0bbb6e8f48d9c5a84e84bdb0ea37cf2f734564 not found: ID does not exist" Oct 14 08:39:22 crc kubenswrapper[4870]: I1014 08:39:22.576024 4870 scope.go:117] "RemoveContainer" containerID="6439581ad9335c27c204b5ea46672e6a7ce578a8085a52123db3911f1dc5daaa" Oct 14 08:39:22 crc kubenswrapper[4870]: E1014 08:39:22.576340 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6439581ad9335c27c204b5ea46672e6a7ce578a8085a52123db3911f1dc5daaa\": container with ID starting with 6439581ad9335c27c204b5ea46672e6a7ce578a8085a52123db3911f1dc5daaa not found: ID does not exist" containerID="6439581ad9335c27c204b5ea46672e6a7ce578a8085a52123db3911f1dc5daaa" Oct 14 08:39:22 crc kubenswrapper[4870]: I1014 08:39:22.576360 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6439581ad9335c27c204b5ea46672e6a7ce578a8085a52123db3911f1dc5daaa"} err="failed to get container status \"6439581ad9335c27c204b5ea46672e6a7ce578a8085a52123db3911f1dc5daaa\": rpc error: code = NotFound desc = could not find container \"6439581ad9335c27c204b5ea46672e6a7ce578a8085a52123db3911f1dc5daaa\": container with ID starting with 6439581ad9335c27c204b5ea46672e6a7ce578a8085a52123db3911f1dc5daaa not found: ID does not exist" Oct 14 08:39:23 crc kubenswrapper[4870]: I1014 08:39:23.054279 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="64d3a546-2e9d-4c15-bcb0-53bbdf6070b3" path="/var/lib/kubelet/pods/64d3a546-2e9d-4c15-bcb0-53bbdf6070b3/volumes" Oct 14 08:39:23 crc kubenswrapper[4870]: I1014 08:39:23.056323 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="98a7db06-05d2-4c54-9321-11ad232caeda" path="/var/lib/kubelet/pods/98a7db06-05d2-4c54-9321-11ad232caeda/volumes" Oct 14 08:39:23 crc kubenswrapper[4870]: I1014 08:39:23.536765 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"a40d8a1a-605f-4db1-bf5f-4e4785efaa4e","Type":"ContainerStarted","Data":"3225f7ca8393f0da6e4316fca5f6d6da69c21b441152cec133f8e32a01eb1e72"} Oct 14 08:39:24 crc kubenswrapper[4870]: I1014 08:39:24.549868 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"e07b718e-6d3b-491e-9229-d30e5228635c","Type":"ContainerStarted","Data":"e7202ee1265f1c08afda1759a5a0b146c605132e054a093841d69ecd13dbfe11"} Oct 14 08:39:56 crc kubenswrapper[4870]: I1014 08:39:56.931864 4870 generic.go:334] "Generic (PLEG): container finished" podID="a40d8a1a-605f-4db1-bf5f-4e4785efaa4e" containerID="3225f7ca8393f0da6e4316fca5f6d6da69c21b441152cec133f8e32a01eb1e72" exitCode=0 Oct 14 08:39:56 crc kubenswrapper[4870]: I1014 08:39:56.932066 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"a40d8a1a-605f-4db1-bf5f-4e4785efaa4e","Type":"ContainerDied","Data":"3225f7ca8393f0da6e4316fca5f6d6da69c21b441152cec133f8e32a01eb1e72"} Oct 14 08:39:57 crc kubenswrapper[4870]: I1014 08:39:57.947450 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"a40d8a1a-605f-4db1-bf5f-4e4785efaa4e","Type":"ContainerStarted","Data":"a60a40f601c43fb58262baecb024c33aebde054a4bb39c369a34292e2a26a428"} Oct 14 08:39:57 crc kubenswrapper[4870]: I1014 08:39:57.947960 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Oct 14 08:39:57 crc kubenswrapper[4870]: I1014 08:39:57.949585 4870 generic.go:334] "Generic (PLEG): container finished" podID="e07b718e-6d3b-491e-9229-d30e5228635c" containerID="e7202ee1265f1c08afda1759a5a0b146c605132e054a093841d69ecd13dbfe11" exitCode=0 Oct 14 08:39:57 crc kubenswrapper[4870]: I1014 08:39:57.949617 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"e07b718e-6d3b-491e-9229-d30e5228635c","Type":"ContainerDied","Data":"e7202ee1265f1c08afda1759a5a0b146c605132e054a093841d69ecd13dbfe11"} Oct 14 08:39:57 crc kubenswrapper[4870]: I1014 08:39:57.994839 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=37.994822221 podStartE2EDuration="37.994822221s" podCreationTimestamp="2025-10-14 08:39:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 08:39:57.990466215 +0000 UTC m=+5933.687826586" watchObservedRunningTime="2025-10-14 08:39:57.994822221 +0000 UTC m=+5933.692182592" Oct 14 08:39:58 crc kubenswrapper[4870]: I1014 08:39:58.959117 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"e07b718e-6d3b-491e-9229-d30e5228635c","Type":"ContainerStarted","Data":"3a9a1dc34f854662d55af3217d6c649bb4f7608d66f308ef9c5c651088f17285"} Oct 14 08:39:58 crc kubenswrapper[4870]: I1014 08:39:58.959848 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:39:58 crc kubenswrapper[4870]: I1014 08:39:58.980797 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=37.980778621 podStartE2EDuration="37.980778621s" podCreationTimestamp="2025-10-14 08:39:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 08:39:58.980049204 +0000 UTC m=+5934.677409595" watchObservedRunningTime="2025-10-14 08:39:58.980778621 +0000 UTC m=+5934.678138992" Oct 14 08:40:04 crc kubenswrapper[4870]: I1014 08:40:04.074732 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-ppxml"] Oct 14 08:40:04 crc kubenswrapper[4870]: E1014 08:40:04.076191 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="64d3a546-2e9d-4c15-bcb0-53bbdf6070b3" containerName="init" Oct 14 08:40:04 crc kubenswrapper[4870]: I1014 08:40:04.076216 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="64d3a546-2e9d-4c15-bcb0-53bbdf6070b3" containerName="init" Oct 14 08:40:04 crc kubenswrapper[4870]: E1014 08:40:04.076283 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="64d3a546-2e9d-4c15-bcb0-53bbdf6070b3" containerName="dnsmasq-dns" Oct 14 08:40:04 crc kubenswrapper[4870]: I1014 08:40:04.076297 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="64d3a546-2e9d-4c15-bcb0-53bbdf6070b3" containerName="dnsmasq-dns" Oct 14 08:40:04 crc kubenswrapper[4870]: I1014 08:40:04.076592 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="64d3a546-2e9d-4c15-bcb0-53bbdf6070b3" containerName="dnsmasq-dns" Oct 14 08:40:04 crc kubenswrapper[4870]: I1014 08:40:04.078373 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ppxml" Oct 14 08:40:04 crc kubenswrapper[4870]: I1014 08:40:04.096406 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-ppxml"] Oct 14 08:40:04 crc kubenswrapper[4870]: I1014 08:40:04.211882 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b2386444-0cbd-489d-bedb-14e8c972906b-utilities\") pod \"redhat-operators-ppxml\" (UID: \"b2386444-0cbd-489d-bedb-14e8c972906b\") " pod="openshift-marketplace/redhat-operators-ppxml" Oct 14 08:40:04 crc kubenswrapper[4870]: I1014 08:40:04.211954 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b2386444-0cbd-489d-bedb-14e8c972906b-catalog-content\") pod \"redhat-operators-ppxml\" (UID: \"b2386444-0cbd-489d-bedb-14e8c972906b\") " pod="openshift-marketplace/redhat-operators-ppxml" Oct 14 08:40:04 crc kubenswrapper[4870]: I1014 08:40:04.211976 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wkqlt\" (UniqueName: \"kubernetes.io/projected/b2386444-0cbd-489d-bedb-14e8c972906b-kube-api-access-wkqlt\") pod \"redhat-operators-ppxml\" (UID: \"b2386444-0cbd-489d-bedb-14e8c972906b\") " pod="openshift-marketplace/redhat-operators-ppxml" Oct 14 08:40:04 crc kubenswrapper[4870]: I1014 08:40:04.313078 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b2386444-0cbd-489d-bedb-14e8c972906b-utilities\") pod \"redhat-operators-ppxml\" (UID: \"b2386444-0cbd-489d-bedb-14e8c972906b\") " pod="openshift-marketplace/redhat-operators-ppxml" Oct 14 08:40:04 crc kubenswrapper[4870]: I1014 08:40:04.313153 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b2386444-0cbd-489d-bedb-14e8c972906b-catalog-content\") pod \"redhat-operators-ppxml\" (UID: \"b2386444-0cbd-489d-bedb-14e8c972906b\") " pod="openshift-marketplace/redhat-operators-ppxml" Oct 14 08:40:04 crc kubenswrapper[4870]: I1014 08:40:04.313177 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wkqlt\" (UniqueName: \"kubernetes.io/projected/b2386444-0cbd-489d-bedb-14e8c972906b-kube-api-access-wkqlt\") pod \"redhat-operators-ppxml\" (UID: \"b2386444-0cbd-489d-bedb-14e8c972906b\") " pod="openshift-marketplace/redhat-operators-ppxml" Oct 14 08:40:04 crc kubenswrapper[4870]: I1014 08:40:04.313810 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b2386444-0cbd-489d-bedb-14e8c972906b-catalog-content\") pod \"redhat-operators-ppxml\" (UID: \"b2386444-0cbd-489d-bedb-14e8c972906b\") " pod="openshift-marketplace/redhat-operators-ppxml" Oct 14 08:40:04 crc kubenswrapper[4870]: I1014 08:40:04.313822 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b2386444-0cbd-489d-bedb-14e8c972906b-utilities\") pod \"redhat-operators-ppxml\" (UID: \"b2386444-0cbd-489d-bedb-14e8c972906b\") " pod="openshift-marketplace/redhat-operators-ppxml" Oct 14 08:40:04 crc kubenswrapper[4870]: I1014 08:40:04.332559 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wkqlt\" (UniqueName: \"kubernetes.io/projected/b2386444-0cbd-489d-bedb-14e8c972906b-kube-api-access-wkqlt\") pod \"redhat-operators-ppxml\" (UID: \"b2386444-0cbd-489d-bedb-14e8c972906b\") " pod="openshift-marketplace/redhat-operators-ppxml" Oct 14 08:40:04 crc kubenswrapper[4870]: I1014 08:40:04.403124 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ppxml" Oct 14 08:40:04 crc kubenswrapper[4870]: I1014 08:40:04.903914 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-ppxml"] Oct 14 08:40:05 crc kubenswrapper[4870]: I1014 08:40:05.021776 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ppxml" event={"ID":"b2386444-0cbd-489d-bedb-14e8c972906b","Type":"ContainerStarted","Data":"b598aaf3b52de990c61de6e8cc8d70f2a8bb83e97a9c1736e94f5a5021d51b1f"} Oct 14 08:40:06 crc kubenswrapper[4870]: I1014 08:40:06.041821 4870 generic.go:334] "Generic (PLEG): container finished" podID="b2386444-0cbd-489d-bedb-14e8c972906b" containerID="1846e694f2ad762ca9bfa1917c34febe09adbc1fd9e920946adec6cd98b7c628" exitCode=0 Oct 14 08:40:06 crc kubenswrapper[4870]: I1014 08:40:06.042120 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ppxml" event={"ID":"b2386444-0cbd-489d-bedb-14e8c972906b","Type":"ContainerDied","Data":"1846e694f2ad762ca9bfa1917c34febe09adbc1fd9e920946adec6cd98b7c628"} Oct 14 08:40:07 crc kubenswrapper[4870]: I1014 08:40:07.050023 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ppxml" event={"ID":"b2386444-0cbd-489d-bedb-14e8c972906b","Type":"ContainerStarted","Data":"911ee52d61f819548d70a097d62828a228cf571dea70f2aff132ce0e6a6d264d"} Oct 14 08:40:08 crc kubenswrapper[4870]: I1014 08:40:08.064288 4870 generic.go:334] "Generic (PLEG): container finished" podID="b2386444-0cbd-489d-bedb-14e8c972906b" containerID="911ee52d61f819548d70a097d62828a228cf571dea70f2aff132ce0e6a6d264d" exitCode=0 Oct 14 08:40:08 crc kubenswrapper[4870]: I1014 08:40:08.064361 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ppxml" event={"ID":"b2386444-0cbd-489d-bedb-14e8c972906b","Type":"ContainerDied","Data":"911ee52d61f819548d70a097d62828a228cf571dea70f2aff132ce0e6a6d264d"} Oct 14 08:40:09 crc kubenswrapper[4870]: I1014 08:40:09.091241 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ppxml" event={"ID":"b2386444-0cbd-489d-bedb-14e8c972906b","Type":"ContainerStarted","Data":"3646038fecc8ace40445e2ea3bec365af7226d660f7d5b6215bcaf28cc80a17d"} Oct 14 08:40:09 crc kubenswrapper[4870]: I1014 08:40:09.124216 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-ppxml" podStartSLOduration=2.33718334 podStartE2EDuration="5.124195093s" podCreationTimestamp="2025-10-14 08:40:04 +0000 UTC" firstStartedPulling="2025-10-14 08:40:06.043805846 +0000 UTC m=+5941.741166217" lastFinishedPulling="2025-10-14 08:40:08.830817559 +0000 UTC m=+5944.528177970" observedRunningTime="2025-10-14 08:40:09.12364152 +0000 UTC m=+5944.821001931" watchObservedRunningTime="2025-10-14 08:40:09.124195093 +0000 UTC m=+5944.821555474" Oct 14 08:40:10 crc kubenswrapper[4870]: I1014 08:40:10.933809 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Oct 14 08:40:11 crc kubenswrapper[4870]: I1014 08:40:11.441829 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-7m2jn"] Oct 14 08:40:11 crc kubenswrapper[4870]: I1014 08:40:11.444574 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7m2jn" Oct 14 08:40:11 crc kubenswrapper[4870]: I1014 08:40:11.456817 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-7m2jn"] Oct 14 08:40:11 crc kubenswrapper[4870]: I1014 08:40:11.529493 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f7f758c5-b38e-4a88-b96e-3cc43bfac9d5-utilities\") pod \"certified-operators-7m2jn\" (UID: \"f7f758c5-b38e-4a88-b96e-3cc43bfac9d5\") " pod="openshift-marketplace/certified-operators-7m2jn" Oct 14 08:40:11 crc kubenswrapper[4870]: I1014 08:40:11.529590 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2whlx\" (UniqueName: \"kubernetes.io/projected/f7f758c5-b38e-4a88-b96e-3cc43bfac9d5-kube-api-access-2whlx\") pod \"certified-operators-7m2jn\" (UID: \"f7f758c5-b38e-4a88-b96e-3cc43bfac9d5\") " pod="openshift-marketplace/certified-operators-7m2jn" Oct 14 08:40:11 crc kubenswrapper[4870]: I1014 08:40:11.529681 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f7f758c5-b38e-4a88-b96e-3cc43bfac9d5-catalog-content\") pod \"certified-operators-7m2jn\" (UID: \"f7f758c5-b38e-4a88-b96e-3cc43bfac9d5\") " pod="openshift-marketplace/certified-operators-7m2jn" Oct 14 08:40:11 crc kubenswrapper[4870]: I1014 08:40:11.631329 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2whlx\" (UniqueName: \"kubernetes.io/projected/f7f758c5-b38e-4a88-b96e-3cc43bfac9d5-kube-api-access-2whlx\") pod \"certified-operators-7m2jn\" (UID: \"f7f758c5-b38e-4a88-b96e-3cc43bfac9d5\") " pod="openshift-marketplace/certified-operators-7m2jn" Oct 14 08:40:11 crc kubenswrapper[4870]: I1014 08:40:11.631536 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f7f758c5-b38e-4a88-b96e-3cc43bfac9d5-catalog-content\") pod \"certified-operators-7m2jn\" (UID: \"f7f758c5-b38e-4a88-b96e-3cc43bfac9d5\") " pod="openshift-marketplace/certified-operators-7m2jn" Oct 14 08:40:11 crc kubenswrapper[4870]: I1014 08:40:11.631611 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f7f758c5-b38e-4a88-b96e-3cc43bfac9d5-utilities\") pod \"certified-operators-7m2jn\" (UID: \"f7f758c5-b38e-4a88-b96e-3cc43bfac9d5\") " pod="openshift-marketplace/certified-operators-7m2jn" Oct 14 08:40:11 crc kubenswrapper[4870]: I1014 08:40:11.632028 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f7f758c5-b38e-4a88-b96e-3cc43bfac9d5-catalog-content\") pod \"certified-operators-7m2jn\" (UID: \"f7f758c5-b38e-4a88-b96e-3cc43bfac9d5\") " pod="openshift-marketplace/certified-operators-7m2jn" Oct 14 08:40:11 crc kubenswrapper[4870]: I1014 08:40:11.632148 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f7f758c5-b38e-4a88-b96e-3cc43bfac9d5-utilities\") pod \"certified-operators-7m2jn\" (UID: \"f7f758c5-b38e-4a88-b96e-3cc43bfac9d5\") " pod="openshift-marketplace/certified-operators-7m2jn" Oct 14 08:40:11 crc kubenswrapper[4870]: I1014 08:40:11.664080 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2whlx\" (UniqueName: \"kubernetes.io/projected/f7f758c5-b38e-4a88-b96e-3cc43bfac9d5-kube-api-access-2whlx\") pod \"certified-operators-7m2jn\" (UID: \"f7f758c5-b38e-4a88-b96e-3cc43bfac9d5\") " pod="openshift-marketplace/certified-operators-7m2jn" Oct 14 08:40:11 crc kubenswrapper[4870]: I1014 08:40:11.767488 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7m2jn" Oct 14 08:40:11 crc kubenswrapper[4870]: I1014 08:40:11.933738 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Oct 14 08:40:12 crc kubenswrapper[4870]: I1014 08:40:12.291988 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-7m2jn"] Oct 14 08:40:12 crc kubenswrapper[4870]: W1014 08:40:12.293551 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf7f758c5_b38e_4a88_b96e_3cc43bfac9d5.slice/crio-121cf51348d580fdfd4d99f61bda6cc167f8379d8a9c08610f04f582d0e1c1f2 WatchSource:0}: Error finding container 121cf51348d580fdfd4d99f61bda6cc167f8379d8a9c08610f04f582d0e1c1f2: Status 404 returned error can't find the container with id 121cf51348d580fdfd4d99f61bda6cc167f8379d8a9c08610f04f582d0e1c1f2 Oct 14 08:40:13 crc kubenswrapper[4870]: I1014 08:40:13.126469 4870 generic.go:334] "Generic (PLEG): container finished" podID="f7f758c5-b38e-4a88-b96e-3cc43bfac9d5" containerID="996723e2f597f544da216d4d3d780ecfc892ca42627f3bc5f2b12c3d9e6e5c37" exitCode=0 Oct 14 08:40:13 crc kubenswrapper[4870]: I1014 08:40:13.126565 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7m2jn" event={"ID":"f7f758c5-b38e-4a88-b96e-3cc43bfac9d5","Type":"ContainerDied","Data":"996723e2f597f544da216d4d3d780ecfc892ca42627f3bc5f2b12c3d9e6e5c37"} Oct 14 08:40:13 crc kubenswrapper[4870]: I1014 08:40:13.126875 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7m2jn" event={"ID":"f7f758c5-b38e-4a88-b96e-3cc43bfac9d5","Type":"ContainerStarted","Data":"121cf51348d580fdfd4d99f61bda6cc167f8379d8a9c08610f04f582d0e1c1f2"} Oct 14 08:40:14 crc kubenswrapper[4870]: I1014 08:40:14.137493 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7m2jn" event={"ID":"f7f758c5-b38e-4a88-b96e-3cc43bfac9d5","Type":"ContainerStarted","Data":"c55801a00ec720b64e71069dd9f28e098d5b04396f71fff0edd2000fd479d621"} Oct 14 08:40:14 crc kubenswrapper[4870]: I1014 08:40:14.403502 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-ppxml" Oct 14 08:40:14 crc kubenswrapper[4870]: I1014 08:40:14.403581 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-ppxml" Oct 14 08:40:14 crc kubenswrapper[4870]: I1014 08:40:14.451591 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-ppxml" Oct 14 08:40:15 crc kubenswrapper[4870]: I1014 08:40:15.151708 4870 generic.go:334] "Generic (PLEG): container finished" podID="f7f758c5-b38e-4a88-b96e-3cc43bfac9d5" containerID="c55801a00ec720b64e71069dd9f28e098d5b04396f71fff0edd2000fd479d621" exitCode=0 Oct 14 08:40:15 crc kubenswrapper[4870]: I1014 08:40:15.151799 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7m2jn" event={"ID":"f7f758c5-b38e-4a88-b96e-3cc43bfac9d5","Type":"ContainerDied","Data":"c55801a00ec720b64e71069dd9f28e098d5b04396f71fff0edd2000fd479d621"} Oct 14 08:40:15 crc kubenswrapper[4870]: I1014 08:40:15.214048 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-ppxml" Oct 14 08:40:16 crc kubenswrapper[4870]: I1014 08:40:16.166745 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7m2jn" event={"ID":"f7f758c5-b38e-4a88-b96e-3cc43bfac9d5","Type":"ContainerStarted","Data":"a2b45ea68eb853f92416531b76e36b85734f43bf69c1bff8aac1f4506060cd59"} Oct 14 08:40:16 crc kubenswrapper[4870]: I1014 08:40:16.195948 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-7m2jn" podStartSLOduration=2.6563456629999997 podStartE2EDuration="5.19592808s" podCreationTimestamp="2025-10-14 08:40:11 +0000 UTC" firstStartedPulling="2025-10-14 08:40:13.129322469 +0000 UTC m=+5948.826682880" lastFinishedPulling="2025-10-14 08:40:15.668904896 +0000 UTC m=+5951.366265297" observedRunningTime="2025-10-14 08:40:16.195041699 +0000 UTC m=+5951.892402090" watchObservedRunningTime="2025-10-14 08:40:16.19592808 +0000 UTC m=+5951.893288461" Oct 14 08:40:16 crc kubenswrapper[4870]: I1014 08:40:16.834944 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-ppxml"] Oct 14 08:40:17 crc kubenswrapper[4870]: I1014 08:40:17.175591 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-ppxml" podUID="b2386444-0cbd-489d-bedb-14e8c972906b" containerName="registry-server" containerID="cri-o://3646038fecc8ace40445e2ea3bec365af7226d660f7d5b6215bcaf28cc80a17d" gracePeriod=2 Oct 14 08:40:17 crc kubenswrapper[4870]: I1014 08:40:17.684075 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ppxml" Oct 14 08:40:17 crc kubenswrapper[4870]: I1014 08:40:17.846414 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b2386444-0cbd-489d-bedb-14e8c972906b-catalog-content\") pod \"b2386444-0cbd-489d-bedb-14e8c972906b\" (UID: \"b2386444-0cbd-489d-bedb-14e8c972906b\") " Oct 14 08:40:17 crc kubenswrapper[4870]: I1014 08:40:17.846581 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wkqlt\" (UniqueName: \"kubernetes.io/projected/b2386444-0cbd-489d-bedb-14e8c972906b-kube-api-access-wkqlt\") pod \"b2386444-0cbd-489d-bedb-14e8c972906b\" (UID: \"b2386444-0cbd-489d-bedb-14e8c972906b\") " Oct 14 08:40:17 crc kubenswrapper[4870]: I1014 08:40:17.846665 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b2386444-0cbd-489d-bedb-14e8c972906b-utilities\") pod \"b2386444-0cbd-489d-bedb-14e8c972906b\" (UID: \"b2386444-0cbd-489d-bedb-14e8c972906b\") " Oct 14 08:40:17 crc kubenswrapper[4870]: I1014 08:40:17.849242 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b2386444-0cbd-489d-bedb-14e8c972906b-utilities" (OuterVolumeSpecName: "utilities") pod "b2386444-0cbd-489d-bedb-14e8c972906b" (UID: "b2386444-0cbd-489d-bedb-14e8c972906b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:40:17 crc kubenswrapper[4870]: I1014 08:40:17.863921 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b2386444-0cbd-489d-bedb-14e8c972906b-kube-api-access-wkqlt" (OuterVolumeSpecName: "kube-api-access-wkqlt") pod "b2386444-0cbd-489d-bedb-14e8c972906b" (UID: "b2386444-0cbd-489d-bedb-14e8c972906b"). InnerVolumeSpecName "kube-api-access-wkqlt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:40:17 crc kubenswrapper[4870]: I1014 08:40:17.948853 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wkqlt\" (UniqueName: \"kubernetes.io/projected/b2386444-0cbd-489d-bedb-14e8c972906b-kube-api-access-wkqlt\") on node \"crc\" DevicePath \"\"" Oct 14 08:40:17 crc kubenswrapper[4870]: I1014 08:40:17.949304 4870 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b2386444-0cbd-489d-bedb-14e8c972906b-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 08:40:18 crc kubenswrapper[4870]: I1014 08:40:18.197331 4870 generic.go:334] "Generic (PLEG): container finished" podID="b2386444-0cbd-489d-bedb-14e8c972906b" containerID="3646038fecc8ace40445e2ea3bec365af7226d660f7d5b6215bcaf28cc80a17d" exitCode=0 Oct 14 08:40:18 crc kubenswrapper[4870]: I1014 08:40:18.197407 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ppxml" event={"ID":"b2386444-0cbd-489d-bedb-14e8c972906b","Type":"ContainerDied","Data":"3646038fecc8ace40445e2ea3bec365af7226d660f7d5b6215bcaf28cc80a17d"} Oct 14 08:40:18 crc kubenswrapper[4870]: I1014 08:40:18.197476 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ppxml" event={"ID":"b2386444-0cbd-489d-bedb-14e8c972906b","Type":"ContainerDied","Data":"b598aaf3b52de990c61de6e8cc8d70f2a8bb83e97a9c1736e94f5a5021d51b1f"} Oct 14 08:40:18 crc kubenswrapper[4870]: I1014 08:40:18.197511 4870 scope.go:117] "RemoveContainer" containerID="3646038fecc8ace40445e2ea3bec365af7226d660f7d5b6215bcaf28cc80a17d" Oct 14 08:40:18 crc kubenswrapper[4870]: I1014 08:40:18.197735 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ppxml" Oct 14 08:40:18 crc kubenswrapper[4870]: I1014 08:40:18.248837 4870 scope.go:117] "RemoveContainer" containerID="911ee52d61f819548d70a097d62828a228cf571dea70f2aff132ce0e6a6d264d" Oct 14 08:40:18 crc kubenswrapper[4870]: I1014 08:40:18.288311 4870 scope.go:117] "RemoveContainer" containerID="1846e694f2ad762ca9bfa1917c34febe09adbc1fd9e920946adec6cd98b7c628" Oct 14 08:40:18 crc kubenswrapper[4870]: I1014 08:40:18.313456 4870 scope.go:117] "RemoveContainer" containerID="3646038fecc8ace40445e2ea3bec365af7226d660f7d5b6215bcaf28cc80a17d" Oct 14 08:40:18 crc kubenswrapper[4870]: E1014 08:40:18.313729 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3646038fecc8ace40445e2ea3bec365af7226d660f7d5b6215bcaf28cc80a17d\": container with ID starting with 3646038fecc8ace40445e2ea3bec365af7226d660f7d5b6215bcaf28cc80a17d not found: ID does not exist" containerID="3646038fecc8ace40445e2ea3bec365af7226d660f7d5b6215bcaf28cc80a17d" Oct 14 08:40:18 crc kubenswrapper[4870]: I1014 08:40:18.313760 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3646038fecc8ace40445e2ea3bec365af7226d660f7d5b6215bcaf28cc80a17d"} err="failed to get container status \"3646038fecc8ace40445e2ea3bec365af7226d660f7d5b6215bcaf28cc80a17d\": rpc error: code = NotFound desc = could not find container \"3646038fecc8ace40445e2ea3bec365af7226d660f7d5b6215bcaf28cc80a17d\": container with ID starting with 3646038fecc8ace40445e2ea3bec365af7226d660f7d5b6215bcaf28cc80a17d not found: ID does not exist" Oct 14 08:40:18 crc kubenswrapper[4870]: I1014 08:40:18.313781 4870 scope.go:117] "RemoveContainer" containerID="911ee52d61f819548d70a097d62828a228cf571dea70f2aff132ce0e6a6d264d" Oct 14 08:40:18 crc kubenswrapper[4870]: E1014 08:40:18.314135 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"911ee52d61f819548d70a097d62828a228cf571dea70f2aff132ce0e6a6d264d\": container with ID starting with 911ee52d61f819548d70a097d62828a228cf571dea70f2aff132ce0e6a6d264d not found: ID does not exist" containerID="911ee52d61f819548d70a097d62828a228cf571dea70f2aff132ce0e6a6d264d" Oct 14 08:40:18 crc kubenswrapper[4870]: I1014 08:40:18.314158 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"911ee52d61f819548d70a097d62828a228cf571dea70f2aff132ce0e6a6d264d"} err="failed to get container status \"911ee52d61f819548d70a097d62828a228cf571dea70f2aff132ce0e6a6d264d\": rpc error: code = NotFound desc = could not find container \"911ee52d61f819548d70a097d62828a228cf571dea70f2aff132ce0e6a6d264d\": container with ID starting with 911ee52d61f819548d70a097d62828a228cf571dea70f2aff132ce0e6a6d264d not found: ID does not exist" Oct 14 08:40:18 crc kubenswrapper[4870]: I1014 08:40:18.314172 4870 scope.go:117] "RemoveContainer" containerID="1846e694f2ad762ca9bfa1917c34febe09adbc1fd9e920946adec6cd98b7c628" Oct 14 08:40:18 crc kubenswrapper[4870]: E1014 08:40:18.314397 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1846e694f2ad762ca9bfa1917c34febe09adbc1fd9e920946adec6cd98b7c628\": container with ID starting with 1846e694f2ad762ca9bfa1917c34febe09adbc1fd9e920946adec6cd98b7c628 not found: ID does not exist" containerID="1846e694f2ad762ca9bfa1917c34febe09adbc1fd9e920946adec6cd98b7c628" Oct 14 08:40:18 crc kubenswrapper[4870]: I1014 08:40:18.314420 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1846e694f2ad762ca9bfa1917c34febe09adbc1fd9e920946adec6cd98b7c628"} err="failed to get container status \"1846e694f2ad762ca9bfa1917c34febe09adbc1fd9e920946adec6cd98b7c628\": rpc error: code = NotFound desc = could not find container \"1846e694f2ad762ca9bfa1917c34febe09adbc1fd9e920946adec6cd98b7c628\": container with ID starting with 1846e694f2ad762ca9bfa1917c34febe09adbc1fd9e920946adec6cd98b7c628 not found: ID does not exist" Oct 14 08:40:18 crc kubenswrapper[4870]: I1014 08:40:18.535234 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-1-default"] Oct 14 08:40:18 crc kubenswrapper[4870]: E1014 08:40:18.535603 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b2386444-0cbd-489d-bedb-14e8c972906b" containerName="registry-server" Oct 14 08:40:18 crc kubenswrapper[4870]: I1014 08:40:18.535618 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="b2386444-0cbd-489d-bedb-14e8c972906b" containerName="registry-server" Oct 14 08:40:18 crc kubenswrapper[4870]: E1014 08:40:18.535648 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b2386444-0cbd-489d-bedb-14e8c972906b" containerName="extract-content" Oct 14 08:40:18 crc kubenswrapper[4870]: I1014 08:40:18.535657 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="b2386444-0cbd-489d-bedb-14e8c972906b" containerName="extract-content" Oct 14 08:40:18 crc kubenswrapper[4870]: E1014 08:40:18.535678 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b2386444-0cbd-489d-bedb-14e8c972906b" containerName="extract-utilities" Oct 14 08:40:18 crc kubenswrapper[4870]: I1014 08:40:18.535687 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="b2386444-0cbd-489d-bedb-14e8c972906b" containerName="extract-utilities" Oct 14 08:40:18 crc kubenswrapper[4870]: I1014 08:40:18.535891 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="b2386444-0cbd-489d-bedb-14e8c972906b" containerName="registry-server" Oct 14 08:40:18 crc kubenswrapper[4870]: I1014 08:40:18.536489 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1-default" Oct 14 08:40:18 crc kubenswrapper[4870]: I1014 08:40:18.544115 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-vnjjw" Oct 14 08:40:18 crc kubenswrapper[4870]: I1014 08:40:18.544672 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-1-default"] Oct 14 08:40:18 crc kubenswrapper[4870]: I1014 08:40:18.660094 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-27zx2\" (UniqueName: \"kubernetes.io/projected/fddf65c1-094d-415d-9d2a-aa7e223c2240-kube-api-access-27zx2\") pod \"mariadb-client-1-default\" (UID: \"fddf65c1-094d-415d-9d2a-aa7e223c2240\") " pod="openstack/mariadb-client-1-default" Oct 14 08:40:18 crc kubenswrapper[4870]: I1014 08:40:18.761668 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-27zx2\" (UniqueName: \"kubernetes.io/projected/fddf65c1-094d-415d-9d2a-aa7e223c2240-kube-api-access-27zx2\") pod \"mariadb-client-1-default\" (UID: \"fddf65c1-094d-415d-9d2a-aa7e223c2240\") " pod="openstack/mariadb-client-1-default" Oct 14 08:40:18 crc kubenswrapper[4870]: I1014 08:40:18.785739 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-27zx2\" (UniqueName: \"kubernetes.io/projected/fddf65c1-094d-415d-9d2a-aa7e223c2240-kube-api-access-27zx2\") pod \"mariadb-client-1-default\" (UID: \"fddf65c1-094d-415d-9d2a-aa7e223c2240\") " pod="openstack/mariadb-client-1-default" Oct 14 08:40:18 crc kubenswrapper[4870]: I1014 08:40:18.789853 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b2386444-0cbd-489d-bedb-14e8c972906b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b2386444-0cbd-489d-bedb-14e8c972906b" (UID: "b2386444-0cbd-489d-bedb-14e8c972906b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:40:18 crc kubenswrapper[4870]: I1014 08:40:18.849388 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-ppxml"] Oct 14 08:40:18 crc kubenswrapper[4870]: I1014 08:40:18.855699 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-ppxml"] Oct 14 08:40:18 crc kubenswrapper[4870]: I1014 08:40:18.855968 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1-default" Oct 14 08:40:18 crc kubenswrapper[4870]: I1014 08:40:18.863873 4870 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b2386444-0cbd-489d-bedb-14e8c972906b-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 08:40:19 crc kubenswrapper[4870]: I1014 08:40:19.043310 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b2386444-0cbd-489d-bedb-14e8c972906b" path="/var/lib/kubelet/pods/b2386444-0cbd-489d-bedb-14e8c972906b/volumes" Oct 14 08:40:19 crc kubenswrapper[4870]: I1014 08:40:19.387972 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-1-default"] Oct 14 08:40:19 crc kubenswrapper[4870]: W1014 08:40:19.401185 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfddf65c1_094d_415d_9d2a_aa7e223c2240.slice/crio-de71a87d07c697c7a0bc41b5b738fe99f673a0f0bc01ba92d8f1e25918aad292 WatchSource:0}: Error finding container de71a87d07c697c7a0bc41b5b738fe99f673a0f0bc01ba92d8f1e25918aad292: Status 404 returned error can't find the container with id de71a87d07c697c7a0bc41b5b738fe99f673a0f0bc01ba92d8f1e25918aad292 Oct 14 08:40:20 crc kubenswrapper[4870]: I1014 08:40:20.216090 4870 generic.go:334] "Generic (PLEG): container finished" podID="fddf65c1-094d-415d-9d2a-aa7e223c2240" containerID="136341b92ea97d5b93fbd06bb520655f22b5916e2ed329bb76c6b35b06d087e5" exitCode=0 Oct 14 08:40:20 crc kubenswrapper[4870]: I1014 08:40:20.216246 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-1-default" event={"ID":"fddf65c1-094d-415d-9d2a-aa7e223c2240","Type":"ContainerDied","Data":"136341b92ea97d5b93fbd06bb520655f22b5916e2ed329bb76c6b35b06d087e5"} Oct 14 08:40:20 crc kubenswrapper[4870]: I1014 08:40:20.216450 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-1-default" event={"ID":"fddf65c1-094d-415d-9d2a-aa7e223c2240","Type":"ContainerStarted","Data":"de71a87d07c697c7a0bc41b5b738fe99f673a0f0bc01ba92d8f1e25918aad292"} Oct 14 08:40:21 crc kubenswrapper[4870]: I1014 08:40:21.740137 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1-default" Oct 14 08:40:21 crc kubenswrapper[4870]: I1014 08:40:21.768544 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-7m2jn" Oct 14 08:40:21 crc kubenswrapper[4870]: I1014 08:40:21.768608 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-7m2jn" Oct 14 08:40:21 crc kubenswrapper[4870]: I1014 08:40:21.771730 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-1-default_fddf65c1-094d-415d-9d2a-aa7e223c2240/mariadb-client-1-default/0.log" Oct 14 08:40:21 crc kubenswrapper[4870]: I1014 08:40:21.797682 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-1-default"] Oct 14 08:40:21 crc kubenswrapper[4870]: I1014 08:40:21.804399 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-1-default"] Oct 14 08:40:21 crc kubenswrapper[4870]: I1014 08:40:21.818065 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-27zx2\" (UniqueName: \"kubernetes.io/projected/fddf65c1-094d-415d-9d2a-aa7e223c2240-kube-api-access-27zx2\") pod \"fddf65c1-094d-415d-9d2a-aa7e223c2240\" (UID: \"fddf65c1-094d-415d-9d2a-aa7e223c2240\") " Oct 14 08:40:21 crc kubenswrapper[4870]: I1014 08:40:21.826649 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fddf65c1-094d-415d-9d2a-aa7e223c2240-kube-api-access-27zx2" (OuterVolumeSpecName: "kube-api-access-27zx2") pod "fddf65c1-094d-415d-9d2a-aa7e223c2240" (UID: "fddf65c1-094d-415d-9d2a-aa7e223c2240"). InnerVolumeSpecName "kube-api-access-27zx2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:40:21 crc kubenswrapper[4870]: I1014 08:40:21.887832 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-7m2jn" Oct 14 08:40:21 crc kubenswrapper[4870]: I1014 08:40:21.920291 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-27zx2\" (UniqueName: \"kubernetes.io/projected/fddf65c1-094d-415d-9d2a-aa7e223c2240-kube-api-access-27zx2\") on node \"crc\" DevicePath \"\"" Oct 14 08:40:22 crc kubenswrapper[4870]: I1014 08:40:22.244527 4870 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="de71a87d07c697c7a0bc41b5b738fe99f673a0f0bc01ba92d8f1e25918aad292" Oct 14 08:40:22 crc kubenswrapper[4870]: I1014 08:40:22.245071 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1-default" Oct 14 08:40:22 crc kubenswrapper[4870]: I1014 08:40:22.252754 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-2-default"] Oct 14 08:40:22 crc kubenswrapper[4870]: E1014 08:40:22.253469 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fddf65c1-094d-415d-9d2a-aa7e223c2240" containerName="mariadb-client-1-default" Oct 14 08:40:22 crc kubenswrapper[4870]: I1014 08:40:22.253493 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="fddf65c1-094d-415d-9d2a-aa7e223c2240" containerName="mariadb-client-1-default" Oct 14 08:40:22 crc kubenswrapper[4870]: I1014 08:40:22.253856 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="fddf65c1-094d-415d-9d2a-aa7e223c2240" containerName="mariadb-client-1-default" Oct 14 08:40:22 crc kubenswrapper[4870]: I1014 08:40:22.256808 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2-default" Oct 14 08:40:22 crc kubenswrapper[4870]: I1014 08:40:22.262006 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-vnjjw" Oct 14 08:40:22 crc kubenswrapper[4870]: I1014 08:40:22.264756 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-2-default"] Oct 14 08:40:22 crc kubenswrapper[4870]: I1014 08:40:22.326036 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-7m2jn" Oct 14 08:40:22 crc kubenswrapper[4870]: I1014 08:40:22.427503 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l7tkp\" (UniqueName: \"kubernetes.io/projected/95f78882-9f4d-4cc8-a77a-ed8aaccef072-kube-api-access-l7tkp\") pod \"mariadb-client-2-default\" (UID: \"95f78882-9f4d-4cc8-a77a-ed8aaccef072\") " pod="openstack/mariadb-client-2-default" Oct 14 08:40:22 crc kubenswrapper[4870]: I1014 08:40:22.529207 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l7tkp\" (UniqueName: \"kubernetes.io/projected/95f78882-9f4d-4cc8-a77a-ed8aaccef072-kube-api-access-l7tkp\") pod \"mariadb-client-2-default\" (UID: \"95f78882-9f4d-4cc8-a77a-ed8aaccef072\") " pod="openstack/mariadb-client-2-default" Oct 14 08:40:22 crc kubenswrapper[4870]: I1014 08:40:22.568876 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l7tkp\" (UniqueName: \"kubernetes.io/projected/95f78882-9f4d-4cc8-a77a-ed8aaccef072-kube-api-access-l7tkp\") pod \"mariadb-client-2-default\" (UID: \"95f78882-9f4d-4cc8-a77a-ed8aaccef072\") " pod="openstack/mariadb-client-2-default" Oct 14 08:40:22 crc kubenswrapper[4870]: I1014 08:40:22.602318 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2-default" Oct 14 08:40:22 crc kubenswrapper[4870]: I1014 08:40:22.830149 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-7m2jn"] Oct 14 08:40:23 crc kubenswrapper[4870]: I1014 08:40:23.049041 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fddf65c1-094d-415d-9d2a-aa7e223c2240" path="/var/lib/kubelet/pods/fddf65c1-094d-415d-9d2a-aa7e223c2240/volumes" Oct 14 08:40:23 crc kubenswrapper[4870]: I1014 08:40:23.137157 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-2-default"] Oct 14 08:40:23 crc kubenswrapper[4870]: W1014 08:40:23.142754 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod95f78882_9f4d_4cc8_a77a_ed8aaccef072.slice/crio-dc68bebb5ea1271276b900bede5356ac169674f85a8f0fa5bd7a31fc5128577e WatchSource:0}: Error finding container dc68bebb5ea1271276b900bede5356ac169674f85a8f0fa5bd7a31fc5128577e: Status 404 returned error can't find the container with id dc68bebb5ea1271276b900bede5356ac169674f85a8f0fa5bd7a31fc5128577e Oct 14 08:40:23 crc kubenswrapper[4870]: I1014 08:40:23.255499 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-2-default" event={"ID":"95f78882-9f4d-4cc8-a77a-ed8aaccef072","Type":"ContainerStarted","Data":"dc68bebb5ea1271276b900bede5356ac169674f85a8f0fa5bd7a31fc5128577e"} Oct 14 08:40:23 crc kubenswrapper[4870]: I1014 08:40:23.950908 4870 patch_prober.go:28] interesting pod/machine-config-daemon-7tvc7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 08:40:23 crc kubenswrapper[4870]: I1014 08:40:23.951494 4870 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 08:40:24 crc kubenswrapper[4870]: I1014 08:40:24.268417 4870 generic.go:334] "Generic (PLEG): container finished" podID="95f78882-9f4d-4cc8-a77a-ed8aaccef072" containerID="0ee442bfaaddf52cde88d1a75c7635cf71b4a8104139f8cb3708e30b46e264a3" exitCode=0 Oct 14 08:40:24 crc kubenswrapper[4870]: I1014 08:40:24.268644 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-2-default" event={"ID":"95f78882-9f4d-4cc8-a77a-ed8aaccef072","Type":"ContainerDied","Data":"0ee442bfaaddf52cde88d1a75c7635cf71b4a8104139f8cb3708e30b46e264a3"} Oct 14 08:40:24 crc kubenswrapper[4870]: I1014 08:40:24.268804 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-7m2jn" podUID="f7f758c5-b38e-4a88-b96e-3cc43bfac9d5" containerName="registry-server" containerID="cri-o://a2b45ea68eb853f92416531b76e36b85734f43bf69c1bff8aac1f4506060cd59" gracePeriod=2 Oct 14 08:40:24 crc kubenswrapper[4870]: I1014 08:40:24.753720 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7m2jn" Oct 14 08:40:24 crc kubenswrapper[4870]: I1014 08:40:24.866911 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f7f758c5-b38e-4a88-b96e-3cc43bfac9d5-utilities\") pod \"f7f758c5-b38e-4a88-b96e-3cc43bfac9d5\" (UID: \"f7f758c5-b38e-4a88-b96e-3cc43bfac9d5\") " Oct 14 08:40:24 crc kubenswrapper[4870]: I1014 08:40:24.867008 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2whlx\" (UniqueName: \"kubernetes.io/projected/f7f758c5-b38e-4a88-b96e-3cc43bfac9d5-kube-api-access-2whlx\") pod \"f7f758c5-b38e-4a88-b96e-3cc43bfac9d5\" (UID: \"f7f758c5-b38e-4a88-b96e-3cc43bfac9d5\") " Oct 14 08:40:24 crc kubenswrapper[4870]: I1014 08:40:24.867033 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f7f758c5-b38e-4a88-b96e-3cc43bfac9d5-catalog-content\") pod \"f7f758c5-b38e-4a88-b96e-3cc43bfac9d5\" (UID: \"f7f758c5-b38e-4a88-b96e-3cc43bfac9d5\") " Oct 14 08:40:24 crc kubenswrapper[4870]: I1014 08:40:24.869724 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f7f758c5-b38e-4a88-b96e-3cc43bfac9d5-utilities" (OuterVolumeSpecName: "utilities") pod "f7f758c5-b38e-4a88-b96e-3cc43bfac9d5" (UID: "f7f758c5-b38e-4a88-b96e-3cc43bfac9d5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:40:24 crc kubenswrapper[4870]: I1014 08:40:24.877366 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f7f758c5-b38e-4a88-b96e-3cc43bfac9d5-kube-api-access-2whlx" (OuterVolumeSpecName: "kube-api-access-2whlx") pod "f7f758c5-b38e-4a88-b96e-3cc43bfac9d5" (UID: "f7f758c5-b38e-4a88-b96e-3cc43bfac9d5"). InnerVolumeSpecName "kube-api-access-2whlx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:40:24 crc kubenswrapper[4870]: I1014 08:40:24.969726 4870 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f7f758c5-b38e-4a88-b96e-3cc43bfac9d5-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 08:40:24 crc kubenswrapper[4870]: I1014 08:40:24.969779 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2whlx\" (UniqueName: \"kubernetes.io/projected/f7f758c5-b38e-4a88-b96e-3cc43bfac9d5-kube-api-access-2whlx\") on node \"crc\" DevicePath \"\"" Oct 14 08:40:25 crc kubenswrapper[4870]: I1014 08:40:25.284096 4870 generic.go:334] "Generic (PLEG): container finished" podID="f7f758c5-b38e-4a88-b96e-3cc43bfac9d5" containerID="a2b45ea68eb853f92416531b76e36b85734f43bf69c1bff8aac1f4506060cd59" exitCode=0 Oct 14 08:40:25 crc kubenswrapper[4870]: I1014 08:40:25.284209 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7m2jn" event={"ID":"f7f758c5-b38e-4a88-b96e-3cc43bfac9d5","Type":"ContainerDied","Data":"a2b45ea68eb853f92416531b76e36b85734f43bf69c1bff8aac1f4506060cd59"} Oct 14 08:40:25 crc kubenswrapper[4870]: I1014 08:40:25.284240 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7m2jn" Oct 14 08:40:25 crc kubenswrapper[4870]: I1014 08:40:25.284293 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7m2jn" event={"ID":"f7f758c5-b38e-4a88-b96e-3cc43bfac9d5","Type":"ContainerDied","Data":"121cf51348d580fdfd4d99f61bda6cc167f8379d8a9c08610f04f582d0e1c1f2"} Oct 14 08:40:25 crc kubenswrapper[4870]: I1014 08:40:25.284354 4870 scope.go:117] "RemoveContainer" containerID="a2b45ea68eb853f92416531b76e36b85734f43bf69c1bff8aac1f4506060cd59" Oct 14 08:40:25 crc kubenswrapper[4870]: I1014 08:40:25.328737 4870 scope.go:117] "RemoveContainer" containerID="c55801a00ec720b64e71069dd9f28e098d5b04396f71fff0edd2000fd479d621" Oct 14 08:40:25 crc kubenswrapper[4870]: I1014 08:40:25.332659 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f7f758c5-b38e-4a88-b96e-3cc43bfac9d5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f7f758c5-b38e-4a88-b96e-3cc43bfac9d5" (UID: "f7f758c5-b38e-4a88-b96e-3cc43bfac9d5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:40:25 crc kubenswrapper[4870]: I1014 08:40:25.357324 4870 scope.go:117] "RemoveContainer" containerID="996723e2f597f544da216d4d3d780ecfc892ca42627f3bc5f2b12c3d9e6e5c37" Oct 14 08:40:25 crc kubenswrapper[4870]: I1014 08:40:25.377977 4870 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f7f758c5-b38e-4a88-b96e-3cc43bfac9d5-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 08:40:25 crc kubenswrapper[4870]: I1014 08:40:25.419190 4870 scope.go:117] "RemoveContainer" containerID="a2b45ea68eb853f92416531b76e36b85734f43bf69c1bff8aac1f4506060cd59" Oct 14 08:40:25 crc kubenswrapper[4870]: E1014 08:40:25.419983 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a2b45ea68eb853f92416531b76e36b85734f43bf69c1bff8aac1f4506060cd59\": container with ID starting with a2b45ea68eb853f92416531b76e36b85734f43bf69c1bff8aac1f4506060cd59 not found: ID does not exist" containerID="a2b45ea68eb853f92416531b76e36b85734f43bf69c1bff8aac1f4506060cd59" Oct 14 08:40:25 crc kubenswrapper[4870]: I1014 08:40:25.420044 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a2b45ea68eb853f92416531b76e36b85734f43bf69c1bff8aac1f4506060cd59"} err="failed to get container status \"a2b45ea68eb853f92416531b76e36b85734f43bf69c1bff8aac1f4506060cd59\": rpc error: code = NotFound desc = could not find container \"a2b45ea68eb853f92416531b76e36b85734f43bf69c1bff8aac1f4506060cd59\": container with ID starting with a2b45ea68eb853f92416531b76e36b85734f43bf69c1bff8aac1f4506060cd59 not found: ID does not exist" Oct 14 08:40:25 crc kubenswrapper[4870]: I1014 08:40:25.420083 4870 scope.go:117] "RemoveContainer" containerID="c55801a00ec720b64e71069dd9f28e098d5b04396f71fff0edd2000fd479d621" Oct 14 08:40:25 crc kubenswrapper[4870]: E1014 08:40:25.420764 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c55801a00ec720b64e71069dd9f28e098d5b04396f71fff0edd2000fd479d621\": container with ID starting with c55801a00ec720b64e71069dd9f28e098d5b04396f71fff0edd2000fd479d621 not found: ID does not exist" containerID="c55801a00ec720b64e71069dd9f28e098d5b04396f71fff0edd2000fd479d621" Oct 14 08:40:25 crc kubenswrapper[4870]: I1014 08:40:25.420808 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c55801a00ec720b64e71069dd9f28e098d5b04396f71fff0edd2000fd479d621"} err="failed to get container status \"c55801a00ec720b64e71069dd9f28e098d5b04396f71fff0edd2000fd479d621\": rpc error: code = NotFound desc = could not find container \"c55801a00ec720b64e71069dd9f28e098d5b04396f71fff0edd2000fd479d621\": container with ID starting with c55801a00ec720b64e71069dd9f28e098d5b04396f71fff0edd2000fd479d621 not found: ID does not exist" Oct 14 08:40:25 crc kubenswrapper[4870]: I1014 08:40:25.420838 4870 scope.go:117] "RemoveContainer" containerID="996723e2f597f544da216d4d3d780ecfc892ca42627f3bc5f2b12c3d9e6e5c37" Oct 14 08:40:25 crc kubenswrapper[4870]: E1014 08:40:25.421371 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"996723e2f597f544da216d4d3d780ecfc892ca42627f3bc5f2b12c3d9e6e5c37\": container with ID starting with 996723e2f597f544da216d4d3d780ecfc892ca42627f3bc5f2b12c3d9e6e5c37 not found: ID does not exist" containerID="996723e2f597f544da216d4d3d780ecfc892ca42627f3bc5f2b12c3d9e6e5c37" Oct 14 08:40:25 crc kubenswrapper[4870]: I1014 08:40:25.421410 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"996723e2f597f544da216d4d3d780ecfc892ca42627f3bc5f2b12c3d9e6e5c37"} err="failed to get container status \"996723e2f597f544da216d4d3d780ecfc892ca42627f3bc5f2b12c3d9e6e5c37\": rpc error: code = NotFound desc = could not find container \"996723e2f597f544da216d4d3d780ecfc892ca42627f3bc5f2b12c3d9e6e5c37\": container with ID starting with 996723e2f597f544da216d4d3d780ecfc892ca42627f3bc5f2b12c3d9e6e5c37 not found: ID does not exist" Oct 14 08:40:25 crc kubenswrapper[4870]: I1014 08:40:25.622742 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-7m2jn"] Oct 14 08:40:25 crc kubenswrapper[4870]: I1014 08:40:25.633521 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-7m2jn"] Oct 14 08:40:25 crc kubenswrapper[4870]: I1014 08:40:25.701319 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2-default" Oct 14 08:40:25 crc kubenswrapper[4870]: I1014 08:40:25.783570 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l7tkp\" (UniqueName: \"kubernetes.io/projected/95f78882-9f4d-4cc8-a77a-ed8aaccef072-kube-api-access-l7tkp\") pod \"95f78882-9f4d-4cc8-a77a-ed8aaccef072\" (UID: \"95f78882-9f4d-4cc8-a77a-ed8aaccef072\") " Oct 14 08:40:25 crc kubenswrapper[4870]: I1014 08:40:25.788770 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/95f78882-9f4d-4cc8-a77a-ed8aaccef072-kube-api-access-l7tkp" (OuterVolumeSpecName: "kube-api-access-l7tkp") pod "95f78882-9f4d-4cc8-a77a-ed8aaccef072" (UID: "95f78882-9f4d-4cc8-a77a-ed8aaccef072"). InnerVolumeSpecName "kube-api-access-l7tkp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:40:25 crc kubenswrapper[4870]: I1014 08:40:25.791068 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-2-default_95f78882-9f4d-4cc8-a77a-ed8aaccef072/mariadb-client-2-default/0.log" Oct 14 08:40:25 crc kubenswrapper[4870]: I1014 08:40:25.817454 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-2-default"] Oct 14 08:40:25 crc kubenswrapper[4870]: I1014 08:40:25.822584 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-2-default"] Oct 14 08:40:25 crc kubenswrapper[4870]: I1014 08:40:25.886077 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l7tkp\" (UniqueName: \"kubernetes.io/projected/95f78882-9f4d-4cc8-a77a-ed8aaccef072-kube-api-access-l7tkp\") on node \"crc\" DevicePath \"\"" Oct 14 08:40:26 crc kubenswrapper[4870]: I1014 08:40:26.297891 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-1"] Oct 14 08:40:26 crc kubenswrapper[4870]: E1014 08:40:26.298543 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7f758c5-b38e-4a88-b96e-3cc43bfac9d5" containerName="extract-content" Oct 14 08:40:26 crc kubenswrapper[4870]: I1014 08:40:26.298578 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7f758c5-b38e-4a88-b96e-3cc43bfac9d5" containerName="extract-content" Oct 14 08:40:26 crc kubenswrapper[4870]: E1014 08:40:26.298612 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7f758c5-b38e-4a88-b96e-3cc43bfac9d5" containerName="registry-server" Oct 14 08:40:26 crc kubenswrapper[4870]: I1014 08:40:26.298630 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7f758c5-b38e-4a88-b96e-3cc43bfac9d5" containerName="registry-server" Oct 14 08:40:26 crc kubenswrapper[4870]: E1014 08:40:26.298665 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95f78882-9f4d-4cc8-a77a-ed8aaccef072" containerName="mariadb-client-2-default" Oct 14 08:40:26 crc kubenswrapper[4870]: I1014 08:40:26.298682 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="95f78882-9f4d-4cc8-a77a-ed8aaccef072" containerName="mariadb-client-2-default" Oct 14 08:40:26 crc kubenswrapper[4870]: E1014 08:40:26.298733 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7f758c5-b38e-4a88-b96e-3cc43bfac9d5" containerName="extract-utilities" Oct 14 08:40:26 crc kubenswrapper[4870]: I1014 08:40:26.298750 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7f758c5-b38e-4a88-b96e-3cc43bfac9d5" containerName="extract-utilities" Oct 14 08:40:26 crc kubenswrapper[4870]: I1014 08:40:26.299133 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="95f78882-9f4d-4cc8-a77a-ed8aaccef072" containerName="mariadb-client-2-default" Oct 14 08:40:26 crc kubenswrapper[4870]: I1014 08:40:26.299170 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="f7f758c5-b38e-4a88-b96e-3cc43bfac9d5" containerName="registry-server" Oct 14 08:40:26 crc kubenswrapper[4870]: I1014 08:40:26.300483 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1" Oct 14 08:40:26 crc kubenswrapper[4870]: I1014 08:40:26.309368 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-1"] Oct 14 08:40:26 crc kubenswrapper[4870]: I1014 08:40:26.346362 4870 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dc68bebb5ea1271276b900bede5356ac169674f85a8f0fa5bd7a31fc5128577e" Oct 14 08:40:26 crc kubenswrapper[4870]: I1014 08:40:26.346658 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2-default" Oct 14 08:40:26 crc kubenswrapper[4870]: I1014 08:40:26.395178 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hqn77\" (UniqueName: \"kubernetes.io/projected/532ce506-a3df-486b-88c3-bcda390f2607-kube-api-access-hqn77\") pod \"mariadb-client-1\" (UID: \"532ce506-a3df-486b-88c3-bcda390f2607\") " pod="openstack/mariadb-client-1" Oct 14 08:40:26 crc kubenswrapper[4870]: I1014 08:40:26.498227 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hqn77\" (UniqueName: \"kubernetes.io/projected/532ce506-a3df-486b-88c3-bcda390f2607-kube-api-access-hqn77\") pod \"mariadb-client-1\" (UID: \"532ce506-a3df-486b-88c3-bcda390f2607\") " pod="openstack/mariadb-client-1" Oct 14 08:40:26 crc kubenswrapper[4870]: I1014 08:40:26.528629 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hqn77\" (UniqueName: \"kubernetes.io/projected/532ce506-a3df-486b-88c3-bcda390f2607-kube-api-access-hqn77\") pod \"mariadb-client-1\" (UID: \"532ce506-a3df-486b-88c3-bcda390f2607\") " pod="openstack/mariadb-client-1" Oct 14 08:40:26 crc kubenswrapper[4870]: I1014 08:40:26.661079 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1" Oct 14 08:40:27 crc kubenswrapper[4870]: I1014 08:40:27.045309 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="95f78882-9f4d-4cc8-a77a-ed8aaccef072" path="/var/lib/kubelet/pods/95f78882-9f4d-4cc8-a77a-ed8aaccef072/volumes" Oct 14 08:40:27 crc kubenswrapper[4870]: I1014 08:40:27.046739 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f7f758c5-b38e-4a88-b96e-3cc43bfac9d5" path="/var/lib/kubelet/pods/f7f758c5-b38e-4a88-b96e-3cc43bfac9d5/volumes" Oct 14 08:40:27 crc kubenswrapper[4870]: I1014 08:40:27.244754 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-1"] Oct 14 08:40:27 crc kubenswrapper[4870]: W1014 08:40:27.247865 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod532ce506_a3df_486b_88c3_bcda390f2607.slice/crio-522e6d29a8709d1fa3b5d7029eba7aa172fb5f5475ade1956dc6123523ce871e WatchSource:0}: Error finding container 522e6d29a8709d1fa3b5d7029eba7aa172fb5f5475ade1956dc6123523ce871e: Status 404 returned error can't find the container with id 522e6d29a8709d1fa3b5d7029eba7aa172fb5f5475ade1956dc6123523ce871e Oct 14 08:40:27 crc kubenswrapper[4870]: I1014 08:40:27.355566 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-1" event={"ID":"532ce506-a3df-486b-88c3-bcda390f2607","Type":"ContainerStarted","Data":"522e6d29a8709d1fa3b5d7029eba7aa172fb5f5475ade1956dc6123523ce871e"} Oct 14 08:40:28 crc kubenswrapper[4870]: I1014 08:40:28.366284 4870 generic.go:334] "Generic (PLEG): container finished" podID="532ce506-a3df-486b-88c3-bcda390f2607" containerID="d5d9f80caad66538a27ffe7ac3d34d1941a8068b4dba1426f38017ce3418878f" exitCode=0 Oct 14 08:40:28 crc kubenswrapper[4870]: I1014 08:40:28.366351 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-1" event={"ID":"532ce506-a3df-486b-88c3-bcda390f2607","Type":"ContainerDied","Data":"d5d9f80caad66538a27ffe7ac3d34d1941a8068b4dba1426f38017ce3418878f"} Oct 14 08:40:29 crc kubenswrapper[4870]: I1014 08:40:29.749982 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1" Oct 14 08:40:29 crc kubenswrapper[4870]: I1014 08:40:29.767268 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-1_532ce506-a3df-486b-88c3-bcda390f2607/mariadb-client-1/0.log" Oct 14 08:40:29 crc kubenswrapper[4870]: I1014 08:40:29.800403 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-1"] Oct 14 08:40:29 crc kubenswrapper[4870]: I1014 08:40:29.805706 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-1"] Oct 14 08:40:29 crc kubenswrapper[4870]: I1014 08:40:29.854059 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hqn77\" (UniqueName: \"kubernetes.io/projected/532ce506-a3df-486b-88c3-bcda390f2607-kube-api-access-hqn77\") pod \"532ce506-a3df-486b-88c3-bcda390f2607\" (UID: \"532ce506-a3df-486b-88c3-bcda390f2607\") " Oct 14 08:40:29 crc kubenswrapper[4870]: I1014 08:40:29.860345 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/532ce506-a3df-486b-88c3-bcda390f2607-kube-api-access-hqn77" (OuterVolumeSpecName: "kube-api-access-hqn77") pod "532ce506-a3df-486b-88c3-bcda390f2607" (UID: "532ce506-a3df-486b-88c3-bcda390f2607"). InnerVolumeSpecName "kube-api-access-hqn77". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:40:29 crc kubenswrapper[4870]: I1014 08:40:29.956327 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hqn77\" (UniqueName: \"kubernetes.io/projected/532ce506-a3df-486b-88c3-bcda390f2607-kube-api-access-hqn77\") on node \"crc\" DevicePath \"\"" Oct 14 08:40:30 crc kubenswrapper[4870]: I1014 08:40:30.274846 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-4-default"] Oct 14 08:40:30 crc kubenswrapper[4870]: E1014 08:40:30.275848 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="532ce506-a3df-486b-88c3-bcda390f2607" containerName="mariadb-client-1" Oct 14 08:40:30 crc kubenswrapper[4870]: I1014 08:40:30.275876 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="532ce506-a3df-486b-88c3-bcda390f2607" containerName="mariadb-client-1" Oct 14 08:40:30 crc kubenswrapper[4870]: I1014 08:40:30.276098 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="532ce506-a3df-486b-88c3-bcda390f2607" containerName="mariadb-client-1" Oct 14 08:40:30 crc kubenswrapper[4870]: I1014 08:40:30.276971 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-4-default" Oct 14 08:40:30 crc kubenswrapper[4870]: I1014 08:40:30.284426 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-4-default"] Oct 14 08:40:30 crc kubenswrapper[4870]: I1014 08:40:30.363848 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dpfb8\" (UniqueName: \"kubernetes.io/projected/beb744a1-8672-4cff-b63b-084b8b6f7f3f-kube-api-access-dpfb8\") pod \"mariadb-client-4-default\" (UID: \"beb744a1-8672-4cff-b63b-084b8b6f7f3f\") " pod="openstack/mariadb-client-4-default" Oct 14 08:40:30 crc kubenswrapper[4870]: I1014 08:40:30.387995 4870 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="522e6d29a8709d1fa3b5d7029eba7aa172fb5f5475ade1956dc6123523ce871e" Oct 14 08:40:30 crc kubenswrapper[4870]: I1014 08:40:30.388410 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1" Oct 14 08:40:30 crc kubenswrapper[4870]: I1014 08:40:30.466041 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dpfb8\" (UniqueName: \"kubernetes.io/projected/beb744a1-8672-4cff-b63b-084b8b6f7f3f-kube-api-access-dpfb8\") pod \"mariadb-client-4-default\" (UID: \"beb744a1-8672-4cff-b63b-084b8b6f7f3f\") " pod="openstack/mariadb-client-4-default" Oct 14 08:40:30 crc kubenswrapper[4870]: I1014 08:40:30.490306 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dpfb8\" (UniqueName: \"kubernetes.io/projected/beb744a1-8672-4cff-b63b-084b8b6f7f3f-kube-api-access-dpfb8\") pod \"mariadb-client-4-default\" (UID: \"beb744a1-8672-4cff-b63b-084b8b6f7f3f\") " pod="openstack/mariadb-client-4-default" Oct 14 08:40:30 crc kubenswrapper[4870]: I1014 08:40:30.615694 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-4-default" Oct 14 08:40:30 crc kubenswrapper[4870]: I1014 08:40:30.962028 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-4-default"] Oct 14 08:40:30 crc kubenswrapper[4870]: W1014 08:40:30.968272 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbeb744a1_8672_4cff_b63b_084b8b6f7f3f.slice/crio-29914b8d08d82ecbd0e4965a0e2bdd2406ac96d4d959651218f44f849750aada WatchSource:0}: Error finding container 29914b8d08d82ecbd0e4965a0e2bdd2406ac96d4d959651218f44f849750aada: Status 404 returned error can't find the container with id 29914b8d08d82ecbd0e4965a0e2bdd2406ac96d4d959651218f44f849750aada Oct 14 08:40:31 crc kubenswrapper[4870]: I1014 08:40:31.061558 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="532ce506-a3df-486b-88c3-bcda390f2607" path="/var/lib/kubelet/pods/532ce506-a3df-486b-88c3-bcda390f2607/volumes" Oct 14 08:40:31 crc kubenswrapper[4870]: I1014 08:40:31.403265 4870 generic.go:334] "Generic (PLEG): container finished" podID="beb744a1-8672-4cff-b63b-084b8b6f7f3f" containerID="3ba6b002b289fc6453f10e43aa0495b0623d6b28259cf20d14c14864fd42a879" exitCode=0 Oct 14 08:40:31 crc kubenswrapper[4870]: I1014 08:40:31.403528 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-4-default" event={"ID":"beb744a1-8672-4cff-b63b-084b8b6f7f3f","Type":"ContainerDied","Data":"3ba6b002b289fc6453f10e43aa0495b0623d6b28259cf20d14c14864fd42a879"} Oct 14 08:40:31 crc kubenswrapper[4870]: I1014 08:40:31.403832 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-4-default" event={"ID":"beb744a1-8672-4cff-b63b-084b8b6f7f3f","Type":"ContainerStarted","Data":"29914b8d08d82ecbd0e4965a0e2bdd2406ac96d4d959651218f44f849750aada"} Oct 14 08:40:32 crc kubenswrapper[4870]: I1014 08:40:32.880689 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-4-default" Oct 14 08:40:32 crc kubenswrapper[4870]: I1014 08:40:32.897818 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-4-default_beb744a1-8672-4cff-b63b-084b8b6f7f3f/mariadb-client-4-default/0.log" Oct 14 08:40:32 crc kubenswrapper[4870]: I1014 08:40:32.926908 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-4-default"] Oct 14 08:40:32 crc kubenswrapper[4870]: I1014 08:40:32.932194 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-4-default"] Oct 14 08:40:33 crc kubenswrapper[4870]: I1014 08:40:33.015285 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dpfb8\" (UniqueName: \"kubernetes.io/projected/beb744a1-8672-4cff-b63b-084b8b6f7f3f-kube-api-access-dpfb8\") pod \"beb744a1-8672-4cff-b63b-084b8b6f7f3f\" (UID: \"beb744a1-8672-4cff-b63b-084b8b6f7f3f\") " Oct 14 08:40:33 crc kubenswrapper[4870]: I1014 08:40:33.021107 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/beb744a1-8672-4cff-b63b-084b8b6f7f3f-kube-api-access-dpfb8" (OuterVolumeSpecName: "kube-api-access-dpfb8") pod "beb744a1-8672-4cff-b63b-084b8b6f7f3f" (UID: "beb744a1-8672-4cff-b63b-084b8b6f7f3f"). InnerVolumeSpecName "kube-api-access-dpfb8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:40:33 crc kubenswrapper[4870]: I1014 08:40:33.046862 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="beb744a1-8672-4cff-b63b-084b8b6f7f3f" path="/var/lib/kubelet/pods/beb744a1-8672-4cff-b63b-084b8b6f7f3f/volumes" Oct 14 08:40:33 crc kubenswrapper[4870]: I1014 08:40:33.117541 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dpfb8\" (UniqueName: \"kubernetes.io/projected/beb744a1-8672-4cff-b63b-084b8b6f7f3f-kube-api-access-dpfb8\") on node \"crc\" DevicePath \"\"" Oct 14 08:40:33 crc kubenswrapper[4870]: I1014 08:40:33.424499 4870 scope.go:117] "RemoveContainer" containerID="3ba6b002b289fc6453f10e43aa0495b0623d6b28259cf20d14c14864fd42a879" Oct 14 08:40:33 crc kubenswrapper[4870]: I1014 08:40:33.424587 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-4-default" Oct 14 08:40:36 crc kubenswrapper[4870]: I1014 08:40:36.551490 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-5-default"] Oct 14 08:40:36 crc kubenswrapper[4870]: E1014 08:40:36.552230 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="beb744a1-8672-4cff-b63b-084b8b6f7f3f" containerName="mariadb-client-4-default" Oct 14 08:40:36 crc kubenswrapper[4870]: I1014 08:40:36.552245 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="beb744a1-8672-4cff-b63b-084b8b6f7f3f" containerName="mariadb-client-4-default" Oct 14 08:40:36 crc kubenswrapper[4870]: I1014 08:40:36.552393 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="beb744a1-8672-4cff-b63b-084b8b6f7f3f" containerName="mariadb-client-4-default" Oct 14 08:40:36 crc kubenswrapper[4870]: I1014 08:40:36.552982 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-5-default" Oct 14 08:40:36 crc kubenswrapper[4870]: I1014 08:40:36.555799 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-vnjjw" Oct 14 08:40:36 crc kubenswrapper[4870]: I1014 08:40:36.560305 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-5-default"] Oct 14 08:40:36 crc kubenswrapper[4870]: I1014 08:40:36.688353 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cbnhn\" (UniqueName: \"kubernetes.io/projected/d0d2982b-a0a6-480c-a572-a82ab529df1f-kube-api-access-cbnhn\") pod \"mariadb-client-5-default\" (UID: \"d0d2982b-a0a6-480c-a572-a82ab529df1f\") " pod="openstack/mariadb-client-5-default" Oct 14 08:40:36 crc kubenswrapper[4870]: I1014 08:40:36.789678 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cbnhn\" (UniqueName: \"kubernetes.io/projected/d0d2982b-a0a6-480c-a572-a82ab529df1f-kube-api-access-cbnhn\") pod \"mariadb-client-5-default\" (UID: \"d0d2982b-a0a6-480c-a572-a82ab529df1f\") " pod="openstack/mariadb-client-5-default" Oct 14 08:40:36 crc kubenswrapper[4870]: I1014 08:40:36.813294 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cbnhn\" (UniqueName: \"kubernetes.io/projected/d0d2982b-a0a6-480c-a572-a82ab529df1f-kube-api-access-cbnhn\") pod \"mariadb-client-5-default\" (UID: \"d0d2982b-a0a6-480c-a572-a82ab529df1f\") " pod="openstack/mariadb-client-5-default" Oct 14 08:40:36 crc kubenswrapper[4870]: I1014 08:40:36.886286 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-5-default" Oct 14 08:40:37 crc kubenswrapper[4870]: I1014 08:40:37.372136 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-5-default"] Oct 14 08:40:37 crc kubenswrapper[4870]: W1014 08:40:37.381146 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd0d2982b_a0a6_480c_a572_a82ab529df1f.slice/crio-d04638141236b790a36ecb8f3eac16027e4e20b2015538613b8e32d1e33194c1 WatchSource:0}: Error finding container d04638141236b790a36ecb8f3eac16027e4e20b2015538613b8e32d1e33194c1: Status 404 returned error can't find the container with id d04638141236b790a36ecb8f3eac16027e4e20b2015538613b8e32d1e33194c1 Oct 14 08:40:37 crc kubenswrapper[4870]: I1014 08:40:37.468133 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-5-default" event={"ID":"d0d2982b-a0a6-480c-a572-a82ab529df1f","Type":"ContainerStarted","Data":"d04638141236b790a36ecb8f3eac16027e4e20b2015538613b8e32d1e33194c1"} Oct 14 08:40:38 crc kubenswrapper[4870]: I1014 08:40:38.485003 4870 generic.go:334] "Generic (PLEG): container finished" podID="d0d2982b-a0a6-480c-a572-a82ab529df1f" containerID="db99a336904c4b049dbc03b9fb8b5549dd5ace6bd3c7f587f281aede43d03bb1" exitCode=0 Oct 14 08:40:38 crc kubenswrapper[4870]: I1014 08:40:38.485072 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-5-default" event={"ID":"d0d2982b-a0a6-480c-a572-a82ab529df1f","Type":"ContainerDied","Data":"db99a336904c4b049dbc03b9fb8b5549dd5ace6bd3c7f587f281aede43d03bb1"} Oct 14 08:40:39 crc kubenswrapper[4870]: I1014 08:40:39.985794 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-5-default" Oct 14 08:40:40 crc kubenswrapper[4870]: I1014 08:40:40.009874 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-5-default_d0d2982b-a0a6-480c-a572-a82ab529df1f/mariadb-client-5-default/0.log" Oct 14 08:40:40 crc kubenswrapper[4870]: I1014 08:40:40.043510 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-5-default"] Oct 14 08:40:40 crc kubenswrapper[4870]: I1014 08:40:40.050766 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-5-default"] Oct 14 08:40:40 crc kubenswrapper[4870]: I1014 08:40:40.143862 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cbnhn\" (UniqueName: \"kubernetes.io/projected/d0d2982b-a0a6-480c-a572-a82ab529df1f-kube-api-access-cbnhn\") pod \"d0d2982b-a0a6-480c-a572-a82ab529df1f\" (UID: \"d0d2982b-a0a6-480c-a572-a82ab529df1f\") " Oct 14 08:40:40 crc kubenswrapper[4870]: I1014 08:40:40.151082 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d0d2982b-a0a6-480c-a572-a82ab529df1f-kube-api-access-cbnhn" (OuterVolumeSpecName: "kube-api-access-cbnhn") pod "d0d2982b-a0a6-480c-a572-a82ab529df1f" (UID: "d0d2982b-a0a6-480c-a572-a82ab529df1f"). InnerVolumeSpecName "kube-api-access-cbnhn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:40:40 crc kubenswrapper[4870]: I1014 08:40:40.176027 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-6-default"] Oct 14 08:40:40 crc kubenswrapper[4870]: E1014 08:40:40.176713 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0d2982b-a0a6-480c-a572-a82ab529df1f" containerName="mariadb-client-5-default" Oct 14 08:40:40 crc kubenswrapper[4870]: I1014 08:40:40.176756 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0d2982b-a0a6-480c-a572-a82ab529df1f" containerName="mariadb-client-5-default" Oct 14 08:40:40 crc kubenswrapper[4870]: I1014 08:40:40.177265 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="d0d2982b-a0a6-480c-a572-a82ab529df1f" containerName="mariadb-client-5-default" Oct 14 08:40:40 crc kubenswrapper[4870]: I1014 08:40:40.178530 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-6-default" Oct 14 08:40:40 crc kubenswrapper[4870]: I1014 08:40:40.186657 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-6-default"] Oct 14 08:40:40 crc kubenswrapper[4870]: I1014 08:40:40.245795 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cbnhn\" (UniqueName: \"kubernetes.io/projected/d0d2982b-a0a6-480c-a572-a82ab529df1f-kube-api-access-cbnhn\") on node \"crc\" DevicePath \"\"" Oct 14 08:40:40 crc kubenswrapper[4870]: I1014 08:40:40.347484 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tgf6x\" (UniqueName: \"kubernetes.io/projected/4a5008f8-ef1e-4a8e-ba68-6c39eacc9fc1-kube-api-access-tgf6x\") pod \"mariadb-client-6-default\" (UID: \"4a5008f8-ef1e-4a8e-ba68-6c39eacc9fc1\") " pod="openstack/mariadb-client-6-default" Oct 14 08:40:40 crc kubenswrapper[4870]: I1014 08:40:40.449390 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tgf6x\" (UniqueName: \"kubernetes.io/projected/4a5008f8-ef1e-4a8e-ba68-6c39eacc9fc1-kube-api-access-tgf6x\") pod \"mariadb-client-6-default\" (UID: \"4a5008f8-ef1e-4a8e-ba68-6c39eacc9fc1\") " pod="openstack/mariadb-client-6-default" Oct 14 08:40:40 crc kubenswrapper[4870]: I1014 08:40:40.484980 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tgf6x\" (UniqueName: \"kubernetes.io/projected/4a5008f8-ef1e-4a8e-ba68-6c39eacc9fc1-kube-api-access-tgf6x\") pod \"mariadb-client-6-default\" (UID: \"4a5008f8-ef1e-4a8e-ba68-6c39eacc9fc1\") " pod="openstack/mariadb-client-6-default" Oct 14 08:40:40 crc kubenswrapper[4870]: I1014 08:40:40.520026 4870 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d04638141236b790a36ecb8f3eac16027e4e20b2015538613b8e32d1e33194c1" Oct 14 08:40:40 crc kubenswrapper[4870]: I1014 08:40:40.520096 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-5-default" Oct 14 08:40:40 crc kubenswrapper[4870]: I1014 08:40:40.521603 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-6-default" Oct 14 08:40:41 crc kubenswrapper[4870]: I1014 08:40:41.058421 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d0d2982b-a0a6-480c-a572-a82ab529df1f" path="/var/lib/kubelet/pods/d0d2982b-a0a6-480c-a572-a82ab529df1f/volumes" Oct 14 08:40:41 crc kubenswrapper[4870]: I1014 08:40:41.090428 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-6-default"] Oct 14 08:40:41 crc kubenswrapper[4870]: W1014 08:40:41.092628 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4a5008f8_ef1e_4a8e_ba68_6c39eacc9fc1.slice/crio-1cf5b4742b5e4256ef5a23ad602eb9e620186697c5195fd1fd67395b94b7a4b8 WatchSource:0}: Error finding container 1cf5b4742b5e4256ef5a23ad602eb9e620186697c5195fd1fd67395b94b7a4b8: Status 404 returned error can't find the container with id 1cf5b4742b5e4256ef5a23ad602eb9e620186697c5195fd1fd67395b94b7a4b8 Oct 14 08:40:41 crc kubenswrapper[4870]: I1014 08:40:41.534177 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-6-default" event={"ID":"4a5008f8-ef1e-4a8e-ba68-6c39eacc9fc1","Type":"ContainerStarted","Data":"5432ef26aa9ae79f0bb86f63f580920bd3fd7ca70233298460b8419e0d4c7106"} Oct 14 08:40:41 crc kubenswrapper[4870]: I1014 08:40:41.534803 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-6-default" event={"ID":"4a5008f8-ef1e-4a8e-ba68-6c39eacc9fc1","Type":"ContainerStarted","Data":"1cf5b4742b5e4256ef5a23ad602eb9e620186697c5195fd1fd67395b94b7a4b8"} Oct 14 08:40:41 crc kubenswrapper[4870]: I1014 08:40:41.562968 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/mariadb-client-6-default" podStartSLOduration=1.5629378539999998 podStartE2EDuration="1.562937854s" podCreationTimestamp="2025-10-14 08:40:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 08:40:41.55454499 +0000 UTC m=+5977.251905391" watchObservedRunningTime="2025-10-14 08:40:41.562937854 +0000 UTC m=+5977.260298255" Oct 14 08:40:42 crc kubenswrapper[4870]: I1014 08:40:42.550409 4870 generic.go:334] "Generic (PLEG): container finished" podID="4a5008f8-ef1e-4a8e-ba68-6c39eacc9fc1" containerID="5432ef26aa9ae79f0bb86f63f580920bd3fd7ca70233298460b8419e0d4c7106" exitCode=0 Oct 14 08:40:42 crc kubenswrapper[4870]: I1014 08:40:42.550509 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-6-default" event={"ID":"4a5008f8-ef1e-4a8e-ba68-6c39eacc9fc1","Type":"ContainerDied","Data":"5432ef26aa9ae79f0bb86f63f580920bd3fd7ca70233298460b8419e0d4c7106"} Oct 14 08:40:43 crc kubenswrapper[4870]: I1014 08:40:43.998504 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-6-default" Oct 14 08:40:44 crc kubenswrapper[4870]: I1014 08:40:44.047880 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-6-default"] Oct 14 08:40:44 crc kubenswrapper[4870]: I1014 08:40:44.054120 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-6-default"] Oct 14 08:40:44 crc kubenswrapper[4870]: I1014 08:40:44.112369 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tgf6x\" (UniqueName: \"kubernetes.io/projected/4a5008f8-ef1e-4a8e-ba68-6c39eacc9fc1-kube-api-access-tgf6x\") pod \"4a5008f8-ef1e-4a8e-ba68-6c39eacc9fc1\" (UID: \"4a5008f8-ef1e-4a8e-ba68-6c39eacc9fc1\") " Oct 14 08:40:44 crc kubenswrapper[4870]: I1014 08:40:44.118380 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4a5008f8-ef1e-4a8e-ba68-6c39eacc9fc1-kube-api-access-tgf6x" (OuterVolumeSpecName: "kube-api-access-tgf6x") pod "4a5008f8-ef1e-4a8e-ba68-6c39eacc9fc1" (UID: "4a5008f8-ef1e-4a8e-ba68-6c39eacc9fc1"). InnerVolumeSpecName "kube-api-access-tgf6x". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:40:44 crc kubenswrapper[4870]: I1014 08:40:44.214499 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tgf6x\" (UniqueName: \"kubernetes.io/projected/4a5008f8-ef1e-4a8e-ba68-6c39eacc9fc1-kube-api-access-tgf6x\") on node \"crc\" DevicePath \"\"" Oct 14 08:40:44 crc kubenswrapper[4870]: I1014 08:40:44.237606 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-7-default"] Oct 14 08:40:44 crc kubenswrapper[4870]: E1014 08:40:44.238241 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4a5008f8-ef1e-4a8e-ba68-6c39eacc9fc1" containerName="mariadb-client-6-default" Oct 14 08:40:44 crc kubenswrapper[4870]: I1014 08:40:44.238272 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="4a5008f8-ef1e-4a8e-ba68-6c39eacc9fc1" containerName="mariadb-client-6-default" Oct 14 08:40:44 crc kubenswrapper[4870]: I1014 08:40:44.238718 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="4a5008f8-ef1e-4a8e-ba68-6c39eacc9fc1" containerName="mariadb-client-6-default" Oct 14 08:40:44 crc kubenswrapper[4870]: I1014 08:40:44.239922 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-7-default" Oct 14 08:40:44 crc kubenswrapper[4870]: I1014 08:40:44.250899 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-7-default"] Oct 14 08:40:44 crc kubenswrapper[4870]: I1014 08:40:44.316770 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qvphf\" (UniqueName: \"kubernetes.io/projected/f3866e2f-0055-4ac3-941b-92c2eccec3b1-kube-api-access-qvphf\") pod \"mariadb-client-7-default\" (UID: \"f3866e2f-0055-4ac3-941b-92c2eccec3b1\") " pod="openstack/mariadb-client-7-default" Oct 14 08:40:44 crc kubenswrapper[4870]: I1014 08:40:44.418261 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qvphf\" (UniqueName: \"kubernetes.io/projected/f3866e2f-0055-4ac3-941b-92c2eccec3b1-kube-api-access-qvphf\") pod \"mariadb-client-7-default\" (UID: \"f3866e2f-0055-4ac3-941b-92c2eccec3b1\") " pod="openstack/mariadb-client-7-default" Oct 14 08:40:44 crc kubenswrapper[4870]: I1014 08:40:44.438645 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qvphf\" (UniqueName: \"kubernetes.io/projected/f3866e2f-0055-4ac3-941b-92c2eccec3b1-kube-api-access-qvphf\") pod \"mariadb-client-7-default\" (UID: \"f3866e2f-0055-4ac3-941b-92c2eccec3b1\") " pod="openstack/mariadb-client-7-default" Oct 14 08:40:44 crc kubenswrapper[4870]: I1014 08:40:44.565159 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-7-default" Oct 14 08:40:44 crc kubenswrapper[4870]: I1014 08:40:44.575920 4870 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1cf5b4742b5e4256ef5a23ad602eb9e620186697c5195fd1fd67395b94b7a4b8" Oct 14 08:40:44 crc kubenswrapper[4870]: I1014 08:40:44.576061 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-6-default" Oct 14 08:40:44 crc kubenswrapper[4870]: I1014 08:40:44.961244 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-7-default"] Oct 14 08:40:45 crc kubenswrapper[4870]: I1014 08:40:45.046703 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4a5008f8-ef1e-4a8e-ba68-6c39eacc9fc1" path="/var/lib/kubelet/pods/4a5008f8-ef1e-4a8e-ba68-6c39eacc9fc1/volumes" Oct 14 08:40:45 crc kubenswrapper[4870]: I1014 08:40:45.588500 4870 generic.go:334] "Generic (PLEG): container finished" podID="f3866e2f-0055-4ac3-941b-92c2eccec3b1" containerID="c21891951192a0896d0b3eb18379a84280ddba192f534243b0f1d9bb36da0128" exitCode=0 Oct 14 08:40:45 crc kubenswrapper[4870]: I1014 08:40:45.588566 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-7-default" event={"ID":"f3866e2f-0055-4ac3-941b-92c2eccec3b1","Type":"ContainerDied","Data":"c21891951192a0896d0b3eb18379a84280ddba192f534243b0f1d9bb36da0128"} Oct 14 08:40:45 crc kubenswrapper[4870]: I1014 08:40:45.589064 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-7-default" event={"ID":"f3866e2f-0055-4ac3-941b-92c2eccec3b1","Type":"ContainerStarted","Data":"e107efb6de80adbe8907304cabbba59ed99c0353a6f5eb50a6911d8693f1fdac"} Oct 14 08:40:47 crc kubenswrapper[4870]: I1014 08:40:47.047409 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-7-default" Oct 14 08:40:47 crc kubenswrapper[4870]: I1014 08:40:47.064850 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-7-default_f3866e2f-0055-4ac3-941b-92c2eccec3b1/mariadb-client-7-default/0.log" Oct 14 08:40:47 crc kubenswrapper[4870]: I1014 08:40:47.105608 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-7-default"] Oct 14 08:40:47 crc kubenswrapper[4870]: I1014 08:40:47.120133 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-7-default"] Oct 14 08:40:47 crc kubenswrapper[4870]: I1014 08:40:47.163388 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qvphf\" (UniqueName: \"kubernetes.io/projected/f3866e2f-0055-4ac3-941b-92c2eccec3b1-kube-api-access-qvphf\") pod \"f3866e2f-0055-4ac3-941b-92c2eccec3b1\" (UID: \"f3866e2f-0055-4ac3-941b-92c2eccec3b1\") " Oct 14 08:40:47 crc kubenswrapper[4870]: I1014 08:40:47.176768 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f3866e2f-0055-4ac3-941b-92c2eccec3b1-kube-api-access-qvphf" (OuterVolumeSpecName: "kube-api-access-qvphf") pod "f3866e2f-0055-4ac3-941b-92c2eccec3b1" (UID: "f3866e2f-0055-4ac3-941b-92c2eccec3b1"). InnerVolumeSpecName "kube-api-access-qvphf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:40:47 crc kubenswrapper[4870]: I1014 08:40:47.260423 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-2"] Oct 14 08:40:47 crc kubenswrapper[4870]: E1014 08:40:47.261128 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f3866e2f-0055-4ac3-941b-92c2eccec3b1" containerName="mariadb-client-7-default" Oct 14 08:40:47 crc kubenswrapper[4870]: I1014 08:40:47.261149 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="f3866e2f-0055-4ac3-941b-92c2eccec3b1" containerName="mariadb-client-7-default" Oct 14 08:40:47 crc kubenswrapper[4870]: I1014 08:40:47.261384 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="f3866e2f-0055-4ac3-941b-92c2eccec3b1" containerName="mariadb-client-7-default" Oct 14 08:40:47 crc kubenswrapper[4870]: I1014 08:40:47.262210 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2" Oct 14 08:40:47 crc kubenswrapper[4870]: I1014 08:40:47.266182 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qvphf\" (UniqueName: \"kubernetes.io/projected/f3866e2f-0055-4ac3-941b-92c2eccec3b1-kube-api-access-qvphf\") on node \"crc\" DevicePath \"\"" Oct 14 08:40:47 crc kubenswrapper[4870]: I1014 08:40:47.270762 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-2"] Oct 14 08:40:47 crc kubenswrapper[4870]: I1014 08:40:47.368130 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-82msg\" (UniqueName: \"kubernetes.io/projected/8bf33996-1df9-46b8-b16b-fcef5615b7dd-kube-api-access-82msg\") pod \"mariadb-client-2\" (UID: \"8bf33996-1df9-46b8-b16b-fcef5615b7dd\") " pod="openstack/mariadb-client-2" Oct 14 08:40:47 crc kubenswrapper[4870]: I1014 08:40:47.473248 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-82msg\" (UniqueName: \"kubernetes.io/projected/8bf33996-1df9-46b8-b16b-fcef5615b7dd-kube-api-access-82msg\") pod \"mariadb-client-2\" (UID: \"8bf33996-1df9-46b8-b16b-fcef5615b7dd\") " pod="openstack/mariadb-client-2" Oct 14 08:40:47 crc kubenswrapper[4870]: I1014 08:40:47.506854 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-82msg\" (UniqueName: \"kubernetes.io/projected/8bf33996-1df9-46b8-b16b-fcef5615b7dd-kube-api-access-82msg\") pod \"mariadb-client-2\" (UID: \"8bf33996-1df9-46b8-b16b-fcef5615b7dd\") " pod="openstack/mariadb-client-2" Oct 14 08:40:47 crc kubenswrapper[4870]: I1014 08:40:47.604554 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2" Oct 14 08:40:47 crc kubenswrapper[4870]: I1014 08:40:47.619278 4870 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e107efb6de80adbe8907304cabbba59ed99c0353a6f5eb50a6911d8693f1fdac" Oct 14 08:40:47 crc kubenswrapper[4870]: I1014 08:40:47.619391 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-7-default" Oct 14 08:40:48 crc kubenswrapper[4870]: I1014 08:40:48.187242 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-2"] Oct 14 08:40:48 crc kubenswrapper[4870]: W1014 08:40:48.187569 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8bf33996_1df9_46b8_b16b_fcef5615b7dd.slice/crio-c95a7e55d9346321ff57d4092d59c631dfd543544bbf96154906d3860d910c35 WatchSource:0}: Error finding container c95a7e55d9346321ff57d4092d59c631dfd543544bbf96154906d3860d910c35: Status 404 returned error can't find the container with id c95a7e55d9346321ff57d4092d59c631dfd543544bbf96154906d3860d910c35 Oct 14 08:40:48 crc kubenswrapper[4870]: I1014 08:40:48.634592 4870 generic.go:334] "Generic (PLEG): container finished" podID="8bf33996-1df9-46b8-b16b-fcef5615b7dd" containerID="cd761dbc6cd5f10b20b13d35bfc1872b207956b1b70f764e5957f1338073e111" exitCode=0 Oct 14 08:40:48 crc kubenswrapper[4870]: I1014 08:40:48.634669 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-2" event={"ID":"8bf33996-1df9-46b8-b16b-fcef5615b7dd","Type":"ContainerDied","Data":"cd761dbc6cd5f10b20b13d35bfc1872b207956b1b70f764e5957f1338073e111"} Oct 14 08:40:48 crc kubenswrapper[4870]: I1014 08:40:48.635039 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-2" event={"ID":"8bf33996-1df9-46b8-b16b-fcef5615b7dd","Type":"ContainerStarted","Data":"c95a7e55d9346321ff57d4092d59c631dfd543544bbf96154906d3860d910c35"} Oct 14 08:40:49 crc kubenswrapper[4870]: I1014 08:40:49.051536 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f3866e2f-0055-4ac3-941b-92c2eccec3b1" path="/var/lib/kubelet/pods/f3866e2f-0055-4ac3-941b-92c2eccec3b1/volumes" Oct 14 08:40:50 crc kubenswrapper[4870]: I1014 08:40:50.078917 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2" Oct 14 08:40:50 crc kubenswrapper[4870]: I1014 08:40:50.097059 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-2_8bf33996-1df9-46b8-b16b-fcef5615b7dd/mariadb-client-2/0.log" Oct 14 08:40:50 crc kubenswrapper[4870]: I1014 08:40:50.118855 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-2"] Oct 14 08:40:50 crc kubenswrapper[4870]: I1014 08:40:50.123597 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-2"] Oct 14 08:40:50 crc kubenswrapper[4870]: I1014 08:40:50.222976 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-82msg\" (UniqueName: \"kubernetes.io/projected/8bf33996-1df9-46b8-b16b-fcef5615b7dd-kube-api-access-82msg\") pod \"8bf33996-1df9-46b8-b16b-fcef5615b7dd\" (UID: \"8bf33996-1df9-46b8-b16b-fcef5615b7dd\") " Oct 14 08:40:50 crc kubenswrapper[4870]: I1014 08:40:50.231923 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8bf33996-1df9-46b8-b16b-fcef5615b7dd-kube-api-access-82msg" (OuterVolumeSpecName: "kube-api-access-82msg") pod "8bf33996-1df9-46b8-b16b-fcef5615b7dd" (UID: "8bf33996-1df9-46b8-b16b-fcef5615b7dd"). InnerVolumeSpecName "kube-api-access-82msg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:40:50 crc kubenswrapper[4870]: I1014 08:40:50.324975 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-82msg\" (UniqueName: \"kubernetes.io/projected/8bf33996-1df9-46b8-b16b-fcef5615b7dd-kube-api-access-82msg\") on node \"crc\" DevicePath \"\"" Oct 14 08:40:50 crc kubenswrapper[4870]: I1014 08:40:50.658934 4870 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c95a7e55d9346321ff57d4092d59c631dfd543544bbf96154906d3860d910c35" Oct 14 08:40:50 crc kubenswrapper[4870]: I1014 08:40:50.659044 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2" Oct 14 08:40:51 crc kubenswrapper[4870]: I1014 08:40:51.056333 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8bf33996-1df9-46b8-b16b-fcef5615b7dd" path="/var/lib/kubelet/pods/8bf33996-1df9-46b8-b16b-fcef5615b7dd/volumes" Oct 14 08:40:53 crc kubenswrapper[4870]: I1014 08:40:53.950572 4870 patch_prober.go:28] interesting pod/machine-config-daemon-7tvc7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 08:40:53 crc kubenswrapper[4870]: I1014 08:40:53.950910 4870 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 08:41:21 crc kubenswrapper[4870]: I1014 08:41:21.442136 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-nbnbr"] Oct 14 08:41:21 crc kubenswrapper[4870]: E1014 08:41:21.444382 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8bf33996-1df9-46b8-b16b-fcef5615b7dd" containerName="mariadb-client-2" Oct 14 08:41:21 crc kubenswrapper[4870]: I1014 08:41:21.444621 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="8bf33996-1df9-46b8-b16b-fcef5615b7dd" containerName="mariadb-client-2" Oct 14 08:41:21 crc kubenswrapper[4870]: I1014 08:41:21.444909 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="8bf33996-1df9-46b8-b16b-fcef5615b7dd" containerName="mariadb-client-2" Oct 14 08:41:21 crc kubenswrapper[4870]: I1014 08:41:21.446307 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nbnbr" Oct 14 08:41:21 crc kubenswrapper[4870]: I1014 08:41:21.455791 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-nbnbr"] Oct 14 08:41:21 crc kubenswrapper[4870]: I1014 08:41:21.549328 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/638a15b7-f8f5-4e0b-aff8-892dcb757a76-catalog-content\") pod \"redhat-marketplace-nbnbr\" (UID: \"638a15b7-f8f5-4e0b-aff8-892dcb757a76\") " pod="openshift-marketplace/redhat-marketplace-nbnbr" Oct 14 08:41:21 crc kubenswrapper[4870]: I1014 08:41:21.549418 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gfjtf\" (UniqueName: \"kubernetes.io/projected/638a15b7-f8f5-4e0b-aff8-892dcb757a76-kube-api-access-gfjtf\") pod \"redhat-marketplace-nbnbr\" (UID: \"638a15b7-f8f5-4e0b-aff8-892dcb757a76\") " pod="openshift-marketplace/redhat-marketplace-nbnbr" Oct 14 08:41:21 crc kubenswrapper[4870]: I1014 08:41:21.549458 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/638a15b7-f8f5-4e0b-aff8-892dcb757a76-utilities\") pod \"redhat-marketplace-nbnbr\" (UID: \"638a15b7-f8f5-4e0b-aff8-892dcb757a76\") " pod="openshift-marketplace/redhat-marketplace-nbnbr" Oct 14 08:41:21 crc kubenswrapper[4870]: I1014 08:41:21.650707 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/638a15b7-f8f5-4e0b-aff8-892dcb757a76-catalog-content\") pod \"redhat-marketplace-nbnbr\" (UID: \"638a15b7-f8f5-4e0b-aff8-892dcb757a76\") " pod="openshift-marketplace/redhat-marketplace-nbnbr" Oct 14 08:41:21 crc kubenswrapper[4870]: I1014 08:41:21.651001 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gfjtf\" (UniqueName: \"kubernetes.io/projected/638a15b7-f8f5-4e0b-aff8-892dcb757a76-kube-api-access-gfjtf\") pod \"redhat-marketplace-nbnbr\" (UID: \"638a15b7-f8f5-4e0b-aff8-892dcb757a76\") " pod="openshift-marketplace/redhat-marketplace-nbnbr" Oct 14 08:41:21 crc kubenswrapper[4870]: I1014 08:41:21.651080 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/638a15b7-f8f5-4e0b-aff8-892dcb757a76-utilities\") pod \"redhat-marketplace-nbnbr\" (UID: \"638a15b7-f8f5-4e0b-aff8-892dcb757a76\") " pod="openshift-marketplace/redhat-marketplace-nbnbr" Oct 14 08:41:21 crc kubenswrapper[4870]: I1014 08:41:21.651266 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/638a15b7-f8f5-4e0b-aff8-892dcb757a76-catalog-content\") pod \"redhat-marketplace-nbnbr\" (UID: \"638a15b7-f8f5-4e0b-aff8-892dcb757a76\") " pod="openshift-marketplace/redhat-marketplace-nbnbr" Oct 14 08:41:21 crc kubenswrapper[4870]: I1014 08:41:21.651488 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/638a15b7-f8f5-4e0b-aff8-892dcb757a76-utilities\") pod \"redhat-marketplace-nbnbr\" (UID: \"638a15b7-f8f5-4e0b-aff8-892dcb757a76\") " pod="openshift-marketplace/redhat-marketplace-nbnbr" Oct 14 08:41:21 crc kubenswrapper[4870]: I1014 08:41:21.672267 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gfjtf\" (UniqueName: \"kubernetes.io/projected/638a15b7-f8f5-4e0b-aff8-892dcb757a76-kube-api-access-gfjtf\") pod \"redhat-marketplace-nbnbr\" (UID: \"638a15b7-f8f5-4e0b-aff8-892dcb757a76\") " pod="openshift-marketplace/redhat-marketplace-nbnbr" Oct 14 08:41:21 crc kubenswrapper[4870]: I1014 08:41:21.776735 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nbnbr" Oct 14 08:41:22 crc kubenswrapper[4870]: I1014 08:41:22.222182 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-nbnbr"] Oct 14 08:41:22 crc kubenswrapper[4870]: W1014 08:41:22.237017 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod638a15b7_f8f5_4e0b_aff8_892dcb757a76.slice/crio-bde9db1d26f7c204bc91f7eaa32253ff6df2dd843889f4adab40c0d826b823a1 WatchSource:0}: Error finding container bde9db1d26f7c204bc91f7eaa32253ff6df2dd843889f4adab40c0d826b823a1: Status 404 returned error can't find the container with id bde9db1d26f7c204bc91f7eaa32253ff6df2dd843889f4adab40c0d826b823a1 Oct 14 08:41:22 crc kubenswrapper[4870]: I1014 08:41:22.957890 4870 generic.go:334] "Generic (PLEG): container finished" podID="638a15b7-f8f5-4e0b-aff8-892dcb757a76" containerID="c10d814c27d65d9b9d9b3f9225c39a0cc325090502be700ddb110bb16f0ce9a4" exitCode=0 Oct 14 08:41:22 crc kubenswrapper[4870]: I1014 08:41:22.957982 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nbnbr" event={"ID":"638a15b7-f8f5-4e0b-aff8-892dcb757a76","Type":"ContainerDied","Data":"c10d814c27d65d9b9d9b3f9225c39a0cc325090502be700ddb110bb16f0ce9a4"} Oct 14 08:41:22 crc kubenswrapper[4870]: I1014 08:41:22.958138 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nbnbr" event={"ID":"638a15b7-f8f5-4e0b-aff8-892dcb757a76","Type":"ContainerStarted","Data":"bde9db1d26f7c204bc91f7eaa32253ff6df2dd843889f4adab40c0d826b823a1"} Oct 14 08:41:22 crc kubenswrapper[4870]: I1014 08:41:22.959535 4870 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 14 08:41:23 crc kubenswrapper[4870]: I1014 08:41:23.950643 4870 patch_prober.go:28] interesting pod/machine-config-daemon-7tvc7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 08:41:23 crc kubenswrapper[4870]: I1014 08:41:23.951040 4870 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 08:41:23 crc kubenswrapper[4870]: I1014 08:41:23.951118 4870 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" Oct 14 08:41:23 crc kubenswrapper[4870]: I1014 08:41:23.953107 4870 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f9f1ebc5d4fe74cf43793adf15e4a8f7b2c05c9c750d6e220918c13c542fb541"} pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 14 08:41:23 crc kubenswrapper[4870]: I1014 08:41:23.953179 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerName="machine-config-daemon" containerID="cri-o://f9f1ebc5d4fe74cf43793adf15e4a8f7b2c05c9c750d6e220918c13c542fb541" gracePeriod=600 Oct 14 08:41:23 crc kubenswrapper[4870]: I1014 08:41:23.967417 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nbnbr" event={"ID":"638a15b7-f8f5-4e0b-aff8-892dcb757a76","Type":"ContainerStarted","Data":"9c3494475be75d4120f700756a775dde470f495b660b04bf11450801d8bb36df"} Oct 14 08:41:24 crc kubenswrapper[4870]: E1014 08:41:24.036361 4870 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod638a15b7_f8f5_4e0b_aff8_892dcb757a76.slice/crio-9c3494475be75d4120f700756a775dde470f495b660b04bf11450801d8bb36df.scope\": RecentStats: unable to find data in memory cache]" Oct 14 08:41:24 crc kubenswrapper[4870]: E1014 08:41:24.076762 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 08:41:24 crc kubenswrapper[4870]: I1014 08:41:24.982952 4870 generic.go:334] "Generic (PLEG): container finished" podID="638a15b7-f8f5-4e0b-aff8-892dcb757a76" containerID="9c3494475be75d4120f700756a775dde470f495b660b04bf11450801d8bb36df" exitCode=0 Oct 14 08:41:24 crc kubenswrapper[4870]: I1014 08:41:24.983036 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nbnbr" event={"ID":"638a15b7-f8f5-4e0b-aff8-892dcb757a76","Type":"ContainerDied","Data":"9c3494475be75d4120f700756a775dde470f495b660b04bf11450801d8bb36df"} Oct 14 08:41:24 crc kubenswrapper[4870]: I1014 08:41:24.987630 4870 generic.go:334] "Generic (PLEG): container finished" podID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerID="f9f1ebc5d4fe74cf43793adf15e4a8f7b2c05c9c750d6e220918c13c542fb541" exitCode=0 Oct 14 08:41:24 crc kubenswrapper[4870]: I1014 08:41:24.987688 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" event={"ID":"874a5e45-dffd-4d17-b609-b7d3ed2eab07","Type":"ContainerDied","Data":"f9f1ebc5d4fe74cf43793adf15e4a8f7b2c05c9c750d6e220918c13c542fb541"} Oct 14 08:41:24 crc kubenswrapper[4870]: I1014 08:41:24.987732 4870 scope.go:117] "RemoveContainer" containerID="6f0398b7e7959f65985e6ee24c4434fc79307d6cce2fb874cfc4c5f50459984b" Oct 14 08:41:24 crc kubenswrapper[4870]: I1014 08:41:24.988550 4870 scope.go:117] "RemoveContainer" containerID="f9f1ebc5d4fe74cf43793adf15e4a8f7b2c05c9c750d6e220918c13c542fb541" Oct 14 08:41:24 crc kubenswrapper[4870]: E1014 08:41:24.989046 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 08:41:25 crc kubenswrapper[4870]: I1014 08:41:25.999245 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nbnbr" event={"ID":"638a15b7-f8f5-4e0b-aff8-892dcb757a76","Type":"ContainerStarted","Data":"8455cd657fa2920db7ab3a5644281f0c773f2d4f5104294604d9e71c3ccf81d8"} Oct 14 08:41:26 crc kubenswrapper[4870]: I1014 08:41:26.028631 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-nbnbr" podStartSLOduration=2.258541366 podStartE2EDuration="5.028612245s" podCreationTimestamp="2025-10-14 08:41:21 +0000 UTC" firstStartedPulling="2025-10-14 08:41:22.959303419 +0000 UTC m=+6018.656663790" lastFinishedPulling="2025-10-14 08:41:25.729374288 +0000 UTC m=+6021.426734669" observedRunningTime="2025-10-14 08:41:26.024977376 +0000 UTC m=+6021.722337787" watchObservedRunningTime="2025-10-14 08:41:26.028612245 +0000 UTC m=+6021.725972616" Oct 14 08:41:31 crc kubenswrapper[4870]: I1014 08:41:31.777174 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-nbnbr" Oct 14 08:41:31 crc kubenswrapper[4870]: I1014 08:41:31.777893 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-nbnbr" Oct 14 08:41:31 crc kubenswrapper[4870]: I1014 08:41:31.845107 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-nbnbr" Oct 14 08:41:32 crc kubenswrapper[4870]: I1014 08:41:32.109989 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-nbnbr" Oct 14 08:41:32 crc kubenswrapper[4870]: I1014 08:41:32.164688 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-nbnbr"] Oct 14 08:41:34 crc kubenswrapper[4870]: I1014 08:41:34.077070 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-nbnbr" podUID="638a15b7-f8f5-4e0b-aff8-892dcb757a76" containerName="registry-server" containerID="cri-o://8455cd657fa2920db7ab3a5644281f0c773f2d4f5104294604d9e71c3ccf81d8" gracePeriod=2 Oct 14 08:41:34 crc kubenswrapper[4870]: E1014 08:41:34.223716 4870 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod638a15b7_f8f5_4e0b_aff8_892dcb757a76.slice/crio-8455cd657fa2920db7ab3a5644281f0c773f2d4f5104294604d9e71c3ccf81d8.scope\": RecentStats: unable to find data in memory cache]" Oct 14 08:41:34 crc kubenswrapper[4870]: I1014 08:41:34.566597 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nbnbr" Oct 14 08:41:34 crc kubenswrapper[4870]: I1014 08:41:34.689334 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/638a15b7-f8f5-4e0b-aff8-892dcb757a76-catalog-content\") pod \"638a15b7-f8f5-4e0b-aff8-892dcb757a76\" (UID: \"638a15b7-f8f5-4e0b-aff8-892dcb757a76\") " Oct 14 08:41:34 crc kubenswrapper[4870]: I1014 08:41:34.689404 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gfjtf\" (UniqueName: \"kubernetes.io/projected/638a15b7-f8f5-4e0b-aff8-892dcb757a76-kube-api-access-gfjtf\") pod \"638a15b7-f8f5-4e0b-aff8-892dcb757a76\" (UID: \"638a15b7-f8f5-4e0b-aff8-892dcb757a76\") " Oct 14 08:41:34 crc kubenswrapper[4870]: I1014 08:41:34.689715 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/638a15b7-f8f5-4e0b-aff8-892dcb757a76-utilities\") pod \"638a15b7-f8f5-4e0b-aff8-892dcb757a76\" (UID: \"638a15b7-f8f5-4e0b-aff8-892dcb757a76\") " Oct 14 08:41:34 crc kubenswrapper[4870]: I1014 08:41:34.691380 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/638a15b7-f8f5-4e0b-aff8-892dcb757a76-utilities" (OuterVolumeSpecName: "utilities") pod "638a15b7-f8f5-4e0b-aff8-892dcb757a76" (UID: "638a15b7-f8f5-4e0b-aff8-892dcb757a76"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:41:34 crc kubenswrapper[4870]: I1014 08:41:34.695949 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/638a15b7-f8f5-4e0b-aff8-892dcb757a76-kube-api-access-gfjtf" (OuterVolumeSpecName: "kube-api-access-gfjtf") pod "638a15b7-f8f5-4e0b-aff8-892dcb757a76" (UID: "638a15b7-f8f5-4e0b-aff8-892dcb757a76"). InnerVolumeSpecName "kube-api-access-gfjtf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:41:34 crc kubenswrapper[4870]: I1014 08:41:34.705455 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/638a15b7-f8f5-4e0b-aff8-892dcb757a76-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "638a15b7-f8f5-4e0b-aff8-892dcb757a76" (UID: "638a15b7-f8f5-4e0b-aff8-892dcb757a76"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:41:34 crc kubenswrapper[4870]: I1014 08:41:34.792359 4870 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/638a15b7-f8f5-4e0b-aff8-892dcb757a76-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 08:41:34 crc kubenswrapper[4870]: I1014 08:41:34.792415 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gfjtf\" (UniqueName: \"kubernetes.io/projected/638a15b7-f8f5-4e0b-aff8-892dcb757a76-kube-api-access-gfjtf\") on node \"crc\" DevicePath \"\"" Oct 14 08:41:34 crc kubenswrapper[4870]: I1014 08:41:34.792465 4870 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/638a15b7-f8f5-4e0b-aff8-892dcb757a76-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 08:41:35 crc kubenswrapper[4870]: I1014 08:41:35.086956 4870 generic.go:334] "Generic (PLEG): container finished" podID="638a15b7-f8f5-4e0b-aff8-892dcb757a76" containerID="8455cd657fa2920db7ab3a5644281f0c773f2d4f5104294604d9e71c3ccf81d8" exitCode=0 Oct 14 08:41:35 crc kubenswrapper[4870]: I1014 08:41:35.087037 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nbnbr" event={"ID":"638a15b7-f8f5-4e0b-aff8-892dcb757a76","Type":"ContainerDied","Data":"8455cd657fa2920db7ab3a5644281f0c773f2d4f5104294604d9e71c3ccf81d8"} Oct 14 08:41:35 crc kubenswrapper[4870]: I1014 08:41:35.087127 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nbnbr" event={"ID":"638a15b7-f8f5-4e0b-aff8-892dcb757a76","Type":"ContainerDied","Data":"bde9db1d26f7c204bc91f7eaa32253ff6df2dd843889f4adab40c0d826b823a1"} Oct 14 08:41:35 crc kubenswrapper[4870]: I1014 08:41:35.087167 4870 scope.go:117] "RemoveContainer" containerID="8455cd657fa2920db7ab3a5644281f0c773f2d4f5104294604d9e71c3ccf81d8" Oct 14 08:41:35 crc kubenswrapper[4870]: I1014 08:41:35.087634 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nbnbr" Oct 14 08:41:35 crc kubenswrapper[4870]: I1014 08:41:35.116001 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-nbnbr"] Oct 14 08:41:35 crc kubenswrapper[4870]: I1014 08:41:35.122613 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-nbnbr"] Oct 14 08:41:35 crc kubenswrapper[4870]: I1014 08:41:35.134325 4870 scope.go:117] "RemoveContainer" containerID="9c3494475be75d4120f700756a775dde470f495b660b04bf11450801d8bb36df" Oct 14 08:41:35 crc kubenswrapper[4870]: I1014 08:41:35.164000 4870 scope.go:117] "RemoveContainer" containerID="c10d814c27d65d9b9d9b3f9225c39a0cc325090502be700ddb110bb16f0ce9a4" Oct 14 08:41:35 crc kubenswrapper[4870]: I1014 08:41:35.197523 4870 scope.go:117] "RemoveContainer" containerID="8455cd657fa2920db7ab3a5644281f0c773f2d4f5104294604d9e71c3ccf81d8" Oct 14 08:41:35 crc kubenswrapper[4870]: E1014 08:41:35.198667 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8455cd657fa2920db7ab3a5644281f0c773f2d4f5104294604d9e71c3ccf81d8\": container with ID starting with 8455cd657fa2920db7ab3a5644281f0c773f2d4f5104294604d9e71c3ccf81d8 not found: ID does not exist" containerID="8455cd657fa2920db7ab3a5644281f0c773f2d4f5104294604d9e71c3ccf81d8" Oct 14 08:41:35 crc kubenswrapper[4870]: I1014 08:41:35.198708 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8455cd657fa2920db7ab3a5644281f0c773f2d4f5104294604d9e71c3ccf81d8"} err="failed to get container status \"8455cd657fa2920db7ab3a5644281f0c773f2d4f5104294604d9e71c3ccf81d8\": rpc error: code = NotFound desc = could not find container \"8455cd657fa2920db7ab3a5644281f0c773f2d4f5104294604d9e71c3ccf81d8\": container with ID starting with 8455cd657fa2920db7ab3a5644281f0c773f2d4f5104294604d9e71c3ccf81d8 not found: ID does not exist" Oct 14 08:41:35 crc kubenswrapper[4870]: I1014 08:41:35.198734 4870 scope.go:117] "RemoveContainer" containerID="9c3494475be75d4120f700756a775dde470f495b660b04bf11450801d8bb36df" Oct 14 08:41:35 crc kubenswrapper[4870]: E1014 08:41:35.199048 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9c3494475be75d4120f700756a775dde470f495b660b04bf11450801d8bb36df\": container with ID starting with 9c3494475be75d4120f700756a775dde470f495b660b04bf11450801d8bb36df not found: ID does not exist" containerID="9c3494475be75d4120f700756a775dde470f495b660b04bf11450801d8bb36df" Oct 14 08:41:35 crc kubenswrapper[4870]: I1014 08:41:35.199069 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9c3494475be75d4120f700756a775dde470f495b660b04bf11450801d8bb36df"} err="failed to get container status \"9c3494475be75d4120f700756a775dde470f495b660b04bf11450801d8bb36df\": rpc error: code = NotFound desc = could not find container \"9c3494475be75d4120f700756a775dde470f495b660b04bf11450801d8bb36df\": container with ID starting with 9c3494475be75d4120f700756a775dde470f495b660b04bf11450801d8bb36df not found: ID does not exist" Oct 14 08:41:35 crc kubenswrapper[4870]: I1014 08:41:35.199084 4870 scope.go:117] "RemoveContainer" containerID="c10d814c27d65d9b9d9b3f9225c39a0cc325090502be700ddb110bb16f0ce9a4" Oct 14 08:41:35 crc kubenswrapper[4870]: E1014 08:41:35.199302 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c10d814c27d65d9b9d9b3f9225c39a0cc325090502be700ddb110bb16f0ce9a4\": container with ID starting with c10d814c27d65d9b9d9b3f9225c39a0cc325090502be700ddb110bb16f0ce9a4 not found: ID does not exist" containerID="c10d814c27d65d9b9d9b3f9225c39a0cc325090502be700ddb110bb16f0ce9a4" Oct 14 08:41:35 crc kubenswrapper[4870]: I1014 08:41:35.199324 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c10d814c27d65d9b9d9b3f9225c39a0cc325090502be700ddb110bb16f0ce9a4"} err="failed to get container status \"c10d814c27d65d9b9d9b3f9225c39a0cc325090502be700ddb110bb16f0ce9a4\": rpc error: code = NotFound desc = could not find container \"c10d814c27d65d9b9d9b3f9225c39a0cc325090502be700ddb110bb16f0ce9a4\": container with ID starting with c10d814c27d65d9b9d9b3f9225c39a0cc325090502be700ddb110bb16f0ce9a4 not found: ID does not exist" Oct 14 08:41:37 crc kubenswrapper[4870]: I1014 08:41:37.044239 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="638a15b7-f8f5-4e0b-aff8-892dcb757a76" path="/var/lib/kubelet/pods/638a15b7-f8f5-4e0b-aff8-892dcb757a76/volumes" Oct 14 08:41:38 crc kubenswrapper[4870]: I1014 08:41:38.035094 4870 scope.go:117] "RemoveContainer" containerID="f9f1ebc5d4fe74cf43793adf15e4a8f7b2c05c9c750d6e220918c13c542fb541" Oct 14 08:41:38 crc kubenswrapper[4870]: E1014 08:41:38.035590 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 08:41:53 crc kubenswrapper[4870]: I1014 08:41:53.034126 4870 scope.go:117] "RemoveContainer" containerID="f9f1ebc5d4fe74cf43793adf15e4a8f7b2c05c9c750d6e220918c13c542fb541" Oct 14 08:41:53 crc kubenswrapper[4870]: E1014 08:41:53.034886 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 08:42:06 crc kubenswrapper[4870]: I1014 08:42:06.034577 4870 scope.go:117] "RemoveContainer" containerID="f9f1ebc5d4fe74cf43793adf15e4a8f7b2c05c9c750d6e220918c13c542fb541" Oct 14 08:42:06 crc kubenswrapper[4870]: E1014 08:42:06.035935 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 08:42:14 crc kubenswrapper[4870]: I1014 08:42:14.459179 4870 scope.go:117] "RemoveContainer" containerID="38057b445ce0f34a4b8f85cd8b85ce6aebbbdae879eb23826923ab6094445017" Oct 14 08:42:17 crc kubenswrapper[4870]: I1014 08:42:17.033554 4870 scope.go:117] "RemoveContainer" containerID="f9f1ebc5d4fe74cf43793adf15e4a8f7b2c05c9c750d6e220918c13c542fb541" Oct 14 08:42:17 crc kubenswrapper[4870]: E1014 08:42:17.035415 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 08:42:29 crc kubenswrapper[4870]: I1014 08:42:29.033892 4870 scope.go:117] "RemoveContainer" containerID="f9f1ebc5d4fe74cf43793adf15e4a8f7b2c05c9c750d6e220918c13c542fb541" Oct 14 08:42:29 crc kubenswrapper[4870]: E1014 08:42:29.034943 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 08:42:40 crc kubenswrapper[4870]: I1014 08:42:40.034337 4870 scope.go:117] "RemoveContainer" containerID="f9f1ebc5d4fe74cf43793adf15e4a8f7b2c05c9c750d6e220918c13c542fb541" Oct 14 08:42:40 crc kubenswrapper[4870]: E1014 08:42:40.035150 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 08:42:52 crc kubenswrapper[4870]: I1014 08:42:52.034739 4870 scope.go:117] "RemoveContainer" containerID="f9f1ebc5d4fe74cf43793adf15e4a8f7b2c05c9c750d6e220918c13c542fb541" Oct 14 08:42:52 crc kubenswrapper[4870]: E1014 08:42:52.035758 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 08:43:07 crc kubenswrapper[4870]: I1014 08:43:07.037106 4870 scope.go:117] "RemoveContainer" containerID="f9f1ebc5d4fe74cf43793adf15e4a8f7b2c05c9c750d6e220918c13c542fb541" Oct 14 08:43:07 crc kubenswrapper[4870]: E1014 08:43:07.039165 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 08:43:18 crc kubenswrapper[4870]: I1014 08:43:18.033848 4870 scope.go:117] "RemoveContainer" containerID="f9f1ebc5d4fe74cf43793adf15e4a8f7b2c05c9c750d6e220918c13c542fb541" Oct 14 08:43:18 crc kubenswrapper[4870]: E1014 08:43:18.034729 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 08:43:31 crc kubenswrapper[4870]: I1014 08:43:31.034957 4870 scope.go:117] "RemoveContainer" containerID="f9f1ebc5d4fe74cf43793adf15e4a8f7b2c05c9c750d6e220918c13c542fb541" Oct 14 08:43:31 crc kubenswrapper[4870]: E1014 08:43:31.036103 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 08:43:46 crc kubenswrapper[4870]: I1014 08:43:46.034478 4870 scope.go:117] "RemoveContainer" containerID="f9f1ebc5d4fe74cf43793adf15e4a8f7b2c05c9c750d6e220918c13c542fb541" Oct 14 08:43:46 crc kubenswrapper[4870]: E1014 08:43:46.035498 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 08:44:01 crc kubenswrapper[4870]: I1014 08:44:01.034955 4870 scope.go:117] "RemoveContainer" containerID="f9f1ebc5d4fe74cf43793adf15e4a8f7b2c05c9c750d6e220918c13c542fb541" Oct 14 08:44:01 crc kubenswrapper[4870]: E1014 08:44:01.036120 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 08:44:12 crc kubenswrapper[4870]: I1014 08:44:12.034844 4870 scope.go:117] "RemoveContainer" containerID="f9f1ebc5d4fe74cf43793adf15e4a8f7b2c05c9c750d6e220918c13c542fb541" Oct 14 08:44:12 crc kubenswrapper[4870]: E1014 08:44:12.036082 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 08:44:23 crc kubenswrapper[4870]: I1014 08:44:23.034525 4870 scope.go:117] "RemoveContainer" containerID="f9f1ebc5d4fe74cf43793adf15e4a8f7b2c05c9c750d6e220918c13c542fb541" Oct 14 08:44:23 crc kubenswrapper[4870]: E1014 08:44:23.036329 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 08:44:36 crc kubenswrapper[4870]: I1014 08:44:36.034832 4870 scope.go:117] "RemoveContainer" containerID="f9f1ebc5d4fe74cf43793adf15e4a8f7b2c05c9c750d6e220918c13c542fb541" Oct 14 08:44:36 crc kubenswrapper[4870]: E1014 08:44:36.036070 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 08:44:51 crc kubenswrapper[4870]: I1014 08:44:51.035211 4870 scope.go:117] "RemoveContainer" containerID="f9f1ebc5d4fe74cf43793adf15e4a8f7b2c05c9c750d6e220918c13c542fb541" Oct 14 08:44:51 crc kubenswrapper[4870]: E1014 08:44:51.036967 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 08:45:00 crc kubenswrapper[4870]: I1014 08:45:00.181968 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29340525-c7ww6"] Oct 14 08:45:00 crc kubenswrapper[4870]: E1014 08:45:00.183019 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="638a15b7-f8f5-4e0b-aff8-892dcb757a76" containerName="extract-content" Oct 14 08:45:00 crc kubenswrapper[4870]: I1014 08:45:00.183040 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="638a15b7-f8f5-4e0b-aff8-892dcb757a76" containerName="extract-content" Oct 14 08:45:00 crc kubenswrapper[4870]: E1014 08:45:00.183053 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="638a15b7-f8f5-4e0b-aff8-892dcb757a76" containerName="registry-server" Oct 14 08:45:00 crc kubenswrapper[4870]: I1014 08:45:00.183062 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="638a15b7-f8f5-4e0b-aff8-892dcb757a76" containerName="registry-server" Oct 14 08:45:00 crc kubenswrapper[4870]: E1014 08:45:00.183077 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="638a15b7-f8f5-4e0b-aff8-892dcb757a76" containerName="extract-utilities" Oct 14 08:45:00 crc kubenswrapper[4870]: I1014 08:45:00.183085 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="638a15b7-f8f5-4e0b-aff8-892dcb757a76" containerName="extract-utilities" Oct 14 08:45:00 crc kubenswrapper[4870]: I1014 08:45:00.183367 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="638a15b7-f8f5-4e0b-aff8-892dcb757a76" containerName="registry-server" Oct 14 08:45:00 crc kubenswrapper[4870]: I1014 08:45:00.183988 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29340525-c7ww6" Oct 14 08:45:00 crc kubenswrapper[4870]: I1014 08:45:00.185906 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 14 08:45:00 crc kubenswrapper[4870]: I1014 08:45:00.187012 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 14 08:45:00 crc kubenswrapper[4870]: I1014 08:45:00.200397 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29340525-c7ww6"] Oct 14 08:45:00 crc kubenswrapper[4870]: I1014 08:45:00.247534 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c0c03bf7-dad0-4518-bf17-08e81691880c-secret-volume\") pod \"collect-profiles-29340525-c7ww6\" (UID: \"c0c03bf7-dad0-4518-bf17-08e81691880c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340525-c7ww6" Oct 14 08:45:00 crc kubenswrapper[4870]: I1014 08:45:00.247621 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5wvfl\" (UniqueName: \"kubernetes.io/projected/c0c03bf7-dad0-4518-bf17-08e81691880c-kube-api-access-5wvfl\") pod \"collect-profiles-29340525-c7ww6\" (UID: \"c0c03bf7-dad0-4518-bf17-08e81691880c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340525-c7ww6" Oct 14 08:45:00 crc kubenswrapper[4870]: I1014 08:45:00.247966 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c0c03bf7-dad0-4518-bf17-08e81691880c-config-volume\") pod \"collect-profiles-29340525-c7ww6\" (UID: \"c0c03bf7-dad0-4518-bf17-08e81691880c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340525-c7ww6" Oct 14 08:45:00 crc kubenswrapper[4870]: I1014 08:45:00.349751 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c0c03bf7-dad0-4518-bf17-08e81691880c-config-volume\") pod \"collect-profiles-29340525-c7ww6\" (UID: \"c0c03bf7-dad0-4518-bf17-08e81691880c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340525-c7ww6" Oct 14 08:45:00 crc kubenswrapper[4870]: I1014 08:45:00.349891 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c0c03bf7-dad0-4518-bf17-08e81691880c-secret-volume\") pod \"collect-profiles-29340525-c7ww6\" (UID: \"c0c03bf7-dad0-4518-bf17-08e81691880c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340525-c7ww6" Oct 14 08:45:00 crc kubenswrapper[4870]: I1014 08:45:00.349941 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5wvfl\" (UniqueName: \"kubernetes.io/projected/c0c03bf7-dad0-4518-bf17-08e81691880c-kube-api-access-5wvfl\") pod \"collect-profiles-29340525-c7ww6\" (UID: \"c0c03bf7-dad0-4518-bf17-08e81691880c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340525-c7ww6" Oct 14 08:45:00 crc kubenswrapper[4870]: I1014 08:45:00.351223 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c0c03bf7-dad0-4518-bf17-08e81691880c-config-volume\") pod \"collect-profiles-29340525-c7ww6\" (UID: \"c0c03bf7-dad0-4518-bf17-08e81691880c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340525-c7ww6" Oct 14 08:45:00 crc kubenswrapper[4870]: I1014 08:45:00.361117 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c0c03bf7-dad0-4518-bf17-08e81691880c-secret-volume\") pod \"collect-profiles-29340525-c7ww6\" (UID: \"c0c03bf7-dad0-4518-bf17-08e81691880c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340525-c7ww6" Oct 14 08:45:00 crc kubenswrapper[4870]: I1014 08:45:00.381329 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5wvfl\" (UniqueName: \"kubernetes.io/projected/c0c03bf7-dad0-4518-bf17-08e81691880c-kube-api-access-5wvfl\") pod \"collect-profiles-29340525-c7ww6\" (UID: \"c0c03bf7-dad0-4518-bf17-08e81691880c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340525-c7ww6" Oct 14 08:45:00 crc kubenswrapper[4870]: I1014 08:45:00.512353 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29340525-c7ww6" Oct 14 08:45:01 crc kubenswrapper[4870]: I1014 08:45:01.029023 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29340525-c7ww6"] Oct 14 08:45:01 crc kubenswrapper[4870]: I1014 08:45:01.114550 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29340525-c7ww6" event={"ID":"c0c03bf7-dad0-4518-bf17-08e81691880c","Type":"ContainerStarted","Data":"8128cbb94f6913fed4bff496127d8591239dcce497032aaa9b3dec72a56d5ecb"} Oct 14 08:45:02 crc kubenswrapper[4870]: I1014 08:45:02.125205 4870 generic.go:334] "Generic (PLEG): container finished" podID="c0c03bf7-dad0-4518-bf17-08e81691880c" containerID="3d6a38840140499b75f90504af412a2dbff358a946941c4e51eab87c2bf193e0" exitCode=0 Oct 14 08:45:02 crc kubenswrapper[4870]: I1014 08:45:02.125264 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29340525-c7ww6" event={"ID":"c0c03bf7-dad0-4518-bf17-08e81691880c","Type":"ContainerDied","Data":"3d6a38840140499b75f90504af412a2dbff358a946941c4e51eab87c2bf193e0"} Oct 14 08:45:03 crc kubenswrapper[4870]: I1014 08:45:03.033642 4870 scope.go:117] "RemoveContainer" containerID="f9f1ebc5d4fe74cf43793adf15e4a8f7b2c05c9c750d6e220918c13c542fb541" Oct 14 08:45:03 crc kubenswrapper[4870]: E1014 08:45:03.034115 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 08:45:03 crc kubenswrapper[4870]: I1014 08:45:03.430372 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29340525-c7ww6" Oct 14 08:45:03 crc kubenswrapper[4870]: I1014 08:45:03.504030 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c0c03bf7-dad0-4518-bf17-08e81691880c-secret-volume\") pod \"c0c03bf7-dad0-4518-bf17-08e81691880c\" (UID: \"c0c03bf7-dad0-4518-bf17-08e81691880c\") " Oct 14 08:45:03 crc kubenswrapper[4870]: I1014 08:45:03.504147 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c0c03bf7-dad0-4518-bf17-08e81691880c-config-volume\") pod \"c0c03bf7-dad0-4518-bf17-08e81691880c\" (UID: \"c0c03bf7-dad0-4518-bf17-08e81691880c\") " Oct 14 08:45:03 crc kubenswrapper[4870]: I1014 08:45:03.504188 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5wvfl\" (UniqueName: \"kubernetes.io/projected/c0c03bf7-dad0-4518-bf17-08e81691880c-kube-api-access-5wvfl\") pod \"c0c03bf7-dad0-4518-bf17-08e81691880c\" (UID: \"c0c03bf7-dad0-4518-bf17-08e81691880c\") " Oct 14 08:45:03 crc kubenswrapper[4870]: I1014 08:45:03.504736 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c0c03bf7-dad0-4518-bf17-08e81691880c-config-volume" (OuterVolumeSpecName: "config-volume") pod "c0c03bf7-dad0-4518-bf17-08e81691880c" (UID: "c0c03bf7-dad0-4518-bf17-08e81691880c"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:45:03 crc kubenswrapper[4870]: I1014 08:45:03.509059 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c0c03bf7-dad0-4518-bf17-08e81691880c-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "c0c03bf7-dad0-4518-bf17-08e81691880c" (UID: "c0c03bf7-dad0-4518-bf17-08e81691880c"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:45:03 crc kubenswrapper[4870]: I1014 08:45:03.509185 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c0c03bf7-dad0-4518-bf17-08e81691880c-kube-api-access-5wvfl" (OuterVolumeSpecName: "kube-api-access-5wvfl") pod "c0c03bf7-dad0-4518-bf17-08e81691880c" (UID: "c0c03bf7-dad0-4518-bf17-08e81691880c"). InnerVolumeSpecName "kube-api-access-5wvfl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:45:03 crc kubenswrapper[4870]: I1014 08:45:03.605788 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5wvfl\" (UniqueName: \"kubernetes.io/projected/c0c03bf7-dad0-4518-bf17-08e81691880c-kube-api-access-5wvfl\") on node \"crc\" DevicePath \"\"" Oct 14 08:45:03 crc kubenswrapper[4870]: I1014 08:45:03.605820 4870 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c0c03bf7-dad0-4518-bf17-08e81691880c-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 14 08:45:03 crc kubenswrapper[4870]: I1014 08:45:03.605831 4870 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c0c03bf7-dad0-4518-bf17-08e81691880c-config-volume\") on node \"crc\" DevicePath \"\"" Oct 14 08:45:04 crc kubenswrapper[4870]: I1014 08:45:04.140140 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29340525-c7ww6" event={"ID":"c0c03bf7-dad0-4518-bf17-08e81691880c","Type":"ContainerDied","Data":"8128cbb94f6913fed4bff496127d8591239dcce497032aaa9b3dec72a56d5ecb"} Oct 14 08:45:04 crc kubenswrapper[4870]: I1014 08:45:04.140182 4870 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8128cbb94f6913fed4bff496127d8591239dcce497032aaa9b3dec72a56d5ecb" Oct 14 08:45:04 crc kubenswrapper[4870]: I1014 08:45:04.140221 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29340525-c7ww6" Oct 14 08:45:04 crc kubenswrapper[4870]: I1014 08:45:04.496631 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29340480-hp2hq"] Oct 14 08:45:04 crc kubenswrapper[4870]: I1014 08:45:04.508184 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29340480-hp2hq"] Oct 14 08:45:05 crc kubenswrapper[4870]: I1014 08:45:05.050717 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="db935c6f-1060-4dfa-a650-d0e894ad8a61" path="/var/lib/kubelet/pods/db935c6f-1060-4dfa-a650-d0e894ad8a61/volumes" Oct 14 08:45:14 crc kubenswrapper[4870]: I1014 08:45:14.604560 4870 scope.go:117] "RemoveContainer" containerID="747359111df3b191bff560e8b74280264f374cc938e54fbee17b306a8cb4bb9e" Oct 14 08:45:16 crc kubenswrapper[4870]: I1014 08:45:16.034797 4870 scope.go:117] "RemoveContainer" containerID="f9f1ebc5d4fe74cf43793adf15e4a8f7b2c05c9c750d6e220918c13c542fb541" Oct 14 08:45:16 crc kubenswrapper[4870]: E1014 08:45:16.035213 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 08:45:28 crc kubenswrapper[4870]: I1014 08:45:28.034426 4870 scope.go:117] "RemoveContainer" containerID="f9f1ebc5d4fe74cf43793adf15e4a8f7b2c05c9c750d6e220918c13c542fb541" Oct 14 08:45:28 crc kubenswrapper[4870]: E1014 08:45:28.035538 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 08:45:42 crc kubenswrapper[4870]: I1014 08:45:42.034559 4870 scope.go:117] "RemoveContainer" containerID="f9f1ebc5d4fe74cf43793adf15e4a8f7b2c05c9c750d6e220918c13c542fb541" Oct 14 08:45:42 crc kubenswrapper[4870]: E1014 08:45:42.035714 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 08:45:56 crc kubenswrapper[4870]: I1014 08:45:56.034223 4870 scope.go:117] "RemoveContainer" containerID="f9f1ebc5d4fe74cf43793adf15e4a8f7b2c05c9c750d6e220918c13c542fb541" Oct 14 08:45:56 crc kubenswrapper[4870]: E1014 08:45:56.035261 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 08:46:11 crc kubenswrapper[4870]: I1014 08:46:11.033770 4870 scope.go:117] "RemoveContainer" containerID="f9f1ebc5d4fe74cf43793adf15e4a8f7b2c05c9c750d6e220918c13c542fb541" Oct 14 08:46:11 crc kubenswrapper[4870]: E1014 08:46:11.034606 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 08:46:23 crc kubenswrapper[4870]: I1014 08:46:23.035194 4870 scope.go:117] "RemoveContainer" containerID="f9f1ebc5d4fe74cf43793adf15e4a8f7b2c05c9c750d6e220918c13c542fb541" Oct 14 08:46:23 crc kubenswrapper[4870]: E1014 08:46:23.036104 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 08:46:37 crc kubenswrapper[4870]: I1014 08:46:37.034755 4870 scope.go:117] "RemoveContainer" containerID="f9f1ebc5d4fe74cf43793adf15e4a8f7b2c05c9c750d6e220918c13c542fb541" Oct 14 08:46:38 crc kubenswrapper[4870]: I1014 08:46:38.033391 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" event={"ID":"874a5e45-dffd-4d17-b609-b7d3ed2eab07","Type":"ContainerStarted","Data":"39c59320b08c47c36fe9f61c79f5d172e92fda2c1c6fa3038de1fbd556880538"} Oct 14 08:47:14 crc kubenswrapper[4870]: I1014 08:47:14.673983 4870 scope.go:117] "RemoveContainer" containerID="d5d9f80caad66538a27ffe7ac3d34d1941a8068b4dba1426f38017ce3418878f" Oct 14 08:47:14 crc kubenswrapper[4870]: I1014 08:47:14.707787 4870 scope.go:117] "RemoveContainer" containerID="c21891951192a0896d0b3eb18379a84280ddba192f534243b0f1d9bb36da0128" Oct 14 08:47:14 crc kubenswrapper[4870]: I1014 08:47:14.750631 4870 scope.go:117] "RemoveContainer" containerID="cd761dbc6cd5f10b20b13d35bfc1872b207956b1b70f764e5957f1338073e111" Oct 14 08:47:14 crc kubenswrapper[4870]: I1014 08:47:14.778837 4870 scope.go:117] "RemoveContainer" containerID="136341b92ea97d5b93fbd06bb520655f22b5916e2ed329bb76c6b35b06d087e5" Oct 14 08:47:14 crc kubenswrapper[4870]: I1014 08:47:14.813764 4870 scope.go:117] "RemoveContainer" containerID="5432ef26aa9ae79f0bb86f63f580920bd3fd7ca70233298460b8419e0d4c7106" Oct 14 08:47:14 crc kubenswrapper[4870]: I1014 08:47:14.838772 4870 scope.go:117] "RemoveContainer" containerID="db99a336904c4b049dbc03b9fb8b5549dd5ace6bd3c7f587f281aede43d03bb1" Oct 14 08:47:14 crc kubenswrapper[4870]: I1014 08:47:14.869153 4870 scope.go:117] "RemoveContainer" containerID="0ee442bfaaddf52cde88d1a75c7635cf71b4a8104139f8cb3708e30b46e264a3" Oct 14 08:47:57 crc kubenswrapper[4870]: I1014 08:47:57.748263 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-f74vl"] Oct 14 08:47:57 crc kubenswrapper[4870]: E1014 08:47:57.749321 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c0c03bf7-dad0-4518-bf17-08e81691880c" containerName="collect-profiles" Oct 14 08:47:57 crc kubenswrapper[4870]: I1014 08:47:57.749337 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="c0c03bf7-dad0-4518-bf17-08e81691880c" containerName="collect-profiles" Oct 14 08:47:57 crc kubenswrapper[4870]: I1014 08:47:57.749685 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="c0c03bf7-dad0-4518-bf17-08e81691880c" containerName="collect-profiles" Oct 14 08:47:57 crc kubenswrapper[4870]: I1014 08:47:57.751634 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-f74vl" Oct 14 08:47:57 crc kubenswrapper[4870]: I1014 08:47:57.765124 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-f74vl"] Oct 14 08:47:57 crc kubenswrapper[4870]: I1014 08:47:57.849531 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/37b1f7d7-63a8-439f-b95c-8baa282ce6aa-utilities\") pod \"community-operators-f74vl\" (UID: \"37b1f7d7-63a8-439f-b95c-8baa282ce6aa\") " pod="openshift-marketplace/community-operators-f74vl" Oct 14 08:47:57 crc kubenswrapper[4870]: I1014 08:47:57.849898 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/37b1f7d7-63a8-439f-b95c-8baa282ce6aa-catalog-content\") pod \"community-operators-f74vl\" (UID: \"37b1f7d7-63a8-439f-b95c-8baa282ce6aa\") " pod="openshift-marketplace/community-operators-f74vl" Oct 14 08:47:57 crc kubenswrapper[4870]: I1014 08:47:57.849945 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s464s\" (UniqueName: \"kubernetes.io/projected/37b1f7d7-63a8-439f-b95c-8baa282ce6aa-kube-api-access-s464s\") pod \"community-operators-f74vl\" (UID: \"37b1f7d7-63a8-439f-b95c-8baa282ce6aa\") " pod="openshift-marketplace/community-operators-f74vl" Oct 14 08:47:57 crc kubenswrapper[4870]: I1014 08:47:57.951634 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/37b1f7d7-63a8-439f-b95c-8baa282ce6aa-utilities\") pod \"community-operators-f74vl\" (UID: \"37b1f7d7-63a8-439f-b95c-8baa282ce6aa\") " pod="openshift-marketplace/community-operators-f74vl" Oct 14 08:47:57 crc kubenswrapper[4870]: I1014 08:47:57.951171 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/37b1f7d7-63a8-439f-b95c-8baa282ce6aa-utilities\") pod \"community-operators-f74vl\" (UID: \"37b1f7d7-63a8-439f-b95c-8baa282ce6aa\") " pod="openshift-marketplace/community-operators-f74vl" Oct 14 08:47:57 crc kubenswrapper[4870]: I1014 08:47:57.951726 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/37b1f7d7-63a8-439f-b95c-8baa282ce6aa-catalog-content\") pod \"community-operators-f74vl\" (UID: \"37b1f7d7-63a8-439f-b95c-8baa282ce6aa\") " pod="openshift-marketplace/community-operators-f74vl" Oct 14 08:47:57 crc kubenswrapper[4870]: I1014 08:47:57.951980 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/37b1f7d7-63a8-439f-b95c-8baa282ce6aa-catalog-content\") pod \"community-operators-f74vl\" (UID: \"37b1f7d7-63a8-439f-b95c-8baa282ce6aa\") " pod="openshift-marketplace/community-operators-f74vl" Oct 14 08:47:57 crc kubenswrapper[4870]: I1014 08:47:57.952032 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s464s\" (UniqueName: \"kubernetes.io/projected/37b1f7d7-63a8-439f-b95c-8baa282ce6aa-kube-api-access-s464s\") pod \"community-operators-f74vl\" (UID: \"37b1f7d7-63a8-439f-b95c-8baa282ce6aa\") " pod="openshift-marketplace/community-operators-f74vl" Oct 14 08:47:57 crc kubenswrapper[4870]: I1014 08:47:57.975181 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s464s\" (UniqueName: \"kubernetes.io/projected/37b1f7d7-63a8-439f-b95c-8baa282ce6aa-kube-api-access-s464s\") pod \"community-operators-f74vl\" (UID: \"37b1f7d7-63a8-439f-b95c-8baa282ce6aa\") " pod="openshift-marketplace/community-operators-f74vl" Oct 14 08:47:58 crc kubenswrapper[4870]: I1014 08:47:58.087729 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-f74vl" Oct 14 08:47:58 crc kubenswrapper[4870]: I1014 08:47:58.657634 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-f74vl"] Oct 14 08:47:58 crc kubenswrapper[4870]: I1014 08:47:58.880295 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f74vl" event={"ID":"37b1f7d7-63a8-439f-b95c-8baa282ce6aa","Type":"ContainerStarted","Data":"c85addea55d8b9fe46f60a1d241eec49938fb245c0bd3053a2c991976ed67c14"} Oct 14 08:47:59 crc kubenswrapper[4870]: I1014 08:47:59.895088 4870 generic.go:334] "Generic (PLEG): container finished" podID="37b1f7d7-63a8-439f-b95c-8baa282ce6aa" containerID="93dceb834f0e8105a3ffff6ae5416beb992397a259f0178977d6247985a08015" exitCode=0 Oct 14 08:47:59 crc kubenswrapper[4870]: I1014 08:47:59.895363 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f74vl" event={"ID":"37b1f7d7-63a8-439f-b95c-8baa282ce6aa","Type":"ContainerDied","Data":"93dceb834f0e8105a3ffff6ae5416beb992397a259f0178977d6247985a08015"} Oct 14 08:47:59 crc kubenswrapper[4870]: I1014 08:47:59.898394 4870 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 14 08:48:00 crc kubenswrapper[4870]: I1014 08:48:00.908632 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f74vl" event={"ID":"37b1f7d7-63a8-439f-b95c-8baa282ce6aa","Type":"ContainerStarted","Data":"10fe8d7132f0a919fac308f1f0bdbc94c24b8830d51592378d5f2582889fd273"} Oct 14 08:48:01 crc kubenswrapper[4870]: I1014 08:48:01.926211 4870 generic.go:334] "Generic (PLEG): container finished" podID="37b1f7d7-63a8-439f-b95c-8baa282ce6aa" containerID="10fe8d7132f0a919fac308f1f0bdbc94c24b8830d51592378d5f2582889fd273" exitCode=0 Oct 14 08:48:01 crc kubenswrapper[4870]: I1014 08:48:01.926345 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f74vl" event={"ID":"37b1f7d7-63a8-439f-b95c-8baa282ce6aa","Type":"ContainerDied","Data":"10fe8d7132f0a919fac308f1f0bdbc94c24b8830d51592378d5f2582889fd273"} Oct 14 08:48:02 crc kubenswrapper[4870]: I1014 08:48:02.937117 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f74vl" event={"ID":"37b1f7d7-63a8-439f-b95c-8baa282ce6aa","Type":"ContainerStarted","Data":"0b36ff96d841ad5f50b7164827395a1ea7eed9dbd0793a2b7753a1be4994d399"} Oct 14 08:48:02 crc kubenswrapper[4870]: I1014 08:48:02.966576 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-f74vl" podStartSLOduration=3.46942353 podStartE2EDuration="5.9665618s" podCreationTimestamp="2025-10-14 08:47:57 +0000 UTC" firstStartedPulling="2025-10-14 08:47:59.897343342 +0000 UTC m=+6415.594703763" lastFinishedPulling="2025-10-14 08:48:02.394481632 +0000 UTC m=+6418.091842033" observedRunningTime="2025-10-14 08:48:02.96310762 +0000 UTC m=+6418.660467991" watchObservedRunningTime="2025-10-14 08:48:02.9665618 +0000 UTC m=+6418.663922171" Oct 14 08:48:08 crc kubenswrapper[4870]: I1014 08:48:08.088077 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-f74vl" Oct 14 08:48:08 crc kubenswrapper[4870]: I1014 08:48:08.088550 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-f74vl" Oct 14 08:48:08 crc kubenswrapper[4870]: I1014 08:48:08.159306 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-f74vl" Oct 14 08:48:09 crc kubenswrapper[4870]: I1014 08:48:09.062965 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-f74vl" Oct 14 08:48:09 crc kubenswrapper[4870]: I1014 08:48:09.127304 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-f74vl"] Oct 14 08:48:11 crc kubenswrapper[4870]: I1014 08:48:11.004041 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-f74vl" podUID="37b1f7d7-63a8-439f-b95c-8baa282ce6aa" containerName="registry-server" containerID="cri-o://0b36ff96d841ad5f50b7164827395a1ea7eed9dbd0793a2b7753a1be4994d399" gracePeriod=2 Oct 14 08:48:11 crc kubenswrapper[4870]: I1014 08:48:11.556223 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-f74vl" Oct 14 08:48:11 crc kubenswrapper[4870]: I1014 08:48:11.684338 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s464s\" (UniqueName: \"kubernetes.io/projected/37b1f7d7-63a8-439f-b95c-8baa282ce6aa-kube-api-access-s464s\") pod \"37b1f7d7-63a8-439f-b95c-8baa282ce6aa\" (UID: \"37b1f7d7-63a8-439f-b95c-8baa282ce6aa\") " Oct 14 08:48:11 crc kubenswrapper[4870]: I1014 08:48:11.684399 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/37b1f7d7-63a8-439f-b95c-8baa282ce6aa-catalog-content\") pod \"37b1f7d7-63a8-439f-b95c-8baa282ce6aa\" (UID: \"37b1f7d7-63a8-439f-b95c-8baa282ce6aa\") " Oct 14 08:48:11 crc kubenswrapper[4870]: I1014 08:48:11.684567 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/37b1f7d7-63a8-439f-b95c-8baa282ce6aa-utilities\") pod \"37b1f7d7-63a8-439f-b95c-8baa282ce6aa\" (UID: \"37b1f7d7-63a8-439f-b95c-8baa282ce6aa\") " Oct 14 08:48:11 crc kubenswrapper[4870]: I1014 08:48:11.686787 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/37b1f7d7-63a8-439f-b95c-8baa282ce6aa-utilities" (OuterVolumeSpecName: "utilities") pod "37b1f7d7-63a8-439f-b95c-8baa282ce6aa" (UID: "37b1f7d7-63a8-439f-b95c-8baa282ce6aa"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:48:11 crc kubenswrapper[4870]: I1014 08:48:11.692404 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/37b1f7d7-63a8-439f-b95c-8baa282ce6aa-kube-api-access-s464s" (OuterVolumeSpecName: "kube-api-access-s464s") pod "37b1f7d7-63a8-439f-b95c-8baa282ce6aa" (UID: "37b1f7d7-63a8-439f-b95c-8baa282ce6aa"). InnerVolumeSpecName "kube-api-access-s464s". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:48:11 crc kubenswrapper[4870]: I1014 08:48:11.770751 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/37b1f7d7-63a8-439f-b95c-8baa282ce6aa-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "37b1f7d7-63a8-439f-b95c-8baa282ce6aa" (UID: "37b1f7d7-63a8-439f-b95c-8baa282ce6aa"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:48:11 crc kubenswrapper[4870]: I1014 08:48:11.786392 4870 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/37b1f7d7-63a8-439f-b95c-8baa282ce6aa-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 08:48:11 crc kubenswrapper[4870]: I1014 08:48:11.786503 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s464s\" (UniqueName: \"kubernetes.io/projected/37b1f7d7-63a8-439f-b95c-8baa282ce6aa-kube-api-access-s464s\") on node \"crc\" DevicePath \"\"" Oct 14 08:48:11 crc kubenswrapper[4870]: I1014 08:48:11.786532 4870 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/37b1f7d7-63a8-439f-b95c-8baa282ce6aa-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 08:48:12 crc kubenswrapper[4870]: I1014 08:48:12.019721 4870 generic.go:334] "Generic (PLEG): container finished" podID="37b1f7d7-63a8-439f-b95c-8baa282ce6aa" containerID="0b36ff96d841ad5f50b7164827395a1ea7eed9dbd0793a2b7753a1be4994d399" exitCode=0 Oct 14 08:48:12 crc kubenswrapper[4870]: I1014 08:48:12.019809 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f74vl" event={"ID":"37b1f7d7-63a8-439f-b95c-8baa282ce6aa","Type":"ContainerDied","Data":"0b36ff96d841ad5f50b7164827395a1ea7eed9dbd0793a2b7753a1be4994d399"} Oct 14 08:48:12 crc kubenswrapper[4870]: I1014 08:48:12.019826 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-f74vl" Oct 14 08:48:12 crc kubenswrapper[4870]: I1014 08:48:12.019880 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f74vl" event={"ID":"37b1f7d7-63a8-439f-b95c-8baa282ce6aa","Type":"ContainerDied","Data":"c85addea55d8b9fe46f60a1d241eec49938fb245c0bd3053a2c991976ed67c14"} Oct 14 08:48:12 crc kubenswrapper[4870]: I1014 08:48:12.019925 4870 scope.go:117] "RemoveContainer" containerID="0b36ff96d841ad5f50b7164827395a1ea7eed9dbd0793a2b7753a1be4994d399" Oct 14 08:48:12 crc kubenswrapper[4870]: I1014 08:48:12.069559 4870 scope.go:117] "RemoveContainer" containerID="10fe8d7132f0a919fac308f1f0bdbc94c24b8830d51592378d5f2582889fd273" Oct 14 08:48:12 crc kubenswrapper[4870]: I1014 08:48:12.071354 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-f74vl"] Oct 14 08:48:12 crc kubenswrapper[4870]: I1014 08:48:12.083852 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-f74vl"] Oct 14 08:48:12 crc kubenswrapper[4870]: I1014 08:48:12.103917 4870 scope.go:117] "RemoveContainer" containerID="93dceb834f0e8105a3ffff6ae5416beb992397a259f0178977d6247985a08015" Oct 14 08:48:12 crc kubenswrapper[4870]: I1014 08:48:12.130201 4870 scope.go:117] "RemoveContainer" containerID="0b36ff96d841ad5f50b7164827395a1ea7eed9dbd0793a2b7753a1be4994d399" Oct 14 08:48:12 crc kubenswrapper[4870]: E1014 08:48:12.130879 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0b36ff96d841ad5f50b7164827395a1ea7eed9dbd0793a2b7753a1be4994d399\": container with ID starting with 0b36ff96d841ad5f50b7164827395a1ea7eed9dbd0793a2b7753a1be4994d399 not found: ID does not exist" containerID="0b36ff96d841ad5f50b7164827395a1ea7eed9dbd0793a2b7753a1be4994d399" Oct 14 08:48:12 crc kubenswrapper[4870]: I1014 08:48:12.130971 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0b36ff96d841ad5f50b7164827395a1ea7eed9dbd0793a2b7753a1be4994d399"} err="failed to get container status \"0b36ff96d841ad5f50b7164827395a1ea7eed9dbd0793a2b7753a1be4994d399\": rpc error: code = NotFound desc = could not find container \"0b36ff96d841ad5f50b7164827395a1ea7eed9dbd0793a2b7753a1be4994d399\": container with ID starting with 0b36ff96d841ad5f50b7164827395a1ea7eed9dbd0793a2b7753a1be4994d399 not found: ID does not exist" Oct 14 08:48:12 crc kubenswrapper[4870]: I1014 08:48:12.131030 4870 scope.go:117] "RemoveContainer" containerID="10fe8d7132f0a919fac308f1f0bdbc94c24b8830d51592378d5f2582889fd273" Oct 14 08:48:12 crc kubenswrapper[4870]: E1014 08:48:12.131566 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"10fe8d7132f0a919fac308f1f0bdbc94c24b8830d51592378d5f2582889fd273\": container with ID starting with 10fe8d7132f0a919fac308f1f0bdbc94c24b8830d51592378d5f2582889fd273 not found: ID does not exist" containerID="10fe8d7132f0a919fac308f1f0bdbc94c24b8830d51592378d5f2582889fd273" Oct 14 08:48:12 crc kubenswrapper[4870]: I1014 08:48:12.131644 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"10fe8d7132f0a919fac308f1f0bdbc94c24b8830d51592378d5f2582889fd273"} err="failed to get container status \"10fe8d7132f0a919fac308f1f0bdbc94c24b8830d51592378d5f2582889fd273\": rpc error: code = NotFound desc = could not find container \"10fe8d7132f0a919fac308f1f0bdbc94c24b8830d51592378d5f2582889fd273\": container with ID starting with 10fe8d7132f0a919fac308f1f0bdbc94c24b8830d51592378d5f2582889fd273 not found: ID does not exist" Oct 14 08:48:12 crc kubenswrapper[4870]: I1014 08:48:12.131688 4870 scope.go:117] "RemoveContainer" containerID="93dceb834f0e8105a3ffff6ae5416beb992397a259f0178977d6247985a08015" Oct 14 08:48:12 crc kubenswrapper[4870]: E1014 08:48:12.132324 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"93dceb834f0e8105a3ffff6ae5416beb992397a259f0178977d6247985a08015\": container with ID starting with 93dceb834f0e8105a3ffff6ae5416beb992397a259f0178977d6247985a08015 not found: ID does not exist" containerID="93dceb834f0e8105a3ffff6ae5416beb992397a259f0178977d6247985a08015" Oct 14 08:48:12 crc kubenswrapper[4870]: I1014 08:48:12.132602 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"93dceb834f0e8105a3ffff6ae5416beb992397a259f0178977d6247985a08015"} err="failed to get container status \"93dceb834f0e8105a3ffff6ae5416beb992397a259f0178977d6247985a08015\": rpc error: code = NotFound desc = could not find container \"93dceb834f0e8105a3ffff6ae5416beb992397a259f0178977d6247985a08015\": container with ID starting with 93dceb834f0e8105a3ffff6ae5416beb992397a259f0178977d6247985a08015 not found: ID does not exist" Oct 14 08:48:13 crc kubenswrapper[4870]: I1014 08:48:13.051672 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="37b1f7d7-63a8-439f-b95c-8baa282ce6aa" path="/var/lib/kubelet/pods/37b1f7d7-63a8-439f-b95c-8baa282ce6aa/volumes" Oct 14 08:48:53 crc kubenswrapper[4870]: I1014 08:48:53.950825 4870 patch_prober.go:28] interesting pod/machine-config-daemon-7tvc7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 08:48:53 crc kubenswrapper[4870]: I1014 08:48:53.951711 4870 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 08:49:23 crc kubenswrapper[4870]: I1014 08:49:23.950930 4870 patch_prober.go:28] interesting pod/machine-config-daemon-7tvc7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 08:49:23 crc kubenswrapper[4870]: I1014 08:49:23.951762 4870 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 08:49:53 crc kubenswrapper[4870]: I1014 08:49:53.951724 4870 patch_prober.go:28] interesting pod/machine-config-daemon-7tvc7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 08:49:53 crc kubenswrapper[4870]: I1014 08:49:53.952423 4870 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 08:49:53 crc kubenswrapper[4870]: I1014 08:49:53.952534 4870 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" Oct 14 08:49:53 crc kubenswrapper[4870]: I1014 08:49:53.953381 4870 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"39c59320b08c47c36fe9f61c79f5d172e92fda2c1c6fa3038de1fbd556880538"} pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 14 08:49:53 crc kubenswrapper[4870]: I1014 08:49:53.953523 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerName="machine-config-daemon" containerID="cri-o://39c59320b08c47c36fe9f61c79f5d172e92fda2c1c6fa3038de1fbd556880538" gracePeriod=600 Oct 14 08:49:55 crc kubenswrapper[4870]: I1014 08:49:55.011001 4870 generic.go:334] "Generic (PLEG): container finished" podID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerID="39c59320b08c47c36fe9f61c79f5d172e92fda2c1c6fa3038de1fbd556880538" exitCode=0 Oct 14 08:49:55 crc kubenswrapper[4870]: I1014 08:49:55.011086 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" event={"ID":"874a5e45-dffd-4d17-b609-b7d3ed2eab07","Type":"ContainerDied","Data":"39c59320b08c47c36fe9f61c79f5d172e92fda2c1c6fa3038de1fbd556880538"} Oct 14 08:49:55 crc kubenswrapper[4870]: I1014 08:49:55.011716 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" event={"ID":"874a5e45-dffd-4d17-b609-b7d3ed2eab07","Type":"ContainerStarted","Data":"8c0f754fe94b7cb7b9bb5dcd613f7e42093968b49509d8ee37e8fd4db89eedd5"} Oct 14 08:49:55 crc kubenswrapper[4870]: I1014 08:49:55.011770 4870 scope.go:117] "RemoveContainer" containerID="f9f1ebc5d4fe74cf43793adf15e4a8f7b2c05c9c750d6e220918c13c542fb541" Oct 14 08:50:40 crc kubenswrapper[4870]: I1014 08:50:40.122858 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-copy-data"] Oct 14 08:50:40 crc kubenswrapper[4870]: E1014 08:50:40.124867 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37b1f7d7-63a8-439f-b95c-8baa282ce6aa" containerName="extract-content" Oct 14 08:50:40 crc kubenswrapper[4870]: I1014 08:50:40.124967 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="37b1f7d7-63a8-439f-b95c-8baa282ce6aa" containerName="extract-content" Oct 14 08:50:40 crc kubenswrapper[4870]: E1014 08:50:40.125069 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37b1f7d7-63a8-439f-b95c-8baa282ce6aa" containerName="registry-server" Oct 14 08:50:40 crc kubenswrapper[4870]: I1014 08:50:40.125145 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="37b1f7d7-63a8-439f-b95c-8baa282ce6aa" containerName="registry-server" Oct 14 08:50:40 crc kubenswrapper[4870]: E1014 08:50:40.125229 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37b1f7d7-63a8-439f-b95c-8baa282ce6aa" containerName="extract-utilities" Oct 14 08:50:40 crc kubenswrapper[4870]: I1014 08:50:40.125301 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="37b1f7d7-63a8-439f-b95c-8baa282ce6aa" containerName="extract-utilities" Oct 14 08:50:40 crc kubenswrapper[4870]: I1014 08:50:40.125568 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="37b1f7d7-63a8-439f-b95c-8baa282ce6aa" containerName="registry-server" Oct 14 08:50:40 crc kubenswrapper[4870]: I1014 08:50:40.126248 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-copy-data" Oct 14 08:50:40 crc kubenswrapper[4870]: I1014 08:50:40.128304 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-vnjjw" Oct 14 08:50:40 crc kubenswrapper[4870]: I1014 08:50:40.129869 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-copy-data"] Oct 14 08:50:40 crc kubenswrapper[4870]: I1014 08:50:40.249409 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qlct8\" (UniqueName: \"kubernetes.io/projected/592ad2e9-ae45-4985-92f8-da4f3a76b5cb-kube-api-access-qlct8\") pod \"mariadb-copy-data\" (UID: \"592ad2e9-ae45-4985-92f8-da4f3a76b5cb\") " pod="openstack/mariadb-copy-data" Oct 14 08:50:40 crc kubenswrapper[4870]: I1014 08:50:40.249503 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-2cf7c6e8-938d-4c3b-9b5b-7b262e943b94\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2cf7c6e8-938d-4c3b-9b5b-7b262e943b94\") pod \"mariadb-copy-data\" (UID: \"592ad2e9-ae45-4985-92f8-da4f3a76b5cb\") " pod="openstack/mariadb-copy-data" Oct 14 08:50:40 crc kubenswrapper[4870]: I1014 08:50:40.350688 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qlct8\" (UniqueName: \"kubernetes.io/projected/592ad2e9-ae45-4985-92f8-da4f3a76b5cb-kube-api-access-qlct8\") pod \"mariadb-copy-data\" (UID: \"592ad2e9-ae45-4985-92f8-da4f3a76b5cb\") " pod="openstack/mariadb-copy-data" Oct 14 08:50:40 crc kubenswrapper[4870]: I1014 08:50:40.350984 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-2cf7c6e8-938d-4c3b-9b5b-7b262e943b94\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2cf7c6e8-938d-4c3b-9b5b-7b262e943b94\") pod \"mariadb-copy-data\" (UID: \"592ad2e9-ae45-4985-92f8-da4f3a76b5cb\") " pod="openstack/mariadb-copy-data" Oct 14 08:50:40 crc kubenswrapper[4870]: I1014 08:50:40.354962 4870 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 14 08:50:40 crc kubenswrapper[4870]: I1014 08:50:40.355023 4870 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-2cf7c6e8-938d-4c3b-9b5b-7b262e943b94\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2cf7c6e8-938d-4c3b-9b5b-7b262e943b94\") pod \"mariadb-copy-data\" (UID: \"592ad2e9-ae45-4985-92f8-da4f3a76b5cb\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/ab01c9fdc58f5754794fe09a5dd4097a9bf664365d2765dacbaa302ebadde619/globalmount\"" pod="openstack/mariadb-copy-data" Oct 14 08:50:40 crc kubenswrapper[4870]: I1014 08:50:40.383650 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qlct8\" (UniqueName: \"kubernetes.io/projected/592ad2e9-ae45-4985-92f8-da4f3a76b5cb-kube-api-access-qlct8\") pod \"mariadb-copy-data\" (UID: \"592ad2e9-ae45-4985-92f8-da4f3a76b5cb\") " pod="openstack/mariadb-copy-data" Oct 14 08:50:40 crc kubenswrapper[4870]: I1014 08:50:40.405382 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-2cf7c6e8-938d-4c3b-9b5b-7b262e943b94\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2cf7c6e8-938d-4c3b-9b5b-7b262e943b94\") pod \"mariadb-copy-data\" (UID: \"592ad2e9-ae45-4985-92f8-da4f3a76b5cb\") " pod="openstack/mariadb-copy-data" Oct 14 08:50:40 crc kubenswrapper[4870]: I1014 08:50:40.445128 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-copy-data" Oct 14 08:50:41 crc kubenswrapper[4870]: I1014 08:50:41.012051 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-copy-data"] Oct 14 08:50:41 crc kubenswrapper[4870]: I1014 08:50:41.441427 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-copy-data" event={"ID":"592ad2e9-ae45-4985-92f8-da4f3a76b5cb","Type":"ContainerStarted","Data":"3dd253c3e0fa1fe373307af5bdeb4a9a5e12f8ee948dd1bc28a8e0c134af8acd"} Oct 14 08:50:41 crc kubenswrapper[4870]: I1014 08:50:41.441510 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-copy-data" event={"ID":"592ad2e9-ae45-4985-92f8-da4f3a76b5cb","Type":"ContainerStarted","Data":"87cf07d511e054223dd4255a5b936e615fb5a362a3619fb51149bd78ff00a135"} Oct 14 08:50:41 crc kubenswrapper[4870]: I1014 08:50:41.463419 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/mariadb-copy-data" podStartSLOduration=2.463395858 podStartE2EDuration="2.463395858s" podCreationTimestamp="2025-10-14 08:50:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 08:50:41.458203783 +0000 UTC m=+6577.155564174" watchObservedRunningTime="2025-10-14 08:50:41.463395858 +0000 UTC m=+6577.160756239" Oct 14 08:50:43 crc kubenswrapper[4870]: I1014 08:50:43.859034 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client"] Oct 14 08:50:43 crc kubenswrapper[4870]: I1014 08:50:43.860591 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Oct 14 08:50:43 crc kubenswrapper[4870]: I1014 08:50:43.869592 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client"] Oct 14 08:50:44 crc kubenswrapper[4870]: I1014 08:50:44.013232 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xxhzq\" (UniqueName: \"kubernetes.io/projected/996eb3e2-befd-4bc9-aadc-a699da548f7d-kube-api-access-xxhzq\") pod \"mariadb-client\" (UID: \"996eb3e2-befd-4bc9-aadc-a699da548f7d\") " pod="openstack/mariadb-client" Oct 14 08:50:44 crc kubenswrapper[4870]: I1014 08:50:44.115818 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xxhzq\" (UniqueName: \"kubernetes.io/projected/996eb3e2-befd-4bc9-aadc-a699da548f7d-kube-api-access-xxhzq\") pod \"mariadb-client\" (UID: \"996eb3e2-befd-4bc9-aadc-a699da548f7d\") " pod="openstack/mariadb-client" Oct 14 08:50:44 crc kubenswrapper[4870]: I1014 08:50:44.153953 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xxhzq\" (UniqueName: \"kubernetes.io/projected/996eb3e2-befd-4bc9-aadc-a699da548f7d-kube-api-access-xxhzq\") pod \"mariadb-client\" (UID: \"996eb3e2-befd-4bc9-aadc-a699da548f7d\") " pod="openstack/mariadb-client" Oct 14 08:50:44 crc kubenswrapper[4870]: I1014 08:50:44.181924 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Oct 14 08:50:44 crc kubenswrapper[4870]: I1014 08:50:44.715826 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client"] Oct 14 08:50:44 crc kubenswrapper[4870]: W1014 08:50:44.719612 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod996eb3e2_befd_4bc9_aadc_a699da548f7d.slice/crio-001d2e781129f1385b3589f1aed7793d1e3a18f1dee83ab754811c6c573fc34e WatchSource:0}: Error finding container 001d2e781129f1385b3589f1aed7793d1e3a18f1dee83ab754811c6c573fc34e: Status 404 returned error can't find the container with id 001d2e781129f1385b3589f1aed7793d1e3a18f1dee83ab754811c6c573fc34e Oct 14 08:50:45 crc kubenswrapper[4870]: I1014 08:50:45.509952 4870 generic.go:334] "Generic (PLEG): container finished" podID="996eb3e2-befd-4bc9-aadc-a699da548f7d" containerID="487e6147b8e1d9521158b2f483a47b2101c4c07eebb3475f7cbcfa1dff9d79c6" exitCode=0 Oct 14 08:50:45 crc kubenswrapper[4870]: I1014 08:50:45.511890 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client" event={"ID":"996eb3e2-befd-4bc9-aadc-a699da548f7d","Type":"ContainerDied","Data":"487e6147b8e1d9521158b2f483a47b2101c4c07eebb3475f7cbcfa1dff9d79c6"} Oct 14 08:50:45 crc kubenswrapper[4870]: I1014 08:50:45.511948 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client" event={"ID":"996eb3e2-befd-4bc9-aadc-a699da548f7d","Type":"ContainerStarted","Data":"001d2e781129f1385b3589f1aed7793d1e3a18f1dee83ab754811c6c573fc34e"} Oct 14 08:50:46 crc kubenswrapper[4870]: I1014 08:50:46.810291 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Oct 14 08:50:46 crc kubenswrapper[4870]: I1014 08:50:46.876708 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xxhzq\" (UniqueName: \"kubernetes.io/projected/996eb3e2-befd-4bc9-aadc-a699da548f7d-kube-api-access-xxhzq\") pod \"996eb3e2-befd-4bc9-aadc-a699da548f7d\" (UID: \"996eb3e2-befd-4bc9-aadc-a699da548f7d\") " Oct 14 08:50:46 crc kubenswrapper[4870]: I1014 08:50:46.878331 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client_996eb3e2-befd-4bc9-aadc-a699da548f7d/mariadb-client/0.log" Oct 14 08:50:46 crc kubenswrapper[4870]: I1014 08:50:46.886790 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/996eb3e2-befd-4bc9-aadc-a699da548f7d-kube-api-access-xxhzq" (OuterVolumeSpecName: "kube-api-access-xxhzq") pod "996eb3e2-befd-4bc9-aadc-a699da548f7d" (UID: "996eb3e2-befd-4bc9-aadc-a699da548f7d"). InnerVolumeSpecName "kube-api-access-xxhzq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:50:46 crc kubenswrapper[4870]: I1014 08:50:46.917308 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client"] Oct 14 08:50:46 crc kubenswrapper[4870]: I1014 08:50:46.925512 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client"] Oct 14 08:50:46 crc kubenswrapper[4870]: I1014 08:50:46.979121 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xxhzq\" (UniqueName: \"kubernetes.io/projected/996eb3e2-befd-4bc9-aadc-a699da548f7d-kube-api-access-xxhzq\") on node \"crc\" DevicePath \"\"" Oct 14 08:50:47 crc kubenswrapper[4870]: I1014 08:50:47.051137 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="996eb3e2-befd-4bc9-aadc-a699da548f7d" path="/var/lib/kubelet/pods/996eb3e2-befd-4bc9-aadc-a699da548f7d/volumes" Oct 14 08:50:47 crc kubenswrapper[4870]: I1014 08:50:47.062031 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-wdr54"] Oct 14 08:50:47 crc kubenswrapper[4870]: E1014 08:50:47.062762 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="996eb3e2-befd-4bc9-aadc-a699da548f7d" containerName="mariadb-client" Oct 14 08:50:47 crc kubenswrapper[4870]: I1014 08:50:47.062787 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="996eb3e2-befd-4bc9-aadc-a699da548f7d" containerName="mariadb-client" Oct 14 08:50:47 crc kubenswrapper[4870]: I1014 08:50:47.063018 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="996eb3e2-befd-4bc9-aadc-a699da548f7d" containerName="mariadb-client" Oct 14 08:50:47 crc kubenswrapper[4870]: I1014 08:50:47.065891 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-wdr54"] Oct 14 08:50:47 crc kubenswrapper[4870]: I1014 08:50:47.066050 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wdr54" Oct 14 08:50:47 crc kubenswrapper[4870]: I1014 08:50:47.082430 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client"] Oct 14 08:50:47 crc kubenswrapper[4870]: I1014 08:50:47.083952 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Oct 14 08:50:47 crc kubenswrapper[4870]: I1014 08:50:47.111693 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client"] Oct 14 08:50:47 crc kubenswrapper[4870]: I1014 08:50:47.187285 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/32664967-2bce-429d-bb04-ebd14b9f03f5-utilities\") pod \"certified-operators-wdr54\" (UID: \"32664967-2bce-429d-bb04-ebd14b9f03f5\") " pod="openshift-marketplace/certified-operators-wdr54" Oct 14 08:50:47 crc kubenswrapper[4870]: I1014 08:50:47.187410 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vmjz2\" (UniqueName: \"kubernetes.io/projected/32664967-2bce-429d-bb04-ebd14b9f03f5-kube-api-access-vmjz2\") pod \"certified-operators-wdr54\" (UID: \"32664967-2bce-429d-bb04-ebd14b9f03f5\") " pod="openshift-marketplace/certified-operators-wdr54" Oct 14 08:50:47 crc kubenswrapper[4870]: I1014 08:50:47.187525 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rvkkg\" (UniqueName: \"kubernetes.io/projected/80e92bad-0578-41c9-b087-89f938be53a5-kube-api-access-rvkkg\") pod \"mariadb-client\" (UID: \"80e92bad-0578-41c9-b087-89f938be53a5\") " pod="openstack/mariadb-client" Oct 14 08:50:47 crc kubenswrapper[4870]: I1014 08:50:47.187550 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/32664967-2bce-429d-bb04-ebd14b9f03f5-catalog-content\") pod \"certified-operators-wdr54\" (UID: \"32664967-2bce-429d-bb04-ebd14b9f03f5\") " pod="openshift-marketplace/certified-operators-wdr54" Oct 14 08:50:47 crc kubenswrapper[4870]: I1014 08:50:47.289530 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rvkkg\" (UniqueName: \"kubernetes.io/projected/80e92bad-0578-41c9-b087-89f938be53a5-kube-api-access-rvkkg\") pod \"mariadb-client\" (UID: \"80e92bad-0578-41c9-b087-89f938be53a5\") " pod="openstack/mariadb-client" Oct 14 08:50:47 crc kubenswrapper[4870]: I1014 08:50:47.289967 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/32664967-2bce-429d-bb04-ebd14b9f03f5-catalog-content\") pod \"certified-operators-wdr54\" (UID: \"32664967-2bce-429d-bb04-ebd14b9f03f5\") " pod="openshift-marketplace/certified-operators-wdr54" Oct 14 08:50:47 crc kubenswrapper[4870]: I1014 08:50:47.290695 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/32664967-2bce-429d-bb04-ebd14b9f03f5-utilities\") pod \"certified-operators-wdr54\" (UID: \"32664967-2bce-429d-bb04-ebd14b9f03f5\") " pod="openshift-marketplace/certified-operators-wdr54" Oct 14 08:50:47 crc kubenswrapper[4870]: I1014 08:50:47.290780 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vmjz2\" (UniqueName: \"kubernetes.io/projected/32664967-2bce-429d-bb04-ebd14b9f03f5-kube-api-access-vmjz2\") pod \"certified-operators-wdr54\" (UID: \"32664967-2bce-429d-bb04-ebd14b9f03f5\") " pod="openshift-marketplace/certified-operators-wdr54" Oct 14 08:50:47 crc kubenswrapper[4870]: I1014 08:50:47.291204 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/32664967-2bce-429d-bb04-ebd14b9f03f5-utilities\") pod \"certified-operators-wdr54\" (UID: \"32664967-2bce-429d-bb04-ebd14b9f03f5\") " pod="openshift-marketplace/certified-operators-wdr54" Oct 14 08:50:47 crc kubenswrapper[4870]: I1014 08:50:47.291575 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/32664967-2bce-429d-bb04-ebd14b9f03f5-catalog-content\") pod \"certified-operators-wdr54\" (UID: \"32664967-2bce-429d-bb04-ebd14b9f03f5\") " pod="openshift-marketplace/certified-operators-wdr54" Oct 14 08:50:47 crc kubenswrapper[4870]: I1014 08:50:47.310731 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vmjz2\" (UniqueName: \"kubernetes.io/projected/32664967-2bce-429d-bb04-ebd14b9f03f5-kube-api-access-vmjz2\") pod \"certified-operators-wdr54\" (UID: \"32664967-2bce-429d-bb04-ebd14b9f03f5\") " pod="openshift-marketplace/certified-operators-wdr54" Oct 14 08:50:47 crc kubenswrapper[4870]: I1014 08:50:47.312278 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rvkkg\" (UniqueName: \"kubernetes.io/projected/80e92bad-0578-41c9-b087-89f938be53a5-kube-api-access-rvkkg\") pod \"mariadb-client\" (UID: \"80e92bad-0578-41c9-b087-89f938be53a5\") " pod="openstack/mariadb-client" Oct 14 08:50:47 crc kubenswrapper[4870]: I1014 08:50:47.395563 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wdr54" Oct 14 08:50:47 crc kubenswrapper[4870]: I1014 08:50:47.407053 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Oct 14 08:50:47 crc kubenswrapper[4870]: I1014 08:50:47.537491 4870 scope.go:117] "RemoveContainer" containerID="487e6147b8e1d9521158b2f483a47b2101c4c07eebb3475f7cbcfa1dff9d79c6" Oct 14 08:50:47 crc kubenswrapper[4870]: I1014 08:50:47.537549 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Oct 14 08:50:47 crc kubenswrapper[4870]: I1014 08:50:47.914374 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-wdr54"] Oct 14 08:50:48 crc kubenswrapper[4870]: I1014 08:50:48.044124 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client"] Oct 14 08:50:48 crc kubenswrapper[4870]: I1014 08:50:48.548997 4870 generic.go:334] "Generic (PLEG): container finished" podID="32664967-2bce-429d-bb04-ebd14b9f03f5" containerID="c9b4104ff3ca2a1b1e28c7dd0deb488fdd313a06645aa94ff88804a992daf2f5" exitCode=0 Oct 14 08:50:48 crc kubenswrapper[4870]: I1014 08:50:48.549140 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wdr54" event={"ID":"32664967-2bce-429d-bb04-ebd14b9f03f5","Type":"ContainerDied","Data":"c9b4104ff3ca2a1b1e28c7dd0deb488fdd313a06645aa94ff88804a992daf2f5"} Oct 14 08:50:48 crc kubenswrapper[4870]: I1014 08:50:48.549182 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wdr54" event={"ID":"32664967-2bce-429d-bb04-ebd14b9f03f5","Type":"ContainerStarted","Data":"551cd2db06b3448d1656fb3eb4ffdfa6434b52a9236c50e25cf4e60b69cdd30b"} Oct 14 08:50:48 crc kubenswrapper[4870]: I1014 08:50:48.564628 4870 generic.go:334] "Generic (PLEG): container finished" podID="80e92bad-0578-41c9-b087-89f938be53a5" containerID="2df76e3cfba0d4fc310c83933f2a1ddf28afd200614e7727c2c55f61c7cf17e2" exitCode=0 Oct 14 08:50:48 crc kubenswrapper[4870]: I1014 08:50:48.564802 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client" event={"ID":"80e92bad-0578-41c9-b087-89f938be53a5","Type":"ContainerDied","Data":"2df76e3cfba0d4fc310c83933f2a1ddf28afd200614e7727c2c55f61c7cf17e2"} Oct 14 08:50:48 crc kubenswrapper[4870]: I1014 08:50:48.564924 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client" event={"ID":"80e92bad-0578-41c9-b087-89f938be53a5","Type":"ContainerStarted","Data":"a27c804bb675f3995bd5e9db534f717fbc4f496aa9b2bb3b703a68b349adbec5"} Oct 14 08:50:49 crc kubenswrapper[4870]: I1014 08:50:49.578569 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wdr54" event={"ID":"32664967-2bce-429d-bb04-ebd14b9f03f5","Type":"ContainerStarted","Data":"79afcd8a0aa446b0dd1e7ae18c6ff3892eb6e5942e4e94b180e1283eeeeef384"} Oct 14 08:50:49 crc kubenswrapper[4870]: I1014 08:50:49.943027 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Oct 14 08:50:49 crc kubenswrapper[4870]: I1014 08:50:49.971672 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client_80e92bad-0578-41c9-b087-89f938be53a5/mariadb-client/0.log" Oct 14 08:50:49 crc kubenswrapper[4870]: I1014 08:50:49.996498 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client"] Oct 14 08:50:50 crc kubenswrapper[4870]: I1014 08:50:50.002003 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client"] Oct 14 08:50:50 crc kubenswrapper[4870]: I1014 08:50:50.036696 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rvkkg\" (UniqueName: \"kubernetes.io/projected/80e92bad-0578-41c9-b087-89f938be53a5-kube-api-access-rvkkg\") pod \"80e92bad-0578-41c9-b087-89f938be53a5\" (UID: \"80e92bad-0578-41c9-b087-89f938be53a5\") " Oct 14 08:50:50 crc kubenswrapper[4870]: I1014 08:50:50.047684 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/80e92bad-0578-41c9-b087-89f938be53a5-kube-api-access-rvkkg" (OuterVolumeSpecName: "kube-api-access-rvkkg") pod "80e92bad-0578-41c9-b087-89f938be53a5" (UID: "80e92bad-0578-41c9-b087-89f938be53a5"). InnerVolumeSpecName "kube-api-access-rvkkg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:50:50 crc kubenswrapper[4870]: I1014 08:50:50.138810 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rvkkg\" (UniqueName: \"kubernetes.io/projected/80e92bad-0578-41c9-b087-89f938be53a5-kube-api-access-rvkkg\") on node \"crc\" DevicePath \"\"" Oct 14 08:50:50 crc kubenswrapper[4870]: I1014 08:50:50.588409 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Oct 14 08:50:50 crc kubenswrapper[4870]: I1014 08:50:50.588495 4870 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a27c804bb675f3995bd5e9db534f717fbc4f496aa9b2bb3b703a68b349adbec5" Oct 14 08:50:50 crc kubenswrapper[4870]: I1014 08:50:50.590105 4870 generic.go:334] "Generic (PLEG): container finished" podID="32664967-2bce-429d-bb04-ebd14b9f03f5" containerID="79afcd8a0aa446b0dd1e7ae18c6ff3892eb6e5942e4e94b180e1283eeeeef384" exitCode=0 Oct 14 08:50:50 crc kubenswrapper[4870]: I1014 08:50:50.590153 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wdr54" event={"ID":"32664967-2bce-429d-bb04-ebd14b9f03f5","Type":"ContainerDied","Data":"79afcd8a0aa446b0dd1e7ae18c6ff3892eb6e5942e4e94b180e1283eeeeef384"} Oct 14 08:50:51 crc kubenswrapper[4870]: I1014 08:50:51.064756 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="80e92bad-0578-41c9-b087-89f938be53a5" path="/var/lib/kubelet/pods/80e92bad-0578-41c9-b087-89f938be53a5/volumes" Oct 14 08:50:51 crc kubenswrapper[4870]: I1014 08:50:51.602023 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wdr54" event={"ID":"32664967-2bce-429d-bb04-ebd14b9f03f5","Type":"ContainerStarted","Data":"a30d157fe7e16e953bf51aa6c962ee186b3cae92e8792a4a64a4d1c4c6f120be"} Oct 14 08:50:51 crc kubenswrapper[4870]: I1014 08:50:51.632271 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-wdr54" podStartSLOduration=2.140893981 podStartE2EDuration="4.632246912s" podCreationTimestamp="2025-10-14 08:50:47 +0000 UTC" firstStartedPulling="2025-10-14 08:50:48.551334202 +0000 UTC m=+6584.248694603" lastFinishedPulling="2025-10-14 08:50:51.042687123 +0000 UTC m=+6586.740047534" observedRunningTime="2025-10-14 08:50:51.627605602 +0000 UTC m=+6587.324966013" watchObservedRunningTime="2025-10-14 08:50:51.632246912 +0000 UTC m=+6587.329607303" Oct 14 08:50:57 crc kubenswrapper[4870]: I1014 08:50:57.396525 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-wdr54" Oct 14 08:50:57 crc kubenswrapper[4870]: I1014 08:50:57.397095 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-wdr54" Oct 14 08:50:57 crc kubenswrapper[4870]: I1014 08:50:57.462399 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-wdr54" Oct 14 08:50:57 crc kubenswrapper[4870]: I1014 08:50:57.715669 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-wdr54" Oct 14 08:50:57 crc kubenswrapper[4870]: I1014 08:50:57.757670 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-wdr54"] Oct 14 08:50:59 crc kubenswrapper[4870]: I1014 08:50:59.674498 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-wdr54" podUID="32664967-2bce-429d-bb04-ebd14b9f03f5" containerName="registry-server" containerID="cri-o://a30d157fe7e16e953bf51aa6c962ee186b3cae92e8792a4a64a4d1c4c6f120be" gracePeriod=2 Oct 14 08:51:00 crc kubenswrapper[4870]: I1014 08:51:00.104551 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wdr54" Oct 14 08:51:00 crc kubenswrapper[4870]: I1014 08:51:00.197670 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vmjz2\" (UniqueName: \"kubernetes.io/projected/32664967-2bce-429d-bb04-ebd14b9f03f5-kube-api-access-vmjz2\") pod \"32664967-2bce-429d-bb04-ebd14b9f03f5\" (UID: \"32664967-2bce-429d-bb04-ebd14b9f03f5\") " Oct 14 08:51:00 crc kubenswrapper[4870]: I1014 08:51:00.197788 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/32664967-2bce-429d-bb04-ebd14b9f03f5-utilities\") pod \"32664967-2bce-429d-bb04-ebd14b9f03f5\" (UID: \"32664967-2bce-429d-bb04-ebd14b9f03f5\") " Oct 14 08:51:00 crc kubenswrapper[4870]: I1014 08:51:00.197986 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/32664967-2bce-429d-bb04-ebd14b9f03f5-catalog-content\") pod \"32664967-2bce-429d-bb04-ebd14b9f03f5\" (UID: \"32664967-2bce-429d-bb04-ebd14b9f03f5\") " Oct 14 08:51:00 crc kubenswrapper[4870]: I1014 08:51:00.198631 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/32664967-2bce-429d-bb04-ebd14b9f03f5-utilities" (OuterVolumeSpecName: "utilities") pod "32664967-2bce-429d-bb04-ebd14b9f03f5" (UID: "32664967-2bce-429d-bb04-ebd14b9f03f5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:51:00 crc kubenswrapper[4870]: I1014 08:51:00.204293 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/32664967-2bce-429d-bb04-ebd14b9f03f5-kube-api-access-vmjz2" (OuterVolumeSpecName: "kube-api-access-vmjz2") pod "32664967-2bce-429d-bb04-ebd14b9f03f5" (UID: "32664967-2bce-429d-bb04-ebd14b9f03f5"). InnerVolumeSpecName "kube-api-access-vmjz2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:51:00 crc kubenswrapper[4870]: I1014 08:51:00.243091 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/32664967-2bce-429d-bb04-ebd14b9f03f5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "32664967-2bce-429d-bb04-ebd14b9f03f5" (UID: "32664967-2bce-429d-bb04-ebd14b9f03f5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:51:00 crc kubenswrapper[4870]: I1014 08:51:00.299665 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vmjz2\" (UniqueName: \"kubernetes.io/projected/32664967-2bce-429d-bb04-ebd14b9f03f5-kube-api-access-vmjz2\") on node \"crc\" DevicePath \"\"" Oct 14 08:51:00 crc kubenswrapper[4870]: I1014 08:51:00.299714 4870 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/32664967-2bce-429d-bb04-ebd14b9f03f5-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 08:51:00 crc kubenswrapper[4870]: I1014 08:51:00.299732 4870 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/32664967-2bce-429d-bb04-ebd14b9f03f5-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 08:51:00 crc kubenswrapper[4870]: I1014 08:51:00.685559 4870 generic.go:334] "Generic (PLEG): container finished" podID="32664967-2bce-429d-bb04-ebd14b9f03f5" containerID="a30d157fe7e16e953bf51aa6c962ee186b3cae92e8792a4a64a4d1c4c6f120be" exitCode=0 Oct 14 08:51:00 crc kubenswrapper[4870]: I1014 08:51:00.685620 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wdr54" event={"ID":"32664967-2bce-429d-bb04-ebd14b9f03f5","Type":"ContainerDied","Data":"a30d157fe7e16e953bf51aa6c962ee186b3cae92e8792a4a64a4d1c4c6f120be"} Oct 14 08:51:00 crc kubenswrapper[4870]: I1014 08:51:00.685657 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wdr54" event={"ID":"32664967-2bce-429d-bb04-ebd14b9f03f5","Type":"ContainerDied","Data":"551cd2db06b3448d1656fb3eb4ffdfa6434b52a9236c50e25cf4e60b69cdd30b"} Oct 14 08:51:00 crc kubenswrapper[4870]: I1014 08:51:00.685686 4870 scope.go:117] "RemoveContainer" containerID="a30d157fe7e16e953bf51aa6c962ee186b3cae92e8792a4a64a4d1c4c6f120be" Oct 14 08:51:00 crc kubenswrapper[4870]: I1014 08:51:00.685863 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wdr54" Oct 14 08:51:00 crc kubenswrapper[4870]: I1014 08:51:00.737129 4870 scope.go:117] "RemoveContainer" containerID="79afcd8a0aa446b0dd1e7ae18c6ff3892eb6e5942e4e94b180e1283eeeeef384" Oct 14 08:51:00 crc kubenswrapper[4870]: I1014 08:51:00.744790 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-wdr54"] Oct 14 08:51:00 crc kubenswrapper[4870]: I1014 08:51:00.754891 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-wdr54"] Oct 14 08:51:00 crc kubenswrapper[4870]: I1014 08:51:00.760172 4870 scope.go:117] "RemoveContainer" containerID="c9b4104ff3ca2a1b1e28c7dd0deb488fdd313a06645aa94ff88804a992daf2f5" Oct 14 08:51:00 crc kubenswrapper[4870]: I1014 08:51:00.805943 4870 scope.go:117] "RemoveContainer" containerID="a30d157fe7e16e953bf51aa6c962ee186b3cae92e8792a4a64a4d1c4c6f120be" Oct 14 08:51:00 crc kubenswrapper[4870]: E1014 08:51:00.806387 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a30d157fe7e16e953bf51aa6c962ee186b3cae92e8792a4a64a4d1c4c6f120be\": container with ID starting with a30d157fe7e16e953bf51aa6c962ee186b3cae92e8792a4a64a4d1c4c6f120be not found: ID does not exist" containerID="a30d157fe7e16e953bf51aa6c962ee186b3cae92e8792a4a64a4d1c4c6f120be" Oct 14 08:51:00 crc kubenswrapper[4870]: I1014 08:51:00.806455 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a30d157fe7e16e953bf51aa6c962ee186b3cae92e8792a4a64a4d1c4c6f120be"} err="failed to get container status \"a30d157fe7e16e953bf51aa6c962ee186b3cae92e8792a4a64a4d1c4c6f120be\": rpc error: code = NotFound desc = could not find container \"a30d157fe7e16e953bf51aa6c962ee186b3cae92e8792a4a64a4d1c4c6f120be\": container with ID starting with a30d157fe7e16e953bf51aa6c962ee186b3cae92e8792a4a64a4d1c4c6f120be not found: ID does not exist" Oct 14 08:51:00 crc kubenswrapper[4870]: I1014 08:51:00.806490 4870 scope.go:117] "RemoveContainer" containerID="79afcd8a0aa446b0dd1e7ae18c6ff3892eb6e5942e4e94b180e1283eeeeef384" Oct 14 08:51:00 crc kubenswrapper[4870]: E1014 08:51:00.806888 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"79afcd8a0aa446b0dd1e7ae18c6ff3892eb6e5942e4e94b180e1283eeeeef384\": container with ID starting with 79afcd8a0aa446b0dd1e7ae18c6ff3892eb6e5942e4e94b180e1283eeeeef384 not found: ID does not exist" containerID="79afcd8a0aa446b0dd1e7ae18c6ff3892eb6e5942e4e94b180e1283eeeeef384" Oct 14 08:51:00 crc kubenswrapper[4870]: I1014 08:51:00.806928 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"79afcd8a0aa446b0dd1e7ae18c6ff3892eb6e5942e4e94b180e1283eeeeef384"} err="failed to get container status \"79afcd8a0aa446b0dd1e7ae18c6ff3892eb6e5942e4e94b180e1283eeeeef384\": rpc error: code = NotFound desc = could not find container \"79afcd8a0aa446b0dd1e7ae18c6ff3892eb6e5942e4e94b180e1283eeeeef384\": container with ID starting with 79afcd8a0aa446b0dd1e7ae18c6ff3892eb6e5942e4e94b180e1283eeeeef384 not found: ID does not exist" Oct 14 08:51:00 crc kubenswrapper[4870]: I1014 08:51:00.806953 4870 scope.go:117] "RemoveContainer" containerID="c9b4104ff3ca2a1b1e28c7dd0deb488fdd313a06645aa94ff88804a992daf2f5" Oct 14 08:51:00 crc kubenswrapper[4870]: E1014 08:51:00.807421 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c9b4104ff3ca2a1b1e28c7dd0deb488fdd313a06645aa94ff88804a992daf2f5\": container with ID starting with c9b4104ff3ca2a1b1e28c7dd0deb488fdd313a06645aa94ff88804a992daf2f5 not found: ID does not exist" containerID="c9b4104ff3ca2a1b1e28c7dd0deb488fdd313a06645aa94ff88804a992daf2f5" Oct 14 08:51:00 crc kubenswrapper[4870]: I1014 08:51:00.807619 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c9b4104ff3ca2a1b1e28c7dd0deb488fdd313a06645aa94ff88804a992daf2f5"} err="failed to get container status \"c9b4104ff3ca2a1b1e28c7dd0deb488fdd313a06645aa94ff88804a992daf2f5\": rpc error: code = NotFound desc = could not find container \"c9b4104ff3ca2a1b1e28c7dd0deb488fdd313a06645aa94ff88804a992daf2f5\": container with ID starting with c9b4104ff3ca2a1b1e28c7dd0deb488fdd313a06645aa94ff88804a992daf2f5 not found: ID does not exist" Oct 14 08:51:01 crc kubenswrapper[4870]: I1014 08:51:01.044868 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="32664967-2bce-429d-bb04-ebd14b9f03f5" path="/var/lib/kubelet/pods/32664967-2bce-429d-bb04-ebd14b9f03f5/volumes" Oct 14 08:51:21 crc kubenswrapper[4870]: I1014 08:51:21.646822 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 14 08:51:21 crc kubenswrapper[4870]: E1014 08:51:21.647889 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80e92bad-0578-41c9-b087-89f938be53a5" containerName="mariadb-client" Oct 14 08:51:21 crc kubenswrapper[4870]: I1014 08:51:21.647911 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="80e92bad-0578-41c9-b087-89f938be53a5" containerName="mariadb-client" Oct 14 08:51:21 crc kubenswrapper[4870]: E1014 08:51:21.647945 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="32664967-2bce-429d-bb04-ebd14b9f03f5" containerName="extract-utilities" Oct 14 08:51:21 crc kubenswrapper[4870]: I1014 08:51:21.647957 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="32664967-2bce-429d-bb04-ebd14b9f03f5" containerName="extract-utilities" Oct 14 08:51:21 crc kubenswrapper[4870]: E1014 08:51:21.647971 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="32664967-2bce-429d-bb04-ebd14b9f03f5" containerName="extract-content" Oct 14 08:51:21 crc kubenswrapper[4870]: I1014 08:51:21.647986 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="32664967-2bce-429d-bb04-ebd14b9f03f5" containerName="extract-content" Oct 14 08:51:21 crc kubenswrapper[4870]: E1014 08:51:21.648048 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="32664967-2bce-429d-bb04-ebd14b9f03f5" containerName="registry-server" Oct 14 08:51:21 crc kubenswrapper[4870]: I1014 08:51:21.648061 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="32664967-2bce-429d-bb04-ebd14b9f03f5" containerName="registry-server" Oct 14 08:51:21 crc kubenswrapper[4870]: I1014 08:51:21.648311 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="32664967-2bce-429d-bb04-ebd14b9f03f5" containerName="registry-server" Oct 14 08:51:21 crc kubenswrapper[4870]: I1014 08:51:21.648337 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="80e92bad-0578-41c9-b087-89f938be53a5" containerName="mariadb-client" Oct 14 08:51:21 crc kubenswrapper[4870]: I1014 08:51:21.649742 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Oct 14 08:51:21 crc kubenswrapper[4870]: I1014 08:51:21.652791 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-sppm7" Oct 14 08:51:21 crc kubenswrapper[4870]: I1014 08:51:21.653265 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Oct 14 08:51:21 crc kubenswrapper[4870]: I1014 08:51:21.653748 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Oct 14 08:51:21 crc kubenswrapper[4870]: I1014 08:51:21.667552 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-2"] Oct 14 08:51:21 crc kubenswrapper[4870]: I1014 08:51:21.670348 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-2" Oct 14 08:51:21 crc kubenswrapper[4870]: I1014 08:51:21.681652 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 14 08:51:21 crc kubenswrapper[4870]: I1014 08:51:21.699106 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-1"] Oct 14 08:51:21 crc kubenswrapper[4870]: I1014 08:51:21.701835 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-1" Oct 14 08:51:21 crc kubenswrapper[4870]: I1014 08:51:21.725293 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-2"] Oct 14 08:51:21 crc kubenswrapper[4870]: I1014 08:51:21.739131 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-1"] Oct 14 08:51:21 crc kubenswrapper[4870]: I1014 08:51:21.775695 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-aa874511-6bbc-44b4-89a3-aa0b8e38ef42\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-aa874511-6bbc-44b4-89a3-aa0b8e38ef42\") pod \"ovsdbserver-nb-2\" (UID: \"7d4f0ff6-a758-4177-a48e-679d3373589f\") " pod="openstack/ovsdbserver-nb-2" Oct 14 08:51:21 crc kubenswrapper[4870]: I1014 08:51:21.775961 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79b5761f-2312-4d15-9ded-d42c9f8f6009-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"79b5761f-2312-4d15-9ded-d42c9f8f6009\") " pod="openstack/ovsdbserver-nb-0" Oct 14 08:51:21 crc kubenswrapper[4870]: I1014 08:51:21.776233 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2f70f991-789c-49bc-bd89-aeb6a69e6ac5-scripts\") pod \"ovsdbserver-nb-1\" (UID: \"2f70f991-789c-49bc-bd89-aeb6a69e6ac5\") " pod="openstack/ovsdbserver-nb-1" Oct 14 08:51:21 crc kubenswrapper[4870]: I1014 08:51:21.776329 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2f70f991-789c-49bc-bd89-aeb6a69e6ac5-config\") pod \"ovsdbserver-nb-1\" (UID: \"2f70f991-789c-49bc-bd89-aeb6a69e6ac5\") " pod="openstack/ovsdbserver-nb-1" Oct 14 08:51:21 crc kubenswrapper[4870]: I1014 08:51:21.776416 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/79b5761f-2312-4d15-9ded-d42c9f8f6009-config\") pod \"ovsdbserver-nb-0\" (UID: \"79b5761f-2312-4d15-9ded-d42c9f8f6009\") " pod="openstack/ovsdbserver-nb-0" Oct 14 08:51:21 crc kubenswrapper[4870]: I1014 08:51:21.776547 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7d4f0ff6-a758-4177-a48e-679d3373589f-scripts\") pod \"ovsdbserver-nb-2\" (UID: \"7d4f0ff6-a758-4177-a48e-679d3373589f\") " pod="openstack/ovsdbserver-nb-2" Oct 14 08:51:21 crc kubenswrapper[4870]: I1014 08:51:21.776659 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/79b5761f-2312-4d15-9ded-d42c9f8f6009-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"79b5761f-2312-4d15-9ded-d42c9f8f6009\") " pod="openstack/ovsdbserver-nb-0" Oct 14 08:51:21 crc kubenswrapper[4870]: I1014 08:51:21.776748 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/2f70f991-789c-49bc-bd89-aeb6a69e6ac5-ovsdb-rundir\") pod \"ovsdbserver-nb-1\" (UID: \"2f70f991-789c-49bc-bd89-aeb6a69e6ac5\") " pod="openstack/ovsdbserver-nb-1" Oct 14 08:51:21 crc kubenswrapper[4870]: I1014 08:51:21.776838 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5kht4\" (UniqueName: \"kubernetes.io/projected/79b5761f-2312-4d15-9ded-d42c9f8f6009-kube-api-access-5kht4\") pod \"ovsdbserver-nb-0\" (UID: \"79b5761f-2312-4d15-9ded-d42c9f8f6009\") " pod="openstack/ovsdbserver-nb-0" Oct 14 08:51:21 crc kubenswrapper[4870]: I1014 08:51:21.776920 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f70f991-789c-49bc-bd89-aeb6a69e6ac5-combined-ca-bundle\") pod \"ovsdbserver-nb-1\" (UID: \"2f70f991-789c-49bc-bd89-aeb6a69e6ac5\") " pod="openstack/ovsdbserver-nb-1" Oct 14 08:51:21 crc kubenswrapper[4870]: I1014 08:51:21.776975 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nzkcm\" (UniqueName: \"kubernetes.io/projected/7d4f0ff6-a758-4177-a48e-679d3373589f-kube-api-access-nzkcm\") pod \"ovsdbserver-nb-2\" (UID: \"7d4f0ff6-a758-4177-a48e-679d3373589f\") " pod="openstack/ovsdbserver-nb-2" Oct 14 08:51:21 crc kubenswrapper[4870]: I1014 08:51:21.777024 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7d4f0ff6-a758-4177-a48e-679d3373589f-config\") pod \"ovsdbserver-nb-2\" (UID: \"7d4f0ff6-a758-4177-a48e-679d3373589f\") " pod="openstack/ovsdbserver-nb-2" Oct 14 08:51:21 crc kubenswrapper[4870]: I1014 08:51:21.777081 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/79b5761f-2312-4d15-9ded-d42c9f8f6009-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"79b5761f-2312-4d15-9ded-d42c9f8f6009\") " pod="openstack/ovsdbserver-nb-0" Oct 14 08:51:21 crc kubenswrapper[4870]: I1014 08:51:21.777157 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/7d4f0ff6-a758-4177-a48e-679d3373589f-ovsdb-rundir\") pod \"ovsdbserver-nb-2\" (UID: \"7d4f0ff6-a758-4177-a48e-679d3373589f\") " pod="openstack/ovsdbserver-nb-2" Oct 14 08:51:21 crc kubenswrapper[4870]: I1014 08:51:21.777336 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-a68485db-dc49-4567-a02f-030fdeeaaf4d\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a68485db-dc49-4567-a02f-030fdeeaaf4d\") pod \"ovsdbserver-nb-1\" (UID: \"2f70f991-789c-49bc-bd89-aeb6a69e6ac5\") " pod="openstack/ovsdbserver-nb-1" Oct 14 08:51:21 crc kubenswrapper[4870]: I1014 08:51:21.777518 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-68d34c00-a49c-4752-9346-35bc16e45c9a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-68d34c00-a49c-4752-9346-35bc16e45c9a\") pod \"ovsdbserver-nb-0\" (UID: \"79b5761f-2312-4d15-9ded-d42c9f8f6009\") " pod="openstack/ovsdbserver-nb-0" Oct 14 08:51:21 crc kubenswrapper[4870]: I1014 08:51:21.777578 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d4f0ff6-a758-4177-a48e-679d3373589f-combined-ca-bundle\") pod \"ovsdbserver-nb-2\" (UID: \"7d4f0ff6-a758-4177-a48e-679d3373589f\") " pod="openstack/ovsdbserver-nb-2" Oct 14 08:51:21 crc kubenswrapper[4870]: I1014 08:51:21.777635 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kkfcv\" (UniqueName: \"kubernetes.io/projected/2f70f991-789c-49bc-bd89-aeb6a69e6ac5-kube-api-access-kkfcv\") pod \"ovsdbserver-nb-1\" (UID: \"2f70f991-789c-49bc-bd89-aeb6a69e6ac5\") " pod="openstack/ovsdbserver-nb-1" Oct 14 08:51:21 crc kubenswrapper[4870]: I1014 08:51:21.840172 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 14 08:51:21 crc kubenswrapper[4870]: I1014 08:51:21.842464 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Oct 14 08:51:21 crc kubenswrapper[4870]: I1014 08:51:21.844812 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Oct 14 08:51:21 crc kubenswrapper[4870]: I1014 08:51:21.845183 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-x94mn" Oct 14 08:51:21 crc kubenswrapper[4870]: I1014 08:51:21.845322 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Oct 14 08:51:21 crc kubenswrapper[4870]: I1014 08:51:21.868401 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-1"] Oct 14 08:51:21 crc kubenswrapper[4870]: I1014 08:51:21.880770 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-a68485db-dc49-4567-a02f-030fdeeaaf4d\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a68485db-dc49-4567-a02f-030fdeeaaf4d\") pod \"ovsdbserver-nb-1\" (UID: \"2f70f991-789c-49bc-bd89-aeb6a69e6ac5\") " pod="openstack/ovsdbserver-nb-1" Oct 14 08:51:21 crc kubenswrapper[4870]: I1014 08:51:21.880858 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-68d34c00-a49c-4752-9346-35bc16e45c9a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-68d34c00-a49c-4752-9346-35bc16e45c9a\") pod \"ovsdbserver-nb-0\" (UID: \"79b5761f-2312-4d15-9ded-d42c9f8f6009\") " pod="openstack/ovsdbserver-nb-0" Oct 14 08:51:21 crc kubenswrapper[4870]: I1014 08:51:21.880899 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d4f0ff6-a758-4177-a48e-679d3373589f-combined-ca-bundle\") pod \"ovsdbserver-nb-2\" (UID: \"7d4f0ff6-a758-4177-a48e-679d3373589f\") " pod="openstack/ovsdbserver-nb-2" Oct 14 08:51:21 crc kubenswrapper[4870]: I1014 08:51:21.880934 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kkfcv\" (UniqueName: \"kubernetes.io/projected/2f70f991-789c-49bc-bd89-aeb6a69e6ac5-kube-api-access-kkfcv\") pod \"ovsdbserver-nb-1\" (UID: \"2f70f991-789c-49bc-bd89-aeb6a69e6ac5\") " pod="openstack/ovsdbserver-nb-1" Oct 14 08:51:21 crc kubenswrapper[4870]: I1014 08:51:21.880987 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-aa874511-6bbc-44b4-89a3-aa0b8e38ef42\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-aa874511-6bbc-44b4-89a3-aa0b8e38ef42\") pod \"ovsdbserver-nb-2\" (UID: \"7d4f0ff6-a758-4177-a48e-679d3373589f\") " pod="openstack/ovsdbserver-nb-2" Oct 14 08:51:21 crc kubenswrapper[4870]: I1014 08:51:21.881037 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79b5761f-2312-4d15-9ded-d42c9f8f6009-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"79b5761f-2312-4d15-9ded-d42c9f8f6009\") " pod="openstack/ovsdbserver-nb-0" Oct 14 08:51:21 crc kubenswrapper[4870]: I1014 08:51:21.881113 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2f70f991-789c-49bc-bd89-aeb6a69e6ac5-scripts\") pod \"ovsdbserver-nb-1\" (UID: \"2f70f991-789c-49bc-bd89-aeb6a69e6ac5\") " pod="openstack/ovsdbserver-nb-1" Oct 14 08:51:21 crc kubenswrapper[4870]: I1014 08:51:21.881147 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2f70f991-789c-49bc-bd89-aeb6a69e6ac5-config\") pod \"ovsdbserver-nb-1\" (UID: \"2f70f991-789c-49bc-bd89-aeb6a69e6ac5\") " pod="openstack/ovsdbserver-nb-1" Oct 14 08:51:21 crc kubenswrapper[4870]: I1014 08:51:21.881182 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/79b5761f-2312-4d15-9ded-d42c9f8f6009-config\") pod \"ovsdbserver-nb-0\" (UID: \"79b5761f-2312-4d15-9ded-d42c9f8f6009\") " pod="openstack/ovsdbserver-nb-0" Oct 14 08:51:21 crc kubenswrapper[4870]: I1014 08:51:21.881224 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7d4f0ff6-a758-4177-a48e-679d3373589f-scripts\") pod \"ovsdbserver-nb-2\" (UID: \"7d4f0ff6-a758-4177-a48e-679d3373589f\") " pod="openstack/ovsdbserver-nb-2" Oct 14 08:51:21 crc kubenswrapper[4870]: I1014 08:51:21.881277 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/79b5761f-2312-4d15-9ded-d42c9f8f6009-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"79b5761f-2312-4d15-9ded-d42c9f8f6009\") " pod="openstack/ovsdbserver-nb-0" Oct 14 08:51:21 crc kubenswrapper[4870]: I1014 08:51:21.881332 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/2f70f991-789c-49bc-bd89-aeb6a69e6ac5-ovsdb-rundir\") pod \"ovsdbserver-nb-1\" (UID: \"2f70f991-789c-49bc-bd89-aeb6a69e6ac5\") " pod="openstack/ovsdbserver-nb-1" Oct 14 08:51:21 crc kubenswrapper[4870]: I1014 08:51:21.881400 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5kht4\" (UniqueName: \"kubernetes.io/projected/79b5761f-2312-4d15-9ded-d42c9f8f6009-kube-api-access-5kht4\") pod \"ovsdbserver-nb-0\" (UID: \"79b5761f-2312-4d15-9ded-d42c9f8f6009\") " pod="openstack/ovsdbserver-nb-0" Oct 14 08:51:21 crc kubenswrapper[4870]: I1014 08:51:21.881432 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f70f991-789c-49bc-bd89-aeb6a69e6ac5-combined-ca-bundle\") pod \"ovsdbserver-nb-1\" (UID: \"2f70f991-789c-49bc-bd89-aeb6a69e6ac5\") " pod="openstack/ovsdbserver-nb-1" Oct 14 08:51:21 crc kubenswrapper[4870]: I1014 08:51:21.881561 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nzkcm\" (UniqueName: \"kubernetes.io/projected/7d4f0ff6-a758-4177-a48e-679d3373589f-kube-api-access-nzkcm\") pod \"ovsdbserver-nb-2\" (UID: \"7d4f0ff6-a758-4177-a48e-679d3373589f\") " pod="openstack/ovsdbserver-nb-2" Oct 14 08:51:21 crc kubenswrapper[4870]: I1014 08:51:21.881594 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7d4f0ff6-a758-4177-a48e-679d3373589f-config\") pod \"ovsdbserver-nb-2\" (UID: \"7d4f0ff6-a758-4177-a48e-679d3373589f\") " pod="openstack/ovsdbserver-nb-2" Oct 14 08:51:21 crc kubenswrapper[4870]: I1014 08:51:21.881621 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/79b5761f-2312-4d15-9ded-d42c9f8f6009-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"79b5761f-2312-4d15-9ded-d42c9f8f6009\") " pod="openstack/ovsdbserver-nb-0" Oct 14 08:51:21 crc kubenswrapper[4870]: I1014 08:51:21.881853 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/7d4f0ff6-a758-4177-a48e-679d3373589f-ovsdb-rundir\") pod \"ovsdbserver-nb-2\" (UID: \"7d4f0ff6-a758-4177-a48e-679d3373589f\") " pod="openstack/ovsdbserver-nb-2" Oct 14 08:51:21 crc kubenswrapper[4870]: I1014 08:51:21.882568 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/7d4f0ff6-a758-4177-a48e-679d3373589f-ovsdb-rundir\") pod \"ovsdbserver-nb-2\" (UID: \"7d4f0ff6-a758-4177-a48e-679d3373589f\") " pod="openstack/ovsdbserver-nb-2" Oct 14 08:51:21 crc kubenswrapper[4870]: I1014 08:51:21.883713 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/79b5761f-2312-4d15-9ded-d42c9f8f6009-config\") pod \"ovsdbserver-nb-0\" (UID: \"79b5761f-2312-4d15-9ded-d42c9f8f6009\") " pod="openstack/ovsdbserver-nb-0" Oct 14 08:51:21 crc kubenswrapper[4870]: I1014 08:51:21.883777 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7d4f0ff6-a758-4177-a48e-679d3373589f-scripts\") pod \"ovsdbserver-nb-2\" (UID: \"7d4f0ff6-a758-4177-a48e-679d3373589f\") " pod="openstack/ovsdbserver-nb-2" Oct 14 08:51:21 crc kubenswrapper[4870]: I1014 08:51:21.901639 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/2f70f991-789c-49bc-bd89-aeb6a69e6ac5-ovsdb-rundir\") pod \"ovsdbserver-nb-1\" (UID: \"2f70f991-789c-49bc-bd89-aeb6a69e6ac5\") " pod="openstack/ovsdbserver-nb-1" Oct 14 08:51:21 crc kubenswrapper[4870]: I1014 08:51:21.901739 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d4f0ff6-a758-4177-a48e-679d3373589f-combined-ca-bundle\") pod \"ovsdbserver-nb-2\" (UID: \"7d4f0ff6-a758-4177-a48e-679d3373589f\") " pod="openstack/ovsdbserver-nb-2" Oct 14 08:51:21 crc kubenswrapper[4870]: I1014 08:51:21.901932 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/79b5761f-2312-4d15-9ded-d42c9f8f6009-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"79b5761f-2312-4d15-9ded-d42c9f8f6009\") " pod="openstack/ovsdbserver-nb-0" Oct 14 08:51:21 crc kubenswrapper[4870]: I1014 08:51:21.921366 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2f70f991-789c-49bc-bd89-aeb6a69e6ac5-config\") pod \"ovsdbserver-nb-1\" (UID: \"2f70f991-789c-49bc-bd89-aeb6a69e6ac5\") " pod="openstack/ovsdbserver-nb-1" Oct 14 08:51:21 crc kubenswrapper[4870]: I1014 08:51:21.922467 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/79b5761f-2312-4d15-9ded-d42c9f8f6009-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"79b5761f-2312-4d15-9ded-d42c9f8f6009\") " pod="openstack/ovsdbserver-nb-0" Oct 14 08:51:21 crc kubenswrapper[4870]: I1014 08:51:21.925532 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2f70f991-789c-49bc-bd89-aeb6a69e6ac5-scripts\") pod \"ovsdbserver-nb-1\" (UID: \"2f70f991-789c-49bc-bd89-aeb6a69e6ac5\") " pod="openstack/ovsdbserver-nb-1" Oct 14 08:51:21 crc kubenswrapper[4870]: I1014 08:51:21.925604 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79b5761f-2312-4d15-9ded-d42c9f8f6009-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"79b5761f-2312-4d15-9ded-d42c9f8f6009\") " pod="openstack/ovsdbserver-nb-0" Oct 14 08:51:21 crc kubenswrapper[4870]: I1014 08:51:21.926093 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 14 08:51:21 crc kubenswrapper[4870]: I1014 08:51:21.926198 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-1" Oct 14 08:51:21 crc kubenswrapper[4870]: I1014 08:51:21.927142 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7d4f0ff6-a758-4177-a48e-679d3373589f-config\") pod \"ovsdbserver-nb-2\" (UID: \"7d4f0ff6-a758-4177-a48e-679d3373589f\") " pod="openstack/ovsdbserver-nb-2" Oct 14 08:51:21 crc kubenswrapper[4870]: I1014 08:51:21.927911 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f70f991-789c-49bc-bd89-aeb6a69e6ac5-combined-ca-bundle\") pod \"ovsdbserver-nb-1\" (UID: \"2f70f991-789c-49bc-bd89-aeb6a69e6ac5\") " pod="openstack/ovsdbserver-nb-1" Oct 14 08:51:21 crc kubenswrapper[4870]: I1014 08:51:21.929319 4870 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 14 08:51:21 crc kubenswrapper[4870]: I1014 08:51:21.929360 4870 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-a68485db-dc49-4567-a02f-030fdeeaaf4d\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a68485db-dc49-4567-a02f-030fdeeaaf4d\") pod \"ovsdbserver-nb-1\" (UID: \"2f70f991-789c-49bc-bd89-aeb6a69e6ac5\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/315ea0409245a893c3dceca574e93da65071f2a24d2939155b22832f3f4cbedc/globalmount\"" pod="openstack/ovsdbserver-nb-1" Oct 14 08:51:21 crc kubenswrapper[4870]: I1014 08:51:21.929520 4870 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 14 08:51:21 crc kubenswrapper[4870]: I1014 08:51:21.929602 4870 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-aa874511-6bbc-44b4-89a3-aa0b8e38ef42\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-aa874511-6bbc-44b4-89a3-aa0b8e38ef42\") pod \"ovsdbserver-nb-2\" (UID: \"7d4f0ff6-a758-4177-a48e-679d3373589f\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/5fcbff340acff9a08600e3cf92ac0a4abf0c8889ba7ac3700682e71ce2db3e1f/globalmount\"" pod="openstack/ovsdbserver-nb-2" Oct 14 08:51:21 crc kubenswrapper[4870]: I1014 08:51:21.930656 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-2"] Oct 14 08:51:21 crc kubenswrapper[4870]: I1014 08:51:21.930770 4870 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 14 08:51:21 crc kubenswrapper[4870]: I1014 08:51:21.930806 4870 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-68d34c00-a49c-4752-9346-35bc16e45c9a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-68d34c00-a49c-4752-9346-35bc16e45c9a\") pod \"ovsdbserver-nb-0\" (UID: \"79b5761f-2312-4d15-9ded-d42c9f8f6009\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/acfd0d19101d51999e4b55d827543e03367fbe6cc91769c90b9fb426e972f270/globalmount\"" pod="openstack/ovsdbserver-nb-0" Oct 14 08:51:21 crc kubenswrapper[4870]: I1014 08:51:21.931474 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nzkcm\" (UniqueName: \"kubernetes.io/projected/7d4f0ff6-a758-4177-a48e-679d3373589f-kube-api-access-nzkcm\") pod \"ovsdbserver-nb-2\" (UID: \"7d4f0ff6-a758-4177-a48e-679d3373589f\") " pod="openstack/ovsdbserver-nb-2" Oct 14 08:51:21 crc kubenswrapper[4870]: I1014 08:51:21.932366 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-2" Oct 14 08:51:21 crc kubenswrapper[4870]: I1014 08:51:21.933725 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5kht4\" (UniqueName: \"kubernetes.io/projected/79b5761f-2312-4d15-9ded-d42c9f8f6009-kube-api-access-5kht4\") pod \"ovsdbserver-nb-0\" (UID: \"79b5761f-2312-4d15-9ded-d42c9f8f6009\") " pod="openstack/ovsdbserver-nb-0" Oct 14 08:51:21 crc kubenswrapper[4870]: I1014 08:51:21.934636 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kkfcv\" (UniqueName: \"kubernetes.io/projected/2f70f991-789c-49bc-bd89-aeb6a69e6ac5-kube-api-access-kkfcv\") pod \"ovsdbserver-nb-1\" (UID: \"2f70f991-789c-49bc-bd89-aeb6a69e6ac5\") " pod="openstack/ovsdbserver-nb-1" Oct 14 08:51:21 crc kubenswrapper[4870]: I1014 08:51:21.940598 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-1"] Oct 14 08:51:21 crc kubenswrapper[4870]: I1014 08:51:21.952664 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-2"] Oct 14 08:51:21 crc kubenswrapper[4870]: I1014 08:51:21.967283 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-a68485db-dc49-4567-a02f-030fdeeaaf4d\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a68485db-dc49-4567-a02f-030fdeeaaf4d\") pod \"ovsdbserver-nb-1\" (UID: \"2f70f991-789c-49bc-bd89-aeb6a69e6ac5\") " pod="openstack/ovsdbserver-nb-1" Oct 14 08:51:21 crc kubenswrapper[4870]: I1014 08:51:21.967283 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-68d34c00-a49c-4752-9346-35bc16e45c9a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-68d34c00-a49c-4752-9346-35bc16e45c9a\") pod \"ovsdbserver-nb-0\" (UID: \"79b5761f-2312-4d15-9ded-d42c9f8f6009\") " pod="openstack/ovsdbserver-nb-0" Oct 14 08:51:21 crc kubenswrapper[4870]: I1014 08:51:21.970328 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-aa874511-6bbc-44b4-89a3-aa0b8e38ef42\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-aa874511-6bbc-44b4-89a3-aa0b8e38ef42\") pod \"ovsdbserver-nb-2\" (UID: \"7d4f0ff6-a758-4177-a48e-679d3373589f\") " pod="openstack/ovsdbserver-nb-2" Oct 14 08:51:21 crc kubenswrapper[4870]: I1014 08:51:21.976449 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Oct 14 08:51:21 crc kubenswrapper[4870]: I1014 08:51:21.983192 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0363cb3b-2556-47c2-a1d4-d1cec83c1354-config\") pod \"ovsdbserver-sb-0\" (UID: \"0363cb3b-2556-47c2-a1d4-d1cec83c1354\") " pod="openstack/ovsdbserver-sb-0" Oct 14 08:51:21 crc kubenswrapper[4870]: I1014 08:51:21.983237 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g96l5\" (UniqueName: \"kubernetes.io/projected/cd6f2350-67d9-417b-90e9-599fd80326ab-kube-api-access-g96l5\") pod \"ovsdbserver-sb-2\" (UID: \"cd6f2350-67d9-417b-90e9-599fd80326ab\") " pod="openstack/ovsdbserver-sb-2" Oct 14 08:51:21 crc kubenswrapper[4870]: I1014 08:51:21.983291 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd6f2350-67d9-417b-90e9-599fd80326ab-combined-ca-bundle\") pod \"ovsdbserver-sb-2\" (UID: \"cd6f2350-67d9-417b-90e9-599fd80326ab\") " pod="openstack/ovsdbserver-sb-2" Oct 14 08:51:21 crc kubenswrapper[4870]: I1014 08:51:21.983316 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tx64l\" (UniqueName: \"kubernetes.io/projected/09981a06-37e3-4dbf-b991-5a5a74824641-kube-api-access-tx64l\") pod \"ovsdbserver-sb-1\" (UID: \"09981a06-37e3-4dbf-b991-5a5a74824641\") " pod="openstack/ovsdbserver-sb-1" Oct 14 08:51:21 crc kubenswrapper[4870]: I1014 08:51:21.983337 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/09981a06-37e3-4dbf-b991-5a5a74824641-combined-ca-bundle\") pod \"ovsdbserver-sb-1\" (UID: \"09981a06-37e3-4dbf-b991-5a5a74824641\") " pod="openstack/ovsdbserver-sb-1" Oct 14 08:51:21 crc kubenswrapper[4870]: I1014 08:51:21.983445 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0363cb3b-2556-47c2-a1d4-d1cec83c1354-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"0363cb3b-2556-47c2-a1d4-d1cec83c1354\") " pod="openstack/ovsdbserver-sb-0" Oct 14 08:51:21 crc kubenswrapper[4870]: I1014 08:51:21.983470 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dbp95\" (UniqueName: \"kubernetes.io/projected/0363cb3b-2556-47c2-a1d4-d1cec83c1354-kube-api-access-dbp95\") pod \"ovsdbserver-sb-0\" (UID: \"0363cb3b-2556-47c2-a1d4-d1cec83c1354\") " pod="openstack/ovsdbserver-sb-0" Oct 14 08:51:21 crc kubenswrapper[4870]: I1014 08:51:21.983695 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-613bf0d1-c1f2-4b27-8dcb-62bd8d253b24\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-613bf0d1-c1f2-4b27-8dcb-62bd8d253b24\") pod \"ovsdbserver-sb-0\" (UID: \"0363cb3b-2556-47c2-a1d4-d1cec83c1354\") " pod="openstack/ovsdbserver-sb-0" Oct 14 08:51:21 crc kubenswrapper[4870]: I1014 08:51:21.983722 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cd6f2350-67d9-417b-90e9-599fd80326ab-scripts\") pod \"ovsdbserver-sb-2\" (UID: \"cd6f2350-67d9-417b-90e9-599fd80326ab\") " pod="openstack/ovsdbserver-sb-2" Oct 14 08:51:21 crc kubenswrapper[4870]: I1014 08:51:21.983753 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-35e6cae4-2e95-4ff2-a290-44ba4d895dd4\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-35e6cae4-2e95-4ff2-a290-44ba4d895dd4\") pod \"ovsdbserver-sb-1\" (UID: \"09981a06-37e3-4dbf-b991-5a5a74824641\") " pod="openstack/ovsdbserver-sb-1" Oct 14 08:51:21 crc kubenswrapper[4870]: I1014 08:51:21.983780 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-162248ee-3209-487a-873b-83143a9851c6\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-162248ee-3209-487a-873b-83143a9851c6\") pod \"ovsdbserver-sb-2\" (UID: \"cd6f2350-67d9-417b-90e9-599fd80326ab\") " pod="openstack/ovsdbserver-sb-2" Oct 14 08:51:21 crc kubenswrapper[4870]: I1014 08:51:21.983799 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/0363cb3b-2556-47c2-a1d4-d1cec83c1354-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"0363cb3b-2556-47c2-a1d4-d1cec83c1354\") " pod="openstack/ovsdbserver-sb-0" Oct 14 08:51:21 crc kubenswrapper[4870]: I1014 08:51:21.983840 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/cd6f2350-67d9-417b-90e9-599fd80326ab-ovsdb-rundir\") pod \"ovsdbserver-sb-2\" (UID: \"cd6f2350-67d9-417b-90e9-599fd80326ab\") " pod="openstack/ovsdbserver-sb-2" Oct 14 08:51:21 crc kubenswrapper[4870]: I1014 08:51:21.983862 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cd6f2350-67d9-417b-90e9-599fd80326ab-config\") pod \"ovsdbserver-sb-2\" (UID: \"cd6f2350-67d9-417b-90e9-599fd80326ab\") " pod="openstack/ovsdbserver-sb-2" Oct 14 08:51:21 crc kubenswrapper[4870]: I1014 08:51:21.983878 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/09981a06-37e3-4dbf-b991-5a5a74824641-scripts\") pod \"ovsdbserver-sb-1\" (UID: \"09981a06-37e3-4dbf-b991-5a5a74824641\") " pod="openstack/ovsdbserver-sb-1" Oct 14 08:51:21 crc kubenswrapper[4870]: I1014 08:51:21.983913 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0363cb3b-2556-47c2-a1d4-d1cec83c1354-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"0363cb3b-2556-47c2-a1d4-d1cec83c1354\") " pod="openstack/ovsdbserver-sb-0" Oct 14 08:51:21 crc kubenswrapper[4870]: I1014 08:51:21.983930 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09981a06-37e3-4dbf-b991-5a5a74824641-config\") pod \"ovsdbserver-sb-1\" (UID: \"09981a06-37e3-4dbf-b991-5a5a74824641\") " pod="openstack/ovsdbserver-sb-1" Oct 14 08:51:21 crc kubenswrapper[4870]: I1014 08:51:21.983948 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/09981a06-37e3-4dbf-b991-5a5a74824641-ovsdb-rundir\") pod \"ovsdbserver-sb-1\" (UID: \"09981a06-37e3-4dbf-b991-5a5a74824641\") " pod="openstack/ovsdbserver-sb-1" Oct 14 08:51:21 crc kubenswrapper[4870]: I1014 08:51:21.989763 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-2" Oct 14 08:51:22 crc kubenswrapper[4870]: I1014 08:51:22.019752 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-1" Oct 14 08:51:22 crc kubenswrapper[4870]: I1014 08:51:22.086716 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/cd6f2350-67d9-417b-90e9-599fd80326ab-ovsdb-rundir\") pod \"ovsdbserver-sb-2\" (UID: \"cd6f2350-67d9-417b-90e9-599fd80326ab\") " pod="openstack/ovsdbserver-sb-2" Oct 14 08:51:22 crc kubenswrapper[4870]: I1014 08:51:22.087053 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cd6f2350-67d9-417b-90e9-599fd80326ab-config\") pod \"ovsdbserver-sb-2\" (UID: \"cd6f2350-67d9-417b-90e9-599fd80326ab\") " pod="openstack/ovsdbserver-sb-2" Oct 14 08:51:22 crc kubenswrapper[4870]: I1014 08:51:22.087079 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/09981a06-37e3-4dbf-b991-5a5a74824641-scripts\") pod \"ovsdbserver-sb-1\" (UID: \"09981a06-37e3-4dbf-b991-5a5a74824641\") " pod="openstack/ovsdbserver-sb-1" Oct 14 08:51:22 crc kubenswrapper[4870]: I1014 08:51:22.087123 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0363cb3b-2556-47c2-a1d4-d1cec83c1354-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"0363cb3b-2556-47c2-a1d4-d1cec83c1354\") " pod="openstack/ovsdbserver-sb-0" Oct 14 08:51:22 crc kubenswrapper[4870]: I1014 08:51:22.087145 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09981a06-37e3-4dbf-b991-5a5a74824641-config\") pod \"ovsdbserver-sb-1\" (UID: \"09981a06-37e3-4dbf-b991-5a5a74824641\") " pod="openstack/ovsdbserver-sb-1" Oct 14 08:51:22 crc kubenswrapper[4870]: I1014 08:51:22.087166 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/09981a06-37e3-4dbf-b991-5a5a74824641-ovsdb-rundir\") pod \"ovsdbserver-sb-1\" (UID: \"09981a06-37e3-4dbf-b991-5a5a74824641\") " pod="openstack/ovsdbserver-sb-1" Oct 14 08:51:22 crc kubenswrapper[4870]: I1014 08:51:22.087186 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0363cb3b-2556-47c2-a1d4-d1cec83c1354-config\") pod \"ovsdbserver-sb-0\" (UID: \"0363cb3b-2556-47c2-a1d4-d1cec83c1354\") " pod="openstack/ovsdbserver-sb-0" Oct 14 08:51:22 crc kubenswrapper[4870]: I1014 08:51:22.087205 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g96l5\" (UniqueName: \"kubernetes.io/projected/cd6f2350-67d9-417b-90e9-599fd80326ab-kube-api-access-g96l5\") pod \"ovsdbserver-sb-2\" (UID: \"cd6f2350-67d9-417b-90e9-599fd80326ab\") " pod="openstack/ovsdbserver-sb-2" Oct 14 08:51:22 crc kubenswrapper[4870]: I1014 08:51:22.087224 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd6f2350-67d9-417b-90e9-599fd80326ab-combined-ca-bundle\") pod \"ovsdbserver-sb-2\" (UID: \"cd6f2350-67d9-417b-90e9-599fd80326ab\") " pod="openstack/ovsdbserver-sb-2" Oct 14 08:51:22 crc kubenswrapper[4870]: I1014 08:51:22.087240 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tx64l\" (UniqueName: \"kubernetes.io/projected/09981a06-37e3-4dbf-b991-5a5a74824641-kube-api-access-tx64l\") pod \"ovsdbserver-sb-1\" (UID: \"09981a06-37e3-4dbf-b991-5a5a74824641\") " pod="openstack/ovsdbserver-sb-1" Oct 14 08:51:22 crc kubenswrapper[4870]: I1014 08:51:22.087259 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/09981a06-37e3-4dbf-b991-5a5a74824641-combined-ca-bundle\") pod \"ovsdbserver-sb-1\" (UID: \"09981a06-37e3-4dbf-b991-5a5a74824641\") " pod="openstack/ovsdbserver-sb-1" Oct 14 08:51:22 crc kubenswrapper[4870]: I1014 08:51:22.087270 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/cd6f2350-67d9-417b-90e9-599fd80326ab-ovsdb-rundir\") pod \"ovsdbserver-sb-2\" (UID: \"cd6f2350-67d9-417b-90e9-599fd80326ab\") " pod="openstack/ovsdbserver-sb-2" Oct 14 08:51:22 crc kubenswrapper[4870]: I1014 08:51:22.087280 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0363cb3b-2556-47c2-a1d4-d1cec83c1354-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"0363cb3b-2556-47c2-a1d4-d1cec83c1354\") " pod="openstack/ovsdbserver-sb-0" Oct 14 08:51:22 crc kubenswrapper[4870]: I1014 08:51:22.087370 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dbp95\" (UniqueName: \"kubernetes.io/projected/0363cb3b-2556-47c2-a1d4-d1cec83c1354-kube-api-access-dbp95\") pod \"ovsdbserver-sb-0\" (UID: \"0363cb3b-2556-47c2-a1d4-d1cec83c1354\") " pod="openstack/ovsdbserver-sb-0" Oct 14 08:51:22 crc kubenswrapper[4870]: I1014 08:51:22.087736 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-613bf0d1-c1f2-4b27-8dcb-62bd8d253b24\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-613bf0d1-c1f2-4b27-8dcb-62bd8d253b24\") pod \"ovsdbserver-sb-0\" (UID: \"0363cb3b-2556-47c2-a1d4-d1cec83c1354\") " pod="openstack/ovsdbserver-sb-0" Oct 14 08:51:22 crc kubenswrapper[4870]: I1014 08:51:22.087787 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cd6f2350-67d9-417b-90e9-599fd80326ab-scripts\") pod \"ovsdbserver-sb-2\" (UID: \"cd6f2350-67d9-417b-90e9-599fd80326ab\") " pod="openstack/ovsdbserver-sb-2" Oct 14 08:51:22 crc kubenswrapper[4870]: I1014 08:51:22.087844 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-35e6cae4-2e95-4ff2-a290-44ba4d895dd4\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-35e6cae4-2e95-4ff2-a290-44ba4d895dd4\") pod \"ovsdbserver-sb-1\" (UID: \"09981a06-37e3-4dbf-b991-5a5a74824641\") " pod="openstack/ovsdbserver-sb-1" Oct 14 08:51:22 crc kubenswrapper[4870]: I1014 08:51:22.087872 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-162248ee-3209-487a-873b-83143a9851c6\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-162248ee-3209-487a-873b-83143a9851c6\") pod \"ovsdbserver-sb-2\" (UID: \"cd6f2350-67d9-417b-90e9-599fd80326ab\") " pod="openstack/ovsdbserver-sb-2" Oct 14 08:51:22 crc kubenswrapper[4870]: I1014 08:51:22.087897 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/0363cb3b-2556-47c2-a1d4-d1cec83c1354-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"0363cb3b-2556-47c2-a1d4-d1cec83c1354\") " pod="openstack/ovsdbserver-sb-0" Oct 14 08:51:22 crc kubenswrapper[4870]: I1014 08:51:22.088365 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0363cb3b-2556-47c2-a1d4-d1cec83c1354-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"0363cb3b-2556-47c2-a1d4-d1cec83c1354\") " pod="openstack/ovsdbserver-sb-0" Oct 14 08:51:22 crc kubenswrapper[4870]: I1014 08:51:22.088380 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/0363cb3b-2556-47c2-a1d4-d1cec83c1354-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"0363cb3b-2556-47c2-a1d4-d1cec83c1354\") " pod="openstack/ovsdbserver-sb-0" Oct 14 08:51:22 crc kubenswrapper[4870]: I1014 08:51:22.088679 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09981a06-37e3-4dbf-b991-5a5a74824641-config\") pod \"ovsdbserver-sb-1\" (UID: \"09981a06-37e3-4dbf-b991-5a5a74824641\") " pod="openstack/ovsdbserver-sb-1" Oct 14 08:51:22 crc kubenswrapper[4870]: I1014 08:51:22.089209 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0363cb3b-2556-47c2-a1d4-d1cec83c1354-config\") pod \"ovsdbserver-sb-0\" (UID: \"0363cb3b-2556-47c2-a1d4-d1cec83c1354\") " pod="openstack/ovsdbserver-sb-0" Oct 14 08:51:22 crc kubenswrapper[4870]: I1014 08:51:22.089211 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/09981a06-37e3-4dbf-b991-5a5a74824641-scripts\") pod \"ovsdbserver-sb-1\" (UID: \"09981a06-37e3-4dbf-b991-5a5a74824641\") " pod="openstack/ovsdbserver-sb-1" Oct 14 08:51:22 crc kubenswrapper[4870]: I1014 08:51:22.091217 4870 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 14 08:51:22 crc kubenswrapper[4870]: I1014 08:51:22.091226 4870 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 14 08:51:22 crc kubenswrapper[4870]: I1014 08:51:22.091252 4870 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-162248ee-3209-487a-873b-83143a9851c6\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-162248ee-3209-487a-873b-83143a9851c6\") pod \"ovsdbserver-sb-2\" (UID: \"cd6f2350-67d9-417b-90e9-599fd80326ab\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/49559efe959499ac38d7e3fbf16d32e63c59df56b885abc59694fade3ef1d842/globalmount\"" pod="openstack/ovsdbserver-sb-2" Oct 14 08:51:22 crc kubenswrapper[4870]: I1014 08:51:22.091259 4870 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-35e6cae4-2e95-4ff2-a290-44ba4d895dd4\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-35e6cae4-2e95-4ff2-a290-44ba4d895dd4\") pod \"ovsdbserver-sb-1\" (UID: \"09981a06-37e3-4dbf-b991-5a5a74824641\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/10d2e73b46b7538fdf8a7b7872c0ef7b2941ea4ad573292ff3d08e962fca6bd8/globalmount\"" pod="openstack/ovsdbserver-sb-1" Oct 14 08:51:22 crc kubenswrapper[4870]: I1014 08:51:22.091534 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cd6f2350-67d9-417b-90e9-599fd80326ab-scripts\") pod \"ovsdbserver-sb-2\" (UID: \"cd6f2350-67d9-417b-90e9-599fd80326ab\") " pod="openstack/ovsdbserver-sb-2" Oct 14 08:51:22 crc kubenswrapper[4870]: I1014 08:51:22.091782 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/09981a06-37e3-4dbf-b991-5a5a74824641-ovsdb-rundir\") pod \"ovsdbserver-sb-1\" (UID: \"09981a06-37e3-4dbf-b991-5a5a74824641\") " pod="openstack/ovsdbserver-sb-1" Oct 14 08:51:22 crc kubenswrapper[4870]: I1014 08:51:22.092563 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cd6f2350-67d9-417b-90e9-599fd80326ab-config\") pod \"ovsdbserver-sb-2\" (UID: \"cd6f2350-67d9-417b-90e9-599fd80326ab\") " pod="openstack/ovsdbserver-sb-2" Oct 14 08:51:22 crc kubenswrapper[4870]: I1014 08:51:22.093983 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd6f2350-67d9-417b-90e9-599fd80326ab-combined-ca-bundle\") pod \"ovsdbserver-sb-2\" (UID: \"cd6f2350-67d9-417b-90e9-599fd80326ab\") " pod="openstack/ovsdbserver-sb-2" Oct 14 08:51:22 crc kubenswrapper[4870]: I1014 08:51:22.094185 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0363cb3b-2556-47c2-a1d4-d1cec83c1354-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"0363cb3b-2556-47c2-a1d4-d1cec83c1354\") " pod="openstack/ovsdbserver-sb-0" Oct 14 08:51:22 crc kubenswrapper[4870]: I1014 08:51:22.094339 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/09981a06-37e3-4dbf-b991-5a5a74824641-combined-ca-bundle\") pod \"ovsdbserver-sb-1\" (UID: \"09981a06-37e3-4dbf-b991-5a5a74824641\") " pod="openstack/ovsdbserver-sb-1" Oct 14 08:51:22 crc kubenswrapper[4870]: I1014 08:51:22.110112 4870 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 14 08:51:22 crc kubenswrapper[4870]: I1014 08:51:22.110158 4870 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-613bf0d1-c1f2-4b27-8dcb-62bd8d253b24\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-613bf0d1-c1f2-4b27-8dcb-62bd8d253b24\") pod \"ovsdbserver-sb-0\" (UID: \"0363cb3b-2556-47c2-a1d4-d1cec83c1354\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/cb686951b0f04345e3250632c9c97648adcf87f2db76aeb79c0bfe9d0d8cd167/globalmount\"" pod="openstack/ovsdbserver-sb-0" Oct 14 08:51:22 crc kubenswrapper[4870]: I1014 08:51:22.110298 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dbp95\" (UniqueName: \"kubernetes.io/projected/0363cb3b-2556-47c2-a1d4-d1cec83c1354-kube-api-access-dbp95\") pod \"ovsdbserver-sb-0\" (UID: \"0363cb3b-2556-47c2-a1d4-d1cec83c1354\") " pod="openstack/ovsdbserver-sb-0" Oct 14 08:51:22 crc kubenswrapper[4870]: I1014 08:51:22.113371 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tx64l\" (UniqueName: \"kubernetes.io/projected/09981a06-37e3-4dbf-b991-5a5a74824641-kube-api-access-tx64l\") pod \"ovsdbserver-sb-1\" (UID: \"09981a06-37e3-4dbf-b991-5a5a74824641\") " pod="openstack/ovsdbserver-sb-1" Oct 14 08:51:22 crc kubenswrapper[4870]: I1014 08:51:22.127739 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g96l5\" (UniqueName: \"kubernetes.io/projected/cd6f2350-67d9-417b-90e9-599fd80326ab-kube-api-access-g96l5\") pod \"ovsdbserver-sb-2\" (UID: \"cd6f2350-67d9-417b-90e9-599fd80326ab\") " pod="openstack/ovsdbserver-sb-2" Oct 14 08:51:22 crc kubenswrapper[4870]: I1014 08:51:22.146384 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-162248ee-3209-487a-873b-83143a9851c6\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-162248ee-3209-487a-873b-83143a9851c6\") pod \"ovsdbserver-sb-2\" (UID: \"cd6f2350-67d9-417b-90e9-599fd80326ab\") " pod="openstack/ovsdbserver-sb-2" Oct 14 08:51:22 crc kubenswrapper[4870]: I1014 08:51:22.152492 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-35e6cae4-2e95-4ff2-a290-44ba4d895dd4\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-35e6cae4-2e95-4ff2-a290-44ba4d895dd4\") pod \"ovsdbserver-sb-1\" (UID: \"09981a06-37e3-4dbf-b991-5a5a74824641\") " pod="openstack/ovsdbserver-sb-1" Oct 14 08:51:22 crc kubenswrapper[4870]: I1014 08:51:22.154570 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-613bf0d1-c1f2-4b27-8dcb-62bd8d253b24\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-613bf0d1-c1f2-4b27-8dcb-62bd8d253b24\") pod \"ovsdbserver-sb-0\" (UID: \"0363cb3b-2556-47c2-a1d4-d1cec83c1354\") " pod="openstack/ovsdbserver-sb-0" Oct 14 08:51:22 crc kubenswrapper[4870]: I1014 08:51:22.188199 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Oct 14 08:51:22 crc kubenswrapper[4870]: I1014 08:51:22.357221 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-1" Oct 14 08:51:22 crc kubenswrapper[4870]: I1014 08:51:22.366558 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-2" Oct 14 08:51:22 crc kubenswrapper[4870]: I1014 08:51:22.515608 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 14 08:51:22 crc kubenswrapper[4870]: W1014 08:51:22.568874 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0363cb3b_2556_47c2_a1d4_d1cec83c1354.slice/crio-169c3f6c17755567c97e0584cc24c164d7b42f99cade1c0d9963e89c89b77aa3 WatchSource:0}: Error finding container 169c3f6c17755567c97e0584cc24c164d7b42f99cade1c0d9963e89c89b77aa3: Status 404 returned error can't find the container with id 169c3f6c17755567c97e0584cc24c164d7b42f99cade1c0d9963e89c89b77aa3 Oct 14 08:51:22 crc kubenswrapper[4870]: I1014 08:51:22.597404 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-2"] Oct 14 08:51:22 crc kubenswrapper[4870]: W1014 08:51:22.607467 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7d4f0ff6_a758_4177_a48e_679d3373589f.slice/crio-bfb501afc3030b567b46294d006a9479112041e5ac1c200cd3432a0ff152e888 WatchSource:0}: Error finding container bfb501afc3030b567b46294d006a9479112041e5ac1c200cd3432a0ff152e888: Status 404 returned error can't find the container with id bfb501afc3030b567b46294d006a9479112041e5ac1c200cd3432a0ff152e888 Oct 14 08:51:22 crc kubenswrapper[4870]: I1014 08:51:22.939308 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-2" event={"ID":"7d4f0ff6-a758-4177-a48e-679d3373589f","Type":"ContainerStarted","Data":"bfb501afc3030b567b46294d006a9479112041e5ac1c200cd3432a0ff152e888"} Oct 14 08:51:22 crc kubenswrapper[4870]: I1014 08:51:22.943895 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"0363cb3b-2556-47c2-a1d4-d1cec83c1354","Type":"ContainerStarted","Data":"169c3f6c17755567c97e0584cc24c164d7b42f99cade1c0d9963e89c89b77aa3"} Oct 14 08:51:22 crc kubenswrapper[4870]: I1014 08:51:22.978945 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-1"] Oct 14 08:51:22 crc kubenswrapper[4870]: W1014 08:51:22.989235 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod09981a06_37e3_4dbf_b991_5a5a74824641.slice/crio-55e263d4afb17ebbd22466e03fbad75d9509b6891857b402230bdec436c67d20 WatchSource:0}: Error finding container 55e263d4afb17ebbd22466e03fbad75d9509b6891857b402230bdec436c67d20: Status 404 returned error can't find the container with id 55e263d4afb17ebbd22466e03fbad75d9509b6891857b402230bdec436c67d20 Oct 14 08:51:23 crc kubenswrapper[4870]: I1014 08:51:23.070555 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-2"] Oct 14 08:51:23 crc kubenswrapper[4870]: W1014 08:51:23.072805 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcd6f2350_67d9_417b_90e9_599fd80326ab.slice/crio-d57b46a4249627a2572ea28d649a9edc18bb780eb763204edeb1787e519109a5 WatchSource:0}: Error finding container d57b46a4249627a2572ea28d649a9edc18bb780eb763204edeb1787e519109a5: Status 404 returned error can't find the container with id d57b46a4249627a2572ea28d649a9edc18bb780eb763204edeb1787e519109a5 Oct 14 08:51:23 crc kubenswrapper[4870]: I1014 08:51:23.172934 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-1"] Oct 14 08:51:23 crc kubenswrapper[4870]: W1014 08:51:23.189349 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2f70f991_789c_49bc_bd89_aeb6a69e6ac5.slice/crio-6518dd9966fee84a8fc58d5bce1cbcf532197db409280ad2399b9aab10922033 WatchSource:0}: Error finding container 6518dd9966fee84a8fc58d5bce1cbcf532197db409280ad2399b9aab10922033: Status 404 returned error can't find the container with id 6518dd9966fee84a8fc58d5bce1cbcf532197db409280ad2399b9aab10922033 Oct 14 08:51:23 crc kubenswrapper[4870]: I1014 08:51:23.637822 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 14 08:51:23 crc kubenswrapper[4870]: W1014 08:51:23.646704 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod79b5761f_2312_4d15_9ded_d42c9f8f6009.slice/crio-c07d481091aad15ce000b32937608662bb8b6568d1ac49d2502996ab0b87ab52 WatchSource:0}: Error finding container c07d481091aad15ce000b32937608662bb8b6568d1ac49d2502996ab0b87ab52: Status 404 returned error can't find the container with id c07d481091aad15ce000b32937608662bb8b6568d1ac49d2502996ab0b87ab52 Oct 14 08:51:23 crc kubenswrapper[4870]: I1014 08:51:23.954037 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"79b5761f-2312-4d15-9ded-d42c9f8f6009","Type":"ContainerStarted","Data":"c07d481091aad15ce000b32937608662bb8b6568d1ac49d2502996ab0b87ab52"} Oct 14 08:51:23 crc kubenswrapper[4870]: I1014 08:51:23.956825 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-1" event={"ID":"2f70f991-789c-49bc-bd89-aeb6a69e6ac5","Type":"ContainerStarted","Data":"6518dd9966fee84a8fc58d5bce1cbcf532197db409280ad2399b9aab10922033"} Oct 14 08:51:23 crc kubenswrapper[4870]: I1014 08:51:23.959956 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-2" event={"ID":"cd6f2350-67d9-417b-90e9-599fd80326ab","Type":"ContainerStarted","Data":"d57b46a4249627a2572ea28d649a9edc18bb780eb763204edeb1787e519109a5"} Oct 14 08:51:23 crc kubenswrapper[4870]: I1014 08:51:23.961844 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-1" event={"ID":"09981a06-37e3-4dbf-b991-5a5a74824641","Type":"ContainerStarted","Data":"55e263d4afb17ebbd22466e03fbad75d9509b6891857b402230bdec436c67d20"} Oct 14 08:51:28 crc kubenswrapper[4870]: I1014 08:51:28.016334 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-2" event={"ID":"7d4f0ff6-a758-4177-a48e-679d3373589f","Type":"ContainerStarted","Data":"b42e8317b1a4758a1727eec346b871aedb65b8248f88b289da1ff65d1ed0a754"} Oct 14 08:51:28 crc kubenswrapper[4870]: I1014 08:51:28.024827 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-1" event={"ID":"2f70f991-789c-49bc-bd89-aeb6a69e6ac5","Type":"ContainerStarted","Data":"7bb7a6f87d9a735be8327fd01546cf53deb849264cc9ba2dae346b9f8046fed7"} Oct 14 08:51:28 crc kubenswrapper[4870]: I1014 08:51:28.026473 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"79b5761f-2312-4d15-9ded-d42c9f8f6009","Type":"ContainerStarted","Data":"a2456bed6d7878cbcf50e039d14e517fdab84269753243604890e4c1308f8ded"} Oct 14 08:51:28 crc kubenswrapper[4870]: I1014 08:51:28.032480 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-2" event={"ID":"cd6f2350-67d9-417b-90e9-599fd80326ab","Type":"ContainerStarted","Data":"58b2314ac02c7d150de0b7c007574f2c473e3b0e1460e592a6681e3000ad25a6"} Oct 14 08:51:28 crc kubenswrapper[4870]: I1014 08:51:28.036454 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"0363cb3b-2556-47c2-a1d4-d1cec83c1354","Type":"ContainerStarted","Data":"31cb4006215d16b8a3bb98330abb926d5813e6f78d2e5843b605b1707470489a"} Oct 14 08:51:29 crc kubenswrapper[4870]: I1014 08:51:29.054875 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-1" event={"ID":"2f70f991-789c-49bc-bd89-aeb6a69e6ac5","Type":"ContainerStarted","Data":"ebcaae570a83458a7b700c41c5fcf0c964b9eeec31597a65d27198fd98ae8bb5"} Oct 14 08:51:29 crc kubenswrapper[4870]: I1014 08:51:29.054966 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"79b5761f-2312-4d15-9ded-d42c9f8f6009","Type":"ContainerStarted","Data":"9370c332a4bbdf3126c5d8356f65b77d8b221dd175e8b8578dd8454826e0bd8b"} Oct 14 08:51:29 crc kubenswrapper[4870]: I1014 08:51:29.059627 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-2" event={"ID":"cd6f2350-67d9-417b-90e9-599fd80326ab","Type":"ContainerStarted","Data":"d632464c4545e5cb28174d496641c5157511984d035b39dbf4cd114a47e25d8c"} Oct 14 08:51:29 crc kubenswrapper[4870]: I1014 08:51:29.062210 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"0363cb3b-2556-47c2-a1d4-d1cec83c1354","Type":"ContainerStarted","Data":"c0bb0bcf10052b5514e0c742852d01b6dc082059ada911450f01ffe94c007170"} Oct 14 08:51:29 crc kubenswrapper[4870]: I1014 08:51:29.064961 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-1" event={"ID":"09981a06-37e3-4dbf-b991-5a5a74824641","Type":"ContainerStarted","Data":"b769de52521c15da45b2cee264acbefc6e1f2c374ff2453ca748aea844330edc"} Oct 14 08:51:29 crc kubenswrapper[4870]: I1014 08:51:29.065187 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-1" event={"ID":"09981a06-37e3-4dbf-b991-5a5a74824641","Type":"ContainerStarted","Data":"d5d07fdf767a5f5050ad3b46ca2855e0e30cbc60a41a527ef02478c3152aeb8e"} Oct 14 08:51:29 crc kubenswrapper[4870]: I1014 08:51:29.068046 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-2" event={"ID":"7d4f0ff6-a758-4177-a48e-679d3373589f","Type":"ContainerStarted","Data":"6df246c15741a1415226c5f9ee37aa048be69220257392e6403c5225293c6322"} Oct 14 08:51:29 crc kubenswrapper[4870]: I1014 08:51:29.087888 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-1" podStartSLOduration=4.6107664360000005 podStartE2EDuration="9.08785818s" podCreationTimestamp="2025-10-14 08:51:20 +0000 UTC" firstStartedPulling="2025-10-14 08:51:23.193269411 +0000 UTC m=+6618.890629782" lastFinishedPulling="2025-10-14 08:51:27.670361155 +0000 UTC m=+6623.367721526" observedRunningTime="2025-10-14 08:51:29.082373118 +0000 UTC m=+6624.779733499" watchObservedRunningTime="2025-10-14 08:51:29.08785818 +0000 UTC m=+6624.785218591" Oct 14 08:51:29 crc kubenswrapper[4870]: I1014 08:51:29.113034 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-2" podStartSLOduration=4.053832151 podStartE2EDuration="9.113001971s" podCreationTimestamp="2025-10-14 08:51:20 +0000 UTC" firstStartedPulling="2025-10-14 08:51:22.610261021 +0000 UTC m=+6618.307621392" lastFinishedPulling="2025-10-14 08:51:27.669430841 +0000 UTC m=+6623.366791212" observedRunningTime="2025-10-14 08:51:29.112669892 +0000 UTC m=+6624.810030323" watchObservedRunningTime="2025-10-14 08:51:29.113001971 +0000 UTC m=+6624.810362382" Oct 14 08:51:29 crc kubenswrapper[4870]: I1014 08:51:29.145158 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=5.12549192 podStartE2EDuration="9.145136252s" podCreationTimestamp="2025-10-14 08:51:20 +0000 UTC" firstStartedPulling="2025-10-14 08:51:23.650514508 +0000 UTC m=+6619.347874879" lastFinishedPulling="2025-10-14 08:51:27.67015884 +0000 UTC m=+6623.367519211" observedRunningTime="2025-10-14 08:51:29.135755759 +0000 UTC m=+6624.833116180" watchObservedRunningTime="2025-10-14 08:51:29.145136252 +0000 UTC m=+6624.842496633" Oct 14 08:51:29 crc kubenswrapper[4870]: I1014 08:51:29.189891 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-1" podStartSLOduration=4.475335754 podStartE2EDuration="9.189871759s" podCreationTimestamp="2025-10-14 08:51:20 +0000 UTC" firstStartedPulling="2025-10-14 08:51:22.998163775 +0000 UTC m=+6618.695524146" lastFinishedPulling="2025-10-14 08:51:27.71269977 +0000 UTC m=+6623.410060151" observedRunningTime="2025-10-14 08:51:29.179951832 +0000 UTC m=+6624.877312223" watchObservedRunningTime="2025-10-14 08:51:29.189871759 +0000 UTC m=+6624.887232140" Oct 14 08:51:29 crc kubenswrapper[4870]: I1014 08:51:29.190531 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=4.09127053 podStartE2EDuration="9.190521456s" podCreationTimestamp="2025-10-14 08:51:20 +0000 UTC" firstStartedPulling="2025-10-14 08:51:22.570998506 +0000 UTC m=+6618.268358867" lastFinishedPulling="2025-10-14 08:51:27.670249382 +0000 UTC m=+6623.367609793" observedRunningTime="2025-10-14 08:51:29.160594502 +0000 UTC m=+6624.857954883" watchObservedRunningTime="2025-10-14 08:51:29.190521456 +0000 UTC m=+6624.887881837" Oct 14 08:51:29 crc kubenswrapper[4870]: I1014 08:51:29.212651 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-2" podStartSLOduration=4.615888859 podStartE2EDuration="9.212635838s" podCreationTimestamp="2025-10-14 08:51:20 +0000 UTC" firstStartedPulling="2025-10-14 08:51:23.075219808 +0000 UTC m=+6618.772580179" lastFinishedPulling="2025-10-14 08:51:27.671966747 +0000 UTC m=+6623.369327158" observedRunningTime="2025-10-14 08:51:29.211832437 +0000 UTC m=+6624.909192808" watchObservedRunningTime="2025-10-14 08:51:29.212635838 +0000 UTC m=+6624.909996209" Oct 14 08:51:30 crc kubenswrapper[4870]: I1014 08:51:30.976791 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Oct 14 08:51:30 crc kubenswrapper[4870]: I1014 08:51:30.990720 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-2" Oct 14 08:51:31 crc kubenswrapper[4870]: I1014 08:51:31.020289 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-1" Oct 14 08:51:31 crc kubenswrapper[4870]: I1014 08:51:31.048111 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-2" Oct 14 08:51:31 crc kubenswrapper[4870]: I1014 08:51:31.048186 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Oct 14 08:51:31 crc kubenswrapper[4870]: I1014 08:51:31.062298 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-1" Oct 14 08:51:31 crc kubenswrapper[4870]: I1014 08:51:31.086143 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-2" Oct 14 08:51:31 crc kubenswrapper[4870]: I1014 08:51:31.086773 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Oct 14 08:51:31 crc kubenswrapper[4870]: I1014 08:51:31.086847 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-1" Oct 14 08:51:31 crc kubenswrapper[4870]: I1014 08:51:31.189128 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Oct 14 08:51:31 crc kubenswrapper[4870]: I1014 08:51:31.243491 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Oct 14 08:51:31 crc kubenswrapper[4870]: I1014 08:51:31.357617 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-1" Oct 14 08:51:31 crc kubenswrapper[4870]: I1014 08:51:31.367583 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-2" Oct 14 08:51:31 crc kubenswrapper[4870]: I1014 08:51:31.428246 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-2" Oct 14 08:51:31 crc kubenswrapper[4870]: I1014 08:51:31.429876 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-1" Oct 14 08:51:32 crc kubenswrapper[4870]: I1014 08:51:32.090166 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-1" Oct 14 08:51:32 crc kubenswrapper[4870]: I1014 08:51:32.096358 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Oct 14 08:51:32 crc kubenswrapper[4870]: I1014 08:51:32.096656 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-2" Oct 14 08:51:32 crc kubenswrapper[4870]: I1014 08:51:32.096764 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-1" Oct 14 08:51:32 crc kubenswrapper[4870]: I1014 08:51:32.153888 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Oct 14 08:51:32 crc kubenswrapper[4870]: I1014 08:51:32.160576 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Oct 14 08:51:32 crc kubenswrapper[4870]: I1014 08:51:32.165015 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-1" Oct 14 08:51:32 crc kubenswrapper[4870]: I1014 08:51:32.190224 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-2" Oct 14 08:51:32 crc kubenswrapper[4870]: I1014 08:51:32.201075 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-2" Oct 14 08:51:32 crc kubenswrapper[4870]: I1014 08:51:32.312023 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6d9d9fc6d9-ql84s"] Oct 14 08:51:32 crc kubenswrapper[4870]: I1014 08:51:32.313269 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d9d9fc6d9-ql84s" Oct 14 08:51:32 crc kubenswrapper[4870]: I1014 08:51:32.315226 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Oct 14 08:51:32 crc kubenswrapper[4870]: I1014 08:51:32.333820 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6d9d9fc6d9-ql84s"] Oct 14 08:51:32 crc kubenswrapper[4870]: I1014 08:51:32.380101 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/935a9115-0077-4a70-a6d7-cd9ba80fec28-dns-svc\") pod \"dnsmasq-dns-6d9d9fc6d9-ql84s\" (UID: \"935a9115-0077-4a70-a6d7-cd9ba80fec28\") " pod="openstack/dnsmasq-dns-6d9d9fc6d9-ql84s" Oct 14 08:51:32 crc kubenswrapper[4870]: I1014 08:51:32.380147 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/935a9115-0077-4a70-a6d7-cd9ba80fec28-config\") pod \"dnsmasq-dns-6d9d9fc6d9-ql84s\" (UID: \"935a9115-0077-4a70-a6d7-cd9ba80fec28\") " pod="openstack/dnsmasq-dns-6d9d9fc6d9-ql84s" Oct 14 08:51:32 crc kubenswrapper[4870]: I1014 08:51:32.380402 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vqmlv\" (UniqueName: \"kubernetes.io/projected/935a9115-0077-4a70-a6d7-cd9ba80fec28-kube-api-access-vqmlv\") pod \"dnsmasq-dns-6d9d9fc6d9-ql84s\" (UID: \"935a9115-0077-4a70-a6d7-cd9ba80fec28\") " pod="openstack/dnsmasq-dns-6d9d9fc6d9-ql84s" Oct 14 08:51:32 crc kubenswrapper[4870]: I1014 08:51:32.380614 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/935a9115-0077-4a70-a6d7-cd9ba80fec28-ovsdbserver-nb\") pod \"dnsmasq-dns-6d9d9fc6d9-ql84s\" (UID: \"935a9115-0077-4a70-a6d7-cd9ba80fec28\") " pod="openstack/dnsmasq-dns-6d9d9fc6d9-ql84s" Oct 14 08:51:32 crc kubenswrapper[4870]: I1014 08:51:32.482086 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vqmlv\" (UniqueName: \"kubernetes.io/projected/935a9115-0077-4a70-a6d7-cd9ba80fec28-kube-api-access-vqmlv\") pod \"dnsmasq-dns-6d9d9fc6d9-ql84s\" (UID: \"935a9115-0077-4a70-a6d7-cd9ba80fec28\") " pod="openstack/dnsmasq-dns-6d9d9fc6d9-ql84s" Oct 14 08:51:32 crc kubenswrapper[4870]: I1014 08:51:32.482176 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/935a9115-0077-4a70-a6d7-cd9ba80fec28-ovsdbserver-nb\") pod \"dnsmasq-dns-6d9d9fc6d9-ql84s\" (UID: \"935a9115-0077-4a70-a6d7-cd9ba80fec28\") " pod="openstack/dnsmasq-dns-6d9d9fc6d9-ql84s" Oct 14 08:51:32 crc kubenswrapper[4870]: I1014 08:51:32.482236 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/935a9115-0077-4a70-a6d7-cd9ba80fec28-dns-svc\") pod \"dnsmasq-dns-6d9d9fc6d9-ql84s\" (UID: \"935a9115-0077-4a70-a6d7-cd9ba80fec28\") " pod="openstack/dnsmasq-dns-6d9d9fc6d9-ql84s" Oct 14 08:51:32 crc kubenswrapper[4870]: I1014 08:51:32.482267 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/935a9115-0077-4a70-a6d7-cd9ba80fec28-config\") pod \"dnsmasq-dns-6d9d9fc6d9-ql84s\" (UID: \"935a9115-0077-4a70-a6d7-cd9ba80fec28\") " pod="openstack/dnsmasq-dns-6d9d9fc6d9-ql84s" Oct 14 08:51:32 crc kubenswrapper[4870]: I1014 08:51:32.483263 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/935a9115-0077-4a70-a6d7-cd9ba80fec28-dns-svc\") pod \"dnsmasq-dns-6d9d9fc6d9-ql84s\" (UID: \"935a9115-0077-4a70-a6d7-cd9ba80fec28\") " pod="openstack/dnsmasq-dns-6d9d9fc6d9-ql84s" Oct 14 08:51:32 crc kubenswrapper[4870]: I1014 08:51:32.483275 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/935a9115-0077-4a70-a6d7-cd9ba80fec28-config\") pod \"dnsmasq-dns-6d9d9fc6d9-ql84s\" (UID: \"935a9115-0077-4a70-a6d7-cd9ba80fec28\") " pod="openstack/dnsmasq-dns-6d9d9fc6d9-ql84s" Oct 14 08:51:32 crc kubenswrapper[4870]: I1014 08:51:32.483276 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/935a9115-0077-4a70-a6d7-cd9ba80fec28-ovsdbserver-nb\") pod \"dnsmasq-dns-6d9d9fc6d9-ql84s\" (UID: \"935a9115-0077-4a70-a6d7-cd9ba80fec28\") " pod="openstack/dnsmasq-dns-6d9d9fc6d9-ql84s" Oct 14 08:51:32 crc kubenswrapper[4870]: I1014 08:51:32.507769 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vqmlv\" (UniqueName: \"kubernetes.io/projected/935a9115-0077-4a70-a6d7-cd9ba80fec28-kube-api-access-vqmlv\") pod \"dnsmasq-dns-6d9d9fc6d9-ql84s\" (UID: \"935a9115-0077-4a70-a6d7-cd9ba80fec28\") " pod="openstack/dnsmasq-dns-6d9d9fc6d9-ql84s" Oct 14 08:51:32 crc kubenswrapper[4870]: I1014 08:51:32.610744 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6d9d9fc6d9-ql84s"] Oct 14 08:51:32 crc kubenswrapper[4870]: I1014 08:51:32.611510 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d9d9fc6d9-ql84s" Oct 14 08:51:32 crc kubenswrapper[4870]: I1014 08:51:32.637065 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7cf94d58c5-f8h59"] Oct 14 08:51:32 crc kubenswrapper[4870]: I1014 08:51:32.639332 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7cf94d58c5-f8h59" Oct 14 08:51:32 crc kubenswrapper[4870]: I1014 08:51:32.641796 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Oct 14 08:51:32 crc kubenswrapper[4870]: I1014 08:51:32.655368 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7cf94d58c5-f8h59"] Oct 14 08:51:32 crc kubenswrapper[4870]: I1014 08:51:32.688739 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/df5834a5-3b83-48c0-b83f-10e97d10a7d6-dns-svc\") pod \"dnsmasq-dns-7cf94d58c5-f8h59\" (UID: \"df5834a5-3b83-48c0-b83f-10e97d10a7d6\") " pod="openstack/dnsmasq-dns-7cf94d58c5-f8h59" Oct 14 08:51:32 crc kubenswrapper[4870]: I1014 08:51:32.688794 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/df5834a5-3b83-48c0-b83f-10e97d10a7d6-ovsdbserver-nb\") pod \"dnsmasq-dns-7cf94d58c5-f8h59\" (UID: \"df5834a5-3b83-48c0-b83f-10e97d10a7d6\") " pod="openstack/dnsmasq-dns-7cf94d58c5-f8h59" Oct 14 08:51:32 crc kubenswrapper[4870]: I1014 08:51:32.688971 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/df5834a5-3b83-48c0-b83f-10e97d10a7d6-ovsdbserver-sb\") pod \"dnsmasq-dns-7cf94d58c5-f8h59\" (UID: \"df5834a5-3b83-48c0-b83f-10e97d10a7d6\") " pod="openstack/dnsmasq-dns-7cf94d58c5-f8h59" Oct 14 08:51:32 crc kubenswrapper[4870]: I1014 08:51:32.689138 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xct86\" (UniqueName: \"kubernetes.io/projected/df5834a5-3b83-48c0-b83f-10e97d10a7d6-kube-api-access-xct86\") pod \"dnsmasq-dns-7cf94d58c5-f8h59\" (UID: \"df5834a5-3b83-48c0-b83f-10e97d10a7d6\") " pod="openstack/dnsmasq-dns-7cf94d58c5-f8h59" Oct 14 08:51:32 crc kubenswrapper[4870]: I1014 08:51:32.689274 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/df5834a5-3b83-48c0-b83f-10e97d10a7d6-config\") pod \"dnsmasq-dns-7cf94d58c5-f8h59\" (UID: \"df5834a5-3b83-48c0-b83f-10e97d10a7d6\") " pod="openstack/dnsmasq-dns-7cf94d58c5-f8h59" Oct 14 08:51:32 crc kubenswrapper[4870]: I1014 08:51:32.790916 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xct86\" (UniqueName: \"kubernetes.io/projected/df5834a5-3b83-48c0-b83f-10e97d10a7d6-kube-api-access-xct86\") pod \"dnsmasq-dns-7cf94d58c5-f8h59\" (UID: \"df5834a5-3b83-48c0-b83f-10e97d10a7d6\") " pod="openstack/dnsmasq-dns-7cf94d58c5-f8h59" Oct 14 08:51:32 crc kubenswrapper[4870]: I1014 08:51:32.791052 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/df5834a5-3b83-48c0-b83f-10e97d10a7d6-config\") pod \"dnsmasq-dns-7cf94d58c5-f8h59\" (UID: \"df5834a5-3b83-48c0-b83f-10e97d10a7d6\") " pod="openstack/dnsmasq-dns-7cf94d58c5-f8h59" Oct 14 08:51:32 crc kubenswrapper[4870]: I1014 08:51:32.791098 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/df5834a5-3b83-48c0-b83f-10e97d10a7d6-dns-svc\") pod \"dnsmasq-dns-7cf94d58c5-f8h59\" (UID: \"df5834a5-3b83-48c0-b83f-10e97d10a7d6\") " pod="openstack/dnsmasq-dns-7cf94d58c5-f8h59" Oct 14 08:51:32 crc kubenswrapper[4870]: I1014 08:51:32.791158 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/df5834a5-3b83-48c0-b83f-10e97d10a7d6-ovsdbserver-nb\") pod \"dnsmasq-dns-7cf94d58c5-f8h59\" (UID: \"df5834a5-3b83-48c0-b83f-10e97d10a7d6\") " pod="openstack/dnsmasq-dns-7cf94d58c5-f8h59" Oct 14 08:51:32 crc kubenswrapper[4870]: I1014 08:51:32.791227 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/df5834a5-3b83-48c0-b83f-10e97d10a7d6-ovsdbserver-sb\") pod \"dnsmasq-dns-7cf94d58c5-f8h59\" (UID: \"df5834a5-3b83-48c0-b83f-10e97d10a7d6\") " pod="openstack/dnsmasq-dns-7cf94d58c5-f8h59" Oct 14 08:51:32 crc kubenswrapper[4870]: I1014 08:51:32.792292 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/df5834a5-3b83-48c0-b83f-10e97d10a7d6-dns-svc\") pod \"dnsmasq-dns-7cf94d58c5-f8h59\" (UID: \"df5834a5-3b83-48c0-b83f-10e97d10a7d6\") " pod="openstack/dnsmasq-dns-7cf94d58c5-f8h59" Oct 14 08:51:32 crc kubenswrapper[4870]: I1014 08:51:32.792957 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/df5834a5-3b83-48c0-b83f-10e97d10a7d6-config\") pod \"dnsmasq-dns-7cf94d58c5-f8h59\" (UID: \"df5834a5-3b83-48c0-b83f-10e97d10a7d6\") " pod="openstack/dnsmasq-dns-7cf94d58c5-f8h59" Oct 14 08:51:32 crc kubenswrapper[4870]: I1014 08:51:32.794757 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/df5834a5-3b83-48c0-b83f-10e97d10a7d6-ovsdbserver-nb\") pod \"dnsmasq-dns-7cf94d58c5-f8h59\" (UID: \"df5834a5-3b83-48c0-b83f-10e97d10a7d6\") " pod="openstack/dnsmasq-dns-7cf94d58c5-f8h59" Oct 14 08:51:32 crc kubenswrapper[4870]: I1014 08:51:32.794809 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/df5834a5-3b83-48c0-b83f-10e97d10a7d6-ovsdbserver-sb\") pod \"dnsmasq-dns-7cf94d58c5-f8h59\" (UID: \"df5834a5-3b83-48c0-b83f-10e97d10a7d6\") " pod="openstack/dnsmasq-dns-7cf94d58c5-f8h59" Oct 14 08:51:32 crc kubenswrapper[4870]: I1014 08:51:32.823092 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xct86\" (UniqueName: \"kubernetes.io/projected/df5834a5-3b83-48c0-b83f-10e97d10a7d6-kube-api-access-xct86\") pod \"dnsmasq-dns-7cf94d58c5-f8h59\" (UID: \"df5834a5-3b83-48c0-b83f-10e97d10a7d6\") " pod="openstack/dnsmasq-dns-7cf94d58c5-f8h59" Oct 14 08:51:33 crc kubenswrapper[4870]: I1014 08:51:33.001223 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7cf94d58c5-f8h59" Oct 14 08:51:33 crc kubenswrapper[4870]: I1014 08:51:33.069161 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6d9d9fc6d9-ql84s"] Oct 14 08:51:33 crc kubenswrapper[4870]: W1014 08:51:33.079800 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod935a9115_0077_4a70_a6d7_cd9ba80fec28.slice/crio-41c02e5607de58d660e450e0c3b3d9ddbdc08eece7b97d3ecd27bc59c5099f7d WatchSource:0}: Error finding container 41c02e5607de58d660e450e0c3b3d9ddbdc08eece7b97d3ecd27bc59c5099f7d: Status 404 returned error can't find the container with id 41c02e5607de58d660e450e0c3b3d9ddbdc08eece7b97d3ecd27bc59c5099f7d Oct 14 08:51:33 crc kubenswrapper[4870]: I1014 08:51:33.108561 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d9d9fc6d9-ql84s" event={"ID":"935a9115-0077-4a70-a6d7-cd9ba80fec28","Type":"ContainerStarted","Data":"41c02e5607de58d660e450e0c3b3d9ddbdc08eece7b97d3ecd27bc59c5099f7d"} Oct 14 08:51:33 crc kubenswrapper[4870]: I1014 08:51:33.242145 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7cf94d58c5-f8h59"] Oct 14 08:51:33 crc kubenswrapper[4870]: W1014 08:51:33.243186 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddf5834a5_3b83_48c0_b83f_10e97d10a7d6.slice/crio-2aeb3be9263ddcb47659f5c6b4d75c4f5b5c649dd1ccd7c97c91990e789fbb95 WatchSource:0}: Error finding container 2aeb3be9263ddcb47659f5c6b4d75c4f5b5c649dd1ccd7c97c91990e789fbb95: Status 404 returned error can't find the container with id 2aeb3be9263ddcb47659f5c6b4d75c4f5b5c649dd1ccd7c97c91990e789fbb95 Oct 14 08:51:34 crc kubenswrapper[4870]: I1014 08:51:34.115851 4870 generic.go:334] "Generic (PLEG): container finished" podID="935a9115-0077-4a70-a6d7-cd9ba80fec28" containerID="e27b7dc7c8f47d651412a88d9e8318d7d9b67fdca7f279a48d84265f27eec35a" exitCode=0 Oct 14 08:51:34 crc kubenswrapper[4870]: I1014 08:51:34.115930 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d9d9fc6d9-ql84s" event={"ID":"935a9115-0077-4a70-a6d7-cd9ba80fec28","Type":"ContainerDied","Data":"e27b7dc7c8f47d651412a88d9e8318d7d9b67fdca7f279a48d84265f27eec35a"} Oct 14 08:51:34 crc kubenswrapper[4870]: I1014 08:51:34.118930 4870 generic.go:334] "Generic (PLEG): container finished" podID="df5834a5-3b83-48c0-b83f-10e97d10a7d6" containerID="90837e89ec3a4b7d740d83eedfc117f7917a1ce9f7580a83b8860970008e99c5" exitCode=0 Oct 14 08:51:34 crc kubenswrapper[4870]: I1014 08:51:34.119100 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cf94d58c5-f8h59" event={"ID":"df5834a5-3b83-48c0-b83f-10e97d10a7d6","Type":"ContainerDied","Data":"90837e89ec3a4b7d740d83eedfc117f7917a1ce9f7580a83b8860970008e99c5"} Oct 14 08:51:34 crc kubenswrapper[4870]: I1014 08:51:34.119147 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cf94d58c5-f8h59" event={"ID":"df5834a5-3b83-48c0-b83f-10e97d10a7d6","Type":"ContainerStarted","Data":"2aeb3be9263ddcb47659f5c6b4d75c4f5b5c649dd1ccd7c97c91990e789fbb95"} Oct 14 08:51:34 crc kubenswrapper[4870]: I1014 08:51:34.410045 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d9d9fc6d9-ql84s" Oct 14 08:51:34 crc kubenswrapper[4870]: I1014 08:51:34.524027 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/935a9115-0077-4a70-a6d7-cd9ba80fec28-ovsdbserver-nb\") pod \"935a9115-0077-4a70-a6d7-cd9ba80fec28\" (UID: \"935a9115-0077-4a70-a6d7-cd9ba80fec28\") " Oct 14 08:51:34 crc kubenswrapper[4870]: I1014 08:51:34.524158 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/935a9115-0077-4a70-a6d7-cd9ba80fec28-dns-svc\") pod \"935a9115-0077-4a70-a6d7-cd9ba80fec28\" (UID: \"935a9115-0077-4a70-a6d7-cd9ba80fec28\") " Oct 14 08:51:34 crc kubenswrapper[4870]: I1014 08:51:34.524307 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/935a9115-0077-4a70-a6d7-cd9ba80fec28-config\") pod \"935a9115-0077-4a70-a6d7-cd9ba80fec28\" (UID: \"935a9115-0077-4a70-a6d7-cd9ba80fec28\") " Oct 14 08:51:34 crc kubenswrapper[4870]: I1014 08:51:34.524348 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vqmlv\" (UniqueName: \"kubernetes.io/projected/935a9115-0077-4a70-a6d7-cd9ba80fec28-kube-api-access-vqmlv\") pod \"935a9115-0077-4a70-a6d7-cd9ba80fec28\" (UID: \"935a9115-0077-4a70-a6d7-cd9ba80fec28\") " Oct 14 08:51:34 crc kubenswrapper[4870]: I1014 08:51:34.530644 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/935a9115-0077-4a70-a6d7-cd9ba80fec28-kube-api-access-vqmlv" (OuterVolumeSpecName: "kube-api-access-vqmlv") pod "935a9115-0077-4a70-a6d7-cd9ba80fec28" (UID: "935a9115-0077-4a70-a6d7-cd9ba80fec28"). InnerVolumeSpecName "kube-api-access-vqmlv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:51:34 crc kubenswrapper[4870]: I1014 08:51:34.566108 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/935a9115-0077-4a70-a6d7-cd9ba80fec28-config" (OuterVolumeSpecName: "config") pod "935a9115-0077-4a70-a6d7-cd9ba80fec28" (UID: "935a9115-0077-4a70-a6d7-cd9ba80fec28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:51:34 crc kubenswrapper[4870]: I1014 08:51:34.569962 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/935a9115-0077-4a70-a6d7-cd9ba80fec28-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "935a9115-0077-4a70-a6d7-cd9ba80fec28" (UID: "935a9115-0077-4a70-a6d7-cd9ba80fec28"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:51:34 crc kubenswrapper[4870]: I1014 08:51:34.575827 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/935a9115-0077-4a70-a6d7-cd9ba80fec28-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "935a9115-0077-4a70-a6d7-cd9ba80fec28" (UID: "935a9115-0077-4a70-a6d7-cd9ba80fec28"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:51:34 crc kubenswrapper[4870]: I1014 08:51:34.626656 4870 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/935a9115-0077-4a70-a6d7-cd9ba80fec28-config\") on node \"crc\" DevicePath \"\"" Oct 14 08:51:34 crc kubenswrapper[4870]: I1014 08:51:34.626687 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vqmlv\" (UniqueName: \"kubernetes.io/projected/935a9115-0077-4a70-a6d7-cd9ba80fec28-kube-api-access-vqmlv\") on node \"crc\" DevicePath \"\"" Oct 14 08:51:34 crc kubenswrapper[4870]: I1014 08:51:34.626700 4870 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/935a9115-0077-4a70-a6d7-cd9ba80fec28-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 14 08:51:34 crc kubenswrapper[4870]: I1014 08:51:34.626708 4870 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/935a9115-0077-4a70-a6d7-cd9ba80fec28-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 14 08:51:35 crc kubenswrapper[4870]: I1014 08:51:35.136179 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d9d9fc6d9-ql84s" event={"ID":"935a9115-0077-4a70-a6d7-cd9ba80fec28","Type":"ContainerDied","Data":"41c02e5607de58d660e450e0c3b3d9ddbdc08eece7b97d3ecd27bc59c5099f7d"} Oct 14 08:51:35 crc kubenswrapper[4870]: I1014 08:51:35.136252 4870 scope.go:117] "RemoveContainer" containerID="e27b7dc7c8f47d651412a88d9e8318d7d9b67fdca7f279a48d84265f27eec35a" Oct 14 08:51:35 crc kubenswrapper[4870]: I1014 08:51:35.136379 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d9d9fc6d9-ql84s" Oct 14 08:51:35 crc kubenswrapper[4870]: I1014 08:51:35.140001 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cf94d58c5-f8h59" event={"ID":"df5834a5-3b83-48c0-b83f-10e97d10a7d6","Type":"ContainerStarted","Data":"51a12e0a1aedb158605abc7125b15a1415e952e0919d0aeec5c10cb86b235ec9"} Oct 14 08:51:35 crc kubenswrapper[4870]: I1014 08:51:35.140823 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7cf94d58c5-f8h59" Oct 14 08:51:35 crc kubenswrapper[4870]: I1014 08:51:35.176383 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7cf94d58c5-f8h59" podStartSLOduration=3.176319912 podStartE2EDuration="3.176319912s" podCreationTimestamp="2025-10-14 08:51:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 08:51:35.161375256 +0000 UTC m=+6630.858735717" watchObservedRunningTime="2025-10-14 08:51:35.176319912 +0000 UTC m=+6630.873680333" Oct 14 08:51:35 crc kubenswrapper[4870]: I1014 08:51:35.231505 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6d9d9fc6d9-ql84s"] Oct 14 08:51:35 crc kubenswrapper[4870]: I1014 08:51:35.240616 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6d9d9fc6d9-ql84s"] Oct 14 08:51:35 crc kubenswrapper[4870]: I1014 08:51:35.416064 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-copy-data"] Oct 14 08:51:35 crc kubenswrapper[4870]: E1014 08:51:35.416599 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="935a9115-0077-4a70-a6d7-cd9ba80fec28" containerName="init" Oct 14 08:51:35 crc kubenswrapper[4870]: I1014 08:51:35.416629 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="935a9115-0077-4a70-a6d7-cd9ba80fec28" containerName="init" Oct 14 08:51:35 crc kubenswrapper[4870]: I1014 08:51:35.416968 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="935a9115-0077-4a70-a6d7-cd9ba80fec28" containerName="init" Oct 14 08:51:35 crc kubenswrapper[4870]: I1014 08:51:35.417930 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-copy-data" Oct 14 08:51:35 crc kubenswrapper[4870]: I1014 08:51:35.422081 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovn-data-cert" Oct 14 08:51:35 crc kubenswrapper[4870]: I1014 08:51:35.429654 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-copy-data"] Oct 14 08:51:35 crc kubenswrapper[4870]: I1014 08:51:35.550692 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-64243ea5-9586-429f-ab5b-4806ef5e2614\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-64243ea5-9586-429f-ab5b-4806ef5e2614\") pod \"ovn-copy-data\" (UID: \"3cfdfc3c-8b85-4145-98ad-64c9f91c6860\") " pod="openstack/ovn-copy-data" Oct 14 08:51:35 crc kubenswrapper[4870]: I1014 08:51:35.551130 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f5zx4\" (UniqueName: \"kubernetes.io/projected/3cfdfc3c-8b85-4145-98ad-64c9f91c6860-kube-api-access-f5zx4\") pod \"ovn-copy-data\" (UID: \"3cfdfc3c-8b85-4145-98ad-64c9f91c6860\") " pod="openstack/ovn-copy-data" Oct 14 08:51:35 crc kubenswrapper[4870]: I1014 08:51:35.551207 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-data-cert\" (UniqueName: \"kubernetes.io/secret/3cfdfc3c-8b85-4145-98ad-64c9f91c6860-ovn-data-cert\") pod \"ovn-copy-data\" (UID: \"3cfdfc3c-8b85-4145-98ad-64c9f91c6860\") " pod="openstack/ovn-copy-data" Oct 14 08:51:35 crc kubenswrapper[4870]: I1014 08:51:35.652860 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f5zx4\" (UniqueName: \"kubernetes.io/projected/3cfdfc3c-8b85-4145-98ad-64c9f91c6860-kube-api-access-f5zx4\") pod \"ovn-copy-data\" (UID: \"3cfdfc3c-8b85-4145-98ad-64c9f91c6860\") " pod="openstack/ovn-copy-data" Oct 14 08:51:35 crc kubenswrapper[4870]: I1014 08:51:35.652959 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-data-cert\" (UniqueName: \"kubernetes.io/secret/3cfdfc3c-8b85-4145-98ad-64c9f91c6860-ovn-data-cert\") pod \"ovn-copy-data\" (UID: \"3cfdfc3c-8b85-4145-98ad-64c9f91c6860\") " pod="openstack/ovn-copy-data" Oct 14 08:51:35 crc kubenswrapper[4870]: I1014 08:51:35.653250 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-64243ea5-9586-429f-ab5b-4806ef5e2614\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-64243ea5-9586-429f-ab5b-4806ef5e2614\") pod \"ovn-copy-data\" (UID: \"3cfdfc3c-8b85-4145-98ad-64c9f91c6860\") " pod="openstack/ovn-copy-data" Oct 14 08:51:35 crc kubenswrapper[4870]: I1014 08:51:35.661067 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-data-cert\" (UniqueName: \"kubernetes.io/secret/3cfdfc3c-8b85-4145-98ad-64c9f91c6860-ovn-data-cert\") pod \"ovn-copy-data\" (UID: \"3cfdfc3c-8b85-4145-98ad-64c9f91c6860\") " pod="openstack/ovn-copy-data" Oct 14 08:51:35 crc kubenswrapper[4870]: I1014 08:51:35.669867 4870 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 14 08:51:35 crc kubenswrapper[4870]: I1014 08:51:35.669920 4870 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-64243ea5-9586-429f-ab5b-4806ef5e2614\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-64243ea5-9586-429f-ab5b-4806ef5e2614\") pod \"ovn-copy-data\" (UID: \"3cfdfc3c-8b85-4145-98ad-64c9f91c6860\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/9e074986601915548681b2609670371ae5061e548a59c98bab2f24dd3c0cc11d/globalmount\"" pod="openstack/ovn-copy-data" Oct 14 08:51:35 crc kubenswrapper[4870]: I1014 08:51:35.671353 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f5zx4\" (UniqueName: \"kubernetes.io/projected/3cfdfc3c-8b85-4145-98ad-64c9f91c6860-kube-api-access-f5zx4\") pod \"ovn-copy-data\" (UID: \"3cfdfc3c-8b85-4145-98ad-64c9f91c6860\") " pod="openstack/ovn-copy-data" Oct 14 08:51:35 crc kubenswrapper[4870]: I1014 08:51:35.727551 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-64243ea5-9586-429f-ab5b-4806ef5e2614\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-64243ea5-9586-429f-ab5b-4806ef5e2614\") pod \"ovn-copy-data\" (UID: \"3cfdfc3c-8b85-4145-98ad-64c9f91c6860\") " pod="openstack/ovn-copy-data" Oct 14 08:51:35 crc kubenswrapper[4870]: I1014 08:51:35.740108 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-copy-data" Oct 14 08:51:36 crc kubenswrapper[4870]: I1014 08:51:36.296652 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-copy-data"] Oct 14 08:51:36 crc kubenswrapper[4870]: W1014 08:51:36.309563 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3cfdfc3c_8b85_4145_98ad_64c9f91c6860.slice/crio-e650e748983644c43cb483fe100167fe33ff6c31e22306a1be4c4a77e889b141 WatchSource:0}: Error finding container e650e748983644c43cb483fe100167fe33ff6c31e22306a1be4c4a77e889b141: Status 404 returned error can't find the container with id e650e748983644c43cb483fe100167fe33ff6c31e22306a1be4c4a77e889b141 Oct 14 08:51:37 crc kubenswrapper[4870]: I1014 08:51:37.051797 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="935a9115-0077-4a70-a6d7-cd9ba80fec28" path="/var/lib/kubelet/pods/935a9115-0077-4a70-a6d7-cd9ba80fec28/volumes" Oct 14 08:51:37 crc kubenswrapper[4870]: I1014 08:51:37.162545 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-copy-data" event={"ID":"3cfdfc3c-8b85-4145-98ad-64c9f91c6860","Type":"ContainerStarted","Data":"e4e6ab17398f4be1ba400d1713cb3dbe0d2c2b8ecdebc9c067675e1465fa9fff"} Oct 14 08:51:37 crc kubenswrapper[4870]: I1014 08:51:37.162634 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-copy-data" event={"ID":"3cfdfc3c-8b85-4145-98ad-64c9f91c6860","Type":"ContainerStarted","Data":"e650e748983644c43cb483fe100167fe33ff6c31e22306a1be4c4a77e889b141"} Oct 14 08:51:37 crc kubenswrapper[4870]: I1014 08:51:37.182291 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-copy-data" podStartSLOduration=2.9878669479999997 podStartE2EDuration="3.182261086s" podCreationTimestamp="2025-10-14 08:51:34 +0000 UTC" firstStartedPulling="2025-10-14 08:51:36.312116901 +0000 UTC m=+6632.009477272" lastFinishedPulling="2025-10-14 08:51:36.506511029 +0000 UTC m=+6632.203871410" observedRunningTime="2025-10-14 08:51:37.182260206 +0000 UTC m=+6632.879620617" watchObservedRunningTime="2025-10-14 08:51:37.182261086 +0000 UTC m=+6632.879621497" Oct 14 08:51:43 crc kubenswrapper[4870]: I1014 08:51:43.002765 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7cf94d58c5-f8h59" Oct 14 08:51:43 crc kubenswrapper[4870]: I1014 08:51:43.096239 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7ccc84877c-2ngf9"] Oct 14 08:51:43 crc kubenswrapper[4870]: I1014 08:51:43.096495 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7ccc84877c-2ngf9" podUID="920f17fd-fc08-4a5e-9e7e-b641cc11528a" containerName="dnsmasq-dns" containerID="cri-o://d1f99638f9a135316382888657472849639a8a7ae1413b7da513e34c1354d997" gracePeriod=10 Oct 14 08:51:43 crc kubenswrapper[4870]: I1014 08:51:43.233971 4870 generic.go:334] "Generic (PLEG): container finished" podID="920f17fd-fc08-4a5e-9e7e-b641cc11528a" containerID="d1f99638f9a135316382888657472849639a8a7ae1413b7da513e34c1354d997" exitCode=0 Oct 14 08:51:43 crc kubenswrapper[4870]: I1014 08:51:43.234009 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7ccc84877c-2ngf9" event={"ID":"920f17fd-fc08-4a5e-9e7e-b641cc11528a","Type":"ContainerDied","Data":"d1f99638f9a135316382888657472849639a8a7ae1413b7da513e34c1354d997"} Oct 14 08:51:43 crc kubenswrapper[4870]: I1014 08:51:43.547968 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7ccc84877c-2ngf9" Oct 14 08:51:43 crc kubenswrapper[4870]: I1014 08:51:43.630223 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rj2jz\" (UniqueName: \"kubernetes.io/projected/920f17fd-fc08-4a5e-9e7e-b641cc11528a-kube-api-access-rj2jz\") pod \"920f17fd-fc08-4a5e-9e7e-b641cc11528a\" (UID: \"920f17fd-fc08-4a5e-9e7e-b641cc11528a\") " Oct 14 08:51:43 crc kubenswrapper[4870]: I1014 08:51:43.630390 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/920f17fd-fc08-4a5e-9e7e-b641cc11528a-config\") pod \"920f17fd-fc08-4a5e-9e7e-b641cc11528a\" (UID: \"920f17fd-fc08-4a5e-9e7e-b641cc11528a\") " Oct 14 08:51:43 crc kubenswrapper[4870]: I1014 08:51:43.630432 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/920f17fd-fc08-4a5e-9e7e-b641cc11528a-dns-svc\") pod \"920f17fd-fc08-4a5e-9e7e-b641cc11528a\" (UID: \"920f17fd-fc08-4a5e-9e7e-b641cc11528a\") " Oct 14 08:51:43 crc kubenswrapper[4870]: I1014 08:51:43.637696 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/920f17fd-fc08-4a5e-9e7e-b641cc11528a-kube-api-access-rj2jz" (OuterVolumeSpecName: "kube-api-access-rj2jz") pod "920f17fd-fc08-4a5e-9e7e-b641cc11528a" (UID: "920f17fd-fc08-4a5e-9e7e-b641cc11528a"). InnerVolumeSpecName "kube-api-access-rj2jz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:51:43 crc kubenswrapper[4870]: I1014 08:51:43.666328 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/920f17fd-fc08-4a5e-9e7e-b641cc11528a-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "920f17fd-fc08-4a5e-9e7e-b641cc11528a" (UID: "920f17fd-fc08-4a5e-9e7e-b641cc11528a"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:51:43 crc kubenswrapper[4870]: I1014 08:51:43.681280 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/920f17fd-fc08-4a5e-9e7e-b641cc11528a-config" (OuterVolumeSpecName: "config") pod "920f17fd-fc08-4a5e-9e7e-b641cc11528a" (UID: "920f17fd-fc08-4a5e-9e7e-b641cc11528a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:51:43 crc kubenswrapper[4870]: I1014 08:51:43.732062 4870 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/920f17fd-fc08-4a5e-9e7e-b641cc11528a-config\") on node \"crc\" DevicePath \"\"" Oct 14 08:51:43 crc kubenswrapper[4870]: I1014 08:51:43.732102 4870 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/920f17fd-fc08-4a5e-9e7e-b641cc11528a-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 14 08:51:43 crc kubenswrapper[4870]: I1014 08:51:43.732115 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rj2jz\" (UniqueName: \"kubernetes.io/projected/920f17fd-fc08-4a5e-9e7e-b641cc11528a-kube-api-access-rj2jz\") on node \"crc\" DevicePath \"\"" Oct 14 08:51:44 crc kubenswrapper[4870]: I1014 08:51:44.253777 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7ccc84877c-2ngf9" event={"ID":"920f17fd-fc08-4a5e-9e7e-b641cc11528a","Type":"ContainerDied","Data":"83a45ecc32ec34f4626f3e65f0dbe7d5ac105a88a6aa1f4dd3d3c217b46b0ea0"} Oct 14 08:51:44 crc kubenswrapper[4870]: I1014 08:51:44.253875 4870 scope.go:117] "RemoveContainer" containerID="d1f99638f9a135316382888657472849639a8a7ae1413b7da513e34c1354d997" Oct 14 08:51:44 crc kubenswrapper[4870]: I1014 08:51:44.253927 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7ccc84877c-2ngf9" Oct 14 08:51:44 crc kubenswrapper[4870]: I1014 08:51:44.297792 4870 scope.go:117] "RemoveContainer" containerID="24710d7457adf6e2a83521401b65bb4a8a96f3a42e3ac761d2a2eb94b1acb0a2" Oct 14 08:51:44 crc kubenswrapper[4870]: I1014 08:51:44.305569 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7ccc84877c-2ngf9"] Oct 14 08:51:44 crc kubenswrapper[4870]: I1014 08:51:44.314415 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7ccc84877c-2ngf9"] Oct 14 08:51:45 crc kubenswrapper[4870]: I1014 08:51:45.065351 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="920f17fd-fc08-4a5e-9e7e-b641cc11528a" path="/var/lib/kubelet/pods/920f17fd-fc08-4a5e-9e7e-b641cc11528a/volumes" Oct 14 08:51:45 crc kubenswrapper[4870]: I1014 08:51:45.357021 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Oct 14 08:51:45 crc kubenswrapper[4870]: E1014 08:51:45.357555 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="920f17fd-fc08-4a5e-9e7e-b641cc11528a" containerName="dnsmasq-dns" Oct 14 08:51:45 crc kubenswrapper[4870]: I1014 08:51:45.357577 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="920f17fd-fc08-4a5e-9e7e-b641cc11528a" containerName="dnsmasq-dns" Oct 14 08:51:45 crc kubenswrapper[4870]: E1014 08:51:45.357598 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="920f17fd-fc08-4a5e-9e7e-b641cc11528a" containerName="init" Oct 14 08:51:45 crc kubenswrapper[4870]: I1014 08:51:45.357608 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="920f17fd-fc08-4a5e-9e7e-b641cc11528a" containerName="init" Oct 14 08:51:45 crc kubenswrapper[4870]: I1014 08:51:45.357888 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="920f17fd-fc08-4a5e-9e7e-b641cc11528a" containerName="dnsmasq-dns" Oct 14 08:51:45 crc kubenswrapper[4870]: I1014 08:51:45.359224 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Oct 14 08:51:45 crc kubenswrapper[4870]: I1014 08:51:45.363703 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Oct 14 08:51:45 crc kubenswrapper[4870]: I1014 08:51:45.363734 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-2jphl" Oct 14 08:51:45 crc kubenswrapper[4870]: I1014 08:51:45.366661 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Oct 14 08:51:45 crc kubenswrapper[4870]: I1014 08:51:45.376161 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Oct 14 08:51:45 crc kubenswrapper[4870]: I1014 08:51:45.480233 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8bd2488-7e3a-4dc2-87c9-dbaf25141f92-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"c8bd2488-7e3a-4dc2-87c9-dbaf25141f92\") " pod="openstack/ovn-northd-0" Oct 14 08:51:45 crc kubenswrapper[4870]: I1014 08:51:45.480362 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c8bd2488-7e3a-4dc2-87c9-dbaf25141f92-scripts\") pod \"ovn-northd-0\" (UID: \"c8bd2488-7e3a-4dc2-87c9-dbaf25141f92\") " pod="openstack/ovn-northd-0" Oct 14 08:51:45 crc kubenswrapper[4870]: I1014 08:51:45.480554 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pgtkx\" (UniqueName: \"kubernetes.io/projected/c8bd2488-7e3a-4dc2-87c9-dbaf25141f92-kube-api-access-pgtkx\") pod \"ovn-northd-0\" (UID: \"c8bd2488-7e3a-4dc2-87c9-dbaf25141f92\") " pod="openstack/ovn-northd-0" Oct 14 08:51:45 crc kubenswrapper[4870]: I1014 08:51:45.480858 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/c8bd2488-7e3a-4dc2-87c9-dbaf25141f92-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"c8bd2488-7e3a-4dc2-87c9-dbaf25141f92\") " pod="openstack/ovn-northd-0" Oct 14 08:51:45 crc kubenswrapper[4870]: I1014 08:51:45.480915 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c8bd2488-7e3a-4dc2-87c9-dbaf25141f92-config\") pod \"ovn-northd-0\" (UID: \"c8bd2488-7e3a-4dc2-87c9-dbaf25141f92\") " pod="openstack/ovn-northd-0" Oct 14 08:51:45 crc kubenswrapper[4870]: I1014 08:51:45.582372 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/c8bd2488-7e3a-4dc2-87c9-dbaf25141f92-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"c8bd2488-7e3a-4dc2-87c9-dbaf25141f92\") " pod="openstack/ovn-northd-0" Oct 14 08:51:45 crc kubenswrapper[4870]: I1014 08:51:45.582423 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c8bd2488-7e3a-4dc2-87c9-dbaf25141f92-config\") pod \"ovn-northd-0\" (UID: \"c8bd2488-7e3a-4dc2-87c9-dbaf25141f92\") " pod="openstack/ovn-northd-0" Oct 14 08:51:45 crc kubenswrapper[4870]: I1014 08:51:45.582487 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8bd2488-7e3a-4dc2-87c9-dbaf25141f92-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"c8bd2488-7e3a-4dc2-87c9-dbaf25141f92\") " pod="openstack/ovn-northd-0" Oct 14 08:51:45 crc kubenswrapper[4870]: I1014 08:51:45.582515 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c8bd2488-7e3a-4dc2-87c9-dbaf25141f92-scripts\") pod \"ovn-northd-0\" (UID: \"c8bd2488-7e3a-4dc2-87c9-dbaf25141f92\") " pod="openstack/ovn-northd-0" Oct 14 08:51:45 crc kubenswrapper[4870]: I1014 08:51:45.582550 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pgtkx\" (UniqueName: \"kubernetes.io/projected/c8bd2488-7e3a-4dc2-87c9-dbaf25141f92-kube-api-access-pgtkx\") pod \"ovn-northd-0\" (UID: \"c8bd2488-7e3a-4dc2-87c9-dbaf25141f92\") " pod="openstack/ovn-northd-0" Oct 14 08:51:45 crc kubenswrapper[4870]: I1014 08:51:45.583053 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/c8bd2488-7e3a-4dc2-87c9-dbaf25141f92-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"c8bd2488-7e3a-4dc2-87c9-dbaf25141f92\") " pod="openstack/ovn-northd-0" Oct 14 08:51:45 crc kubenswrapper[4870]: I1014 08:51:45.583619 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c8bd2488-7e3a-4dc2-87c9-dbaf25141f92-config\") pod \"ovn-northd-0\" (UID: \"c8bd2488-7e3a-4dc2-87c9-dbaf25141f92\") " pod="openstack/ovn-northd-0" Oct 14 08:51:45 crc kubenswrapper[4870]: I1014 08:51:45.584239 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c8bd2488-7e3a-4dc2-87c9-dbaf25141f92-scripts\") pod \"ovn-northd-0\" (UID: \"c8bd2488-7e3a-4dc2-87c9-dbaf25141f92\") " pod="openstack/ovn-northd-0" Oct 14 08:51:45 crc kubenswrapper[4870]: I1014 08:51:45.589110 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8bd2488-7e3a-4dc2-87c9-dbaf25141f92-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"c8bd2488-7e3a-4dc2-87c9-dbaf25141f92\") " pod="openstack/ovn-northd-0" Oct 14 08:51:45 crc kubenswrapper[4870]: I1014 08:51:45.598870 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pgtkx\" (UniqueName: \"kubernetes.io/projected/c8bd2488-7e3a-4dc2-87c9-dbaf25141f92-kube-api-access-pgtkx\") pod \"ovn-northd-0\" (UID: \"c8bd2488-7e3a-4dc2-87c9-dbaf25141f92\") " pod="openstack/ovn-northd-0" Oct 14 08:51:45 crc kubenswrapper[4870]: I1014 08:51:45.687776 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Oct 14 08:51:46 crc kubenswrapper[4870]: I1014 08:51:46.175184 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Oct 14 08:51:46 crc kubenswrapper[4870]: W1014 08:51:46.176627 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc8bd2488_7e3a_4dc2_87c9_dbaf25141f92.slice/crio-07849914d49d770b6ba944eb0a5b851cf38b7bc831030d1bb86b1bae03f5faad WatchSource:0}: Error finding container 07849914d49d770b6ba944eb0a5b851cf38b7bc831030d1bb86b1bae03f5faad: Status 404 returned error can't find the container with id 07849914d49d770b6ba944eb0a5b851cf38b7bc831030d1bb86b1bae03f5faad Oct 14 08:51:46 crc kubenswrapper[4870]: I1014 08:51:46.275109 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"c8bd2488-7e3a-4dc2-87c9-dbaf25141f92","Type":"ContainerStarted","Data":"07849914d49d770b6ba944eb0a5b851cf38b7bc831030d1bb86b1bae03f5faad"} Oct 14 08:51:47 crc kubenswrapper[4870]: I1014 08:51:47.284583 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"c8bd2488-7e3a-4dc2-87c9-dbaf25141f92","Type":"ContainerStarted","Data":"72e130e62cf6081ad149a79644d97b67db0559eb0b05b75ee39b3d90bb494095"} Oct 14 08:51:47 crc kubenswrapper[4870]: I1014 08:51:47.284868 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"c8bd2488-7e3a-4dc2-87c9-dbaf25141f92","Type":"ContainerStarted","Data":"cfa3fc49edcda5a23d04f27c9ca1c768de841a709c4527b939086c424edea733"} Oct 14 08:51:47 crc kubenswrapper[4870]: I1014 08:51:47.284886 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Oct 14 08:51:47 crc kubenswrapper[4870]: I1014 08:51:47.313213 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=1.577087463 podStartE2EDuration="2.313192139s" podCreationTimestamp="2025-10-14 08:51:45 +0000 UTC" firstStartedPulling="2025-10-14 08:51:46.180814358 +0000 UTC m=+6641.878174739" lastFinishedPulling="2025-10-14 08:51:46.916919034 +0000 UTC m=+6642.614279415" observedRunningTime="2025-10-14 08:51:47.304884285 +0000 UTC m=+6643.002244676" watchObservedRunningTime="2025-10-14 08:51:47.313192139 +0000 UTC m=+6643.010552520" Oct 14 08:51:53 crc kubenswrapper[4870]: I1014 08:51:53.116371 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-jwfxl"] Oct 14 08:51:53 crc kubenswrapper[4870]: I1014 08:51:53.118114 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-jwfxl" Oct 14 08:51:53 crc kubenswrapper[4870]: I1014 08:51:53.130884 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-jwfxl"] Oct 14 08:51:53 crc kubenswrapper[4870]: I1014 08:51:53.224704 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jrjsx\" (UniqueName: \"kubernetes.io/projected/13680632-67ca-40ab-b71a-569c99f3e424-kube-api-access-jrjsx\") pod \"keystone-db-create-jwfxl\" (UID: \"13680632-67ca-40ab-b71a-569c99f3e424\") " pod="openstack/keystone-db-create-jwfxl" Oct 14 08:51:53 crc kubenswrapper[4870]: I1014 08:51:53.325859 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jrjsx\" (UniqueName: \"kubernetes.io/projected/13680632-67ca-40ab-b71a-569c99f3e424-kube-api-access-jrjsx\") pod \"keystone-db-create-jwfxl\" (UID: \"13680632-67ca-40ab-b71a-569c99f3e424\") " pod="openstack/keystone-db-create-jwfxl" Oct 14 08:51:53 crc kubenswrapper[4870]: I1014 08:51:53.353320 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jrjsx\" (UniqueName: \"kubernetes.io/projected/13680632-67ca-40ab-b71a-569c99f3e424-kube-api-access-jrjsx\") pod \"keystone-db-create-jwfxl\" (UID: \"13680632-67ca-40ab-b71a-569c99f3e424\") " pod="openstack/keystone-db-create-jwfxl" Oct 14 08:51:53 crc kubenswrapper[4870]: I1014 08:51:53.490265 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-jwfxl" Oct 14 08:51:53 crc kubenswrapper[4870]: I1014 08:51:53.968195 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-jwfxl"] Oct 14 08:51:53 crc kubenswrapper[4870]: W1014 08:51:53.982360 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod13680632_67ca_40ab_b71a_569c99f3e424.slice/crio-e360ee91ce62b7f561403d67e94000f718dad6bb0a7d0dce49f8f68825212656 WatchSource:0}: Error finding container e360ee91ce62b7f561403d67e94000f718dad6bb0a7d0dce49f8f68825212656: Status 404 returned error can't find the container with id e360ee91ce62b7f561403d67e94000f718dad6bb0a7d0dce49f8f68825212656 Oct 14 08:51:54 crc kubenswrapper[4870]: I1014 08:51:54.360396 4870 generic.go:334] "Generic (PLEG): container finished" podID="13680632-67ca-40ab-b71a-569c99f3e424" containerID="ad7dfa6a7348cf4fe7ce935499a7c6f4e55c1a1c92056375846ad1c8d0fac2f7" exitCode=0 Oct 14 08:51:54 crc kubenswrapper[4870]: I1014 08:51:54.360466 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-jwfxl" event={"ID":"13680632-67ca-40ab-b71a-569c99f3e424","Type":"ContainerDied","Data":"ad7dfa6a7348cf4fe7ce935499a7c6f4e55c1a1c92056375846ad1c8d0fac2f7"} Oct 14 08:51:54 crc kubenswrapper[4870]: I1014 08:51:54.360509 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-jwfxl" event={"ID":"13680632-67ca-40ab-b71a-569c99f3e424","Type":"ContainerStarted","Data":"e360ee91ce62b7f561403d67e94000f718dad6bb0a7d0dce49f8f68825212656"} Oct 14 08:51:55 crc kubenswrapper[4870]: I1014 08:51:55.723017 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-jwfxl" Oct 14 08:51:55 crc kubenswrapper[4870]: I1014 08:51:55.765189 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jrjsx\" (UniqueName: \"kubernetes.io/projected/13680632-67ca-40ab-b71a-569c99f3e424-kube-api-access-jrjsx\") pod \"13680632-67ca-40ab-b71a-569c99f3e424\" (UID: \"13680632-67ca-40ab-b71a-569c99f3e424\") " Oct 14 08:51:55 crc kubenswrapper[4870]: I1014 08:51:55.771409 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/13680632-67ca-40ab-b71a-569c99f3e424-kube-api-access-jrjsx" (OuterVolumeSpecName: "kube-api-access-jrjsx") pod "13680632-67ca-40ab-b71a-569c99f3e424" (UID: "13680632-67ca-40ab-b71a-569c99f3e424"). InnerVolumeSpecName "kube-api-access-jrjsx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:51:55 crc kubenswrapper[4870]: I1014 08:51:55.867306 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jrjsx\" (UniqueName: \"kubernetes.io/projected/13680632-67ca-40ab-b71a-569c99f3e424-kube-api-access-jrjsx\") on node \"crc\" DevicePath \"\"" Oct 14 08:51:56 crc kubenswrapper[4870]: I1014 08:51:56.383790 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-jwfxl" event={"ID":"13680632-67ca-40ab-b71a-569c99f3e424","Type":"ContainerDied","Data":"e360ee91ce62b7f561403d67e94000f718dad6bb0a7d0dce49f8f68825212656"} Oct 14 08:51:56 crc kubenswrapper[4870]: I1014 08:51:56.383837 4870 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e360ee91ce62b7f561403d67e94000f718dad6bb0a7d0dce49f8f68825212656" Oct 14 08:51:56 crc kubenswrapper[4870]: I1014 08:51:56.383920 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-jwfxl" Oct 14 08:52:00 crc kubenswrapper[4870]: I1014 08:52:00.792664 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Oct 14 08:52:03 crc kubenswrapper[4870]: I1014 08:52:03.212584 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-5ede-account-create-gfcc4"] Oct 14 08:52:03 crc kubenswrapper[4870]: E1014 08:52:03.212968 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="13680632-67ca-40ab-b71a-569c99f3e424" containerName="mariadb-database-create" Oct 14 08:52:03 crc kubenswrapper[4870]: I1014 08:52:03.212982 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="13680632-67ca-40ab-b71a-569c99f3e424" containerName="mariadb-database-create" Oct 14 08:52:03 crc kubenswrapper[4870]: I1014 08:52:03.213150 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="13680632-67ca-40ab-b71a-569c99f3e424" containerName="mariadb-database-create" Oct 14 08:52:03 crc kubenswrapper[4870]: I1014 08:52:03.213705 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-5ede-account-create-gfcc4" Oct 14 08:52:03 crc kubenswrapper[4870]: I1014 08:52:03.215966 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Oct 14 08:52:03 crc kubenswrapper[4870]: I1014 08:52:03.219241 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-5ede-account-create-gfcc4"] Oct 14 08:52:03 crc kubenswrapper[4870]: I1014 08:52:03.323553 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pwzxl\" (UniqueName: \"kubernetes.io/projected/f08148e6-ef78-4157-aaf5-1c6432956583-kube-api-access-pwzxl\") pod \"keystone-5ede-account-create-gfcc4\" (UID: \"f08148e6-ef78-4157-aaf5-1c6432956583\") " pod="openstack/keystone-5ede-account-create-gfcc4" Oct 14 08:52:03 crc kubenswrapper[4870]: I1014 08:52:03.425469 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pwzxl\" (UniqueName: \"kubernetes.io/projected/f08148e6-ef78-4157-aaf5-1c6432956583-kube-api-access-pwzxl\") pod \"keystone-5ede-account-create-gfcc4\" (UID: \"f08148e6-ef78-4157-aaf5-1c6432956583\") " pod="openstack/keystone-5ede-account-create-gfcc4" Oct 14 08:52:03 crc kubenswrapper[4870]: I1014 08:52:03.465668 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pwzxl\" (UniqueName: \"kubernetes.io/projected/f08148e6-ef78-4157-aaf5-1c6432956583-kube-api-access-pwzxl\") pod \"keystone-5ede-account-create-gfcc4\" (UID: \"f08148e6-ef78-4157-aaf5-1c6432956583\") " pod="openstack/keystone-5ede-account-create-gfcc4" Oct 14 08:52:03 crc kubenswrapper[4870]: I1014 08:52:03.562095 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-5ede-account-create-gfcc4" Oct 14 08:52:04 crc kubenswrapper[4870]: I1014 08:52:04.036012 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-5ede-account-create-gfcc4"] Oct 14 08:52:04 crc kubenswrapper[4870]: I1014 08:52:04.457254 4870 generic.go:334] "Generic (PLEG): container finished" podID="f08148e6-ef78-4157-aaf5-1c6432956583" containerID="d1ca5583cb67929436e14fd7ee24f07c74a1e72bcdded1a788b5dbda1ef79a48" exitCode=0 Oct 14 08:52:04 crc kubenswrapper[4870]: I1014 08:52:04.457300 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-5ede-account-create-gfcc4" event={"ID":"f08148e6-ef78-4157-aaf5-1c6432956583","Type":"ContainerDied","Data":"d1ca5583cb67929436e14fd7ee24f07c74a1e72bcdded1a788b5dbda1ef79a48"} Oct 14 08:52:04 crc kubenswrapper[4870]: I1014 08:52:04.457334 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-5ede-account-create-gfcc4" event={"ID":"f08148e6-ef78-4157-aaf5-1c6432956583","Type":"ContainerStarted","Data":"63fd25cf9a7043f6cd93b6b9c1e7322ba237503c78eb57355e063e829ff9701c"} Oct 14 08:52:05 crc kubenswrapper[4870]: I1014 08:52:05.848909 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-5ede-account-create-gfcc4" Oct 14 08:52:05 crc kubenswrapper[4870]: I1014 08:52:05.969387 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pwzxl\" (UniqueName: \"kubernetes.io/projected/f08148e6-ef78-4157-aaf5-1c6432956583-kube-api-access-pwzxl\") pod \"f08148e6-ef78-4157-aaf5-1c6432956583\" (UID: \"f08148e6-ef78-4157-aaf5-1c6432956583\") " Oct 14 08:52:05 crc kubenswrapper[4870]: I1014 08:52:05.974374 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f08148e6-ef78-4157-aaf5-1c6432956583-kube-api-access-pwzxl" (OuterVolumeSpecName: "kube-api-access-pwzxl") pod "f08148e6-ef78-4157-aaf5-1c6432956583" (UID: "f08148e6-ef78-4157-aaf5-1c6432956583"). InnerVolumeSpecName "kube-api-access-pwzxl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:52:06 crc kubenswrapper[4870]: I1014 08:52:06.071203 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pwzxl\" (UniqueName: \"kubernetes.io/projected/f08148e6-ef78-4157-aaf5-1c6432956583-kube-api-access-pwzxl\") on node \"crc\" DevicePath \"\"" Oct 14 08:52:06 crc kubenswrapper[4870]: I1014 08:52:06.481779 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-5ede-account-create-gfcc4" event={"ID":"f08148e6-ef78-4157-aaf5-1c6432956583","Type":"ContainerDied","Data":"63fd25cf9a7043f6cd93b6b9c1e7322ba237503c78eb57355e063e829ff9701c"} Oct 14 08:52:06 crc kubenswrapper[4870]: I1014 08:52:06.481838 4870 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="63fd25cf9a7043f6cd93b6b9c1e7322ba237503c78eb57355e063e829ff9701c" Oct 14 08:52:06 crc kubenswrapper[4870]: I1014 08:52:06.481921 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-5ede-account-create-gfcc4" Oct 14 08:52:08 crc kubenswrapper[4870]: I1014 08:52:08.666032 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-xzttw"] Oct 14 08:52:08 crc kubenswrapper[4870]: E1014 08:52:08.666615 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f08148e6-ef78-4157-aaf5-1c6432956583" containerName="mariadb-account-create" Oct 14 08:52:08 crc kubenswrapper[4870]: I1014 08:52:08.666628 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="f08148e6-ef78-4157-aaf5-1c6432956583" containerName="mariadb-account-create" Oct 14 08:52:08 crc kubenswrapper[4870]: I1014 08:52:08.666811 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="f08148e6-ef78-4157-aaf5-1c6432956583" containerName="mariadb-account-create" Oct 14 08:52:08 crc kubenswrapper[4870]: I1014 08:52:08.667401 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-xzttw" Oct 14 08:52:08 crc kubenswrapper[4870]: I1014 08:52:08.669604 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 14 08:52:08 crc kubenswrapper[4870]: I1014 08:52:08.669635 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 14 08:52:08 crc kubenswrapper[4870]: I1014 08:52:08.669698 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 14 08:52:08 crc kubenswrapper[4870]: I1014 08:52:08.669798 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-mx4tj" Oct 14 08:52:08 crc kubenswrapper[4870]: I1014 08:52:08.676056 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-xzttw"] Oct 14 08:52:08 crc kubenswrapper[4870]: I1014 08:52:08.821261 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b024cec4-79e9-415b-aa23-725f5d15aacf-combined-ca-bundle\") pod \"keystone-db-sync-xzttw\" (UID: \"b024cec4-79e9-415b-aa23-725f5d15aacf\") " pod="openstack/keystone-db-sync-xzttw" Oct 14 08:52:08 crc kubenswrapper[4870]: I1014 08:52:08.821346 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4qtst\" (UniqueName: \"kubernetes.io/projected/b024cec4-79e9-415b-aa23-725f5d15aacf-kube-api-access-4qtst\") pod \"keystone-db-sync-xzttw\" (UID: \"b024cec4-79e9-415b-aa23-725f5d15aacf\") " pod="openstack/keystone-db-sync-xzttw" Oct 14 08:52:08 crc kubenswrapper[4870]: I1014 08:52:08.821396 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b024cec4-79e9-415b-aa23-725f5d15aacf-config-data\") pod \"keystone-db-sync-xzttw\" (UID: \"b024cec4-79e9-415b-aa23-725f5d15aacf\") " pod="openstack/keystone-db-sync-xzttw" Oct 14 08:52:08 crc kubenswrapper[4870]: I1014 08:52:08.923738 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b024cec4-79e9-415b-aa23-725f5d15aacf-combined-ca-bundle\") pod \"keystone-db-sync-xzttw\" (UID: \"b024cec4-79e9-415b-aa23-725f5d15aacf\") " pod="openstack/keystone-db-sync-xzttw" Oct 14 08:52:08 crc kubenswrapper[4870]: I1014 08:52:08.923850 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4qtst\" (UniqueName: \"kubernetes.io/projected/b024cec4-79e9-415b-aa23-725f5d15aacf-kube-api-access-4qtst\") pod \"keystone-db-sync-xzttw\" (UID: \"b024cec4-79e9-415b-aa23-725f5d15aacf\") " pod="openstack/keystone-db-sync-xzttw" Oct 14 08:52:08 crc kubenswrapper[4870]: I1014 08:52:08.923928 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b024cec4-79e9-415b-aa23-725f5d15aacf-config-data\") pod \"keystone-db-sync-xzttw\" (UID: \"b024cec4-79e9-415b-aa23-725f5d15aacf\") " pod="openstack/keystone-db-sync-xzttw" Oct 14 08:52:08 crc kubenswrapper[4870]: I1014 08:52:08.930454 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b024cec4-79e9-415b-aa23-725f5d15aacf-config-data\") pod \"keystone-db-sync-xzttw\" (UID: \"b024cec4-79e9-415b-aa23-725f5d15aacf\") " pod="openstack/keystone-db-sync-xzttw" Oct 14 08:52:08 crc kubenswrapper[4870]: I1014 08:52:08.936207 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b024cec4-79e9-415b-aa23-725f5d15aacf-combined-ca-bundle\") pod \"keystone-db-sync-xzttw\" (UID: \"b024cec4-79e9-415b-aa23-725f5d15aacf\") " pod="openstack/keystone-db-sync-xzttw" Oct 14 08:52:08 crc kubenswrapper[4870]: I1014 08:52:08.940506 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4qtst\" (UniqueName: \"kubernetes.io/projected/b024cec4-79e9-415b-aa23-725f5d15aacf-kube-api-access-4qtst\") pod \"keystone-db-sync-xzttw\" (UID: \"b024cec4-79e9-415b-aa23-725f5d15aacf\") " pod="openstack/keystone-db-sync-xzttw" Oct 14 08:52:08 crc kubenswrapper[4870]: I1014 08:52:08.989483 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-xzttw" Oct 14 08:52:09 crc kubenswrapper[4870]: I1014 08:52:09.418208 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-xzttw"] Oct 14 08:52:09 crc kubenswrapper[4870]: W1014 08:52:09.423730 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb024cec4_79e9_415b_aa23_725f5d15aacf.slice/crio-a50ece30d739547202228cb68a64cf075c5fd977ad07e1f7444a7cd7272bd6e1 WatchSource:0}: Error finding container a50ece30d739547202228cb68a64cf075c5fd977ad07e1f7444a7cd7272bd6e1: Status 404 returned error can't find the container with id a50ece30d739547202228cb68a64cf075c5fd977ad07e1f7444a7cd7272bd6e1 Oct 14 08:52:09 crc kubenswrapper[4870]: I1014 08:52:09.509723 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-xzttw" event={"ID":"b024cec4-79e9-415b-aa23-725f5d15aacf","Type":"ContainerStarted","Data":"a50ece30d739547202228cb68a64cf075c5fd977ad07e1f7444a7cd7272bd6e1"} Oct 14 08:52:14 crc kubenswrapper[4870]: I1014 08:52:14.550094 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-xzttw" event={"ID":"b024cec4-79e9-415b-aa23-725f5d15aacf","Type":"ContainerStarted","Data":"41d1c1da64483937ac1c1d1375ba7897ac6f748fde7cb100e83f2fbbd1051084"} Oct 14 08:52:14 crc kubenswrapper[4870]: I1014 08:52:14.576097 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-xzttw" podStartSLOduration=2.001004595 podStartE2EDuration="6.576080765s" podCreationTimestamp="2025-10-14 08:52:08 +0000 UTC" firstStartedPulling="2025-10-14 08:52:09.42770693 +0000 UTC m=+6665.125067301" lastFinishedPulling="2025-10-14 08:52:14.00278311 +0000 UTC m=+6669.700143471" observedRunningTime="2025-10-14 08:52:14.569895603 +0000 UTC m=+6670.267255984" watchObservedRunningTime="2025-10-14 08:52:14.576080765 +0000 UTC m=+6670.273441136" Oct 14 08:52:16 crc kubenswrapper[4870]: I1014 08:52:16.571959 4870 generic.go:334] "Generic (PLEG): container finished" podID="b024cec4-79e9-415b-aa23-725f5d15aacf" containerID="41d1c1da64483937ac1c1d1375ba7897ac6f748fde7cb100e83f2fbbd1051084" exitCode=0 Oct 14 08:52:16 crc kubenswrapper[4870]: I1014 08:52:16.572098 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-xzttw" event={"ID":"b024cec4-79e9-415b-aa23-725f5d15aacf","Type":"ContainerDied","Data":"41d1c1da64483937ac1c1d1375ba7897ac6f748fde7cb100e83f2fbbd1051084"} Oct 14 08:52:17 crc kubenswrapper[4870]: I1014 08:52:17.946852 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-xzttw" Oct 14 08:52:18 crc kubenswrapper[4870]: I1014 08:52:18.095584 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4qtst\" (UniqueName: \"kubernetes.io/projected/b024cec4-79e9-415b-aa23-725f5d15aacf-kube-api-access-4qtst\") pod \"b024cec4-79e9-415b-aa23-725f5d15aacf\" (UID: \"b024cec4-79e9-415b-aa23-725f5d15aacf\") " Oct 14 08:52:18 crc kubenswrapper[4870]: I1014 08:52:18.095664 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b024cec4-79e9-415b-aa23-725f5d15aacf-config-data\") pod \"b024cec4-79e9-415b-aa23-725f5d15aacf\" (UID: \"b024cec4-79e9-415b-aa23-725f5d15aacf\") " Oct 14 08:52:18 crc kubenswrapper[4870]: I1014 08:52:18.096318 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b024cec4-79e9-415b-aa23-725f5d15aacf-combined-ca-bundle\") pod \"b024cec4-79e9-415b-aa23-725f5d15aacf\" (UID: \"b024cec4-79e9-415b-aa23-725f5d15aacf\") " Oct 14 08:52:18 crc kubenswrapper[4870]: I1014 08:52:18.110855 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b024cec4-79e9-415b-aa23-725f5d15aacf-kube-api-access-4qtst" (OuterVolumeSpecName: "kube-api-access-4qtst") pod "b024cec4-79e9-415b-aa23-725f5d15aacf" (UID: "b024cec4-79e9-415b-aa23-725f5d15aacf"). InnerVolumeSpecName "kube-api-access-4qtst". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:52:18 crc kubenswrapper[4870]: I1014 08:52:18.133231 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b024cec4-79e9-415b-aa23-725f5d15aacf-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b024cec4-79e9-415b-aa23-725f5d15aacf" (UID: "b024cec4-79e9-415b-aa23-725f5d15aacf"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:52:18 crc kubenswrapper[4870]: I1014 08:52:18.176576 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b024cec4-79e9-415b-aa23-725f5d15aacf-config-data" (OuterVolumeSpecName: "config-data") pod "b024cec4-79e9-415b-aa23-725f5d15aacf" (UID: "b024cec4-79e9-415b-aa23-725f5d15aacf"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:52:18 crc kubenswrapper[4870]: I1014 08:52:18.198592 4870 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b024cec4-79e9-415b-aa23-725f5d15aacf-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 08:52:18 crc kubenswrapper[4870]: I1014 08:52:18.198627 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4qtst\" (UniqueName: \"kubernetes.io/projected/b024cec4-79e9-415b-aa23-725f5d15aacf-kube-api-access-4qtst\") on node \"crc\" DevicePath \"\"" Oct 14 08:52:18 crc kubenswrapper[4870]: I1014 08:52:18.198643 4870 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b024cec4-79e9-415b-aa23-725f5d15aacf-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 08:52:18 crc kubenswrapper[4870]: I1014 08:52:18.596212 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-xzttw" event={"ID":"b024cec4-79e9-415b-aa23-725f5d15aacf","Type":"ContainerDied","Data":"a50ece30d739547202228cb68a64cf075c5fd977ad07e1f7444a7cd7272bd6e1"} Oct 14 08:52:18 crc kubenswrapper[4870]: I1014 08:52:18.596266 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-xzttw" Oct 14 08:52:18 crc kubenswrapper[4870]: I1014 08:52:18.596288 4870 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a50ece30d739547202228cb68a64cf075c5fd977ad07e1f7444a7cd7272bd6e1" Oct 14 08:52:18 crc kubenswrapper[4870]: I1014 08:52:18.763988 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5b8d755fbc-ps7fd"] Oct 14 08:52:18 crc kubenswrapper[4870]: E1014 08:52:18.764335 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b024cec4-79e9-415b-aa23-725f5d15aacf" containerName="keystone-db-sync" Oct 14 08:52:18 crc kubenswrapper[4870]: I1014 08:52:18.764352 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="b024cec4-79e9-415b-aa23-725f5d15aacf" containerName="keystone-db-sync" Oct 14 08:52:18 crc kubenswrapper[4870]: I1014 08:52:18.764573 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="b024cec4-79e9-415b-aa23-725f5d15aacf" containerName="keystone-db-sync" Oct 14 08:52:18 crc kubenswrapper[4870]: I1014 08:52:18.765412 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b8d755fbc-ps7fd" Oct 14 08:52:18 crc kubenswrapper[4870]: I1014 08:52:18.788238 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5b8d755fbc-ps7fd"] Oct 14 08:52:18 crc kubenswrapper[4870]: I1014 08:52:18.796137 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-tz678"] Oct 14 08:52:18 crc kubenswrapper[4870]: I1014 08:52:18.798272 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-tz678" Oct 14 08:52:18 crc kubenswrapper[4870]: I1014 08:52:18.804741 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-tz678"] Oct 14 08:52:18 crc kubenswrapper[4870]: I1014 08:52:18.806634 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 14 08:52:18 crc kubenswrapper[4870]: I1014 08:52:18.807190 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 14 08:52:18 crc kubenswrapper[4870]: I1014 08:52:18.808066 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 14 08:52:18 crc kubenswrapper[4870]: I1014 08:52:18.814167 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-mx4tj" Oct 14 08:52:18 crc kubenswrapper[4870]: I1014 08:52:18.912355 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tv5kf\" (UniqueName: \"kubernetes.io/projected/d072b6d9-f87a-4019-acec-907231a700b2-kube-api-access-tv5kf\") pod \"keystone-bootstrap-tz678\" (UID: \"d072b6d9-f87a-4019-acec-907231a700b2\") " pod="openstack/keystone-bootstrap-tz678" Oct 14 08:52:18 crc kubenswrapper[4870]: I1014 08:52:18.912453 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d072b6d9-f87a-4019-acec-907231a700b2-config-data\") pod \"keystone-bootstrap-tz678\" (UID: \"d072b6d9-f87a-4019-acec-907231a700b2\") " pod="openstack/keystone-bootstrap-tz678" Oct 14 08:52:18 crc kubenswrapper[4870]: I1014 08:52:18.912509 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/eb9a9a5e-a99d-4fae-9b28-006252710ab1-ovsdbserver-sb\") pod \"dnsmasq-dns-5b8d755fbc-ps7fd\" (UID: \"eb9a9a5e-a99d-4fae-9b28-006252710ab1\") " pod="openstack/dnsmasq-dns-5b8d755fbc-ps7fd" Oct 14 08:52:18 crc kubenswrapper[4870]: I1014 08:52:18.917016 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/d072b6d9-f87a-4019-acec-907231a700b2-credential-keys\") pod \"keystone-bootstrap-tz678\" (UID: \"d072b6d9-f87a-4019-acec-907231a700b2\") " pod="openstack/keystone-bootstrap-tz678" Oct 14 08:52:18 crc kubenswrapper[4870]: I1014 08:52:18.917072 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d072b6d9-f87a-4019-acec-907231a700b2-combined-ca-bundle\") pod \"keystone-bootstrap-tz678\" (UID: \"d072b6d9-f87a-4019-acec-907231a700b2\") " pod="openstack/keystone-bootstrap-tz678" Oct 14 08:52:18 crc kubenswrapper[4870]: I1014 08:52:18.917163 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/eb9a9a5e-a99d-4fae-9b28-006252710ab1-ovsdbserver-nb\") pod \"dnsmasq-dns-5b8d755fbc-ps7fd\" (UID: \"eb9a9a5e-a99d-4fae-9b28-006252710ab1\") " pod="openstack/dnsmasq-dns-5b8d755fbc-ps7fd" Oct 14 08:52:18 crc kubenswrapper[4870]: I1014 08:52:18.917229 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d072b6d9-f87a-4019-acec-907231a700b2-scripts\") pod \"keystone-bootstrap-tz678\" (UID: \"d072b6d9-f87a-4019-acec-907231a700b2\") " pod="openstack/keystone-bootstrap-tz678" Oct 14 08:52:18 crc kubenswrapper[4870]: I1014 08:52:18.917295 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/d072b6d9-f87a-4019-acec-907231a700b2-fernet-keys\") pod \"keystone-bootstrap-tz678\" (UID: \"d072b6d9-f87a-4019-acec-907231a700b2\") " pod="openstack/keystone-bootstrap-tz678" Oct 14 08:52:18 crc kubenswrapper[4870]: I1014 08:52:18.917337 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pvxvm\" (UniqueName: \"kubernetes.io/projected/eb9a9a5e-a99d-4fae-9b28-006252710ab1-kube-api-access-pvxvm\") pod \"dnsmasq-dns-5b8d755fbc-ps7fd\" (UID: \"eb9a9a5e-a99d-4fae-9b28-006252710ab1\") " pod="openstack/dnsmasq-dns-5b8d755fbc-ps7fd" Oct 14 08:52:18 crc kubenswrapper[4870]: I1014 08:52:18.917383 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eb9a9a5e-a99d-4fae-9b28-006252710ab1-config\") pod \"dnsmasq-dns-5b8d755fbc-ps7fd\" (UID: \"eb9a9a5e-a99d-4fae-9b28-006252710ab1\") " pod="openstack/dnsmasq-dns-5b8d755fbc-ps7fd" Oct 14 08:52:18 crc kubenswrapper[4870]: I1014 08:52:18.917498 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/eb9a9a5e-a99d-4fae-9b28-006252710ab1-dns-svc\") pod \"dnsmasq-dns-5b8d755fbc-ps7fd\" (UID: \"eb9a9a5e-a99d-4fae-9b28-006252710ab1\") " pod="openstack/dnsmasq-dns-5b8d755fbc-ps7fd" Oct 14 08:52:19 crc kubenswrapper[4870]: I1014 08:52:19.018269 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/d072b6d9-f87a-4019-acec-907231a700b2-credential-keys\") pod \"keystone-bootstrap-tz678\" (UID: \"d072b6d9-f87a-4019-acec-907231a700b2\") " pod="openstack/keystone-bootstrap-tz678" Oct 14 08:52:19 crc kubenswrapper[4870]: I1014 08:52:19.018525 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d072b6d9-f87a-4019-acec-907231a700b2-combined-ca-bundle\") pod \"keystone-bootstrap-tz678\" (UID: \"d072b6d9-f87a-4019-acec-907231a700b2\") " pod="openstack/keystone-bootstrap-tz678" Oct 14 08:52:19 crc kubenswrapper[4870]: I1014 08:52:19.018567 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/eb9a9a5e-a99d-4fae-9b28-006252710ab1-ovsdbserver-nb\") pod \"dnsmasq-dns-5b8d755fbc-ps7fd\" (UID: \"eb9a9a5e-a99d-4fae-9b28-006252710ab1\") " pod="openstack/dnsmasq-dns-5b8d755fbc-ps7fd" Oct 14 08:52:19 crc kubenswrapper[4870]: I1014 08:52:19.018594 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d072b6d9-f87a-4019-acec-907231a700b2-scripts\") pod \"keystone-bootstrap-tz678\" (UID: \"d072b6d9-f87a-4019-acec-907231a700b2\") " pod="openstack/keystone-bootstrap-tz678" Oct 14 08:52:19 crc kubenswrapper[4870]: I1014 08:52:19.018625 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/d072b6d9-f87a-4019-acec-907231a700b2-fernet-keys\") pod \"keystone-bootstrap-tz678\" (UID: \"d072b6d9-f87a-4019-acec-907231a700b2\") " pod="openstack/keystone-bootstrap-tz678" Oct 14 08:52:19 crc kubenswrapper[4870]: I1014 08:52:19.018648 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pvxvm\" (UniqueName: \"kubernetes.io/projected/eb9a9a5e-a99d-4fae-9b28-006252710ab1-kube-api-access-pvxvm\") pod \"dnsmasq-dns-5b8d755fbc-ps7fd\" (UID: \"eb9a9a5e-a99d-4fae-9b28-006252710ab1\") " pod="openstack/dnsmasq-dns-5b8d755fbc-ps7fd" Oct 14 08:52:19 crc kubenswrapper[4870]: I1014 08:52:19.018669 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eb9a9a5e-a99d-4fae-9b28-006252710ab1-config\") pod \"dnsmasq-dns-5b8d755fbc-ps7fd\" (UID: \"eb9a9a5e-a99d-4fae-9b28-006252710ab1\") " pod="openstack/dnsmasq-dns-5b8d755fbc-ps7fd" Oct 14 08:52:19 crc kubenswrapper[4870]: I1014 08:52:19.018697 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/eb9a9a5e-a99d-4fae-9b28-006252710ab1-dns-svc\") pod \"dnsmasq-dns-5b8d755fbc-ps7fd\" (UID: \"eb9a9a5e-a99d-4fae-9b28-006252710ab1\") " pod="openstack/dnsmasq-dns-5b8d755fbc-ps7fd" Oct 14 08:52:19 crc kubenswrapper[4870]: I1014 08:52:19.018720 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tv5kf\" (UniqueName: \"kubernetes.io/projected/d072b6d9-f87a-4019-acec-907231a700b2-kube-api-access-tv5kf\") pod \"keystone-bootstrap-tz678\" (UID: \"d072b6d9-f87a-4019-acec-907231a700b2\") " pod="openstack/keystone-bootstrap-tz678" Oct 14 08:52:19 crc kubenswrapper[4870]: I1014 08:52:19.018741 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d072b6d9-f87a-4019-acec-907231a700b2-config-data\") pod \"keystone-bootstrap-tz678\" (UID: \"d072b6d9-f87a-4019-acec-907231a700b2\") " pod="openstack/keystone-bootstrap-tz678" Oct 14 08:52:19 crc kubenswrapper[4870]: I1014 08:52:19.018763 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/eb9a9a5e-a99d-4fae-9b28-006252710ab1-ovsdbserver-sb\") pod \"dnsmasq-dns-5b8d755fbc-ps7fd\" (UID: \"eb9a9a5e-a99d-4fae-9b28-006252710ab1\") " pod="openstack/dnsmasq-dns-5b8d755fbc-ps7fd" Oct 14 08:52:19 crc kubenswrapper[4870]: I1014 08:52:19.019580 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/eb9a9a5e-a99d-4fae-9b28-006252710ab1-dns-svc\") pod \"dnsmasq-dns-5b8d755fbc-ps7fd\" (UID: \"eb9a9a5e-a99d-4fae-9b28-006252710ab1\") " pod="openstack/dnsmasq-dns-5b8d755fbc-ps7fd" Oct 14 08:52:19 crc kubenswrapper[4870]: I1014 08:52:19.019611 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/eb9a9a5e-a99d-4fae-9b28-006252710ab1-ovsdbserver-sb\") pod \"dnsmasq-dns-5b8d755fbc-ps7fd\" (UID: \"eb9a9a5e-a99d-4fae-9b28-006252710ab1\") " pod="openstack/dnsmasq-dns-5b8d755fbc-ps7fd" Oct 14 08:52:19 crc kubenswrapper[4870]: I1014 08:52:19.019704 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eb9a9a5e-a99d-4fae-9b28-006252710ab1-config\") pod \"dnsmasq-dns-5b8d755fbc-ps7fd\" (UID: \"eb9a9a5e-a99d-4fae-9b28-006252710ab1\") " pod="openstack/dnsmasq-dns-5b8d755fbc-ps7fd" Oct 14 08:52:19 crc kubenswrapper[4870]: I1014 08:52:19.020246 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/eb9a9a5e-a99d-4fae-9b28-006252710ab1-ovsdbserver-nb\") pod \"dnsmasq-dns-5b8d755fbc-ps7fd\" (UID: \"eb9a9a5e-a99d-4fae-9b28-006252710ab1\") " pod="openstack/dnsmasq-dns-5b8d755fbc-ps7fd" Oct 14 08:52:19 crc kubenswrapper[4870]: I1014 08:52:19.022022 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d072b6d9-f87a-4019-acec-907231a700b2-scripts\") pod \"keystone-bootstrap-tz678\" (UID: \"d072b6d9-f87a-4019-acec-907231a700b2\") " pod="openstack/keystone-bootstrap-tz678" Oct 14 08:52:19 crc kubenswrapper[4870]: I1014 08:52:19.024092 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d072b6d9-f87a-4019-acec-907231a700b2-config-data\") pod \"keystone-bootstrap-tz678\" (UID: \"d072b6d9-f87a-4019-acec-907231a700b2\") " pod="openstack/keystone-bootstrap-tz678" Oct 14 08:52:19 crc kubenswrapper[4870]: I1014 08:52:19.038103 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tv5kf\" (UniqueName: \"kubernetes.io/projected/d072b6d9-f87a-4019-acec-907231a700b2-kube-api-access-tv5kf\") pod \"keystone-bootstrap-tz678\" (UID: \"d072b6d9-f87a-4019-acec-907231a700b2\") " pod="openstack/keystone-bootstrap-tz678" Oct 14 08:52:19 crc kubenswrapper[4870]: I1014 08:52:19.040198 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d072b6d9-f87a-4019-acec-907231a700b2-combined-ca-bundle\") pod \"keystone-bootstrap-tz678\" (UID: \"d072b6d9-f87a-4019-acec-907231a700b2\") " pod="openstack/keystone-bootstrap-tz678" Oct 14 08:52:19 crc kubenswrapper[4870]: I1014 08:52:19.040557 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pvxvm\" (UniqueName: \"kubernetes.io/projected/eb9a9a5e-a99d-4fae-9b28-006252710ab1-kube-api-access-pvxvm\") pod \"dnsmasq-dns-5b8d755fbc-ps7fd\" (UID: \"eb9a9a5e-a99d-4fae-9b28-006252710ab1\") " pod="openstack/dnsmasq-dns-5b8d755fbc-ps7fd" Oct 14 08:52:19 crc kubenswrapper[4870]: I1014 08:52:19.041622 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/d072b6d9-f87a-4019-acec-907231a700b2-credential-keys\") pod \"keystone-bootstrap-tz678\" (UID: \"d072b6d9-f87a-4019-acec-907231a700b2\") " pod="openstack/keystone-bootstrap-tz678" Oct 14 08:52:19 crc kubenswrapper[4870]: I1014 08:52:19.042153 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/d072b6d9-f87a-4019-acec-907231a700b2-fernet-keys\") pod \"keystone-bootstrap-tz678\" (UID: \"d072b6d9-f87a-4019-acec-907231a700b2\") " pod="openstack/keystone-bootstrap-tz678" Oct 14 08:52:19 crc kubenswrapper[4870]: I1014 08:52:19.107464 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b8d755fbc-ps7fd" Oct 14 08:52:19 crc kubenswrapper[4870]: I1014 08:52:19.140238 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-tz678" Oct 14 08:52:19 crc kubenswrapper[4870]: I1014 08:52:19.554954 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5b8d755fbc-ps7fd"] Oct 14 08:52:19 crc kubenswrapper[4870]: I1014 08:52:19.619827 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b8d755fbc-ps7fd" event={"ID":"eb9a9a5e-a99d-4fae-9b28-006252710ab1","Type":"ContainerStarted","Data":"4d969c51a004112f47a7c2a7594a95d29c9eff5ab482639f4f6c0be2c52d6db9"} Oct 14 08:52:19 crc kubenswrapper[4870]: I1014 08:52:19.639108 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-tz678"] Oct 14 08:52:20 crc kubenswrapper[4870]: I1014 08:52:20.630587 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-tz678" event={"ID":"d072b6d9-f87a-4019-acec-907231a700b2","Type":"ContainerStarted","Data":"63f04964d083df370a45f2e9b29453b85180d36eb0ed7ccd7ecd06cbfc634472"} Oct 14 08:52:20 crc kubenswrapper[4870]: I1014 08:52:20.630903 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-tz678" event={"ID":"d072b6d9-f87a-4019-acec-907231a700b2","Type":"ContainerStarted","Data":"5c4bf563c2c0a44ca494d657914311a793591fe95ee013b664c450ccf58347a2"} Oct 14 08:52:20 crc kubenswrapper[4870]: I1014 08:52:20.632538 4870 generic.go:334] "Generic (PLEG): container finished" podID="eb9a9a5e-a99d-4fae-9b28-006252710ab1" containerID="fd0d21dfd2262dbc3250c7c479f0ad007aab08376da9399b80c824ab62724be5" exitCode=0 Oct 14 08:52:20 crc kubenswrapper[4870]: I1014 08:52:20.632608 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b8d755fbc-ps7fd" event={"ID":"eb9a9a5e-a99d-4fae-9b28-006252710ab1","Type":"ContainerDied","Data":"fd0d21dfd2262dbc3250c7c479f0ad007aab08376da9399b80c824ab62724be5"} Oct 14 08:52:20 crc kubenswrapper[4870]: I1014 08:52:20.648333 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-tz678" podStartSLOduration=2.6483139959999997 podStartE2EDuration="2.648313996s" podCreationTimestamp="2025-10-14 08:52:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 08:52:20.648148832 +0000 UTC m=+6676.345509203" watchObservedRunningTime="2025-10-14 08:52:20.648313996 +0000 UTC m=+6676.345674367" Oct 14 08:52:21 crc kubenswrapper[4870]: I1014 08:52:21.643904 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b8d755fbc-ps7fd" event={"ID":"eb9a9a5e-a99d-4fae-9b28-006252710ab1","Type":"ContainerStarted","Data":"9063c16715734d3fc47a99adf4dfbc8c8f48caa9b3a33c4c1e9a934a7bb6e507"} Oct 14 08:52:21 crc kubenswrapper[4870]: I1014 08:52:21.643961 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5b8d755fbc-ps7fd" Oct 14 08:52:21 crc kubenswrapper[4870]: I1014 08:52:21.688134 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5b8d755fbc-ps7fd" podStartSLOduration=3.688108478 podStartE2EDuration="3.688108478s" podCreationTimestamp="2025-10-14 08:52:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 08:52:21.678607375 +0000 UTC m=+6677.375967816" watchObservedRunningTime="2025-10-14 08:52:21.688108478 +0000 UTC m=+6677.385468879" Oct 14 08:52:23 crc kubenswrapper[4870]: I1014 08:52:23.663152 4870 generic.go:334] "Generic (PLEG): container finished" podID="d072b6d9-f87a-4019-acec-907231a700b2" containerID="63f04964d083df370a45f2e9b29453b85180d36eb0ed7ccd7ecd06cbfc634472" exitCode=0 Oct 14 08:52:23 crc kubenswrapper[4870]: I1014 08:52:23.663301 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-tz678" event={"ID":"d072b6d9-f87a-4019-acec-907231a700b2","Type":"ContainerDied","Data":"63f04964d083df370a45f2e9b29453b85180d36eb0ed7ccd7ecd06cbfc634472"} Oct 14 08:52:23 crc kubenswrapper[4870]: I1014 08:52:23.950970 4870 patch_prober.go:28] interesting pod/machine-config-daemon-7tvc7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 08:52:23 crc kubenswrapper[4870]: I1014 08:52:23.951056 4870 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 08:52:25 crc kubenswrapper[4870]: I1014 08:52:24.999836 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-tz678" Oct 14 08:52:25 crc kubenswrapper[4870]: I1014 08:52:25.126694 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d072b6d9-f87a-4019-acec-907231a700b2-config-data\") pod \"d072b6d9-f87a-4019-acec-907231a700b2\" (UID: \"d072b6d9-f87a-4019-acec-907231a700b2\") " Oct 14 08:52:25 crc kubenswrapper[4870]: I1014 08:52:25.127032 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tv5kf\" (UniqueName: \"kubernetes.io/projected/d072b6d9-f87a-4019-acec-907231a700b2-kube-api-access-tv5kf\") pod \"d072b6d9-f87a-4019-acec-907231a700b2\" (UID: \"d072b6d9-f87a-4019-acec-907231a700b2\") " Oct 14 08:52:25 crc kubenswrapper[4870]: I1014 08:52:25.127163 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/d072b6d9-f87a-4019-acec-907231a700b2-credential-keys\") pod \"d072b6d9-f87a-4019-acec-907231a700b2\" (UID: \"d072b6d9-f87a-4019-acec-907231a700b2\") " Oct 14 08:52:25 crc kubenswrapper[4870]: I1014 08:52:25.127343 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d072b6d9-f87a-4019-acec-907231a700b2-scripts\") pod \"d072b6d9-f87a-4019-acec-907231a700b2\" (UID: \"d072b6d9-f87a-4019-acec-907231a700b2\") " Oct 14 08:52:25 crc kubenswrapper[4870]: I1014 08:52:25.127399 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d072b6d9-f87a-4019-acec-907231a700b2-combined-ca-bundle\") pod \"d072b6d9-f87a-4019-acec-907231a700b2\" (UID: \"d072b6d9-f87a-4019-acec-907231a700b2\") " Oct 14 08:52:25 crc kubenswrapper[4870]: I1014 08:52:25.127428 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/d072b6d9-f87a-4019-acec-907231a700b2-fernet-keys\") pod \"d072b6d9-f87a-4019-acec-907231a700b2\" (UID: \"d072b6d9-f87a-4019-acec-907231a700b2\") " Oct 14 08:52:25 crc kubenswrapper[4870]: I1014 08:52:25.131980 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d072b6d9-f87a-4019-acec-907231a700b2-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "d072b6d9-f87a-4019-acec-907231a700b2" (UID: "d072b6d9-f87a-4019-acec-907231a700b2"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:52:25 crc kubenswrapper[4870]: I1014 08:52:25.132059 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d072b6d9-f87a-4019-acec-907231a700b2-scripts" (OuterVolumeSpecName: "scripts") pod "d072b6d9-f87a-4019-acec-907231a700b2" (UID: "d072b6d9-f87a-4019-acec-907231a700b2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:52:25 crc kubenswrapper[4870]: I1014 08:52:25.132114 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d072b6d9-f87a-4019-acec-907231a700b2-kube-api-access-tv5kf" (OuterVolumeSpecName: "kube-api-access-tv5kf") pod "d072b6d9-f87a-4019-acec-907231a700b2" (UID: "d072b6d9-f87a-4019-acec-907231a700b2"). InnerVolumeSpecName "kube-api-access-tv5kf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:52:25 crc kubenswrapper[4870]: I1014 08:52:25.133626 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d072b6d9-f87a-4019-acec-907231a700b2-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "d072b6d9-f87a-4019-acec-907231a700b2" (UID: "d072b6d9-f87a-4019-acec-907231a700b2"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:52:25 crc kubenswrapper[4870]: I1014 08:52:25.149412 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d072b6d9-f87a-4019-acec-907231a700b2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d072b6d9-f87a-4019-acec-907231a700b2" (UID: "d072b6d9-f87a-4019-acec-907231a700b2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:52:25 crc kubenswrapper[4870]: I1014 08:52:25.156087 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d072b6d9-f87a-4019-acec-907231a700b2-config-data" (OuterVolumeSpecName: "config-data") pod "d072b6d9-f87a-4019-acec-907231a700b2" (UID: "d072b6d9-f87a-4019-acec-907231a700b2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:52:25 crc kubenswrapper[4870]: I1014 08:52:25.229658 4870 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d072b6d9-f87a-4019-acec-907231a700b2-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 08:52:25 crc kubenswrapper[4870]: I1014 08:52:25.229701 4870 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d072b6d9-f87a-4019-acec-907231a700b2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 08:52:25 crc kubenswrapper[4870]: I1014 08:52:25.229717 4870 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/d072b6d9-f87a-4019-acec-907231a700b2-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 14 08:52:25 crc kubenswrapper[4870]: I1014 08:52:25.229730 4870 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d072b6d9-f87a-4019-acec-907231a700b2-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 08:52:25 crc kubenswrapper[4870]: I1014 08:52:25.229742 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tv5kf\" (UniqueName: \"kubernetes.io/projected/d072b6d9-f87a-4019-acec-907231a700b2-kube-api-access-tv5kf\") on node \"crc\" DevicePath \"\"" Oct 14 08:52:25 crc kubenswrapper[4870]: I1014 08:52:25.229755 4870 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/d072b6d9-f87a-4019-acec-907231a700b2-credential-keys\") on node \"crc\" DevicePath \"\"" Oct 14 08:52:25 crc kubenswrapper[4870]: I1014 08:52:25.685796 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-tz678" Oct 14 08:52:25 crc kubenswrapper[4870]: I1014 08:52:25.686389 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-tz678" event={"ID":"d072b6d9-f87a-4019-acec-907231a700b2","Type":"ContainerDied","Data":"5c4bf563c2c0a44ca494d657914311a793591fe95ee013b664c450ccf58347a2"} Oct 14 08:52:25 crc kubenswrapper[4870]: I1014 08:52:25.686495 4870 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5c4bf563c2c0a44ca494d657914311a793591fe95ee013b664c450ccf58347a2" Oct 14 08:52:25 crc kubenswrapper[4870]: I1014 08:52:25.894578 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-tz678"] Oct 14 08:52:25 crc kubenswrapper[4870]: I1014 08:52:25.908613 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-tz678"] Oct 14 08:52:25 crc kubenswrapper[4870]: I1014 08:52:25.975617 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-vf77d"] Oct 14 08:52:25 crc kubenswrapper[4870]: E1014 08:52:25.976035 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d072b6d9-f87a-4019-acec-907231a700b2" containerName="keystone-bootstrap" Oct 14 08:52:25 crc kubenswrapper[4870]: I1014 08:52:25.976063 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="d072b6d9-f87a-4019-acec-907231a700b2" containerName="keystone-bootstrap" Oct 14 08:52:25 crc kubenswrapper[4870]: I1014 08:52:25.976397 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="d072b6d9-f87a-4019-acec-907231a700b2" containerName="keystone-bootstrap" Oct 14 08:52:25 crc kubenswrapper[4870]: I1014 08:52:25.977173 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-vf77d" Oct 14 08:52:25 crc kubenswrapper[4870]: I1014 08:52:25.980500 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 14 08:52:25 crc kubenswrapper[4870]: I1014 08:52:25.980503 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 14 08:52:25 crc kubenswrapper[4870]: I1014 08:52:25.980985 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-mx4tj" Oct 14 08:52:25 crc kubenswrapper[4870]: I1014 08:52:25.983499 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 14 08:52:25 crc kubenswrapper[4870]: I1014 08:52:25.987635 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-vf77d"] Oct 14 08:52:26 crc kubenswrapper[4870]: I1014 08:52:26.044738 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4068efdd-8071-411b-9302-3e26f1bb44c2-combined-ca-bundle\") pod \"keystone-bootstrap-vf77d\" (UID: \"4068efdd-8071-411b-9302-3e26f1bb44c2\") " pod="openstack/keystone-bootstrap-vf77d" Oct 14 08:52:26 crc kubenswrapper[4870]: I1014 08:52:26.044805 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/4068efdd-8071-411b-9302-3e26f1bb44c2-fernet-keys\") pod \"keystone-bootstrap-vf77d\" (UID: \"4068efdd-8071-411b-9302-3e26f1bb44c2\") " pod="openstack/keystone-bootstrap-vf77d" Oct 14 08:52:26 crc kubenswrapper[4870]: I1014 08:52:26.045118 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4068efdd-8071-411b-9302-3e26f1bb44c2-config-data\") pod \"keystone-bootstrap-vf77d\" (UID: \"4068efdd-8071-411b-9302-3e26f1bb44c2\") " pod="openstack/keystone-bootstrap-vf77d" Oct 14 08:52:26 crc kubenswrapper[4870]: I1014 08:52:26.045222 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7p7vw\" (UniqueName: \"kubernetes.io/projected/4068efdd-8071-411b-9302-3e26f1bb44c2-kube-api-access-7p7vw\") pod \"keystone-bootstrap-vf77d\" (UID: \"4068efdd-8071-411b-9302-3e26f1bb44c2\") " pod="openstack/keystone-bootstrap-vf77d" Oct 14 08:52:26 crc kubenswrapper[4870]: I1014 08:52:26.045258 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/4068efdd-8071-411b-9302-3e26f1bb44c2-credential-keys\") pod \"keystone-bootstrap-vf77d\" (UID: \"4068efdd-8071-411b-9302-3e26f1bb44c2\") " pod="openstack/keystone-bootstrap-vf77d" Oct 14 08:52:26 crc kubenswrapper[4870]: I1014 08:52:26.045329 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4068efdd-8071-411b-9302-3e26f1bb44c2-scripts\") pod \"keystone-bootstrap-vf77d\" (UID: \"4068efdd-8071-411b-9302-3e26f1bb44c2\") " pod="openstack/keystone-bootstrap-vf77d" Oct 14 08:52:26 crc kubenswrapper[4870]: I1014 08:52:26.146140 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4068efdd-8071-411b-9302-3e26f1bb44c2-config-data\") pod \"keystone-bootstrap-vf77d\" (UID: \"4068efdd-8071-411b-9302-3e26f1bb44c2\") " pod="openstack/keystone-bootstrap-vf77d" Oct 14 08:52:26 crc kubenswrapper[4870]: I1014 08:52:26.146230 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7p7vw\" (UniqueName: \"kubernetes.io/projected/4068efdd-8071-411b-9302-3e26f1bb44c2-kube-api-access-7p7vw\") pod \"keystone-bootstrap-vf77d\" (UID: \"4068efdd-8071-411b-9302-3e26f1bb44c2\") " pod="openstack/keystone-bootstrap-vf77d" Oct 14 08:52:26 crc kubenswrapper[4870]: I1014 08:52:26.146265 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/4068efdd-8071-411b-9302-3e26f1bb44c2-credential-keys\") pod \"keystone-bootstrap-vf77d\" (UID: \"4068efdd-8071-411b-9302-3e26f1bb44c2\") " pod="openstack/keystone-bootstrap-vf77d" Oct 14 08:52:26 crc kubenswrapper[4870]: I1014 08:52:26.146295 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4068efdd-8071-411b-9302-3e26f1bb44c2-scripts\") pod \"keystone-bootstrap-vf77d\" (UID: \"4068efdd-8071-411b-9302-3e26f1bb44c2\") " pod="openstack/keystone-bootstrap-vf77d" Oct 14 08:52:26 crc kubenswrapper[4870]: I1014 08:52:26.146343 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4068efdd-8071-411b-9302-3e26f1bb44c2-combined-ca-bundle\") pod \"keystone-bootstrap-vf77d\" (UID: \"4068efdd-8071-411b-9302-3e26f1bb44c2\") " pod="openstack/keystone-bootstrap-vf77d" Oct 14 08:52:26 crc kubenswrapper[4870]: I1014 08:52:26.146374 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/4068efdd-8071-411b-9302-3e26f1bb44c2-fernet-keys\") pod \"keystone-bootstrap-vf77d\" (UID: \"4068efdd-8071-411b-9302-3e26f1bb44c2\") " pod="openstack/keystone-bootstrap-vf77d" Oct 14 08:52:26 crc kubenswrapper[4870]: I1014 08:52:26.151648 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4068efdd-8071-411b-9302-3e26f1bb44c2-config-data\") pod \"keystone-bootstrap-vf77d\" (UID: \"4068efdd-8071-411b-9302-3e26f1bb44c2\") " pod="openstack/keystone-bootstrap-vf77d" Oct 14 08:52:26 crc kubenswrapper[4870]: I1014 08:52:26.151666 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4068efdd-8071-411b-9302-3e26f1bb44c2-combined-ca-bundle\") pod \"keystone-bootstrap-vf77d\" (UID: \"4068efdd-8071-411b-9302-3e26f1bb44c2\") " pod="openstack/keystone-bootstrap-vf77d" Oct 14 08:52:26 crc kubenswrapper[4870]: I1014 08:52:26.151932 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/4068efdd-8071-411b-9302-3e26f1bb44c2-credential-keys\") pod \"keystone-bootstrap-vf77d\" (UID: \"4068efdd-8071-411b-9302-3e26f1bb44c2\") " pod="openstack/keystone-bootstrap-vf77d" Oct 14 08:52:26 crc kubenswrapper[4870]: I1014 08:52:26.152144 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/4068efdd-8071-411b-9302-3e26f1bb44c2-fernet-keys\") pod \"keystone-bootstrap-vf77d\" (UID: \"4068efdd-8071-411b-9302-3e26f1bb44c2\") " pod="openstack/keystone-bootstrap-vf77d" Oct 14 08:52:26 crc kubenswrapper[4870]: I1014 08:52:26.152934 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4068efdd-8071-411b-9302-3e26f1bb44c2-scripts\") pod \"keystone-bootstrap-vf77d\" (UID: \"4068efdd-8071-411b-9302-3e26f1bb44c2\") " pod="openstack/keystone-bootstrap-vf77d" Oct 14 08:52:26 crc kubenswrapper[4870]: I1014 08:52:26.165921 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7p7vw\" (UniqueName: \"kubernetes.io/projected/4068efdd-8071-411b-9302-3e26f1bb44c2-kube-api-access-7p7vw\") pod \"keystone-bootstrap-vf77d\" (UID: \"4068efdd-8071-411b-9302-3e26f1bb44c2\") " pod="openstack/keystone-bootstrap-vf77d" Oct 14 08:52:26 crc kubenswrapper[4870]: I1014 08:52:26.300906 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-vf77d" Oct 14 08:52:26 crc kubenswrapper[4870]: I1014 08:52:26.787605 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-vf77d"] Oct 14 08:52:27 crc kubenswrapper[4870]: I1014 08:52:27.045217 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d072b6d9-f87a-4019-acec-907231a700b2" path="/var/lib/kubelet/pods/d072b6d9-f87a-4019-acec-907231a700b2/volumes" Oct 14 08:52:27 crc kubenswrapper[4870]: I1014 08:52:27.706999 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-vf77d" event={"ID":"4068efdd-8071-411b-9302-3e26f1bb44c2","Type":"ContainerStarted","Data":"88870aca262234f0dd719a72dad60be2f2c1437bb35f0d0e9af2e0596d07e866"} Oct 14 08:52:27 crc kubenswrapper[4870]: I1014 08:52:27.707074 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-vf77d" event={"ID":"4068efdd-8071-411b-9302-3e26f1bb44c2","Type":"ContainerStarted","Data":"c58b211768430a9896f2c00226f788f493188b847047136e5ed2e96083862c99"} Oct 14 08:52:27 crc kubenswrapper[4870]: I1014 08:52:27.745193 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-vf77d" podStartSLOduration=2.745167678 podStartE2EDuration="2.745167678s" podCreationTimestamp="2025-10-14 08:52:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 08:52:27.732942438 +0000 UTC m=+6683.430302839" watchObservedRunningTime="2025-10-14 08:52:27.745167678 +0000 UTC m=+6683.442528079" Oct 14 08:52:29 crc kubenswrapper[4870]: I1014 08:52:29.120820 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5b8d755fbc-ps7fd" Oct 14 08:52:29 crc kubenswrapper[4870]: I1014 08:52:29.188928 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7cf94d58c5-f8h59"] Oct 14 08:52:29 crc kubenswrapper[4870]: I1014 08:52:29.189243 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7cf94d58c5-f8h59" podUID="df5834a5-3b83-48c0-b83f-10e97d10a7d6" containerName="dnsmasq-dns" containerID="cri-o://51a12e0a1aedb158605abc7125b15a1415e952e0919d0aeec5c10cb86b235ec9" gracePeriod=10 Oct 14 08:52:29 crc kubenswrapper[4870]: I1014 08:52:29.679343 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7cf94d58c5-f8h59" Oct 14 08:52:29 crc kubenswrapper[4870]: I1014 08:52:29.722481 4870 generic.go:334] "Generic (PLEG): container finished" podID="df5834a5-3b83-48c0-b83f-10e97d10a7d6" containerID="51a12e0a1aedb158605abc7125b15a1415e952e0919d0aeec5c10cb86b235ec9" exitCode=0 Oct 14 08:52:29 crc kubenswrapper[4870]: I1014 08:52:29.722541 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cf94d58c5-f8h59" event={"ID":"df5834a5-3b83-48c0-b83f-10e97d10a7d6","Type":"ContainerDied","Data":"51a12e0a1aedb158605abc7125b15a1415e952e0919d0aeec5c10cb86b235ec9"} Oct 14 08:52:29 crc kubenswrapper[4870]: I1014 08:52:29.722575 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cf94d58c5-f8h59" event={"ID":"df5834a5-3b83-48c0-b83f-10e97d10a7d6","Type":"ContainerDied","Data":"2aeb3be9263ddcb47659f5c6b4d75c4f5b5c649dd1ccd7c97c91990e789fbb95"} Oct 14 08:52:29 crc kubenswrapper[4870]: I1014 08:52:29.722578 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7cf94d58c5-f8h59" Oct 14 08:52:29 crc kubenswrapper[4870]: I1014 08:52:29.722603 4870 scope.go:117] "RemoveContainer" containerID="51a12e0a1aedb158605abc7125b15a1415e952e0919d0aeec5c10cb86b235ec9" Oct 14 08:52:29 crc kubenswrapper[4870]: I1014 08:52:29.754596 4870 scope.go:117] "RemoveContainer" containerID="90837e89ec3a4b7d740d83eedfc117f7917a1ce9f7580a83b8860970008e99c5" Oct 14 08:52:29 crc kubenswrapper[4870]: I1014 08:52:29.814610 4870 scope.go:117] "RemoveContainer" containerID="51a12e0a1aedb158605abc7125b15a1415e952e0919d0aeec5c10cb86b235ec9" Oct 14 08:52:29 crc kubenswrapper[4870]: E1014 08:52:29.815042 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"51a12e0a1aedb158605abc7125b15a1415e952e0919d0aeec5c10cb86b235ec9\": container with ID starting with 51a12e0a1aedb158605abc7125b15a1415e952e0919d0aeec5c10cb86b235ec9 not found: ID does not exist" containerID="51a12e0a1aedb158605abc7125b15a1415e952e0919d0aeec5c10cb86b235ec9" Oct 14 08:52:29 crc kubenswrapper[4870]: I1014 08:52:29.815140 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"51a12e0a1aedb158605abc7125b15a1415e952e0919d0aeec5c10cb86b235ec9"} err="failed to get container status \"51a12e0a1aedb158605abc7125b15a1415e952e0919d0aeec5c10cb86b235ec9\": rpc error: code = NotFound desc = could not find container \"51a12e0a1aedb158605abc7125b15a1415e952e0919d0aeec5c10cb86b235ec9\": container with ID starting with 51a12e0a1aedb158605abc7125b15a1415e952e0919d0aeec5c10cb86b235ec9 not found: ID does not exist" Oct 14 08:52:29 crc kubenswrapper[4870]: I1014 08:52:29.815169 4870 scope.go:117] "RemoveContainer" containerID="90837e89ec3a4b7d740d83eedfc117f7917a1ce9f7580a83b8860970008e99c5" Oct 14 08:52:29 crc kubenswrapper[4870]: E1014 08:52:29.815710 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"90837e89ec3a4b7d740d83eedfc117f7917a1ce9f7580a83b8860970008e99c5\": container with ID starting with 90837e89ec3a4b7d740d83eedfc117f7917a1ce9f7580a83b8860970008e99c5 not found: ID does not exist" containerID="90837e89ec3a4b7d740d83eedfc117f7917a1ce9f7580a83b8860970008e99c5" Oct 14 08:52:29 crc kubenswrapper[4870]: I1014 08:52:29.815732 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"90837e89ec3a4b7d740d83eedfc117f7917a1ce9f7580a83b8860970008e99c5"} err="failed to get container status \"90837e89ec3a4b7d740d83eedfc117f7917a1ce9f7580a83b8860970008e99c5\": rpc error: code = NotFound desc = could not find container \"90837e89ec3a4b7d740d83eedfc117f7917a1ce9f7580a83b8860970008e99c5\": container with ID starting with 90837e89ec3a4b7d740d83eedfc117f7917a1ce9f7580a83b8860970008e99c5 not found: ID does not exist" Oct 14 08:52:29 crc kubenswrapper[4870]: I1014 08:52:29.859626 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/df5834a5-3b83-48c0-b83f-10e97d10a7d6-ovsdbserver-nb\") pod \"df5834a5-3b83-48c0-b83f-10e97d10a7d6\" (UID: \"df5834a5-3b83-48c0-b83f-10e97d10a7d6\") " Oct 14 08:52:29 crc kubenswrapper[4870]: I1014 08:52:29.859796 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/df5834a5-3b83-48c0-b83f-10e97d10a7d6-dns-svc\") pod \"df5834a5-3b83-48c0-b83f-10e97d10a7d6\" (UID: \"df5834a5-3b83-48c0-b83f-10e97d10a7d6\") " Oct 14 08:52:29 crc kubenswrapper[4870]: I1014 08:52:29.859859 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xct86\" (UniqueName: \"kubernetes.io/projected/df5834a5-3b83-48c0-b83f-10e97d10a7d6-kube-api-access-xct86\") pod \"df5834a5-3b83-48c0-b83f-10e97d10a7d6\" (UID: \"df5834a5-3b83-48c0-b83f-10e97d10a7d6\") " Oct 14 08:52:29 crc kubenswrapper[4870]: I1014 08:52:29.859895 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/df5834a5-3b83-48c0-b83f-10e97d10a7d6-ovsdbserver-sb\") pod \"df5834a5-3b83-48c0-b83f-10e97d10a7d6\" (UID: \"df5834a5-3b83-48c0-b83f-10e97d10a7d6\") " Oct 14 08:52:29 crc kubenswrapper[4870]: I1014 08:52:29.859933 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/df5834a5-3b83-48c0-b83f-10e97d10a7d6-config\") pod \"df5834a5-3b83-48c0-b83f-10e97d10a7d6\" (UID: \"df5834a5-3b83-48c0-b83f-10e97d10a7d6\") " Oct 14 08:52:29 crc kubenswrapper[4870]: I1014 08:52:29.865124 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/df5834a5-3b83-48c0-b83f-10e97d10a7d6-kube-api-access-xct86" (OuterVolumeSpecName: "kube-api-access-xct86") pod "df5834a5-3b83-48c0-b83f-10e97d10a7d6" (UID: "df5834a5-3b83-48c0-b83f-10e97d10a7d6"). InnerVolumeSpecName "kube-api-access-xct86". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:52:29 crc kubenswrapper[4870]: I1014 08:52:29.917116 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/df5834a5-3b83-48c0-b83f-10e97d10a7d6-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "df5834a5-3b83-48c0-b83f-10e97d10a7d6" (UID: "df5834a5-3b83-48c0-b83f-10e97d10a7d6"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:52:29 crc kubenswrapper[4870]: I1014 08:52:29.919631 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/df5834a5-3b83-48c0-b83f-10e97d10a7d6-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "df5834a5-3b83-48c0-b83f-10e97d10a7d6" (UID: "df5834a5-3b83-48c0-b83f-10e97d10a7d6"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:52:29 crc kubenswrapper[4870]: I1014 08:52:29.923778 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/df5834a5-3b83-48c0-b83f-10e97d10a7d6-config" (OuterVolumeSpecName: "config") pod "df5834a5-3b83-48c0-b83f-10e97d10a7d6" (UID: "df5834a5-3b83-48c0-b83f-10e97d10a7d6"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:52:29 crc kubenswrapper[4870]: I1014 08:52:29.942495 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/df5834a5-3b83-48c0-b83f-10e97d10a7d6-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "df5834a5-3b83-48c0-b83f-10e97d10a7d6" (UID: "df5834a5-3b83-48c0-b83f-10e97d10a7d6"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:52:29 crc kubenswrapper[4870]: I1014 08:52:29.961476 4870 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/df5834a5-3b83-48c0-b83f-10e97d10a7d6-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 14 08:52:29 crc kubenswrapper[4870]: I1014 08:52:29.961507 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xct86\" (UniqueName: \"kubernetes.io/projected/df5834a5-3b83-48c0-b83f-10e97d10a7d6-kube-api-access-xct86\") on node \"crc\" DevicePath \"\"" Oct 14 08:52:29 crc kubenswrapper[4870]: I1014 08:52:29.961526 4870 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/df5834a5-3b83-48c0-b83f-10e97d10a7d6-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 14 08:52:29 crc kubenswrapper[4870]: I1014 08:52:29.961538 4870 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/df5834a5-3b83-48c0-b83f-10e97d10a7d6-config\") on node \"crc\" DevicePath \"\"" Oct 14 08:52:29 crc kubenswrapper[4870]: I1014 08:52:29.961548 4870 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/df5834a5-3b83-48c0-b83f-10e97d10a7d6-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 14 08:52:30 crc kubenswrapper[4870]: I1014 08:52:30.058404 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7cf94d58c5-f8h59"] Oct 14 08:52:30 crc kubenswrapper[4870]: I1014 08:52:30.064132 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7cf94d58c5-f8h59"] Oct 14 08:52:30 crc kubenswrapper[4870]: I1014 08:52:30.735497 4870 generic.go:334] "Generic (PLEG): container finished" podID="4068efdd-8071-411b-9302-3e26f1bb44c2" containerID="88870aca262234f0dd719a72dad60be2f2c1437bb35f0d0e9af2e0596d07e866" exitCode=0 Oct 14 08:52:30 crc kubenswrapper[4870]: I1014 08:52:30.736801 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-vf77d" event={"ID":"4068efdd-8071-411b-9302-3e26f1bb44c2","Type":"ContainerDied","Data":"88870aca262234f0dd719a72dad60be2f2c1437bb35f0d0e9af2e0596d07e866"} Oct 14 08:52:31 crc kubenswrapper[4870]: I1014 08:52:31.056964 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="df5834a5-3b83-48c0-b83f-10e97d10a7d6" path="/var/lib/kubelet/pods/df5834a5-3b83-48c0-b83f-10e97d10a7d6/volumes" Oct 14 08:52:32 crc kubenswrapper[4870]: I1014 08:52:32.156401 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-vf77d" Oct 14 08:52:32 crc kubenswrapper[4870]: I1014 08:52:32.303011 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4068efdd-8071-411b-9302-3e26f1bb44c2-scripts\") pod \"4068efdd-8071-411b-9302-3e26f1bb44c2\" (UID: \"4068efdd-8071-411b-9302-3e26f1bb44c2\") " Oct 14 08:52:32 crc kubenswrapper[4870]: I1014 08:52:32.303348 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7p7vw\" (UniqueName: \"kubernetes.io/projected/4068efdd-8071-411b-9302-3e26f1bb44c2-kube-api-access-7p7vw\") pod \"4068efdd-8071-411b-9302-3e26f1bb44c2\" (UID: \"4068efdd-8071-411b-9302-3e26f1bb44c2\") " Oct 14 08:52:32 crc kubenswrapper[4870]: I1014 08:52:32.303473 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/4068efdd-8071-411b-9302-3e26f1bb44c2-fernet-keys\") pod \"4068efdd-8071-411b-9302-3e26f1bb44c2\" (UID: \"4068efdd-8071-411b-9302-3e26f1bb44c2\") " Oct 14 08:52:32 crc kubenswrapper[4870]: I1014 08:52:32.303513 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4068efdd-8071-411b-9302-3e26f1bb44c2-config-data\") pod \"4068efdd-8071-411b-9302-3e26f1bb44c2\" (UID: \"4068efdd-8071-411b-9302-3e26f1bb44c2\") " Oct 14 08:52:32 crc kubenswrapper[4870]: I1014 08:52:32.303564 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/4068efdd-8071-411b-9302-3e26f1bb44c2-credential-keys\") pod \"4068efdd-8071-411b-9302-3e26f1bb44c2\" (UID: \"4068efdd-8071-411b-9302-3e26f1bb44c2\") " Oct 14 08:52:32 crc kubenswrapper[4870]: I1014 08:52:32.303638 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4068efdd-8071-411b-9302-3e26f1bb44c2-combined-ca-bundle\") pod \"4068efdd-8071-411b-9302-3e26f1bb44c2\" (UID: \"4068efdd-8071-411b-9302-3e26f1bb44c2\") " Oct 14 08:52:32 crc kubenswrapper[4870]: I1014 08:52:32.308375 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4068efdd-8071-411b-9302-3e26f1bb44c2-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "4068efdd-8071-411b-9302-3e26f1bb44c2" (UID: "4068efdd-8071-411b-9302-3e26f1bb44c2"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:52:32 crc kubenswrapper[4870]: I1014 08:52:32.308809 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4068efdd-8071-411b-9302-3e26f1bb44c2-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "4068efdd-8071-411b-9302-3e26f1bb44c2" (UID: "4068efdd-8071-411b-9302-3e26f1bb44c2"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:52:32 crc kubenswrapper[4870]: I1014 08:52:32.309189 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4068efdd-8071-411b-9302-3e26f1bb44c2-kube-api-access-7p7vw" (OuterVolumeSpecName: "kube-api-access-7p7vw") pod "4068efdd-8071-411b-9302-3e26f1bb44c2" (UID: "4068efdd-8071-411b-9302-3e26f1bb44c2"). InnerVolumeSpecName "kube-api-access-7p7vw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:52:32 crc kubenswrapper[4870]: I1014 08:52:32.309875 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4068efdd-8071-411b-9302-3e26f1bb44c2-scripts" (OuterVolumeSpecName: "scripts") pod "4068efdd-8071-411b-9302-3e26f1bb44c2" (UID: "4068efdd-8071-411b-9302-3e26f1bb44c2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:52:32 crc kubenswrapper[4870]: I1014 08:52:32.327527 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4068efdd-8071-411b-9302-3e26f1bb44c2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4068efdd-8071-411b-9302-3e26f1bb44c2" (UID: "4068efdd-8071-411b-9302-3e26f1bb44c2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:52:32 crc kubenswrapper[4870]: I1014 08:52:32.331381 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4068efdd-8071-411b-9302-3e26f1bb44c2-config-data" (OuterVolumeSpecName: "config-data") pod "4068efdd-8071-411b-9302-3e26f1bb44c2" (UID: "4068efdd-8071-411b-9302-3e26f1bb44c2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:52:32 crc kubenswrapper[4870]: I1014 08:52:32.405126 4870 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4068efdd-8071-411b-9302-3e26f1bb44c2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 08:52:32 crc kubenswrapper[4870]: I1014 08:52:32.405173 4870 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4068efdd-8071-411b-9302-3e26f1bb44c2-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 08:52:32 crc kubenswrapper[4870]: I1014 08:52:32.405183 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7p7vw\" (UniqueName: \"kubernetes.io/projected/4068efdd-8071-411b-9302-3e26f1bb44c2-kube-api-access-7p7vw\") on node \"crc\" DevicePath \"\"" Oct 14 08:52:32 crc kubenswrapper[4870]: I1014 08:52:32.405193 4870 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/4068efdd-8071-411b-9302-3e26f1bb44c2-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 14 08:52:32 crc kubenswrapper[4870]: I1014 08:52:32.405202 4870 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4068efdd-8071-411b-9302-3e26f1bb44c2-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 08:52:32 crc kubenswrapper[4870]: I1014 08:52:32.405210 4870 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/4068efdd-8071-411b-9302-3e26f1bb44c2-credential-keys\") on node \"crc\" DevicePath \"\"" Oct 14 08:52:32 crc kubenswrapper[4870]: I1014 08:52:32.761818 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-vf77d" event={"ID":"4068efdd-8071-411b-9302-3e26f1bb44c2","Type":"ContainerDied","Data":"c58b211768430a9896f2c00226f788f493188b847047136e5ed2e96083862c99"} Oct 14 08:52:32 crc kubenswrapper[4870]: I1014 08:52:32.761956 4870 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c58b211768430a9896f2c00226f788f493188b847047136e5ed2e96083862c99" Oct 14 08:52:32 crc kubenswrapper[4870]: I1014 08:52:32.762351 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-vf77d" Oct 14 08:52:32 crc kubenswrapper[4870]: I1014 08:52:32.851339 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-58b648bb8f-6s7rq"] Oct 14 08:52:32 crc kubenswrapper[4870]: E1014 08:52:32.851843 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4068efdd-8071-411b-9302-3e26f1bb44c2" containerName="keystone-bootstrap" Oct 14 08:52:32 crc kubenswrapper[4870]: I1014 08:52:32.851863 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="4068efdd-8071-411b-9302-3e26f1bb44c2" containerName="keystone-bootstrap" Oct 14 08:52:32 crc kubenswrapper[4870]: E1014 08:52:32.851916 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df5834a5-3b83-48c0-b83f-10e97d10a7d6" containerName="init" Oct 14 08:52:32 crc kubenswrapper[4870]: I1014 08:52:32.851926 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="df5834a5-3b83-48c0-b83f-10e97d10a7d6" containerName="init" Oct 14 08:52:32 crc kubenswrapper[4870]: E1014 08:52:32.851944 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df5834a5-3b83-48c0-b83f-10e97d10a7d6" containerName="dnsmasq-dns" Oct 14 08:52:32 crc kubenswrapper[4870]: I1014 08:52:32.851954 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="df5834a5-3b83-48c0-b83f-10e97d10a7d6" containerName="dnsmasq-dns" Oct 14 08:52:32 crc kubenswrapper[4870]: I1014 08:52:32.852148 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="df5834a5-3b83-48c0-b83f-10e97d10a7d6" containerName="dnsmasq-dns" Oct 14 08:52:32 crc kubenswrapper[4870]: I1014 08:52:32.852169 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="4068efdd-8071-411b-9302-3e26f1bb44c2" containerName="keystone-bootstrap" Oct 14 08:52:32 crc kubenswrapper[4870]: I1014 08:52:32.853169 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-58b648bb8f-6s7rq" Oct 14 08:52:32 crc kubenswrapper[4870]: I1014 08:52:32.854946 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-mx4tj" Oct 14 08:52:32 crc kubenswrapper[4870]: I1014 08:52:32.854973 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 14 08:52:32 crc kubenswrapper[4870]: I1014 08:52:32.855251 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 14 08:52:32 crc kubenswrapper[4870]: I1014 08:52:32.856126 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 14 08:52:32 crc kubenswrapper[4870]: I1014 08:52:32.862728 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-58b648bb8f-6s7rq"] Oct 14 08:52:33 crc kubenswrapper[4870]: I1014 08:52:33.013638 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5df1fce8-53fe-42c0-a90c-683eb730b653-config-data\") pod \"keystone-58b648bb8f-6s7rq\" (UID: \"5df1fce8-53fe-42c0-a90c-683eb730b653\") " pod="openstack/keystone-58b648bb8f-6s7rq" Oct 14 08:52:33 crc kubenswrapper[4870]: I1014 08:52:33.013679 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/5df1fce8-53fe-42c0-a90c-683eb730b653-credential-keys\") pod \"keystone-58b648bb8f-6s7rq\" (UID: \"5df1fce8-53fe-42c0-a90c-683eb730b653\") " pod="openstack/keystone-58b648bb8f-6s7rq" Oct 14 08:52:33 crc kubenswrapper[4870]: I1014 08:52:33.013747 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/5df1fce8-53fe-42c0-a90c-683eb730b653-fernet-keys\") pod \"keystone-58b648bb8f-6s7rq\" (UID: \"5df1fce8-53fe-42c0-a90c-683eb730b653\") " pod="openstack/keystone-58b648bb8f-6s7rq" Oct 14 08:52:33 crc kubenswrapper[4870]: I1014 08:52:33.013835 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5df1fce8-53fe-42c0-a90c-683eb730b653-combined-ca-bundle\") pod \"keystone-58b648bb8f-6s7rq\" (UID: \"5df1fce8-53fe-42c0-a90c-683eb730b653\") " pod="openstack/keystone-58b648bb8f-6s7rq" Oct 14 08:52:33 crc kubenswrapper[4870]: I1014 08:52:33.013878 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5df1fce8-53fe-42c0-a90c-683eb730b653-scripts\") pod \"keystone-58b648bb8f-6s7rq\" (UID: \"5df1fce8-53fe-42c0-a90c-683eb730b653\") " pod="openstack/keystone-58b648bb8f-6s7rq" Oct 14 08:52:33 crc kubenswrapper[4870]: I1014 08:52:33.013930 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8c7hj\" (UniqueName: \"kubernetes.io/projected/5df1fce8-53fe-42c0-a90c-683eb730b653-kube-api-access-8c7hj\") pod \"keystone-58b648bb8f-6s7rq\" (UID: \"5df1fce8-53fe-42c0-a90c-683eb730b653\") " pod="openstack/keystone-58b648bb8f-6s7rq" Oct 14 08:52:33 crc kubenswrapper[4870]: I1014 08:52:33.115794 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5df1fce8-53fe-42c0-a90c-683eb730b653-config-data\") pod \"keystone-58b648bb8f-6s7rq\" (UID: \"5df1fce8-53fe-42c0-a90c-683eb730b653\") " pod="openstack/keystone-58b648bb8f-6s7rq" Oct 14 08:52:33 crc kubenswrapper[4870]: I1014 08:52:33.115840 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/5df1fce8-53fe-42c0-a90c-683eb730b653-credential-keys\") pod \"keystone-58b648bb8f-6s7rq\" (UID: \"5df1fce8-53fe-42c0-a90c-683eb730b653\") " pod="openstack/keystone-58b648bb8f-6s7rq" Oct 14 08:52:33 crc kubenswrapper[4870]: I1014 08:52:33.115891 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/5df1fce8-53fe-42c0-a90c-683eb730b653-fernet-keys\") pod \"keystone-58b648bb8f-6s7rq\" (UID: \"5df1fce8-53fe-42c0-a90c-683eb730b653\") " pod="openstack/keystone-58b648bb8f-6s7rq" Oct 14 08:52:33 crc kubenswrapper[4870]: I1014 08:52:33.115915 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5df1fce8-53fe-42c0-a90c-683eb730b653-combined-ca-bundle\") pod \"keystone-58b648bb8f-6s7rq\" (UID: \"5df1fce8-53fe-42c0-a90c-683eb730b653\") " pod="openstack/keystone-58b648bb8f-6s7rq" Oct 14 08:52:33 crc kubenswrapper[4870]: I1014 08:52:33.115948 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5df1fce8-53fe-42c0-a90c-683eb730b653-scripts\") pod \"keystone-58b648bb8f-6s7rq\" (UID: \"5df1fce8-53fe-42c0-a90c-683eb730b653\") " pod="openstack/keystone-58b648bb8f-6s7rq" Oct 14 08:52:33 crc kubenswrapper[4870]: I1014 08:52:33.115991 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8c7hj\" (UniqueName: \"kubernetes.io/projected/5df1fce8-53fe-42c0-a90c-683eb730b653-kube-api-access-8c7hj\") pod \"keystone-58b648bb8f-6s7rq\" (UID: \"5df1fce8-53fe-42c0-a90c-683eb730b653\") " pod="openstack/keystone-58b648bb8f-6s7rq" Oct 14 08:52:33 crc kubenswrapper[4870]: I1014 08:52:33.120148 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/5df1fce8-53fe-42c0-a90c-683eb730b653-credential-keys\") pod \"keystone-58b648bb8f-6s7rq\" (UID: \"5df1fce8-53fe-42c0-a90c-683eb730b653\") " pod="openstack/keystone-58b648bb8f-6s7rq" Oct 14 08:52:33 crc kubenswrapper[4870]: I1014 08:52:33.120267 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/5df1fce8-53fe-42c0-a90c-683eb730b653-fernet-keys\") pod \"keystone-58b648bb8f-6s7rq\" (UID: \"5df1fce8-53fe-42c0-a90c-683eb730b653\") " pod="openstack/keystone-58b648bb8f-6s7rq" Oct 14 08:52:33 crc kubenswrapper[4870]: I1014 08:52:33.120479 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5df1fce8-53fe-42c0-a90c-683eb730b653-scripts\") pod \"keystone-58b648bb8f-6s7rq\" (UID: \"5df1fce8-53fe-42c0-a90c-683eb730b653\") " pod="openstack/keystone-58b648bb8f-6s7rq" Oct 14 08:52:33 crc kubenswrapper[4870]: I1014 08:52:33.120714 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5df1fce8-53fe-42c0-a90c-683eb730b653-combined-ca-bundle\") pod \"keystone-58b648bb8f-6s7rq\" (UID: \"5df1fce8-53fe-42c0-a90c-683eb730b653\") " pod="openstack/keystone-58b648bb8f-6s7rq" Oct 14 08:52:33 crc kubenswrapper[4870]: I1014 08:52:33.121430 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5df1fce8-53fe-42c0-a90c-683eb730b653-config-data\") pod \"keystone-58b648bb8f-6s7rq\" (UID: \"5df1fce8-53fe-42c0-a90c-683eb730b653\") " pod="openstack/keystone-58b648bb8f-6s7rq" Oct 14 08:52:33 crc kubenswrapper[4870]: I1014 08:52:33.137148 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8c7hj\" (UniqueName: \"kubernetes.io/projected/5df1fce8-53fe-42c0-a90c-683eb730b653-kube-api-access-8c7hj\") pod \"keystone-58b648bb8f-6s7rq\" (UID: \"5df1fce8-53fe-42c0-a90c-683eb730b653\") " pod="openstack/keystone-58b648bb8f-6s7rq" Oct 14 08:52:33 crc kubenswrapper[4870]: I1014 08:52:33.213183 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-58b648bb8f-6s7rq" Oct 14 08:52:33 crc kubenswrapper[4870]: I1014 08:52:33.661050 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-58b648bb8f-6s7rq"] Oct 14 08:52:33 crc kubenswrapper[4870]: I1014 08:52:33.772633 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-58b648bb8f-6s7rq" event={"ID":"5df1fce8-53fe-42c0-a90c-683eb730b653","Type":"ContainerStarted","Data":"00e6027e8f0dbb10857f11bbd61dd2e112333589e7cb4d884fdf6b93c8760a6f"} Oct 14 08:52:34 crc kubenswrapper[4870]: I1014 08:52:34.785094 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-58b648bb8f-6s7rq" event={"ID":"5df1fce8-53fe-42c0-a90c-683eb730b653","Type":"ContainerStarted","Data":"d8c3dbeaa2097e26873829db1e59842b06a466b286d26a7120190a28e3cafe15"} Oct 14 08:52:34 crc kubenswrapper[4870]: I1014 08:52:34.785662 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-58b648bb8f-6s7rq" Oct 14 08:52:34 crc kubenswrapper[4870]: I1014 08:52:34.822648 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-58b648bb8f-6s7rq" podStartSLOduration=2.822621454 podStartE2EDuration="2.822621454s" podCreationTimestamp="2025-10-14 08:52:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 08:52:34.812376472 +0000 UTC m=+6690.509736883" watchObservedRunningTime="2025-10-14 08:52:34.822621454 +0000 UTC m=+6690.519981855" Oct 14 08:52:47 crc kubenswrapper[4870]: I1014 08:52:47.394606 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-mg6xs"] Oct 14 08:52:47 crc kubenswrapper[4870]: I1014 08:52:47.398759 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mg6xs" Oct 14 08:52:47 crc kubenswrapper[4870]: I1014 08:52:47.406963 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-mg6xs"] Oct 14 08:52:47 crc kubenswrapper[4870]: I1014 08:52:47.580106 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gpcqb\" (UniqueName: \"kubernetes.io/projected/05c579b3-c8dc-4ad0-8333-7e41e09d4690-kube-api-access-gpcqb\") pod \"redhat-marketplace-mg6xs\" (UID: \"05c579b3-c8dc-4ad0-8333-7e41e09d4690\") " pod="openshift-marketplace/redhat-marketplace-mg6xs" Oct 14 08:52:47 crc kubenswrapper[4870]: I1014 08:52:47.580589 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/05c579b3-c8dc-4ad0-8333-7e41e09d4690-utilities\") pod \"redhat-marketplace-mg6xs\" (UID: \"05c579b3-c8dc-4ad0-8333-7e41e09d4690\") " pod="openshift-marketplace/redhat-marketplace-mg6xs" Oct 14 08:52:47 crc kubenswrapper[4870]: I1014 08:52:47.580723 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/05c579b3-c8dc-4ad0-8333-7e41e09d4690-catalog-content\") pod \"redhat-marketplace-mg6xs\" (UID: \"05c579b3-c8dc-4ad0-8333-7e41e09d4690\") " pod="openshift-marketplace/redhat-marketplace-mg6xs" Oct 14 08:52:47 crc kubenswrapper[4870]: I1014 08:52:47.682395 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gpcqb\" (UniqueName: \"kubernetes.io/projected/05c579b3-c8dc-4ad0-8333-7e41e09d4690-kube-api-access-gpcqb\") pod \"redhat-marketplace-mg6xs\" (UID: \"05c579b3-c8dc-4ad0-8333-7e41e09d4690\") " pod="openshift-marketplace/redhat-marketplace-mg6xs" Oct 14 08:52:47 crc kubenswrapper[4870]: I1014 08:52:47.682764 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/05c579b3-c8dc-4ad0-8333-7e41e09d4690-utilities\") pod \"redhat-marketplace-mg6xs\" (UID: \"05c579b3-c8dc-4ad0-8333-7e41e09d4690\") " pod="openshift-marketplace/redhat-marketplace-mg6xs" Oct 14 08:52:47 crc kubenswrapper[4870]: I1014 08:52:47.682924 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/05c579b3-c8dc-4ad0-8333-7e41e09d4690-catalog-content\") pod \"redhat-marketplace-mg6xs\" (UID: \"05c579b3-c8dc-4ad0-8333-7e41e09d4690\") " pod="openshift-marketplace/redhat-marketplace-mg6xs" Oct 14 08:52:47 crc kubenswrapper[4870]: I1014 08:52:47.683191 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/05c579b3-c8dc-4ad0-8333-7e41e09d4690-utilities\") pod \"redhat-marketplace-mg6xs\" (UID: \"05c579b3-c8dc-4ad0-8333-7e41e09d4690\") " pod="openshift-marketplace/redhat-marketplace-mg6xs" Oct 14 08:52:47 crc kubenswrapper[4870]: I1014 08:52:47.683328 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/05c579b3-c8dc-4ad0-8333-7e41e09d4690-catalog-content\") pod \"redhat-marketplace-mg6xs\" (UID: \"05c579b3-c8dc-4ad0-8333-7e41e09d4690\") " pod="openshift-marketplace/redhat-marketplace-mg6xs" Oct 14 08:52:47 crc kubenswrapper[4870]: I1014 08:52:47.701692 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gpcqb\" (UniqueName: \"kubernetes.io/projected/05c579b3-c8dc-4ad0-8333-7e41e09d4690-kube-api-access-gpcqb\") pod \"redhat-marketplace-mg6xs\" (UID: \"05c579b3-c8dc-4ad0-8333-7e41e09d4690\") " pod="openshift-marketplace/redhat-marketplace-mg6xs" Oct 14 08:52:47 crc kubenswrapper[4870]: I1014 08:52:47.730814 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mg6xs" Oct 14 08:52:48 crc kubenswrapper[4870]: I1014 08:52:48.260147 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-mg6xs"] Oct 14 08:52:48 crc kubenswrapper[4870]: I1014 08:52:48.919960 4870 generic.go:334] "Generic (PLEG): container finished" podID="05c579b3-c8dc-4ad0-8333-7e41e09d4690" containerID="f12e09bc462ce07a7268b6a54e0ed5519b418d883768e8d9afa132c4b045fa74" exitCode=0 Oct 14 08:52:48 crc kubenswrapper[4870]: I1014 08:52:48.920053 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mg6xs" event={"ID":"05c579b3-c8dc-4ad0-8333-7e41e09d4690","Type":"ContainerDied","Data":"f12e09bc462ce07a7268b6a54e0ed5519b418d883768e8d9afa132c4b045fa74"} Oct 14 08:52:48 crc kubenswrapper[4870]: I1014 08:52:48.920475 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mg6xs" event={"ID":"05c579b3-c8dc-4ad0-8333-7e41e09d4690","Type":"ContainerStarted","Data":"3116acbfe6dda9e3d4c8025c272b138ca63bbfeed3bea67a5e818297cd674c98"} Oct 14 08:52:49 crc kubenswrapper[4870]: I1014 08:52:49.757746 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-8z46d"] Oct 14 08:52:49 crc kubenswrapper[4870]: I1014 08:52:49.764911 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8z46d" Oct 14 08:52:49 crc kubenswrapper[4870]: I1014 08:52:49.779905 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-8z46d"] Oct 14 08:52:49 crc kubenswrapper[4870]: I1014 08:52:49.948455 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3ca3a92e-2bef-4ad4-a44d-12aee242999e-catalog-content\") pod \"redhat-operators-8z46d\" (UID: \"3ca3a92e-2bef-4ad4-a44d-12aee242999e\") " pod="openshift-marketplace/redhat-operators-8z46d" Oct 14 08:52:49 crc kubenswrapper[4870]: I1014 08:52:49.948625 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kg956\" (UniqueName: \"kubernetes.io/projected/3ca3a92e-2bef-4ad4-a44d-12aee242999e-kube-api-access-kg956\") pod \"redhat-operators-8z46d\" (UID: \"3ca3a92e-2bef-4ad4-a44d-12aee242999e\") " pod="openshift-marketplace/redhat-operators-8z46d" Oct 14 08:52:49 crc kubenswrapper[4870]: I1014 08:52:49.948712 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3ca3a92e-2bef-4ad4-a44d-12aee242999e-utilities\") pod \"redhat-operators-8z46d\" (UID: \"3ca3a92e-2bef-4ad4-a44d-12aee242999e\") " pod="openshift-marketplace/redhat-operators-8z46d" Oct 14 08:52:50 crc kubenswrapper[4870]: I1014 08:52:50.049926 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kg956\" (UniqueName: \"kubernetes.io/projected/3ca3a92e-2bef-4ad4-a44d-12aee242999e-kube-api-access-kg956\") pod \"redhat-operators-8z46d\" (UID: \"3ca3a92e-2bef-4ad4-a44d-12aee242999e\") " pod="openshift-marketplace/redhat-operators-8z46d" Oct 14 08:52:50 crc kubenswrapper[4870]: I1014 08:52:50.050205 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3ca3a92e-2bef-4ad4-a44d-12aee242999e-utilities\") pod \"redhat-operators-8z46d\" (UID: \"3ca3a92e-2bef-4ad4-a44d-12aee242999e\") " pod="openshift-marketplace/redhat-operators-8z46d" Oct 14 08:52:50 crc kubenswrapper[4870]: I1014 08:52:50.050263 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3ca3a92e-2bef-4ad4-a44d-12aee242999e-catalog-content\") pod \"redhat-operators-8z46d\" (UID: \"3ca3a92e-2bef-4ad4-a44d-12aee242999e\") " pod="openshift-marketplace/redhat-operators-8z46d" Oct 14 08:52:50 crc kubenswrapper[4870]: I1014 08:52:50.050718 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3ca3a92e-2bef-4ad4-a44d-12aee242999e-utilities\") pod \"redhat-operators-8z46d\" (UID: \"3ca3a92e-2bef-4ad4-a44d-12aee242999e\") " pod="openshift-marketplace/redhat-operators-8z46d" Oct 14 08:52:50 crc kubenswrapper[4870]: I1014 08:52:50.050788 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3ca3a92e-2bef-4ad4-a44d-12aee242999e-catalog-content\") pod \"redhat-operators-8z46d\" (UID: \"3ca3a92e-2bef-4ad4-a44d-12aee242999e\") " pod="openshift-marketplace/redhat-operators-8z46d" Oct 14 08:52:50 crc kubenswrapper[4870]: I1014 08:52:50.068246 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kg956\" (UniqueName: \"kubernetes.io/projected/3ca3a92e-2bef-4ad4-a44d-12aee242999e-kube-api-access-kg956\") pod \"redhat-operators-8z46d\" (UID: \"3ca3a92e-2bef-4ad4-a44d-12aee242999e\") " pod="openshift-marketplace/redhat-operators-8z46d" Oct 14 08:52:50 crc kubenswrapper[4870]: I1014 08:52:50.105888 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8z46d" Oct 14 08:52:50 crc kubenswrapper[4870]: I1014 08:52:50.548844 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-8z46d"] Oct 14 08:52:50 crc kubenswrapper[4870]: I1014 08:52:50.939857 4870 generic.go:334] "Generic (PLEG): container finished" podID="3ca3a92e-2bef-4ad4-a44d-12aee242999e" containerID="86b45ee47d660589fbf59edd0c2257a29b297228858e7e9dcdeea817356dcfb7" exitCode=0 Oct 14 08:52:50 crc kubenswrapper[4870]: I1014 08:52:50.939997 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8z46d" event={"ID":"3ca3a92e-2bef-4ad4-a44d-12aee242999e","Type":"ContainerDied","Data":"86b45ee47d660589fbf59edd0c2257a29b297228858e7e9dcdeea817356dcfb7"} Oct 14 08:52:50 crc kubenswrapper[4870]: I1014 08:52:50.940497 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8z46d" event={"ID":"3ca3a92e-2bef-4ad4-a44d-12aee242999e","Type":"ContainerStarted","Data":"3ac623f609c613e2685f63771bc1a5970604014ac46aa0fe1d5d1b7fe9d1ea63"} Oct 14 08:52:50 crc kubenswrapper[4870]: I1014 08:52:50.943054 4870 generic.go:334] "Generic (PLEG): container finished" podID="05c579b3-c8dc-4ad0-8333-7e41e09d4690" containerID="6139c99953429ec9df30367d17871f1b6ff10cdab5f1b6838a4b2b8dcd8f7fba" exitCode=0 Oct 14 08:52:50 crc kubenswrapper[4870]: I1014 08:52:50.943108 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mg6xs" event={"ID":"05c579b3-c8dc-4ad0-8333-7e41e09d4690","Type":"ContainerDied","Data":"6139c99953429ec9df30367d17871f1b6ff10cdab5f1b6838a4b2b8dcd8f7fba"} Oct 14 08:52:51 crc kubenswrapper[4870]: I1014 08:52:51.953633 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8z46d" event={"ID":"3ca3a92e-2bef-4ad4-a44d-12aee242999e","Type":"ContainerStarted","Data":"1a9e77c6eb92cdbbce52ffa9521b887a642c016b22e0c3a4a6b8033850bfd1ed"} Oct 14 08:52:51 crc kubenswrapper[4870]: I1014 08:52:51.956701 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mg6xs" event={"ID":"05c579b3-c8dc-4ad0-8333-7e41e09d4690","Type":"ContainerStarted","Data":"dcbf74927fd8cb9114851ef0ee0c7cfba27555c697a54bf90aeaf917b6b7bb5b"} Oct 14 08:52:51 crc kubenswrapper[4870]: I1014 08:52:51.996425 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-mg6xs" podStartSLOduration=2.5320322280000003 podStartE2EDuration="4.996394273s" podCreationTimestamp="2025-10-14 08:52:47 +0000 UTC" firstStartedPulling="2025-10-14 08:52:48.92349402 +0000 UTC m=+6704.620854401" lastFinishedPulling="2025-10-14 08:52:51.387856045 +0000 UTC m=+6707.085216446" observedRunningTime="2025-10-14 08:52:51.992590079 +0000 UTC m=+6707.689950450" watchObservedRunningTime="2025-10-14 08:52:51.996394273 +0000 UTC m=+6707.693754684" Oct 14 08:52:52 crc kubenswrapper[4870]: I1014 08:52:52.971109 4870 generic.go:334] "Generic (PLEG): container finished" podID="3ca3a92e-2bef-4ad4-a44d-12aee242999e" containerID="1a9e77c6eb92cdbbce52ffa9521b887a642c016b22e0c3a4a6b8033850bfd1ed" exitCode=0 Oct 14 08:52:52 crc kubenswrapper[4870]: I1014 08:52:52.971162 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8z46d" event={"ID":"3ca3a92e-2bef-4ad4-a44d-12aee242999e","Type":"ContainerDied","Data":"1a9e77c6eb92cdbbce52ffa9521b887a642c016b22e0c3a4a6b8033850bfd1ed"} Oct 14 08:52:53 crc kubenswrapper[4870]: I1014 08:52:53.950637 4870 patch_prober.go:28] interesting pod/machine-config-daemon-7tvc7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 08:52:53 crc kubenswrapper[4870]: I1014 08:52:53.950723 4870 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 08:52:53 crc kubenswrapper[4870]: I1014 08:52:53.982225 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8z46d" event={"ID":"3ca3a92e-2bef-4ad4-a44d-12aee242999e","Type":"ContainerStarted","Data":"c85fdc2f34da32a3b4af97037618abc67b8a19870ca8684801fc3cce89ddcf40"} Oct 14 08:52:54 crc kubenswrapper[4870]: I1014 08:52:54.007547 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-8z46d" podStartSLOduration=2.440994117 podStartE2EDuration="5.007521356s" podCreationTimestamp="2025-10-14 08:52:49 +0000 UTC" firstStartedPulling="2025-10-14 08:52:50.941889162 +0000 UTC m=+6706.639249573" lastFinishedPulling="2025-10-14 08:52:53.508416401 +0000 UTC m=+6709.205776812" observedRunningTime="2025-10-14 08:52:54.005778554 +0000 UTC m=+6709.703138965" watchObservedRunningTime="2025-10-14 08:52:54.007521356 +0000 UTC m=+6709.704881767" Oct 14 08:52:57 crc kubenswrapper[4870]: I1014 08:52:57.731103 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-mg6xs" Oct 14 08:52:57 crc kubenswrapper[4870]: I1014 08:52:57.731827 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-mg6xs" Oct 14 08:52:57 crc kubenswrapper[4870]: I1014 08:52:57.776069 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-mg6xs" Oct 14 08:52:58 crc kubenswrapper[4870]: I1014 08:52:58.097124 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-mg6xs" Oct 14 08:52:59 crc kubenswrapper[4870]: I1014 08:52:59.150734 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-mg6xs"] Oct 14 08:53:00 crc kubenswrapper[4870]: I1014 08:53:00.042050 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-mg6xs" podUID="05c579b3-c8dc-4ad0-8333-7e41e09d4690" containerName="registry-server" containerID="cri-o://dcbf74927fd8cb9114851ef0ee0c7cfba27555c697a54bf90aeaf917b6b7bb5b" gracePeriod=2 Oct 14 08:53:00 crc kubenswrapper[4870]: I1014 08:53:00.106424 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-8z46d" Oct 14 08:53:00 crc kubenswrapper[4870]: I1014 08:53:00.106483 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-8z46d" Oct 14 08:53:00 crc kubenswrapper[4870]: I1014 08:53:00.176171 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-8z46d" Oct 14 08:53:00 crc kubenswrapper[4870]: I1014 08:53:00.564331 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mg6xs" Oct 14 08:53:00 crc kubenswrapper[4870]: I1014 08:53:00.712980 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gpcqb\" (UniqueName: \"kubernetes.io/projected/05c579b3-c8dc-4ad0-8333-7e41e09d4690-kube-api-access-gpcqb\") pod \"05c579b3-c8dc-4ad0-8333-7e41e09d4690\" (UID: \"05c579b3-c8dc-4ad0-8333-7e41e09d4690\") " Oct 14 08:53:00 crc kubenswrapper[4870]: I1014 08:53:00.713195 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/05c579b3-c8dc-4ad0-8333-7e41e09d4690-catalog-content\") pod \"05c579b3-c8dc-4ad0-8333-7e41e09d4690\" (UID: \"05c579b3-c8dc-4ad0-8333-7e41e09d4690\") " Oct 14 08:53:00 crc kubenswrapper[4870]: I1014 08:53:00.713379 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/05c579b3-c8dc-4ad0-8333-7e41e09d4690-utilities\") pod \"05c579b3-c8dc-4ad0-8333-7e41e09d4690\" (UID: \"05c579b3-c8dc-4ad0-8333-7e41e09d4690\") " Oct 14 08:53:00 crc kubenswrapper[4870]: I1014 08:53:00.715183 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/05c579b3-c8dc-4ad0-8333-7e41e09d4690-utilities" (OuterVolumeSpecName: "utilities") pod "05c579b3-c8dc-4ad0-8333-7e41e09d4690" (UID: "05c579b3-c8dc-4ad0-8333-7e41e09d4690"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:53:00 crc kubenswrapper[4870]: I1014 08:53:00.731623 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/05c579b3-c8dc-4ad0-8333-7e41e09d4690-kube-api-access-gpcqb" (OuterVolumeSpecName: "kube-api-access-gpcqb") pod "05c579b3-c8dc-4ad0-8333-7e41e09d4690" (UID: "05c579b3-c8dc-4ad0-8333-7e41e09d4690"). InnerVolumeSpecName "kube-api-access-gpcqb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:53:00 crc kubenswrapper[4870]: I1014 08:53:00.746353 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/05c579b3-c8dc-4ad0-8333-7e41e09d4690-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "05c579b3-c8dc-4ad0-8333-7e41e09d4690" (UID: "05c579b3-c8dc-4ad0-8333-7e41e09d4690"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:53:00 crc kubenswrapper[4870]: I1014 08:53:00.816001 4870 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/05c579b3-c8dc-4ad0-8333-7e41e09d4690-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 08:53:00 crc kubenswrapper[4870]: I1014 08:53:00.816047 4870 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/05c579b3-c8dc-4ad0-8333-7e41e09d4690-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 08:53:00 crc kubenswrapper[4870]: I1014 08:53:00.816068 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gpcqb\" (UniqueName: \"kubernetes.io/projected/05c579b3-c8dc-4ad0-8333-7e41e09d4690-kube-api-access-gpcqb\") on node \"crc\" DevicePath \"\"" Oct 14 08:53:01 crc kubenswrapper[4870]: I1014 08:53:01.055416 4870 generic.go:334] "Generic (PLEG): container finished" podID="05c579b3-c8dc-4ad0-8333-7e41e09d4690" containerID="dcbf74927fd8cb9114851ef0ee0c7cfba27555c697a54bf90aeaf917b6b7bb5b" exitCode=0 Oct 14 08:53:01 crc kubenswrapper[4870]: I1014 08:53:01.055550 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mg6xs" Oct 14 08:53:01 crc kubenswrapper[4870]: I1014 08:53:01.056550 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mg6xs" event={"ID":"05c579b3-c8dc-4ad0-8333-7e41e09d4690","Type":"ContainerDied","Data":"dcbf74927fd8cb9114851ef0ee0c7cfba27555c697a54bf90aeaf917b6b7bb5b"} Oct 14 08:53:01 crc kubenswrapper[4870]: I1014 08:53:01.056629 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mg6xs" event={"ID":"05c579b3-c8dc-4ad0-8333-7e41e09d4690","Type":"ContainerDied","Data":"3116acbfe6dda9e3d4c8025c272b138ca63bbfeed3bea67a5e818297cd674c98"} Oct 14 08:53:01 crc kubenswrapper[4870]: I1014 08:53:01.056671 4870 scope.go:117] "RemoveContainer" containerID="dcbf74927fd8cb9114851ef0ee0c7cfba27555c697a54bf90aeaf917b6b7bb5b" Oct 14 08:53:01 crc kubenswrapper[4870]: I1014 08:53:01.112528 4870 scope.go:117] "RemoveContainer" containerID="6139c99953429ec9df30367d17871f1b6ff10cdab5f1b6838a4b2b8dcd8f7fba" Oct 14 08:53:01 crc kubenswrapper[4870]: I1014 08:53:01.122758 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-mg6xs"] Oct 14 08:53:01 crc kubenswrapper[4870]: I1014 08:53:01.135367 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-mg6xs"] Oct 14 08:53:01 crc kubenswrapper[4870]: I1014 08:53:01.144759 4870 scope.go:117] "RemoveContainer" containerID="f12e09bc462ce07a7268b6a54e0ed5519b418d883768e8d9afa132c4b045fa74" Oct 14 08:53:01 crc kubenswrapper[4870]: I1014 08:53:01.151989 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-8z46d" Oct 14 08:53:01 crc kubenswrapper[4870]: I1014 08:53:01.185242 4870 scope.go:117] "RemoveContainer" containerID="dcbf74927fd8cb9114851ef0ee0c7cfba27555c697a54bf90aeaf917b6b7bb5b" Oct 14 08:53:01 crc kubenswrapper[4870]: E1014 08:53:01.185732 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dcbf74927fd8cb9114851ef0ee0c7cfba27555c697a54bf90aeaf917b6b7bb5b\": container with ID starting with dcbf74927fd8cb9114851ef0ee0c7cfba27555c697a54bf90aeaf917b6b7bb5b not found: ID does not exist" containerID="dcbf74927fd8cb9114851ef0ee0c7cfba27555c697a54bf90aeaf917b6b7bb5b" Oct 14 08:53:01 crc kubenswrapper[4870]: I1014 08:53:01.185773 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dcbf74927fd8cb9114851ef0ee0c7cfba27555c697a54bf90aeaf917b6b7bb5b"} err="failed to get container status \"dcbf74927fd8cb9114851ef0ee0c7cfba27555c697a54bf90aeaf917b6b7bb5b\": rpc error: code = NotFound desc = could not find container \"dcbf74927fd8cb9114851ef0ee0c7cfba27555c697a54bf90aeaf917b6b7bb5b\": container with ID starting with dcbf74927fd8cb9114851ef0ee0c7cfba27555c697a54bf90aeaf917b6b7bb5b not found: ID does not exist" Oct 14 08:53:01 crc kubenswrapper[4870]: I1014 08:53:01.185806 4870 scope.go:117] "RemoveContainer" containerID="6139c99953429ec9df30367d17871f1b6ff10cdab5f1b6838a4b2b8dcd8f7fba" Oct 14 08:53:01 crc kubenswrapper[4870]: E1014 08:53:01.186414 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6139c99953429ec9df30367d17871f1b6ff10cdab5f1b6838a4b2b8dcd8f7fba\": container with ID starting with 6139c99953429ec9df30367d17871f1b6ff10cdab5f1b6838a4b2b8dcd8f7fba not found: ID does not exist" containerID="6139c99953429ec9df30367d17871f1b6ff10cdab5f1b6838a4b2b8dcd8f7fba" Oct 14 08:53:01 crc kubenswrapper[4870]: I1014 08:53:01.186552 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6139c99953429ec9df30367d17871f1b6ff10cdab5f1b6838a4b2b8dcd8f7fba"} err="failed to get container status \"6139c99953429ec9df30367d17871f1b6ff10cdab5f1b6838a4b2b8dcd8f7fba\": rpc error: code = NotFound desc = could not find container \"6139c99953429ec9df30367d17871f1b6ff10cdab5f1b6838a4b2b8dcd8f7fba\": container with ID starting with 6139c99953429ec9df30367d17871f1b6ff10cdab5f1b6838a4b2b8dcd8f7fba not found: ID does not exist" Oct 14 08:53:01 crc kubenswrapper[4870]: I1014 08:53:01.186597 4870 scope.go:117] "RemoveContainer" containerID="f12e09bc462ce07a7268b6a54e0ed5519b418d883768e8d9afa132c4b045fa74" Oct 14 08:53:01 crc kubenswrapper[4870]: E1014 08:53:01.187011 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f12e09bc462ce07a7268b6a54e0ed5519b418d883768e8d9afa132c4b045fa74\": container with ID starting with f12e09bc462ce07a7268b6a54e0ed5519b418d883768e8d9afa132c4b045fa74 not found: ID does not exist" containerID="f12e09bc462ce07a7268b6a54e0ed5519b418d883768e8d9afa132c4b045fa74" Oct 14 08:53:01 crc kubenswrapper[4870]: I1014 08:53:01.187113 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f12e09bc462ce07a7268b6a54e0ed5519b418d883768e8d9afa132c4b045fa74"} err="failed to get container status \"f12e09bc462ce07a7268b6a54e0ed5519b418d883768e8d9afa132c4b045fa74\": rpc error: code = NotFound desc = could not find container \"f12e09bc462ce07a7268b6a54e0ed5519b418d883768e8d9afa132c4b045fa74\": container with ID starting with f12e09bc462ce07a7268b6a54e0ed5519b418d883768e8d9afa132c4b045fa74 not found: ID does not exist" Oct 14 08:53:02 crc kubenswrapper[4870]: I1014 08:53:02.349026 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-8z46d"] Oct 14 08:53:03 crc kubenswrapper[4870]: I1014 08:53:03.049801 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="05c579b3-c8dc-4ad0-8333-7e41e09d4690" path="/var/lib/kubelet/pods/05c579b3-c8dc-4ad0-8333-7e41e09d4690/volumes" Oct 14 08:53:03 crc kubenswrapper[4870]: I1014 08:53:03.078126 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-8z46d" podUID="3ca3a92e-2bef-4ad4-a44d-12aee242999e" containerName="registry-server" containerID="cri-o://c85fdc2f34da32a3b4af97037618abc67b8a19870ca8684801fc3cce89ddcf40" gracePeriod=2 Oct 14 08:53:03 crc kubenswrapper[4870]: I1014 08:53:03.502767 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8z46d" Oct 14 08:53:03 crc kubenswrapper[4870]: I1014 08:53:03.675711 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3ca3a92e-2bef-4ad4-a44d-12aee242999e-catalog-content\") pod \"3ca3a92e-2bef-4ad4-a44d-12aee242999e\" (UID: \"3ca3a92e-2bef-4ad4-a44d-12aee242999e\") " Oct 14 08:53:03 crc kubenswrapper[4870]: I1014 08:53:03.675797 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kg956\" (UniqueName: \"kubernetes.io/projected/3ca3a92e-2bef-4ad4-a44d-12aee242999e-kube-api-access-kg956\") pod \"3ca3a92e-2bef-4ad4-a44d-12aee242999e\" (UID: \"3ca3a92e-2bef-4ad4-a44d-12aee242999e\") " Oct 14 08:53:03 crc kubenswrapper[4870]: I1014 08:53:03.675877 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3ca3a92e-2bef-4ad4-a44d-12aee242999e-utilities\") pod \"3ca3a92e-2bef-4ad4-a44d-12aee242999e\" (UID: \"3ca3a92e-2bef-4ad4-a44d-12aee242999e\") " Oct 14 08:53:03 crc kubenswrapper[4870]: I1014 08:53:03.677052 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3ca3a92e-2bef-4ad4-a44d-12aee242999e-utilities" (OuterVolumeSpecName: "utilities") pod "3ca3a92e-2bef-4ad4-a44d-12aee242999e" (UID: "3ca3a92e-2bef-4ad4-a44d-12aee242999e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:53:03 crc kubenswrapper[4870]: I1014 08:53:03.685129 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ca3a92e-2bef-4ad4-a44d-12aee242999e-kube-api-access-kg956" (OuterVolumeSpecName: "kube-api-access-kg956") pod "3ca3a92e-2bef-4ad4-a44d-12aee242999e" (UID: "3ca3a92e-2bef-4ad4-a44d-12aee242999e"). InnerVolumeSpecName "kube-api-access-kg956". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:53:03 crc kubenswrapper[4870]: I1014 08:53:03.777590 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3ca3a92e-2bef-4ad4-a44d-12aee242999e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3ca3a92e-2bef-4ad4-a44d-12aee242999e" (UID: "3ca3a92e-2bef-4ad4-a44d-12aee242999e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:53:03 crc kubenswrapper[4870]: I1014 08:53:03.777692 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kg956\" (UniqueName: \"kubernetes.io/projected/3ca3a92e-2bef-4ad4-a44d-12aee242999e-kube-api-access-kg956\") on node \"crc\" DevicePath \"\"" Oct 14 08:53:03 crc kubenswrapper[4870]: I1014 08:53:03.777941 4870 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3ca3a92e-2bef-4ad4-a44d-12aee242999e-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 08:53:03 crc kubenswrapper[4870]: I1014 08:53:03.879644 4870 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3ca3a92e-2bef-4ad4-a44d-12aee242999e-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 08:53:04 crc kubenswrapper[4870]: I1014 08:53:04.089538 4870 generic.go:334] "Generic (PLEG): container finished" podID="3ca3a92e-2bef-4ad4-a44d-12aee242999e" containerID="c85fdc2f34da32a3b4af97037618abc67b8a19870ca8684801fc3cce89ddcf40" exitCode=0 Oct 14 08:53:04 crc kubenswrapper[4870]: I1014 08:53:04.089596 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8z46d" event={"ID":"3ca3a92e-2bef-4ad4-a44d-12aee242999e","Type":"ContainerDied","Data":"c85fdc2f34da32a3b4af97037618abc67b8a19870ca8684801fc3cce89ddcf40"} Oct 14 08:53:04 crc kubenswrapper[4870]: I1014 08:53:04.089685 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8z46d" event={"ID":"3ca3a92e-2bef-4ad4-a44d-12aee242999e","Type":"ContainerDied","Data":"3ac623f609c613e2685f63771bc1a5970604014ac46aa0fe1d5d1b7fe9d1ea63"} Oct 14 08:53:04 crc kubenswrapper[4870]: I1014 08:53:04.089754 4870 scope.go:117] "RemoveContainer" containerID="c85fdc2f34da32a3b4af97037618abc67b8a19870ca8684801fc3cce89ddcf40" Oct 14 08:53:04 crc kubenswrapper[4870]: I1014 08:53:04.090326 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8z46d" Oct 14 08:53:04 crc kubenswrapper[4870]: I1014 08:53:04.116038 4870 scope.go:117] "RemoveContainer" containerID="1a9e77c6eb92cdbbce52ffa9521b887a642c016b22e0c3a4a6b8033850bfd1ed" Oct 14 08:53:04 crc kubenswrapper[4870]: I1014 08:53:04.129080 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-8z46d"] Oct 14 08:53:04 crc kubenswrapper[4870]: I1014 08:53:04.140341 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-8z46d"] Oct 14 08:53:04 crc kubenswrapper[4870]: I1014 08:53:04.153124 4870 scope.go:117] "RemoveContainer" containerID="86b45ee47d660589fbf59edd0c2257a29b297228858e7e9dcdeea817356dcfb7" Oct 14 08:53:04 crc kubenswrapper[4870]: I1014 08:53:04.182389 4870 scope.go:117] "RemoveContainer" containerID="c85fdc2f34da32a3b4af97037618abc67b8a19870ca8684801fc3cce89ddcf40" Oct 14 08:53:04 crc kubenswrapper[4870]: E1014 08:53:04.182852 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c85fdc2f34da32a3b4af97037618abc67b8a19870ca8684801fc3cce89ddcf40\": container with ID starting with c85fdc2f34da32a3b4af97037618abc67b8a19870ca8684801fc3cce89ddcf40 not found: ID does not exist" containerID="c85fdc2f34da32a3b4af97037618abc67b8a19870ca8684801fc3cce89ddcf40" Oct 14 08:53:04 crc kubenswrapper[4870]: I1014 08:53:04.182969 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c85fdc2f34da32a3b4af97037618abc67b8a19870ca8684801fc3cce89ddcf40"} err="failed to get container status \"c85fdc2f34da32a3b4af97037618abc67b8a19870ca8684801fc3cce89ddcf40\": rpc error: code = NotFound desc = could not find container \"c85fdc2f34da32a3b4af97037618abc67b8a19870ca8684801fc3cce89ddcf40\": container with ID starting with c85fdc2f34da32a3b4af97037618abc67b8a19870ca8684801fc3cce89ddcf40 not found: ID does not exist" Oct 14 08:53:04 crc kubenswrapper[4870]: I1014 08:53:04.183054 4870 scope.go:117] "RemoveContainer" containerID="1a9e77c6eb92cdbbce52ffa9521b887a642c016b22e0c3a4a6b8033850bfd1ed" Oct 14 08:53:04 crc kubenswrapper[4870]: E1014 08:53:04.183336 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1a9e77c6eb92cdbbce52ffa9521b887a642c016b22e0c3a4a6b8033850bfd1ed\": container with ID starting with 1a9e77c6eb92cdbbce52ffa9521b887a642c016b22e0c3a4a6b8033850bfd1ed not found: ID does not exist" containerID="1a9e77c6eb92cdbbce52ffa9521b887a642c016b22e0c3a4a6b8033850bfd1ed" Oct 14 08:53:04 crc kubenswrapper[4870]: I1014 08:53:04.183362 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1a9e77c6eb92cdbbce52ffa9521b887a642c016b22e0c3a4a6b8033850bfd1ed"} err="failed to get container status \"1a9e77c6eb92cdbbce52ffa9521b887a642c016b22e0c3a4a6b8033850bfd1ed\": rpc error: code = NotFound desc = could not find container \"1a9e77c6eb92cdbbce52ffa9521b887a642c016b22e0c3a4a6b8033850bfd1ed\": container with ID starting with 1a9e77c6eb92cdbbce52ffa9521b887a642c016b22e0c3a4a6b8033850bfd1ed not found: ID does not exist" Oct 14 08:53:04 crc kubenswrapper[4870]: I1014 08:53:04.183395 4870 scope.go:117] "RemoveContainer" containerID="86b45ee47d660589fbf59edd0c2257a29b297228858e7e9dcdeea817356dcfb7" Oct 14 08:53:04 crc kubenswrapper[4870]: E1014 08:53:04.183810 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"86b45ee47d660589fbf59edd0c2257a29b297228858e7e9dcdeea817356dcfb7\": container with ID starting with 86b45ee47d660589fbf59edd0c2257a29b297228858e7e9dcdeea817356dcfb7 not found: ID does not exist" containerID="86b45ee47d660589fbf59edd0c2257a29b297228858e7e9dcdeea817356dcfb7" Oct 14 08:53:04 crc kubenswrapper[4870]: I1014 08:53:04.183912 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"86b45ee47d660589fbf59edd0c2257a29b297228858e7e9dcdeea817356dcfb7"} err="failed to get container status \"86b45ee47d660589fbf59edd0c2257a29b297228858e7e9dcdeea817356dcfb7\": rpc error: code = NotFound desc = could not find container \"86b45ee47d660589fbf59edd0c2257a29b297228858e7e9dcdeea817356dcfb7\": container with ID starting with 86b45ee47d660589fbf59edd0c2257a29b297228858e7e9dcdeea817356dcfb7 not found: ID does not exist" Oct 14 08:53:04 crc kubenswrapper[4870]: I1014 08:53:04.685560 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-58b648bb8f-6s7rq" Oct 14 08:53:05 crc kubenswrapper[4870]: I1014 08:53:05.048856 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ca3a92e-2bef-4ad4-a44d-12aee242999e" path="/var/lib/kubelet/pods/3ca3a92e-2bef-4ad4-a44d-12aee242999e/volumes" Oct 14 08:53:08 crc kubenswrapper[4870]: I1014 08:53:08.308208 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Oct 14 08:53:08 crc kubenswrapper[4870]: E1014 08:53:08.309189 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="05c579b3-c8dc-4ad0-8333-7e41e09d4690" containerName="extract-content" Oct 14 08:53:08 crc kubenswrapper[4870]: I1014 08:53:08.309218 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="05c579b3-c8dc-4ad0-8333-7e41e09d4690" containerName="extract-content" Oct 14 08:53:08 crc kubenswrapper[4870]: E1014 08:53:08.309248 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="05c579b3-c8dc-4ad0-8333-7e41e09d4690" containerName="registry-server" Oct 14 08:53:08 crc kubenswrapper[4870]: I1014 08:53:08.309264 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="05c579b3-c8dc-4ad0-8333-7e41e09d4690" containerName="registry-server" Oct 14 08:53:08 crc kubenswrapper[4870]: E1014 08:53:08.309291 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3ca3a92e-2bef-4ad4-a44d-12aee242999e" containerName="extract-content" Oct 14 08:53:08 crc kubenswrapper[4870]: I1014 08:53:08.309303 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="3ca3a92e-2bef-4ad4-a44d-12aee242999e" containerName="extract-content" Oct 14 08:53:08 crc kubenswrapper[4870]: E1014 08:53:08.309326 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="05c579b3-c8dc-4ad0-8333-7e41e09d4690" containerName="extract-utilities" Oct 14 08:53:08 crc kubenswrapper[4870]: I1014 08:53:08.309338 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="05c579b3-c8dc-4ad0-8333-7e41e09d4690" containerName="extract-utilities" Oct 14 08:53:08 crc kubenswrapper[4870]: E1014 08:53:08.309385 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3ca3a92e-2bef-4ad4-a44d-12aee242999e" containerName="extract-utilities" Oct 14 08:53:08 crc kubenswrapper[4870]: I1014 08:53:08.309397 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="3ca3a92e-2bef-4ad4-a44d-12aee242999e" containerName="extract-utilities" Oct 14 08:53:08 crc kubenswrapper[4870]: E1014 08:53:08.309412 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3ca3a92e-2bef-4ad4-a44d-12aee242999e" containerName="registry-server" Oct 14 08:53:08 crc kubenswrapper[4870]: I1014 08:53:08.309424 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="3ca3a92e-2bef-4ad4-a44d-12aee242999e" containerName="registry-server" Oct 14 08:53:08 crc kubenswrapper[4870]: I1014 08:53:08.309754 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="05c579b3-c8dc-4ad0-8333-7e41e09d4690" containerName="registry-server" Oct 14 08:53:08 crc kubenswrapper[4870]: I1014 08:53:08.309808 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="3ca3a92e-2bef-4ad4-a44d-12aee242999e" containerName="registry-server" Oct 14 08:53:08 crc kubenswrapper[4870]: I1014 08:53:08.310859 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 14 08:53:08 crc kubenswrapper[4870]: I1014 08:53:08.313873 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-zv9fk" Oct 14 08:53:08 crc kubenswrapper[4870]: I1014 08:53:08.315644 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Oct 14 08:53:08 crc kubenswrapper[4870]: I1014 08:53:08.315698 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Oct 14 08:53:08 crc kubenswrapper[4870]: I1014 08:53:08.319821 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 14 08:53:08 crc kubenswrapper[4870]: I1014 08:53:08.475974 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/fff3c418-2e18-4393-99c2-f24ad07746bb-openstack-config-secret\") pod \"openstackclient\" (UID: \"fff3c418-2e18-4393-99c2-f24ad07746bb\") " pod="openstack/openstackclient" Oct 14 08:53:08 crc kubenswrapper[4870]: I1014 08:53:08.476052 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rgcgw\" (UniqueName: \"kubernetes.io/projected/fff3c418-2e18-4393-99c2-f24ad07746bb-kube-api-access-rgcgw\") pod \"openstackclient\" (UID: \"fff3c418-2e18-4393-99c2-f24ad07746bb\") " pod="openstack/openstackclient" Oct 14 08:53:08 crc kubenswrapper[4870]: I1014 08:53:08.476103 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/fff3c418-2e18-4393-99c2-f24ad07746bb-openstack-config\") pod \"openstackclient\" (UID: \"fff3c418-2e18-4393-99c2-f24ad07746bb\") " pod="openstack/openstackclient" Oct 14 08:53:08 crc kubenswrapper[4870]: I1014 08:53:08.577877 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/fff3c418-2e18-4393-99c2-f24ad07746bb-openstack-config-secret\") pod \"openstackclient\" (UID: \"fff3c418-2e18-4393-99c2-f24ad07746bb\") " pod="openstack/openstackclient" Oct 14 08:53:08 crc kubenswrapper[4870]: I1014 08:53:08.577925 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rgcgw\" (UniqueName: \"kubernetes.io/projected/fff3c418-2e18-4393-99c2-f24ad07746bb-kube-api-access-rgcgw\") pod \"openstackclient\" (UID: \"fff3c418-2e18-4393-99c2-f24ad07746bb\") " pod="openstack/openstackclient" Oct 14 08:53:08 crc kubenswrapper[4870]: I1014 08:53:08.577953 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/fff3c418-2e18-4393-99c2-f24ad07746bb-openstack-config\") pod \"openstackclient\" (UID: \"fff3c418-2e18-4393-99c2-f24ad07746bb\") " pod="openstack/openstackclient" Oct 14 08:53:08 crc kubenswrapper[4870]: I1014 08:53:08.578805 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/fff3c418-2e18-4393-99c2-f24ad07746bb-openstack-config\") pod \"openstackclient\" (UID: \"fff3c418-2e18-4393-99c2-f24ad07746bb\") " pod="openstack/openstackclient" Oct 14 08:53:08 crc kubenswrapper[4870]: I1014 08:53:08.583900 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/fff3c418-2e18-4393-99c2-f24ad07746bb-openstack-config-secret\") pod \"openstackclient\" (UID: \"fff3c418-2e18-4393-99c2-f24ad07746bb\") " pod="openstack/openstackclient" Oct 14 08:53:08 crc kubenswrapper[4870]: I1014 08:53:08.598175 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rgcgw\" (UniqueName: \"kubernetes.io/projected/fff3c418-2e18-4393-99c2-f24ad07746bb-kube-api-access-rgcgw\") pod \"openstackclient\" (UID: \"fff3c418-2e18-4393-99c2-f24ad07746bb\") " pod="openstack/openstackclient" Oct 14 08:53:08 crc kubenswrapper[4870]: I1014 08:53:08.642292 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 14 08:53:09 crc kubenswrapper[4870]: I1014 08:53:09.156575 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 14 08:53:09 crc kubenswrapper[4870]: W1014 08:53:09.159767 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfff3c418_2e18_4393_99c2_f24ad07746bb.slice/crio-bf68aa4a836b6a89ad20ca3a4d9beaf5e761ceabe12a7a77f3d426fc22496edb WatchSource:0}: Error finding container bf68aa4a836b6a89ad20ca3a4d9beaf5e761ceabe12a7a77f3d426fc22496edb: Status 404 returned error can't find the container with id bf68aa4a836b6a89ad20ca3a4d9beaf5e761ceabe12a7a77f3d426fc22496edb Oct 14 08:53:09 crc kubenswrapper[4870]: I1014 08:53:09.163648 4870 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 14 08:53:10 crc kubenswrapper[4870]: I1014 08:53:10.158930 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"fff3c418-2e18-4393-99c2-f24ad07746bb","Type":"ContainerStarted","Data":"bf68aa4a836b6a89ad20ca3a4d9beaf5e761ceabe12a7a77f3d426fc22496edb"} Oct 14 08:53:20 crc kubenswrapper[4870]: I1014 08:53:20.264568 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"fff3c418-2e18-4393-99c2-f24ad07746bb","Type":"ContainerStarted","Data":"5f2890b7d74b88dc7653d308e99317cec8d57d4714978e696efc0b53a6d9c3e9"} Oct 14 08:53:20 crc kubenswrapper[4870]: I1014 08:53:20.279370 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=1.767881026 podStartE2EDuration="12.279348288s" podCreationTimestamp="2025-10-14 08:53:08 +0000 UTC" firstStartedPulling="2025-10-14 08:53:09.163171851 +0000 UTC m=+6724.860532252" lastFinishedPulling="2025-10-14 08:53:19.674639133 +0000 UTC m=+6735.371999514" observedRunningTime="2025-10-14 08:53:20.277609746 +0000 UTC m=+6735.974970117" watchObservedRunningTime="2025-10-14 08:53:20.279348288 +0000 UTC m=+6735.976708659" Oct 14 08:53:23 crc kubenswrapper[4870]: I1014 08:53:23.952333 4870 patch_prober.go:28] interesting pod/machine-config-daemon-7tvc7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 08:53:23 crc kubenswrapper[4870]: I1014 08:53:23.953003 4870 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 08:53:23 crc kubenswrapper[4870]: I1014 08:53:23.953090 4870 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" Oct 14 08:53:23 crc kubenswrapper[4870]: I1014 08:53:23.954036 4870 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"8c0f754fe94b7cb7b9bb5dcd613f7e42093968b49509d8ee37e8fd4db89eedd5"} pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 14 08:53:23 crc kubenswrapper[4870]: I1014 08:53:23.954138 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerName="machine-config-daemon" containerID="cri-o://8c0f754fe94b7cb7b9bb5dcd613f7e42093968b49509d8ee37e8fd4db89eedd5" gracePeriod=600 Oct 14 08:53:24 crc kubenswrapper[4870]: E1014 08:53:24.078920 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 08:53:24 crc kubenswrapper[4870]: I1014 08:53:24.302632 4870 generic.go:334] "Generic (PLEG): container finished" podID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerID="8c0f754fe94b7cb7b9bb5dcd613f7e42093968b49509d8ee37e8fd4db89eedd5" exitCode=0 Oct 14 08:53:24 crc kubenswrapper[4870]: I1014 08:53:24.302704 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" event={"ID":"874a5e45-dffd-4d17-b609-b7d3ed2eab07","Type":"ContainerDied","Data":"8c0f754fe94b7cb7b9bb5dcd613f7e42093968b49509d8ee37e8fd4db89eedd5"} Oct 14 08:53:24 crc kubenswrapper[4870]: I1014 08:53:24.303046 4870 scope.go:117] "RemoveContainer" containerID="39c59320b08c47c36fe9f61c79f5d172e92fda2c1c6fa3038de1fbd556880538" Oct 14 08:53:24 crc kubenswrapper[4870]: I1014 08:53:24.303592 4870 scope.go:117] "RemoveContainer" containerID="8c0f754fe94b7cb7b9bb5dcd613f7e42093968b49509d8ee37e8fd4db89eedd5" Oct 14 08:53:24 crc kubenswrapper[4870]: E1014 08:53:24.303850 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 08:53:35 crc kubenswrapper[4870]: I1014 08:53:35.042978 4870 scope.go:117] "RemoveContainer" containerID="8c0f754fe94b7cb7b9bb5dcd613f7e42093968b49509d8ee37e8fd4db89eedd5" Oct 14 08:53:35 crc kubenswrapper[4870]: E1014 08:53:35.047070 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 08:53:50 crc kubenswrapper[4870]: I1014 08:53:50.035389 4870 scope.go:117] "RemoveContainer" containerID="8c0f754fe94b7cb7b9bb5dcd613f7e42093968b49509d8ee37e8fd4db89eedd5" Oct 14 08:53:50 crc kubenswrapper[4870]: E1014 08:53:50.036508 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 08:54:05 crc kubenswrapper[4870]: I1014 08:54:05.045148 4870 scope.go:117] "RemoveContainer" containerID="8c0f754fe94b7cb7b9bb5dcd613f7e42093968b49509d8ee37e8fd4db89eedd5" Oct 14 08:54:05 crc kubenswrapper[4870]: E1014 08:54:05.046141 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 08:54:17 crc kubenswrapper[4870]: I1014 08:54:17.034048 4870 scope.go:117] "RemoveContainer" containerID="8c0f754fe94b7cb7b9bb5dcd613f7e42093968b49509d8ee37e8fd4db89eedd5" Oct 14 08:54:17 crc kubenswrapper[4870]: E1014 08:54:17.035272 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 08:54:31 crc kubenswrapper[4870]: I1014 08:54:31.035012 4870 scope.go:117] "RemoveContainer" containerID="8c0f754fe94b7cb7b9bb5dcd613f7e42093968b49509d8ee37e8fd4db89eedd5" Oct 14 08:54:31 crc kubenswrapper[4870]: E1014 08:54:31.035821 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 08:54:37 crc kubenswrapper[4870]: E1014 08:54:37.562784 4870 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.23:50328->38.102.83.23:45053: write tcp 38.102.83.23:50328->38.102.83.23:45053: write: broken pipe Oct 14 08:54:45 crc kubenswrapper[4870]: I1014 08:54:45.040576 4870 scope.go:117] "RemoveContainer" containerID="8c0f754fe94b7cb7b9bb5dcd613f7e42093968b49509d8ee37e8fd4db89eedd5" Oct 14 08:54:45 crc kubenswrapper[4870]: E1014 08:54:45.041396 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 08:54:49 crc kubenswrapper[4870]: I1014 08:54:49.877874 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-8ctvq"] Oct 14 08:54:49 crc kubenswrapper[4870]: I1014 08:54:49.879243 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-8ctvq" Oct 14 08:54:49 crc kubenswrapper[4870]: I1014 08:54:49.887748 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-8ctvq"] Oct 14 08:54:49 crc kubenswrapper[4870]: I1014 08:54:49.992535 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gjk59\" (UniqueName: \"kubernetes.io/projected/803b85ed-0a7c-475b-bbfd-23265017a113-kube-api-access-gjk59\") pod \"barbican-db-create-8ctvq\" (UID: \"803b85ed-0a7c-475b-bbfd-23265017a113\") " pod="openstack/barbican-db-create-8ctvq" Oct 14 08:54:50 crc kubenswrapper[4870]: I1014 08:54:50.094933 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gjk59\" (UniqueName: \"kubernetes.io/projected/803b85ed-0a7c-475b-bbfd-23265017a113-kube-api-access-gjk59\") pod \"barbican-db-create-8ctvq\" (UID: \"803b85ed-0a7c-475b-bbfd-23265017a113\") " pod="openstack/barbican-db-create-8ctvq" Oct 14 08:54:50 crc kubenswrapper[4870]: I1014 08:54:50.129898 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gjk59\" (UniqueName: \"kubernetes.io/projected/803b85ed-0a7c-475b-bbfd-23265017a113-kube-api-access-gjk59\") pod \"barbican-db-create-8ctvq\" (UID: \"803b85ed-0a7c-475b-bbfd-23265017a113\") " pod="openstack/barbican-db-create-8ctvq" Oct 14 08:54:50 crc kubenswrapper[4870]: I1014 08:54:50.205571 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-8ctvq" Oct 14 08:54:50 crc kubenswrapper[4870]: I1014 08:54:50.706719 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-8ctvq"] Oct 14 08:54:50 crc kubenswrapper[4870]: W1014 08:54:50.715090 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod803b85ed_0a7c_475b_bbfd_23265017a113.slice/crio-a46daa71e527a8025b6b38cc4f2dd13a40d6f3fe9df75a7fff6790c3bb69deb9 WatchSource:0}: Error finding container a46daa71e527a8025b6b38cc4f2dd13a40d6f3fe9df75a7fff6790c3bb69deb9: Status 404 returned error can't find the container with id a46daa71e527a8025b6b38cc4f2dd13a40d6f3fe9df75a7fff6790c3bb69deb9 Oct 14 08:54:51 crc kubenswrapper[4870]: I1014 08:54:51.119308 4870 generic.go:334] "Generic (PLEG): container finished" podID="803b85ed-0a7c-475b-bbfd-23265017a113" containerID="b99681a3b1ae47c5d44defc5cbe2dfbe66c07469f3187c7620644febe6ff30bc" exitCode=0 Oct 14 08:54:51 crc kubenswrapper[4870]: I1014 08:54:51.119467 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-8ctvq" event={"ID":"803b85ed-0a7c-475b-bbfd-23265017a113","Type":"ContainerDied","Data":"b99681a3b1ae47c5d44defc5cbe2dfbe66c07469f3187c7620644febe6ff30bc"} Oct 14 08:54:51 crc kubenswrapper[4870]: I1014 08:54:51.119639 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-8ctvq" event={"ID":"803b85ed-0a7c-475b-bbfd-23265017a113","Type":"ContainerStarted","Data":"a46daa71e527a8025b6b38cc4f2dd13a40d6f3fe9df75a7fff6790c3bb69deb9"} Oct 14 08:54:52 crc kubenswrapper[4870]: I1014 08:54:52.529415 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-8ctvq" Oct 14 08:54:52 crc kubenswrapper[4870]: I1014 08:54:52.705650 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gjk59\" (UniqueName: \"kubernetes.io/projected/803b85ed-0a7c-475b-bbfd-23265017a113-kube-api-access-gjk59\") pod \"803b85ed-0a7c-475b-bbfd-23265017a113\" (UID: \"803b85ed-0a7c-475b-bbfd-23265017a113\") " Oct 14 08:54:52 crc kubenswrapper[4870]: I1014 08:54:52.711773 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/803b85ed-0a7c-475b-bbfd-23265017a113-kube-api-access-gjk59" (OuterVolumeSpecName: "kube-api-access-gjk59") pod "803b85ed-0a7c-475b-bbfd-23265017a113" (UID: "803b85ed-0a7c-475b-bbfd-23265017a113"). InnerVolumeSpecName "kube-api-access-gjk59". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:54:52 crc kubenswrapper[4870]: I1014 08:54:52.808479 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gjk59\" (UniqueName: \"kubernetes.io/projected/803b85ed-0a7c-475b-bbfd-23265017a113-kube-api-access-gjk59\") on node \"crc\" DevicePath \"\"" Oct 14 08:54:53 crc kubenswrapper[4870]: I1014 08:54:53.138274 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-8ctvq" event={"ID":"803b85ed-0a7c-475b-bbfd-23265017a113","Type":"ContainerDied","Data":"a46daa71e527a8025b6b38cc4f2dd13a40d6f3fe9df75a7fff6790c3bb69deb9"} Oct 14 08:54:53 crc kubenswrapper[4870]: I1014 08:54:53.138359 4870 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a46daa71e527a8025b6b38cc4f2dd13a40d6f3fe9df75a7fff6790c3bb69deb9" Oct 14 08:54:53 crc kubenswrapper[4870]: I1014 08:54:53.138314 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-8ctvq" Oct 14 08:54:58 crc kubenswrapper[4870]: I1014 08:54:58.034104 4870 scope.go:117] "RemoveContainer" containerID="8c0f754fe94b7cb7b9bb5dcd613f7e42093968b49509d8ee37e8fd4db89eedd5" Oct 14 08:54:58 crc kubenswrapper[4870]: E1014 08:54:58.034912 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 08:55:00 crc kubenswrapper[4870]: I1014 08:55:00.023310 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-9d9f-account-create-wr5dt"] Oct 14 08:55:00 crc kubenswrapper[4870]: E1014 08:55:00.024468 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="803b85ed-0a7c-475b-bbfd-23265017a113" containerName="mariadb-database-create" Oct 14 08:55:00 crc kubenswrapper[4870]: I1014 08:55:00.024502 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="803b85ed-0a7c-475b-bbfd-23265017a113" containerName="mariadb-database-create" Oct 14 08:55:00 crc kubenswrapper[4870]: I1014 08:55:00.024873 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="803b85ed-0a7c-475b-bbfd-23265017a113" containerName="mariadb-database-create" Oct 14 08:55:00 crc kubenswrapper[4870]: I1014 08:55:00.025859 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-9d9f-account-create-wr5dt" Oct 14 08:55:00 crc kubenswrapper[4870]: I1014 08:55:00.029887 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Oct 14 08:55:00 crc kubenswrapper[4870]: I1014 08:55:00.038568 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-9d9f-account-create-wr5dt"] Oct 14 08:55:00 crc kubenswrapper[4870]: I1014 08:55:00.159056 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zmdtk\" (UniqueName: \"kubernetes.io/projected/efb802fd-6646-4bed-842e-4208b746bf7f-kube-api-access-zmdtk\") pod \"barbican-9d9f-account-create-wr5dt\" (UID: \"efb802fd-6646-4bed-842e-4208b746bf7f\") " pod="openstack/barbican-9d9f-account-create-wr5dt" Oct 14 08:55:00 crc kubenswrapper[4870]: I1014 08:55:00.260997 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zmdtk\" (UniqueName: \"kubernetes.io/projected/efb802fd-6646-4bed-842e-4208b746bf7f-kube-api-access-zmdtk\") pod \"barbican-9d9f-account-create-wr5dt\" (UID: \"efb802fd-6646-4bed-842e-4208b746bf7f\") " pod="openstack/barbican-9d9f-account-create-wr5dt" Oct 14 08:55:00 crc kubenswrapper[4870]: I1014 08:55:00.296246 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zmdtk\" (UniqueName: \"kubernetes.io/projected/efb802fd-6646-4bed-842e-4208b746bf7f-kube-api-access-zmdtk\") pod \"barbican-9d9f-account-create-wr5dt\" (UID: \"efb802fd-6646-4bed-842e-4208b746bf7f\") " pod="openstack/barbican-9d9f-account-create-wr5dt" Oct 14 08:55:00 crc kubenswrapper[4870]: I1014 08:55:00.357881 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-9d9f-account-create-wr5dt" Oct 14 08:55:00 crc kubenswrapper[4870]: I1014 08:55:00.830394 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-9d9f-account-create-wr5dt"] Oct 14 08:55:01 crc kubenswrapper[4870]: I1014 08:55:01.219538 4870 generic.go:334] "Generic (PLEG): container finished" podID="efb802fd-6646-4bed-842e-4208b746bf7f" containerID="265bb2cb99c9bcf3c27b483692e9d6d16286d098bd5f2f4414e975b6e72ffaa5" exitCode=0 Oct 14 08:55:01 crc kubenswrapper[4870]: I1014 08:55:01.219627 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-9d9f-account-create-wr5dt" event={"ID":"efb802fd-6646-4bed-842e-4208b746bf7f","Type":"ContainerDied","Data":"265bb2cb99c9bcf3c27b483692e9d6d16286d098bd5f2f4414e975b6e72ffaa5"} Oct 14 08:55:01 crc kubenswrapper[4870]: I1014 08:55:01.219963 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-9d9f-account-create-wr5dt" event={"ID":"efb802fd-6646-4bed-842e-4208b746bf7f","Type":"ContainerStarted","Data":"34519066e67afd238e1eb900aef47873534ee3901461003c3078ba75fe0a723b"} Oct 14 08:55:02 crc kubenswrapper[4870]: I1014 08:55:02.588734 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-9d9f-account-create-wr5dt" Oct 14 08:55:02 crc kubenswrapper[4870]: I1014 08:55:02.705929 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zmdtk\" (UniqueName: \"kubernetes.io/projected/efb802fd-6646-4bed-842e-4208b746bf7f-kube-api-access-zmdtk\") pod \"efb802fd-6646-4bed-842e-4208b746bf7f\" (UID: \"efb802fd-6646-4bed-842e-4208b746bf7f\") " Oct 14 08:55:02 crc kubenswrapper[4870]: I1014 08:55:02.717288 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efb802fd-6646-4bed-842e-4208b746bf7f-kube-api-access-zmdtk" (OuterVolumeSpecName: "kube-api-access-zmdtk") pod "efb802fd-6646-4bed-842e-4208b746bf7f" (UID: "efb802fd-6646-4bed-842e-4208b746bf7f"). InnerVolumeSpecName "kube-api-access-zmdtk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:55:02 crc kubenswrapper[4870]: I1014 08:55:02.807493 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zmdtk\" (UniqueName: \"kubernetes.io/projected/efb802fd-6646-4bed-842e-4208b746bf7f-kube-api-access-zmdtk\") on node \"crc\" DevicePath \"\"" Oct 14 08:55:03 crc kubenswrapper[4870]: I1014 08:55:03.269208 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-9d9f-account-create-wr5dt" event={"ID":"efb802fd-6646-4bed-842e-4208b746bf7f","Type":"ContainerDied","Data":"34519066e67afd238e1eb900aef47873534ee3901461003c3078ba75fe0a723b"} Oct 14 08:55:03 crc kubenswrapper[4870]: I1014 08:55:03.269869 4870 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="34519066e67afd238e1eb900aef47873534ee3901461003c3078ba75fe0a723b" Oct 14 08:55:03 crc kubenswrapper[4870]: I1014 08:55:03.269359 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-9d9f-account-create-wr5dt" Oct 14 08:55:05 crc kubenswrapper[4870]: I1014 08:55:05.278340 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-hqrsx"] Oct 14 08:55:05 crc kubenswrapper[4870]: E1014 08:55:05.278789 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="efb802fd-6646-4bed-842e-4208b746bf7f" containerName="mariadb-account-create" Oct 14 08:55:05 crc kubenswrapper[4870]: I1014 08:55:05.278808 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="efb802fd-6646-4bed-842e-4208b746bf7f" containerName="mariadb-account-create" Oct 14 08:55:05 crc kubenswrapper[4870]: I1014 08:55:05.279004 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="efb802fd-6646-4bed-842e-4208b746bf7f" containerName="mariadb-account-create" Oct 14 08:55:05 crc kubenswrapper[4870]: I1014 08:55:05.279850 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-hqrsx" Oct 14 08:55:05 crc kubenswrapper[4870]: I1014 08:55:05.283212 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-jt62z" Oct 14 08:55:05 crc kubenswrapper[4870]: I1014 08:55:05.283628 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Oct 14 08:55:05 crc kubenswrapper[4870]: I1014 08:55:05.296049 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-hqrsx"] Oct 14 08:55:05 crc kubenswrapper[4870]: I1014 08:55:05.461853 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/019b2f50-511d-44f0-9a63-05bbed8c08e8-db-sync-config-data\") pod \"barbican-db-sync-hqrsx\" (UID: \"019b2f50-511d-44f0-9a63-05bbed8c08e8\") " pod="openstack/barbican-db-sync-hqrsx" Oct 14 08:55:05 crc kubenswrapper[4870]: I1014 08:55:05.462568 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r5llk\" (UniqueName: \"kubernetes.io/projected/019b2f50-511d-44f0-9a63-05bbed8c08e8-kube-api-access-r5llk\") pod \"barbican-db-sync-hqrsx\" (UID: \"019b2f50-511d-44f0-9a63-05bbed8c08e8\") " pod="openstack/barbican-db-sync-hqrsx" Oct 14 08:55:05 crc kubenswrapper[4870]: I1014 08:55:05.462679 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/019b2f50-511d-44f0-9a63-05bbed8c08e8-combined-ca-bundle\") pod \"barbican-db-sync-hqrsx\" (UID: \"019b2f50-511d-44f0-9a63-05bbed8c08e8\") " pod="openstack/barbican-db-sync-hqrsx" Oct 14 08:55:05 crc kubenswrapper[4870]: I1014 08:55:05.563986 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r5llk\" (UniqueName: \"kubernetes.io/projected/019b2f50-511d-44f0-9a63-05bbed8c08e8-kube-api-access-r5llk\") pod \"barbican-db-sync-hqrsx\" (UID: \"019b2f50-511d-44f0-9a63-05bbed8c08e8\") " pod="openstack/barbican-db-sync-hqrsx" Oct 14 08:55:05 crc kubenswrapper[4870]: I1014 08:55:05.564086 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/019b2f50-511d-44f0-9a63-05bbed8c08e8-combined-ca-bundle\") pod \"barbican-db-sync-hqrsx\" (UID: \"019b2f50-511d-44f0-9a63-05bbed8c08e8\") " pod="openstack/barbican-db-sync-hqrsx" Oct 14 08:55:05 crc kubenswrapper[4870]: I1014 08:55:05.564258 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/019b2f50-511d-44f0-9a63-05bbed8c08e8-db-sync-config-data\") pod \"barbican-db-sync-hqrsx\" (UID: \"019b2f50-511d-44f0-9a63-05bbed8c08e8\") " pod="openstack/barbican-db-sync-hqrsx" Oct 14 08:55:05 crc kubenswrapper[4870]: I1014 08:55:05.570122 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/019b2f50-511d-44f0-9a63-05bbed8c08e8-db-sync-config-data\") pod \"barbican-db-sync-hqrsx\" (UID: \"019b2f50-511d-44f0-9a63-05bbed8c08e8\") " pod="openstack/barbican-db-sync-hqrsx" Oct 14 08:55:05 crc kubenswrapper[4870]: I1014 08:55:05.570789 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/019b2f50-511d-44f0-9a63-05bbed8c08e8-combined-ca-bundle\") pod \"barbican-db-sync-hqrsx\" (UID: \"019b2f50-511d-44f0-9a63-05bbed8c08e8\") " pod="openstack/barbican-db-sync-hqrsx" Oct 14 08:55:05 crc kubenswrapper[4870]: I1014 08:55:05.580641 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r5llk\" (UniqueName: \"kubernetes.io/projected/019b2f50-511d-44f0-9a63-05bbed8c08e8-kube-api-access-r5llk\") pod \"barbican-db-sync-hqrsx\" (UID: \"019b2f50-511d-44f0-9a63-05bbed8c08e8\") " pod="openstack/barbican-db-sync-hqrsx" Oct 14 08:55:05 crc kubenswrapper[4870]: I1014 08:55:05.600719 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-hqrsx" Oct 14 08:55:06 crc kubenswrapper[4870]: I1014 08:55:06.093961 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-hqrsx"] Oct 14 08:55:06 crc kubenswrapper[4870]: I1014 08:55:06.304334 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-hqrsx" event={"ID":"019b2f50-511d-44f0-9a63-05bbed8c08e8","Type":"ContainerStarted","Data":"1fe90f425320e8f66b06706832fca7ff34fde61d8dc382a38c24beedc2ec91c0"} Oct 14 08:55:10 crc kubenswrapper[4870]: I1014 08:55:10.351799 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-hqrsx" event={"ID":"019b2f50-511d-44f0-9a63-05bbed8c08e8","Type":"ContainerStarted","Data":"e446213e53ea1d18c3e066b85dce2df279affa0d0e817f3cbc1e3252be59fe91"} Oct 14 08:55:10 crc kubenswrapper[4870]: I1014 08:55:10.378068 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-hqrsx" podStartSLOduration=1.383281807 podStartE2EDuration="5.37804685s" podCreationTimestamp="2025-10-14 08:55:05 +0000 UTC" firstStartedPulling="2025-10-14 08:55:06.102473433 +0000 UTC m=+6841.799833804" lastFinishedPulling="2025-10-14 08:55:10.097238476 +0000 UTC m=+6845.794598847" observedRunningTime="2025-10-14 08:55:10.376737558 +0000 UTC m=+6846.074097939" watchObservedRunningTime="2025-10-14 08:55:10.37804685 +0000 UTC m=+6846.075407241" Oct 14 08:55:11 crc kubenswrapper[4870]: I1014 08:55:11.034771 4870 scope.go:117] "RemoveContainer" containerID="8c0f754fe94b7cb7b9bb5dcd613f7e42093968b49509d8ee37e8fd4db89eedd5" Oct 14 08:55:11 crc kubenswrapper[4870]: E1014 08:55:11.035138 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 08:55:12 crc kubenswrapper[4870]: E1014 08:55:12.530694 4870 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod019b2f50_511d_44f0_9a63_05bbed8c08e8.slice/crio-conmon-e446213e53ea1d18c3e066b85dce2df279affa0d0e817f3cbc1e3252be59fe91.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod019b2f50_511d_44f0_9a63_05bbed8c08e8.slice/crio-e446213e53ea1d18c3e066b85dce2df279affa0d0e817f3cbc1e3252be59fe91.scope\": RecentStats: unable to find data in memory cache]" Oct 14 08:55:13 crc kubenswrapper[4870]: I1014 08:55:13.387545 4870 generic.go:334] "Generic (PLEG): container finished" podID="019b2f50-511d-44f0-9a63-05bbed8c08e8" containerID="e446213e53ea1d18c3e066b85dce2df279affa0d0e817f3cbc1e3252be59fe91" exitCode=0 Oct 14 08:55:13 crc kubenswrapper[4870]: I1014 08:55:13.387626 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-hqrsx" event={"ID":"019b2f50-511d-44f0-9a63-05bbed8c08e8","Type":"ContainerDied","Data":"e446213e53ea1d18c3e066b85dce2df279affa0d0e817f3cbc1e3252be59fe91"} Oct 14 08:55:14 crc kubenswrapper[4870]: I1014 08:55:14.727666 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-hqrsx" Oct 14 08:55:14 crc kubenswrapper[4870]: I1014 08:55:14.835344 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/019b2f50-511d-44f0-9a63-05bbed8c08e8-db-sync-config-data\") pod \"019b2f50-511d-44f0-9a63-05bbed8c08e8\" (UID: \"019b2f50-511d-44f0-9a63-05bbed8c08e8\") " Oct 14 08:55:14 crc kubenswrapper[4870]: I1014 08:55:14.835510 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r5llk\" (UniqueName: \"kubernetes.io/projected/019b2f50-511d-44f0-9a63-05bbed8c08e8-kube-api-access-r5llk\") pod \"019b2f50-511d-44f0-9a63-05bbed8c08e8\" (UID: \"019b2f50-511d-44f0-9a63-05bbed8c08e8\") " Oct 14 08:55:14 crc kubenswrapper[4870]: I1014 08:55:14.835585 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/019b2f50-511d-44f0-9a63-05bbed8c08e8-combined-ca-bundle\") pod \"019b2f50-511d-44f0-9a63-05bbed8c08e8\" (UID: \"019b2f50-511d-44f0-9a63-05bbed8c08e8\") " Oct 14 08:55:14 crc kubenswrapper[4870]: I1014 08:55:14.841126 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/019b2f50-511d-44f0-9a63-05bbed8c08e8-kube-api-access-r5llk" (OuterVolumeSpecName: "kube-api-access-r5llk") pod "019b2f50-511d-44f0-9a63-05bbed8c08e8" (UID: "019b2f50-511d-44f0-9a63-05bbed8c08e8"). InnerVolumeSpecName "kube-api-access-r5llk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:55:14 crc kubenswrapper[4870]: I1014 08:55:14.841362 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/019b2f50-511d-44f0-9a63-05bbed8c08e8-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "019b2f50-511d-44f0-9a63-05bbed8c08e8" (UID: "019b2f50-511d-44f0-9a63-05bbed8c08e8"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:55:14 crc kubenswrapper[4870]: I1014 08:55:14.860204 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/019b2f50-511d-44f0-9a63-05bbed8c08e8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "019b2f50-511d-44f0-9a63-05bbed8c08e8" (UID: "019b2f50-511d-44f0-9a63-05bbed8c08e8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:55:14 crc kubenswrapper[4870]: I1014 08:55:14.937272 4870 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/019b2f50-511d-44f0-9a63-05bbed8c08e8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 08:55:14 crc kubenswrapper[4870]: I1014 08:55:14.937307 4870 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/019b2f50-511d-44f0-9a63-05bbed8c08e8-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 08:55:14 crc kubenswrapper[4870]: I1014 08:55:14.937320 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r5llk\" (UniqueName: \"kubernetes.io/projected/019b2f50-511d-44f0-9a63-05bbed8c08e8-kube-api-access-r5llk\") on node \"crc\" DevicePath \"\"" Oct 14 08:55:15 crc kubenswrapper[4870]: I1014 08:55:15.415659 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-hqrsx" event={"ID":"019b2f50-511d-44f0-9a63-05bbed8c08e8","Type":"ContainerDied","Data":"1fe90f425320e8f66b06706832fca7ff34fde61d8dc382a38c24beedc2ec91c0"} Oct 14 08:55:15 crc kubenswrapper[4870]: I1014 08:55:15.415707 4870 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1fe90f425320e8f66b06706832fca7ff34fde61d8dc382a38c24beedc2ec91c0" Oct 14 08:55:15 crc kubenswrapper[4870]: I1014 08:55:15.415755 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-hqrsx" Oct 14 08:55:15 crc kubenswrapper[4870]: I1014 08:55:15.624092 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-7c9b74fd49-gbcpm"] Oct 14 08:55:15 crc kubenswrapper[4870]: E1014 08:55:15.624694 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="019b2f50-511d-44f0-9a63-05bbed8c08e8" containerName="barbican-db-sync" Oct 14 08:55:15 crc kubenswrapper[4870]: I1014 08:55:15.624729 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="019b2f50-511d-44f0-9a63-05bbed8c08e8" containerName="barbican-db-sync" Oct 14 08:55:15 crc kubenswrapper[4870]: I1014 08:55:15.625244 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="019b2f50-511d-44f0-9a63-05bbed8c08e8" containerName="barbican-db-sync" Oct 14 08:55:15 crc kubenswrapper[4870]: I1014 08:55:15.626937 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-7c9b74fd49-gbcpm" Oct 14 08:55:15 crc kubenswrapper[4870]: I1014 08:55:15.629085 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Oct 14 08:55:15 crc kubenswrapper[4870]: I1014 08:55:15.629279 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-jt62z" Oct 14 08:55:15 crc kubenswrapper[4870]: I1014 08:55:15.629572 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Oct 14 08:55:15 crc kubenswrapper[4870]: I1014 08:55:15.635299 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-7c9b74fd49-gbcpm"] Oct 14 08:55:15 crc kubenswrapper[4870]: I1014 08:55:15.697311 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-68859d4c94-fx6jt"] Oct 14 08:55:15 crc kubenswrapper[4870]: I1014 08:55:15.700339 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-68859d4c94-fx6jt" Oct 14 08:55:15 crc kubenswrapper[4870]: I1014 08:55:15.706289 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Oct 14 08:55:15 crc kubenswrapper[4870]: I1014 08:55:15.729661 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-68859d4c94-fx6jt"] Oct 14 08:55:15 crc kubenswrapper[4870]: I1014 08:55:15.755917 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4446db55-22b0-4800-9ebc-f4b6b6578679-combined-ca-bundle\") pod \"barbican-worker-7c9b74fd49-gbcpm\" (UID: \"4446db55-22b0-4800-9ebc-f4b6b6578679\") " pod="openstack/barbican-worker-7c9b74fd49-gbcpm" Oct 14 08:55:15 crc kubenswrapper[4870]: I1014 08:55:15.756011 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4446db55-22b0-4800-9ebc-f4b6b6578679-logs\") pod \"barbican-worker-7c9b74fd49-gbcpm\" (UID: \"4446db55-22b0-4800-9ebc-f4b6b6578679\") " pod="openstack/barbican-worker-7c9b74fd49-gbcpm" Oct 14 08:55:15 crc kubenswrapper[4870]: I1014 08:55:15.756029 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4446db55-22b0-4800-9ebc-f4b6b6578679-config-data\") pod \"barbican-worker-7c9b74fd49-gbcpm\" (UID: \"4446db55-22b0-4800-9ebc-f4b6b6578679\") " pod="openstack/barbican-worker-7c9b74fd49-gbcpm" Oct 14 08:55:15 crc kubenswrapper[4870]: I1014 08:55:15.756048 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rnh4p\" (UniqueName: \"kubernetes.io/projected/4446db55-22b0-4800-9ebc-f4b6b6578679-kube-api-access-rnh4p\") pod \"barbican-worker-7c9b74fd49-gbcpm\" (UID: \"4446db55-22b0-4800-9ebc-f4b6b6578679\") " pod="openstack/barbican-worker-7c9b74fd49-gbcpm" Oct 14 08:55:15 crc kubenswrapper[4870]: I1014 08:55:15.756107 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4446db55-22b0-4800-9ebc-f4b6b6578679-config-data-custom\") pod \"barbican-worker-7c9b74fd49-gbcpm\" (UID: \"4446db55-22b0-4800-9ebc-f4b6b6578679\") " pod="openstack/barbican-worker-7c9b74fd49-gbcpm" Oct 14 08:55:15 crc kubenswrapper[4870]: I1014 08:55:15.804662 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-69f7d94959-2sm7c"] Oct 14 08:55:15 crc kubenswrapper[4870]: I1014 08:55:15.813294 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-69f7d94959-2sm7c"] Oct 14 08:55:15 crc kubenswrapper[4870]: I1014 08:55:15.813409 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-69f7d94959-2sm7c" Oct 14 08:55:15 crc kubenswrapper[4870]: I1014 08:55:15.860320 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4446db55-22b0-4800-9ebc-f4b6b6578679-logs\") pod \"barbican-worker-7c9b74fd49-gbcpm\" (UID: \"4446db55-22b0-4800-9ebc-f4b6b6578679\") " pod="openstack/barbican-worker-7c9b74fd49-gbcpm" Oct 14 08:55:15 crc kubenswrapper[4870]: I1014 08:55:15.860380 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4446db55-22b0-4800-9ebc-f4b6b6578679-config-data\") pod \"barbican-worker-7c9b74fd49-gbcpm\" (UID: \"4446db55-22b0-4800-9ebc-f4b6b6578679\") " pod="openstack/barbican-worker-7c9b74fd49-gbcpm" Oct 14 08:55:15 crc kubenswrapper[4870]: I1014 08:55:15.860411 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rnh4p\" (UniqueName: \"kubernetes.io/projected/4446db55-22b0-4800-9ebc-f4b6b6578679-kube-api-access-rnh4p\") pod \"barbican-worker-7c9b74fd49-gbcpm\" (UID: \"4446db55-22b0-4800-9ebc-f4b6b6578679\") " pod="openstack/barbican-worker-7c9b74fd49-gbcpm" Oct 14 08:55:15 crc kubenswrapper[4870]: I1014 08:55:15.860466 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-84tgb\" (UniqueName: \"kubernetes.io/projected/cc348a20-302d-407f-be5c-b690c39f12cf-kube-api-access-84tgb\") pod \"barbican-keystone-listener-68859d4c94-fx6jt\" (UID: \"cc348a20-302d-407f-be5c-b690c39f12cf\") " pod="openstack/barbican-keystone-listener-68859d4c94-fx6jt" Oct 14 08:55:15 crc kubenswrapper[4870]: I1014 08:55:15.860508 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cc348a20-302d-407f-be5c-b690c39f12cf-logs\") pod \"barbican-keystone-listener-68859d4c94-fx6jt\" (UID: \"cc348a20-302d-407f-be5c-b690c39f12cf\") " pod="openstack/barbican-keystone-listener-68859d4c94-fx6jt" Oct 14 08:55:15 crc kubenswrapper[4870]: I1014 08:55:15.860547 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc348a20-302d-407f-be5c-b690c39f12cf-combined-ca-bundle\") pod \"barbican-keystone-listener-68859d4c94-fx6jt\" (UID: \"cc348a20-302d-407f-be5c-b690c39f12cf\") " pod="openstack/barbican-keystone-listener-68859d4c94-fx6jt" Oct 14 08:55:15 crc kubenswrapper[4870]: I1014 08:55:15.860571 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4446db55-22b0-4800-9ebc-f4b6b6578679-config-data-custom\") pod \"barbican-worker-7c9b74fd49-gbcpm\" (UID: \"4446db55-22b0-4800-9ebc-f4b6b6578679\") " pod="openstack/barbican-worker-7c9b74fd49-gbcpm" Oct 14 08:55:15 crc kubenswrapper[4870]: I1014 08:55:15.860605 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cc348a20-302d-407f-be5c-b690c39f12cf-config-data\") pod \"barbican-keystone-listener-68859d4c94-fx6jt\" (UID: \"cc348a20-302d-407f-be5c-b690c39f12cf\") " pod="openstack/barbican-keystone-listener-68859d4c94-fx6jt" Oct 14 08:55:15 crc kubenswrapper[4870]: I1014 08:55:15.860636 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4446db55-22b0-4800-9ebc-f4b6b6578679-combined-ca-bundle\") pod \"barbican-worker-7c9b74fd49-gbcpm\" (UID: \"4446db55-22b0-4800-9ebc-f4b6b6578679\") " pod="openstack/barbican-worker-7c9b74fd49-gbcpm" Oct 14 08:55:15 crc kubenswrapper[4870]: I1014 08:55:15.860710 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cc348a20-302d-407f-be5c-b690c39f12cf-config-data-custom\") pod \"barbican-keystone-listener-68859d4c94-fx6jt\" (UID: \"cc348a20-302d-407f-be5c-b690c39f12cf\") " pod="openstack/barbican-keystone-listener-68859d4c94-fx6jt" Oct 14 08:55:15 crc kubenswrapper[4870]: I1014 08:55:15.860854 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4446db55-22b0-4800-9ebc-f4b6b6578679-logs\") pod \"barbican-worker-7c9b74fd49-gbcpm\" (UID: \"4446db55-22b0-4800-9ebc-f4b6b6578679\") " pod="openstack/barbican-worker-7c9b74fd49-gbcpm" Oct 14 08:55:15 crc kubenswrapper[4870]: I1014 08:55:15.866755 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4446db55-22b0-4800-9ebc-f4b6b6578679-combined-ca-bundle\") pod \"barbican-worker-7c9b74fd49-gbcpm\" (UID: \"4446db55-22b0-4800-9ebc-f4b6b6578679\") " pod="openstack/barbican-worker-7c9b74fd49-gbcpm" Oct 14 08:55:15 crc kubenswrapper[4870]: I1014 08:55:15.866763 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-d46667c8-kddgf"] Oct 14 08:55:15 crc kubenswrapper[4870]: I1014 08:55:15.869462 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-d46667c8-kddgf" Oct 14 08:55:15 crc kubenswrapper[4870]: I1014 08:55:15.872774 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Oct 14 08:55:15 crc kubenswrapper[4870]: I1014 08:55:15.873829 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4446db55-22b0-4800-9ebc-f4b6b6578679-config-data\") pod \"barbican-worker-7c9b74fd49-gbcpm\" (UID: \"4446db55-22b0-4800-9ebc-f4b6b6578679\") " pod="openstack/barbican-worker-7c9b74fd49-gbcpm" Oct 14 08:55:15 crc kubenswrapper[4870]: I1014 08:55:15.877615 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-d46667c8-kddgf"] Oct 14 08:55:15 crc kubenswrapper[4870]: I1014 08:55:15.882734 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4446db55-22b0-4800-9ebc-f4b6b6578679-config-data-custom\") pod \"barbican-worker-7c9b74fd49-gbcpm\" (UID: \"4446db55-22b0-4800-9ebc-f4b6b6578679\") " pod="openstack/barbican-worker-7c9b74fd49-gbcpm" Oct 14 08:55:15 crc kubenswrapper[4870]: I1014 08:55:15.885013 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rnh4p\" (UniqueName: \"kubernetes.io/projected/4446db55-22b0-4800-9ebc-f4b6b6578679-kube-api-access-rnh4p\") pod \"barbican-worker-7c9b74fd49-gbcpm\" (UID: \"4446db55-22b0-4800-9ebc-f4b6b6578679\") " pod="openstack/barbican-worker-7c9b74fd49-gbcpm" Oct 14 08:55:15 crc kubenswrapper[4870]: I1014 08:55:15.942589 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-7c9b74fd49-gbcpm" Oct 14 08:55:15 crc kubenswrapper[4870]: I1014 08:55:15.962638 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lblql\" (UniqueName: \"kubernetes.io/projected/2491fadf-97f6-4e7e-932b-3ca607c015ab-kube-api-access-lblql\") pod \"dnsmasq-dns-69f7d94959-2sm7c\" (UID: \"2491fadf-97f6-4e7e-932b-3ca607c015ab\") " pod="openstack/dnsmasq-dns-69f7d94959-2sm7c" Oct 14 08:55:15 crc kubenswrapper[4870]: I1014 08:55:15.962955 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/658a9360-4e8e-40ca-9575-5501fb52be26-combined-ca-bundle\") pod \"barbican-api-d46667c8-kddgf\" (UID: \"658a9360-4e8e-40ca-9575-5501fb52be26\") " pod="openstack/barbican-api-d46667c8-kddgf" Oct 14 08:55:15 crc kubenswrapper[4870]: I1014 08:55:15.962983 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cc348a20-302d-407f-be5c-b690c39f12cf-config-data-custom\") pod \"barbican-keystone-listener-68859d4c94-fx6jt\" (UID: \"cc348a20-302d-407f-be5c-b690c39f12cf\") " pod="openstack/barbican-keystone-listener-68859d4c94-fx6jt" Oct 14 08:55:15 crc kubenswrapper[4870]: I1014 08:55:15.963014 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-84tgb\" (UniqueName: \"kubernetes.io/projected/cc348a20-302d-407f-be5c-b690c39f12cf-kube-api-access-84tgb\") pod \"barbican-keystone-listener-68859d4c94-fx6jt\" (UID: \"cc348a20-302d-407f-be5c-b690c39f12cf\") " pod="openstack/barbican-keystone-listener-68859d4c94-fx6jt" Oct 14 08:55:15 crc kubenswrapper[4870]: I1014 08:55:15.963032 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2491fadf-97f6-4e7e-932b-3ca607c015ab-config\") pod \"dnsmasq-dns-69f7d94959-2sm7c\" (UID: \"2491fadf-97f6-4e7e-932b-3ca607c015ab\") " pod="openstack/dnsmasq-dns-69f7d94959-2sm7c" Oct 14 08:55:15 crc kubenswrapper[4870]: I1014 08:55:15.963059 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cc348a20-302d-407f-be5c-b690c39f12cf-logs\") pod \"barbican-keystone-listener-68859d4c94-fx6jt\" (UID: \"cc348a20-302d-407f-be5c-b690c39f12cf\") " pod="openstack/barbican-keystone-listener-68859d4c94-fx6jt" Oct 14 08:55:15 crc kubenswrapper[4870]: I1014 08:55:15.963078 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2491fadf-97f6-4e7e-932b-3ca607c015ab-dns-svc\") pod \"dnsmasq-dns-69f7d94959-2sm7c\" (UID: \"2491fadf-97f6-4e7e-932b-3ca607c015ab\") " pod="openstack/dnsmasq-dns-69f7d94959-2sm7c" Oct 14 08:55:15 crc kubenswrapper[4870]: I1014 08:55:15.963096 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/658a9360-4e8e-40ca-9575-5501fb52be26-logs\") pod \"barbican-api-d46667c8-kddgf\" (UID: \"658a9360-4e8e-40ca-9575-5501fb52be26\") " pod="openstack/barbican-api-d46667c8-kddgf" Oct 14 08:55:15 crc kubenswrapper[4870]: I1014 08:55:15.963120 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2491fadf-97f6-4e7e-932b-3ca607c015ab-ovsdbserver-nb\") pod \"dnsmasq-dns-69f7d94959-2sm7c\" (UID: \"2491fadf-97f6-4e7e-932b-3ca607c015ab\") " pod="openstack/dnsmasq-dns-69f7d94959-2sm7c" Oct 14 08:55:15 crc kubenswrapper[4870]: I1014 08:55:15.963144 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc348a20-302d-407f-be5c-b690c39f12cf-combined-ca-bundle\") pod \"barbican-keystone-listener-68859d4c94-fx6jt\" (UID: \"cc348a20-302d-407f-be5c-b690c39f12cf\") " pod="openstack/barbican-keystone-listener-68859d4c94-fx6jt" Oct 14 08:55:15 crc kubenswrapper[4870]: I1014 08:55:15.963174 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cc348a20-302d-407f-be5c-b690c39f12cf-config-data\") pod \"barbican-keystone-listener-68859d4c94-fx6jt\" (UID: \"cc348a20-302d-407f-be5c-b690c39f12cf\") " pod="openstack/barbican-keystone-listener-68859d4c94-fx6jt" Oct 14 08:55:15 crc kubenswrapper[4870]: I1014 08:55:15.963210 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/658a9360-4e8e-40ca-9575-5501fb52be26-config-data-custom\") pod \"barbican-api-d46667c8-kddgf\" (UID: \"658a9360-4e8e-40ca-9575-5501fb52be26\") " pod="openstack/barbican-api-d46667c8-kddgf" Oct 14 08:55:15 crc kubenswrapper[4870]: I1014 08:55:15.963231 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2491fadf-97f6-4e7e-932b-3ca607c015ab-ovsdbserver-sb\") pod \"dnsmasq-dns-69f7d94959-2sm7c\" (UID: \"2491fadf-97f6-4e7e-932b-3ca607c015ab\") " pod="openstack/dnsmasq-dns-69f7d94959-2sm7c" Oct 14 08:55:15 crc kubenswrapper[4870]: I1014 08:55:15.963249 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-frzh7\" (UniqueName: \"kubernetes.io/projected/658a9360-4e8e-40ca-9575-5501fb52be26-kube-api-access-frzh7\") pod \"barbican-api-d46667c8-kddgf\" (UID: \"658a9360-4e8e-40ca-9575-5501fb52be26\") " pod="openstack/barbican-api-d46667c8-kddgf" Oct 14 08:55:15 crc kubenswrapper[4870]: I1014 08:55:15.963264 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/658a9360-4e8e-40ca-9575-5501fb52be26-config-data\") pod \"barbican-api-d46667c8-kddgf\" (UID: \"658a9360-4e8e-40ca-9575-5501fb52be26\") " pod="openstack/barbican-api-d46667c8-kddgf" Oct 14 08:55:15 crc kubenswrapper[4870]: I1014 08:55:15.966938 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cc348a20-302d-407f-be5c-b690c39f12cf-config-data-custom\") pod \"barbican-keystone-listener-68859d4c94-fx6jt\" (UID: \"cc348a20-302d-407f-be5c-b690c39f12cf\") " pod="openstack/barbican-keystone-listener-68859d4c94-fx6jt" Oct 14 08:55:15 crc kubenswrapper[4870]: I1014 08:55:15.967415 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cc348a20-302d-407f-be5c-b690c39f12cf-logs\") pod \"barbican-keystone-listener-68859d4c94-fx6jt\" (UID: \"cc348a20-302d-407f-be5c-b690c39f12cf\") " pod="openstack/barbican-keystone-listener-68859d4c94-fx6jt" Oct 14 08:55:15 crc kubenswrapper[4870]: I1014 08:55:15.974162 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc348a20-302d-407f-be5c-b690c39f12cf-combined-ca-bundle\") pod \"barbican-keystone-listener-68859d4c94-fx6jt\" (UID: \"cc348a20-302d-407f-be5c-b690c39f12cf\") " pod="openstack/barbican-keystone-listener-68859d4c94-fx6jt" Oct 14 08:55:15 crc kubenswrapper[4870]: I1014 08:55:15.979365 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cc348a20-302d-407f-be5c-b690c39f12cf-config-data\") pod \"barbican-keystone-listener-68859d4c94-fx6jt\" (UID: \"cc348a20-302d-407f-be5c-b690c39f12cf\") " pod="openstack/barbican-keystone-listener-68859d4c94-fx6jt" Oct 14 08:55:15 crc kubenswrapper[4870]: I1014 08:55:15.986088 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-84tgb\" (UniqueName: \"kubernetes.io/projected/cc348a20-302d-407f-be5c-b690c39f12cf-kube-api-access-84tgb\") pod \"barbican-keystone-listener-68859d4c94-fx6jt\" (UID: \"cc348a20-302d-407f-be5c-b690c39f12cf\") " pod="openstack/barbican-keystone-listener-68859d4c94-fx6jt" Oct 14 08:55:16 crc kubenswrapper[4870]: I1014 08:55:16.033505 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-68859d4c94-fx6jt" Oct 14 08:55:16 crc kubenswrapper[4870]: I1014 08:55:16.067680 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2491fadf-97f6-4e7e-932b-3ca607c015ab-config\") pod \"dnsmasq-dns-69f7d94959-2sm7c\" (UID: \"2491fadf-97f6-4e7e-932b-3ca607c015ab\") " pod="openstack/dnsmasq-dns-69f7d94959-2sm7c" Oct 14 08:55:16 crc kubenswrapper[4870]: I1014 08:55:16.067778 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2491fadf-97f6-4e7e-932b-3ca607c015ab-dns-svc\") pod \"dnsmasq-dns-69f7d94959-2sm7c\" (UID: \"2491fadf-97f6-4e7e-932b-3ca607c015ab\") " pod="openstack/dnsmasq-dns-69f7d94959-2sm7c" Oct 14 08:55:16 crc kubenswrapper[4870]: I1014 08:55:16.067845 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/658a9360-4e8e-40ca-9575-5501fb52be26-logs\") pod \"barbican-api-d46667c8-kddgf\" (UID: \"658a9360-4e8e-40ca-9575-5501fb52be26\") " pod="openstack/barbican-api-d46667c8-kddgf" Oct 14 08:55:16 crc kubenswrapper[4870]: I1014 08:55:16.067935 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2491fadf-97f6-4e7e-932b-3ca607c015ab-ovsdbserver-nb\") pod \"dnsmasq-dns-69f7d94959-2sm7c\" (UID: \"2491fadf-97f6-4e7e-932b-3ca607c015ab\") " pod="openstack/dnsmasq-dns-69f7d94959-2sm7c" Oct 14 08:55:16 crc kubenswrapper[4870]: I1014 08:55:16.068129 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/658a9360-4e8e-40ca-9575-5501fb52be26-config-data-custom\") pod \"barbican-api-d46667c8-kddgf\" (UID: \"658a9360-4e8e-40ca-9575-5501fb52be26\") " pod="openstack/barbican-api-d46667c8-kddgf" Oct 14 08:55:16 crc kubenswrapper[4870]: I1014 08:55:16.068170 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2491fadf-97f6-4e7e-932b-3ca607c015ab-ovsdbserver-sb\") pod \"dnsmasq-dns-69f7d94959-2sm7c\" (UID: \"2491fadf-97f6-4e7e-932b-3ca607c015ab\") " pod="openstack/dnsmasq-dns-69f7d94959-2sm7c" Oct 14 08:55:16 crc kubenswrapper[4870]: I1014 08:55:16.068194 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-frzh7\" (UniqueName: \"kubernetes.io/projected/658a9360-4e8e-40ca-9575-5501fb52be26-kube-api-access-frzh7\") pod \"barbican-api-d46667c8-kddgf\" (UID: \"658a9360-4e8e-40ca-9575-5501fb52be26\") " pod="openstack/barbican-api-d46667c8-kddgf" Oct 14 08:55:16 crc kubenswrapper[4870]: I1014 08:55:16.068213 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/658a9360-4e8e-40ca-9575-5501fb52be26-config-data\") pod \"barbican-api-d46667c8-kddgf\" (UID: \"658a9360-4e8e-40ca-9575-5501fb52be26\") " pod="openstack/barbican-api-d46667c8-kddgf" Oct 14 08:55:16 crc kubenswrapper[4870]: I1014 08:55:16.068276 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lblql\" (UniqueName: \"kubernetes.io/projected/2491fadf-97f6-4e7e-932b-3ca607c015ab-kube-api-access-lblql\") pod \"dnsmasq-dns-69f7d94959-2sm7c\" (UID: \"2491fadf-97f6-4e7e-932b-3ca607c015ab\") " pod="openstack/dnsmasq-dns-69f7d94959-2sm7c" Oct 14 08:55:16 crc kubenswrapper[4870]: I1014 08:55:16.068308 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/658a9360-4e8e-40ca-9575-5501fb52be26-combined-ca-bundle\") pod \"barbican-api-d46667c8-kddgf\" (UID: \"658a9360-4e8e-40ca-9575-5501fb52be26\") " pod="openstack/barbican-api-d46667c8-kddgf" Oct 14 08:55:16 crc kubenswrapper[4870]: I1014 08:55:16.070450 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2491fadf-97f6-4e7e-932b-3ca607c015ab-config\") pod \"dnsmasq-dns-69f7d94959-2sm7c\" (UID: \"2491fadf-97f6-4e7e-932b-3ca607c015ab\") " pod="openstack/dnsmasq-dns-69f7d94959-2sm7c" Oct 14 08:55:16 crc kubenswrapper[4870]: I1014 08:55:16.071031 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2491fadf-97f6-4e7e-932b-3ca607c015ab-dns-svc\") pod \"dnsmasq-dns-69f7d94959-2sm7c\" (UID: \"2491fadf-97f6-4e7e-932b-3ca607c015ab\") " pod="openstack/dnsmasq-dns-69f7d94959-2sm7c" Oct 14 08:55:16 crc kubenswrapper[4870]: I1014 08:55:16.071265 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/658a9360-4e8e-40ca-9575-5501fb52be26-logs\") pod \"barbican-api-d46667c8-kddgf\" (UID: \"658a9360-4e8e-40ca-9575-5501fb52be26\") " pod="openstack/barbican-api-d46667c8-kddgf" Oct 14 08:55:16 crc kubenswrapper[4870]: I1014 08:55:16.071775 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2491fadf-97f6-4e7e-932b-3ca607c015ab-ovsdbserver-nb\") pod \"dnsmasq-dns-69f7d94959-2sm7c\" (UID: \"2491fadf-97f6-4e7e-932b-3ca607c015ab\") " pod="openstack/dnsmasq-dns-69f7d94959-2sm7c" Oct 14 08:55:16 crc kubenswrapper[4870]: I1014 08:55:16.077579 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/658a9360-4e8e-40ca-9575-5501fb52be26-config-data\") pod \"barbican-api-d46667c8-kddgf\" (UID: \"658a9360-4e8e-40ca-9575-5501fb52be26\") " pod="openstack/barbican-api-d46667c8-kddgf" Oct 14 08:55:16 crc kubenswrapper[4870]: I1014 08:55:16.078155 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2491fadf-97f6-4e7e-932b-3ca607c015ab-ovsdbserver-sb\") pod \"dnsmasq-dns-69f7d94959-2sm7c\" (UID: \"2491fadf-97f6-4e7e-932b-3ca607c015ab\") " pod="openstack/dnsmasq-dns-69f7d94959-2sm7c" Oct 14 08:55:16 crc kubenswrapper[4870]: I1014 08:55:16.078651 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/658a9360-4e8e-40ca-9575-5501fb52be26-combined-ca-bundle\") pod \"barbican-api-d46667c8-kddgf\" (UID: \"658a9360-4e8e-40ca-9575-5501fb52be26\") " pod="openstack/barbican-api-d46667c8-kddgf" Oct 14 08:55:16 crc kubenswrapper[4870]: I1014 08:55:16.078769 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/658a9360-4e8e-40ca-9575-5501fb52be26-config-data-custom\") pod \"barbican-api-d46667c8-kddgf\" (UID: \"658a9360-4e8e-40ca-9575-5501fb52be26\") " pod="openstack/barbican-api-d46667c8-kddgf" Oct 14 08:55:16 crc kubenswrapper[4870]: I1014 08:55:16.094348 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-frzh7\" (UniqueName: \"kubernetes.io/projected/658a9360-4e8e-40ca-9575-5501fb52be26-kube-api-access-frzh7\") pod \"barbican-api-d46667c8-kddgf\" (UID: \"658a9360-4e8e-40ca-9575-5501fb52be26\") " pod="openstack/barbican-api-d46667c8-kddgf" Oct 14 08:55:16 crc kubenswrapper[4870]: I1014 08:55:16.110002 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lblql\" (UniqueName: \"kubernetes.io/projected/2491fadf-97f6-4e7e-932b-3ca607c015ab-kube-api-access-lblql\") pod \"dnsmasq-dns-69f7d94959-2sm7c\" (UID: \"2491fadf-97f6-4e7e-932b-3ca607c015ab\") " pod="openstack/dnsmasq-dns-69f7d94959-2sm7c" Oct 14 08:55:16 crc kubenswrapper[4870]: I1014 08:55:16.131554 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-69f7d94959-2sm7c" Oct 14 08:55:16 crc kubenswrapper[4870]: I1014 08:55:16.263184 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-d46667c8-kddgf" Oct 14 08:55:16 crc kubenswrapper[4870]: I1014 08:55:16.416503 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-7c9b74fd49-gbcpm"] Oct 14 08:55:16 crc kubenswrapper[4870]: I1014 08:55:16.444846 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-7c9b74fd49-gbcpm" event={"ID":"4446db55-22b0-4800-9ebc-f4b6b6578679","Type":"ContainerStarted","Data":"3b5b3e5599b6d64168420e3f18401839bc75b0005c77401b55a09b62773ed34d"} Oct 14 08:55:16 crc kubenswrapper[4870]: I1014 08:55:16.528491 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-68859d4c94-fx6jt"] Oct 14 08:55:16 crc kubenswrapper[4870]: I1014 08:55:16.625310 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-69f7d94959-2sm7c"] Oct 14 08:55:16 crc kubenswrapper[4870]: W1014 08:55:16.633126 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2491fadf_97f6_4e7e_932b_3ca607c015ab.slice/crio-79987c8cdafa09d6fdd69831214d775d2707d58ad4de6306ca9c3d02e93f1cd2 WatchSource:0}: Error finding container 79987c8cdafa09d6fdd69831214d775d2707d58ad4de6306ca9c3d02e93f1cd2: Status 404 returned error can't find the container with id 79987c8cdafa09d6fdd69831214d775d2707d58ad4de6306ca9c3d02e93f1cd2 Oct 14 08:55:16 crc kubenswrapper[4870]: I1014 08:55:16.798294 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-d46667c8-kddgf"] Oct 14 08:55:17 crc kubenswrapper[4870]: I1014 08:55:17.454167 4870 generic.go:334] "Generic (PLEG): container finished" podID="2491fadf-97f6-4e7e-932b-3ca607c015ab" containerID="f54d4714558468e367cb7334b879aa0eafd61a63e0c81c28a522ef4839f64d39" exitCode=0 Oct 14 08:55:17 crc kubenswrapper[4870]: I1014 08:55:17.454339 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-69f7d94959-2sm7c" event={"ID":"2491fadf-97f6-4e7e-932b-3ca607c015ab","Type":"ContainerDied","Data":"f54d4714558468e367cb7334b879aa0eafd61a63e0c81c28a522ef4839f64d39"} Oct 14 08:55:17 crc kubenswrapper[4870]: I1014 08:55:17.454618 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-69f7d94959-2sm7c" event={"ID":"2491fadf-97f6-4e7e-932b-3ca607c015ab","Type":"ContainerStarted","Data":"79987c8cdafa09d6fdd69831214d775d2707d58ad4de6306ca9c3d02e93f1cd2"} Oct 14 08:55:17 crc kubenswrapper[4870]: I1014 08:55:17.456949 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-68859d4c94-fx6jt" event={"ID":"cc348a20-302d-407f-be5c-b690c39f12cf","Type":"ContainerStarted","Data":"3735e690b11fe11faf097da03a03f1c091290d3f8b193ef8a4e914ccf0ce84fd"} Oct 14 08:55:17 crc kubenswrapper[4870]: I1014 08:55:17.460886 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-d46667c8-kddgf" event={"ID":"658a9360-4e8e-40ca-9575-5501fb52be26","Type":"ContainerStarted","Data":"19d7ef1f441103c9afce6e4d4c2b3db1595e3831eeae84f02b19634e14b7eacf"} Oct 14 08:55:17 crc kubenswrapper[4870]: I1014 08:55:17.460916 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-d46667c8-kddgf" event={"ID":"658a9360-4e8e-40ca-9575-5501fb52be26","Type":"ContainerStarted","Data":"4a4a0d9dfedbaa26d0c084ae7dc75ea2aa19e03ebcf7721713a81a38c75355a3"} Oct 14 08:55:18 crc kubenswrapper[4870]: I1014 08:55:18.471951 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-69f7d94959-2sm7c" event={"ID":"2491fadf-97f6-4e7e-932b-3ca607c015ab","Type":"ContainerStarted","Data":"f9ef13564e588ee8e9f11fa2197497d04269bcfbcca40379a8d841340a5750b6"} Oct 14 08:55:18 crc kubenswrapper[4870]: I1014 08:55:18.472313 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-69f7d94959-2sm7c" Oct 14 08:55:18 crc kubenswrapper[4870]: I1014 08:55:18.475755 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-68859d4c94-fx6jt" event={"ID":"cc348a20-302d-407f-be5c-b690c39f12cf","Type":"ContainerStarted","Data":"80af7bd81951443fee62b1cfcef25e2daace5c14a3c4b092a985181c1db3b4b8"} Oct 14 08:55:18 crc kubenswrapper[4870]: I1014 08:55:18.475784 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-68859d4c94-fx6jt" event={"ID":"cc348a20-302d-407f-be5c-b690c39f12cf","Type":"ContainerStarted","Data":"f9bc7325250d538a0ce37342466d5f1c1705fb512c8d724d5175ae77d40758b6"} Oct 14 08:55:18 crc kubenswrapper[4870]: I1014 08:55:18.478832 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-7c9b74fd49-gbcpm" event={"ID":"4446db55-22b0-4800-9ebc-f4b6b6578679","Type":"ContainerStarted","Data":"e9b24eb43c71e76d112ce988acb9910c209e48fde74de15467cf99bd1996760f"} Oct 14 08:55:18 crc kubenswrapper[4870]: I1014 08:55:18.478892 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-7c9b74fd49-gbcpm" event={"ID":"4446db55-22b0-4800-9ebc-f4b6b6578679","Type":"ContainerStarted","Data":"a9cefe4ae44015b10e0d5dc0a6a1c8d41542450b6c9d02745ca90d6e963daf3b"} Oct 14 08:55:18 crc kubenswrapper[4870]: I1014 08:55:18.481136 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-d46667c8-kddgf" event={"ID":"658a9360-4e8e-40ca-9575-5501fb52be26","Type":"ContainerStarted","Data":"ade1a70e1481a848d55823ff6ee0e2adf30d2fb2d33e7e88d53775e6516a770e"} Oct 14 08:55:18 crc kubenswrapper[4870]: I1014 08:55:18.481329 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-d46667c8-kddgf" Oct 14 08:55:18 crc kubenswrapper[4870]: I1014 08:55:18.481364 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-d46667c8-kddgf" Oct 14 08:55:18 crc kubenswrapper[4870]: I1014 08:55:18.516870 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-69f7d94959-2sm7c" podStartSLOduration=3.516843365 podStartE2EDuration="3.516843365s" podCreationTimestamp="2025-10-14 08:55:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 08:55:18.504169005 +0000 UTC m=+6854.201529386" watchObservedRunningTime="2025-10-14 08:55:18.516843365 +0000 UTC m=+6854.214203776" Oct 14 08:55:18 crc kubenswrapper[4870]: I1014 08:55:18.534102 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-d46667c8-kddgf" podStartSLOduration=3.534073058 podStartE2EDuration="3.534073058s" podCreationTimestamp="2025-10-14 08:55:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 08:55:18.52394869 +0000 UTC m=+6854.221309071" watchObservedRunningTime="2025-10-14 08:55:18.534073058 +0000 UTC m=+6854.231433449" Oct 14 08:55:18 crc kubenswrapper[4870]: I1014 08:55:18.569602 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-7c9b74fd49-gbcpm" podStartSLOduration=2.38677801 podStartE2EDuration="3.569575777s" podCreationTimestamp="2025-10-14 08:55:15 +0000 UTC" firstStartedPulling="2025-10-14 08:55:16.41960435 +0000 UTC m=+6852.116964721" lastFinishedPulling="2025-10-14 08:55:17.602402117 +0000 UTC m=+6853.299762488" observedRunningTime="2025-10-14 08:55:18.558600228 +0000 UTC m=+6854.255960609" watchObservedRunningTime="2025-10-14 08:55:18.569575777 +0000 UTC m=+6854.266936158" Oct 14 08:55:18 crc kubenswrapper[4870]: I1014 08:55:18.580630 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-68859d4c94-fx6jt" podStartSLOduration=2.511388245 podStartE2EDuration="3.580604507s" podCreationTimestamp="2025-10-14 08:55:15 +0000 UTC" firstStartedPulling="2025-10-14 08:55:16.533671397 +0000 UTC m=+6852.231031768" lastFinishedPulling="2025-10-14 08:55:17.602887649 +0000 UTC m=+6853.300248030" observedRunningTime="2025-10-14 08:55:18.57538416 +0000 UTC m=+6854.272744551" watchObservedRunningTime="2025-10-14 08:55:18.580604507 +0000 UTC m=+6854.277964888" Oct 14 08:55:23 crc kubenswrapper[4870]: I1014 08:55:23.713422 4870 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-d46667c8-kddgf" podUID="658a9360-4e8e-40ca-9575-5501fb52be26" containerName="barbican-api-log" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 14 08:55:24 crc kubenswrapper[4870]: I1014 08:55:24.039763 4870 scope.go:117] "RemoveContainer" containerID="8c0f754fe94b7cb7b9bb5dcd613f7e42093968b49509d8ee37e8fd4db89eedd5" Oct 14 08:55:24 crc kubenswrapper[4870]: E1014 08:55:24.040397 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 08:55:26 crc kubenswrapper[4870]: I1014 08:55:26.133628 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-69f7d94959-2sm7c" Oct 14 08:55:26 crc kubenswrapper[4870]: I1014 08:55:26.194459 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5b8d755fbc-ps7fd"] Oct 14 08:55:26 crc kubenswrapper[4870]: I1014 08:55:26.194715 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5b8d755fbc-ps7fd" podUID="eb9a9a5e-a99d-4fae-9b28-006252710ab1" containerName="dnsmasq-dns" containerID="cri-o://9063c16715734d3fc47a99adf4dfbc8c8f48caa9b3a33c4c1e9a934a7bb6e507" gracePeriod=10 Oct 14 08:55:26 crc kubenswrapper[4870]: I1014 08:55:26.557734 4870 generic.go:334] "Generic (PLEG): container finished" podID="eb9a9a5e-a99d-4fae-9b28-006252710ab1" containerID="9063c16715734d3fc47a99adf4dfbc8c8f48caa9b3a33c4c1e9a934a7bb6e507" exitCode=0 Oct 14 08:55:26 crc kubenswrapper[4870]: I1014 08:55:26.557832 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b8d755fbc-ps7fd" event={"ID":"eb9a9a5e-a99d-4fae-9b28-006252710ab1","Type":"ContainerDied","Data":"9063c16715734d3fc47a99adf4dfbc8c8f48caa9b3a33c4c1e9a934a7bb6e507"} Oct 14 08:55:26 crc kubenswrapper[4870]: I1014 08:55:26.670060 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b8d755fbc-ps7fd" Oct 14 08:55:26 crc kubenswrapper[4870]: I1014 08:55:26.789837 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/eb9a9a5e-a99d-4fae-9b28-006252710ab1-dns-svc\") pod \"eb9a9a5e-a99d-4fae-9b28-006252710ab1\" (UID: \"eb9a9a5e-a99d-4fae-9b28-006252710ab1\") " Oct 14 08:55:26 crc kubenswrapper[4870]: I1014 08:55:26.790239 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pvxvm\" (UniqueName: \"kubernetes.io/projected/eb9a9a5e-a99d-4fae-9b28-006252710ab1-kube-api-access-pvxvm\") pod \"eb9a9a5e-a99d-4fae-9b28-006252710ab1\" (UID: \"eb9a9a5e-a99d-4fae-9b28-006252710ab1\") " Oct 14 08:55:26 crc kubenswrapper[4870]: I1014 08:55:26.790270 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eb9a9a5e-a99d-4fae-9b28-006252710ab1-config\") pod \"eb9a9a5e-a99d-4fae-9b28-006252710ab1\" (UID: \"eb9a9a5e-a99d-4fae-9b28-006252710ab1\") " Oct 14 08:55:26 crc kubenswrapper[4870]: I1014 08:55:26.790412 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/eb9a9a5e-a99d-4fae-9b28-006252710ab1-ovsdbserver-nb\") pod \"eb9a9a5e-a99d-4fae-9b28-006252710ab1\" (UID: \"eb9a9a5e-a99d-4fae-9b28-006252710ab1\") " Oct 14 08:55:26 crc kubenswrapper[4870]: I1014 08:55:26.790461 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/eb9a9a5e-a99d-4fae-9b28-006252710ab1-ovsdbserver-sb\") pod \"eb9a9a5e-a99d-4fae-9b28-006252710ab1\" (UID: \"eb9a9a5e-a99d-4fae-9b28-006252710ab1\") " Oct 14 08:55:26 crc kubenswrapper[4870]: I1014 08:55:26.799334 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eb9a9a5e-a99d-4fae-9b28-006252710ab1-kube-api-access-pvxvm" (OuterVolumeSpecName: "kube-api-access-pvxvm") pod "eb9a9a5e-a99d-4fae-9b28-006252710ab1" (UID: "eb9a9a5e-a99d-4fae-9b28-006252710ab1"). InnerVolumeSpecName "kube-api-access-pvxvm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:55:26 crc kubenswrapper[4870]: I1014 08:55:26.834282 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eb9a9a5e-a99d-4fae-9b28-006252710ab1-config" (OuterVolumeSpecName: "config") pod "eb9a9a5e-a99d-4fae-9b28-006252710ab1" (UID: "eb9a9a5e-a99d-4fae-9b28-006252710ab1"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:55:26 crc kubenswrapper[4870]: I1014 08:55:26.838803 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eb9a9a5e-a99d-4fae-9b28-006252710ab1-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "eb9a9a5e-a99d-4fae-9b28-006252710ab1" (UID: "eb9a9a5e-a99d-4fae-9b28-006252710ab1"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:55:26 crc kubenswrapper[4870]: I1014 08:55:26.839182 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eb9a9a5e-a99d-4fae-9b28-006252710ab1-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "eb9a9a5e-a99d-4fae-9b28-006252710ab1" (UID: "eb9a9a5e-a99d-4fae-9b28-006252710ab1"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:55:26 crc kubenswrapper[4870]: I1014 08:55:26.845974 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eb9a9a5e-a99d-4fae-9b28-006252710ab1-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "eb9a9a5e-a99d-4fae-9b28-006252710ab1" (UID: "eb9a9a5e-a99d-4fae-9b28-006252710ab1"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:55:26 crc kubenswrapper[4870]: I1014 08:55:26.892388 4870 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/eb9a9a5e-a99d-4fae-9b28-006252710ab1-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 14 08:55:26 crc kubenswrapper[4870]: I1014 08:55:26.892425 4870 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/eb9a9a5e-a99d-4fae-9b28-006252710ab1-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 14 08:55:26 crc kubenswrapper[4870]: I1014 08:55:26.892464 4870 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/eb9a9a5e-a99d-4fae-9b28-006252710ab1-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 14 08:55:26 crc kubenswrapper[4870]: I1014 08:55:26.892478 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pvxvm\" (UniqueName: \"kubernetes.io/projected/eb9a9a5e-a99d-4fae-9b28-006252710ab1-kube-api-access-pvxvm\") on node \"crc\" DevicePath \"\"" Oct 14 08:55:26 crc kubenswrapper[4870]: I1014 08:55:26.892490 4870 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eb9a9a5e-a99d-4fae-9b28-006252710ab1-config\") on node \"crc\" DevicePath \"\"" Oct 14 08:55:27 crc kubenswrapper[4870]: I1014 08:55:27.591760 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b8d755fbc-ps7fd" event={"ID":"eb9a9a5e-a99d-4fae-9b28-006252710ab1","Type":"ContainerDied","Data":"4d969c51a004112f47a7c2a7594a95d29c9eff5ab482639f4f6c0be2c52d6db9"} Oct 14 08:55:27 crc kubenswrapper[4870]: I1014 08:55:27.591821 4870 scope.go:117] "RemoveContainer" containerID="9063c16715734d3fc47a99adf4dfbc8c8f48caa9b3a33c4c1e9a934a7bb6e507" Oct 14 08:55:27 crc kubenswrapper[4870]: I1014 08:55:27.593405 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b8d755fbc-ps7fd" Oct 14 08:55:27 crc kubenswrapper[4870]: I1014 08:55:27.636536 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5b8d755fbc-ps7fd"] Oct 14 08:55:27 crc kubenswrapper[4870]: I1014 08:55:27.637672 4870 scope.go:117] "RemoveContainer" containerID="fd0d21dfd2262dbc3250c7c479f0ad007aab08376da9399b80c824ab62724be5" Oct 14 08:55:27 crc kubenswrapper[4870]: I1014 08:55:27.651713 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5b8d755fbc-ps7fd"] Oct 14 08:55:27 crc kubenswrapper[4870]: I1014 08:55:27.773718 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-d46667c8-kddgf" Oct 14 08:55:27 crc kubenswrapper[4870]: I1014 08:55:27.807928 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-d46667c8-kddgf" Oct 14 08:55:29 crc kubenswrapper[4870]: I1014 08:55:29.045707 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eb9a9a5e-a99d-4fae-9b28-006252710ab1" path="/var/lib/kubelet/pods/eb9a9a5e-a99d-4fae-9b28-006252710ab1/volumes" Oct 14 08:55:35 crc kubenswrapper[4870]: I1014 08:55:35.576373 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-l2b5s"] Oct 14 08:55:35 crc kubenswrapper[4870]: E1014 08:55:35.577187 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb9a9a5e-a99d-4fae-9b28-006252710ab1" containerName="dnsmasq-dns" Oct 14 08:55:35 crc kubenswrapper[4870]: I1014 08:55:35.577200 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb9a9a5e-a99d-4fae-9b28-006252710ab1" containerName="dnsmasq-dns" Oct 14 08:55:35 crc kubenswrapper[4870]: E1014 08:55:35.577220 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb9a9a5e-a99d-4fae-9b28-006252710ab1" containerName="init" Oct 14 08:55:35 crc kubenswrapper[4870]: I1014 08:55:35.577227 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb9a9a5e-a99d-4fae-9b28-006252710ab1" containerName="init" Oct 14 08:55:35 crc kubenswrapper[4870]: I1014 08:55:35.577396 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="eb9a9a5e-a99d-4fae-9b28-006252710ab1" containerName="dnsmasq-dns" Oct 14 08:55:35 crc kubenswrapper[4870]: I1014 08:55:35.577981 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-l2b5s" Oct 14 08:55:35 crc kubenswrapper[4870]: I1014 08:55:35.598730 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-l2b5s"] Oct 14 08:55:35 crc kubenswrapper[4870]: I1014 08:55:35.676988 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b5c2g\" (UniqueName: \"kubernetes.io/projected/76690419-68ac-42b3-a6a0-3fb21543786a-kube-api-access-b5c2g\") pod \"neutron-db-create-l2b5s\" (UID: \"76690419-68ac-42b3-a6a0-3fb21543786a\") " pod="openstack/neutron-db-create-l2b5s" Oct 14 08:55:35 crc kubenswrapper[4870]: I1014 08:55:35.778851 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b5c2g\" (UniqueName: \"kubernetes.io/projected/76690419-68ac-42b3-a6a0-3fb21543786a-kube-api-access-b5c2g\") pod \"neutron-db-create-l2b5s\" (UID: \"76690419-68ac-42b3-a6a0-3fb21543786a\") " pod="openstack/neutron-db-create-l2b5s" Oct 14 08:55:35 crc kubenswrapper[4870]: I1014 08:55:35.803756 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b5c2g\" (UniqueName: \"kubernetes.io/projected/76690419-68ac-42b3-a6a0-3fb21543786a-kube-api-access-b5c2g\") pod \"neutron-db-create-l2b5s\" (UID: \"76690419-68ac-42b3-a6a0-3fb21543786a\") " pod="openstack/neutron-db-create-l2b5s" Oct 14 08:55:35 crc kubenswrapper[4870]: I1014 08:55:35.905120 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-l2b5s" Oct 14 08:55:36 crc kubenswrapper[4870]: W1014 08:55:36.459501 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod76690419_68ac_42b3_a6a0_3fb21543786a.slice/crio-ffc66e198dabb2be78af580ed0b3c335489d1a43a7b94781938277a099823191 WatchSource:0}: Error finding container ffc66e198dabb2be78af580ed0b3c335489d1a43a7b94781938277a099823191: Status 404 returned error can't find the container with id ffc66e198dabb2be78af580ed0b3c335489d1a43a7b94781938277a099823191 Oct 14 08:55:36 crc kubenswrapper[4870]: I1014 08:55:36.463628 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-l2b5s"] Oct 14 08:55:36 crc kubenswrapper[4870]: I1014 08:55:36.695956 4870 generic.go:334] "Generic (PLEG): container finished" podID="76690419-68ac-42b3-a6a0-3fb21543786a" containerID="580060d7584a6e8e9e9af94c1f92c8c6ad81baaab6f9d16525ef2919c61c2b31" exitCode=0 Oct 14 08:55:36 crc kubenswrapper[4870]: I1014 08:55:36.696227 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-l2b5s" event={"ID":"76690419-68ac-42b3-a6a0-3fb21543786a","Type":"ContainerDied","Data":"580060d7584a6e8e9e9af94c1f92c8c6ad81baaab6f9d16525ef2919c61c2b31"} Oct 14 08:55:36 crc kubenswrapper[4870]: I1014 08:55:36.696256 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-l2b5s" event={"ID":"76690419-68ac-42b3-a6a0-3fb21543786a","Type":"ContainerStarted","Data":"ffc66e198dabb2be78af580ed0b3c335489d1a43a7b94781938277a099823191"} Oct 14 08:55:37 crc kubenswrapper[4870]: I1014 08:55:37.033985 4870 scope.go:117] "RemoveContainer" containerID="8c0f754fe94b7cb7b9bb5dcd613f7e42093968b49509d8ee37e8fd4db89eedd5" Oct 14 08:55:37 crc kubenswrapper[4870]: E1014 08:55:37.034213 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 08:55:38 crc kubenswrapper[4870]: I1014 08:55:38.032084 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-l2b5s" Oct 14 08:55:38 crc kubenswrapper[4870]: I1014 08:55:38.119197 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b5c2g\" (UniqueName: \"kubernetes.io/projected/76690419-68ac-42b3-a6a0-3fb21543786a-kube-api-access-b5c2g\") pod \"76690419-68ac-42b3-a6a0-3fb21543786a\" (UID: \"76690419-68ac-42b3-a6a0-3fb21543786a\") " Oct 14 08:55:38 crc kubenswrapper[4870]: I1014 08:55:38.125548 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/76690419-68ac-42b3-a6a0-3fb21543786a-kube-api-access-b5c2g" (OuterVolumeSpecName: "kube-api-access-b5c2g") pod "76690419-68ac-42b3-a6a0-3fb21543786a" (UID: "76690419-68ac-42b3-a6a0-3fb21543786a"). InnerVolumeSpecName "kube-api-access-b5c2g". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:55:38 crc kubenswrapper[4870]: I1014 08:55:38.221449 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b5c2g\" (UniqueName: \"kubernetes.io/projected/76690419-68ac-42b3-a6a0-3fb21543786a-kube-api-access-b5c2g\") on node \"crc\" DevicePath \"\"" Oct 14 08:55:38 crc kubenswrapper[4870]: I1014 08:55:38.723546 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-l2b5s" Oct 14 08:55:38 crc kubenswrapper[4870]: I1014 08:55:38.723534 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-l2b5s" event={"ID":"76690419-68ac-42b3-a6a0-3fb21543786a","Type":"ContainerDied","Data":"ffc66e198dabb2be78af580ed0b3c335489d1a43a7b94781938277a099823191"} Oct 14 08:55:38 crc kubenswrapper[4870]: I1014 08:55:38.723832 4870 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ffc66e198dabb2be78af580ed0b3c335489d1a43a7b94781938277a099823191" Oct 14 08:55:45 crc kubenswrapper[4870]: I1014 08:55:45.624640 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-750f-account-create-wg5jr"] Oct 14 08:55:45 crc kubenswrapper[4870]: E1014 08:55:45.625813 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76690419-68ac-42b3-a6a0-3fb21543786a" containerName="mariadb-database-create" Oct 14 08:55:45 crc kubenswrapper[4870]: I1014 08:55:45.625840 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="76690419-68ac-42b3-a6a0-3fb21543786a" containerName="mariadb-database-create" Oct 14 08:55:45 crc kubenswrapper[4870]: I1014 08:55:45.626176 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="76690419-68ac-42b3-a6a0-3fb21543786a" containerName="mariadb-database-create" Oct 14 08:55:45 crc kubenswrapper[4870]: I1014 08:55:45.626935 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-750f-account-create-wg5jr" Oct 14 08:55:45 crc kubenswrapper[4870]: I1014 08:55:45.629484 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Oct 14 08:55:45 crc kubenswrapper[4870]: I1014 08:55:45.647526 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-750f-account-create-wg5jr"] Oct 14 08:55:45 crc kubenswrapper[4870]: I1014 08:55:45.667982 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wkg6f\" (UniqueName: \"kubernetes.io/projected/5fd63d29-5666-442a-82b5-1e054c10b29c-kube-api-access-wkg6f\") pod \"neutron-750f-account-create-wg5jr\" (UID: \"5fd63d29-5666-442a-82b5-1e054c10b29c\") " pod="openstack/neutron-750f-account-create-wg5jr" Oct 14 08:55:45 crc kubenswrapper[4870]: I1014 08:55:45.770594 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wkg6f\" (UniqueName: \"kubernetes.io/projected/5fd63d29-5666-442a-82b5-1e054c10b29c-kube-api-access-wkg6f\") pod \"neutron-750f-account-create-wg5jr\" (UID: \"5fd63d29-5666-442a-82b5-1e054c10b29c\") " pod="openstack/neutron-750f-account-create-wg5jr" Oct 14 08:55:45 crc kubenswrapper[4870]: I1014 08:55:45.802569 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wkg6f\" (UniqueName: \"kubernetes.io/projected/5fd63d29-5666-442a-82b5-1e054c10b29c-kube-api-access-wkg6f\") pod \"neutron-750f-account-create-wg5jr\" (UID: \"5fd63d29-5666-442a-82b5-1e054c10b29c\") " pod="openstack/neutron-750f-account-create-wg5jr" Oct 14 08:55:45 crc kubenswrapper[4870]: I1014 08:55:45.949975 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-750f-account-create-wg5jr" Oct 14 08:55:46 crc kubenswrapper[4870]: I1014 08:55:46.469661 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-750f-account-create-wg5jr"] Oct 14 08:55:46 crc kubenswrapper[4870]: I1014 08:55:46.801885 4870 generic.go:334] "Generic (PLEG): container finished" podID="5fd63d29-5666-442a-82b5-1e054c10b29c" containerID="ed6f882b5aaf3efccc07a4239089b7fe033432102b1bbc861f8b1ab3577a69c4" exitCode=0 Oct 14 08:55:46 crc kubenswrapper[4870]: I1014 08:55:46.802182 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-750f-account-create-wg5jr" event={"ID":"5fd63d29-5666-442a-82b5-1e054c10b29c","Type":"ContainerDied","Data":"ed6f882b5aaf3efccc07a4239089b7fe033432102b1bbc861f8b1ab3577a69c4"} Oct 14 08:55:46 crc kubenswrapper[4870]: I1014 08:55:46.802434 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-750f-account-create-wg5jr" event={"ID":"5fd63d29-5666-442a-82b5-1e054c10b29c","Type":"ContainerStarted","Data":"b5b51196f05f8468692f12bb3fa3f4f41b878872807ead81d4e19f3a35cfcdbb"} Oct 14 08:55:48 crc kubenswrapper[4870]: I1014 08:55:48.220956 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-750f-account-create-wg5jr" Oct 14 08:55:48 crc kubenswrapper[4870]: I1014 08:55:48.318789 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wkg6f\" (UniqueName: \"kubernetes.io/projected/5fd63d29-5666-442a-82b5-1e054c10b29c-kube-api-access-wkg6f\") pod \"5fd63d29-5666-442a-82b5-1e054c10b29c\" (UID: \"5fd63d29-5666-442a-82b5-1e054c10b29c\") " Oct 14 08:55:48 crc kubenswrapper[4870]: I1014 08:55:48.324153 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fd63d29-5666-442a-82b5-1e054c10b29c-kube-api-access-wkg6f" (OuterVolumeSpecName: "kube-api-access-wkg6f") pod "5fd63d29-5666-442a-82b5-1e054c10b29c" (UID: "5fd63d29-5666-442a-82b5-1e054c10b29c"). InnerVolumeSpecName "kube-api-access-wkg6f". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:55:48 crc kubenswrapper[4870]: I1014 08:55:48.421106 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wkg6f\" (UniqueName: \"kubernetes.io/projected/5fd63d29-5666-442a-82b5-1e054c10b29c-kube-api-access-wkg6f\") on node \"crc\" DevicePath \"\"" Oct 14 08:55:48 crc kubenswrapper[4870]: I1014 08:55:48.820025 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-750f-account-create-wg5jr" event={"ID":"5fd63d29-5666-442a-82b5-1e054c10b29c","Type":"ContainerDied","Data":"b5b51196f05f8468692f12bb3fa3f4f41b878872807ead81d4e19f3a35cfcdbb"} Oct 14 08:55:48 crc kubenswrapper[4870]: I1014 08:55:48.820066 4870 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b5b51196f05f8468692f12bb3fa3f4f41b878872807ead81d4e19f3a35cfcdbb" Oct 14 08:55:48 crc kubenswrapper[4870]: I1014 08:55:48.820070 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-750f-account-create-wg5jr" Oct 14 08:55:50 crc kubenswrapper[4870]: I1014 08:55:50.035994 4870 scope.go:117] "RemoveContainer" containerID="8c0f754fe94b7cb7b9bb5dcd613f7e42093968b49509d8ee37e8fd4db89eedd5" Oct 14 08:55:50 crc kubenswrapper[4870]: E1014 08:55:50.036685 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 08:55:50 crc kubenswrapper[4870]: I1014 08:55:50.783236 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-qfmh6"] Oct 14 08:55:50 crc kubenswrapper[4870]: E1014 08:55:50.783865 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5fd63d29-5666-442a-82b5-1e054c10b29c" containerName="mariadb-account-create" Oct 14 08:55:50 crc kubenswrapper[4870]: I1014 08:55:50.783891 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="5fd63d29-5666-442a-82b5-1e054c10b29c" containerName="mariadb-account-create" Oct 14 08:55:50 crc kubenswrapper[4870]: I1014 08:55:50.784198 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="5fd63d29-5666-442a-82b5-1e054c10b29c" containerName="mariadb-account-create" Oct 14 08:55:50 crc kubenswrapper[4870]: I1014 08:55:50.785111 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-qfmh6" Oct 14 08:55:50 crc kubenswrapper[4870]: I1014 08:55:50.791890 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-qfmh6"] Oct 14 08:55:50 crc kubenswrapper[4870]: I1014 08:55:50.834681 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Oct 14 08:55:50 crc kubenswrapper[4870]: I1014 08:55:50.834946 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Oct 14 08:55:50 crc kubenswrapper[4870]: I1014 08:55:50.839211 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-p2ckg" Oct 14 08:55:50 crc kubenswrapper[4870]: I1014 08:55:50.863487 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d51a0c5-822f-47d5-af81-b812ac44fe67-combined-ca-bundle\") pod \"neutron-db-sync-qfmh6\" (UID: \"8d51a0c5-822f-47d5-af81-b812ac44fe67\") " pod="openstack/neutron-db-sync-qfmh6" Oct 14 08:55:50 crc kubenswrapper[4870]: I1014 08:55:50.863602 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/8d51a0c5-822f-47d5-af81-b812ac44fe67-config\") pod \"neutron-db-sync-qfmh6\" (UID: \"8d51a0c5-822f-47d5-af81-b812ac44fe67\") " pod="openstack/neutron-db-sync-qfmh6" Oct 14 08:55:50 crc kubenswrapper[4870]: I1014 08:55:50.863660 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w9lxp\" (UniqueName: \"kubernetes.io/projected/8d51a0c5-822f-47d5-af81-b812ac44fe67-kube-api-access-w9lxp\") pod \"neutron-db-sync-qfmh6\" (UID: \"8d51a0c5-822f-47d5-af81-b812ac44fe67\") " pod="openstack/neutron-db-sync-qfmh6" Oct 14 08:55:50 crc kubenswrapper[4870]: I1014 08:55:50.966233 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d51a0c5-822f-47d5-af81-b812ac44fe67-combined-ca-bundle\") pod \"neutron-db-sync-qfmh6\" (UID: \"8d51a0c5-822f-47d5-af81-b812ac44fe67\") " pod="openstack/neutron-db-sync-qfmh6" Oct 14 08:55:50 crc kubenswrapper[4870]: I1014 08:55:50.966353 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/8d51a0c5-822f-47d5-af81-b812ac44fe67-config\") pod \"neutron-db-sync-qfmh6\" (UID: \"8d51a0c5-822f-47d5-af81-b812ac44fe67\") " pod="openstack/neutron-db-sync-qfmh6" Oct 14 08:55:50 crc kubenswrapper[4870]: I1014 08:55:50.966408 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w9lxp\" (UniqueName: \"kubernetes.io/projected/8d51a0c5-822f-47d5-af81-b812ac44fe67-kube-api-access-w9lxp\") pod \"neutron-db-sync-qfmh6\" (UID: \"8d51a0c5-822f-47d5-af81-b812ac44fe67\") " pod="openstack/neutron-db-sync-qfmh6" Oct 14 08:55:50 crc kubenswrapper[4870]: I1014 08:55:50.975082 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/8d51a0c5-822f-47d5-af81-b812ac44fe67-config\") pod \"neutron-db-sync-qfmh6\" (UID: \"8d51a0c5-822f-47d5-af81-b812ac44fe67\") " pod="openstack/neutron-db-sync-qfmh6" Oct 14 08:55:50 crc kubenswrapper[4870]: I1014 08:55:50.976072 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d51a0c5-822f-47d5-af81-b812ac44fe67-combined-ca-bundle\") pod \"neutron-db-sync-qfmh6\" (UID: \"8d51a0c5-822f-47d5-af81-b812ac44fe67\") " pod="openstack/neutron-db-sync-qfmh6" Oct 14 08:55:50 crc kubenswrapper[4870]: I1014 08:55:50.997491 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w9lxp\" (UniqueName: \"kubernetes.io/projected/8d51a0c5-822f-47d5-af81-b812ac44fe67-kube-api-access-w9lxp\") pod \"neutron-db-sync-qfmh6\" (UID: \"8d51a0c5-822f-47d5-af81-b812ac44fe67\") " pod="openstack/neutron-db-sync-qfmh6" Oct 14 08:55:51 crc kubenswrapper[4870]: I1014 08:55:51.146989 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-qfmh6" Oct 14 08:55:51 crc kubenswrapper[4870]: I1014 08:55:51.610428 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-qfmh6"] Oct 14 08:55:51 crc kubenswrapper[4870]: I1014 08:55:51.858230 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-qfmh6" event={"ID":"8d51a0c5-822f-47d5-af81-b812ac44fe67","Type":"ContainerStarted","Data":"b8baaeb44c9d5a477c37b5e781b3be584d0acddcb67f90363b3985344c6c16a5"} Oct 14 08:55:51 crc kubenswrapper[4870]: I1014 08:55:51.858755 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-qfmh6" event={"ID":"8d51a0c5-822f-47d5-af81-b812ac44fe67","Type":"ContainerStarted","Data":"264848824f7329af5e476a8b8c198139f82718e803fd993a8baa608f111c594f"} Oct 14 08:55:51 crc kubenswrapper[4870]: I1014 08:55:51.875117 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-qfmh6" podStartSLOduration=1.8750921699999998 podStartE2EDuration="1.87509217s" podCreationTimestamp="2025-10-14 08:55:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 08:55:51.872668381 +0000 UTC m=+6887.570028772" watchObservedRunningTime="2025-10-14 08:55:51.87509217 +0000 UTC m=+6887.572452571" Oct 14 08:55:55 crc kubenswrapper[4870]: I1014 08:55:55.899173 4870 generic.go:334] "Generic (PLEG): container finished" podID="8d51a0c5-822f-47d5-af81-b812ac44fe67" containerID="b8baaeb44c9d5a477c37b5e781b3be584d0acddcb67f90363b3985344c6c16a5" exitCode=0 Oct 14 08:55:55 crc kubenswrapper[4870]: I1014 08:55:55.899287 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-qfmh6" event={"ID":"8d51a0c5-822f-47d5-af81-b812ac44fe67","Type":"ContainerDied","Data":"b8baaeb44c9d5a477c37b5e781b3be584d0acddcb67f90363b3985344c6c16a5"} Oct 14 08:55:57 crc kubenswrapper[4870]: I1014 08:55:57.276252 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-qfmh6" Oct 14 08:55:57 crc kubenswrapper[4870]: I1014 08:55:57.391467 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d51a0c5-822f-47d5-af81-b812ac44fe67-combined-ca-bundle\") pod \"8d51a0c5-822f-47d5-af81-b812ac44fe67\" (UID: \"8d51a0c5-822f-47d5-af81-b812ac44fe67\") " Oct 14 08:55:57 crc kubenswrapper[4870]: I1014 08:55:57.391518 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9lxp\" (UniqueName: \"kubernetes.io/projected/8d51a0c5-822f-47d5-af81-b812ac44fe67-kube-api-access-w9lxp\") pod \"8d51a0c5-822f-47d5-af81-b812ac44fe67\" (UID: \"8d51a0c5-822f-47d5-af81-b812ac44fe67\") " Oct 14 08:55:57 crc kubenswrapper[4870]: I1014 08:55:57.391562 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/8d51a0c5-822f-47d5-af81-b812ac44fe67-config\") pod \"8d51a0c5-822f-47d5-af81-b812ac44fe67\" (UID: \"8d51a0c5-822f-47d5-af81-b812ac44fe67\") " Oct 14 08:55:57 crc kubenswrapper[4870]: I1014 08:55:57.415039 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8d51a0c5-822f-47d5-af81-b812ac44fe67-kube-api-access-w9lxp" (OuterVolumeSpecName: "kube-api-access-w9lxp") pod "8d51a0c5-822f-47d5-af81-b812ac44fe67" (UID: "8d51a0c5-822f-47d5-af81-b812ac44fe67"). InnerVolumeSpecName "kube-api-access-w9lxp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:55:57 crc kubenswrapper[4870]: I1014 08:55:57.443559 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8d51a0c5-822f-47d5-af81-b812ac44fe67-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8d51a0c5-822f-47d5-af81-b812ac44fe67" (UID: "8d51a0c5-822f-47d5-af81-b812ac44fe67"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:55:57 crc kubenswrapper[4870]: I1014 08:55:57.461709 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8d51a0c5-822f-47d5-af81-b812ac44fe67-config" (OuterVolumeSpecName: "config") pod "8d51a0c5-822f-47d5-af81-b812ac44fe67" (UID: "8d51a0c5-822f-47d5-af81-b812ac44fe67"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:55:57 crc kubenswrapper[4870]: I1014 08:55:57.494069 4870 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d51a0c5-822f-47d5-af81-b812ac44fe67-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 08:55:57 crc kubenswrapper[4870]: I1014 08:55:57.494104 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9lxp\" (UniqueName: \"kubernetes.io/projected/8d51a0c5-822f-47d5-af81-b812ac44fe67-kube-api-access-w9lxp\") on node \"crc\" DevicePath \"\"" Oct 14 08:55:57 crc kubenswrapper[4870]: I1014 08:55:57.494116 4870 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/8d51a0c5-822f-47d5-af81-b812ac44fe67-config\") on node \"crc\" DevicePath \"\"" Oct 14 08:55:57 crc kubenswrapper[4870]: I1014 08:55:57.926153 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-qfmh6" event={"ID":"8d51a0c5-822f-47d5-af81-b812ac44fe67","Type":"ContainerDied","Data":"264848824f7329af5e476a8b8c198139f82718e803fd993a8baa608f111c594f"} Oct 14 08:55:57 crc kubenswrapper[4870]: I1014 08:55:57.926584 4870 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="264848824f7329af5e476a8b8c198139f82718e803fd993a8baa608f111c594f" Oct 14 08:55:57 crc kubenswrapper[4870]: I1014 08:55:57.926237 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-qfmh6" Oct 14 08:55:58 crc kubenswrapper[4870]: I1014 08:55:58.204909 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6fcb7f9899-zxtfl"] Oct 14 08:55:58 crc kubenswrapper[4870]: E1014 08:55:58.205394 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d51a0c5-822f-47d5-af81-b812ac44fe67" containerName="neutron-db-sync" Oct 14 08:55:58 crc kubenswrapper[4870]: I1014 08:55:58.205412 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d51a0c5-822f-47d5-af81-b812ac44fe67" containerName="neutron-db-sync" Oct 14 08:55:58 crc kubenswrapper[4870]: I1014 08:55:58.205691 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="8d51a0c5-822f-47d5-af81-b812ac44fe67" containerName="neutron-db-sync" Oct 14 08:55:58 crc kubenswrapper[4870]: I1014 08:55:58.206910 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6fcb7f9899-zxtfl" Oct 14 08:55:58 crc kubenswrapper[4870]: I1014 08:55:58.213272 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6fcb7f9899-zxtfl"] Oct 14 08:55:58 crc kubenswrapper[4870]: I1014 08:55:58.314930 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b3691846-4d55-4390-8cf4-94a9b28e4440-ovsdbserver-nb\") pod \"dnsmasq-dns-6fcb7f9899-zxtfl\" (UID: \"b3691846-4d55-4390-8cf4-94a9b28e4440\") " pod="openstack/dnsmasq-dns-6fcb7f9899-zxtfl" Oct 14 08:55:58 crc kubenswrapper[4870]: I1014 08:55:58.316165 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b3691846-4d55-4390-8cf4-94a9b28e4440-config\") pod \"dnsmasq-dns-6fcb7f9899-zxtfl\" (UID: \"b3691846-4d55-4390-8cf4-94a9b28e4440\") " pod="openstack/dnsmasq-dns-6fcb7f9899-zxtfl" Oct 14 08:55:58 crc kubenswrapper[4870]: I1014 08:55:58.316284 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b3691846-4d55-4390-8cf4-94a9b28e4440-ovsdbserver-sb\") pod \"dnsmasq-dns-6fcb7f9899-zxtfl\" (UID: \"b3691846-4d55-4390-8cf4-94a9b28e4440\") " pod="openstack/dnsmasq-dns-6fcb7f9899-zxtfl" Oct 14 08:55:58 crc kubenswrapper[4870]: I1014 08:55:58.316393 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wq65z\" (UniqueName: \"kubernetes.io/projected/b3691846-4d55-4390-8cf4-94a9b28e4440-kube-api-access-wq65z\") pod \"dnsmasq-dns-6fcb7f9899-zxtfl\" (UID: \"b3691846-4d55-4390-8cf4-94a9b28e4440\") " pod="openstack/dnsmasq-dns-6fcb7f9899-zxtfl" Oct 14 08:55:58 crc kubenswrapper[4870]: I1014 08:55:58.316579 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b3691846-4d55-4390-8cf4-94a9b28e4440-dns-svc\") pod \"dnsmasq-dns-6fcb7f9899-zxtfl\" (UID: \"b3691846-4d55-4390-8cf4-94a9b28e4440\") " pod="openstack/dnsmasq-dns-6fcb7f9899-zxtfl" Oct 14 08:55:58 crc kubenswrapper[4870]: I1014 08:55:58.418347 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b3691846-4d55-4390-8cf4-94a9b28e4440-config\") pod \"dnsmasq-dns-6fcb7f9899-zxtfl\" (UID: \"b3691846-4d55-4390-8cf4-94a9b28e4440\") " pod="openstack/dnsmasq-dns-6fcb7f9899-zxtfl" Oct 14 08:55:58 crc kubenswrapper[4870]: I1014 08:55:58.418409 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b3691846-4d55-4390-8cf4-94a9b28e4440-ovsdbserver-sb\") pod \"dnsmasq-dns-6fcb7f9899-zxtfl\" (UID: \"b3691846-4d55-4390-8cf4-94a9b28e4440\") " pod="openstack/dnsmasq-dns-6fcb7f9899-zxtfl" Oct 14 08:55:58 crc kubenswrapper[4870]: I1014 08:55:58.418467 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wq65z\" (UniqueName: \"kubernetes.io/projected/b3691846-4d55-4390-8cf4-94a9b28e4440-kube-api-access-wq65z\") pod \"dnsmasq-dns-6fcb7f9899-zxtfl\" (UID: \"b3691846-4d55-4390-8cf4-94a9b28e4440\") " pod="openstack/dnsmasq-dns-6fcb7f9899-zxtfl" Oct 14 08:55:58 crc kubenswrapper[4870]: I1014 08:55:58.418526 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b3691846-4d55-4390-8cf4-94a9b28e4440-dns-svc\") pod \"dnsmasq-dns-6fcb7f9899-zxtfl\" (UID: \"b3691846-4d55-4390-8cf4-94a9b28e4440\") " pod="openstack/dnsmasq-dns-6fcb7f9899-zxtfl" Oct 14 08:55:58 crc kubenswrapper[4870]: I1014 08:55:58.418562 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b3691846-4d55-4390-8cf4-94a9b28e4440-ovsdbserver-nb\") pod \"dnsmasq-dns-6fcb7f9899-zxtfl\" (UID: \"b3691846-4d55-4390-8cf4-94a9b28e4440\") " pod="openstack/dnsmasq-dns-6fcb7f9899-zxtfl" Oct 14 08:55:58 crc kubenswrapper[4870]: I1014 08:55:58.419911 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b3691846-4d55-4390-8cf4-94a9b28e4440-ovsdbserver-nb\") pod \"dnsmasq-dns-6fcb7f9899-zxtfl\" (UID: \"b3691846-4d55-4390-8cf4-94a9b28e4440\") " pod="openstack/dnsmasq-dns-6fcb7f9899-zxtfl" Oct 14 08:55:58 crc kubenswrapper[4870]: I1014 08:55:58.421256 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b3691846-4d55-4390-8cf4-94a9b28e4440-dns-svc\") pod \"dnsmasq-dns-6fcb7f9899-zxtfl\" (UID: \"b3691846-4d55-4390-8cf4-94a9b28e4440\") " pod="openstack/dnsmasq-dns-6fcb7f9899-zxtfl" Oct 14 08:55:58 crc kubenswrapper[4870]: I1014 08:55:58.421284 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b3691846-4d55-4390-8cf4-94a9b28e4440-ovsdbserver-sb\") pod \"dnsmasq-dns-6fcb7f9899-zxtfl\" (UID: \"b3691846-4d55-4390-8cf4-94a9b28e4440\") " pod="openstack/dnsmasq-dns-6fcb7f9899-zxtfl" Oct 14 08:55:58 crc kubenswrapper[4870]: I1014 08:55:58.421256 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b3691846-4d55-4390-8cf4-94a9b28e4440-config\") pod \"dnsmasq-dns-6fcb7f9899-zxtfl\" (UID: \"b3691846-4d55-4390-8cf4-94a9b28e4440\") " pod="openstack/dnsmasq-dns-6fcb7f9899-zxtfl" Oct 14 08:55:58 crc kubenswrapper[4870]: I1014 08:55:58.425758 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-75695669c7-rj984"] Oct 14 08:55:58 crc kubenswrapper[4870]: I1014 08:55:58.427152 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-75695669c7-rj984" Oct 14 08:55:58 crc kubenswrapper[4870]: I1014 08:55:58.432213 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Oct 14 08:55:58 crc kubenswrapper[4870]: I1014 08:55:58.432536 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-p2ckg" Oct 14 08:55:58 crc kubenswrapper[4870]: I1014 08:55:58.432718 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Oct 14 08:55:58 crc kubenswrapper[4870]: I1014 08:55:58.452953 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-75695669c7-rj984"] Oct 14 08:55:58 crc kubenswrapper[4870]: I1014 08:55:58.459220 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wq65z\" (UniqueName: \"kubernetes.io/projected/b3691846-4d55-4390-8cf4-94a9b28e4440-kube-api-access-wq65z\") pod \"dnsmasq-dns-6fcb7f9899-zxtfl\" (UID: \"b3691846-4d55-4390-8cf4-94a9b28e4440\") " pod="openstack/dnsmasq-dns-6fcb7f9899-zxtfl" Oct 14 08:55:58 crc kubenswrapper[4870]: I1014 08:55:58.520286 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47c05f0e-09b2-453f-b777-72836ec666f3-combined-ca-bundle\") pod \"neutron-75695669c7-rj984\" (UID: \"47c05f0e-09b2-453f-b777-72836ec666f3\") " pod="openstack/neutron-75695669c7-rj984" Oct 14 08:55:58 crc kubenswrapper[4870]: I1014 08:55:58.520393 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g5sdn\" (UniqueName: \"kubernetes.io/projected/47c05f0e-09b2-453f-b777-72836ec666f3-kube-api-access-g5sdn\") pod \"neutron-75695669c7-rj984\" (UID: \"47c05f0e-09b2-453f-b777-72836ec666f3\") " pod="openstack/neutron-75695669c7-rj984" Oct 14 08:55:58 crc kubenswrapper[4870]: I1014 08:55:58.520427 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/47c05f0e-09b2-453f-b777-72836ec666f3-httpd-config\") pod \"neutron-75695669c7-rj984\" (UID: \"47c05f0e-09b2-453f-b777-72836ec666f3\") " pod="openstack/neutron-75695669c7-rj984" Oct 14 08:55:58 crc kubenswrapper[4870]: I1014 08:55:58.520509 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/47c05f0e-09b2-453f-b777-72836ec666f3-config\") pod \"neutron-75695669c7-rj984\" (UID: \"47c05f0e-09b2-453f-b777-72836ec666f3\") " pod="openstack/neutron-75695669c7-rj984" Oct 14 08:55:58 crc kubenswrapper[4870]: I1014 08:55:58.551689 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6fcb7f9899-zxtfl" Oct 14 08:55:58 crc kubenswrapper[4870]: I1014 08:55:58.621724 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/47c05f0e-09b2-453f-b777-72836ec666f3-httpd-config\") pod \"neutron-75695669c7-rj984\" (UID: \"47c05f0e-09b2-453f-b777-72836ec666f3\") " pod="openstack/neutron-75695669c7-rj984" Oct 14 08:55:58 crc kubenswrapper[4870]: I1014 08:55:58.621829 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/47c05f0e-09b2-453f-b777-72836ec666f3-config\") pod \"neutron-75695669c7-rj984\" (UID: \"47c05f0e-09b2-453f-b777-72836ec666f3\") " pod="openstack/neutron-75695669c7-rj984" Oct 14 08:55:58 crc kubenswrapper[4870]: I1014 08:55:58.621888 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47c05f0e-09b2-453f-b777-72836ec666f3-combined-ca-bundle\") pod \"neutron-75695669c7-rj984\" (UID: \"47c05f0e-09b2-453f-b777-72836ec666f3\") " pod="openstack/neutron-75695669c7-rj984" Oct 14 08:55:58 crc kubenswrapper[4870]: I1014 08:55:58.621951 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g5sdn\" (UniqueName: \"kubernetes.io/projected/47c05f0e-09b2-453f-b777-72836ec666f3-kube-api-access-g5sdn\") pod \"neutron-75695669c7-rj984\" (UID: \"47c05f0e-09b2-453f-b777-72836ec666f3\") " pod="openstack/neutron-75695669c7-rj984" Oct 14 08:55:58 crc kubenswrapper[4870]: I1014 08:55:58.626565 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/47c05f0e-09b2-453f-b777-72836ec666f3-httpd-config\") pod \"neutron-75695669c7-rj984\" (UID: \"47c05f0e-09b2-453f-b777-72836ec666f3\") " pod="openstack/neutron-75695669c7-rj984" Oct 14 08:55:58 crc kubenswrapper[4870]: I1014 08:55:58.636960 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47c05f0e-09b2-453f-b777-72836ec666f3-combined-ca-bundle\") pod \"neutron-75695669c7-rj984\" (UID: \"47c05f0e-09b2-453f-b777-72836ec666f3\") " pod="openstack/neutron-75695669c7-rj984" Oct 14 08:55:58 crc kubenswrapper[4870]: I1014 08:55:58.637758 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/47c05f0e-09b2-453f-b777-72836ec666f3-config\") pod \"neutron-75695669c7-rj984\" (UID: \"47c05f0e-09b2-453f-b777-72836ec666f3\") " pod="openstack/neutron-75695669c7-rj984" Oct 14 08:55:58 crc kubenswrapper[4870]: I1014 08:55:58.639324 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g5sdn\" (UniqueName: \"kubernetes.io/projected/47c05f0e-09b2-453f-b777-72836ec666f3-kube-api-access-g5sdn\") pod \"neutron-75695669c7-rj984\" (UID: \"47c05f0e-09b2-453f-b777-72836ec666f3\") " pod="openstack/neutron-75695669c7-rj984" Oct 14 08:55:58 crc kubenswrapper[4870]: I1014 08:55:58.786007 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-75695669c7-rj984" Oct 14 08:55:59 crc kubenswrapper[4870]: I1014 08:55:59.070226 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6fcb7f9899-zxtfl"] Oct 14 08:55:59 crc kubenswrapper[4870]: W1014 08:55:59.082598 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb3691846_4d55_4390_8cf4_94a9b28e4440.slice/crio-e8e29e5abfff49fa7c476afdddc9aa532b1d27ce2e4574ddf5318dbd0f116dbe WatchSource:0}: Error finding container e8e29e5abfff49fa7c476afdddc9aa532b1d27ce2e4574ddf5318dbd0f116dbe: Status 404 returned error can't find the container with id e8e29e5abfff49fa7c476afdddc9aa532b1d27ce2e4574ddf5318dbd0f116dbe Oct 14 08:55:59 crc kubenswrapper[4870]: I1014 08:55:59.383899 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-75695669c7-rj984"] Oct 14 08:55:59 crc kubenswrapper[4870]: W1014 08:55:59.391265 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod47c05f0e_09b2_453f_b777_72836ec666f3.slice/crio-92ab1786da232418eb42862bc5abb98e6eec09e2f6ec0fd26fafd4c8817cb238 WatchSource:0}: Error finding container 92ab1786da232418eb42862bc5abb98e6eec09e2f6ec0fd26fafd4c8817cb238: Status 404 returned error can't find the container with id 92ab1786da232418eb42862bc5abb98e6eec09e2f6ec0fd26fafd4c8817cb238 Oct 14 08:55:59 crc kubenswrapper[4870]: I1014 08:55:59.953510 4870 generic.go:334] "Generic (PLEG): container finished" podID="b3691846-4d55-4390-8cf4-94a9b28e4440" containerID="eb3cbe7ff45c41a4d8c72ac6527f2f9d1676951bee76d6f7d0832963a48fc451" exitCode=0 Oct 14 08:55:59 crc kubenswrapper[4870]: I1014 08:55:59.953622 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6fcb7f9899-zxtfl" event={"ID":"b3691846-4d55-4390-8cf4-94a9b28e4440","Type":"ContainerDied","Data":"eb3cbe7ff45c41a4d8c72ac6527f2f9d1676951bee76d6f7d0832963a48fc451"} Oct 14 08:55:59 crc kubenswrapper[4870]: I1014 08:55:59.953977 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6fcb7f9899-zxtfl" event={"ID":"b3691846-4d55-4390-8cf4-94a9b28e4440","Type":"ContainerStarted","Data":"e8e29e5abfff49fa7c476afdddc9aa532b1d27ce2e4574ddf5318dbd0f116dbe"} Oct 14 08:55:59 crc kubenswrapper[4870]: I1014 08:55:59.957125 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-75695669c7-rj984" event={"ID":"47c05f0e-09b2-453f-b777-72836ec666f3","Type":"ContainerStarted","Data":"79e0a3fffce0aa7b925d3966e8fc094fed35140c4e54de41a04a2245aa0122e6"} Oct 14 08:55:59 crc kubenswrapper[4870]: I1014 08:55:59.957309 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-75695669c7-rj984" Oct 14 08:55:59 crc kubenswrapper[4870]: I1014 08:55:59.957327 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-75695669c7-rj984" event={"ID":"47c05f0e-09b2-453f-b777-72836ec666f3","Type":"ContainerStarted","Data":"0862d2a9b1ccad23f3bc9e8a1a993c22d093acf8399bfa959b89c2f25da3177e"} Oct 14 08:55:59 crc kubenswrapper[4870]: I1014 08:55:59.957356 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-75695669c7-rj984" event={"ID":"47c05f0e-09b2-453f-b777-72836ec666f3","Type":"ContainerStarted","Data":"92ab1786da232418eb42862bc5abb98e6eec09e2f6ec0fd26fafd4c8817cb238"} Oct 14 08:56:00 crc kubenswrapper[4870]: I1014 08:56:00.991955 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6fcb7f9899-zxtfl" event={"ID":"b3691846-4d55-4390-8cf4-94a9b28e4440","Type":"ContainerStarted","Data":"522faffedac332784e3353c8f156a89f2734f5295fd2d562bad530cf310dcc75"} Oct 14 08:56:00 crc kubenswrapper[4870]: I1014 08:56:00.992348 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6fcb7f9899-zxtfl" Oct 14 08:56:01 crc kubenswrapper[4870]: I1014 08:56:01.011202 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-75695669c7-rj984" podStartSLOduration=3.011176775 podStartE2EDuration="3.011176775s" podCreationTimestamp="2025-10-14 08:55:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 08:56:00.012743588 +0000 UTC m=+6895.710103959" watchObservedRunningTime="2025-10-14 08:56:01.011176775 +0000 UTC m=+6896.708537176" Oct 14 08:56:01 crc kubenswrapper[4870]: I1014 08:56:01.017969 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6fcb7f9899-zxtfl" podStartSLOduration=3.017948941 podStartE2EDuration="3.017948941s" podCreationTimestamp="2025-10-14 08:55:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 08:56:01.010783415 +0000 UTC m=+6896.708143826" watchObservedRunningTime="2025-10-14 08:56:01.017948941 +0000 UTC m=+6896.715309312" Oct 14 08:56:03 crc kubenswrapper[4870]: I1014 08:56:03.034571 4870 scope.go:117] "RemoveContainer" containerID="8c0f754fe94b7cb7b9bb5dcd613f7e42093968b49509d8ee37e8fd4db89eedd5" Oct 14 08:56:03 crc kubenswrapper[4870]: E1014 08:56:03.035129 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 08:56:08 crc kubenswrapper[4870]: I1014 08:56:08.553636 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6fcb7f9899-zxtfl" Oct 14 08:56:08 crc kubenswrapper[4870]: I1014 08:56:08.615205 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-69f7d94959-2sm7c"] Oct 14 08:56:08 crc kubenswrapper[4870]: I1014 08:56:08.615451 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-69f7d94959-2sm7c" podUID="2491fadf-97f6-4e7e-932b-3ca607c015ab" containerName="dnsmasq-dns" containerID="cri-o://f9ef13564e588ee8e9f11fa2197497d04269bcfbcca40379a8d841340a5750b6" gracePeriod=10 Oct 14 08:56:09 crc kubenswrapper[4870]: I1014 08:56:09.079294 4870 generic.go:334] "Generic (PLEG): container finished" podID="2491fadf-97f6-4e7e-932b-3ca607c015ab" containerID="f9ef13564e588ee8e9f11fa2197497d04269bcfbcca40379a8d841340a5750b6" exitCode=0 Oct 14 08:56:09 crc kubenswrapper[4870]: I1014 08:56:09.079369 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-69f7d94959-2sm7c" event={"ID":"2491fadf-97f6-4e7e-932b-3ca607c015ab","Type":"ContainerDied","Data":"f9ef13564e588ee8e9f11fa2197497d04269bcfbcca40379a8d841340a5750b6"} Oct 14 08:56:09 crc kubenswrapper[4870]: I1014 08:56:09.080028 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-69f7d94959-2sm7c" event={"ID":"2491fadf-97f6-4e7e-932b-3ca607c015ab","Type":"ContainerDied","Data":"79987c8cdafa09d6fdd69831214d775d2707d58ad4de6306ca9c3d02e93f1cd2"} Oct 14 08:56:09 crc kubenswrapper[4870]: I1014 08:56:09.080150 4870 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="79987c8cdafa09d6fdd69831214d775d2707d58ad4de6306ca9c3d02e93f1cd2" Oct 14 08:56:09 crc kubenswrapper[4870]: I1014 08:56:09.111148 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-69f7d94959-2sm7c" Oct 14 08:56:09 crc kubenswrapper[4870]: I1014 08:56:09.245998 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2491fadf-97f6-4e7e-932b-3ca607c015ab-ovsdbserver-sb\") pod \"2491fadf-97f6-4e7e-932b-3ca607c015ab\" (UID: \"2491fadf-97f6-4e7e-932b-3ca607c015ab\") " Oct 14 08:56:09 crc kubenswrapper[4870]: I1014 08:56:09.246115 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2491fadf-97f6-4e7e-932b-3ca607c015ab-config\") pod \"2491fadf-97f6-4e7e-932b-3ca607c015ab\" (UID: \"2491fadf-97f6-4e7e-932b-3ca607c015ab\") " Oct 14 08:56:09 crc kubenswrapper[4870]: I1014 08:56:09.246194 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2491fadf-97f6-4e7e-932b-3ca607c015ab-ovsdbserver-nb\") pod \"2491fadf-97f6-4e7e-932b-3ca607c015ab\" (UID: \"2491fadf-97f6-4e7e-932b-3ca607c015ab\") " Oct 14 08:56:09 crc kubenswrapper[4870]: I1014 08:56:09.246247 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lblql\" (UniqueName: \"kubernetes.io/projected/2491fadf-97f6-4e7e-932b-3ca607c015ab-kube-api-access-lblql\") pod \"2491fadf-97f6-4e7e-932b-3ca607c015ab\" (UID: \"2491fadf-97f6-4e7e-932b-3ca607c015ab\") " Oct 14 08:56:09 crc kubenswrapper[4870]: I1014 08:56:09.246291 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2491fadf-97f6-4e7e-932b-3ca607c015ab-dns-svc\") pod \"2491fadf-97f6-4e7e-932b-3ca607c015ab\" (UID: \"2491fadf-97f6-4e7e-932b-3ca607c015ab\") " Oct 14 08:56:09 crc kubenswrapper[4870]: I1014 08:56:09.254811 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2491fadf-97f6-4e7e-932b-3ca607c015ab-kube-api-access-lblql" (OuterVolumeSpecName: "kube-api-access-lblql") pod "2491fadf-97f6-4e7e-932b-3ca607c015ab" (UID: "2491fadf-97f6-4e7e-932b-3ca607c015ab"). InnerVolumeSpecName "kube-api-access-lblql". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:56:09 crc kubenswrapper[4870]: I1014 08:56:09.288941 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2491fadf-97f6-4e7e-932b-3ca607c015ab-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "2491fadf-97f6-4e7e-932b-3ca607c015ab" (UID: "2491fadf-97f6-4e7e-932b-3ca607c015ab"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:56:09 crc kubenswrapper[4870]: I1014 08:56:09.295752 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2491fadf-97f6-4e7e-932b-3ca607c015ab-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "2491fadf-97f6-4e7e-932b-3ca607c015ab" (UID: "2491fadf-97f6-4e7e-932b-3ca607c015ab"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:56:09 crc kubenswrapper[4870]: I1014 08:56:09.306076 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2491fadf-97f6-4e7e-932b-3ca607c015ab-config" (OuterVolumeSpecName: "config") pod "2491fadf-97f6-4e7e-932b-3ca607c015ab" (UID: "2491fadf-97f6-4e7e-932b-3ca607c015ab"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:56:09 crc kubenswrapper[4870]: I1014 08:56:09.308787 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2491fadf-97f6-4e7e-932b-3ca607c015ab-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "2491fadf-97f6-4e7e-932b-3ca607c015ab" (UID: "2491fadf-97f6-4e7e-932b-3ca607c015ab"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:56:09 crc kubenswrapper[4870]: I1014 08:56:09.348129 4870 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2491fadf-97f6-4e7e-932b-3ca607c015ab-config\") on node \"crc\" DevicePath \"\"" Oct 14 08:56:09 crc kubenswrapper[4870]: I1014 08:56:09.348161 4870 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2491fadf-97f6-4e7e-932b-3ca607c015ab-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 14 08:56:09 crc kubenswrapper[4870]: I1014 08:56:09.348174 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lblql\" (UniqueName: \"kubernetes.io/projected/2491fadf-97f6-4e7e-932b-3ca607c015ab-kube-api-access-lblql\") on node \"crc\" DevicePath \"\"" Oct 14 08:56:09 crc kubenswrapper[4870]: I1014 08:56:09.348184 4870 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2491fadf-97f6-4e7e-932b-3ca607c015ab-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 14 08:56:09 crc kubenswrapper[4870]: I1014 08:56:09.348194 4870 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2491fadf-97f6-4e7e-932b-3ca607c015ab-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 14 08:56:10 crc kubenswrapper[4870]: I1014 08:56:10.086057 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-69f7d94959-2sm7c" Oct 14 08:56:10 crc kubenswrapper[4870]: I1014 08:56:10.121882 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-69f7d94959-2sm7c"] Oct 14 08:56:10 crc kubenswrapper[4870]: I1014 08:56:10.131957 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-69f7d94959-2sm7c"] Oct 14 08:56:11 crc kubenswrapper[4870]: I1014 08:56:11.049940 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2491fadf-97f6-4e7e-932b-3ca607c015ab" path="/var/lib/kubelet/pods/2491fadf-97f6-4e7e-932b-3ca607c015ab/volumes" Oct 14 08:56:17 crc kubenswrapper[4870]: I1014 08:56:17.034178 4870 scope.go:117] "RemoveContainer" containerID="8c0f754fe94b7cb7b9bb5dcd613f7e42093968b49509d8ee37e8fd4db89eedd5" Oct 14 08:56:17 crc kubenswrapper[4870]: E1014 08:56:17.035010 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 08:56:28 crc kubenswrapper[4870]: I1014 08:56:28.800238 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-75695669c7-rj984" Oct 14 08:56:32 crc kubenswrapper[4870]: I1014 08:56:32.034610 4870 scope.go:117] "RemoveContainer" containerID="8c0f754fe94b7cb7b9bb5dcd613f7e42093968b49509d8ee37e8fd4db89eedd5" Oct 14 08:56:32 crc kubenswrapper[4870]: E1014 08:56:32.035422 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 08:56:36 crc kubenswrapper[4870]: I1014 08:56:36.708507 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-8h9wb"] Oct 14 08:56:36 crc kubenswrapper[4870]: E1014 08:56:36.710244 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2491fadf-97f6-4e7e-932b-3ca607c015ab" containerName="init" Oct 14 08:56:36 crc kubenswrapper[4870]: I1014 08:56:36.710332 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="2491fadf-97f6-4e7e-932b-3ca607c015ab" containerName="init" Oct 14 08:56:36 crc kubenswrapper[4870]: E1014 08:56:36.710419 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2491fadf-97f6-4e7e-932b-3ca607c015ab" containerName="dnsmasq-dns" Oct 14 08:56:36 crc kubenswrapper[4870]: I1014 08:56:36.710506 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="2491fadf-97f6-4e7e-932b-3ca607c015ab" containerName="dnsmasq-dns" Oct 14 08:56:36 crc kubenswrapper[4870]: I1014 08:56:36.710727 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="2491fadf-97f6-4e7e-932b-3ca607c015ab" containerName="dnsmasq-dns" Oct 14 08:56:36 crc kubenswrapper[4870]: I1014 08:56:36.711420 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-8h9wb" Oct 14 08:56:36 crc kubenswrapper[4870]: I1014 08:56:36.717468 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-8h9wb"] Oct 14 08:56:36 crc kubenswrapper[4870]: I1014 08:56:36.868588 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nkrh2\" (UniqueName: \"kubernetes.io/projected/673ae6d4-e844-4bda-8be7-0cc398c456c9-kube-api-access-nkrh2\") pod \"glance-db-create-8h9wb\" (UID: \"673ae6d4-e844-4bda-8be7-0cc398c456c9\") " pod="openstack/glance-db-create-8h9wb" Oct 14 08:56:36 crc kubenswrapper[4870]: I1014 08:56:36.971101 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nkrh2\" (UniqueName: \"kubernetes.io/projected/673ae6d4-e844-4bda-8be7-0cc398c456c9-kube-api-access-nkrh2\") pod \"glance-db-create-8h9wb\" (UID: \"673ae6d4-e844-4bda-8be7-0cc398c456c9\") " pod="openstack/glance-db-create-8h9wb" Oct 14 08:56:36 crc kubenswrapper[4870]: I1014 08:56:36.999536 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nkrh2\" (UniqueName: \"kubernetes.io/projected/673ae6d4-e844-4bda-8be7-0cc398c456c9-kube-api-access-nkrh2\") pod \"glance-db-create-8h9wb\" (UID: \"673ae6d4-e844-4bda-8be7-0cc398c456c9\") " pod="openstack/glance-db-create-8h9wb" Oct 14 08:56:37 crc kubenswrapper[4870]: I1014 08:56:37.034176 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-8h9wb" Oct 14 08:56:37 crc kubenswrapper[4870]: I1014 08:56:37.522171 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-8h9wb"] Oct 14 08:56:38 crc kubenswrapper[4870]: I1014 08:56:38.409203 4870 generic.go:334] "Generic (PLEG): container finished" podID="673ae6d4-e844-4bda-8be7-0cc398c456c9" containerID="df6172ec65f91a6d51dc77c1a71e028a0ae2e669355403f5e0db3860d68e8a0c" exitCode=0 Oct 14 08:56:38 crc kubenswrapper[4870]: I1014 08:56:38.409348 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-8h9wb" event={"ID":"673ae6d4-e844-4bda-8be7-0cc398c456c9","Type":"ContainerDied","Data":"df6172ec65f91a6d51dc77c1a71e028a0ae2e669355403f5e0db3860d68e8a0c"} Oct 14 08:56:38 crc kubenswrapper[4870]: I1014 08:56:38.409669 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-8h9wb" event={"ID":"673ae6d4-e844-4bda-8be7-0cc398c456c9","Type":"ContainerStarted","Data":"3d733978ea0cc22911465a6fdc73626c8adbcbc535f32f099d1ac8b49366e087"} Oct 14 08:56:39 crc kubenswrapper[4870]: I1014 08:56:39.819096 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-8h9wb" Oct 14 08:56:39 crc kubenswrapper[4870]: I1014 08:56:39.926814 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nkrh2\" (UniqueName: \"kubernetes.io/projected/673ae6d4-e844-4bda-8be7-0cc398c456c9-kube-api-access-nkrh2\") pod \"673ae6d4-e844-4bda-8be7-0cc398c456c9\" (UID: \"673ae6d4-e844-4bda-8be7-0cc398c456c9\") " Oct 14 08:56:39 crc kubenswrapper[4870]: I1014 08:56:39.936726 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/673ae6d4-e844-4bda-8be7-0cc398c456c9-kube-api-access-nkrh2" (OuterVolumeSpecName: "kube-api-access-nkrh2") pod "673ae6d4-e844-4bda-8be7-0cc398c456c9" (UID: "673ae6d4-e844-4bda-8be7-0cc398c456c9"). InnerVolumeSpecName "kube-api-access-nkrh2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:56:40 crc kubenswrapper[4870]: I1014 08:56:40.029264 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nkrh2\" (UniqueName: \"kubernetes.io/projected/673ae6d4-e844-4bda-8be7-0cc398c456c9-kube-api-access-nkrh2\") on node \"crc\" DevicePath \"\"" Oct 14 08:56:40 crc kubenswrapper[4870]: I1014 08:56:40.433669 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-8h9wb" event={"ID":"673ae6d4-e844-4bda-8be7-0cc398c456c9","Type":"ContainerDied","Data":"3d733978ea0cc22911465a6fdc73626c8adbcbc535f32f099d1ac8b49366e087"} Oct 14 08:56:40 crc kubenswrapper[4870]: I1014 08:56:40.434203 4870 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3d733978ea0cc22911465a6fdc73626c8adbcbc535f32f099d1ac8b49366e087" Oct 14 08:56:40 crc kubenswrapper[4870]: I1014 08:56:40.433714 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-8h9wb" Oct 14 08:56:45 crc kubenswrapper[4870]: I1014 08:56:45.041784 4870 scope.go:117] "RemoveContainer" containerID="8c0f754fe94b7cb7b9bb5dcd613f7e42093968b49509d8ee37e8fd4db89eedd5" Oct 14 08:56:45 crc kubenswrapper[4870]: E1014 08:56:45.042405 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 08:56:46 crc kubenswrapper[4870]: I1014 08:56:46.751084 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-6d91-account-create-7mzvp"] Oct 14 08:56:46 crc kubenswrapper[4870]: E1014 08:56:46.751599 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="673ae6d4-e844-4bda-8be7-0cc398c456c9" containerName="mariadb-database-create" Oct 14 08:56:46 crc kubenswrapper[4870]: I1014 08:56:46.751618 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="673ae6d4-e844-4bda-8be7-0cc398c456c9" containerName="mariadb-database-create" Oct 14 08:56:46 crc kubenswrapper[4870]: I1014 08:56:46.751893 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="673ae6d4-e844-4bda-8be7-0cc398c456c9" containerName="mariadb-database-create" Oct 14 08:56:46 crc kubenswrapper[4870]: I1014 08:56:46.752592 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-6d91-account-create-7mzvp" Oct 14 08:56:46 crc kubenswrapper[4870]: I1014 08:56:46.755681 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Oct 14 08:56:46 crc kubenswrapper[4870]: I1014 08:56:46.764126 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-6d91-account-create-7mzvp"] Oct 14 08:56:46 crc kubenswrapper[4870]: I1014 08:56:46.873121 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vwrzc\" (UniqueName: \"kubernetes.io/projected/4e4335da-b836-4ca3-b683-f658d7085175-kube-api-access-vwrzc\") pod \"glance-6d91-account-create-7mzvp\" (UID: \"4e4335da-b836-4ca3-b683-f658d7085175\") " pod="openstack/glance-6d91-account-create-7mzvp" Oct 14 08:56:46 crc kubenswrapper[4870]: I1014 08:56:46.975828 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vwrzc\" (UniqueName: \"kubernetes.io/projected/4e4335da-b836-4ca3-b683-f658d7085175-kube-api-access-vwrzc\") pod \"glance-6d91-account-create-7mzvp\" (UID: \"4e4335da-b836-4ca3-b683-f658d7085175\") " pod="openstack/glance-6d91-account-create-7mzvp" Oct 14 08:56:47 crc kubenswrapper[4870]: I1014 08:56:47.002370 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vwrzc\" (UniqueName: \"kubernetes.io/projected/4e4335da-b836-4ca3-b683-f658d7085175-kube-api-access-vwrzc\") pod \"glance-6d91-account-create-7mzvp\" (UID: \"4e4335da-b836-4ca3-b683-f658d7085175\") " pod="openstack/glance-6d91-account-create-7mzvp" Oct 14 08:56:47 crc kubenswrapper[4870]: I1014 08:56:47.088407 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-6d91-account-create-7mzvp" Oct 14 08:56:47 crc kubenswrapper[4870]: I1014 08:56:47.573270 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-6d91-account-create-7mzvp"] Oct 14 08:56:47 crc kubenswrapper[4870]: W1014 08:56:47.590653 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4e4335da_b836_4ca3_b683_f658d7085175.slice/crio-be4a52a4f042e0b1806ba974010212d12828dd5312fbd0c59963baa9de8f06dc WatchSource:0}: Error finding container be4a52a4f042e0b1806ba974010212d12828dd5312fbd0c59963baa9de8f06dc: Status 404 returned error can't find the container with id be4a52a4f042e0b1806ba974010212d12828dd5312fbd0c59963baa9de8f06dc Oct 14 08:56:48 crc kubenswrapper[4870]: I1014 08:56:48.528489 4870 generic.go:334] "Generic (PLEG): container finished" podID="4e4335da-b836-4ca3-b683-f658d7085175" containerID="f2b6e0bf724555237b504d3bf97ca25246546d718a3200ab1c570224f37f8f59" exitCode=0 Oct 14 08:56:48 crc kubenswrapper[4870]: I1014 08:56:48.528529 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-6d91-account-create-7mzvp" event={"ID":"4e4335da-b836-4ca3-b683-f658d7085175","Type":"ContainerDied","Data":"f2b6e0bf724555237b504d3bf97ca25246546d718a3200ab1c570224f37f8f59"} Oct 14 08:56:48 crc kubenswrapper[4870]: I1014 08:56:48.528569 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-6d91-account-create-7mzvp" event={"ID":"4e4335da-b836-4ca3-b683-f658d7085175","Type":"ContainerStarted","Data":"be4a52a4f042e0b1806ba974010212d12828dd5312fbd0c59963baa9de8f06dc"} Oct 14 08:56:49 crc kubenswrapper[4870]: I1014 08:56:49.895174 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-6d91-account-create-7mzvp" Oct 14 08:56:50 crc kubenswrapper[4870]: I1014 08:56:50.034841 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vwrzc\" (UniqueName: \"kubernetes.io/projected/4e4335da-b836-4ca3-b683-f658d7085175-kube-api-access-vwrzc\") pod \"4e4335da-b836-4ca3-b683-f658d7085175\" (UID: \"4e4335da-b836-4ca3-b683-f658d7085175\") " Oct 14 08:56:50 crc kubenswrapper[4870]: I1014 08:56:50.040750 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4e4335da-b836-4ca3-b683-f658d7085175-kube-api-access-vwrzc" (OuterVolumeSpecName: "kube-api-access-vwrzc") pod "4e4335da-b836-4ca3-b683-f658d7085175" (UID: "4e4335da-b836-4ca3-b683-f658d7085175"). InnerVolumeSpecName "kube-api-access-vwrzc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:56:50 crc kubenswrapper[4870]: I1014 08:56:50.136940 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vwrzc\" (UniqueName: \"kubernetes.io/projected/4e4335da-b836-4ca3-b683-f658d7085175-kube-api-access-vwrzc\") on node \"crc\" DevicePath \"\"" Oct 14 08:56:50 crc kubenswrapper[4870]: I1014 08:56:50.551671 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-6d91-account-create-7mzvp" event={"ID":"4e4335da-b836-4ca3-b683-f658d7085175","Type":"ContainerDied","Data":"be4a52a4f042e0b1806ba974010212d12828dd5312fbd0c59963baa9de8f06dc"} Oct 14 08:56:50 crc kubenswrapper[4870]: I1014 08:56:50.551713 4870 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="be4a52a4f042e0b1806ba974010212d12828dd5312fbd0c59963baa9de8f06dc" Oct 14 08:56:50 crc kubenswrapper[4870]: I1014 08:56:50.551761 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-6d91-account-create-7mzvp" Oct 14 08:56:51 crc kubenswrapper[4870]: I1014 08:56:51.811681 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-4p4kx"] Oct 14 08:56:51 crc kubenswrapper[4870]: E1014 08:56:51.812032 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4e4335da-b836-4ca3-b683-f658d7085175" containerName="mariadb-account-create" Oct 14 08:56:51 crc kubenswrapper[4870]: I1014 08:56:51.812044 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="4e4335da-b836-4ca3-b683-f658d7085175" containerName="mariadb-account-create" Oct 14 08:56:51 crc kubenswrapper[4870]: I1014 08:56:51.812187 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="4e4335da-b836-4ca3-b683-f658d7085175" containerName="mariadb-account-create" Oct 14 08:56:51 crc kubenswrapper[4870]: I1014 08:56:51.812732 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-4p4kx" Oct 14 08:56:51 crc kubenswrapper[4870]: I1014 08:56:51.814983 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Oct 14 08:56:51 crc kubenswrapper[4870]: I1014 08:56:51.818304 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-f98hg" Oct 14 08:56:51 crc kubenswrapper[4870]: I1014 08:56:51.837575 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-4p4kx"] Oct 14 08:56:51 crc kubenswrapper[4870]: I1014 08:56:51.982305 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f7a69635-d246-4920-b96b-6b232fa4e813-db-sync-config-data\") pod \"glance-db-sync-4p4kx\" (UID: \"f7a69635-d246-4920-b96b-6b232fa4e813\") " pod="openstack/glance-db-sync-4p4kx" Oct 14 08:56:51 crc kubenswrapper[4870]: I1014 08:56:51.982389 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7a69635-d246-4920-b96b-6b232fa4e813-config-data\") pod \"glance-db-sync-4p4kx\" (UID: \"f7a69635-d246-4920-b96b-6b232fa4e813\") " pod="openstack/glance-db-sync-4p4kx" Oct 14 08:56:51 crc kubenswrapper[4870]: I1014 08:56:51.982452 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7a69635-d246-4920-b96b-6b232fa4e813-combined-ca-bundle\") pod \"glance-db-sync-4p4kx\" (UID: \"f7a69635-d246-4920-b96b-6b232fa4e813\") " pod="openstack/glance-db-sync-4p4kx" Oct 14 08:56:51 crc kubenswrapper[4870]: I1014 08:56:51.982476 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k7zxc\" (UniqueName: \"kubernetes.io/projected/f7a69635-d246-4920-b96b-6b232fa4e813-kube-api-access-k7zxc\") pod \"glance-db-sync-4p4kx\" (UID: \"f7a69635-d246-4920-b96b-6b232fa4e813\") " pod="openstack/glance-db-sync-4p4kx" Oct 14 08:56:52 crc kubenswrapper[4870]: I1014 08:56:52.084342 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f7a69635-d246-4920-b96b-6b232fa4e813-db-sync-config-data\") pod \"glance-db-sync-4p4kx\" (UID: \"f7a69635-d246-4920-b96b-6b232fa4e813\") " pod="openstack/glance-db-sync-4p4kx" Oct 14 08:56:52 crc kubenswrapper[4870]: I1014 08:56:52.084468 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7a69635-d246-4920-b96b-6b232fa4e813-config-data\") pod \"glance-db-sync-4p4kx\" (UID: \"f7a69635-d246-4920-b96b-6b232fa4e813\") " pod="openstack/glance-db-sync-4p4kx" Oct 14 08:56:52 crc kubenswrapper[4870]: I1014 08:56:52.084532 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7a69635-d246-4920-b96b-6b232fa4e813-combined-ca-bundle\") pod \"glance-db-sync-4p4kx\" (UID: \"f7a69635-d246-4920-b96b-6b232fa4e813\") " pod="openstack/glance-db-sync-4p4kx" Oct 14 08:56:52 crc kubenswrapper[4870]: I1014 08:56:52.084570 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k7zxc\" (UniqueName: \"kubernetes.io/projected/f7a69635-d246-4920-b96b-6b232fa4e813-kube-api-access-k7zxc\") pod \"glance-db-sync-4p4kx\" (UID: \"f7a69635-d246-4920-b96b-6b232fa4e813\") " pod="openstack/glance-db-sync-4p4kx" Oct 14 08:56:52 crc kubenswrapper[4870]: I1014 08:56:52.092113 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f7a69635-d246-4920-b96b-6b232fa4e813-db-sync-config-data\") pod \"glance-db-sync-4p4kx\" (UID: \"f7a69635-d246-4920-b96b-6b232fa4e813\") " pod="openstack/glance-db-sync-4p4kx" Oct 14 08:56:52 crc kubenswrapper[4870]: I1014 08:56:52.092116 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7a69635-d246-4920-b96b-6b232fa4e813-config-data\") pod \"glance-db-sync-4p4kx\" (UID: \"f7a69635-d246-4920-b96b-6b232fa4e813\") " pod="openstack/glance-db-sync-4p4kx" Oct 14 08:56:52 crc kubenswrapper[4870]: I1014 08:56:52.092332 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7a69635-d246-4920-b96b-6b232fa4e813-combined-ca-bundle\") pod \"glance-db-sync-4p4kx\" (UID: \"f7a69635-d246-4920-b96b-6b232fa4e813\") " pod="openstack/glance-db-sync-4p4kx" Oct 14 08:56:52 crc kubenswrapper[4870]: I1014 08:56:52.100016 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k7zxc\" (UniqueName: \"kubernetes.io/projected/f7a69635-d246-4920-b96b-6b232fa4e813-kube-api-access-k7zxc\") pod \"glance-db-sync-4p4kx\" (UID: \"f7a69635-d246-4920-b96b-6b232fa4e813\") " pod="openstack/glance-db-sync-4p4kx" Oct 14 08:56:52 crc kubenswrapper[4870]: I1014 08:56:52.134386 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-4p4kx" Oct 14 08:56:53 crc kubenswrapper[4870]: I1014 08:56:53.185071 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-4p4kx"] Oct 14 08:56:53 crc kubenswrapper[4870]: I1014 08:56:53.583641 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-4p4kx" event={"ID":"f7a69635-d246-4920-b96b-6b232fa4e813","Type":"ContainerStarted","Data":"15524e006051bc007ba24f6a1734e8ebb7b8453fd733127cfc09d974a4f1a363"} Oct 14 08:57:00 crc kubenswrapper[4870]: I1014 08:57:00.034190 4870 scope.go:117] "RemoveContainer" containerID="8c0f754fe94b7cb7b9bb5dcd613f7e42093968b49509d8ee37e8fd4db89eedd5" Oct 14 08:57:00 crc kubenswrapper[4870]: E1014 08:57:00.035318 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 08:57:09 crc kubenswrapper[4870]: I1014 08:57:09.720467 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-4p4kx" event={"ID":"f7a69635-d246-4920-b96b-6b232fa4e813","Type":"ContainerStarted","Data":"8e6799c323a696d937ce6adae78594ef16d1caaf215d3b849a6c746b0dccc342"} Oct 14 08:57:09 crc kubenswrapper[4870]: I1014 08:57:09.759393 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-4p4kx" podStartSLOduration=3.217019747 podStartE2EDuration="18.759363182s" podCreationTimestamp="2025-10-14 08:56:51 +0000 UTC" firstStartedPulling="2025-10-14 08:56:53.191080686 +0000 UTC m=+6948.888441057" lastFinishedPulling="2025-10-14 08:57:08.733424121 +0000 UTC m=+6964.430784492" observedRunningTime="2025-10-14 08:57:09.750822762 +0000 UTC m=+6965.448183173" watchObservedRunningTime="2025-10-14 08:57:09.759363182 +0000 UTC m=+6965.456723583" Oct 14 08:57:12 crc kubenswrapper[4870]: I1014 08:57:12.761907 4870 generic.go:334] "Generic (PLEG): container finished" podID="f7a69635-d246-4920-b96b-6b232fa4e813" containerID="8e6799c323a696d937ce6adae78594ef16d1caaf215d3b849a6c746b0dccc342" exitCode=0 Oct 14 08:57:12 crc kubenswrapper[4870]: I1014 08:57:12.761975 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-4p4kx" event={"ID":"f7a69635-d246-4920-b96b-6b232fa4e813","Type":"ContainerDied","Data":"8e6799c323a696d937ce6adae78594ef16d1caaf215d3b849a6c746b0dccc342"} Oct 14 08:57:13 crc kubenswrapper[4870]: I1014 08:57:13.035290 4870 scope.go:117] "RemoveContainer" containerID="8c0f754fe94b7cb7b9bb5dcd613f7e42093968b49509d8ee37e8fd4db89eedd5" Oct 14 08:57:13 crc kubenswrapper[4870]: E1014 08:57:13.035602 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 08:57:14 crc kubenswrapper[4870]: I1014 08:57:14.266398 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-4p4kx" Oct 14 08:57:14 crc kubenswrapper[4870]: I1014 08:57:14.380114 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7a69635-d246-4920-b96b-6b232fa4e813-config-data\") pod \"f7a69635-d246-4920-b96b-6b232fa4e813\" (UID: \"f7a69635-d246-4920-b96b-6b232fa4e813\") " Oct 14 08:57:14 crc kubenswrapper[4870]: I1014 08:57:14.380258 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7a69635-d246-4920-b96b-6b232fa4e813-combined-ca-bundle\") pod \"f7a69635-d246-4920-b96b-6b232fa4e813\" (UID: \"f7a69635-d246-4920-b96b-6b232fa4e813\") " Oct 14 08:57:14 crc kubenswrapper[4870]: I1014 08:57:14.380378 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f7a69635-d246-4920-b96b-6b232fa4e813-db-sync-config-data\") pod \"f7a69635-d246-4920-b96b-6b232fa4e813\" (UID: \"f7a69635-d246-4920-b96b-6b232fa4e813\") " Oct 14 08:57:14 crc kubenswrapper[4870]: I1014 08:57:14.381580 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k7zxc\" (UniqueName: \"kubernetes.io/projected/f7a69635-d246-4920-b96b-6b232fa4e813-kube-api-access-k7zxc\") pod \"f7a69635-d246-4920-b96b-6b232fa4e813\" (UID: \"f7a69635-d246-4920-b96b-6b232fa4e813\") " Oct 14 08:57:14 crc kubenswrapper[4870]: I1014 08:57:14.387742 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f7a69635-d246-4920-b96b-6b232fa4e813-kube-api-access-k7zxc" (OuterVolumeSpecName: "kube-api-access-k7zxc") pod "f7a69635-d246-4920-b96b-6b232fa4e813" (UID: "f7a69635-d246-4920-b96b-6b232fa4e813"). InnerVolumeSpecName "kube-api-access-k7zxc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:57:14 crc kubenswrapper[4870]: I1014 08:57:14.388600 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7a69635-d246-4920-b96b-6b232fa4e813-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "f7a69635-d246-4920-b96b-6b232fa4e813" (UID: "f7a69635-d246-4920-b96b-6b232fa4e813"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:57:14 crc kubenswrapper[4870]: I1014 08:57:14.404574 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7a69635-d246-4920-b96b-6b232fa4e813-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f7a69635-d246-4920-b96b-6b232fa4e813" (UID: "f7a69635-d246-4920-b96b-6b232fa4e813"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:57:14 crc kubenswrapper[4870]: I1014 08:57:14.436723 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7a69635-d246-4920-b96b-6b232fa4e813-config-data" (OuterVolumeSpecName: "config-data") pod "f7a69635-d246-4920-b96b-6b232fa4e813" (UID: "f7a69635-d246-4920-b96b-6b232fa4e813"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:57:14 crc kubenswrapper[4870]: I1014 08:57:14.484643 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k7zxc\" (UniqueName: \"kubernetes.io/projected/f7a69635-d246-4920-b96b-6b232fa4e813-kube-api-access-k7zxc\") on node \"crc\" DevicePath \"\"" Oct 14 08:57:14 crc kubenswrapper[4870]: I1014 08:57:14.484721 4870 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7a69635-d246-4920-b96b-6b232fa4e813-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 08:57:14 crc kubenswrapper[4870]: I1014 08:57:14.484753 4870 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7a69635-d246-4920-b96b-6b232fa4e813-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 08:57:14 crc kubenswrapper[4870]: I1014 08:57:14.484819 4870 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f7a69635-d246-4920-b96b-6b232fa4e813-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 08:57:14 crc kubenswrapper[4870]: I1014 08:57:14.788745 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-4p4kx" event={"ID":"f7a69635-d246-4920-b96b-6b232fa4e813","Type":"ContainerDied","Data":"15524e006051bc007ba24f6a1734e8ebb7b8453fd733127cfc09d974a4f1a363"} Oct 14 08:57:14 crc kubenswrapper[4870]: I1014 08:57:14.788806 4870 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="15524e006051bc007ba24f6a1734e8ebb7b8453fd733127cfc09d974a4f1a363" Oct 14 08:57:14 crc kubenswrapper[4870]: I1014 08:57:14.788879 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-4p4kx" Oct 14 08:57:15 crc kubenswrapper[4870]: I1014 08:57:15.100728 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 14 08:57:15 crc kubenswrapper[4870]: E1014 08:57:15.101104 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7a69635-d246-4920-b96b-6b232fa4e813" containerName="glance-db-sync" Oct 14 08:57:15 crc kubenswrapper[4870]: I1014 08:57:15.101126 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7a69635-d246-4920-b96b-6b232fa4e813" containerName="glance-db-sync" Oct 14 08:57:15 crc kubenswrapper[4870]: I1014 08:57:15.101313 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="f7a69635-d246-4920-b96b-6b232fa4e813" containerName="glance-db-sync" Oct 14 08:57:15 crc kubenswrapper[4870]: I1014 08:57:15.104056 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 14 08:57:15 crc kubenswrapper[4870]: I1014 08:57:15.111045 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Oct 14 08:57:15 crc kubenswrapper[4870]: I1014 08:57:15.111266 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 14 08:57:15 crc kubenswrapper[4870]: I1014 08:57:15.111418 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-f98hg" Oct 14 08:57:15 crc kubenswrapper[4870]: I1014 08:57:15.111605 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Oct 14 08:57:15 crc kubenswrapper[4870]: I1014 08:57:15.120304 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 14 08:57:15 crc kubenswrapper[4870]: I1014 08:57:15.189782 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6695878787-8ngcl"] Oct 14 08:57:15 crc kubenswrapper[4870]: I1014 08:57:15.191564 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6695878787-8ngcl" Oct 14 08:57:15 crc kubenswrapper[4870]: I1014 08:57:15.200369 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/60dbc30c-de4d-4574-b658-e5d43fc32288-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"60dbc30c-de4d-4574-b658-e5d43fc32288\") " pod="openstack/glance-default-external-api-0" Oct 14 08:57:15 crc kubenswrapper[4870]: I1014 08:57:15.200414 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/60dbc30c-de4d-4574-b658-e5d43fc32288-scripts\") pod \"glance-default-external-api-0\" (UID: \"60dbc30c-de4d-4574-b658-e5d43fc32288\") " pod="openstack/glance-default-external-api-0" Oct 14 08:57:15 crc kubenswrapper[4870]: I1014 08:57:15.200485 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60dbc30c-de4d-4574-b658-e5d43fc32288-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"60dbc30c-de4d-4574-b658-e5d43fc32288\") " pod="openstack/glance-default-external-api-0" Oct 14 08:57:15 crc kubenswrapper[4870]: I1014 08:57:15.200516 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/60dbc30c-de4d-4574-b658-e5d43fc32288-logs\") pod \"glance-default-external-api-0\" (UID: \"60dbc30c-de4d-4574-b658-e5d43fc32288\") " pod="openstack/glance-default-external-api-0" Oct 14 08:57:15 crc kubenswrapper[4870]: I1014 08:57:15.200533 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/60dbc30c-de4d-4574-b658-e5d43fc32288-ceph\") pod \"glance-default-external-api-0\" (UID: \"60dbc30c-de4d-4574-b658-e5d43fc32288\") " pod="openstack/glance-default-external-api-0" Oct 14 08:57:15 crc kubenswrapper[4870]: I1014 08:57:15.200552 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tzdg9\" (UniqueName: \"kubernetes.io/projected/60dbc30c-de4d-4574-b658-e5d43fc32288-kube-api-access-tzdg9\") pod \"glance-default-external-api-0\" (UID: \"60dbc30c-de4d-4574-b658-e5d43fc32288\") " pod="openstack/glance-default-external-api-0" Oct 14 08:57:15 crc kubenswrapper[4870]: I1014 08:57:15.200584 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/60dbc30c-de4d-4574-b658-e5d43fc32288-config-data\") pod \"glance-default-external-api-0\" (UID: \"60dbc30c-de4d-4574-b658-e5d43fc32288\") " pod="openstack/glance-default-external-api-0" Oct 14 08:57:15 crc kubenswrapper[4870]: I1014 08:57:15.216624 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6695878787-8ngcl"] Oct 14 08:57:15 crc kubenswrapper[4870]: I1014 08:57:15.302989 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60dbc30c-de4d-4574-b658-e5d43fc32288-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"60dbc30c-de4d-4574-b658-e5d43fc32288\") " pod="openstack/glance-default-external-api-0" Oct 14 08:57:15 crc kubenswrapper[4870]: I1014 08:57:15.303058 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/60dbc30c-de4d-4574-b658-e5d43fc32288-logs\") pod \"glance-default-external-api-0\" (UID: \"60dbc30c-de4d-4574-b658-e5d43fc32288\") " pod="openstack/glance-default-external-api-0" Oct 14 08:57:15 crc kubenswrapper[4870]: I1014 08:57:15.303088 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/60dbc30c-de4d-4574-b658-e5d43fc32288-ceph\") pod \"glance-default-external-api-0\" (UID: \"60dbc30c-de4d-4574-b658-e5d43fc32288\") " pod="openstack/glance-default-external-api-0" Oct 14 08:57:15 crc kubenswrapper[4870]: I1014 08:57:15.303115 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nhmld\" (UniqueName: \"kubernetes.io/projected/37b9bd38-e85a-452c-97f6-67d5d8d3958f-kube-api-access-nhmld\") pod \"dnsmasq-dns-6695878787-8ngcl\" (UID: \"37b9bd38-e85a-452c-97f6-67d5d8d3958f\") " pod="openstack/dnsmasq-dns-6695878787-8ngcl" Oct 14 08:57:15 crc kubenswrapper[4870]: I1014 08:57:15.303141 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tzdg9\" (UniqueName: \"kubernetes.io/projected/60dbc30c-de4d-4574-b658-e5d43fc32288-kube-api-access-tzdg9\") pod \"glance-default-external-api-0\" (UID: \"60dbc30c-de4d-4574-b658-e5d43fc32288\") " pod="openstack/glance-default-external-api-0" Oct 14 08:57:15 crc kubenswrapper[4870]: I1014 08:57:15.303176 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/60dbc30c-de4d-4574-b658-e5d43fc32288-config-data\") pod \"glance-default-external-api-0\" (UID: \"60dbc30c-de4d-4574-b658-e5d43fc32288\") " pod="openstack/glance-default-external-api-0" Oct 14 08:57:15 crc kubenswrapper[4870]: I1014 08:57:15.303248 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/37b9bd38-e85a-452c-97f6-67d5d8d3958f-dns-svc\") pod \"dnsmasq-dns-6695878787-8ngcl\" (UID: \"37b9bd38-e85a-452c-97f6-67d5d8d3958f\") " pod="openstack/dnsmasq-dns-6695878787-8ngcl" Oct 14 08:57:15 crc kubenswrapper[4870]: I1014 08:57:15.303289 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/60dbc30c-de4d-4574-b658-e5d43fc32288-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"60dbc30c-de4d-4574-b658-e5d43fc32288\") " pod="openstack/glance-default-external-api-0" Oct 14 08:57:15 crc kubenswrapper[4870]: I1014 08:57:15.303312 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/60dbc30c-de4d-4574-b658-e5d43fc32288-scripts\") pod \"glance-default-external-api-0\" (UID: \"60dbc30c-de4d-4574-b658-e5d43fc32288\") " pod="openstack/glance-default-external-api-0" Oct 14 08:57:15 crc kubenswrapper[4870]: I1014 08:57:15.303335 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/37b9bd38-e85a-452c-97f6-67d5d8d3958f-ovsdbserver-nb\") pod \"dnsmasq-dns-6695878787-8ngcl\" (UID: \"37b9bd38-e85a-452c-97f6-67d5d8d3958f\") " pod="openstack/dnsmasq-dns-6695878787-8ngcl" Oct 14 08:57:15 crc kubenswrapper[4870]: I1014 08:57:15.303368 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/37b9bd38-e85a-452c-97f6-67d5d8d3958f-ovsdbserver-sb\") pod \"dnsmasq-dns-6695878787-8ngcl\" (UID: \"37b9bd38-e85a-452c-97f6-67d5d8d3958f\") " pod="openstack/dnsmasq-dns-6695878787-8ngcl" Oct 14 08:57:15 crc kubenswrapper[4870]: I1014 08:57:15.303402 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/37b9bd38-e85a-452c-97f6-67d5d8d3958f-config\") pod \"dnsmasq-dns-6695878787-8ngcl\" (UID: \"37b9bd38-e85a-452c-97f6-67d5d8d3958f\") " pod="openstack/dnsmasq-dns-6695878787-8ngcl" Oct 14 08:57:15 crc kubenswrapper[4870]: I1014 08:57:15.304246 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/60dbc30c-de4d-4574-b658-e5d43fc32288-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"60dbc30c-de4d-4574-b658-e5d43fc32288\") " pod="openstack/glance-default-external-api-0" Oct 14 08:57:15 crc kubenswrapper[4870]: I1014 08:57:15.304548 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/60dbc30c-de4d-4574-b658-e5d43fc32288-logs\") pod \"glance-default-external-api-0\" (UID: \"60dbc30c-de4d-4574-b658-e5d43fc32288\") " pod="openstack/glance-default-external-api-0" Oct 14 08:57:15 crc kubenswrapper[4870]: I1014 08:57:15.308334 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/60dbc30c-de4d-4574-b658-e5d43fc32288-config-data\") pod \"glance-default-external-api-0\" (UID: \"60dbc30c-de4d-4574-b658-e5d43fc32288\") " pod="openstack/glance-default-external-api-0" Oct 14 08:57:15 crc kubenswrapper[4870]: I1014 08:57:15.319994 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/60dbc30c-de4d-4574-b658-e5d43fc32288-scripts\") pod \"glance-default-external-api-0\" (UID: \"60dbc30c-de4d-4574-b658-e5d43fc32288\") " pod="openstack/glance-default-external-api-0" Oct 14 08:57:15 crc kubenswrapper[4870]: I1014 08:57:15.322987 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/60dbc30c-de4d-4574-b658-e5d43fc32288-ceph\") pod \"glance-default-external-api-0\" (UID: \"60dbc30c-de4d-4574-b658-e5d43fc32288\") " pod="openstack/glance-default-external-api-0" Oct 14 08:57:15 crc kubenswrapper[4870]: I1014 08:57:15.327127 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tzdg9\" (UniqueName: \"kubernetes.io/projected/60dbc30c-de4d-4574-b658-e5d43fc32288-kube-api-access-tzdg9\") pod \"glance-default-external-api-0\" (UID: \"60dbc30c-de4d-4574-b658-e5d43fc32288\") " pod="openstack/glance-default-external-api-0" Oct 14 08:57:15 crc kubenswrapper[4870]: I1014 08:57:15.328086 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60dbc30c-de4d-4574-b658-e5d43fc32288-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"60dbc30c-de4d-4574-b658-e5d43fc32288\") " pod="openstack/glance-default-external-api-0" Oct 14 08:57:15 crc kubenswrapper[4870]: I1014 08:57:15.384850 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 14 08:57:15 crc kubenswrapper[4870]: I1014 08:57:15.388539 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 14 08:57:15 crc kubenswrapper[4870]: I1014 08:57:15.390689 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 14 08:57:15 crc kubenswrapper[4870]: I1014 08:57:15.427129 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 14 08:57:15 crc kubenswrapper[4870]: I1014 08:57:15.435506 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/37b9bd38-e85a-452c-97f6-67d5d8d3958f-dns-svc\") pod \"dnsmasq-dns-6695878787-8ngcl\" (UID: \"37b9bd38-e85a-452c-97f6-67d5d8d3958f\") " pod="openstack/dnsmasq-dns-6695878787-8ngcl" Oct 14 08:57:15 crc kubenswrapper[4870]: I1014 08:57:15.435587 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f94fd4d7-586c-4efd-8329-2b3667fd8d58-config-data\") pod \"glance-default-internal-api-0\" (UID: \"f94fd4d7-586c-4efd-8329-2b3667fd8d58\") " pod="openstack/glance-default-internal-api-0" Oct 14 08:57:15 crc kubenswrapper[4870]: I1014 08:57:15.435595 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 14 08:57:15 crc kubenswrapper[4870]: I1014 08:57:15.435631 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/37b9bd38-e85a-452c-97f6-67d5d8d3958f-ovsdbserver-nb\") pod \"dnsmasq-dns-6695878787-8ngcl\" (UID: \"37b9bd38-e85a-452c-97f6-67d5d8d3958f\") " pod="openstack/dnsmasq-dns-6695878787-8ngcl" Oct 14 08:57:15 crc kubenswrapper[4870]: I1014 08:57:15.435663 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/37b9bd38-e85a-452c-97f6-67d5d8d3958f-ovsdbserver-sb\") pod \"dnsmasq-dns-6695878787-8ngcl\" (UID: \"37b9bd38-e85a-452c-97f6-67d5d8d3958f\") " pod="openstack/dnsmasq-dns-6695878787-8ngcl" Oct 14 08:57:15 crc kubenswrapper[4870]: I1014 08:57:15.435691 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/f94fd4d7-586c-4efd-8329-2b3667fd8d58-ceph\") pod \"glance-default-internal-api-0\" (UID: \"f94fd4d7-586c-4efd-8329-2b3667fd8d58\") " pod="openstack/glance-default-internal-api-0" Oct 14 08:57:15 crc kubenswrapper[4870]: I1014 08:57:15.435721 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/37b9bd38-e85a-452c-97f6-67d5d8d3958f-config\") pod \"dnsmasq-dns-6695878787-8ngcl\" (UID: \"37b9bd38-e85a-452c-97f6-67d5d8d3958f\") " pod="openstack/dnsmasq-dns-6695878787-8ngcl" Oct 14 08:57:15 crc kubenswrapper[4870]: I1014 08:57:15.435747 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f94fd4d7-586c-4efd-8329-2b3667fd8d58-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"f94fd4d7-586c-4efd-8329-2b3667fd8d58\") " pod="openstack/glance-default-internal-api-0" Oct 14 08:57:15 crc kubenswrapper[4870]: I1014 08:57:15.435783 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f94fd4d7-586c-4efd-8329-2b3667fd8d58-scripts\") pod \"glance-default-internal-api-0\" (UID: \"f94fd4d7-586c-4efd-8329-2b3667fd8d58\") " pod="openstack/glance-default-internal-api-0" Oct 14 08:57:15 crc kubenswrapper[4870]: I1014 08:57:15.435810 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f94fd4d7-586c-4efd-8329-2b3667fd8d58-logs\") pod \"glance-default-internal-api-0\" (UID: \"f94fd4d7-586c-4efd-8329-2b3667fd8d58\") " pod="openstack/glance-default-internal-api-0" Oct 14 08:57:15 crc kubenswrapper[4870]: I1014 08:57:15.435848 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f94fd4d7-586c-4efd-8329-2b3667fd8d58-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"f94fd4d7-586c-4efd-8329-2b3667fd8d58\") " pod="openstack/glance-default-internal-api-0" Oct 14 08:57:15 crc kubenswrapper[4870]: I1014 08:57:15.435873 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sxt7b\" (UniqueName: \"kubernetes.io/projected/f94fd4d7-586c-4efd-8329-2b3667fd8d58-kube-api-access-sxt7b\") pod \"glance-default-internal-api-0\" (UID: \"f94fd4d7-586c-4efd-8329-2b3667fd8d58\") " pod="openstack/glance-default-internal-api-0" Oct 14 08:57:15 crc kubenswrapper[4870]: I1014 08:57:15.435919 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nhmld\" (UniqueName: \"kubernetes.io/projected/37b9bd38-e85a-452c-97f6-67d5d8d3958f-kube-api-access-nhmld\") pod \"dnsmasq-dns-6695878787-8ngcl\" (UID: \"37b9bd38-e85a-452c-97f6-67d5d8d3958f\") " pod="openstack/dnsmasq-dns-6695878787-8ngcl" Oct 14 08:57:15 crc kubenswrapper[4870]: I1014 08:57:15.444451 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/37b9bd38-e85a-452c-97f6-67d5d8d3958f-dns-svc\") pod \"dnsmasq-dns-6695878787-8ngcl\" (UID: \"37b9bd38-e85a-452c-97f6-67d5d8d3958f\") " pod="openstack/dnsmasq-dns-6695878787-8ngcl" Oct 14 08:57:15 crc kubenswrapper[4870]: I1014 08:57:15.444738 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/37b9bd38-e85a-452c-97f6-67d5d8d3958f-config\") pod \"dnsmasq-dns-6695878787-8ngcl\" (UID: \"37b9bd38-e85a-452c-97f6-67d5d8d3958f\") " pod="openstack/dnsmasq-dns-6695878787-8ngcl" Oct 14 08:57:15 crc kubenswrapper[4870]: I1014 08:57:15.444837 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/37b9bd38-e85a-452c-97f6-67d5d8d3958f-ovsdbserver-sb\") pod \"dnsmasq-dns-6695878787-8ngcl\" (UID: \"37b9bd38-e85a-452c-97f6-67d5d8d3958f\") " pod="openstack/dnsmasq-dns-6695878787-8ngcl" Oct 14 08:57:15 crc kubenswrapper[4870]: I1014 08:57:15.445025 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/37b9bd38-e85a-452c-97f6-67d5d8d3958f-ovsdbserver-nb\") pod \"dnsmasq-dns-6695878787-8ngcl\" (UID: \"37b9bd38-e85a-452c-97f6-67d5d8d3958f\") " pod="openstack/dnsmasq-dns-6695878787-8ngcl" Oct 14 08:57:15 crc kubenswrapper[4870]: I1014 08:57:15.482263 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nhmld\" (UniqueName: \"kubernetes.io/projected/37b9bd38-e85a-452c-97f6-67d5d8d3958f-kube-api-access-nhmld\") pod \"dnsmasq-dns-6695878787-8ngcl\" (UID: \"37b9bd38-e85a-452c-97f6-67d5d8d3958f\") " pod="openstack/dnsmasq-dns-6695878787-8ngcl" Oct 14 08:57:15 crc kubenswrapper[4870]: I1014 08:57:15.516591 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6695878787-8ngcl" Oct 14 08:57:15 crc kubenswrapper[4870]: I1014 08:57:15.538313 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f94fd4d7-586c-4efd-8329-2b3667fd8d58-config-data\") pod \"glance-default-internal-api-0\" (UID: \"f94fd4d7-586c-4efd-8329-2b3667fd8d58\") " pod="openstack/glance-default-internal-api-0" Oct 14 08:57:15 crc kubenswrapper[4870]: I1014 08:57:15.538379 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/f94fd4d7-586c-4efd-8329-2b3667fd8d58-ceph\") pod \"glance-default-internal-api-0\" (UID: \"f94fd4d7-586c-4efd-8329-2b3667fd8d58\") " pod="openstack/glance-default-internal-api-0" Oct 14 08:57:15 crc kubenswrapper[4870]: I1014 08:57:15.538406 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f94fd4d7-586c-4efd-8329-2b3667fd8d58-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"f94fd4d7-586c-4efd-8329-2b3667fd8d58\") " pod="openstack/glance-default-internal-api-0" Oct 14 08:57:15 crc kubenswrapper[4870]: I1014 08:57:15.538432 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f94fd4d7-586c-4efd-8329-2b3667fd8d58-scripts\") pod \"glance-default-internal-api-0\" (UID: \"f94fd4d7-586c-4efd-8329-2b3667fd8d58\") " pod="openstack/glance-default-internal-api-0" Oct 14 08:57:15 crc kubenswrapper[4870]: I1014 08:57:15.538454 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f94fd4d7-586c-4efd-8329-2b3667fd8d58-logs\") pod \"glance-default-internal-api-0\" (UID: \"f94fd4d7-586c-4efd-8329-2b3667fd8d58\") " pod="openstack/glance-default-internal-api-0" Oct 14 08:57:15 crc kubenswrapper[4870]: I1014 08:57:15.541399 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f94fd4d7-586c-4efd-8329-2b3667fd8d58-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"f94fd4d7-586c-4efd-8329-2b3667fd8d58\") " pod="openstack/glance-default-internal-api-0" Oct 14 08:57:15 crc kubenswrapper[4870]: I1014 08:57:15.541440 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sxt7b\" (UniqueName: \"kubernetes.io/projected/f94fd4d7-586c-4efd-8329-2b3667fd8d58-kube-api-access-sxt7b\") pod \"glance-default-internal-api-0\" (UID: \"f94fd4d7-586c-4efd-8329-2b3667fd8d58\") " pod="openstack/glance-default-internal-api-0" Oct 14 08:57:15 crc kubenswrapper[4870]: I1014 08:57:15.544157 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f94fd4d7-586c-4efd-8329-2b3667fd8d58-logs\") pod \"glance-default-internal-api-0\" (UID: \"f94fd4d7-586c-4efd-8329-2b3667fd8d58\") " pod="openstack/glance-default-internal-api-0" Oct 14 08:57:15 crc kubenswrapper[4870]: I1014 08:57:15.544240 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f94fd4d7-586c-4efd-8329-2b3667fd8d58-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"f94fd4d7-586c-4efd-8329-2b3667fd8d58\") " pod="openstack/glance-default-internal-api-0" Oct 14 08:57:15 crc kubenswrapper[4870]: I1014 08:57:15.545232 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/f94fd4d7-586c-4efd-8329-2b3667fd8d58-ceph\") pod \"glance-default-internal-api-0\" (UID: \"f94fd4d7-586c-4efd-8329-2b3667fd8d58\") " pod="openstack/glance-default-internal-api-0" Oct 14 08:57:15 crc kubenswrapper[4870]: I1014 08:57:15.547283 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f94fd4d7-586c-4efd-8329-2b3667fd8d58-scripts\") pod \"glance-default-internal-api-0\" (UID: \"f94fd4d7-586c-4efd-8329-2b3667fd8d58\") " pod="openstack/glance-default-internal-api-0" Oct 14 08:57:15 crc kubenswrapper[4870]: I1014 08:57:15.547839 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f94fd4d7-586c-4efd-8329-2b3667fd8d58-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"f94fd4d7-586c-4efd-8329-2b3667fd8d58\") " pod="openstack/glance-default-internal-api-0" Oct 14 08:57:15 crc kubenswrapper[4870]: I1014 08:57:15.547843 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f94fd4d7-586c-4efd-8329-2b3667fd8d58-config-data\") pod \"glance-default-internal-api-0\" (UID: \"f94fd4d7-586c-4efd-8329-2b3667fd8d58\") " pod="openstack/glance-default-internal-api-0" Oct 14 08:57:15 crc kubenswrapper[4870]: I1014 08:57:15.570807 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sxt7b\" (UniqueName: \"kubernetes.io/projected/f94fd4d7-586c-4efd-8329-2b3667fd8d58-kube-api-access-sxt7b\") pod \"glance-default-internal-api-0\" (UID: \"f94fd4d7-586c-4efd-8329-2b3667fd8d58\") " pod="openstack/glance-default-internal-api-0" Oct 14 08:57:15 crc kubenswrapper[4870]: I1014 08:57:15.752993 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 14 08:57:16 crc kubenswrapper[4870]: W1014 08:57:16.075681 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod37b9bd38_e85a_452c_97f6_67d5d8d3958f.slice/crio-c49a35593276d0c662a20854447b8fcbf40fddf54f7566401f170f0024820608 WatchSource:0}: Error finding container c49a35593276d0c662a20854447b8fcbf40fddf54f7566401f170f0024820608: Status 404 returned error can't find the container with id c49a35593276d0c662a20854447b8fcbf40fddf54f7566401f170f0024820608 Oct 14 08:57:16 crc kubenswrapper[4870]: I1014 08:57:16.083545 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6695878787-8ngcl"] Oct 14 08:57:16 crc kubenswrapper[4870]: I1014 08:57:16.090419 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 14 08:57:16 crc kubenswrapper[4870]: I1014 08:57:16.425566 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 14 08:57:16 crc kubenswrapper[4870]: I1014 08:57:16.470273 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 14 08:57:16 crc kubenswrapper[4870]: I1014 08:57:16.821427 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"60dbc30c-de4d-4574-b658-e5d43fc32288","Type":"ContainerStarted","Data":"1cfc46d02e2a200441f75ab42bdbc56afce8c182aa35abbf6490c152c62996f2"} Oct 14 08:57:16 crc kubenswrapper[4870]: I1014 08:57:16.821763 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"60dbc30c-de4d-4574-b658-e5d43fc32288","Type":"ContainerStarted","Data":"b274cfcdb5293c8225020e85e98bb265bc973be4efe90109c6f318e49d67945a"} Oct 14 08:57:16 crc kubenswrapper[4870]: I1014 08:57:16.825157 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f94fd4d7-586c-4efd-8329-2b3667fd8d58","Type":"ContainerStarted","Data":"6d003e5e9ad8663a2a36de730e70c5ce9ae9f17f809c449cf5a5e44429256edf"} Oct 14 08:57:16 crc kubenswrapper[4870]: I1014 08:57:16.826738 4870 generic.go:334] "Generic (PLEG): container finished" podID="37b9bd38-e85a-452c-97f6-67d5d8d3958f" containerID="9a6b519117c6b878a44d820c3e19189970be4558a4f74d53ac33c5ecca6fbee4" exitCode=0 Oct 14 08:57:16 crc kubenswrapper[4870]: I1014 08:57:16.826784 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6695878787-8ngcl" event={"ID":"37b9bd38-e85a-452c-97f6-67d5d8d3958f","Type":"ContainerDied","Data":"9a6b519117c6b878a44d820c3e19189970be4558a4f74d53ac33c5ecca6fbee4"} Oct 14 08:57:16 crc kubenswrapper[4870]: I1014 08:57:16.826807 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6695878787-8ngcl" event={"ID":"37b9bd38-e85a-452c-97f6-67d5d8d3958f","Type":"ContainerStarted","Data":"c49a35593276d0c662a20854447b8fcbf40fddf54f7566401f170f0024820608"} Oct 14 08:57:17 crc kubenswrapper[4870]: I1014 08:57:17.836335 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6695878787-8ngcl" event={"ID":"37b9bd38-e85a-452c-97f6-67d5d8d3958f","Type":"ContainerStarted","Data":"436e31f0c5e68792f3f32f7723f0da2b5d815ce570cdc562c44058b35d0bc9df"} Oct 14 08:57:17 crc kubenswrapper[4870]: I1014 08:57:17.836868 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6695878787-8ngcl" Oct 14 08:57:17 crc kubenswrapper[4870]: I1014 08:57:17.838048 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"60dbc30c-de4d-4574-b658-e5d43fc32288","Type":"ContainerStarted","Data":"c818cc0257089bc638ae8a03602c2c4939e1ec5115ea406f0fa7b0a4a6172e12"} Oct 14 08:57:17 crc kubenswrapper[4870]: I1014 08:57:17.838138 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="60dbc30c-de4d-4574-b658-e5d43fc32288" containerName="glance-log" containerID="cri-o://1cfc46d02e2a200441f75ab42bdbc56afce8c182aa35abbf6490c152c62996f2" gracePeriod=30 Oct 14 08:57:17 crc kubenswrapper[4870]: I1014 08:57:17.838261 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="60dbc30c-de4d-4574-b658-e5d43fc32288" containerName="glance-httpd" containerID="cri-o://c818cc0257089bc638ae8a03602c2c4939e1ec5115ea406f0fa7b0a4a6172e12" gracePeriod=30 Oct 14 08:57:17 crc kubenswrapper[4870]: I1014 08:57:17.841299 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f94fd4d7-586c-4efd-8329-2b3667fd8d58","Type":"ContainerStarted","Data":"a36a0cf52fe028fb39868934fe28934070222f110f220e1dda8bf75588ca0950"} Oct 14 08:57:17 crc kubenswrapper[4870]: I1014 08:57:17.841329 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f94fd4d7-586c-4efd-8329-2b3667fd8d58","Type":"ContainerStarted","Data":"e80efe033ae130b7eee883c22976e3984670f7779d8adbc52b6bef2d3839ad2c"} Oct 14 08:57:17 crc kubenswrapper[4870]: I1014 08:57:17.861249 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6695878787-8ngcl" podStartSLOduration=2.861231691 podStartE2EDuration="2.861231691s" podCreationTimestamp="2025-10-14 08:57:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 08:57:17.85993278 +0000 UTC m=+6973.557293151" watchObservedRunningTime="2025-10-14 08:57:17.861231691 +0000 UTC m=+6973.558592082" Oct 14 08:57:17 crc kubenswrapper[4870]: I1014 08:57:17.881790 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=2.881774355 podStartE2EDuration="2.881774355s" podCreationTimestamp="2025-10-14 08:57:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 08:57:17.877034179 +0000 UTC m=+6973.574394550" watchObservedRunningTime="2025-10-14 08:57:17.881774355 +0000 UTC m=+6973.579134726" Oct 14 08:57:17 crc kubenswrapper[4870]: I1014 08:57:17.902837 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=2.902818091 podStartE2EDuration="2.902818091s" podCreationTimestamp="2025-10-14 08:57:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 08:57:17.895871261 +0000 UTC m=+6973.593231632" watchObservedRunningTime="2025-10-14 08:57:17.902818091 +0000 UTC m=+6973.600178462" Oct 14 08:57:18 crc kubenswrapper[4870]: I1014 08:57:18.416964 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 14 08:57:18 crc kubenswrapper[4870]: I1014 08:57:18.507323 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/60dbc30c-de4d-4574-b658-e5d43fc32288-logs\") pod \"60dbc30c-de4d-4574-b658-e5d43fc32288\" (UID: \"60dbc30c-de4d-4574-b658-e5d43fc32288\") " Oct 14 08:57:18 crc kubenswrapper[4870]: I1014 08:57:18.507434 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60dbc30c-de4d-4574-b658-e5d43fc32288-combined-ca-bundle\") pod \"60dbc30c-de4d-4574-b658-e5d43fc32288\" (UID: \"60dbc30c-de4d-4574-b658-e5d43fc32288\") " Oct 14 08:57:18 crc kubenswrapper[4870]: I1014 08:57:18.507533 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/60dbc30c-de4d-4574-b658-e5d43fc32288-config-data\") pod \"60dbc30c-de4d-4574-b658-e5d43fc32288\" (UID: \"60dbc30c-de4d-4574-b658-e5d43fc32288\") " Oct 14 08:57:18 crc kubenswrapper[4870]: I1014 08:57:18.507612 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tzdg9\" (UniqueName: \"kubernetes.io/projected/60dbc30c-de4d-4574-b658-e5d43fc32288-kube-api-access-tzdg9\") pod \"60dbc30c-de4d-4574-b658-e5d43fc32288\" (UID: \"60dbc30c-de4d-4574-b658-e5d43fc32288\") " Oct 14 08:57:18 crc kubenswrapper[4870]: I1014 08:57:18.507644 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/60dbc30c-de4d-4574-b658-e5d43fc32288-httpd-run\") pod \"60dbc30c-de4d-4574-b658-e5d43fc32288\" (UID: \"60dbc30c-de4d-4574-b658-e5d43fc32288\") " Oct 14 08:57:18 crc kubenswrapper[4870]: I1014 08:57:18.507660 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/60dbc30c-de4d-4574-b658-e5d43fc32288-ceph\") pod \"60dbc30c-de4d-4574-b658-e5d43fc32288\" (UID: \"60dbc30c-de4d-4574-b658-e5d43fc32288\") " Oct 14 08:57:18 crc kubenswrapper[4870]: I1014 08:57:18.507749 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/60dbc30c-de4d-4574-b658-e5d43fc32288-scripts\") pod \"60dbc30c-de4d-4574-b658-e5d43fc32288\" (UID: \"60dbc30c-de4d-4574-b658-e5d43fc32288\") " Oct 14 08:57:18 crc kubenswrapper[4870]: I1014 08:57:18.507920 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/60dbc30c-de4d-4574-b658-e5d43fc32288-logs" (OuterVolumeSpecName: "logs") pod "60dbc30c-de4d-4574-b658-e5d43fc32288" (UID: "60dbc30c-de4d-4574-b658-e5d43fc32288"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:57:18 crc kubenswrapper[4870]: I1014 08:57:18.508141 4870 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/60dbc30c-de4d-4574-b658-e5d43fc32288-logs\") on node \"crc\" DevicePath \"\"" Oct 14 08:57:18 crc kubenswrapper[4870]: I1014 08:57:18.508617 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/60dbc30c-de4d-4574-b658-e5d43fc32288-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "60dbc30c-de4d-4574-b658-e5d43fc32288" (UID: "60dbc30c-de4d-4574-b658-e5d43fc32288"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:57:18 crc kubenswrapper[4870]: I1014 08:57:18.512881 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/60dbc30c-de4d-4574-b658-e5d43fc32288-ceph" (OuterVolumeSpecName: "ceph") pod "60dbc30c-de4d-4574-b658-e5d43fc32288" (UID: "60dbc30c-de4d-4574-b658-e5d43fc32288"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:57:18 crc kubenswrapper[4870]: I1014 08:57:18.513199 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/60dbc30c-de4d-4574-b658-e5d43fc32288-scripts" (OuterVolumeSpecName: "scripts") pod "60dbc30c-de4d-4574-b658-e5d43fc32288" (UID: "60dbc30c-de4d-4574-b658-e5d43fc32288"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:57:18 crc kubenswrapper[4870]: I1014 08:57:18.514901 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/60dbc30c-de4d-4574-b658-e5d43fc32288-kube-api-access-tzdg9" (OuterVolumeSpecName: "kube-api-access-tzdg9") pod "60dbc30c-de4d-4574-b658-e5d43fc32288" (UID: "60dbc30c-de4d-4574-b658-e5d43fc32288"). InnerVolumeSpecName "kube-api-access-tzdg9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:57:18 crc kubenswrapper[4870]: I1014 08:57:18.531433 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/60dbc30c-de4d-4574-b658-e5d43fc32288-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "60dbc30c-de4d-4574-b658-e5d43fc32288" (UID: "60dbc30c-de4d-4574-b658-e5d43fc32288"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:57:18 crc kubenswrapper[4870]: I1014 08:57:18.565938 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/60dbc30c-de4d-4574-b658-e5d43fc32288-config-data" (OuterVolumeSpecName: "config-data") pod "60dbc30c-de4d-4574-b658-e5d43fc32288" (UID: "60dbc30c-de4d-4574-b658-e5d43fc32288"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:57:18 crc kubenswrapper[4870]: I1014 08:57:18.610116 4870 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/60dbc30c-de4d-4574-b658-e5d43fc32288-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 08:57:18 crc kubenswrapper[4870]: I1014 08:57:18.610157 4870 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60dbc30c-de4d-4574-b658-e5d43fc32288-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 08:57:18 crc kubenswrapper[4870]: I1014 08:57:18.610177 4870 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/60dbc30c-de4d-4574-b658-e5d43fc32288-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 08:57:18 crc kubenswrapper[4870]: I1014 08:57:18.610188 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tzdg9\" (UniqueName: \"kubernetes.io/projected/60dbc30c-de4d-4574-b658-e5d43fc32288-kube-api-access-tzdg9\") on node \"crc\" DevicePath \"\"" Oct 14 08:57:18 crc kubenswrapper[4870]: I1014 08:57:18.610198 4870 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/60dbc30c-de4d-4574-b658-e5d43fc32288-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 14 08:57:18 crc kubenswrapper[4870]: I1014 08:57:18.610210 4870 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/60dbc30c-de4d-4574-b658-e5d43fc32288-ceph\") on node \"crc\" DevicePath \"\"" Oct 14 08:57:18 crc kubenswrapper[4870]: I1014 08:57:18.731208 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 14 08:57:18 crc kubenswrapper[4870]: I1014 08:57:18.852871 4870 generic.go:334] "Generic (PLEG): container finished" podID="60dbc30c-de4d-4574-b658-e5d43fc32288" containerID="c818cc0257089bc638ae8a03602c2c4939e1ec5115ea406f0fa7b0a4a6172e12" exitCode=0 Oct 14 08:57:18 crc kubenswrapper[4870]: I1014 08:57:18.852918 4870 generic.go:334] "Generic (PLEG): container finished" podID="60dbc30c-de4d-4574-b658-e5d43fc32288" containerID="1cfc46d02e2a200441f75ab42bdbc56afce8c182aa35abbf6490c152c62996f2" exitCode=143 Oct 14 08:57:18 crc kubenswrapper[4870]: I1014 08:57:18.853852 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 14 08:57:18 crc kubenswrapper[4870]: I1014 08:57:18.858082 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"60dbc30c-de4d-4574-b658-e5d43fc32288","Type":"ContainerDied","Data":"c818cc0257089bc638ae8a03602c2c4939e1ec5115ea406f0fa7b0a4a6172e12"} Oct 14 08:57:18 crc kubenswrapper[4870]: I1014 08:57:18.858140 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"60dbc30c-de4d-4574-b658-e5d43fc32288","Type":"ContainerDied","Data":"1cfc46d02e2a200441f75ab42bdbc56afce8c182aa35abbf6490c152c62996f2"} Oct 14 08:57:18 crc kubenswrapper[4870]: I1014 08:57:18.858153 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"60dbc30c-de4d-4574-b658-e5d43fc32288","Type":"ContainerDied","Data":"b274cfcdb5293c8225020e85e98bb265bc973be4efe90109c6f318e49d67945a"} Oct 14 08:57:18 crc kubenswrapper[4870]: I1014 08:57:18.858171 4870 scope.go:117] "RemoveContainer" containerID="c818cc0257089bc638ae8a03602c2c4939e1ec5115ea406f0fa7b0a4a6172e12" Oct 14 08:57:18 crc kubenswrapper[4870]: I1014 08:57:18.907403 4870 scope.go:117] "RemoveContainer" containerID="1cfc46d02e2a200441f75ab42bdbc56afce8c182aa35abbf6490c152c62996f2" Oct 14 08:57:18 crc kubenswrapper[4870]: I1014 08:57:18.930827 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 14 08:57:18 crc kubenswrapper[4870]: I1014 08:57:18.950490 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 14 08:57:18 crc kubenswrapper[4870]: I1014 08:57:18.977546 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 14 08:57:18 crc kubenswrapper[4870]: E1014 08:57:18.977929 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="60dbc30c-de4d-4574-b658-e5d43fc32288" containerName="glance-log" Oct 14 08:57:18 crc kubenswrapper[4870]: I1014 08:57:18.977940 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="60dbc30c-de4d-4574-b658-e5d43fc32288" containerName="glance-log" Oct 14 08:57:18 crc kubenswrapper[4870]: E1014 08:57:18.977958 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="60dbc30c-de4d-4574-b658-e5d43fc32288" containerName="glance-httpd" Oct 14 08:57:18 crc kubenswrapper[4870]: I1014 08:57:18.977964 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="60dbc30c-de4d-4574-b658-e5d43fc32288" containerName="glance-httpd" Oct 14 08:57:18 crc kubenswrapper[4870]: I1014 08:57:18.978128 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="60dbc30c-de4d-4574-b658-e5d43fc32288" containerName="glance-httpd" Oct 14 08:57:18 crc kubenswrapper[4870]: I1014 08:57:18.978147 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="60dbc30c-de4d-4574-b658-e5d43fc32288" containerName="glance-log" Oct 14 08:57:18 crc kubenswrapper[4870]: I1014 08:57:18.979021 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 14 08:57:18 crc kubenswrapper[4870]: I1014 08:57:18.984179 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 14 08:57:19 crc kubenswrapper[4870]: I1014 08:57:19.009095 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 14 08:57:19 crc kubenswrapper[4870]: I1014 08:57:19.017461 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gqdcb\" (UniqueName: \"kubernetes.io/projected/829bce4c-6fe7-46fa-a4d8-1558220a039f-kube-api-access-gqdcb\") pod \"glance-default-external-api-0\" (UID: \"829bce4c-6fe7-46fa-a4d8-1558220a039f\") " pod="openstack/glance-default-external-api-0" Oct 14 08:57:19 crc kubenswrapper[4870]: I1014 08:57:19.017512 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/829bce4c-6fe7-46fa-a4d8-1558220a039f-ceph\") pod \"glance-default-external-api-0\" (UID: \"829bce4c-6fe7-46fa-a4d8-1558220a039f\") " pod="openstack/glance-default-external-api-0" Oct 14 08:57:19 crc kubenswrapper[4870]: I1014 08:57:19.018155 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/829bce4c-6fe7-46fa-a4d8-1558220a039f-logs\") pod \"glance-default-external-api-0\" (UID: \"829bce4c-6fe7-46fa-a4d8-1558220a039f\") " pod="openstack/glance-default-external-api-0" Oct 14 08:57:19 crc kubenswrapper[4870]: I1014 08:57:19.018311 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/829bce4c-6fe7-46fa-a4d8-1558220a039f-scripts\") pod \"glance-default-external-api-0\" (UID: \"829bce4c-6fe7-46fa-a4d8-1558220a039f\") " pod="openstack/glance-default-external-api-0" Oct 14 08:57:19 crc kubenswrapper[4870]: I1014 08:57:19.018384 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/829bce4c-6fe7-46fa-a4d8-1558220a039f-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"829bce4c-6fe7-46fa-a4d8-1558220a039f\") " pod="openstack/glance-default-external-api-0" Oct 14 08:57:19 crc kubenswrapper[4870]: I1014 08:57:19.018424 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/829bce4c-6fe7-46fa-a4d8-1558220a039f-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"829bce4c-6fe7-46fa-a4d8-1558220a039f\") " pod="openstack/glance-default-external-api-0" Oct 14 08:57:19 crc kubenswrapper[4870]: I1014 08:57:19.018647 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/829bce4c-6fe7-46fa-a4d8-1558220a039f-config-data\") pod \"glance-default-external-api-0\" (UID: \"829bce4c-6fe7-46fa-a4d8-1558220a039f\") " pod="openstack/glance-default-external-api-0" Oct 14 08:57:19 crc kubenswrapper[4870]: I1014 08:57:19.018856 4870 scope.go:117] "RemoveContainer" containerID="c818cc0257089bc638ae8a03602c2c4939e1ec5115ea406f0fa7b0a4a6172e12" Oct 14 08:57:19 crc kubenswrapper[4870]: E1014 08:57:19.023293 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c818cc0257089bc638ae8a03602c2c4939e1ec5115ea406f0fa7b0a4a6172e12\": container with ID starting with c818cc0257089bc638ae8a03602c2c4939e1ec5115ea406f0fa7b0a4a6172e12 not found: ID does not exist" containerID="c818cc0257089bc638ae8a03602c2c4939e1ec5115ea406f0fa7b0a4a6172e12" Oct 14 08:57:19 crc kubenswrapper[4870]: I1014 08:57:19.023375 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c818cc0257089bc638ae8a03602c2c4939e1ec5115ea406f0fa7b0a4a6172e12"} err="failed to get container status \"c818cc0257089bc638ae8a03602c2c4939e1ec5115ea406f0fa7b0a4a6172e12\": rpc error: code = NotFound desc = could not find container \"c818cc0257089bc638ae8a03602c2c4939e1ec5115ea406f0fa7b0a4a6172e12\": container with ID starting with c818cc0257089bc638ae8a03602c2c4939e1ec5115ea406f0fa7b0a4a6172e12 not found: ID does not exist" Oct 14 08:57:19 crc kubenswrapper[4870]: I1014 08:57:19.023471 4870 scope.go:117] "RemoveContainer" containerID="1cfc46d02e2a200441f75ab42bdbc56afce8c182aa35abbf6490c152c62996f2" Oct 14 08:57:19 crc kubenswrapper[4870]: E1014 08:57:19.039053 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1cfc46d02e2a200441f75ab42bdbc56afce8c182aa35abbf6490c152c62996f2\": container with ID starting with 1cfc46d02e2a200441f75ab42bdbc56afce8c182aa35abbf6490c152c62996f2 not found: ID does not exist" containerID="1cfc46d02e2a200441f75ab42bdbc56afce8c182aa35abbf6490c152c62996f2" Oct 14 08:57:19 crc kubenswrapper[4870]: I1014 08:57:19.039181 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1cfc46d02e2a200441f75ab42bdbc56afce8c182aa35abbf6490c152c62996f2"} err="failed to get container status \"1cfc46d02e2a200441f75ab42bdbc56afce8c182aa35abbf6490c152c62996f2\": rpc error: code = NotFound desc = could not find container \"1cfc46d02e2a200441f75ab42bdbc56afce8c182aa35abbf6490c152c62996f2\": container with ID starting with 1cfc46d02e2a200441f75ab42bdbc56afce8c182aa35abbf6490c152c62996f2 not found: ID does not exist" Oct 14 08:57:19 crc kubenswrapper[4870]: I1014 08:57:19.039274 4870 scope.go:117] "RemoveContainer" containerID="c818cc0257089bc638ae8a03602c2c4939e1ec5115ea406f0fa7b0a4a6172e12" Oct 14 08:57:19 crc kubenswrapper[4870]: I1014 08:57:19.054559 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c818cc0257089bc638ae8a03602c2c4939e1ec5115ea406f0fa7b0a4a6172e12"} err="failed to get container status \"c818cc0257089bc638ae8a03602c2c4939e1ec5115ea406f0fa7b0a4a6172e12\": rpc error: code = NotFound desc = could not find container \"c818cc0257089bc638ae8a03602c2c4939e1ec5115ea406f0fa7b0a4a6172e12\": container with ID starting with c818cc0257089bc638ae8a03602c2c4939e1ec5115ea406f0fa7b0a4a6172e12 not found: ID does not exist" Oct 14 08:57:19 crc kubenswrapper[4870]: I1014 08:57:19.054778 4870 scope.go:117] "RemoveContainer" containerID="1cfc46d02e2a200441f75ab42bdbc56afce8c182aa35abbf6490c152c62996f2" Oct 14 08:57:19 crc kubenswrapper[4870]: I1014 08:57:19.064851 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1cfc46d02e2a200441f75ab42bdbc56afce8c182aa35abbf6490c152c62996f2"} err="failed to get container status \"1cfc46d02e2a200441f75ab42bdbc56afce8c182aa35abbf6490c152c62996f2\": rpc error: code = NotFound desc = could not find container \"1cfc46d02e2a200441f75ab42bdbc56afce8c182aa35abbf6490c152c62996f2\": container with ID starting with 1cfc46d02e2a200441f75ab42bdbc56afce8c182aa35abbf6490c152c62996f2 not found: ID does not exist" Oct 14 08:57:19 crc kubenswrapper[4870]: I1014 08:57:19.069777 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="60dbc30c-de4d-4574-b658-e5d43fc32288" path="/var/lib/kubelet/pods/60dbc30c-de4d-4574-b658-e5d43fc32288/volumes" Oct 14 08:57:19 crc kubenswrapper[4870]: I1014 08:57:19.120390 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/829bce4c-6fe7-46fa-a4d8-1558220a039f-logs\") pod \"glance-default-external-api-0\" (UID: \"829bce4c-6fe7-46fa-a4d8-1558220a039f\") " pod="openstack/glance-default-external-api-0" Oct 14 08:57:19 crc kubenswrapper[4870]: I1014 08:57:19.120691 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/829bce4c-6fe7-46fa-a4d8-1558220a039f-scripts\") pod \"glance-default-external-api-0\" (UID: \"829bce4c-6fe7-46fa-a4d8-1558220a039f\") " pod="openstack/glance-default-external-api-0" Oct 14 08:57:19 crc kubenswrapper[4870]: I1014 08:57:19.120810 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/829bce4c-6fe7-46fa-a4d8-1558220a039f-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"829bce4c-6fe7-46fa-a4d8-1558220a039f\") " pod="openstack/glance-default-external-api-0" Oct 14 08:57:19 crc kubenswrapper[4870]: I1014 08:57:19.120880 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/829bce4c-6fe7-46fa-a4d8-1558220a039f-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"829bce4c-6fe7-46fa-a4d8-1558220a039f\") " pod="openstack/glance-default-external-api-0" Oct 14 08:57:19 crc kubenswrapper[4870]: I1014 08:57:19.121001 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/829bce4c-6fe7-46fa-a4d8-1558220a039f-config-data\") pod \"glance-default-external-api-0\" (UID: \"829bce4c-6fe7-46fa-a4d8-1558220a039f\") " pod="openstack/glance-default-external-api-0" Oct 14 08:57:19 crc kubenswrapper[4870]: I1014 08:57:19.121100 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gqdcb\" (UniqueName: \"kubernetes.io/projected/829bce4c-6fe7-46fa-a4d8-1558220a039f-kube-api-access-gqdcb\") pod \"glance-default-external-api-0\" (UID: \"829bce4c-6fe7-46fa-a4d8-1558220a039f\") " pod="openstack/glance-default-external-api-0" Oct 14 08:57:19 crc kubenswrapper[4870]: I1014 08:57:19.121173 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/829bce4c-6fe7-46fa-a4d8-1558220a039f-ceph\") pod \"glance-default-external-api-0\" (UID: \"829bce4c-6fe7-46fa-a4d8-1558220a039f\") " pod="openstack/glance-default-external-api-0" Oct 14 08:57:19 crc kubenswrapper[4870]: I1014 08:57:19.120895 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/829bce4c-6fe7-46fa-a4d8-1558220a039f-logs\") pod \"glance-default-external-api-0\" (UID: \"829bce4c-6fe7-46fa-a4d8-1558220a039f\") " pod="openstack/glance-default-external-api-0" Oct 14 08:57:19 crc kubenswrapper[4870]: I1014 08:57:19.121120 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/829bce4c-6fe7-46fa-a4d8-1558220a039f-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"829bce4c-6fe7-46fa-a4d8-1558220a039f\") " pod="openstack/glance-default-external-api-0" Oct 14 08:57:19 crc kubenswrapper[4870]: I1014 08:57:19.126187 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/829bce4c-6fe7-46fa-a4d8-1558220a039f-scripts\") pod \"glance-default-external-api-0\" (UID: \"829bce4c-6fe7-46fa-a4d8-1558220a039f\") " pod="openstack/glance-default-external-api-0" Oct 14 08:57:19 crc kubenswrapper[4870]: I1014 08:57:19.126500 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/829bce4c-6fe7-46fa-a4d8-1558220a039f-ceph\") pod \"glance-default-external-api-0\" (UID: \"829bce4c-6fe7-46fa-a4d8-1558220a039f\") " pod="openstack/glance-default-external-api-0" Oct 14 08:57:19 crc kubenswrapper[4870]: I1014 08:57:19.127009 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/829bce4c-6fe7-46fa-a4d8-1558220a039f-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"829bce4c-6fe7-46fa-a4d8-1558220a039f\") " pod="openstack/glance-default-external-api-0" Oct 14 08:57:19 crc kubenswrapper[4870]: I1014 08:57:19.127835 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/829bce4c-6fe7-46fa-a4d8-1558220a039f-config-data\") pod \"glance-default-external-api-0\" (UID: \"829bce4c-6fe7-46fa-a4d8-1558220a039f\") " pod="openstack/glance-default-external-api-0" Oct 14 08:57:19 crc kubenswrapper[4870]: I1014 08:57:19.142766 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gqdcb\" (UniqueName: \"kubernetes.io/projected/829bce4c-6fe7-46fa-a4d8-1558220a039f-kube-api-access-gqdcb\") pod \"glance-default-external-api-0\" (UID: \"829bce4c-6fe7-46fa-a4d8-1558220a039f\") " pod="openstack/glance-default-external-api-0" Oct 14 08:57:19 crc kubenswrapper[4870]: I1014 08:57:19.333997 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 14 08:57:19 crc kubenswrapper[4870]: I1014 08:57:19.745006 4870 scope.go:117] "RemoveContainer" containerID="2df76e3cfba0d4fc310c83933f2a1ddf28afd200614e7727c2c55f61c7cf17e2" Oct 14 08:57:19 crc kubenswrapper[4870]: I1014 08:57:19.860012 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="f94fd4d7-586c-4efd-8329-2b3667fd8d58" containerName="glance-log" containerID="cri-o://e80efe033ae130b7eee883c22976e3984670f7779d8adbc52b6bef2d3839ad2c" gracePeriod=30 Oct 14 08:57:19 crc kubenswrapper[4870]: I1014 08:57:19.860097 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="f94fd4d7-586c-4efd-8329-2b3667fd8d58" containerName="glance-httpd" containerID="cri-o://a36a0cf52fe028fb39868934fe28934070222f110f220e1dda8bf75588ca0950" gracePeriod=30 Oct 14 08:57:19 crc kubenswrapper[4870]: I1014 08:57:19.888274 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 14 08:57:19 crc kubenswrapper[4870]: W1014 08:57:19.914757 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod829bce4c_6fe7_46fa_a4d8_1558220a039f.slice/crio-853d87787afccaeff5d595dbd4610fa6450723bc1f68c6e77468ead695db7bd7 WatchSource:0}: Error finding container 853d87787afccaeff5d595dbd4610fa6450723bc1f68c6e77468ead695db7bd7: Status 404 returned error can't find the container with id 853d87787afccaeff5d595dbd4610fa6450723bc1f68c6e77468ead695db7bd7 Oct 14 08:57:20 crc kubenswrapper[4870]: I1014 08:57:20.618697 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 14 08:57:20 crc kubenswrapper[4870]: I1014 08:57:20.649112 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f94fd4d7-586c-4efd-8329-2b3667fd8d58-combined-ca-bundle\") pod \"f94fd4d7-586c-4efd-8329-2b3667fd8d58\" (UID: \"f94fd4d7-586c-4efd-8329-2b3667fd8d58\") " Oct 14 08:57:20 crc kubenswrapper[4870]: I1014 08:57:20.649184 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sxt7b\" (UniqueName: \"kubernetes.io/projected/f94fd4d7-586c-4efd-8329-2b3667fd8d58-kube-api-access-sxt7b\") pod \"f94fd4d7-586c-4efd-8329-2b3667fd8d58\" (UID: \"f94fd4d7-586c-4efd-8329-2b3667fd8d58\") " Oct 14 08:57:20 crc kubenswrapper[4870]: I1014 08:57:20.649286 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f94fd4d7-586c-4efd-8329-2b3667fd8d58-logs\") pod \"f94fd4d7-586c-4efd-8329-2b3667fd8d58\" (UID: \"f94fd4d7-586c-4efd-8329-2b3667fd8d58\") " Oct 14 08:57:20 crc kubenswrapper[4870]: I1014 08:57:20.649371 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f94fd4d7-586c-4efd-8329-2b3667fd8d58-config-data\") pod \"f94fd4d7-586c-4efd-8329-2b3667fd8d58\" (UID: \"f94fd4d7-586c-4efd-8329-2b3667fd8d58\") " Oct 14 08:57:20 crc kubenswrapper[4870]: I1014 08:57:20.649408 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f94fd4d7-586c-4efd-8329-2b3667fd8d58-scripts\") pod \"f94fd4d7-586c-4efd-8329-2b3667fd8d58\" (UID: \"f94fd4d7-586c-4efd-8329-2b3667fd8d58\") " Oct 14 08:57:20 crc kubenswrapper[4870]: I1014 08:57:20.649463 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f94fd4d7-586c-4efd-8329-2b3667fd8d58-httpd-run\") pod \"f94fd4d7-586c-4efd-8329-2b3667fd8d58\" (UID: \"f94fd4d7-586c-4efd-8329-2b3667fd8d58\") " Oct 14 08:57:20 crc kubenswrapper[4870]: I1014 08:57:20.649489 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/f94fd4d7-586c-4efd-8329-2b3667fd8d58-ceph\") pod \"f94fd4d7-586c-4efd-8329-2b3667fd8d58\" (UID: \"f94fd4d7-586c-4efd-8329-2b3667fd8d58\") " Oct 14 08:57:20 crc kubenswrapper[4870]: I1014 08:57:20.650312 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f94fd4d7-586c-4efd-8329-2b3667fd8d58-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "f94fd4d7-586c-4efd-8329-2b3667fd8d58" (UID: "f94fd4d7-586c-4efd-8329-2b3667fd8d58"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:57:20 crc kubenswrapper[4870]: I1014 08:57:20.650525 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f94fd4d7-586c-4efd-8329-2b3667fd8d58-logs" (OuterVolumeSpecName: "logs") pod "f94fd4d7-586c-4efd-8329-2b3667fd8d58" (UID: "f94fd4d7-586c-4efd-8329-2b3667fd8d58"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:57:20 crc kubenswrapper[4870]: I1014 08:57:20.653706 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f94fd4d7-586c-4efd-8329-2b3667fd8d58-kube-api-access-sxt7b" (OuterVolumeSpecName: "kube-api-access-sxt7b") pod "f94fd4d7-586c-4efd-8329-2b3667fd8d58" (UID: "f94fd4d7-586c-4efd-8329-2b3667fd8d58"). InnerVolumeSpecName "kube-api-access-sxt7b". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:57:20 crc kubenswrapper[4870]: I1014 08:57:20.654181 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f94fd4d7-586c-4efd-8329-2b3667fd8d58-ceph" (OuterVolumeSpecName: "ceph") pod "f94fd4d7-586c-4efd-8329-2b3667fd8d58" (UID: "f94fd4d7-586c-4efd-8329-2b3667fd8d58"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:57:20 crc kubenswrapper[4870]: I1014 08:57:20.658667 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f94fd4d7-586c-4efd-8329-2b3667fd8d58-scripts" (OuterVolumeSpecName: "scripts") pod "f94fd4d7-586c-4efd-8329-2b3667fd8d58" (UID: "f94fd4d7-586c-4efd-8329-2b3667fd8d58"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:57:20 crc kubenswrapper[4870]: I1014 08:57:20.694581 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f94fd4d7-586c-4efd-8329-2b3667fd8d58-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f94fd4d7-586c-4efd-8329-2b3667fd8d58" (UID: "f94fd4d7-586c-4efd-8329-2b3667fd8d58"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:57:20 crc kubenswrapper[4870]: I1014 08:57:20.695613 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f94fd4d7-586c-4efd-8329-2b3667fd8d58-config-data" (OuterVolumeSpecName: "config-data") pod "f94fd4d7-586c-4efd-8329-2b3667fd8d58" (UID: "f94fd4d7-586c-4efd-8329-2b3667fd8d58"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:57:20 crc kubenswrapper[4870]: I1014 08:57:20.751864 4870 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f94fd4d7-586c-4efd-8329-2b3667fd8d58-logs\") on node \"crc\" DevicePath \"\"" Oct 14 08:57:20 crc kubenswrapper[4870]: I1014 08:57:20.752165 4870 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f94fd4d7-586c-4efd-8329-2b3667fd8d58-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 08:57:20 crc kubenswrapper[4870]: I1014 08:57:20.752187 4870 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f94fd4d7-586c-4efd-8329-2b3667fd8d58-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 08:57:20 crc kubenswrapper[4870]: I1014 08:57:20.752207 4870 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f94fd4d7-586c-4efd-8329-2b3667fd8d58-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 14 08:57:20 crc kubenswrapper[4870]: I1014 08:57:20.752228 4870 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/f94fd4d7-586c-4efd-8329-2b3667fd8d58-ceph\") on node \"crc\" DevicePath \"\"" Oct 14 08:57:20 crc kubenswrapper[4870]: I1014 08:57:20.752247 4870 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f94fd4d7-586c-4efd-8329-2b3667fd8d58-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 08:57:20 crc kubenswrapper[4870]: I1014 08:57:20.752268 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sxt7b\" (UniqueName: \"kubernetes.io/projected/f94fd4d7-586c-4efd-8329-2b3667fd8d58-kube-api-access-sxt7b\") on node \"crc\" DevicePath \"\"" Oct 14 08:57:20 crc kubenswrapper[4870]: I1014 08:57:20.872022 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"829bce4c-6fe7-46fa-a4d8-1558220a039f","Type":"ContainerStarted","Data":"dcdf64ab0812d5c47611df139478fcdb35fb58c52140a0bfaf26de928cff6bc0"} Oct 14 08:57:20 crc kubenswrapper[4870]: I1014 08:57:20.872067 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"829bce4c-6fe7-46fa-a4d8-1558220a039f","Type":"ContainerStarted","Data":"853d87787afccaeff5d595dbd4610fa6450723bc1f68c6e77468ead695db7bd7"} Oct 14 08:57:20 crc kubenswrapper[4870]: I1014 08:57:20.873935 4870 generic.go:334] "Generic (PLEG): container finished" podID="f94fd4d7-586c-4efd-8329-2b3667fd8d58" containerID="a36a0cf52fe028fb39868934fe28934070222f110f220e1dda8bf75588ca0950" exitCode=0 Oct 14 08:57:20 crc kubenswrapper[4870]: I1014 08:57:20.873990 4870 generic.go:334] "Generic (PLEG): container finished" podID="f94fd4d7-586c-4efd-8329-2b3667fd8d58" containerID="e80efe033ae130b7eee883c22976e3984670f7779d8adbc52b6bef2d3839ad2c" exitCode=143 Oct 14 08:57:20 crc kubenswrapper[4870]: I1014 08:57:20.874006 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f94fd4d7-586c-4efd-8329-2b3667fd8d58","Type":"ContainerDied","Data":"a36a0cf52fe028fb39868934fe28934070222f110f220e1dda8bf75588ca0950"} Oct 14 08:57:20 crc kubenswrapper[4870]: I1014 08:57:20.874024 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f94fd4d7-586c-4efd-8329-2b3667fd8d58","Type":"ContainerDied","Data":"e80efe033ae130b7eee883c22976e3984670f7779d8adbc52b6bef2d3839ad2c"} Oct 14 08:57:20 crc kubenswrapper[4870]: I1014 08:57:20.874036 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f94fd4d7-586c-4efd-8329-2b3667fd8d58","Type":"ContainerDied","Data":"6d003e5e9ad8663a2a36de730e70c5ce9ae9f17f809c449cf5a5e44429256edf"} Oct 14 08:57:20 crc kubenswrapper[4870]: I1014 08:57:20.874086 4870 scope.go:117] "RemoveContainer" containerID="a36a0cf52fe028fb39868934fe28934070222f110f220e1dda8bf75588ca0950" Oct 14 08:57:20 crc kubenswrapper[4870]: I1014 08:57:20.874107 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 14 08:57:20 crc kubenswrapper[4870]: I1014 08:57:20.908345 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 14 08:57:20 crc kubenswrapper[4870]: I1014 08:57:20.911639 4870 scope.go:117] "RemoveContainer" containerID="e80efe033ae130b7eee883c22976e3984670f7779d8adbc52b6bef2d3839ad2c" Oct 14 08:57:20 crc kubenswrapper[4870]: I1014 08:57:20.916367 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 14 08:57:20 crc kubenswrapper[4870]: I1014 08:57:20.930386 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 14 08:57:20 crc kubenswrapper[4870]: E1014 08:57:20.930797 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f94fd4d7-586c-4efd-8329-2b3667fd8d58" containerName="glance-log" Oct 14 08:57:20 crc kubenswrapper[4870]: I1014 08:57:20.930821 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="f94fd4d7-586c-4efd-8329-2b3667fd8d58" containerName="glance-log" Oct 14 08:57:20 crc kubenswrapper[4870]: E1014 08:57:20.930838 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f94fd4d7-586c-4efd-8329-2b3667fd8d58" containerName="glance-httpd" Oct 14 08:57:20 crc kubenswrapper[4870]: I1014 08:57:20.930844 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="f94fd4d7-586c-4efd-8329-2b3667fd8d58" containerName="glance-httpd" Oct 14 08:57:20 crc kubenswrapper[4870]: I1014 08:57:20.931025 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="f94fd4d7-586c-4efd-8329-2b3667fd8d58" containerName="glance-log" Oct 14 08:57:20 crc kubenswrapper[4870]: I1014 08:57:20.931120 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="f94fd4d7-586c-4efd-8329-2b3667fd8d58" containerName="glance-httpd" Oct 14 08:57:20 crc kubenswrapper[4870]: I1014 08:57:20.932623 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 14 08:57:20 crc kubenswrapper[4870]: I1014 08:57:20.941383 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 14 08:57:20 crc kubenswrapper[4870]: I1014 08:57:20.947378 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 14 08:57:20 crc kubenswrapper[4870]: I1014 08:57:20.956686 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/de9df7d3-70ec-4abd-9cd0-55038d9ab2df-scripts\") pod \"glance-default-internal-api-0\" (UID: \"de9df7d3-70ec-4abd-9cd0-55038d9ab2df\") " pod="openstack/glance-default-internal-api-0" Oct 14 08:57:20 crc kubenswrapper[4870]: I1014 08:57:20.956730 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de9df7d3-70ec-4abd-9cd0-55038d9ab2df-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"de9df7d3-70ec-4abd-9cd0-55038d9ab2df\") " pod="openstack/glance-default-internal-api-0" Oct 14 08:57:20 crc kubenswrapper[4870]: I1014 08:57:20.956764 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/de9df7d3-70ec-4abd-9cd0-55038d9ab2df-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"de9df7d3-70ec-4abd-9cd0-55038d9ab2df\") " pod="openstack/glance-default-internal-api-0" Oct 14 08:57:20 crc kubenswrapper[4870]: I1014 08:57:20.956808 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/de9df7d3-70ec-4abd-9cd0-55038d9ab2df-config-data\") pod \"glance-default-internal-api-0\" (UID: \"de9df7d3-70ec-4abd-9cd0-55038d9ab2df\") " pod="openstack/glance-default-internal-api-0" Oct 14 08:57:20 crc kubenswrapper[4870]: I1014 08:57:20.956997 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/de9df7d3-70ec-4abd-9cd0-55038d9ab2df-ceph\") pod \"glance-default-internal-api-0\" (UID: \"de9df7d3-70ec-4abd-9cd0-55038d9ab2df\") " pod="openstack/glance-default-internal-api-0" Oct 14 08:57:20 crc kubenswrapper[4870]: I1014 08:57:20.957071 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/de9df7d3-70ec-4abd-9cd0-55038d9ab2df-logs\") pod \"glance-default-internal-api-0\" (UID: \"de9df7d3-70ec-4abd-9cd0-55038d9ab2df\") " pod="openstack/glance-default-internal-api-0" Oct 14 08:57:20 crc kubenswrapper[4870]: I1014 08:57:20.957104 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wpsvd\" (UniqueName: \"kubernetes.io/projected/de9df7d3-70ec-4abd-9cd0-55038d9ab2df-kube-api-access-wpsvd\") pod \"glance-default-internal-api-0\" (UID: \"de9df7d3-70ec-4abd-9cd0-55038d9ab2df\") " pod="openstack/glance-default-internal-api-0" Oct 14 08:57:20 crc kubenswrapper[4870]: I1014 08:57:20.978553 4870 scope.go:117] "RemoveContainer" containerID="a36a0cf52fe028fb39868934fe28934070222f110f220e1dda8bf75588ca0950" Oct 14 08:57:20 crc kubenswrapper[4870]: E1014 08:57:20.979033 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a36a0cf52fe028fb39868934fe28934070222f110f220e1dda8bf75588ca0950\": container with ID starting with a36a0cf52fe028fb39868934fe28934070222f110f220e1dda8bf75588ca0950 not found: ID does not exist" containerID="a36a0cf52fe028fb39868934fe28934070222f110f220e1dda8bf75588ca0950" Oct 14 08:57:20 crc kubenswrapper[4870]: I1014 08:57:20.979113 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a36a0cf52fe028fb39868934fe28934070222f110f220e1dda8bf75588ca0950"} err="failed to get container status \"a36a0cf52fe028fb39868934fe28934070222f110f220e1dda8bf75588ca0950\": rpc error: code = NotFound desc = could not find container \"a36a0cf52fe028fb39868934fe28934070222f110f220e1dda8bf75588ca0950\": container with ID starting with a36a0cf52fe028fb39868934fe28934070222f110f220e1dda8bf75588ca0950 not found: ID does not exist" Oct 14 08:57:20 crc kubenswrapper[4870]: I1014 08:57:20.979152 4870 scope.go:117] "RemoveContainer" containerID="e80efe033ae130b7eee883c22976e3984670f7779d8adbc52b6bef2d3839ad2c" Oct 14 08:57:20 crc kubenswrapper[4870]: E1014 08:57:20.979471 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e80efe033ae130b7eee883c22976e3984670f7779d8adbc52b6bef2d3839ad2c\": container with ID starting with e80efe033ae130b7eee883c22976e3984670f7779d8adbc52b6bef2d3839ad2c not found: ID does not exist" containerID="e80efe033ae130b7eee883c22976e3984670f7779d8adbc52b6bef2d3839ad2c" Oct 14 08:57:20 crc kubenswrapper[4870]: I1014 08:57:20.979504 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e80efe033ae130b7eee883c22976e3984670f7779d8adbc52b6bef2d3839ad2c"} err="failed to get container status \"e80efe033ae130b7eee883c22976e3984670f7779d8adbc52b6bef2d3839ad2c\": rpc error: code = NotFound desc = could not find container \"e80efe033ae130b7eee883c22976e3984670f7779d8adbc52b6bef2d3839ad2c\": container with ID starting with e80efe033ae130b7eee883c22976e3984670f7779d8adbc52b6bef2d3839ad2c not found: ID does not exist" Oct 14 08:57:20 crc kubenswrapper[4870]: I1014 08:57:20.979525 4870 scope.go:117] "RemoveContainer" containerID="a36a0cf52fe028fb39868934fe28934070222f110f220e1dda8bf75588ca0950" Oct 14 08:57:20 crc kubenswrapper[4870]: I1014 08:57:20.979740 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a36a0cf52fe028fb39868934fe28934070222f110f220e1dda8bf75588ca0950"} err="failed to get container status \"a36a0cf52fe028fb39868934fe28934070222f110f220e1dda8bf75588ca0950\": rpc error: code = NotFound desc = could not find container \"a36a0cf52fe028fb39868934fe28934070222f110f220e1dda8bf75588ca0950\": container with ID starting with a36a0cf52fe028fb39868934fe28934070222f110f220e1dda8bf75588ca0950 not found: ID does not exist" Oct 14 08:57:20 crc kubenswrapper[4870]: I1014 08:57:20.979762 4870 scope.go:117] "RemoveContainer" containerID="e80efe033ae130b7eee883c22976e3984670f7779d8adbc52b6bef2d3839ad2c" Oct 14 08:57:20 crc kubenswrapper[4870]: I1014 08:57:20.979981 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e80efe033ae130b7eee883c22976e3984670f7779d8adbc52b6bef2d3839ad2c"} err="failed to get container status \"e80efe033ae130b7eee883c22976e3984670f7779d8adbc52b6bef2d3839ad2c\": rpc error: code = NotFound desc = could not find container \"e80efe033ae130b7eee883c22976e3984670f7779d8adbc52b6bef2d3839ad2c\": container with ID starting with e80efe033ae130b7eee883c22976e3984670f7779d8adbc52b6bef2d3839ad2c not found: ID does not exist" Oct 14 08:57:21 crc kubenswrapper[4870]: I1014 08:57:21.045893 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f94fd4d7-586c-4efd-8329-2b3667fd8d58" path="/var/lib/kubelet/pods/f94fd4d7-586c-4efd-8329-2b3667fd8d58/volumes" Oct 14 08:57:21 crc kubenswrapper[4870]: I1014 08:57:21.058446 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de9df7d3-70ec-4abd-9cd0-55038d9ab2df-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"de9df7d3-70ec-4abd-9cd0-55038d9ab2df\") " pod="openstack/glance-default-internal-api-0" Oct 14 08:57:21 crc kubenswrapper[4870]: I1014 08:57:21.058549 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/de9df7d3-70ec-4abd-9cd0-55038d9ab2df-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"de9df7d3-70ec-4abd-9cd0-55038d9ab2df\") " pod="openstack/glance-default-internal-api-0" Oct 14 08:57:21 crc kubenswrapper[4870]: I1014 08:57:21.058666 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/de9df7d3-70ec-4abd-9cd0-55038d9ab2df-config-data\") pod \"glance-default-internal-api-0\" (UID: \"de9df7d3-70ec-4abd-9cd0-55038d9ab2df\") " pod="openstack/glance-default-internal-api-0" Oct 14 08:57:21 crc kubenswrapper[4870]: I1014 08:57:21.058782 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/de9df7d3-70ec-4abd-9cd0-55038d9ab2df-ceph\") pod \"glance-default-internal-api-0\" (UID: \"de9df7d3-70ec-4abd-9cd0-55038d9ab2df\") " pod="openstack/glance-default-internal-api-0" Oct 14 08:57:21 crc kubenswrapper[4870]: I1014 08:57:21.058823 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/de9df7d3-70ec-4abd-9cd0-55038d9ab2df-logs\") pod \"glance-default-internal-api-0\" (UID: \"de9df7d3-70ec-4abd-9cd0-55038d9ab2df\") " pod="openstack/glance-default-internal-api-0" Oct 14 08:57:21 crc kubenswrapper[4870]: I1014 08:57:21.058848 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wpsvd\" (UniqueName: \"kubernetes.io/projected/de9df7d3-70ec-4abd-9cd0-55038d9ab2df-kube-api-access-wpsvd\") pod \"glance-default-internal-api-0\" (UID: \"de9df7d3-70ec-4abd-9cd0-55038d9ab2df\") " pod="openstack/glance-default-internal-api-0" Oct 14 08:57:21 crc kubenswrapper[4870]: I1014 08:57:21.058903 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/de9df7d3-70ec-4abd-9cd0-55038d9ab2df-scripts\") pod \"glance-default-internal-api-0\" (UID: \"de9df7d3-70ec-4abd-9cd0-55038d9ab2df\") " pod="openstack/glance-default-internal-api-0" Oct 14 08:57:21 crc kubenswrapper[4870]: I1014 08:57:21.061011 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/de9df7d3-70ec-4abd-9cd0-55038d9ab2df-logs\") pod \"glance-default-internal-api-0\" (UID: \"de9df7d3-70ec-4abd-9cd0-55038d9ab2df\") " pod="openstack/glance-default-internal-api-0" Oct 14 08:57:21 crc kubenswrapper[4870]: I1014 08:57:21.061715 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/de9df7d3-70ec-4abd-9cd0-55038d9ab2df-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"de9df7d3-70ec-4abd-9cd0-55038d9ab2df\") " pod="openstack/glance-default-internal-api-0" Oct 14 08:57:21 crc kubenswrapper[4870]: I1014 08:57:21.063656 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de9df7d3-70ec-4abd-9cd0-55038d9ab2df-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"de9df7d3-70ec-4abd-9cd0-55038d9ab2df\") " pod="openstack/glance-default-internal-api-0" Oct 14 08:57:21 crc kubenswrapper[4870]: I1014 08:57:21.064128 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/de9df7d3-70ec-4abd-9cd0-55038d9ab2df-ceph\") pod \"glance-default-internal-api-0\" (UID: \"de9df7d3-70ec-4abd-9cd0-55038d9ab2df\") " pod="openstack/glance-default-internal-api-0" Oct 14 08:57:21 crc kubenswrapper[4870]: I1014 08:57:21.064381 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/de9df7d3-70ec-4abd-9cd0-55038d9ab2df-scripts\") pod \"glance-default-internal-api-0\" (UID: \"de9df7d3-70ec-4abd-9cd0-55038d9ab2df\") " pod="openstack/glance-default-internal-api-0" Oct 14 08:57:21 crc kubenswrapper[4870]: I1014 08:57:21.068320 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/de9df7d3-70ec-4abd-9cd0-55038d9ab2df-config-data\") pod \"glance-default-internal-api-0\" (UID: \"de9df7d3-70ec-4abd-9cd0-55038d9ab2df\") " pod="openstack/glance-default-internal-api-0" Oct 14 08:57:21 crc kubenswrapper[4870]: I1014 08:57:21.079502 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wpsvd\" (UniqueName: \"kubernetes.io/projected/de9df7d3-70ec-4abd-9cd0-55038d9ab2df-kube-api-access-wpsvd\") pod \"glance-default-internal-api-0\" (UID: \"de9df7d3-70ec-4abd-9cd0-55038d9ab2df\") " pod="openstack/glance-default-internal-api-0" Oct 14 08:57:21 crc kubenswrapper[4870]: I1014 08:57:21.265660 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 14 08:57:21 crc kubenswrapper[4870]: I1014 08:57:21.838212 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 14 08:57:21 crc kubenswrapper[4870]: I1014 08:57:21.886768 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"829bce4c-6fe7-46fa-a4d8-1558220a039f","Type":"ContainerStarted","Data":"6d39e737807423f661bed2bc5c4c82a9315751ab881b785991a4633e1b26b216"} Oct 14 08:57:21 crc kubenswrapper[4870]: I1014 08:57:21.887979 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"de9df7d3-70ec-4abd-9cd0-55038d9ab2df","Type":"ContainerStarted","Data":"9a39ca8808b54188910904d6b5204d25404c01fd88ea125f6c7620b7becdc941"} Oct 14 08:57:21 crc kubenswrapper[4870]: I1014 08:57:21.910322 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=3.9102979060000003 podStartE2EDuration="3.910297906s" podCreationTimestamp="2025-10-14 08:57:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 08:57:21.90556405 +0000 UTC m=+6977.602924421" watchObservedRunningTime="2025-10-14 08:57:21.910297906 +0000 UTC m=+6977.607658287" Oct 14 08:57:22 crc kubenswrapper[4870]: I1014 08:57:22.904630 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"de9df7d3-70ec-4abd-9cd0-55038d9ab2df","Type":"ContainerStarted","Data":"d5be18839f4b371ce0db12541e6bfc5e5ccd1903925a9840da4265c4d2692924"} Oct 14 08:57:22 crc kubenswrapper[4870]: I1014 08:57:22.904943 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"de9df7d3-70ec-4abd-9cd0-55038d9ab2df","Type":"ContainerStarted","Data":"8eb5a5a2c1544b193bef195717e303075859ef64fbf449f7c5c4eb693af477f6"} Oct 14 08:57:22 crc kubenswrapper[4870]: I1014 08:57:22.930589 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=2.930563138 podStartE2EDuration="2.930563138s" podCreationTimestamp="2025-10-14 08:57:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 08:57:22.921905306 +0000 UTC m=+6978.619265687" watchObservedRunningTime="2025-10-14 08:57:22.930563138 +0000 UTC m=+6978.627923509" Oct 14 08:57:24 crc kubenswrapper[4870]: I1014 08:57:24.033975 4870 scope.go:117] "RemoveContainer" containerID="8c0f754fe94b7cb7b9bb5dcd613f7e42093968b49509d8ee37e8fd4db89eedd5" Oct 14 08:57:24 crc kubenswrapper[4870]: E1014 08:57:24.034258 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 08:57:25 crc kubenswrapper[4870]: I1014 08:57:25.518884 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6695878787-8ngcl" Oct 14 08:57:25 crc kubenswrapper[4870]: I1014 08:57:25.625605 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6fcb7f9899-zxtfl"] Oct 14 08:57:25 crc kubenswrapper[4870]: I1014 08:57:25.625954 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6fcb7f9899-zxtfl" podUID="b3691846-4d55-4390-8cf4-94a9b28e4440" containerName="dnsmasq-dns" containerID="cri-o://522faffedac332784e3353c8f156a89f2734f5295fd2d562bad530cf310dcc75" gracePeriod=10 Oct 14 08:57:25 crc kubenswrapper[4870]: E1014 08:57:25.841122 4870 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb3691846_4d55_4390_8cf4_94a9b28e4440.slice/crio-conmon-522faffedac332784e3353c8f156a89f2734f5295fd2d562bad530cf310dcc75.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb3691846_4d55_4390_8cf4_94a9b28e4440.slice/crio-522faffedac332784e3353c8f156a89f2734f5295fd2d562bad530cf310dcc75.scope\": RecentStats: unable to find data in memory cache]" Oct 14 08:57:25 crc kubenswrapper[4870]: I1014 08:57:25.975221 4870 generic.go:334] "Generic (PLEG): container finished" podID="b3691846-4d55-4390-8cf4-94a9b28e4440" containerID="522faffedac332784e3353c8f156a89f2734f5295fd2d562bad530cf310dcc75" exitCode=0 Oct 14 08:57:25 crc kubenswrapper[4870]: I1014 08:57:25.975539 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6fcb7f9899-zxtfl" event={"ID":"b3691846-4d55-4390-8cf4-94a9b28e4440","Type":"ContainerDied","Data":"522faffedac332784e3353c8f156a89f2734f5295fd2d562bad530cf310dcc75"} Oct 14 08:57:26 crc kubenswrapper[4870]: I1014 08:57:26.196534 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6fcb7f9899-zxtfl" Oct 14 08:57:26 crc kubenswrapper[4870]: I1014 08:57:26.317660 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b3691846-4d55-4390-8cf4-94a9b28e4440-config\") pod \"b3691846-4d55-4390-8cf4-94a9b28e4440\" (UID: \"b3691846-4d55-4390-8cf4-94a9b28e4440\") " Oct 14 08:57:26 crc kubenswrapper[4870]: I1014 08:57:26.317734 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b3691846-4d55-4390-8cf4-94a9b28e4440-dns-svc\") pod \"b3691846-4d55-4390-8cf4-94a9b28e4440\" (UID: \"b3691846-4d55-4390-8cf4-94a9b28e4440\") " Oct 14 08:57:26 crc kubenswrapper[4870]: I1014 08:57:26.317781 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b3691846-4d55-4390-8cf4-94a9b28e4440-ovsdbserver-nb\") pod \"b3691846-4d55-4390-8cf4-94a9b28e4440\" (UID: \"b3691846-4d55-4390-8cf4-94a9b28e4440\") " Oct 14 08:57:26 crc kubenswrapper[4870]: I1014 08:57:26.317804 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wq65z\" (UniqueName: \"kubernetes.io/projected/b3691846-4d55-4390-8cf4-94a9b28e4440-kube-api-access-wq65z\") pod \"b3691846-4d55-4390-8cf4-94a9b28e4440\" (UID: \"b3691846-4d55-4390-8cf4-94a9b28e4440\") " Oct 14 08:57:26 crc kubenswrapper[4870]: I1014 08:57:26.317915 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b3691846-4d55-4390-8cf4-94a9b28e4440-ovsdbserver-sb\") pod \"b3691846-4d55-4390-8cf4-94a9b28e4440\" (UID: \"b3691846-4d55-4390-8cf4-94a9b28e4440\") " Oct 14 08:57:26 crc kubenswrapper[4870]: I1014 08:57:26.323131 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b3691846-4d55-4390-8cf4-94a9b28e4440-kube-api-access-wq65z" (OuterVolumeSpecName: "kube-api-access-wq65z") pod "b3691846-4d55-4390-8cf4-94a9b28e4440" (UID: "b3691846-4d55-4390-8cf4-94a9b28e4440"). InnerVolumeSpecName "kube-api-access-wq65z". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:57:26 crc kubenswrapper[4870]: I1014 08:57:26.362843 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b3691846-4d55-4390-8cf4-94a9b28e4440-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "b3691846-4d55-4390-8cf4-94a9b28e4440" (UID: "b3691846-4d55-4390-8cf4-94a9b28e4440"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:57:26 crc kubenswrapper[4870]: I1014 08:57:26.365992 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b3691846-4d55-4390-8cf4-94a9b28e4440-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "b3691846-4d55-4390-8cf4-94a9b28e4440" (UID: "b3691846-4d55-4390-8cf4-94a9b28e4440"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:57:26 crc kubenswrapper[4870]: I1014 08:57:26.372028 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b3691846-4d55-4390-8cf4-94a9b28e4440-config" (OuterVolumeSpecName: "config") pod "b3691846-4d55-4390-8cf4-94a9b28e4440" (UID: "b3691846-4d55-4390-8cf4-94a9b28e4440"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:57:26 crc kubenswrapper[4870]: I1014 08:57:26.380469 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b3691846-4d55-4390-8cf4-94a9b28e4440-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "b3691846-4d55-4390-8cf4-94a9b28e4440" (UID: "b3691846-4d55-4390-8cf4-94a9b28e4440"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:57:26 crc kubenswrapper[4870]: I1014 08:57:26.419779 4870 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b3691846-4d55-4390-8cf4-94a9b28e4440-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 14 08:57:26 crc kubenswrapper[4870]: I1014 08:57:26.419813 4870 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b3691846-4d55-4390-8cf4-94a9b28e4440-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 14 08:57:26 crc kubenswrapper[4870]: I1014 08:57:26.419825 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wq65z\" (UniqueName: \"kubernetes.io/projected/b3691846-4d55-4390-8cf4-94a9b28e4440-kube-api-access-wq65z\") on node \"crc\" DevicePath \"\"" Oct 14 08:57:26 crc kubenswrapper[4870]: I1014 08:57:26.419834 4870 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b3691846-4d55-4390-8cf4-94a9b28e4440-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 14 08:57:26 crc kubenswrapper[4870]: I1014 08:57:26.419842 4870 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b3691846-4d55-4390-8cf4-94a9b28e4440-config\") on node \"crc\" DevicePath \"\"" Oct 14 08:57:26 crc kubenswrapper[4870]: I1014 08:57:26.991174 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6fcb7f9899-zxtfl" event={"ID":"b3691846-4d55-4390-8cf4-94a9b28e4440","Type":"ContainerDied","Data":"e8e29e5abfff49fa7c476afdddc9aa532b1d27ce2e4574ddf5318dbd0f116dbe"} Oct 14 08:57:26 crc kubenswrapper[4870]: I1014 08:57:26.991250 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6fcb7f9899-zxtfl" Oct 14 08:57:26 crc kubenswrapper[4870]: I1014 08:57:26.992631 4870 scope.go:117] "RemoveContainer" containerID="522faffedac332784e3353c8f156a89f2734f5295fd2d562bad530cf310dcc75" Oct 14 08:57:27 crc kubenswrapper[4870]: I1014 08:57:27.027195 4870 scope.go:117] "RemoveContainer" containerID="eb3cbe7ff45c41a4d8c72ac6527f2f9d1676951bee76d6f7d0832963a48fc451" Oct 14 08:57:27 crc kubenswrapper[4870]: I1014 08:57:27.049831 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6fcb7f9899-zxtfl"] Oct 14 08:57:27 crc kubenswrapper[4870]: I1014 08:57:27.055281 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6fcb7f9899-zxtfl"] Oct 14 08:57:29 crc kubenswrapper[4870]: I1014 08:57:29.048895 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b3691846-4d55-4390-8cf4-94a9b28e4440" path="/var/lib/kubelet/pods/b3691846-4d55-4390-8cf4-94a9b28e4440/volumes" Oct 14 08:57:29 crc kubenswrapper[4870]: I1014 08:57:29.334730 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 14 08:57:29 crc kubenswrapper[4870]: I1014 08:57:29.334803 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 14 08:57:29 crc kubenswrapper[4870]: I1014 08:57:29.385713 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 14 08:57:29 crc kubenswrapper[4870]: I1014 08:57:29.407024 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 14 08:57:30 crc kubenswrapper[4870]: I1014 08:57:30.023128 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 14 08:57:30 crc kubenswrapper[4870]: I1014 08:57:30.023171 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 14 08:57:31 crc kubenswrapper[4870]: I1014 08:57:31.265908 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 14 08:57:31 crc kubenswrapper[4870]: I1014 08:57:31.267218 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 14 08:57:31 crc kubenswrapper[4870]: I1014 08:57:31.305268 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 14 08:57:31 crc kubenswrapper[4870]: I1014 08:57:31.322964 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 14 08:57:31 crc kubenswrapper[4870]: I1014 08:57:31.915308 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 14 08:57:31 crc kubenswrapper[4870]: I1014 08:57:31.920586 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 14 08:57:32 crc kubenswrapper[4870]: I1014 08:57:32.043107 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 14 08:57:32 crc kubenswrapper[4870]: I1014 08:57:32.043528 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 14 08:57:33 crc kubenswrapper[4870]: I1014 08:57:33.869201 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 14 08:57:33 crc kubenswrapper[4870]: I1014 08:57:33.869774 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 14 08:57:37 crc kubenswrapper[4870]: I1014 08:57:37.035950 4870 scope.go:117] "RemoveContainer" containerID="8c0f754fe94b7cb7b9bb5dcd613f7e42093968b49509d8ee37e8fd4db89eedd5" Oct 14 08:57:37 crc kubenswrapper[4870]: E1014 08:57:37.036264 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 08:57:40 crc kubenswrapper[4870]: I1014 08:57:40.237500 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-v5vqn"] Oct 14 08:57:40 crc kubenswrapper[4870]: E1014 08:57:40.238182 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3691846-4d55-4390-8cf4-94a9b28e4440" containerName="dnsmasq-dns" Oct 14 08:57:40 crc kubenswrapper[4870]: I1014 08:57:40.238195 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3691846-4d55-4390-8cf4-94a9b28e4440" containerName="dnsmasq-dns" Oct 14 08:57:40 crc kubenswrapper[4870]: E1014 08:57:40.238210 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3691846-4d55-4390-8cf4-94a9b28e4440" containerName="init" Oct 14 08:57:40 crc kubenswrapper[4870]: I1014 08:57:40.238216 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3691846-4d55-4390-8cf4-94a9b28e4440" containerName="init" Oct 14 08:57:40 crc kubenswrapper[4870]: I1014 08:57:40.238384 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="b3691846-4d55-4390-8cf4-94a9b28e4440" containerName="dnsmasq-dns" Oct 14 08:57:40 crc kubenswrapper[4870]: I1014 08:57:40.239003 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-v5vqn" Oct 14 08:57:40 crc kubenswrapper[4870]: I1014 08:57:40.267242 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-v5vqn"] Oct 14 08:57:40 crc kubenswrapper[4870]: I1014 08:57:40.319960 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2jmgx\" (UniqueName: \"kubernetes.io/projected/6e1e7452-1d68-46e7-872f-973667e7d3d6-kube-api-access-2jmgx\") pod \"placement-db-create-v5vqn\" (UID: \"6e1e7452-1d68-46e7-872f-973667e7d3d6\") " pod="openstack/placement-db-create-v5vqn" Oct 14 08:57:40 crc kubenswrapper[4870]: I1014 08:57:40.421314 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2jmgx\" (UniqueName: \"kubernetes.io/projected/6e1e7452-1d68-46e7-872f-973667e7d3d6-kube-api-access-2jmgx\") pod \"placement-db-create-v5vqn\" (UID: \"6e1e7452-1d68-46e7-872f-973667e7d3d6\") " pod="openstack/placement-db-create-v5vqn" Oct 14 08:57:40 crc kubenswrapper[4870]: I1014 08:57:40.461063 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2jmgx\" (UniqueName: \"kubernetes.io/projected/6e1e7452-1d68-46e7-872f-973667e7d3d6-kube-api-access-2jmgx\") pod \"placement-db-create-v5vqn\" (UID: \"6e1e7452-1d68-46e7-872f-973667e7d3d6\") " pod="openstack/placement-db-create-v5vqn" Oct 14 08:57:40 crc kubenswrapper[4870]: I1014 08:57:40.572812 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-v5vqn" Oct 14 08:57:41 crc kubenswrapper[4870]: I1014 08:57:41.029949 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-v5vqn"] Oct 14 08:57:41 crc kubenswrapper[4870]: W1014 08:57:41.033194 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6e1e7452_1d68_46e7_872f_973667e7d3d6.slice/crio-af4f5c3ed751258120a7155602a74425015df371e9eb934f318567b48fbee3c9 WatchSource:0}: Error finding container af4f5c3ed751258120a7155602a74425015df371e9eb934f318567b48fbee3c9: Status 404 returned error can't find the container with id af4f5c3ed751258120a7155602a74425015df371e9eb934f318567b48fbee3c9 Oct 14 08:57:41 crc kubenswrapper[4870]: I1014 08:57:41.126776 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-v5vqn" event={"ID":"6e1e7452-1d68-46e7-872f-973667e7d3d6","Type":"ContainerStarted","Data":"af4f5c3ed751258120a7155602a74425015df371e9eb934f318567b48fbee3c9"} Oct 14 08:57:42 crc kubenswrapper[4870]: I1014 08:57:42.139738 4870 generic.go:334] "Generic (PLEG): container finished" podID="6e1e7452-1d68-46e7-872f-973667e7d3d6" containerID="a26882018748ab7886af431748d8380dfd8592cd004d6110b333d865a49e851c" exitCode=0 Oct 14 08:57:42 crc kubenswrapper[4870]: I1014 08:57:42.139787 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-v5vqn" event={"ID":"6e1e7452-1d68-46e7-872f-973667e7d3d6","Type":"ContainerDied","Data":"a26882018748ab7886af431748d8380dfd8592cd004d6110b333d865a49e851c"} Oct 14 08:57:43 crc kubenswrapper[4870]: I1014 08:57:43.535739 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-v5vqn" Oct 14 08:57:43 crc kubenswrapper[4870]: I1014 08:57:43.578068 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2jmgx\" (UniqueName: \"kubernetes.io/projected/6e1e7452-1d68-46e7-872f-973667e7d3d6-kube-api-access-2jmgx\") pod \"6e1e7452-1d68-46e7-872f-973667e7d3d6\" (UID: \"6e1e7452-1d68-46e7-872f-973667e7d3d6\") " Oct 14 08:57:43 crc kubenswrapper[4870]: I1014 08:57:43.584806 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6e1e7452-1d68-46e7-872f-973667e7d3d6-kube-api-access-2jmgx" (OuterVolumeSpecName: "kube-api-access-2jmgx") pod "6e1e7452-1d68-46e7-872f-973667e7d3d6" (UID: "6e1e7452-1d68-46e7-872f-973667e7d3d6"). InnerVolumeSpecName "kube-api-access-2jmgx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:57:43 crc kubenswrapper[4870]: I1014 08:57:43.680163 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2jmgx\" (UniqueName: \"kubernetes.io/projected/6e1e7452-1d68-46e7-872f-973667e7d3d6-kube-api-access-2jmgx\") on node \"crc\" DevicePath \"\"" Oct 14 08:57:44 crc kubenswrapper[4870]: I1014 08:57:44.163632 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-v5vqn" event={"ID":"6e1e7452-1d68-46e7-872f-973667e7d3d6","Type":"ContainerDied","Data":"af4f5c3ed751258120a7155602a74425015df371e9eb934f318567b48fbee3c9"} Oct 14 08:57:44 crc kubenswrapper[4870]: I1014 08:57:44.163678 4870 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="af4f5c3ed751258120a7155602a74425015df371e9eb934f318567b48fbee3c9" Oct 14 08:57:44 crc kubenswrapper[4870]: I1014 08:57:44.163765 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-v5vqn" Oct 14 08:57:49 crc kubenswrapper[4870]: I1014 08:57:49.034605 4870 scope.go:117] "RemoveContainer" containerID="8c0f754fe94b7cb7b9bb5dcd613f7e42093968b49509d8ee37e8fd4db89eedd5" Oct 14 08:57:49 crc kubenswrapper[4870]: E1014 08:57:49.035298 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 08:57:50 crc kubenswrapper[4870]: I1014 08:57:50.371952 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-bfc7-account-create-9n2mf"] Oct 14 08:57:50 crc kubenswrapper[4870]: E1014 08:57:50.372638 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e1e7452-1d68-46e7-872f-973667e7d3d6" containerName="mariadb-database-create" Oct 14 08:57:50 crc kubenswrapper[4870]: I1014 08:57:50.372654 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e1e7452-1d68-46e7-872f-973667e7d3d6" containerName="mariadb-database-create" Oct 14 08:57:50 crc kubenswrapper[4870]: I1014 08:57:50.372835 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="6e1e7452-1d68-46e7-872f-973667e7d3d6" containerName="mariadb-database-create" Oct 14 08:57:50 crc kubenswrapper[4870]: I1014 08:57:50.373382 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-bfc7-account-create-9n2mf" Oct 14 08:57:50 crc kubenswrapper[4870]: I1014 08:57:50.375319 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Oct 14 08:57:50 crc kubenswrapper[4870]: I1014 08:57:50.385265 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-bfc7-account-create-9n2mf"] Oct 14 08:57:50 crc kubenswrapper[4870]: I1014 08:57:50.404401 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rqfm2\" (UniqueName: \"kubernetes.io/projected/21b0ad50-1397-43e9-97d1-2bd8a4aff236-kube-api-access-rqfm2\") pod \"placement-bfc7-account-create-9n2mf\" (UID: \"21b0ad50-1397-43e9-97d1-2bd8a4aff236\") " pod="openstack/placement-bfc7-account-create-9n2mf" Oct 14 08:57:50 crc kubenswrapper[4870]: I1014 08:57:50.505832 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rqfm2\" (UniqueName: \"kubernetes.io/projected/21b0ad50-1397-43e9-97d1-2bd8a4aff236-kube-api-access-rqfm2\") pod \"placement-bfc7-account-create-9n2mf\" (UID: \"21b0ad50-1397-43e9-97d1-2bd8a4aff236\") " pod="openstack/placement-bfc7-account-create-9n2mf" Oct 14 08:57:50 crc kubenswrapper[4870]: I1014 08:57:50.523376 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rqfm2\" (UniqueName: \"kubernetes.io/projected/21b0ad50-1397-43e9-97d1-2bd8a4aff236-kube-api-access-rqfm2\") pod \"placement-bfc7-account-create-9n2mf\" (UID: \"21b0ad50-1397-43e9-97d1-2bd8a4aff236\") " pod="openstack/placement-bfc7-account-create-9n2mf" Oct 14 08:57:50 crc kubenswrapper[4870]: I1014 08:57:50.728844 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-bfc7-account-create-9n2mf" Oct 14 08:57:51 crc kubenswrapper[4870]: I1014 08:57:51.202611 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-bfc7-account-create-9n2mf"] Oct 14 08:57:51 crc kubenswrapper[4870]: W1014 08:57:51.212026 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod21b0ad50_1397_43e9_97d1_2bd8a4aff236.slice/crio-b53c1ce51476395b663c935b3ce81f6ee4472a20583c276ff03e683fe2cb83e6 WatchSource:0}: Error finding container b53c1ce51476395b663c935b3ce81f6ee4472a20583c276ff03e683fe2cb83e6: Status 404 returned error can't find the container with id b53c1ce51476395b663c935b3ce81f6ee4472a20583c276ff03e683fe2cb83e6 Oct 14 08:57:51 crc kubenswrapper[4870]: I1014 08:57:51.222959 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-bfc7-account-create-9n2mf" event={"ID":"21b0ad50-1397-43e9-97d1-2bd8a4aff236","Type":"ContainerStarted","Data":"b53c1ce51476395b663c935b3ce81f6ee4472a20583c276ff03e683fe2cb83e6"} Oct 14 08:57:52 crc kubenswrapper[4870]: I1014 08:57:52.235717 4870 generic.go:334] "Generic (PLEG): container finished" podID="21b0ad50-1397-43e9-97d1-2bd8a4aff236" containerID="adacb51ba823d60acf3fb6ca55a5842d901122fccb80169da2fe7d33d7baff7c" exitCode=0 Oct 14 08:57:52 crc kubenswrapper[4870]: I1014 08:57:52.236043 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-bfc7-account-create-9n2mf" event={"ID":"21b0ad50-1397-43e9-97d1-2bd8a4aff236","Type":"ContainerDied","Data":"adacb51ba823d60acf3fb6ca55a5842d901122fccb80169da2fe7d33d7baff7c"} Oct 14 08:57:53 crc kubenswrapper[4870]: I1014 08:57:53.662478 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-bfc7-account-create-9n2mf" Oct 14 08:57:53 crc kubenswrapper[4870]: I1014 08:57:53.764500 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rqfm2\" (UniqueName: \"kubernetes.io/projected/21b0ad50-1397-43e9-97d1-2bd8a4aff236-kube-api-access-rqfm2\") pod \"21b0ad50-1397-43e9-97d1-2bd8a4aff236\" (UID: \"21b0ad50-1397-43e9-97d1-2bd8a4aff236\") " Oct 14 08:57:53 crc kubenswrapper[4870]: I1014 08:57:53.771978 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/21b0ad50-1397-43e9-97d1-2bd8a4aff236-kube-api-access-rqfm2" (OuterVolumeSpecName: "kube-api-access-rqfm2") pod "21b0ad50-1397-43e9-97d1-2bd8a4aff236" (UID: "21b0ad50-1397-43e9-97d1-2bd8a4aff236"). InnerVolumeSpecName "kube-api-access-rqfm2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:57:53 crc kubenswrapper[4870]: I1014 08:57:53.867748 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rqfm2\" (UniqueName: \"kubernetes.io/projected/21b0ad50-1397-43e9-97d1-2bd8a4aff236-kube-api-access-rqfm2\") on node \"crc\" DevicePath \"\"" Oct 14 08:57:54 crc kubenswrapper[4870]: I1014 08:57:54.261399 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-bfc7-account-create-9n2mf" event={"ID":"21b0ad50-1397-43e9-97d1-2bd8a4aff236","Type":"ContainerDied","Data":"b53c1ce51476395b663c935b3ce81f6ee4472a20583c276ff03e683fe2cb83e6"} Oct 14 08:57:54 crc kubenswrapper[4870]: I1014 08:57:54.261496 4870 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b53c1ce51476395b663c935b3ce81f6ee4472a20583c276ff03e683fe2cb83e6" Oct 14 08:57:54 crc kubenswrapper[4870]: I1014 08:57:54.261601 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-bfc7-account-create-9n2mf" Oct 14 08:57:55 crc kubenswrapper[4870]: I1014 08:57:55.718181 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-kfqr4"] Oct 14 08:57:55 crc kubenswrapper[4870]: E1014 08:57:55.719863 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="21b0ad50-1397-43e9-97d1-2bd8a4aff236" containerName="mariadb-account-create" Oct 14 08:57:55 crc kubenswrapper[4870]: I1014 08:57:55.719942 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="21b0ad50-1397-43e9-97d1-2bd8a4aff236" containerName="mariadb-account-create" Oct 14 08:57:55 crc kubenswrapper[4870]: I1014 08:57:55.720266 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="21b0ad50-1397-43e9-97d1-2bd8a4aff236" containerName="mariadb-account-create" Oct 14 08:57:55 crc kubenswrapper[4870]: I1014 08:57:55.720953 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-kfqr4" Oct 14 08:57:55 crc kubenswrapper[4870]: I1014 08:57:55.729062 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Oct 14 08:57:55 crc kubenswrapper[4870]: I1014 08:57:55.729249 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Oct 14 08:57:55 crc kubenswrapper[4870]: I1014 08:57:55.729393 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-sx8ps" Oct 14 08:57:55 crc kubenswrapper[4870]: I1014 08:57:55.758330 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-kfqr4"] Oct 14 08:57:55 crc kubenswrapper[4870]: I1014 08:57:55.768542 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7d6744f579-msx66"] Oct 14 08:57:55 crc kubenswrapper[4870]: I1014 08:57:55.770163 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7d6744f579-msx66" Oct 14 08:57:55 crc kubenswrapper[4870]: I1014 08:57:55.785852 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7d6744f579-msx66"] Oct 14 08:57:55 crc kubenswrapper[4870]: I1014 08:57:55.840086 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a04eef9c-afd1-41ae-af51-9d4d97a40a60-ovsdbserver-sb\") pod \"dnsmasq-dns-7d6744f579-msx66\" (UID: \"a04eef9c-afd1-41ae-af51-9d4d97a40a60\") " pod="openstack/dnsmasq-dns-7d6744f579-msx66" Oct 14 08:57:55 crc kubenswrapper[4870]: I1014 08:57:55.840130 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hqwsz\" (UniqueName: \"kubernetes.io/projected/a04eef9c-afd1-41ae-af51-9d4d97a40a60-kube-api-access-hqwsz\") pod \"dnsmasq-dns-7d6744f579-msx66\" (UID: \"a04eef9c-afd1-41ae-af51-9d4d97a40a60\") " pod="openstack/dnsmasq-dns-7d6744f579-msx66" Oct 14 08:57:55 crc kubenswrapper[4870]: I1014 08:57:55.840165 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/32cde066-88bb-4e2b-b276-227282560585-scripts\") pod \"placement-db-sync-kfqr4\" (UID: \"32cde066-88bb-4e2b-b276-227282560585\") " pod="openstack/placement-db-sync-kfqr4" Oct 14 08:57:55 crc kubenswrapper[4870]: I1014 08:57:55.840192 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a04eef9c-afd1-41ae-af51-9d4d97a40a60-dns-svc\") pod \"dnsmasq-dns-7d6744f579-msx66\" (UID: \"a04eef9c-afd1-41ae-af51-9d4d97a40a60\") " pod="openstack/dnsmasq-dns-7d6744f579-msx66" Oct 14 08:57:55 crc kubenswrapper[4870]: I1014 08:57:55.840402 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a04eef9c-afd1-41ae-af51-9d4d97a40a60-config\") pod \"dnsmasq-dns-7d6744f579-msx66\" (UID: \"a04eef9c-afd1-41ae-af51-9d4d97a40a60\") " pod="openstack/dnsmasq-dns-7d6744f579-msx66" Oct 14 08:57:55 crc kubenswrapper[4870]: I1014 08:57:55.840635 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h7f7h\" (UniqueName: \"kubernetes.io/projected/32cde066-88bb-4e2b-b276-227282560585-kube-api-access-h7f7h\") pod \"placement-db-sync-kfqr4\" (UID: \"32cde066-88bb-4e2b-b276-227282560585\") " pod="openstack/placement-db-sync-kfqr4" Oct 14 08:57:55 crc kubenswrapper[4870]: I1014 08:57:55.840706 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a04eef9c-afd1-41ae-af51-9d4d97a40a60-ovsdbserver-nb\") pod \"dnsmasq-dns-7d6744f579-msx66\" (UID: \"a04eef9c-afd1-41ae-af51-9d4d97a40a60\") " pod="openstack/dnsmasq-dns-7d6744f579-msx66" Oct 14 08:57:55 crc kubenswrapper[4870]: I1014 08:57:55.840865 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/32cde066-88bb-4e2b-b276-227282560585-logs\") pod \"placement-db-sync-kfqr4\" (UID: \"32cde066-88bb-4e2b-b276-227282560585\") " pod="openstack/placement-db-sync-kfqr4" Oct 14 08:57:55 crc kubenswrapper[4870]: I1014 08:57:55.840946 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/32cde066-88bb-4e2b-b276-227282560585-config-data\") pod \"placement-db-sync-kfqr4\" (UID: \"32cde066-88bb-4e2b-b276-227282560585\") " pod="openstack/placement-db-sync-kfqr4" Oct 14 08:57:55 crc kubenswrapper[4870]: I1014 08:57:55.841031 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32cde066-88bb-4e2b-b276-227282560585-combined-ca-bundle\") pod \"placement-db-sync-kfqr4\" (UID: \"32cde066-88bb-4e2b-b276-227282560585\") " pod="openstack/placement-db-sync-kfqr4" Oct 14 08:57:55 crc kubenswrapper[4870]: I1014 08:57:55.942699 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/32cde066-88bb-4e2b-b276-227282560585-config-data\") pod \"placement-db-sync-kfqr4\" (UID: \"32cde066-88bb-4e2b-b276-227282560585\") " pod="openstack/placement-db-sync-kfqr4" Oct 14 08:57:55 crc kubenswrapper[4870]: I1014 08:57:55.943087 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32cde066-88bb-4e2b-b276-227282560585-combined-ca-bundle\") pod \"placement-db-sync-kfqr4\" (UID: \"32cde066-88bb-4e2b-b276-227282560585\") " pod="openstack/placement-db-sync-kfqr4" Oct 14 08:57:55 crc kubenswrapper[4870]: I1014 08:57:55.943219 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a04eef9c-afd1-41ae-af51-9d4d97a40a60-ovsdbserver-sb\") pod \"dnsmasq-dns-7d6744f579-msx66\" (UID: \"a04eef9c-afd1-41ae-af51-9d4d97a40a60\") " pod="openstack/dnsmasq-dns-7d6744f579-msx66" Oct 14 08:57:55 crc kubenswrapper[4870]: I1014 08:57:55.943345 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hqwsz\" (UniqueName: \"kubernetes.io/projected/a04eef9c-afd1-41ae-af51-9d4d97a40a60-kube-api-access-hqwsz\") pod \"dnsmasq-dns-7d6744f579-msx66\" (UID: \"a04eef9c-afd1-41ae-af51-9d4d97a40a60\") " pod="openstack/dnsmasq-dns-7d6744f579-msx66" Oct 14 08:57:55 crc kubenswrapper[4870]: I1014 08:57:55.943461 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/32cde066-88bb-4e2b-b276-227282560585-scripts\") pod \"placement-db-sync-kfqr4\" (UID: \"32cde066-88bb-4e2b-b276-227282560585\") " pod="openstack/placement-db-sync-kfqr4" Oct 14 08:57:55 crc kubenswrapper[4870]: I1014 08:57:55.943559 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a04eef9c-afd1-41ae-af51-9d4d97a40a60-dns-svc\") pod \"dnsmasq-dns-7d6744f579-msx66\" (UID: \"a04eef9c-afd1-41ae-af51-9d4d97a40a60\") " pod="openstack/dnsmasq-dns-7d6744f579-msx66" Oct 14 08:57:55 crc kubenswrapper[4870]: I1014 08:57:55.943713 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a04eef9c-afd1-41ae-af51-9d4d97a40a60-config\") pod \"dnsmasq-dns-7d6744f579-msx66\" (UID: \"a04eef9c-afd1-41ae-af51-9d4d97a40a60\") " pod="openstack/dnsmasq-dns-7d6744f579-msx66" Oct 14 08:57:55 crc kubenswrapper[4870]: I1014 08:57:55.943864 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h7f7h\" (UniqueName: \"kubernetes.io/projected/32cde066-88bb-4e2b-b276-227282560585-kube-api-access-h7f7h\") pod \"placement-db-sync-kfqr4\" (UID: \"32cde066-88bb-4e2b-b276-227282560585\") " pod="openstack/placement-db-sync-kfqr4" Oct 14 08:57:55 crc kubenswrapper[4870]: I1014 08:57:55.943968 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a04eef9c-afd1-41ae-af51-9d4d97a40a60-ovsdbserver-nb\") pod \"dnsmasq-dns-7d6744f579-msx66\" (UID: \"a04eef9c-afd1-41ae-af51-9d4d97a40a60\") " pod="openstack/dnsmasq-dns-7d6744f579-msx66" Oct 14 08:57:55 crc kubenswrapper[4870]: I1014 08:57:55.944105 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a04eef9c-afd1-41ae-af51-9d4d97a40a60-ovsdbserver-sb\") pod \"dnsmasq-dns-7d6744f579-msx66\" (UID: \"a04eef9c-afd1-41ae-af51-9d4d97a40a60\") " pod="openstack/dnsmasq-dns-7d6744f579-msx66" Oct 14 08:57:55 crc kubenswrapper[4870]: I1014 08:57:55.944122 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/32cde066-88bb-4e2b-b276-227282560585-logs\") pod \"placement-db-sync-kfqr4\" (UID: \"32cde066-88bb-4e2b-b276-227282560585\") " pod="openstack/placement-db-sync-kfqr4" Oct 14 08:57:55 crc kubenswrapper[4870]: I1014 08:57:55.944472 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a04eef9c-afd1-41ae-af51-9d4d97a40a60-dns-svc\") pod \"dnsmasq-dns-7d6744f579-msx66\" (UID: \"a04eef9c-afd1-41ae-af51-9d4d97a40a60\") " pod="openstack/dnsmasq-dns-7d6744f579-msx66" Oct 14 08:57:55 crc kubenswrapper[4870]: I1014 08:57:55.944552 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a04eef9c-afd1-41ae-af51-9d4d97a40a60-config\") pod \"dnsmasq-dns-7d6744f579-msx66\" (UID: \"a04eef9c-afd1-41ae-af51-9d4d97a40a60\") " pod="openstack/dnsmasq-dns-7d6744f579-msx66" Oct 14 08:57:55 crc kubenswrapper[4870]: I1014 08:57:55.944788 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a04eef9c-afd1-41ae-af51-9d4d97a40a60-ovsdbserver-nb\") pod \"dnsmasq-dns-7d6744f579-msx66\" (UID: \"a04eef9c-afd1-41ae-af51-9d4d97a40a60\") " pod="openstack/dnsmasq-dns-7d6744f579-msx66" Oct 14 08:57:55 crc kubenswrapper[4870]: I1014 08:57:55.944973 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/32cde066-88bb-4e2b-b276-227282560585-logs\") pod \"placement-db-sync-kfqr4\" (UID: \"32cde066-88bb-4e2b-b276-227282560585\") " pod="openstack/placement-db-sync-kfqr4" Oct 14 08:57:55 crc kubenswrapper[4870]: I1014 08:57:55.951026 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/32cde066-88bb-4e2b-b276-227282560585-scripts\") pod \"placement-db-sync-kfqr4\" (UID: \"32cde066-88bb-4e2b-b276-227282560585\") " pod="openstack/placement-db-sync-kfqr4" Oct 14 08:57:55 crc kubenswrapper[4870]: I1014 08:57:55.953128 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32cde066-88bb-4e2b-b276-227282560585-combined-ca-bundle\") pod \"placement-db-sync-kfqr4\" (UID: \"32cde066-88bb-4e2b-b276-227282560585\") " pod="openstack/placement-db-sync-kfqr4" Oct 14 08:57:55 crc kubenswrapper[4870]: I1014 08:57:55.959287 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/32cde066-88bb-4e2b-b276-227282560585-config-data\") pod \"placement-db-sync-kfqr4\" (UID: \"32cde066-88bb-4e2b-b276-227282560585\") " pod="openstack/placement-db-sync-kfqr4" Oct 14 08:57:55 crc kubenswrapper[4870]: I1014 08:57:55.960465 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hqwsz\" (UniqueName: \"kubernetes.io/projected/a04eef9c-afd1-41ae-af51-9d4d97a40a60-kube-api-access-hqwsz\") pod \"dnsmasq-dns-7d6744f579-msx66\" (UID: \"a04eef9c-afd1-41ae-af51-9d4d97a40a60\") " pod="openstack/dnsmasq-dns-7d6744f579-msx66" Oct 14 08:57:55 crc kubenswrapper[4870]: I1014 08:57:55.970755 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h7f7h\" (UniqueName: \"kubernetes.io/projected/32cde066-88bb-4e2b-b276-227282560585-kube-api-access-h7f7h\") pod \"placement-db-sync-kfqr4\" (UID: \"32cde066-88bb-4e2b-b276-227282560585\") " pod="openstack/placement-db-sync-kfqr4" Oct 14 08:57:56 crc kubenswrapper[4870]: I1014 08:57:56.056956 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-kfqr4" Oct 14 08:57:56 crc kubenswrapper[4870]: I1014 08:57:56.091764 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7d6744f579-msx66" Oct 14 08:57:56 crc kubenswrapper[4870]: I1014 08:57:56.549367 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-kfqr4"] Oct 14 08:57:56 crc kubenswrapper[4870]: I1014 08:57:56.555621 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7d6744f579-msx66"] Oct 14 08:57:57 crc kubenswrapper[4870]: I1014 08:57:57.296909 4870 generic.go:334] "Generic (PLEG): container finished" podID="a04eef9c-afd1-41ae-af51-9d4d97a40a60" containerID="a973acaa26add648ee12bf8f9f22a327b5650bca2d8d38945dbf06f00f26b639" exitCode=0 Oct 14 08:57:57 crc kubenswrapper[4870]: I1014 08:57:57.297044 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d6744f579-msx66" event={"ID":"a04eef9c-afd1-41ae-af51-9d4d97a40a60","Type":"ContainerDied","Data":"a973acaa26add648ee12bf8f9f22a327b5650bca2d8d38945dbf06f00f26b639"} Oct 14 08:57:57 crc kubenswrapper[4870]: I1014 08:57:57.297948 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d6744f579-msx66" event={"ID":"a04eef9c-afd1-41ae-af51-9d4d97a40a60","Type":"ContainerStarted","Data":"3372c8de6d8608ed6d47e15fc528ec3abf81705c7c79f25cde49586b5998cff1"} Oct 14 08:57:57 crc kubenswrapper[4870]: I1014 08:57:57.301077 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-kfqr4" event={"ID":"32cde066-88bb-4e2b-b276-227282560585","Type":"ContainerStarted","Data":"a5b9e6865b0fbb6dd389820c006c1aeb5e9690e79e926c27b3daeaf2680d4cf5"} Oct 14 08:57:58 crc kubenswrapper[4870]: I1014 08:57:58.314491 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d6744f579-msx66" event={"ID":"a04eef9c-afd1-41ae-af51-9d4d97a40a60","Type":"ContainerStarted","Data":"6271a06ee44d0217f821148cf6af1d6eee1c6f9527fc3bf11f0f48dd83e8facd"} Oct 14 08:57:58 crc kubenswrapper[4870]: I1014 08:57:58.315894 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7d6744f579-msx66" Oct 14 08:57:58 crc kubenswrapper[4870]: I1014 08:57:58.341870 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7d6744f579-msx66" podStartSLOduration=3.341849523 podStartE2EDuration="3.341849523s" podCreationTimestamp="2025-10-14 08:57:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 08:57:58.328435784 +0000 UTC m=+7014.025796165" watchObservedRunningTime="2025-10-14 08:57:58.341849523 +0000 UTC m=+7014.039209894" Oct 14 08:58:00 crc kubenswrapper[4870]: I1014 08:58:00.330558 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-kfqr4" event={"ID":"32cde066-88bb-4e2b-b276-227282560585","Type":"ContainerStarted","Data":"94d794a728ce92614be836c095267b61dda5f462fbda98a73e90746bc7ee4b6a"} Oct 14 08:58:00 crc kubenswrapper[4870]: I1014 08:58:00.350058 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-kfqr4" podStartSLOduration=1.997898555 podStartE2EDuration="5.350034134s" podCreationTimestamp="2025-10-14 08:57:55 +0000 UTC" firstStartedPulling="2025-10-14 08:57:56.559400315 +0000 UTC m=+7012.256760686" lastFinishedPulling="2025-10-14 08:57:59.911535894 +0000 UTC m=+7015.608896265" observedRunningTime="2025-10-14 08:58:00.342732075 +0000 UTC m=+7016.040092446" watchObservedRunningTime="2025-10-14 08:58:00.350034134 +0000 UTC m=+7016.047394505" Oct 14 08:58:01 crc kubenswrapper[4870]: I1014 08:58:01.342491 4870 generic.go:334] "Generic (PLEG): container finished" podID="32cde066-88bb-4e2b-b276-227282560585" containerID="94d794a728ce92614be836c095267b61dda5f462fbda98a73e90746bc7ee4b6a" exitCode=0 Oct 14 08:58:01 crc kubenswrapper[4870]: I1014 08:58:01.342556 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-kfqr4" event={"ID":"32cde066-88bb-4e2b-b276-227282560585","Type":"ContainerDied","Data":"94d794a728ce92614be836c095267b61dda5f462fbda98a73e90746bc7ee4b6a"} Oct 14 08:58:02 crc kubenswrapper[4870]: I1014 08:58:02.699387 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-kfqr4" Oct 14 08:58:02 crc kubenswrapper[4870]: I1014 08:58:02.770862 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/32cde066-88bb-4e2b-b276-227282560585-logs\") pod \"32cde066-88bb-4e2b-b276-227282560585\" (UID: \"32cde066-88bb-4e2b-b276-227282560585\") " Oct 14 08:58:02 crc kubenswrapper[4870]: I1014 08:58:02.770951 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/32cde066-88bb-4e2b-b276-227282560585-scripts\") pod \"32cde066-88bb-4e2b-b276-227282560585\" (UID: \"32cde066-88bb-4e2b-b276-227282560585\") " Oct 14 08:58:02 crc kubenswrapper[4870]: I1014 08:58:02.771171 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32cde066-88bb-4e2b-b276-227282560585-combined-ca-bundle\") pod \"32cde066-88bb-4e2b-b276-227282560585\" (UID: \"32cde066-88bb-4e2b-b276-227282560585\") " Oct 14 08:58:02 crc kubenswrapper[4870]: I1014 08:58:02.771211 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h7f7h\" (UniqueName: \"kubernetes.io/projected/32cde066-88bb-4e2b-b276-227282560585-kube-api-access-h7f7h\") pod \"32cde066-88bb-4e2b-b276-227282560585\" (UID: \"32cde066-88bb-4e2b-b276-227282560585\") " Oct 14 08:58:02 crc kubenswrapper[4870]: I1014 08:58:02.771242 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/32cde066-88bb-4e2b-b276-227282560585-config-data\") pod \"32cde066-88bb-4e2b-b276-227282560585\" (UID: \"32cde066-88bb-4e2b-b276-227282560585\") " Oct 14 08:58:02 crc kubenswrapper[4870]: I1014 08:58:02.772936 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/32cde066-88bb-4e2b-b276-227282560585-logs" (OuterVolumeSpecName: "logs") pod "32cde066-88bb-4e2b-b276-227282560585" (UID: "32cde066-88bb-4e2b-b276-227282560585"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:58:02 crc kubenswrapper[4870]: I1014 08:58:02.777339 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/32cde066-88bb-4e2b-b276-227282560585-scripts" (OuterVolumeSpecName: "scripts") pod "32cde066-88bb-4e2b-b276-227282560585" (UID: "32cde066-88bb-4e2b-b276-227282560585"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:58:02 crc kubenswrapper[4870]: I1014 08:58:02.777900 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/32cde066-88bb-4e2b-b276-227282560585-kube-api-access-h7f7h" (OuterVolumeSpecName: "kube-api-access-h7f7h") pod "32cde066-88bb-4e2b-b276-227282560585" (UID: "32cde066-88bb-4e2b-b276-227282560585"). InnerVolumeSpecName "kube-api-access-h7f7h". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:58:02 crc kubenswrapper[4870]: I1014 08:58:02.803370 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/32cde066-88bb-4e2b-b276-227282560585-config-data" (OuterVolumeSpecName: "config-data") pod "32cde066-88bb-4e2b-b276-227282560585" (UID: "32cde066-88bb-4e2b-b276-227282560585"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:58:02 crc kubenswrapper[4870]: I1014 08:58:02.805344 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/32cde066-88bb-4e2b-b276-227282560585-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "32cde066-88bb-4e2b-b276-227282560585" (UID: "32cde066-88bb-4e2b-b276-227282560585"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:58:02 crc kubenswrapper[4870]: I1014 08:58:02.872853 4870 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32cde066-88bb-4e2b-b276-227282560585-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 08:58:02 crc kubenswrapper[4870]: I1014 08:58:02.872889 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h7f7h\" (UniqueName: \"kubernetes.io/projected/32cde066-88bb-4e2b-b276-227282560585-kube-api-access-h7f7h\") on node \"crc\" DevicePath \"\"" Oct 14 08:58:02 crc kubenswrapper[4870]: I1014 08:58:02.872902 4870 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/32cde066-88bb-4e2b-b276-227282560585-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 08:58:02 crc kubenswrapper[4870]: I1014 08:58:02.872914 4870 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/32cde066-88bb-4e2b-b276-227282560585-logs\") on node \"crc\" DevicePath \"\"" Oct 14 08:58:02 crc kubenswrapper[4870]: I1014 08:58:02.872923 4870 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/32cde066-88bb-4e2b-b276-227282560585-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 08:58:03 crc kubenswrapper[4870]: I1014 08:58:03.034793 4870 scope.go:117] "RemoveContainer" containerID="8c0f754fe94b7cb7b9bb5dcd613f7e42093968b49509d8ee37e8fd4db89eedd5" Oct 14 08:58:03 crc kubenswrapper[4870]: E1014 08:58:03.035666 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 08:58:03 crc kubenswrapper[4870]: I1014 08:58:03.365137 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-kfqr4" event={"ID":"32cde066-88bb-4e2b-b276-227282560585","Type":"ContainerDied","Data":"a5b9e6865b0fbb6dd389820c006c1aeb5e9690e79e926c27b3daeaf2680d4cf5"} Oct 14 08:58:03 crc kubenswrapper[4870]: I1014 08:58:03.365524 4870 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a5b9e6865b0fbb6dd389820c006c1aeb5e9690e79e926c27b3daeaf2680d4cf5" Oct 14 08:58:03 crc kubenswrapper[4870]: I1014 08:58:03.365263 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-kfqr4" Oct 14 08:58:03 crc kubenswrapper[4870]: I1014 08:58:03.456981 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-594cc89896-9lvzl"] Oct 14 08:58:03 crc kubenswrapper[4870]: E1014 08:58:03.457463 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="32cde066-88bb-4e2b-b276-227282560585" containerName="placement-db-sync" Oct 14 08:58:03 crc kubenswrapper[4870]: I1014 08:58:03.457491 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="32cde066-88bb-4e2b-b276-227282560585" containerName="placement-db-sync" Oct 14 08:58:03 crc kubenswrapper[4870]: I1014 08:58:03.457748 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="32cde066-88bb-4e2b-b276-227282560585" containerName="placement-db-sync" Oct 14 08:58:03 crc kubenswrapper[4870]: I1014 08:58:03.458918 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-594cc89896-9lvzl" Oct 14 08:58:03 crc kubenswrapper[4870]: I1014 08:58:03.462194 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-sx8ps" Oct 14 08:58:03 crc kubenswrapper[4870]: I1014 08:58:03.465931 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Oct 14 08:58:03 crc kubenswrapper[4870]: I1014 08:58:03.468824 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Oct 14 08:58:03 crc kubenswrapper[4870]: I1014 08:58:03.485524 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/21457f84-984b-4745-823d-d4fb6bb2a6c0-config-data\") pod \"placement-594cc89896-9lvzl\" (UID: \"21457f84-984b-4745-823d-d4fb6bb2a6c0\") " pod="openstack/placement-594cc89896-9lvzl" Oct 14 08:58:03 crc kubenswrapper[4870]: I1014 08:58:03.485743 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-clb82\" (UniqueName: \"kubernetes.io/projected/21457f84-984b-4745-823d-d4fb6bb2a6c0-kube-api-access-clb82\") pod \"placement-594cc89896-9lvzl\" (UID: \"21457f84-984b-4745-823d-d4fb6bb2a6c0\") " pod="openstack/placement-594cc89896-9lvzl" Oct 14 08:58:03 crc kubenswrapper[4870]: I1014 08:58:03.485794 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21457f84-984b-4745-823d-d4fb6bb2a6c0-combined-ca-bundle\") pod \"placement-594cc89896-9lvzl\" (UID: \"21457f84-984b-4745-823d-d4fb6bb2a6c0\") " pod="openstack/placement-594cc89896-9lvzl" Oct 14 08:58:03 crc kubenswrapper[4870]: I1014 08:58:03.485912 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/21457f84-984b-4745-823d-d4fb6bb2a6c0-scripts\") pod \"placement-594cc89896-9lvzl\" (UID: \"21457f84-984b-4745-823d-d4fb6bb2a6c0\") " pod="openstack/placement-594cc89896-9lvzl" Oct 14 08:58:03 crc kubenswrapper[4870]: I1014 08:58:03.485949 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/21457f84-984b-4745-823d-d4fb6bb2a6c0-logs\") pod \"placement-594cc89896-9lvzl\" (UID: \"21457f84-984b-4745-823d-d4fb6bb2a6c0\") " pod="openstack/placement-594cc89896-9lvzl" Oct 14 08:58:03 crc kubenswrapper[4870]: I1014 08:58:03.490196 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-594cc89896-9lvzl"] Oct 14 08:58:03 crc kubenswrapper[4870]: I1014 08:58:03.588502 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/21457f84-984b-4745-823d-d4fb6bb2a6c0-config-data\") pod \"placement-594cc89896-9lvzl\" (UID: \"21457f84-984b-4745-823d-d4fb6bb2a6c0\") " pod="openstack/placement-594cc89896-9lvzl" Oct 14 08:58:03 crc kubenswrapper[4870]: I1014 08:58:03.588615 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-clb82\" (UniqueName: \"kubernetes.io/projected/21457f84-984b-4745-823d-d4fb6bb2a6c0-kube-api-access-clb82\") pod \"placement-594cc89896-9lvzl\" (UID: \"21457f84-984b-4745-823d-d4fb6bb2a6c0\") " pod="openstack/placement-594cc89896-9lvzl" Oct 14 08:58:03 crc kubenswrapper[4870]: I1014 08:58:03.588650 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21457f84-984b-4745-823d-d4fb6bb2a6c0-combined-ca-bundle\") pod \"placement-594cc89896-9lvzl\" (UID: \"21457f84-984b-4745-823d-d4fb6bb2a6c0\") " pod="openstack/placement-594cc89896-9lvzl" Oct 14 08:58:03 crc kubenswrapper[4870]: I1014 08:58:03.588719 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/21457f84-984b-4745-823d-d4fb6bb2a6c0-scripts\") pod \"placement-594cc89896-9lvzl\" (UID: \"21457f84-984b-4745-823d-d4fb6bb2a6c0\") " pod="openstack/placement-594cc89896-9lvzl" Oct 14 08:58:03 crc kubenswrapper[4870]: I1014 08:58:03.588747 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/21457f84-984b-4745-823d-d4fb6bb2a6c0-logs\") pod \"placement-594cc89896-9lvzl\" (UID: \"21457f84-984b-4745-823d-d4fb6bb2a6c0\") " pod="openstack/placement-594cc89896-9lvzl" Oct 14 08:58:03 crc kubenswrapper[4870]: I1014 08:58:03.589339 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/21457f84-984b-4745-823d-d4fb6bb2a6c0-logs\") pod \"placement-594cc89896-9lvzl\" (UID: \"21457f84-984b-4745-823d-d4fb6bb2a6c0\") " pod="openstack/placement-594cc89896-9lvzl" Oct 14 08:58:03 crc kubenswrapper[4870]: I1014 08:58:03.602230 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/21457f84-984b-4745-823d-d4fb6bb2a6c0-scripts\") pod \"placement-594cc89896-9lvzl\" (UID: \"21457f84-984b-4745-823d-d4fb6bb2a6c0\") " pod="openstack/placement-594cc89896-9lvzl" Oct 14 08:58:03 crc kubenswrapper[4870]: I1014 08:58:03.613635 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/21457f84-984b-4745-823d-d4fb6bb2a6c0-config-data\") pod \"placement-594cc89896-9lvzl\" (UID: \"21457f84-984b-4745-823d-d4fb6bb2a6c0\") " pod="openstack/placement-594cc89896-9lvzl" Oct 14 08:58:03 crc kubenswrapper[4870]: I1014 08:58:03.616723 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-clb82\" (UniqueName: \"kubernetes.io/projected/21457f84-984b-4745-823d-d4fb6bb2a6c0-kube-api-access-clb82\") pod \"placement-594cc89896-9lvzl\" (UID: \"21457f84-984b-4745-823d-d4fb6bb2a6c0\") " pod="openstack/placement-594cc89896-9lvzl" Oct 14 08:58:03 crc kubenswrapper[4870]: I1014 08:58:03.617338 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21457f84-984b-4745-823d-d4fb6bb2a6c0-combined-ca-bundle\") pod \"placement-594cc89896-9lvzl\" (UID: \"21457f84-984b-4745-823d-d4fb6bb2a6c0\") " pod="openstack/placement-594cc89896-9lvzl" Oct 14 08:58:03 crc kubenswrapper[4870]: I1014 08:58:03.775668 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-594cc89896-9lvzl" Oct 14 08:58:04 crc kubenswrapper[4870]: W1014 08:58:04.287692 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod21457f84_984b_4745_823d_d4fb6bb2a6c0.slice/crio-024a6dfd9955c0f97b06be753ac90ee50d67e4d5ef3b199e44bf9468dc3bf09c WatchSource:0}: Error finding container 024a6dfd9955c0f97b06be753ac90ee50d67e4d5ef3b199e44bf9468dc3bf09c: Status 404 returned error can't find the container with id 024a6dfd9955c0f97b06be753ac90ee50d67e4d5ef3b199e44bf9468dc3bf09c Oct 14 08:58:04 crc kubenswrapper[4870]: I1014 08:58:04.294263 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-594cc89896-9lvzl"] Oct 14 08:58:04 crc kubenswrapper[4870]: I1014 08:58:04.375894 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-594cc89896-9lvzl" event={"ID":"21457f84-984b-4745-823d-d4fb6bb2a6c0","Type":"ContainerStarted","Data":"024a6dfd9955c0f97b06be753ac90ee50d67e4d5ef3b199e44bf9468dc3bf09c"} Oct 14 08:58:05 crc kubenswrapper[4870]: I1014 08:58:05.387507 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-594cc89896-9lvzl" event={"ID":"21457f84-984b-4745-823d-d4fb6bb2a6c0","Type":"ContainerStarted","Data":"14cca7b7b2c271a2076ae20a82d63561b84ab4263932b5e0ecf3116d274d5de7"} Oct 14 08:58:05 crc kubenswrapper[4870]: I1014 08:58:05.387908 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-594cc89896-9lvzl" event={"ID":"21457f84-984b-4745-823d-d4fb6bb2a6c0","Type":"ContainerStarted","Data":"94d15c3f71175a8bc24c30d298fb7ac286fe9577ef59271495de82d24ae04c3f"} Oct 14 08:58:05 crc kubenswrapper[4870]: I1014 08:58:05.389583 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-594cc89896-9lvzl" Oct 14 08:58:05 crc kubenswrapper[4870]: I1014 08:58:05.389624 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-594cc89896-9lvzl" Oct 14 08:58:05 crc kubenswrapper[4870]: I1014 08:58:05.409940 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-594cc89896-9lvzl" podStartSLOduration=2.409917319 podStartE2EDuration="2.409917319s" podCreationTimestamp="2025-10-14 08:58:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 08:58:05.407873938 +0000 UTC m=+7021.105234349" watchObservedRunningTime="2025-10-14 08:58:05.409917319 +0000 UTC m=+7021.107277720" Oct 14 08:58:06 crc kubenswrapper[4870]: I1014 08:58:06.093789 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7d6744f579-msx66" Oct 14 08:58:06 crc kubenswrapper[4870]: I1014 08:58:06.178063 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6695878787-8ngcl"] Oct 14 08:58:06 crc kubenswrapper[4870]: I1014 08:58:06.178708 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6695878787-8ngcl" podUID="37b9bd38-e85a-452c-97f6-67d5d8d3958f" containerName="dnsmasq-dns" containerID="cri-o://436e31f0c5e68792f3f32f7723f0da2b5d815ce570cdc562c44058b35d0bc9df" gracePeriod=10 Oct 14 08:58:06 crc kubenswrapper[4870]: I1014 08:58:06.397833 4870 generic.go:334] "Generic (PLEG): container finished" podID="37b9bd38-e85a-452c-97f6-67d5d8d3958f" containerID="436e31f0c5e68792f3f32f7723f0da2b5d815ce570cdc562c44058b35d0bc9df" exitCode=0 Oct 14 08:58:06 crc kubenswrapper[4870]: I1014 08:58:06.397925 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6695878787-8ngcl" event={"ID":"37b9bd38-e85a-452c-97f6-67d5d8d3958f","Type":"ContainerDied","Data":"436e31f0c5e68792f3f32f7723f0da2b5d815ce570cdc562c44058b35d0bc9df"} Oct 14 08:58:06 crc kubenswrapper[4870]: I1014 08:58:06.635094 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6695878787-8ngcl" Oct 14 08:58:06 crc kubenswrapper[4870]: I1014 08:58:06.752894 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/37b9bd38-e85a-452c-97f6-67d5d8d3958f-ovsdbserver-nb\") pod \"37b9bd38-e85a-452c-97f6-67d5d8d3958f\" (UID: \"37b9bd38-e85a-452c-97f6-67d5d8d3958f\") " Oct 14 08:58:06 crc kubenswrapper[4870]: I1014 08:58:06.752932 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/37b9bd38-e85a-452c-97f6-67d5d8d3958f-dns-svc\") pod \"37b9bd38-e85a-452c-97f6-67d5d8d3958f\" (UID: \"37b9bd38-e85a-452c-97f6-67d5d8d3958f\") " Oct 14 08:58:06 crc kubenswrapper[4870]: I1014 08:58:06.752979 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nhmld\" (UniqueName: \"kubernetes.io/projected/37b9bd38-e85a-452c-97f6-67d5d8d3958f-kube-api-access-nhmld\") pod \"37b9bd38-e85a-452c-97f6-67d5d8d3958f\" (UID: \"37b9bd38-e85a-452c-97f6-67d5d8d3958f\") " Oct 14 08:58:06 crc kubenswrapper[4870]: I1014 08:58:06.752998 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/37b9bd38-e85a-452c-97f6-67d5d8d3958f-ovsdbserver-sb\") pod \"37b9bd38-e85a-452c-97f6-67d5d8d3958f\" (UID: \"37b9bd38-e85a-452c-97f6-67d5d8d3958f\") " Oct 14 08:58:06 crc kubenswrapper[4870]: I1014 08:58:06.753069 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/37b9bd38-e85a-452c-97f6-67d5d8d3958f-config\") pod \"37b9bd38-e85a-452c-97f6-67d5d8d3958f\" (UID: \"37b9bd38-e85a-452c-97f6-67d5d8d3958f\") " Oct 14 08:58:06 crc kubenswrapper[4870]: I1014 08:58:06.759158 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/37b9bd38-e85a-452c-97f6-67d5d8d3958f-kube-api-access-nhmld" (OuterVolumeSpecName: "kube-api-access-nhmld") pod "37b9bd38-e85a-452c-97f6-67d5d8d3958f" (UID: "37b9bd38-e85a-452c-97f6-67d5d8d3958f"). InnerVolumeSpecName "kube-api-access-nhmld". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:58:06 crc kubenswrapper[4870]: I1014 08:58:06.796936 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/37b9bd38-e85a-452c-97f6-67d5d8d3958f-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "37b9bd38-e85a-452c-97f6-67d5d8d3958f" (UID: "37b9bd38-e85a-452c-97f6-67d5d8d3958f"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:58:06 crc kubenswrapper[4870]: I1014 08:58:06.810259 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/37b9bd38-e85a-452c-97f6-67d5d8d3958f-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "37b9bd38-e85a-452c-97f6-67d5d8d3958f" (UID: "37b9bd38-e85a-452c-97f6-67d5d8d3958f"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:58:06 crc kubenswrapper[4870]: I1014 08:58:06.811594 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/37b9bd38-e85a-452c-97f6-67d5d8d3958f-config" (OuterVolumeSpecName: "config") pod "37b9bd38-e85a-452c-97f6-67d5d8d3958f" (UID: "37b9bd38-e85a-452c-97f6-67d5d8d3958f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:58:06 crc kubenswrapper[4870]: I1014 08:58:06.822796 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/37b9bd38-e85a-452c-97f6-67d5d8d3958f-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "37b9bd38-e85a-452c-97f6-67d5d8d3958f" (UID: "37b9bd38-e85a-452c-97f6-67d5d8d3958f"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:58:06 crc kubenswrapper[4870]: I1014 08:58:06.854415 4870 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/37b9bd38-e85a-452c-97f6-67d5d8d3958f-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 14 08:58:06 crc kubenswrapper[4870]: I1014 08:58:06.854600 4870 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/37b9bd38-e85a-452c-97f6-67d5d8d3958f-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 14 08:58:06 crc kubenswrapper[4870]: I1014 08:58:06.854614 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nhmld\" (UniqueName: \"kubernetes.io/projected/37b9bd38-e85a-452c-97f6-67d5d8d3958f-kube-api-access-nhmld\") on node \"crc\" DevicePath \"\"" Oct 14 08:58:06 crc kubenswrapper[4870]: I1014 08:58:06.854627 4870 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/37b9bd38-e85a-452c-97f6-67d5d8d3958f-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 14 08:58:06 crc kubenswrapper[4870]: I1014 08:58:06.854636 4870 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/37b9bd38-e85a-452c-97f6-67d5d8d3958f-config\") on node \"crc\" DevicePath \"\"" Oct 14 08:58:07 crc kubenswrapper[4870]: I1014 08:58:07.410020 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6695878787-8ngcl" event={"ID":"37b9bd38-e85a-452c-97f6-67d5d8d3958f","Type":"ContainerDied","Data":"c49a35593276d0c662a20854447b8fcbf40fddf54f7566401f170f0024820608"} Oct 14 08:58:07 crc kubenswrapper[4870]: I1014 08:58:07.410061 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6695878787-8ngcl" Oct 14 08:58:07 crc kubenswrapper[4870]: I1014 08:58:07.410083 4870 scope.go:117] "RemoveContainer" containerID="436e31f0c5e68792f3f32f7723f0da2b5d815ce570cdc562c44058b35d0bc9df" Oct 14 08:58:07 crc kubenswrapper[4870]: I1014 08:58:07.449411 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6695878787-8ngcl"] Oct 14 08:58:07 crc kubenswrapper[4870]: I1014 08:58:07.449775 4870 scope.go:117] "RemoveContainer" containerID="9a6b519117c6b878a44d820c3e19189970be4558a4f74d53ac33c5ecca6fbee4" Oct 14 08:58:07 crc kubenswrapper[4870]: I1014 08:58:07.471320 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6695878787-8ngcl"] Oct 14 08:58:09 crc kubenswrapper[4870]: I1014 08:58:09.051401 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="37b9bd38-e85a-452c-97f6-67d5d8d3958f" path="/var/lib/kubelet/pods/37b9bd38-e85a-452c-97f6-67d5d8d3958f/volumes" Oct 14 08:58:14 crc kubenswrapper[4870]: I1014 08:58:14.034530 4870 scope.go:117] "RemoveContainer" containerID="8c0f754fe94b7cb7b9bb5dcd613f7e42093968b49509d8ee37e8fd4db89eedd5" Oct 14 08:58:14 crc kubenswrapper[4870]: E1014 08:58:14.035664 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 08:58:19 crc kubenswrapper[4870]: I1014 08:58:19.812868 4870 scope.go:117] "RemoveContainer" containerID="63f04964d083df370a45f2e9b29453b85180d36eb0ed7ccd7ecd06cbfc634472" Oct 14 08:58:27 crc kubenswrapper[4870]: I1014 08:58:27.035523 4870 scope.go:117] "RemoveContainer" containerID="8c0f754fe94b7cb7b9bb5dcd613f7e42093968b49509d8ee37e8fd4db89eedd5" Oct 14 08:58:27 crc kubenswrapper[4870]: I1014 08:58:27.713706 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" event={"ID":"874a5e45-dffd-4d17-b609-b7d3ed2eab07","Type":"ContainerStarted","Data":"9426cde7309378345e6129005b55744e852dbf5b2ef5a92cc907fc57d1dc3ae8"} Oct 14 08:58:34 crc kubenswrapper[4870]: I1014 08:58:34.791904 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-594cc89896-9lvzl" Oct 14 08:58:34 crc kubenswrapper[4870]: I1014 08:58:34.854999 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-594cc89896-9lvzl" Oct 14 08:59:00 crc kubenswrapper[4870]: I1014 08:59:00.499597 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-9htzx"] Oct 14 08:59:00 crc kubenswrapper[4870]: E1014 08:59:00.500377 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37b9bd38-e85a-452c-97f6-67d5d8d3958f" containerName="init" Oct 14 08:59:00 crc kubenswrapper[4870]: I1014 08:59:00.500389 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="37b9bd38-e85a-452c-97f6-67d5d8d3958f" containerName="init" Oct 14 08:59:00 crc kubenswrapper[4870]: E1014 08:59:00.500411 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37b9bd38-e85a-452c-97f6-67d5d8d3958f" containerName="dnsmasq-dns" Oct 14 08:59:00 crc kubenswrapper[4870]: I1014 08:59:00.500418 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="37b9bd38-e85a-452c-97f6-67d5d8d3958f" containerName="dnsmasq-dns" Oct 14 08:59:00 crc kubenswrapper[4870]: I1014 08:59:00.500625 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="37b9bd38-e85a-452c-97f6-67d5d8d3958f" containerName="dnsmasq-dns" Oct 14 08:59:00 crc kubenswrapper[4870]: I1014 08:59:00.501236 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-9htzx" Oct 14 08:59:00 crc kubenswrapper[4870]: I1014 08:59:00.518380 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-9htzx"] Oct 14 08:59:00 crc kubenswrapper[4870]: I1014 08:59:00.592531 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-scw7l"] Oct 14 08:59:00 crc kubenswrapper[4870]: I1014 08:59:00.593789 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-scw7l" Oct 14 08:59:00 crc kubenswrapper[4870]: I1014 08:59:00.603411 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-scw7l"] Oct 14 08:59:00 crc kubenswrapper[4870]: I1014 08:59:00.640919 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dmphw\" (UniqueName: \"kubernetes.io/projected/16fe9c99-c775-4b99-8132-535f9daa9a9f-kube-api-access-dmphw\") pod \"nova-api-db-create-9htzx\" (UID: \"16fe9c99-c775-4b99-8132-535f9daa9a9f\") " pod="openstack/nova-api-db-create-9htzx" Oct 14 08:59:00 crc kubenswrapper[4870]: I1014 08:59:00.698520 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-p9h2b"] Oct 14 08:59:00 crc kubenswrapper[4870]: I1014 08:59:00.699957 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-p9h2b" Oct 14 08:59:00 crc kubenswrapper[4870]: I1014 08:59:00.709641 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-p9h2b"] Oct 14 08:59:00 crc kubenswrapper[4870]: I1014 08:59:00.742747 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-76jpt\" (UniqueName: \"kubernetes.io/projected/8fba7d20-e48c-4e8e-b0a6-21ed0aea5ab1-kube-api-access-76jpt\") pod \"nova-cell0-db-create-scw7l\" (UID: \"8fba7d20-e48c-4e8e-b0a6-21ed0aea5ab1\") " pod="openstack/nova-cell0-db-create-scw7l" Oct 14 08:59:00 crc kubenswrapper[4870]: I1014 08:59:00.742809 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dmphw\" (UniqueName: \"kubernetes.io/projected/16fe9c99-c775-4b99-8132-535f9daa9a9f-kube-api-access-dmphw\") pod \"nova-api-db-create-9htzx\" (UID: \"16fe9c99-c775-4b99-8132-535f9daa9a9f\") " pod="openstack/nova-api-db-create-9htzx" Oct 14 08:59:00 crc kubenswrapper[4870]: I1014 08:59:00.766795 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dmphw\" (UniqueName: \"kubernetes.io/projected/16fe9c99-c775-4b99-8132-535f9daa9a9f-kube-api-access-dmphw\") pod \"nova-api-db-create-9htzx\" (UID: \"16fe9c99-c775-4b99-8132-535f9daa9a9f\") " pod="openstack/nova-api-db-create-9htzx" Oct 14 08:59:00 crc kubenswrapper[4870]: I1014 08:59:00.818005 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-9htzx" Oct 14 08:59:00 crc kubenswrapper[4870]: I1014 08:59:00.844472 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-76jpt\" (UniqueName: \"kubernetes.io/projected/8fba7d20-e48c-4e8e-b0a6-21ed0aea5ab1-kube-api-access-76jpt\") pod \"nova-cell0-db-create-scw7l\" (UID: \"8fba7d20-e48c-4e8e-b0a6-21ed0aea5ab1\") " pod="openstack/nova-cell0-db-create-scw7l" Oct 14 08:59:00 crc kubenswrapper[4870]: I1014 08:59:00.844520 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nm52f\" (UniqueName: \"kubernetes.io/projected/2ea8030c-8337-467b-be99-b8eb2e48ea56-kube-api-access-nm52f\") pod \"nova-cell1-db-create-p9h2b\" (UID: \"2ea8030c-8337-467b-be99-b8eb2e48ea56\") " pod="openstack/nova-cell1-db-create-p9h2b" Oct 14 08:59:00 crc kubenswrapper[4870]: I1014 08:59:00.867427 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-76jpt\" (UniqueName: \"kubernetes.io/projected/8fba7d20-e48c-4e8e-b0a6-21ed0aea5ab1-kube-api-access-76jpt\") pod \"nova-cell0-db-create-scw7l\" (UID: \"8fba7d20-e48c-4e8e-b0a6-21ed0aea5ab1\") " pod="openstack/nova-cell0-db-create-scw7l" Oct 14 08:59:00 crc kubenswrapper[4870]: I1014 08:59:00.907026 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-scw7l" Oct 14 08:59:00 crc kubenswrapper[4870]: I1014 08:59:00.946626 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nm52f\" (UniqueName: \"kubernetes.io/projected/2ea8030c-8337-467b-be99-b8eb2e48ea56-kube-api-access-nm52f\") pod \"nova-cell1-db-create-p9h2b\" (UID: \"2ea8030c-8337-467b-be99-b8eb2e48ea56\") " pod="openstack/nova-cell1-db-create-p9h2b" Oct 14 08:59:00 crc kubenswrapper[4870]: I1014 08:59:00.962700 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nm52f\" (UniqueName: \"kubernetes.io/projected/2ea8030c-8337-467b-be99-b8eb2e48ea56-kube-api-access-nm52f\") pod \"nova-cell1-db-create-p9h2b\" (UID: \"2ea8030c-8337-467b-be99-b8eb2e48ea56\") " pod="openstack/nova-cell1-db-create-p9h2b" Oct 14 08:59:01 crc kubenswrapper[4870]: I1014 08:59:01.020130 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-p9h2b" Oct 14 08:59:01 crc kubenswrapper[4870]: I1014 08:59:01.282624 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-9htzx"] Oct 14 08:59:01 crc kubenswrapper[4870]: I1014 08:59:01.401401 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-scw7l"] Oct 14 08:59:01 crc kubenswrapper[4870]: W1014 08:59:01.407697 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8fba7d20_e48c_4e8e_b0a6_21ed0aea5ab1.slice/crio-0ba73f1017923a5f264ce76f2e7c5e914b1fdb5b9341d9237adc38ff4fe8f5c0 WatchSource:0}: Error finding container 0ba73f1017923a5f264ce76f2e7c5e914b1fdb5b9341d9237adc38ff4fe8f5c0: Status 404 returned error can't find the container with id 0ba73f1017923a5f264ce76f2e7c5e914b1fdb5b9341d9237adc38ff4fe8f5c0 Oct 14 08:59:01 crc kubenswrapper[4870]: I1014 08:59:01.516292 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-p9h2b"] Oct 14 08:59:01 crc kubenswrapper[4870]: W1014 08:59:01.640899 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2ea8030c_8337_467b_be99_b8eb2e48ea56.slice/crio-1870288ecad1238bc8da08e20d5e2205faec9baa807b9f88e1a6de58141e98f3 WatchSource:0}: Error finding container 1870288ecad1238bc8da08e20d5e2205faec9baa807b9f88e1a6de58141e98f3: Status 404 returned error can't find the container with id 1870288ecad1238bc8da08e20d5e2205faec9baa807b9f88e1a6de58141e98f3 Oct 14 08:59:02 crc kubenswrapper[4870]: I1014 08:59:02.082138 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-scw7l" event={"ID":"8fba7d20-e48c-4e8e-b0a6-21ed0aea5ab1","Type":"ContainerDied","Data":"e5c32fb72855303655e866115cac1797680ed1edbe001700b324e72837a8647f"} Oct 14 08:59:02 crc kubenswrapper[4870]: I1014 08:59:02.082581 4870 generic.go:334] "Generic (PLEG): container finished" podID="8fba7d20-e48c-4e8e-b0a6-21ed0aea5ab1" containerID="e5c32fb72855303655e866115cac1797680ed1edbe001700b324e72837a8647f" exitCode=0 Oct 14 08:59:02 crc kubenswrapper[4870]: I1014 08:59:02.082741 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-scw7l" event={"ID":"8fba7d20-e48c-4e8e-b0a6-21ed0aea5ab1","Type":"ContainerStarted","Data":"0ba73f1017923a5f264ce76f2e7c5e914b1fdb5b9341d9237adc38ff4fe8f5c0"} Oct 14 08:59:02 crc kubenswrapper[4870]: I1014 08:59:02.084479 4870 generic.go:334] "Generic (PLEG): container finished" podID="2ea8030c-8337-467b-be99-b8eb2e48ea56" containerID="0e314a6abc2dab8dca15f3a5e845c1abbd2ca7e8beb7c42bd2a1e0e2ee0f0790" exitCode=0 Oct 14 08:59:02 crc kubenswrapper[4870]: I1014 08:59:02.084637 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-p9h2b" event={"ID":"2ea8030c-8337-467b-be99-b8eb2e48ea56","Type":"ContainerDied","Data":"0e314a6abc2dab8dca15f3a5e845c1abbd2ca7e8beb7c42bd2a1e0e2ee0f0790"} Oct 14 08:59:02 crc kubenswrapper[4870]: I1014 08:59:02.084793 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-p9h2b" event={"ID":"2ea8030c-8337-467b-be99-b8eb2e48ea56","Type":"ContainerStarted","Data":"1870288ecad1238bc8da08e20d5e2205faec9baa807b9f88e1a6de58141e98f3"} Oct 14 08:59:02 crc kubenswrapper[4870]: I1014 08:59:02.088157 4870 generic.go:334] "Generic (PLEG): container finished" podID="16fe9c99-c775-4b99-8132-535f9daa9a9f" containerID="c44545800af6e2b89c3a53d3ede0edff2502d2bed187c89bf6e629d76d11e590" exitCode=0 Oct 14 08:59:02 crc kubenswrapper[4870]: I1014 08:59:02.088202 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-9htzx" event={"ID":"16fe9c99-c775-4b99-8132-535f9daa9a9f","Type":"ContainerDied","Data":"c44545800af6e2b89c3a53d3ede0edff2502d2bed187c89bf6e629d76d11e590"} Oct 14 08:59:02 crc kubenswrapper[4870]: I1014 08:59:02.088226 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-9htzx" event={"ID":"16fe9c99-c775-4b99-8132-535f9daa9a9f","Type":"ContainerStarted","Data":"8e4981832345541eb4b2d101eb952fb0a6b26f2c6f6ab68635dacc88b958e97e"} Oct 14 08:59:03 crc kubenswrapper[4870]: I1014 08:59:03.588691 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-p9h2b" Oct 14 08:59:03 crc kubenswrapper[4870]: I1014 08:59:03.594575 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-scw7l" Oct 14 08:59:03 crc kubenswrapper[4870]: I1014 08:59:03.600485 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-9htzx" Oct 14 08:59:03 crc kubenswrapper[4870]: I1014 08:59:03.695738 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-76jpt\" (UniqueName: \"kubernetes.io/projected/8fba7d20-e48c-4e8e-b0a6-21ed0aea5ab1-kube-api-access-76jpt\") pod \"8fba7d20-e48c-4e8e-b0a6-21ed0aea5ab1\" (UID: \"8fba7d20-e48c-4e8e-b0a6-21ed0aea5ab1\") " Oct 14 08:59:03 crc kubenswrapper[4870]: I1014 08:59:03.695783 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nm52f\" (UniqueName: \"kubernetes.io/projected/2ea8030c-8337-467b-be99-b8eb2e48ea56-kube-api-access-nm52f\") pod \"2ea8030c-8337-467b-be99-b8eb2e48ea56\" (UID: \"2ea8030c-8337-467b-be99-b8eb2e48ea56\") " Oct 14 08:59:03 crc kubenswrapper[4870]: I1014 08:59:03.695811 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dmphw\" (UniqueName: \"kubernetes.io/projected/16fe9c99-c775-4b99-8132-535f9daa9a9f-kube-api-access-dmphw\") pod \"16fe9c99-c775-4b99-8132-535f9daa9a9f\" (UID: \"16fe9c99-c775-4b99-8132-535f9daa9a9f\") " Oct 14 08:59:03 crc kubenswrapper[4870]: I1014 08:59:03.700835 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2ea8030c-8337-467b-be99-b8eb2e48ea56-kube-api-access-nm52f" (OuterVolumeSpecName: "kube-api-access-nm52f") pod "2ea8030c-8337-467b-be99-b8eb2e48ea56" (UID: "2ea8030c-8337-467b-be99-b8eb2e48ea56"). InnerVolumeSpecName "kube-api-access-nm52f". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:59:03 crc kubenswrapper[4870]: I1014 08:59:03.701312 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8fba7d20-e48c-4e8e-b0a6-21ed0aea5ab1-kube-api-access-76jpt" (OuterVolumeSpecName: "kube-api-access-76jpt") pod "8fba7d20-e48c-4e8e-b0a6-21ed0aea5ab1" (UID: "8fba7d20-e48c-4e8e-b0a6-21ed0aea5ab1"). InnerVolumeSpecName "kube-api-access-76jpt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:59:03 crc kubenswrapper[4870]: I1014 08:59:03.702603 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/16fe9c99-c775-4b99-8132-535f9daa9a9f-kube-api-access-dmphw" (OuterVolumeSpecName: "kube-api-access-dmphw") pod "16fe9c99-c775-4b99-8132-535f9daa9a9f" (UID: "16fe9c99-c775-4b99-8132-535f9daa9a9f"). InnerVolumeSpecName "kube-api-access-dmphw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:59:03 crc kubenswrapper[4870]: I1014 08:59:03.798238 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-76jpt\" (UniqueName: \"kubernetes.io/projected/8fba7d20-e48c-4e8e-b0a6-21ed0aea5ab1-kube-api-access-76jpt\") on node \"crc\" DevicePath \"\"" Oct 14 08:59:03 crc kubenswrapper[4870]: I1014 08:59:03.798289 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nm52f\" (UniqueName: \"kubernetes.io/projected/2ea8030c-8337-467b-be99-b8eb2e48ea56-kube-api-access-nm52f\") on node \"crc\" DevicePath \"\"" Oct 14 08:59:03 crc kubenswrapper[4870]: I1014 08:59:03.798313 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dmphw\" (UniqueName: \"kubernetes.io/projected/16fe9c99-c775-4b99-8132-535f9daa9a9f-kube-api-access-dmphw\") on node \"crc\" DevicePath \"\"" Oct 14 08:59:04 crc kubenswrapper[4870]: I1014 08:59:04.111132 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-9htzx" event={"ID":"16fe9c99-c775-4b99-8132-535f9daa9a9f","Type":"ContainerDied","Data":"8e4981832345541eb4b2d101eb952fb0a6b26f2c6f6ab68635dacc88b958e97e"} Oct 14 08:59:04 crc kubenswrapper[4870]: I1014 08:59:04.111176 4870 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8e4981832345541eb4b2d101eb952fb0a6b26f2c6f6ab68635dacc88b958e97e" Oct 14 08:59:04 crc kubenswrapper[4870]: I1014 08:59:04.111172 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-9htzx" Oct 14 08:59:04 crc kubenswrapper[4870]: I1014 08:59:04.113340 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-scw7l" event={"ID":"8fba7d20-e48c-4e8e-b0a6-21ed0aea5ab1","Type":"ContainerDied","Data":"0ba73f1017923a5f264ce76f2e7c5e914b1fdb5b9341d9237adc38ff4fe8f5c0"} Oct 14 08:59:04 crc kubenswrapper[4870]: I1014 08:59:04.113360 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-scw7l" Oct 14 08:59:04 crc kubenswrapper[4870]: I1014 08:59:04.113363 4870 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0ba73f1017923a5f264ce76f2e7c5e914b1fdb5b9341d9237adc38ff4fe8f5c0" Oct 14 08:59:04 crc kubenswrapper[4870]: I1014 08:59:04.114998 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-p9h2b" event={"ID":"2ea8030c-8337-467b-be99-b8eb2e48ea56","Type":"ContainerDied","Data":"1870288ecad1238bc8da08e20d5e2205faec9baa807b9f88e1a6de58141e98f3"} Oct 14 08:59:04 crc kubenswrapper[4870]: I1014 08:59:04.115023 4870 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1870288ecad1238bc8da08e20d5e2205faec9baa807b9f88e1a6de58141e98f3" Oct 14 08:59:04 crc kubenswrapper[4870]: I1014 08:59:04.115074 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-p9h2b" Oct 14 08:59:10 crc kubenswrapper[4870]: I1014 08:59:10.743076 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-e4ea-account-create-8t749"] Oct 14 08:59:10 crc kubenswrapper[4870]: E1014 08:59:10.744049 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8fba7d20-e48c-4e8e-b0a6-21ed0aea5ab1" containerName="mariadb-database-create" Oct 14 08:59:10 crc kubenswrapper[4870]: I1014 08:59:10.744066 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="8fba7d20-e48c-4e8e-b0a6-21ed0aea5ab1" containerName="mariadb-database-create" Oct 14 08:59:10 crc kubenswrapper[4870]: E1014 08:59:10.744085 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ea8030c-8337-467b-be99-b8eb2e48ea56" containerName="mariadb-database-create" Oct 14 08:59:10 crc kubenswrapper[4870]: I1014 08:59:10.744093 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ea8030c-8337-467b-be99-b8eb2e48ea56" containerName="mariadb-database-create" Oct 14 08:59:10 crc kubenswrapper[4870]: E1014 08:59:10.744362 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="16fe9c99-c775-4b99-8132-535f9daa9a9f" containerName="mariadb-database-create" Oct 14 08:59:10 crc kubenswrapper[4870]: I1014 08:59:10.744378 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="16fe9c99-c775-4b99-8132-535f9daa9a9f" containerName="mariadb-database-create" Oct 14 08:59:10 crc kubenswrapper[4870]: I1014 08:59:10.744632 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="2ea8030c-8337-467b-be99-b8eb2e48ea56" containerName="mariadb-database-create" Oct 14 08:59:10 crc kubenswrapper[4870]: I1014 08:59:10.744651 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="16fe9c99-c775-4b99-8132-535f9daa9a9f" containerName="mariadb-database-create" Oct 14 08:59:10 crc kubenswrapper[4870]: I1014 08:59:10.744676 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="8fba7d20-e48c-4e8e-b0a6-21ed0aea5ab1" containerName="mariadb-database-create" Oct 14 08:59:10 crc kubenswrapper[4870]: I1014 08:59:10.745406 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-e4ea-account-create-8t749" Oct 14 08:59:10 crc kubenswrapper[4870]: I1014 08:59:10.748863 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Oct 14 08:59:10 crc kubenswrapper[4870]: I1014 08:59:10.768150 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-e4ea-account-create-8t749"] Oct 14 08:59:10 crc kubenswrapper[4870]: I1014 08:59:10.848974 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-45qfs\" (UniqueName: \"kubernetes.io/projected/ee31f93f-00ce-45ea-8e8c-4c12036e3d11-kube-api-access-45qfs\") pod \"nova-api-e4ea-account-create-8t749\" (UID: \"ee31f93f-00ce-45ea-8e8c-4c12036e3d11\") " pod="openstack/nova-api-e4ea-account-create-8t749" Oct 14 08:59:10 crc kubenswrapper[4870]: I1014 08:59:10.932834 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-f10d-account-create-pkpj9"] Oct 14 08:59:10 crc kubenswrapper[4870]: I1014 08:59:10.934170 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-f10d-account-create-pkpj9" Oct 14 08:59:10 crc kubenswrapper[4870]: I1014 08:59:10.936268 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Oct 14 08:59:10 crc kubenswrapper[4870]: I1014 08:59:10.942170 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-f10d-account-create-pkpj9"] Oct 14 08:59:10 crc kubenswrapper[4870]: I1014 08:59:10.950389 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-45qfs\" (UniqueName: \"kubernetes.io/projected/ee31f93f-00ce-45ea-8e8c-4c12036e3d11-kube-api-access-45qfs\") pod \"nova-api-e4ea-account-create-8t749\" (UID: \"ee31f93f-00ce-45ea-8e8c-4c12036e3d11\") " pod="openstack/nova-api-e4ea-account-create-8t749" Oct 14 08:59:10 crc kubenswrapper[4870]: I1014 08:59:10.987214 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-45qfs\" (UniqueName: \"kubernetes.io/projected/ee31f93f-00ce-45ea-8e8c-4c12036e3d11-kube-api-access-45qfs\") pod \"nova-api-e4ea-account-create-8t749\" (UID: \"ee31f93f-00ce-45ea-8e8c-4c12036e3d11\") " pod="openstack/nova-api-e4ea-account-create-8t749" Oct 14 08:59:11 crc kubenswrapper[4870]: I1014 08:59:11.051798 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-glp6f\" (UniqueName: \"kubernetes.io/projected/f022b544-37f1-4e1e-a1bc-614f4f36b8c3-kube-api-access-glp6f\") pod \"nova-cell0-f10d-account-create-pkpj9\" (UID: \"f022b544-37f1-4e1e-a1bc-614f4f36b8c3\") " pod="openstack/nova-cell0-f10d-account-create-pkpj9" Oct 14 08:59:11 crc kubenswrapper[4870]: I1014 08:59:11.082494 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-e4ea-account-create-8t749" Oct 14 08:59:11 crc kubenswrapper[4870]: I1014 08:59:11.139673 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-66ee-account-create-wjnkl"] Oct 14 08:59:11 crc kubenswrapper[4870]: I1014 08:59:11.141561 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-66ee-account-create-wjnkl" Oct 14 08:59:11 crc kubenswrapper[4870]: I1014 08:59:11.149998 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Oct 14 08:59:11 crc kubenswrapper[4870]: I1014 08:59:11.153606 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-glp6f\" (UniqueName: \"kubernetes.io/projected/f022b544-37f1-4e1e-a1bc-614f4f36b8c3-kube-api-access-glp6f\") pod \"nova-cell0-f10d-account-create-pkpj9\" (UID: \"f022b544-37f1-4e1e-a1bc-614f4f36b8c3\") " pod="openstack/nova-cell0-f10d-account-create-pkpj9" Oct 14 08:59:11 crc kubenswrapper[4870]: I1014 08:59:11.156885 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-66ee-account-create-wjnkl"] Oct 14 08:59:11 crc kubenswrapper[4870]: I1014 08:59:11.175335 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-glp6f\" (UniqueName: \"kubernetes.io/projected/f022b544-37f1-4e1e-a1bc-614f4f36b8c3-kube-api-access-glp6f\") pod \"nova-cell0-f10d-account-create-pkpj9\" (UID: \"f022b544-37f1-4e1e-a1bc-614f4f36b8c3\") " pod="openstack/nova-cell0-f10d-account-create-pkpj9" Oct 14 08:59:11 crc kubenswrapper[4870]: I1014 08:59:11.255516 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k9prf\" (UniqueName: \"kubernetes.io/projected/984f8c9e-ebdd-4c8f-b093-8d30b38b5a2a-kube-api-access-k9prf\") pod \"nova-cell1-66ee-account-create-wjnkl\" (UID: \"984f8c9e-ebdd-4c8f-b093-8d30b38b5a2a\") " pod="openstack/nova-cell1-66ee-account-create-wjnkl" Oct 14 08:59:11 crc kubenswrapper[4870]: I1014 08:59:11.258909 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-f10d-account-create-pkpj9" Oct 14 08:59:11 crc kubenswrapper[4870]: I1014 08:59:11.357359 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k9prf\" (UniqueName: \"kubernetes.io/projected/984f8c9e-ebdd-4c8f-b093-8d30b38b5a2a-kube-api-access-k9prf\") pod \"nova-cell1-66ee-account-create-wjnkl\" (UID: \"984f8c9e-ebdd-4c8f-b093-8d30b38b5a2a\") " pod="openstack/nova-cell1-66ee-account-create-wjnkl" Oct 14 08:59:11 crc kubenswrapper[4870]: I1014 08:59:11.378852 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k9prf\" (UniqueName: \"kubernetes.io/projected/984f8c9e-ebdd-4c8f-b093-8d30b38b5a2a-kube-api-access-k9prf\") pod \"nova-cell1-66ee-account-create-wjnkl\" (UID: \"984f8c9e-ebdd-4c8f-b093-8d30b38b5a2a\") " pod="openstack/nova-cell1-66ee-account-create-wjnkl" Oct 14 08:59:11 crc kubenswrapper[4870]: I1014 08:59:11.560379 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-66ee-account-create-wjnkl" Oct 14 08:59:11 crc kubenswrapper[4870]: I1014 08:59:11.573327 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-e4ea-account-create-8t749"] Oct 14 08:59:11 crc kubenswrapper[4870]: I1014 08:59:11.698758 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-f10d-account-create-pkpj9"] Oct 14 08:59:11 crc kubenswrapper[4870]: W1014 08:59:11.698982 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf022b544_37f1_4e1e_a1bc_614f4f36b8c3.slice/crio-ad20357b51a3f6b32afa686870704703c0a6af7bcaf2e0152613fc0093cde891 WatchSource:0}: Error finding container ad20357b51a3f6b32afa686870704703c0a6af7bcaf2e0152613fc0093cde891: Status 404 returned error can't find the container with id ad20357b51a3f6b32afa686870704703c0a6af7bcaf2e0152613fc0093cde891 Oct 14 08:59:12 crc kubenswrapper[4870]: I1014 08:59:12.004674 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-66ee-account-create-wjnkl"] Oct 14 08:59:12 crc kubenswrapper[4870]: I1014 08:59:12.194208 4870 generic.go:334] "Generic (PLEG): container finished" podID="f022b544-37f1-4e1e-a1bc-614f4f36b8c3" containerID="12da0dee2ef976c06be85dab716622747e95edc7d4700015b5f01575ba23bd3e" exitCode=0 Oct 14 08:59:12 crc kubenswrapper[4870]: I1014 08:59:12.194272 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-f10d-account-create-pkpj9" event={"ID":"f022b544-37f1-4e1e-a1bc-614f4f36b8c3","Type":"ContainerDied","Data":"12da0dee2ef976c06be85dab716622747e95edc7d4700015b5f01575ba23bd3e"} Oct 14 08:59:12 crc kubenswrapper[4870]: I1014 08:59:12.194339 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-f10d-account-create-pkpj9" event={"ID":"f022b544-37f1-4e1e-a1bc-614f4f36b8c3","Type":"ContainerStarted","Data":"ad20357b51a3f6b32afa686870704703c0a6af7bcaf2e0152613fc0093cde891"} Oct 14 08:59:12 crc kubenswrapper[4870]: I1014 08:59:12.195633 4870 generic.go:334] "Generic (PLEG): container finished" podID="ee31f93f-00ce-45ea-8e8c-4c12036e3d11" containerID="4e79e333d58ceed9e942c90eccb0fcc1d37daa5266b237942c698e6f1e9bb5b1" exitCode=0 Oct 14 08:59:12 crc kubenswrapper[4870]: I1014 08:59:12.195689 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-e4ea-account-create-8t749" event={"ID":"ee31f93f-00ce-45ea-8e8c-4c12036e3d11","Type":"ContainerDied","Data":"4e79e333d58ceed9e942c90eccb0fcc1d37daa5266b237942c698e6f1e9bb5b1"} Oct 14 08:59:12 crc kubenswrapper[4870]: I1014 08:59:12.195754 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-e4ea-account-create-8t749" event={"ID":"ee31f93f-00ce-45ea-8e8c-4c12036e3d11","Type":"ContainerStarted","Data":"bd8c0690e618de166daef5c432e174de51735b9b18ea722eb856906a94fdbac0"} Oct 14 08:59:12 crc kubenswrapper[4870]: I1014 08:59:12.197326 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-66ee-account-create-wjnkl" event={"ID":"984f8c9e-ebdd-4c8f-b093-8d30b38b5a2a","Type":"ContainerStarted","Data":"b2fc13652a9b4083bea6fb724b4947403d0eb1c717c668647708ae1d806b15e9"} Oct 14 08:59:13 crc kubenswrapper[4870]: I1014 08:59:13.208151 4870 generic.go:334] "Generic (PLEG): container finished" podID="984f8c9e-ebdd-4c8f-b093-8d30b38b5a2a" containerID="fc13475221487a46b31d10b88df54efde5603c4ed03902033b2e560f96f5278c" exitCode=0 Oct 14 08:59:13 crc kubenswrapper[4870]: I1014 08:59:13.208206 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-66ee-account-create-wjnkl" event={"ID":"984f8c9e-ebdd-4c8f-b093-8d30b38b5a2a","Type":"ContainerDied","Data":"fc13475221487a46b31d10b88df54efde5603c4ed03902033b2e560f96f5278c"} Oct 14 08:59:13 crc kubenswrapper[4870]: I1014 08:59:13.642916 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-e4ea-account-create-8t749" Oct 14 08:59:13 crc kubenswrapper[4870]: I1014 08:59:13.648506 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-f10d-account-create-pkpj9" Oct 14 08:59:13 crc kubenswrapper[4870]: I1014 08:59:13.704480 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-45qfs\" (UniqueName: \"kubernetes.io/projected/ee31f93f-00ce-45ea-8e8c-4c12036e3d11-kube-api-access-45qfs\") pod \"ee31f93f-00ce-45ea-8e8c-4c12036e3d11\" (UID: \"ee31f93f-00ce-45ea-8e8c-4c12036e3d11\") " Oct 14 08:59:13 crc kubenswrapper[4870]: I1014 08:59:13.704852 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-glp6f\" (UniqueName: \"kubernetes.io/projected/f022b544-37f1-4e1e-a1bc-614f4f36b8c3-kube-api-access-glp6f\") pod \"f022b544-37f1-4e1e-a1bc-614f4f36b8c3\" (UID: \"f022b544-37f1-4e1e-a1bc-614f4f36b8c3\") " Oct 14 08:59:13 crc kubenswrapper[4870]: I1014 08:59:13.712215 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ee31f93f-00ce-45ea-8e8c-4c12036e3d11-kube-api-access-45qfs" (OuterVolumeSpecName: "kube-api-access-45qfs") pod "ee31f93f-00ce-45ea-8e8c-4c12036e3d11" (UID: "ee31f93f-00ce-45ea-8e8c-4c12036e3d11"). InnerVolumeSpecName "kube-api-access-45qfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:59:13 crc kubenswrapper[4870]: I1014 08:59:13.712290 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f022b544-37f1-4e1e-a1bc-614f4f36b8c3-kube-api-access-glp6f" (OuterVolumeSpecName: "kube-api-access-glp6f") pod "f022b544-37f1-4e1e-a1bc-614f4f36b8c3" (UID: "f022b544-37f1-4e1e-a1bc-614f4f36b8c3"). InnerVolumeSpecName "kube-api-access-glp6f". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:59:13 crc kubenswrapper[4870]: I1014 08:59:13.807359 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-glp6f\" (UniqueName: \"kubernetes.io/projected/f022b544-37f1-4e1e-a1bc-614f4f36b8c3-kube-api-access-glp6f\") on node \"crc\" DevicePath \"\"" Oct 14 08:59:13 crc kubenswrapper[4870]: I1014 08:59:13.807425 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-45qfs\" (UniqueName: \"kubernetes.io/projected/ee31f93f-00ce-45ea-8e8c-4c12036e3d11-kube-api-access-45qfs\") on node \"crc\" DevicePath \"\"" Oct 14 08:59:14 crc kubenswrapper[4870]: I1014 08:59:14.220297 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-f10d-account-create-pkpj9" event={"ID":"f022b544-37f1-4e1e-a1bc-614f4f36b8c3","Type":"ContainerDied","Data":"ad20357b51a3f6b32afa686870704703c0a6af7bcaf2e0152613fc0093cde891"} Oct 14 08:59:14 crc kubenswrapper[4870]: I1014 08:59:14.220350 4870 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ad20357b51a3f6b32afa686870704703c0a6af7bcaf2e0152613fc0093cde891" Oct 14 08:59:14 crc kubenswrapper[4870]: I1014 08:59:14.220324 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-f10d-account-create-pkpj9" Oct 14 08:59:14 crc kubenswrapper[4870]: I1014 08:59:14.222287 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-e4ea-account-create-8t749" Oct 14 08:59:14 crc kubenswrapper[4870]: I1014 08:59:14.222334 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-e4ea-account-create-8t749" event={"ID":"ee31f93f-00ce-45ea-8e8c-4c12036e3d11","Type":"ContainerDied","Data":"bd8c0690e618de166daef5c432e174de51735b9b18ea722eb856906a94fdbac0"} Oct 14 08:59:14 crc kubenswrapper[4870]: I1014 08:59:14.222359 4870 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bd8c0690e618de166daef5c432e174de51735b9b18ea722eb856906a94fdbac0" Oct 14 08:59:14 crc kubenswrapper[4870]: I1014 08:59:14.596740 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-66ee-account-create-wjnkl" Oct 14 08:59:14 crc kubenswrapper[4870]: I1014 08:59:14.727900 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k9prf\" (UniqueName: \"kubernetes.io/projected/984f8c9e-ebdd-4c8f-b093-8d30b38b5a2a-kube-api-access-k9prf\") pod \"984f8c9e-ebdd-4c8f-b093-8d30b38b5a2a\" (UID: \"984f8c9e-ebdd-4c8f-b093-8d30b38b5a2a\") " Oct 14 08:59:14 crc kubenswrapper[4870]: I1014 08:59:14.733736 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/984f8c9e-ebdd-4c8f-b093-8d30b38b5a2a-kube-api-access-k9prf" (OuterVolumeSpecName: "kube-api-access-k9prf") pod "984f8c9e-ebdd-4c8f-b093-8d30b38b5a2a" (UID: "984f8c9e-ebdd-4c8f-b093-8d30b38b5a2a"). InnerVolumeSpecName "kube-api-access-k9prf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:59:14 crc kubenswrapper[4870]: I1014 08:59:14.830740 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k9prf\" (UniqueName: \"kubernetes.io/projected/984f8c9e-ebdd-4c8f-b093-8d30b38b5a2a-kube-api-access-k9prf\") on node \"crc\" DevicePath \"\"" Oct 14 08:59:15 crc kubenswrapper[4870]: I1014 08:59:15.235464 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-66ee-account-create-wjnkl" event={"ID":"984f8c9e-ebdd-4c8f-b093-8d30b38b5a2a","Type":"ContainerDied","Data":"b2fc13652a9b4083bea6fb724b4947403d0eb1c717c668647708ae1d806b15e9"} Oct 14 08:59:15 crc kubenswrapper[4870]: I1014 08:59:15.235524 4870 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b2fc13652a9b4083bea6fb724b4947403d0eb1c717c668647708ae1d806b15e9" Oct 14 08:59:15 crc kubenswrapper[4870]: I1014 08:59:15.235539 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-66ee-account-create-wjnkl" Oct 14 08:59:16 crc kubenswrapper[4870]: I1014 08:59:16.031000 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-kxsd4"] Oct 14 08:59:16 crc kubenswrapper[4870]: E1014 08:59:16.031791 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="984f8c9e-ebdd-4c8f-b093-8d30b38b5a2a" containerName="mariadb-account-create" Oct 14 08:59:16 crc kubenswrapper[4870]: I1014 08:59:16.031814 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="984f8c9e-ebdd-4c8f-b093-8d30b38b5a2a" containerName="mariadb-account-create" Oct 14 08:59:16 crc kubenswrapper[4870]: E1014 08:59:16.031840 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f022b544-37f1-4e1e-a1bc-614f4f36b8c3" containerName="mariadb-account-create" Oct 14 08:59:16 crc kubenswrapper[4870]: I1014 08:59:16.031850 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="f022b544-37f1-4e1e-a1bc-614f4f36b8c3" containerName="mariadb-account-create" Oct 14 08:59:16 crc kubenswrapper[4870]: E1014 08:59:16.031865 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee31f93f-00ce-45ea-8e8c-4c12036e3d11" containerName="mariadb-account-create" Oct 14 08:59:16 crc kubenswrapper[4870]: I1014 08:59:16.031873 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee31f93f-00ce-45ea-8e8c-4c12036e3d11" containerName="mariadb-account-create" Oct 14 08:59:16 crc kubenswrapper[4870]: I1014 08:59:16.032163 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="984f8c9e-ebdd-4c8f-b093-8d30b38b5a2a" containerName="mariadb-account-create" Oct 14 08:59:16 crc kubenswrapper[4870]: I1014 08:59:16.032188 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee31f93f-00ce-45ea-8e8c-4c12036e3d11" containerName="mariadb-account-create" Oct 14 08:59:16 crc kubenswrapper[4870]: I1014 08:59:16.032215 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="f022b544-37f1-4e1e-a1bc-614f4f36b8c3" containerName="mariadb-account-create" Oct 14 08:59:16 crc kubenswrapper[4870]: I1014 08:59:16.032958 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-kxsd4" Oct 14 08:59:16 crc kubenswrapper[4870]: I1014 08:59:16.038022 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Oct 14 08:59:16 crc kubenswrapper[4870]: I1014 08:59:16.038286 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Oct 14 08:59:16 crc kubenswrapper[4870]: I1014 08:59:16.038463 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-7qvv5" Oct 14 08:59:16 crc kubenswrapper[4870]: I1014 08:59:16.049539 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-kxsd4"] Oct 14 08:59:16 crc kubenswrapper[4870]: I1014 08:59:16.153933 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/57647978-3492-4dcf-85ae-ad3109e97980-scripts\") pod \"nova-cell0-conductor-db-sync-kxsd4\" (UID: \"57647978-3492-4dcf-85ae-ad3109e97980\") " pod="openstack/nova-cell0-conductor-db-sync-kxsd4" Oct 14 08:59:16 crc kubenswrapper[4870]: I1014 08:59:16.154110 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-27zp2\" (UniqueName: \"kubernetes.io/projected/57647978-3492-4dcf-85ae-ad3109e97980-kube-api-access-27zp2\") pod \"nova-cell0-conductor-db-sync-kxsd4\" (UID: \"57647978-3492-4dcf-85ae-ad3109e97980\") " pod="openstack/nova-cell0-conductor-db-sync-kxsd4" Oct 14 08:59:16 crc kubenswrapper[4870]: I1014 08:59:16.154186 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57647978-3492-4dcf-85ae-ad3109e97980-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-kxsd4\" (UID: \"57647978-3492-4dcf-85ae-ad3109e97980\") " pod="openstack/nova-cell0-conductor-db-sync-kxsd4" Oct 14 08:59:16 crc kubenswrapper[4870]: I1014 08:59:16.154305 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/57647978-3492-4dcf-85ae-ad3109e97980-config-data\") pod \"nova-cell0-conductor-db-sync-kxsd4\" (UID: \"57647978-3492-4dcf-85ae-ad3109e97980\") " pod="openstack/nova-cell0-conductor-db-sync-kxsd4" Oct 14 08:59:16 crc kubenswrapper[4870]: I1014 08:59:16.256194 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/57647978-3492-4dcf-85ae-ad3109e97980-config-data\") pod \"nova-cell0-conductor-db-sync-kxsd4\" (UID: \"57647978-3492-4dcf-85ae-ad3109e97980\") " pod="openstack/nova-cell0-conductor-db-sync-kxsd4" Oct 14 08:59:16 crc kubenswrapper[4870]: I1014 08:59:16.256746 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/57647978-3492-4dcf-85ae-ad3109e97980-scripts\") pod \"nova-cell0-conductor-db-sync-kxsd4\" (UID: \"57647978-3492-4dcf-85ae-ad3109e97980\") " pod="openstack/nova-cell0-conductor-db-sync-kxsd4" Oct 14 08:59:16 crc kubenswrapper[4870]: I1014 08:59:16.256869 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-27zp2\" (UniqueName: \"kubernetes.io/projected/57647978-3492-4dcf-85ae-ad3109e97980-kube-api-access-27zp2\") pod \"nova-cell0-conductor-db-sync-kxsd4\" (UID: \"57647978-3492-4dcf-85ae-ad3109e97980\") " pod="openstack/nova-cell0-conductor-db-sync-kxsd4" Oct 14 08:59:16 crc kubenswrapper[4870]: I1014 08:59:16.256970 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57647978-3492-4dcf-85ae-ad3109e97980-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-kxsd4\" (UID: \"57647978-3492-4dcf-85ae-ad3109e97980\") " pod="openstack/nova-cell0-conductor-db-sync-kxsd4" Oct 14 08:59:16 crc kubenswrapper[4870]: I1014 08:59:16.260723 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/57647978-3492-4dcf-85ae-ad3109e97980-scripts\") pod \"nova-cell0-conductor-db-sync-kxsd4\" (UID: \"57647978-3492-4dcf-85ae-ad3109e97980\") " pod="openstack/nova-cell0-conductor-db-sync-kxsd4" Oct 14 08:59:16 crc kubenswrapper[4870]: I1014 08:59:16.260917 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57647978-3492-4dcf-85ae-ad3109e97980-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-kxsd4\" (UID: \"57647978-3492-4dcf-85ae-ad3109e97980\") " pod="openstack/nova-cell0-conductor-db-sync-kxsd4" Oct 14 08:59:16 crc kubenswrapper[4870]: I1014 08:59:16.261716 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/57647978-3492-4dcf-85ae-ad3109e97980-config-data\") pod \"nova-cell0-conductor-db-sync-kxsd4\" (UID: \"57647978-3492-4dcf-85ae-ad3109e97980\") " pod="openstack/nova-cell0-conductor-db-sync-kxsd4" Oct 14 08:59:16 crc kubenswrapper[4870]: I1014 08:59:16.273138 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-27zp2\" (UniqueName: \"kubernetes.io/projected/57647978-3492-4dcf-85ae-ad3109e97980-kube-api-access-27zp2\") pod \"nova-cell0-conductor-db-sync-kxsd4\" (UID: \"57647978-3492-4dcf-85ae-ad3109e97980\") " pod="openstack/nova-cell0-conductor-db-sync-kxsd4" Oct 14 08:59:16 crc kubenswrapper[4870]: I1014 08:59:16.364196 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-kxsd4" Oct 14 08:59:16 crc kubenswrapper[4870]: I1014 08:59:16.809588 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-kxsd4"] Oct 14 08:59:16 crc kubenswrapper[4870]: I1014 08:59:16.823040 4870 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 14 08:59:17 crc kubenswrapper[4870]: I1014 08:59:17.255014 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-kxsd4" event={"ID":"57647978-3492-4dcf-85ae-ad3109e97980","Type":"ContainerStarted","Data":"7f94e2bb24fc9cdb260a0f8a8db42d80f65190fe615587820bf5160d89acaa0d"} Oct 14 08:59:21 crc kubenswrapper[4870]: I1014 08:59:21.473554 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-j9wl4"] Oct 14 08:59:21 crc kubenswrapper[4870]: I1014 08:59:21.476532 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-j9wl4" Oct 14 08:59:21 crc kubenswrapper[4870]: I1014 08:59:21.482729 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-j9wl4"] Oct 14 08:59:21 crc kubenswrapper[4870]: I1014 08:59:21.553877 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gq2h9\" (UniqueName: \"kubernetes.io/projected/34086778-4805-47e2-9d8d-7b9a327dcd41-kube-api-access-gq2h9\") pod \"community-operators-j9wl4\" (UID: \"34086778-4805-47e2-9d8d-7b9a327dcd41\") " pod="openshift-marketplace/community-operators-j9wl4" Oct 14 08:59:21 crc kubenswrapper[4870]: I1014 08:59:21.554139 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/34086778-4805-47e2-9d8d-7b9a327dcd41-catalog-content\") pod \"community-operators-j9wl4\" (UID: \"34086778-4805-47e2-9d8d-7b9a327dcd41\") " pod="openshift-marketplace/community-operators-j9wl4" Oct 14 08:59:21 crc kubenswrapper[4870]: I1014 08:59:21.554203 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/34086778-4805-47e2-9d8d-7b9a327dcd41-utilities\") pod \"community-operators-j9wl4\" (UID: \"34086778-4805-47e2-9d8d-7b9a327dcd41\") " pod="openshift-marketplace/community-operators-j9wl4" Oct 14 08:59:21 crc kubenswrapper[4870]: I1014 08:59:21.656361 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gq2h9\" (UniqueName: \"kubernetes.io/projected/34086778-4805-47e2-9d8d-7b9a327dcd41-kube-api-access-gq2h9\") pod \"community-operators-j9wl4\" (UID: \"34086778-4805-47e2-9d8d-7b9a327dcd41\") " pod="openshift-marketplace/community-operators-j9wl4" Oct 14 08:59:21 crc kubenswrapper[4870]: I1014 08:59:21.656578 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/34086778-4805-47e2-9d8d-7b9a327dcd41-catalog-content\") pod \"community-operators-j9wl4\" (UID: \"34086778-4805-47e2-9d8d-7b9a327dcd41\") " pod="openshift-marketplace/community-operators-j9wl4" Oct 14 08:59:21 crc kubenswrapper[4870]: I1014 08:59:21.656616 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/34086778-4805-47e2-9d8d-7b9a327dcd41-utilities\") pod \"community-operators-j9wl4\" (UID: \"34086778-4805-47e2-9d8d-7b9a327dcd41\") " pod="openshift-marketplace/community-operators-j9wl4" Oct 14 08:59:21 crc kubenswrapper[4870]: I1014 08:59:21.657163 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/34086778-4805-47e2-9d8d-7b9a327dcd41-utilities\") pod \"community-operators-j9wl4\" (UID: \"34086778-4805-47e2-9d8d-7b9a327dcd41\") " pod="openshift-marketplace/community-operators-j9wl4" Oct 14 08:59:21 crc kubenswrapper[4870]: I1014 08:59:21.658159 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/34086778-4805-47e2-9d8d-7b9a327dcd41-catalog-content\") pod \"community-operators-j9wl4\" (UID: \"34086778-4805-47e2-9d8d-7b9a327dcd41\") " pod="openshift-marketplace/community-operators-j9wl4" Oct 14 08:59:21 crc kubenswrapper[4870]: I1014 08:59:21.676959 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gq2h9\" (UniqueName: \"kubernetes.io/projected/34086778-4805-47e2-9d8d-7b9a327dcd41-kube-api-access-gq2h9\") pod \"community-operators-j9wl4\" (UID: \"34086778-4805-47e2-9d8d-7b9a327dcd41\") " pod="openshift-marketplace/community-operators-j9wl4" Oct 14 08:59:21 crc kubenswrapper[4870]: I1014 08:59:21.807703 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-j9wl4" Oct 14 08:59:25 crc kubenswrapper[4870]: W1014 08:59:25.647332 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod34086778_4805_47e2_9d8d_7b9a327dcd41.slice/crio-f866e32f1f542e6a2ed826d13a13059c5bb85151c0188dd5090d68b000b7d6fb WatchSource:0}: Error finding container f866e32f1f542e6a2ed826d13a13059c5bb85151c0188dd5090d68b000b7d6fb: Status 404 returned error can't find the container with id f866e32f1f542e6a2ed826d13a13059c5bb85151c0188dd5090d68b000b7d6fb Oct 14 08:59:25 crc kubenswrapper[4870]: I1014 08:59:25.655723 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-j9wl4"] Oct 14 08:59:26 crc kubenswrapper[4870]: I1014 08:59:26.340929 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-kxsd4" event={"ID":"57647978-3492-4dcf-85ae-ad3109e97980","Type":"ContainerStarted","Data":"be40d5b0502c566a97b58ff8206fb2ff8e99acc212f7cc433e012aa49735f5cf"} Oct 14 08:59:26 crc kubenswrapper[4870]: I1014 08:59:26.342649 4870 generic.go:334] "Generic (PLEG): container finished" podID="34086778-4805-47e2-9d8d-7b9a327dcd41" containerID="77dd272ab1479a5ca28bbd599e63483d892f3c0a381d1c090dc5bcf3dde539b0" exitCode=0 Oct 14 08:59:26 crc kubenswrapper[4870]: I1014 08:59:26.342778 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-j9wl4" event={"ID":"34086778-4805-47e2-9d8d-7b9a327dcd41","Type":"ContainerDied","Data":"77dd272ab1479a5ca28bbd599e63483d892f3c0a381d1c090dc5bcf3dde539b0"} Oct 14 08:59:26 crc kubenswrapper[4870]: I1014 08:59:26.342885 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-j9wl4" event={"ID":"34086778-4805-47e2-9d8d-7b9a327dcd41","Type":"ContainerStarted","Data":"f866e32f1f542e6a2ed826d13a13059c5bb85151c0188dd5090d68b000b7d6fb"} Oct 14 08:59:26 crc kubenswrapper[4870]: I1014 08:59:26.365121 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-kxsd4" podStartSLOduration=1.89641335 podStartE2EDuration="10.365097608s" podCreationTimestamp="2025-10-14 08:59:16 +0000 UTC" firstStartedPulling="2025-10-14 08:59:16.822779131 +0000 UTC m=+7092.520139512" lastFinishedPulling="2025-10-14 08:59:25.291463399 +0000 UTC m=+7100.988823770" observedRunningTime="2025-10-14 08:59:26.362253618 +0000 UTC m=+7102.059613989" watchObservedRunningTime="2025-10-14 08:59:26.365097608 +0000 UTC m=+7102.062457999" Oct 14 08:59:28 crc kubenswrapper[4870]: I1014 08:59:28.364087 4870 generic.go:334] "Generic (PLEG): container finished" podID="34086778-4805-47e2-9d8d-7b9a327dcd41" containerID="e1cccedd2884abb09c6c3e1a1f1ae35b632d9ab7cc81be04828a640ae6b257ff" exitCode=0 Oct 14 08:59:28 crc kubenswrapper[4870]: I1014 08:59:28.364167 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-j9wl4" event={"ID":"34086778-4805-47e2-9d8d-7b9a327dcd41","Type":"ContainerDied","Data":"e1cccedd2884abb09c6c3e1a1f1ae35b632d9ab7cc81be04828a640ae6b257ff"} Oct 14 08:59:29 crc kubenswrapper[4870]: I1014 08:59:29.377152 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-j9wl4" event={"ID":"34086778-4805-47e2-9d8d-7b9a327dcd41","Type":"ContainerStarted","Data":"a9acd3cff7d1bdae5ee636f689792077f7163dfde6ef599a991f19dc2c4ae8e8"} Oct 14 08:59:30 crc kubenswrapper[4870]: I1014 08:59:30.392823 4870 generic.go:334] "Generic (PLEG): container finished" podID="57647978-3492-4dcf-85ae-ad3109e97980" containerID="be40d5b0502c566a97b58ff8206fb2ff8e99acc212f7cc433e012aa49735f5cf" exitCode=0 Oct 14 08:59:30 crc kubenswrapper[4870]: I1014 08:59:30.392922 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-kxsd4" event={"ID":"57647978-3492-4dcf-85ae-ad3109e97980","Type":"ContainerDied","Data":"be40d5b0502c566a97b58ff8206fb2ff8e99acc212f7cc433e012aa49735f5cf"} Oct 14 08:59:30 crc kubenswrapper[4870]: I1014 08:59:30.413378 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-j9wl4" podStartSLOduration=6.796574704 podStartE2EDuration="9.413356308s" podCreationTimestamp="2025-10-14 08:59:21 +0000 UTC" firstStartedPulling="2025-10-14 08:59:26.345540966 +0000 UTC m=+7102.042901357" lastFinishedPulling="2025-10-14 08:59:28.96232259 +0000 UTC m=+7104.659682961" observedRunningTime="2025-10-14 08:59:29.409830639 +0000 UTC m=+7105.107191010" watchObservedRunningTime="2025-10-14 08:59:30.413356308 +0000 UTC m=+7106.110716689" Oct 14 08:59:31 crc kubenswrapper[4870]: I1014 08:59:31.739193 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-kxsd4" Oct 14 08:59:31 crc kubenswrapper[4870]: I1014 08:59:31.753648 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57647978-3492-4dcf-85ae-ad3109e97980-combined-ca-bundle\") pod \"57647978-3492-4dcf-85ae-ad3109e97980\" (UID: \"57647978-3492-4dcf-85ae-ad3109e97980\") " Oct 14 08:59:31 crc kubenswrapper[4870]: I1014 08:59:31.753722 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-27zp2\" (UniqueName: \"kubernetes.io/projected/57647978-3492-4dcf-85ae-ad3109e97980-kube-api-access-27zp2\") pod \"57647978-3492-4dcf-85ae-ad3109e97980\" (UID: \"57647978-3492-4dcf-85ae-ad3109e97980\") " Oct 14 08:59:31 crc kubenswrapper[4870]: I1014 08:59:31.753752 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/57647978-3492-4dcf-85ae-ad3109e97980-scripts\") pod \"57647978-3492-4dcf-85ae-ad3109e97980\" (UID: \"57647978-3492-4dcf-85ae-ad3109e97980\") " Oct 14 08:59:31 crc kubenswrapper[4870]: I1014 08:59:31.753824 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/57647978-3492-4dcf-85ae-ad3109e97980-config-data\") pod \"57647978-3492-4dcf-85ae-ad3109e97980\" (UID: \"57647978-3492-4dcf-85ae-ad3109e97980\") " Oct 14 08:59:31 crc kubenswrapper[4870]: I1014 08:59:31.759148 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/57647978-3492-4dcf-85ae-ad3109e97980-scripts" (OuterVolumeSpecName: "scripts") pod "57647978-3492-4dcf-85ae-ad3109e97980" (UID: "57647978-3492-4dcf-85ae-ad3109e97980"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:59:31 crc kubenswrapper[4870]: I1014 08:59:31.761390 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57647978-3492-4dcf-85ae-ad3109e97980-kube-api-access-27zp2" (OuterVolumeSpecName: "kube-api-access-27zp2") pod "57647978-3492-4dcf-85ae-ad3109e97980" (UID: "57647978-3492-4dcf-85ae-ad3109e97980"). InnerVolumeSpecName "kube-api-access-27zp2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:59:31 crc kubenswrapper[4870]: I1014 08:59:31.782773 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/57647978-3492-4dcf-85ae-ad3109e97980-config-data" (OuterVolumeSpecName: "config-data") pod "57647978-3492-4dcf-85ae-ad3109e97980" (UID: "57647978-3492-4dcf-85ae-ad3109e97980"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:59:31 crc kubenswrapper[4870]: I1014 08:59:31.787742 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/57647978-3492-4dcf-85ae-ad3109e97980-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "57647978-3492-4dcf-85ae-ad3109e97980" (UID: "57647978-3492-4dcf-85ae-ad3109e97980"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:59:31 crc kubenswrapper[4870]: I1014 08:59:31.809177 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-j9wl4" Oct 14 08:59:31 crc kubenswrapper[4870]: I1014 08:59:31.809240 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-j9wl4" Oct 14 08:59:31 crc kubenswrapper[4870]: I1014 08:59:31.856334 4870 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57647978-3492-4dcf-85ae-ad3109e97980-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 08:59:31 crc kubenswrapper[4870]: I1014 08:59:31.856375 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-27zp2\" (UniqueName: \"kubernetes.io/projected/57647978-3492-4dcf-85ae-ad3109e97980-kube-api-access-27zp2\") on node \"crc\" DevicePath \"\"" Oct 14 08:59:31 crc kubenswrapper[4870]: I1014 08:59:31.856415 4870 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/57647978-3492-4dcf-85ae-ad3109e97980-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 08:59:31 crc kubenswrapper[4870]: I1014 08:59:31.856429 4870 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/57647978-3492-4dcf-85ae-ad3109e97980-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 08:59:31 crc kubenswrapper[4870]: I1014 08:59:31.875839 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-j9wl4" Oct 14 08:59:32 crc kubenswrapper[4870]: I1014 08:59:32.415751 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-kxsd4" event={"ID":"57647978-3492-4dcf-85ae-ad3109e97980","Type":"ContainerDied","Data":"7f94e2bb24fc9cdb260a0f8a8db42d80f65190fe615587820bf5160d89acaa0d"} Oct 14 08:59:32 crc kubenswrapper[4870]: I1014 08:59:32.415816 4870 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7f94e2bb24fc9cdb260a0f8a8db42d80f65190fe615587820bf5160d89acaa0d" Oct 14 08:59:32 crc kubenswrapper[4870]: I1014 08:59:32.415772 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-kxsd4" Oct 14 08:59:32 crc kubenswrapper[4870]: I1014 08:59:32.504067 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 14 08:59:32 crc kubenswrapper[4870]: E1014 08:59:32.504553 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="57647978-3492-4dcf-85ae-ad3109e97980" containerName="nova-cell0-conductor-db-sync" Oct 14 08:59:32 crc kubenswrapper[4870]: I1014 08:59:32.504575 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="57647978-3492-4dcf-85ae-ad3109e97980" containerName="nova-cell0-conductor-db-sync" Oct 14 08:59:32 crc kubenswrapper[4870]: I1014 08:59:32.504809 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="57647978-3492-4dcf-85ae-ad3109e97980" containerName="nova-cell0-conductor-db-sync" Oct 14 08:59:32 crc kubenswrapper[4870]: I1014 08:59:32.505591 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 14 08:59:32 crc kubenswrapper[4870]: I1014 08:59:32.513313 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 14 08:59:32 crc kubenswrapper[4870]: I1014 08:59:32.555405 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Oct 14 08:59:32 crc kubenswrapper[4870]: I1014 08:59:32.555527 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-7qvv5" Oct 14 08:59:32 crc kubenswrapper[4870]: I1014 08:59:32.577317 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/200f8a16-57ae-44d5-bfb3-ba41de932542-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"200f8a16-57ae-44d5-bfb3-ba41de932542\") " pod="openstack/nova-cell0-conductor-0" Oct 14 08:59:32 crc kubenswrapper[4870]: I1014 08:59:32.577372 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r552c\" (UniqueName: \"kubernetes.io/projected/200f8a16-57ae-44d5-bfb3-ba41de932542-kube-api-access-r552c\") pod \"nova-cell0-conductor-0\" (UID: \"200f8a16-57ae-44d5-bfb3-ba41de932542\") " pod="openstack/nova-cell0-conductor-0" Oct 14 08:59:32 crc kubenswrapper[4870]: I1014 08:59:32.577510 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/200f8a16-57ae-44d5-bfb3-ba41de932542-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"200f8a16-57ae-44d5-bfb3-ba41de932542\") " pod="openstack/nova-cell0-conductor-0" Oct 14 08:59:32 crc kubenswrapper[4870]: I1014 08:59:32.679349 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/200f8a16-57ae-44d5-bfb3-ba41de932542-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"200f8a16-57ae-44d5-bfb3-ba41de932542\") " pod="openstack/nova-cell0-conductor-0" Oct 14 08:59:32 crc kubenswrapper[4870]: I1014 08:59:32.679963 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/200f8a16-57ae-44d5-bfb3-ba41de932542-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"200f8a16-57ae-44d5-bfb3-ba41de932542\") " pod="openstack/nova-cell0-conductor-0" Oct 14 08:59:32 crc kubenswrapper[4870]: I1014 08:59:32.679996 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r552c\" (UniqueName: \"kubernetes.io/projected/200f8a16-57ae-44d5-bfb3-ba41de932542-kube-api-access-r552c\") pod \"nova-cell0-conductor-0\" (UID: \"200f8a16-57ae-44d5-bfb3-ba41de932542\") " pod="openstack/nova-cell0-conductor-0" Oct 14 08:59:32 crc kubenswrapper[4870]: I1014 08:59:32.688755 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/200f8a16-57ae-44d5-bfb3-ba41de932542-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"200f8a16-57ae-44d5-bfb3-ba41de932542\") " pod="openstack/nova-cell0-conductor-0" Oct 14 08:59:32 crc kubenswrapper[4870]: I1014 08:59:32.689773 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/200f8a16-57ae-44d5-bfb3-ba41de932542-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"200f8a16-57ae-44d5-bfb3-ba41de932542\") " pod="openstack/nova-cell0-conductor-0" Oct 14 08:59:32 crc kubenswrapper[4870]: I1014 08:59:32.697654 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r552c\" (UniqueName: \"kubernetes.io/projected/200f8a16-57ae-44d5-bfb3-ba41de932542-kube-api-access-r552c\") pod \"nova-cell0-conductor-0\" (UID: \"200f8a16-57ae-44d5-bfb3-ba41de932542\") " pod="openstack/nova-cell0-conductor-0" Oct 14 08:59:32 crc kubenswrapper[4870]: I1014 08:59:32.878788 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 14 08:59:33 crc kubenswrapper[4870]: I1014 08:59:33.337158 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 14 08:59:33 crc kubenswrapper[4870]: I1014 08:59:33.430128 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"200f8a16-57ae-44d5-bfb3-ba41de932542","Type":"ContainerStarted","Data":"f3d2abdea93d4ec269b0a52cd31cd60f1839468d1a0f9c3845b4c4ff82d8e64d"} Oct 14 08:59:34 crc kubenswrapper[4870]: I1014 08:59:34.439042 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"200f8a16-57ae-44d5-bfb3-ba41de932542","Type":"ContainerStarted","Data":"07e0333d969b88cecad4b5dfbe928bb218113231bd5e69a5da72d866efa3ff49"} Oct 14 08:59:34 crc kubenswrapper[4870]: I1014 08:59:34.439533 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Oct 14 08:59:41 crc kubenswrapper[4870]: I1014 08:59:41.864401 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-j9wl4" Oct 14 08:59:41 crc kubenswrapper[4870]: I1014 08:59:41.908771 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=9.908741652 podStartE2EDuration="9.908741652s" podCreationTimestamp="2025-10-14 08:59:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 08:59:34.461424742 +0000 UTC m=+7110.158785113" watchObservedRunningTime="2025-10-14 08:59:41.908741652 +0000 UTC m=+7117.606102063" Oct 14 08:59:41 crc kubenswrapper[4870]: I1014 08:59:41.928151 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-j9wl4"] Oct 14 08:59:42 crc kubenswrapper[4870]: I1014 08:59:42.538427 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-j9wl4" podUID="34086778-4805-47e2-9d8d-7b9a327dcd41" containerName="registry-server" containerID="cri-o://a9acd3cff7d1bdae5ee636f689792077f7163dfde6ef599a991f19dc2c4ae8e8" gracePeriod=2 Oct 14 08:59:42 crc kubenswrapper[4870]: I1014 08:59:42.916074 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Oct 14 08:59:43 crc kubenswrapper[4870]: I1014 08:59:43.047420 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-j9wl4" Oct 14 08:59:43 crc kubenswrapper[4870]: I1014 08:59:43.086415 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/34086778-4805-47e2-9d8d-7b9a327dcd41-utilities\") pod \"34086778-4805-47e2-9d8d-7b9a327dcd41\" (UID: \"34086778-4805-47e2-9d8d-7b9a327dcd41\") " Oct 14 08:59:43 crc kubenswrapper[4870]: I1014 08:59:43.086520 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gq2h9\" (UniqueName: \"kubernetes.io/projected/34086778-4805-47e2-9d8d-7b9a327dcd41-kube-api-access-gq2h9\") pod \"34086778-4805-47e2-9d8d-7b9a327dcd41\" (UID: \"34086778-4805-47e2-9d8d-7b9a327dcd41\") " Oct 14 08:59:43 crc kubenswrapper[4870]: I1014 08:59:43.086673 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/34086778-4805-47e2-9d8d-7b9a327dcd41-catalog-content\") pod \"34086778-4805-47e2-9d8d-7b9a327dcd41\" (UID: \"34086778-4805-47e2-9d8d-7b9a327dcd41\") " Oct 14 08:59:43 crc kubenswrapper[4870]: I1014 08:59:43.087402 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/34086778-4805-47e2-9d8d-7b9a327dcd41-utilities" (OuterVolumeSpecName: "utilities") pod "34086778-4805-47e2-9d8d-7b9a327dcd41" (UID: "34086778-4805-47e2-9d8d-7b9a327dcd41"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:59:43 crc kubenswrapper[4870]: I1014 08:59:43.091593 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/34086778-4805-47e2-9d8d-7b9a327dcd41-kube-api-access-gq2h9" (OuterVolumeSpecName: "kube-api-access-gq2h9") pod "34086778-4805-47e2-9d8d-7b9a327dcd41" (UID: "34086778-4805-47e2-9d8d-7b9a327dcd41"). InnerVolumeSpecName "kube-api-access-gq2h9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:59:43 crc kubenswrapper[4870]: I1014 08:59:43.133157 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/34086778-4805-47e2-9d8d-7b9a327dcd41-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "34086778-4805-47e2-9d8d-7b9a327dcd41" (UID: "34086778-4805-47e2-9d8d-7b9a327dcd41"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:59:43 crc kubenswrapper[4870]: I1014 08:59:43.188986 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gq2h9\" (UniqueName: \"kubernetes.io/projected/34086778-4805-47e2-9d8d-7b9a327dcd41-kube-api-access-gq2h9\") on node \"crc\" DevicePath \"\"" Oct 14 08:59:43 crc kubenswrapper[4870]: I1014 08:59:43.189266 4870 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/34086778-4805-47e2-9d8d-7b9a327dcd41-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 08:59:43 crc kubenswrapper[4870]: I1014 08:59:43.189371 4870 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/34086778-4805-47e2-9d8d-7b9a327dcd41-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 08:59:43 crc kubenswrapper[4870]: I1014 08:59:43.447566 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-mvlrf"] Oct 14 08:59:43 crc kubenswrapper[4870]: E1014 08:59:43.448103 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34086778-4805-47e2-9d8d-7b9a327dcd41" containerName="extract-content" Oct 14 08:59:43 crc kubenswrapper[4870]: I1014 08:59:43.448132 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="34086778-4805-47e2-9d8d-7b9a327dcd41" containerName="extract-content" Oct 14 08:59:43 crc kubenswrapper[4870]: E1014 08:59:43.448164 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34086778-4805-47e2-9d8d-7b9a327dcd41" containerName="extract-utilities" Oct 14 08:59:43 crc kubenswrapper[4870]: I1014 08:59:43.448173 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="34086778-4805-47e2-9d8d-7b9a327dcd41" containerName="extract-utilities" Oct 14 08:59:43 crc kubenswrapper[4870]: E1014 08:59:43.448198 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34086778-4805-47e2-9d8d-7b9a327dcd41" containerName="registry-server" Oct 14 08:59:43 crc kubenswrapper[4870]: I1014 08:59:43.448206 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="34086778-4805-47e2-9d8d-7b9a327dcd41" containerName="registry-server" Oct 14 08:59:43 crc kubenswrapper[4870]: I1014 08:59:43.448518 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="34086778-4805-47e2-9d8d-7b9a327dcd41" containerName="registry-server" Oct 14 08:59:43 crc kubenswrapper[4870]: I1014 08:59:43.449293 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-mvlrf" Oct 14 08:59:43 crc kubenswrapper[4870]: I1014 08:59:43.455755 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Oct 14 08:59:43 crc kubenswrapper[4870]: I1014 08:59:43.463165 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Oct 14 08:59:43 crc kubenswrapper[4870]: I1014 08:59:43.486384 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-mvlrf"] Oct 14 08:59:43 crc kubenswrapper[4870]: I1014 08:59:43.493332 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ds85t\" (UniqueName: \"kubernetes.io/projected/c0c2d021-038b-4127-bceb-bbe26c1f3070-kube-api-access-ds85t\") pod \"nova-cell0-cell-mapping-mvlrf\" (UID: \"c0c2d021-038b-4127-bceb-bbe26c1f3070\") " pod="openstack/nova-cell0-cell-mapping-mvlrf" Oct 14 08:59:43 crc kubenswrapper[4870]: I1014 08:59:43.493554 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c0c2d021-038b-4127-bceb-bbe26c1f3070-config-data\") pod \"nova-cell0-cell-mapping-mvlrf\" (UID: \"c0c2d021-038b-4127-bceb-bbe26c1f3070\") " pod="openstack/nova-cell0-cell-mapping-mvlrf" Oct 14 08:59:43 crc kubenswrapper[4870]: I1014 08:59:43.493631 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c0c2d021-038b-4127-bceb-bbe26c1f3070-scripts\") pod \"nova-cell0-cell-mapping-mvlrf\" (UID: \"c0c2d021-038b-4127-bceb-bbe26c1f3070\") " pod="openstack/nova-cell0-cell-mapping-mvlrf" Oct 14 08:59:43 crc kubenswrapper[4870]: I1014 08:59:43.493677 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c0c2d021-038b-4127-bceb-bbe26c1f3070-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-mvlrf\" (UID: \"c0c2d021-038b-4127-bceb-bbe26c1f3070\") " pod="openstack/nova-cell0-cell-mapping-mvlrf" Oct 14 08:59:43 crc kubenswrapper[4870]: I1014 08:59:43.550676 4870 generic.go:334] "Generic (PLEG): container finished" podID="34086778-4805-47e2-9d8d-7b9a327dcd41" containerID="a9acd3cff7d1bdae5ee636f689792077f7163dfde6ef599a991f19dc2c4ae8e8" exitCode=0 Oct 14 08:59:43 crc kubenswrapper[4870]: I1014 08:59:43.550728 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-j9wl4" event={"ID":"34086778-4805-47e2-9d8d-7b9a327dcd41","Type":"ContainerDied","Data":"a9acd3cff7d1bdae5ee636f689792077f7163dfde6ef599a991f19dc2c4ae8e8"} Oct 14 08:59:43 crc kubenswrapper[4870]: I1014 08:59:43.550760 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-j9wl4" event={"ID":"34086778-4805-47e2-9d8d-7b9a327dcd41","Type":"ContainerDied","Data":"f866e32f1f542e6a2ed826d13a13059c5bb85151c0188dd5090d68b000b7d6fb"} Oct 14 08:59:43 crc kubenswrapper[4870]: I1014 08:59:43.550779 4870 scope.go:117] "RemoveContainer" containerID="a9acd3cff7d1bdae5ee636f689792077f7163dfde6ef599a991f19dc2c4ae8e8" Oct 14 08:59:43 crc kubenswrapper[4870]: I1014 08:59:43.550950 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-j9wl4" Oct 14 08:59:43 crc kubenswrapper[4870]: I1014 08:59:43.594401 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ds85t\" (UniqueName: \"kubernetes.io/projected/c0c2d021-038b-4127-bceb-bbe26c1f3070-kube-api-access-ds85t\") pod \"nova-cell0-cell-mapping-mvlrf\" (UID: \"c0c2d021-038b-4127-bceb-bbe26c1f3070\") " pod="openstack/nova-cell0-cell-mapping-mvlrf" Oct 14 08:59:43 crc kubenswrapper[4870]: I1014 08:59:43.594559 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c0c2d021-038b-4127-bceb-bbe26c1f3070-config-data\") pod \"nova-cell0-cell-mapping-mvlrf\" (UID: \"c0c2d021-038b-4127-bceb-bbe26c1f3070\") " pod="openstack/nova-cell0-cell-mapping-mvlrf" Oct 14 08:59:43 crc kubenswrapper[4870]: I1014 08:59:43.594630 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c0c2d021-038b-4127-bceb-bbe26c1f3070-scripts\") pod \"nova-cell0-cell-mapping-mvlrf\" (UID: \"c0c2d021-038b-4127-bceb-bbe26c1f3070\") " pod="openstack/nova-cell0-cell-mapping-mvlrf" Oct 14 08:59:43 crc kubenswrapper[4870]: I1014 08:59:43.594661 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c0c2d021-038b-4127-bceb-bbe26c1f3070-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-mvlrf\" (UID: \"c0c2d021-038b-4127-bceb-bbe26c1f3070\") " pod="openstack/nova-cell0-cell-mapping-mvlrf" Oct 14 08:59:43 crc kubenswrapper[4870]: I1014 08:59:43.596109 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-j9wl4"] Oct 14 08:59:43 crc kubenswrapper[4870]: I1014 08:59:43.601702 4870 scope.go:117] "RemoveContainer" containerID="e1cccedd2884abb09c6c3e1a1f1ae35b632d9ab7cc81be04828a640ae6b257ff" Oct 14 08:59:43 crc kubenswrapper[4870]: I1014 08:59:43.603004 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c0c2d021-038b-4127-bceb-bbe26c1f3070-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-mvlrf\" (UID: \"c0c2d021-038b-4127-bceb-bbe26c1f3070\") " pod="openstack/nova-cell0-cell-mapping-mvlrf" Oct 14 08:59:43 crc kubenswrapper[4870]: I1014 08:59:43.614539 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ds85t\" (UniqueName: \"kubernetes.io/projected/c0c2d021-038b-4127-bceb-bbe26c1f3070-kube-api-access-ds85t\") pod \"nova-cell0-cell-mapping-mvlrf\" (UID: \"c0c2d021-038b-4127-bceb-bbe26c1f3070\") " pod="openstack/nova-cell0-cell-mapping-mvlrf" Oct 14 08:59:43 crc kubenswrapper[4870]: I1014 08:59:43.615274 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c0c2d021-038b-4127-bceb-bbe26c1f3070-config-data\") pod \"nova-cell0-cell-mapping-mvlrf\" (UID: \"c0c2d021-038b-4127-bceb-bbe26c1f3070\") " pod="openstack/nova-cell0-cell-mapping-mvlrf" Oct 14 08:59:43 crc kubenswrapper[4870]: I1014 08:59:43.624058 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-j9wl4"] Oct 14 08:59:43 crc kubenswrapper[4870]: I1014 08:59:43.625818 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c0c2d021-038b-4127-bceb-bbe26c1f3070-scripts\") pod \"nova-cell0-cell-mapping-mvlrf\" (UID: \"c0c2d021-038b-4127-bceb-bbe26c1f3070\") " pod="openstack/nova-cell0-cell-mapping-mvlrf" Oct 14 08:59:43 crc kubenswrapper[4870]: I1014 08:59:43.653515 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 14 08:59:43 crc kubenswrapper[4870]: I1014 08:59:43.655452 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 14 08:59:43 crc kubenswrapper[4870]: I1014 08:59:43.660904 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 14 08:59:43 crc kubenswrapper[4870]: I1014 08:59:43.675398 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 14 08:59:43 crc kubenswrapper[4870]: I1014 08:59:43.687113 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 14 08:59:43 crc kubenswrapper[4870]: I1014 08:59:43.691064 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 14 08:59:43 crc kubenswrapper[4870]: I1014 08:59:43.696070 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 14 08:59:43 crc kubenswrapper[4870]: I1014 08:59:43.701524 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 14 08:59:43 crc kubenswrapper[4870]: I1014 08:59:43.758563 4870 scope.go:117] "RemoveContainer" containerID="77dd272ab1479a5ca28bbd599e63483d892f3c0a381d1c090dc5bcf3dde539b0" Oct 14 08:59:43 crc kubenswrapper[4870]: I1014 08:59:43.769967 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-mvlrf" Oct 14 08:59:43 crc kubenswrapper[4870]: I1014 08:59:43.779677 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 14 08:59:43 crc kubenswrapper[4870]: I1014 08:59:43.785536 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 14 08:59:43 crc kubenswrapper[4870]: I1014 08:59:43.787102 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 14 08:59:43 crc kubenswrapper[4870]: I1014 08:59:43.791671 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 14 08:59:43 crc kubenswrapper[4870]: I1014 08:59:43.799658 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p7kxv\" (UniqueName: \"kubernetes.io/projected/057b319f-9b0f-4c6d-ae9c-3a18ded7d377-kube-api-access-p7kxv\") pod \"nova-metadata-0\" (UID: \"057b319f-9b0f-4c6d-ae9c-3a18ded7d377\") " pod="openstack/nova-metadata-0" Oct 14 08:59:43 crc kubenswrapper[4870]: I1014 08:59:43.799745 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3b91331d-394b-4f04-be17-96fd0a71fc6e-config-data\") pod \"nova-api-0\" (UID: \"3b91331d-394b-4f04-be17-96fd0a71fc6e\") " pod="openstack/nova-api-0" Oct 14 08:59:43 crc kubenswrapper[4870]: I1014 08:59:43.799831 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3b91331d-394b-4f04-be17-96fd0a71fc6e-logs\") pod \"nova-api-0\" (UID: \"3b91331d-394b-4f04-be17-96fd0a71fc6e\") " pod="openstack/nova-api-0" Oct 14 08:59:43 crc kubenswrapper[4870]: I1014 08:59:43.799861 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tvtft\" (UniqueName: \"kubernetes.io/projected/3b91331d-394b-4f04-be17-96fd0a71fc6e-kube-api-access-tvtft\") pod \"nova-api-0\" (UID: \"3b91331d-394b-4f04-be17-96fd0a71fc6e\") " pod="openstack/nova-api-0" Oct 14 08:59:43 crc kubenswrapper[4870]: I1014 08:59:43.799885 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/057b319f-9b0f-4c6d-ae9c-3a18ded7d377-logs\") pod \"nova-metadata-0\" (UID: \"057b319f-9b0f-4c6d-ae9c-3a18ded7d377\") " pod="openstack/nova-metadata-0" Oct 14 08:59:43 crc kubenswrapper[4870]: I1014 08:59:43.799912 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b91331d-394b-4f04-be17-96fd0a71fc6e-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"3b91331d-394b-4f04-be17-96fd0a71fc6e\") " pod="openstack/nova-api-0" Oct 14 08:59:43 crc kubenswrapper[4870]: I1014 08:59:43.799970 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/057b319f-9b0f-4c6d-ae9c-3a18ded7d377-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"057b319f-9b0f-4c6d-ae9c-3a18ded7d377\") " pod="openstack/nova-metadata-0" Oct 14 08:59:43 crc kubenswrapper[4870]: I1014 08:59:43.800015 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/057b319f-9b0f-4c6d-ae9c-3a18ded7d377-config-data\") pod \"nova-metadata-0\" (UID: \"057b319f-9b0f-4c6d-ae9c-3a18ded7d377\") " pod="openstack/nova-metadata-0" Oct 14 08:59:43 crc kubenswrapper[4870]: I1014 08:59:43.820050 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 14 08:59:43 crc kubenswrapper[4870]: I1014 08:59:43.821319 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 14 08:59:43 crc kubenswrapper[4870]: I1014 08:59:43.843006 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Oct 14 08:59:43 crc kubenswrapper[4870]: I1014 08:59:43.863146 4870 scope.go:117] "RemoveContainer" containerID="a9acd3cff7d1bdae5ee636f689792077f7163dfde6ef599a991f19dc2c4ae8e8" Oct 14 08:59:43 crc kubenswrapper[4870]: E1014 08:59:43.866864 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a9acd3cff7d1bdae5ee636f689792077f7163dfde6ef599a991f19dc2c4ae8e8\": container with ID starting with a9acd3cff7d1bdae5ee636f689792077f7163dfde6ef599a991f19dc2c4ae8e8 not found: ID does not exist" containerID="a9acd3cff7d1bdae5ee636f689792077f7163dfde6ef599a991f19dc2c4ae8e8" Oct 14 08:59:43 crc kubenswrapper[4870]: I1014 08:59:43.866930 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a9acd3cff7d1bdae5ee636f689792077f7163dfde6ef599a991f19dc2c4ae8e8"} err="failed to get container status \"a9acd3cff7d1bdae5ee636f689792077f7163dfde6ef599a991f19dc2c4ae8e8\": rpc error: code = NotFound desc = could not find container \"a9acd3cff7d1bdae5ee636f689792077f7163dfde6ef599a991f19dc2c4ae8e8\": container with ID starting with a9acd3cff7d1bdae5ee636f689792077f7163dfde6ef599a991f19dc2c4ae8e8 not found: ID does not exist" Oct 14 08:59:43 crc kubenswrapper[4870]: I1014 08:59:43.866961 4870 scope.go:117] "RemoveContainer" containerID="e1cccedd2884abb09c6c3e1a1f1ae35b632d9ab7cc81be04828a640ae6b257ff" Oct 14 08:59:43 crc kubenswrapper[4870]: E1014 08:59:43.867262 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e1cccedd2884abb09c6c3e1a1f1ae35b632d9ab7cc81be04828a640ae6b257ff\": container with ID starting with e1cccedd2884abb09c6c3e1a1f1ae35b632d9ab7cc81be04828a640ae6b257ff not found: ID does not exist" containerID="e1cccedd2884abb09c6c3e1a1f1ae35b632d9ab7cc81be04828a640ae6b257ff" Oct 14 08:59:43 crc kubenswrapper[4870]: I1014 08:59:43.867294 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e1cccedd2884abb09c6c3e1a1f1ae35b632d9ab7cc81be04828a640ae6b257ff"} err="failed to get container status \"e1cccedd2884abb09c6c3e1a1f1ae35b632d9ab7cc81be04828a640ae6b257ff\": rpc error: code = NotFound desc = could not find container \"e1cccedd2884abb09c6c3e1a1f1ae35b632d9ab7cc81be04828a640ae6b257ff\": container with ID starting with e1cccedd2884abb09c6c3e1a1f1ae35b632d9ab7cc81be04828a640ae6b257ff not found: ID does not exist" Oct 14 08:59:43 crc kubenswrapper[4870]: I1014 08:59:43.867315 4870 scope.go:117] "RemoveContainer" containerID="77dd272ab1479a5ca28bbd599e63483d892f3c0a381d1c090dc5bcf3dde539b0" Oct 14 08:59:43 crc kubenswrapper[4870]: E1014 08:59:43.867591 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"77dd272ab1479a5ca28bbd599e63483d892f3c0a381d1c090dc5bcf3dde539b0\": container with ID starting with 77dd272ab1479a5ca28bbd599e63483d892f3c0a381d1c090dc5bcf3dde539b0 not found: ID does not exist" containerID="77dd272ab1479a5ca28bbd599e63483d892f3c0a381d1c090dc5bcf3dde539b0" Oct 14 08:59:43 crc kubenswrapper[4870]: I1014 08:59:43.867621 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"77dd272ab1479a5ca28bbd599e63483d892f3c0a381d1c090dc5bcf3dde539b0"} err="failed to get container status \"77dd272ab1479a5ca28bbd599e63483d892f3c0a381d1c090dc5bcf3dde539b0\": rpc error: code = NotFound desc = could not find container \"77dd272ab1479a5ca28bbd599e63483d892f3c0a381d1c090dc5bcf3dde539b0\": container with ID starting with 77dd272ab1479a5ca28bbd599e63483d892f3c0a381d1c090dc5bcf3dde539b0 not found: ID does not exist" Oct 14 08:59:43 crc kubenswrapper[4870]: I1014 08:59:43.873844 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7c8cbf976c-khp98"] Oct 14 08:59:43 crc kubenswrapper[4870]: I1014 08:59:43.901367 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7c8cbf976c-khp98" Oct 14 08:59:43 crc kubenswrapper[4870]: I1014 08:59:43.909052 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/057b319f-9b0f-4c6d-ae9c-3a18ded7d377-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"057b319f-9b0f-4c6d-ae9c-3a18ded7d377\") " pod="openstack/nova-metadata-0" Oct 14 08:59:43 crc kubenswrapper[4870]: I1014 08:59:43.909158 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/057b319f-9b0f-4c6d-ae9c-3a18ded7d377-config-data\") pod \"nova-metadata-0\" (UID: \"057b319f-9b0f-4c6d-ae9c-3a18ded7d377\") " pod="openstack/nova-metadata-0" Oct 14 08:59:43 crc kubenswrapper[4870]: I1014 08:59:43.909263 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p7kxv\" (UniqueName: \"kubernetes.io/projected/057b319f-9b0f-4c6d-ae9c-3a18ded7d377-kube-api-access-p7kxv\") pod \"nova-metadata-0\" (UID: \"057b319f-9b0f-4c6d-ae9c-3a18ded7d377\") " pod="openstack/nova-metadata-0" Oct 14 08:59:43 crc kubenswrapper[4870]: I1014 08:59:43.909356 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93eea128-7b79-4505-b492-9c34adf49cdf-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"93eea128-7b79-4505-b492-9c34adf49cdf\") " pod="openstack/nova-scheduler-0" Oct 14 08:59:43 crc kubenswrapper[4870]: I1014 08:59:43.909402 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3b91331d-394b-4f04-be17-96fd0a71fc6e-config-data\") pod \"nova-api-0\" (UID: \"3b91331d-394b-4f04-be17-96fd0a71fc6e\") " pod="openstack/nova-api-0" Oct 14 08:59:43 crc kubenswrapper[4870]: I1014 08:59:43.909502 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/93eea128-7b79-4505-b492-9c34adf49cdf-config-data\") pod \"nova-scheduler-0\" (UID: \"93eea128-7b79-4505-b492-9c34adf49cdf\") " pod="openstack/nova-scheduler-0" Oct 14 08:59:43 crc kubenswrapper[4870]: I1014 08:59:43.909585 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3b91331d-394b-4f04-be17-96fd0a71fc6e-logs\") pod \"nova-api-0\" (UID: \"3b91331d-394b-4f04-be17-96fd0a71fc6e\") " pod="openstack/nova-api-0" Oct 14 08:59:43 crc kubenswrapper[4870]: I1014 08:59:43.909632 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tvtft\" (UniqueName: \"kubernetes.io/projected/3b91331d-394b-4f04-be17-96fd0a71fc6e-kube-api-access-tvtft\") pod \"nova-api-0\" (UID: \"3b91331d-394b-4f04-be17-96fd0a71fc6e\") " pod="openstack/nova-api-0" Oct 14 08:59:43 crc kubenswrapper[4870]: I1014 08:59:43.909665 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pmdjc\" (UniqueName: \"kubernetes.io/projected/93eea128-7b79-4505-b492-9c34adf49cdf-kube-api-access-pmdjc\") pod \"nova-scheduler-0\" (UID: \"93eea128-7b79-4505-b492-9c34adf49cdf\") " pod="openstack/nova-scheduler-0" Oct 14 08:59:43 crc kubenswrapper[4870]: I1014 08:59:43.909696 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/057b319f-9b0f-4c6d-ae9c-3a18ded7d377-logs\") pod \"nova-metadata-0\" (UID: \"057b319f-9b0f-4c6d-ae9c-3a18ded7d377\") " pod="openstack/nova-metadata-0" Oct 14 08:59:43 crc kubenswrapper[4870]: I1014 08:59:43.909741 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b91331d-394b-4f04-be17-96fd0a71fc6e-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"3b91331d-394b-4f04-be17-96fd0a71fc6e\") " pod="openstack/nova-api-0" Oct 14 08:59:43 crc kubenswrapper[4870]: I1014 08:59:43.919171 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3b91331d-394b-4f04-be17-96fd0a71fc6e-logs\") pod \"nova-api-0\" (UID: \"3b91331d-394b-4f04-be17-96fd0a71fc6e\") " pod="openstack/nova-api-0" Oct 14 08:59:43 crc kubenswrapper[4870]: I1014 08:59:43.922135 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/057b319f-9b0f-4c6d-ae9c-3a18ded7d377-logs\") pod \"nova-metadata-0\" (UID: \"057b319f-9b0f-4c6d-ae9c-3a18ded7d377\") " pod="openstack/nova-metadata-0" Oct 14 08:59:43 crc kubenswrapper[4870]: I1014 08:59:43.925632 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b91331d-394b-4f04-be17-96fd0a71fc6e-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"3b91331d-394b-4f04-be17-96fd0a71fc6e\") " pod="openstack/nova-api-0" Oct 14 08:59:43 crc kubenswrapper[4870]: I1014 08:59:43.929011 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3b91331d-394b-4f04-be17-96fd0a71fc6e-config-data\") pod \"nova-api-0\" (UID: \"3b91331d-394b-4f04-be17-96fd0a71fc6e\") " pod="openstack/nova-api-0" Oct 14 08:59:43 crc kubenswrapper[4870]: I1014 08:59:43.928972 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/057b319f-9b0f-4c6d-ae9c-3a18ded7d377-config-data\") pod \"nova-metadata-0\" (UID: \"057b319f-9b0f-4c6d-ae9c-3a18ded7d377\") " pod="openstack/nova-metadata-0" Oct 14 08:59:43 crc kubenswrapper[4870]: I1014 08:59:43.939611 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tvtft\" (UniqueName: \"kubernetes.io/projected/3b91331d-394b-4f04-be17-96fd0a71fc6e-kube-api-access-tvtft\") pod \"nova-api-0\" (UID: \"3b91331d-394b-4f04-be17-96fd0a71fc6e\") " pod="openstack/nova-api-0" Oct 14 08:59:43 crc kubenswrapper[4870]: I1014 08:59:43.947190 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/057b319f-9b0f-4c6d-ae9c-3a18ded7d377-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"057b319f-9b0f-4c6d-ae9c-3a18ded7d377\") " pod="openstack/nova-metadata-0" Oct 14 08:59:43 crc kubenswrapper[4870]: I1014 08:59:43.947779 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p7kxv\" (UniqueName: \"kubernetes.io/projected/057b319f-9b0f-4c6d-ae9c-3a18ded7d377-kube-api-access-p7kxv\") pod \"nova-metadata-0\" (UID: \"057b319f-9b0f-4c6d-ae9c-3a18ded7d377\") " pod="openstack/nova-metadata-0" Oct 14 08:59:43 crc kubenswrapper[4870]: I1014 08:59:43.968281 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 14 08:59:43 crc kubenswrapper[4870]: I1014 08:59:43.980508 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7c8cbf976c-khp98"] Oct 14 08:59:44 crc kubenswrapper[4870]: I1014 08:59:44.031052 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k8lx7\" (UniqueName: \"kubernetes.io/projected/36c98e9a-274d-466a-8e68-b2804be28695-kube-api-access-k8lx7\") pod \"dnsmasq-dns-7c8cbf976c-khp98\" (UID: \"36c98e9a-274d-466a-8e68-b2804be28695\") " pod="openstack/dnsmasq-dns-7c8cbf976c-khp98" Oct 14 08:59:44 crc kubenswrapper[4870]: I1014 08:59:44.031623 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a29f723-aa05-46bd-a2f9-a4422fe7646f-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"1a29f723-aa05-46bd-a2f9-a4422fe7646f\") " pod="openstack/nova-cell1-novncproxy-0" Oct 14 08:59:44 crc kubenswrapper[4870]: I1014 08:59:44.031715 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93eea128-7b79-4505-b492-9c34adf49cdf-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"93eea128-7b79-4505-b492-9c34adf49cdf\") " pod="openstack/nova-scheduler-0" Oct 14 08:59:44 crc kubenswrapper[4870]: I1014 08:59:44.031759 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4qxfd\" (UniqueName: \"kubernetes.io/projected/1a29f723-aa05-46bd-a2f9-a4422fe7646f-kube-api-access-4qxfd\") pod \"nova-cell1-novncproxy-0\" (UID: \"1a29f723-aa05-46bd-a2f9-a4422fe7646f\") " pod="openstack/nova-cell1-novncproxy-0" Oct 14 08:59:44 crc kubenswrapper[4870]: I1014 08:59:44.031789 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/36c98e9a-274d-466a-8e68-b2804be28695-config\") pod \"dnsmasq-dns-7c8cbf976c-khp98\" (UID: \"36c98e9a-274d-466a-8e68-b2804be28695\") " pod="openstack/dnsmasq-dns-7c8cbf976c-khp98" Oct 14 08:59:44 crc kubenswrapper[4870]: I1014 08:59:44.031817 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/36c98e9a-274d-466a-8e68-b2804be28695-dns-svc\") pod \"dnsmasq-dns-7c8cbf976c-khp98\" (UID: \"36c98e9a-274d-466a-8e68-b2804be28695\") " pod="openstack/dnsmasq-dns-7c8cbf976c-khp98" Oct 14 08:59:44 crc kubenswrapper[4870]: I1014 08:59:44.031948 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/93eea128-7b79-4505-b492-9c34adf49cdf-config-data\") pod \"nova-scheduler-0\" (UID: \"93eea128-7b79-4505-b492-9c34adf49cdf\") " pod="openstack/nova-scheduler-0" Oct 14 08:59:44 crc kubenswrapper[4870]: I1014 08:59:44.032048 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/36c98e9a-274d-466a-8e68-b2804be28695-ovsdbserver-nb\") pod \"dnsmasq-dns-7c8cbf976c-khp98\" (UID: \"36c98e9a-274d-466a-8e68-b2804be28695\") " pod="openstack/dnsmasq-dns-7c8cbf976c-khp98" Oct 14 08:59:44 crc kubenswrapper[4870]: I1014 08:59:44.032094 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pmdjc\" (UniqueName: \"kubernetes.io/projected/93eea128-7b79-4505-b492-9c34adf49cdf-kube-api-access-pmdjc\") pod \"nova-scheduler-0\" (UID: \"93eea128-7b79-4505-b492-9c34adf49cdf\") " pod="openstack/nova-scheduler-0" Oct 14 08:59:44 crc kubenswrapper[4870]: I1014 08:59:44.032158 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1a29f723-aa05-46bd-a2f9-a4422fe7646f-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"1a29f723-aa05-46bd-a2f9-a4422fe7646f\") " pod="openstack/nova-cell1-novncproxy-0" Oct 14 08:59:44 crc kubenswrapper[4870]: I1014 08:59:44.032199 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/36c98e9a-274d-466a-8e68-b2804be28695-ovsdbserver-sb\") pod \"dnsmasq-dns-7c8cbf976c-khp98\" (UID: \"36c98e9a-274d-466a-8e68-b2804be28695\") " pod="openstack/dnsmasq-dns-7c8cbf976c-khp98" Oct 14 08:59:44 crc kubenswrapper[4870]: I1014 08:59:44.035186 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/93eea128-7b79-4505-b492-9c34adf49cdf-config-data\") pod \"nova-scheduler-0\" (UID: \"93eea128-7b79-4505-b492-9c34adf49cdf\") " pod="openstack/nova-scheduler-0" Oct 14 08:59:44 crc kubenswrapper[4870]: I1014 08:59:44.038132 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93eea128-7b79-4505-b492-9c34adf49cdf-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"93eea128-7b79-4505-b492-9c34adf49cdf\") " pod="openstack/nova-scheduler-0" Oct 14 08:59:44 crc kubenswrapper[4870]: I1014 08:59:44.055361 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pmdjc\" (UniqueName: \"kubernetes.io/projected/93eea128-7b79-4505-b492-9c34adf49cdf-kube-api-access-pmdjc\") pod \"nova-scheduler-0\" (UID: \"93eea128-7b79-4505-b492-9c34adf49cdf\") " pod="openstack/nova-scheduler-0" Oct 14 08:59:44 crc kubenswrapper[4870]: I1014 08:59:44.083175 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 14 08:59:44 crc kubenswrapper[4870]: I1014 08:59:44.134339 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/36c98e9a-274d-466a-8e68-b2804be28695-ovsdbserver-sb\") pod \"dnsmasq-dns-7c8cbf976c-khp98\" (UID: \"36c98e9a-274d-466a-8e68-b2804be28695\") " pod="openstack/dnsmasq-dns-7c8cbf976c-khp98" Oct 14 08:59:44 crc kubenswrapper[4870]: I1014 08:59:44.134688 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k8lx7\" (UniqueName: \"kubernetes.io/projected/36c98e9a-274d-466a-8e68-b2804be28695-kube-api-access-k8lx7\") pod \"dnsmasq-dns-7c8cbf976c-khp98\" (UID: \"36c98e9a-274d-466a-8e68-b2804be28695\") " pod="openstack/dnsmasq-dns-7c8cbf976c-khp98" Oct 14 08:59:44 crc kubenswrapper[4870]: I1014 08:59:44.134785 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a29f723-aa05-46bd-a2f9-a4422fe7646f-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"1a29f723-aa05-46bd-a2f9-a4422fe7646f\") " pod="openstack/nova-cell1-novncproxy-0" Oct 14 08:59:44 crc kubenswrapper[4870]: I1014 08:59:44.134872 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4qxfd\" (UniqueName: \"kubernetes.io/projected/1a29f723-aa05-46bd-a2f9-a4422fe7646f-kube-api-access-4qxfd\") pod \"nova-cell1-novncproxy-0\" (UID: \"1a29f723-aa05-46bd-a2f9-a4422fe7646f\") " pod="openstack/nova-cell1-novncproxy-0" Oct 14 08:59:44 crc kubenswrapper[4870]: I1014 08:59:44.134972 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/36c98e9a-274d-466a-8e68-b2804be28695-config\") pod \"dnsmasq-dns-7c8cbf976c-khp98\" (UID: \"36c98e9a-274d-466a-8e68-b2804be28695\") " pod="openstack/dnsmasq-dns-7c8cbf976c-khp98" Oct 14 08:59:44 crc kubenswrapper[4870]: I1014 08:59:44.135047 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/36c98e9a-274d-466a-8e68-b2804be28695-dns-svc\") pod \"dnsmasq-dns-7c8cbf976c-khp98\" (UID: \"36c98e9a-274d-466a-8e68-b2804be28695\") " pod="openstack/dnsmasq-dns-7c8cbf976c-khp98" Oct 14 08:59:44 crc kubenswrapper[4870]: I1014 08:59:44.135170 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/36c98e9a-274d-466a-8e68-b2804be28695-ovsdbserver-nb\") pod \"dnsmasq-dns-7c8cbf976c-khp98\" (UID: \"36c98e9a-274d-466a-8e68-b2804be28695\") " pod="openstack/dnsmasq-dns-7c8cbf976c-khp98" Oct 14 08:59:44 crc kubenswrapper[4870]: I1014 08:59:44.135278 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1a29f723-aa05-46bd-a2f9-a4422fe7646f-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"1a29f723-aa05-46bd-a2f9-a4422fe7646f\") " pod="openstack/nova-cell1-novncproxy-0" Oct 14 08:59:44 crc kubenswrapper[4870]: I1014 08:59:44.137228 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/36c98e9a-274d-466a-8e68-b2804be28695-ovsdbserver-sb\") pod \"dnsmasq-dns-7c8cbf976c-khp98\" (UID: \"36c98e9a-274d-466a-8e68-b2804be28695\") " pod="openstack/dnsmasq-dns-7c8cbf976c-khp98" Oct 14 08:59:44 crc kubenswrapper[4870]: I1014 08:59:44.138419 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/36c98e9a-274d-466a-8e68-b2804be28695-ovsdbserver-nb\") pod \"dnsmasq-dns-7c8cbf976c-khp98\" (UID: \"36c98e9a-274d-466a-8e68-b2804be28695\") " pod="openstack/dnsmasq-dns-7c8cbf976c-khp98" Oct 14 08:59:44 crc kubenswrapper[4870]: I1014 08:59:44.138786 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/36c98e9a-274d-466a-8e68-b2804be28695-dns-svc\") pod \"dnsmasq-dns-7c8cbf976c-khp98\" (UID: \"36c98e9a-274d-466a-8e68-b2804be28695\") " pod="openstack/dnsmasq-dns-7c8cbf976c-khp98" Oct 14 08:59:44 crc kubenswrapper[4870]: I1014 08:59:44.139269 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1a29f723-aa05-46bd-a2f9-a4422fe7646f-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"1a29f723-aa05-46bd-a2f9-a4422fe7646f\") " pod="openstack/nova-cell1-novncproxy-0" Oct 14 08:59:44 crc kubenswrapper[4870]: I1014 08:59:44.140725 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a29f723-aa05-46bd-a2f9-a4422fe7646f-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"1a29f723-aa05-46bd-a2f9-a4422fe7646f\") " pod="openstack/nova-cell1-novncproxy-0" Oct 14 08:59:44 crc kubenswrapper[4870]: I1014 08:59:44.142082 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/36c98e9a-274d-466a-8e68-b2804be28695-config\") pod \"dnsmasq-dns-7c8cbf976c-khp98\" (UID: \"36c98e9a-274d-466a-8e68-b2804be28695\") " pod="openstack/dnsmasq-dns-7c8cbf976c-khp98" Oct 14 08:59:44 crc kubenswrapper[4870]: I1014 08:59:44.152750 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k8lx7\" (UniqueName: \"kubernetes.io/projected/36c98e9a-274d-466a-8e68-b2804be28695-kube-api-access-k8lx7\") pod \"dnsmasq-dns-7c8cbf976c-khp98\" (UID: \"36c98e9a-274d-466a-8e68-b2804be28695\") " pod="openstack/dnsmasq-dns-7c8cbf976c-khp98" Oct 14 08:59:44 crc kubenswrapper[4870]: I1014 08:59:44.157914 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4qxfd\" (UniqueName: \"kubernetes.io/projected/1a29f723-aa05-46bd-a2f9-a4422fe7646f-kube-api-access-4qxfd\") pod \"nova-cell1-novncproxy-0\" (UID: \"1a29f723-aa05-46bd-a2f9-a4422fe7646f\") " pod="openstack/nova-cell1-novncproxy-0" Oct 14 08:59:44 crc kubenswrapper[4870]: I1014 08:59:44.240012 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 14 08:59:44 crc kubenswrapper[4870]: I1014 08:59:44.261166 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 14 08:59:44 crc kubenswrapper[4870]: I1014 08:59:44.303009 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 14 08:59:44 crc kubenswrapper[4870]: W1014 08:59:44.303084 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3b91331d_394b_4f04_be17_96fd0a71fc6e.slice/crio-915fcc628853744c6d938dd4d8142db6210f2629affb798f068922df1658c292 WatchSource:0}: Error finding container 915fcc628853744c6d938dd4d8142db6210f2629affb798f068922df1658c292: Status 404 returned error can't find the container with id 915fcc628853744c6d938dd4d8142db6210f2629affb798f068922df1658c292 Oct 14 08:59:44 crc kubenswrapper[4870]: I1014 08:59:44.314351 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 14 08:59:44 crc kubenswrapper[4870]: I1014 08:59:44.336169 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7c8cbf976c-khp98" Oct 14 08:59:44 crc kubenswrapper[4870]: I1014 08:59:44.394137 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-mvlrf"] Oct 14 08:59:44 crc kubenswrapper[4870]: I1014 08:59:44.526704 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 14 08:59:44 crc kubenswrapper[4870]: I1014 08:59:44.578298 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-mvlrf" event={"ID":"c0c2d021-038b-4127-bceb-bbe26c1f3070","Type":"ContainerStarted","Data":"8c1213cfaf5ebc515f799606685aa0193cbe26bb6414a92b41d6c715ccd2c06c"} Oct 14 08:59:44 crc kubenswrapper[4870]: I1014 08:59:44.581314 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"3b91331d-394b-4f04-be17-96fd0a71fc6e","Type":"ContainerStarted","Data":"915fcc628853744c6d938dd4d8142db6210f2629affb798f068922df1658c292"} Oct 14 08:59:44 crc kubenswrapper[4870]: I1014 08:59:44.582563 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"057b319f-9b0f-4c6d-ae9c-3a18ded7d377","Type":"ContainerStarted","Data":"01fc9cb4cee14e4574ca529caa4585c265a7f7ba8eba957fb2e8b143961e3294"} Oct 14 08:59:44 crc kubenswrapper[4870]: I1014 08:59:44.645424 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-pl2ff"] Oct 14 08:59:44 crc kubenswrapper[4870]: I1014 08:59:44.646829 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-pl2ff" Oct 14 08:59:44 crc kubenswrapper[4870]: I1014 08:59:44.648629 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Oct 14 08:59:44 crc kubenswrapper[4870]: I1014 08:59:44.649271 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Oct 14 08:59:44 crc kubenswrapper[4870]: I1014 08:59:44.656657 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8w9mt\" (UniqueName: \"kubernetes.io/projected/cbd7ca07-b13e-421e-8e09-1b96dc55132a-kube-api-access-8w9mt\") pod \"nova-cell1-conductor-db-sync-pl2ff\" (UID: \"cbd7ca07-b13e-421e-8e09-1b96dc55132a\") " pod="openstack/nova-cell1-conductor-db-sync-pl2ff" Oct 14 08:59:44 crc kubenswrapper[4870]: I1014 08:59:44.656749 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cbd7ca07-b13e-421e-8e09-1b96dc55132a-config-data\") pod \"nova-cell1-conductor-db-sync-pl2ff\" (UID: \"cbd7ca07-b13e-421e-8e09-1b96dc55132a\") " pod="openstack/nova-cell1-conductor-db-sync-pl2ff" Oct 14 08:59:44 crc kubenswrapper[4870]: I1014 08:59:44.656805 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cbd7ca07-b13e-421e-8e09-1b96dc55132a-scripts\") pod \"nova-cell1-conductor-db-sync-pl2ff\" (UID: \"cbd7ca07-b13e-421e-8e09-1b96dc55132a\") " pod="openstack/nova-cell1-conductor-db-sync-pl2ff" Oct 14 08:59:44 crc kubenswrapper[4870]: I1014 08:59:44.656825 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cbd7ca07-b13e-421e-8e09-1b96dc55132a-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-pl2ff\" (UID: \"cbd7ca07-b13e-421e-8e09-1b96dc55132a\") " pod="openstack/nova-cell1-conductor-db-sync-pl2ff" Oct 14 08:59:44 crc kubenswrapper[4870]: I1014 08:59:44.656964 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-pl2ff"] Oct 14 08:59:44 crc kubenswrapper[4870]: I1014 08:59:44.760936 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8w9mt\" (UniqueName: \"kubernetes.io/projected/cbd7ca07-b13e-421e-8e09-1b96dc55132a-kube-api-access-8w9mt\") pod \"nova-cell1-conductor-db-sync-pl2ff\" (UID: \"cbd7ca07-b13e-421e-8e09-1b96dc55132a\") " pod="openstack/nova-cell1-conductor-db-sync-pl2ff" Oct 14 08:59:44 crc kubenswrapper[4870]: I1014 08:59:44.761506 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cbd7ca07-b13e-421e-8e09-1b96dc55132a-config-data\") pod \"nova-cell1-conductor-db-sync-pl2ff\" (UID: \"cbd7ca07-b13e-421e-8e09-1b96dc55132a\") " pod="openstack/nova-cell1-conductor-db-sync-pl2ff" Oct 14 08:59:44 crc kubenswrapper[4870]: I1014 08:59:44.761553 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cbd7ca07-b13e-421e-8e09-1b96dc55132a-scripts\") pod \"nova-cell1-conductor-db-sync-pl2ff\" (UID: \"cbd7ca07-b13e-421e-8e09-1b96dc55132a\") " pod="openstack/nova-cell1-conductor-db-sync-pl2ff" Oct 14 08:59:44 crc kubenswrapper[4870]: I1014 08:59:44.761586 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cbd7ca07-b13e-421e-8e09-1b96dc55132a-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-pl2ff\" (UID: \"cbd7ca07-b13e-421e-8e09-1b96dc55132a\") " pod="openstack/nova-cell1-conductor-db-sync-pl2ff" Oct 14 08:59:44 crc kubenswrapper[4870]: I1014 08:59:44.771832 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cbd7ca07-b13e-421e-8e09-1b96dc55132a-scripts\") pod \"nova-cell1-conductor-db-sync-pl2ff\" (UID: \"cbd7ca07-b13e-421e-8e09-1b96dc55132a\") " pod="openstack/nova-cell1-conductor-db-sync-pl2ff" Oct 14 08:59:44 crc kubenswrapper[4870]: I1014 08:59:44.771920 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cbd7ca07-b13e-421e-8e09-1b96dc55132a-config-data\") pod \"nova-cell1-conductor-db-sync-pl2ff\" (UID: \"cbd7ca07-b13e-421e-8e09-1b96dc55132a\") " pod="openstack/nova-cell1-conductor-db-sync-pl2ff" Oct 14 08:59:44 crc kubenswrapper[4870]: I1014 08:59:44.781375 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8w9mt\" (UniqueName: \"kubernetes.io/projected/cbd7ca07-b13e-421e-8e09-1b96dc55132a-kube-api-access-8w9mt\") pod \"nova-cell1-conductor-db-sync-pl2ff\" (UID: \"cbd7ca07-b13e-421e-8e09-1b96dc55132a\") " pod="openstack/nova-cell1-conductor-db-sync-pl2ff" Oct 14 08:59:44 crc kubenswrapper[4870]: I1014 08:59:44.785532 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cbd7ca07-b13e-421e-8e09-1b96dc55132a-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-pl2ff\" (UID: \"cbd7ca07-b13e-421e-8e09-1b96dc55132a\") " pod="openstack/nova-cell1-conductor-db-sync-pl2ff" Oct 14 08:59:44 crc kubenswrapper[4870]: I1014 08:59:44.843428 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 14 08:59:44 crc kubenswrapper[4870]: W1014 08:59:44.855557 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod93eea128_7b79_4505_b492_9c34adf49cdf.slice/crio-6ff34cc4cb8acdeb81adc100abf2b12e95129c9ffddd32c743f4aba60ee786d7 WatchSource:0}: Error finding container 6ff34cc4cb8acdeb81adc100abf2b12e95129c9ffddd32c743f4aba60ee786d7: Status 404 returned error can't find the container with id 6ff34cc4cb8acdeb81adc100abf2b12e95129c9ffddd32c743f4aba60ee786d7 Oct 14 08:59:44 crc kubenswrapper[4870]: I1014 08:59:44.910589 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 14 08:59:44 crc kubenswrapper[4870]: W1014 08:59:44.919154 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1a29f723_aa05_46bd_a2f9_a4422fe7646f.slice/crio-47ceb2fb298fcb88f9ecd4fa43fde70f42312cc43de78369f1a6487d47c1605d WatchSource:0}: Error finding container 47ceb2fb298fcb88f9ecd4fa43fde70f42312cc43de78369f1a6487d47c1605d: Status 404 returned error can't find the container with id 47ceb2fb298fcb88f9ecd4fa43fde70f42312cc43de78369f1a6487d47c1605d Oct 14 08:59:44 crc kubenswrapper[4870]: I1014 08:59:44.973920 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7c8cbf976c-khp98"] Oct 14 08:59:45 crc kubenswrapper[4870]: I1014 08:59:45.047845 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="34086778-4805-47e2-9d8d-7b9a327dcd41" path="/var/lib/kubelet/pods/34086778-4805-47e2-9d8d-7b9a327dcd41/volumes" Oct 14 08:59:45 crc kubenswrapper[4870]: I1014 08:59:45.047975 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-pl2ff" Oct 14 08:59:45 crc kubenswrapper[4870]: I1014 08:59:45.551856 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-pl2ff"] Oct 14 08:59:45 crc kubenswrapper[4870]: I1014 08:59:45.594457 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-mvlrf" event={"ID":"c0c2d021-038b-4127-bceb-bbe26c1f3070","Type":"ContainerStarted","Data":"65c074e8a549fa46c3006560067a6b54b95899f9654c6d1d2b04c1ed63aabd3f"} Oct 14 08:59:45 crc kubenswrapper[4870]: I1014 08:59:45.597257 4870 generic.go:334] "Generic (PLEG): container finished" podID="36c98e9a-274d-466a-8e68-b2804be28695" containerID="99c14090d95b5b3aa6788c08b7baf81b4f6c73cd221330803a8d489320dd6cf9" exitCode=0 Oct 14 08:59:45 crc kubenswrapper[4870]: I1014 08:59:45.597339 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7c8cbf976c-khp98" event={"ID":"36c98e9a-274d-466a-8e68-b2804be28695","Type":"ContainerDied","Data":"99c14090d95b5b3aa6788c08b7baf81b4f6c73cd221330803a8d489320dd6cf9"} Oct 14 08:59:45 crc kubenswrapper[4870]: I1014 08:59:45.597381 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7c8cbf976c-khp98" event={"ID":"36c98e9a-274d-466a-8e68-b2804be28695","Type":"ContainerStarted","Data":"4be6160942e68ba9cfc1c81fa07a993f4ceacf23291f9760a26e029e847ff36d"} Oct 14 08:59:45 crc kubenswrapper[4870]: I1014 08:59:45.598803 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"1a29f723-aa05-46bd-a2f9-a4422fe7646f","Type":"ContainerStarted","Data":"47ceb2fb298fcb88f9ecd4fa43fde70f42312cc43de78369f1a6487d47c1605d"} Oct 14 08:59:45 crc kubenswrapper[4870]: I1014 08:59:45.599706 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"93eea128-7b79-4505-b492-9c34adf49cdf","Type":"ContainerStarted","Data":"6ff34cc4cb8acdeb81adc100abf2b12e95129c9ffddd32c743f4aba60ee786d7"} Oct 14 08:59:45 crc kubenswrapper[4870]: I1014 08:59:45.612186 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-mvlrf" podStartSLOduration=2.612162957 podStartE2EDuration="2.612162957s" podCreationTimestamp="2025-10-14 08:59:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 08:59:45.60862433 +0000 UTC m=+7121.305984711" watchObservedRunningTime="2025-10-14 08:59:45.612162957 +0000 UTC m=+7121.309523328" Oct 14 08:59:46 crc kubenswrapper[4870]: W1014 08:59:46.038822 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcbd7ca07_b13e_421e_8e09_1b96dc55132a.slice/crio-a9c4d07e5b41ddd39391a7a1ed6e0efd884042f4caf44e6652276f31ba90a02e WatchSource:0}: Error finding container a9c4d07e5b41ddd39391a7a1ed6e0efd884042f4caf44e6652276f31ba90a02e: Status 404 returned error can't find the container with id a9c4d07e5b41ddd39391a7a1ed6e0efd884042f4caf44e6652276f31ba90a02e Oct 14 08:59:46 crc kubenswrapper[4870]: I1014 08:59:46.611421 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-pl2ff" event={"ID":"cbd7ca07-b13e-421e-8e09-1b96dc55132a","Type":"ContainerStarted","Data":"a9c4d07e5b41ddd39391a7a1ed6e0efd884042f4caf44e6652276f31ba90a02e"} Oct 14 08:59:47 crc kubenswrapper[4870]: I1014 08:59:47.624808 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-pl2ff" event={"ID":"cbd7ca07-b13e-421e-8e09-1b96dc55132a","Type":"ContainerStarted","Data":"fd26f0ddc6114ecb8fa4383933073a6afec3038dbefa14de68ea6c388abc85cc"} Oct 14 08:59:47 crc kubenswrapper[4870]: I1014 08:59:47.631057 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"057b319f-9b0f-4c6d-ae9c-3a18ded7d377","Type":"ContainerStarted","Data":"2a827006db00f6e2578579289b4615ddfbc20e427874936003dcc922a29bcdce"} Oct 14 08:59:47 crc kubenswrapper[4870]: I1014 08:59:47.631100 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"057b319f-9b0f-4c6d-ae9c-3a18ded7d377","Type":"ContainerStarted","Data":"077100ca854f931a799d88f903ddd3fd21ed6619196f49a986e4b31ece4712e5"} Oct 14 08:59:47 crc kubenswrapper[4870]: I1014 08:59:47.633346 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"1a29f723-aa05-46bd-a2f9-a4422fe7646f","Type":"ContainerStarted","Data":"6344e90ea632210985754aaf112afea23aaeab912ffe8717f5dd2f5ee1b46948"} Oct 14 08:59:47 crc kubenswrapper[4870]: I1014 08:59:47.635208 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"93eea128-7b79-4505-b492-9c34adf49cdf","Type":"ContainerStarted","Data":"90ad58e4ecdb8f55072333851a3ce5d4c55c7e8e9c11bb2d18e8c77dc85b0d56"} Oct 14 08:59:47 crc kubenswrapper[4870]: I1014 08:59:47.637343 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"3b91331d-394b-4f04-be17-96fd0a71fc6e","Type":"ContainerStarted","Data":"e30e539c18f4802d52a6d08cc9d481a425430a4eb4e49cf61c90f147eb29330c"} Oct 14 08:59:47 crc kubenswrapper[4870]: I1014 08:59:47.637388 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"3b91331d-394b-4f04-be17-96fd0a71fc6e","Type":"ContainerStarted","Data":"3e4c6635911e376780a55963977f2060ede48f178692dd405d3881bc6d535969"} Oct 14 08:59:47 crc kubenswrapper[4870]: I1014 08:59:47.643329 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-pl2ff" podStartSLOduration=3.64331651 podStartE2EDuration="3.64331651s" podCreationTimestamp="2025-10-14 08:59:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 08:59:47.639156757 +0000 UTC m=+7123.336517138" watchObservedRunningTime="2025-10-14 08:59:47.64331651 +0000 UTC m=+7123.340676891" Oct 14 08:59:47 crc kubenswrapper[4870]: I1014 08:59:47.644278 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7c8cbf976c-khp98" event={"ID":"36c98e9a-274d-466a-8e68-b2804be28695","Type":"ContainerStarted","Data":"b92b8e6d69e662f349f55344ecfc601712b19204e2f344064d23ad7c0b0fecce"} Oct 14 08:59:47 crc kubenswrapper[4870]: I1014 08:59:47.645328 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7c8cbf976c-khp98" Oct 14 08:59:47 crc kubenswrapper[4870]: I1014 08:59:47.661350 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.803627594 podStartE2EDuration="4.661316374s" podCreationTimestamp="2025-10-14 08:59:43 +0000 UTC" firstStartedPulling="2025-10-14 08:59:44.92338831 +0000 UTC m=+7120.620748681" lastFinishedPulling="2025-10-14 08:59:46.78107709 +0000 UTC m=+7122.478437461" observedRunningTime="2025-10-14 08:59:47.65427038 +0000 UTC m=+7123.351630751" watchObservedRunningTime="2025-10-14 08:59:47.661316374 +0000 UTC m=+7123.358676755" Oct 14 08:59:47 crc kubenswrapper[4870]: I1014 08:59:47.678826 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.727546115 podStartE2EDuration="4.678808916s" podCreationTimestamp="2025-10-14 08:59:43 +0000 UTC" firstStartedPulling="2025-10-14 08:59:44.861097322 +0000 UTC m=+7120.558457693" lastFinishedPulling="2025-10-14 08:59:46.812360123 +0000 UTC m=+7122.509720494" observedRunningTime="2025-10-14 08:59:47.671766092 +0000 UTC m=+7123.369126473" watchObservedRunningTime="2025-10-14 08:59:47.678808916 +0000 UTC m=+7123.376169287" Oct 14 08:59:47 crc kubenswrapper[4870]: I1014 08:59:47.693803 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.4603958280000002 podStartE2EDuration="4.693771806s" podCreationTimestamp="2025-10-14 08:59:43 +0000 UTC" firstStartedPulling="2025-10-14 08:59:44.548649986 +0000 UTC m=+7120.246010357" lastFinishedPulling="2025-10-14 08:59:46.782025964 +0000 UTC m=+7122.479386335" observedRunningTime="2025-10-14 08:59:47.690994157 +0000 UTC m=+7123.388354528" watchObservedRunningTime="2025-10-14 08:59:47.693771806 +0000 UTC m=+7123.391132187" Oct 14 08:59:47 crc kubenswrapper[4870]: I1014 08:59:47.749579 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7c8cbf976c-khp98" podStartSLOduration=4.749555063 podStartE2EDuration="4.749555063s" podCreationTimestamp="2025-10-14 08:59:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 08:59:47.744839867 +0000 UTC m=+7123.442200238" watchObservedRunningTime="2025-10-14 08:59:47.749555063 +0000 UTC m=+7123.446915454" Oct 14 08:59:47 crc kubenswrapper[4870]: I1014 08:59:47.750732 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.275310549 podStartE2EDuration="4.750723522s" podCreationTimestamp="2025-10-14 08:59:43 +0000 UTC" firstStartedPulling="2025-10-14 08:59:44.30537875 +0000 UTC m=+7120.002739121" lastFinishedPulling="2025-10-14 08:59:46.780791703 +0000 UTC m=+7122.478152094" observedRunningTime="2025-10-14 08:59:47.72312071 +0000 UTC m=+7123.420481091" watchObservedRunningTime="2025-10-14 08:59:47.750723522 +0000 UTC m=+7123.448083903" Oct 14 08:59:49 crc kubenswrapper[4870]: I1014 08:59:49.240134 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 14 08:59:49 crc kubenswrapper[4870]: I1014 08:59:49.241661 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 14 08:59:49 crc kubenswrapper[4870]: I1014 08:59:49.262025 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 14 08:59:49 crc kubenswrapper[4870]: I1014 08:59:49.314602 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Oct 14 08:59:49 crc kubenswrapper[4870]: E1014 08:59:49.323170 4870 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc0c2d021_038b_4127_bceb_bbe26c1f3070.slice/crio-conmon-65c074e8a549fa46c3006560067a6b54b95899f9654c6d1d2b04c1ed63aabd3f.scope\": RecentStats: unable to find data in memory cache]" Oct 14 08:59:49 crc kubenswrapper[4870]: I1014 08:59:49.674901 4870 generic.go:334] "Generic (PLEG): container finished" podID="c0c2d021-038b-4127-bceb-bbe26c1f3070" containerID="65c074e8a549fa46c3006560067a6b54b95899f9654c6d1d2b04c1ed63aabd3f" exitCode=0 Oct 14 08:59:49 crc kubenswrapper[4870]: I1014 08:59:49.675030 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-mvlrf" event={"ID":"c0c2d021-038b-4127-bceb-bbe26c1f3070","Type":"ContainerDied","Data":"65c074e8a549fa46c3006560067a6b54b95899f9654c6d1d2b04c1ed63aabd3f"} Oct 14 08:59:50 crc kubenswrapper[4870]: I1014 08:59:50.688689 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-pl2ff" event={"ID":"cbd7ca07-b13e-421e-8e09-1b96dc55132a","Type":"ContainerDied","Data":"fd26f0ddc6114ecb8fa4383933073a6afec3038dbefa14de68ea6c388abc85cc"} Oct 14 08:59:50 crc kubenswrapper[4870]: I1014 08:59:50.689579 4870 generic.go:334] "Generic (PLEG): container finished" podID="cbd7ca07-b13e-421e-8e09-1b96dc55132a" containerID="fd26f0ddc6114ecb8fa4383933073a6afec3038dbefa14de68ea6c388abc85cc" exitCode=0 Oct 14 08:59:51 crc kubenswrapper[4870]: I1014 08:59:51.094162 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-mvlrf" Oct 14 08:59:51 crc kubenswrapper[4870]: I1014 08:59:51.195195 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c0c2d021-038b-4127-bceb-bbe26c1f3070-scripts\") pod \"c0c2d021-038b-4127-bceb-bbe26c1f3070\" (UID: \"c0c2d021-038b-4127-bceb-bbe26c1f3070\") " Oct 14 08:59:51 crc kubenswrapper[4870]: I1014 08:59:51.195351 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c0c2d021-038b-4127-bceb-bbe26c1f3070-combined-ca-bundle\") pod \"c0c2d021-038b-4127-bceb-bbe26c1f3070\" (UID: \"c0c2d021-038b-4127-bceb-bbe26c1f3070\") " Oct 14 08:59:51 crc kubenswrapper[4870]: I1014 08:59:51.195415 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ds85t\" (UniqueName: \"kubernetes.io/projected/c0c2d021-038b-4127-bceb-bbe26c1f3070-kube-api-access-ds85t\") pod \"c0c2d021-038b-4127-bceb-bbe26c1f3070\" (UID: \"c0c2d021-038b-4127-bceb-bbe26c1f3070\") " Oct 14 08:59:51 crc kubenswrapper[4870]: I1014 08:59:51.195540 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c0c2d021-038b-4127-bceb-bbe26c1f3070-config-data\") pod \"c0c2d021-038b-4127-bceb-bbe26c1f3070\" (UID: \"c0c2d021-038b-4127-bceb-bbe26c1f3070\") " Oct 14 08:59:51 crc kubenswrapper[4870]: I1014 08:59:51.201016 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c0c2d021-038b-4127-bceb-bbe26c1f3070-scripts" (OuterVolumeSpecName: "scripts") pod "c0c2d021-038b-4127-bceb-bbe26c1f3070" (UID: "c0c2d021-038b-4127-bceb-bbe26c1f3070"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:59:51 crc kubenswrapper[4870]: I1014 08:59:51.201737 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c0c2d021-038b-4127-bceb-bbe26c1f3070-kube-api-access-ds85t" (OuterVolumeSpecName: "kube-api-access-ds85t") pod "c0c2d021-038b-4127-bceb-bbe26c1f3070" (UID: "c0c2d021-038b-4127-bceb-bbe26c1f3070"). InnerVolumeSpecName "kube-api-access-ds85t". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:59:51 crc kubenswrapper[4870]: I1014 08:59:51.222787 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c0c2d021-038b-4127-bceb-bbe26c1f3070-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c0c2d021-038b-4127-bceb-bbe26c1f3070" (UID: "c0c2d021-038b-4127-bceb-bbe26c1f3070"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:59:51 crc kubenswrapper[4870]: I1014 08:59:51.248174 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c0c2d021-038b-4127-bceb-bbe26c1f3070-config-data" (OuterVolumeSpecName: "config-data") pod "c0c2d021-038b-4127-bceb-bbe26c1f3070" (UID: "c0c2d021-038b-4127-bceb-bbe26c1f3070"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:59:51 crc kubenswrapper[4870]: I1014 08:59:51.297762 4870 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c0c2d021-038b-4127-bceb-bbe26c1f3070-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 08:59:51 crc kubenswrapper[4870]: I1014 08:59:51.297805 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ds85t\" (UniqueName: \"kubernetes.io/projected/c0c2d021-038b-4127-bceb-bbe26c1f3070-kube-api-access-ds85t\") on node \"crc\" DevicePath \"\"" Oct 14 08:59:51 crc kubenswrapper[4870]: I1014 08:59:51.297815 4870 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c0c2d021-038b-4127-bceb-bbe26c1f3070-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 08:59:51 crc kubenswrapper[4870]: I1014 08:59:51.297823 4870 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c0c2d021-038b-4127-bceb-bbe26c1f3070-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 08:59:51 crc kubenswrapper[4870]: I1014 08:59:51.703301 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-mvlrf" Oct 14 08:59:51 crc kubenswrapper[4870]: I1014 08:59:51.703314 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-mvlrf" event={"ID":"c0c2d021-038b-4127-bceb-bbe26c1f3070","Type":"ContainerDied","Data":"8c1213cfaf5ebc515f799606685aa0193cbe26bb6414a92b41d6c715ccd2c06c"} Oct 14 08:59:51 crc kubenswrapper[4870]: I1014 08:59:51.703509 4870 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8c1213cfaf5ebc515f799606685aa0193cbe26bb6414a92b41d6c715ccd2c06c" Oct 14 08:59:51 crc kubenswrapper[4870]: I1014 08:59:51.917351 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 14 08:59:51 crc kubenswrapper[4870]: I1014 08:59:51.917601 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="3b91331d-394b-4f04-be17-96fd0a71fc6e" containerName="nova-api-log" containerID="cri-o://3e4c6635911e376780a55963977f2060ede48f178692dd405d3881bc6d535969" gracePeriod=30 Oct 14 08:59:51 crc kubenswrapper[4870]: I1014 08:59:51.918062 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="3b91331d-394b-4f04-be17-96fd0a71fc6e" containerName="nova-api-api" containerID="cri-o://e30e539c18f4802d52a6d08cc9d481a425430a4eb4e49cf61c90f147eb29330c" gracePeriod=30 Oct 14 08:59:51 crc kubenswrapper[4870]: I1014 08:59:51.950653 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 14 08:59:51 crc kubenswrapper[4870]: I1014 08:59:51.950927 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="93eea128-7b79-4505-b492-9c34adf49cdf" containerName="nova-scheduler-scheduler" containerID="cri-o://90ad58e4ecdb8f55072333851a3ce5d4c55c7e8e9c11bb2d18e8c77dc85b0d56" gracePeriod=30 Oct 14 08:59:51 crc kubenswrapper[4870]: I1014 08:59:51.965499 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 14 08:59:51 crc kubenswrapper[4870]: I1014 08:59:51.965710 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="057b319f-9b0f-4c6d-ae9c-3a18ded7d377" containerName="nova-metadata-log" containerID="cri-o://077100ca854f931a799d88f903ddd3fd21ed6619196f49a986e4b31ece4712e5" gracePeriod=30 Oct 14 08:59:51 crc kubenswrapper[4870]: I1014 08:59:51.965874 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="057b319f-9b0f-4c6d-ae9c-3a18ded7d377" containerName="nova-metadata-metadata" containerID="cri-o://2a827006db00f6e2578579289b4615ddfbc20e427874936003dcc922a29bcdce" gracePeriod=30 Oct 14 08:59:52 crc kubenswrapper[4870]: I1014 08:59:52.308146 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-pl2ff" Oct 14 08:59:52 crc kubenswrapper[4870]: I1014 08:59:52.423875 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cbd7ca07-b13e-421e-8e09-1b96dc55132a-scripts\") pod \"cbd7ca07-b13e-421e-8e09-1b96dc55132a\" (UID: \"cbd7ca07-b13e-421e-8e09-1b96dc55132a\") " Oct 14 08:59:52 crc kubenswrapper[4870]: I1014 08:59:52.423933 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cbd7ca07-b13e-421e-8e09-1b96dc55132a-combined-ca-bundle\") pod \"cbd7ca07-b13e-421e-8e09-1b96dc55132a\" (UID: \"cbd7ca07-b13e-421e-8e09-1b96dc55132a\") " Oct 14 08:59:52 crc kubenswrapper[4870]: I1014 08:59:52.423997 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cbd7ca07-b13e-421e-8e09-1b96dc55132a-config-data\") pod \"cbd7ca07-b13e-421e-8e09-1b96dc55132a\" (UID: \"cbd7ca07-b13e-421e-8e09-1b96dc55132a\") " Oct 14 08:59:52 crc kubenswrapper[4870]: I1014 08:59:52.424032 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8w9mt\" (UniqueName: \"kubernetes.io/projected/cbd7ca07-b13e-421e-8e09-1b96dc55132a-kube-api-access-8w9mt\") pod \"cbd7ca07-b13e-421e-8e09-1b96dc55132a\" (UID: \"cbd7ca07-b13e-421e-8e09-1b96dc55132a\") " Oct 14 08:59:52 crc kubenswrapper[4870]: I1014 08:59:52.429927 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cbd7ca07-b13e-421e-8e09-1b96dc55132a-kube-api-access-8w9mt" (OuterVolumeSpecName: "kube-api-access-8w9mt") pod "cbd7ca07-b13e-421e-8e09-1b96dc55132a" (UID: "cbd7ca07-b13e-421e-8e09-1b96dc55132a"). InnerVolumeSpecName "kube-api-access-8w9mt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:59:52 crc kubenswrapper[4870]: I1014 08:59:52.430646 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cbd7ca07-b13e-421e-8e09-1b96dc55132a-scripts" (OuterVolumeSpecName: "scripts") pod "cbd7ca07-b13e-421e-8e09-1b96dc55132a" (UID: "cbd7ca07-b13e-421e-8e09-1b96dc55132a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:59:52 crc kubenswrapper[4870]: I1014 08:59:52.462087 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cbd7ca07-b13e-421e-8e09-1b96dc55132a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cbd7ca07-b13e-421e-8e09-1b96dc55132a" (UID: "cbd7ca07-b13e-421e-8e09-1b96dc55132a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:59:52 crc kubenswrapper[4870]: I1014 08:59:52.472854 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cbd7ca07-b13e-421e-8e09-1b96dc55132a-config-data" (OuterVolumeSpecName: "config-data") pod "cbd7ca07-b13e-421e-8e09-1b96dc55132a" (UID: "cbd7ca07-b13e-421e-8e09-1b96dc55132a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:59:52 crc kubenswrapper[4870]: I1014 08:59:52.519404 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 14 08:59:52 crc kubenswrapper[4870]: I1014 08:59:52.526554 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 14 08:59:52 crc kubenswrapper[4870]: I1014 08:59:52.527459 4870 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cbd7ca07-b13e-421e-8e09-1b96dc55132a-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 08:59:52 crc kubenswrapper[4870]: I1014 08:59:52.527487 4870 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cbd7ca07-b13e-421e-8e09-1b96dc55132a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 08:59:52 crc kubenswrapper[4870]: I1014 08:59:52.527502 4870 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cbd7ca07-b13e-421e-8e09-1b96dc55132a-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 08:59:52 crc kubenswrapper[4870]: I1014 08:59:52.527547 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8w9mt\" (UniqueName: \"kubernetes.io/projected/cbd7ca07-b13e-421e-8e09-1b96dc55132a-kube-api-access-8w9mt\") on node \"crc\" DevicePath \"\"" Oct 14 08:59:52 crc kubenswrapper[4870]: I1014 08:59:52.628392 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3b91331d-394b-4f04-be17-96fd0a71fc6e-logs\") pod \"3b91331d-394b-4f04-be17-96fd0a71fc6e\" (UID: \"3b91331d-394b-4f04-be17-96fd0a71fc6e\") " Oct 14 08:59:52 crc kubenswrapper[4870]: I1014 08:59:52.628479 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/057b319f-9b0f-4c6d-ae9c-3a18ded7d377-logs\") pod \"057b319f-9b0f-4c6d-ae9c-3a18ded7d377\" (UID: \"057b319f-9b0f-4c6d-ae9c-3a18ded7d377\") " Oct 14 08:59:52 crc kubenswrapper[4870]: I1014 08:59:52.628527 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tvtft\" (UniqueName: \"kubernetes.io/projected/3b91331d-394b-4f04-be17-96fd0a71fc6e-kube-api-access-tvtft\") pod \"3b91331d-394b-4f04-be17-96fd0a71fc6e\" (UID: \"3b91331d-394b-4f04-be17-96fd0a71fc6e\") " Oct 14 08:59:52 crc kubenswrapper[4870]: I1014 08:59:52.628565 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p7kxv\" (UniqueName: \"kubernetes.io/projected/057b319f-9b0f-4c6d-ae9c-3a18ded7d377-kube-api-access-p7kxv\") pod \"057b319f-9b0f-4c6d-ae9c-3a18ded7d377\" (UID: \"057b319f-9b0f-4c6d-ae9c-3a18ded7d377\") " Oct 14 08:59:52 crc kubenswrapper[4870]: I1014 08:59:52.628628 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3b91331d-394b-4f04-be17-96fd0a71fc6e-config-data\") pod \"3b91331d-394b-4f04-be17-96fd0a71fc6e\" (UID: \"3b91331d-394b-4f04-be17-96fd0a71fc6e\") " Oct 14 08:59:52 crc kubenswrapper[4870]: I1014 08:59:52.628672 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/057b319f-9b0f-4c6d-ae9c-3a18ded7d377-config-data\") pod \"057b319f-9b0f-4c6d-ae9c-3a18ded7d377\" (UID: \"057b319f-9b0f-4c6d-ae9c-3a18ded7d377\") " Oct 14 08:59:52 crc kubenswrapper[4870]: I1014 08:59:52.628723 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b91331d-394b-4f04-be17-96fd0a71fc6e-combined-ca-bundle\") pod \"3b91331d-394b-4f04-be17-96fd0a71fc6e\" (UID: \"3b91331d-394b-4f04-be17-96fd0a71fc6e\") " Oct 14 08:59:52 crc kubenswrapper[4870]: I1014 08:59:52.628755 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/057b319f-9b0f-4c6d-ae9c-3a18ded7d377-combined-ca-bundle\") pod \"057b319f-9b0f-4c6d-ae9c-3a18ded7d377\" (UID: \"057b319f-9b0f-4c6d-ae9c-3a18ded7d377\") " Oct 14 08:59:52 crc kubenswrapper[4870]: I1014 08:59:52.628893 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/057b319f-9b0f-4c6d-ae9c-3a18ded7d377-logs" (OuterVolumeSpecName: "logs") pod "057b319f-9b0f-4c6d-ae9c-3a18ded7d377" (UID: "057b319f-9b0f-4c6d-ae9c-3a18ded7d377"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:59:52 crc kubenswrapper[4870]: I1014 08:59:52.628913 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3b91331d-394b-4f04-be17-96fd0a71fc6e-logs" (OuterVolumeSpecName: "logs") pod "3b91331d-394b-4f04-be17-96fd0a71fc6e" (UID: "3b91331d-394b-4f04-be17-96fd0a71fc6e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 08:59:52 crc kubenswrapper[4870]: I1014 08:59:52.629798 4870 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3b91331d-394b-4f04-be17-96fd0a71fc6e-logs\") on node \"crc\" DevicePath \"\"" Oct 14 08:59:52 crc kubenswrapper[4870]: I1014 08:59:52.629823 4870 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/057b319f-9b0f-4c6d-ae9c-3a18ded7d377-logs\") on node \"crc\" DevicePath \"\"" Oct 14 08:59:52 crc kubenswrapper[4870]: I1014 08:59:52.632765 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/057b319f-9b0f-4c6d-ae9c-3a18ded7d377-kube-api-access-p7kxv" (OuterVolumeSpecName: "kube-api-access-p7kxv") pod "057b319f-9b0f-4c6d-ae9c-3a18ded7d377" (UID: "057b319f-9b0f-4c6d-ae9c-3a18ded7d377"). InnerVolumeSpecName "kube-api-access-p7kxv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:59:52 crc kubenswrapper[4870]: I1014 08:59:52.633163 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3b91331d-394b-4f04-be17-96fd0a71fc6e-kube-api-access-tvtft" (OuterVolumeSpecName: "kube-api-access-tvtft") pod "3b91331d-394b-4f04-be17-96fd0a71fc6e" (UID: "3b91331d-394b-4f04-be17-96fd0a71fc6e"). InnerVolumeSpecName "kube-api-access-tvtft". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:59:52 crc kubenswrapper[4870]: I1014 08:59:52.655982 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3b91331d-394b-4f04-be17-96fd0a71fc6e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3b91331d-394b-4f04-be17-96fd0a71fc6e" (UID: "3b91331d-394b-4f04-be17-96fd0a71fc6e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:59:52 crc kubenswrapper[4870]: I1014 08:59:52.656581 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/057b319f-9b0f-4c6d-ae9c-3a18ded7d377-config-data" (OuterVolumeSpecName: "config-data") pod "057b319f-9b0f-4c6d-ae9c-3a18ded7d377" (UID: "057b319f-9b0f-4c6d-ae9c-3a18ded7d377"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:59:52 crc kubenswrapper[4870]: I1014 08:59:52.660636 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/057b319f-9b0f-4c6d-ae9c-3a18ded7d377-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "057b319f-9b0f-4c6d-ae9c-3a18ded7d377" (UID: "057b319f-9b0f-4c6d-ae9c-3a18ded7d377"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:59:52 crc kubenswrapper[4870]: I1014 08:59:52.661758 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3b91331d-394b-4f04-be17-96fd0a71fc6e-config-data" (OuterVolumeSpecName: "config-data") pod "3b91331d-394b-4f04-be17-96fd0a71fc6e" (UID: "3b91331d-394b-4f04-be17-96fd0a71fc6e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:59:52 crc kubenswrapper[4870]: I1014 08:59:52.712501 4870 generic.go:334] "Generic (PLEG): container finished" podID="057b319f-9b0f-4c6d-ae9c-3a18ded7d377" containerID="2a827006db00f6e2578579289b4615ddfbc20e427874936003dcc922a29bcdce" exitCode=0 Oct 14 08:59:52 crc kubenswrapper[4870]: I1014 08:59:52.712529 4870 generic.go:334] "Generic (PLEG): container finished" podID="057b319f-9b0f-4c6d-ae9c-3a18ded7d377" containerID="077100ca854f931a799d88f903ddd3fd21ed6619196f49a986e4b31ece4712e5" exitCode=143 Oct 14 08:59:52 crc kubenswrapper[4870]: I1014 08:59:52.712629 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 14 08:59:52 crc kubenswrapper[4870]: I1014 08:59:52.713728 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"057b319f-9b0f-4c6d-ae9c-3a18ded7d377","Type":"ContainerDied","Data":"2a827006db00f6e2578579289b4615ddfbc20e427874936003dcc922a29bcdce"} Oct 14 08:59:52 crc kubenswrapper[4870]: I1014 08:59:52.713804 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"057b319f-9b0f-4c6d-ae9c-3a18ded7d377","Type":"ContainerDied","Data":"077100ca854f931a799d88f903ddd3fd21ed6619196f49a986e4b31ece4712e5"} Oct 14 08:59:52 crc kubenswrapper[4870]: I1014 08:59:52.713821 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"057b319f-9b0f-4c6d-ae9c-3a18ded7d377","Type":"ContainerDied","Data":"01fc9cb4cee14e4574ca529caa4585c265a7f7ba8eba957fb2e8b143961e3294"} Oct 14 08:59:52 crc kubenswrapper[4870]: I1014 08:59:52.713859 4870 scope.go:117] "RemoveContainer" containerID="2a827006db00f6e2578579289b4615ddfbc20e427874936003dcc922a29bcdce" Oct 14 08:59:52 crc kubenswrapper[4870]: I1014 08:59:52.716581 4870 generic.go:334] "Generic (PLEG): container finished" podID="3b91331d-394b-4f04-be17-96fd0a71fc6e" containerID="e30e539c18f4802d52a6d08cc9d481a425430a4eb4e49cf61c90f147eb29330c" exitCode=0 Oct 14 08:59:52 crc kubenswrapper[4870]: I1014 08:59:52.716604 4870 generic.go:334] "Generic (PLEG): container finished" podID="3b91331d-394b-4f04-be17-96fd0a71fc6e" containerID="3e4c6635911e376780a55963977f2060ede48f178692dd405d3881bc6d535969" exitCode=143 Oct 14 08:59:52 crc kubenswrapper[4870]: I1014 08:59:52.716664 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"3b91331d-394b-4f04-be17-96fd0a71fc6e","Type":"ContainerDied","Data":"e30e539c18f4802d52a6d08cc9d481a425430a4eb4e49cf61c90f147eb29330c"} Oct 14 08:59:52 crc kubenswrapper[4870]: I1014 08:59:52.716728 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"3b91331d-394b-4f04-be17-96fd0a71fc6e","Type":"ContainerDied","Data":"3e4c6635911e376780a55963977f2060ede48f178692dd405d3881bc6d535969"} Oct 14 08:59:52 crc kubenswrapper[4870]: I1014 08:59:52.716739 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"3b91331d-394b-4f04-be17-96fd0a71fc6e","Type":"ContainerDied","Data":"915fcc628853744c6d938dd4d8142db6210f2629affb798f068922df1658c292"} Oct 14 08:59:52 crc kubenswrapper[4870]: I1014 08:59:52.716733 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 14 08:59:52 crc kubenswrapper[4870]: I1014 08:59:52.718566 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-pl2ff" event={"ID":"cbd7ca07-b13e-421e-8e09-1b96dc55132a","Type":"ContainerDied","Data":"a9c4d07e5b41ddd39391a7a1ed6e0efd884042f4caf44e6652276f31ba90a02e"} Oct 14 08:59:52 crc kubenswrapper[4870]: I1014 08:59:52.718588 4870 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a9c4d07e5b41ddd39391a7a1ed6e0efd884042f4caf44e6652276f31ba90a02e" Oct 14 08:59:52 crc kubenswrapper[4870]: I1014 08:59:52.718661 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-pl2ff" Oct 14 08:59:52 crc kubenswrapper[4870]: I1014 08:59:52.742336 4870 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/057b319f-9b0f-4c6d-ae9c-3a18ded7d377-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 08:59:52 crc kubenswrapper[4870]: I1014 08:59:52.742367 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tvtft\" (UniqueName: \"kubernetes.io/projected/3b91331d-394b-4f04-be17-96fd0a71fc6e-kube-api-access-tvtft\") on node \"crc\" DevicePath \"\"" Oct 14 08:59:52 crc kubenswrapper[4870]: I1014 08:59:52.742381 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p7kxv\" (UniqueName: \"kubernetes.io/projected/057b319f-9b0f-4c6d-ae9c-3a18ded7d377-kube-api-access-p7kxv\") on node \"crc\" DevicePath \"\"" Oct 14 08:59:52 crc kubenswrapper[4870]: I1014 08:59:52.742391 4870 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3b91331d-394b-4f04-be17-96fd0a71fc6e-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 08:59:52 crc kubenswrapper[4870]: I1014 08:59:52.742401 4870 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/057b319f-9b0f-4c6d-ae9c-3a18ded7d377-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 08:59:52 crc kubenswrapper[4870]: I1014 08:59:52.742409 4870 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b91331d-394b-4f04-be17-96fd0a71fc6e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 08:59:52 crc kubenswrapper[4870]: I1014 08:59:52.746050 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 14 08:59:52 crc kubenswrapper[4870]: I1014 08:59:52.747288 4870 scope.go:117] "RemoveContainer" containerID="077100ca854f931a799d88f903ddd3fd21ed6619196f49a986e4b31ece4712e5" Oct 14 08:59:52 crc kubenswrapper[4870]: I1014 08:59:52.758213 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 14 08:59:52 crc kubenswrapper[4870]: I1014 08:59:52.783869 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 14 08:59:52 crc kubenswrapper[4870]: I1014 08:59:52.784002 4870 scope.go:117] "RemoveContainer" containerID="2a827006db00f6e2578579289b4615ddfbc20e427874936003dcc922a29bcdce" Oct 14 08:59:52 crc kubenswrapper[4870]: E1014 08:59:52.784243 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="057b319f-9b0f-4c6d-ae9c-3a18ded7d377" containerName="nova-metadata-log" Oct 14 08:59:52 crc kubenswrapper[4870]: I1014 08:59:52.784256 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="057b319f-9b0f-4c6d-ae9c-3a18ded7d377" containerName="nova-metadata-log" Oct 14 08:59:52 crc kubenswrapper[4870]: E1014 08:59:52.784275 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b91331d-394b-4f04-be17-96fd0a71fc6e" containerName="nova-api-api" Oct 14 08:59:52 crc kubenswrapper[4870]: I1014 08:59:52.784282 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b91331d-394b-4f04-be17-96fd0a71fc6e" containerName="nova-api-api" Oct 14 08:59:52 crc kubenswrapper[4870]: E1014 08:59:52.784293 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cbd7ca07-b13e-421e-8e09-1b96dc55132a" containerName="nova-cell1-conductor-db-sync" Oct 14 08:59:52 crc kubenswrapper[4870]: I1014 08:59:52.784301 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="cbd7ca07-b13e-421e-8e09-1b96dc55132a" containerName="nova-cell1-conductor-db-sync" Oct 14 08:59:52 crc kubenswrapper[4870]: E1014 08:59:52.784311 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b91331d-394b-4f04-be17-96fd0a71fc6e" containerName="nova-api-log" Oct 14 08:59:52 crc kubenswrapper[4870]: I1014 08:59:52.784318 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b91331d-394b-4f04-be17-96fd0a71fc6e" containerName="nova-api-log" Oct 14 08:59:52 crc kubenswrapper[4870]: E1014 08:59:52.784337 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c0c2d021-038b-4127-bceb-bbe26c1f3070" containerName="nova-manage" Oct 14 08:59:52 crc kubenswrapper[4870]: I1014 08:59:52.784343 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="c0c2d021-038b-4127-bceb-bbe26c1f3070" containerName="nova-manage" Oct 14 08:59:52 crc kubenswrapper[4870]: E1014 08:59:52.784353 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="057b319f-9b0f-4c6d-ae9c-3a18ded7d377" containerName="nova-metadata-metadata" Oct 14 08:59:52 crc kubenswrapper[4870]: I1014 08:59:52.784358 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="057b319f-9b0f-4c6d-ae9c-3a18ded7d377" containerName="nova-metadata-metadata" Oct 14 08:59:52 crc kubenswrapper[4870]: I1014 08:59:52.784554 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="057b319f-9b0f-4c6d-ae9c-3a18ded7d377" containerName="nova-metadata-metadata" Oct 14 08:59:52 crc kubenswrapper[4870]: I1014 08:59:52.784566 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="057b319f-9b0f-4c6d-ae9c-3a18ded7d377" containerName="nova-metadata-log" Oct 14 08:59:52 crc kubenswrapper[4870]: I1014 08:59:52.784579 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="cbd7ca07-b13e-421e-8e09-1b96dc55132a" containerName="nova-cell1-conductor-db-sync" Oct 14 08:59:52 crc kubenswrapper[4870]: I1014 08:59:52.784589 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="3b91331d-394b-4f04-be17-96fd0a71fc6e" containerName="nova-api-log" Oct 14 08:59:52 crc kubenswrapper[4870]: I1014 08:59:52.784600 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="3b91331d-394b-4f04-be17-96fd0a71fc6e" containerName="nova-api-api" Oct 14 08:59:52 crc kubenswrapper[4870]: I1014 08:59:52.784611 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="c0c2d021-038b-4127-bceb-bbe26c1f3070" containerName="nova-manage" Oct 14 08:59:52 crc kubenswrapper[4870]: I1014 08:59:52.787773 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 14 08:59:52 crc kubenswrapper[4870]: E1014 08:59:52.789751 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2a827006db00f6e2578579289b4615ddfbc20e427874936003dcc922a29bcdce\": container with ID starting with 2a827006db00f6e2578579289b4615ddfbc20e427874936003dcc922a29bcdce not found: ID does not exist" containerID="2a827006db00f6e2578579289b4615ddfbc20e427874936003dcc922a29bcdce" Oct 14 08:59:52 crc kubenswrapper[4870]: I1014 08:59:52.789810 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2a827006db00f6e2578579289b4615ddfbc20e427874936003dcc922a29bcdce"} err="failed to get container status \"2a827006db00f6e2578579289b4615ddfbc20e427874936003dcc922a29bcdce\": rpc error: code = NotFound desc = could not find container \"2a827006db00f6e2578579289b4615ddfbc20e427874936003dcc922a29bcdce\": container with ID starting with 2a827006db00f6e2578579289b4615ddfbc20e427874936003dcc922a29bcdce not found: ID does not exist" Oct 14 08:59:52 crc kubenswrapper[4870]: I1014 08:59:52.789844 4870 scope.go:117] "RemoveContainer" containerID="077100ca854f931a799d88f903ddd3fd21ed6619196f49a986e4b31ece4712e5" Oct 14 08:59:52 crc kubenswrapper[4870]: E1014 08:59:52.792842 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"077100ca854f931a799d88f903ddd3fd21ed6619196f49a986e4b31ece4712e5\": container with ID starting with 077100ca854f931a799d88f903ddd3fd21ed6619196f49a986e4b31ece4712e5 not found: ID does not exist" containerID="077100ca854f931a799d88f903ddd3fd21ed6619196f49a986e4b31ece4712e5" Oct 14 08:59:52 crc kubenswrapper[4870]: I1014 08:59:52.792902 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"077100ca854f931a799d88f903ddd3fd21ed6619196f49a986e4b31ece4712e5"} err="failed to get container status \"077100ca854f931a799d88f903ddd3fd21ed6619196f49a986e4b31ece4712e5\": rpc error: code = NotFound desc = could not find container \"077100ca854f931a799d88f903ddd3fd21ed6619196f49a986e4b31ece4712e5\": container with ID starting with 077100ca854f931a799d88f903ddd3fd21ed6619196f49a986e4b31ece4712e5 not found: ID does not exist" Oct 14 08:59:52 crc kubenswrapper[4870]: I1014 08:59:52.792939 4870 scope.go:117] "RemoveContainer" containerID="2a827006db00f6e2578579289b4615ddfbc20e427874936003dcc922a29bcdce" Oct 14 08:59:52 crc kubenswrapper[4870]: I1014 08:59:52.793613 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2a827006db00f6e2578579289b4615ddfbc20e427874936003dcc922a29bcdce"} err="failed to get container status \"2a827006db00f6e2578579289b4615ddfbc20e427874936003dcc922a29bcdce\": rpc error: code = NotFound desc = could not find container \"2a827006db00f6e2578579289b4615ddfbc20e427874936003dcc922a29bcdce\": container with ID starting with 2a827006db00f6e2578579289b4615ddfbc20e427874936003dcc922a29bcdce not found: ID does not exist" Oct 14 08:59:52 crc kubenswrapper[4870]: I1014 08:59:52.793647 4870 scope.go:117] "RemoveContainer" containerID="077100ca854f931a799d88f903ddd3fd21ed6619196f49a986e4b31ece4712e5" Oct 14 08:59:52 crc kubenswrapper[4870]: I1014 08:59:52.793899 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"077100ca854f931a799d88f903ddd3fd21ed6619196f49a986e4b31ece4712e5"} err="failed to get container status \"077100ca854f931a799d88f903ddd3fd21ed6619196f49a986e4b31ece4712e5\": rpc error: code = NotFound desc = could not find container \"077100ca854f931a799d88f903ddd3fd21ed6619196f49a986e4b31ece4712e5\": container with ID starting with 077100ca854f931a799d88f903ddd3fd21ed6619196f49a986e4b31ece4712e5 not found: ID does not exist" Oct 14 08:59:52 crc kubenswrapper[4870]: I1014 08:59:52.793990 4870 scope.go:117] "RemoveContainer" containerID="e30e539c18f4802d52a6d08cc9d481a425430a4eb4e49cf61c90f147eb29330c" Oct 14 08:59:52 crc kubenswrapper[4870]: I1014 08:59:52.806889 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 14 08:59:52 crc kubenswrapper[4870]: I1014 08:59:52.812678 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 14 08:59:52 crc kubenswrapper[4870]: I1014 08:59:52.822908 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 14 08:59:52 crc kubenswrapper[4870]: I1014 08:59:52.838250 4870 scope.go:117] "RemoveContainer" containerID="3e4c6635911e376780a55963977f2060ede48f178692dd405d3881bc6d535969" Oct 14 08:59:52 crc kubenswrapper[4870]: I1014 08:59:52.840465 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 14 08:59:52 crc kubenswrapper[4870]: I1014 08:59:52.849569 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 14 08:59:52 crc kubenswrapper[4870]: I1014 08:59:52.856046 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 14 08:59:52 crc kubenswrapper[4870]: I1014 08:59:52.856070 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 14 08:59:52 crc kubenswrapper[4870]: I1014 08:59:52.857933 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 14 08:59:52 crc kubenswrapper[4870]: I1014 08:59:52.858783 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 14 08:59:52 crc kubenswrapper[4870]: I1014 08:59:52.859122 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Oct 14 08:59:52 crc kubenswrapper[4870]: I1014 08:59:52.870861 4870 scope.go:117] "RemoveContainer" containerID="e30e539c18f4802d52a6d08cc9d481a425430a4eb4e49cf61c90f147eb29330c" Oct 14 08:59:52 crc kubenswrapper[4870]: I1014 08:59:52.870978 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 14 08:59:52 crc kubenswrapper[4870]: E1014 08:59:52.871309 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e30e539c18f4802d52a6d08cc9d481a425430a4eb4e49cf61c90f147eb29330c\": container with ID starting with e30e539c18f4802d52a6d08cc9d481a425430a4eb4e49cf61c90f147eb29330c not found: ID does not exist" containerID="e30e539c18f4802d52a6d08cc9d481a425430a4eb4e49cf61c90f147eb29330c" Oct 14 08:59:52 crc kubenswrapper[4870]: I1014 08:59:52.871338 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e30e539c18f4802d52a6d08cc9d481a425430a4eb4e49cf61c90f147eb29330c"} err="failed to get container status \"e30e539c18f4802d52a6d08cc9d481a425430a4eb4e49cf61c90f147eb29330c\": rpc error: code = NotFound desc = could not find container \"e30e539c18f4802d52a6d08cc9d481a425430a4eb4e49cf61c90f147eb29330c\": container with ID starting with e30e539c18f4802d52a6d08cc9d481a425430a4eb4e49cf61c90f147eb29330c not found: ID does not exist" Oct 14 08:59:52 crc kubenswrapper[4870]: I1014 08:59:52.871362 4870 scope.go:117] "RemoveContainer" containerID="3e4c6635911e376780a55963977f2060ede48f178692dd405d3881bc6d535969" Oct 14 08:59:52 crc kubenswrapper[4870]: E1014 08:59:52.872557 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3e4c6635911e376780a55963977f2060ede48f178692dd405d3881bc6d535969\": container with ID starting with 3e4c6635911e376780a55963977f2060ede48f178692dd405d3881bc6d535969 not found: ID does not exist" containerID="3e4c6635911e376780a55963977f2060ede48f178692dd405d3881bc6d535969" Oct 14 08:59:52 crc kubenswrapper[4870]: I1014 08:59:52.872577 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3e4c6635911e376780a55963977f2060ede48f178692dd405d3881bc6d535969"} err="failed to get container status \"3e4c6635911e376780a55963977f2060ede48f178692dd405d3881bc6d535969\": rpc error: code = NotFound desc = could not find container \"3e4c6635911e376780a55963977f2060ede48f178692dd405d3881bc6d535969\": container with ID starting with 3e4c6635911e376780a55963977f2060ede48f178692dd405d3881bc6d535969 not found: ID does not exist" Oct 14 08:59:52 crc kubenswrapper[4870]: I1014 08:59:52.872592 4870 scope.go:117] "RemoveContainer" containerID="e30e539c18f4802d52a6d08cc9d481a425430a4eb4e49cf61c90f147eb29330c" Oct 14 08:59:52 crc kubenswrapper[4870]: I1014 08:59:52.875195 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e30e539c18f4802d52a6d08cc9d481a425430a4eb4e49cf61c90f147eb29330c"} err="failed to get container status \"e30e539c18f4802d52a6d08cc9d481a425430a4eb4e49cf61c90f147eb29330c\": rpc error: code = NotFound desc = could not find container \"e30e539c18f4802d52a6d08cc9d481a425430a4eb4e49cf61c90f147eb29330c\": container with ID starting with e30e539c18f4802d52a6d08cc9d481a425430a4eb4e49cf61c90f147eb29330c not found: ID does not exist" Oct 14 08:59:52 crc kubenswrapper[4870]: I1014 08:59:52.875234 4870 scope.go:117] "RemoveContainer" containerID="3e4c6635911e376780a55963977f2060ede48f178692dd405d3881bc6d535969" Oct 14 08:59:52 crc kubenswrapper[4870]: I1014 08:59:52.875653 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3e4c6635911e376780a55963977f2060ede48f178692dd405d3881bc6d535969"} err="failed to get container status \"3e4c6635911e376780a55963977f2060ede48f178692dd405d3881bc6d535969\": rpc error: code = NotFound desc = could not find container \"3e4c6635911e376780a55963977f2060ede48f178692dd405d3881bc6d535969\": container with ID starting with 3e4c6635911e376780a55963977f2060ede48f178692dd405d3881bc6d535969 not found: ID does not exist" Oct 14 08:59:52 crc kubenswrapper[4870]: I1014 08:59:52.883592 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 14 08:59:52 crc kubenswrapper[4870]: I1014 08:59:52.947906 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fd2f3bc5-f9f6-434c-bf5a-364b7d8a05b9-logs\") pod \"nova-metadata-0\" (UID: \"fd2f3bc5-f9f6-434c-bf5a-364b7d8a05b9\") " pod="openstack/nova-metadata-0" Oct 14 08:59:52 crc kubenswrapper[4870]: I1014 08:59:52.947964 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7ffeeb3-4944-42e5-8789-a4922ab6e2c8-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"e7ffeeb3-4944-42e5-8789-a4922ab6e2c8\") " pod="openstack/nova-api-0" Oct 14 08:59:52 crc kubenswrapper[4870]: I1014 08:59:52.948012 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e7ffeeb3-4944-42e5-8789-a4922ab6e2c8-logs\") pod \"nova-api-0\" (UID: \"e7ffeeb3-4944-42e5-8789-a4922ab6e2c8\") " pod="openstack/nova-api-0" Oct 14 08:59:52 crc kubenswrapper[4870]: I1014 08:59:52.948194 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd2f3bc5-f9f6-434c-bf5a-364b7d8a05b9-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"fd2f3bc5-f9f6-434c-bf5a-364b7d8a05b9\") " pod="openstack/nova-metadata-0" Oct 14 08:59:52 crc kubenswrapper[4870]: I1014 08:59:52.948276 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fd2f3bc5-f9f6-434c-bf5a-364b7d8a05b9-config-data\") pod \"nova-metadata-0\" (UID: \"fd2f3bc5-f9f6-434c-bf5a-364b7d8a05b9\") " pod="openstack/nova-metadata-0" Oct 14 08:59:52 crc kubenswrapper[4870]: I1014 08:59:52.948306 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w5qxs\" (UniqueName: \"kubernetes.io/projected/9c8809d3-35d1-436b-b384-7114eb392a92-kube-api-access-w5qxs\") pod \"nova-cell1-conductor-0\" (UID: \"9c8809d3-35d1-436b-b384-7114eb392a92\") " pod="openstack/nova-cell1-conductor-0" Oct 14 08:59:52 crc kubenswrapper[4870]: I1014 08:59:52.948420 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c8809d3-35d1-436b-b384-7114eb392a92-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"9c8809d3-35d1-436b-b384-7114eb392a92\") " pod="openstack/nova-cell1-conductor-0" Oct 14 08:59:52 crc kubenswrapper[4870]: I1014 08:59:52.948555 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7ddfw\" (UniqueName: \"kubernetes.io/projected/e7ffeeb3-4944-42e5-8789-a4922ab6e2c8-kube-api-access-7ddfw\") pod \"nova-api-0\" (UID: \"e7ffeeb3-4944-42e5-8789-a4922ab6e2c8\") " pod="openstack/nova-api-0" Oct 14 08:59:52 crc kubenswrapper[4870]: I1014 08:59:52.948591 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cn8n9\" (UniqueName: \"kubernetes.io/projected/fd2f3bc5-f9f6-434c-bf5a-364b7d8a05b9-kube-api-access-cn8n9\") pod \"nova-metadata-0\" (UID: \"fd2f3bc5-f9f6-434c-bf5a-364b7d8a05b9\") " pod="openstack/nova-metadata-0" Oct 14 08:59:52 crc kubenswrapper[4870]: I1014 08:59:52.948743 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c8809d3-35d1-436b-b384-7114eb392a92-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"9c8809d3-35d1-436b-b384-7114eb392a92\") " pod="openstack/nova-cell1-conductor-0" Oct 14 08:59:52 crc kubenswrapper[4870]: I1014 08:59:52.948804 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e7ffeeb3-4944-42e5-8789-a4922ab6e2c8-config-data\") pod \"nova-api-0\" (UID: \"e7ffeeb3-4944-42e5-8789-a4922ab6e2c8\") " pod="openstack/nova-api-0" Oct 14 08:59:53 crc kubenswrapper[4870]: I1014 08:59:53.050075 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd2f3bc5-f9f6-434c-bf5a-364b7d8a05b9-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"fd2f3bc5-f9f6-434c-bf5a-364b7d8a05b9\") " pod="openstack/nova-metadata-0" Oct 14 08:59:53 crc kubenswrapper[4870]: I1014 08:59:53.050147 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fd2f3bc5-f9f6-434c-bf5a-364b7d8a05b9-config-data\") pod \"nova-metadata-0\" (UID: \"fd2f3bc5-f9f6-434c-bf5a-364b7d8a05b9\") " pod="openstack/nova-metadata-0" Oct 14 08:59:53 crc kubenswrapper[4870]: I1014 08:59:53.050183 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w5qxs\" (UniqueName: \"kubernetes.io/projected/9c8809d3-35d1-436b-b384-7114eb392a92-kube-api-access-w5qxs\") pod \"nova-cell1-conductor-0\" (UID: \"9c8809d3-35d1-436b-b384-7114eb392a92\") " pod="openstack/nova-cell1-conductor-0" Oct 14 08:59:53 crc kubenswrapper[4870]: I1014 08:59:53.050218 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c8809d3-35d1-436b-b384-7114eb392a92-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"9c8809d3-35d1-436b-b384-7114eb392a92\") " pod="openstack/nova-cell1-conductor-0" Oct 14 08:59:53 crc kubenswrapper[4870]: I1014 08:59:53.050269 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7ddfw\" (UniqueName: \"kubernetes.io/projected/e7ffeeb3-4944-42e5-8789-a4922ab6e2c8-kube-api-access-7ddfw\") pod \"nova-api-0\" (UID: \"e7ffeeb3-4944-42e5-8789-a4922ab6e2c8\") " pod="openstack/nova-api-0" Oct 14 08:59:53 crc kubenswrapper[4870]: I1014 08:59:53.050298 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cn8n9\" (UniqueName: \"kubernetes.io/projected/fd2f3bc5-f9f6-434c-bf5a-364b7d8a05b9-kube-api-access-cn8n9\") pod \"nova-metadata-0\" (UID: \"fd2f3bc5-f9f6-434c-bf5a-364b7d8a05b9\") " pod="openstack/nova-metadata-0" Oct 14 08:59:53 crc kubenswrapper[4870]: I1014 08:59:53.050365 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c8809d3-35d1-436b-b384-7114eb392a92-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"9c8809d3-35d1-436b-b384-7114eb392a92\") " pod="openstack/nova-cell1-conductor-0" Oct 14 08:59:53 crc kubenswrapper[4870]: I1014 08:59:53.050401 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e7ffeeb3-4944-42e5-8789-a4922ab6e2c8-config-data\") pod \"nova-api-0\" (UID: \"e7ffeeb3-4944-42e5-8789-a4922ab6e2c8\") " pod="openstack/nova-api-0" Oct 14 08:59:53 crc kubenswrapper[4870]: I1014 08:59:53.050454 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fd2f3bc5-f9f6-434c-bf5a-364b7d8a05b9-logs\") pod \"nova-metadata-0\" (UID: \"fd2f3bc5-f9f6-434c-bf5a-364b7d8a05b9\") " pod="openstack/nova-metadata-0" Oct 14 08:59:53 crc kubenswrapper[4870]: I1014 08:59:53.050484 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7ffeeb3-4944-42e5-8789-a4922ab6e2c8-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"e7ffeeb3-4944-42e5-8789-a4922ab6e2c8\") " pod="openstack/nova-api-0" Oct 14 08:59:53 crc kubenswrapper[4870]: I1014 08:59:53.050534 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e7ffeeb3-4944-42e5-8789-a4922ab6e2c8-logs\") pod \"nova-api-0\" (UID: \"e7ffeeb3-4944-42e5-8789-a4922ab6e2c8\") " pod="openstack/nova-api-0" Oct 14 08:59:53 crc kubenswrapper[4870]: I1014 08:59:53.050977 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e7ffeeb3-4944-42e5-8789-a4922ab6e2c8-logs\") pod \"nova-api-0\" (UID: \"e7ffeeb3-4944-42e5-8789-a4922ab6e2c8\") " pod="openstack/nova-api-0" Oct 14 08:59:53 crc kubenswrapper[4870]: I1014 08:59:53.052011 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fd2f3bc5-f9f6-434c-bf5a-364b7d8a05b9-logs\") pod \"nova-metadata-0\" (UID: \"fd2f3bc5-f9f6-434c-bf5a-364b7d8a05b9\") " pod="openstack/nova-metadata-0" Oct 14 08:59:53 crc kubenswrapper[4870]: I1014 08:59:53.057600 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7ffeeb3-4944-42e5-8789-a4922ab6e2c8-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"e7ffeeb3-4944-42e5-8789-a4922ab6e2c8\") " pod="openstack/nova-api-0" Oct 14 08:59:53 crc kubenswrapper[4870]: I1014 08:59:53.058466 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="057b319f-9b0f-4c6d-ae9c-3a18ded7d377" path="/var/lib/kubelet/pods/057b319f-9b0f-4c6d-ae9c-3a18ded7d377/volumes" Oct 14 08:59:53 crc kubenswrapper[4870]: I1014 08:59:53.060289 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e7ffeeb3-4944-42e5-8789-a4922ab6e2c8-config-data\") pod \"nova-api-0\" (UID: \"e7ffeeb3-4944-42e5-8789-a4922ab6e2c8\") " pod="openstack/nova-api-0" Oct 14 08:59:53 crc kubenswrapper[4870]: I1014 08:59:53.060414 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3b91331d-394b-4f04-be17-96fd0a71fc6e" path="/var/lib/kubelet/pods/3b91331d-394b-4f04-be17-96fd0a71fc6e/volumes" Oct 14 08:59:53 crc kubenswrapper[4870]: I1014 08:59:53.066558 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fd2f3bc5-f9f6-434c-bf5a-364b7d8a05b9-config-data\") pod \"nova-metadata-0\" (UID: \"fd2f3bc5-f9f6-434c-bf5a-364b7d8a05b9\") " pod="openstack/nova-metadata-0" Oct 14 08:59:53 crc kubenswrapper[4870]: I1014 08:59:53.068800 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd2f3bc5-f9f6-434c-bf5a-364b7d8a05b9-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"fd2f3bc5-f9f6-434c-bf5a-364b7d8a05b9\") " pod="openstack/nova-metadata-0" Oct 14 08:59:53 crc kubenswrapper[4870]: I1014 08:59:53.078522 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c8809d3-35d1-436b-b384-7114eb392a92-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"9c8809d3-35d1-436b-b384-7114eb392a92\") " pod="openstack/nova-cell1-conductor-0" Oct 14 08:59:53 crc kubenswrapper[4870]: I1014 08:59:53.080038 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c8809d3-35d1-436b-b384-7114eb392a92-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"9c8809d3-35d1-436b-b384-7114eb392a92\") " pod="openstack/nova-cell1-conductor-0" Oct 14 08:59:53 crc kubenswrapper[4870]: I1014 08:59:53.080527 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7ddfw\" (UniqueName: \"kubernetes.io/projected/e7ffeeb3-4944-42e5-8789-a4922ab6e2c8-kube-api-access-7ddfw\") pod \"nova-api-0\" (UID: \"e7ffeeb3-4944-42e5-8789-a4922ab6e2c8\") " pod="openstack/nova-api-0" Oct 14 08:59:53 crc kubenswrapper[4870]: I1014 08:59:53.092296 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cn8n9\" (UniqueName: \"kubernetes.io/projected/fd2f3bc5-f9f6-434c-bf5a-364b7d8a05b9-kube-api-access-cn8n9\") pod \"nova-metadata-0\" (UID: \"fd2f3bc5-f9f6-434c-bf5a-364b7d8a05b9\") " pod="openstack/nova-metadata-0" Oct 14 08:59:53 crc kubenswrapper[4870]: I1014 08:59:53.124789 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w5qxs\" (UniqueName: \"kubernetes.io/projected/9c8809d3-35d1-436b-b384-7114eb392a92-kube-api-access-w5qxs\") pod \"nova-cell1-conductor-0\" (UID: \"9c8809d3-35d1-436b-b384-7114eb392a92\") " pod="openstack/nova-cell1-conductor-0" Oct 14 08:59:53 crc kubenswrapper[4870]: I1014 08:59:53.126653 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 14 08:59:53 crc kubenswrapper[4870]: I1014 08:59:53.183171 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 14 08:59:53 crc kubenswrapper[4870]: I1014 08:59:53.193047 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 14 08:59:53 crc kubenswrapper[4870]: I1014 08:59:53.645723 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 14 08:59:53 crc kubenswrapper[4870]: I1014 08:59:53.736772 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"fd2f3bc5-f9f6-434c-bf5a-364b7d8a05b9","Type":"ContainerStarted","Data":"637cce04ef35ae06562b0ad8fc9768602a9d7d1f4b720c68246cf6173eb6f71b"} Oct 14 08:59:53 crc kubenswrapper[4870]: I1014 08:59:53.737386 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 14 08:59:53 crc kubenswrapper[4870]: W1014 08:59:53.752539 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode7ffeeb3_4944_42e5_8789_a4922ab6e2c8.slice/crio-266d7eb76949e8fe2775fffc7aa4b782b73952f107e16bda5d19cc3e52a91aa8 WatchSource:0}: Error finding container 266d7eb76949e8fe2775fffc7aa4b782b73952f107e16bda5d19cc3e52a91aa8: Status 404 returned error can't find the container with id 266d7eb76949e8fe2775fffc7aa4b782b73952f107e16bda5d19cc3e52a91aa8 Oct 14 08:59:53 crc kubenswrapper[4870]: I1014 08:59:53.903263 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 14 08:59:54 crc kubenswrapper[4870]: I1014 08:59:54.314875 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Oct 14 08:59:54 crc kubenswrapper[4870]: I1014 08:59:54.328651 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Oct 14 08:59:54 crc kubenswrapper[4870]: I1014 08:59:54.338384 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7c8cbf976c-khp98" Oct 14 08:59:54 crc kubenswrapper[4870]: I1014 08:59:54.400472 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7d6744f579-msx66"] Oct 14 08:59:54 crc kubenswrapper[4870]: I1014 08:59:54.401166 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7d6744f579-msx66" podUID="a04eef9c-afd1-41ae-af51-9d4d97a40a60" containerName="dnsmasq-dns" containerID="cri-o://6271a06ee44d0217f821148cf6af1d6eee1c6f9527fc3bf11f0f48dd83e8facd" gracePeriod=10 Oct 14 08:59:54 crc kubenswrapper[4870]: I1014 08:59:54.748300 4870 generic.go:334] "Generic (PLEG): container finished" podID="93eea128-7b79-4505-b492-9c34adf49cdf" containerID="90ad58e4ecdb8f55072333851a3ce5d4c55c7e8e9c11bb2d18e8c77dc85b0d56" exitCode=0 Oct 14 08:59:54 crc kubenswrapper[4870]: I1014 08:59:54.748364 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"93eea128-7b79-4505-b492-9c34adf49cdf","Type":"ContainerDied","Data":"90ad58e4ecdb8f55072333851a3ce5d4c55c7e8e9c11bb2d18e8c77dc85b0d56"} Oct 14 08:59:54 crc kubenswrapper[4870]: I1014 08:59:54.749900 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"9c8809d3-35d1-436b-b384-7114eb392a92","Type":"ContainerStarted","Data":"15973d9dc9aabc88d034bacb72c7d2fecc6418b3360a7881625e8c9c98f02e5d"} Oct 14 08:59:54 crc kubenswrapper[4870]: I1014 08:59:54.749921 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"9c8809d3-35d1-436b-b384-7114eb392a92","Type":"ContainerStarted","Data":"222da9312d9c0ffe5bbd7ec4e4fbf2f8e44c6f88e2357e309e2a7d8de8a5fce5"} Oct 14 08:59:54 crc kubenswrapper[4870]: I1014 08:59:54.751138 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Oct 14 08:59:54 crc kubenswrapper[4870]: I1014 08:59:54.752662 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e7ffeeb3-4944-42e5-8789-a4922ab6e2c8","Type":"ContainerStarted","Data":"fe1fbc1e1ac5645d510a2c4bd1d21d900c44d1f5c7e01e6926224a3ef748981b"} Oct 14 08:59:54 crc kubenswrapper[4870]: I1014 08:59:54.752681 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e7ffeeb3-4944-42e5-8789-a4922ab6e2c8","Type":"ContainerStarted","Data":"21712480245d9483d0084dd3bac8b74bd6478ff05aefecc24a739a42f104e3b0"} Oct 14 08:59:54 crc kubenswrapper[4870]: I1014 08:59:54.752691 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e7ffeeb3-4944-42e5-8789-a4922ab6e2c8","Type":"ContainerStarted","Data":"266d7eb76949e8fe2775fffc7aa4b782b73952f107e16bda5d19cc3e52a91aa8"} Oct 14 08:59:54 crc kubenswrapper[4870]: I1014 08:59:54.755995 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"fd2f3bc5-f9f6-434c-bf5a-364b7d8a05b9","Type":"ContainerStarted","Data":"2a61496f82dca223ef233b4d38c74d2e577859ab091f9959e058e706b27e7029"} Oct 14 08:59:54 crc kubenswrapper[4870]: I1014 08:59:54.756038 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"fd2f3bc5-f9f6-434c-bf5a-364b7d8a05b9","Type":"ContainerStarted","Data":"87f9c0df03f215e3cf69750e3466233c1f0f00ec14013ce4ec59d7f6ba7044d6"} Oct 14 08:59:54 crc kubenswrapper[4870]: I1014 08:59:54.760851 4870 generic.go:334] "Generic (PLEG): container finished" podID="a04eef9c-afd1-41ae-af51-9d4d97a40a60" containerID="6271a06ee44d0217f821148cf6af1d6eee1c6f9527fc3bf11f0f48dd83e8facd" exitCode=0 Oct 14 08:59:54 crc kubenswrapper[4870]: I1014 08:59:54.763132 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d6744f579-msx66" event={"ID":"a04eef9c-afd1-41ae-af51-9d4d97a40a60","Type":"ContainerDied","Data":"6271a06ee44d0217f821148cf6af1d6eee1c6f9527fc3bf11f0f48dd83e8facd"} Oct 14 08:59:54 crc kubenswrapper[4870]: I1014 08:59:54.781681 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Oct 14 08:59:54 crc kubenswrapper[4870]: I1014 08:59:54.781921 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.781900965 podStartE2EDuration="2.781900965s" podCreationTimestamp="2025-10-14 08:59:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 08:59:54.767348066 +0000 UTC m=+7130.464708437" watchObservedRunningTime="2025-10-14 08:59:54.781900965 +0000 UTC m=+7130.479261336" Oct 14 08:59:54 crc kubenswrapper[4870]: I1014 08:59:54.795157 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.795138562 podStartE2EDuration="2.795138562s" podCreationTimestamp="2025-10-14 08:59:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 08:59:54.788527199 +0000 UTC m=+7130.485887580" watchObservedRunningTime="2025-10-14 08:59:54.795138562 +0000 UTC m=+7130.492498933" Oct 14 08:59:54 crc kubenswrapper[4870]: I1014 08:59:54.831615 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.8315910520000003 podStartE2EDuration="2.831591052s" podCreationTimestamp="2025-10-14 08:59:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 08:59:54.817794081 +0000 UTC m=+7130.515154462" watchObservedRunningTime="2025-10-14 08:59:54.831591052 +0000 UTC m=+7130.528951433" Oct 14 08:59:54 crc kubenswrapper[4870]: I1014 08:59:54.880665 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 14 08:59:54 crc kubenswrapper[4870]: I1014 08:59:54.966096 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7d6744f579-msx66" Oct 14 08:59:54 crc kubenswrapper[4870]: I1014 08:59:54.993926 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/93eea128-7b79-4505-b492-9c34adf49cdf-config-data\") pod \"93eea128-7b79-4505-b492-9c34adf49cdf\" (UID: \"93eea128-7b79-4505-b492-9c34adf49cdf\") " Oct 14 08:59:54 crc kubenswrapper[4870]: I1014 08:59:54.994790 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93eea128-7b79-4505-b492-9c34adf49cdf-combined-ca-bundle\") pod \"93eea128-7b79-4505-b492-9c34adf49cdf\" (UID: \"93eea128-7b79-4505-b492-9c34adf49cdf\") " Oct 14 08:59:54 crc kubenswrapper[4870]: I1014 08:59:54.994866 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pmdjc\" (UniqueName: \"kubernetes.io/projected/93eea128-7b79-4505-b492-9c34adf49cdf-kube-api-access-pmdjc\") pod \"93eea128-7b79-4505-b492-9c34adf49cdf\" (UID: \"93eea128-7b79-4505-b492-9c34adf49cdf\") " Oct 14 08:59:55 crc kubenswrapper[4870]: I1014 08:59:55.003364 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/93eea128-7b79-4505-b492-9c34adf49cdf-kube-api-access-pmdjc" (OuterVolumeSpecName: "kube-api-access-pmdjc") pod "93eea128-7b79-4505-b492-9c34adf49cdf" (UID: "93eea128-7b79-4505-b492-9c34adf49cdf"). InnerVolumeSpecName "kube-api-access-pmdjc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:59:55 crc kubenswrapper[4870]: I1014 08:59:55.029653 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/93eea128-7b79-4505-b492-9c34adf49cdf-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "93eea128-7b79-4505-b492-9c34adf49cdf" (UID: "93eea128-7b79-4505-b492-9c34adf49cdf"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:59:55 crc kubenswrapper[4870]: I1014 08:59:55.037467 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/93eea128-7b79-4505-b492-9c34adf49cdf-config-data" (OuterVolumeSpecName: "config-data") pod "93eea128-7b79-4505-b492-9c34adf49cdf" (UID: "93eea128-7b79-4505-b492-9c34adf49cdf"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 08:59:55 crc kubenswrapper[4870]: I1014 08:59:55.098771 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a04eef9c-afd1-41ae-af51-9d4d97a40a60-dns-svc\") pod \"a04eef9c-afd1-41ae-af51-9d4d97a40a60\" (UID: \"a04eef9c-afd1-41ae-af51-9d4d97a40a60\") " Oct 14 08:59:55 crc kubenswrapper[4870]: I1014 08:59:55.098808 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a04eef9c-afd1-41ae-af51-9d4d97a40a60-ovsdbserver-nb\") pod \"a04eef9c-afd1-41ae-af51-9d4d97a40a60\" (UID: \"a04eef9c-afd1-41ae-af51-9d4d97a40a60\") " Oct 14 08:59:55 crc kubenswrapper[4870]: I1014 08:59:55.098828 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hqwsz\" (UniqueName: \"kubernetes.io/projected/a04eef9c-afd1-41ae-af51-9d4d97a40a60-kube-api-access-hqwsz\") pod \"a04eef9c-afd1-41ae-af51-9d4d97a40a60\" (UID: \"a04eef9c-afd1-41ae-af51-9d4d97a40a60\") " Oct 14 08:59:55 crc kubenswrapper[4870]: I1014 08:59:55.099021 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a04eef9c-afd1-41ae-af51-9d4d97a40a60-ovsdbserver-sb\") pod \"a04eef9c-afd1-41ae-af51-9d4d97a40a60\" (UID: \"a04eef9c-afd1-41ae-af51-9d4d97a40a60\") " Oct 14 08:59:55 crc kubenswrapper[4870]: I1014 08:59:55.099129 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a04eef9c-afd1-41ae-af51-9d4d97a40a60-config\") pod \"a04eef9c-afd1-41ae-af51-9d4d97a40a60\" (UID: \"a04eef9c-afd1-41ae-af51-9d4d97a40a60\") " Oct 14 08:59:55 crc kubenswrapper[4870]: I1014 08:59:55.099711 4870 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/93eea128-7b79-4505-b492-9c34adf49cdf-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 08:59:55 crc kubenswrapper[4870]: I1014 08:59:55.099724 4870 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93eea128-7b79-4505-b492-9c34adf49cdf-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 08:59:55 crc kubenswrapper[4870]: I1014 08:59:55.099735 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pmdjc\" (UniqueName: \"kubernetes.io/projected/93eea128-7b79-4505-b492-9c34adf49cdf-kube-api-access-pmdjc\") on node \"crc\" DevicePath \"\"" Oct 14 08:59:55 crc kubenswrapper[4870]: I1014 08:59:55.105702 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a04eef9c-afd1-41ae-af51-9d4d97a40a60-kube-api-access-hqwsz" (OuterVolumeSpecName: "kube-api-access-hqwsz") pod "a04eef9c-afd1-41ae-af51-9d4d97a40a60" (UID: "a04eef9c-afd1-41ae-af51-9d4d97a40a60"). InnerVolumeSpecName "kube-api-access-hqwsz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 08:59:55 crc kubenswrapper[4870]: I1014 08:59:55.156601 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a04eef9c-afd1-41ae-af51-9d4d97a40a60-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "a04eef9c-afd1-41ae-af51-9d4d97a40a60" (UID: "a04eef9c-afd1-41ae-af51-9d4d97a40a60"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:59:55 crc kubenswrapper[4870]: I1014 08:59:55.159404 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a04eef9c-afd1-41ae-af51-9d4d97a40a60-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "a04eef9c-afd1-41ae-af51-9d4d97a40a60" (UID: "a04eef9c-afd1-41ae-af51-9d4d97a40a60"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:59:55 crc kubenswrapper[4870]: I1014 08:59:55.166730 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a04eef9c-afd1-41ae-af51-9d4d97a40a60-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "a04eef9c-afd1-41ae-af51-9d4d97a40a60" (UID: "a04eef9c-afd1-41ae-af51-9d4d97a40a60"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:59:55 crc kubenswrapper[4870]: I1014 08:59:55.201845 4870 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a04eef9c-afd1-41ae-af51-9d4d97a40a60-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 14 08:59:55 crc kubenswrapper[4870]: I1014 08:59:55.201879 4870 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a04eef9c-afd1-41ae-af51-9d4d97a40a60-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 14 08:59:55 crc kubenswrapper[4870]: I1014 08:59:55.201889 4870 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a04eef9c-afd1-41ae-af51-9d4d97a40a60-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 14 08:59:55 crc kubenswrapper[4870]: I1014 08:59:55.201897 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hqwsz\" (UniqueName: \"kubernetes.io/projected/a04eef9c-afd1-41ae-af51-9d4d97a40a60-kube-api-access-hqwsz\") on node \"crc\" DevicePath \"\"" Oct 14 08:59:55 crc kubenswrapper[4870]: I1014 08:59:55.204264 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a04eef9c-afd1-41ae-af51-9d4d97a40a60-config" (OuterVolumeSpecName: "config") pod "a04eef9c-afd1-41ae-af51-9d4d97a40a60" (UID: "a04eef9c-afd1-41ae-af51-9d4d97a40a60"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 08:59:55 crc kubenswrapper[4870]: I1014 08:59:55.304388 4870 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a04eef9c-afd1-41ae-af51-9d4d97a40a60-config\") on node \"crc\" DevicePath \"\"" Oct 14 08:59:55 crc kubenswrapper[4870]: I1014 08:59:55.772231 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d6744f579-msx66" event={"ID":"a04eef9c-afd1-41ae-af51-9d4d97a40a60","Type":"ContainerDied","Data":"3372c8de6d8608ed6d47e15fc528ec3abf81705c7c79f25cde49586b5998cff1"} Oct 14 08:59:55 crc kubenswrapper[4870]: I1014 08:59:55.772280 4870 scope.go:117] "RemoveContainer" containerID="6271a06ee44d0217f821148cf6af1d6eee1c6f9527fc3bf11f0f48dd83e8facd" Oct 14 08:59:55 crc kubenswrapper[4870]: I1014 08:59:55.772390 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7d6744f579-msx66" Oct 14 08:59:55 crc kubenswrapper[4870]: I1014 08:59:55.786346 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 14 08:59:55 crc kubenswrapper[4870]: I1014 08:59:55.786556 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"93eea128-7b79-4505-b492-9c34adf49cdf","Type":"ContainerDied","Data":"6ff34cc4cb8acdeb81adc100abf2b12e95129c9ffddd32c743f4aba60ee786d7"} Oct 14 08:59:55 crc kubenswrapper[4870]: I1014 08:59:55.808413 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7d6744f579-msx66"] Oct 14 08:59:55 crc kubenswrapper[4870]: I1014 08:59:55.815018 4870 scope.go:117] "RemoveContainer" containerID="a973acaa26add648ee12bf8f9f22a327b5650bca2d8d38945dbf06f00f26b639" Oct 14 08:59:55 crc kubenswrapper[4870]: I1014 08:59:55.816576 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7d6744f579-msx66"] Oct 14 08:59:55 crc kubenswrapper[4870]: I1014 08:59:55.829780 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 14 08:59:55 crc kubenswrapper[4870]: I1014 08:59:55.836234 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Oct 14 08:59:55 crc kubenswrapper[4870]: I1014 08:59:55.846014 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 14 08:59:55 crc kubenswrapper[4870]: E1014 08:59:55.846670 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93eea128-7b79-4505-b492-9c34adf49cdf" containerName="nova-scheduler-scheduler" Oct 14 08:59:55 crc kubenswrapper[4870]: I1014 08:59:55.846694 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="93eea128-7b79-4505-b492-9c34adf49cdf" containerName="nova-scheduler-scheduler" Oct 14 08:59:55 crc kubenswrapper[4870]: E1014 08:59:55.846731 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a04eef9c-afd1-41ae-af51-9d4d97a40a60" containerName="init" Oct 14 08:59:55 crc kubenswrapper[4870]: I1014 08:59:55.846743 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="a04eef9c-afd1-41ae-af51-9d4d97a40a60" containerName="init" Oct 14 08:59:55 crc kubenswrapper[4870]: E1014 08:59:55.846766 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a04eef9c-afd1-41ae-af51-9d4d97a40a60" containerName="dnsmasq-dns" Oct 14 08:59:55 crc kubenswrapper[4870]: I1014 08:59:55.846775 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="a04eef9c-afd1-41ae-af51-9d4d97a40a60" containerName="dnsmasq-dns" Oct 14 08:59:55 crc kubenswrapper[4870]: I1014 08:59:55.846946 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="a04eef9c-afd1-41ae-af51-9d4d97a40a60" containerName="dnsmasq-dns" Oct 14 08:59:55 crc kubenswrapper[4870]: I1014 08:59:55.846964 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="93eea128-7b79-4505-b492-9c34adf49cdf" containerName="nova-scheduler-scheduler" Oct 14 08:59:55 crc kubenswrapper[4870]: I1014 08:59:55.847631 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 14 08:59:55 crc kubenswrapper[4870]: I1014 08:59:55.849480 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 14 08:59:55 crc kubenswrapper[4870]: I1014 08:59:55.855848 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 14 08:59:55 crc kubenswrapper[4870]: I1014 08:59:55.857247 4870 scope.go:117] "RemoveContainer" containerID="90ad58e4ecdb8f55072333851a3ce5d4c55c7e8e9c11bb2d18e8c77dc85b0d56" Oct 14 08:59:55 crc kubenswrapper[4870]: I1014 08:59:55.915067 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/77e2dbb6-3f73-4617-994b-cacf4a839808-config-data\") pod \"nova-scheduler-0\" (UID: \"77e2dbb6-3f73-4617-994b-cacf4a839808\") " pod="openstack/nova-scheduler-0" Oct 14 08:59:55 crc kubenswrapper[4870]: I1014 08:59:55.915227 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77e2dbb6-3f73-4617-994b-cacf4a839808-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"77e2dbb6-3f73-4617-994b-cacf4a839808\") " pod="openstack/nova-scheduler-0" Oct 14 08:59:55 crc kubenswrapper[4870]: I1014 08:59:55.915272 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wzdhh\" (UniqueName: \"kubernetes.io/projected/77e2dbb6-3f73-4617-994b-cacf4a839808-kube-api-access-wzdhh\") pod \"nova-scheduler-0\" (UID: \"77e2dbb6-3f73-4617-994b-cacf4a839808\") " pod="openstack/nova-scheduler-0" Oct 14 08:59:56 crc kubenswrapper[4870]: I1014 08:59:56.017510 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/77e2dbb6-3f73-4617-994b-cacf4a839808-config-data\") pod \"nova-scheduler-0\" (UID: \"77e2dbb6-3f73-4617-994b-cacf4a839808\") " pod="openstack/nova-scheduler-0" Oct 14 08:59:56 crc kubenswrapper[4870]: I1014 08:59:56.017645 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77e2dbb6-3f73-4617-994b-cacf4a839808-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"77e2dbb6-3f73-4617-994b-cacf4a839808\") " pod="openstack/nova-scheduler-0" Oct 14 08:59:56 crc kubenswrapper[4870]: I1014 08:59:56.017682 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wzdhh\" (UniqueName: \"kubernetes.io/projected/77e2dbb6-3f73-4617-994b-cacf4a839808-kube-api-access-wzdhh\") pod \"nova-scheduler-0\" (UID: \"77e2dbb6-3f73-4617-994b-cacf4a839808\") " pod="openstack/nova-scheduler-0" Oct 14 08:59:56 crc kubenswrapper[4870]: I1014 08:59:56.022087 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77e2dbb6-3f73-4617-994b-cacf4a839808-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"77e2dbb6-3f73-4617-994b-cacf4a839808\") " pod="openstack/nova-scheduler-0" Oct 14 08:59:56 crc kubenswrapper[4870]: I1014 08:59:56.024863 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/77e2dbb6-3f73-4617-994b-cacf4a839808-config-data\") pod \"nova-scheduler-0\" (UID: \"77e2dbb6-3f73-4617-994b-cacf4a839808\") " pod="openstack/nova-scheduler-0" Oct 14 08:59:56 crc kubenswrapper[4870]: I1014 08:59:56.045191 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wzdhh\" (UniqueName: \"kubernetes.io/projected/77e2dbb6-3f73-4617-994b-cacf4a839808-kube-api-access-wzdhh\") pod \"nova-scheduler-0\" (UID: \"77e2dbb6-3f73-4617-994b-cacf4a839808\") " pod="openstack/nova-scheduler-0" Oct 14 08:59:56 crc kubenswrapper[4870]: I1014 08:59:56.170407 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 14 08:59:56 crc kubenswrapper[4870]: W1014 08:59:56.615499 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod77e2dbb6_3f73_4617_994b_cacf4a839808.slice/crio-6514c581a068725908b3af7bcdcb1553f42fcec2956cfaaa8b54045dc4c231e2 WatchSource:0}: Error finding container 6514c581a068725908b3af7bcdcb1553f42fcec2956cfaaa8b54045dc4c231e2: Status 404 returned error can't find the container with id 6514c581a068725908b3af7bcdcb1553f42fcec2956cfaaa8b54045dc4c231e2 Oct 14 08:59:56 crc kubenswrapper[4870]: I1014 08:59:56.619620 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 14 08:59:56 crc kubenswrapper[4870]: I1014 08:59:56.797709 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"77e2dbb6-3f73-4617-994b-cacf4a839808","Type":"ContainerStarted","Data":"6514c581a068725908b3af7bcdcb1553f42fcec2956cfaaa8b54045dc4c231e2"} Oct 14 08:59:57 crc kubenswrapper[4870]: I1014 08:59:57.074428 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="93eea128-7b79-4505-b492-9c34adf49cdf" path="/var/lib/kubelet/pods/93eea128-7b79-4505-b492-9c34adf49cdf/volumes" Oct 14 08:59:57 crc kubenswrapper[4870]: I1014 08:59:57.081175 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a04eef9c-afd1-41ae-af51-9d4d97a40a60" path="/var/lib/kubelet/pods/a04eef9c-afd1-41ae-af51-9d4d97a40a60/volumes" Oct 14 08:59:57 crc kubenswrapper[4870]: I1014 08:59:57.829749 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"77e2dbb6-3f73-4617-994b-cacf4a839808","Type":"ContainerStarted","Data":"eb61e04aa194a1edd6e70017933438f95a86807b7f7bce94ffaf66a42d4a8799"} Oct 14 08:59:57 crc kubenswrapper[4870]: I1014 08:59:57.853038 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.853010626 podStartE2EDuration="2.853010626s" podCreationTimestamp="2025-10-14 08:59:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 08:59:57.849777527 +0000 UTC m=+7133.547137908" watchObservedRunningTime="2025-10-14 08:59:57.853010626 +0000 UTC m=+7133.550371027" Oct 14 08:59:58 crc kubenswrapper[4870]: I1014 08:59:58.127966 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 14 08:59:58 crc kubenswrapper[4870]: I1014 08:59:58.128045 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 14 09:00:00 crc kubenswrapper[4870]: I1014 09:00:00.169995 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29340540-jw8p5"] Oct 14 09:00:00 crc kubenswrapper[4870]: I1014 09:00:00.171586 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29340540-jw8p5" Oct 14 09:00:00 crc kubenswrapper[4870]: I1014 09:00:00.184067 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29340540-jw8p5"] Oct 14 09:00:00 crc kubenswrapper[4870]: I1014 09:00:00.216276 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 14 09:00:00 crc kubenswrapper[4870]: I1014 09:00:00.216343 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 14 09:00:00 crc kubenswrapper[4870]: I1014 09:00:00.315847 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4c777f9d-8c71-4d17-ac52-ae341bd066a2-config-volume\") pod \"collect-profiles-29340540-jw8p5\" (UID: \"4c777f9d-8c71-4d17-ac52-ae341bd066a2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340540-jw8p5" Oct 14 09:00:00 crc kubenswrapper[4870]: I1014 09:00:00.316050 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4c777f9d-8c71-4d17-ac52-ae341bd066a2-secret-volume\") pod \"collect-profiles-29340540-jw8p5\" (UID: \"4c777f9d-8c71-4d17-ac52-ae341bd066a2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340540-jw8p5" Oct 14 09:00:00 crc kubenswrapper[4870]: I1014 09:00:00.316143 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-598gr\" (UniqueName: \"kubernetes.io/projected/4c777f9d-8c71-4d17-ac52-ae341bd066a2-kube-api-access-598gr\") pod \"collect-profiles-29340540-jw8p5\" (UID: \"4c777f9d-8c71-4d17-ac52-ae341bd066a2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340540-jw8p5" Oct 14 09:00:00 crc kubenswrapper[4870]: I1014 09:00:00.417963 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4c777f9d-8c71-4d17-ac52-ae341bd066a2-config-volume\") pod \"collect-profiles-29340540-jw8p5\" (UID: \"4c777f9d-8c71-4d17-ac52-ae341bd066a2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340540-jw8p5" Oct 14 09:00:00 crc kubenswrapper[4870]: I1014 09:00:00.418037 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4c777f9d-8c71-4d17-ac52-ae341bd066a2-secret-volume\") pod \"collect-profiles-29340540-jw8p5\" (UID: \"4c777f9d-8c71-4d17-ac52-ae341bd066a2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340540-jw8p5" Oct 14 09:00:00 crc kubenswrapper[4870]: I1014 09:00:00.418121 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-598gr\" (UniqueName: \"kubernetes.io/projected/4c777f9d-8c71-4d17-ac52-ae341bd066a2-kube-api-access-598gr\") pod \"collect-profiles-29340540-jw8p5\" (UID: \"4c777f9d-8c71-4d17-ac52-ae341bd066a2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340540-jw8p5" Oct 14 09:00:00 crc kubenswrapper[4870]: I1014 09:00:00.419255 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4c777f9d-8c71-4d17-ac52-ae341bd066a2-config-volume\") pod \"collect-profiles-29340540-jw8p5\" (UID: \"4c777f9d-8c71-4d17-ac52-ae341bd066a2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340540-jw8p5" Oct 14 09:00:00 crc kubenswrapper[4870]: I1014 09:00:00.427964 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4c777f9d-8c71-4d17-ac52-ae341bd066a2-secret-volume\") pod \"collect-profiles-29340540-jw8p5\" (UID: \"4c777f9d-8c71-4d17-ac52-ae341bd066a2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340540-jw8p5" Oct 14 09:00:00 crc kubenswrapper[4870]: I1014 09:00:00.433750 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-598gr\" (UniqueName: \"kubernetes.io/projected/4c777f9d-8c71-4d17-ac52-ae341bd066a2-kube-api-access-598gr\") pod \"collect-profiles-29340540-jw8p5\" (UID: \"4c777f9d-8c71-4d17-ac52-ae341bd066a2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340540-jw8p5" Oct 14 09:00:00 crc kubenswrapper[4870]: I1014 09:00:00.546268 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29340540-jw8p5" Oct 14 09:00:01 crc kubenswrapper[4870]: I1014 09:00:01.034486 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29340540-jw8p5"] Oct 14 09:00:01 crc kubenswrapper[4870]: I1014 09:00:01.171745 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 14 09:00:01 crc kubenswrapper[4870]: I1014 09:00:01.875774 4870 generic.go:334] "Generic (PLEG): container finished" podID="4c777f9d-8c71-4d17-ac52-ae341bd066a2" containerID="6d6e738bef449215a9a338332d1ae8ffd723ed0b9e5443e70020c01178dcd790" exitCode=0 Oct 14 09:00:01 crc kubenswrapper[4870]: I1014 09:00:01.875854 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29340540-jw8p5" event={"ID":"4c777f9d-8c71-4d17-ac52-ae341bd066a2","Type":"ContainerDied","Data":"6d6e738bef449215a9a338332d1ae8ffd723ed0b9e5443e70020c01178dcd790"} Oct 14 09:00:01 crc kubenswrapper[4870]: I1014 09:00:01.875982 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29340540-jw8p5" event={"ID":"4c777f9d-8c71-4d17-ac52-ae341bd066a2","Type":"ContainerStarted","Data":"2755740d7ce689c73b7bbb1b15bac463b790ae4fc48132c3e9331d0be2e7feb1"} Oct 14 09:00:03 crc kubenswrapper[4870]: I1014 09:00:03.127619 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 14 09:00:03 crc kubenswrapper[4870]: I1014 09:00:03.128297 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 14 09:00:03 crc kubenswrapper[4870]: I1014 09:00:03.184937 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 14 09:00:03 crc kubenswrapper[4870]: I1014 09:00:03.184980 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 14 09:00:03 crc kubenswrapper[4870]: I1014 09:00:03.215370 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29340540-jw8p5" Oct 14 09:00:03 crc kubenswrapper[4870]: I1014 09:00:03.231924 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Oct 14 09:00:03 crc kubenswrapper[4870]: I1014 09:00:03.371654 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-598gr\" (UniqueName: \"kubernetes.io/projected/4c777f9d-8c71-4d17-ac52-ae341bd066a2-kube-api-access-598gr\") pod \"4c777f9d-8c71-4d17-ac52-ae341bd066a2\" (UID: \"4c777f9d-8c71-4d17-ac52-ae341bd066a2\") " Oct 14 09:00:03 crc kubenswrapper[4870]: I1014 09:00:03.372005 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4c777f9d-8c71-4d17-ac52-ae341bd066a2-secret-volume\") pod \"4c777f9d-8c71-4d17-ac52-ae341bd066a2\" (UID: \"4c777f9d-8c71-4d17-ac52-ae341bd066a2\") " Oct 14 09:00:03 crc kubenswrapper[4870]: I1014 09:00:03.372182 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4c777f9d-8c71-4d17-ac52-ae341bd066a2-config-volume\") pod \"4c777f9d-8c71-4d17-ac52-ae341bd066a2\" (UID: \"4c777f9d-8c71-4d17-ac52-ae341bd066a2\") " Oct 14 09:00:03 crc kubenswrapper[4870]: I1014 09:00:03.373618 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4c777f9d-8c71-4d17-ac52-ae341bd066a2-config-volume" (OuterVolumeSpecName: "config-volume") pod "4c777f9d-8c71-4d17-ac52-ae341bd066a2" (UID: "4c777f9d-8c71-4d17-ac52-ae341bd066a2"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 09:00:03 crc kubenswrapper[4870]: I1014 09:00:03.377930 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c777f9d-8c71-4d17-ac52-ae341bd066a2-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "4c777f9d-8c71-4d17-ac52-ae341bd066a2" (UID: "4c777f9d-8c71-4d17-ac52-ae341bd066a2"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:00:03 crc kubenswrapper[4870]: I1014 09:00:03.386125 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4c777f9d-8c71-4d17-ac52-ae341bd066a2-kube-api-access-598gr" (OuterVolumeSpecName: "kube-api-access-598gr") pod "4c777f9d-8c71-4d17-ac52-ae341bd066a2" (UID: "4c777f9d-8c71-4d17-ac52-ae341bd066a2"). InnerVolumeSpecName "kube-api-access-598gr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:00:03 crc kubenswrapper[4870]: I1014 09:00:03.473794 4870 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4c777f9d-8c71-4d17-ac52-ae341bd066a2-config-volume\") on node \"crc\" DevicePath \"\"" Oct 14 09:00:03 crc kubenswrapper[4870]: I1014 09:00:03.473833 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-598gr\" (UniqueName: \"kubernetes.io/projected/4c777f9d-8c71-4d17-ac52-ae341bd066a2-kube-api-access-598gr\") on node \"crc\" DevicePath \"\"" Oct 14 09:00:03 crc kubenswrapper[4870]: I1014 09:00:03.473848 4870 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4c777f9d-8c71-4d17-ac52-ae341bd066a2-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 14 09:00:03 crc kubenswrapper[4870]: I1014 09:00:03.851612 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-fbdnv"] Oct 14 09:00:03 crc kubenswrapper[4870]: E1014 09:00:03.852083 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c777f9d-8c71-4d17-ac52-ae341bd066a2" containerName="collect-profiles" Oct 14 09:00:03 crc kubenswrapper[4870]: I1014 09:00:03.852103 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c777f9d-8c71-4d17-ac52-ae341bd066a2" containerName="collect-profiles" Oct 14 09:00:03 crc kubenswrapper[4870]: I1014 09:00:03.852343 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="4c777f9d-8c71-4d17-ac52-ae341bd066a2" containerName="collect-profiles" Oct 14 09:00:03 crc kubenswrapper[4870]: I1014 09:00:03.853079 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-fbdnv" Oct 14 09:00:03 crc kubenswrapper[4870]: I1014 09:00:03.855307 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Oct 14 09:00:03 crc kubenswrapper[4870]: I1014 09:00:03.855619 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Oct 14 09:00:03 crc kubenswrapper[4870]: I1014 09:00:03.861878 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-fbdnv"] Oct 14 09:00:03 crc kubenswrapper[4870]: I1014 09:00:03.898628 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29340540-jw8p5" event={"ID":"4c777f9d-8c71-4d17-ac52-ae341bd066a2","Type":"ContainerDied","Data":"2755740d7ce689c73b7bbb1b15bac463b790ae4fc48132c3e9331d0be2e7feb1"} Oct 14 09:00:03 crc kubenswrapper[4870]: I1014 09:00:03.898670 4870 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2755740d7ce689c73b7bbb1b15bac463b790ae4fc48132c3e9331d0be2e7feb1" Oct 14 09:00:03 crc kubenswrapper[4870]: I1014 09:00:03.898675 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29340540-jw8p5" Oct 14 09:00:03 crc kubenswrapper[4870]: I1014 09:00:03.987324 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2447a450-aee2-4e46-9bac-fa754b69ecaf-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-fbdnv\" (UID: \"2447a450-aee2-4e46-9bac-fa754b69ecaf\") " pod="openstack/nova-cell1-cell-mapping-fbdnv" Oct 14 09:00:03 crc kubenswrapper[4870]: I1014 09:00:03.987392 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jwscp\" (UniqueName: \"kubernetes.io/projected/2447a450-aee2-4e46-9bac-fa754b69ecaf-kube-api-access-jwscp\") pod \"nova-cell1-cell-mapping-fbdnv\" (UID: \"2447a450-aee2-4e46-9bac-fa754b69ecaf\") " pod="openstack/nova-cell1-cell-mapping-fbdnv" Oct 14 09:00:03 crc kubenswrapper[4870]: I1014 09:00:03.987625 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2447a450-aee2-4e46-9bac-fa754b69ecaf-scripts\") pod \"nova-cell1-cell-mapping-fbdnv\" (UID: \"2447a450-aee2-4e46-9bac-fa754b69ecaf\") " pod="openstack/nova-cell1-cell-mapping-fbdnv" Oct 14 09:00:03 crc kubenswrapper[4870]: I1014 09:00:03.987709 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2447a450-aee2-4e46-9bac-fa754b69ecaf-config-data\") pod \"nova-cell1-cell-mapping-fbdnv\" (UID: \"2447a450-aee2-4e46-9bac-fa754b69ecaf\") " pod="openstack/nova-cell1-cell-mapping-fbdnv" Oct 14 09:00:04 crc kubenswrapper[4870]: I1014 09:00:04.089412 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2447a450-aee2-4e46-9bac-fa754b69ecaf-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-fbdnv\" (UID: \"2447a450-aee2-4e46-9bac-fa754b69ecaf\") " pod="openstack/nova-cell1-cell-mapping-fbdnv" Oct 14 09:00:04 crc kubenswrapper[4870]: I1014 09:00:04.089497 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jwscp\" (UniqueName: \"kubernetes.io/projected/2447a450-aee2-4e46-9bac-fa754b69ecaf-kube-api-access-jwscp\") pod \"nova-cell1-cell-mapping-fbdnv\" (UID: \"2447a450-aee2-4e46-9bac-fa754b69ecaf\") " pod="openstack/nova-cell1-cell-mapping-fbdnv" Oct 14 09:00:04 crc kubenswrapper[4870]: I1014 09:00:04.089593 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2447a450-aee2-4e46-9bac-fa754b69ecaf-scripts\") pod \"nova-cell1-cell-mapping-fbdnv\" (UID: \"2447a450-aee2-4e46-9bac-fa754b69ecaf\") " pod="openstack/nova-cell1-cell-mapping-fbdnv" Oct 14 09:00:04 crc kubenswrapper[4870]: I1014 09:00:04.089647 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2447a450-aee2-4e46-9bac-fa754b69ecaf-config-data\") pod \"nova-cell1-cell-mapping-fbdnv\" (UID: \"2447a450-aee2-4e46-9bac-fa754b69ecaf\") " pod="openstack/nova-cell1-cell-mapping-fbdnv" Oct 14 09:00:04 crc kubenswrapper[4870]: I1014 09:00:04.094342 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2447a450-aee2-4e46-9bac-fa754b69ecaf-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-fbdnv\" (UID: \"2447a450-aee2-4e46-9bac-fa754b69ecaf\") " pod="openstack/nova-cell1-cell-mapping-fbdnv" Oct 14 09:00:04 crc kubenswrapper[4870]: I1014 09:00:04.095021 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2447a450-aee2-4e46-9bac-fa754b69ecaf-config-data\") pod \"nova-cell1-cell-mapping-fbdnv\" (UID: \"2447a450-aee2-4e46-9bac-fa754b69ecaf\") " pod="openstack/nova-cell1-cell-mapping-fbdnv" Oct 14 09:00:04 crc kubenswrapper[4870]: I1014 09:00:04.099023 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2447a450-aee2-4e46-9bac-fa754b69ecaf-scripts\") pod \"nova-cell1-cell-mapping-fbdnv\" (UID: \"2447a450-aee2-4e46-9bac-fa754b69ecaf\") " pod="openstack/nova-cell1-cell-mapping-fbdnv" Oct 14 09:00:04 crc kubenswrapper[4870]: I1014 09:00:04.110937 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jwscp\" (UniqueName: \"kubernetes.io/projected/2447a450-aee2-4e46-9bac-fa754b69ecaf-kube-api-access-jwscp\") pod \"nova-cell1-cell-mapping-fbdnv\" (UID: \"2447a450-aee2-4e46-9bac-fa754b69ecaf\") " pod="openstack/nova-cell1-cell-mapping-fbdnv" Oct 14 09:00:04 crc kubenswrapper[4870]: I1014 09:00:04.169185 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-fbdnv" Oct 14 09:00:04 crc kubenswrapper[4870]: I1014 09:00:04.214651 4870 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="fd2f3bc5-f9f6-434c-bf5a-364b7d8a05b9" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"http://10.217.1.76:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 14 09:00:04 crc kubenswrapper[4870]: I1014 09:00:04.214741 4870 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="fd2f3bc5-f9f6-434c-bf5a-364b7d8a05b9" containerName="nova-metadata-log" probeResult="failure" output="Get \"http://10.217.1.76:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 14 09:00:04 crc kubenswrapper[4870]: I1014 09:00:04.298080 4870 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="e7ffeeb3-4944-42e5-8789-a4922ab6e2c8" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.1.77:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 14 09:00:04 crc kubenswrapper[4870]: I1014 09:00:04.298085 4870 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="e7ffeeb3-4944-42e5-8789-a4922ab6e2c8" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.1.77:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 14 09:00:04 crc kubenswrapper[4870]: I1014 09:00:04.298406 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29340495-ldzr5"] Oct 14 09:00:04 crc kubenswrapper[4870]: I1014 09:00:04.306998 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29340495-ldzr5"] Oct 14 09:00:04 crc kubenswrapper[4870]: I1014 09:00:04.507422 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-fbdnv"] Oct 14 09:00:04 crc kubenswrapper[4870]: I1014 09:00:04.907383 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-fbdnv" event={"ID":"2447a450-aee2-4e46-9bac-fa754b69ecaf","Type":"ContainerStarted","Data":"ffcade20088eb5ff83f1f617deef2c3d30eced0aac5f3c80b87cd3a1131acbae"} Oct 14 09:00:04 crc kubenswrapper[4870]: I1014 09:00:04.907467 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-fbdnv" event={"ID":"2447a450-aee2-4e46-9bac-fa754b69ecaf","Type":"ContainerStarted","Data":"fd161c77e336d8392328dc9d1b503086ff67b4cc22a2428fa88d7bb17a4ffe98"} Oct 14 09:00:04 crc kubenswrapper[4870]: I1014 09:00:04.950822 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-fbdnv" podStartSLOduration=1.950791604 podStartE2EDuration="1.950791604s" podCreationTimestamp="2025-10-14 09:00:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 09:00:04.934372299 +0000 UTC m=+7140.631732670" watchObservedRunningTime="2025-10-14 09:00:04.950791604 +0000 UTC m=+7140.648151975" Oct 14 09:00:05 crc kubenswrapper[4870]: I1014 09:00:05.089333 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="26e2587a-73e2-48c4-9330-09d968b5a9c4" path="/var/lib/kubelet/pods/26e2587a-73e2-48c4-9330-09d968b5a9c4/volumes" Oct 14 09:00:06 crc kubenswrapper[4870]: I1014 09:00:06.171127 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 14 09:00:06 crc kubenswrapper[4870]: I1014 09:00:06.198243 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 14 09:00:06 crc kubenswrapper[4870]: I1014 09:00:06.954259 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 14 09:00:09 crc kubenswrapper[4870]: I1014 09:00:09.969978 4870 generic.go:334] "Generic (PLEG): container finished" podID="2447a450-aee2-4e46-9bac-fa754b69ecaf" containerID="ffcade20088eb5ff83f1f617deef2c3d30eced0aac5f3c80b87cd3a1131acbae" exitCode=0 Oct 14 09:00:09 crc kubenswrapper[4870]: I1014 09:00:09.970075 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-fbdnv" event={"ID":"2447a450-aee2-4e46-9bac-fa754b69ecaf","Type":"ContainerDied","Data":"ffcade20088eb5ff83f1f617deef2c3d30eced0aac5f3c80b87cd3a1131acbae"} Oct 14 09:00:11 crc kubenswrapper[4870]: I1014 09:00:11.409815 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-fbdnv" Oct 14 09:00:11 crc kubenswrapper[4870]: I1014 09:00:11.550223 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2447a450-aee2-4e46-9bac-fa754b69ecaf-scripts\") pod \"2447a450-aee2-4e46-9bac-fa754b69ecaf\" (UID: \"2447a450-aee2-4e46-9bac-fa754b69ecaf\") " Oct 14 09:00:11 crc kubenswrapper[4870]: I1014 09:00:11.550278 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jwscp\" (UniqueName: \"kubernetes.io/projected/2447a450-aee2-4e46-9bac-fa754b69ecaf-kube-api-access-jwscp\") pod \"2447a450-aee2-4e46-9bac-fa754b69ecaf\" (UID: \"2447a450-aee2-4e46-9bac-fa754b69ecaf\") " Oct 14 09:00:11 crc kubenswrapper[4870]: I1014 09:00:11.550343 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2447a450-aee2-4e46-9bac-fa754b69ecaf-combined-ca-bundle\") pod \"2447a450-aee2-4e46-9bac-fa754b69ecaf\" (UID: \"2447a450-aee2-4e46-9bac-fa754b69ecaf\") " Oct 14 09:00:11 crc kubenswrapper[4870]: I1014 09:00:11.550474 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2447a450-aee2-4e46-9bac-fa754b69ecaf-config-data\") pod \"2447a450-aee2-4e46-9bac-fa754b69ecaf\" (UID: \"2447a450-aee2-4e46-9bac-fa754b69ecaf\") " Oct 14 09:00:11 crc kubenswrapper[4870]: I1014 09:00:11.556547 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2447a450-aee2-4e46-9bac-fa754b69ecaf-scripts" (OuterVolumeSpecName: "scripts") pod "2447a450-aee2-4e46-9bac-fa754b69ecaf" (UID: "2447a450-aee2-4e46-9bac-fa754b69ecaf"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:00:11 crc kubenswrapper[4870]: I1014 09:00:11.556555 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2447a450-aee2-4e46-9bac-fa754b69ecaf-kube-api-access-jwscp" (OuterVolumeSpecName: "kube-api-access-jwscp") pod "2447a450-aee2-4e46-9bac-fa754b69ecaf" (UID: "2447a450-aee2-4e46-9bac-fa754b69ecaf"). InnerVolumeSpecName "kube-api-access-jwscp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:00:11 crc kubenswrapper[4870]: I1014 09:00:11.585760 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2447a450-aee2-4e46-9bac-fa754b69ecaf-config-data" (OuterVolumeSpecName: "config-data") pod "2447a450-aee2-4e46-9bac-fa754b69ecaf" (UID: "2447a450-aee2-4e46-9bac-fa754b69ecaf"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:00:11 crc kubenswrapper[4870]: I1014 09:00:11.587225 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2447a450-aee2-4e46-9bac-fa754b69ecaf-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2447a450-aee2-4e46-9bac-fa754b69ecaf" (UID: "2447a450-aee2-4e46-9bac-fa754b69ecaf"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:00:11 crc kubenswrapper[4870]: I1014 09:00:11.652544 4870 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2447a450-aee2-4e46-9bac-fa754b69ecaf-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 09:00:11 crc kubenswrapper[4870]: I1014 09:00:11.652576 4870 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2447a450-aee2-4e46-9bac-fa754b69ecaf-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 09:00:11 crc kubenswrapper[4870]: I1014 09:00:11.652589 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jwscp\" (UniqueName: \"kubernetes.io/projected/2447a450-aee2-4e46-9bac-fa754b69ecaf-kube-api-access-jwscp\") on node \"crc\" DevicePath \"\"" Oct 14 09:00:11 crc kubenswrapper[4870]: I1014 09:00:11.652600 4870 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2447a450-aee2-4e46-9bac-fa754b69ecaf-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 09:00:12 crc kubenswrapper[4870]: I1014 09:00:12.001629 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-fbdnv" event={"ID":"2447a450-aee2-4e46-9bac-fa754b69ecaf","Type":"ContainerDied","Data":"fd161c77e336d8392328dc9d1b503086ff67b4cc22a2428fa88d7bb17a4ffe98"} Oct 14 09:00:12 crc kubenswrapper[4870]: I1014 09:00:12.001674 4870 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fd161c77e336d8392328dc9d1b503086ff67b4cc22a2428fa88d7bb17a4ffe98" Oct 14 09:00:12 crc kubenswrapper[4870]: I1014 09:00:12.001777 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-fbdnv" Oct 14 09:00:12 crc kubenswrapper[4870]: I1014 09:00:12.177291 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 14 09:00:12 crc kubenswrapper[4870]: I1014 09:00:12.178072 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="e7ffeeb3-4944-42e5-8789-a4922ab6e2c8" containerName="nova-api-log" containerID="cri-o://21712480245d9483d0084dd3bac8b74bd6478ff05aefecc24a739a42f104e3b0" gracePeriod=30 Oct 14 09:00:12 crc kubenswrapper[4870]: I1014 09:00:12.178209 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="e7ffeeb3-4944-42e5-8789-a4922ab6e2c8" containerName="nova-api-api" containerID="cri-o://fe1fbc1e1ac5645d510a2c4bd1d21d900c44d1f5c7e01e6926224a3ef748981b" gracePeriod=30 Oct 14 09:00:12 crc kubenswrapper[4870]: I1014 09:00:12.197412 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 14 09:00:12 crc kubenswrapper[4870]: I1014 09:00:12.201380 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="77e2dbb6-3f73-4617-994b-cacf4a839808" containerName="nova-scheduler-scheduler" containerID="cri-o://eb61e04aa194a1edd6e70017933438f95a86807b7f7bce94ffaf66a42d4a8799" gracePeriod=30 Oct 14 09:00:12 crc kubenswrapper[4870]: I1014 09:00:12.209507 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 14 09:00:12 crc kubenswrapper[4870]: I1014 09:00:12.209719 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="fd2f3bc5-f9f6-434c-bf5a-364b7d8a05b9" containerName="nova-metadata-log" containerID="cri-o://87f9c0df03f215e3cf69750e3466233c1f0f00ec14013ce4ec59d7f6ba7044d6" gracePeriod=30 Oct 14 09:00:12 crc kubenswrapper[4870]: I1014 09:00:12.209839 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="fd2f3bc5-f9f6-434c-bf5a-364b7d8a05b9" containerName="nova-metadata-metadata" containerID="cri-o://2a61496f82dca223ef233b4d38c74d2e577859ab091f9959e058e706b27e7029" gracePeriod=30 Oct 14 09:00:13 crc kubenswrapper[4870]: I1014 09:00:13.014104 4870 generic.go:334] "Generic (PLEG): container finished" podID="fd2f3bc5-f9f6-434c-bf5a-364b7d8a05b9" containerID="87f9c0df03f215e3cf69750e3466233c1f0f00ec14013ce4ec59d7f6ba7044d6" exitCode=143 Oct 14 09:00:13 crc kubenswrapper[4870]: I1014 09:00:13.014205 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"fd2f3bc5-f9f6-434c-bf5a-364b7d8a05b9","Type":"ContainerDied","Data":"87f9c0df03f215e3cf69750e3466233c1f0f00ec14013ce4ec59d7f6ba7044d6"} Oct 14 09:00:13 crc kubenswrapper[4870]: I1014 09:00:13.015932 4870 generic.go:334] "Generic (PLEG): container finished" podID="e7ffeeb3-4944-42e5-8789-a4922ab6e2c8" containerID="21712480245d9483d0084dd3bac8b74bd6478ff05aefecc24a739a42f104e3b0" exitCode=143 Oct 14 09:00:13 crc kubenswrapper[4870]: I1014 09:00:13.015969 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e7ffeeb3-4944-42e5-8789-a4922ab6e2c8","Type":"ContainerDied","Data":"21712480245d9483d0084dd3bac8b74bd6478ff05aefecc24a739a42f104e3b0"} Oct 14 09:00:13 crc kubenswrapper[4870]: I1014 09:00:13.371425 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 14 09:00:13 crc kubenswrapper[4870]: I1014 09:00:13.480134 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/77e2dbb6-3f73-4617-994b-cacf4a839808-config-data\") pod \"77e2dbb6-3f73-4617-994b-cacf4a839808\" (UID: \"77e2dbb6-3f73-4617-994b-cacf4a839808\") " Oct 14 09:00:13 crc kubenswrapper[4870]: I1014 09:00:13.480177 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77e2dbb6-3f73-4617-994b-cacf4a839808-combined-ca-bundle\") pod \"77e2dbb6-3f73-4617-994b-cacf4a839808\" (UID: \"77e2dbb6-3f73-4617-994b-cacf4a839808\") " Oct 14 09:00:13 crc kubenswrapper[4870]: I1014 09:00:13.480274 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wzdhh\" (UniqueName: \"kubernetes.io/projected/77e2dbb6-3f73-4617-994b-cacf4a839808-kube-api-access-wzdhh\") pod \"77e2dbb6-3f73-4617-994b-cacf4a839808\" (UID: \"77e2dbb6-3f73-4617-994b-cacf4a839808\") " Oct 14 09:00:13 crc kubenswrapper[4870]: I1014 09:00:13.488910 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/77e2dbb6-3f73-4617-994b-cacf4a839808-kube-api-access-wzdhh" (OuterVolumeSpecName: "kube-api-access-wzdhh") pod "77e2dbb6-3f73-4617-994b-cacf4a839808" (UID: "77e2dbb6-3f73-4617-994b-cacf4a839808"). InnerVolumeSpecName "kube-api-access-wzdhh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:00:13 crc kubenswrapper[4870]: I1014 09:00:13.508020 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/77e2dbb6-3f73-4617-994b-cacf4a839808-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "77e2dbb6-3f73-4617-994b-cacf4a839808" (UID: "77e2dbb6-3f73-4617-994b-cacf4a839808"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:00:13 crc kubenswrapper[4870]: I1014 09:00:13.510129 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/77e2dbb6-3f73-4617-994b-cacf4a839808-config-data" (OuterVolumeSpecName: "config-data") pod "77e2dbb6-3f73-4617-994b-cacf4a839808" (UID: "77e2dbb6-3f73-4617-994b-cacf4a839808"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:00:13 crc kubenswrapper[4870]: I1014 09:00:13.581752 4870 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77e2dbb6-3f73-4617-994b-cacf4a839808-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 09:00:13 crc kubenswrapper[4870]: I1014 09:00:13.581788 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wzdhh\" (UniqueName: \"kubernetes.io/projected/77e2dbb6-3f73-4617-994b-cacf4a839808-kube-api-access-wzdhh\") on node \"crc\" DevicePath \"\"" Oct 14 09:00:13 crc kubenswrapper[4870]: I1014 09:00:13.581799 4870 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/77e2dbb6-3f73-4617-994b-cacf4a839808-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 09:00:14 crc kubenswrapper[4870]: I1014 09:00:14.027663 4870 generic.go:334] "Generic (PLEG): container finished" podID="77e2dbb6-3f73-4617-994b-cacf4a839808" containerID="eb61e04aa194a1edd6e70017933438f95a86807b7f7bce94ffaf66a42d4a8799" exitCode=0 Oct 14 09:00:14 crc kubenswrapper[4870]: I1014 09:00:14.027709 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 14 09:00:14 crc kubenswrapper[4870]: I1014 09:00:14.027727 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"77e2dbb6-3f73-4617-994b-cacf4a839808","Type":"ContainerDied","Data":"eb61e04aa194a1edd6e70017933438f95a86807b7f7bce94ffaf66a42d4a8799"} Oct 14 09:00:14 crc kubenswrapper[4870]: I1014 09:00:14.027758 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"77e2dbb6-3f73-4617-994b-cacf4a839808","Type":"ContainerDied","Data":"6514c581a068725908b3af7bcdcb1553f42fcec2956cfaaa8b54045dc4c231e2"} Oct 14 09:00:14 crc kubenswrapper[4870]: I1014 09:00:14.027780 4870 scope.go:117] "RemoveContainer" containerID="eb61e04aa194a1edd6e70017933438f95a86807b7f7bce94ffaf66a42d4a8799" Oct 14 09:00:14 crc kubenswrapper[4870]: I1014 09:00:14.061196 4870 scope.go:117] "RemoveContainer" containerID="eb61e04aa194a1edd6e70017933438f95a86807b7f7bce94ffaf66a42d4a8799" Oct 14 09:00:14 crc kubenswrapper[4870]: E1014 09:00:14.066177 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eb61e04aa194a1edd6e70017933438f95a86807b7f7bce94ffaf66a42d4a8799\": container with ID starting with eb61e04aa194a1edd6e70017933438f95a86807b7f7bce94ffaf66a42d4a8799 not found: ID does not exist" containerID="eb61e04aa194a1edd6e70017933438f95a86807b7f7bce94ffaf66a42d4a8799" Oct 14 09:00:14 crc kubenswrapper[4870]: I1014 09:00:14.066220 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eb61e04aa194a1edd6e70017933438f95a86807b7f7bce94ffaf66a42d4a8799"} err="failed to get container status \"eb61e04aa194a1edd6e70017933438f95a86807b7f7bce94ffaf66a42d4a8799\": rpc error: code = NotFound desc = could not find container \"eb61e04aa194a1edd6e70017933438f95a86807b7f7bce94ffaf66a42d4a8799\": container with ID starting with eb61e04aa194a1edd6e70017933438f95a86807b7f7bce94ffaf66a42d4a8799 not found: ID does not exist" Oct 14 09:00:14 crc kubenswrapper[4870]: I1014 09:00:14.074075 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 14 09:00:14 crc kubenswrapper[4870]: I1014 09:00:14.081912 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Oct 14 09:00:14 crc kubenswrapper[4870]: I1014 09:00:14.100719 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 14 09:00:14 crc kubenswrapper[4870]: E1014 09:00:14.101209 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2447a450-aee2-4e46-9bac-fa754b69ecaf" containerName="nova-manage" Oct 14 09:00:14 crc kubenswrapper[4870]: I1014 09:00:14.101230 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="2447a450-aee2-4e46-9bac-fa754b69ecaf" containerName="nova-manage" Oct 14 09:00:14 crc kubenswrapper[4870]: E1014 09:00:14.101245 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="77e2dbb6-3f73-4617-994b-cacf4a839808" containerName="nova-scheduler-scheduler" Oct 14 09:00:14 crc kubenswrapper[4870]: I1014 09:00:14.101254 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="77e2dbb6-3f73-4617-994b-cacf4a839808" containerName="nova-scheduler-scheduler" Oct 14 09:00:14 crc kubenswrapper[4870]: I1014 09:00:14.101551 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="77e2dbb6-3f73-4617-994b-cacf4a839808" containerName="nova-scheduler-scheduler" Oct 14 09:00:14 crc kubenswrapper[4870]: I1014 09:00:14.101607 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="2447a450-aee2-4e46-9bac-fa754b69ecaf" containerName="nova-manage" Oct 14 09:00:14 crc kubenswrapper[4870]: I1014 09:00:14.102456 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 14 09:00:14 crc kubenswrapper[4870]: I1014 09:00:14.108708 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 14 09:00:14 crc kubenswrapper[4870]: I1014 09:00:14.112702 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 14 09:00:14 crc kubenswrapper[4870]: I1014 09:00:14.192924 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-942pr\" (UniqueName: \"kubernetes.io/projected/4eabbefe-21bd-4eba-aab2-5f9123db96fa-kube-api-access-942pr\") pod \"nova-scheduler-0\" (UID: \"4eabbefe-21bd-4eba-aab2-5f9123db96fa\") " pod="openstack/nova-scheduler-0" Oct 14 09:00:14 crc kubenswrapper[4870]: I1014 09:00:14.193384 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4eabbefe-21bd-4eba-aab2-5f9123db96fa-config-data\") pod \"nova-scheduler-0\" (UID: \"4eabbefe-21bd-4eba-aab2-5f9123db96fa\") " pod="openstack/nova-scheduler-0" Oct 14 09:00:14 crc kubenswrapper[4870]: I1014 09:00:14.193559 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4eabbefe-21bd-4eba-aab2-5f9123db96fa-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"4eabbefe-21bd-4eba-aab2-5f9123db96fa\") " pod="openstack/nova-scheduler-0" Oct 14 09:00:14 crc kubenswrapper[4870]: I1014 09:00:14.295770 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4eabbefe-21bd-4eba-aab2-5f9123db96fa-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"4eabbefe-21bd-4eba-aab2-5f9123db96fa\") " pod="openstack/nova-scheduler-0" Oct 14 09:00:14 crc kubenswrapper[4870]: I1014 09:00:14.295897 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-942pr\" (UniqueName: \"kubernetes.io/projected/4eabbefe-21bd-4eba-aab2-5f9123db96fa-kube-api-access-942pr\") pod \"nova-scheduler-0\" (UID: \"4eabbefe-21bd-4eba-aab2-5f9123db96fa\") " pod="openstack/nova-scheduler-0" Oct 14 09:00:14 crc kubenswrapper[4870]: I1014 09:00:14.296055 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4eabbefe-21bd-4eba-aab2-5f9123db96fa-config-data\") pod \"nova-scheduler-0\" (UID: \"4eabbefe-21bd-4eba-aab2-5f9123db96fa\") " pod="openstack/nova-scheduler-0" Oct 14 09:00:14 crc kubenswrapper[4870]: I1014 09:00:14.302670 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4eabbefe-21bd-4eba-aab2-5f9123db96fa-config-data\") pod \"nova-scheduler-0\" (UID: \"4eabbefe-21bd-4eba-aab2-5f9123db96fa\") " pod="openstack/nova-scheduler-0" Oct 14 09:00:14 crc kubenswrapper[4870]: I1014 09:00:14.302857 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4eabbefe-21bd-4eba-aab2-5f9123db96fa-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"4eabbefe-21bd-4eba-aab2-5f9123db96fa\") " pod="openstack/nova-scheduler-0" Oct 14 09:00:14 crc kubenswrapper[4870]: I1014 09:00:14.321555 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-942pr\" (UniqueName: \"kubernetes.io/projected/4eabbefe-21bd-4eba-aab2-5f9123db96fa-kube-api-access-942pr\") pod \"nova-scheduler-0\" (UID: \"4eabbefe-21bd-4eba-aab2-5f9123db96fa\") " pod="openstack/nova-scheduler-0" Oct 14 09:00:14 crc kubenswrapper[4870]: I1014 09:00:14.424753 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 14 09:00:14 crc kubenswrapper[4870]: I1014 09:00:14.897744 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 14 09:00:14 crc kubenswrapper[4870]: W1014 09:00:14.900642 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4eabbefe_21bd_4eba_aab2_5f9123db96fa.slice/crio-a6988420bd4775a3a0b784381d4e6615df53417a5438e804260a50423cabc929 WatchSource:0}: Error finding container a6988420bd4775a3a0b784381d4e6615df53417a5438e804260a50423cabc929: Status 404 returned error can't find the container with id a6988420bd4775a3a0b784381d4e6615df53417a5438e804260a50423cabc929 Oct 14 09:00:15 crc kubenswrapper[4870]: I1014 09:00:15.046006 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="77e2dbb6-3f73-4617-994b-cacf4a839808" path="/var/lib/kubelet/pods/77e2dbb6-3f73-4617-994b-cacf4a839808/volumes" Oct 14 09:00:15 crc kubenswrapper[4870]: I1014 09:00:15.048480 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"4eabbefe-21bd-4eba-aab2-5f9123db96fa","Type":"ContainerStarted","Data":"a6988420bd4775a3a0b784381d4e6615df53417a5438e804260a50423cabc929"} Oct 14 09:00:15 crc kubenswrapper[4870]: I1014 09:00:15.781863 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 14 09:00:15 crc kubenswrapper[4870]: I1014 09:00:15.826245 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e7ffeeb3-4944-42e5-8789-a4922ab6e2c8-config-data\") pod \"e7ffeeb3-4944-42e5-8789-a4922ab6e2c8\" (UID: \"e7ffeeb3-4944-42e5-8789-a4922ab6e2c8\") " Oct 14 09:00:15 crc kubenswrapper[4870]: I1014 09:00:15.826342 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7ffeeb3-4944-42e5-8789-a4922ab6e2c8-combined-ca-bundle\") pod \"e7ffeeb3-4944-42e5-8789-a4922ab6e2c8\" (UID: \"e7ffeeb3-4944-42e5-8789-a4922ab6e2c8\") " Oct 14 09:00:15 crc kubenswrapper[4870]: I1014 09:00:15.826502 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e7ffeeb3-4944-42e5-8789-a4922ab6e2c8-logs\") pod \"e7ffeeb3-4944-42e5-8789-a4922ab6e2c8\" (UID: \"e7ffeeb3-4944-42e5-8789-a4922ab6e2c8\") " Oct 14 09:00:15 crc kubenswrapper[4870]: I1014 09:00:15.826524 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7ddfw\" (UniqueName: \"kubernetes.io/projected/e7ffeeb3-4944-42e5-8789-a4922ab6e2c8-kube-api-access-7ddfw\") pod \"e7ffeeb3-4944-42e5-8789-a4922ab6e2c8\" (UID: \"e7ffeeb3-4944-42e5-8789-a4922ab6e2c8\") " Oct 14 09:00:15 crc kubenswrapper[4870]: I1014 09:00:15.837367 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7ffeeb3-4944-42e5-8789-a4922ab6e2c8-kube-api-access-7ddfw" (OuterVolumeSpecName: "kube-api-access-7ddfw") pod "e7ffeeb3-4944-42e5-8789-a4922ab6e2c8" (UID: "e7ffeeb3-4944-42e5-8789-a4922ab6e2c8"). InnerVolumeSpecName "kube-api-access-7ddfw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:00:15 crc kubenswrapper[4870]: I1014 09:00:15.837869 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e7ffeeb3-4944-42e5-8789-a4922ab6e2c8-logs" (OuterVolumeSpecName: "logs") pod "e7ffeeb3-4944-42e5-8789-a4922ab6e2c8" (UID: "e7ffeeb3-4944-42e5-8789-a4922ab6e2c8"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 09:00:15 crc kubenswrapper[4870]: I1014 09:00:15.866200 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7ffeeb3-4944-42e5-8789-a4922ab6e2c8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e7ffeeb3-4944-42e5-8789-a4922ab6e2c8" (UID: "e7ffeeb3-4944-42e5-8789-a4922ab6e2c8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:00:15 crc kubenswrapper[4870]: I1014 09:00:15.866882 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7ffeeb3-4944-42e5-8789-a4922ab6e2c8-config-data" (OuterVolumeSpecName: "config-data") pod "e7ffeeb3-4944-42e5-8789-a4922ab6e2c8" (UID: "e7ffeeb3-4944-42e5-8789-a4922ab6e2c8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:00:15 crc kubenswrapper[4870]: I1014 09:00:15.890130 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 14 09:00:15 crc kubenswrapper[4870]: I1014 09:00:15.928920 4870 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e7ffeeb3-4944-42e5-8789-a4922ab6e2c8-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 09:00:15 crc kubenswrapper[4870]: I1014 09:00:15.928959 4870 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7ffeeb3-4944-42e5-8789-a4922ab6e2c8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 09:00:15 crc kubenswrapper[4870]: I1014 09:00:15.928974 4870 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e7ffeeb3-4944-42e5-8789-a4922ab6e2c8-logs\") on node \"crc\" DevicePath \"\"" Oct 14 09:00:15 crc kubenswrapper[4870]: I1014 09:00:15.928988 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7ddfw\" (UniqueName: \"kubernetes.io/projected/e7ffeeb3-4944-42e5-8789-a4922ab6e2c8-kube-api-access-7ddfw\") on node \"crc\" DevicePath \"\"" Oct 14 09:00:16 crc kubenswrapper[4870]: I1014 09:00:16.030351 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fd2f3bc5-f9f6-434c-bf5a-364b7d8a05b9-config-data\") pod \"fd2f3bc5-f9f6-434c-bf5a-364b7d8a05b9\" (UID: \"fd2f3bc5-f9f6-434c-bf5a-364b7d8a05b9\") " Oct 14 09:00:16 crc kubenswrapper[4870]: I1014 09:00:16.030459 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd2f3bc5-f9f6-434c-bf5a-364b7d8a05b9-combined-ca-bundle\") pod \"fd2f3bc5-f9f6-434c-bf5a-364b7d8a05b9\" (UID: \"fd2f3bc5-f9f6-434c-bf5a-364b7d8a05b9\") " Oct 14 09:00:16 crc kubenswrapper[4870]: I1014 09:00:16.030648 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cn8n9\" (UniqueName: \"kubernetes.io/projected/fd2f3bc5-f9f6-434c-bf5a-364b7d8a05b9-kube-api-access-cn8n9\") pod \"fd2f3bc5-f9f6-434c-bf5a-364b7d8a05b9\" (UID: \"fd2f3bc5-f9f6-434c-bf5a-364b7d8a05b9\") " Oct 14 09:00:16 crc kubenswrapper[4870]: I1014 09:00:16.030723 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fd2f3bc5-f9f6-434c-bf5a-364b7d8a05b9-logs\") pod \"fd2f3bc5-f9f6-434c-bf5a-364b7d8a05b9\" (UID: \"fd2f3bc5-f9f6-434c-bf5a-364b7d8a05b9\") " Oct 14 09:00:16 crc kubenswrapper[4870]: I1014 09:00:16.031323 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fd2f3bc5-f9f6-434c-bf5a-364b7d8a05b9-logs" (OuterVolumeSpecName: "logs") pod "fd2f3bc5-f9f6-434c-bf5a-364b7d8a05b9" (UID: "fd2f3bc5-f9f6-434c-bf5a-364b7d8a05b9"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 09:00:16 crc kubenswrapper[4870]: I1014 09:00:16.034634 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fd2f3bc5-f9f6-434c-bf5a-364b7d8a05b9-kube-api-access-cn8n9" (OuterVolumeSpecName: "kube-api-access-cn8n9") pod "fd2f3bc5-f9f6-434c-bf5a-364b7d8a05b9" (UID: "fd2f3bc5-f9f6-434c-bf5a-364b7d8a05b9"). InnerVolumeSpecName "kube-api-access-cn8n9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:00:16 crc kubenswrapper[4870]: I1014 09:00:16.050965 4870 generic.go:334] "Generic (PLEG): container finished" podID="fd2f3bc5-f9f6-434c-bf5a-364b7d8a05b9" containerID="2a61496f82dca223ef233b4d38c74d2e577859ab091f9959e058e706b27e7029" exitCode=0 Oct 14 09:00:16 crc kubenswrapper[4870]: I1014 09:00:16.051107 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 14 09:00:16 crc kubenswrapper[4870]: I1014 09:00:16.051143 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"fd2f3bc5-f9f6-434c-bf5a-364b7d8a05b9","Type":"ContainerDied","Data":"2a61496f82dca223ef233b4d38c74d2e577859ab091f9959e058e706b27e7029"} Oct 14 09:00:16 crc kubenswrapper[4870]: I1014 09:00:16.051680 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"fd2f3bc5-f9f6-434c-bf5a-364b7d8a05b9","Type":"ContainerDied","Data":"637cce04ef35ae06562b0ad8fc9768602a9d7d1f4b720c68246cf6173eb6f71b"} Oct 14 09:00:16 crc kubenswrapper[4870]: I1014 09:00:16.051725 4870 scope.go:117] "RemoveContainer" containerID="2a61496f82dca223ef233b4d38c74d2e577859ab091f9959e058e706b27e7029" Oct 14 09:00:16 crc kubenswrapper[4870]: I1014 09:00:16.053208 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fd2f3bc5-f9f6-434c-bf5a-364b7d8a05b9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fd2f3bc5-f9f6-434c-bf5a-364b7d8a05b9" (UID: "fd2f3bc5-f9f6-434c-bf5a-364b7d8a05b9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:00:16 crc kubenswrapper[4870]: I1014 09:00:16.054889 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fd2f3bc5-f9f6-434c-bf5a-364b7d8a05b9-config-data" (OuterVolumeSpecName: "config-data") pod "fd2f3bc5-f9f6-434c-bf5a-364b7d8a05b9" (UID: "fd2f3bc5-f9f6-434c-bf5a-364b7d8a05b9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:00:16 crc kubenswrapper[4870]: I1014 09:00:16.054930 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"4eabbefe-21bd-4eba-aab2-5f9123db96fa","Type":"ContainerStarted","Data":"f22756363c89da54b7a27a7e9a3a91c6a3191a7f2ac7cd6b298026c27f2df43d"} Oct 14 09:00:16 crc kubenswrapper[4870]: I1014 09:00:16.058805 4870 generic.go:334] "Generic (PLEG): container finished" podID="e7ffeeb3-4944-42e5-8789-a4922ab6e2c8" containerID="fe1fbc1e1ac5645d510a2c4bd1d21d900c44d1f5c7e01e6926224a3ef748981b" exitCode=0 Oct 14 09:00:16 crc kubenswrapper[4870]: I1014 09:00:16.058847 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e7ffeeb3-4944-42e5-8789-a4922ab6e2c8","Type":"ContainerDied","Data":"fe1fbc1e1ac5645d510a2c4bd1d21d900c44d1f5c7e01e6926224a3ef748981b"} Oct 14 09:00:16 crc kubenswrapper[4870]: I1014 09:00:16.058893 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e7ffeeb3-4944-42e5-8789-a4922ab6e2c8","Type":"ContainerDied","Data":"266d7eb76949e8fe2775fffc7aa4b782b73952f107e16bda5d19cc3e52a91aa8"} Oct 14 09:00:16 crc kubenswrapper[4870]: I1014 09:00:16.058855 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 14 09:00:16 crc kubenswrapper[4870]: I1014 09:00:16.079784 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.07976553 podStartE2EDuration="2.07976553s" podCreationTimestamp="2025-10-14 09:00:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 09:00:16.068908632 +0000 UTC m=+7151.766268993" watchObservedRunningTime="2025-10-14 09:00:16.07976553 +0000 UTC m=+7151.777125901" Oct 14 09:00:16 crc kubenswrapper[4870]: I1014 09:00:16.092859 4870 scope.go:117] "RemoveContainer" containerID="87f9c0df03f215e3cf69750e3466233c1f0f00ec14013ce4ec59d7f6ba7044d6" Oct 14 09:00:16 crc kubenswrapper[4870]: I1014 09:00:16.134160 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cn8n9\" (UniqueName: \"kubernetes.io/projected/fd2f3bc5-f9f6-434c-bf5a-364b7d8a05b9-kube-api-access-cn8n9\") on node \"crc\" DevicePath \"\"" Oct 14 09:00:16 crc kubenswrapper[4870]: I1014 09:00:16.134199 4870 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fd2f3bc5-f9f6-434c-bf5a-364b7d8a05b9-logs\") on node \"crc\" DevicePath \"\"" Oct 14 09:00:16 crc kubenswrapper[4870]: I1014 09:00:16.134209 4870 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fd2f3bc5-f9f6-434c-bf5a-364b7d8a05b9-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 09:00:16 crc kubenswrapper[4870]: I1014 09:00:16.134224 4870 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd2f3bc5-f9f6-434c-bf5a-364b7d8a05b9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 09:00:16 crc kubenswrapper[4870]: I1014 09:00:16.144462 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 14 09:00:16 crc kubenswrapper[4870]: I1014 09:00:16.154809 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 14 09:00:16 crc kubenswrapper[4870]: I1014 09:00:16.162468 4870 scope.go:117] "RemoveContainer" containerID="2a61496f82dca223ef233b4d38c74d2e577859ab091f9959e058e706b27e7029" Oct 14 09:00:16 crc kubenswrapper[4870]: E1014 09:00:16.162847 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2a61496f82dca223ef233b4d38c74d2e577859ab091f9959e058e706b27e7029\": container with ID starting with 2a61496f82dca223ef233b4d38c74d2e577859ab091f9959e058e706b27e7029 not found: ID does not exist" containerID="2a61496f82dca223ef233b4d38c74d2e577859ab091f9959e058e706b27e7029" Oct 14 09:00:16 crc kubenswrapper[4870]: I1014 09:00:16.162897 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2a61496f82dca223ef233b4d38c74d2e577859ab091f9959e058e706b27e7029"} err="failed to get container status \"2a61496f82dca223ef233b4d38c74d2e577859ab091f9959e058e706b27e7029\": rpc error: code = NotFound desc = could not find container \"2a61496f82dca223ef233b4d38c74d2e577859ab091f9959e058e706b27e7029\": container with ID starting with 2a61496f82dca223ef233b4d38c74d2e577859ab091f9959e058e706b27e7029 not found: ID does not exist" Oct 14 09:00:16 crc kubenswrapper[4870]: I1014 09:00:16.162925 4870 scope.go:117] "RemoveContainer" containerID="87f9c0df03f215e3cf69750e3466233c1f0f00ec14013ce4ec59d7f6ba7044d6" Oct 14 09:00:16 crc kubenswrapper[4870]: E1014 09:00:16.163202 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"87f9c0df03f215e3cf69750e3466233c1f0f00ec14013ce4ec59d7f6ba7044d6\": container with ID starting with 87f9c0df03f215e3cf69750e3466233c1f0f00ec14013ce4ec59d7f6ba7044d6 not found: ID does not exist" containerID="87f9c0df03f215e3cf69750e3466233c1f0f00ec14013ce4ec59d7f6ba7044d6" Oct 14 09:00:16 crc kubenswrapper[4870]: I1014 09:00:16.163233 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"87f9c0df03f215e3cf69750e3466233c1f0f00ec14013ce4ec59d7f6ba7044d6"} err="failed to get container status \"87f9c0df03f215e3cf69750e3466233c1f0f00ec14013ce4ec59d7f6ba7044d6\": rpc error: code = NotFound desc = could not find container \"87f9c0df03f215e3cf69750e3466233c1f0f00ec14013ce4ec59d7f6ba7044d6\": container with ID starting with 87f9c0df03f215e3cf69750e3466233c1f0f00ec14013ce4ec59d7f6ba7044d6 not found: ID does not exist" Oct 14 09:00:16 crc kubenswrapper[4870]: I1014 09:00:16.163254 4870 scope.go:117] "RemoveContainer" containerID="fe1fbc1e1ac5645d510a2c4bd1d21d900c44d1f5c7e01e6926224a3ef748981b" Oct 14 09:00:16 crc kubenswrapper[4870]: I1014 09:00:16.163684 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 14 09:00:16 crc kubenswrapper[4870]: E1014 09:00:16.164058 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd2f3bc5-f9f6-434c-bf5a-364b7d8a05b9" containerName="nova-metadata-log" Oct 14 09:00:16 crc kubenswrapper[4870]: I1014 09:00:16.164079 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd2f3bc5-f9f6-434c-bf5a-364b7d8a05b9" containerName="nova-metadata-log" Oct 14 09:00:16 crc kubenswrapper[4870]: E1014 09:00:16.164102 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd2f3bc5-f9f6-434c-bf5a-364b7d8a05b9" containerName="nova-metadata-metadata" Oct 14 09:00:16 crc kubenswrapper[4870]: I1014 09:00:16.164111 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd2f3bc5-f9f6-434c-bf5a-364b7d8a05b9" containerName="nova-metadata-metadata" Oct 14 09:00:16 crc kubenswrapper[4870]: E1014 09:00:16.164127 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7ffeeb3-4944-42e5-8789-a4922ab6e2c8" containerName="nova-api-api" Oct 14 09:00:16 crc kubenswrapper[4870]: I1014 09:00:16.164135 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7ffeeb3-4944-42e5-8789-a4922ab6e2c8" containerName="nova-api-api" Oct 14 09:00:16 crc kubenswrapper[4870]: E1014 09:00:16.164169 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7ffeeb3-4944-42e5-8789-a4922ab6e2c8" containerName="nova-api-log" Oct 14 09:00:16 crc kubenswrapper[4870]: I1014 09:00:16.164176 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7ffeeb3-4944-42e5-8789-a4922ab6e2c8" containerName="nova-api-log" Oct 14 09:00:16 crc kubenswrapper[4870]: I1014 09:00:16.164380 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="e7ffeeb3-4944-42e5-8789-a4922ab6e2c8" containerName="nova-api-log" Oct 14 09:00:16 crc kubenswrapper[4870]: I1014 09:00:16.164400 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="fd2f3bc5-f9f6-434c-bf5a-364b7d8a05b9" containerName="nova-metadata-log" Oct 14 09:00:16 crc kubenswrapper[4870]: I1014 09:00:16.164411 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="fd2f3bc5-f9f6-434c-bf5a-364b7d8a05b9" containerName="nova-metadata-metadata" Oct 14 09:00:16 crc kubenswrapper[4870]: I1014 09:00:16.164423 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="e7ffeeb3-4944-42e5-8789-a4922ab6e2c8" containerName="nova-api-api" Oct 14 09:00:16 crc kubenswrapper[4870]: I1014 09:00:16.165577 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 14 09:00:16 crc kubenswrapper[4870]: I1014 09:00:16.167785 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 14 09:00:16 crc kubenswrapper[4870]: I1014 09:00:16.173135 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 14 09:00:16 crc kubenswrapper[4870]: I1014 09:00:16.189595 4870 scope.go:117] "RemoveContainer" containerID="21712480245d9483d0084dd3bac8b74bd6478ff05aefecc24a739a42f104e3b0" Oct 14 09:00:16 crc kubenswrapper[4870]: I1014 09:00:16.206497 4870 scope.go:117] "RemoveContainer" containerID="fe1fbc1e1ac5645d510a2c4bd1d21d900c44d1f5c7e01e6926224a3ef748981b" Oct 14 09:00:16 crc kubenswrapper[4870]: E1014 09:00:16.206915 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fe1fbc1e1ac5645d510a2c4bd1d21d900c44d1f5c7e01e6926224a3ef748981b\": container with ID starting with fe1fbc1e1ac5645d510a2c4bd1d21d900c44d1f5c7e01e6926224a3ef748981b not found: ID does not exist" containerID="fe1fbc1e1ac5645d510a2c4bd1d21d900c44d1f5c7e01e6926224a3ef748981b" Oct 14 09:00:16 crc kubenswrapper[4870]: I1014 09:00:16.206958 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fe1fbc1e1ac5645d510a2c4bd1d21d900c44d1f5c7e01e6926224a3ef748981b"} err="failed to get container status \"fe1fbc1e1ac5645d510a2c4bd1d21d900c44d1f5c7e01e6926224a3ef748981b\": rpc error: code = NotFound desc = could not find container \"fe1fbc1e1ac5645d510a2c4bd1d21d900c44d1f5c7e01e6926224a3ef748981b\": container with ID starting with fe1fbc1e1ac5645d510a2c4bd1d21d900c44d1f5c7e01e6926224a3ef748981b not found: ID does not exist" Oct 14 09:00:16 crc kubenswrapper[4870]: I1014 09:00:16.206985 4870 scope.go:117] "RemoveContainer" containerID="21712480245d9483d0084dd3bac8b74bd6478ff05aefecc24a739a42f104e3b0" Oct 14 09:00:16 crc kubenswrapper[4870]: E1014 09:00:16.207329 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"21712480245d9483d0084dd3bac8b74bd6478ff05aefecc24a739a42f104e3b0\": container with ID starting with 21712480245d9483d0084dd3bac8b74bd6478ff05aefecc24a739a42f104e3b0 not found: ID does not exist" containerID="21712480245d9483d0084dd3bac8b74bd6478ff05aefecc24a739a42f104e3b0" Oct 14 09:00:16 crc kubenswrapper[4870]: I1014 09:00:16.207375 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"21712480245d9483d0084dd3bac8b74bd6478ff05aefecc24a739a42f104e3b0"} err="failed to get container status \"21712480245d9483d0084dd3bac8b74bd6478ff05aefecc24a739a42f104e3b0\": rpc error: code = NotFound desc = could not find container \"21712480245d9483d0084dd3bac8b74bd6478ff05aefecc24a739a42f104e3b0\": container with ID starting with 21712480245d9483d0084dd3bac8b74bd6478ff05aefecc24a739a42f104e3b0 not found: ID does not exist" Oct 14 09:00:16 crc kubenswrapper[4870]: I1014 09:00:16.235991 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hg454\" (UniqueName: \"kubernetes.io/projected/d3db1cbf-b57a-4f03-bde9-25dc20acd3bb-kube-api-access-hg454\") pod \"nova-api-0\" (UID: \"d3db1cbf-b57a-4f03-bde9-25dc20acd3bb\") " pod="openstack/nova-api-0" Oct 14 09:00:16 crc kubenswrapper[4870]: I1014 09:00:16.236067 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3db1cbf-b57a-4f03-bde9-25dc20acd3bb-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"d3db1cbf-b57a-4f03-bde9-25dc20acd3bb\") " pod="openstack/nova-api-0" Oct 14 09:00:16 crc kubenswrapper[4870]: I1014 09:00:16.236258 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d3db1cbf-b57a-4f03-bde9-25dc20acd3bb-config-data\") pod \"nova-api-0\" (UID: \"d3db1cbf-b57a-4f03-bde9-25dc20acd3bb\") " pod="openstack/nova-api-0" Oct 14 09:00:16 crc kubenswrapper[4870]: I1014 09:00:16.236330 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d3db1cbf-b57a-4f03-bde9-25dc20acd3bb-logs\") pod \"nova-api-0\" (UID: \"d3db1cbf-b57a-4f03-bde9-25dc20acd3bb\") " pod="openstack/nova-api-0" Oct 14 09:00:16 crc kubenswrapper[4870]: I1014 09:00:16.338747 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hg454\" (UniqueName: \"kubernetes.io/projected/d3db1cbf-b57a-4f03-bde9-25dc20acd3bb-kube-api-access-hg454\") pod \"nova-api-0\" (UID: \"d3db1cbf-b57a-4f03-bde9-25dc20acd3bb\") " pod="openstack/nova-api-0" Oct 14 09:00:16 crc kubenswrapper[4870]: I1014 09:00:16.338885 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3db1cbf-b57a-4f03-bde9-25dc20acd3bb-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"d3db1cbf-b57a-4f03-bde9-25dc20acd3bb\") " pod="openstack/nova-api-0" Oct 14 09:00:16 crc kubenswrapper[4870]: I1014 09:00:16.338995 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d3db1cbf-b57a-4f03-bde9-25dc20acd3bb-config-data\") pod \"nova-api-0\" (UID: \"d3db1cbf-b57a-4f03-bde9-25dc20acd3bb\") " pod="openstack/nova-api-0" Oct 14 09:00:16 crc kubenswrapper[4870]: I1014 09:00:16.339038 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d3db1cbf-b57a-4f03-bde9-25dc20acd3bb-logs\") pod \"nova-api-0\" (UID: \"d3db1cbf-b57a-4f03-bde9-25dc20acd3bb\") " pod="openstack/nova-api-0" Oct 14 09:00:16 crc kubenswrapper[4870]: I1014 09:00:16.339739 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d3db1cbf-b57a-4f03-bde9-25dc20acd3bb-logs\") pod \"nova-api-0\" (UID: \"d3db1cbf-b57a-4f03-bde9-25dc20acd3bb\") " pod="openstack/nova-api-0" Oct 14 09:00:16 crc kubenswrapper[4870]: I1014 09:00:16.344492 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3db1cbf-b57a-4f03-bde9-25dc20acd3bb-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"d3db1cbf-b57a-4f03-bde9-25dc20acd3bb\") " pod="openstack/nova-api-0" Oct 14 09:00:16 crc kubenswrapper[4870]: I1014 09:00:16.345592 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d3db1cbf-b57a-4f03-bde9-25dc20acd3bb-config-data\") pod \"nova-api-0\" (UID: \"d3db1cbf-b57a-4f03-bde9-25dc20acd3bb\") " pod="openstack/nova-api-0" Oct 14 09:00:16 crc kubenswrapper[4870]: I1014 09:00:16.358513 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hg454\" (UniqueName: \"kubernetes.io/projected/d3db1cbf-b57a-4f03-bde9-25dc20acd3bb-kube-api-access-hg454\") pod \"nova-api-0\" (UID: \"d3db1cbf-b57a-4f03-bde9-25dc20acd3bb\") " pod="openstack/nova-api-0" Oct 14 09:00:16 crc kubenswrapper[4870]: I1014 09:00:16.416648 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 14 09:00:16 crc kubenswrapper[4870]: I1014 09:00:16.427614 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 14 09:00:16 crc kubenswrapper[4870]: I1014 09:00:16.451354 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 14 09:00:16 crc kubenswrapper[4870]: I1014 09:00:16.452973 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 14 09:00:16 crc kubenswrapper[4870]: I1014 09:00:16.457280 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 14 09:00:16 crc kubenswrapper[4870]: I1014 09:00:16.463836 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 14 09:00:16 crc kubenswrapper[4870]: I1014 09:00:16.487607 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 14 09:00:16 crc kubenswrapper[4870]: I1014 09:00:16.542067 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/337148f2-88d7-4694-947e-de1aa4375d33-logs\") pod \"nova-metadata-0\" (UID: \"337148f2-88d7-4694-947e-de1aa4375d33\") " pod="openstack/nova-metadata-0" Oct 14 09:00:16 crc kubenswrapper[4870]: I1014 09:00:16.542402 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/337148f2-88d7-4694-947e-de1aa4375d33-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"337148f2-88d7-4694-947e-de1aa4375d33\") " pod="openstack/nova-metadata-0" Oct 14 09:00:16 crc kubenswrapper[4870]: I1014 09:00:16.542583 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q2w2n\" (UniqueName: \"kubernetes.io/projected/337148f2-88d7-4694-947e-de1aa4375d33-kube-api-access-q2w2n\") pod \"nova-metadata-0\" (UID: \"337148f2-88d7-4694-947e-de1aa4375d33\") " pod="openstack/nova-metadata-0" Oct 14 09:00:16 crc kubenswrapper[4870]: I1014 09:00:16.542763 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/337148f2-88d7-4694-947e-de1aa4375d33-config-data\") pod \"nova-metadata-0\" (UID: \"337148f2-88d7-4694-947e-de1aa4375d33\") " pod="openstack/nova-metadata-0" Oct 14 09:00:16 crc kubenswrapper[4870]: I1014 09:00:16.644725 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/337148f2-88d7-4694-947e-de1aa4375d33-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"337148f2-88d7-4694-947e-de1aa4375d33\") " pod="openstack/nova-metadata-0" Oct 14 09:00:16 crc kubenswrapper[4870]: I1014 09:00:16.645140 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q2w2n\" (UniqueName: \"kubernetes.io/projected/337148f2-88d7-4694-947e-de1aa4375d33-kube-api-access-q2w2n\") pod \"nova-metadata-0\" (UID: \"337148f2-88d7-4694-947e-de1aa4375d33\") " pod="openstack/nova-metadata-0" Oct 14 09:00:16 crc kubenswrapper[4870]: I1014 09:00:16.645228 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/337148f2-88d7-4694-947e-de1aa4375d33-config-data\") pod \"nova-metadata-0\" (UID: \"337148f2-88d7-4694-947e-de1aa4375d33\") " pod="openstack/nova-metadata-0" Oct 14 09:00:16 crc kubenswrapper[4870]: I1014 09:00:16.645327 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/337148f2-88d7-4694-947e-de1aa4375d33-logs\") pod \"nova-metadata-0\" (UID: \"337148f2-88d7-4694-947e-de1aa4375d33\") " pod="openstack/nova-metadata-0" Oct 14 09:00:16 crc kubenswrapper[4870]: I1014 09:00:16.645704 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/337148f2-88d7-4694-947e-de1aa4375d33-logs\") pod \"nova-metadata-0\" (UID: \"337148f2-88d7-4694-947e-de1aa4375d33\") " pod="openstack/nova-metadata-0" Oct 14 09:00:16 crc kubenswrapper[4870]: I1014 09:00:16.655196 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/337148f2-88d7-4694-947e-de1aa4375d33-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"337148f2-88d7-4694-947e-de1aa4375d33\") " pod="openstack/nova-metadata-0" Oct 14 09:00:16 crc kubenswrapper[4870]: I1014 09:00:16.657089 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/337148f2-88d7-4694-947e-de1aa4375d33-config-data\") pod \"nova-metadata-0\" (UID: \"337148f2-88d7-4694-947e-de1aa4375d33\") " pod="openstack/nova-metadata-0" Oct 14 09:00:16 crc kubenswrapper[4870]: I1014 09:00:16.668308 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q2w2n\" (UniqueName: \"kubernetes.io/projected/337148f2-88d7-4694-947e-de1aa4375d33-kube-api-access-q2w2n\") pod \"nova-metadata-0\" (UID: \"337148f2-88d7-4694-947e-de1aa4375d33\") " pod="openstack/nova-metadata-0" Oct 14 09:00:16 crc kubenswrapper[4870]: I1014 09:00:16.773797 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 14 09:00:16 crc kubenswrapper[4870]: I1014 09:00:16.938504 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 14 09:00:17 crc kubenswrapper[4870]: I1014 09:00:17.047235 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7ffeeb3-4944-42e5-8789-a4922ab6e2c8" path="/var/lib/kubelet/pods/e7ffeeb3-4944-42e5-8789-a4922ab6e2c8/volumes" Oct 14 09:00:17 crc kubenswrapper[4870]: I1014 09:00:17.048008 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fd2f3bc5-f9f6-434c-bf5a-364b7d8a05b9" path="/var/lib/kubelet/pods/fd2f3bc5-f9f6-434c-bf5a-364b7d8a05b9/volumes" Oct 14 09:00:17 crc kubenswrapper[4870]: I1014 09:00:17.074061 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d3db1cbf-b57a-4f03-bde9-25dc20acd3bb","Type":"ContainerStarted","Data":"de798eeb6172573679a9b418bf5642ab3594f8a209529f091eff0cd0fcc6e30d"} Oct 14 09:00:17 crc kubenswrapper[4870]: I1014 09:00:17.199545 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 14 09:00:17 crc kubenswrapper[4870]: W1014 09:00:17.211387 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod337148f2_88d7_4694_947e_de1aa4375d33.slice/crio-048bd1f7b10b4aa1a4bb1b8923b6c77fd20c3f166aba1eda399448f63091c675 WatchSource:0}: Error finding container 048bd1f7b10b4aa1a4bb1b8923b6c77fd20c3f166aba1eda399448f63091c675: Status 404 returned error can't find the container with id 048bd1f7b10b4aa1a4bb1b8923b6c77fd20c3f166aba1eda399448f63091c675 Oct 14 09:00:18 crc kubenswrapper[4870]: I1014 09:00:18.091673 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"337148f2-88d7-4694-947e-de1aa4375d33","Type":"ContainerStarted","Data":"afb0153dee62d6c5510c9ee41556a4075a587da0d89753f84a48becc83a89047"} Oct 14 09:00:18 crc kubenswrapper[4870]: I1014 09:00:18.092514 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"337148f2-88d7-4694-947e-de1aa4375d33","Type":"ContainerStarted","Data":"5968c04b35057ebe36c338a812baba14f9fa27c198995f3c2f308c83f14b2950"} Oct 14 09:00:18 crc kubenswrapper[4870]: I1014 09:00:18.092550 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"337148f2-88d7-4694-947e-de1aa4375d33","Type":"ContainerStarted","Data":"048bd1f7b10b4aa1a4bb1b8923b6c77fd20c3f166aba1eda399448f63091c675"} Oct 14 09:00:18 crc kubenswrapper[4870]: I1014 09:00:18.095175 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d3db1cbf-b57a-4f03-bde9-25dc20acd3bb","Type":"ContainerStarted","Data":"dcba0cf698d44da1b6af963fe253650289ce35521a5481babbc0ed0eac4ac7d8"} Oct 14 09:00:18 crc kubenswrapper[4870]: I1014 09:00:18.095232 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d3db1cbf-b57a-4f03-bde9-25dc20acd3bb","Type":"ContainerStarted","Data":"f01114c13eca471447c03dcafa83c119dfd75fd57f5778b0f86074ae10416aac"} Oct 14 09:00:18 crc kubenswrapper[4870]: I1014 09:00:18.128955 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.128929158 podStartE2EDuration="2.128929158s" podCreationTimestamp="2025-10-14 09:00:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 09:00:18.119358222 +0000 UTC m=+7153.816718623" watchObservedRunningTime="2025-10-14 09:00:18.128929158 +0000 UTC m=+7153.826289569" Oct 14 09:00:19 crc kubenswrapper[4870]: I1014 09:00:19.425593 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 14 09:00:19 crc kubenswrapper[4870]: I1014 09:00:19.983245 4870 scope.go:117] "RemoveContainer" containerID="81cd4308735a167d73e072e83aa982ef464c8eb82a7354700badc7d2297f1f09" Oct 14 09:00:21 crc kubenswrapper[4870]: I1014 09:00:21.774875 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 14 09:00:21 crc kubenswrapper[4870]: I1014 09:00:21.775288 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 14 09:00:24 crc kubenswrapper[4870]: I1014 09:00:24.425633 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 14 09:00:24 crc kubenswrapper[4870]: I1014 09:00:24.468287 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 14 09:00:24 crc kubenswrapper[4870]: I1014 09:00:24.497864 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=8.49784279 podStartE2EDuration="8.49784279s" podCreationTimestamp="2025-10-14 09:00:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 09:00:18.155124745 +0000 UTC m=+7153.852485156" watchObservedRunningTime="2025-10-14 09:00:24.49784279 +0000 UTC m=+7160.195203171" Oct 14 09:00:25 crc kubenswrapper[4870]: I1014 09:00:25.221301 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 14 09:00:26 crc kubenswrapper[4870]: I1014 09:00:26.488587 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 14 09:00:26 crc kubenswrapper[4870]: I1014 09:00:26.488657 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 14 09:00:26 crc kubenswrapper[4870]: I1014 09:00:26.774571 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 14 09:00:26 crc kubenswrapper[4870]: I1014 09:00:26.775021 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 14 09:00:27 crc kubenswrapper[4870]: I1014 09:00:27.570641 4870 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="d3db1cbf-b57a-4f03-bde9-25dc20acd3bb" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.1.83:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 14 09:00:27 crc kubenswrapper[4870]: I1014 09:00:27.571000 4870 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="d3db1cbf-b57a-4f03-bde9-25dc20acd3bb" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.1.83:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 14 09:00:27 crc kubenswrapper[4870]: I1014 09:00:27.856709 4870 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="337148f2-88d7-4694-947e-de1aa4375d33" containerName="nova-metadata-log" probeResult="failure" output="Get \"http://10.217.1.84:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 14 09:00:27 crc kubenswrapper[4870]: I1014 09:00:27.857201 4870 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="337148f2-88d7-4694-947e-de1aa4375d33" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"http://10.217.1.84:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 14 09:00:36 crc kubenswrapper[4870]: I1014 09:00:36.493399 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 14 09:00:36 crc kubenswrapper[4870]: I1014 09:00:36.493897 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 14 09:00:36 crc kubenswrapper[4870]: I1014 09:00:36.494369 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 14 09:00:36 crc kubenswrapper[4870]: I1014 09:00:36.494489 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 14 09:00:36 crc kubenswrapper[4870]: I1014 09:00:36.497682 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 14 09:00:36 crc kubenswrapper[4870]: I1014 09:00:36.499750 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 14 09:00:36 crc kubenswrapper[4870]: I1014 09:00:36.708117 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5b4495fb7f-cw5zs"] Oct 14 09:00:36 crc kubenswrapper[4870]: I1014 09:00:36.731007 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b4495fb7f-cw5zs" Oct 14 09:00:36 crc kubenswrapper[4870]: I1014 09:00:36.760430 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5b4495fb7f-cw5zs"] Oct 14 09:00:36 crc kubenswrapper[4870]: I1014 09:00:36.779086 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 14 09:00:36 crc kubenswrapper[4870]: I1014 09:00:36.785067 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 14 09:00:36 crc kubenswrapper[4870]: I1014 09:00:36.788395 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 14 09:00:36 crc kubenswrapper[4870]: I1014 09:00:36.887810 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/935482d1-d09a-4c56-9b4d-cfd2d7897820-config\") pod \"dnsmasq-dns-5b4495fb7f-cw5zs\" (UID: \"935482d1-d09a-4c56-9b4d-cfd2d7897820\") " pod="openstack/dnsmasq-dns-5b4495fb7f-cw5zs" Oct 14 09:00:36 crc kubenswrapper[4870]: I1014 09:00:36.887870 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/935482d1-d09a-4c56-9b4d-cfd2d7897820-dns-svc\") pod \"dnsmasq-dns-5b4495fb7f-cw5zs\" (UID: \"935482d1-d09a-4c56-9b4d-cfd2d7897820\") " pod="openstack/dnsmasq-dns-5b4495fb7f-cw5zs" Oct 14 09:00:36 crc kubenswrapper[4870]: I1014 09:00:36.887911 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/935482d1-d09a-4c56-9b4d-cfd2d7897820-ovsdbserver-sb\") pod \"dnsmasq-dns-5b4495fb7f-cw5zs\" (UID: \"935482d1-d09a-4c56-9b4d-cfd2d7897820\") " pod="openstack/dnsmasq-dns-5b4495fb7f-cw5zs" Oct 14 09:00:36 crc kubenswrapper[4870]: I1014 09:00:36.887934 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2rjmz\" (UniqueName: \"kubernetes.io/projected/935482d1-d09a-4c56-9b4d-cfd2d7897820-kube-api-access-2rjmz\") pod \"dnsmasq-dns-5b4495fb7f-cw5zs\" (UID: \"935482d1-d09a-4c56-9b4d-cfd2d7897820\") " pod="openstack/dnsmasq-dns-5b4495fb7f-cw5zs" Oct 14 09:00:36 crc kubenswrapper[4870]: I1014 09:00:36.888033 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/935482d1-d09a-4c56-9b4d-cfd2d7897820-ovsdbserver-nb\") pod \"dnsmasq-dns-5b4495fb7f-cw5zs\" (UID: \"935482d1-d09a-4c56-9b4d-cfd2d7897820\") " pod="openstack/dnsmasq-dns-5b4495fb7f-cw5zs" Oct 14 09:00:36 crc kubenswrapper[4870]: I1014 09:00:36.990109 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/935482d1-d09a-4c56-9b4d-cfd2d7897820-dns-svc\") pod \"dnsmasq-dns-5b4495fb7f-cw5zs\" (UID: \"935482d1-d09a-4c56-9b4d-cfd2d7897820\") " pod="openstack/dnsmasq-dns-5b4495fb7f-cw5zs" Oct 14 09:00:36 crc kubenswrapper[4870]: I1014 09:00:36.990179 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/935482d1-d09a-4c56-9b4d-cfd2d7897820-ovsdbserver-sb\") pod \"dnsmasq-dns-5b4495fb7f-cw5zs\" (UID: \"935482d1-d09a-4c56-9b4d-cfd2d7897820\") " pod="openstack/dnsmasq-dns-5b4495fb7f-cw5zs" Oct 14 09:00:36 crc kubenswrapper[4870]: I1014 09:00:36.990203 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2rjmz\" (UniqueName: \"kubernetes.io/projected/935482d1-d09a-4c56-9b4d-cfd2d7897820-kube-api-access-2rjmz\") pod \"dnsmasq-dns-5b4495fb7f-cw5zs\" (UID: \"935482d1-d09a-4c56-9b4d-cfd2d7897820\") " pod="openstack/dnsmasq-dns-5b4495fb7f-cw5zs" Oct 14 09:00:36 crc kubenswrapper[4870]: I1014 09:00:36.990278 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/935482d1-d09a-4c56-9b4d-cfd2d7897820-ovsdbserver-nb\") pod \"dnsmasq-dns-5b4495fb7f-cw5zs\" (UID: \"935482d1-d09a-4c56-9b4d-cfd2d7897820\") " pod="openstack/dnsmasq-dns-5b4495fb7f-cw5zs" Oct 14 09:00:36 crc kubenswrapper[4870]: I1014 09:00:36.990337 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/935482d1-d09a-4c56-9b4d-cfd2d7897820-config\") pod \"dnsmasq-dns-5b4495fb7f-cw5zs\" (UID: \"935482d1-d09a-4c56-9b4d-cfd2d7897820\") " pod="openstack/dnsmasq-dns-5b4495fb7f-cw5zs" Oct 14 09:00:36 crc kubenswrapper[4870]: I1014 09:00:36.991265 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/935482d1-d09a-4c56-9b4d-cfd2d7897820-dns-svc\") pod \"dnsmasq-dns-5b4495fb7f-cw5zs\" (UID: \"935482d1-d09a-4c56-9b4d-cfd2d7897820\") " pod="openstack/dnsmasq-dns-5b4495fb7f-cw5zs" Oct 14 09:00:36 crc kubenswrapper[4870]: I1014 09:00:36.991327 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/935482d1-d09a-4c56-9b4d-cfd2d7897820-ovsdbserver-sb\") pod \"dnsmasq-dns-5b4495fb7f-cw5zs\" (UID: \"935482d1-d09a-4c56-9b4d-cfd2d7897820\") " pod="openstack/dnsmasq-dns-5b4495fb7f-cw5zs" Oct 14 09:00:36 crc kubenswrapper[4870]: I1014 09:00:36.991610 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/935482d1-d09a-4c56-9b4d-cfd2d7897820-config\") pod \"dnsmasq-dns-5b4495fb7f-cw5zs\" (UID: \"935482d1-d09a-4c56-9b4d-cfd2d7897820\") " pod="openstack/dnsmasq-dns-5b4495fb7f-cw5zs" Oct 14 09:00:36 crc kubenswrapper[4870]: I1014 09:00:36.998596 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/935482d1-d09a-4c56-9b4d-cfd2d7897820-ovsdbserver-nb\") pod \"dnsmasq-dns-5b4495fb7f-cw5zs\" (UID: \"935482d1-d09a-4c56-9b4d-cfd2d7897820\") " pod="openstack/dnsmasq-dns-5b4495fb7f-cw5zs" Oct 14 09:00:37 crc kubenswrapper[4870]: I1014 09:00:37.024497 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2rjmz\" (UniqueName: \"kubernetes.io/projected/935482d1-d09a-4c56-9b4d-cfd2d7897820-kube-api-access-2rjmz\") pod \"dnsmasq-dns-5b4495fb7f-cw5zs\" (UID: \"935482d1-d09a-4c56-9b4d-cfd2d7897820\") " pod="openstack/dnsmasq-dns-5b4495fb7f-cw5zs" Oct 14 09:00:37 crc kubenswrapper[4870]: I1014 09:00:37.067016 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b4495fb7f-cw5zs" Oct 14 09:00:37 crc kubenswrapper[4870]: I1014 09:00:37.321791 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 14 09:00:37 crc kubenswrapper[4870]: W1014 09:00:37.554462 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod935482d1_d09a_4c56_9b4d_cfd2d7897820.slice/crio-e8d0775880cb586d7c87ede08d0c5960cc045508f80101cc9226fea6d0987291 WatchSource:0}: Error finding container e8d0775880cb586d7c87ede08d0c5960cc045508f80101cc9226fea6d0987291: Status 404 returned error can't find the container with id e8d0775880cb586d7c87ede08d0c5960cc045508f80101cc9226fea6d0987291 Oct 14 09:00:37 crc kubenswrapper[4870]: I1014 09:00:37.567390 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5b4495fb7f-cw5zs"] Oct 14 09:00:38 crc kubenswrapper[4870]: I1014 09:00:38.326622 4870 generic.go:334] "Generic (PLEG): container finished" podID="935482d1-d09a-4c56-9b4d-cfd2d7897820" containerID="dab6722d8bb2837aa5d068e5d4d12b75bd61194b8d7ab4c71e41dd75604954b7" exitCode=0 Oct 14 09:00:38 crc kubenswrapper[4870]: I1014 09:00:38.326656 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b4495fb7f-cw5zs" event={"ID":"935482d1-d09a-4c56-9b4d-cfd2d7897820","Type":"ContainerDied","Data":"dab6722d8bb2837aa5d068e5d4d12b75bd61194b8d7ab4c71e41dd75604954b7"} Oct 14 09:00:38 crc kubenswrapper[4870]: I1014 09:00:38.327007 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b4495fb7f-cw5zs" event={"ID":"935482d1-d09a-4c56-9b4d-cfd2d7897820","Type":"ContainerStarted","Data":"e8d0775880cb586d7c87ede08d0c5960cc045508f80101cc9226fea6d0987291"} Oct 14 09:00:39 crc kubenswrapper[4870]: I1014 09:00:39.340232 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b4495fb7f-cw5zs" event={"ID":"935482d1-d09a-4c56-9b4d-cfd2d7897820","Type":"ContainerStarted","Data":"f6a6d0a40aa24e08080ff7782a1374691a10405c3eb48f5a12139f0887eec7e5"} Oct 14 09:00:39 crc kubenswrapper[4870]: I1014 09:00:39.340797 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5b4495fb7f-cw5zs" Oct 14 09:00:39 crc kubenswrapper[4870]: I1014 09:00:39.368965 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5b4495fb7f-cw5zs" podStartSLOduration=3.368944676 podStartE2EDuration="3.368944676s" podCreationTimestamp="2025-10-14 09:00:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 09:00:39.362758344 +0000 UTC m=+7175.060118715" watchObservedRunningTime="2025-10-14 09:00:39.368944676 +0000 UTC m=+7175.066305047" Oct 14 09:00:47 crc kubenswrapper[4870]: I1014 09:00:47.069012 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5b4495fb7f-cw5zs" Oct 14 09:00:47 crc kubenswrapper[4870]: I1014 09:00:47.175417 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7c8cbf976c-khp98"] Oct 14 09:00:47 crc kubenswrapper[4870]: I1014 09:00:47.175789 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7c8cbf976c-khp98" podUID="36c98e9a-274d-466a-8e68-b2804be28695" containerName="dnsmasq-dns" containerID="cri-o://b92b8e6d69e662f349f55344ecfc601712b19204e2f344064d23ad7c0b0fecce" gracePeriod=10 Oct 14 09:00:47 crc kubenswrapper[4870]: I1014 09:00:47.429168 4870 generic.go:334] "Generic (PLEG): container finished" podID="36c98e9a-274d-466a-8e68-b2804be28695" containerID="b92b8e6d69e662f349f55344ecfc601712b19204e2f344064d23ad7c0b0fecce" exitCode=0 Oct 14 09:00:47 crc kubenswrapper[4870]: I1014 09:00:47.429298 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7c8cbf976c-khp98" event={"ID":"36c98e9a-274d-466a-8e68-b2804be28695","Type":"ContainerDied","Data":"b92b8e6d69e662f349f55344ecfc601712b19204e2f344064d23ad7c0b0fecce"} Oct 14 09:00:47 crc kubenswrapper[4870]: I1014 09:00:47.664164 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7c8cbf976c-khp98" Oct 14 09:00:47 crc kubenswrapper[4870]: I1014 09:00:47.808205 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/36c98e9a-274d-466a-8e68-b2804be28695-ovsdbserver-nb\") pod \"36c98e9a-274d-466a-8e68-b2804be28695\" (UID: \"36c98e9a-274d-466a-8e68-b2804be28695\") " Oct 14 09:00:47 crc kubenswrapper[4870]: I1014 09:00:47.808333 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/36c98e9a-274d-466a-8e68-b2804be28695-config\") pod \"36c98e9a-274d-466a-8e68-b2804be28695\" (UID: \"36c98e9a-274d-466a-8e68-b2804be28695\") " Oct 14 09:00:47 crc kubenswrapper[4870]: I1014 09:00:47.808426 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/36c98e9a-274d-466a-8e68-b2804be28695-ovsdbserver-sb\") pod \"36c98e9a-274d-466a-8e68-b2804be28695\" (UID: \"36c98e9a-274d-466a-8e68-b2804be28695\") " Oct 14 09:00:47 crc kubenswrapper[4870]: I1014 09:00:47.808518 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/36c98e9a-274d-466a-8e68-b2804be28695-dns-svc\") pod \"36c98e9a-274d-466a-8e68-b2804be28695\" (UID: \"36c98e9a-274d-466a-8e68-b2804be28695\") " Oct 14 09:00:47 crc kubenswrapper[4870]: I1014 09:00:47.808576 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k8lx7\" (UniqueName: \"kubernetes.io/projected/36c98e9a-274d-466a-8e68-b2804be28695-kube-api-access-k8lx7\") pod \"36c98e9a-274d-466a-8e68-b2804be28695\" (UID: \"36c98e9a-274d-466a-8e68-b2804be28695\") " Oct 14 09:00:47 crc kubenswrapper[4870]: I1014 09:00:47.813288 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/36c98e9a-274d-466a-8e68-b2804be28695-kube-api-access-k8lx7" (OuterVolumeSpecName: "kube-api-access-k8lx7") pod "36c98e9a-274d-466a-8e68-b2804be28695" (UID: "36c98e9a-274d-466a-8e68-b2804be28695"). InnerVolumeSpecName "kube-api-access-k8lx7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:00:47 crc kubenswrapper[4870]: I1014 09:00:47.859032 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/36c98e9a-274d-466a-8e68-b2804be28695-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "36c98e9a-274d-466a-8e68-b2804be28695" (UID: "36c98e9a-274d-466a-8e68-b2804be28695"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 09:00:47 crc kubenswrapper[4870]: I1014 09:00:47.859562 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/36c98e9a-274d-466a-8e68-b2804be28695-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "36c98e9a-274d-466a-8e68-b2804be28695" (UID: "36c98e9a-274d-466a-8e68-b2804be28695"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 09:00:47 crc kubenswrapper[4870]: I1014 09:00:47.873489 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/36c98e9a-274d-466a-8e68-b2804be28695-config" (OuterVolumeSpecName: "config") pod "36c98e9a-274d-466a-8e68-b2804be28695" (UID: "36c98e9a-274d-466a-8e68-b2804be28695"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 09:00:47 crc kubenswrapper[4870]: I1014 09:00:47.881860 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/36c98e9a-274d-466a-8e68-b2804be28695-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "36c98e9a-274d-466a-8e68-b2804be28695" (UID: "36c98e9a-274d-466a-8e68-b2804be28695"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 09:00:47 crc kubenswrapper[4870]: I1014 09:00:47.910598 4870 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/36c98e9a-274d-466a-8e68-b2804be28695-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 14 09:00:47 crc kubenswrapper[4870]: I1014 09:00:47.910625 4870 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/36c98e9a-274d-466a-8e68-b2804be28695-config\") on node \"crc\" DevicePath \"\"" Oct 14 09:00:47 crc kubenswrapper[4870]: I1014 09:00:47.910636 4870 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/36c98e9a-274d-466a-8e68-b2804be28695-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 14 09:00:47 crc kubenswrapper[4870]: I1014 09:00:47.910646 4870 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/36c98e9a-274d-466a-8e68-b2804be28695-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 14 09:00:47 crc kubenswrapper[4870]: I1014 09:00:47.910655 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k8lx7\" (UniqueName: \"kubernetes.io/projected/36c98e9a-274d-466a-8e68-b2804be28695-kube-api-access-k8lx7\") on node \"crc\" DevicePath \"\"" Oct 14 09:00:48 crc kubenswrapper[4870]: I1014 09:00:48.454466 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7c8cbf976c-khp98" event={"ID":"36c98e9a-274d-466a-8e68-b2804be28695","Type":"ContainerDied","Data":"4be6160942e68ba9cfc1c81fa07a993f4ceacf23291f9760a26e029e847ff36d"} Oct 14 09:00:48 crc kubenswrapper[4870]: I1014 09:00:48.454534 4870 scope.go:117] "RemoveContainer" containerID="b92b8e6d69e662f349f55344ecfc601712b19204e2f344064d23ad7c0b0fecce" Oct 14 09:00:48 crc kubenswrapper[4870]: I1014 09:00:48.454557 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7c8cbf976c-khp98" Oct 14 09:00:48 crc kubenswrapper[4870]: I1014 09:00:48.481166 4870 scope.go:117] "RemoveContainer" containerID="99c14090d95b5b3aa6788c08b7baf81b4f6c73cd221330803a8d489320dd6cf9" Oct 14 09:00:48 crc kubenswrapper[4870]: I1014 09:00:48.513997 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7c8cbf976c-khp98"] Oct 14 09:00:48 crc kubenswrapper[4870]: I1014 09:00:48.515493 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7c8cbf976c-khp98"] Oct 14 09:00:49 crc kubenswrapper[4870]: I1014 09:00:49.045674 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="36c98e9a-274d-466a-8e68-b2804be28695" path="/var/lib/kubelet/pods/36c98e9a-274d-466a-8e68-b2804be28695/volumes" Oct 14 09:00:51 crc kubenswrapper[4870]: I1014 09:00:51.597182 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-c7pv6"] Oct 14 09:00:51 crc kubenswrapper[4870]: E1014 09:00:51.597945 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="36c98e9a-274d-466a-8e68-b2804be28695" containerName="init" Oct 14 09:00:51 crc kubenswrapper[4870]: I1014 09:00:51.597959 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="36c98e9a-274d-466a-8e68-b2804be28695" containerName="init" Oct 14 09:00:51 crc kubenswrapper[4870]: E1014 09:00:51.597976 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="36c98e9a-274d-466a-8e68-b2804be28695" containerName="dnsmasq-dns" Oct 14 09:00:51 crc kubenswrapper[4870]: I1014 09:00:51.597982 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="36c98e9a-274d-466a-8e68-b2804be28695" containerName="dnsmasq-dns" Oct 14 09:00:51 crc kubenswrapper[4870]: I1014 09:00:51.598201 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="36c98e9a-274d-466a-8e68-b2804be28695" containerName="dnsmasq-dns" Oct 14 09:00:51 crc kubenswrapper[4870]: I1014 09:00:51.598897 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-c7pv6" Oct 14 09:00:51 crc kubenswrapper[4870]: I1014 09:00:51.605981 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-c7pv6"] Oct 14 09:00:51 crc kubenswrapper[4870]: I1014 09:00:51.687424 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xrqzl\" (UniqueName: \"kubernetes.io/projected/651aa59d-e6a6-48b2-806d-8a2e2531814f-kube-api-access-xrqzl\") pod \"cinder-db-create-c7pv6\" (UID: \"651aa59d-e6a6-48b2-806d-8a2e2531814f\") " pod="openstack/cinder-db-create-c7pv6" Oct 14 09:00:51 crc kubenswrapper[4870]: I1014 09:00:51.789522 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xrqzl\" (UniqueName: \"kubernetes.io/projected/651aa59d-e6a6-48b2-806d-8a2e2531814f-kube-api-access-xrqzl\") pod \"cinder-db-create-c7pv6\" (UID: \"651aa59d-e6a6-48b2-806d-8a2e2531814f\") " pod="openstack/cinder-db-create-c7pv6" Oct 14 09:00:51 crc kubenswrapper[4870]: I1014 09:00:51.811259 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xrqzl\" (UniqueName: \"kubernetes.io/projected/651aa59d-e6a6-48b2-806d-8a2e2531814f-kube-api-access-xrqzl\") pod \"cinder-db-create-c7pv6\" (UID: \"651aa59d-e6a6-48b2-806d-8a2e2531814f\") " pod="openstack/cinder-db-create-c7pv6" Oct 14 09:00:51 crc kubenswrapper[4870]: I1014 09:00:51.922766 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-c7pv6" Oct 14 09:00:52 crc kubenswrapper[4870]: I1014 09:00:52.364628 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-c7pv6"] Oct 14 09:00:52 crc kubenswrapper[4870]: I1014 09:00:52.496690 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-c7pv6" event={"ID":"651aa59d-e6a6-48b2-806d-8a2e2531814f","Type":"ContainerStarted","Data":"4107bc355433d11a9b3c13a4c89ef9dd67b8422804d6d8161ed1190203ad943f"} Oct 14 09:00:53 crc kubenswrapper[4870]: I1014 09:00:53.509521 4870 generic.go:334] "Generic (PLEG): container finished" podID="651aa59d-e6a6-48b2-806d-8a2e2531814f" containerID="7221c8a48ccda4afcf145ed2303bc545cf64e8407ec3f7f95bbf196913e791c2" exitCode=0 Oct 14 09:00:53 crc kubenswrapper[4870]: I1014 09:00:53.509585 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-c7pv6" event={"ID":"651aa59d-e6a6-48b2-806d-8a2e2531814f","Type":"ContainerDied","Data":"7221c8a48ccda4afcf145ed2303bc545cf64e8407ec3f7f95bbf196913e791c2"} Oct 14 09:00:53 crc kubenswrapper[4870]: I1014 09:00:53.951254 4870 patch_prober.go:28] interesting pod/machine-config-daemon-7tvc7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 09:00:53 crc kubenswrapper[4870]: I1014 09:00:53.951350 4870 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 09:00:54 crc kubenswrapper[4870]: I1014 09:00:54.883066 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-c7pv6" Oct 14 09:00:54 crc kubenswrapper[4870]: I1014 09:00:54.951711 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xrqzl\" (UniqueName: \"kubernetes.io/projected/651aa59d-e6a6-48b2-806d-8a2e2531814f-kube-api-access-xrqzl\") pod \"651aa59d-e6a6-48b2-806d-8a2e2531814f\" (UID: \"651aa59d-e6a6-48b2-806d-8a2e2531814f\") " Oct 14 09:00:54 crc kubenswrapper[4870]: I1014 09:00:54.963939 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/651aa59d-e6a6-48b2-806d-8a2e2531814f-kube-api-access-xrqzl" (OuterVolumeSpecName: "kube-api-access-xrqzl") pod "651aa59d-e6a6-48b2-806d-8a2e2531814f" (UID: "651aa59d-e6a6-48b2-806d-8a2e2531814f"). InnerVolumeSpecName "kube-api-access-xrqzl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:00:55 crc kubenswrapper[4870]: I1014 09:00:55.053958 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xrqzl\" (UniqueName: \"kubernetes.io/projected/651aa59d-e6a6-48b2-806d-8a2e2531814f-kube-api-access-xrqzl\") on node \"crc\" DevicePath \"\"" Oct 14 09:00:55 crc kubenswrapper[4870]: I1014 09:00:55.534913 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-c7pv6" event={"ID":"651aa59d-e6a6-48b2-806d-8a2e2531814f","Type":"ContainerDied","Data":"4107bc355433d11a9b3c13a4c89ef9dd67b8422804d6d8161ed1190203ad943f"} Oct 14 09:00:55 crc kubenswrapper[4870]: I1014 09:00:55.535372 4870 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4107bc355433d11a9b3c13a4c89ef9dd67b8422804d6d8161ed1190203ad943f" Oct 14 09:00:55 crc kubenswrapper[4870]: I1014 09:00:55.534961 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-c7pv6" Oct 14 09:01:00 crc kubenswrapper[4870]: I1014 09:01:00.135096 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29340541-wtcff"] Oct 14 09:01:00 crc kubenswrapper[4870]: E1014 09:01:00.136260 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="651aa59d-e6a6-48b2-806d-8a2e2531814f" containerName="mariadb-database-create" Oct 14 09:01:00 crc kubenswrapper[4870]: I1014 09:01:00.136278 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="651aa59d-e6a6-48b2-806d-8a2e2531814f" containerName="mariadb-database-create" Oct 14 09:01:00 crc kubenswrapper[4870]: I1014 09:01:00.136565 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="651aa59d-e6a6-48b2-806d-8a2e2531814f" containerName="mariadb-database-create" Oct 14 09:01:00 crc kubenswrapper[4870]: I1014 09:01:00.137435 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29340541-wtcff" Oct 14 09:01:00 crc kubenswrapper[4870]: I1014 09:01:00.148039 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29340541-wtcff"] Oct 14 09:01:00 crc kubenswrapper[4870]: I1014 09:01:00.250797 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/ec2b912e-ff75-44ed-b862-6bffd46f0ddf-fernet-keys\") pod \"keystone-cron-29340541-wtcff\" (UID: \"ec2b912e-ff75-44ed-b862-6bffd46f0ddf\") " pod="openstack/keystone-cron-29340541-wtcff" Oct 14 09:01:00 crc kubenswrapper[4870]: I1014 09:01:00.250866 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ec2b912e-ff75-44ed-b862-6bffd46f0ddf-config-data\") pod \"keystone-cron-29340541-wtcff\" (UID: \"ec2b912e-ff75-44ed-b862-6bffd46f0ddf\") " pod="openstack/keystone-cron-29340541-wtcff" Oct 14 09:01:00 crc kubenswrapper[4870]: I1014 09:01:00.250896 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v9ph4\" (UniqueName: \"kubernetes.io/projected/ec2b912e-ff75-44ed-b862-6bffd46f0ddf-kube-api-access-v9ph4\") pod \"keystone-cron-29340541-wtcff\" (UID: \"ec2b912e-ff75-44ed-b862-6bffd46f0ddf\") " pod="openstack/keystone-cron-29340541-wtcff" Oct 14 09:01:00 crc kubenswrapper[4870]: I1014 09:01:00.250939 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec2b912e-ff75-44ed-b862-6bffd46f0ddf-combined-ca-bundle\") pod \"keystone-cron-29340541-wtcff\" (UID: \"ec2b912e-ff75-44ed-b862-6bffd46f0ddf\") " pod="openstack/keystone-cron-29340541-wtcff" Oct 14 09:01:00 crc kubenswrapper[4870]: I1014 09:01:00.352773 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/ec2b912e-ff75-44ed-b862-6bffd46f0ddf-fernet-keys\") pod \"keystone-cron-29340541-wtcff\" (UID: \"ec2b912e-ff75-44ed-b862-6bffd46f0ddf\") " pod="openstack/keystone-cron-29340541-wtcff" Oct 14 09:01:00 crc kubenswrapper[4870]: I1014 09:01:00.352870 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ec2b912e-ff75-44ed-b862-6bffd46f0ddf-config-data\") pod \"keystone-cron-29340541-wtcff\" (UID: \"ec2b912e-ff75-44ed-b862-6bffd46f0ddf\") " pod="openstack/keystone-cron-29340541-wtcff" Oct 14 09:01:00 crc kubenswrapper[4870]: I1014 09:01:00.352903 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v9ph4\" (UniqueName: \"kubernetes.io/projected/ec2b912e-ff75-44ed-b862-6bffd46f0ddf-kube-api-access-v9ph4\") pod \"keystone-cron-29340541-wtcff\" (UID: \"ec2b912e-ff75-44ed-b862-6bffd46f0ddf\") " pod="openstack/keystone-cron-29340541-wtcff" Oct 14 09:01:00 crc kubenswrapper[4870]: I1014 09:01:00.352946 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec2b912e-ff75-44ed-b862-6bffd46f0ddf-combined-ca-bundle\") pod \"keystone-cron-29340541-wtcff\" (UID: \"ec2b912e-ff75-44ed-b862-6bffd46f0ddf\") " pod="openstack/keystone-cron-29340541-wtcff" Oct 14 09:01:00 crc kubenswrapper[4870]: I1014 09:01:00.358720 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/ec2b912e-ff75-44ed-b862-6bffd46f0ddf-fernet-keys\") pod \"keystone-cron-29340541-wtcff\" (UID: \"ec2b912e-ff75-44ed-b862-6bffd46f0ddf\") " pod="openstack/keystone-cron-29340541-wtcff" Oct 14 09:01:00 crc kubenswrapper[4870]: I1014 09:01:00.359702 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec2b912e-ff75-44ed-b862-6bffd46f0ddf-combined-ca-bundle\") pod \"keystone-cron-29340541-wtcff\" (UID: \"ec2b912e-ff75-44ed-b862-6bffd46f0ddf\") " pod="openstack/keystone-cron-29340541-wtcff" Oct 14 09:01:00 crc kubenswrapper[4870]: I1014 09:01:00.364014 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ec2b912e-ff75-44ed-b862-6bffd46f0ddf-config-data\") pod \"keystone-cron-29340541-wtcff\" (UID: \"ec2b912e-ff75-44ed-b862-6bffd46f0ddf\") " pod="openstack/keystone-cron-29340541-wtcff" Oct 14 09:01:00 crc kubenswrapper[4870]: I1014 09:01:00.381889 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v9ph4\" (UniqueName: \"kubernetes.io/projected/ec2b912e-ff75-44ed-b862-6bffd46f0ddf-kube-api-access-v9ph4\") pod \"keystone-cron-29340541-wtcff\" (UID: \"ec2b912e-ff75-44ed-b862-6bffd46f0ddf\") " pod="openstack/keystone-cron-29340541-wtcff" Oct 14 09:01:00 crc kubenswrapper[4870]: I1014 09:01:00.476119 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29340541-wtcff" Oct 14 09:01:00 crc kubenswrapper[4870]: W1014 09:01:00.766169 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podec2b912e_ff75_44ed_b862_6bffd46f0ddf.slice/crio-b6082526502633c4616c7d1c6c04c206343f36187cb51218c3be545c61f6410e WatchSource:0}: Error finding container b6082526502633c4616c7d1c6c04c206343f36187cb51218c3be545c61f6410e: Status 404 returned error can't find the container with id b6082526502633c4616c7d1c6c04c206343f36187cb51218c3be545c61f6410e Oct 14 09:01:00 crc kubenswrapper[4870]: I1014 09:01:00.770785 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29340541-wtcff"] Oct 14 09:01:01 crc kubenswrapper[4870]: I1014 09:01:01.598601 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29340541-wtcff" event={"ID":"ec2b912e-ff75-44ed-b862-6bffd46f0ddf","Type":"ContainerStarted","Data":"dfd1476d94de8928925042d38747d1660d3ae8e9ab3288c27e997278ea500bb9"} Oct 14 09:01:01 crc kubenswrapper[4870]: I1014 09:01:01.600575 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29340541-wtcff" event={"ID":"ec2b912e-ff75-44ed-b862-6bffd46f0ddf","Type":"ContainerStarted","Data":"b6082526502633c4616c7d1c6c04c206343f36187cb51218c3be545c61f6410e"} Oct 14 09:01:01 crc kubenswrapper[4870]: I1014 09:01:01.627378 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29340541-wtcff" podStartSLOduration=1.627361 podStartE2EDuration="1.627361s" podCreationTimestamp="2025-10-14 09:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 09:01:01.618932192 +0000 UTC m=+7197.316292593" watchObservedRunningTime="2025-10-14 09:01:01.627361 +0000 UTC m=+7197.324721371" Oct 14 09:01:01 crc kubenswrapper[4870]: I1014 09:01:01.735990 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-fafa-account-create-758c7"] Oct 14 09:01:01 crc kubenswrapper[4870]: I1014 09:01:01.737886 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-fafa-account-create-758c7" Oct 14 09:01:01 crc kubenswrapper[4870]: I1014 09:01:01.740593 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Oct 14 09:01:01 crc kubenswrapper[4870]: I1014 09:01:01.750848 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-fafa-account-create-758c7"] Oct 14 09:01:01 crc kubenswrapper[4870]: I1014 09:01:01.817653 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kkw4l\" (UniqueName: \"kubernetes.io/projected/ad2fe341-5486-4fec-b744-e294b5d034bc-kube-api-access-kkw4l\") pod \"cinder-fafa-account-create-758c7\" (UID: \"ad2fe341-5486-4fec-b744-e294b5d034bc\") " pod="openstack/cinder-fafa-account-create-758c7" Oct 14 09:01:01 crc kubenswrapper[4870]: I1014 09:01:01.919693 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kkw4l\" (UniqueName: \"kubernetes.io/projected/ad2fe341-5486-4fec-b744-e294b5d034bc-kube-api-access-kkw4l\") pod \"cinder-fafa-account-create-758c7\" (UID: \"ad2fe341-5486-4fec-b744-e294b5d034bc\") " pod="openstack/cinder-fafa-account-create-758c7" Oct 14 09:01:01 crc kubenswrapper[4870]: I1014 09:01:01.941918 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kkw4l\" (UniqueName: \"kubernetes.io/projected/ad2fe341-5486-4fec-b744-e294b5d034bc-kube-api-access-kkw4l\") pod \"cinder-fafa-account-create-758c7\" (UID: \"ad2fe341-5486-4fec-b744-e294b5d034bc\") " pod="openstack/cinder-fafa-account-create-758c7" Oct 14 09:01:02 crc kubenswrapper[4870]: I1014 09:01:02.064164 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-fafa-account-create-758c7" Oct 14 09:01:02 crc kubenswrapper[4870]: W1014 09:01:02.508930 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podad2fe341_5486_4fec_b744_e294b5d034bc.slice/crio-6be5faac56c1b17f64399233f9881a4f9e2394f4749c763336ff389b578315be WatchSource:0}: Error finding container 6be5faac56c1b17f64399233f9881a4f9e2394f4749c763336ff389b578315be: Status 404 returned error can't find the container with id 6be5faac56c1b17f64399233f9881a4f9e2394f4749c763336ff389b578315be Oct 14 09:01:02 crc kubenswrapper[4870]: I1014 09:01:02.509296 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-fafa-account-create-758c7"] Oct 14 09:01:02 crc kubenswrapper[4870]: I1014 09:01:02.608083 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-fafa-account-create-758c7" event={"ID":"ad2fe341-5486-4fec-b744-e294b5d034bc","Type":"ContainerStarted","Data":"6be5faac56c1b17f64399233f9881a4f9e2394f4749c763336ff389b578315be"} Oct 14 09:01:03 crc kubenswrapper[4870]: I1014 09:01:03.617173 4870 generic.go:334] "Generic (PLEG): container finished" podID="ad2fe341-5486-4fec-b744-e294b5d034bc" containerID="5bff5d9b13861ce498c2037fffc163d2c8ac9a090f60d4678677965ee8fdfec9" exitCode=0 Oct 14 09:01:03 crc kubenswrapper[4870]: I1014 09:01:03.617258 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-fafa-account-create-758c7" event={"ID":"ad2fe341-5486-4fec-b744-e294b5d034bc","Type":"ContainerDied","Data":"5bff5d9b13861ce498c2037fffc163d2c8ac9a090f60d4678677965ee8fdfec9"} Oct 14 09:01:03 crc kubenswrapper[4870]: I1014 09:01:03.619559 4870 generic.go:334] "Generic (PLEG): container finished" podID="ec2b912e-ff75-44ed-b862-6bffd46f0ddf" containerID="dfd1476d94de8928925042d38747d1660d3ae8e9ab3288c27e997278ea500bb9" exitCode=0 Oct 14 09:01:03 crc kubenswrapper[4870]: I1014 09:01:03.619801 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29340541-wtcff" event={"ID":"ec2b912e-ff75-44ed-b862-6bffd46f0ddf","Type":"ContainerDied","Data":"dfd1476d94de8928925042d38747d1660d3ae8e9ab3288c27e997278ea500bb9"} Oct 14 09:01:05 crc kubenswrapper[4870]: I1014 09:01:05.072319 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-fafa-account-create-758c7" Oct 14 09:01:05 crc kubenswrapper[4870]: I1014 09:01:05.084245 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29340541-wtcff" Oct 14 09:01:05 crc kubenswrapper[4870]: I1014 09:01:05.176208 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/ec2b912e-ff75-44ed-b862-6bffd46f0ddf-fernet-keys\") pod \"ec2b912e-ff75-44ed-b862-6bffd46f0ddf\" (UID: \"ec2b912e-ff75-44ed-b862-6bffd46f0ddf\") " Oct 14 09:01:05 crc kubenswrapper[4870]: I1014 09:01:05.176294 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kkw4l\" (UniqueName: \"kubernetes.io/projected/ad2fe341-5486-4fec-b744-e294b5d034bc-kube-api-access-kkw4l\") pod \"ad2fe341-5486-4fec-b744-e294b5d034bc\" (UID: \"ad2fe341-5486-4fec-b744-e294b5d034bc\") " Oct 14 09:01:05 crc kubenswrapper[4870]: I1014 09:01:05.176380 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ec2b912e-ff75-44ed-b862-6bffd46f0ddf-config-data\") pod \"ec2b912e-ff75-44ed-b862-6bffd46f0ddf\" (UID: \"ec2b912e-ff75-44ed-b862-6bffd46f0ddf\") " Oct 14 09:01:05 crc kubenswrapper[4870]: I1014 09:01:05.176819 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v9ph4\" (UniqueName: \"kubernetes.io/projected/ec2b912e-ff75-44ed-b862-6bffd46f0ddf-kube-api-access-v9ph4\") pod \"ec2b912e-ff75-44ed-b862-6bffd46f0ddf\" (UID: \"ec2b912e-ff75-44ed-b862-6bffd46f0ddf\") " Oct 14 09:01:05 crc kubenswrapper[4870]: I1014 09:01:05.176852 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec2b912e-ff75-44ed-b862-6bffd46f0ddf-combined-ca-bundle\") pod \"ec2b912e-ff75-44ed-b862-6bffd46f0ddf\" (UID: \"ec2b912e-ff75-44ed-b862-6bffd46f0ddf\") " Oct 14 09:01:05 crc kubenswrapper[4870]: I1014 09:01:05.181587 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ad2fe341-5486-4fec-b744-e294b5d034bc-kube-api-access-kkw4l" (OuterVolumeSpecName: "kube-api-access-kkw4l") pod "ad2fe341-5486-4fec-b744-e294b5d034bc" (UID: "ad2fe341-5486-4fec-b744-e294b5d034bc"). InnerVolumeSpecName "kube-api-access-kkw4l". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:01:05 crc kubenswrapper[4870]: I1014 09:01:05.181683 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ec2b912e-ff75-44ed-b862-6bffd46f0ddf-kube-api-access-v9ph4" (OuterVolumeSpecName: "kube-api-access-v9ph4") pod "ec2b912e-ff75-44ed-b862-6bffd46f0ddf" (UID: "ec2b912e-ff75-44ed-b862-6bffd46f0ddf"). InnerVolumeSpecName "kube-api-access-v9ph4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:01:05 crc kubenswrapper[4870]: I1014 09:01:05.183565 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ec2b912e-ff75-44ed-b862-6bffd46f0ddf-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "ec2b912e-ff75-44ed-b862-6bffd46f0ddf" (UID: "ec2b912e-ff75-44ed-b862-6bffd46f0ddf"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:01:05 crc kubenswrapper[4870]: I1014 09:01:05.211693 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ec2b912e-ff75-44ed-b862-6bffd46f0ddf-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ec2b912e-ff75-44ed-b862-6bffd46f0ddf" (UID: "ec2b912e-ff75-44ed-b862-6bffd46f0ddf"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:01:05 crc kubenswrapper[4870]: I1014 09:01:05.227351 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ec2b912e-ff75-44ed-b862-6bffd46f0ddf-config-data" (OuterVolumeSpecName: "config-data") pod "ec2b912e-ff75-44ed-b862-6bffd46f0ddf" (UID: "ec2b912e-ff75-44ed-b862-6bffd46f0ddf"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:01:05 crc kubenswrapper[4870]: I1014 09:01:05.280355 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v9ph4\" (UniqueName: \"kubernetes.io/projected/ec2b912e-ff75-44ed-b862-6bffd46f0ddf-kube-api-access-v9ph4\") on node \"crc\" DevicePath \"\"" Oct 14 09:01:05 crc kubenswrapper[4870]: I1014 09:01:05.280480 4870 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec2b912e-ff75-44ed-b862-6bffd46f0ddf-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 09:01:05 crc kubenswrapper[4870]: I1014 09:01:05.280546 4870 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/ec2b912e-ff75-44ed-b862-6bffd46f0ddf-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 14 09:01:05 crc kubenswrapper[4870]: I1014 09:01:05.280568 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kkw4l\" (UniqueName: \"kubernetes.io/projected/ad2fe341-5486-4fec-b744-e294b5d034bc-kube-api-access-kkw4l\") on node \"crc\" DevicePath \"\"" Oct 14 09:01:05 crc kubenswrapper[4870]: I1014 09:01:05.280587 4870 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ec2b912e-ff75-44ed-b862-6bffd46f0ddf-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 09:01:05 crc kubenswrapper[4870]: I1014 09:01:05.641218 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29340541-wtcff" Oct 14 09:01:05 crc kubenswrapper[4870]: I1014 09:01:05.641254 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29340541-wtcff" event={"ID":"ec2b912e-ff75-44ed-b862-6bffd46f0ddf","Type":"ContainerDied","Data":"b6082526502633c4616c7d1c6c04c206343f36187cb51218c3be545c61f6410e"} Oct 14 09:01:05 crc kubenswrapper[4870]: I1014 09:01:05.641327 4870 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b6082526502633c4616c7d1c6c04c206343f36187cb51218c3be545c61f6410e" Oct 14 09:01:05 crc kubenswrapper[4870]: I1014 09:01:05.642765 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-fafa-account-create-758c7" event={"ID":"ad2fe341-5486-4fec-b744-e294b5d034bc","Type":"ContainerDied","Data":"6be5faac56c1b17f64399233f9881a4f9e2394f4749c763336ff389b578315be"} Oct 14 09:01:05 crc kubenswrapper[4870]: I1014 09:01:05.642803 4870 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6be5faac56c1b17f64399233f9881a4f9e2394f4749c763336ff389b578315be" Oct 14 09:01:05 crc kubenswrapper[4870]: I1014 09:01:05.642867 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-fafa-account-create-758c7" Oct 14 09:01:06 crc kubenswrapper[4870]: I1014 09:01:06.963425 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-x7dzz"] Oct 14 09:01:06 crc kubenswrapper[4870]: E1014 09:01:06.964205 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ec2b912e-ff75-44ed-b862-6bffd46f0ddf" containerName="keystone-cron" Oct 14 09:01:06 crc kubenswrapper[4870]: I1014 09:01:06.964222 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="ec2b912e-ff75-44ed-b862-6bffd46f0ddf" containerName="keystone-cron" Oct 14 09:01:06 crc kubenswrapper[4870]: E1014 09:01:06.964251 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad2fe341-5486-4fec-b744-e294b5d034bc" containerName="mariadb-account-create" Oct 14 09:01:06 crc kubenswrapper[4870]: I1014 09:01:06.964260 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad2fe341-5486-4fec-b744-e294b5d034bc" containerName="mariadb-account-create" Oct 14 09:01:06 crc kubenswrapper[4870]: I1014 09:01:06.964478 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="ec2b912e-ff75-44ed-b862-6bffd46f0ddf" containerName="keystone-cron" Oct 14 09:01:06 crc kubenswrapper[4870]: I1014 09:01:06.964511 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="ad2fe341-5486-4fec-b744-e294b5d034bc" containerName="mariadb-account-create" Oct 14 09:01:06 crc kubenswrapper[4870]: I1014 09:01:06.965239 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-x7dzz" Oct 14 09:01:06 crc kubenswrapper[4870]: I1014 09:01:06.969556 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Oct 14 09:01:06 crc kubenswrapper[4870]: I1014 09:01:06.969867 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-4nb8d" Oct 14 09:01:06 crc kubenswrapper[4870]: I1014 09:01:06.970131 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Oct 14 09:01:06 crc kubenswrapper[4870]: I1014 09:01:06.982664 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-x7dzz"] Oct 14 09:01:07 crc kubenswrapper[4870]: I1014 09:01:07.132197 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/a5cfe53c-f2e3-4660-a84b-e33429acfc0f-db-sync-config-data\") pod \"cinder-db-sync-x7dzz\" (UID: \"a5cfe53c-f2e3-4660-a84b-e33429acfc0f\") " pod="openstack/cinder-db-sync-x7dzz" Oct 14 09:01:07 crc kubenswrapper[4870]: I1014 09:01:07.132399 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a5cfe53c-f2e3-4660-a84b-e33429acfc0f-scripts\") pod \"cinder-db-sync-x7dzz\" (UID: \"a5cfe53c-f2e3-4660-a84b-e33429acfc0f\") " pod="openstack/cinder-db-sync-x7dzz" Oct 14 09:01:07 crc kubenswrapper[4870]: I1014 09:01:07.132550 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a5cfe53c-f2e3-4660-a84b-e33429acfc0f-etc-machine-id\") pod \"cinder-db-sync-x7dzz\" (UID: \"a5cfe53c-f2e3-4660-a84b-e33429acfc0f\") " pod="openstack/cinder-db-sync-x7dzz" Oct 14 09:01:07 crc kubenswrapper[4870]: I1014 09:01:07.132694 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hb4f4\" (UniqueName: \"kubernetes.io/projected/a5cfe53c-f2e3-4660-a84b-e33429acfc0f-kube-api-access-hb4f4\") pod \"cinder-db-sync-x7dzz\" (UID: \"a5cfe53c-f2e3-4660-a84b-e33429acfc0f\") " pod="openstack/cinder-db-sync-x7dzz" Oct 14 09:01:07 crc kubenswrapper[4870]: I1014 09:01:07.132794 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a5cfe53c-f2e3-4660-a84b-e33429acfc0f-config-data\") pod \"cinder-db-sync-x7dzz\" (UID: \"a5cfe53c-f2e3-4660-a84b-e33429acfc0f\") " pod="openstack/cinder-db-sync-x7dzz" Oct 14 09:01:07 crc kubenswrapper[4870]: I1014 09:01:07.132823 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5cfe53c-f2e3-4660-a84b-e33429acfc0f-combined-ca-bundle\") pod \"cinder-db-sync-x7dzz\" (UID: \"a5cfe53c-f2e3-4660-a84b-e33429acfc0f\") " pod="openstack/cinder-db-sync-x7dzz" Oct 14 09:01:07 crc kubenswrapper[4870]: I1014 09:01:07.234685 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hb4f4\" (UniqueName: \"kubernetes.io/projected/a5cfe53c-f2e3-4660-a84b-e33429acfc0f-kube-api-access-hb4f4\") pod \"cinder-db-sync-x7dzz\" (UID: \"a5cfe53c-f2e3-4660-a84b-e33429acfc0f\") " pod="openstack/cinder-db-sync-x7dzz" Oct 14 09:01:07 crc kubenswrapper[4870]: I1014 09:01:07.234810 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a5cfe53c-f2e3-4660-a84b-e33429acfc0f-config-data\") pod \"cinder-db-sync-x7dzz\" (UID: \"a5cfe53c-f2e3-4660-a84b-e33429acfc0f\") " pod="openstack/cinder-db-sync-x7dzz" Oct 14 09:01:07 crc kubenswrapper[4870]: I1014 09:01:07.234847 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5cfe53c-f2e3-4660-a84b-e33429acfc0f-combined-ca-bundle\") pod \"cinder-db-sync-x7dzz\" (UID: \"a5cfe53c-f2e3-4660-a84b-e33429acfc0f\") " pod="openstack/cinder-db-sync-x7dzz" Oct 14 09:01:07 crc kubenswrapper[4870]: I1014 09:01:07.234882 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/a5cfe53c-f2e3-4660-a84b-e33429acfc0f-db-sync-config-data\") pod \"cinder-db-sync-x7dzz\" (UID: \"a5cfe53c-f2e3-4660-a84b-e33429acfc0f\") " pod="openstack/cinder-db-sync-x7dzz" Oct 14 09:01:07 crc kubenswrapper[4870]: I1014 09:01:07.234929 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a5cfe53c-f2e3-4660-a84b-e33429acfc0f-scripts\") pod \"cinder-db-sync-x7dzz\" (UID: \"a5cfe53c-f2e3-4660-a84b-e33429acfc0f\") " pod="openstack/cinder-db-sync-x7dzz" Oct 14 09:01:07 crc kubenswrapper[4870]: I1014 09:01:07.234975 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a5cfe53c-f2e3-4660-a84b-e33429acfc0f-etc-machine-id\") pod \"cinder-db-sync-x7dzz\" (UID: \"a5cfe53c-f2e3-4660-a84b-e33429acfc0f\") " pod="openstack/cinder-db-sync-x7dzz" Oct 14 09:01:07 crc kubenswrapper[4870]: I1014 09:01:07.235086 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a5cfe53c-f2e3-4660-a84b-e33429acfc0f-etc-machine-id\") pod \"cinder-db-sync-x7dzz\" (UID: \"a5cfe53c-f2e3-4660-a84b-e33429acfc0f\") " pod="openstack/cinder-db-sync-x7dzz" Oct 14 09:01:07 crc kubenswrapper[4870]: I1014 09:01:07.242081 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a5cfe53c-f2e3-4660-a84b-e33429acfc0f-scripts\") pod \"cinder-db-sync-x7dzz\" (UID: \"a5cfe53c-f2e3-4660-a84b-e33429acfc0f\") " pod="openstack/cinder-db-sync-x7dzz" Oct 14 09:01:07 crc kubenswrapper[4870]: I1014 09:01:07.242428 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5cfe53c-f2e3-4660-a84b-e33429acfc0f-combined-ca-bundle\") pod \"cinder-db-sync-x7dzz\" (UID: \"a5cfe53c-f2e3-4660-a84b-e33429acfc0f\") " pod="openstack/cinder-db-sync-x7dzz" Oct 14 09:01:07 crc kubenswrapper[4870]: I1014 09:01:07.242483 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/a5cfe53c-f2e3-4660-a84b-e33429acfc0f-db-sync-config-data\") pod \"cinder-db-sync-x7dzz\" (UID: \"a5cfe53c-f2e3-4660-a84b-e33429acfc0f\") " pod="openstack/cinder-db-sync-x7dzz" Oct 14 09:01:07 crc kubenswrapper[4870]: I1014 09:01:07.243831 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a5cfe53c-f2e3-4660-a84b-e33429acfc0f-config-data\") pod \"cinder-db-sync-x7dzz\" (UID: \"a5cfe53c-f2e3-4660-a84b-e33429acfc0f\") " pod="openstack/cinder-db-sync-x7dzz" Oct 14 09:01:07 crc kubenswrapper[4870]: I1014 09:01:07.256636 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hb4f4\" (UniqueName: \"kubernetes.io/projected/a5cfe53c-f2e3-4660-a84b-e33429acfc0f-kube-api-access-hb4f4\") pod \"cinder-db-sync-x7dzz\" (UID: \"a5cfe53c-f2e3-4660-a84b-e33429acfc0f\") " pod="openstack/cinder-db-sync-x7dzz" Oct 14 09:01:07 crc kubenswrapper[4870]: I1014 09:01:07.285399 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-x7dzz" Oct 14 09:01:07 crc kubenswrapper[4870]: I1014 09:01:07.631769 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-x7dzz"] Oct 14 09:01:07 crc kubenswrapper[4870]: I1014 09:01:07.665789 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-x7dzz" event={"ID":"a5cfe53c-f2e3-4660-a84b-e33429acfc0f","Type":"ContainerStarted","Data":"b3379c4c512c8bd1fc20d41addd3de57e6679456c184dfa0207a4ed6141b8b92"} Oct 14 09:01:20 crc kubenswrapper[4870]: I1014 09:01:20.178593 4870 scope.go:117] "RemoveContainer" containerID="f54d4714558468e367cb7334b879aa0eafd61a63e0c81c28a522ef4839f64d39" Oct 14 09:01:23 crc kubenswrapper[4870]: I1014 09:01:23.957273 4870 patch_prober.go:28] interesting pod/machine-config-daemon-7tvc7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 09:01:23 crc kubenswrapper[4870]: I1014 09:01:23.957687 4870 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 09:01:26 crc kubenswrapper[4870]: I1014 09:01:26.974994 4870 scope.go:117] "RemoveContainer" containerID="f9ef13564e588ee8e9f11fa2197497d04269bcfbcca40379a8d841340a5750b6" Oct 14 09:01:28 crc kubenswrapper[4870]: I1014 09:01:28.863679 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-x7dzz" event={"ID":"a5cfe53c-f2e3-4660-a84b-e33429acfc0f","Type":"ContainerStarted","Data":"fcf2557eeea7af11b86a0d1b9f17fde60af84369c41c9200aed16412b714bcd1"} Oct 14 09:01:28 crc kubenswrapper[4870]: I1014 09:01:28.885309 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-x7dzz" podStartSLOduration=2.476442437 podStartE2EDuration="22.88528879s" podCreationTimestamp="2025-10-14 09:01:06 +0000 UTC" firstStartedPulling="2025-10-14 09:01:07.64390729 +0000 UTC m=+7203.341267661" lastFinishedPulling="2025-10-14 09:01:28.052753593 +0000 UTC m=+7223.750114014" observedRunningTime="2025-10-14 09:01:28.879099837 +0000 UTC m=+7224.576460288" watchObservedRunningTime="2025-10-14 09:01:28.88528879 +0000 UTC m=+7224.582649171" Oct 14 09:01:31 crc kubenswrapper[4870]: I1014 09:01:31.891668 4870 generic.go:334] "Generic (PLEG): container finished" podID="a5cfe53c-f2e3-4660-a84b-e33429acfc0f" containerID="fcf2557eeea7af11b86a0d1b9f17fde60af84369c41c9200aed16412b714bcd1" exitCode=0 Oct 14 09:01:31 crc kubenswrapper[4870]: I1014 09:01:31.891766 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-x7dzz" event={"ID":"a5cfe53c-f2e3-4660-a84b-e33429acfc0f","Type":"ContainerDied","Data":"fcf2557eeea7af11b86a0d1b9f17fde60af84369c41c9200aed16412b714bcd1"} Oct 14 09:01:33 crc kubenswrapper[4870]: I1014 09:01:33.303130 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-x7dzz" Oct 14 09:01:33 crc kubenswrapper[4870]: I1014 09:01:33.370502 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a5cfe53c-f2e3-4660-a84b-e33429acfc0f-config-data\") pod \"a5cfe53c-f2e3-4660-a84b-e33429acfc0f\" (UID: \"a5cfe53c-f2e3-4660-a84b-e33429acfc0f\") " Oct 14 09:01:33 crc kubenswrapper[4870]: I1014 09:01:33.370555 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5cfe53c-f2e3-4660-a84b-e33429acfc0f-combined-ca-bundle\") pod \"a5cfe53c-f2e3-4660-a84b-e33429acfc0f\" (UID: \"a5cfe53c-f2e3-4660-a84b-e33429acfc0f\") " Oct 14 09:01:33 crc kubenswrapper[4870]: I1014 09:01:33.370586 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/a5cfe53c-f2e3-4660-a84b-e33429acfc0f-db-sync-config-data\") pod \"a5cfe53c-f2e3-4660-a84b-e33429acfc0f\" (UID: \"a5cfe53c-f2e3-4660-a84b-e33429acfc0f\") " Oct 14 09:01:33 crc kubenswrapper[4870]: I1014 09:01:33.370639 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a5cfe53c-f2e3-4660-a84b-e33429acfc0f-scripts\") pod \"a5cfe53c-f2e3-4660-a84b-e33429acfc0f\" (UID: \"a5cfe53c-f2e3-4660-a84b-e33429acfc0f\") " Oct 14 09:01:33 crc kubenswrapper[4870]: I1014 09:01:33.370663 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a5cfe53c-f2e3-4660-a84b-e33429acfc0f-etc-machine-id\") pod \"a5cfe53c-f2e3-4660-a84b-e33429acfc0f\" (UID: \"a5cfe53c-f2e3-4660-a84b-e33429acfc0f\") " Oct 14 09:01:33 crc kubenswrapper[4870]: I1014 09:01:33.370751 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hb4f4\" (UniqueName: \"kubernetes.io/projected/a5cfe53c-f2e3-4660-a84b-e33429acfc0f-kube-api-access-hb4f4\") pod \"a5cfe53c-f2e3-4660-a84b-e33429acfc0f\" (UID: \"a5cfe53c-f2e3-4660-a84b-e33429acfc0f\") " Oct 14 09:01:33 crc kubenswrapper[4870]: I1014 09:01:33.371569 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a5cfe53c-f2e3-4660-a84b-e33429acfc0f-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "a5cfe53c-f2e3-4660-a84b-e33429acfc0f" (UID: "a5cfe53c-f2e3-4660-a84b-e33429acfc0f"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 14 09:01:33 crc kubenswrapper[4870]: I1014 09:01:33.377850 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a5cfe53c-f2e3-4660-a84b-e33429acfc0f-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "a5cfe53c-f2e3-4660-a84b-e33429acfc0f" (UID: "a5cfe53c-f2e3-4660-a84b-e33429acfc0f"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:01:33 crc kubenswrapper[4870]: I1014 09:01:33.377896 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a5cfe53c-f2e3-4660-a84b-e33429acfc0f-scripts" (OuterVolumeSpecName: "scripts") pod "a5cfe53c-f2e3-4660-a84b-e33429acfc0f" (UID: "a5cfe53c-f2e3-4660-a84b-e33429acfc0f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:01:33 crc kubenswrapper[4870]: I1014 09:01:33.378687 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a5cfe53c-f2e3-4660-a84b-e33429acfc0f-kube-api-access-hb4f4" (OuterVolumeSpecName: "kube-api-access-hb4f4") pod "a5cfe53c-f2e3-4660-a84b-e33429acfc0f" (UID: "a5cfe53c-f2e3-4660-a84b-e33429acfc0f"). InnerVolumeSpecName "kube-api-access-hb4f4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:01:33 crc kubenswrapper[4870]: I1014 09:01:33.408436 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a5cfe53c-f2e3-4660-a84b-e33429acfc0f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a5cfe53c-f2e3-4660-a84b-e33429acfc0f" (UID: "a5cfe53c-f2e3-4660-a84b-e33429acfc0f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:01:33 crc kubenswrapper[4870]: I1014 09:01:33.439488 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a5cfe53c-f2e3-4660-a84b-e33429acfc0f-config-data" (OuterVolumeSpecName: "config-data") pod "a5cfe53c-f2e3-4660-a84b-e33429acfc0f" (UID: "a5cfe53c-f2e3-4660-a84b-e33429acfc0f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:01:33 crc kubenswrapper[4870]: I1014 09:01:33.473520 4870 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a5cfe53c-f2e3-4660-a84b-e33429acfc0f-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 09:01:33 crc kubenswrapper[4870]: I1014 09:01:33.473576 4870 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5cfe53c-f2e3-4660-a84b-e33429acfc0f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 09:01:33 crc kubenswrapper[4870]: I1014 09:01:33.473605 4870 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/a5cfe53c-f2e3-4660-a84b-e33429acfc0f-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 09:01:33 crc kubenswrapper[4870]: I1014 09:01:33.473624 4870 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a5cfe53c-f2e3-4660-a84b-e33429acfc0f-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 09:01:33 crc kubenswrapper[4870]: I1014 09:01:33.473641 4870 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a5cfe53c-f2e3-4660-a84b-e33429acfc0f-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 14 09:01:33 crc kubenswrapper[4870]: I1014 09:01:33.473659 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hb4f4\" (UniqueName: \"kubernetes.io/projected/a5cfe53c-f2e3-4660-a84b-e33429acfc0f-kube-api-access-hb4f4\") on node \"crc\" DevicePath \"\"" Oct 14 09:01:33 crc kubenswrapper[4870]: I1014 09:01:33.915789 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-x7dzz" event={"ID":"a5cfe53c-f2e3-4660-a84b-e33429acfc0f","Type":"ContainerDied","Data":"b3379c4c512c8bd1fc20d41addd3de57e6679456c184dfa0207a4ed6141b8b92"} Oct 14 09:01:33 crc kubenswrapper[4870]: I1014 09:01:33.915835 4870 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b3379c4c512c8bd1fc20d41addd3de57e6679456c184dfa0207a4ed6141b8b92" Oct 14 09:01:33 crc kubenswrapper[4870]: I1014 09:01:33.915915 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-x7dzz" Oct 14 09:01:34 crc kubenswrapper[4870]: I1014 09:01:34.205159 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6b44c8b5f5-47x86"] Oct 14 09:01:34 crc kubenswrapper[4870]: E1014 09:01:34.206096 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5cfe53c-f2e3-4660-a84b-e33429acfc0f" containerName="cinder-db-sync" Oct 14 09:01:34 crc kubenswrapper[4870]: I1014 09:01:34.206112 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5cfe53c-f2e3-4660-a84b-e33429acfc0f" containerName="cinder-db-sync" Oct 14 09:01:34 crc kubenswrapper[4870]: I1014 09:01:34.206287 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="a5cfe53c-f2e3-4660-a84b-e33429acfc0f" containerName="cinder-db-sync" Oct 14 09:01:34 crc kubenswrapper[4870]: I1014 09:01:34.207293 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b44c8b5f5-47x86" Oct 14 09:01:34 crc kubenswrapper[4870]: I1014 09:01:34.229411 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6b44c8b5f5-47x86"] Oct 14 09:01:34 crc kubenswrapper[4870]: I1014 09:01:34.288556 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d666d204-b974-4c9a-aaba-6d02302fc332-config\") pod \"dnsmasq-dns-6b44c8b5f5-47x86\" (UID: \"d666d204-b974-4c9a-aaba-6d02302fc332\") " pod="openstack/dnsmasq-dns-6b44c8b5f5-47x86" Oct 14 09:01:34 crc kubenswrapper[4870]: I1014 09:01:34.288644 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d666d204-b974-4c9a-aaba-6d02302fc332-ovsdbserver-nb\") pod \"dnsmasq-dns-6b44c8b5f5-47x86\" (UID: \"d666d204-b974-4c9a-aaba-6d02302fc332\") " pod="openstack/dnsmasq-dns-6b44c8b5f5-47x86" Oct 14 09:01:34 crc kubenswrapper[4870]: I1014 09:01:34.288785 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d666d204-b974-4c9a-aaba-6d02302fc332-dns-svc\") pod \"dnsmasq-dns-6b44c8b5f5-47x86\" (UID: \"d666d204-b974-4c9a-aaba-6d02302fc332\") " pod="openstack/dnsmasq-dns-6b44c8b5f5-47x86" Oct 14 09:01:34 crc kubenswrapper[4870]: I1014 09:01:34.288931 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2qzvg\" (UniqueName: \"kubernetes.io/projected/d666d204-b974-4c9a-aaba-6d02302fc332-kube-api-access-2qzvg\") pod \"dnsmasq-dns-6b44c8b5f5-47x86\" (UID: \"d666d204-b974-4c9a-aaba-6d02302fc332\") " pod="openstack/dnsmasq-dns-6b44c8b5f5-47x86" Oct 14 09:01:34 crc kubenswrapper[4870]: I1014 09:01:34.288972 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d666d204-b974-4c9a-aaba-6d02302fc332-ovsdbserver-sb\") pod \"dnsmasq-dns-6b44c8b5f5-47x86\" (UID: \"d666d204-b974-4c9a-aaba-6d02302fc332\") " pod="openstack/dnsmasq-dns-6b44c8b5f5-47x86" Oct 14 09:01:34 crc kubenswrapper[4870]: I1014 09:01:34.390243 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2qzvg\" (UniqueName: \"kubernetes.io/projected/d666d204-b974-4c9a-aaba-6d02302fc332-kube-api-access-2qzvg\") pod \"dnsmasq-dns-6b44c8b5f5-47x86\" (UID: \"d666d204-b974-4c9a-aaba-6d02302fc332\") " pod="openstack/dnsmasq-dns-6b44c8b5f5-47x86" Oct 14 09:01:34 crc kubenswrapper[4870]: I1014 09:01:34.390310 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d666d204-b974-4c9a-aaba-6d02302fc332-ovsdbserver-sb\") pod \"dnsmasq-dns-6b44c8b5f5-47x86\" (UID: \"d666d204-b974-4c9a-aaba-6d02302fc332\") " pod="openstack/dnsmasq-dns-6b44c8b5f5-47x86" Oct 14 09:01:34 crc kubenswrapper[4870]: I1014 09:01:34.390394 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d666d204-b974-4c9a-aaba-6d02302fc332-config\") pod \"dnsmasq-dns-6b44c8b5f5-47x86\" (UID: \"d666d204-b974-4c9a-aaba-6d02302fc332\") " pod="openstack/dnsmasq-dns-6b44c8b5f5-47x86" Oct 14 09:01:34 crc kubenswrapper[4870]: I1014 09:01:34.390460 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d666d204-b974-4c9a-aaba-6d02302fc332-ovsdbserver-nb\") pod \"dnsmasq-dns-6b44c8b5f5-47x86\" (UID: \"d666d204-b974-4c9a-aaba-6d02302fc332\") " pod="openstack/dnsmasq-dns-6b44c8b5f5-47x86" Oct 14 09:01:34 crc kubenswrapper[4870]: I1014 09:01:34.390494 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d666d204-b974-4c9a-aaba-6d02302fc332-dns-svc\") pod \"dnsmasq-dns-6b44c8b5f5-47x86\" (UID: \"d666d204-b974-4c9a-aaba-6d02302fc332\") " pod="openstack/dnsmasq-dns-6b44c8b5f5-47x86" Oct 14 09:01:34 crc kubenswrapper[4870]: I1014 09:01:34.391509 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d666d204-b974-4c9a-aaba-6d02302fc332-config\") pod \"dnsmasq-dns-6b44c8b5f5-47x86\" (UID: \"d666d204-b974-4c9a-aaba-6d02302fc332\") " pod="openstack/dnsmasq-dns-6b44c8b5f5-47x86" Oct 14 09:01:34 crc kubenswrapper[4870]: I1014 09:01:34.391778 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d666d204-b974-4c9a-aaba-6d02302fc332-dns-svc\") pod \"dnsmasq-dns-6b44c8b5f5-47x86\" (UID: \"d666d204-b974-4c9a-aaba-6d02302fc332\") " pod="openstack/dnsmasq-dns-6b44c8b5f5-47x86" Oct 14 09:01:34 crc kubenswrapper[4870]: I1014 09:01:34.391983 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d666d204-b974-4c9a-aaba-6d02302fc332-ovsdbserver-nb\") pod \"dnsmasq-dns-6b44c8b5f5-47x86\" (UID: \"d666d204-b974-4c9a-aaba-6d02302fc332\") " pod="openstack/dnsmasq-dns-6b44c8b5f5-47x86" Oct 14 09:01:34 crc kubenswrapper[4870]: I1014 09:01:34.392021 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d666d204-b974-4c9a-aaba-6d02302fc332-ovsdbserver-sb\") pod \"dnsmasq-dns-6b44c8b5f5-47x86\" (UID: \"d666d204-b974-4c9a-aaba-6d02302fc332\") " pod="openstack/dnsmasq-dns-6b44c8b5f5-47x86" Oct 14 09:01:34 crc kubenswrapper[4870]: I1014 09:01:34.416391 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Oct 14 09:01:34 crc kubenswrapper[4870]: I1014 09:01:34.418328 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 14 09:01:34 crc kubenswrapper[4870]: I1014 09:01:34.419327 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2qzvg\" (UniqueName: \"kubernetes.io/projected/d666d204-b974-4c9a-aaba-6d02302fc332-kube-api-access-2qzvg\") pod \"dnsmasq-dns-6b44c8b5f5-47x86\" (UID: \"d666d204-b974-4c9a-aaba-6d02302fc332\") " pod="openstack/dnsmasq-dns-6b44c8b5f5-47x86" Oct 14 09:01:34 crc kubenswrapper[4870]: I1014 09:01:34.424229 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Oct 14 09:01:34 crc kubenswrapper[4870]: I1014 09:01:34.424449 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Oct 14 09:01:34 crc kubenswrapper[4870]: I1014 09:01:34.424593 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Oct 14 09:01:34 crc kubenswrapper[4870]: I1014 09:01:34.424864 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-4nb8d" Oct 14 09:01:34 crc kubenswrapper[4870]: I1014 09:01:34.432770 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 14 09:01:34 crc kubenswrapper[4870]: I1014 09:01:34.540210 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b44c8b5f5-47x86" Oct 14 09:01:34 crc kubenswrapper[4870]: I1014 09:01:34.593511 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a662858e-76ac-4bf6-949d-4bc17f6a3aaf-etc-machine-id\") pod \"cinder-api-0\" (UID: \"a662858e-76ac-4bf6-949d-4bc17f6a3aaf\") " pod="openstack/cinder-api-0" Oct 14 09:01:34 crc kubenswrapper[4870]: I1014 09:01:34.593885 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a662858e-76ac-4bf6-949d-4bc17f6a3aaf-logs\") pod \"cinder-api-0\" (UID: \"a662858e-76ac-4bf6-949d-4bc17f6a3aaf\") " pod="openstack/cinder-api-0" Oct 14 09:01:34 crc kubenswrapper[4870]: I1014 09:01:34.593928 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a662858e-76ac-4bf6-949d-4bc17f6a3aaf-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"a662858e-76ac-4bf6-949d-4bc17f6a3aaf\") " pod="openstack/cinder-api-0" Oct 14 09:01:34 crc kubenswrapper[4870]: I1014 09:01:34.593992 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a662858e-76ac-4bf6-949d-4bc17f6a3aaf-config-data-custom\") pod \"cinder-api-0\" (UID: \"a662858e-76ac-4bf6-949d-4bc17f6a3aaf\") " pod="openstack/cinder-api-0" Oct 14 09:01:34 crc kubenswrapper[4870]: I1014 09:01:34.594043 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a662858e-76ac-4bf6-949d-4bc17f6a3aaf-config-data\") pod \"cinder-api-0\" (UID: \"a662858e-76ac-4bf6-949d-4bc17f6a3aaf\") " pod="openstack/cinder-api-0" Oct 14 09:01:34 crc kubenswrapper[4870]: I1014 09:01:34.594060 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a662858e-76ac-4bf6-949d-4bc17f6a3aaf-scripts\") pod \"cinder-api-0\" (UID: \"a662858e-76ac-4bf6-949d-4bc17f6a3aaf\") " pod="openstack/cinder-api-0" Oct 14 09:01:34 crc kubenswrapper[4870]: I1014 09:01:34.594115 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g7jbh\" (UniqueName: \"kubernetes.io/projected/a662858e-76ac-4bf6-949d-4bc17f6a3aaf-kube-api-access-g7jbh\") pod \"cinder-api-0\" (UID: \"a662858e-76ac-4bf6-949d-4bc17f6a3aaf\") " pod="openstack/cinder-api-0" Oct 14 09:01:34 crc kubenswrapper[4870]: I1014 09:01:34.696527 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a662858e-76ac-4bf6-949d-4bc17f6a3aaf-config-data-custom\") pod \"cinder-api-0\" (UID: \"a662858e-76ac-4bf6-949d-4bc17f6a3aaf\") " pod="openstack/cinder-api-0" Oct 14 09:01:34 crc kubenswrapper[4870]: I1014 09:01:34.696598 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a662858e-76ac-4bf6-949d-4bc17f6a3aaf-config-data\") pod \"cinder-api-0\" (UID: \"a662858e-76ac-4bf6-949d-4bc17f6a3aaf\") " pod="openstack/cinder-api-0" Oct 14 09:01:34 crc kubenswrapper[4870]: I1014 09:01:34.696615 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a662858e-76ac-4bf6-949d-4bc17f6a3aaf-scripts\") pod \"cinder-api-0\" (UID: \"a662858e-76ac-4bf6-949d-4bc17f6a3aaf\") " pod="openstack/cinder-api-0" Oct 14 09:01:34 crc kubenswrapper[4870]: I1014 09:01:34.696667 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g7jbh\" (UniqueName: \"kubernetes.io/projected/a662858e-76ac-4bf6-949d-4bc17f6a3aaf-kube-api-access-g7jbh\") pod \"cinder-api-0\" (UID: \"a662858e-76ac-4bf6-949d-4bc17f6a3aaf\") " pod="openstack/cinder-api-0" Oct 14 09:01:34 crc kubenswrapper[4870]: I1014 09:01:34.696690 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a662858e-76ac-4bf6-949d-4bc17f6a3aaf-etc-machine-id\") pod \"cinder-api-0\" (UID: \"a662858e-76ac-4bf6-949d-4bc17f6a3aaf\") " pod="openstack/cinder-api-0" Oct 14 09:01:34 crc kubenswrapper[4870]: I1014 09:01:34.696704 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a662858e-76ac-4bf6-949d-4bc17f6a3aaf-logs\") pod \"cinder-api-0\" (UID: \"a662858e-76ac-4bf6-949d-4bc17f6a3aaf\") " pod="openstack/cinder-api-0" Oct 14 09:01:34 crc kubenswrapper[4870]: I1014 09:01:34.696731 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a662858e-76ac-4bf6-949d-4bc17f6a3aaf-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"a662858e-76ac-4bf6-949d-4bc17f6a3aaf\") " pod="openstack/cinder-api-0" Oct 14 09:01:34 crc kubenswrapper[4870]: I1014 09:01:34.699550 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a662858e-76ac-4bf6-949d-4bc17f6a3aaf-etc-machine-id\") pod \"cinder-api-0\" (UID: \"a662858e-76ac-4bf6-949d-4bc17f6a3aaf\") " pod="openstack/cinder-api-0" Oct 14 09:01:34 crc kubenswrapper[4870]: I1014 09:01:34.700150 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a662858e-76ac-4bf6-949d-4bc17f6a3aaf-logs\") pod \"cinder-api-0\" (UID: \"a662858e-76ac-4bf6-949d-4bc17f6a3aaf\") " pod="openstack/cinder-api-0" Oct 14 09:01:34 crc kubenswrapper[4870]: I1014 09:01:34.703105 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a662858e-76ac-4bf6-949d-4bc17f6a3aaf-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"a662858e-76ac-4bf6-949d-4bc17f6a3aaf\") " pod="openstack/cinder-api-0" Oct 14 09:01:34 crc kubenswrapper[4870]: I1014 09:01:34.703572 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a662858e-76ac-4bf6-949d-4bc17f6a3aaf-config-data\") pod \"cinder-api-0\" (UID: \"a662858e-76ac-4bf6-949d-4bc17f6a3aaf\") " pod="openstack/cinder-api-0" Oct 14 09:01:34 crc kubenswrapper[4870]: I1014 09:01:34.704275 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a662858e-76ac-4bf6-949d-4bc17f6a3aaf-config-data-custom\") pod \"cinder-api-0\" (UID: \"a662858e-76ac-4bf6-949d-4bc17f6a3aaf\") " pod="openstack/cinder-api-0" Oct 14 09:01:34 crc kubenswrapper[4870]: I1014 09:01:34.712911 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a662858e-76ac-4bf6-949d-4bc17f6a3aaf-scripts\") pod \"cinder-api-0\" (UID: \"a662858e-76ac-4bf6-949d-4bc17f6a3aaf\") " pod="openstack/cinder-api-0" Oct 14 09:01:34 crc kubenswrapper[4870]: I1014 09:01:34.715900 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g7jbh\" (UniqueName: \"kubernetes.io/projected/a662858e-76ac-4bf6-949d-4bc17f6a3aaf-kube-api-access-g7jbh\") pod \"cinder-api-0\" (UID: \"a662858e-76ac-4bf6-949d-4bc17f6a3aaf\") " pod="openstack/cinder-api-0" Oct 14 09:01:34 crc kubenswrapper[4870]: I1014 09:01:34.768034 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 14 09:01:35 crc kubenswrapper[4870]: I1014 09:01:35.056252 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6b44c8b5f5-47x86"] Oct 14 09:01:35 crc kubenswrapper[4870]: I1014 09:01:35.237631 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 14 09:01:35 crc kubenswrapper[4870]: W1014 09:01:35.242143 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda662858e_76ac_4bf6_949d_4bc17f6a3aaf.slice/crio-f1305639c736356705e48b686bd47788e0f4c6e5c9376968a8b98d594517b5c3 WatchSource:0}: Error finding container f1305639c736356705e48b686bd47788e0f4c6e5c9376968a8b98d594517b5c3: Status 404 returned error can't find the container with id f1305639c736356705e48b686bd47788e0f4c6e5c9376968a8b98d594517b5c3 Oct 14 09:01:36 crc kubenswrapper[4870]: I1014 09:01:36.005229 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"a662858e-76ac-4bf6-949d-4bc17f6a3aaf","Type":"ContainerStarted","Data":"8747381c748307eddcef853ed8b5d3128671bd63c249dac6069447d3067eabb8"} Oct 14 09:01:36 crc kubenswrapper[4870]: I1014 09:01:36.005955 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"a662858e-76ac-4bf6-949d-4bc17f6a3aaf","Type":"ContainerStarted","Data":"f1305639c736356705e48b686bd47788e0f4c6e5c9376968a8b98d594517b5c3"} Oct 14 09:01:36 crc kubenswrapper[4870]: I1014 09:01:36.007682 4870 generic.go:334] "Generic (PLEG): container finished" podID="d666d204-b974-4c9a-aaba-6d02302fc332" containerID="99dca1d602683adf2005164f939b7d374fb1c110208d96e73dd1db9ad628d9d1" exitCode=0 Oct 14 09:01:36 crc kubenswrapper[4870]: I1014 09:01:36.007894 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b44c8b5f5-47x86" event={"ID":"d666d204-b974-4c9a-aaba-6d02302fc332","Type":"ContainerDied","Data":"99dca1d602683adf2005164f939b7d374fb1c110208d96e73dd1db9ad628d9d1"} Oct 14 09:01:36 crc kubenswrapper[4870]: I1014 09:01:36.007921 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b44c8b5f5-47x86" event={"ID":"d666d204-b974-4c9a-aaba-6d02302fc332","Type":"ContainerStarted","Data":"2f090c359b4dfbaf97b1ac18905a081b8a4a20d8ac2fb08f4290e4ec192971e3"} Oct 14 09:01:37 crc kubenswrapper[4870]: I1014 09:01:37.030835 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"a662858e-76ac-4bf6-949d-4bc17f6a3aaf","Type":"ContainerStarted","Data":"9f71bfbc3b3893910374afed0b2400760747c481120e966d946e4c2dab9b0550"} Oct 14 09:01:37 crc kubenswrapper[4870]: I1014 09:01:37.031205 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Oct 14 09:01:37 crc kubenswrapper[4870]: I1014 09:01:37.056749 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=3.056722499 podStartE2EDuration="3.056722499s" podCreationTimestamp="2025-10-14 09:01:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 09:01:37.048939817 +0000 UTC m=+7232.746300188" watchObservedRunningTime="2025-10-14 09:01:37.056722499 +0000 UTC m=+7232.754082870" Oct 14 09:01:37 crc kubenswrapper[4870]: I1014 09:01:37.063765 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b44c8b5f5-47x86" event={"ID":"d666d204-b974-4c9a-aaba-6d02302fc332","Type":"ContainerStarted","Data":"654085a705d3044a54f19852baa77fbbc5a4c6cb8bff2383483e23651297a478"} Oct 14 09:01:37 crc kubenswrapper[4870]: I1014 09:01:37.063949 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6b44c8b5f5-47x86" Oct 14 09:01:37 crc kubenswrapper[4870]: I1014 09:01:37.073176 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6b44c8b5f5-47x86" podStartSLOduration=3.073149275 podStartE2EDuration="3.073149275s" podCreationTimestamp="2025-10-14 09:01:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 09:01:37.067758412 +0000 UTC m=+7232.765118783" watchObservedRunningTime="2025-10-14 09:01:37.073149275 +0000 UTC m=+7232.770509646" Oct 14 09:01:44 crc kubenswrapper[4870]: I1014 09:01:44.541678 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6b44c8b5f5-47x86" Oct 14 09:01:44 crc kubenswrapper[4870]: I1014 09:01:44.640874 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5b4495fb7f-cw5zs"] Oct 14 09:01:44 crc kubenswrapper[4870]: I1014 09:01:44.641216 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5b4495fb7f-cw5zs" podUID="935482d1-d09a-4c56-9b4d-cfd2d7897820" containerName="dnsmasq-dns" containerID="cri-o://f6a6d0a40aa24e08080ff7782a1374691a10405c3eb48f5a12139f0887eec7e5" gracePeriod=10 Oct 14 09:01:45 crc kubenswrapper[4870]: I1014 09:01:45.142933 4870 generic.go:334] "Generic (PLEG): container finished" podID="935482d1-d09a-4c56-9b4d-cfd2d7897820" containerID="f6a6d0a40aa24e08080ff7782a1374691a10405c3eb48f5a12139f0887eec7e5" exitCode=0 Oct 14 09:01:45 crc kubenswrapper[4870]: I1014 09:01:45.143305 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b4495fb7f-cw5zs" event={"ID":"935482d1-d09a-4c56-9b4d-cfd2d7897820","Type":"ContainerDied","Data":"f6a6d0a40aa24e08080ff7782a1374691a10405c3eb48f5a12139f0887eec7e5"} Oct 14 09:01:45 crc kubenswrapper[4870]: I1014 09:01:45.143333 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b4495fb7f-cw5zs" event={"ID":"935482d1-d09a-4c56-9b4d-cfd2d7897820","Type":"ContainerDied","Data":"e8d0775880cb586d7c87ede08d0c5960cc045508f80101cc9226fea6d0987291"} Oct 14 09:01:45 crc kubenswrapper[4870]: I1014 09:01:45.143344 4870 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e8d0775880cb586d7c87ede08d0c5960cc045508f80101cc9226fea6d0987291" Oct 14 09:01:45 crc kubenswrapper[4870]: I1014 09:01:45.174125 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b4495fb7f-cw5zs" Oct 14 09:01:45 crc kubenswrapper[4870]: I1014 09:01:45.250632 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/935482d1-d09a-4c56-9b4d-cfd2d7897820-config\") pod \"935482d1-d09a-4c56-9b4d-cfd2d7897820\" (UID: \"935482d1-d09a-4c56-9b4d-cfd2d7897820\") " Oct 14 09:01:45 crc kubenswrapper[4870]: I1014 09:01:45.250690 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/935482d1-d09a-4c56-9b4d-cfd2d7897820-dns-svc\") pod \"935482d1-d09a-4c56-9b4d-cfd2d7897820\" (UID: \"935482d1-d09a-4c56-9b4d-cfd2d7897820\") " Oct 14 09:01:45 crc kubenswrapper[4870]: I1014 09:01:45.250746 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/935482d1-d09a-4c56-9b4d-cfd2d7897820-ovsdbserver-nb\") pod \"935482d1-d09a-4c56-9b4d-cfd2d7897820\" (UID: \"935482d1-d09a-4c56-9b4d-cfd2d7897820\") " Oct 14 09:01:45 crc kubenswrapper[4870]: I1014 09:01:45.250801 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2rjmz\" (UniqueName: \"kubernetes.io/projected/935482d1-d09a-4c56-9b4d-cfd2d7897820-kube-api-access-2rjmz\") pod \"935482d1-d09a-4c56-9b4d-cfd2d7897820\" (UID: \"935482d1-d09a-4c56-9b4d-cfd2d7897820\") " Oct 14 09:01:45 crc kubenswrapper[4870]: I1014 09:01:45.250921 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/935482d1-d09a-4c56-9b4d-cfd2d7897820-ovsdbserver-sb\") pod \"935482d1-d09a-4c56-9b4d-cfd2d7897820\" (UID: \"935482d1-d09a-4c56-9b4d-cfd2d7897820\") " Oct 14 09:01:45 crc kubenswrapper[4870]: I1014 09:01:45.265668 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/935482d1-d09a-4c56-9b4d-cfd2d7897820-kube-api-access-2rjmz" (OuterVolumeSpecName: "kube-api-access-2rjmz") pod "935482d1-d09a-4c56-9b4d-cfd2d7897820" (UID: "935482d1-d09a-4c56-9b4d-cfd2d7897820"). InnerVolumeSpecName "kube-api-access-2rjmz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:01:45 crc kubenswrapper[4870]: I1014 09:01:45.309617 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/935482d1-d09a-4c56-9b4d-cfd2d7897820-config" (OuterVolumeSpecName: "config") pod "935482d1-d09a-4c56-9b4d-cfd2d7897820" (UID: "935482d1-d09a-4c56-9b4d-cfd2d7897820"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 09:01:45 crc kubenswrapper[4870]: I1014 09:01:45.310326 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/935482d1-d09a-4c56-9b4d-cfd2d7897820-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "935482d1-d09a-4c56-9b4d-cfd2d7897820" (UID: "935482d1-d09a-4c56-9b4d-cfd2d7897820"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 09:01:45 crc kubenswrapper[4870]: I1014 09:01:45.313159 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/935482d1-d09a-4c56-9b4d-cfd2d7897820-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "935482d1-d09a-4c56-9b4d-cfd2d7897820" (UID: "935482d1-d09a-4c56-9b4d-cfd2d7897820"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 09:01:45 crc kubenswrapper[4870]: I1014 09:01:45.330327 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/935482d1-d09a-4c56-9b4d-cfd2d7897820-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "935482d1-d09a-4c56-9b4d-cfd2d7897820" (UID: "935482d1-d09a-4c56-9b4d-cfd2d7897820"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 09:01:45 crc kubenswrapper[4870]: I1014 09:01:45.356392 4870 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/935482d1-d09a-4c56-9b4d-cfd2d7897820-config\") on node \"crc\" DevicePath \"\"" Oct 14 09:01:45 crc kubenswrapper[4870]: I1014 09:01:45.356863 4870 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/935482d1-d09a-4c56-9b4d-cfd2d7897820-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 14 09:01:45 crc kubenswrapper[4870]: I1014 09:01:45.356882 4870 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/935482d1-d09a-4c56-9b4d-cfd2d7897820-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 14 09:01:45 crc kubenswrapper[4870]: I1014 09:01:45.356894 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2rjmz\" (UniqueName: \"kubernetes.io/projected/935482d1-d09a-4c56-9b4d-cfd2d7897820-kube-api-access-2rjmz\") on node \"crc\" DevicePath \"\"" Oct 14 09:01:45 crc kubenswrapper[4870]: I1014 09:01:45.356905 4870 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/935482d1-d09a-4c56-9b4d-cfd2d7897820-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 14 09:01:46 crc kubenswrapper[4870]: I1014 09:01:46.165562 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b4495fb7f-cw5zs" Oct 14 09:01:46 crc kubenswrapper[4870]: I1014 09:01:46.243510 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5b4495fb7f-cw5zs"] Oct 14 09:01:46 crc kubenswrapper[4870]: I1014 09:01:46.262892 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5b4495fb7f-cw5zs"] Oct 14 09:01:46 crc kubenswrapper[4870]: I1014 09:01:46.460955 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 14 09:01:46 crc kubenswrapper[4870]: I1014 09:01:46.461167 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="1a29f723-aa05-46bd-a2f9-a4422fe7646f" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://6344e90ea632210985754aaf112afea23aaeab912ffe8717f5dd2f5ee1b46948" gracePeriod=30 Oct 14 09:01:46 crc kubenswrapper[4870]: I1014 09:01:46.480495 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 14 09:01:46 crc kubenswrapper[4870]: I1014 09:01:46.480737 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell0-conductor-0" podUID="200f8a16-57ae-44d5-bfb3-ba41de932542" containerName="nova-cell0-conductor-conductor" containerID="cri-o://07e0333d969b88cecad4b5dfbe928bb218113231bd5e69a5da72d866efa3ff49" gracePeriod=30 Oct 14 09:01:46 crc kubenswrapper[4870]: I1014 09:01:46.483499 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 14 09:01:46 crc kubenswrapper[4870]: I1014 09:01:46.483876 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="337148f2-88d7-4694-947e-de1aa4375d33" containerName="nova-metadata-log" containerID="cri-o://5968c04b35057ebe36c338a812baba14f9fa27c198995f3c2f308c83f14b2950" gracePeriod=30 Oct 14 09:01:46 crc kubenswrapper[4870]: I1014 09:01:46.485819 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="337148f2-88d7-4694-947e-de1aa4375d33" containerName="nova-metadata-metadata" containerID="cri-o://afb0153dee62d6c5510c9ee41556a4075a587da0d89753f84a48becc83a89047" gracePeriod=30 Oct 14 09:01:46 crc kubenswrapper[4870]: I1014 09:01:46.496696 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 14 09:01:46 crc kubenswrapper[4870]: I1014 09:01:46.497138 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="d3db1cbf-b57a-4f03-bde9-25dc20acd3bb" containerName="nova-api-log" containerID="cri-o://f01114c13eca471447c03dcafa83c119dfd75fd57f5778b0f86074ae10416aac" gracePeriod=30 Oct 14 09:01:46 crc kubenswrapper[4870]: I1014 09:01:46.497581 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="d3db1cbf-b57a-4f03-bde9-25dc20acd3bb" containerName="nova-api-api" containerID="cri-o://dcba0cf698d44da1b6af963fe253650289ce35521a5481babbc0ed0eac4ac7d8" gracePeriod=30 Oct 14 09:01:46 crc kubenswrapper[4870]: I1014 09:01:46.506985 4870 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-api-0" podUID="d3db1cbf-b57a-4f03-bde9-25dc20acd3bb" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.1.83:8774/\": EOF" Oct 14 09:01:46 crc kubenswrapper[4870]: I1014 09:01:46.507159 4870 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/nova-api-0" podUID="d3db1cbf-b57a-4f03-bde9-25dc20acd3bb" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.1.83:8774/\": EOF" Oct 14 09:01:46 crc kubenswrapper[4870]: I1014 09:01:46.507327 4870 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-api-0" podUID="d3db1cbf-b57a-4f03-bde9-25dc20acd3bb" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.1.83:8774/\": EOF" Oct 14 09:01:46 crc kubenswrapper[4870]: I1014 09:01:46.507577 4870 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/nova-api-0" podUID="d3db1cbf-b57a-4f03-bde9-25dc20acd3bb" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.1.83:8774/\": EOF" Oct 14 09:01:46 crc kubenswrapper[4870]: I1014 09:01:46.527632 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 14 09:01:46 crc kubenswrapper[4870]: I1014 09:01:46.527879 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="4eabbefe-21bd-4eba-aab2-5f9123db96fa" containerName="nova-scheduler-scheduler" containerID="cri-o://f22756363c89da54b7a27a7e9a3a91c6a3191a7f2ac7cd6b298026c27f2df43d" gracePeriod=30 Oct 14 09:01:46 crc kubenswrapper[4870]: I1014 09:01:46.728359 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Oct 14 09:01:47 crc kubenswrapper[4870]: I1014 09:01:47.052484 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="935482d1-d09a-4c56-9b4d-cfd2d7897820" path="/var/lib/kubelet/pods/935482d1-d09a-4c56-9b4d-cfd2d7897820/volumes" Oct 14 09:01:47 crc kubenswrapper[4870]: I1014 09:01:47.187216 4870 generic.go:334] "Generic (PLEG): container finished" podID="d3db1cbf-b57a-4f03-bde9-25dc20acd3bb" containerID="f01114c13eca471447c03dcafa83c119dfd75fd57f5778b0f86074ae10416aac" exitCode=143 Oct 14 09:01:47 crc kubenswrapper[4870]: I1014 09:01:47.187278 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d3db1cbf-b57a-4f03-bde9-25dc20acd3bb","Type":"ContainerDied","Data":"f01114c13eca471447c03dcafa83c119dfd75fd57f5778b0f86074ae10416aac"} Oct 14 09:01:47 crc kubenswrapper[4870]: I1014 09:01:47.192289 4870 generic.go:334] "Generic (PLEG): container finished" podID="1a29f723-aa05-46bd-a2f9-a4422fe7646f" containerID="6344e90ea632210985754aaf112afea23aaeab912ffe8717f5dd2f5ee1b46948" exitCode=0 Oct 14 09:01:47 crc kubenswrapper[4870]: I1014 09:01:47.192368 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"1a29f723-aa05-46bd-a2f9-a4422fe7646f","Type":"ContainerDied","Data":"6344e90ea632210985754aaf112afea23aaeab912ffe8717f5dd2f5ee1b46948"} Oct 14 09:01:47 crc kubenswrapper[4870]: I1014 09:01:47.192404 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"1a29f723-aa05-46bd-a2f9-a4422fe7646f","Type":"ContainerDied","Data":"47ceb2fb298fcb88f9ecd4fa43fde70f42312cc43de78369f1a6487d47c1605d"} Oct 14 09:01:47 crc kubenswrapper[4870]: I1014 09:01:47.192420 4870 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="47ceb2fb298fcb88f9ecd4fa43fde70f42312cc43de78369f1a6487d47c1605d" Oct 14 09:01:47 crc kubenswrapper[4870]: I1014 09:01:47.197540 4870 generic.go:334] "Generic (PLEG): container finished" podID="337148f2-88d7-4694-947e-de1aa4375d33" containerID="5968c04b35057ebe36c338a812baba14f9fa27c198995f3c2f308c83f14b2950" exitCode=143 Oct 14 09:01:47 crc kubenswrapper[4870]: I1014 09:01:47.197588 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"337148f2-88d7-4694-947e-de1aa4375d33","Type":"ContainerDied","Data":"5968c04b35057ebe36c338a812baba14f9fa27c198995f3c2f308c83f14b2950"} Oct 14 09:01:47 crc kubenswrapper[4870]: I1014 09:01:47.241551 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 14 09:01:47 crc kubenswrapper[4870]: I1014 09:01:47.298174 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a29f723-aa05-46bd-a2f9-a4422fe7646f-combined-ca-bundle\") pod \"1a29f723-aa05-46bd-a2f9-a4422fe7646f\" (UID: \"1a29f723-aa05-46bd-a2f9-a4422fe7646f\") " Oct 14 09:01:47 crc kubenswrapper[4870]: I1014 09:01:47.298269 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4qxfd\" (UniqueName: \"kubernetes.io/projected/1a29f723-aa05-46bd-a2f9-a4422fe7646f-kube-api-access-4qxfd\") pod \"1a29f723-aa05-46bd-a2f9-a4422fe7646f\" (UID: \"1a29f723-aa05-46bd-a2f9-a4422fe7646f\") " Oct 14 09:01:47 crc kubenswrapper[4870]: I1014 09:01:47.298335 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1a29f723-aa05-46bd-a2f9-a4422fe7646f-config-data\") pod \"1a29f723-aa05-46bd-a2f9-a4422fe7646f\" (UID: \"1a29f723-aa05-46bd-a2f9-a4422fe7646f\") " Oct 14 09:01:47 crc kubenswrapper[4870]: I1014 09:01:47.320603 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1a29f723-aa05-46bd-a2f9-a4422fe7646f-kube-api-access-4qxfd" (OuterVolumeSpecName: "kube-api-access-4qxfd") pod "1a29f723-aa05-46bd-a2f9-a4422fe7646f" (UID: "1a29f723-aa05-46bd-a2f9-a4422fe7646f"). InnerVolumeSpecName "kube-api-access-4qxfd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:01:47 crc kubenswrapper[4870]: I1014 09:01:47.337359 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a29f723-aa05-46bd-a2f9-a4422fe7646f-config-data" (OuterVolumeSpecName: "config-data") pod "1a29f723-aa05-46bd-a2f9-a4422fe7646f" (UID: "1a29f723-aa05-46bd-a2f9-a4422fe7646f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:01:47 crc kubenswrapper[4870]: I1014 09:01:47.341642 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a29f723-aa05-46bd-a2f9-a4422fe7646f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1a29f723-aa05-46bd-a2f9-a4422fe7646f" (UID: "1a29f723-aa05-46bd-a2f9-a4422fe7646f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:01:47 crc kubenswrapper[4870]: I1014 09:01:47.400544 4870 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a29f723-aa05-46bd-a2f9-a4422fe7646f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 09:01:47 crc kubenswrapper[4870]: I1014 09:01:47.400581 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4qxfd\" (UniqueName: \"kubernetes.io/projected/1a29f723-aa05-46bd-a2f9-a4422fe7646f-kube-api-access-4qxfd\") on node \"crc\" DevicePath \"\"" Oct 14 09:01:47 crc kubenswrapper[4870]: I1014 09:01:47.400593 4870 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1a29f723-aa05-46bd-a2f9-a4422fe7646f-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 09:01:47 crc kubenswrapper[4870]: E1014 09:01:47.882048 4870 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="07e0333d969b88cecad4b5dfbe928bb218113231bd5e69a5da72d866efa3ff49" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 14 09:01:47 crc kubenswrapper[4870]: E1014 09:01:47.883414 4870 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="07e0333d969b88cecad4b5dfbe928bb218113231bd5e69a5da72d866efa3ff49" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 14 09:01:47 crc kubenswrapper[4870]: E1014 09:01:47.885130 4870 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="07e0333d969b88cecad4b5dfbe928bb218113231bd5e69a5da72d866efa3ff49" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 14 09:01:47 crc kubenswrapper[4870]: E1014 09:01:47.885177 4870 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-cell0-conductor-0" podUID="200f8a16-57ae-44d5-bfb3-ba41de932542" containerName="nova-cell0-conductor-conductor" Oct 14 09:01:48 crc kubenswrapper[4870]: I1014 09:01:48.206568 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 14 09:01:48 crc kubenswrapper[4870]: I1014 09:01:48.254495 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 14 09:01:48 crc kubenswrapper[4870]: I1014 09:01:48.267767 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 14 09:01:48 crc kubenswrapper[4870]: I1014 09:01:48.277499 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 14 09:01:48 crc kubenswrapper[4870]: E1014 09:01:48.277933 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a29f723-aa05-46bd-a2f9-a4422fe7646f" containerName="nova-cell1-novncproxy-novncproxy" Oct 14 09:01:48 crc kubenswrapper[4870]: I1014 09:01:48.277956 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a29f723-aa05-46bd-a2f9-a4422fe7646f" containerName="nova-cell1-novncproxy-novncproxy" Oct 14 09:01:48 crc kubenswrapper[4870]: E1014 09:01:48.277968 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="935482d1-d09a-4c56-9b4d-cfd2d7897820" containerName="dnsmasq-dns" Oct 14 09:01:48 crc kubenswrapper[4870]: I1014 09:01:48.277975 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="935482d1-d09a-4c56-9b4d-cfd2d7897820" containerName="dnsmasq-dns" Oct 14 09:01:48 crc kubenswrapper[4870]: E1014 09:01:48.277983 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="935482d1-d09a-4c56-9b4d-cfd2d7897820" containerName="init" Oct 14 09:01:48 crc kubenswrapper[4870]: I1014 09:01:48.277989 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="935482d1-d09a-4c56-9b4d-cfd2d7897820" containerName="init" Oct 14 09:01:48 crc kubenswrapper[4870]: I1014 09:01:48.278184 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="1a29f723-aa05-46bd-a2f9-a4422fe7646f" containerName="nova-cell1-novncproxy-novncproxy" Oct 14 09:01:48 crc kubenswrapper[4870]: I1014 09:01:48.278199 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="935482d1-d09a-4c56-9b4d-cfd2d7897820" containerName="dnsmasq-dns" Oct 14 09:01:48 crc kubenswrapper[4870]: I1014 09:01:48.278776 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 14 09:01:48 crc kubenswrapper[4870]: I1014 09:01:48.285060 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Oct 14 09:01:48 crc kubenswrapper[4870]: I1014 09:01:48.290530 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 14 09:01:48 crc kubenswrapper[4870]: I1014 09:01:48.315060 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m76n9\" (UniqueName: \"kubernetes.io/projected/becafd06-ed0f-4ecf-b9f1-2027b349da1b-kube-api-access-m76n9\") pod \"nova-cell1-novncproxy-0\" (UID: \"becafd06-ed0f-4ecf-b9f1-2027b349da1b\") " pod="openstack/nova-cell1-novncproxy-0" Oct 14 09:01:48 crc kubenswrapper[4870]: I1014 09:01:48.315449 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/becafd06-ed0f-4ecf-b9f1-2027b349da1b-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"becafd06-ed0f-4ecf-b9f1-2027b349da1b\") " pod="openstack/nova-cell1-novncproxy-0" Oct 14 09:01:48 crc kubenswrapper[4870]: I1014 09:01:48.315698 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/becafd06-ed0f-4ecf-b9f1-2027b349da1b-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"becafd06-ed0f-4ecf-b9f1-2027b349da1b\") " pod="openstack/nova-cell1-novncproxy-0" Oct 14 09:01:48 crc kubenswrapper[4870]: I1014 09:01:48.417780 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m76n9\" (UniqueName: \"kubernetes.io/projected/becafd06-ed0f-4ecf-b9f1-2027b349da1b-kube-api-access-m76n9\") pod \"nova-cell1-novncproxy-0\" (UID: \"becafd06-ed0f-4ecf-b9f1-2027b349da1b\") " pod="openstack/nova-cell1-novncproxy-0" Oct 14 09:01:48 crc kubenswrapper[4870]: I1014 09:01:48.417866 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/becafd06-ed0f-4ecf-b9f1-2027b349da1b-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"becafd06-ed0f-4ecf-b9f1-2027b349da1b\") " pod="openstack/nova-cell1-novncproxy-0" Oct 14 09:01:48 crc kubenswrapper[4870]: I1014 09:01:48.417913 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/becafd06-ed0f-4ecf-b9f1-2027b349da1b-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"becafd06-ed0f-4ecf-b9f1-2027b349da1b\") " pod="openstack/nova-cell1-novncproxy-0" Oct 14 09:01:48 crc kubenswrapper[4870]: I1014 09:01:48.426257 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/becafd06-ed0f-4ecf-b9f1-2027b349da1b-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"becafd06-ed0f-4ecf-b9f1-2027b349da1b\") " pod="openstack/nova-cell1-novncproxy-0" Oct 14 09:01:48 crc kubenswrapper[4870]: I1014 09:01:48.434469 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/becafd06-ed0f-4ecf-b9f1-2027b349da1b-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"becafd06-ed0f-4ecf-b9f1-2027b349da1b\") " pod="openstack/nova-cell1-novncproxy-0" Oct 14 09:01:48 crc kubenswrapper[4870]: I1014 09:01:48.436660 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m76n9\" (UniqueName: \"kubernetes.io/projected/becafd06-ed0f-4ecf-b9f1-2027b349da1b-kube-api-access-m76n9\") pod \"nova-cell1-novncproxy-0\" (UID: \"becafd06-ed0f-4ecf-b9f1-2027b349da1b\") " pod="openstack/nova-cell1-novncproxy-0" Oct 14 09:01:48 crc kubenswrapper[4870]: I1014 09:01:48.605374 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 14 09:01:49 crc kubenswrapper[4870]: I1014 09:01:49.051024 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1a29f723-aa05-46bd-a2f9-a4422fe7646f" path="/var/lib/kubelet/pods/1a29f723-aa05-46bd-a2f9-a4422fe7646f/volumes" Oct 14 09:01:49 crc kubenswrapper[4870]: I1014 09:01:49.099594 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 14 09:01:49 crc kubenswrapper[4870]: I1014 09:01:49.216404 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"becafd06-ed0f-4ecf-b9f1-2027b349da1b","Type":"ContainerStarted","Data":"661f878eca345290f9ac03c02c490ab52b42b5c2869663850e35c451869867dc"} Oct 14 09:01:49 crc kubenswrapper[4870]: E1014 09:01:49.426222 4870 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="f22756363c89da54b7a27a7e9a3a91c6a3191a7f2ac7cd6b298026c27f2df43d" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 14 09:01:49 crc kubenswrapper[4870]: E1014 09:01:49.427702 4870 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="f22756363c89da54b7a27a7e9a3a91c6a3191a7f2ac7cd6b298026c27f2df43d" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 14 09:01:49 crc kubenswrapper[4870]: E1014 09:01:49.428917 4870 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="f22756363c89da54b7a27a7e9a3a91c6a3191a7f2ac7cd6b298026c27f2df43d" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 14 09:01:49 crc kubenswrapper[4870]: E1014 09:01:49.428948 4870 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="4eabbefe-21bd-4eba-aab2-5f9123db96fa" containerName="nova-scheduler-scheduler" Oct 14 09:01:49 crc kubenswrapper[4870]: I1014 09:01:49.669573 4870 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="337148f2-88d7-4694-947e-de1aa4375d33" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"http://10.217.1.84:8775/\": read tcp 10.217.0.2:40574->10.217.1.84:8775: read: connection reset by peer" Oct 14 09:01:49 crc kubenswrapper[4870]: I1014 09:01:49.669597 4870 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="337148f2-88d7-4694-947e-de1aa4375d33" containerName="nova-metadata-log" probeResult="failure" output="Get \"http://10.217.1.84:8775/\": read tcp 10.217.0.2:40570->10.217.1.84:8775: read: connection reset by peer" Oct 14 09:01:49 crc kubenswrapper[4870]: I1014 09:01:49.698845 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 14 09:01:49 crc kubenswrapper[4870]: I1014 09:01:49.699547 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-conductor-0" podUID="9c8809d3-35d1-436b-b384-7114eb392a92" containerName="nova-cell1-conductor-conductor" containerID="cri-o://15973d9dc9aabc88d034bacb72c7d2fecc6418b3360a7881625e8c9c98f02e5d" gracePeriod=30 Oct 14 09:01:50 crc kubenswrapper[4870]: I1014 09:01:50.171163 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 14 09:01:50 crc kubenswrapper[4870]: I1014 09:01:50.178544 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 14 09:01:50 crc kubenswrapper[4870]: I1014 09:01:50.237113 4870 generic.go:334] "Generic (PLEG): container finished" podID="d3db1cbf-b57a-4f03-bde9-25dc20acd3bb" containerID="dcba0cf698d44da1b6af963fe253650289ce35521a5481babbc0ed0eac4ac7d8" exitCode=0 Oct 14 09:01:50 crc kubenswrapper[4870]: I1014 09:01:50.237170 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d3db1cbf-b57a-4f03-bde9-25dc20acd3bb","Type":"ContainerDied","Data":"dcba0cf698d44da1b6af963fe253650289ce35521a5481babbc0ed0eac4ac7d8"} Oct 14 09:01:50 crc kubenswrapper[4870]: I1014 09:01:50.237195 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d3db1cbf-b57a-4f03-bde9-25dc20acd3bb","Type":"ContainerDied","Data":"de798eeb6172573679a9b418bf5642ab3594f8a209529f091eff0cd0fcc6e30d"} Oct 14 09:01:50 crc kubenswrapper[4870]: I1014 09:01:50.237210 4870 scope.go:117] "RemoveContainer" containerID="dcba0cf698d44da1b6af963fe253650289ce35521a5481babbc0ed0eac4ac7d8" Oct 14 09:01:50 crc kubenswrapper[4870]: I1014 09:01:50.237332 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 14 09:01:50 crc kubenswrapper[4870]: I1014 09:01:50.245140 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"becafd06-ed0f-4ecf-b9f1-2027b349da1b","Type":"ContainerStarted","Data":"9bbb216a3c7dac6f8ec16ed73d655cd5911261d191748f9525edd5e27c89e293"} Oct 14 09:01:50 crc kubenswrapper[4870]: I1014 09:01:50.250222 4870 generic.go:334] "Generic (PLEG): container finished" podID="337148f2-88d7-4694-947e-de1aa4375d33" containerID="afb0153dee62d6c5510c9ee41556a4075a587da0d89753f84a48becc83a89047" exitCode=0 Oct 14 09:01:50 crc kubenswrapper[4870]: I1014 09:01:50.250269 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"337148f2-88d7-4694-947e-de1aa4375d33","Type":"ContainerDied","Data":"afb0153dee62d6c5510c9ee41556a4075a587da0d89753f84a48becc83a89047"} Oct 14 09:01:50 crc kubenswrapper[4870]: I1014 09:01:50.250297 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"337148f2-88d7-4694-947e-de1aa4375d33","Type":"ContainerDied","Data":"048bd1f7b10b4aa1a4bb1b8923b6c77fd20c3f166aba1eda399448f63091c675"} Oct 14 09:01:50 crc kubenswrapper[4870]: I1014 09:01:50.250359 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 14 09:01:50 crc kubenswrapper[4870]: I1014 09:01:50.255843 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q2w2n\" (UniqueName: \"kubernetes.io/projected/337148f2-88d7-4694-947e-de1aa4375d33-kube-api-access-q2w2n\") pod \"337148f2-88d7-4694-947e-de1aa4375d33\" (UID: \"337148f2-88d7-4694-947e-de1aa4375d33\") " Oct 14 09:01:50 crc kubenswrapper[4870]: I1014 09:01:50.255952 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/337148f2-88d7-4694-947e-de1aa4375d33-logs\") pod \"337148f2-88d7-4694-947e-de1aa4375d33\" (UID: \"337148f2-88d7-4694-947e-de1aa4375d33\") " Oct 14 09:01:50 crc kubenswrapper[4870]: I1014 09:01:50.255988 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/337148f2-88d7-4694-947e-de1aa4375d33-combined-ca-bundle\") pod \"337148f2-88d7-4694-947e-de1aa4375d33\" (UID: \"337148f2-88d7-4694-947e-de1aa4375d33\") " Oct 14 09:01:50 crc kubenswrapper[4870]: I1014 09:01:50.258300 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/337148f2-88d7-4694-947e-de1aa4375d33-logs" (OuterVolumeSpecName: "logs") pod "337148f2-88d7-4694-947e-de1aa4375d33" (UID: "337148f2-88d7-4694-947e-de1aa4375d33"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 09:01:50 crc kubenswrapper[4870]: I1014 09:01:50.263338 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/337148f2-88d7-4694-947e-de1aa4375d33-kube-api-access-q2w2n" (OuterVolumeSpecName: "kube-api-access-q2w2n") pod "337148f2-88d7-4694-947e-de1aa4375d33" (UID: "337148f2-88d7-4694-947e-de1aa4375d33"). InnerVolumeSpecName "kube-api-access-q2w2n". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:01:50 crc kubenswrapper[4870]: I1014 09:01:50.272070 4870 scope.go:117] "RemoveContainer" containerID="f01114c13eca471447c03dcafa83c119dfd75fd57f5778b0f86074ae10416aac" Oct 14 09:01:50 crc kubenswrapper[4870]: I1014 09:01:50.279025 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.279003452 podStartE2EDuration="2.279003452s" podCreationTimestamp="2025-10-14 09:01:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 09:01:50.271256671 +0000 UTC m=+7245.968617062" watchObservedRunningTime="2025-10-14 09:01:50.279003452 +0000 UTC m=+7245.976363823" Oct 14 09:01:50 crc kubenswrapper[4870]: I1014 09:01:50.304322 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/337148f2-88d7-4694-947e-de1aa4375d33-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "337148f2-88d7-4694-947e-de1aa4375d33" (UID: "337148f2-88d7-4694-947e-de1aa4375d33"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:01:50 crc kubenswrapper[4870]: I1014 09:01:50.306637 4870 scope.go:117] "RemoveContainer" containerID="dcba0cf698d44da1b6af963fe253650289ce35521a5481babbc0ed0eac4ac7d8" Oct 14 09:01:50 crc kubenswrapper[4870]: E1014 09:01:50.307168 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dcba0cf698d44da1b6af963fe253650289ce35521a5481babbc0ed0eac4ac7d8\": container with ID starting with dcba0cf698d44da1b6af963fe253650289ce35521a5481babbc0ed0eac4ac7d8 not found: ID does not exist" containerID="dcba0cf698d44da1b6af963fe253650289ce35521a5481babbc0ed0eac4ac7d8" Oct 14 09:01:50 crc kubenswrapper[4870]: I1014 09:01:50.307199 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dcba0cf698d44da1b6af963fe253650289ce35521a5481babbc0ed0eac4ac7d8"} err="failed to get container status \"dcba0cf698d44da1b6af963fe253650289ce35521a5481babbc0ed0eac4ac7d8\": rpc error: code = NotFound desc = could not find container \"dcba0cf698d44da1b6af963fe253650289ce35521a5481babbc0ed0eac4ac7d8\": container with ID starting with dcba0cf698d44da1b6af963fe253650289ce35521a5481babbc0ed0eac4ac7d8 not found: ID does not exist" Oct 14 09:01:50 crc kubenswrapper[4870]: I1014 09:01:50.307218 4870 scope.go:117] "RemoveContainer" containerID="f01114c13eca471447c03dcafa83c119dfd75fd57f5778b0f86074ae10416aac" Oct 14 09:01:50 crc kubenswrapper[4870]: E1014 09:01:50.307695 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f01114c13eca471447c03dcafa83c119dfd75fd57f5778b0f86074ae10416aac\": container with ID starting with f01114c13eca471447c03dcafa83c119dfd75fd57f5778b0f86074ae10416aac not found: ID does not exist" containerID="f01114c13eca471447c03dcafa83c119dfd75fd57f5778b0f86074ae10416aac" Oct 14 09:01:50 crc kubenswrapper[4870]: I1014 09:01:50.307728 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f01114c13eca471447c03dcafa83c119dfd75fd57f5778b0f86074ae10416aac"} err="failed to get container status \"f01114c13eca471447c03dcafa83c119dfd75fd57f5778b0f86074ae10416aac\": rpc error: code = NotFound desc = could not find container \"f01114c13eca471447c03dcafa83c119dfd75fd57f5778b0f86074ae10416aac\": container with ID starting with f01114c13eca471447c03dcafa83c119dfd75fd57f5778b0f86074ae10416aac not found: ID does not exist" Oct 14 09:01:50 crc kubenswrapper[4870]: I1014 09:01:50.307748 4870 scope.go:117] "RemoveContainer" containerID="afb0153dee62d6c5510c9ee41556a4075a587da0d89753f84a48becc83a89047" Oct 14 09:01:50 crc kubenswrapper[4870]: I1014 09:01:50.329681 4870 scope.go:117] "RemoveContainer" containerID="5968c04b35057ebe36c338a812baba14f9fa27c198995f3c2f308c83f14b2950" Oct 14 09:01:50 crc kubenswrapper[4870]: I1014 09:01:50.346833 4870 scope.go:117] "RemoveContainer" containerID="afb0153dee62d6c5510c9ee41556a4075a587da0d89753f84a48becc83a89047" Oct 14 09:01:50 crc kubenswrapper[4870]: E1014 09:01:50.347172 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"afb0153dee62d6c5510c9ee41556a4075a587da0d89753f84a48becc83a89047\": container with ID starting with afb0153dee62d6c5510c9ee41556a4075a587da0d89753f84a48becc83a89047 not found: ID does not exist" containerID="afb0153dee62d6c5510c9ee41556a4075a587da0d89753f84a48becc83a89047" Oct 14 09:01:50 crc kubenswrapper[4870]: I1014 09:01:50.347201 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"afb0153dee62d6c5510c9ee41556a4075a587da0d89753f84a48becc83a89047"} err="failed to get container status \"afb0153dee62d6c5510c9ee41556a4075a587da0d89753f84a48becc83a89047\": rpc error: code = NotFound desc = could not find container \"afb0153dee62d6c5510c9ee41556a4075a587da0d89753f84a48becc83a89047\": container with ID starting with afb0153dee62d6c5510c9ee41556a4075a587da0d89753f84a48becc83a89047 not found: ID does not exist" Oct 14 09:01:50 crc kubenswrapper[4870]: I1014 09:01:50.347220 4870 scope.go:117] "RemoveContainer" containerID="5968c04b35057ebe36c338a812baba14f9fa27c198995f3c2f308c83f14b2950" Oct 14 09:01:50 crc kubenswrapper[4870]: E1014 09:01:50.347384 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5968c04b35057ebe36c338a812baba14f9fa27c198995f3c2f308c83f14b2950\": container with ID starting with 5968c04b35057ebe36c338a812baba14f9fa27c198995f3c2f308c83f14b2950 not found: ID does not exist" containerID="5968c04b35057ebe36c338a812baba14f9fa27c198995f3c2f308c83f14b2950" Oct 14 09:01:50 crc kubenswrapper[4870]: I1014 09:01:50.347407 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5968c04b35057ebe36c338a812baba14f9fa27c198995f3c2f308c83f14b2950"} err="failed to get container status \"5968c04b35057ebe36c338a812baba14f9fa27c198995f3c2f308c83f14b2950\": rpc error: code = NotFound desc = could not find container \"5968c04b35057ebe36c338a812baba14f9fa27c198995f3c2f308c83f14b2950\": container with ID starting with 5968c04b35057ebe36c338a812baba14f9fa27c198995f3c2f308c83f14b2950 not found: ID does not exist" Oct 14 09:01:50 crc kubenswrapper[4870]: I1014 09:01:50.356930 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d3db1cbf-b57a-4f03-bde9-25dc20acd3bb-config-data\") pod \"d3db1cbf-b57a-4f03-bde9-25dc20acd3bb\" (UID: \"d3db1cbf-b57a-4f03-bde9-25dc20acd3bb\") " Oct 14 09:01:50 crc kubenswrapper[4870]: I1014 09:01:50.356998 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d3db1cbf-b57a-4f03-bde9-25dc20acd3bb-logs\") pod \"d3db1cbf-b57a-4f03-bde9-25dc20acd3bb\" (UID: \"d3db1cbf-b57a-4f03-bde9-25dc20acd3bb\") " Oct 14 09:01:50 crc kubenswrapper[4870]: I1014 09:01:50.357029 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/337148f2-88d7-4694-947e-de1aa4375d33-config-data\") pod \"337148f2-88d7-4694-947e-de1aa4375d33\" (UID: \"337148f2-88d7-4694-947e-de1aa4375d33\") " Oct 14 09:01:50 crc kubenswrapper[4870]: I1014 09:01:50.357074 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3db1cbf-b57a-4f03-bde9-25dc20acd3bb-combined-ca-bundle\") pod \"d3db1cbf-b57a-4f03-bde9-25dc20acd3bb\" (UID: \"d3db1cbf-b57a-4f03-bde9-25dc20acd3bb\") " Oct 14 09:01:50 crc kubenswrapper[4870]: I1014 09:01:50.357103 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hg454\" (UniqueName: \"kubernetes.io/projected/d3db1cbf-b57a-4f03-bde9-25dc20acd3bb-kube-api-access-hg454\") pod \"d3db1cbf-b57a-4f03-bde9-25dc20acd3bb\" (UID: \"d3db1cbf-b57a-4f03-bde9-25dc20acd3bb\") " Oct 14 09:01:50 crc kubenswrapper[4870]: I1014 09:01:50.357580 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q2w2n\" (UniqueName: \"kubernetes.io/projected/337148f2-88d7-4694-947e-de1aa4375d33-kube-api-access-q2w2n\") on node \"crc\" DevicePath \"\"" Oct 14 09:01:50 crc kubenswrapper[4870]: I1014 09:01:50.357596 4870 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/337148f2-88d7-4694-947e-de1aa4375d33-logs\") on node \"crc\" DevicePath \"\"" Oct 14 09:01:50 crc kubenswrapper[4870]: I1014 09:01:50.357608 4870 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/337148f2-88d7-4694-947e-de1aa4375d33-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 09:01:50 crc kubenswrapper[4870]: I1014 09:01:50.358155 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d3db1cbf-b57a-4f03-bde9-25dc20acd3bb-logs" (OuterVolumeSpecName: "logs") pod "d3db1cbf-b57a-4f03-bde9-25dc20acd3bb" (UID: "d3db1cbf-b57a-4f03-bde9-25dc20acd3bb"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 09:01:50 crc kubenswrapper[4870]: I1014 09:01:50.361968 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d3db1cbf-b57a-4f03-bde9-25dc20acd3bb-kube-api-access-hg454" (OuterVolumeSpecName: "kube-api-access-hg454") pod "d3db1cbf-b57a-4f03-bde9-25dc20acd3bb" (UID: "d3db1cbf-b57a-4f03-bde9-25dc20acd3bb"). InnerVolumeSpecName "kube-api-access-hg454". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:01:50 crc kubenswrapper[4870]: I1014 09:01:50.384932 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d3db1cbf-b57a-4f03-bde9-25dc20acd3bb-config-data" (OuterVolumeSpecName: "config-data") pod "d3db1cbf-b57a-4f03-bde9-25dc20acd3bb" (UID: "d3db1cbf-b57a-4f03-bde9-25dc20acd3bb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:01:50 crc kubenswrapper[4870]: I1014 09:01:50.392664 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/337148f2-88d7-4694-947e-de1aa4375d33-config-data" (OuterVolumeSpecName: "config-data") pod "337148f2-88d7-4694-947e-de1aa4375d33" (UID: "337148f2-88d7-4694-947e-de1aa4375d33"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:01:50 crc kubenswrapper[4870]: I1014 09:01:50.401639 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d3db1cbf-b57a-4f03-bde9-25dc20acd3bb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d3db1cbf-b57a-4f03-bde9-25dc20acd3bb" (UID: "d3db1cbf-b57a-4f03-bde9-25dc20acd3bb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:01:50 crc kubenswrapper[4870]: I1014 09:01:50.459488 4870 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3db1cbf-b57a-4f03-bde9-25dc20acd3bb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 09:01:50 crc kubenswrapper[4870]: I1014 09:01:50.459538 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hg454\" (UniqueName: \"kubernetes.io/projected/d3db1cbf-b57a-4f03-bde9-25dc20acd3bb-kube-api-access-hg454\") on node \"crc\" DevicePath \"\"" Oct 14 09:01:50 crc kubenswrapper[4870]: I1014 09:01:50.459555 4870 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d3db1cbf-b57a-4f03-bde9-25dc20acd3bb-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 09:01:50 crc kubenswrapper[4870]: I1014 09:01:50.459569 4870 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d3db1cbf-b57a-4f03-bde9-25dc20acd3bb-logs\") on node \"crc\" DevicePath \"\"" Oct 14 09:01:50 crc kubenswrapper[4870]: I1014 09:01:50.459583 4870 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/337148f2-88d7-4694-947e-de1aa4375d33-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 09:01:50 crc kubenswrapper[4870]: I1014 09:01:50.577002 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 14 09:01:50 crc kubenswrapper[4870]: I1014 09:01:50.604466 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 14 09:01:50 crc kubenswrapper[4870]: I1014 09:01:50.619340 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 14 09:01:50 crc kubenswrapper[4870]: I1014 09:01:50.632616 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 14 09:01:50 crc kubenswrapper[4870]: I1014 09:01:50.639725 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 14 09:01:50 crc kubenswrapper[4870]: E1014 09:01:50.640187 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="337148f2-88d7-4694-947e-de1aa4375d33" containerName="nova-metadata-log" Oct 14 09:01:50 crc kubenswrapper[4870]: I1014 09:01:50.640211 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="337148f2-88d7-4694-947e-de1aa4375d33" containerName="nova-metadata-log" Oct 14 09:01:50 crc kubenswrapper[4870]: E1014 09:01:50.640241 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="337148f2-88d7-4694-947e-de1aa4375d33" containerName="nova-metadata-metadata" Oct 14 09:01:50 crc kubenswrapper[4870]: I1014 09:01:50.640251 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="337148f2-88d7-4694-947e-de1aa4375d33" containerName="nova-metadata-metadata" Oct 14 09:01:50 crc kubenswrapper[4870]: E1014 09:01:50.640264 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3db1cbf-b57a-4f03-bde9-25dc20acd3bb" containerName="nova-api-log" Oct 14 09:01:50 crc kubenswrapper[4870]: I1014 09:01:50.640270 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3db1cbf-b57a-4f03-bde9-25dc20acd3bb" containerName="nova-api-log" Oct 14 09:01:50 crc kubenswrapper[4870]: E1014 09:01:50.640280 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3db1cbf-b57a-4f03-bde9-25dc20acd3bb" containerName="nova-api-api" Oct 14 09:01:50 crc kubenswrapper[4870]: I1014 09:01:50.640286 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3db1cbf-b57a-4f03-bde9-25dc20acd3bb" containerName="nova-api-api" Oct 14 09:01:50 crc kubenswrapper[4870]: I1014 09:01:50.640484 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="d3db1cbf-b57a-4f03-bde9-25dc20acd3bb" containerName="nova-api-api" Oct 14 09:01:50 crc kubenswrapper[4870]: I1014 09:01:50.640504 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="337148f2-88d7-4694-947e-de1aa4375d33" containerName="nova-metadata-metadata" Oct 14 09:01:50 crc kubenswrapper[4870]: I1014 09:01:50.640512 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="d3db1cbf-b57a-4f03-bde9-25dc20acd3bb" containerName="nova-api-log" Oct 14 09:01:50 crc kubenswrapper[4870]: I1014 09:01:50.640526 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="337148f2-88d7-4694-947e-de1aa4375d33" containerName="nova-metadata-log" Oct 14 09:01:50 crc kubenswrapper[4870]: I1014 09:01:50.641611 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 14 09:01:50 crc kubenswrapper[4870]: I1014 09:01:50.643655 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 14 09:01:50 crc kubenswrapper[4870]: I1014 09:01:50.646488 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 14 09:01:50 crc kubenswrapper[4870]: I1014 09:01:50.648127 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 14 09:01:50 crc kubenswrapper[4870]: I1014 09:01:50.652115 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 14 09:01:50 crc kubenswrapper[4870]: I1014 09:01:50.655779 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 14 09:01:50 crc kubenswrapper[4870]: I1014 09:01:50.661449 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 14 09:01:50 crc kubenswrapper[4870]: I1014 09:01:50.665411 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e1055bc1-24f3-4893-ad77-c520a36428bb-logs\") pod \"nova-api-0\" (UID: \"e1055bc1-24f3-4893-ad77-c520a36428bb\") " pod="openstack/nova-api-0" Oct 14 09:01:50 crc kubenswrapper[4870]: I1014 09:01:50.665480 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tzv26\" (UniqueName: \"kubernetes.io/projected/e1055bc1-24f3-4893-ad77-c520a36428bb-kube-api-access-tzv26\") pod \"nova-api-0\" (UID: \"e1055bc1-24f3-4893-ad77-c520a36428bb\") " pod="openstack/nova-api-0" Oct 14 09:01:50 crc kubenswrapper[4870]: I1014 09:01:50.665505 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/17f1a688-1bd1-4b88-ad00-51caf1af0ce0-config-data\") pod \"nova-metadata-0\" (UID: \"17f1a688-1bd1-4b88-ad00-51caf1af0ce0\") " pod="openstack/nova-metadata-0" Oct 14 09:01:50 crc kubenswrapper[4870]: I1014 09:01:50.665521 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e1055bc1-24f3-4893-ad77-c520a36428bb-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"e1055bc1-24f3-4893-ad77-c520a36428bb\") " pod="openstack/nova-api-0" Oct 14 09:01:50 crc kubenswrapper[4870]: I1014 09:01:50.665542 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e1055bc1-24f3-4893-ad77-c520a36428bb-config-data\") pod \"nova-api-0\" (UID: \"e1055bc1-24f3-4893-ad77-c520a36428bb\") " pod="openstack/nova-api-0" Oct 14 09:01:50 crc kubenswrapper[4870]: I1014 09:01:50.665583 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b6p95\" (UniqueName: \"kubernetes.io/projected/17f1a688-1bd1-4b88-ad00-51caf1af0ce0-kube-api-access-b6p95\") pod \"nova-metadata-0\" (UID: \"17f1a688-1bd1-4b88-ad00-51caf1af0ce0\") " pod="openstack/nova-metadata-0" Oct 14 09:01:50 crc kubenswrapper[4870]: I1014 09:01:50.665612 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/17f1a688-1bd1-4b88-ad00-51caf1af0ce0-logs\") pod \"nova-metadata-0\" (UID: \"17f1a688-1bd1-4b88-ad00-51caf1af0ce0\") " pod="openstack/nova-metadata-0" Oct 14 09:01:50 crc kubenswrapper[4870]: I1014 09:01:50.665821 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17f1a688-1bd1-4b88-ad00-51caf1af0ce0-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"17f1a688-1bd1-4b88-ad00-51caf1af0ce0\") " pod="openstack/nova-metadata-0" Oct 14 09:01:50 crc kubenswrapper[4870]: I1014 09:01:50.767891 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e1055bc1-24f3-4893-ad77-c520a36428bb-logs\") pod \"nova-api-0\" (UID: \"e1055bc1-24f3-4893-ad77-c520a36428bb\") " pod="openstack/nova-api-0" Oct 14 09:01:50 crc kubenswrapper[4870]: I1014 09:01:50.767949 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tzv26\" (UniqueName: \"kubernetes.io/projected/e1055bc1-24f3-4893-ad77-c520a36428bb-kube-api-access-tzv26\") pod \"nova-api-0\" (UID: \"e1055bc1-24f3-4893-ad77-c520a36428bb\") " pod="openstack/nova-api-0" Oct 14 09:01:50 crc kubenswrapper[4870]: I1014 09:01:50.767980 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/17f1a688-1bd1-4b88-ad00-51caf1af0ce0-config-data\") pod \"nova-metadata-0\" (UID: \"17f1a688-1bd1-4b88-ad00-51caf1af0ce0\") " pod="openstack/nova-metadata-0" Oct 14 09:01:50 crc kubenswrapper[4870]: I1014 09:01:50.767999 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e1055bc1-24f3-4893-ad77-c520a36428bb-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"e1055bc1-24f3-4893-ad77-c520a36428bb\") " pod="openstack/nova-api-0" Oct 14 09:01:50 crc kubenswrapper[4870]: I1014 09:01:50.768019 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e1055bc1-24f3-4893-ad77-c520a36428bb-config-data\") pod \"nova-api-0\" (UID: \"e1055bc1-24f3-4893-ad77-c520a36428bb\") " pod="openstack/nova-api-0" Oct 14 09:01:50 crc kubenswrapper[4870]: I1014 09:01:50.768064 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b6p95\" (UniqueName: \"kubernetes.io/projected/17f1a688-1bd1-4b88-ad00-51caf1af0ce0-kube-api-access-b6p95\") pod \"nova-metadata-0\" (UID: \"17f1a688-1bd1-4b88-ad00-51caf1af0ce0\") " pod="openstack/nova-metadata-0" Oct 14 09:01:50 crc kubenswrapper[4870]: I1014 09:01:50.768083 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/17f1a688-1bd1-4b88-ad00-51caf1af0ce0-logs\") pod \"nova-metadata-0\" (UID: \"17f1a688-1bd1-4b88-ad00-51caf1af0ce0\") " pod="openstack/nova-metadata-0" Oct 14 09:01:50 crc kubenswrapper[4870]: I1014 09:01:50.768139 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17f1a688-1bd1-4b88-ad00-51caf1af0ce0-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"17f1a688-1bd1-4b88-ad00-51caf1af0ce0\") " pod="openstack/nova-metadata-0" Oct 14 09:01:50 crc kubenswrapper[4870]: I1014 09:01:50.768349 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e1055bc1-24f3-4893-ad77-c520a36428bb-logs\") pod \"nova-api-0\" (UID: \"e1055bc1-24f3-4893-ad77-c520a36428bb\") " pod="openstack/nova-api-0" Oct 14 09:01:50 crc kubenswrapper[4870]: I1014 09:01:50.768705 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/17f1a688-1bd1-4b88-ad00-51caf1af0ce0-logs\") pod \"nova-metadata-0\" (UID: \"17f1a688-1bd1-4b88-ad00-51caf1af0ce0\") " pod="openstack/nova-metadata-0" Oct 14 09:01:50 crc kubenswrapper[4870]: I1014 09:01:50.771627 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e1055bc1-24f3-4893-ad77-c520a36428bb-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"e1055bc1-24f3-4893-ad77-c520a36428bb\") " pod="openstack/nova-api-0" Oct 14 09:01:50 crc kubenswrapper[4870]: I1014 09:01:50.773136 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/17f1a688-1bd1-4b88-ad00-51caf1af0ce0-config-data\") pod \"nova-metadata-0\" (UID: \"17f1a688-1bd1-4b88-ad00-51caf1af0ce0\") " pod="openstack/nova-metadata-0" Oct 14 09:01:50 crc kubenswrapper[4870]: I1014 09:01:50.774025 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17f1a688-1bd1-4b88-ad00-51caf1af0ce0-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"17f1a688-1bd1-4b88-ad00-51caf1af0ce0\") " pod="openstack/nova-metadata-0" Oct 14 09:01:50 crc kubenswrapper[4870]: I1014 09:01:50.780802 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e1055bc1-24f3-4893-ad77-c520a36428bb-config-data\") pod \"nova-api-0\" (UID: \"e1055bc1-24f3-4893-ad77-c520a36428bb\") " pod="openstack/nova-api-0" Oct 14 09:01:50 crc kubenswrapper[4870]: I1014 09:01:50.787587 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tzv26\" (UniqueName: \"kubernetes.io/projected/e1055bc1-24f3-4893-ad77-c520a36428bb-kube-api-access-tzv26\") pod \"nova-api-0\" (UID: \"e1055bc1-24f3-4893-ad77-c520a36428bb\") " pod="openstack/nova-api-0" Oct 14 09:01:50 crc kubenswrapper[4870]: I1014 09:01:50.792042 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b6p95\" (UniqueName: \"kubernetes.io/projected/17f1a688-1bd1-4b88-ad00-51caf1af0ce0-kube-api-access-b6p95\") pod \"nova-metadata-0\" (UID: \"17f1a688-1bd1-4b88-ad00-51caf1af0ce0\") " pod="openstack/nova-metadata-0" Oct 14 09:01:50 crc kubenswrapper[4870]: I1014 09:01:50.967469 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 14 09:01:50 crc kubenswrapper[4870]: I1014 09:01:50.978300 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 14 09:01:51 crc kubenswrapper[4870]: I1014 09:01:51.061364 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="337148f2-88d7-4694-947e-de1aa4375d33" path="/var/lib/kubelet/pods/337148f2-88d7-4694-947e-de1aa4375d33/volumes" Oct 14 09:01:51 crc kubenswrapper[4870]: I1014 09:01:51.062009 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d3db1cbf-b57a-4f03-bde9-25dc20acd3bb" path="/var/lib/kubelet/pods/d3db1cbf-b57a-4f03-bde9-25dc20acd3bb/volumes" Oct 14 09:01:51 crc kubenswrapper[4870]: I1014 09:01:51.520835 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 14 09:01:51 crc kubenswrapper[4870]: I1014 09:01:51.609708 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 14 09:01:51 crc kubenswrapper[4870]: W1014 09:01:51.637727 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod17f1a688_1bd1_4b88_ad00_51caf1af0ce0.slice/crio-4445c716ef7f432ea4d5783ad7075d6d63bcdc2920320f42fcc9bb29df12107a WatchSource:0}: Error finding container 4445c716ef7f432ea4d5783ad7075d6d63bcdc2920320f42fcc9bb29df12107a: Status 404 returned error can't find the container with id 4445c716ef7f432ea4d5783ad7075d6d63bcdc2920320f42fcc9bb29df12107a Oct 14 09:01:51 crc kubenswrapper[4870]: I1014 09:01:51.934854 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 14 09:01:51 crc kubenswrapper[4870]: I1014 09:01:51.988169 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/200f8a16-57ae-44d5-bfb3-ba41de932542-combined-ca-bundle\") pod \"200f8a16-57ae-44d5-bfb3-ba41de932542\" (UID: \"200f8a16-57ae-44d5-bfb3-ba41de932542\") " Oct 14 09:01:51 crc kubenswrapper[4870]: I1014 09:01:51.988279 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/200f8a16-57ae-44d5-bfb3-ba41de932542-config-data\") pod \"200f8a16-57ae-44d5-bfb3-ba41de932542\" (UID: \"200f8a16-57ae-44d5-bfb3-ba41de932542\") " Oct 14 09:01:51 crc kubenswrapper[4870]: I1014 09:01:51.988367 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r552c\" (UniqueName: \"kubernetes.io/projected/200f8a16-57ae-44d5-bfb3-ba41de932542-kube-api-access-r552c\") pod \"200f8a16-57ae-44d5-bfb3-ba41de932542\" (UID: \"200f8a16-57ae-44d5-bfb3-ba41de932542\") " Oct 14 09:01:52 crc kubenswrapper[4870]: I1014 09:01:52.009170 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/200f8a16-57ae-44d5-bfb3-ba41de932542-kube-api-access-r552c" (OuterVolumeSpecName: "kube-api-access-r552c") pod "200f8a16-57ae-44d5-bfb3-ba41de932542" (UID: "200f8a16-57ae-44d5-bfb3-ba41de932542"). InnerVolumeSpecName "kube-api-access-r552c". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:01:52 crc kubenswrapper[4870]: I1014 09:01:52.076285 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/200f8a16-57ae-44d5-bfb3-ba41de932542-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "200f8a16-57ae-44d5-bfb3-ba41de932542" (UID: "200f8a16-57ae-44d5-bfb3-ba41de932542"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:01:52 crc kubenswrapper[4870]: I1014 09:01:52.090177 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r552c\" (UniqueName: \"kubernetes.io/projected/200f8a16-57ae-44d5-bfb3-ba41de932542-kube-api-access-r552c\") on node \"crc\" DevicePath \"\"" Oct 14 09:01:52 crc kubenswrapper[4870]: I1014 09:01:52.090387 4870 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/200f8a16-57ae-44d5-bfb3-ba41de932542-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 09:01:52 crc kubenswrapper[4870]: I1014 09:01:52.130485 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/200f8a16-57ae-44d5-bfb3-ba41de932542-config-data" (OuterVolumeSpecName: "config-data") pod "200f8a16-57ae-44d5-bfb3-ba41de932542" (UID: "200f8a16-57ae-44d5-bfb3-ba41de932542"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:01:52 crc kubenswrapper[4870]: I1014 09:01:52.191205 4870 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/200f8a16-57ae-44d5-bfb3-ba41de932542-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 09:01:52 crc kubenswrapper[4870]: I1014 09:01:52.288614 4870 generic.go:334] "Generic (PLEG): container finished" podID="200f8a16-57ae-44d5-bfb3-ba41de932542" containerID="07e0333d969b88cecad4b5dfbe928bb218113231bd5e69a5da72d866efa3ff49" exitCode=0 Oct 14 09:01:52 crc kubenswrapper[4870]: I1014 09:01:52.288661 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"200f8a16-57ae-44d5-bfb3-ba41de932542","Type":"ContainerDied","Data":"07e0333d969b88cecad4b5dfbe928bb218113231bd5e69a5da72d866efa3ff49"} Oct 14 09:01:52 crc kubenswrapper[4870]: I1014 09:01:52.288700 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"200f8a16-57ae-44d5-bfb3-ba41de932542","Type":"ContainerDied","Data":"f3d2abdea93d4ec269b0a52cd31cd60f1839468d1a0f9c3845b4c4ff82d8e64d"} Oct 14 09:01:52 crc kubenswrapper[4870]: I1014 09:01:52.288715 4870 scope.go:117] "RemoveContainer" containerID="07e0333d969b88cecad4b5dfbe928bb218113231bd5e69a5da72d866efa3ff49" Oct 14 09:01:52 crc kubenswrapper[4870]: I1014 09:01:52.288724 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 14 09:01:52 crc kubenswrapper[4870]: I1014 09:01:52.291655 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e1055bc1-24f3-4893-ad77-c520a36428bb","Type":"ContainerStarted","Data":"36057428133bd01d5a6c0e75037e7dff3a30c1c00aa184489e5fee2e48ca9c66"} Oct 14 09:01:52 crc kubenswrapper[4870]: I1014 09:01:52.291682 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e1055bc1-24f3-4893-ad77-c520a36428bb","Type":"ContainerStarted","Data":"f61fa490fab929fb5e534efb5255c5f168fdc3b95da9665beae2c6c2d3ba5803"} Oct 14 09:01:52 crc kubenswrapper[4870]: I1014 09:01:52.291695 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e1055bc1-24f3-4893-ad77-c520a36428bb","Type":"ContainerStarted","Data":"2e6295cdaa37c572684a9ab5e125e12fef5dbe0be1f1d9f3c30c40b8c4a44e04"} Oct 14 09:01:52 crc kubenswrapper[4870]: I1014 09:01:52.298161 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"17f1a688-1bd1-4b88-ad00-51caf1af0ce0","Type":"ContainerStarted","Data":"1bc76a1a314ba9d308a310397f670dc886227d98b6f2d40ae039086e08114b2d"} Oct 14 09:01:52 crc kubenswrapper[4870]: I1014 09:01:52.298214 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"17f1a688-1bd1-4b88-ad00-51caf1af0ce0","Type":"ContainerStarted","Data":"990d26c8691a9651642a973ab6b25f404d460d767a6e4471a24b9aaf9d964606"} Oct 14 09:01:52 crc kubenswrapper[4870]: I1014 09:01:52.298226 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"17f1a688-1bd1-4b88-ad00-51caf1af0ce0","Type":"ContainerStarted","Data":"4445c716ef7f432ea4d5783ad7075d6d63bcdc2920320f42fcc9bb29df12107a"} Oct 14 09:01:52 crc kubenswrapper[4870]: I1014 09:01:52.313974 4870 scope.go:117] "RemoveContainer" containerID="07e0333d969b88cecad4b5dfbe928bb218113231bd5e69a5da72d866efa3ff49" Oct 14 09:01:52 crc kubenswrapper[4870]: E1014 09:01:52.320804 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"07e0333d969b88cecad4b5dfbe928bb218113231bd5e69a5da72d866efa3ff49\": container with ID starting with 07e0333d969b88cecad4b5dfbe928bb218113231bd5e69a5da72d866efa3ff49 not found: ID does not exist" containerID="07e0333d969b88cecad4b5dfbe928bb218113231bd5e69a5da72d866efa3ff49" Oct 14 09:01:52 crc kubenswrapper[4870]: I1014 09:01:52.320856 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"07e0333d969b88cecad4b5dfbe928bb218113231bd5e69a5da72d866efa3ff49"} err="failed to get container status \"07e0333d969b88cecad4b5dfbe928bb218113231bd5e69a5da72d866efa3ff49\": rpc error: code = NotFound desc = could not find container \"07e0333d969b88cecad4b5dfbe928bb218113231bd5e69a5da72d866efa3ff49\": container with ID starting with 07e0333d969b88cecad4b5dfbe928bb218113231bd5e69a5da72d866efa3ff49 not found: ID does not exist" Oct 14 09:01:52 crc kubenswrapper[4870]: I1014 09:01:52.323180 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.323162166 podStartE2EDuration="2.323162166s" podCreationTimestamp="2025-10-14 09:01:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 09:01:52.309333454 +0000 UTC m=+7248.006693835" watchObservedRunningTime="2025-10-14 09:01:52.323162166 +0000 UTC m=+7248.020522537" Oct 14 09:01:52 crc kubenswrapper[4870]: I1014 09:01:52.339423 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 14 09:01:52 crc kubenswrapper[4870]: I1014 09:01:52.347097 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 14 09:01:52 crc kubenswrapper[4870]: I1014 09:01:52.353652 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 14 09:01:52 crc kubenswrapper[4870]: E1014 09:01:52.354169 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="200f8a16-57ae-44d5-bfb3-ba41de932542" containerName="nova-cell0-conductor-conductor" Oct 14 09:01:52 crc kubenswrapper[4870]: I1014 09:01:52.354190 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="200f8a16-57ae-44d5-bfb3-ba41de932542" containerName="nova-cell0-conductor-conductor" Oct 14 09:01:52 crc kubenswrapper[4870]: I1014 09:01:52.354424 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="200f8a16-57ae-44d5-bfb3-ba41de932542" containerName="nova-cell0-conductor-conductor" Oct 14 09:01:52 crc kubenswrapper[4870]: I1014 09:01:52.354994 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.354975961 podStartE2EDuration="2.354975961s" podCreationTimestamp="2025-10-14 09:01:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 09:01:52.34278102 +0000 UTC m=+7248.040141391" watchObservedRunningTime="2025-10-14 09:01:52.354975961 +0000 UTC m=+7248.052336332" Oct 14 09:01:52 crc kubenswrapper[4870]: I1014 09:01:52.355123 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 14 09:01:52 crc kubenswrapper[4870]: I1014 09:01:52.358236 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Oct 14 09:01:52 crc kubenswrapper[4870]: I1014 09:01:52.369422 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 14 09:01:52 crc kubenswrapper[4870]: E1014 09:01:52.481143 4870 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod200f8a16_57ae_44d5_bfb3_ba41de932542.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod200f8a16_57ae_44d5_bfb3_ba41de932542.slice/crio-f3d2abdea93d4ec269b0a52cd31cd60f1839468d1a0f9c3845b4c4ff82d8e64d\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9c8809d3_35d1_436b_b384_7114eb392a92.slice/crio-15973d9dc9aabc88d034bacb72c7d2fecc6418b3360a7881625e8c9c98f02e5d.scope\": RecentStats: unable to find data in memory cache]" Oct 14 09:01:52 crc kubenswrapper[4870]: I1014 09:01:52.496796 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee5dcbc8-b98e-48b9-a089-a9f5614249d3-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"ee5dcbc8-b98e-48b9-a089-a9f5614249d3\") " pod="openstack/nova-cell0-conductor-0" Oct 14 09:01:52 crc kubenswrapper[4870]: I1014 09:01:52.496914 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ee5dcbc8-b98e-48b9-a089-a9f5614249d3-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"ee5dcbc8-b98e-48b9-a089-a9f5614249d3\") " pod="openstack/nova-cell0-conductor-0" Oct 14 09:01:52 crc kubenswrapper[4870]: I1014 09:01:52.496951 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x2lf8\" (UniqueName: \"kubernetes.io/projected/ee5dcbc8-b98e-48b9-a089-a9f5614249d3-kube-api-access-x2lf8\") pod \"nova-cell0-conductor-0\" (UID: \"ee5dcbc8-b98e-48b9-a089-a9f5614249d3\") " pod="openstack/nova-cell0-conductor-0" Oct 14 09:01:52 crc kubenswrapper[4870]: I1014 09:01:52.599081 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ee5dcbc8-b98e-48b9-a089-a9f5614249d3-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"ee5dcbc8-b98e-48b9-a089-a9f5614249d3\") " pod="openstack/nova-cell0-conductor-0" Oct 14 09:01:52 crc kubenswrapper[4870]: I1014 09:01:52.599566 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x2lf8\" (UniqueName: \"kubernetes.io/projected/ee5dcbc8-b98e-48b9-a089-a9f5614249d3-kube-api-access-x2lf8\") pod \"nova-cell0-conductor-0\" (UID: \"ee5dcbc8-b98e-48b9-a089-a9f5614249d3\") " pod="openstack/nova-cell0-conductor-0" Oct 14 09:01:52 crc kubenswrapper[4870]: I1014 09:01:52.599645 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee5dcbc8-b98e-48b9-a089-a9f5614249d3-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"ee5dcbc8-b98e-48b9-a089-a9f5614249d3\") " pod="openstack/nova-cell0-conductor-0" Oct 14 09:01:52 crc kubenswrapper[4870]: I1014 09:01:52.603360 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee5dcbc8-b98e-48b9-a089-a9f5614249d3-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"ee5dcbc8-b98e-48b9-a089-a9f5614249d3\") " pod="openstack/nova-cell0-conductor-0" Oct 14 09:01:52 crc kubenswrapper[4870]: I1014 09:01:52.612136 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ee5dcbc8-b98e-48b9-a089-a9f5614249d3-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"ee5dcbc8-b98e-48b9-a089-a9f5614249d3\") " pod="openstack/nova-cell0-conductor-0" Oct 14 09:01:52 crc kubenswrapper[4870]: I1014 09:01:52.620203 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x2lf8\" (UniqueName: \"kubernetes.io/projected/ee5dcbc8-b98e-48b9-a089-a9f5614249d3-kube-api-access-x2lf8\") pod \"nova-cell0-conductor-0\" (UID: \"ee5dcbc8-b98e-48b9-a089-a9f5614249d3\") " pod="openstack/nova-cell0-conductor-0" Oct 14 09:01:52 crc kubenswrapper[4870]: I1014 09:01:52.690962 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 14 09:01:52 crc kubenswrapper[4870]: I1014 09:01:52.886993 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 14 09:01:53 crc kubenswrapper[4870]: I1014 09:01:53.006962 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c8809d3-35d1-436b-b384-7114eb392a92-config-data\") pod \"9c8809d3-35d1-436b-b384-7114eb392a92\" (UID: \"9c8809d3-35d1-436b-b384-7114eb392a92\") " Oct 14 09:01:53 crc kubenswrapper[4870]: I1014 09:01:53.007066 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c8809d3-35d1-436b-b384-7114eb392a92-combined-ca-bundle\") pod \"9c8809d3-35d1-436b-b384-7114eb392a92\" (UID: \"9c8809d3-35d1-436b-b384-7114eb392a92\") " Oct 14 09:01:53 crc kubenswrapper[4870]: I1014 09:01:53.007112 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w5qxs\" (UniqueName: \"kubernetes.io/projected/9c8809d3-35d1-436b-b384-7114eb392a92-kube-api-access-w5qxs\") pod \"9c8809d3-35d1-436b-b384-7114eb392a92\" (UID: \"9c8809d3-35d1-436b-b384-7114eb392a92\") " Oct 14 09:01:53 crc kubenswrapper[4870]: I1014 09:01:53.012963 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9c8809d3-35d1-436b-b384-7114eb392a92-kube-api-access-w5qxs" (OuterVolumeSpecName: "kube-api-access-w5qxs") pod "9c8809d3-35d1-436b-b384-7114eb392a92" (UID: "9c8809d3-35d1-436b-b384-7114eb392a92"). InnerVolumeSpecName "kube-api-access-w5qxs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:01:53 crc kubenswrapper[4870]: I1014 09:01:53.031969 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9c8809d3-35d1-436b-b384-7114eb392a92-config-data" (OuterVolumeSpecName: "config-data") pod "9c8809d3-35d1-436b-b384-7114eb392a92" (UID: "9c8809d3-35d1-436b-b384-7114eb392a92"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:01:53 crc kubenswrapper[4870]: I1014 09:01:53.039802 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9c8809d3-35d1-436b-b384-7114eb392a92-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9c8809d3-35d1-436b-b384-7114eb392a92" (UID: "9c8809d3-35d1-436b-b384-7114eb392a92"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:01:53 crc kubenswrapper[4870]: I1014 09:01:53.046358 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="200f8a16-57ae-44d5-bfb3-ba41de932542" path="/var/lib/kubelet/pods/200f8a16-57ae-44d5-bfb3-ba41de932542/volumes" Oct 14 09:01:53 crc kubenswrapper[4870]: I1014 09:01:53.109457 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w5qxs\" (UniqueName: \"kubernetes.io/projected/9c8809d3-35d1-436b-b384-7114eb392a92-kube-api-access-w5qxs\") on node \"crc\" DevicePath \"\"" Oct 14 09:01:53 crc kubenswrapper[4870]: I1014 09:01:53.109487 4870 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c8809d3-35d1-436b-b384-7114eb392a92-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 09:01:53 crc kubenswrapper[4870]: I1014 09:01:53.109498 4870 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c8809d3-35d1-436b-b384-7114eb392a92-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 09:01:53 crc kubenswrapper[4870]: I1014 09:01:53.184766 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 14 09:01:53 crc kubenswrapper[4870]: W1014 09:01:53.217901 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podee5dcbc8_b98e_48b9_a089_a9f5614249d3.slice/crio-a7fc468608113d657aebbc527900ae007bc57e5d1082a93586d12cda783ca7a6 WatchSource:0}: Error finding container a7fc468608113d657aebbc527900ae007bc57e5d1082a93586d12cda783ca7a6: Status 404 returned error can't find the container with id a7fc468608113d657aebbc527900ae007bc57e5d1082a93586d12cda783ca7a6 Oct 14 09:01:53 crc kubenswrapper[4870]: I1014 09:01:53.310347 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"ee5dcbc8-b98e-48b9-a089-a9f5614249d3","Type":"ContainerStarted","Data":"a7fc468608113d657aebbc527900ae007bc57e5d1082a93586d12cda783ca7a6"} Oct 14 09:01:53 crc kubenswrapper[4870]: I1014 09:01:53.312111 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"4eabbefe-21bd-4eba-aab2-5f9123db96fa","Type":"ContainerDied","Data":"f22756363c89da54b7a27a7e9a3a91c6a3191a7f2ac7cd6b298026c27f2df43d"} Oct 14 09:01:53 crc kubenswrapper[4870]: I1014 09:01:53.312125 4870 generic.go:334] "Generic (PLEG): container finished" podID="4eabbefe-21bd-4eba-aab2-5f9123db96fa" containerID="f22756363c89da54b7a27a7e9a3a91c6a3191a7f2ac7cd6b298026c27f2df43d" exitCode=0 Oct 14 09:01:53 crc kubenswrapper[4870]: I1014 09:01:53.314420 4870 generic.go:334] "Generic (PLEG): container finished" podID="9c8809d3-35d1-436b-b384-7114eb392a92" containerID="15973d9dc9aabc88d034bacb72c7d2fecc6418b3360a7881625e8c9c98f02e5d" exitCode=0 Oct 14 09:01:53 crc kubenswrapper[4870]: I1014 09:01:53.314832 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 14 09:01:53 crc kubenswrapper[4870]: I1014 09:01:53.314882 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"9c8809d3-35d1-436b-b384-7114eb392a92","Type":"ContainerDied","Data":"15973d9dc9aabc88d034bacb72c7d2fecc6418b3360a7881625e8c9c98f02e5d"} Oct 14 09:01:53 crc kubenswrapper[4870]: I1014 09:01:53.314912 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"9c8809d3-35d1-436b-b384-7114eb392a92","Type":"ContainerDied","Data":"222da9312d9c0ffe5bbd7ec4e4fbf2f8e44c6f88e2357e309e2a7d8de8a5fce5"} Oct 14 09:01:53 crc kubenswrapper[4870]: I1014 09:01:53.314934 4870 scope.go:117] "RemoveContainer" containerID="15973d9dc9aabc88d034bacb72c7d2fecc6418b3360a7881625e8c9c98f02e5d" Oct 14 09:01:53 crc kubenswrapper[4870]: I1014 09:01:53.345827 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 14 09:01:53 crc kubenswrapper[4870]: I1014 09:01:53.355775 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 14 09:01:53 crc kubenswrapper[4870]: I1014 09:01:53.359370 4870 scope.go:117] "RemoveContainer" containerID="15973d9dc9aabc88d034bacb72c7d2fecc6418b3360a7881625e8c9c98f02e5d" Oct 14 09:01:53 crc kubenswrapper[4870]: E1014 09:01:53.360480 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"15973d9dc9aabc88d034bacb72c7d2fecc6418b3360a7881625e8c9c98f02e5d\": container with ID starting with 15973d9dc9aabc88d034bacb72c7d2fecc6418b3360a7881625e8c9c98f02e5d not found: ID does not exist" containerID="15973d9dc9aabc88d034bacb72c7d2fecc6418b3360a7881625e8c9c98f02e5d" Oct 14 09:01:53 crc kubenswrapper[4870]: I1014 09:01:53.360541 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"15973d9dc9aabc88d034bacb72c7d2fecc6418b3360a7881625e8c9c98f02e5d"} err="failed to get container status \"15973d9dc9aabc88d034bacb72c7d2fecc6418b3360a7881625e8c9c98f02e5d\": rpc error: code = NotFound desc = could not find container \"15973d9dc9aabc88d034bacb72c7d2fecc6418b3360a7881625e8c9c98f02e5d\": container with ID starting with 15973d9dc9aabc88d034bacb72c7d2fecc6418b3360a7881625e8c9c98f02e5d not found: ID does not exist" Oct 14 09:01:53 crc kubenswrapper[4870]: I1014 09:01:53.377105 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 14 09:01:53 crc kubenswrapper[4870]: E1014 09:01:53.378790 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c8809d3-35d1-436b-b384-7114eb392a92" containerName="nova-cell1-conductor-conductor" Oct 14 09:01:53 crc kubenswrapper[4870]: I1014 09:01:53.378836 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c8809d3-35d1-436b-b384-7114eb392a92" containerName="nova-cell1-conductor-conductor" Oct 14 09:01:53 crc kubenswrapper[4870]: I1014 09:01:53.379185 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="9c8809d3-35d1-436b-b384-7114eb392a92" containerName="nova-cell1-conductor-conductor" Oct 14 09:01:53 crc kubenswrapper[4870]: I1014 09:01:53.380352 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 14 09:01:53 crc kubenswrapper[4870]: I1014 09:01:53.384300 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Oct 14 09:01:53 crc kubenswrapper[4870]: I1014 09:01:53.405346 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 14 09:01:53 crc kubenswrapper[4870]: I1014 09:01:53.516094 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p9bkx\" (UniqueName: \"kubernetes.io/projected/02ebb80d-7ee3-48e5-99e4-1d9e061d6e06-kube-api-access-p9bkx\") pod \"nova-cell1-conductor-0\" (UID: \"02ebb80d-7ee3-48e5-99e4-1d9e061d6e06\") " pod="openstack/nova-cell1-conductor-0" Oct 14 09:01:53 crc kubenswrapper[4870]: I1014 09:01:53.516534 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02ebb80d-7ee3-48e5-99e4-1d9e061d6e06-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"02ebb80d-7ee3-48e5-99e4-1d9e061d6e06\") " pod="openstack/nova-cell1-conductor-0" Oct 14 09:01:53 crc kubenswrapper[4870]: I1014 09:01:53.516585 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/02ebb80d-7ee3-48e5-99e4-1d9e061d6e06-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"02ebb80d-7ee3-48e5-99e4-1d9e061d6e06\") " pod="openstack/nova-cell1-conductor-0" Oct 14 09:01:53 crc kubenswrapper[4870]: I1014 09:01:53.606329 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Oct 14 09:01:53 crc kubenswrapper[4870]: I1014 09:01:53.617616 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p9bkx\" (UniqueName: \"kubernetes.io/projected/02ebb80d-7ee3-48e5-99e4-1d9e061d6e06-kube-api-access-p9bkx\") pod \"nova-cell1-conductor-0\" (UID: \"02ebb80d-7ee3-48e5-99e4-1d9e061d6e06\") " pod="openstack/nova-cell1-conductor-0" Oct 14 09:01:53 crc kubenswrapper[4870]: I1014 09:01:53.617708 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02ebb80d-7ee3-48e5-99e4-1d9e061d6e06-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"02ebb80d-7ee3-48e5-99e4-1d9e061d6e06\") " pod="openstack/nova-cell1-conductor-0" Oct 14 09:01:53 crc kubenswrapper[4870]: I1014 09:01:53.617746 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/02ebb80d-7ee3-48e5-99e4-1d9e061d6e06-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"02ebb80d-7ee3-48e5-99e4-1d9e061d6e06\") " pod="openstack/nova-cell1-conductor-0" Oct 14 09:01:53 crc kubenswrapper[4870]: I1014 09:01:53.622258 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02ebb80d-7ee3-48e5-99e4-1d9e061d6e06-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"02ebb80d-7ee3-48e5-99e4-1d9e061d6e06\") " pod="openstack/nova-cell1-conductor-0" Oct 14 09:01:53 crc kubenswrapper[4870]: I1014 09:01:53.622954 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/02ebb80d-7ee3-48e5-99e4-1d9e061d6e06-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"02ebb80d-7ee3-48e5-99e4-1d9e061d6e06\") " pod="openstack/nova-cell1-conductor-0" Oct 14 09:01:53 crc kubenswrapper[4870]: I1014 09:01:53.637130 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p9bkx\" (UniqueName: \"kubernetes.io/projected/02ebb80d-7ee3-48e5-99e4-1d9e061d6e06-kube-api-access-p9bkx\") pod \"nova-cell1-conductor-0\" (UID: \"02ebb80d-7ee3-48e5-99e4-1d9e061d6e06\") " pod="openstack/nova-cell1-conductor-0" Oct 14 09:01:53 crc kubenswrapper[4870]: I1014 09:01:53.705680 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 14 09:01:53 crc kubenswrapper[4870]: I1014 09:01:53.713776 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 14 09:01:53 crc kubenswrapper[4870]: I1014 09:01:53.820698 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-942pr\" (UniqueName: \"kubernetes.io/projected/4eabbefe-21bd-4eba-aab2-5f9123db96fa-kube-api-access-942pr\") pod \"4eabbefe-21bd-4eba-aab2-5f9123db96fa\" (UID: \"4eabbefe-21bd-4eba-aab2-5f9123db96fa\") " Oct 14 09:01:53 crc kubenswrapper[4870]: I1014 09:01:53.820998 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4eabbefe-21bd-4eba-aab2-5f9123db96fa-config-data\") pod \"4eabbefe-21bd-4eba-aab2-5f9123db96fa\" (UID: \"4eabbefe-21bd-4eba-aab2-5f9123db96fa\") " Oct 14 09:01:53 crc kubenswrapper[4870]: I1014 09:01:53.821068 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4eabbefe-21bd-4eba-aab2-5f9123db96fa-combined-ca-bundle\") pod \"4eabbefe-21bd-4eba-aab2-5f9123db96fa\" (UID: \"4eabbefe-21bd-4eba-aab2-5f9123db96fa\") " Oct 14 09:01:53 crc kubenswrapper[4870]: I1014 09:01:53.831703 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4eabbefe-21bd-4eba-aab2-5f9123db96fa-kube-api-access-942pr" (OuterVolumeSpecName: "kube-api-access-942pr") pod "4eabbefe-21bd-4eba-aab2-5f9123db96fa" (UID: "4eabbefe-21bd-4eba-aab2-5f9123db96fa"). InnerVolumeSpecName "kube-api-access-942pr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:01:53 crc kubenswrapper[4870]: I1014 09:01:53.849978 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4eabbefe-21bd-4eba-aab2-5f9123db96fa-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4eabbefe-21bd-4eba-aab2-5f9123db96fa" (UID: "4eabbefe-21bd-4eba-aab2-5f9123db96fa"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:01:53 crc kubenswrapper[4870]: I1014 09:01:53.853725 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4eabbefe-21bd-4eba-aab2-5f9123db96fa-config-data" (OuterVolumeSpecName: "config-data") pod "4eabbefe-21bd-4eba-aab2-5f9123db96fa" (UID: "4eabbefe-21bd-4eba-aab2-5f9123db96fa"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:01:53 crc kubenswrapper[4870]: I1014 09:01:53.924036 4870 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4eabbefe-21bd-4eba-aab2-5f9123db96fa-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 09:01:53 crc kubenswrapper[4870]: I1014 09:01:53.924078 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-942pr\" (UniqueName: \"kubernetes.io/projected/4eabbefe-21bd-4eba-aab2-5f9123db96fa-kube-api-access-942pr\") on node \"crc\" DevicePath \"\"" Oct 14 09:01:53 crc kubenswrapper[4870]: I1014 09:01:53.924093 4870 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4eabbefe-21bd-4eba-aab2-5f9123db96fa-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 09:01:53 crc kubenswrapper[4870]: I1014 09:01:53.951700 4870 patch_prober.go:28] interesting pod/machine-config-daemon-7tvc7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 09:01:53 crc kubenswrapper[4870]: I1014 09:01:53.951767 4870 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 09:01:53 crc kubenswrapper[4870]: I1014 09:01:53.951822 4870 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" Oct 14 09:01:53 crc kubenswrapper[4870]: I1014 09:01:53.952598 4870 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"9426cde7309378345e6129005b55744e852dbf5b2ef5a92cc907fc57d1dc3ae8"} pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 14 09:01:53 crc kubenswrapper[4870]: I1014 09:01:53.952663 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerName="machine-config-daemon" containerID="cri-o://9426cde7309378345e6129005b55744e852dbf5b2ef5a92cc907fc57d1dc3ae8" gracePeriod=600 Oct 14 09:01:54 crc kubenswrapper[4870]: I1014 09:01:54.157857 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 14 09:01:54 crc kubenswrapper[4870]: I1014 09:01:54.324030 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"ee5dcbc8-b98e-48b9-a089-a9f5614249d3","Type":"ContainerStarted","Data":"910ed7bad34d9a7163b841990638ec4c809ae09c71d936618c95878703298b7e"} Oct 14 09:01:54 crc kubenswrapper[4870]: I1014 09:01:54.324543 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Oct 14 09:01:54 crc kubenswrapper[4870]: I1014 09:01:54.326834 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"02ebb80d-7ee3-48e5-99e4-1d9e061d6e06","Type":"ContainerStarted","Data":"7de2685da46eb5150b97e8f6b62c03e1e3311ee185474328abef90e2a3d2e359"} Oct 14 09:01:54 crc kubenswrapper[4870]: I1014 09:01:54.329356 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 14 09:01:54 crc kubenswrapper[4870]: I1014 09:01:54.329569 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"4eabbefe-21bd-4eba-aab2-5f9123db96fa","Type":"ContainerDied","Data":"a6988420bd4775a3a0b784381d4e6615df53417a5438e804260a50423cabc929"} Oct 14 09:01:54 crc kubenswrapper[4870]: I1014 09:01:54.329600 4870 scope.go:117] "RemoveContainer" containerID="f22756363c89da54b7a27a7e9a3a91c6a3191a7f2ac7cd6b298026c27f2df43d" Oct 14 09:01:54 crc kubenswrapper[4870]: I1014 09:01:54.337994 4870 generic.go:334] "Generic (PLEG): container finished" podID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerID="9426cde7309378345e6129005b55744e852dbf5b2ef5a92cc907fc57d1dc3ae8" exitCode=0 Oct 14 09:01:54 crc kubenswrapper[4870]: I1014 09:01:54.338027 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" event={"ID":"874a5e45-dffd-4d17-b609-b7d3ed2eab07","Type":"ContainerDied","Data":"9426cde7309378345e6129005b55744e852dbf5b2ef5a92cc907fc57d1dc3ae8"} Oct 14 09:01:54 crc kubenswrapper[4870]: I1014 09:01:54.338047 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" event={"ID":"874a5e45-dffd-4d17-b609-b7d3ed2eab07","Type":"ContainerStarted","Data":"8fcaeb3090f0c89cba1e66db524b4a246072daa35592012664aca450fa09ddc9"} Oct 14 09:01:54 crc kubenswrapper[4870]: I1014 09:01:54.340830 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.3408160159999998 podStartE2EDuration="2.340816016s" podCreationTimestamp="2025-10-14 09:01:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 09:01:54.340124009 +0000 UTC m=+7250.037484380" watchObservedRunningTime="2025-10-14 09:01:54.340816016 +0000 UTC m=+7250.038176387" Oct 14 09:01:54 crc kubenswrapper[4870]: I1014 09:01:54.373400 4870 scope.go:117] "RemoveContainer" containerID="8c0f754fe94b7cb7b9bb5dcd613f7e42093968b49509d8ee37e8fd4db89eedd5" Oct 14 09:01:54 crc kubenswrapper[4870]: I1014 09:01:54.398972 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 14 09:01:54 crc kubenswrapper[4870]: I1014 09:01:54.417848 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Oct 14 09:01:54 crc kubenswrapper[4870]: I1014 09:01:54.423838 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 14 09:01:54 crc kubenswrapper[4870]: E1014 09:01:54.424313 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4eabbefe-21bd-4eba-aab2-5f9123db96fa" containerName="nova-scheduler-scheduler" Oct 14 09:01:54 crc kubenswrapper[4870]: I1014 09:01:54.424337 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="4eabbefe-21bd-4eba-aab2-5f9123db96fa" containerName="nova-scheduler-scheduler" Oct 14 09:01:54 crc kubenswrapper[4870]: I1014 09:01:54.424554 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="4eabbefe-21bd-4eba-aab2-5f9123db96fa" containerName="nova-scheduler-scheduler" Oct 14 09:01:54 crc kubenswrapper[4870]: I1014 09:01:54.426480 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 14 09:01:54 crc kubenswrapper[4870]: I1014 09:01:54.432727 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 14 09:01:54 crc kubenswrapper[4870]: I1014 09:01:54.433669 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 14 09:01:54 crc kubenswrapper[4870]: I1014 09:01:54.533659 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/73281705-4198-40a4-b523-c0bfc322c833-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"73281705-4198-40a4-b523-c0bfc322c833\") " pod="openstack/nova-scheduler-0" Oct 14 09:01:54 crc kubenswrapper[4870]: I1014 09:01:54.533731 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/73281705-4198-40a4-b523-c0bfc322c833-config-data\") pod \"nova-scheduler-0\" (UID: \"73281705-4198-40a4-b523-c0bfc322c833\") " pod="openstack/nova-scheduler-0" Oct 14 09:01:54 crc kubenswrapper[4870]: I1014 09:01:54.533897 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vscgf\" (UniqueName: \"kubernetes.io/projected/73281705-4198-40a4-b523-c0bfc322c833-kube-api-access-vscgf\") pod \"nova-scheduler-0\" (UID: \"73281705-4198-40a4-b523-c0bfc322c833\") " pod="openstack/nova-scheduler-0" Oct 14 09:01:54 crc kubenswrapper[4870]: I1014 09:01:54.635132 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vscgf\" (UniqueName: \"kubernetes.io/projected/73281705-4198-40a4-b523-c0bfc322c833-kube-api-access-vscgf\") pod \"nova-scheduler-0\" (UID: \"73281705-4198-40a4-b523-c0bfc322c833\") " pod="openstack/nova-scheduler-0" Oct 14 09:01:54 crc kubenswrapper[4870]: I1014 09:01:54.635910 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/73281705-4198-40a4-b523-c0bfc322c833-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"73281705-4198-40a4-b523-c0bfc322c833\") " pod="openstack/nova-scheduler-0" Oct 14 09:01:54 crc kubenswrapper[4870]: I1014 09:01:54.636064 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/73281705-4198-40a4-b523-c0bfc322c833-config-data\") pod \"nova-scheduler-0\" (UID: \"73281705-4198-40a4-b523-c0bfc322c833\") " pod="openstack/nova-scheduler-0" Oct 14 09:01:54 crc kubenswrapper[4870]: I1014 09:01:54.643411 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/73281705-4198-40a4-b523-c0bfc322c833-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"73281705-4198-40a4-b523-c0bfc322c833\") " pod="openstack/nova-scheduler-0" Oct 14 09:01:54 crc kubenswrapper[4870]: I1014 09:01:54.644515 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/73281705-4198-40a4-b523-c0bfc322c833-config-data\") pod \"nova-scheduler-0\" (UID: \"73281705-4198-40a4-b523-c0bfc322c833\") " pod="openstack/nova-scheduler-0" Oct 14 09:01:54 crc kubenswrapper[4870]: I1014 09:01:54.651172 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vscgf\" (UniqueName: \"kubernetes.io/projected/73281705-4198-40a4-b523-c0bfc322c833-kube-api-access-vscgf\") pod \"nova-scheduler-0\" (UID: \"73281705-4198-40a4-b523-c0bfc322c833\") " pod="openstack/nova-scheduler-0" Oct 14 09:01:54 crc kubenswrapper[4870]: I1014 09:01:54.744284 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 14 09:01:55 crc kubenswrapper[4870]: I1014 09:01:55.047745 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4eabbefe-21bd-4eba-aab2-5f9123db96fa" path="/var/lib/kubelet/pods/4eabbefe-21bd-4eba-aab2-5f9123db96fa/volumes" Oct 14 09:01:55 crc kubenswrapper[4870]: I1014 09:01:55.048552 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9c8809d3-35d1-436b-b384-7114eb392a92" path="/var/lib/kubelet/pods/9c8809d3-35d1-436b-b384-7114eb392a92/volumes" Oct 14 09:01:55 crc kubenswrapper[4870]: I1014 09:01:55.238325 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 14 09:01:55 crc kubenswrapper[4870]: I1014 09:01:55.358840 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"73281705-4198-40a4-b523-c0bfc322c833","Type":"ContainerStarted","Data":"fef838a7c02ea978f46e98ee10f6903ba186b375270b82222309e4108eb295ce"} Oct 14 09:01:55 crc kubenswrapper[4870]: I1014 09:01:55.362575 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"02ebb80d-7ee3-48e5-99e4-1d9e061d6e06","Type":"ContainerStarted","Data":"652700e216f5422f63850496eceac88c82acc847b709d8a19b42bc4103863ad9"} Oct 14 09:01:55 crc kubenswrapper[4870]: I1014 09:01:55.367201 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Oct 14 09:01:55 crc kubenswrapper[4870]: I1014 09:01:55.391480 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.390760361 podStartE2EDuration="2.390760361s" podCreationTimestamp="2025-10-14 09:01:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 09:01:55.380431416 +0000 UTC m=+7251.077791787" watchObservedRunningTime="2025-10-14 09:01:55.390760361 +0000 UTC m=+7251.088120742" Oct 14 09:01:55 crc kubenswrapper[4870]: I1014 09:01:55.979198 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 14 09:01:55 crc kubenswrapper[4870]: I1014 09:01:55.979550 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 14 09:01:56 crc kubenswrapper[4870]: I1014 09:01:56.053141 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-jwfxl"] Oct 14 09:01:56 crc kubenswrapper[4870]: I1014 09:01:56.063955 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-jwfxl"] Oct 14 09:01:56 crc kubenswrapper[4870]: I1014 09:01:56.384208 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"73281705-4198-40a4-b523-c0bfc322c833","Type":"ContainerStarted","Data":"c0020e7320b47c052dc0efb0f61570b5a39b13c7d39e5b5d0bc9ebec23c7b744"} Oct 14 09:01:56 crc kubenswrapper[4870]: I1014 09:01:56.404873 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.404855741 podStartE2EDuration="2.404855741s" podCreationTimestamp="2025-10-14 09:01:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 09:01:56.395593882 +0000 UTC m=+7252.092954273" watchObservedRunningTime="2025-10-14 09:01:56.404855741 +0000 UTC m=+7252.102216112" Oct 14 09:01:57 crc kubenswrapper[4870]: I1014 09:01:57.063102 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="13680632-67ca-40ab-b71a-569c99f3e424" path="/var/lib/kubelet/pods/13680632-67ca-40ab-b71a-569c99f3e424/volumes" Oct 14 09:01:58 crc kubenswrapper[4870]: I1014 09:01:58.605957 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Oct 14 09:01:58 crc kubenswrapper[4870]: I1014 09:01:58.619599 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Oct 14 09:01:59 crc kubenswrapper[4870]: I1014 09:01:59.424484 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Oct 14 09:01:59 crc kubenswrapper[4870]: I1014 09:01:59.744709 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 14 09:02:00 crc kubenswrapper[4870]: I1014 09:02:00.968167 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 14 09:02:00 crc kubenswrapper[4870]: I1014 09:02:00.968587 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 14 09:02:00 crc kubenswrapper[4870]: I1014 09:02:00.979620 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 14 09:02:00 crc kubenswrapper[4870]: I1014 09:02:00.979670 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 14 09:02:02 crc kubenswrapper[4870]: I1014 09:02:02.093814 4870 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="e1055bc1-24f3-4893-ad77-c520a36428bb" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.1.93:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 14 09:02:02 crc kubenswrapper[4870]: I1014 09:02:02.093864 4870 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="e1055bc1-24f3-4893-ad77-c520a36428bb" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.1.93:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 14 09:02:02 crc kubenswrapper[4870]: I1014 09:02:02.094850 4870 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="17f1a688-1bd1-4b88-ad00-51caf1af0ce0" containerName="nova-metadata-log" probeResult="failure" output="Get \"http://10.217.1.94:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 14 09:02:02 crc kubenswrapper[4870]: I1014 09:02:02.095129 4870 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="17f1a688-1bd1-4b88-ad00-51caf1af0ce0" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"http://10.217.1.94:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 14 09:02:02 crc kubenswrapper[4870]: I1014 09:02:02.725679 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Oct 14 09:02:03 crc kubenswrapper[4870]: I1014 09:02:03.749739 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Oct 14 09:02:04 crc kubenswrapper[4870]: I1014 09:02:04.745352 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 14 09:02:04 crc kubenswrapper[4870]: I1014 09:02:04.800258 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 14 09:02:05 crc kubenswrapper[4870]: I1014 09:02:05.502611 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 14 09:02:06 crc kubenswrapper[4870]: I1014 09:02:06.032065 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-5ede-account-create-gfcc4"] Oct 14 09:02:06 crc kubenswrapper[4870]: I1014 09:02:06.043867 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-5ede-account-create-gfcc4"] Oct 14 09:02:07 crc kubenswrapper[4870]: I1014 09:02:07.058818 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f08148e6-ef78-4157-aaf5-1c6432956583" path="/var/lib/kubelet/pods/f08148e6-ef78-4157-aaf5-1c6432956583/volumes" Oct 14 09:02:09 crc kubenswrapper[4870]: I1014 09:02:09.828291 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Oct 14 09:02:09 crc kubenswrapper[4870]: I1014 09:02:09.832012 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 14 09:02:09 crc kubenswrapper[4870]: I1014 09:02:09.835781 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 14 09:02:09 crc kubenswrapper[4870]: I1014 09:02:09.836768 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Oct 14 09:02:09 crc kubenswrapper[4870]: I1014 09:02:09.884454 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/cb4814fd-7bd7-430e-a390-159f8f9756e8-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"cb4814fd-7bd7-430e-a390-159f8f9756e8\") " pod="openstack/cinder-scheduler-0" Oct 14 09:02:09 crc kubenswrapper[4870]: I1014 09:02:09.884508 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p6dq5\" (UniqueName: \"kubernetes.io/projected/cb4814fd-7bd7-430e-a390-159f8f9756e8-kube-api-access-p6dq5\") pod \"cinder-scheduler-0\" (UID: \"cb4814fd-7bd7-430e-a390-159f8f9756e8\") " pod="openstack/cinder-scheduler-0" Oct 14 09:02:09 crc kubenswrapper[4870]: I1014 09:02:09.884540 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cb4814fd-7bd7-430e-a390-159f8f9756e8-config-data\") pod \"cinder-scheduler-0\" (UID: \"cb4814fd-7bd7-430e-a390-159f8f9756e8\") " pod="openstack/cinder-scheduler-0" Oct 14 09:02:09 crc kubenswrapper[4870]: I1014 09:02:09.884615 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cb4814fd-7bd7-430e-a390-159f8f9756e8-scripts\") pod \"cinder-scheduler-0\" (UID: \"cb4814fd-7bd7-430e-a390-159f8f9756e8\") " pod="openstack/cinder-scheduler-0" Oct 14 09:02:09 crc kubenswrapper[4870]: I1014 09:02:09.884641 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb4814fd-7bd7-430e-a390-159f8f9756e8-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"cb4814fd-7bd7-430e-a390-159f8f9756e8\") " pod="openstack/cinder-scheduler-0" Oct 14 09:02:09 crc kubenswrapper[4870]: I1014 09:02:09.884668 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cb4814fd-7bd7-430e-a390-159f8f9756e8-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"cb4814fd-7bd7-430e-a390-159f8f9756e8\") " pod="openstack/cinder-scheduler-0" Oct 14 09:02:09 crc kubenswrapper[4870]: I1014 09:02:09.986913 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cb4814fd-7bd7-430e-a390-159f8f9756e8-scripts\") pod \"cinder-scheduler-0\" (UID: \"cb4814fd-7bd7-430e-a390-159f8f9756e8\") " pod="openstack/cinder-scheduler-0" Oct 14 09:02:09 crc kubenswrapper[4870]: I1014 09:02:09.987000 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb4814fd-7bd7-430e-a390-159f8f9756e8-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"cb4814fd-7bd7-430e-a390-159f8f9756e8\") " pod="openstack/cinder-scheduler-0" Oct 14 09:02:09 crc kubenswrapper[4870]: I1014 09:02:09.987039 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cb4814fd-7bd7-430e-a390-159f8f9756e8-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"cb4814fd-7bd7-430e-a390-159f8f9756e8\") " pod="openstack/cinder-scheduler-0" Oct 14 09:02:09 crc kubenswrapper[4870]: I1014 09:02:09.987105 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/cb4814fd-7bd7-430e-a390-159f8f9756e8-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"cb4814fd-7bd7-430e-a390-159f8f9756e8\") " pod="openstack/cinder-scheduler-0" Oct 14 09:02:09 crc kubenswrapper[4870]: I1014 09:02:09.987141 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p6dq5\" (UniqueName: \"kubernetes.io/projected/cb4814fd-7bd7-430e-a390-159f8f9756e8-kube-api-access-p6dq5\") pod \"cinder-scheduler-0\" (UID: \"cb4814fd-7bd7-430e-a390-159f8f9756e8\") " pod="openstack/cinder-scheduler-0" Oct 14 09:02:09 crc kubenswrapper[4870]: I1014 09:02:09.987166 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cb4814fd-7bd7-430e-a390-159f8f9756e8-config-data\") pod \"cinder-scheduler-0\" (UID: \"cb4814fd-7bd7-430e-a390-159f8f9756e8\") " pod="openstack/cinder-scheduler-0" Oct 14 09:02:09 crc kubenswrapper[4870]: I1014 09:02:09.987311 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/cb4814fd-7bd7-430e-a390-159f8f9756e8-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"cb4814fd-7bd7-430e-a390-159f8f9756e8\") " pod="openstack/cinder-scheduler-0" Oct 14 09:02:09 crc kubenswrapper[4870]: I1014 09:02:09.995313 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cb4814fd-7bd7-430e-a390-159f8f9756e8-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"cb4814fd-7bd7-430e-a390-159f8f9756e8\") " pod="openstack/cinder-scheduler-0" Oct 14 09:02:09 crc kubenswrapper[4870]: I1014 09:02:09.995590 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cb4814fd-7bd7-430e-a390-159f8f9756e8-config-data\") pod \"cinder-scheduler-0\" (UID: \"cb4814fd-7bd7-430e-a390-159f8f9756e8\") " pod="openstack/cinder-scheduler-0" Oct 14 09:02:09 crc kubenswrapper[4870]: I1014 09:02:09.995635 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb4814fd-7bd7-430e-a390-159f8f9756e8-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"cb4814fd-7bd7-430e-a390-159f8f9756e8\") " pod="openstack/cinder-scheduler-0" Oct 14 09:02:09 crc kubenswrapper[4870]: I1014 09:02:09.999564 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cb4814fd-7bd7-430e-a390-159f8f9756e8-scripts\") pod \"cinder-scheduler-0\" (UID: \"cb4814fd-7bd7-430e-a390-159f8f9756e8\") " pod="openstack/cinder-scheduler-0" Oct 14 09:02:10 crc kubenswrapper[4870]: I1014 09:02:10.008921 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p6dq5\" (UniqueName: \"kubernetes.io/projected/cb4814fd-7bd7-430e-a390-159f8f9756e8-kube-api-access-p6dq5\") pod \"cinder-scheduler-0\" (UID: \"cb4814fd-7bd7-430e-a390-159f8f9756e8\") " pod="openstack/cinder-scheduler-0" Oct 14 09:02:10 crc kubenswrapper[4870]: I1014 09:02:10.190366 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 14 09:02:10 crc kubenswrapper[4870]: I1014 09:02:10.644862 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 14 09:02:10 crc kubenswrapper[4870]: I1014 09:02:10.971806 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 14 09:02:10 crc kubenswrapper[4870]: I1014 09:02:10.973147 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 14 09:02:10 crc kubenswrapper[4870]: I1014 09:02:10.984334 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 14 09:02:10 crc kubenswrapper[4870]: I1014 09:02:10.987230 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 14 09:02:10 crc kubenswrapper[4870]: I1014 09:02:10.987985 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 14 09:02:10 crc kubenswrapper[4870]: I1014 09:02:10.988222 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 14 09:02:10 crc kubenswrapper[4870]: I1014 09:02:10.992427 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 14 09:02:11 crc kubenswrapper[4870]: I1014 09:02:11.007934 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 14 09:02:11 crc kubenswrapper[4870]: I1014 09:02:11.389636 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Oct 14 09:02:11 crc kubenswrapper[4870]: I1014 09:02:11.390234 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="a662858e-76ac-4bf6-949d-4bc17f6a3aaf" containerName="cinder-api-log" containerID="cri-o://8747381c748307eddcef853ed8b5d3128671bd63c249dac6069447d3067eabb8" gracePeriod=30 Oct 14 09:02:11 crc kubenswrapper[4870]: I1014 09:02:11.390385 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="a662858e-76ac-4bf6-949d-4bc17f6a3aaf" containerName="cinder-api" containerID="cri-o://9f71bfbc3b3893910374afed0b2400760747c481120e966d946e4c2dab9b0550" gracePeriod=30 Oct 14 09:02:11 crc kubenswrapper[4870]: I1014 09:02:11.528188 4870 generic.go:334] "Generic (PLEG): container finished" podID="a662858e-76ac-4bf6-949d-4bc17f6a3aaf" containerID="8747381c748307eddcef853ed8b5d3128671bd63c249dac6069447d3067eabb8" exitCode=143 Oct 14 09:02:11 crc kubenswrapper[4870]: I1014 09:02:11.528256 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"a662858e-76ac-4bf6-949d-4bc17f6a3aaf","Type":"ContainerDied","Data":"8747381c748307eddcef853ed8b5d3128671bd63c249dac6069447d3067eabb8"} Oct 14 09:02:11 crc kubenswrapper[4870]: I1014 09:02:11.530253 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"cb4814fd-7bd7-430e-a390-159f8f9756e8","Type":"ContainerStarted","Data":"26a061f76ae099e9a8894ac4d3accbb909fd7b04bc926e7b21b048407549517d"} Oct 14 09:02:11 crc kubenswrapper[4870]: I1014 09:02:11.531007 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 14 09:02:11 crc kubenswrapper[4870]: I1014 09:02:11.537891 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 14 09:02:11 crc kubenswrapper[4870]: I1014 09:02:11.972199 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-volume-volume1-0"] Oct 14 09:02:11 crc kubenswrapper[4870]: I1014 09:02:11.973896 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-volume-volume1-0" Oct 14 09:02:11 crc kubenswrapper[4870]: I1014 09:02:11.983501 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-volume-volume1-config-data" Oct 14 09:02:11 crc kubenswrapper[4870]: I1014 09:02:11.985202 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-volume-volume1-0"] Oct 14 09:02:12 crc kubenswrapper[4870]: I1014 09:02:12.151869 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/599f0358-dd66-498a-94c2-b0cfe3053ac8-dev\") pod \"cinder-volume-volume1-0\" (UID: \"599f0358-dd66-498a-94c2-b0cfe3053ac8\") " pod="openstack/cinder-volume-volume1-0" Oct 14 09:02:12 crc kubenswrapper[4870]: I1014 09:02:12.151937 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bbmxm\" (UniqueName: \"kubernetes.io/projected/599f0358-dd66-498a-94c2-b0cfe3053ac8-kube-api-access-bbmxm\") pod \"cinder-volume-volume1-0\" (UID: \"599f0358-dd66-498a-94c2-b0cfe3053ac8\") " pod="openstack/cinder-volume-volume1-0" Oct 14 09:02:12 crc kubenswrapper[4870]: I1014 09:02:12.151987 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/599f0358-dd66-498a-94c2-b0cfe3053ac8-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"599f0358-dd66-498a-94c2-b0cfe3053ac8\") " pod="openstack/cinder-volume-volume1-0" Oct 14 09:02:12 crc kubenswrapper[4870]: I1014 09:02:12.152076 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/599f0358-dd66-498a-94c2-b0cfe3053ac8-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"599f0358-dd66-498a-94c2-b0cfe3053ac8\") " pod="openstack/cinder-volume-volume1-0" Oct 14 09:02:12 crc kubenswrapper[4870]: I1014 09:02:12.152118 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/599f0358-dd66-498a-94c2-b0cfe3053ac8-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"599f0358-dd66-498a-94c2-b0cfe3053ac8\") " pod="openstack/cinder-volume-volume1-0" Oct 14 09:02:12 crc kubenswrapper[4870]: I1014 09:02:12.152167 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/599f0358-dd66-498a-94c2-b0cfe3053ac8-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"599f0358-dd66-498a-94c2-b0cfe3053ac8\") " pod="openstack/cinder-volume-volume1-0" Oct 14 09:02:12 crc kubenswrapper[4870]: I1014 09:02:12.152212 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/599f0358-dd66-498a-94c2-b0cfe3053ac8-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"599f0358-dd66-498a-94c2-b0cfe3053ac8\") " pod="openstack/cinder-volume-volume1-0" Oct 14 09:02:12 crc kubenswrapper[4870]: I1014 09:02:12.152240 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/599f0358-dd66-498a-94c2-b0cfe3053ac8-ceph\") pod \"cinder-volume-volume1-0\" (UID: \"599f0358-dd66-498a-94c2-b0cfe3053ac8\") " pod="openstack/cinder-volume-volume1-0" Oct 14 09:02:12 crc kubenswrapper[4870]: I1014 09:02:12.152272 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/599f0358-dd66-498a-94c2-b0cfe3053ac8-combined-ca-bundle\") pod \"cinder-volume-volume1-0\" (UID: \"599f0358-dd66-498a-94c2-b0cfe3053ac8\") " pod="openstack/cinder-volume-volume1-0" Oct 14 09:02:12 crc kubenswrapper[4870]: I1014 09:02:12.152304 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/599f0358-dd66-498a-94c2-b0cfe3053ac8-run\") pod \"cinder-volume-volume1-0\" (UID: \"599f0358-dd66-498a-94c2-b0cfe3053ac8\") " pod="openstack/cinder-volume-volume1-0" Oct 14 09:02:12 crc kubenswrapper[4870]: I1014 09:02:12.152340 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/599f0358-dd66-498a-94c2-b0cfe3053ac8-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"599f0358-dd66-498a-94c2-b0cfe3053ac8\") " pod="openstack/cinder-volume-volume1-0" Oct 14 09:02:12 crc kubenswrapper[4870]: I1014 09:02:12.152361 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/599f0358-dd66-498a-94c2-b0cfe3053ac8-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"599f0358-dd66-498a-94c2-b0cfe3053ac8\") " pod="openstack/cinder-volume-volume1-0" Oct 14 09:02:12 crc kubenswrapper[4870]: I1014 09:02:12.152381 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/599f0358-dd66-498a-94c2-b0cfe3053ac8-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"599f0358-dd66-498a-94c2-b0cfe3053ac8\") " pod="openstack/cinder-volume-volume1-0" Oct 14 09:02:12 crc kubenswrapper[4870]: I1014 09:02:12.152407 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/599f0358-dd66-498a-94c2-b0cfe3053ac8-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"599f0358-dd66-498a-94c2-b0cfe3053ac8\") " pod="openstack/cinder-volume-volume1-0" Oct 14 09:02:12 crc kubenswrapper[4870]: I1014 09:02:12.152435 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/599f0358-dd66-498a-94c2-b0cfe3053ac8-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"599f0358-dd66-498a-94c2-b0cfe3053ac8\") " pod="openstack/cinder-volume-volume1-0" Oct 14 09:02:12 crc kubenswrapper[4870]: I1014 09:02:12.152479 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/599f0358-dd66-498a-94c2-b0cfe3053ac8-sys\") pod \"cinder-volume-volume1-0\" (UID: \"599f0358-dd66-498a-94c2-b0cfe3053ac8\") " pod="openstack/cinder-volume-volume1-0" Oct 14 09:02:12 crc kubenswrapper[4870]: I1014 09:02:12.253674 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/599f0358-dd66-498a-94c2-b0cfe3053ac8-dev\") pod \"cinder-volume-volume1-0\" (UID: \"599f0358-dd66-498a-94c2-b0cfe3053ac8\") " pod="openstack/cinder-volume-volume1-0" Oct 14 09:02:12 crc kubenswrapper[4870]: I1014 09:02:12.253722 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bbmxm\" (UniqueName: \"kubernetes.io/projected/599f0358-dd66-498a-94c2-b0cfe3053ac8-kube-api-access-bbmxm\") pod \"cinder-volume-volume1-0\" (UID: \"599f0358-dd66-498a-94c2-b0cfe3053ac8\") " pod="openstack/cinder-volume-volume1-0" Oct 14 09:02:12 crc kubenswrapper[4870]: I1014 09:02:12.253752 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/599f0358-dd66-498a-94c2-b0cfe3053ac8-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"599f0358-dd66-498a-94c2-b0cfe3053ac8\") " pod="openstack/cinder-volume-volume1-0" Oct 14 09:02:12 crc kubenswrapper[4870]: I1014 09:02:12.253781 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/599f0358-dd66-498a-94c2-b0cfe3053ac8-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"599f0358-dd66-498a-94c2-b0cfe3053ac8\") " pod="openstack/cinder-volume-volume1-0" Oct 14 09:02:12 crc kubenswrapper[4870]: I1014 09:02:12.253805 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/599f0358-dd66-498a-94c2-b0cfe3053ac8-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"599f0358-dd66-498a-94c2-b0cfe3053ac8\") " pod="openstack/cinder-volume-volume1-0" Oct 14 09:02:12 crc kubenswrapper[4870]: I1014 09:02:12.253815 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/599f0358-dd66-498a-94c2-b0cfe3053ac8-dev\") pod \"cinder-volume-volume1-0\" (UID: \"599f0358-dd66-498a-94c2-b0cfe3053ac8\") " pod="openstack/cinder-volume-volume1-0" Oct 14 09:02:12 crc kubenswrapper[4870]: I1014 09:02:12.253838 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/599f0358-dd66-498a-94c2-b0cfe3053ac8-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"599f0358-dd66-498a-94c2-b0cfe3053ac8\") " pod="openstack/cinder-volume-volume1-0" Oct 14 09:02:12 crc kubenswrapper[4870]: I1014 09:02:12.253874 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/599f0358-dd66-498a-94c2-b0cfe3053ac8-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"599f0358-dd66-498a-94c2-b0cfe3053ac8\") " pod="openstack/cinder-volume-volume1-0" Oct 14 09:02:12 crc kubenswrapper[4870]: I1014 09:02:12.253903 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/599f0358-dd66-498a-94c2-b0cfe3053ac8-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"599f0358-dd66-498a-94c2-b0cfe3053ac8\") " pod="openstack/cinder-volume-volume1-0" Oct 14 09:02:12 crc kubenswrapper[4870]: I1014 09:02:12.253950 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/599f0358-dd66-498a-94c2-b0cfe3053ac8-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"599f0358-dd66-498a-94c2-b0cfe3053ac8\") " pod="openstack/cinder-volume-volume1-0" Oct 14 09:02:12 crc kubenswrapper[4870]: I1014 09:02:12.253981 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/599f0358-dd66-498a-94c2-b0cfe3053ac8-ceph\") pod \"cinder-volume-volume1-0\" (UID: \"599f0358-dd66-498a-94c2-b0cfe3053ac8\") " pod="openstack/cinder-volume-volume1-0" Oct 14 09:02:12 crc kubenswrapper[4870]: I1014 09:02:12.254036 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/599f0358-dd66-498a-94c2-b0cfe3053ac8-combined-ca-bundle\") pod \"cinder-volume-volume1-0\" (UID: \"599f0358-dd66-498a-94c2-b0cfe3053ac8\") " pod="openstack/cinder-volume-volume1-0" Oct 14 09:02:12 crc kubenswrapper[4870]: I1014 09:02:12.254078 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/599f0358-dd66-498a-94c2-b0cfe3053ac8-run\") pod \"cinder-volume-volume1-0\" (UID: \"599f0358-dd66-498a-94c2-b0cfe3053ac8\") " pod="openstack/cinder-volume-volume1-0" Oct 14 09:02:12 crc kubenswrapper[4870]: I1014 09:02:12.254126 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/599f0358-dd66-498a-94c2-b0cfe3053ac8-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"599f0358-dd66-498a-94c2-b0cfe3053ac8\") " pod="openstack/cinder-volume-volume1-0" Oct 14 09:02:12 crc kubenswrapper[4870]: I1014 09:02:12.254142 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/599f0358-dd66-498a-94c2-b0cfe3053ac8-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"599f0358-dd66-498a-94c2-b0cfe3053ac8\") " pod="openstack/cinder-volume-volume1-0" Oct 14 09:02:12 crc kubenswrapper[4870]: I1014 09:02:12.254155 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/599f0358-dd66-498a-94c2-b0cfe3053ac8-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"599f0358-dd66-498a-94c2-b0cfe3053ac8\") " pod="openstack/cinder-volume-volume1-0" Oct 14 09:02:12 crc kubenswrapper[4870]: I1014 09:02:12.254182 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/599f0358-dd66-498a-94c2-b0cfe3053ac8-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"599f0358-dd66-498a-94c2-b0cfe3053ac8\") " pod="openstack/cinder-volume-volume1-0" Oct 14 09:02:12 crc kubenswrapper[4870]: I1014 09:02:12.254241 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/599f0358-dd66-498a-94c2-b0cfe3053ac8-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"599f0358-dd66-498a-94c2-b0cfe3053ac8\") " pod="openstack/cinder-volume-volume1-0" Oct 14 09:02:12 crc kubenswrapper[4870]: I1014 09:02:12.254259 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/599f0358-dd66-498a-94c2-b0cfe3053ac8-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"599f0358-dd66-498a-94c2-b0cfe3053ac8\") " pod="openstack/cinder-volume-volume1-0" Oct 14 09:02:12 crc kubenswrapper[4870]: I1014 09:02:12.254189 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/599f0358-dd66-498a-94c2-b0cfe3053ac8-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"599f0358-dd66-498a-94c2-b0cfe3053ac8\") " pod="openstack/cinder-volume-volume1-0" Oct 14 09:02:12 crc kubenswrapper[4870]: I1014 09:02:12.254281 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/599f0358-dd66-498a-94c2-b0cfe3053ac8-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"599f0358-dd66-498a-94c2-b0cfe3053ac8\") " pod="openstack/cinder-volume-volume1-0" Oct 14 09:02:12 crc kubenswrapper[4870]: I1014 09:02:12.254295 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/599f0358-dd66-498a-94c2-b0cfe3053ac8-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"599f0358-dd66-498a-94c2-b0cfe3053ac8\") " pod="openstack/cinder-volume-volume1-0" Oct 14 09:02:12 crc kubenswrapper[4870]: I1014 09:02:12.254314 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/599f0358-dd66-498a-94c2-b0cfe3053ac8-sys\") pod \"cinder-volume-volume1-0\" (UID: \"599f0358-dd66-498a-94c2-b0cfe3053ac8\") " pod="openstack/cinder-volume-volume1-0" Oct 14 09:02:12 crc kubenswrapper[4870]: I1014 09:02:12.254335 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/599f0358-dd66-498a-94c2-b0cfe3053ac8-sys\") pod \"cinder-volume-volume1-0\" (UID: \"599f0358-dd66-498a-94c2-b0cfe3053ac8\") " pod="openstack/cinder-volume-volume1-0" Oct 14 09:02:12 crc kubenswrapper[4870]: I1014 09:02:12.254297 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/599f0358-dd66-498a-94c2-b0cfe3053ac8-run\") pod \"cinder-volume-volume1-0\" (UID: \"599f0358-dd66-498a-94c2-b0cfe3053ac8\") " pod="openstack/cinder-volume-volume1-0" Oct 14 09:02:12 crc kubenswrapper[4870]: I1014 09:02:12.254315 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/599f0358-dd66-498a-94c2-b0cfe3053ac8-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"599f0358-dd66-498a-94c2-b0cfe3053ac8\") " pod="openstack/cinder-volume-volume1-0" Oct 14 09:02:12 crc kubenswrapper[4870]: I1014 09:02:12.258009 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/599f0358-dd66-498a-94c2-b0cfe3053ac8-ceph\") pod \"cinder-volume-volume1-0\" (UID: \"599f0358-dd66-498a-94c2-b0cfe3053ac8\") " pod="openstack/cinder-volume-volume1-0" Oct 14 09:02:12 crc kubenswrapper[4870]: I1014 09:02:12.258246 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/599f0358-dd66-498a-94c2-b0cfe3053ac8-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"599f0358-dd66-498a-94c2-b0cfe3053ac8\") " pod="openstack/cinder-volume-volume1-0" Oct 14 09:02:12 crc kubenswrapper[4870]: I1014 09:02:12.258943 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/599f0358-dd66-498a-94c2-b0cfe3053ac8-combined-ca-bundle\") pod \"cinder-volume-volume1-0\" (UID: \"599f0358-dd66-498a-94c2-b0cfe3053ac8\") " pod="openstack/cinder-volume-volume1-0" Oct 14 09:02:12 crc kubenswrapper[4870]: I1014 09:02:12.260273 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/599f0358-dd66-498a-94c2-b0cfe3053ac8-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"599f0358-dd66-498a-94c2-b0cfe3053ac8\") " pod="openstack/cinder-volume-volume1-0" Oct 14 09:02:12 crc kubenswrapper[4870]: I1014 09:02:12.267687 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/599f0358-dd66-498a-94c2-b0cfe3053ac8-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"599f0358-dd66-498a-94c2-b0cfe3053ac8\") " pod="openstack/cinder-volume-volume1-0" Oct 14 09:02:12 crc kubenswrapper[4870]: I1014 09:02:12.275634 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bbmxm\" (UniqueName: \"kubernetes.io/projected/599f0358-dd66-498a-94c2-b0cfe3053ac8-kube-api-access-bbmxm\") pod \"cinder-volume-volume1-0\" (UID: \"599f0358-dd66-498a-94c2-b0cfe3053ac8\") " pod="openstack/cinder-volume-volume1-0" Oct 14 09:02:12 crc kubenswrapper[4870]: I1014 09:02:12.335917 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-volume-volume1-0" Oct 14 09:02:12 crc kubenswrapper[4870]: I1014 09:02:12.500465 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-backup-0"] Oct 14 09:02:12 crc kubenswrapper[4870]: I1014 09:02:12.502409 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-backup-0" Oct 14 09:02:12 crc kubenswrapper[4870]: I1014 09:02:12.508374 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-backup-config-data" Oct 14 09:02:12 crc kubenswrapper[4870]: I1014 09:02:12.516179 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-backup-0"] Oct 14 09:02:12 crc kubenswrapper[4870]: I1014 09:02:12.543791 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"cb4814fd-7bd7-430e-a390-159f8f9756e8","Type":"ContainerStarted","Data":"25557807f08792a43a72386bbb6a7562a89ca90f729c8be080136a3ecb98e8f4"} Oct 14 09:02:12 crc kubenswrapper[4870]: I1014 09:02:12.543828 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"cb4814fd-7bd7-430e-a390-159f8f9756e8","Type":"ContainerStarted","Data":"ff04ef7fa291cdae82a8079bd5e68607f24a1612e3e1305db27347d575e692fc"} Oct 14 09:02:12 crc kubenswrapper[4870]: I1014 09:02:12.567771 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.291469092 podStartE2EDuration="3.567750374s" podCreationTimestamp="2025-10-14 09:02:09 +0000 UTC" firstStartedPulling="2025-10-14 09:02:10.654366988 +0000 UTC m=+7266.351727379" lastFinishedPulling="2025-10-14 09:02:10.93064829 +0000 UTC m=+7266.628008661" observedRunningTime="2025-10-14 09:02:12.562723669 +0000 UTC m=+7268.260084040" watchObservedRunningTime="2025-10-14 09:02:12.567750374 +0000 UTC m=+7268.265110745" Oct 14 09:02:12 crc kubenswrapper[4870]: I1014 09:02:12.663402 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/0852377f-00a7-406c-91e0-83310c8c4abc-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"0852377f-00a7-406c-91e0-83310c8c4abc\") " pod="openstack/cinder-backup-0" Oct 14 09:02:12 crc kubenswrapper[4870]: I1014 09:02:12.663809 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/0852377f-00a7-406c-91e0-83310c8c4abc-run\") pod \"cinder-backup-0\" (UID: \"0852377f-00a7-406c-91e0-83310c8c4abc\") " pod="openstack/cinder-backup-0" Oct 14 09:02:12 crc kubenswrapper[4870]: I1014 09:02:12.663851 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0852377f-00a7-406c-91e0-83310c8c4abc-scripts\") pod \"cinder-backup-0\" (UID: \"0852377f-00a7-406c-91e0-83310c8c4abc\") " pod="openstack/cinder-backup-0" Oct 14 09:02:12 crc kubenswrapper[4870]: I1014 09:02:12.663892 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/0852377f-00a7-406c-91e0-83310c8c4abc-dev\") pod \"cinder-backup-0\" (UID: \"0852377f-00a7-406c-91e0-83310c8c4abc\") " pod="openstack/cinder-backup-0" Oct 14 09:02:12 crc kubenswrapper[4870]: I1014 09:02:12.663910 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/0852377f-00a7-406c-91e0-83310c8c4abc-sys\") pod \"cinder-backup-0\" (UID: \"0852377f-00a7-406c-91e0-83310c8c4abc\") " pod="openstack/cinder-backup-0" Oct 14 09:02:12 crc kubenswrapper[4870]: I1014 09:02:12.663940 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0852377f-00a7-406c-91e0-83310c8c4abc-config-data-custom\") pod \"cinder-backup-0\" (UID: \"0852377f-00a7-406c-91e0-83310c8c4abc\") " pod="openstack/cinder-backup-0" Oct 14 09:02:12 crc kubenswrapper[4870]: I1014 09:02:12.663965 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/0852377f-00a7-406c-91e0-83310c8c4abc-etc-nvme\") pod \"cinder-backup-0\" (UID: \"0852377f-00a7-406c-91e0-83310c8c4abc\") " pod="openstack/cinder-backup-0" Oct 14 09:02:12 crc kubenswrapper[4870]: I1014 09:02:12.663980 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/0852377f-00a7-406c-91e0-83310c8c4abc-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"0852377f-00a7-406c-91e0-83310c8c4abc\") " pod="openstack/cinder-backup-0" Oct 14 09:02:12 crc kubenswrapper[4870]: I1014 09:02:12.664019 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0852377f-00a7-406c-91e0-83310c8c4abc-config-data\") pod \"cinder-backup-0\" (UID: \"0852377f-00a7-406c-91e0-83310c8c4abc\") " pod="openstack/cinder-backup-0" Oct 14 09:02:12 crc kubenswrapper[4870]: I1014 09:02:12.664037 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/0852377f-00a7-406c-91e0-83310c8c4abc-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"0852377f-00a7-406c-91e0-83310c8c4abc\") " pod="openstack/cinder-backup-0" Oct 14 09:02:12 crc kubenswrapper[4870]: I1014 09:02:12.664075 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/0852377f-00a7-406c-91e0-83310c8c4abc-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"0852377f-00a7-406c-91e0-83310c8c4abc\") " pod="openstack/cinder-backup-0" Oct 14 09:02:12 crc kubenswrapper[4870]: I1014 09:02:12.664125 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0852377f-00a7-406c-91e0-83310c8c4abc-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"0852377f-00a7-406c-91e0-83310c8c4abc\") " pod="openstack/cinder-backup-0" Oct 14 09:02:12 crc kubenswrapper[4870]: I1014 09:02:12.664140 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/0852377f-00a7-406c-91e0-83310c8c4abc-ceph\") pod \"cinder-backup-0\" (UID: \"0852377f-00a7-406c-91e0-83310c8c4abc\") " pod="openstack/cinder-backup-0" Oct 14 09:02:12 crc kubenswrapper[4870]: I1014 09:02:12.664163 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0852377f-00a7-406c-91e0-83310c8c4abc-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"0852377f-00a7-406c-91e0-83310c8c4abc\") " pod="openstack/cinder-backup-0" Oct 14 09:02:12 crc kubenswrapper[4870]: I1014 09:02:12.664191 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/0852377f-00a7-406c-91e0-83310c8c4abc-lib-modules\") pod \"cinder-backup-0\" (UID: \"0852377f-00a7-406c-91e0-83310c8c4abc\") " pod="openstack/cinder-backup-0" Oct 14 09:02:12 crc kubenswrapper[4870]: I1014 09:02:12.664249 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dxl88\" (UniqueName: \"kubernetes.io/projected/0852377f-00a7-406c-91e0-83310c8c4abc-kube-api-access-dxl88\") pod \"cinder-backup-0\" (UID: \"0852377f-00a7-406c-91e0-83310c8c4abc\") " pod="openstack/cinder-backup-0" Oct 14 09:02:12 crc kubenswrapper[4870]: I1014 09:02:12.735817 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-volume-volume1-0"] Oct 14 09:02:12 crc kubenswrapper[4870]: W1014 09:02:12.736905 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod599f0358_dd66_498a_94c2_b0cfe3053ac8.slice/crio-d538ebee4f51ab0e04e82cc464576fdc90851c083483821380d943758eb04aeb WatchSource:0}: Error finding container d538ebee4f51ab0e04e82cc464576fdc90851c083483821380d943758eb04aeb: Status 404 returned error can't find the container with id d538ebee4f51ab0e04e82cc464576fdc90851c083483821380d943758eb04aeb Oct 14 09:02:12 crc kubenswrapper[4870]: I1014 09:02:12.766461 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/0852377f-00a7-406c-91e0-83310c8c4abc-dev\") pod \"cinder-backup-0\" (UID: \"0852377f-00a7-406c-91e0-83310c8c4abc\") " pod="openstack/cinder-backup-0" Oct 14 09:02:12 crc kubenswrapper[4870]: I1014 09:02:12.766525 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/0852377f-00a7-406c-91e0-83310c8c4abc-sys\") pod \"cinder-backup-0\" (UID: \"0852377f-00a7-406c-91e0-83310c8c4abc\") " pod="openstack/cinder-backup-0" Oct 14 09:02:12 crc kubenswrapper[4870]: I1014 09:02:12.766563 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0852377f-00a7-406c-91e0-83310c8c4abc-config-data-custom\") pod \"cinder-backup-0\" (UID: \"0852377f-00a7-406c-91e0-83310c8c4abc\") " pod="openstack/cinder-backup-0" Oct 14 09:02:12 crc kubenswrapper[4870]: I1014 09:02:12.766588 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/0852377f-00a7-406c-91e0-83310c8c4abc-dev\") pod \"cinder-backup-0\" (UID: \"0852377f-00a7-406c-91e0-83310c8c4abc\") " pod="openstack/cinder-backup-0" Oct 14 09:02:12 crc kubenswrapper[4870]: I1014 09:02:12.766593 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/0852377f-00a7-406c-91e0-83310c8c4abc-etc-nvme\") pod \"cinder-backup-0\" (UID: \"0852377f-00a7-406c-91e0-83310c8c4abc\") " pod="openstack/cinder-backup-0" Oct 14 09:02:12 crc kubenswrapper[4870]: I1014 09:02:12.766651 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/0852377f-00a7-406c-91e0-83310c8c4abc-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"0852377f-00a7-406c-91e0-83310c8c4abc\") " pod="openstack/cinder-backup-0" Oct 14 09:02:12 crc kubenswrapper[4870]: I1014 09:02:12.766656 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/0852377f-00a7-406c-91e0-83310c8c4abc-sys\") pod \"cinder-backup-0\" (UID: \"0852377f-00a7-406c-91e0-83310c8c4abc\") " pod="openstack/cinder-backup-0" Oct 14 09:02:12 crc kubenswrapper[4870]: I1014 09:02:12.766660 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/0852377f-00a7-406c-91e0-83310c8c4abc-etc-nvme\") pod \"cinder-backup-0\" (UID: \"0852377f-00a7-406c-91e0-83310c8c4abc\") " pod="openstack/cinder-backup-0" Oct 14 09:02:12 crc kubenswrapper[4870]: I1014 09:02:12.766699 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/0852377f-00a7-406c-91e0-83310c8c4abc-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"0852377f-00a7-406c-91e0-83310c8c4abc\") " pod="openstack/cinder-backup-0" Oct 14 09:02:12 crc kubenswrapper[4870]: I1014 09:02:12.766724 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0852377f-00a7-406c-91e0-83310c8c4abc-config-data\") pod \"cinder-backup-0\" (UID: \"0852377f-00a7-406c-91e0-83310c8c4abc\") " pod="openstack/cinder-backup-0" Oct 14 09:02:12 crc kubenswrapper[4870]: I1014 09:02:12.766755 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/0852377f-00a7-406c-91e0-83310c8c4abc-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"0852377f-00a7-406c-91e0-83310c8c4abc\") " pod="openstack/cinder-backup-0" Oct 14 09:02:12 crc kubenswrapper[4870]: I1014 09:02:12.766797 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/0852377f-00a7-406c-91e0-83310c8c4abc-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"0852377f-00a7-406c-91e0-83310c8c4abc\") " pod="openstack/cinder-backup-0" Oct 14 09:02:12 crc kubenswrapper[4870]: I1014 09:02:12.766838 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/0852377f-00a7-406c-91e0-83310c8c4abc-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"0852377f-00a7-406c-91e0-83310c8c4abc\") " pod="openstack/cinder-backup-0" Oct 14 09:02:12 crc kubenswrapper[4870]: I1014 09:02:12.766858 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0852377f-00a7-406c-91e0-83310c8c4abc-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"0852377f-00a7-406c-91e0-83310c8c4abc\") " pod="openstack/cinder-backup-0" Oct 14 09:02:12 crc kubenswrapper[4870]: I1014 09:02:12.766886 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/0852377f-00a7-406c-91e0-83310c8c4abc-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"0852377f-00a7-406c-91e0-83310c8c4abc\") " pod="openstack/cinder-backup-0" Oct 14 09:02:12 crc kubenswrapper[4870]: I1014 09:02:12.766889 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/0852377f-00a7-406c-91e0-83310c8c4abc-ceph\") pod \"cinder-backup-0\" (UID: \"0852377f-00a7-406c-91e0-83310c8c4abc\") " pod="openstack/cinder-backup-0" Oct 14 09:02:12 crc kubenswrapper[4870]: I1014 09:02:12.766940 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0852377f-00a7-406c-91e0-83310c8c4abc-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"0852377f-00a7-406c-91e0-83310c8c4abc\") " pod="openstack/cinder-backup-0" Oct 14 09:02:12 crc kubenswrapper[4870]: I1014 09:02:12.766979 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/0852377f-00a7-406c-91e0-83310c8c4abc-lib-modules\") pod \"cinder-backup-0\" (UID: \"0852377f-00a7-406c-91e0-83310c8c4abc\") " pod="openstack/cinder-backup-0" Oct 14 09:02:12 crc kubenswrapper[4870]: I1014 09:02:12.767008 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dxl88\" (UniqueName: \"kubernetes.io/projected/0852377f-00a7-406c-91e0-83310c8c4abc-kube-api-access-dxl88\") pod \"cinder-backup-0\" (UID: \"0852377f-00a7-406c-91e0-83310c8c4abc\") " pod="openstack/cinder-backup-0" Oct 14 09:02:12 crc kubenswrapper[4870]: I1014 09:02:12.767036 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/0852377f-00a7-406c-91e0-83310c8c4abc-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"0852377f-00a7-406c-91e0-83310c8c4abc\") " pod="openstack/cinder-backup-0" Oct 14 09:02:12 crc kubenswrapper[4870]: I1014 09:02:12.767047 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0852377f-00a7-406c-91e0-83310c8c4abc-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"0852377f-00a7-406c-91e0-83310c8c4abc\") " pod="openstack/cinder-backup-0" Oct 14 09:02:12 crc kubenswrapper[4870]: I1014 09:02:12.767082 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/0852377f-00a7-406c-91e0-83310c8c4abc-run\") pod \"cinder-backup-0\" (UID: \"0852377f-00a7-406c-91e0-83310c8c4abc\") " pod="openstack/cinder-backup-0" Oct 14 09:02:12 crc kubenswrapper[4870]: I1014 09:02:12.767105 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/0852377f-00a7-406c-91e0-83310c8c4abc-lib-modules\") pod \"cinder-backup-0\" (UID: \"0852377f-00a7-406c-91e0-83310c8c4abc\") " pod="openstack/cinder-backup-0" Oct 14 09:02:12 crc kubenswrapper[4870]: I1014 09:02:12.767121 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0852377f-00a7-406c-91e0-83310c8c4abc-scripts\") pod \"cinder-backup-0\" (UID: \"0852377f-00a7-406c-91e0-83310c8c4abc\") " pod="openstack/cinder-backup-0" Oct 14 09:02:12 crc kubenswrapper[4870]: I1014 09:02:12.767304 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/0852377f-00a7-406c-91e0-83310c8c4abc-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"0852377f-00a7-406c-91e0-83310c8c4abc\") " pod="openstack/cinder-backup-0" Oct 14 09:02:12 crc kubenswrapper[4870]: I1014 09:02:12.767335 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/0852377f-00a7-406c-91e0-83310c8c4abc-run\") pod \"cinder-backup-0\" (UID: \"0852377f-00a7-406c-91e0-83310c8c4abc\") " pod="openstack/cinder-backup-0" Oct 14 09:02:12 crc kubenswrapper[4870]: I1014 09:02:12.790106 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0852377f-00a7-406c-91e0-83310c8c4abc-scripts\") pod \"cinder-backup-0\" (UID: \"0852377f-00a7-406c-91e0-83310c8c4abc\") " pod="openstack/cinder-backup-0" Oct 14 09:02:12 crc kubenswrapper[4870]: I1014 09:02:12.790408 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0852377f-00a7-406c-91e0-83310c8c4abc-config-data-custom\") pod \"cinder-backup-0\" (UID: \"0852377f-00a7-406c-91e0-83310c8c4abc\") " pod="openstack/cinder-backup-0" Oct 14 09:02:12 crc kubenswrapper[4870]: I1014 09:02:12.793523 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0852377f-00a7-406c-91e0-83310c8c4abc-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"0852377f-00a7-406c-91e0-83310c8c4abc\") " pod="openstack/cinder-backup-0" Oct 14 09:02:12 crc kubenswrapper[4870]: I1014 09:02:12.796798 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/0852377f-00a7-406c-91e0-83310c8c4abc-ceph\") pod \"cinder-backup-0\" (UID: \"0852377f-00a7-406c-91e0-83310c8c4abc\") " pod="openstack/cinder-backup-0" Oct 14 09:02:12 crc kubenswrapper[4870]: I1014 09:02:12.798540 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dxl88\" (UniqueName: \"kubernetes.io/projected/0852377f-00a7-406c-91e0-83310c8c4abc-kube-api-access-dxl88\") pod \"cinder-backup-0\" (UID: \"0852377f-00a7-406c-91e0-83310c8c4abc\") " pod="openstack/cinder-backup-0" Oct 14 09:02:12 crc kubenswrapper[4870]: I1014 09:02:12.800583 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0852377f-00a7-406c-91e0-83310c8c4abc-config-data\") pod \"cinder-backup-0\" (UID: \"0852377f-00a7-406c-91e0-83310c8c4abc\") " pod="openstack/cinder-backup-0" Oct 14 09:02:12 crc kubenswrapper[4870]: I1014 09:02:12.832958 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-backup-0" Oct 14 09:02:13 crc kubenswrapper[4870]: I1014 09:02:13.391976 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-backup-0"] Oct 14 09:02:13 crc kubenswrapper[4870]: W1014 09:02:13.400870 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0852377f_00a7_406c_91e0_83310c8c4abc.slice/crio-01a1ca81a9bf4202ab659e052361ad02a4a09eb82c4426d43a257ec60ee0781b WatchSource:0}: Error finding container 01a1ca81a9bf4202ab659e052361ad02a4a09eb82c4426d43a257ec60ee0781b: Status 404 returned error can't find the container with id 01a1ca81a9bf4202ab659e052361ad02a4a09eb82c4426d43a257ec60ee0781b Oct 14 09:02:13 crc kubenswrapper[4870]: I1014 09:02:13.556244 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-backup-0" event={"ID":"0852377f-00a7-406c-91e0-83310c8c4abc","Type":"ContainerStarted","Data":"01a1ca81a9bf4202ab659e052361ad02a4a09eb82c4426d43a257ec60ee0781b"} Oct 14 09:02:13 crc kubenswrapper[4870]: I1014 09:02:13.558512 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-volume1-0" event={"ID":"599f0358-dd66-498a-94c2-b0cfe3053ac8","Type":"ContainerStarted","Data":"1421781d9b902987964207c61e7911d9537e6a35b5ef8e9c7e24cd582eff5652"} Oct 14 09:02:13 crc kubenswrapper[4870]: I1014 09:02:13.558557 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-volume1-0" event={"ID":"599f0358-dd66-498a-94c2-b0cfe3053ac8","Type":"ContainerStarted","Data":"d538ebee4f51ab0e04e82cc464576fdc90851c083483821380d943758eb04aeb"} Oct 14 09:02:14 crc kubenswrapper[4870]: I1014 09:02:14.571025 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-backup-0" event={"ID":"0852377f-00a7-406c-91e0-83310c8c4abc","Type":"ContainerStarted","Data":"f692b261951dc6f66ec8e3504437b10f6127d3f769a154413a1b27e33e0acf61"} Oct 14 09:02:14 crc kubenswrapper[4870]: I1014 09:02:14.571761 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-backup-0" event={"ID":"0852377f-00a7-406c-91e0-83310c8c4abc","Type":"ContainerStarted","Data":"98a8d665e8a43004478f764b0fea35e710d21d04e0fb63d5a2f9846e81706977"} Oct 14 09:02:14 crc kubenswrapper[4870]: I1014 09:02:14.578610 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-volume1-0" event={"ID":"599f0358-dd66-498a-94c2-b0cfe3053ac8","Type":"ContainerStarted","Data":"6c55a37d60d524b30f8c5b93e5dd4aedaaeaf9962af019dead7a2d0eef8298ea"} Oct 14 09:02:14 crc kubenswrapper[4870]: I1014 09:02:14.623919 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-backup-0" podStartSLOduration=2.307610963 podStartE2EDuration="2.623894793s" podCreationTimestamp="2025-10-14 09:02:12 +0000 UTC" firstStartedPulling="2025-10-14 09:02:13.402657139 +0000 UTC m=+7269.100017510" lastFinishedPulling="2025-10-14 09:02:13.718940969 +0000 UTC m=+7269.416301340" observedRunningTime="2025-10-14 09:02:14.61930332 +0000 UTC m=+7270.316663691" watchObservedRunningTime="2025-10-14 09:02:14.623894793 +0000 UTC m=+7270.321255164" Oct 14 09:02:14 crc kubenswrapper[4870]: I1014 09:02:14.662306 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-volume-volume1-0" podStartSLOduration=3.246115215 podStartE2EDuration="3.662288811s" podCreationTimestamp="2025-10-14 09:02:11 +0000 UTC" firstStartedPulling="2025-10-14 09:02:12.742926949 +0000 UTC m=+7268.440287320" lastFinishedPulling="2025-10-14 09:02:13.159100535 +0000 UTC m=+7268.856460916" observedRunningTime="2025-10-14 09:02:14.651525126 +0000 UTC m=+7270.348885497" watchObservedRunningTime="2025-10-14 09:02:14.662288811 +0000 UTC m=+7270.359649182" Oct 14 09:02:14 crc kubenswrapper[4870]: I1014 09:02:14.790338 4870 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/cinder-api-0" podUID="a662858e-76ac-4bf6-949d-4bc17f6a3aaf" containerName="cinder-api" probeResult="failure" output="Get \"http://10.217.1.91:8776/healthcheck\": read tcp 10.217.0.2:37626->10.217.1.91:8776: read: connection reset by peer" Oct 14 09:02:15 crc kubenswrapper[4870]: I1014 09:02:15.186705 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 14 09:02:15 crc kubenswrapper[4870]: I1014 09:02:15.190851 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Oct 14 09:02:15 crc kubenswrapper[4870]: I1014 09:02:15.327814 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a662858e-76ac-4bf6-949d-4bc17f6a3aaf-logs\") pod \"a662858e-76ac-4bf6-949d-4bc17f6a3aaf\" (UID: \"a662858e-76ac-4bf6-949d-4bc17f6a3aaf\") " Oct 14 09:02:15 crc kubenswrapper[4870]: I1014 09:02:15.328209 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a662858e-76ac-4bf6-949d-4bc17f6a3aaf-scripts\") pod \"a662858e-76ac-4bf6-949d-4bc17f6a3aaf\" (UID: \"a662858e-76ac-4bf6-949d-4bc17f6a3aaf\") " Oct 14 09:02:15 crc kubenswrapper[4870]: I1014 09:02:15.328313 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a662858e-76ac-4bf6-949d-4bc17f6a3aaf-config-data\") pod \"a662858e-76ac-4bf6-949d-4bc17f6a3aaf\" (UID: \"a662858e-76ac-4bf6-949d-4bc17f6a3aaf\") " Oct 14 09:02:15 crc kubenswrapper[4870]: I1014 09:02:15.328408 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g7jbh\" (UniqueName: \"kubernetes.io/projected/a662858e-76ac-4bf6-949d-4bc17f6a3aaf-kube-api-access-g7jbh\") pod \"a662858e-76ac-4bf6-949d-4bc17f6a3aaf\" (UID: \"a662858e-76ac-4bf6-949d-4bc17f6a3aaf\") " Oct 14 09:02:15 crc kubenswrapper[4870]: I1014 09:02:15.328474 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a662858e-76ac-4bf6-949d-4bc17f6a3aaf-combined-ca-bundle\") pod \"a662858e-76ac-4bf6-949d-4bc17f6a3aaf\" (UID: \"a662858e-76ac-4bf6-949d-4bc17f6a3aaf\") " Oct 14 09:02:15 crc kubenswrapper[4870]: I1014 09:02:15.328875 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a662858e-76ac-4bf6-949d-4bc17f6a3aaf-logs" (OuterVolumeSpecName: "logs") pod "a662858e-76ac-4bf6-949d-4bc17f6a3aaf" (UID: "a662858e-76ac-4bf6-949d-4bc17f6a3aaf"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 09:02:15 crc kubenswrapper[4870]: I1014 09:02:15.329573 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a662858e-76ac-4bf6-949d-4bc17f6a3aaf-etc-machine-id\") pod \"a662858e-76ac-4bf6-949d-4bc17f6a3aaf\" (UID: \"a662858e-76ac-4bf6-949d-4bc17f6a3aaf\") " Oct 14 09:02:15 crc kubenswrapper[4870]: I1014 09:02:15.329690 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a662858e-76ac-4bf6-949d-4bc17f6a3aaf-config-data-custom\") pod \"a662858e-76ac-4bf6-949d-4bc17f6a3aaf\" (UID: \"a662858e-76ac-4bf6-949d-4bc17f6a3aaf\") " Oct 14 09:02:15 crc kubenswrapper[4870]: I1014 09:02:15.330540 4870 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a662858e-76ac-4bf6-949d-4bc17f6a3aaf-logs\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:15 crc kubenswrapper[4870]: I1014 09:02:15.332022 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a662858e-76ac-4bf6-949d-4bc17f6a3aaf-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "a662858e-76ac-4bf6-949d-4bc17f6a3aaf" (UID: "a662858e-76ac-4bf6-949d-4bc17f6a3aaf"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 14 09:02:15 crc kubenswrapper[4870]: I1014 09:02:15.335634 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a662858e-76ac-4bf6-949d-4bc17f6a3aaf-kube-api-access-g7jbh" (OuterVolumeSpecName: "kube-api-access-g7jbh") pod "a662858e-76ac-4bf6-949d-4bc17f6a3aaf" (UID: "a662858e-76ac-4bf6-949d-4bc17f6a3aaf"). InnerVolumeSpecName "kube-api-access-g7jbh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:02:15 crc kubenswrapper[4870]: I1014 09:02:15.343174 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a662858e-76ac-4bf6-949d-4bc17f6a3aaf-scripts" (OuterVolumeSpecName: "scripts") pod "a662858e-76ac-4bf6-949d-4bc17f6a3aaf" (UID: "a662858e-76ac-4bf6-949d-4bc17f6a3aaf"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:02:15 crc kubenswrapper[4870]: I1014 09:02:15.375697 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a662858e-76ac-4bf6-949d-4bc17f6a3aaf-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "a662858e-76ac-4bf6-949d-4bc17f6a3aaf" (UID: "a662858e-76ac-4bf6-949d-4bc17f6a3aaf"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:02:15 crc kubenswrapper[4870]: I1014 09:02:15.380904 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a662858e-76ac-4bf6-949d-4bc17f6a3aaf-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a662858e-76ac-4bf6-949d-4bc17f6a3aaf" (UID: "a662858e-76ac-4bf6-949d-4bc17f6a3aaf"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:02:15 crc kubenswrapper[4870]: I1014 09:02:15.390356 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a662858e-76ac-4bf6-949d-4bc17f6a3aaf-config-data" (OuterVolumeSpecName: "config-data") pod "a662858e-76ac-4bf6-949d-4bc17f6a3aaf" (UID: "a662858e-76ac-4bf6-949d-4bc17f6a3aaf"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:02:15 crc kubenswrapper[4870]: I1014 09:02:15.432745 4870 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a662858e-76ac-4bf6-949d-4bc17f6a3aaf-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:15 crc kubenswrapper[4870]: I1014 09:02:15.432786 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g7jbh\" (UniqueName: \"kubernetes.io/projected/a662858e-76ac-4bf6-949d-4bc17f6a3aaf-kube-api-access-g7jbh\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:15 crc kubenswrapper[4870]: I1014 09:02:15.432802 4870 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a662858e-76ac-4bf6-949d-4bc17f6a3aaf-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:15 crc kubenswrapper[4870]: I1014 09:02:15.432813 4870 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a662858e-76ac-4bf6-949d-4bc17f6a3aaf-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:15 crc kubenswrapper[4870]: I1014 09:02:15.432828 4870 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a662858e-76ac-4bf6-949d-4bc17f6a3aaf-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:15 crc kubenswrapper[4870]: I1014 09:02:15.432840 4870 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a662858e-76ac-4bf6-949d-4bc17f6a3aaf-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:15 crc kubenswrapper[4870]: I1014 09:02:15.589646 4870 generic.go:334] "Generic (PLEG): container finished" podID="a662858e-76ac-4bf6-949d-4bc17f6a3aaf" containerID="9f71bfbc3b3893910374afed0b2400760747c481120e966d946e4c2dab9b0550" exitCode=0 Oct 14 09:02:15 crc kubenswrapper[4870]: I1014 09:02:15.589719 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 14 09:02:15 crc kubenswrapper[4870]: I1014 09:02:15.589779 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"a662858e-76ac-4bf6-949d-4bc17f6a3aaf","Type":"ContainerDied","Data":"9f71bfbc3b3893910374afed0b2400760747c481120e966d946e4c2dab9b0550"} Oct 14 09:02:15 crc kubenswrapper[4870]: I1014 09:02:15.589838 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"a662858e-76ac-4bf6-949d-4bc17f6a3aaf","Type":"ContainerDied","Data":"f1305639c736356705e48b686bd47788e0f4c6e5c9376968a8b98d594517b5c3"} Oct 14 09:02:15 crc kubenswrapper[4870]: I1014 09:02:15.589861 4870 scope.go:117] "RemoveContainer" containerID="9f71bfbc3b3893910374afed0b2400760747c481120e966d946e4c2dab9b0550" Oct 14 09:02:15 crc kubenswrapper[4870]: I1014 09:02:15.626687 4870 scope.go:117] "RemoveContainer" containerID="8747381c748307eddcef853ed8b5d3128671bd63c249dac6069447d3067eabb8" Oct 14 09:02:15 crc kubenswrapper[4870]: I1014 09:02:15.630069 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Oct 14 09:02:15 crc kubenswrapper[4870]: I1014 09:02:15.651865 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Oct 14 09:02:15 crc kubenswrapper[4870]: I1014 09:02:15.656434 4870 scope.go:117] "RemoveContainer" containerID="9f71bfbc3b3893910374afed0b2400760747c481120e966d946e4c2dab9b0550" Oct 14 09:02:15 crc kubenswrapper[4870]: E1014 09:02:15.656931 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9f71bfbc3b3893910374afed0b2400760747c481120e966d946e4c2dab9b0550\": container with ID starting with 9f71bfbc3b3893910374afed0b2400760747c481120e966d946e4c2dab9b0550 not found: ID does not exist" containerID="9f71bfbc3b3893910374afed0b2400760747c481120e966d946e4c2dab9b0550" Oct 14 09:02:15 crc kubenswrapper[4870]: I1014 09:02:15.656974 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9f71bfbc3b3893910374afed0b2400760747c481120e966d946e4c2dab9b0550"} err="failed to get container status \"9f71bfbc3b3893910374afed0b2400760747c481120e966d946e4c2dab9b0550\": rpc error: code = NotFound desc = could not find container \"9f71bfbc3b3893910374afed0b2400760747c481120e966d946e4c2dab9b0550\": container with ID starting with 9f71bfbc3b3893910374afed0b2400760747c481120e966d946e4c2dab9b0550 not found: ID does not exist" Oct 14 09:02:15 crc kubenswrapper[4870]: I1014 09:02:15.656999 4870 scope.go:117] "RemoveContainer" containerID="8747381c748307eddcef853ed8b5d3128671bd63c249dac6069447d3067eabb8" Oct 14 09:02:15 crc kubenswrapper[4870]: E1014 09:02:15.657301 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8747381c748307eddcef853ed8b5d3128671bd63c249dac6069447d3067eabb8\": container with ID starting with 8747381c748307eddcef853ed8b5d3128671bd63c249dac6069447d3067eabb8 not found: ID does not exist" containerID="8747381c748307eddcef853ed8b5d3128671bd63c249dac6069447d3067eabb8" Oct 14 09:02:15 crc kubenswrapper[4870]: I1014 09:02:15.657335 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8747381c748307eddcef853ed8b5d3128671bd63c249dac6069447d3067eabb8"} err="failed to get container status \"8747381c748307eddcef853ed8b5d3128671bd63c249dac6069447d3067eabb8\": rpc error: code = NotFound desc = could not find container \"8747381c748307eddcef853ed8b5d3128671bd63c249dac6069447d3067eabb8\": container with ID starting with 8747381c748307eddcef853ed8b5d3128671bd63c249dac6069447d3067eabb8 not found: ID does not exist" Oct 14 09:02:15 crc kubenswrapper[4870]: I1014 09:02:15.666332 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Oct 14 09:02:15 crc kubenswrapper[4870]: E1014 09:02:15.666831 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a662858e-76ac-4bf6-949d-4bc17f6a3aaf" containerName="cinder-api-log" Oct 14 09:02:15 crc kubenswrapper[4870]: I1014 09:02:15.666856 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="a662858e-76ac-4bf6-949d-4bc17f6a3aaf" containerName="cinder-api-log" Oct 14 09:02:15 crc kubenswrapper[4870]: E1014 09:02:15.666880 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a662858e-76ac-4bf6-949d-4bc17f6a3aaf" containerName="cinder-api" Oct 14 09:02:15 crc kubenswrapper[4870]: I1014 09:02:15.666890 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="a662858e-76ac-4bf6-949d-4bc17f6a3aaf" containerName="cinder-api" Oct 14 09:02:15 crc kubenswrapper[4870]: I1014 09:02:15.667161 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="a662858e-76ac-4bf6-949d-4bc17f6a3aaf" containerName="cinder-api-log" Oct 14 09:02:15 crc kubenswrapper[4870]: I1014 09:02:15.667195 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="a662858e-76ac-4bf6-949d-4bc17f6a3aaf" containerName="cinder-api" Oct 14 09:02:15 crc kubenswrapper[4870]: I1014 09:02:15.691647 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 14 09:02:15 crc kubenswrapper[4870]: I1014 09:02:15.694763 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Oct 14 09:02:15 crc kubenswrapper[4870]: I1014 09:02:15.710669 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 14 09:02:15 crc kubenswrapper[4870]: I1014 09:02:15.838304 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l79xk\" (UniqueName: \"kubernetes.io/projected/0a387a03-3520-4d50-b7c7-5e3abb17978d-kube-api-access-l79xk\") pod \"cinder-api-0\" (UID: \"0a387a03-3520-4d50-b7c7-5e3abb17978d\") " pod="openstack/cinder-api-0" Oct 14 09:02:15 crc kubenswrapper[4870]: I1014 09:02:15.838635 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a387a03-3520-4d50-b7c7-5e3abb17978d-config-data\") pod \"cinder-api-0\" (UID: \"0a387a03-3520-4d50-b7c7-5e3abb17978d\") " pod="openstack/cinder-api-0" Oct 14 09:02:15 crc kubenswrapper[4870]: I1014 09:02:15.838686 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a387a03-3520-4d50-b7c7-5e3abb17978d-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"0a387a03-3520-4d50-b7c7-5e3abb17978d\") " pod="openstack/cinder-api-0" Oct 14 09:02:15 crc kubenswrapper[4870]: I1014 09:02:15.838705 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0a387a03-3520-4d50-b7c7-5e3abb17978d-etc-machine-id\") pod \"cinder-api-0\" (UID: \"0a387a03-3520-4d50-b7c7-5e3abb17978d\") " pod="openstack/cinder-api-0" Oct 14 09:02:15 crc kubenswrapper[4870]: I1014 09:02:15.838761 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0a387a03-3520-4d50-b7c7-5e3abb17978d-logs\") pod \"cinder-api-0\" (UID: \"0a387a03-3520-4d50-b7c7-5e3abb17978d\") " pod="openstack/cinder-api-0" Oct 14 09:02:15 crc kubenswrapper[4870]: I1014 09:02:15.838837 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0a387a03-3520-4d50-b7c7-5e3abb17978d-config-data-custom\") pod \"cinder-api-0\" (UID: \"0a387a03-3520-4d50-b7c7-5e3abb17978d\") " pod="openstack/cinder-api-0" Oct 14 09:02:15 crc kubenswrapper[4870]: I1014 09:02:15.838866 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0a387a03-3520-4d50-b7c7-5e3abb17978d-scripts\") pod \"cinder-api-0\" (UID: \"0a387a03-3520-4d50-b7c7-5e3abb17978d\") " pod="openstack/cinder-api-0" Oct 14 09:02:15 crc kubenswrapper[4870]: I1014 09:02:15.941981 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a387a03-3520-4d50-b7c7-5e3abb17978d-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"0a387a03-3520-4d50-b7c7-5e3abb17978d\") " pod="openstack/cinder-api-0" Oct 14 09:02:15 crc kubenswrapper[4870]: I1014 09:02:15.942045 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0a387a03-3520-4d50-b7c7-5e3abb17978d-etc-machine-id\") pod \"cinder-api-0\" (UID: \"0a387a03-3520-4d50-b7c7-5e3abb17978d\") " pod="openstack/cinder-api-0" Oct 14 09:02:15 crc kubenswrapper[4870]: I1014 09:02:15.942155 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0a387a03-3520-4d50-b7c7-5e3abb17978d-logs\") pod \"cinder-api-0\" (UID: \"0a387a03-3520-4d50-b7c7-5e3abb17978d\") " pod="openstack/cinder-api-0" Oct 14 09:02:15 crc kubenswrapper[4870]: I1014 09:02:15.942265 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0a387a03-3520-4d50-b7c7-5e3abb17978d-config-data-custom\") pod \"cinder-api-0\" (UID: \"0a387a03-3520-4d50-b7c7-5e3abb17978d\") " pod="openstack/cinder-api-0" Oct 14 09:02:15 crc kubenswrapper[4870]: I1014 09:02:15.942498 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0a387a03-3520-4d50-b7c7-5e3abb17978d-scripts\") pod \"cinder-api-0\" (UID: \"0a387a03-3520-4d50-b7c7-5e3abb17978d\") " pod="openstack/cinder-api-0" Oct 14 09:02:15 crc kubenswrapper[4870]: I1014 09:02:15.942590 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l79xk\" (UniqueName: \"kubernetes.io/projected/0a387a03-3520-4d50-b7c7-5e3abb17978d-kube-api-access-l79xk\") pod \"cinder-api-0\" (UID: \"0a387a03-3520-4d50-b7c7-5e3abb17978d\") " pod="openstack/cinder-api-0" Oct 14 09:02:15 crc kubenswrapper[4870]: I1014 09:02:15.942672 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a387a03-3520-4d50-b7c7-5e3abb17978d-config-data\") pod \"cinder-api-0\" (UID: \"0a387a03-3520-4d50-b7c7-5e3abb17978d\") " pod="openstack/cinder-api-0" Oct 14 09:02:15 crc kubenswrapper[4870]: I1014 09:02:15.943474 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0a387a03-3520-4d50-b7c7-5e3abb17978d-etc-machine-id\") pod \"cinder-api-0\" (UID: \"0a387a03-3520-4d50-b7c7-5e3abb17978d\") " pod="openstack/cinder-api-0" Oct 14 09:02:15 crc kubenswrapper[4870]: I1014 09:02:15.943941 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0a387a03-3520-4d50-b7c7-5e3abb17978d-logs\") pod \"cinder-api-0\" (UID: \"0a387a03-3520-4d50-b7c7-5e3abb17978d\") " pod="openstack/cinder-api-0" Oct 14 09:02:15 crc kubenswrapper[4870]: I1014 09:02:15.948466 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0a387a03-3520-4d50-b7c7-5e3abb17978d-scripts\") pod \"cinder-api-0\" (UID: \"0a387a03-3520-4d50-b7c7-5e3abb17978d\") " pod="openstack/cinder-api-0" Oct 14 09:02:15 crc kubenswrapper[4870]: I1014 09:02:15.948675 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a387a03-3520-4d50-b7c7-5e3abb17978d-config-data\") pod \"cinder-api-0\" (UID: \"0a387a03-3520-4d50-b7c7-5e3abb17978d\") " pod="openstack/cinder-api-0" Oct 14 09:02:15 crc kubenswrapper[4870]: I1014 09:02:15.948763 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0a387a03-3520-4d50-b7c7-5e3abb17978d-config-data-custom\") pod \"cinder-api-0\" (UID: \"0a387a03-3520-4d50-b7c7-5e3abb17978d\") " pod="openstack/cinder-api-0" Oct 14 09:02:15 crc kubenswrapper[4870]: I1014 09:02:15.949215 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a387a03-3520-4d50-b7c7-5e3abb17978d-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"0a387a03-3520-4d50-b7c7-5e3abb17978d\") " pod="openstack/cinder-api-0" Oct 14 09:02:15 crc kubenswrapper[4870]: I1014 09:02:15.967031 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l79xk\" (UniqueName: \"kubernetes.io/projected/0a387a03-3520-4d50-b7c7-5e3abb17978d-kube-api-access-l79xk\") pod \"cinder-api-0\" (UID: \"0a387a03-3520-4d50-b7c7-5e3abb17978d\") " pod="openstack/cinder-api-0" Oct 14 09:02:16 crc kubenswrapper[4870]: I1014 09:02:16.020812 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 14 09:02:16 crc kubenswrapper[4870]: I1014 09:02:16.488393 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 14 09:02:16 crc kubenswrapper[4870]: I1014 09:02:16.601630 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"0a387a03-3520-4d50-b7c7-5e3abb17978d","Type":"ContainerStarted","Data":"3d6055d5f94d484a79f37ec13c40727fecbfb08c2db6e760f87dd83b9564d714"} Oct 14 09:02:17 crc kubenswrapper[4870]: I1014 09:02:17.048372 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a662858e-76ac-4bf6-949d-4bc17f6a3aaf" path="/var/lib/kubelet/pods/a662858e-76ac-4bf6-949d-4bc17f6a3aaf/volumes" Oct 14 09:02:17 crc kubenswrapper[4870]: I1014 09:02:17.336622 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-volume-volume1-0" Oct 14 09:02:17 crc kubenswrapper[4870]: I1014 09:02:17.610871 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"0a387a03-3520-4d50-b7c7-5e3abb17978d","Type":"ContainerStarted","Data":"1178a1fe92c0e746865441d200003f2d7d1a1e01a0da65ac322579b546aa1073"} Oct 14 09:02:17 crc kubenswrapper[4870]: I1014 09:02:17.833730 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-backup-0" Oct 14 09:02:18 crc kubenswrapper[4870]: I1014 09:02:18.041511 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-xzttw"] Oct 14 09:02:18 crc kubenswrapper[4870]: I1014 09:02:18.056968 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-xzttw"] Oct 14 09:02:18 crc kubenswrapper[4870]: I1014 09:02:18.624092 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"0a387a03-3520-4d50-b7c7-5e3abb17978d","Type":"ContainerStarted","Data":"3b18dbf574c9d0130815aea77dcbbe733c24bacc99055f471553fe9dbade093b"} Oct 14 09:02:18 crc kubenswrapper[4870]: I1014 09:02:18.624605 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Oct 14 09:02:18 crc kubenswrapper[4870]: I1014 09:02:18.658829 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=3.658800334 podStartE2EDuration="3.658800334s" podCreationTimestamp="2025-10-14 09:02:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 09:02:18.638955044 +0000 UTC m=+7274.336315485" watchObservedRunningTime="2025-10-14 09:02:18.658800334 +0000 UTC m=+7274.356160745" Oct 14 09:02:19 crc kubenswrapper[4870]: I1014 09:02:19.049525 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b024cec4-79e9-415b-aa23-725f5d15aacf" path="/var/lib/kubelet/pods/b024cec4-79e9-415b-aa23-725f5d15aacf/volumes" Oct 14 09:02:20 crc kubenswrapper[4870]: I1014 09:02:20.404822 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Oct 14 09:02:20 crc kubenswrapper[4870]: I1014 09:02:20.478198 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 14 09:02:20 crc kubenswrapper[4870]: I1014 09:02:20.645724 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="cb4814fd-7bd7-430e-a390-159f8f9756e8" containerName="cinder-scheduler" containerID="cri-o://ff04ef7fa291cdae82a8079bd5e68607f24a1612e3e1305db27347d575e692fc" gracePeriod=30 Oct 14 09:02:20 crc kubenswrapper[4870]: I1014 09:02:20.646212 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="cb4814fd-7bd7-430e-a390-159f8f9756e8" containerName="probe" containerID="cri-o://25557807f08792a43a72386bbb6a7562a89ca90f729c8be080136a3ecb98e8f4" gracePeriod=30 Oct 14 09:02:21 crc kubenswrapper[4870]: I1014 09:02:21.657077 4870 generic.go:334] "Generic (PLEG): container finished" podID="cb4814fd-7bd7-430e-a390-159f8f9756e8" containerID="25557807f08792a43a72386bbb6a7562a89ca90f729c8be080136a3ecb98e8f4" exitCode=0 Oct 14 09:02:21 crc kubenswrapper[4870]: I1014 09:02:21.657135 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"cb4814fd-7bd7-430e-a390-159f8f9756e8","Type":"ContainerDied","Data":"25557807f08792a43a72386bbb6a7562a89ca90f729c8be080136a3ecb98e8f4"} Oct 14 09:02:22 crc kubenswrapper[4870]: I1014 09:02:22.343011 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 14 09:02:22 crc kubenswrapper[4870]: I1014 09:02:22.498640 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p6dq5\" (UniqueName: \"kubernetes.io/projected/cb4814fd-7bd7-430e-a390-159f8f9756e8-kube-api-access-p6dq5\") pod \"cb4814fd-7bd7-430e-a390-159f8f9756e8\" (UID: \"cb4814fd-7bd7-430e-a390-159f8f9756e8\") " Oct 14 09:02:22 crc kubenswrapper[4870]: I1014 09:02:22.498699 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb4814fd-7bd7-430e-a390-159f8f9756e8-combined-ca-bundle\") pod \"cb4814fd-7bd7-430e-a390-159f8f9756e8\" (UID: \"cb4814fd-7bd7-430e-a390-159f8f9756e8\") " Oct 14 09:02:22 crc kubenswrapper[4870]: I1014 09:02:22.498728 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cb4814fd-7bd7-430e-a390-159f8f9756e8-scripts\") pod \"cb4814fd-7bd7-430e-a390-159f8f9756e8\" (UID: \"cb4814fd-7bd7-430e-a390-159f8f9756e8\") " Oct 14 09:02:22 crc kubenswrapper[4870]: I1014 09:02:22.498890 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cb4814fd-7bd7-430e-a390-159f8f9756e8-config-data\") pod \"cb4814fd-7bd7-430e-a390-159f8f9756e8\" (UID: \"cb4814fd-7bd7-430e-a390-159f8f9756e8\") " Oct 14 09:02:22 crc kubenswrapper[4870]: I1014 09:02:22.499112 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cb4814fd-7bd7-430e-a390-159f8f9756e8-config-data-custom\") pod \"cb4814fd-7bd7-430e-a390-159f8f9756e8\" (UID: \"cb4814fd-7bd7-430e-a390-159f8f9756e8\") " Oct 14 09:02:22 crc kubenswrapper[4870]: I1014 09:02:22.499281 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/cb4814fd-7bd7-430e-a390-159f8f9756e8-etc-machine-id\") pod \"cb4814fd-7bd7-430e-a390-159f8f9756e8\" (UID: \"cb4814fd-7bd7-430e-a390-159f8f9756e8\") " Oct 14 09:02:22 crc kubenswrapper[4870]: I1014 09:02:22.499801 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/cb4814fd-7bd7-430e-a390-159f8f9756e8-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "cb4814fd-7bd7-430e-a390-159f8f9756e8" (UID: "cb4814fd-7bd7-430e-a390-159f8f9756e8"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 14 09:02:22 crc kubenswrapper[4870]: I1014 09:02:22.503855 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cb4814fd-7bd7-430e-a390-159f8f9756e8-kube-api-access-p6dq5" (OuterVolumeSpecName: "kube-api-access-p6dq5") pod "cb4814fd-7bd7-430e-a390-159f8f9756e8" (UID: "cb4814fd-7bd7-430e-a390-159f8f9756e8"). InnerVolumeSpecName "kube-api-access-p6dq5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:02:22 crc kubenswrapper[4870]: I1014 09:02:22.511897 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cb4814fd-7bd7-430e-a390-159f8f9756e8-scripts" (OuterVolumeSpecName: "scripts") pod "cb4814fd-7bd7-430e-a390-159f8f9756e8" (UID: "cb4814fd-7bd7-430e-a390-159f8f9756e8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:02:22 crc kubenswrapper[4870]: I1014 09:02:22.517973 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cb4814fd-7bd7-430e-a390-159f8f9756e8-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "cb4814fd-7bd7-430e-a390-159f8f9756e8" (UID: "cb4814fd-7bd7-430e-a390-159f8f9756e8"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:02:22 crc kubenswrapper[4870]: I1014 09:02:22.553833 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cb4814fd-7bd7-430e-a390-159f8f9756e8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cb4814fd-7bd7-430e-a390-159f8f9756e8" (UID: "cb4814fd-7bd7-430e-a390-159f8f9756e8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:02:22 crc kubenswrapper[4870]: I1014 09:02:22.587320 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-volume-volume1-0" Oct 14 09:02:22 crc kubenswrapper[4870]: I1014 09:02:22.592026 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cb4814fd-7bd7-430e-a390-159f8f9756e8-config-data" (OuterVolumeSpecName: "config-data") pod "cb4814fd-7bd7-430e-a390-159f8f9756e8" (UID: "cb4814fd-7bd7-430e-a390-159f8f9756e8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:02:22 crc kubenswrapper[4870]: I1014 09:02:22.602103 4870 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/cb4814fd-7bd7-430e-a390-159f8f9756e8-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:22 crc kubenswrapper[4870]: I1014 09:02:22.602138 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p6dq5\" (UniqueName: \"kubernetes.io/projected/cb4814fd-7bd7-430e-a390-159f8f9756e8-kube-api-access-p6dq5\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:22 crc kubenswrapper[4870]: I1014 09:02:22.602150 4870 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb4814fd-7bd7-430e-a390-159f8f9756e8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:22 crc kubenswrapper[4870]: I1014 09:02:22.602160 4870 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cb4814fd-7bd7-430e-a390-159f8f9756e8-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:22 crc kubenswrapper[4870]: I1014 09:02:22.602170 4870 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cb4814fd-7bd7-430e-a390-159f8f9756e8-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:22 crc kubenswrapper[4870]: I1014 09:02:22.602182 4870 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cb4814fd-7bd7-430e-a390-159f8f9756e8-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 14 09:02:22 crc kubenswrapper[4870]: I1014 09:02:22.669994 4870 generic.go:334] "Generic (PLEG): container finished" podID="cb4814fd-7bd7-430e-a390-159f8f9756e8" containerID="ff04ef7fa291cdae82a8079bd5e68607f24a1612e3e1305db27347d575e692fc" exitCode=0 Oct 14 09:02:22 crc kubenswrapper[4870]: I1014 09:02:22.670047 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"cb4814fd-7bd7-430e-a390-159f8f9756e8","Type":"ContainerDied","Data":"ff04ef7fa291cdae82a8079bd5e68607f24a1612e3e1305db27347d575e692fc"} Oct 14 09:02:22 crc kubenswrapper[4870]: I1014 09:02:22.670078 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"cb4814fd-7bd7-430e-a390-159f8f9756e8","Type":"ContainerDied","Data":"26a061f76ae099e9a8894ac4d3accbb909fd7b04bc926e7b21b048407549517d"} Oct 14 09:02:22 crc kubenswrapper[4870]: I1014 09:02:22.670099 4870 scope.go:117] "RemoveContainer" containerID="25557807f08792a43a72386bbb6a7562a89ca90f729c8be080136a3ecb98e8f4" Oct 14 09:02:22 crc kubenswrapper[4870]: I1014 09:02:22.670197 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 14 09:02:22 crc kubenswrapper[4870]: I1014 09:02:22.694791 4870 scope.go:117] "RemoveContainer" containerID="ff04ef7fa291cdae82a8079bd5e68607f24a1612e3e1305db27347d575e692fc" Oct 14 09:02:22 crc kubenswrapper[4870]: I1014 09:02:22.705678 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 14 09:02:22 crc kubenswrapper[4870]: I1014 09:02:22.715939 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 14 09:02:22 crc kubenswrapper[4870]: I1014 09:02:22.740174 4870 scope.go:117] "RemoveContainer" containerID="25557807f08792a43a72386bbb6a7562a89ca90f729c8be080136a3ecb98e8f4" Oct 14 09:02:22 crc kubenswrapper[4870]: I1014 09:02:22.740832 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Oct 14 09:02:22 crc kubenswrapper[4870]: E1014 09:02:22.741272 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"25557807f08792a43a72386bbb6a7562a89ca90f729c8be080136a3ecb98e8f4\": container with ID starting with 25557807f08792a43a72386bbb6a7562a89ca90f729c8be080136a3ecb98e8f4 not found: ID does not exist" containerID="25557807f08792a43a72386bbb6a7562a89ca90f729c8be080136a3ecb98e8f4" Oct 14 09:02:22 crc kubenswrapper[4870]: I1014 09:02:22.741362 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"25557807f08792a43a72386bbb6a7562a89ca90f729c8be080136a3ecb98e8f4"} err="failed to get container status \"25557807f08792a43a72386bbb6a7562a89ca90f729c8be080136a3ecb98e8f4\": rpc error: code = NotFound desc = could not find container \"25557807f08792a43a72386bbb6a7562a89ca90f729c8be080136a3ecb98e8f4\": container with ID starting with 25557807f08792a43a72386bbb6a7562a89ca90f729c8be080136a3ecb98e8f4 not found: ID does not exist" Oct 14 09:02:22 crc kubenswrapper[4870]: I1014 09:02:22.741456 4870 scope.go:117] "RemoveContainer" containerID="ff04ef7fa291cdae82a8079bd5e68607f24a1612e3e1305db27347d575e692fc" Oct 14 09:02:22 crc kubenswrapper[4870]: E1014 09:02:22.741489 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb4814fd-7bd7-430e-a390-159f8f9756e8" containerName="probe" Oct 14 09:02:22 crc kubenswrapper[4870]: I1014 09:02:22.741740 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb4814fd-7bd7-430e-a390-159f8f9756e8" containerName="probe" Oct 14 09:02:22 crc kubenswrapper[4870]: E1014 09:02:22.741758 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb4814fd-7bd7-430e-a390-159f8f9756e8" containerName="cinder-scheduler" Oct 14 09:02:22 crc kubenswrapper[4870]: I1014 09:02:22.741765 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb4814fd-7bd7-430e-a390-159f8f9756e8" containerName="cinder-scheduler" Oct 14 09:02:22 crc kubenswrapper[4870]: I1014 09:02:22.741976 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="cb4814fd-7bd7-430e-a390-159f8f9756e8" containerName="probe" Oct 14 09:02:22 crc kubenswrapper[4870]: I1014 09:02:22.741992 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="cb4814fd-7bd7-430e-a390-159f8f9756e8" containerName="cinder-scheduler" Oct 14 09:02:22 crc kubenswrapper[4870]: E1014 09:02:22.742753 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ff04ef7fa291cdae82a8079bd5e68607f24a1612e3e1305db27347d575e692fc\": container with ID starting with ff04ef7fa291cdae82a8079bd5e68607f24a1612e3e1305db27347d575e692fc not found: ID does not exist" containerID="ff04ef7fa291cdae82a8079bd5e68607f24a1612e3e1305db27347d575e692fc" Oct 14 09:02:22 crc kubenswrapper[4870]: I1014 09:02:22.742814 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ff04ef7fa291cdae82a8079bd5e68607f24a1612e3e1305db27347d575e692fc"} err="failed to get container status \"ff04ef7fa291cdae82a8079bd5e68607f24a1612e3e1305db27347d575e692fc\": rpc error: code = NotFound desc = could not find container \"ff04ef7fa291cdae82a8079bd5e68607f24a1612e3e1305db27347d575e692fc\": container with ID starting with ff04ef7fa291cdae82a8079bd5e68607f24a1612e3e1305db27347d575e692fc not found: ID does not exist" Oct 14 09:02:22 crc kubenswrapper[4870]: I1014 09:02:22.743040 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 14 09:02:22 crc kubenswrapper[4870]: I1014 09:02:22.744634 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Oct 14 09:02:22 crc kubenswrapper[4870]: I1014 09:02:22.762740 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 14 09:02:22 crc kubenswrapper[4870]: I1014 09:02:22.906915 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/dad751a8-e411-4244-8bae-004bcad630a0-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"dad751a8-e411-4244-8bae-004bcad630a0\") " pod="openstack/cinder-scheduler-0" Oct 14 09:02:22 crc kubenswrapper[4870]: I1014 09:02:22.907201 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/dad751a8-e411-4244-8bae-004bcad630a0-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"dad751a8-e411-4244-8bae-004bcad630a0\") " pod="openstack/cinder-scheduler-0" Oct 14 09:02:22 crc kubenswrapper[4870]: I1014 09:02:22.907369 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q4pmn\" (UniqueName: \"kubernetes.io/projected/dad751a8-e411-4244-8bae-004bcad630a0-kube-api-access-q4pmn\") pod \"cinder-scheduler-0\" (UID: \"dad751a8-e411-4244-8bae-004bcad630a0\") " pod="openstack/cinder-scheduler-0" Oct 14 09:02:22 crc kubenswrapper[4870]: I1014 09:02:22.907548 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dad751a8-e411-4244-8bae-004bcad630a0-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"dad751a8-e411-4244-8bae-004bcad630a0\") " pod="openstack/cinder-scheduler-0" Oct 14 09:02:22 crc kubenswrapper[4870]: I1014 09:02:22.907648 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dad751a8-e411-4244-8bae-004bcad630a0-scripts\") pod \"cinder-scheduler-0\" (UID: \"dad751a8-e411-4244-8bae-004bcad630a0\") " pod="openstack/cinder-scheduler-0" Oct 14 09:02:22 crc kubenswrapper[4870]: I1014 09:02:22.907726 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dad751a8-e411-4244-8bae-004bcad630a0-config-data\") pod \"cinder-scheduler-0\" (UID: \"dad751a8-e411-4244-8bae-004bcad630a0\") " pod="openstack/cinder-scheduler-0" Oct 14 09:02:23 crc kubenswrapper[4870]: I1014 09:02:23.009500 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dad751a8-e411-4244-8bae-004bcad630a0-scripts\") pod \"cinder-scheduler-0\" (UID: \"dad751a8-e411-4244-8bae-004bcad630a0\") " pod="openstack/cinder-scheduler-0" Oct 14 09:02:23 crc kubenswrapper[4870]: I1014 09:02:23.009550 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dad751a8-e411-4244-8bae-004bcad630a0-config-data\") pod \"cinder-scheduler-0\" (UID: \"dad751a8-e411-4244-8bae-004bcad630a0\") " pod="openstack/cinder-scheduler-0" Oct 14 09:02:23 crc kubenswrapper[4870]: I1014 09:02:23.009583 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/dad751a8-e411-4244-8bae-004bcad630a0-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"dad751a8-e411-4244-8bae-004bcad630a0\") " pod="openstack/cinder-scheduler-0" Oct 14 09:02:23 crc kubenswrapper[4870]: I1014 09:02:23.009627 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/dad751a8-e411-4244-8bae-004bcad630a0-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"dad751a8-e411-4244-8bae-004bcad630a0\") " pod="openstack/cinder-scheduler-0" Oct 14 09:02:23 crc kubenswrapper[4870]: I1014 09:02:23.009696 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q4pmn\" (UniqueName: \"kubernetes.io/projected/dad751a8-e411-4244-8bae-004bcad630a0-kube-api-access-q4pmn\") pod \"cinder-scheduler-0\" (UID: \"dad751a8-e411-4244-8bae-004bcad630a0\") " pod="openstack/cinder-scheduler-0" Oct 14 09:02:23 crc kubenswrapper[4870]: I1014 09:02:23.009731 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dad751a8-e411-4244-8bae-004bcad630a0-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"dad751a8-e411-4244-8bae-004bcad630a0\") " pod="openstack/cinder-scheduler-0" Oct 14 09:02:23 crc kubenswrapper[4870]: I1014 09:02:23.010870 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/dad751a8-e411-4244-8bae-004bcad630a0-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"dad751a8-e411-4244-8bae-004bcad630a0\") " pod="openstack/cinder-scheduler-0" Oct 14 09:02:23 crc kubenswrapper[4870]: I1014 09:02:23.014499 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dad751a8-e411-4244-8bae-004bcad630a0-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"dad751a8-e411-4244-8bae-004bcad630a0\") " pod="openstack/cinder-scheduler-0" Oct 14 09:02:23 crc kubenswrapper[4870]: I1014 09:02:23.015080 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dad751a8-e411-4244-8bae-004bcad630a0-scripts\") pod \"cinder-scheduler-0\" (UID: \"dad751a8-e411-4244-8bae-004bcad630a0\") " pod="openstack/cinder-scheduler-0" Oct 14 09:02:23 crc kubenswrapper[4870]: I1014 09:02:23.015131 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/dad751a8-e411-4244-8bae-004bcad630a0-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"dad751a8-e411-4244-8bae-004bcad630a0\") " pod="openstack/cinder-scheduler-0" Oct 14 09:02:23 crc kubenswrapper[4870]: I1014 09:02:23.015569 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dad751a8-e411-4244-8bae-004bcad630a0-config-data\") pod \"cinder-scheduler-0\" (UID: \"dad751a8-e411-4244-8bae-004bcad630a0\") " pod="openstack/cinder-scheduler-0" Oct 14 09:02:23 crc kubenswrapper[4870]: I1014 09:02:23.027787 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q4pmn\" (UniqueName: \"kubernetes.io/projected/dad751a8-e411-4244-8bae-004bcad630a0-kube-api-access-q4pmn\") pod \"cinder-scheduler-0\" (UID: \"dad751a8-e411-4244-8bae-004bcad630a0\") " pod="openstack/cinder-scheduler-0" Oct 14 09:02:23 crc kubenswrapper[4870]: I1014 09:02:23.048448 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cb4814fd-7bd7-430e-a390-159f8f9756e8" path="/var/lib/kubelet/pods/cb4814fd-7bd7-430e-a390-159f8f9756e8/volumes" Oct 14 09:02:23 crc kubenswrapper[4870]: I1014 09:02:23.049611 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-backup-0" Oct 14 09:02:23 crc kubenswrapper[4870]: I1014 09:02:23.076808 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 14 09:02:23 crc kubenswrapper[4870]: I1014 09:02:23.645348 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 14 09:02:23 crc kubenswrapper[4870]: I1014 09:02:23.686357 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"dad751a8-e411-4244-8bae-004bcad630a0","Type":"ContainerStarted","Data":"bf90737b687bbbd0e5bcecee89842bb897e191cd443a9d0424252a9e17ae15e6"} Oct 14 09:02:24 crc kubenswrapper[4870]: I1014 09:02:24.698160 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"dad751a8-e411-4244-8bae-004bcad630a0","Type":"ContainerStarted","Data":"29e2a65e840f3dcc3bf6988232fecb3b8d49208122f2d7e8ac0aae3767a6a789"} Oct 14 09:02:25 crc kubenswrapper[4870]: I1014 09:02:25.731927 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"dad751a8-e411-4244-8bae-004bcad630a0","Type":"ContainerStarted","Data":"16fce691adc58e647916a6ae283fc95565466b911cb3dae0c884ebdc8f494706"} Oct 14 09:02:25 crc kubenswrapper[4870]: I1014 09:02:25.765977 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.7659493939999997 podStartE2EDuration="3.765949394s" podCreationTimestamp="2025-10-14 09:02:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 09:02:25.756280505 +0000 UTC m=+7281.453640896" watchObservedRunningTime="2025-10-14 09:02:25.765949394 +0000 UTC m=+7281.463309765" Oct 14 09:02:27 crc kubenswrapper[4870]: I1014 09:02:27.870468 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Oct 14 09:02:28 crc kubenswrapper[4870]: I1014 09:02:28.040563 4870 scope.go:117] "RemoveContainer" containerID="41d1c1da64483937ac1c1d1375ba7897ac6f748fde7cb100e83f2fbbd1051084" Oct 14 09:02:28 crc kubenswrapper[4870]: I1014 09:02:28.077876 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Oct 14 09:02:28 crc kubenswrapper[4870]: I1014 09:02:28.082490 4870 scope.go:117] "RemoveContainer" containerID="ad7dfa6a7348cf4fe7ce935499a7c6f4e55c1a1c92056375846ad1c8d0fac2f7" Oct 14 09:02:28 crc kubenswrapper[4870]: I1014 09:02:28.113040 4870 scope.go:117] "RemoveContainer" containerID="d1ca5583cb67929436e14fd7ee24f07c74a1e72bcdded1a788b5dbda1ef79a48" Oct 14 09:02:32 crc kubenswrapper[4870]: I1014 09:02:32.051243 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-vf77d"] Oct 14 09:02:32 crc kubenswrapper[4870]: I1014 09:02:32.066476 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-vf77d"] Oct 14 09:02:33 crc kubenswrapper[4870]: I1014 09:02:33.047332 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4068efdd-8071-411b-9302-3e26f1bb44c2" path="/var/lib/kubelet/pods/4068efdd-8071-411b-9302-3e26f1bb44c2/volumes" Oct 14 09:02:33 crc kubenswrapper[4870]: I1014 09:02:33.304573 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Oct 14 09:03:28 crc kubenswrapper[4870]: I1014 09:03:28.356566 4870 scope.go:117] "RemoveContainer" containerID="88870aca262234f0dd719a72dad60be2f2c1437bb35f0d0e9af2e0596d07e866" Oct 14 09:03:48 crc kubenswrapper[4870]: I1014 09:03:48.271155 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-2tk5n"] Oct 14 09:03:48 crc kubenswrapper[4870]: I1014 09:03:48.274637 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2tk5n" Oct 14 09:03:48 crc kubenswrapper[4870]: I1014 09:03:48.283153 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-2tk5n"] Oct 14 09:03:48 crc kubenswrapper[4870]: I1014 09:03:48.365982 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fa58de33-a6ea-4970-bdc5-9b947a069050-catalog-content\") pod \"redhat-marketplace-2tk5n\" (UID: \"fa58de33-a6ea-4970-bdc5-9b947a069050\") " pod="openshift-marketplace/redhat-marketplace-2tk5n" Oct 14 09:03:48 crc kubenswrapper[4870]: I1014 09:03:48.366229 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pxvp6\" (UniqueName: \"kubernetes.io/projected/fa58de33-a6ea-4970-bdc5-9b947a069050-kube-api-access-pxvp6\") pod \"redhat-marketplace-2tk5n\" (UID: \"fa58de33-a6ea-4970-bdc5-9b947a069050\") " pod="openshift-marketplace/redhat-marketplace-2tk5n" Oct 14 09:03:48 crc kubenswrapper[4870]: I1014 09:03:48.366368 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fa58de33-a6ea-4970-bdc5-9b947a069050-utilities\") pod \"redhat-marketplace-2tk5n\" (UID: \"fa58de33-a6ea-4970-bdc5-9b947a069050\") " pod="openshift-marketplace/redhat-marketplace-2tk5n" Oct 14 09:03:48 crc kubenswrapper[4870]: I1014 09:03:48.468680 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pxvp6\" (UniqueName: \"kubernetes.io/projected/fa58de33-a6ea-4970-bdc5-9b947a069050-kube-api-access-pxvp6\") pod \"redhat-marketplace-2tk5n\" (UID: \"fa58de33-a6ea-4970-bdc5-9b947a069050\") " pod="openshift-marketplace/redhat-marketplace-2tk5n" Oct 14 09:03:48 crc kubenswrapper[4870]: I1014 09:03:48.468765 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fa58de33-a6ea-4970-bdc5-9b947a069050-utilities\") pod \"redhat-marketplace-2tk5n\" (UID: \"fa58de33-a6ea-4970-bdc5-9b947a069050\") " pod="openshift-marketplace/redhat-marketplace-2tk5n" Oct 14 09:03:48 crc kubenswrapper[4870]: I1014 09:03:48.468917 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fa58de33-a6ea-4970-bdc5-9b947a069050-catalog-content\") pod \"redhat-marketplace-2tk5n\" (UID: \"fa58de33-a6ea-4970-bdc5-9b947a069050\") " pod="openshift-marketplace/redhat-marketplace-2tk5n" Oct 14 09:03:48 crc kubenswrapper[4870]: I1014 09:03:48.469497 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fa58de33-a6ea-4970-bdc5-9b947a069050-utilities\") pod \"redhat-marketplace-2tk5n\" (UID: \"fa58de33-a6ea-4970-bdc5-9b947a069050\") " pod="openshift-marketplace/redhat-marketplace-2tk5n" Oct 14 09:03:48 crc kubenswrapper[4870]: I1014 09:03:48.469595 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fa58de33-a6ea-4970-bdc5-9b947a069050-catalog-content\") pod \"redhat-marketplace-2tk5n\" (UID: \"fa58de33-a6ea-4970-bdc5-9b947a069050\") " pod="openshift-marketplace/redhat-marketplace-2tk5n" Oct 14 09:03:48 crc kubenswrapper[4870]: I1014 09:03:48.496504 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pxvp6\" (UniqueName: \"kubernetes.io/projected/fa58de33-a6ea-4970-bdc5-9b947a069050-kube-api-access-pxvp6\") pod \"redhat-marketplace-2tk5n\" (UID: \"fa58de33-a6ea-4970-bdc5-9b947a069050\") " pod="openshift-marketplace/redhat-marketplace-2tk5n" Oct 14 09:03:48 crc kubenswrapper[4870]: I1014 09:03:48.606338 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2tk5n" Oct 14 09:03:49 crc kubenswrapper[4870]: I1014 09:03:49.075639 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-2tk5n"] Oct 14 09:03:49 crc kubenswrapper[4870]: I1014 09:03:49.615253 4870 generic.go:334] "Generic (PLEG): container finished" podID="fa58de33-a6ea-4970-bdc5-9b947a069050" containerID="47a4267b8e4fbb61bf2c690cb4746d4a415f517803cbbabd4ec4d51af7d67270" exitCode=0 Oct 14 09:03:49 crc kubenswrapper[4870]: I1014 09:03:49.615320 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2tk5n" event={"ID":"fa58de33-a6ea-4970-bdc5-9b947a069050","Type":"ContainerDied","Data":"47a4267b8e4fbb61bf2c690cb4746d4a415f517803cbbabd4ec4d51af7d67270"} Oct 14 09:03:49 crc kubenswrapper[4870]: I1014 09:03:49.615380 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2tk5n" event={"ID":"fa58de33-a6ea-4970-bdc5-9b947a069050","Type":"ContainerStarted","Data":"104f2ac6669878bed5f71b6867d186e76a6ef05c1a657e2f60c4b8d31fc6f749"} Oct 14 09:03:50 crc kubenswrapper[4870]: I1014 09:03:50.630624 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2tk5n" event={"ID":"fa58de33-a6ea-4970-bdc5-9b947a069050","Type":"ContainerStarted","Data":"83a908aed744b6473217d1eb21648c8d4c8fe1dc41b189b5f53fd4770ea97cc0"} Oct 14 09:03:51 crc kubenswrapper[4870]: I1014 09:03:51.643711 4870 generic.go:334] "Generic (PLEG): container finished" podID="fa58de33-a6ea-4970-bdc5-9b947a069050" containerID="83a908aed744b6473217d1eb21648c8d4c8fe1dc41b189b5f53fd4770ea97cc0" exitCode=0 Oct 14 09:03:51 crc kubenswrapper[4870]: I1014 09:03:51.643773 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2tk5n" event={"ID":"fa58de33-a6ea-4970-bdc5-9b947a069050","Type":"ContainerDied","Data":"83a908aed744b6473217d1eb21648c8d4c8fe1dc41b189b5f53fd4770ea97cc0"} Oct 14 09:03:53 crc kubenswrapper[4870]: I1014 09:03:53.695676 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2tk5n" event={"ID":"fa58de33-a6ea-4970-bdc5-9b947a069050","Type":"ContainerStarted","Data":"ba70abae180c7a12d643a2fd8348db0b6e7c8552010c7041e1f86c5fb6ad89ad"} Oct 14 09:03:53 crc kubenswrapper[4870]: I1014 09:03:53.719626 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-2tk5n" podStartSLOduration=2.863285755 podStartE2EDuration="5.719598502s" podCreationTimestamp="2025-10-14 09:03:48 +0000 UTC" firstStartedPulling="2025-10-14 09:03:49.618106408 +0000 UTC m=+7365.315466779" lastFinishedPulling="2025-10-14 09:03:52.474419145 +0000 UTC m=+7368.171779526" observedRunningTime="2025-10-14 09:03:53.71630984 +0000 UTC m=+7369.413670261" watchObservedRunningTime="2025-10-14 09:03:53.719598502 +0000 UTC m=+7369.416958913" Oct 14 09:03:58 crc kubenswrapper[4870]: I1014 09:03:58.606618 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-2tk5n" Oct 14 09:03:58 crc kubenswrapper[4870]: I1014 09:03:58.607526 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-2tk5n" Oct 14 09:03:58 crc kubenswrapper[4870]: I1014 09:03:58.681009 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-2tk5n" Oct 14 09:03:58 crc kubenswrapper[4870]: I1014 09:03:58.818500 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-2tk5n" Oct 14 09:03:58 crc kubenswrapper[4870]: I1014 09:03:58.927857 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-2tk5n"] Oct 14 09:04:00 crc kubenswrapper[4870]: I1014 09:04:00.782831 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-2tk5n" podUID="fa58de33-a6ea-4970-bdc5-9b947a069050" containerName="registry-server" containerID="cri-o://ba70abae180c7a12d643a2fd8348db0b6e7c8552010c7041e1f86c5fb6ad89ad" gracePeriod=2 Oct 14 09:04:01 crc kubenswrapper[4870]: I1014 09:04:01.240424 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2tk5n" Oct 14 09:04:01 crc kubenswrapper[4870]: I1014 09:04:01.361309 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fa58de33-a6ea-4970-bdc5-9b947a069050-utilities\") pod \"fa58de33-a6ea-4970-bdc5-9b947a069050\" (UID: \"fa58de33-a6ea-4970-bdc5-9b947a069050\") " Oct 14 09:04:01 crc kubenswrapper[4870]: I1014 09:04:01.361632 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pxvp6\" (UniqueName: \"kubernetes.io/projected/fa58de33-a6ea-4970-bdc5-9b947a069050-kube-api-access-pxvp6\") pod \"fa58de33-a6ea-4970-bdc5-9b947a069050\" (UID: \"fa58de33-a6ea-4970-bdc5-9b947a069050\") " Oct 14 09:04:01 crc kubenswrapper[4870]: I1014 09:04:01.361852 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fa58de33-a6ea-4970-bdc5-9b947a069050-catalog-content\") pod \"fa58de33-a6ea-4970-bdc5-9b947a069050\" (UID: \"fa58de33-a6ea-4970-bdc5-9b947a069050\") " Oct 14 09:04:01 crc kubenswrapper[4870]: I1014 09:04:01.362433 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fa58de33-a6ea-4970-bdc5-9b947a069050-utilities" (OuterVolumeSpecName: "utilities") pod "fa58de33-a6ea-4970-bdc5-9b947a069050" (UID: "fa58de33-a6ea-4970-bdc5-9b947a069050"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 09:04:01 crc kubenswrapper[4870]: I1014 09:04:01.362706 4870 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fa58de33-a6ea-4970-bdc5-9b947a069050-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 09:04:01 crc kubenswrapper[4870]: I1014 09:04:01.377732 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fa58de33-a6ea-4970-bdc5-9b947a069050-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "fa58de33-a6ea-4970-bdc5-9b947a069050" (UID: "fa58de33-a6ea-4970-bdc5-9b947a069050"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 09:04:01 crc kubenswrapper[4870]: I1014 09:04:01.378722 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fa58de33-a6ea-4970-bdc5-9b947a069050-kube-api-access-pxvp6" (OuterVolumeSpecName: "kube-api-access-pxvp6") pod "fa58de33-a6ea-4970-bdc5-9b947a069050" (UID: "fa58de33-a6ea-4970-bdc5-9b947a069050"). InnerVolumeSpecName "kube-api-access-pxvp6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:04:01 crc kubenswrapper[4870]: I1014 09:04:01.463978 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pxvp6\" (UniqueName: \"kubernetes.io/projected/fa58de33-a6ea-4970-bdc5-9b947a069050-kube-api-access-pxvp6\") on node \"crc\" DevicePath \"\"" Oct 14 09:04:01 crc kubenswrapper[4870]: I1014 09:04:01.464021 4870 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fa58de33-a6ea-4970-bdc5-9b947a069050-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 09:04:01 crc kubenswrapper[4870]: I1014 09:04:01.791242 4870 generic.go:334] "Generic (PLEG): container finished" podID="fa58de33-a6ea-4970-bdc5-9b947a069050" containerID="ba70abae180c7a12d643a2fd8348db0b6e7c8552010c7041e1f86c5fb6ad89ad" exitCode=0 Oct 14 09:04:01 crc kubenswrapper[4870]: I1014 09:04:01.791287 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2tk5n" event={"ID":"fa58de33-a6ea-4970-bdc5-9b947a069050","Type":"ContainerDied","Data":"ba70abae180c7a12d643a2fd8348db0b6e7c8552010c7041e1f86c5fb6ad89ad"} Oct 14 09:04:01 crc kubenswrapper[4870]: I1014 09:04:01.791332 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2tk5n" Oct 14 09:04:01 crc kubenswrapper[4870]: I1014 09:04:01.791351 4870 scope.go:117] "RemoveContainer" containerID="ba70abae180c7a12d643a2fd8348db0b6e7c8552010c7041e1f86c5fb6ad89ad" Oct 14 09:04:01 crc kubenswrapper[4870]: I1014 09:04:01.791333 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2tk5n" event={"ID":"fa58de33-a6ea-4970-bdc5-9b947a069050","Type":"ContainerDied","Data":"104f2ac6669878bed5f71b6867d186e76a6ef05c1a657e2f60c4b8d31fc6f749"} Oct 14 09:04:01 crc kubenswrapper[4870]: I1014 09:04:01.811362 4870 scope.go:117] "RemoveContainer" containerID="83a908aed744b6473217d1eb21648c8d4c8fe1dc41b189b5f53fd4770ea97cc0" Oct 14 09:04:01 crc kubenswrapper[4870]: I1014 09:04:01.832587 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-2tk5n"] Oct 14 09:04:01 crc kubenswrapper[4870]: I1014 09:04:01.835674 4870 scope.go:117] "RemoveContainer" containerID="47a4267b8e4fbb61bf2c690cb4746d4a415f517803cbbabd4ec4d51af7d67270" Oct 14 09:04:01 crc kubenswrapper[4870]: I1014 09:04:01.840687 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-2tk5n"] Oct 14 09:04:01 crc kubenswrapper[4870]: I1014 09:04:01.892993 4870 scope.go:117] "RemoveContainer" containerID="ba70abae180c7a12d643a2fd8348db0b6e7c8552010c7041e1f86c5fb6ad89ad" Oct 14 09:04:01 crc kubenswrapper[4870]: E1014 09:04:01.893586 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ba70abae180c7a12d643a2fd8348db0b6e7c8552010c7041e1f86c5fb6ad89ad\": container with ID starting with ba70abae180c7a12d643a2fd8348db0b6e7c8552010c7041e1f86c5fb6ad89ad not found: ID does not exist" containerID="ba70abae180c7a12d643a2fd8348db0b6e7c8552010c7041e1f86c5fb6ad89ad" Oct 14 09:04:01 crc kubenswrapper[4870]: I1014 09:04:01.893687 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ba70abae180c7a12d643a2fd8348db0b6e7c8552010c7041e1f86c5fb6ad89ad"} err="failed to get container status \"ba70abae180c7a12d643a2fd8348db0b6e7c8552010c7041e1f86c5fb6ad89ad\": rpc error: code = NotFound desc = could not find container \"ba70abae180c7a12d643a2fd8348db0b6e7c8552010c7041e1f86c5fb6ad89ad\": container with ID starting with ba70abae180c7a12d643a2fd8348db0b6e7c8552010c7041e1f86c5fb6ad89ad not found: ID does not exist" Oct 14 09:04:01 crc kubenswrapper[4870]: I1014 09:04:01.893714 4870 scope.go:117] "RemoveContainer" containerID="83a908aed744b6473217d1eb21648c8d4c8fe1dc41b189b5f53fd4770ea97cc0" Oct 14 09:04:01 crc kubenswrapper[4870]: E1014 09:04:01.894298 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"83a908aed744b6473217d1eb21648c8d4c8fe1dc41b189b5f53fd4770ea97cc0\": container with ID starting with 83a908aed744b6473217d1eb21648c8d4c8fe1dc41b189b5f53fd4770ea97cc0 not found: ID does not exist" containerID="83a908aed744b6473217d1eb21648c8d4c8fe1dc41b189b5f53fd4770ea97cc0" Oct 14 09:04:01 crc kubenswrapper[4870]: I1014 09:04:01.894347 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"83a908aed744b6473217d1eb21648c8d4c8fe1dc41b189b5f53fd4770ea97cc0"} err="failed to get container status \"83a908aed744b6473217d1eb21648c8d4c8fe1dc41b189b5f53fd4770ea97cc0\": rpc error: code = NotFound desc = could not find container \"83a908aed744b6473217d1eb21648c8d4c8fe1dc41b189b5f53fd4770ea97cc0\": container with ID starting with 83a908aed744b6473217d1eb21648c8d4c8fe1dc41b189b5f53fd4770ea97cc0 not found: ID does not exist" Oct 14 09:04:01 crc kubenswrapper[4870]: I1014 09:04:01.894363 4870 scope.go:117] "RemoveContainer" containerID="47a4267b8e4fbb61bf2c690cb4746d4a415f517803cbbabd4ec4d51af7d67270" Oct 14 09:04:01 crc kubenswrapper[4870]: E1014 09:04:01.894672 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"47a4267b8e4fbb61bf2c690cb4746d4a415f517803cbbabd4ec4d51af7d67270\": container with ID starting with 47a4267b8e4fbb61bf2c690cb4746d4a415f517803cbbabd4ec4d51af7d67270 not found: ID does not exist" containerID="47a4267b8e4fbb61bf2c690cb4746d4a415f517803cbbabd4ec4d51af7d67270" Oct 14 09:04:01 crc kubenswrapper[4870]: I1014 09:04:01.894728 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"47a4267b8e4fbb61bf2c690cb4746d4a415f517803cbbabd4ec4d51af7d67270"} err="failed to get container status \"47a4267b8e4fbb61bf2c690cb4746d4a415f517803cbbabd4ec4d51af7d67270\": rpc error: code = NotFound desc = could not find container \"47a4267b8e4fbb61bf2c690cb4746d4a415f517803cbbabd4ec4d51af7d67270\": container with ID starting with 47a4267b8e4fbb61bf2c690cb4746d4a415f517803cbbabd4ec4d51af7d67270 not found: ID does not exist" Oct 14 09:04:03 crc kubenswrapper[4870]: I1014 09:04:03.055866 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fa58de33-a6ea-4970-bdc5-9b947a069050" path="/var/lib/kubelet/pods/fa58de33-a6ea-4970-bdc5-9b947a069050/volumes" Oct 14 09:04:07 crc kubenswrapper[4870]: I1014 09:04:07.428111 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-869b6669bf-lc5s8"] Oct 14 09:04:07 crc kubenswrapper[4870]: E1014 09:04:07.429195 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa58de33-a6ea-4970-bdc5-9b947a069050" containerName="extract-content" Oct 14 09:04:07 crc kubenswrapper[4870]: I1014 09:04:07.429213 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa58de33-a6ea-4970-bdc5-9b947a069050" containerName="extract-content" Oct 14 09:04:07 crc kubenswrapper[4870]: E1014 09:04:07.429245 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa58de33-a6ea-4970-bdc5-9b947a069050" containerName="registry-server" Oct 14 09:04:07 crc kubenswrapper[4870]: I1014 09:04:07.429254 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa58de33-a6ea-4970-bdc5-9b947a069050" containerName="registry-server" Oct 14 09:04:07 crc kubenswrapper[4870]: E1014 09:04:07.429272 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa58de33-a6ea-4970-bdc5-9b947a069050" containerName="extract-utilities" Oct 14 09:04:07 crc kubenswrapper[4870]: I1014 09:04:07.429284 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa58de33-a6ea-4970-bdc5-9b947a069050" containerName="extract-utilities" Oct 14 09:04:07 crc kubenswrapper[4870]: I1014 09:04:07.429556 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="fa58de33-a6ea-4970-bdc5-9b947a069050" containerName="registry-server" Oct 14 09:04:07 crc kubenswrapper[4870]: I1014 09:04:07.430821 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-869b6669bf-lc5s8" Oct 14 09:04:07 crc kubenswrapper[4870]: I1014 09:04:07.437230 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon" Oct 14 09:04:07 crc kubenswrapper[4870]: I1014 09:04:07.437274 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-scripts" Oct 14 09:04:07 crc kubenswrapper[4870]: I1014 09:04:07.437242 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon-horizon-dockercfg-2h6ck" Oct 14 09:04:07 crc kubenswrapper[4870]: I1014 09:04:07.437390 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-config-data" Oct 14 09:04:07 crc kubenswrapper[4870]: I1014 09:04:07.444248 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-869b6669bf-lc5s8"] Oct 14 09:04:07 crc kubenswrapper[4870]: I1014 09:04:07.537634 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 14 09:04:07 crc kubenswrapper[4870]: I1014 09:04:07.538158 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="de9df7d3-70ec-4abd-9cd0-55038d9ab2df" containerName="glance-log" containerID="cri-o://8eb5a5a2c1544b193bef195717e303075859ef64fbf449f7c5c4eb693af477f6" gracePeriod=30 Oct 14 09:04:07 crc kubenswrapper[4870]: I1014 09:04:07.539407 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="de9df7d3-70ec-4abd-9cd0-55038d9ab2df" containerName="glance-httpd" containerID="cri-o://d5be18839f4b371ce0db12541e6bfc5e5ccd1903925a9840da4265c4d2692924" gracePeriod=30 Oct 14 09:04:07 crc kubenswrapper[4870]: I1014 09:04:07.568632 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 14 09:04:07 crc kubenswrapper[4870]: I1014 09:04:07.568895 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="829bce4c-6fe7-46fa-a4d8-1558220a039f" containerName="glance-log" containerID="cri-o://dcdf64ab0812d5c47611df139478fcdb35fb58c52140a0bfaf26de928cff6bc0" gracePeriod=30 Oct 14 09:04:07 crc kubenswrapper[4870]: I1014 09:04:07.569364 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="829bce4c-6fe7-46fa-a4d8-1558220a039f" containerName="glance-httpd" containerID="cri-o://6d39e737807423f661bed2bc5c4c82a9315751ab881b785991a4633e1b26b216" gracePeriod=30 Oct 14 09:04:07 crc kubenswrapper[4870]: I1014 09:04:07.580745 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rnhnl\" (UniqueName: \"kubernetes.io/projected/2ea687b1-b04d-476a-a333-a731a71e7a66-kube-api-access-rnhnl\") pod \"horizon-869b6669bf-lc5s8\" (UID: \"2ea687b1-b04d-476a-a333-a731a71e7a66\") " pod="openstack/horizon-869b6669bf-lc5s8" Oct 14 09:04:07 crc kubenswrapper[4870]: I1014 09:04:07.580797 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/2ea687b1-b04d-476a-a333-a731a71e7a66-horizon-secret-key\") pod \"horizon-869b6669bf-lc5s8\" (UID: \"2ea687b1-b04d-476a-a333-a731a71e7a66\") " pod="openstack/horizon-869b6669bf-lc5s8" Oct 14 09:04:07 crc kubenswrapper[4870]: I1014 09:04:07.580814 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2ea687b1-b04d-476a-a333-a731a71e7a66-logs\") pod \"horizon-869b6669bf-lc5s8\" (UID: \"2ea687b1-b04d-476a-a333-a731a71e7a66\") " pod="openstack/horizon-869b6669bf-lc5s8" Oct 14 09:04:07 crc kubenswrapper[4870]: I1014 09:04:07.580883 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2ea687b1-b04d-476a-a333-a731a71e7a66-scripts\") pod \"horizon-869b6669bf-lc5s8\" (UID: \"2ea687b1-b04d-476a-a333-a731a71e7a66\") " pod="openstack/horizon-869b6669bf-lc5s8" Oct 14 09:04:07 crc kubenswrapper[4870]: I1014 09:04:07.580902 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2ea687b1-b04d-476a-a333-a731a71e7a66-config-data\") pod \"horizon-869b6669bf-lc5s8\" (UID: \"2ea687b1-b04d-476a-a333-a731a71e7a66\") " pod="openstack/horizon-869b6669bf-lc5s8" Oct 14 09:04:07 crc kubenswrapper[4870]: I1014 09:04:07.612035 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-7fbc647f5-9ph8d"] Oct 14 09:04:07 crc kubenswrapper[4870]: I1014 09:04:07.614078 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7fbc647f5-9ph8d" Oct 14 09:04:07 crc kubenswrapper[4870]: I1014 09:04:07.624498 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7fbc647f5-9ph8d"] Oct 14 09:04:07 crc kubenswrapper[4870]: I1014 09:04:07.683553 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rnhnl\" (UniqueName: \"kubernetes.io/projected/2ea687b1-b04d-476a-a333-a731a71e7a66-kube-api-access-rnhnl\") pod \"horizon-869b6669bf-lc5s8\" (UID: \"2ea687b1-b04d-476a-a333-a731a71e7a66\") " pod="openstack/horizon-869b6669bf-lc5s8" Oct 14 09:04:07 crc kubenswrapper[4870]: I1014 09:04:07.683615 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/2ea687b1-b04d-476a-a333-a731a71e7a66-horizon-secret-key\") pod \"horizon-869b6669bf-lc5s8\" (UID: \"2ea687b1-b04d-476a-a333-a731a71e7a66\") " pod="openstack/horizon-869b6669bf-lc5s8" Oct 14 09:04:07 crc kubenswrapper[4870]: I1014 09:04:07.683635 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2ea687b1-b04d-476a-a333-a731a71e7a66-logs\") pod \"horizon-869b6669bf-lc5s8\" (UID: \"2ea687b1-b04d-476a-a333-a731a71e7a66\") " pod="openstack/horizon-869b6669bf-lc5s8" Oct 14 09:04:07 crc kubenswrapper[4870]: I1014 09:04:07.683705 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2ea687b1-b04d-476a-a333-a731a71e7a66-scripts\") pod \"horizon-869b6669bf-lc5s8\" (UID: \"2ea687b1-b04d-476a-a333-a731a71e7a66\") " pod="openstack/horizon-869b6669bf-lc5s8" Oct 14 09:04:07 crc kubenswrapper[4870]: I1014 09:04:07.683727 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2ea687b1-b04d-476a-a333-a731a71e7a66-config-data\") pod \"horizon-869b6669bf-lc5s8\" (UID: \"2ea687b1-b04d-476a-a333-a731a71e7a66\") " pod="openstack/horizon-869b6669bf-lc5s8" Oct 14 09:04:07 crc kubenswrapper[4870]: I1014 09:04:07.685077 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2ea687b1-b04d-476a-a333-a731a71e7a66-config-data\") pod \"horizon-869b6669bf-lc5s8\" (UID: \"2ea687b1-b04d-476a-a333-a731a71e7a66\") " pod="openstack/horizon-869b6669bf-lc5s8" Oct 14 09:04:07 crc kubenswrapper[4870]: I1014 09:04:07.694712 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2ea687b1-b04d-476a-a333-a731a71e7a66-logs\") pod \"horizon-869b6669bf-lc5s8\" (UID: \"2ea687b1-b04d-476a-a333-a731a71e7a66\") " pod="openstack/horizon-869b6669bf-lc5s8" Oct 14 09:04:07 crc kubenswrapper[4870]: I1014 09:04:07.697405 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2ea687b1-b04d-476a-a333-a731a71e7a66-scripts\") pod \"horizon-869b6669bf-lc5s8\" (UID: \"2ea687b1-b04d-476a-a333-a731a71e7a66\") " pod="openstack/horizon-869b6669bf-lc5s8" Oct 14 09:04:07 crc kubenswrapper[4870]: I1014 09:04:07.709639 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rnhnl\" (UniqueName: \"kubernetes.io/projected/2ea687b1-b04d-476a-a333-a731a71e7a66-kube-api-access-rnhnl\") pod \"horizon-869b6669bf-lc5s8\" (UID: \"2ea687b1-b04d-476a-a333-a731a71e7a66\") " pod="openstack/horizon-869b6669bf-lc5s8" Oct 14 09:04:07 crc kubenswrapper[4870]: I1014 09:04:07.714093 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/2ea687b1-b04d-476a-a333-a731a71e7a66-horizon-secret-key\") pod \"horizon-869b6669bf-lc5s8\" (UID: \"2ea687b1-b04d-476a-a333-a731a71e7a66\") " pod="openstack/horizon-869b6669bf-lc5s8" Oct 14 09:04:07 crc kubenswrapper[4870]: I1014 09:04:07.785609 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/2c148c06-20aa-42c4-8214-575bbd7ce58b-horizon-secret-key\") pod \"horizon-7fbc647f5-9ph8d\" (UID: \"2c148c06-20aa-42c4-8214-575bbd7ce58b\") " pod="openstack/horizon-7fbc647f5-9ph8d" Oct 14 09:04:07 crc kubenswrapper[4870]: I1014 09:04:07.785686 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2c148c06-20aa-42c4-8214-575bbd7ce58b-scripts\") pod \"horizon-7fbc647f5-9ph8d\" (UID: \"2c148c06-20aa-42c4-8214-575bbd7ce58b\") " pod="openstack/horizon-7fbc647f5-9ph8d" Oct 14 09:04:07 crc kubenswrapper[4870]: I1014 09:04:07.785728 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2c148c06-20aa-42c4-8214-575bbd7ce58b-config-data\") pod \"horizon-7fbc647f5-9ph8d\" (UID: \"2c148c06-20aa-42c4-8214-575bbd7ce58b\") " pod="openstack/horizon-7fbc647f5-9ph8d" Oct 14 09:04:07 crc kubenswrapper[4870]: I1014 09:04:07.785828 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2c148c06-20aa-42c4-8214-575bbd7ce58b-logs\") pod \"horizon-7fbc647f5-9ph8d\" (UID: \"2c148c06-20aa-42c4-8214-575bbd7ce58b\") " pod="openstack/horizon-7fbc647f5-9ph8d" Oct 14 09:04:07 crc kubenswrapper[4870]: I1014 09:04:07.785867 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gkz7d\" (UniqueName: \"kubernetes.io/projected/2c148c06-20aa-42c4-8214-575bbd7ce58b-kube-api-access-gkz7d\") pod \"horizon-7fbc647f5-9ph8d\" (UID: \"2c148c06-20aa-42c4-8214-575bbd7ce58b\") " pod="openstack/horizon-7fbc647f5-9ph8d" Oct 14 09:04:07 crc kubenswrapper[4870]: I1014 09:04:07.800822 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-869b6669bf-lc5s8" Oct 14 09:04:07 crc kubenswrapper[4870]: I1014 09:04:07.857092 4870 generic.go:334] "Generic (PLEG): container finished" podID="829bce4c-6fe7-46fa-a4d8-1558220a039f" containerID="dcdf64ab0812d5c47611df139478fcdb35fb58c52140a0bfaf26de928cff6bc0" exitCode=143 Oct 14 09:04:07 crc kubenswrapper[4870]: I1014 09:04:07.857201 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"829bce4c-6fe7-46fa-a4d8-1558220a039f","Type":"ContainerDied","Data":"dcdf64ab0812d5c47611df139478fcdb35fb58c52140a0bfaf26de928cff6bc0"} Oct 14 09:04:07 crc kubenswrapper[4870]: I1014 09:04:07.880195 4870 generic.go:334] "Generic (PLEG): container finished" podID="de9df7d3-70ec-4abd-9cd0-55038d9ab2df" containerID="8eb5a5a2c1544b193bef195717e303075859ef64fbf449f7c5c4eb693af477f6" exitCode=143 Oct 14 09:04:07 crc kubenswrapper[4870]: I1014 09:04:07.880260 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"de9df7d3-70ec-4abd-9cd0-55038d9ab2df","Type":"ContainerDied","Data":"8eb5a5a2c1544b193bef195717e303075859ef64fbf449f7c5c4eb693af477f6"} Oct 14 09:04:07 crc kubenswrapper[4870]: I1014 09:04:07.889571 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2c148c06-20aa-42c4-8214-575bbd7ce58b-logs\") pod \"horizon-7fbc647f5-9ph8d\" (UID: \"2c148c06-20aa-42c4-8214-575bbd7ce58b\") " pod="openstack/horizon-7fbc647f5-9ph8d" Oct 14 09:04:07 crc kubenswrapper[4870]: I1014 09:04:07.889652 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gkz7d\" (UniqueName: \"kubernetes.io/projected/2c148c06-20aa-42c4-8214-575bbd7ce58b-kube-api-access-gkz7d\") pod \"horizon-7fbc647f5-9ph8d\" (UID: \"2c148c06-20aa-42c4-8214-575bbd7ce58b\") " pod="openstack/horizon-7fbc647f5-9ph8d" Oct 14 09:04:07 crc kubenswrapper[4870]: I1014 09:04:07.889722 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/2c148c06-20aa-42c4-8214-575bbd7ce58b-horizon-secret-key\") pod \"horizon-7fbc647f5-9ph8d\" (UID: \"2c148c06-20aa-42c4-8214-575bbd7ce58b\") " pod="openstack/horizon-7fbc647f5-9ph8d" Oct 14 09:04:07 crc kubenswrapper[4870]: I1014 09:04:07.889768 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2c148c06-20aa-42c4-8214-575bbd7ce58b-scripts\") pod \"horizon-7fbc647f5-9ph8d\" (UID: \"2c148c06-20aa-42c4-8214-575bbd7ce58b\") " pod="openstack/horizon-7fbc647f5-9ph8d" Oct 14 09:04:07 crc kubenswrapper[4870]: I1014 09:04:07.889817 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2c148c06-20aa-42c4-8214-575bbd7ce58b-config-data\") pod \"horizon-7fbc647f5-9ph8d\" (UID: \"2c148c06-20aa-42c4-8214-575bbd7ce58b\") " pod="openstack/horizon-7fbc647f5-9ph8d" Oct 14 09:04:07 crc kubenswrapper[4870]: I1014 09:04:07.890952 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2c148c06-20aa-42c4-8214-575bbd7ce58b-logs\") pod \"horizon-7fbc647f5-9ph8d\" (UID: \"2c148c06-20aa-42c4-8214-575bbd7ce58b\") " pod="openstack/horizon-7fbc647f5-9ph8d" Oct 14 09:04:07 crc kubenswrapper[4870]: I1014 09:04:07.892162 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2c148c06-20aa-42c4-8214-575bbd7ce58b-scripts\") pod \"horizon-7fbc647f5-9ph8d\" (UID: \"2c148c06-20aa-42c4-8214-575bbd7ce58b\") " pod="openstack/horizon-7fbc647f5-9ph8d" Oct 14 09:04:07 crc kubenswrapper[4870]: I1014 09:04:07.892826 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2c148c06-20aa-42c4-8214-575bbd7ce58b-config-data\") pod \"horizon-7fbc647f5-9ph8d\" (UID: \"2c148c06-20aa-42c4-8214-575bbd7ce58b\") " pod="openstack/horizon-7fbc647f5-9ph8d" Oct 14 09:04:07 crc kubenswrapper[4870]: I1014 09:04:07.900239 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/2c148c06-20aa-42c4-8214-575bbd7ce58b-horizon-secret-key\") pod \"horizon-7fbc647f5-9ph8d\" (UID: \"2c148c06-20aa-42c4-8214-575bbd7ce58b\") " pod="openstack/horizon-7fbc647f5-9ph8d" Oct 14 09:04:07 crc kubenswrapper[4870]: I1014 09:04:07.910261 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gkz7d\" (UniqueName: \"kubernetes.io/projected/2c148c06-20aa-42c4-8214-575bbd7ce58b-kube-api-access-gkz7d\") pod \"horizon-7fbc647f5-9ph8d\" (UID: \"2c148c06-20aa-42c4-8214-575bbd7ce58b\") " pod="openstack/horizon-7fbc647f5-9ph8d" Oct 14 09:04:07 crc kubenswrapper[4870]: I1014 09:04:07.942637 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7fbc647f5-9ph8d" Oct 14 09:04:08 crc kubenswrapper[4870]: I1014 09:04:08.126208 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7fbc647f5-9ph8d"] Oct 14 09:04:08 crc kubenswrapper[4870]: I1014 09:04:08.170868 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-7cff98d7df-2fx9m"] Oct 14 09:04:08 crc kubenswrapper[4870]: I1014 09:04:08.172647 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7cff98d7df-2fx9m" Oct 14 09:04:08 crc kubenswrapper[4870]: I1014 09:04:08.207332 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7cff98d7df-2fx9m"] Oct 14 09:04:08 crc kubenswrapper[4870]: I1014 09:04:08.300927 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/7ad77931-c302-4ce8-81f0-78767bb80443-horizon-secret-key\") pod \"horizon-7cff98d7df-2fx9m\" (UID: \"7ad77931-c302-4ce8-81f0-78767bb80443\") " pod="openstack/horizon-7cff98d7df-2fx9m" Oct 14 09:04:08 crc kubenswrapper[4870]: I1014 09:04:08.301527 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7ad77931-c302-4ce8-81f0-78767bb80443-scripts\") pod \"horizon-7cff98d7df-2fx9m\" (UID: \"7ad77931-c302-4ce8-81f0-78767bb80443\") " pod="openstack/horizon-7cff98d7df-2fx9m" Oct 14 09:04:08 crc kubenswrapper[4870]: I1014 09:04:08.301571 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l2c8j\" (UniqueName: \"kubernetes.io/projected/7ad77931-c302-4ce8-81f0-78767bb80443-kube-api-access-l2c8j\") pod \"horizon-7cff98d7df-2fx9m\" (UID: \"7ad77931-c302-4ce8-81f0-78767bb80443\") " pod="openstack/horizon-7cff98d7df-2fx9m" Oct 14 09:04:08 crc kubenswrapper[4870]: I1014 09:04:08.301617 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7ad77931-c302-4ce8-81f0-78767bb80443-config-data\") pod \"horizon-7cff98d7df-2fx9m\" (UID: \"7ad77931-c302-4ce8-81f0-78767bb80443\") " pod="openstack/horizon-7cff98d7df-2fx9m" Oct 14 09:04:08 crc kubenswrapper[4870]: I1014 09:04:08.301651 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7ad77931-c302-4ce8-81f0-78767bb80443-logs\") pod \"horizon-7cff98d7df-2fx9m\" (UID: \"7ad77931-c302-4ce8-81f0-78767bb80443\") " pod="openstack/horizon-7cff98d7df-2fx9m" Oct 14 09:04:08 crc kubenswrapper[4870]: I1014 09:04:08.361356 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-869b6669bf-lc5s8"] Oct 14 09:04:08 crc kubenswrapper[4870]: I1014 09:04:08.403093 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7ad77931-c302-4ce8-81f0-78767bb80443-config-data\") pod \"horizon-7cff98d7df-2fx9m\" (UID: \"7ad77931-c302-4ce8-81f0-78767bb80443\") " pod="openstack/horizon-7cff98d7df-2fx9m" Oct 14 09:04:08 crc kubenswrapper[4870]: I1014 09:04:08.403158 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7ad77931-c302-4ce8-81f0-78767bb80443-logs\") pod \"horizon-7cff98d7df-2fx9m\" (UID: \"7ad77931-c302-4ce8-81f0-78767bb80443\") " pod="openstack/horizon-7cff98d7df-2fx9m" Oct 14 09:04:08 crc kubenswrapper[4870]: I1014 09:04:08.403796 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/7ad77931-c302-4ce8-81f0-78767bb80443-horizon-secret-key\") pod \"horizon-7cff98d7df-2fx9m\" (UID: \"7ad77931-c302-4ce8-81f0-78767bb80443\") " pod="openstack/horizon-7cff98d7df-2fx9m" Oct 14 09:04:08 crc kubenswrapper[4870]: I1014 09:04:08.403857 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7ad77931-c302-4ce8-81f0-78767bb80443-scripts\") pod \"horizon-7cff98d7df-2fx9m\" (UID: \"7ad77931-c302-4ce8-81f0-78767bb80443\") " pod="openstack/horizon-7cff98d7df-2fx9m" Oct 14 09:04:08 crc kubenswrapper[4870]: I1014 09:04:08.403888 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l2c8j\" (UniqueName: \"kubernetes.io/projected/7ad77931-c302-4ce8-81f0-78767bb80443-kube-api-access-l2c8j\") pod \"horizon-7cff98d7df-2fx9m\" (UID: \"7ad77931-c302-4ce8-81f0-78767bb80443\") " pod="openstack/horizon-7cff98d7df-2fx9m" Oct 14 09:04:08 crc kubenswrapper[4870]: I1014 09:04:08.404451 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7ad77931-c302-4ce8-81f0-78767bb80443-logs\") pod \"horizon-7cff98d7df-2fx9m\" (UID: \"7ad77931-c302-4ce8-81f0-78767bb80443\") " pod="openstack/horizon-7cff98d7df-2fx9m" Oct 14 09:04:08 crc kubenswrapper[4870]: I1014 09:04:08.405002 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7ad77931-c302-4ce8-81f0-78767bb80443-scripts\") pod \"horizon-7cff98d7df-2fx9m\" (UID: \"7ad77931-c302-4ce8-81f0-78767bb80443\") " pod="openstack/horizon-7cff98d7df-2fx9m" Oct 14 09:04:08 crc kubenswrapper[4870]: I1014 09:04:08.406429 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7ad77931-c302-4ce8-81f0-78767bb80443-config-data\") pod \"horizon-7cff98d7df-2fx9m\" (UID: \"7ad77931-c302-4ce8-81f0-78767bb80443\") " pod="openstack/horizon-7cff98d7df-2fx9m" Oct 14 09:04:08 crc kubenswrapper[4870]: I1014 09:04:08.408331 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/7ad77931-c302-4ce8-81f0-78767bb80443-horizon-secret-key\") pod \"horizon-7cff98d7df-2fx9m\" (UID: \"7ad77931-c302-4ce8-81f0-78767bb80443\") " pod="openstack/horizon-7cff98d7df-2fx9m" Oct 14 09:04:08 crc kubenswrapper[4870]: I1014 09:04:08.423678 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l2c8j\" (UniqueName: \"kubernetes.io/projected/7ad77931-c302-4ce8-81f0-78767bb80443-kube-api-access-l2c8j\") pod \"horizon-7cff98d7df-2fx9m\" (UID: \"7ad77931-c302-4ce8-81f0-78767bb80443\") " pod="openstack/horizon-7cff98d7df-2fx9m" Oct 14 09:04:08 crc kubenswrapper[4870]: I1014 09:04:08.497707 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7cff98d7df-2fx9m" Oct 14 09:04:08 crc kubenswrapper[4870]: I1014 09:04:08.545194 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7fbc647f5-9ph8d"] Oct 14 09:04:08 crc kubenswrapper[4870]: I1014 09:04:08.891896 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-869b6669bf-lc5s8" event={"ID":"2ea687b1-b04d-476a-a333-a731a71e7a66","Type":"ContainerStarted","Data":"52c3a13da1b0e7af4b4cd912cf7b283d670168315a1ecdd839c1f94c4c0b74ce"} Oct 14 09:04:08 crc kubenswrapper[4870]: I1014 09:04:08.893191 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7fbc647f5-9ph8d" event={"ID":"2c148c06-20aa-42c4-8214-575bbd7ce58b","Type":"ContainerStarted","Data":"1fa60abd1a023611ea23394b6a0c85a280a323a5f1dcafede77d19200ab455a8"} Oct 14 09:04:08 crc kubenswrapper[4870]: I1014 09:04:08.988911 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7cff98d7df-2fx9m"] Oct 14 09:04:09 crc kubenswrapper[4870]: I1014 09:04:09.903600 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7cff98d7df-2fx9m" event={"ID":"7ad77931-c302-4ce8-81f0-78767bb80443","Type":"ContainerStarted","Data":"75811ed35ae717f13ca70040f547a8131fa29f5d02af80e4da49519a86594f06"} Oct 14 09:04:11 crc kubenswrapper[4870]: I1014 09:04:11.523457 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 14 09:04:11 crc kubenswrapper[4870]: I1014 09:04:11.671924 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/de9df7d3-70ec-4abd-9cd0-55038d9ab2df-httpd-run\") pod \"de9df7d3-70ec-4abd-9cd0-55038d9ab2df\" (UID: \"de9df7d3-70ec-4abd-9cd0-55038d9ab2df\") " Oct 14 09:04:11 crc kubenswrapper[4870]: I1014 09:04:11.671969 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/de9df7d3-70ec-4abd-9cd0-55038d9ab2df-scripts\") pod \"de9df7d3-70ec-4abd-9cd0-55038d9ab2df\" (UID: \"de9df7d3-70ec-4abd-9cd0-55038d9ab2df\") " Oct 14 09:04:11 crc kubenswrapper[4870]: I1014 09:04:11.672044 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de9df7d3-70ec-4abd-9cd0-55038d9ab2df-combined-ca-bundle\") pod \"de9df7d3-70ec-4abd-9cd0-55038d9ab2df\" (UID: \"de9df7d3-70ec-4abd-9cd0-55038d9ab2df\") " Oct 14 09:04:11 crc kubenswrapper[4870]: I1014 09:04:11.672062 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/de9df7d3-70ec-4abd-9cd0-55038d9ab2df-logs\") pod \"de9df7d3-70ec-4abd-9cd0-55038d9ab2df\" (UID: \"de9df7d3-70ec-4abd-9cd0-55038d9ab2df\") " Oct 14 09:04:11 crc kubenswrapper[4870]: I1014 09:04:11.672086 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/de9df7d3-70ec-4abd-9cd0-55038d9ab2df-config-data\") pod \"de9df7d3-70ec-4abd-9cd0-55038d9ab2df\" (UID: \"de9df7d3-70ec-4abd-9cd0-55038d9ab2df\") " Oct 14 09:04:11 crc kubenswrapper[4870]: I1014 09:04:11.672112 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wpsvd\" (UniqueName: \"kubernetes.io/projected/de9df7d3-70ec-4abd-9cd0-55038d9ab2df-kube-api-access-wpsvd\") pod \"de9df7d3-70ec-4abd-9cd0-55038d9ab2df\" (UID: \"de9df7d3-70ec-4abd-9cd0-55038d9ab2df\") " Oct 14 09:04:11 crc kubenswrapper[4870]: I1014 09:04:11.672148 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/de9df7d3-70ec-4abd-9cd0-55038d9ab2df-ceph\") pod \"de9df7d3-70ec-4abd-9cd0-55038d9ab2df\" (UID: \"de9df7d3-70ec-4abd-9cd0-55038d9ab2df\") " Oct 14 09:04:11 crc kubenswrapper[4870]: I1014 09:04:11.672555 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/de9df7d3-70ec-4abd-9cd0-55038d9ab2df-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "de9df7d3-70ec-4abd-9cd0-55038d9ab2df" (UID: "de9df7d3-70ec-4abd-9cd0-55038d9ab2df"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 09:04:11 crc kubenswrapper[4870]: I1014 09:04:11.672592 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/de9df7d3-70ec-4abd-9cd0-55038d9ab2df-logs" (OuterVolumeSpecName: "logs") pod "de9df7d3-70ec-4abd-9cd0-55038d9ab2df" (UID: "de9df7d3-70ec-4abd-9cd0-55038d9ab2df"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 09:04:11 crc kubenswrapper[4870]: I1014 09:04:11.673411 4870 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/de9df7d3-70ec-4abd-9cd0-55038d9ab2df-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 14 09:04:11 crc kubenswrapper[4870]: I1014 09:04:11.673497 4870 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/de9df7d3-70ec-4abd-9cd0-55038d9ab2df-logs\") on node \"crc\" DevicePath \"\"" Oct 14 09:04:11 crc kubenswrapper[4870]: I1014 09:04:11.683255 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/de9df7d3-70ec-4abd-9cd0-55038d9ab2df-kube-api-access-wpsvd" (OuterVolumeSpecName: "kube-api-access-wpsvd") pod "de9df7d3-70ec-4abd-9cd0-55038d9ab2df" (UID: "de9df7d3-70ec-4abd-9cd0-55038d9ab2df"). InnerVolumeSpecName "kube-api-access-wpsvd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:04:11 crc kubenswrapper[4870]: I1014 09:04:11.685488 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de9df7d3-70ec-4abd-9cd0-55038d9ab2df-scripts" (OuterVolumeSpecName: "scripts") pod "de9df7d3-70ec-4abd-9cd0-55038d9ab2df" (UID: "de9df7d3-70ec-4abd-9cd0-55038d9ab2df"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:04:11 crc kubenswrapper[4870]: I1014 09:04:11.695070 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/de9df7d3-70ec-4abd-9cd0-55038d9ab2df-ceph" (OuterVolumeSpecName: "ceph") pod "de9df7d3-70ec-4abd-9cd0-55038d9ab2df" (UID: "de9df7d3-70ec-4abd-9cd0-55038d9ab2df"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:04:11 crc kubenswrapper[4870]: I1014 09:04:11.696313 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 14 09:04:11 crc kubenswrapper[4870]: I1014 09:04:11.706105 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-lnlhg"] Oct 14 09:04:11 crc kubenswrapper[4870]: E1014 09:04:11.706730 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de9df7d3-70ec-4abd-9cd0-55038d9ab2df" containerName="glance-log" Oct 14 09:04:11 crc kubenswrapper[4870]: I1014 09:04:11.706752 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="de9df7d3-70ec-4abd-9cd0-55038d9ab2df" containerName="glance-log" Oct 14 09:04:11 crc kubenswrapper[4870]: E1014 09:04:11.706766 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de9df7d3-70ec-4abd-9cd0-55038d9ab2df" containerName="glance-httpd" Oct 14 09:04:11 crc kubenswrapper[4870]: I1014 09:04:11.706775 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="de9df7d3-70ec-4abd-9cd0-55038d9ab2df" containerName="glance-httpd" Oct 14 09:04:11 crc kubenswrapper[4870]: E1014 09:04:11.706797 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="829bce4c-6fe7-46fa-a4d8-1558220a039f" containerName="glance-httpd" Oct 14 09:04:11 crc kubenswrapper[4870]: I1014 09:04:11.706805 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="829bce4c-6fe7-46fa-a4d8-1558220a039f" containerName="glance-httpd" Oct 14 09:04:11 crc kubenswrapper[4870]: E1014 09:04:11.706847 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="829bce4c-6fe7-46fa-a4d8-1558220a039f" containerName="glance-log" Oct 14 09:04:11 crc kubenswrapper[4870]: I1014 09:04:11.706854 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="829bce4c-6fe7-46fa-a4d8-1558220a039f" containerName="glance-log" Oct 14 09:04:11 crc kubenswrapper[4870]: I1014 09:04:11.707077 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="829bce4c-6fe7-46fa-a4d8-1558220a039f" containerName="glance-httpd" Oct 14 09:04:11 crc kubenswrapper[4870]: I1014 09:04:11.707107 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="829bce4c-6fe7-46fa-a4d8-1558220a039f" containerName="glance-log" Oct 14 09:04:11 crc kubenswrapper[4870]: I1014 09:04:11.707119 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="de9df7d3-70ec-4abd-9cd0-55038d9ab2df" containerName="glance-log" Oct 14 09:04:11 crc kubenswrapper[4870]: I1014 09:04:11.707134 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="de9df7d3-70ec-4abd-9cd0-55038d9ab2df" containerName="glance-httpd" Oct 14 09:04:11 crc kubenswrapper[4870]: I1014 09:04:11.709702 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lnlhg" Oct 14 09:04:11 crc kubenswrapper[4870]: I1014 09:04:11.718417 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-lnlhg"] Oct 14 09:04:11 crc kubenswrapper[4870]: I1014 09:04:11.739055 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de9df7d3-70ec-4abd-9cd0-55038d9ab2df-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "de9df7d3-70ec-4abd-9cd0-55038d9ab2df" (UID: "de9df7d3-70ec-4abd-9cd0-55038d9ab2df"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:04:11 crc kubenswrapper[4870]: I1014 09:04:11.775260 4870 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de9df7d3-70ec-4abd-9cd0-55038d9ab2df-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 09:04:11 crc kubenswrapper[4870]: I1014 09:04:11.775290 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wpsvd\" (UniqueName: \"kubernetes.io/projected/de9df7d3-70ec-4abd-9cd0-55038d9ab2df-kube-api-access-wpsvd\") on node \"crc\" DevicePath \"\"" Oct 14 09:04:11 crc kubenswrapper[4870]: I1014 09:04:11.775305 4870 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/de9df7d3-70ec-4abd-9cd0-55038d9ab2df-ceph\") on node \"crc\" DevicePath \"\"" Oct 14 09:04:11 crc kubenswrapper[4870]: I1014 09:04:11.775333 4870 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/de9df7d3-70ec-4abd-9cd0-55038d9ab2df-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 09:04:11 crc kubenswrapper[4870]: I1014 09:04:11.794510 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de9df7d3-70ec-4abd-9cd0-55038d9ab2df-config-data" (OuterVolumeSpecName: "config-data") pod "de9df7d3-70ec-4abd-9cd0-55038d9ab2df" (UID: "de9df7d3-70ec-4abd-9cd0-55038d9ab2df"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:04:11 crc kubenswrapper[4870]: I1014 09:04:11.876314 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/829bce4c-6fe7-46fa-a4d8-1558220a039f-config-data\") pod \"829bce4c-6fe7-46fa-a4d8-1558220a039f\" (UID: \"829bce4c-6fe7-46fa-a4d8-1558220a039f\") " Oct 14 09:04:11 crc kubenswrapper[4870]: I1014 09:04:11.876531 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/829bce4c-6fe7-46fa-a4d8-1558220a039f-httpd-run\") pod \"829bce4c-6fe7-46fa-a4d8-1558220a039f\" (UID: \"829bce4c-6fe7-46fa-a4d8-1558220a039f\") " Oct 14 09:04:11 crc kubenswrapper[4870]: I1014 09:04:11.876558 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/829bce4c-6fe7-46fa-a4d8-1558220a039f-ceph\") pod \"829bce4c-6fe7-46fa-a4d8-1558220a039f\" (UID: \"829bce4c-6fe7-46fa-a4d8-1558220a039f\") " Oct 14 09:04:11 crc kubenswrapper[4870]: I1014 09:04:11.876587 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/829bce4c-6fe7-46fa-a4d8-1558220a039f-combined-ca-bundle\") pod \"829bce4c-6fe7-46fa-a4d8-1558220a039f\" (UID: \"829bce4c-6fe7-46fa-a4d8-1558220a039f\") " Oct 14 09:04:11 crc kubenswrapper[4870]: I1014 09:04:11.876664 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gqdcb\" (UniqueName: \"kubernetes.io/projected/829bce4c-6fe7-46fa-a4d8-1558220a039f-kube-api-access-gqdcb\") pod \"829bce4c-6fe7-46fa-a4d8-1558220a039f\" (UID: \"829bce4c-6fe7-46fa-a4d8-1558220a039f\") " Oct 14 09:04:11 crc kubenswrapper[4870]: I1014 09:04:11.876702 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/829bce4c-6fe7-46fa-a4d8-1558220a039f-scripts\") pod \"829bce4c-6fe7-46fa-a4d8-1558220a039f\" (UID: \"829bce4c-6fe7-46fa-a4d8-1558220a039f\") " Oct 14 09:04:11 crc kubenswrapper[4870]: I1014 09:04:11.876737 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/829bce4c-6fe7-46fa-a4d8-1558220a039f-logs\") pod \"829bce4c-6fe7-46fa-a4d8-1558220a039f\" (UID: \"829bce4c-6fe7-46fa-a4d8-1558220a039f\") " Oct 14 09:04:11 crc kubenswrapper[4870]: I1014 09:04:11.877173 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kw8lf\" (UniqueName: \"kubernetes.io/projected/e71ba358-f793-409b-ad26-ba1be98ad917-kube-api-access-kw8lf\") pod \"certified-operators-lnlhg\" (UID: \"e71ba358-f793-409b-ad26-ba1be98ad917\") " pod="openshift-marketplace/certified-operators-lnlhg" Oct 14 09:04:11 crc kubenswrapper[4870]: I1014 09:04:11.877607 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/829bce4c-6fe7-46fa-a4d8-1558220a039f-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "829bce4c-6fe7-46fa-a4d8-1558220a039f" (UID: "829bce4c-6fe7-46fa-a4d8-1558220a039f"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 09:04:11 crc kubenswrapper[4870]: I1014 09:04:11.877647 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e71ba358-f793-409b-ad26-ba1be98ad917-utilities\") pod \"certified-operators-lnlhg\" (UID: \"e71ba358-f793-409b-ad26-ba1be98ad917\") " pod="openshift-marketplace/certified-operators-lnlhg" Oct 14 09:04:11 crc kubenswrapper[4870]: I1014 09:04:11.877873 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e71ba358-f793-409b-ad26-ba1be98ad917-catalog-content\") pod \"certified-operators-lnlhg\" (UID: \"e71ba358-f793-409b-ad26-ba1be98ad917\") " pod="openshift-marketplace/certified-operators-lnlhg" Oct 14 09:04:11 crc kubenswrapper[4870]: I1014 09:04:11.877977 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/829bce4c-6fe7-46fa-a4d8-1558220a039f-logs" (OuterVolumeSpecName: "logs") pod "829bce4c-6fe7-46fa-a4d8-1558220a039f" (UID: "829bce4c-6fe7-46fa-a4d8-1558220a039f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 09:04:11 crc kubenswrapper[4870]: I1014 09:04:11.878067 4870 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/829bce4c-6fe7-46fa-a4d8-1558220a039f-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 14 09:04:11 crc kubenswrapper[4870]: I1014 09:04:11.878097 4870 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/de9df7d3-70ec-4abd-9cd0-55038d9ab2df-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 09:04:11 crc kubenswrapper[4870]: I1014 09:04:11.880370 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/829bce4c-6fe7-46fa-a4d8-1558220a039f-scripts" (OuterVolumeSpecName: "scripts") pod "829bce4c-6fe7-46fa-a4d8-1558220a039f" (UID: "829bce4c-6fe7-46fa-a4d8-1558220a039f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:04:11 crc kubenswrapper[4870]: I1014 09:04:11.880544 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/829bce4c-6fe7-46fa-a4d8-1558220a039f-ceph" (OuterVolumeSpecName: "ceph") pod "829bce4c-6fe7-46fa-a4d8-1558220a039f" (UID: "829bce4c-6fe7-46fa-a4d8-1558220a039f"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:04:11 crc kubenswrapper[4870]: I1014 09:04:11.881014 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/829bce4c-6fe7-46fa-a4d8-1558220a039f-kube-api-access-gqdcb" (OuterVolumeSpecName: "kube-api-access-gqdcb") pod "829bce4c-6fe7-46fa-a4d8-1558220a039f" (UID: "829bce4c-6fe7-46fa-a4d8-1558220a039f"). InnerVolumeSpecName "kube-api-access-gqdcb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:04:11 crc kubenswrapper[4870]: I1014 09:04:11.900741 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/829bce4c-6fe7-46fa-a4d8-1558220a039f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "829bce4c-6fe7-46fa-a4d8-1558220a039f" (UID: "829bce4c-6fe7-46fa-a4d8-1558220a039f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:04:11 crc kubenswrapper[4870]: I1014 09:04:11.935117 4870 generic.go:334] "Generic (PLEG): container finished" podID="de9df7d3-70ec-4abd-9cd0-55038d9ab2df" containerID="d5be18839f4b371ce0db12541e6bfc5e5ccd1903925a9840da4265c4d2692924" exitCode=0 Oct 14 09:04:11 crc kubenswrapper[4870]: I1014 09:04:11.935176 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 14 09:04:11 crc kubenswrapper[4870]: I1014 09:04:11.935199 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"de9df7d3-70ec-4abd-9cd0-55038d9ab2df","Type":"ContainerDied","Data":"d5be18839f4b371ce0db12541e6bfc5e5ccd1903925a9840da4265c4d2692924"} Oct 14 09:04:11 crc kubenswrapper[4870]: I1014 09:04:11.935230 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"de9df7d3-70ec-4abd-9cd0-55038d9ab2df","Type":"ContainerDied","Data":"9a39ca8808b54188910904d6b5204d25404c01fd88ea125f6c7620b7becdc941"} Oct 14 09:04:11 crc kubenswrapper[4870]: I1014 09:04:11.935251 4870 scope.go:117] "RemoveContainer" containerID="d5be18839f4b371ce0db12541e6bfc5e5ccd1903925a9840da4265c4d2692924" Oct 14 09:04:11 crc kubenswrapper[4870]: I1014 09:04:11.939183 4870 generic.go:334] "Generic (PLEG): container finished" podID="829bce4c-6fe7-46fa-a4d8-1558220a039f" containerID="6d39e737807423f661bed2bc5c4c82a9315751ab881b785991a4633e1b26b216" exitCode=0 Oct 14 09:04:11 crc kubenswrapper[4870]: I1014 09:04:11.939227 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"829bce4c-6fe7-46fa-a4d8-1558220a039f","Type":"ContainerDied","Data":"6d39e737807423f661bed2bc5c4c82a9315751ab881b785991a4633e1b26b216"} Oct 14 09:04:11 crc kubenswrapper[4870]: I1014 09:04:11.939254 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"829bce4c-6fe7-46fa-a4d8-1558220a039f","Type":"ContainerDied","Data":"853d87787afccaeff5d595dbd4610fa6450723bc1f68c6e77468ead695db7bd7"} Oct 14 09:04:11 crc kubenswrapper[4870]: I1014 09:04:11.939311 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 14 09:04:11 crc kubenswrapper[4870]: I1014 09:04:11.944979 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/829bce4c-6fe7-46fa-a4d8-1558220a039f-config-data" (OuterVolumeSpecName: "config-data") pod "829bce4c-6fe7-46fa-a4d8-1558220a039f" (UID: "829bce4c-6fe7-46fa-a4d8-1558220a039f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:04:11 crc kubenswrapper[4870]: I1014 09:04:11.982393 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e71ba358-f793-409b-ad26-ba1be98ad917-utilities\") pod \"certified-operators-lnlhg\" (UID: \"e71ba358-f793-409b-ad26-ba1be98ad917\") " pod="openshift-marketplace/certified-operators-lnlhg" Oct 14 09:04:11 crc kubenswrapper[4870]: I1014 09:04:11.982477 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e71ba358-f793-409b-ad26-ba1be98ad917-catalog-content\") pod \"certified-operators-lnlhg\" (UID: \"e71ba358-f793-409b-ad26-ba1be98ad917\") " pod="openshift-marketplace/certified-operators-lnlhg" Oct 14 09:04:11 crc kubenswrapper[4870]: I1014 09:04:11.982532 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kw8lf\" (UniqueName: \"kubernetes.io/projected/e71ba358-f793-409b-ad26-ba1be98ad917-kube-api-access-kw8lf\") pod \"certified-operators-lnlhg\" (UID: \"e71ba358-f793-409b-ad26-ba1be98ad917\") " pod="openshift-marketplace/certified-operators-lnlhg" Oct 14 09:04:11 crc kubenswrapper[4870]: I1014 09:04:11.982615 4870 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/829bce4c-6fe7-46fa-a4d8-1558220a039f-ceph\") on node \"crc\" DevicePath \"\"" Oct 14 09:04:11 crc kubenswrapper[4870]: I1014 09:04:11.982630 4870 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/829bce4c-6fe7-46fa-a4d8-1558220a039f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 09:04:11 crc kubenswrapper[4870]: I1014 09:04:11.982644 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gqdcb\" (UniqueName: \"kubernetes.io/projected/829bce4c-6fe7-46fa-a4d8-1558220a039f-kube-api-access-gqdcb\") on node \"crc\" DevicePath \"\"" Oct 14 09:04:11 crc kubenswrapper[4870]: I1014 09:04:11.982655 4870 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/829bce4c-6fe7-46fa-a4d8-1558220a039f-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 09:04:11 crc kubenswrapper[4870]: I1014 09:04:11.982665 4870 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/829bce4c-6fe7-46fa-a4d8-1558220a039f-logs\") on node \"crc\" DevicePath \"\"" Oct 14 09:04:11 crc kubenswrapper[4870]: I1014 09:04:11.982677 4870 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/829bce4c-6fe7-46fa-a4d8-1558220a039f-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 09:04:11 crc kubenswrapper[4870]: I1014 09:04:11.983703 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e71ba358-f793-409b-ad26-ba1be98ad917-catalog-content\") pod \"certified-operators-lnlhg\" (UID: \"e71ba358-f793-409b-ad26-ba1be98ad917\") " pod="openshift-marketplace/certified-operators-lnlhg" Oct 14 09:04:11 crc kubenswrapper[4870]: I1014 09:04:11.983782 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e71ba358-f793-409b-ad26-ba1be98ad917-utilities\") pod \"certified-operators-lnlhg\" (UID: \"e71ba358-f793-409b-ad26-ba1be98ad917\") " pod="openshift-marketplace/certified-operators-lnlhg" Oct 14 09:04:12 crc kubenswrapper[4870]: I1014 09:04:12.003556 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 14 09:04:12 crc kubenswrapper[4870]: I1014 09:04:12.006891 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kw8lf\" (UniqueName: \"kubernetes.io/projected/e71ba358-f793-409b-ad26-ba1be98ad917-kube-api-access-kw8lf\") pod \"certified-operators-lnlhg\" (UID: \"e71ba358-f793-409b-ad26-ba1be98ad917\") " pod="openshift-marketplace/certified-operators-lnlhg" Oct 14 09:04:12 crc kubenswrapper[4870]: I1014 09:04:12.016292 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 14 09:04:12 crc kubenswrapper[4870]: I1014 09:04:12.044032 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 14 09:04:12 crc kubenswrapper[4870]: I1014 09:04:12.046471 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 14 09:04:12 crc kubenswrapper[4870]: I1014 09:04:12.059506 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lnlhg" Oct 14 09:04:12 crc kubenswrapper[4870]: I1014 09:04:12.060399 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 14 09:04:12 crc kubenswrapper[4870]: I1014 09:04:12.062571 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 14 09:04:12 crc kubenswrapper[4870]: I1014 09:04:12.187333 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/bdd40f40-57d1-4dbf-8b03-f447844dbcd3-ceph\") pod \"glance-default-internal-api-0\" (UID: \"bdd40f40-57d1-4dbf-8b03-f447844dbcd3\") " pod="openstack/glance-default-internal-api-0" Oct 14 09:04:12 crc kubenswrapper[4870]: I1014 09:04:12.187433 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/bdd40f40-57d1-4dbf-8b03-f447844dbcd3-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"bdd40f40-57d1-4dbf-8b03-f447844dbcd3\") " pod="openstack/glance-default-internal-api-0" Oct 14 09:04:12 crc kubenswrapper[4870]: I1014 09:04:12.187519 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bdd40f40-57d1-4dbf-8b03-f447844dbcd3-logs\") pod \"glance-default-internal-api-0\" (UID: \"bdd40f40-57d1-4dbf-8b03-f447844dbcd3\") " pod="openstack/glance-default-internal-api-0" Oct 14 09:04:12 crc kubenswrapper[4870]: I1014 09:04:12.188625 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bdd40f40-57d1-4dbf-8b03-f447844dbcd3-scripts\") pod \"glance-default-internal-api-0\" (UID: \"bdd40f40-57d1-4dbf-8b03-f447844dbcd3\") " pod="openstack/glance-default-internal-api-0" Oct 14 09:04:12 crc kubenswrapper[4870]: I1014 09:04:12.188660 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bdd40f40-57d1-4dbf-8b03-f447844dbcd3-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"bdd40f40-57d1-4dbf-8b03-f447844dbcd3\") " pod="openstack/glance-default-internal-api-0" Oct 14 09:04:12 crc kubenswrapper[4870]: I1014 09:04:12.188682 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bdd40f40-57d1-4dbf-8b03-f447844dbcd3-config-data\") pod \"glance-default-internal-api-0\" (UID: \"bdd40f40-57d1-4dbf-8b03-f447844dbcd3\") " pod="openstack/glance-default-internal-api-0" Oct 14 09:04:12 crc kubenswrapper[4870]: I1014 09:04:12.188794 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vql97\" (UniqueName: \"kubernetes.io/projected/bdd40f40-57d1-4dbf-8b03-f447844dbcd3-kube-api-access-vql97\") pod \"glance-default-internal-api-0\" (UID: \"bdd40f40-57d1-4dbf-8b03-f447844dbcd3\") " pod="openstack/glance-default-internal-api-0" Oct 14 09:04:12 crc kubenswrapper[4870]: I1014 09:04:12.281567 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 14 09:04:12 crc kubenswrapper[4870]: I1014 09:04:12.292469 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/bdd40f40-57d1-4dbf-8b03-f447844dbcd3-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"bdd40f40-57d1-4dbf-8b03-f447844dbcd3\") " pod="openstack/glance-default-internal-api-0" Oct 14 09:04:12 crc kubenswrapper[4870]: I1014 09:04:12.292750 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bdd40f40-57d1-4dbf-8b03-f447844dbcd3-logs\") pod \"glance-default-internal-api-0\" (UID: \"bdd40f40-57d1-4dbf-8b03-f447844dbcd3\") " pod="openstack/glance-default-internal-api-0" Oct 14 09:04:12 crc kubenswrapper[4870]: I1014 09:04:12.292885 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bdd40f40-57d1-4dbf-8b03-f447844dbcd3-scripts\") pod \"glance-default-internal-api-0\" (UID: \"bdd40f40-57d1-4dbf-8b03-f447844dbcd3\") " pod="openstack/glance-default-internal-api-0" Oct 14 09:04:12 crc kubenswrapper[4870]: I1014 09:04:12.292975 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bdd40f40-57d1-4dbf-8b03-f447844dbcd3-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"bdd40f40-57d1-4dbf-8b03-f447844dbcd3\") " pod="openstack/glance-default-internal-api-0" Oct 14 09:04:12 crc kubenswrapper[4870]: I1014 09:04:12.293043 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bdd40f40-57d1-4dbf-8b03-f447844dbcd3-config-data\") pod \"glance-default-internal-api-0\" (UID: \"bdd40f40-57d1-4dbf-8b03-f447844dbcd3\") " pod="openstack/glance-default-internal-api-0" Oct 14 09:04:12 crc kubenswrapper[4870]: I1014 09:04:12.293135 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vql97\" (UniqueName: \"kubernetes.io/projected/bdd40f40-57d1-4dbf-8b03-f447844dbcd3-kube-api-access-vql97\") pod \"glance-default-internal-api-0\" (UID: \"bdd40f40-57d1-4dbf-8b03-f447844dbcd3\") " pod="openstack/glance-default-internal-api-0" Oct 14 09:04:12 crc kubenswrapper[4870]: I1014 09:04:12.293246 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/bdd40f40-57d1-4dbf-8b03-f447844dbcd3-ceph\") pod \"glance-default-internal-api-0\" (UID: \"bdd40f40-57d1-4dbf-8b03-f447844dbcd3\") " pod="openstack/glance-default-internal-api-0" Oct 14 09:04:12 crc kubenswrapper[4870]: I1014 09:04:12.303618 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 14 09:04:12 crc kubenswrapper[4870]: I1014 09:04:12.306713 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/bdd40f40-57d1-4dbf-8b03-f447844dbcd3-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"bdd40f40-57d1-4dbf-8b03-f447844dbcd3\") " pod="openstack/glance-default-internal-api-0" Oct 14 09:04:12 crc kubenswrapper[4870]: I1014 09:04:12.315035 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 14 09:04:12 crc kubenswrapper[4870]: I1014 09:04:12.316013 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bdd40f40-57d1-4dbf-8b03-f447844dbcd3-logs\") pod \"glance-default-internal-api-0\" (UID: \"bdd40f40-57d1-4dbf-8b03-f447844dbcd3\") " pod="openstack/glance-default-internal-api-0" Oct 14 09:04:12 crc kubenswrapper[4870]: I1014 09:04:12.317051 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 14 09:04:12 crc kubenswrapper[4870]: I1014 09:04:12.320412 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vql97\" (UniqueName: \"kubernetes.io/projected/bdd40f40-57d1-4dbf-8b03-f447844dbcd3-kube-api-access-vql97\") pod \"glance-default-internal-api-0\" (UID: \"bdd40f40-57d1-4dbf-8b03-f447844dbcd3\") " pod="openstack/glance-default-internal-api-0" Oct 14 09:04:12 crc kubenswrapper[4870]: I1014 09:04:12.321750 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bdd40f40-57d1-4dbf-8b03-f447844dbcd3-config-data\") pod \"glance-default-internal-api-0\" (UID: \"bdd40f40-57d1-4dbf-8b03-f447844dbcd3\") " pod="openstack/glance-default-internal-api-0" Oct 14 09:04:12 crc kubenswrapper[4870]: I1014 09:04:12.322095 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/bdd40f40-57d1-4dbf-8b03-f447844dbcd3-ceph\") pod \"glance-default-internal-api-0\" (UID: \"bdd40f40-57d1-4dbf-8b03-f447844dbcd3\") " pod="openstack/glance-default-internal-api-0" Oct 14 09:04:12 crc kubenswrapper[4870]: I1014 09:04:12.322428 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 14 09:04:12 crc kubenswrapper[4870]: I1014 09:04:12.327293 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bdd40f40-57d1-4dbf-8b03-f447844dbcd3-scripts\") pod \"glance-default-internal-api-0\" (UID: \"bdd40f40-57d1-4dbf-8b03-f447844dbcd3\") " pod="openstack/glance-default-internal-api-0" Oct 14 09:04:12 crc kubenswrapper[4870]: I1014 09:04:12.327409 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bdd40f40-57d1-4dbf-8b03-f447844dbcd3-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"bdd40f40-57d1-4dbf-8b03-f447844dbcd3\") " pod="openstack/glance-default-internal-api-0" Oct 14 09:04:12 crc kubenswrapper[4870]: I1014 09:04:12.370987 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 14 09:04:12 crc kubenswrapper[4870]: I1014 09:04:12.389612 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 14 09:04:12 crc kubenswrapper[4870]: I1014 09:04:12.496709 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/95cfd5a9-9dfc-44a4-bb94-0d60963dffd7-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"95cfd5a9-9dfc-44a4-bb94-0d60963dffd7\") " pod="openstack/glance-default-external-api-0" Oct 14 09:04:12 crc kubenswrapper[4870]: I1014 09:04:12.496811 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/95cfd5a9-9dfc-44a4-bb94-0d60963dffd7-logs\") pod \"glance-default-external-api-0\" (UID: \"95cfd5a9-9dfc-44a4-bb94-0d60963dffd7\") " pod="openstack/glance-default-external-api-0" Oct 14 09:04:12 crc kubenswrapper[4870]: I1014 09:04:12.496970 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/95cfd5a9-9dfc-44a4-bb94-0d60963dffd7-scripts\") pod \"glance-default-external-api-0\" (UID: \"95cfd5a9-9dfc-44a4-bb94-0d60963dffd7\") " pod="openstack/glance-default-external-api-0" Oct 14 09:04:12 crc kubenswrapper[4870]: I1014 09:04:12.497012 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/95cfd5a9-9dfc-44a4-bb94-0d60963dffd7-ceph\") pod \"glance-default-external-api-0\" (UID: \"95cfd5a9-9dfc-44a4-bb94-0d60963dffd7\") " pod="openstack/glance-default-external-api-0" Oct 14 09:04:12 crc kubenswrapper[4870]: I1014 09:04:12.497131 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7bp58\" (UniqueName: \"kubernetes.io/projected/95cfd5a9-9dfc-44a4-bb94-0d60963dffd7-kube-api-access-7bp58\") pod \"glance-default-external-api-0\" (UID: \"95cfd5a9-9dfc-44a4-bb94-0d60963dffd7\") " pod="openstack/glance-default-external-api-0" Oct 14 09:04:12 crc kubenswrapper[4870]: I1014 09:04:12.497284 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/95cfd5a9-9dfc-44a4-bb94-0d60963dffd7-config-data\") pod \"glance-default-external-api-0\" (UID: \"95cfd5a9-9dfc-44a4-bb94-0d60963dffd7\") " pod="openstack/glance-default-external-api-0" Oct 14 09:04:12 crc kubenswrapper[4870]: I1014 09:04:12.497317 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95cfd5a9-9dfc-44a4-bb94-0d60963dffd7-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"95cfd5a9-9dfc-44a4-bb94-0d60963dffd7\") " pod="openstack/glance-default-external-api-0" Oct 14 09:04:12 crc kubenswrapper[4870]: I1014 09:04:12.599041 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/95cfd5a9-9dfc-44a4-bb94-0d60963dffd7-scripts\") pod \"glance-default-external-api-0\" (UID: \"95cfd5a9-9dfc-44a4-bb94-0d60963dffd7\") " pod="openstack/glance-default-external-api-0" Oct 14 09:04:12 crc kubenswrapper[4870]: I1014 09:04:12.599111 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/95cfd5a9-9dfc-44a4-bb94-0d60963dffd7-ceph\") pod \"glance-default-external-api-0\" (UID: \"95cfd5a9-9dfc-44a4-bb94-0d60963dffd7\") " pod="openstack/glance-default-external-api-0" Oct 14 09:04:12 crc kubenswrapper[4870]: I1014 09:04:12.599143 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7bp58\" (UniqueName: \"kubernetes.io/projected/95cfd5a9-9dfc-44a4-bb94-0d60963dffd7-kube-api-access-7bp58\") pod \"glance-default-external-api-0\" (UID: \"95cfd5a9-9dfc-44a4-bb94-0d60963dffd7\") " pod="openstack/glance-default-external-api-0" Oct 14 09:04:12 crc kubenswrapper[4870]: I1014 09:04:12.599198 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/95cfd5a9-9dfc-44a4-bb94-0d60963dffd7-config-data\") pod \"glance-default-external-api-0\" (UID: \"95cfd5a9-9dfc-44a4-bb94-0d60963dffd7\") " pod="openstack/glance-default-external-api-0" Oct 14 09:04:12 crc kubenswrapper[4870]: I1014 09:04:12.599215 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95cfd5a9-9dfc-44a4-bb94-0d60963dffd7-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"95cfd5a9-9dfc-44a4-bb94-0d60963dffd7\") " pod="openstack/glance-default-external-api-0" Oct 14 09:04:12 crc kubenswrapper[4870]: I1014 09:04:12.599287 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/95cfd5a9-9dfc-44a4-bb94-0d60963dffd7-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"95cfd5a9-9dfc-44a4-bb94-0d60963dffd7\") " pod="openstack/glance-default-external-api-0" Oct 14 09:04:12 crc kubenswrapper[4870]: I1014 09:04:12.599313 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/95cfd5a9-9dfc-44a4-bb94-0d60963dffd7-logs\") pod \"glance-default-external-api-0\" (UID: \"95cfd5a9-9dfc-44a4-bb94-0d60963dffd7\") " pod="openstack/glance-default-external-api-0" Oct 14 09:04:12 crc kubenswrapper[4870]: I1014 09:04:12.600263 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/95cfd5a9-9dfc-44a4-bb94-0d60963dffd7-logs\") pod \"glance-default-external-api-0\" (UID: \"95cfd5a9-9dfc-44a4-bb94-0d60963dffd7\") " pod="openstack/glance-default-external-api-0" Oct 14 09:04:12 crc kubenswrapper[4870]: I1014 09:04:12.600484 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/95cfd5a9-9dfc-44a4-bb94-0d60963dffd7-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"95cfd5a9-9dfc-44a4-bb94-0d60963dffd7\") " pod="openstack/glance-default-external-api-0" Oct 14 09:04:12 crc kubenswrapper[4870]: I1014 09:04:12.620155 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/95cfd5a9-9dfc-44a4-bb94-0d60963dffd7-ceph\") pod \"glance-default-external-api-0\" (UID: \"95cfd5a9-9dfc-44a4-bb94-0d60963dffd7\") " pod="openstack/glance-default-external-api-0" Oct 14 09:04:12 crc kubenswrapper[4870]: I1014 09:04:12.620821 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/95cfd5a9-9dfc-44a4-bb94-0d60963dffd7-config-data\") pod \"glance-default-external-api-0\" (UID: \"95cfd5a9-9dfc-44a4-bb94-0d60963dffd7\") " pod="openstack/glance-default-external-api-0" Oct 14 09:04:12 crc kubenswrapper[4870]: I1014 09:04:12.621178 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95cfd5a9-9dfc-44a4-bb94-0d60963dffd7-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"95cfd5a9-9dfc-44a4-bb94-0d60963dffd7\") " pod="openstack/glance-default-external-api-0" Oct 14 09:04:12 crc kubenswrapper[4870]: I1014 09:04:12.621388 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7bp58\" (UniqueName: \"kubernetes.io/projected/95cfd5a9-9dfc-44a4-bb94-0d60963dffd7-kube-api-access-7bp58\") pod \"glance-default-external-api-0\" (UID: \"95cfd5a9-9dfc-44a4-bb94-0d60963dffd7\") " pod="openstack/glance-default-external-api-0" Oct 14 09:04:12 crc kubenswrapper[4870]: I1014 09:04:12.623285 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/95cfd5a9-9dfc-44a4-bb94-0d60963dffd7-scripts\") pod \"glance-default-external-api-0\" (UID: \"95cfd5a9-9dfc-44a4-bb94-0d60963dffd7\") " pod="openstack/glance-default-external-api-0" Oct 14 09:04:12 crc kubenswrapper[4870]: I1014 09:04:12.724326 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 14 09:04:13 crc kubenswrapper[4870]: I1014 09:04:13.055494 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="829bce4c-6fe7-46fa-a4d8-1558220a039f" path="/var/lib/kubelet/pods/829bce4c-6fe7-46fa-a4d8-1558220a039f/volumes" Oct 14 09:04:13 crc kubenswrapper[4870]: I1014 09:04:13.056298 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="de9df7d3-70ec-4abd-9cd0-55038d9ab2df" path="/var/lib/kubelet/pods/de9df7d3-70ec-4abd-9cd0-55038d9ab2df/volumes" Oct 14 09:04:16 crc kubenswrapper[4870]: I1014 09:04:16.714052 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-8cfqn"] Oct 14 09:04:16 crc kubenswrapper[4870]: I1014 09:04:16.716740 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8cfqn" Oct 14 09:04:16 crc kubenswrapper[4870]: I1014 09:04:16.721382 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-8cfqn"] Oct 14 09:04:16 crc kubenswrapper[4870]: I1014 09:04:16.855554 4870 scope.go:117] "RemoveContainer" containerID="8eb5a5a2c1544b193bef195717e303075859ef64fbf449f7c5c4eb693af477f6" Oct 14 09:04:16 crc kubenswrapper[4870]: I1014 09:04:16.885395 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v8vrw\" (UniqueName: \"kubernetes.io/projected/6e46d0e5-9fdd-449b-80b6-21cbb12034bd-kube-api-access-v8vrw\") pod \"redhat-operators-8cfqn\" (UID: \"6e46d0e5-9fdd-449b-80b6-21cbb12034bd\") " pod="openshift-marketplace/redhat-operators-8cfqn" Oct 14 09:04:16 crc kubenswrapper[4870]: I1014 09:04:16.885595 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6e46d0e5-9fdd-449b-80b6-21cbb12034bd-utilities\") pod \"redhat-operators-8cfqn\" (UID: \"6e46d0e5-9fdd-449b-80b6-21cbb12034bd\") " pod="openshift-marketplace/redhat-operators-8cfqn" Oct 14 09:04:16 crc kubenswrapper[4870]: I1014 09:04:16.885966 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6e46d0e5-9fdd-449b-80b6-21cbb12034bd-catalog-content\") pod \"redhat-operators-8cfqn\" (UID: \"6e46d0e5-9fdd-449b-80b6-21cbb12034bd\") " pod="openshift-marketplace/redhat-operators-8cfqn" Oct 14 09:04:16 crc kubenswrapper[4870]: I1014 09:04:16.975190 4870 scope.go:117] "RemoveContainer" containerID="d5be18839f4b371ce0db12541e6bfc5e5ccd1903925a9840da4265c4d2692924" Oct 14 09:04:16 crc kubenswrapper[4870]: E1014 09:04:16.975750 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d5be18839f4b371ce0db12541e6bfc5e5ccd1903925a9840da4265c4d2692924\": container with ID starting with d5be18839f4b371ce0db12541e6bfc5e5ccd1903925a9840da4265c4d2692924 not found: ID does not exist" containerID="d5be18839f4b371ce0db12541e6bfc5e5ccd1903925a9840da4265c4d2692924" Oct 14 09:04:16 crc kubenswrapper[4870]: I1014 09:04:16.975801 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d5be18839f4b371ce0db12541e6bfc5e5ccd1903925a9840da4265c4d2692924"} err="failed to get container status \"d5be18839f4b371ce0db12541e6bfc5e5ccd1903925a9840da4265c4d2692924\": rpc error: code = NotFound desc = could not find container \"d5be18839f4b371ce0db12541e6bfc5e5ccd1903925a9840da4265c4d2692924\": container with ID starting with d5be18839f4b371ce0db12541e6bfc5e5ccd1903925a9840da4265c4d2692924 not found: ID does not exist" Oct 14 09:04:16 crc kubenswrapper[4870]: I1014 09:04:16.975831 4870 scope.go:117] "RemoveContainer" containerID="8eb5a5a2c1544b193bef195717e303075859ef64fbf449f7c5c4eb693af477f6" Oct 14 09:04:16 crc kubenswrapper[4870]: E1014 09:04:16.976202 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8eb5a5a2c1544b193bef195717e303075859ef64fbf449f7c5c4eb693af477f6\": container with ID starting with 8eb5a5a2c1544b193bef195717e303075859ef64fbf449f7c5c4eb693af477f6 not found: ID does not exist" containerID="8eb5a5a2c1544b193bef195717e303075859ef64fbf449f7c5c4eb693af477f6" Oct 14 09:04:16 crc kubenswrapper[4870]: I1014 09:04:16.976242 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8eb5a5a2c1544b193bef195717e303075859ef64fbf449f7c5c4eb693af477f6"} err="failed to get container status \"8eb5a5a2c1544b193bef195717e303075859ef64fbf449f7c5c4eb693af477f6\": rpc error: code = NotFound desc = could not find container \"8eb5a5a2c1544b193bef195717e303075859ef64fbf449f7c5c4eb693af477f6\": container with ID starting with 8eb5a5a2c1544b193bef195717e303075859ef64fbf449f7c5c4eb693af477f6 not found: ID does not exist" Oct 14 09:04:16 crc kubenswrapper[4870]: I1014 09:04:16.976265 4870 scope.go:117] "RemoveContainer" containerID="6d39e737807423f661bed2bc5c4c82a9315751ab881b785991a4633e1b26b216" Oct 14 09:04:16 crc kubenswrapper[4870]: I1014 09:04:16.987888 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v8vrw\" (UniqueName: \"kubernetes.io/projected/6e46d0e5-9fdd-449b-80b6-21cbb12034bd-kube-api-access-v8vrw\") pod \"redhat-operators-8cfqn\" (UID: \"6e46d0e5-9fdd-449b-80b6-21cbb12034bd\") " pod="openshift-marketplace/redhat-operators-8cfqn" Oct 14 09:04:16 crc kubenswrapper[4870]: I1014 09:04:16.987960 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6e46d0e5-9fdd-449b-80b6-21cbb12034bd-utilities\") pod \"redhat-operators-8cfqn\" (UID: \"6e46d0e5-9fdd-449b-80b6-21cbb12034bd\") " pod="openshift-marketplace/redhat-operators-8cfqn" Oct 14 09:04:16 crc kubenswrapper[4870]: I1014 09:04:16.988133 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6e46d0e5-9fdd-449b-80b6-21cbb12034bd-catalog-content\") pod \"redhat-operators-8cfqn\" (UID: \"6e46d0e5-9fdd-449b-80b6-21cbb12034bd\") " pod="openshift-marketplace/redhat-operators-8cfqn" Oct 14 09:04:16 crc kubenswrapper[4870]: I1014 09:04:16.988528 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6e46d0e5-9fdd-449b-80b6-21cbb12034bd-utilities\") pod \"redhat-operators-8cfqn\" (UID: \"6e46d0e5-9fdd-449b-80b6-21cbb12034bd\") " pod="openshift-marketplace/redhat-operators-8cfqn" Oct 14 09:04:16 crc kubenswrapper[4870]: I1014 09:04:16.988567 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6e46d0e5-9fdd-449b-80b6-21cbb12034bd-catalog-content\") pod \"redhat-operators-8cfqn\" (UID: \"6e46d0e5-9fdd-449b-80b6-21cbb12034bd\") " pod="openshift-marketplace/redhat-operators-8cfqn" Oct 14 09:04:17 crc kubenswrapper[4870]: I1014 09:04:17.012428 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v8vrw\" (UniqueName: \"kubernetes.io/projected/6e46d0e5-9fdd-449b-80b6-21cbb12034bd-kube-api-access-v8vrw\") pod \"redhat-operators-8cfqn\" (UID: \"6e46d0e5-9fdd-449b-80b6-21cbb12034bd\") " pod="openshift-marketplace/redhat-operators-8cfqn" Oct 14 09:04:17 crc kubenswrapper[4870]: I1014 09:04:17.080984 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8cfqn" Oct 14 09:04:17 crc kubenswrapper[4870]: I1014 09:04:17.161994 4870 scope.go:117] "RemoveContainer" containerID="dcdf64ab0812d5c47611df139478fcdb35fb58c52140a0bfaf26de928cff6bc0" Oct 14 09:04:17 crc kubenswrapper[4870]: I1014 09:04:17.221730 4870 scope.go:117] "RemoveContainer" containerID="6d39e737807423f661bed2bc5c4c82a9315751ab881b785991a4633e1b26b216" Oct 14 09:04:17 crc kubenswrapper[4870]: E1014 09:04:17.222669 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6d39e737807423f661bed2bc5c4c82a9315751ab881b785991a4633e1b26b216\": container with ID starting with 6d39e737807423f661bed2bc5c4c82a9315751ab881b785991a4633e1b26b216 not found: ID does not exist" containerID="6d39e737807423f661bed2bc5c4c82a9315751ab881b785991a4633e1b26b216" Oct 14 09:04:17 crc kubenswrapper[4870]: I1014 09:04:17.222700 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6d39e737807423f661bed2bc5c4c82a9315751ab881b785991a4633e1b26b216"} err="failed to get container status \"6d39e737807423f661bed2bc5c4c82a9315751ab881b785991a4633e1b26b216\": rpc error: code = NotFound desc = could not find container \"6d39e737807423f661bed2bc5c4c82a9315751ab881b785991a4633e1b26b216\": container with ID starting with 6d39e737807423f661bed2bc5c4c82a9315751ab881b785991a4633e1b26b216 not found: ID does not exist" Oct 14 09:04:17 crc kubenswrapper[4870]: I1014 09:04:17.222722 4870 scope.go:117] "RemoveContainer" containerID="dcdf64ab0812d5c47611df139478fcdb35fb58c52140a0bfaf26de928cff6bc0" Oct 14 09:04:17 crc kubenswrapper[4870]: E1014 09:04:17.223001 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dcdf64ab0812d5c47611df139478fcdb35fb58c52140a0bfaf26de928cff6bc0\": container with ID starting with dcdf64ab0812d5c47611df139478fcdb35fb58c52140a0bfaf26de928cff6bc0 not found: ID does not exist" containerID="dcdf64ab0812d5c47611df139478fcdb35fb58c52140a0bfaf26de928cff6bc0" Oct 14 09:04:17 crc kubenswrapper[4870]: I1014 09:04:17.223023 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dcdf64ab0812d5c47611df139478fcdb35fb58c52140a0bfaf26de928cff6bc0"} err="failed to get container status \"dcdf64ab0812d5c47611df139478fcdb35fb58c52140a0bfaf26de928cff6bc0\": rpc error: code = NotFound desc = could not find container \"dcdf64ab0812d5c47611df139478fcdb35fb58c52140a0bfaf26de928cff6bc0\": container with ID starting with dcdf64ab0812d5c47611df139478fcdb35fb58c52140a0bfaf26de928cff6bc0 not found: ID does not exist" Oct 14 09:04:17 crc kubenswrapper[4870]: I1014 09:04:17.579302 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 14 09:04:17 crc kubenswrapper[4870]: W1014 09:04:17.606096 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbdd40f40_57d1_4dbf_8b03_f447844dbcd3.slice/crio-35eae101dbdfe2800d03a4f0f2cd7c51e5ce2f0d2d819509039a7c2c8bc46ecb WatchSource:0}: Error finding container 35eae101dbdfe2800d03a4f0f2cd7c51e5ce2f0d2d819509039a7c2c8bc46ecb: Status 404 returned error can't find the container with id 35eae101dbdfe2800d03a4f0f2cd7c51e5ce2f0d2d819509039a7c2c8bc46ecb Oct 14 09:04:17 crc kubenswrapper[4870]: I1014 09:04:17.740882 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-lnlhg"] Oct 14 09:04:17 crc kubenswrapper[4870]: I1014 09:04:17.757825 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 14 09:04:17 crc kubenswrapper[4870]: I1014 09:04:17.889968 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-8cfqn"] Oct 14 09:04:18 crc kubenswrapper[4870]: I1014 09:04:18.018806 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7cff98d7df-2fx9m" event={"ID":"7ad77931-c302-4ce8-81f0-78767bb80443","Type":"ContainerStarted","Data":"d898a5c61cd9c202ebe91686a9d143d06fbb01cb5ba6341691ff053cbe036f0c"} Oct 14 09:04:18 crc kubenswrapper[4870]: I1014 09:04:18.019103 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7cff98d7df-2fx9m" event={"ID":"7ad77931-c302-4ce8-81f0-78767bb80443","Type":"ContainerStarted","Data":"4a68c4125fa0d27922a302bd38cd7ca956c788c5062a558fbf2fe6050cd6ae9a"} Oct 14 09:04:18 crc kubenswrapper[4870]: I1014 09:04:18.028592 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"bdd40f40-57d1-4dbf-8b03-f447844dbcd3","Type":"ContainerStarted","Data":"35eae101dbdfe2800d03a4f0f2cd7c51e5ce2f0d2d819509039a7c2c8bc46ecb"} Oct 14 09:04:18 crc kubenswrapper[4870]: I1014 09:04:18.030605 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lnlhg" event={"ID":"e71ba358-f793-409b-ad26-ba1be98ad917","Type":"ContainerStarted","Data":"09c460e024e6e99bb605b491b86f27cc944734d05a199315a25f4162d4c5194c"} Oct 14 09:04:18 crc kubenswrapper[4870]: I1014 09:04:18.052427 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-7cff98d7df-2fx9m" podStartSLOduration=1.897358361 podStartE2EDuration="10.052410945s" podCreationTimestamp="2025-10-14 09:04:08 +0000 UTC" firstStartedPulling="2025-10-14 09:04:08.992519498 +0000 UTC m=+7384.689879869" lastFinishedPulling="2025-10-14 09:04:17.147572082 +0000 UTC m=+7392.844932453" observedRunningTime="2025-10-14 09:04:18.047838922 +0000 UTC m=+7393.745199283" watchObservedRunningTime="2025-10-14 09:04:18.052410945 +0000 UTC m=+7393.749771306" Oct 14 09:04:18 crc kubenswrapper[4870]: I1014 09:04:18.072649 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8cfqn" event={"ID":"6e46d0e5-9fdd-449b-80b6-21cbb12034bd","Type":"ContainerStarted","Data":"98856dee49b5a7dd704ad52235616341fc0746313f76c36e8189eece4f0dfd2f"} Oct 14 09:04:18 crc kubenswrapper[4870]: I1014 09:04:18.077852 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-869b6669bf-lc5s8" event={"ID":"2ea687b1-b04d-476a-a333-a731a71e7a66","Type":"ContainerStarted","Data":"06e259540492daf3152e3782aabf72bd64d5af26cb809cbb6104992128b775bf"} Oct 14 09:04:18 crc kubenswrapper[4870]: I1014 09:04:18.077882 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-869b6669bf-lc5s8" event={"ID":"2ea687b1-b04d-476a-a333-a731a71e7a66","Type":"ContainerStarted","Data":"b923b85ec0aabb8c092a290ac924282aeefe155dde9b6e97d4241d10dee37081"} Oct 14 09:04:18 crc kubenswrapper[4870]: I1014 09:04:18.083301 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7fbc647f5-9ph8d" event={"ID":"2c148c06-20aa-42c4-8214-575bbd7ce58b","Type":"ContainerStarted","Data":"d96697d8d135d62e8516d08211e951ebbd5b03db0de02038b259e5f52e2aa8c3"} Oct 14 09:04:18 crc kubenswrapper[4870]: I1014 09:04:18.083337 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7fbc647f5-9ph8d" event={"ID":"2c148c06-20aa-42c4-8214-575bbd7ce58b","Type":"ContainerStarted","Data":"56bbe467d4c1379fd0b140aec08cccaffb367f9bbe2482f5c9f68117f91d1ac7"} Oct 14 09:04:18 crc kubenswrapper[4870]: I1014 09:04:18.083446 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-7fbc647f5-9ph8d" podUID="2c148c06-20aa-42c4-8214-575bbd7ce58b" containerName="horizon-log" containerID="cri-o://56bbe467d4c1379fd0b140aec08cccaffb367f9bbe2482f5c9f68117f91d1ac7" gracePeriod=30 Oct 14 09:04:18 crc kubenswrapper[4870]: I1014 09:04:18.083535 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-7fbc647f5-9ph8d" podUID="2c148c06-20aa-42c4-8214-575bbd7ce58b" containerName="horizon" containerID="cri-o://d96697d8d135d62e8516d08211e951ebbd5b03db0de02038b259e5f52e2aa8c3" gracePeriod=30 Oct 14 09:04:18 crc kubenswrapper[4870]: I1014 09:04:18.087130 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"95cfd5a9-9dfc-44a4-bb94-0d60963dffd7","Type":"ContainerStarted","Data":"5183e22070564c7c67e31f46b645e2b760e223f907ef294adac0851d85154e73"} Oct 14 09:04:18 crc kubenswrapper[4870]: I1014 09:04:18.118997 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-869b6669bf-lc5s8" podStartSLOduration=2.468097731 podStartE2EDuration="11.118978239s" podCreationTimestamp="2025-10-14 09:04:07 +0000 UTC" firstStartedPulling="2025-10-14 09:04:08.378397354 +0000 UTC m=+7384.075757725" lastFinishedPulling="2025-10-14 09:04:17.029277862 +0000 UTC m=+7392.726638233" observedRunningTime="2025-10-14 09:04:18.100466052 +0000 UTC m=+7393.797826423" watchObservedRunningTime="2025-10-14 09:04:18.118978239 +0000 UTC m=+7393.816338600" Oct 14 09:04:18 crc kubenswrapper[4870]: I1014 09:04:18.128034 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-7fbc647f5-9ph8d" podStartSLOduration=2.65434084 podStartE2EDuration="11.128013272s" podCreationTimestamp="2025-10-14 09:04:07 +0000 UTC" firstStartedPulling="2025-10-14 09:04:08.555949268 +0000 UTC m=+7384.253309639" lastFinishedPulling="2025-10-14 09:04:17.02962171 +0000 UTC m=+7392.726982071" observedRunningTime="2025-10-14 09:04:18.121944762 +0000 UTC m=+7393.819305133" watchObservedRunningTime="2025-10-14 09:04:18.128013272 +0000 UTC m=+7393.825373633" Oct 14 09:04:18 crc kubenswrapper[4870]: I1014 09:04:18.498640 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-7cff98d7df-2fx9m" Oct 14 09:04:18 crc kubenswrapper[4870]: I1014 09:04:18.500573 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-7cff98d7df-2fx9m" Oct 14 09:04:19 crc kubenswrapper[4870]: I1014 09:04:19.100644 4870 generic.go:334] "Generic (PLEG): container finished" podID="6e46d0e5-9fdd-449b-80b6-21cbb12034bd" containerID="3acf47a771c9fe423e7498d6d5e68924f3e40eeae4fce101462346bfb812ba6b" exitCode=0 Oct 14 09:04:19 crc kubenswrapper[4870]: I1014 09:04:19.100692 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8cfqn" event={"ID":"6e46d0e5-9fdd-449b-80b6-21cbb12034bd","Type":"ContainerDied","Data":"3acf47a771c9fe423e7498d6d5e68924f3e40eeae4fce101462346bfb812ba6b"} Oct 14 09:04:19 crc kubenswrapper[4870]: I1014 09:04:19.102942 4870 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 14 09:04:19 crc kubenswrapper[4870]: I1014 09:04:19.103598 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"95cfd5a9-9dfc-44a4-bb94-0d60963dffd7","Type":"ContainerStarted","Data":"fe2acc5d544daf5b4396b6bc620f66f0b1ecd60cc80096d63326c795bb74a694"} Oct 14 09:04:19 crc kubenswrapper[4870]: I1014 09:04:19.108816 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"bdd40f40-57d1-4dbf-8b03-f447844dbcd3","Type":"ContainerStarted","Data":"1d476f3e706c6350617aa4cfe52c9cfe4a32ca48adb4f7b35f3d523b4a0356ba"} Oct 14 09:04:19 crc kubenswrapper[4870]: I1014 09:04:19.113395 4870 generic.go:334] "Generic (PLEG): container finished" podID="e71ba358-f793-409b-ad26-ba1be98ad917" containerID="8df473df5d7cac61191326d176896e1a6301501b4e79e00507d08c3a99f19241" exitCode=0 Oct 14 09:04:19 crc kubenswrapper[4870]: I1014 09:04:19.113591 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lnlhg" event={"ID":"e71ba358-f793-409b-ad26-ba1be98ad917","Type":"ContainerDied","Data":"8df473df5d7cac61191326d176896e1a6301501b4e79e00507d08c3a99f19241"} Oct 14 09:04:20 crc kubenswrapper[4870]: I1014 09:04:20.132961 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"bdd40f40-57d1-4dbf-8b03-f447844dbcd3","Type":"ContainerStarted","Data":"965a2ed3abf9c4e6398f4949f7b6fd24405591b58d563c93cb44adf8ee0377c8"} Oct 14 09:04:20 crc kubenswrapper[4870]: I1014 09:04:20.140603 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"95cfd5a9-9dfc-44a4-bb94-0d60963dffd7","Type":"ContainerStarted","Data":"b64cf7b139e107b52f7d20140f68309baf94c4e574316399e876b57c1c978d95"} Oct 14 09:04:20 crc kubenswrapper[4870]: I1014 09:04:20.184850 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=9.184817458 podStartE2EDuration="9.184817458s" podCreationTimestamp="2025-10-14 09:04:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 09:04:20.154803047 +0000 UTC m=+7395.852163458" watchObservedRunningTime="2025-10-14 09:04:20.184817458 +0000 UTC m=+7395.882177849" Oct 14 09:04:20 crc kubenswrapper[4870]: I1014 09:04:20.206393 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=8.20636912 podStartE2EDuration="8.20636912s" podCreationTimestamp="2025-10-14 09:04:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 09:04:20.192040096 +0000 UTC m=+7395.889400477" watchObservedRunningTime="2025-10-14 09:04:20.20636912 +0000 UTC m=+7395.903729501" Oct 14 09:04:21 crc kubenswrapper[4870]: I1014 09:04:21.155114 4870 generic.go:334] "Generic (PLEG): container finished" podID="e71ba358-f793-409b-ad26-ba1be98ad917" containerID="2887f23568558a0ee07b1de44b992866061e47935b145ffb56de2bd31720c1f6" exitCode=0 Oct 14 09:04:21 crc kubenswrapper[4870]: I1014 09:04:21.155171 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lnlhg" event={"ID":"e71ba358-f793-409b-ad26-ba1be98ad917","Type":"ContainerDied","Data":"2887f23568558a0ee07b1de44b992866061e47935b145ffb56de2bd31720c1f6"} Oct 14 09:04:21 crc kubenswrapper[4870]: I1014 09:04:21.163954 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8cfqn" event={"ID":"6e46d0e5-9fdd-449b-80b6-21cbb12034bd","Type":"ContainerStarted","Data":"d2ba368f31f0295498018ad8f24fd0d53e0e6cb74759c68fc7f90afd52a2f09a"} Oct 14 09:04:22 crc kubenswrapper[4870]: I1014 09:04:22.371842 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 14 09:04:22 crc kubenswrapper[4870]: I1014 09:04:22.372083 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 14 09:04:22 crc kubenswrapper[4870]: I1014 09:04:22.423626 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 14 09:04:22 crc kubenswrapper[4870]: I1014 09:04:22.434978 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 14 09:04:22 crc kubenswrapper[4870]: I1014 09:04:22.724801 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 14 09:04:22 crc kubenswrapper[4870]: I1014 09:04:22.724889 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 14 09:04:22 crc kubenswrapper[4870]: I1014 09:04:22.755242 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 14 09:04:22 crc kubenswrapper[4870]: I1014 09:04:22.763983 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 14 09:04:23 crc kubenswrapper[4870]: I1014 09:04:23.185023 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lnlhg" event={"ID":"e71ba358-f793-409b-ad26-ba1be98ad917","Type":"ContainerStarted","Data":"ecef19eab11d1dc906e43d04d5df39661da9ca6a50acb504f0d0fe98184a17b6"} Oct 14 09:04:23 crc kubenswrapper[4870]: I1014 09:04:23.185061 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 14 09:04:23 crc kubenswrapper[4870]: I1014 09:04:23.185078 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 14 09:04:23 crc kubenswrapper[4870]: I1014 09:04:23.185097 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 14 09:04:23 crc kubenswrapper[4870]: I1014 09:04:23.185130 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 14 09:04:23 crc kubenswrapper[4870]: I1014 09:04:23.203530 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-lnlhg" podStartSLOduration=9.170316309 podStartE2EDuration="12.203513455s" podCreationTimestamp="2025-10-14 09:04:11 +0000 UTC" firstStartedPulling="2025-10-14 09:04:19.116336535 +0000 UTC m=+7394.813696906" lastFinishedPulling="2025-10-14 09:04:22.149533681 +0000 UTC m=+7397.846894052" observedRunningTime="2025-10-14 09:04:23.198943142 +0000 UTC m=+7398.896303513" watchObservedRunningTime="2025-10-14 09:04:23.203513455 +0000 UTC m=+7398.900873826" Oct 14 09:04:23 crc kubenswrapper[4870]: I1014 09:04:23.950655 4870 patch_prober.go:28] interesting pod/machine-config-daemon-7tvc7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 09:04:23 crc kubenswrapper[4870]: I1014 09:04:23.950969 4870 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 09:04:25 crc kubenswrapper[4870]: I1014 09:04:25.208074 4870 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 14 09:04:25 crc kubenswrapper[4870]: I1014 09:04:25.208118 4870 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 14 09:04:25 crc kubenswrapper[4870]: I1014 09:04:25.435157 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 14 09:04:25 crc kubenswrapper[4870]: I1014 09:04:25.563479 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 14 09:04:26 crc kubenswrapper[4870]: I1014 09:04:26.219125 4870 generic.go:334] "Generic (PLEG): container finished" podID="6e46d0e5-9fdd-449b-80b6-21cbb12034bd" containerID="d2ba368f31f0295498018ad8f24fd0d53e0e6cb74759c68fc7f90afd52a2f09a" exitCode=0 Oct 14 09:04:26 crc kubenswrapper[4870]: I1014 09:04:26.219199 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8cfqn" event={"ID":"6e46d0e5-9fdd-449b-80b6-21cbb12034bd","Type":"ContainerDied","Data":"d2ba368f31f0295498018ad8f24fd0d53e0e6cb74759c68fc7f90afd52a2f09a"} Oct 14 09:04:26 crc kubenswrapper[4870]: I1014 09:04:26.219418 4870 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 14 09:04:26 crc kubenswrapper[4870]: I1014 09:04:26.220179 4870 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 14 09:04:26 crc kubenswrapper[4870]: I1014 09:04:26.352433 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 14 09:04:26 crc kubenswrapper[4870]: I1014 09:04:26.444144 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 14 09:04:27 crc kubenswrapper[4870]: I1014 09:04:27.801712 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-869b6669bf-lc5s8" Oct 14 09:04:27 crc kubenswrapper[4870]: I1014 09:04:27.802171 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-869b6669bf-lc5s8" Oct 14 09:04:27 crc kubenswrapper[4870]: I1014 09:04:27.803210 4870 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-869b6669bf-lc5s8" podUID="2ea687b1-b04d-476a-a333-a731a71e7a66" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.104:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.104:8080: connect: connection refused" Oct 14 09:04:27 crc kubenswrapper[4870]: I1014 09:04:27.944027 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-7fbc647f5-9ph8d" Oct 14 09:04:28 crc kubenswrapper[4870]: I1014 09:04:28.239858 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8cfqn" event={"ID":"6e46d0e5-9fdd-449b-80b6-21cbb12034bd","Type":"ContainerStarted","Data":"f7f6dd96fe79989f88753444093ac7e2b47b327de58cb8e833cf3fe38ce0903b"} Oct 14 09:04:28 crc kubenswrapper[4870]: I1014 09:04:28.264916 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-8cfqn" podStartSLOduration=4.3995387 podStartE2EDuration="12.264899791s" podCreationTimestamp="2025-10-14 09:04:16 +0000 UTC" firstStartedPulling="2025-10-14 09:04:19.102663667 +0000 UTC m=+7394.800024038" lastFinishedPulling="2025-10-14 09:04:26.968024758 +0000 UTC m=+7402.665385129" observedRunningTime="2025-10-14 09:04:28.257520229 +0000 UTC m=+7403.954880600" watchObservedRunningTime="2025-10-14 09:04:28.264899791 +0000 UTC m=+7403.962260162" Oct 14 09:04:28 crc kubenswrapper[4870]: I1014 09:04:28.499546 4870 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-7cff98d7df-2fx9m" podUID="7ad77931-c302-4ce8-81f0-78767bb80443" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.106:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.106:8080: connect: connection refused" Oct 14 09:04:32 crc kubenswrapper[4870]: I1014 09:04:32.060427 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-lnlhg" Oct 14 09:04:32 crc kubenswrapper[4870]: I1014 09:04:32.060931 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-lnlhg" Oct 14 09:04:32 crc kubenswrapper[4870]: I1014 09:04:32.131560 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-lnlhg" Oct 14 09:04:32 crc kubenswrapper[4870]: I1014 09:04:32.324219 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-lnlhg" Oct 14 09:04:32 crc kubenswrapper[4870]: I1014 09:04:32.385735 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-lnlhg"] Oct 14 09:04:34 crc kubenswrapper[4870]: I1014 09:04:34.291504 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-lnlhg" podUID="e71ba358-f793-409b-ad26-ba1be98ad917" containerName="registry-server" containerID="cri-o://ecef19eab11d1dc906e43d04d5df39661da9ca6a50acb504f0d0fe98184a17b6" gracePeriod=2 Oct 14 09:04:34 crc kubenswrapper[4870]: I1014 09:04:34.811797 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lnlhg" Oct 14 09:04:34 crc kubenswrapper[4870]: I1014 09:04:34.856634 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kw8lf\" (UniqueName: \"kubernetes.io/projected/e71ba358-f793-409b-ad26-ba1be98ad917-kube-api-access-kw8lf\") pod \"e71ba358-f793-409b-ad26-ba1be98ad917\" (UID: \"e71ba358-f793-409b-ad26-ba1be98ad917\") " Oct 14 09:04:34 crc kubenswrapper[4870]: I1014 09:04:34.856827 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e71ba358-f793-409b-ad26-ba1be98ad917-catalog-content\") pod \"e71ba358-f793-409b-ad26-ba1be98ad917\" (UID: \"e71ba358-f793-409b-ad26-ba1be98ad917\") " Oct 14 09:04:34 crc kubenswrapper[4870]: I1014 09:04:34.857651 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e71ba358-f793-409b-ad26-ba1be98ad917-utilities\") pod \"e71ba358-f793-409b-ad26-ba1be98ad917\" (UID: \"e71ba358-f793-409b-ad26-ba1be98ad917\") " Oct 14 09:04:34 crc kubenswrapper[4870]: I1014 09:04:34.858278 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e71ba358-f793-409b-ad26-ba1be98ad917-utilities" (OuterVolumeSpecName: "utilities") pod "e71ba358-f793-409b-ad26-ba1be98ad917" (UID: "e71ba358-f793-409b-ad26-ba1be98ad917"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 09:04:34 crc kubenswrapper[4870]: I1014 09:04:34.862634 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e71ba358-f793-409b-ad26-ba1be98ad917-kube-api-access-kw8lf" (OuterVolumeSpecName: "kube-api-access-kw8lf") pod "e71ba358-f793-409b-ad26-ba1be98ad917" (UID: "e71ba358-f793-409b-ad26-ba1be98ad917"). InnerVolumeSpecName "kube-api-access-kw8lf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:04:34 crc kubenswrapper[4870]: I1014 09:04:34.903052 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e71ba358-f793-409b-ad26-ba1be98ad917-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e71ba358-f793-409b-ad26-ba1be98ad917" (UID: "e71ba358-f793-409b-ad26-ba1be98ad917"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 09:04:34 crc kubenswrapper[4870]: I1014 09:04:34.959167 4870 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e71ba358-f793-409b-ad26-ba1be98ad917-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 09:04:34 crc kubenswrapper[4870]: I1014 09:04:34.959203 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kw8lf\" (UniqueName: \"kubernetes.io/projected/e71ba358-f793-409b-ad26-ba1be98ad917-kube-api-access-kw8lf\") on node \"crc\" DevicePath \"\"" Oct 14 09:04:34 crc kubenswrapper[4870]: I1014 09:04:34.959215 4870 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e71ba358-f793-409b-ad26-ba1be98ad917-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 09:04:35 crc kubenswrapper[4870]: I1014 09:04:35.305149 4870 generic.go:334] "Generic (PLEG): container finished" podID="e71ba358-f793-409b-ad26-ba1be98ad917" containerID="ecef19eab11d1dc906e43d04d5df39661da9ca6a50acb504f0d0fe98184a17b6" exitCode=0 Oct 14 09:04:35 crc kubenswrapper[4870]: I1014 09:04:35.305203 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lnlhg" event={"ID":"e71ba358-f793-409b-ad26-ba1be98ad917","Type":"ContainerDied","Data":"ecef19eab11d1dc906e43d04d5df39661da9ca6a50acb504f0d0fe98184a17b6"} Oct 14 09:04:35 crc kubenswrapper[4870]: I1014 09:04:35.305232 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lnlhg" Oct 14 09:04:35 crc kubenswrapper[4870]: I1014 09:04:35.305262 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lnlhg" event={"ID":"e71ba358-f793-409b-ad26-ba1be98ad917","Type":"ContainerDied","Data":"09c460e024e6e99bb605b491b86f27cc944734d05a199315a25f4162d4c5194c"} Oct 14 09:04:35 crc kubenswrapper[4870]: I1014 09:04:35.305302 4870 scope.go:117] "RemoveContainer" containerID="ecef19eab11d1dc906e43d04d5df39661da9ca6a50acb504f0d0fe98184a17b6" Oct 14 09:04:35 crc kubenswrapper[4870]: I1014 09:04:35.336357 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-lnlhg"] Oct 14 09:04:35 crc kubenswrapper[4870]: I1014 09:04:35.338223 4870 scope.go:117] "RemoveContainer" containerID="2887f23568558a0ee07b1de44b992866061e47935b145ffb56de2bd31720c1f6" Oct 14 09:04:35 crc kubenswrapper[4870]: I1014 09:04:35.347256 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-lnlhg"] Oct 14 09:04:35 crc kubenswrapper[4870]: I1014 09:04:35.362914 4870 scope.go:117] "RemoveContainer" containerID="8df473df5d7cac61191326d176896e1a6301501b4e79e00507d08c3a99f19241" Oct 14 09:04:35 crc kubenswrapper[4870]: I1014 09:04:35.407511 4870 scope.go:117] "RemoveContainer" containerID="ecef19eab11d1dc906e43d04d5df39661da9ca6a50acb504f0d0fe98184a17b6" Oct 14 09:04:35 crc kubenswrapper[4870]: E1014 09:04:35.414532 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ecef19eab11d1dc906e43d04d5df39661da9ca6a50acb504f0d0fe98184a17b6\": container with ID starting with ecef19eab11d1dc906e43d04d5df39661da9ca6a50acb504f0d0fe98184a17b6 not found: ID does not exist" containerID="ecef19eab11d1dc906e43d04d5df39661da9ca6a50acb504f0d0fe98184a17b6" Oct 14 09:04:35 crc kubenswrapper[4870]: I1014 09:04:35.414619 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ecef19eab11d1dc906e43d04d5df39661da9ca6a50acb504f0d0fe98184a17b6"} err="failed to get container status \"ecef19eab11d1dc906e43d04d5df39661da9ca6a50acb504f0d0fe98184a17b6\": rpc error: code = NotFound desc = could not find container \"ecef19eab11d1dc906e43d04d5df39661da9ca6a50acb504f0d0fe98184a17b6\": container with ID starting with ecef19eab11d1dc906e43d04d5df39661da9ca6a50acb504f0d0fe98184a17b6 not found: ID does not exist" Oct 14 09:04:35 crc kubenswrapper[4870]: I1014 09:04:35.414661 4870 scope.go:117] "RemoveContainer" containerID="2887f23568558a0ee07b1de44b992866061e47935b145ffb56de2bd31720c1f6" Oct 14 09:04:35 crc kubenswrapper[4870]: E1014 09:04:35.415127 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2887f23568558a0ee07b1de44b992866061e47935b145ffb56de2bd31720c1f6\": container with ID starting with 2887f23568558a0ee07b1de44b992866061e47935b145ffb56de2bd31720c1f6 not found: ID does not exist" containerID="2887f23568558a0ee07b1de44b992866061e47935b145ffb56de2bd31720c1f6" Oct 14 09:04:35 crc kubenswrapper[4870]: I1014 09:04:35.415172 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2887f23568558a0ee07b1de44b992866061e47935b145ffb56de2bd31720c1f6"} err="failed to get container status \"2887f23568558a0ee07b1de44b992866061e47935b145ffb56de2bd31720c1f6\": rpc error: code = NotFound desc = could not find container \"2887f23568558a0ee07b1de44b992866061e47935b145ffb56de2bd31720c1f6\": container with ID starting with 2887f23568558a0ee07b1de44b992866061e47935b145ffb56de2bd31720c1f6 not found: ID does not exist" Oct 14 09:04:35 crc kubenswrapper[4870]: I1014 09:04:35.415198 4870 scope.go:117] "RemoveContainer" containerID="8df473df5d7cac61191326d176896e1a6301501b4e79e00507d08c3a99f19241" Oct 14 09:04:35 crc kubenswrapper[4870]: E1014 09:04:35.416490 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8df473df5d7cac61191326d176896e1a6301501b4e79e00507d08c3a99f19241\": container with ID starting with 8df473df5d7cac61191326d176896e1a6301501b4e79e00507d08c3a99f19241 not found: ID does not exist" containerID="8df473df5d7cac61191326d176896e1a6301501b4e79e00507d08c3a99f19241" Oct 14 09:04:35 crc kubenswrapper[4870]: I1014 09:04:35.416534 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8df473df5d7cac61191326d176896e1a6301501b4e79e00507d08c3a99f19241"} err="failed to get container status \"8df473df5d7cac61191326d176896e1a6301501b4e79e00507d08c3a99f19241\": rpc error: code = NotFound desc = could not find container \"8df473df5d7cac61191326d176896e1a6301501b4e79e00507d08c3a99f19241\": container with ID starting with 8df473df5d7cac61191326d176896e1a6301501b4e79e00507d08c3a99f19241 not found: ID does not exist" Oct 14 09:04:37 crc kubenswrapper[4870]: I1014 09:04:37.049468 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e71ba358-f793-409b-ad26-ba1be98ad917" path="/var/lib/kubelet/pods/e71ba358-f793-409b-ad26-ba1be98ad917/volumes" Oct 14 09:04:37 crc kubenswrapper[4870]: I1014 09:04:37.082540 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-8cfqn" Oct 14 09:04:37 crc kubenswrapper[4870]: I1014 09:04:37.082845 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-8cfqn" Oct 14 09:04:38 crc kubenswrapper[4870]: I1014 09:04:38.127474 4870 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-8cfqn" podUID="6e46d0e5-9fdd-449b-80b6-21cbb12034bd" containerName="registry-server" probeResult="failure" output=< Oct 14 09:04:38 crc kubenswrapper[4870]: timeout: failed to connect service ":50051" within 1s Oct 14 09:04:38 crc kubenswrapper[4870]: > Oct 14 09:04:39 crc kubenswrapper[4870]: I1014 09:04:39.707195 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-869b6669bf-lc5s8" Oct 14 09:04:40 crc kubenswrapper[4870]: I1014 09:04:40.251762 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-7cff98d7df-2fx9m" Oct 14 09:04:41 crc kubenswrapper[4870]: I1014 09:04:41.293418 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-869b6669bf-lc5s8" Oct 14 09:04:41 crc kubenswrapper[4870]: I1014 09:04:41.946060 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-7cff98d7df-2fx9m" Oct 14 09:04:42 crc kubenswrapper[4870]: I1014 09:04:42.035666 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-869b6669bf-lc5s8"] Oct 14 09:04:42 crc kubenswrapper[4870]: I1014 09:04:42.035928 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-869b6669bf-lc5s8" podUID="2ea687b1-b04d-476a-a333-a731a71e7a66" containerName="horizon-log" containerID="cri-o://b923b85ec0aabb8c092a290ac924282aeefe155dde9b6e97d4241d10dee37081" gracePeriod=30 Oct 14 09:04:42 crc kubenswrapper[4870]: I1014 09:04:42.036016 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-869b6669bf-lc5s8" podUID="2ea687b1-b04d-476a-a333-a731a71e7a66" containerName="horizon" containerID="cri-o://06e259540492daf3152e3782aabf72bd64d5af26cb809cbb6104992128b775bf" gracePeriod=30 Oct 14 09:04:45 crc kubenswrapper[4870]: I1014 09:04:45.438728 4870 generic.go:334] "Generic (PLEG): container finished" podID="2ea687b1-b04d-476a-a333-a731a71e7a66" containerID="06e259540492daf3152e3782aabf72bd64d5af26cb809cbb6104992128b775bf" exitCode=0 Oct 14 09:04:45 crc kubenswrapper[4870]: I1014 09:04:45.438837 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-869b6669bf-lc5s8" event={"ID":"2ea687b1-b04d-476a-a333-a731a71e7a66","Type":"ContainerDied","Data":"06e259540492daf3152e3782aabf72bd64d5af26cb809cbb6104992128b775bf"} Oct 14 09:04:47 crc kubenswrapper[4870]: I1014 09:04:47.801658 4870 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-869b6669bf-lc5s8" podUID="2ea687b1-b04d-476a-a333-a731a71e7a66" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.104:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.104:8080: connect: connection refused" Oct 14 09:04:48 crc kubenswrapper[4870]: I1014 09:04:48.156917 4870 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-8cfqn" podUID="6e46d0e5-9fdd-449b-80b6-21cbb12034bd" containerName="registry-server" probeResult="failure" output=< Oct 14 09:04:48 crc kubenswrapper[4870]: timeout: failed to connect service ":50051" within 1s Oct 14 09:04:48 crc kubenswrapper[4870]: > Oct 14 09:04:48 crc kubenswrapper[4870]: I1014 09:04:48.473973 4870 generic.go:334] "Generic (PLEG): container finished" podID="2c148c06-20aa-42c4-8214-575bbd7ce58b" containerID="d96697d8d135d62e8516d08211e951ebbd5b03db0de02038b259e5f52e2aa8c3" exitCode=137 Oct 14 09:04:48 crc kubenswrapper[4870]: I1014 09:04:48.474293 4870 generic.go:334] "Generic (PLEG): container finished" podID="2c148c06-20aa-42c4-8214-575bbd7ce58b" containerID="56bbe467d4c1379fd0b140aec08cccaffb367f9bbe2482f5c9f68117f91d1ac7" exitCode=137 Oct 14 09:04:48 crc kubenswrapper[4870]: I1014 09:04:48.474073 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7fbc647f5-9ph8d" event={"ID":"2c148c06-20aa-42c4-8214-575bbd7ce58b","Type":"ContainerDied","Data":"d96697d8d135d62e8516d08211e951ebbd5b03db0de02038b259e5f52e2aa8c3"} Oct 14 09:04:48 crc kubenswrapper[4870]: I1014 09:04:48.474343 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7fbc647f5-9ph8d" event={"ID":"2c148c06-20aa-42c4-8214-575bbd7ce58b","Type":"ContainerDied","Data":"56bbe467d4c1379fd0b140aec08cccaffb367f9bbe2482f5c9f68117f91d1ac7"} Oct 14 09:04:48 crc kubenswrapper[4870]: I1014 09:04:48.474355 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7fbc647f5-9ph8d" event={"ID":"2c148c06-20aa-42c4-8214-575bbd7ce58b","Type":"ContainerDied","Data":"1fa60abd1a023611ea23394b6a0c85a280a323a5f1dcafede77d19200ab455a8"} Oct 14 09:04:48 crc kubenswrapper[4870]: I1014 09:04:48.474366 4870 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1fa60abd1a023611ea23394b6a0c85a280a323a5f1dcafede77d19200ab455a8" Oct 14 09:04:48 crc kubenswrapper[4870]: I1014 09:04:48.517473 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7fbc647f5-9ph8d" Oct 14 09:04:48 crc kubenswrapper[4870]: I1014 09:04:48.539895 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2c148c06-20aa-42c4-8214-575bbd7ce58b-scripts\") pod \"2c148c06-20aa-42c4-8214-575bbd7ce58b\" (UID: \"2c148c06-20aa-42c4-8214-575bbd7ce58b\") " Oct 14 09:04:48 crc kubenswrapper[4870]: I1014 09:04:48.539962 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gkz7d\" (UniqueName: \"kubernetes.io/projected/2c148c06-20aa-42c4-8214-575bbd7ce58b-kube-api-access-gkz7d\") pod \"2c148c06-20aa-42c4-8214-575bbd7ce58b\" (UID: \"2c148c06-20aa-42c4-8214-575bbd7ce58b\") " Oct 14 09:04:48 crc kubenswrapper[4870]: I1014 09:04:48.540071 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2c148c06-20aa-42c4-8214-575bbd7ce58b-logs\") pod \"2c148c06-20aa-42c4-8214-575bbd7ce58b\" (UID: \"2c148c06-20aa-42c4-8214-575bbd7ce58b\") " Oct 14 09:04:48 crc kubenswrapper[4870]: I1014 09:04:48.540188 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/2c148c06-20aa-42c4-8214-575bbd7ce58b-horizon-secret-key\") pod \"2c148c06-20aa-42c4-8214-575bbd7ce58b\" (UID: \"2c148c06-20aa-42c4-8214-575bbd7ce58b\") " Oct 14 09:04:48 crc kubenswrapper[4870]: I1014 09:04:48.540257 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2c148c06-20aa-42c4-8214-575bbd7ce58b-config-data\") pod \"2c148c06-20aa-42c4-8214-575bbd7ce58b\" (UID: \"2c148c06-20aa-42c4-8214-575bbd7ce58b\") " Oct 14 09:04:48 crc kubenswrapper[4870]: I1014 09:04:48.541727 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2c148c06-20aa-42c4-8214-575bbd7ce58b-logs" (OuterVolumeSpecName: "logs") pod "2c148c06-20aa-42c4-8214-575bbd7ce58b" (UID: "2c148c06-20aa-42c4-8214-575bbd7ce58b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 09:04:48 crc kubenswrapper[4870]: I1014 09:04:48.542367 4870 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2c148c06-20aa-42c4-8214-575bbd7ce58b-logs\") on node \"crc\" DevicePath \"\"" Oct 14 09:04:48 crc kubenswrapper[4870]: I1014 09:04:48.549853 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2c148c06-20aa-42c4-8214-575bbd7ce58b-kube-api-access-gkz7d" (OuterVolumeSpecName: "kube-api-access-gkz7d") pod "2c148c06-20aa-42c4-8214-575bbd7ce58b" (UID: "2c148c06-20aa-42c4-8214-575bbd7ce58b"). InnerVolumeSpecName "kube-api-access-gkz7d". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:04:48 crc kubenswrapper[4870]: I1014 09:04:48.552276 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2c148c06-20aa-42c4-8214-575bbd7ce58b-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "2c148c06-20aa-42c4-8214-575bbd7ce58b" (UID: "2c148c06-20aa-42c4-8214-575bbd7ce58b"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:04:48 crc kubenswrapper[4870]: I1014 09:04:48.586845 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2c148c06-20aa-42c4-8214-575bbd7ce58b-config-data" (OuterVolumeSpecName: "config-data") pod "2c148c06-20aa-42c4-8214-575bbd7ce58b" (UID: "2c148c06-20aa-42c4-8214-575bbd7ce58b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 09:04:48 crc kubenswrapper[4870]: I1014 09:04:48.588616 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2c148c06-20aa-42c4-8214-575bbd7ce58b-scripts" (OuterVolumeSpecName: "scripts") pod "2c148c06-20aa-42c4-8214-575bbd7ce58b" (UID: "2c148c06-20aa-42c4-8214-575bbd7ce58b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 09:04:48 crc kubenswrapper[4870]: I1014 09:04:48.643622 4870 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2c148c06-20aa-42c4-8214-575bbd7ce58b-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 09:04:48 crc kubenswrapper[4870]: I1014 09:04:48.643657 4870 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2c148c06-20aa-42c4-8214-575bbd7ce58b-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 09:04:48 crc kubenswrapper[4870]: I1014 09:04:48.643668 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gkz7d\" (UniqueName: \"kubernetes.io/projected/2c148c06-20aa-42c4-8214-575bbd7ce58b-kube-api-access-gkz7d\") on node \"crc\" DevicePath \"\"" Oct 14 09:04:48 crc kubenswrapper[4870]: I1014 09:04:48.643680 4870 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/2c148c06-20aa-42c4-8214-575bbd7ce58b-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Oct 14 09:04:49 crc kubenswrapper[4870]: I1014 09:04:49.483083 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7fbc647f5-9ph8d" Oct 14 09:04:49 crc kubenswrapper[4870]: I1014 09:04:49.522920 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7fbc647f5-9ph8d"] Oct 14 09:04:49 crc kubenswrapper[4870]: I1014 09:04:49.533865 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-7fbc647f5-9ph8d"] Oct 14 09:04:51 crc kubenswrapper[4870]: I1014 09:04:51.055169 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2c148c06-20aa-42c4-8214-575bbd7ce58b" path="/var/lib/kubelet/pods/2c148c06-20aa-42c4-8214-575bbd7ce58b/volumes" Oct 14 09:04:53 crc kubenswrapper[4870]: I1014 09:04:53.066973 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-8ctvq"] Oct 14 09:04:53 crc kubenswrapper[4870]: I1014 09:04:53.080596 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-8ctvq"] Oct 14 09:04:53 crc kubenswrapper[4870]: I1014 09:04:53.950879 4870 patch_prober.go:28] interesting pod/machine-config-daemon-7tvc7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 09:04:53 crc kubenswrapper[4870]: I1014 09:04:53.951398 4870 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 09:04:55 crc kubenswrapper[4870]: I1014 09:04:55.072662 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="803b85ed-0a7c-475b-bbfd-23265017a113" path="/var/lib/kubelet/pods/803b85ed-0a7c-475b-bbfd-23265017a113/volumes" Oct 14 09:04:57 crc kubenswrapper[4870]: I1014 09:04:57.147504 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-8cfqn" Oct 14 09:04:57 crc kubenswrapper[4870]: I1014 09:04:57.230554 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-8cfqn" Oct 14 09:04:57 crc kubenswrapper[4870]: I1014 09:04:57.388341 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-8cfqn"] Oct 14 09:04:57 crc kubenswrapper[4870]: I1014 09:04:57.802398 4870 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-869b6669bf-lc5s8" podUID="2ea687b1-b04d-476a-a333-a731a71e7a66" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.104:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.104:8080: connect: connection refused" Oct 14 09:04:58 crc kubenswrapper[4870]: I1014 09:04:58.589352 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-8cfqn" podUID="6e46d0e5-9fdd-449b-80b6-21cbb12034bd" containerName="registry-server" containerID="cri-o://f7f6dd96fe79989f88753444093ac7e2b47b327de58cb8e833cf3fe38ce0903b" gracePeriod=2 Oct 14 09:04:59 crc kubenswrapper[4870]: I1014 09:04:59.091366 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8cfqn" Oct 14 09:04:59 crc kubenswrapper[4870]: I1014 09:04:59.185690 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6e46d0e5-9fdd-449b-80b6-21cbb12034bd-catalog-content\") pod \"6e46d0e5-9fdd-449b-80b6-21cbb12034bd\" (UID: \"6e46d0e5-9fdd-449b-80b6-21cbb12034bd\") " Oct 14 09:04:59 crc kubenswrapper[4870]: I1014 09:04:59.185813 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6e46d0e5-9fdd-449b-80b6-21cbb12034bd-utilities\") pod \"6e46d0e5-9fdd-449b-80b6-21cbb12034bd\" (UID: \"6e46d0e5-9fdd-449b-80b6-21cbb12034bd\") " Oct 14 09:04:59 crc kubenswrapper[4870]: I1014 09:04:59.187343 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6e46d0e5-9fdd-449b-80b6-21cbb12034bd-utilities" (OuterVolumeSpecName: "utilities") pod "6e46d0e5-9fdd-449b-80b6-21cbb12034bd" (UID: "6e46d0e5-9fdd-449b-80b6-21cbb12034bd"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 09:04:59 crc kubenswrapper[4870]: I1014 09:04:59.264042 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6e46d0e5-9fdd-449b-80b6-21cbb12034bd-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6e46d0e5-9fdd-449b-80b6-21cbb12034bd" (UID: "6e46d0e5-9fdd-449b-80b6-21cbb12034bd"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 09:04:59 crc kubenswrapper[4870]: I1014 09:04:59.287824 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v8vrw\" (UniqueName: \"kubernetes.io/projected/6e46d0e5-9fdd-449b-80b6-21cbb12034bd-kube-api-access-v8vrw\") pod \"6e46d0e5-9fdd-449b-80b6-21cbb12034bd\" (UID: \"6e46d0e5-9fdd-449b-80b6-21cbb12034bd\") " Oct 14 09:04:59 crc kubenswrapper[4870]: I1014 09:04:59.288832 4870 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6e46d0e5-9fdd-449b-80b6-21cbb12034bd-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 09:04:59 crc kubenswrapper[4870]: I1014 09:04:59.288861 4870 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6e46d0e5-9fdd-449b-80b6-21cbb12034bd-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 09:04:59 crc kubenswrapper[4870]: I1014 09:04:59.295004 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6e46d0e5-9fdd-449b-80b6-21cbb12034bd-kube-api-access-v8vrw" (OuterVolumeSpecName: "kube-api-access-v8vrw") pod "6e46d0e5-9fdd-449b-80b6-21cbb12034bd" (UID: "6e46d0e5-9fdd-449b-80b6-21cbb12034bd"). InnerVolumeSpecName "kube-api-access-v8vrw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:04:59 crc kubenswrapper[4870]: I1014 09:04:59.391069 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v8vrw\" (UniqueName: \"kubernetes.io/projected/6e46d0e5-9fdd-449b-80b6-21cbb12034bd-kube-api-access-v8vrw\") on node \"crc\" DevicePath \"\"" Oct 14 09:04:59 crc kubenswrapper[4870]: I1014 09:04:59.605573 4870 generic.go:334] "Generic (PLEG): container finished" podID="6e46d0e5-9fdd-449b-80b6-21cbb12034bd" containerID="f7f6dd96fe79989f88753444093ac7e2b47b327de58cb8e833cf3fe38ce0903b" exitCode=0 Oct 14 09:04:59 crc kubenswrapper[4870]: I1014 09:04:59.605629 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8cfqn" event={"ID":"6e46d0e5-9fdd-449b-80b6-21cbb12034bd","Type":"ContainerDied","Data":"f7f6dd96fe79989f88753444093ac7e2b47b327de58cb8e833cf3fe38ce0903b"} Oct 14 09:04:59 crc kubenswrapper[4870]: I1014 09:04:59.605669 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8cfqn" event={"ID":"6e46d0e5-9fdd-449b-80b6-21cbb12034bd","Type":"ContainerDied","Data":"98856dee49b5a7dd704ad52235616341fc0746313f76c36e8189eece4f0dfd2f"} Oct 14 09:04:59 crc kubenswrapper[4870]: I1014 09:04:59.605755 4870 scope.go:117] "RemoveContainer" containerID="f7f6dd96fe79989f88753444093ac7e2b47b327de58cb8e833cf3fe38ce0903b" Oct 14 09:04:59 crc kubenswrapper[4870]: I1014 09:04:59.605873 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8cfqn" Oct 14 09:04:59 crc kubenswrapper[4870]: I1014 09:04:59.640843 4870 scope.go:117] "RemoveContainer" containerID="d2ba368f31f0295498018ad8f24fd0d53e0e6cb74759c68fc7f90afd52a2f09a" Oct 14 09:04:59 crc kubenswrapper[4870]: I1014 09:04:59.665543 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-8cfqn"] Oct 14 09:04:59 crc kubenswrapper[4870]: I1014 09:04:59.675546 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-8cfqn"] Oct 14 09:04:59 crc kubenswrapper[4870]: I1014 09:04:59.683362 4870 scope.go:117] "RemoveContainer" containerID="3acf47a771c9fe423e7498d6d5e68924f3e40eeae4fce101462346bfb812ba6b" Oct 14 09:04:59 crc kubenswrapper[4870]: I1014 09:04:59.744164 4870 scope.go:117] "RemoveContainer" containerID="f7f6dd96fe79989f88753444093ac7e2b47b327de58cb8e833cf3fe38ce0903b" Oct 14 09:04:59 crc kubenswrapper[4870]: E1014 09:04:59.745330 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f7f6dd96fe79989f88753444093ac7e2b47b327de58cb8e833cf3fe38ce0903b\": container with ID starting with f7f6dd96fe79989f88753444093ac7e2b47b327de58cb8e833cf3fe38ce0903b not found: ID does not exist" containerID="f7f6dd96fe79989f88753444093ac7e2b47b327de58cb8e833cf3fe38ce0903b" Oct 14 09:04:59 crc kubenswrapper[4870]: I1014 09:04:59.745374 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f7f6dd96fe79989f88753444093ac7e2b47b327de58cb8e833cf3fe38ce0903b"} err="failed to get container status \"f7f6dd96fe79989f88753444093ac7e2b47b327de58cb8e833cf3fe38ce0903b\": rpc error: code = NotFound desc = could not find container \"f7f6dd96fe79989f88753444093ac7e2b47b327de58cb8e833cf3fe38ce0903b\": container with ID starting with f7f6dd96fe79989f88753444093ac7e2b47b327de58cb8e833cf3fe38ce0903b not found: ID does not exist" Oct 14 09:04:59 crc kubenswrapper[4870]: I1014 09:04:59.745403 4870 scope.go:117] "RemoveContainer" containerID="d2ba368f31f0295498018ad8f24fd0d53e0e6cb74759c68fc7f90afd52a2f09a" Oct 14 09:04:59 crc kubenswrapper[4870]: E1014 09:04:59.745718 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d2ba368f31f0295498018ad8f24fd0d53e0e6cb74759c68fc7f90afd52a2f09a\": container with ID starting with d2ba368f31f0295498018ad8f24fd0d53e0e6cb74759c68fc7f90afd52a2f09a not found: ID does not exist" containerID="d2ba368f31f0295498018ad8f24fd0d53e0e6cb74759c68fc7f90afd52a2f09a" Oct 14 09:04:59 crc kubenswrapper[4870]: I1014 09:04:59.745765 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d2ba368f31f0295498018ad8f24fd0d53e0e6cb74759c68fc7f90afd52a2f09a"} err="failed to get container status \"d2ba368f31f0295498018ad8f24fd0d53e0e6cb74759c68fc7f90afd52a2f09a\": rpc error: code = NotFound desc = could not find container \"d2ba368f31f0295498018ad8f24fd0d53e0e6cb74759c68fc7f90afd52a2f09a\": container with ID starting with d2ba368f31f0295498018ad8f24fd0d53e0e6cb74759c68fc7f90afd52a2f09a not found: ID does not exist" Oct 14 09:04:59 crc kubenswrapper[4870]: I1014 09:04:59.745784 4870 scope.go:117] "RemoveContainer" containerID="3acf47a771c9fe423e7498d6d5e68924f3e40eeae4fce101462346bfb812ba6b" Oct 14 09:04:59 crc kubenswrapper[4870]: E1014 09:04:59.746138 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3acf47a771c9fe423e7498d6d5e68924f3e40eeae4fce101462346bfb812ba6b\": container with ID starting with 3acf47a771c9fe423e7498d6d5e68924f3e40eeae4fce101462346bfb812ba6b not found: ID does not exist" containerID="3acf47a771c9fe423e7498d6d5e68924f3e40eeae4fce101462346bfb812ba6b" Oct 14 09:04:59 crc kubenswrapper[4870]: I1014 09:04:59.746184 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3acf47a771c9fe423e7498d6d5e68924f3e40eeae4fce101462346bfb812ba6b"} err="failed to get container status \"3acf47a771c9fe423e7498d6d5e68924f3e40eeae4fce101462346bfb812ba6b\": rpc error: code = NotFound desc = could not find container \"3acf47a771c9fe423e7498d6d5e68924f3e40eeae4fce101462346bfb812ba6b\": container with ID starting with 3acf47a771c9fe423e7498d6d5e68924f3e40eeae4fce101462346bfb812ba6b not found: ID does not exist" Oct 14 09:05:01 crc kubenswrapper[4870]: I1014 09:05:01.049019 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6e46d0e5-9fdd-449b-80b6-21cbb12034bd" path="/var/lib/kubelet/pods/6e46d0e5-9fdd-449b-80b6-21cbb12034bd/volumes" Oct 14 09:05:03 crc kubenswrapper[4870]: I1014 09:05:03.032155 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-9d9f-account-create-wr5dt"] Oct 14 09:05:03 crc kubenswrapper[4870]: I1014 09:05:03.053967 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-9d9f-account-create-wr5dt"] Oct 14 09:05:05 crc kubenswrapper[4870]: I1014 09:05:05.053859 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efb802fd-6646-4bed-842e-4208b746bf7f" path="/var/lib/kubelet/pods/efb802fd-6646-4bed-842e-4208b746bf7f/volumes" Oct 14 09:05:07 crc kubenswrapper[4870]: I1014 09:05:07.801794 4870 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-869b6669bf-lc5s8" podUID="2ea687b1-b04d-476a-a333-a731a71e7a66" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.104:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.104:8080: connect: connection refused" Oct 14 09:05:07 crc kubenswrapper[4870]: I1014 09:05:07.802565 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-869b6669bf-lc5s8" Oct 14 09:05:12 crc kubenswrapper[4870]: W1014 09:05:12.095191 4870 container.go:586] Failed to update stats for container "/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2ea687b1_b04d_476a_a333_a731a71e7a66.slice/crio-52c3a13da1b0e7af4b4cd912cf7b283d670168315a1ecdd839c1f94c4c0b74ce": error while statting cgroup v2: [unable to parse /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2ea687b1_b04d_476a_a333_a731a71e7a66.slice/crio-52c3a13da1b0e7af4b4cd912cf7b283d670168315a1ecdd839c1f94c4c0b74ce/io.stat: read /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2ea687b1_b04d_476a_a333_a731a71e7a66.slice/crio-52c3a13da1b0e7af4b4cd912cf7b283d670168315a1ecdd839c1f94c4c0b74ce/io.stat: no such device], continuing to push stats Oct 14 09:05:12 crc kubenswrapper[4870]: I1014 09:05:12.534905 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-869b6669bf-lc5s8" Oct 14 09:05:12 crc kubenswrapper[4870]: I1014 09:05:12.602611 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnhnl\" (UniqueName: \"kubernetes.io/projected/2ea687b1-b04d-476a-a333-a731a71e7a66-kube-api-access-rnhnl\") pod \"2ea687b1-b04d-476a-a333-a731a71e7a66\" (UID: \"2ea687b1-b04d-476a-a333-a731a71e7a66\") " Oct 14 09:05:12 crc kubenswrapper[4870]: I1014 09:05:12.602681 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/2ea687b1-b04d-476a-a333-a731a71e7a66-horizon-secret-key\") pod \"2ea687b1-b04d-476a-a333-a731a71e7a66\" (UID: \"2ea687b1-b04d-476a-a333-a731a71e7a66\") " Oct 14 09:05:12 crc kubenswrapper[4870]: I1014 09:05:12.602735 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2ea687b1-b04d-476a-a333-a731a71e7a66-scripts\") pod \"2ea687b1-b04d-476a-a333-a731a71e7a66\" (UID: \"2ea687b1-b04d-476a-a333-a731a71e7a66\") " Oct 14 09:05:12 crc kubenswrapper[4870]: I1014 09:05:12.602829 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2ea687b1-b04d-476a-a333-a731a71e7a66-config-data\") pod \"2ea687b1-b04d-476a-a333-a731a71e7a66\" (UID: \"2ea687b1-b04d-476a-a333-a731a71e7a66\") " Oct 14 09:05:12 crc kubenswrapper[4870]: I1014 09:05:12.602913 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2ea687b1-b04d-476a-a333-a731a71e7a66-logs\") pod \"2ea687b1-b04d-476a-a333-a731a71e7a66\" (UID: \"2ea687b1-b04d-476a-a333-a731a71e7a66\") " Oct 14 09:05:12 crc kubenswrapper[4870]: I1014 09:05:12.603861 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2ea687b1-b04d-476a-a333-a731a71e7a66-logs" (OuterVolumeSpecName: "logs") pod "2ea687b1-b04d-476a-a333-a731a71e7a66" (UID: "2ea687b1-b04d-476a-a333-a731a71e7a66"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 09:05:12 crc kubenswrapper[4870]: I1014 09:05:12.609564 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2ea687b1-b04d-476a-a333-a731a71e7a66-kube-api-access-rnhnl" (OuterVolumeSpecName: "kube-api-access-rnhnl") pod "2ea687b1-b04d-476a-a333-a731a71e7a66" (UID: "2ea687b1-b04d-476a-a333-a731a71e7a66"). InnerVolumeSpecName "kube-api-access-rnhnl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:05:12 crc kubenswrapper[4870]: I1014 09:05:12.613993 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2ea687b1-b04d-476a-a333-a731a71e7a66-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "2ea687b1-b04d-476a-a333-a731a71e7a66" (UID: "2ea687b1-b04d-476a-a333-a731a71e7a66"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:05:12 crc kubenswrapper[4870]: I1014 09:05:12.634729 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2ea687b1-b04d-476a-a333-a731a71e7a66-scripts" (OuterVolumeSpecName: "scripts") pod "2ea687b1-b04d-476a-a333-a731a71e7a66" (UID: "2ea687b1-b04d-476a-a333-a731a71e7a66"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 09:05:12 crc kubenswrapper[4870]: I1014 09:05:12.647171 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2ea687b1-b04d-476a-a333-a731a71e7a66-config-data" (OuterVolumeSpecName: "config-data") pod "2ea687b1-b04d-476a-a333-a731a71e7a66" (UID: "2ea687b1-b04d-476a-a333-a731a71e7a66"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 09:05:12 crc kubenswrapper[4870]: I1014 09:05:12.705658 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnhnl\" (UniqueName: \"kubernetes.io/projected/2ea687b1-b04d-476a-a333-a731a71e7a66-kube-api-access-rnhnl\") on node \"crc\" DevicePath \"\"" Oct 14 09:05:12 crc kubenswrapper[4870]: I1014 09:05:12.705693 4870 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/2ea687b1-b04d-476a-a333-a731a71e7a66-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Oct 14 09:05:12 crc kubenswrapper[4870]: I1014 09:05:12.705702 4870 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2ea687b1-b04d-476a-a333-a731a71e7a66-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 09:05:12 crc kubenswrapper[4870]: I1014 09:05:12.705712 4870 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2ea687b1-b04d-476a-a333-a731a71e7a66-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 09:05:12 crc kubenswrapper[4870]: I1014 09:05:12.705720 4870 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2ea687b1-b04d-476a-a333-a731a71e7a66-logs\") on node \"crc\" DevicePath \"\"" Oct 14 09:05:12 crc kubenswrapper[4870]: I1014 09:05:12.754358 4870 generic.go:334] "Generic (PLEG): container finished" podID="2ea687b1-b04d-476a-a333-a731a71e7a66" containerID="b923b85ec0aabb8c092a290ac924282aeefe155dde9b6e97d4241d10dee37081" exitCode=137 Oct 14 09:05:12 crc kubenswrapper[4870]: I1014 09:05:12.754403 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-869b6669bf-lc5s8" Oct 14 09:05:12 crc kubenswrapper[4870]: I1014 09:05:12.754410 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-869b6669bf-lc5s8" event={"ID":"2ea687b1-b04d-476a-a333-a731a71e7a66","Type":"ContainerDied","Data":"b923b85ec0aabb8c092a290ac924282aeefe155dde9b6e97d4241d10dee37081"} Oct 14 09:05:12 crc kubenswrapper[4870]: I1014 09:05:12.754487 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-869b6669bf-lc5s8" event={"ID":"2ea687b1-b04d-476a-a333-a731a71e7a66","Type":"ContainerDied","Data":"52c3a13da1b0e7af4b4cd912cf7b283d670168315a1ecdd839c1f94c4c0b74ce"} Oct 14 09:05:12 crc kubenswrapper[4870]: I1014 09:05:12.754510 4870 scope.go:117] "RemoveContainer" containerID="06e259540492daf3152e3782aabf72bd64d5af26cb809cbb6104992128b775bf" Oct 14 09:05:12 crc kubenswrapper[4870]: I1014 09:05:12.786073 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-869b6669bf-lc5s8"] Oct 14 09:05:12 crc kubenswrapper[4870]: I1014 09:05:12.792987 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-869b6669bf-lc5s8"] Oct 14 09:05:12 crc kubenswrapper[4870]: I1014 09:05:12.944099 4870 scope.go:117] "RemoveContainer" containerID="b923b85ec0aabb8c092a290ac924282aeefe155dde9b6e97d4241d10dee37081" Oct 14 09:05:12 crc kubenswrapper[4870]: I1014 09:05:12.967691 4870 scope.go:117] "RemoveContainer" containerID="06e259540492daf3152e3782aabf72bd64d5af26cb809cbb6104992128b775bf" Oct 14 09:05:12 crc kubenswrapper[4870]: E1014 09:05:12.969878 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"06e259540492daf3152e3782aabf72bd64d5af26cb809cbb6104992128b775bf\": container with ID starting with 06e259540492daf3152e3782aabf72bd64d5af26cb809cbb6104992128b775bf not found: ID does not exist" containerID="06e259540492daf3152e3782aabf72bd64d5af26cb809cbb6104992128b775bf" Oct 14 09:05:12 crc kubenswrapper[4870]: I1014 09:05:12.969929 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"06e259540492daf3152e3782aabf72bd64d5af26cb809cbb6104992128b775bf"} err="failed to get container status \"06e259540492daf3152e3782aabf72bd64d5af26cb809cbb6104992128b775bf\": rpc error: code = NotFound desc = could not find container \"06e259540492daf3152e3782aabf72bd64d5af26cb809cbb6104992128b775bf\": container with ID starting with 06e259540492daf3152e3782aabf72bd64d5af26cb809cbb6104992128b775bf not found: ID does not exist" Oct 14 09:05:12 crc kubenswrapper[4870]: I1014 09:05:12.969957 4870 scope.go:117] "RemoveContainer" containerID="b923b85ec0aabb8c092a290ac924282aeefe155dde9b6e97d4241d10dee37081" Oct 14 09:05:12 crc kubenswrapper[4870]: E1014 09:05:12.970330 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b923b85ec0aabb8c092a290ac924282aeefe155dde9b6e97d4241d10dee37081\": container with ID starting with b923b85ec0aabb8c092a290ac924282aeefe155dde9b6e97d4241d10dee37081 not found: ID does not exist" containerID="b923b85ec0aabb8c092a290ac924282aeefe155dde9b6e97d4241d10dee37081" Oct 14 09:05:12 crc kubenswrapper[4870]: I1014 09:05:12.970365 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b923b85ec0aabb8c092a290ac924282aeefe155dde9b6e97d4241d10dee37081"} err="failed to get container status \"b923b85ec0aabb8c092a290ac924282aeefe155dde9b6e97d4241d10dee37081\": rpc error: code = NotFound desc = could not find container \"b923b85ec0aabb8c092a290ac924282aeefe155dde9b6e97d4241d10dee37081\": container with ID starting with b923b85ec0aabb8c092a290ac924282aeefe155dde9b6e97d4241d10dee37081 not found: ID does not exist" Oct 14 09:05:13 crc kubenswrapper[4870]: I1014 09:05:13.050319 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2ea687b1-b04d-476a-a333-a731a71e7a66" path="/var/lib/kubelet/pods/2ea687b1-b04d-476a-a333-a731a71e7a66/volumes" Oct 14 09:05:15 crc kubenswrapper[4870]: I1014 09:05:15.051385 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-hqrsx"] Oct 14 09:05:15 crc kubenswrapper[4870]: I1014 09:05:15.052251 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-hqrsx"] Oct 14 09:05:17 crc kubenswrapper[4870]: I1014 09:05:17.057317 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="019b2f50-511d-44f0-9a63-05bbed8c08e8" path="/var/lib/kubelet/pods/019b2f50-511d-44f0-9a63-05bbed8c08e8/volumes" Oct 14 09:05:23 crc kubenswrapper[4870]: I1014 09:05:23.951246 4870 patch_prober.go:28] interesting pod/machine-config-daemon-7tvc7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 09:05:23 crc kubenswrapper[4870]: I1014 09:05:23.951853 4870 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 09:05:23 crc kubenswrapper[4870]: I1014 09:05:23.951913 4870 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" Oct 14 09:05:23 crc kubenswrapper[4870]: I1014 09:05:23.952624 4870 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"8fcaeb3090f0c89cba1e66db524b4a246072daa35592012664aca450fa09ddc9"} pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 14 09:05:23 crc kubenswrapper[4870]: I1014 09:05:23.952724 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerName="machine-config-daemon" containerID="cri-o://8fcaeb3090f0c89cba1e66db524b4a246072daa35592012664aca450fa09ddc9" gracePeriod=600 Oct 14 09:05:24 crc kubenswrapper[4870]: E1014 09:05:24.082304 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 09:05:24 crc kubenswrapper[4870]: I1014 09:05:24.919493 4870 generic.go:334] "Generic (PLEG): container finished" podID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerID="8fcaeb3090f0c89cba1e66db524b4a246072daa35592012664aca450fa09ddc9" exitCode=0 Oct 14 09:05:24 crc kubenswrapper[4870]: I1014 09:05:24.919595 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" event={"ID":"874a5e45-dffd-4d17-b609-b7d3ed2eab07","Type":"ContainerDied","Data":"8fcaeb3090f0c89cba1e66db524b4a246072daa35592012664aca450fa09ddc9"} Oct 14 09:05:24 crc kubenswrapper[4870]: I1014 09:05:24.919924 4870 scope.go:117] "RemoveContainer" containerID="9426cde7309378345e6129005b55744e852dbf5b2ef5a92cc907fc57d1dc3ae8" Oct 14 09:05:24 crc kubenswrapper[4870]: I1014 09:05:24.921173 4870 scope.go:117] "RemoveContainer" containerID="8fcaeb3090f0c89cba1e66db524b4a246072daa35592012664aca450fa09ddc9" Oct 14 09:05:24 crc kubenswrapper[4870]: E1014 09:05:24.922040 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 09:05:25 crc kubenswrapper[4870]: I1014 09:05:25.065284 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-6fb9fc4bf-jmsxf"] Oct 14 09:05:25 crc kubenswrapper[4870]: E1014 09:05:25.066246 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e71ba358-f793-409b-ad26-ba1be98ad917" containerName="registry-server" Oct 14 09:05:25 crc kubenswrapper[4870]: I1014 09:05:25.066265 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="e71ba358-f793-409b-ad26-ba1be98ad917" containerName="registry-server" Oct 14 09:05:25 crc kubenswrapper[4870]: E1014 09:05:25.066294 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e46d0e5-9fdd-449b-80b6-21cbb12034bd" containerName="extract-utilities" Oct 14 09:05:25 crc kubenswrapper[4870]: I1014 09:05:25.066301 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e46d0e5-9fdd-449b-80b6-21cbb12034bd" containerName="extract-utilities" Oct 14 09:05:25 crc kubenswrapper[4870]: E1014 09:05:25.066325 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c148c06-20aa-42c4-8214-575bbd7ce58b" containerName="horizon-log" Oct 14 09:05:25 crc kubenswrapper[4870]: I1014 09:05:25.066332 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c148c06-20aa-42c4-8214-575bbd7ce58b" containerName="horizon-log" Oct 14 09:05:25 crc kubenswrapper[4870]: E1014 09:05:25.066356 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c148c06-20aa-42c4-8214-575bbd7ce58b" containerName="horizon" Oct 14 09:05:25 crc kubenswrapper[4870]: I1014 09:05:25.066362 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c148c06-20aa-42c4-8214-575bbd7ce58b" containerName="horizon" Oct 14 09:05:25 crc kubenswrapper[4870]: E1014 09:05:25.066374 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ea687b1-b04d-476a-a333-a731a71e7a66" containerName="horizon-log" Oct 14 09:05:25 crc kubenswrapper[4870]: I1014 09:05:25.066380 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ea687b1-b04d-476a-a333-a731a71e7a66" containerName="horizon-log" Oct 14 09:05:25 crc kubenswrapper[4870]: E1014 09:05:25.066394 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ea687b1-b04d-476a-a333-a731a71e7a66" containerName="horizon" Oct 14 09:05:25 crc kubenswrapper[4870]: I1014 09:05:25.066399 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ea687b1-b04d-476a-a333-a731a71e7a66" containerName="horizon" Oct 14 09:05:25 crc kubenswrapper[4870]: E1014 09:05:25.066418 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e46d0e5-9fdd-449b-80b6-21cbb12034bd" containerName="registry-server" Oct 14 09:05:25 crc kubenswrapper[4870]: I1014 09:05:25.066424 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e46d0e5-9fdd-449b-80b6-21cbb12034bd" containerName="registry-server" Oct 14 09:05:25 crc kubenswrapper[4870]: E1014 09:05:25.066457 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e71ba358-f793-409b-ad26-ba1be98ad917" containerName="extract-utilities" Oct 14 09:05:25 crc kubenswrapper[4870]: I1014 09:05:25.066466 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="e71ba358-f793-409b-ad26-ba1be98ad917" containerName="extract-utilities" Oct 14 09:05:25 crc kubenswrapper[4870]: E1014 09:05:25.066487 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e46d0e5-9fdd-449b-80b6-21cbb12034bd" containerName="extract-content" Oct 14 09:05:25 crc kubenswrapper[4870]: I1014 09:05:25.066495 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e46d0e5-9fdd-449b-80b6-21cbb12034bd" containerName="extract-content" Oct 14 09:05:25 crc kubenswrapper[4870]: E1014 09:05:25.066549 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e71ba358-f793-409b-ad26-ba1be98ad917" containerName="extract-content" Oct 14 09:05:25 crc kubenswrapper[4870]: I1014 09:05:25.066556 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="e71ba358-f793-409b-ad26-ba1be98ad917" containerName="extract-content" Oct 14 09:05:25 crc kubenswrapper[4870]: I1014 09:05:25.068397 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="2ea687b1-b04d-476a-a333-a731a71e7a66" containerName="horizon-log" Oct 14 09:05:25 crc kubenswrapper[4870]: I1014 09:05:25.068450 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="6e46d0e5-9fdd-449b-80b6-21cbb12034bd" containerName="registry-server" Oct 14 09:05:25 crc kubenswrapper[4870]: I1014 09:05:25.068466 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="2ea687b1-b04d-476a-a333-a731a71e7a66" containerName="horizon" Oct 14 09:05:25 crc kubenswrapper[4870]: I1014 09:05:25.068496 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="e71ba358-f793-409b-ad26-ba1be98ad917" containerName="registry-server" Oct 14 09:05:25 crc kubenswrapper[4870]: I1014 09:05:25.068520 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="2c148c06-20aa-42c4-8214-575bbd7ce58b" containerName="horizon" Oct 14 09:05:25 crc kubenswrapper[4870]: I1014 09:05:25.068533 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="2c148c06-20aa-42c4-8214-575bbd7ce58b" containerName="horizon-log" Oct 14 09:05:25 crc kubenswrapper[4870]: I1014 09:05:25.078670 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6fb9fc4bf-jmsxf" Oct 14 09:05:25 crc kubenswrapper[4870]: I1014 09:05:25.079411 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-6fb9fc4bf-jmsxf"] Oct 14 09:05:25 crc kubenswrapper[4870]: I1014 09:05:25.180301 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/73bab307-e3b8-47df-bba3-fa624f4c4261-scripts\") pod \"horizon-6fb9fc4bf-jmsxf\" (UID: \"73bab307-e3b8-47df-bba3-fa624f4c4261\") " pod="openstack/horizon-6fb9fc4bf-jmsxf" Oct 14 09:05:25 crc kubenswrapper[4870]: I1014 09:05:25.180631 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/73bab307-e3b8-47df-bba3-fa624f4c4261-horizon-secret-key\") pod \"horizon-6fb9fc4bf-jmsxf\" (UID: \"73bab307-e3b8-47df-bba3-fa624f4c4261\") " pod="openstack/horizon-6fb9fc4bf-jmsxf" Oct 14 09:05:25 crc kubenswrapper[4870]: I1014 09:05:25.181194 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/73bab307-e3b8-47df-bba3-fa624f4c4261-logs\") pod \"horizon-6fb9fc4bf-jmsxf\" (UID: \"73bab307-e3b8-47df-bba3-fa624f4c4261\") " pod="openstack/horizon-6fb9fc4bf-jmsxf" Oct 14 09:05:25 crc kubenswrapper[4870]: I1014 09:05:25.181349 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cxgbg\" (UniqueName: \"kubernetes.io/projected/73bab307-e3b8-47df-bba3-fa624f4c4261-kube-api-access-cxgbg\") pod \"horizon-6fb9fc4bf-jmsxf\" (UID: \"73bab307-e3b8-47df-bba3-fa624f4c4261\") " pod="openstack/horizon-6fb9fc4bf-jmsxf" Oct 14 09:05:25 crc kubenswrapper[4870]: I1014 09:05:25.181471 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/73bab307-e3b8-47df-bba3-fa624f4c4261-config-data\") pod \"horizon-6fb9fc4bf-jmsxf\" (UID: \"73bab307-e3b8-47df-bba3-fa624f4c4261\") " pod="openstack/horizon-6fb9fc4bf-jmsxf" Oct 14 09:05:25 crc kubenswrapper[4870]: I1014 09:05:25.282905 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/73bab307-e3b8-47df-bba3-fa624f4c4261-scripts\") pod \"horizon-6fb9fc4bf-jmsxf\" (UID: \"73bab307-e3b8-47df-bba3-fa624f4c4261\") " pod="openstack/horizon-6fb9fc4bf-jmsxf" Oct 14 09:05:25 crc kubenswrapper[4870]: I1014 09:05:25.283041 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/73bab307-e3b8-47df-bba3-fa624f4c4261-horizon-secret-key\") pod \"horizon-6fb9fc4bf-jmsxf\" (UID: \"73bab307-e3b8-47df-bba3-fa624f4c4261\") " pod="openstack/horizon-6fb9fc4bf-jmsxf" Oct 14 09:05:25 crc kubenswrapper[4870]: I1014 09:05:25.283073 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/73bab307-e3b8-47df-bba3-fa624f4c4261-logs\") pod \"horizon-6fb9fc4bf-jmsxf\" (UID: \"73bab307-e3b8-47df-bba3-fa624f4c4261\") " pod="openstack/horizon-6fb9fc4bf-jmsxf" Oct 14 09:05:25 crc kubenswrapper[4870]: I1014 09:05:25.283117 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cxgbg\" (UniqueName: \"kubernetes.io/projected/73bab307-e3b8-47df-bba3-fa624f4c4261-kube-api-access-cxgbg\") pod \"horizon-6fb9fc4bf-jmsxf\" (UID: \"73bab307-e3b8-47df-bba3-fa624f4c4261\") " pod="openstack/horizon-6fb9fc4bf-jmsxf" Oct 14 09:05:25 crc kubenswrapper[4870]: I1014 09:05:25.283148 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/73bab307-e3b8-47df-bba3-fa624f4c4261-config-data\") pod \"horizon-6fb9fc4bf-jmsxf\" (UID: \"73bab307-e3b8-47df-bba3-fa624f4c4261\") " pod="openstack/horizon-6fb9fc4bf-jmsxf" Oct 14 09:05:25 crc kubenswrapper[4870]: I1014 09:05:25.283914 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/73bab307-e3b8-47df-bba3-fa624f4c4261-scripts\") pod \"horizon-6fb9fc4bf-jmsxf\" (UID: \"73bab307-e3b8-47df-bba3-fa624f4c4261\") " pod="openstack/horizon-6fb9fc4bf-jmsxf" Oct 14 09:05:25 crc kubenswrapper[4870]: I1014 09:05:25.284795 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/73bab307-e3b8-47df-bba3-fa624f4c4261-config-data\") pod \"horizon-6fb9fc4bf-jmsxf\" (UID: \"73bab307-e3b8-47df-bba3-fa624f4c4261\") " pod="openstack/horizon-6fb9fc4bf-jmsxf" Oct 14 09:05:25 crc kubenswrapper[4870]: I1014 09:05:25.284948 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/73bab307-e3b8-47df-bba3-fa624f4c4261-logs\") pod \"horizon-6fb9fc4bf-jmsxf\" (UID: \"73bab307-e3b8-47df-bba3-fa624f4c4261\") " pod="openstack/horizon-6fb9fc4bf-jmsxf" Oct 14 09:05:25 crc kubenswrapper[4870]: I1014 09:05:25.289569 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/73bab307-e3b8-47df-bba3-fa624f4c4261-horizon-secret-key\") pod \"horizon-6fb9fc4bf-jmsxf\" (UID: \"73bab307-e3b8-47df-bba3-fa624f4c4261\") " pod="openstack/horizon-6fb9fc4bf-jmsxf" Oct 14 09:05:25 crc kubenswrapper[4870]: I1014 09:05:25.307845 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cxgbg\" (UniqueName: \"kubernetes.io/projected/73bab307-e3b8-47df-bba3-fa624f4c4261-kube-api-access-cxgbg\") pod \"horizon-6fb9fc4bf-jmsxf\" (UID: \"73bab307-e3b8-47df-bba3-fa624f4c4261\") " pod="openstack/horizon-6fb9fc4bf-jmsxf" Oct 14 09:05:25 crc kubenswrapper[4870]: I1014 09:05:25.403182 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6fb9fc4bf-jmsxf" Oct 14 09:05:25 crc kubenswrapper[4870]: I1014 09:05:25.851732 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-6fb9fc4bf-jmsxf"] Oct 14 09:05:25 crc kubenswrapper[4870]: I1014 09:05:25.933806 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6fb9fc4bf-jmsxf" event={"ID":"73bab307-e3b8-47df-bba3-fa624f4c4261","Type":"ContainerStarted","Data":"4cb88123d713f9e11600e75fe0b50a01a3c1e0eb51c8f398332a6487fb1d3059"} Oct 14 09:05:26 crc kubenswrapper[4870]: I1014 09:05:26.154100 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-db-create-njft6"] Oct 14 09:05:26 crc kubenswrapper[4870]: I1014 09:05:26.155618 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-njft6" Oct 14 09:05:26 crc kubenswrapper[4870]: I1014 09:05:26.165742 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-create-njft6"] Oct 14 09:05:26 crc kubenswrapper[4870]: I1014 09:05:26.300657 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xmlzl\" (UniqueName: \"kubernetes.io/projected/e10cd527-69c5-4c6c-bd52-774cdb2dc9e3-kube-api-access-xmlzl\") pod \"heat-db-create-njft6\" (UID: \"e10cd527-69c5-4c6c-bd52-774cdb2dc9e3\") " pod="openstack/heat-db-create-njft6" Oct 14 09:05:26 crc kubenswrapper[4870]: I1014 09:05:26.403193 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xmlzl\" (UniqueName: \"kubernetes.io/projected/e10cd527-69c5-4c6c-bd52-774cdb2dc9e3-kube-api-access-xmlzl\") pod \"heat-db-create-njft6\" (UID: \"e10cd527-69c5-4c6c-bd52-774cdb2dc9e3\") " pod="openstack/heat-db-create-njft6" Oct 14 09:05:26 crc kubenswrapper[4870]: I1014 09:05:26.424434 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xmlzl\" (UniqueName: \"kubernetes.io/projected/e10cd527-69c5-4c6c-bd52-774cdb2dc9e3-kube-api-access-xmlzl\") pod \"heat-db-create-njft6\" (UID: \"e10cd527-69c5-4c6c-bd52-774cdb2dc9e3\") " pod="openstack/heat-db-create-njft6" Oct 14 09:05:26 crc kubenswrapper[4870]: I1014 09:05:26.499581 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-njft6" Oct 14 09:05:26 crc kubenswrapper[4870]: I1014 09:05:26.947830 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6fb9fc4bf-jmsxf" event={"ID":"73bab307-e3b8-47df-bba3-fa624f4c4261","Type":"ContainerStarted","Data":"50382b4becd0f5924ec409cd7cf5a5ecdf6ce09d7c8b6bda5f34e09e6be64e15"} Oct 14 09:05:26 crc kubenswrapper[4870]: I1014 09:05:26.947871 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6fb9fc4bf-jmsxf" event={"ID":"73bab307-e3b8-47df-bba3-fa624f4c4261","Type":"ContainerStarted","Data":"75ee191ace8dd309e7dd12a52acd942f71087d84cad384d421ce516e579b8e64"} Oct 14 09:05:26 crc kubenswrapper[4870]: I1014 09:05:26.976938 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-create-njft6"] Oct 14 09:05:26 crc kubenswrapper[4870]: I1014 09:05:26.984115 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-6fb9fc4bf-jmsxf" podStartSLOduration=1.984096291 podStartE2EDuration="1.984096291s" podCreationTimestamp="2025-10-14 09:05:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 09:05:26.982404429 +0000 UTC m=+7462.679764790" watchObservedRunningTime="2025-10-14 09:05:26.984096291 +0000 UTC m=+7462.681456662" Oct 14 09:05:27 crc kubenswrapper[4870]: I1014 09:05:27.967795 4870 generic.go:334] "Generic (PLEG): container finished" podID="e10cd527-69c5-4c6c-bd52-774cdb2dc9e3" containerID="39257698f6cdcb9408754a91d15b0bc347d48f9bd53382fceab1014d43df1138" exitCode=0 Oct 14 09:05:27 crc kubenswrapper[4870]: I1014 09:05:27.967866 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-njft6" event={"ID":"e10cd527-69c5-4c6c-bd52-774cdb2dc9e3","Type":"ContainerDied","Data":"39257698f6cdcb9408754a91d15b0bc347d48f9bd53382fceab1014d43df1138"} Oct 14 09:05:27 crc kubenswrapper[4870]: I1014 09:05:27.968360 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-njft6" event={"ID":"e10cd527-69c5-4c6c-bd52-774cdb2dc9e3","Type":"ContainerStarted","Data":"bf1725f612eb595b1eca002690633e77a34906f0050aee4048d917d7cb2e9c60"} Oct 14 09:05:28 crc kubenswrapper[4870]: I1014 09:05:28.502772 4870 scope.go:117] "RemoveContainer" containerID="265bb2cb99c9bcf3c27b483692e9d6d16286d098bd5f2f4414e975b6e72ffaa5" Oct 14 09:05:28 crc kubenswrapper[4870]: I1014 09:05:28.541738 4870 scope.go:117] "RemoveContainer" containerID="e446213e53ea1d18c3e066b85dce2df279affa0d0e817f3cbc1e3252be59fe91" Oct 14 09:05:28 crc kubenswrapper[4870]: I1014 09:05:28.615098 4870 scope.go:117] "RemoveContainer" containerID="b99681a3b1ae47c5d44defc5cbe2dfbe66c07469f3187c7620644febe6ff30bc" Oct 14 09:05:29 crc kubenswrapper[4870]: I1014 09:05:29.370062 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-njft6" Oct 14 09:05:29 crc kubenswrapper[4870]: I1014 09:05:29.471549 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xmlzl\" (UniqueName: \"kubernetes.io/projected/e10cd527-69c5-4c6c-bd52-774cdb2dc9e3-kube-api-access-xmlzl\") pod \"e10cd527-69c5-4c6c-bd52-774cdb2dc9e3\" (UID: \"e10cd527-69c5-4c6c-bd52-774cdb2dc9e3\") " Oct 14 09:05:29 crc kubenswrapper[4870]: I1014 09:05:29.478948 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e10cd527-69c5-4c6c-bd52-774cdb2dc9e3-kube-api-access-xmlzl" (OuterVolumeSpecName: "kube-api-access-xmlzl") pod "e10cd527-69c5-4c6c-bd52-774cdb2dc9e3" (UID: "e10cd527-69c5-4c6c-bd52-774cdb2dc9e3"). InnerVolumeSpecName "kube-api-access-xmlzl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:05:29 crc kubenswrapper[4870]: I1014 09:05:29.573606 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xmlzl\" (UniqueName: \"kubernetes.io/projected/e10cd527-69c5-4c6c-bd52-774cdb2dc9e3-kube-api-access-xmlzl\") on node \"crc\" DevicePath \"\"" Oct 14 09:05:29 crc kubenswrapper[4870]: I1014 09:05:29.987580 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-njft6" event={"ID":"e10cd527-69c5-4c6c-bd52-774cdb2dc9e3","Type":"ContainerDied","Data":"bf1725f612eb595b1eca002690633e77a34906f0050aee4048d917d7cb2e9c60"} Oct 14 09:05:29 crc kubenswrapper[4870]: I1014 09:05:29.987635 4870 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bf1725f612eb595b1eca002690633e77a34906f0050aee4048d917d7cb2e9c60" Oct 14 09:05:29 crc kubenswrapper[4870]: I1014 09:05:29.987716 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-njft6" Oct 14 09:05:35 crc kubenswrapper[4870]: I1014 09:05:35.403976 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-6fb9fc4bf-jmsxf" Oct 14 09:05:35 crc kubenswrapper[4870]: I1014 09:05:35.404639 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-6fb9fc4bf-jmsxf" Oct 14 09:05:36 crc kubenswrapper[4870]: I1014 09:05:36.034729 4870 scope.go:117] "RemoveContainer" containerID="8fcaeb3090f0c89cba1e66db524b4a246072daa35592012664aca450fa09ddc9" Oct 14 09:05:36 crc kubenswrapper[4870]: E1014 09:05:36.034957 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 09:05:36 crc kubenswrapper[4870]: I1014 09:05:36.274541 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-c54c-account-create-vgr7d"] Oct 14 09:05:36 crc kubenswrapper[4870]: E1014 09:05:36.275112 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e10cd527-69c5-4c6c-bd52-774cdb2dc9e3" containerName="mariadb-database-create" Oct 14 09:05:36 crc kubenswrapper[4870]: I1014 09:05:36.275142 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="e10cd527-69c5-4c6c-bd52-774cdb2dc9e3" containerName="mariadb-database-create" Oct 14 09:05:36 crc kubenswrapper[4870]: I1014 09:05:36.275529 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="e10cd527-69c5-4c6c-bd52-774cdb2dc9e3" containerName="mariadb-database-create" Oct 14 09:05:36 crc kubenswrapper[4870]: I1014 09:05:36.276628 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-c54c-account-create-vgr7d" Oct 14 09:05:36 crc kubenswrapper[4870]: I1014 09:05:36.279699 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-db-secret" Oct 14 09:05:36 crc kubenswrapper[4870]: I1014 09:05:36.294291 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-c54c-account-create-vgr7d"] Oct 14 09:05:36 crc kubenswrapper[4870]: I1014 09:05:36.433174 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gr5ht\" (UniqueName: \"kubernetes.io/projected/2c21a773-0454-49b9-8a63-5312e7a64826-kube-api-access-gr5ht\") pod \"heat-c54c-account-create-vgr7d\" (UID: \"2c21a773-0454-49b9-8a63-5312e7a64826\") " pod="openstack/heat-c54c-account-create-vgr7d" Oct 14 09:05:36 crc kubenswrapper[4870]: I1014 09:05:36.534653 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gr5ht\" (UniqueName: \"kubernetes.io/projected/2c21a773-0454-49b9-8a63-5312e7a64826-kube-api-access-gr5ht\") pod \"heat-c54c-account-create-vgr7d\" (UID: \"2c21a773-0454-49b9-8a63-5312e7a64826\") " pod="openstack/heat-c54c-account-create-vgr7d" Oct 14 09:05:36 crc kubenswrapper[4870]: I1014 09:05:36.552887 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gr5ht\" (UniqueName: \"kubernetes.io/projected/2c21a773-0454-49b9-8a63-5312e7a64826-kube-api-access-gr5ht\") pod \"heat-c54c-account-create-vgr7d\" (UID: \"2c21a773-0454-49b9-8a63-5312e7a64826\") " pod="openstack/heat-c54c-account-create-vgr7d" Oct 14 09:05:36 crc kubenswrapper[4870]: I1014 09:05:36.595185 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-c54c-account-create-vgr7d" Oct 14 09:05:37 crc kubenswrapper[4870]: I1014 09:05:37.158392 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-c54c-account-create-vgr7d"] Oct 14 09:05:38 crc kubenswrapper[4870]: I1014 09:05:38.087964 4870 generic.go:334] "Generic (PLEG): container finished" podID="2c21a773-0454-49b9-8a63-5312e7a64826" containerID="6bfdccb12a2aa548d1f24fc373dd026d9c8de828f33bf089aa12dc1b00379fe4" exitCode=0 Oct 14 09:05:38 crc kubenswrapper[4870]: I1014 09:05:38.088067 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-c54c-account-create-vgr7d" event={"ID":"2c21a773-0454-49b9-8a63-5312e7a64826","Type":"ContainerDied","Data":"6bfdccb12a2aa548d1f24fc373dd026d9c8de828f33bf089aa12dc1b00379fe4"} Oct 14 09:05:38 crc kubenswrapper[4870]: I1014 09:05:38.088341 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-c54c-account-create-vgr7d" event={"ID":"2c21a773-0454-49b9-8a63-5312e7a64826","Type":"ContainerStarted","Data":"cf345b36140edd000ccaaca2aa3fbcc144e8eca724c373365eff00b167969650"} Oct 14 09:05:39 crc kubenswrapper[4870]: I1014 09:05:39.055109 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-l2b5s"] Oct 14 09:05:39 crc kubenswrapper[4870]: I1014 09:05:39.068772 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-l2b5s"] Oct 14 09:05:39 crc kubenswrapper[4870]: I1014 09:05:39.460722 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-c54c-account-create-vgr7d" Oct 14 09:05:39 crc kubenswrapper[4870]: I1014 09:05:39.627473 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gr5ht\" (UniqueName: \"kubernetes.io/projected/2c21a773-0454-49b9-8a63-5312e7a64826-kube-api-access-gr5ht\") pod \"2c21a773-0454-49b9-8a63-5312e7a64826\" (UID: \"2c21a773-0454-49b9-8a63-5312e7a64826\") " Oct 14 09:05:39 crc kubenswrapper[4870]: I1014 09:05:39.635612 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2c21a773-0454-49b9-8a63-5312e7a64826-kube-api-access-gr5ht" (OuterVolumeSpecName: "kube-api-access-gr5ht") pod "2c21a773-0454-49b9-8a63-5312e7a64826" (UID: "2c21a773-0454-49b9-8a63-5312e7a64826"). InnerVolumeSpecName "kube-api-access-gr5ht". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:05:39 crc kubenswrapper[4870]: I1014 09:05:39.729911 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gr5ht\" (UniqueName: \"kubernetes.io/projected/2c21a773-0454-49b9-8a63-5312e7a64826-kube-api-access-gr5ht\") on node \"crc\" DevicePath \"\"" Oct 14 09:05:40 crc kubenswrapper[4870]: I1014 09:05:40.115816 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-c54c-account-create-vgr7d" event={"ID":"2c21a773-0454-49b9-8a63-5312e7a64826","Type":"ContainerDied","Data":"cf345b36140edd000ccaaca2aa3fbcc144e8eca724c373365eff00b167969650"} Oct 14 09:05:40 crc kubenswrapper[4870]: I1014 09:05:40.116250 4870 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cf345b36140edd000ccaaca2aa3fbcc144e8eca724c373365eff00b167969650" Oct 14 09:05:40 crc kubenswrapper[4870]: I1014 09:05:40.115923 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-c54c-account-create-vgr7d" Oct 14 09:05:41 crc kubenswrapper[4870]: I1014 09:05:41.050101 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="76690419-68ac-42b3-a6a0-3fb21543786a" path="/var/lib/kubelet/pods/76690419-68ac-42b3-a6a0-3fb21543786a/volumes" Oct 14 09:05:41 crc kubenswrapper[4870]: I1014 09:05:41.326489 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-db-sync-lrs97"] Oct 14 09:05:41 crc kubenswrapper[4870]: E1014 09:05:41.326879 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c21a773-0454-49b9-8a63-5312e7a64826" containerName="mariadb-account-create" Oct 14 09:05:41 crc kubenswrapper[4870]: I1014 09:05:41.326899 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c21a773-0454-49b9-8a63-5312e7a64826" containerName="mariadb-account-create" Oct 14 09:05:41 crc kubenswrapper[4870]: I1014 09:05:41.327114 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="2c21a773-0454-49b9-8a63-5312e7a64826" containerName="mariadb-account-create" Oct 14 09:05:41 crc kubenswrapper[4870]: I1014 09:05:41.327827 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-lrs97" Oct 14 09:05:41 crc kubenswrapper[4870]: I1014 09:05:41.330201 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-config-data" Oct 14 09:05:41 crc kubenswrapper[4870]: I1014 09:05:41.330988 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-heat-dockercfg-4n6jj" Oct 14 09:05:41 crc kubenswrapper[4870]: I1014 09:05:41.351754 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-sync-lrs97"] Oct 14 09:05:41 crc kubenswrapper[4870]: I1014 09:05:41.471522 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ddf1fd10-61d6-41a6-a261-5fa5388602bb-config-data\") pod \"heat-db-sync-lrs97\" (UID: \"ddf1fd10-61d6-41a6-a261-5fa5388602bb\") " pod="openstack/heat-db-sync-lrs97" Oct 14 09:05:41 crc kubenswrapper[4870]: I1014 09:05:41.471929 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ddf1fd10-61d6-41a6-a261-5fa5388602bb-combined-ca-bundle\") pod \"heat-db-sync-lrs97\" (UID: \"ddf1fd10-61d6-41a6-a261-5fa5388602bb\") " pod="openstack/heat-db-sync-lrs97" Oct 14 09:05:41 crc kubenswrapper[4870]: I1014 09:05:41.472016 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-84js8\" (UniqueName: \"kubernetes.io/projected/ddf1fd10-61d6-41a6-a261-5fa5388602bb-kube-api-access-84js8\") pod \"heat-db-sync-lrs97\" (UID: \"ddf1fd10-61d6-41a6-a261-5fa5388602bb\") " pod="openstack/heat-db-sync-lrs97" Oct 14 09:05:41 crc kubenswrapper[4870]: I1014 09:05:41.574269 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ddf1fd10-61d6-41a6-a261-5fa5388602bb-config-data\") pod \"heat-db-sync-lrs97\" (UID: \"ddf1fd10-61d6-41a6-a261-5fa5388602bb\") " pod="openstack/heat-db-sync-lrs97" Oct 14 09:05:41 crc kubenswrapper[4870]: I1014 09:05:41.574384 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ddf1fd10-61d6-41a6-a261-5fa5388602bb-combined-ca-bundle\") pod \"heat-db-sync-lrs97\" (UID: \"ddf1fd10-61d6-41a6-a261-5fa5388602bb\") " pod="openstack/heat-db-sync-lrs97" Oct 14 09:05:41 crc kubenswrapper[4870]: I1014 09:05:41.574516 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-84js8\" (UniqueName: \"kubernetes.io/projected/ddf1fd10-61d6-41a6-a261-5fa5388602bb-kube-api-access-84js8\") pod \"heat-db-sync-lrs97\" (UID: \"ddf1fd10-61d6-41a6-a261-5fa5388602bb\") " pod="openstack/heat-db-sync-lrs97" Oct 14 09:05:41 crc kubenswrapper[4870]: I1014 09:05:41.584665 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ddf1fd10-61d6-41a6-a261-5fa5388602bb-config-data\") pod \"heat-db-sync-lrs97\" (UID: \"ddf1fd10-61d6-41a6-a261-5fa5388602bb\") " pod="openstack/heat-db-sync-lrs97" Oct 14 09:05:41 crc kubenswrapper[4870]: I1014 09:05:41.591115 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ddf1fd10-61d6-41a6-a261-5fa5388602bb-combined-ca-bundle\") pod \"heat-db-sync-lrs97\" (UID: \"ddf1fd10-61d6-41a6-a261-5fa5388602bb\") " pod="openstack/heat-db-sync-lrs97" Oct 14 09:05:41 crc kubenswrapper[4870]: I1014 09:05:41.595879 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-84js8\" (UniqueName: \"kubernetes.io/projected/ddf1fd10-61d6-41a6-a261-5fa5388602bb-kube-api-access-84js8\") pod \"heat-db-sync-lrs97\" (UID: \"ddf1fd10-61d6-41a6-a261-5fa5388602bb\") " pod="openstack/heat-db-sync-lrs97" Oct 14 09:05:41 crc kubenswrapper[4870]: I1014 09:05:41.669681 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-lrs97" Oct 14 09:05:42 crc kubenswrapper[4870]: I1014 09:05:42.131553 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-sync-lrs97"] Oct 14 09:05:43 crc kubenswrapper[4870]: I1014 09:05:43.169751 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-lrs97" event={"ID":"ddf1fd10-61d6-41a6-a261-5fa5388602bb","Type":"ContainerStarted","Data":"5b143e9e9fc5ebdcbba35aceb4dcf663bbbdd480f7987072cadf18e65c0cad67"} Oct 14 09:05:45 crc kubenswrapper[4870]: I1014 09:05:45.405805 4870 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-6fb9fc4bf-jmsxf" podUID="73bab307-e3b8-47df-bba3-fa624f4c4261" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.111:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.111:8080: connect: connection refused" Oct 14 09:05:47 crc kubenswrapper[4870]: I1014 09:05:47.036584 4870 scope.go:117] "RemoveContainer" containerID="8fcaeb3090f0c89cba1e66db524b4a246072daa35592012664aca450fa09ddc9" Oct 14 09:05:47 crc kubenswrapper[4870]: E1014 09:05:47.036985 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 09:05:49 crc kubenswrapper[4870]: I1014 09:05:49.046317 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-750f-account-create-wg5jr"] Oct 14 09:05:49 crc kubenswrapper[4870]: I1014 09:05:49.050123 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-750f-account-create-wg5jr"] Oct 14 09:05:50 crc kubenswrapper[4870]: I1014 09:05:50.233931 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-lrs97" event={"ID":"ddf1fd10-61d6-41a6-a261-5fa5388602bb","Type":"ContainerStarted","Data":"1b75db4079b0f6952970dc908c691c7371d927a7b96d26ad3cdd316012caef08"} Oct 14 09:05:50 crc kubenswrapper[4870]: I1014 09:05:50.256875 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-db-sync-lrs97" podStartSLOduration=1.5505522859999998 podStartE2EDuration="9.256843579s" podCreationTimestamp="2025-10-14 09:05:41 +0000 UTC" firstStartedPulling="2025-10-14 09:05:42.140253065 +0000 UTC m=+7477.837613456" lastFinishedPulling="2025-10-14 09:05:49.846544378 +0000 UTC m=+7485.543904749" observedRunningTime="2025-10-14 09:05:50.248412591 +0000 UTC m=+7485.945773002" watchObservedRunningTime="2025-10-14 09:05:50.256843579 +0000 UTC m=+7485.954203990" Oct 14 09:05:51 crc kubenswrapper[4870]: I1014 09:05:51.056530 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fd63d29-5666-442a-82b5-1e054c10b29c" path="/var/lib/kubelet/pods/5fd63d29-5666-442a-82b5-1e054c10b29c/volumes" Oct 14 09:05:53 crc kubenswrapper[4870]: I1014 09:05:53.275729 4870 generic.go:334] "Generic (PLEG): container finished" podID="ddf1fd10-61d6-41a6-a261-5fa5388602bb" containerID="1b75db4079b0f6952970dc908c691c7371d927a7b96d26ad3cdd316012caef08" exitCode=0 Oct 14 09:05:53 crc kubenswrapper[4870]: I1014 09:05:53.275821 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-lrs97" event={"ID":"ddf1fd10-61d6-41a6-a261-5fa5388602bb","Type":"ContainerDied","Data":"1b75db4079b0f6952970dc908c691c7371d927a7b96d26ad3cdd316012caef08"} Oct 14 09:05:54 crc kubenswrapper[4870]: I1014 09:05:54.730939 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-lrs97" Oct 14 09:05:54 crc kubenswrapper[4870]: I1014 09:05:54.855726 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-84js8\" (UniqueName: \"kubernetes.io/projected/ddf1fd10-61d6-41a6-a261-5fa5388602bb-kube-api-access-84js8\") pod \"ddf1fd10-61d6-41a6-a261-5fa5388602bb\" (UID: \"ddf1fd10-61d6-41a6-a261-5fa5388602bb\") " Oct 14 09:05:54 crc kubenswrapper[4870]: I1014 09:05:54.855779 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ddf1fd10-61d6-41a6-a261-5fa5388602bb-config-data\") pod \"ddf1fd10-61d6-41a6-a261-5fa5388602bb\" (UID: \"ddf1fd10-61d6-41a6-a261-5fa5388602bb\") " Oct 14 09:05:54 crc kubenswrapper[4870]: I1014 09:05:54.855857 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ddf1fd10-61d6-41a6-a261-5fa5388602bb-combined-ca-bundle\") pod \"ddf1fd10-61d6-41a6-a261-5fa5388602bb\" (UID: \"ddf1fd10-61d6-41a6-a261-5fa5388602bb\") " Oct 14 09:05:54 crc kubenswrapper[4870]: I1014 09:05:54.863998 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ddf1fd10-61d6-41a6-a261-5fa5388602bb-kube-api-access-84js8" (OuterVolumeSpecName: "kube-api-access-84js8") pod "ddf1fd10-61d6-41a6-a261-5fa5388602bb" (UID: "ddf1fd10-61d6-41a6-a261-5fa5388602bb"). InnerVolumeSpecName "kube-api-access-84js8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:05:54 crc kubenswrapper[4870]: I1014 09:05:54.891033 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ddf1fd10-61d6-41a6-a261-5fa5388602bb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ddf1fd10-61d6-41a6-a261-5fa5388602bb" (UID: "ddf1fd10-61d6-41a6-a261-5fa5388602bb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:05:54 crc kubenswrapper[4870]: I1014 09:05:54.968190 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-84js8\" (UniqueName: \"kubernetes.io/projected/ddf1fd10-61d6-41a6-a261-5fa5388602bb-kube-api-access-84js8\") on node \"crc\" DevicePath \"\"" Oct 14 09:05:54 crc kubenswrapper[4870]: I1014 09:05:54.968253 4870 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ddf1fd10-61d6-41a6-a261-5fa5388602bb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 09:05:55 crc kubenswrapper[4870]: I1014 09:05:55.014550 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ddf1fd10-61d6-41a6-a261-5fa5388602bb-config-data" (OuterVolumeSpecName: "config-data") pod "ddf1fd10-61d6-41a6-a261-5fa5388602bb" (UID: "ddf1fd10-61d6-41a6-a261-5fa5388602bb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:05:55 crc kubenswrapper[4870]: I1014 09:05:55.069736 4870 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ddf1fd10-61d6-41a6-a261-5fa5388602bb-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 09:05:55 crc kubenswrapper[4870]: I1014 09:05:55.304176 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-lrs97" event={"ID":"ddf1fd10-61d6-41a6-a261-5fa5388602bb","Type":"ContainerDied","Data":"5b143e9e9fc5ebdcbba35aceb4dcf663bbbdd480f7987072cadf18e65c0cad67"} Oct 14 09:05:55 crc kubenswrapper[4870]: I1014 09:05:55.304228 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-lrs97" Oct 14 09:05:55 crc kubenswrapper[4870]: I1014 09:05:55.304232 4870 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5b143e9e9fc5ebdcbba35aceb4dcf663bbbdd480f7987072cadf18e65c0cad67" Oct 14 09:05:56 crc kubenswrapper[4870]: I1014 09:05:56.723539 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-engine-6bb7df6545-4x5wj"] Oct 14 09:05:56 crc kubenswrapper[4870]: E1014 09:05:56.724518 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ddf1fd10-61d6-41a6-a261-5fa5388602bb" containerName="heat-db-sync" Oct 14 09:05:56 crc kubenswrapper[4870]: I1014 09:05:56.724545 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="ddf1fd10-61d6-41a6-a261-5fa5388602bb" containerName="heat-db-sync" Oct 14 09:05:56 crc kubenswrapper[4870]: I1014 09:05:56.724884 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="ddf1fd10-61d6-41a6-a261-5fa5388602bb" containerName="heat-db-sync" Oct 14 09:05:56 crc kubenswrapper[4870]: I1014 09:05:56.725921 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-6bb7df6545-4x5wj" Oct 14 09:05:56 crc kubenswrapper[4870]: I1014 09:05:56.731188 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-config-data" Oct 14 09:05:56 crc kubenswrapper[4870]: I1014 09:05:56.731475 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-engine-config-data" Oct 14 09:05:56 crc kubenswrapper[4870]: I1014 09:05:56.731621 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-heat-dockercfg-4n6jj" Oct 14 09:05:56 crc kubenswrapper[4870]: I1014 09:05:56.737016 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-6bb7df6545-4x5wj"] Oct 14 09:05:56 crc kubenswrapper[4870]: I1014 09:05:56.810063 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3ff05517-1df4-4a38-ad5c-c293546a56e3-config-data\") pod \"heat-engine-6bb7df6545-4x5wj\" (UID: \"3ff05517-1df4-4a38-ad5c-c293546a56e3\") " pod="openstack/heat-engine-6bb7df6545-4x5wj" Oct 14 09:05:56 crc kubenswrapper[4870]: I1014 09:05:56.810175 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ff05517-1df4-4a38-ad5c-c293546a56e3-combined-ca-bundle\") pod \"heat-engine-6bb7df6545-4x5wj\" (UID: \"3ff05517-1df4-4a38-ad5c-c293546a56e3\") " pod="openstack/heat-engine-6bb7df6545-4x5wj" Oct 14 09:05:56 crc kubenswrapper[4870]: I1014 09:05:56.810226 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3ff05517-1df4-4a38-ad5c-c293546a56e3-config-data-custom\") pod \"heat-engine-6bb7df6545-4x5wj\" (UID: \"3ff05517-1df4-4a38-ad5c-c293546a56e3\") " pod="openstack/heat-engine-6bb7df6545-4x5wj" Oct 14 09:05:56 crc kubenswrapper[4870]: I1014 09:05:56.810270 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2p99b\" (UniqueName: \"kubernetes.io/projected/3ff05517-1df4-4a38-ad5c-c293546a56e3-kube-api-access-2p99b\") pod \"heat-engine-6bb7df6545-4x5wj\" (UID: \"3ff05517-1df4-4a38-ad5c-c293546a56e3\") " pod="openstack/heat-engine-6bb7df6545-4x5wj" Oct 14 09:05:56 crc kubenswrapper[4870]: I1014 09:05:56.821080 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-api-f4c547d9b-gjq77"] Oct 14 09:05:56 crc kubenswrapper[4870]: I1014 09:05:56.822737 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-f4c547d9b-gjq77" Oct 14 09:05:56 crc kubenswrapper[4870]: I1014 09:05:56.827277 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-api-config-data" Oct 14 09:05:56 crc kubenswrapper[4870]: I1014 09:05:56.833805 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-f4c547d9b-gjq77"] Oct 14 09:05:56 crc kubenswrapper[4870]: I1014 09:05:56.907714 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-cfnapi-77b878966d-fm2hf"] Oct 14 09:05:56 crc kubenswrapper[4870]: I1014 09:05:56.909373 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-77b878966d-fm2hf" Oct 14 09:05:56 crc kubenswrapper[4870]: I1014 09:05:56.911843 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e232c83-c81a-4c64-ad90-74b771732688-combined-ca-bundle\") pod \"heat-api-f4c547d9b-gjq77\" (UID: \"0e232c83-c81a-4c64-ad90-74b771732688\") " pod="openstack/heat-api-f4c547d9b-gjq77" Oct 14 09:05:56 crc kubenswrapper[4870]: I1014 09:05:56.911918 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3ff05517-1df4-4a38-ad5c-c293546a56e3-config-data-custom\") pod \"heat-engine-6bb7df6545-4x5wj\" (UID: \"3ff05517-1df4-4a38-ad5c-c293546a56e3\") " pod="openstack/heat-engine-6bb7df6545-4x5wj" Oct 14 09:05:56 crc kubenswrapper[4870]: I1014 09:05:56.911953 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2p99b\" (UniqueName: \"kubernetes.io/projected/3ff05517-1df4-4a38-ad5c-c293546a56e3-kube-api-access-2p99b\") pod \"heat-engine-6bb7df6545-4x5wj\" (UID: \"3ff05517-1df4-4a38-ad5c-c293546a56e3\") " pod="openstack/heat-engine-6bb7df6545-4x5wj" Oct 14 09:05:56 crc kubenswrapper[4870]: I1014 09:05:56.912009 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ptptt\" (UniqueName: \"kubernetes.io/projected/0e232c83-c81a-4c64-ad90-74b771732688-kube-api-access-ptptt\") pod \"heat-api-f4c547d9b-gjq77\" (UID: \"0e232c83-c81a-4c64-ad90-74b771732688\") " pod="openstack/heat-api-f4c547d9b-gjq77" Oct 14 09:05:56 crc kubenswrapper[4870]: I1014 09:05:56.912028 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0e232c83-c81a-4c64-ad90-74b771732688-config-data-custom\") pod \"heat-api-f4c547d9b-gjq77\" (UID: \"0e232c83-c81a-4c64-ad90-74b771732688\") " pod="openstack/heat-api-f4c547d9b-gjq77" Oct 14 09:05:56 crc kubenswrapper[4870]: I1014 09:05:56.912069 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3ff05517-1df4-4a38-ad5c-c293546a56e3-config-data\") pod \"heat-engine-6bb7df6545-4x5wj\" (UID: \"3ff05517-1df4-4a38-ad5c-c293546a56e3\") " pod="openstack/heat-engine-6bb7df6545-4x5wj" Oct 14 09:05:56 crc kubenswrapper[4870]: I1014 09:05:56.912097 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0e232c83-c81a-4c64-ad90-74b771732688-config-data\") pod \"heat-api-f4c547d9b-gjq77\" (UID: \"0e232c83-c81a-4c64-ad90-74b771732688\") " pod="openstack/heat-api-f4c547d9b-gjq77" Oct 14 09:05:56 crc kubenswrapper[4870]: I1014 09:05:56.912130 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ff05517-1df4-4a38-ad5c-c293546a56e3-combined-ca-bundle\") pod \"heat-engine-6bb7df6545-4x5wj\" (UID: \"3ff05517-1df4-4a38-ad5c-c293546a56e3\") " pod="openstack/heat-engine-6bb7df6545-4x5wj" Oct 14 09:05:56 crc kubenswrapper[4870]: I1014 09:05:56.915760 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-cfnapi-config-data" Oct 14 09:05:56 crc kubenswrapper[4870]: I1014 09:05:56.918489 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3ff05517-1df4-4a38-ad5c-c293546a56e3-config-data-custom\") pod \"heat-engine-6bb7df6545-4x5wj\" (UID: \"3ff05517-1df4-4a38-ad5c-c293546a56e3\") " pod="openstack/heat-engine-6bb7df6545-4x5wj" Oct 14 09:05:56 crc kubenswrapper[4870]: I1014 09:05:56.918786 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ff05517-1df4-4a38-ad5c-c293546a56e3-combined-ca-bundle\") pod \"heat-engine-6bb7df6545-4x5wj\" (UID: \"3ff05517-1df4-4a38-ad5c-c293546a56e3\") " pod="openstack/heat-engine-6bb7df6545-4x5wj" Oct 14 09:05:56 crc kubenswrapper[4870]: I1014 09:05:56.919699 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3ff05517-1df4-4a38-ad5c-c293546a56e3-config-data\") pod \"heat-engine-6bb7df6545-4x5wj\" (UID: \"3ff05517-1df4-4a38-ad5c-c293546a56e3\") " pod="openstack/heat-engine-6bb7df6545-4x5wj" Oct 14 09:05:56 crc kubenswrapper[4870]: I1014 09:05:56.921939 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-77b878966d-fm2hf"] Oct 14 09:05:56 crc kubenswrapper[4870]: I1014 09:05:56.937086 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2p99b\" (UniqueName: \"kubernetes.io/projected/3ff05517-1df4-4a38-ad5c-c293546a56e3-kube-api-access-2p99b\") pod \"heat-engine-6bb7df6545-4x5wj\" (UID: \"3ff05517-1df4-4a38-ad5c-c293546a56e3\") " pod="openstack/heat-engine-6bb7df6545-4x5wj" Oct 14 09:05:57 crc kubenswrapper[4870]: I1014 09:05:57.013652 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0e232c83-c81a-4c64-ad90-74b771732688-config-data\") pod \"heat-api-f4c547d9b-gjq77\" (UID: \"0e232c83-c81a-4c64-ad90-74b771732688\") " pod="openstack/heat-api-f4c547d9b-gjq77" Oct 14 09:05:57 crc kubenswrapper[4870]: I1014 09:05:57.013736 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e232c83-c81a-4c64-ad90-74b771732688-combined-ca-bundle\") pod \"heat-api-f4c547d9b-gjq77\" (UID: \"0e232c83-c81a-4c64-ad90-74b771732688\") " pod="openstack/heat-api-f4c547d9b-gjq77" Oct 14 09:05:57 crc kubenswrapper[4870]: I1014 09:05:57.013792 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1aa3b494-7370-4251-afd5-01109e40cc85-combined-ca-bundle\") pod \"heat-cfnapi-77b878966d-fm2hf\" (UID: \"1aa3b494-7370-4251-afd5-01109e40cc85\") " pod="openstack/heat-cfnapi-77b878966d-fm2hf" Oct 14 09:05:57 crc kubenswrapper[4870]: I1014 09:05:57.013820 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1aa3b494-7370-4251-afd5-01109e40cc85-config-data-custom\") pod \"heat-cfnapi-77b878966d-fm2hf\" (UID: \"1aa3b494-7370-4251-afd5-01109e40cc85\") " pod="openstack/heat-cfnapi-77b878966d-fm2hf" Oct 14 09:05:57 crc kubenswrapper[4870]: I1014 09:05:57.013838 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ctnpg\" (UniqueName: \"kubernetes.io/projected/1aa3b494-7370-4251-afd5-01109e40cc85-kube-api-access-ctnpg\") pod \"heat-cfnapi-77b878966d-fm2hf\" (UID: \"1aa3b494-7370-4251-afd5-01109e40cc85\") " pod="openstack/heat-cfnapi-77b878966d-fm2hf" Oct 14 09:05:57 crc kubenswrapper[4870]: I1014 09:05:57.013878 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ptptt\" (UniqueName: \"kubernetes.io/projected/0e232c83-c81a-4c64-ad90-74b771732688-kube-api-access-ptptt\") pod \"heat-api-f4c547d9b-gjq77\" (UID: \"0e232c83-c81a-4c64-ad90-74b771732688\") " pod="openstack/heat-api-f4c547d9b-gjq77" Oct 14 09:05:57 crc kubenswrapper[4870]: I1014 09:05:57.013903 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0e232c83-c81a-4c64-ad90-74b771732688-config-data-custom\") pod \"heat-api-f4c547d9b-gjq77\" (UID: \"0e232c83-c81a-4c64-ad90-74b771732688\") " pod="openstack/heat-api-f4c547d9b-gjq77" Oct 14 09:05:57 crc kubenswrapper[4870]: I1014 09:05:57.013951 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1aa3b494-7370-4251-afd5-01109e40cc85-config-data\") pod \"heat-cfnapi-77b878966d-fm2hf\" (UID: \"1aa3b494-7370-4251-afd5-01109e40cc85\") " pod="openstack/heat-cfnapi-77b878966d-fm2hf" Oct 14 09:05:57 crc kubenswrapper[4870]: I1014 09:05:57.020644 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0e232c83-c81a-4c64-ad90-74b771732688-config-data\") pod \"heat-api-f4c547d9b-gjq77\" (UID: \"0e232c83-c81a-4c64-ad90-74b771732688\") " pod="openstack/heat-api-f4c547d9b-gjq77" Oct 14 09:05:57 crc kubenswrapper[4870]: I1014 09:05:57.024387 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e232c83-c81a-4c64-ad90-74b771732688-combined-ca-bundle\") pod \"heat-api-f4c547d9b-gjq77\" (UID: \"0e232c83-c81a-4c64-ad90-74b771732688\") " pod="openstack/heat-api-f4c547d9b-gjq77" Oct 14 09:05:57 crc kubenswrapper[4870]: I1014 09:05:57.029787 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0e232c83-c81a-4c64-ad90-74b771732688-config-data-custom\") pod \"heat-api-f4c547d9b-gjq77\" (UID: \"0e232c83-c81a-4c64-ad90-74b771732688\") " pod="openstack/heat-api-f4c547d9b-gjq77" Oct 14 09:05:57 crc kubenswrapper[4870]: I1014 09:05:57.043262 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ptptt\" (UniqueName: \"kubernetes.io/projected/0e232c83-c81a-4c64-ad90-74b771732688-kube-api-access-ptptt\") pod \"heat-api-f4c547d9b-gjq77\" (UID: \"0e232c83-c81a-4c64-ad90-74b771732688\") " pod="openstack/heat-api-f4c547d9b-gjq77" Oct 14 09:05:57 crc kubenswrapper[4870]: I1014 09:05:57.054104 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-6bb7df6545-4x5wj" Oct 14 09:05:57 crc kubenswrapper[4870]: I1014 09:05:57.056529 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-qfmh6"] Oct 14 09:05:57 crc kubenswrapper[4870]: I1014 09:05:57.067192 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-qfmh6"] Oct 14 09:05:57 crc kubenswrapper[4870]: I1014 09:05:57.115934 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1aa3b494-7370-4251-afd5-01109e40cc85-combined-ca-bundle\") pod \"heat-cfnapi-77b878966d-fm2hf\" (UID: \"1aa3b494-7370-4251-afd5-01109e40cc85\") " pod="openstack/heat-cfnapi-77b878966d-fm2hf" Oct 14 09:05:57 crc kubenswrapper[4870]: I1014 09:05:57.116058 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1aa3b494-7370-4251-afd5-01109e40cc85-config-data-custom\") pod \"heat-cfnapi-77b878966d-fm2hf\" (UID: \"1aa3b494-7370-4251-afd5-01109e40cc85\") " pod="openstack/heat-cfnapi-77b878966d-fm2hf" Oct 14 09:05:57 crc kubenswrapper[4870]: I1014 09:05:57.116090 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ctnpg\" (UniqueName: \"kubernetes.io/projected/1aa3b494-7370-4251-afd5-01109e40cc85-kube-api-access-ctnpg\") pod \"heat-cfnapi-77b878966d-fm2hf\" (UID: \"1aa3b494-7370-4251-afd5-01109e40cc85\") " pod="openstack/heat-cfnapi-77b878966d-fm2hf" Oct 14 09:05:57 crc kubenswrapper[4870]: I1014 09:05:57.116221 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1aa3b494-7370-4251-afd5-01109e40cc85-config-data\") pod \"heat-cfnapi-77b878966d-fm2hf\" (UID: \"1aa3b494-7370-4251-afd5-01109e40cc85\") " pod="openstack/heat-cfnapi-77b878966d-fm2hf" Oct 14 09:05:57 crc kubenswrapper[4870]: I1014 09:05:57.121998 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1aa3b494-7370-4251-afd5-01109e40cc85-combined-ca-bundle\") pod \"heat-cfnapi-77b878966d-fm2hf\" (UID: \"1aa3b494-7370-4251-afd5-01109e40cc85\") " pod="openstack/heat-cfnapi-77b878966d-fm2hf" Oct 14 09:05:57 crc kubenswrapper[4870]: I1014 09:05:57.125454 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1aa3b494-7370-4251-afd5-01109e40cc85-config-data\") pod \"heat-cfnapi-77b878966d-fm2hf\" (UID: \"1aa3b494-7370-4251-afd5-01109e40cc85\") " pod="openstack/heat-cfnapi-77b878966d-fm2hf" Oct 14 09:05:57 crc kubenswrapper[4870]: I1014 09:05:57.132232 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1aa3b494-7370-4251-afd5-01109e40cc85-config-data-custom\") pod \"heat-cfnapi-77b878966d-fm2hf\" (UID: \"1aa3b494-7370-4251-afd5-01109e40cc85\") " pod="openstack/heat-cfnapi-77b878966d-fm2hf" Oct 14 09:05:57 crc kubenswrapper[4870]: I1014 09:05:57.134627 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ctnpg\" (UniqueName: \"kubernetes.io/projected/1aa3b494-7370-4251-afd5-01109e40cc85-kube-api-access-ctnpg\") pod \"heat-cfnapi-77b878966d-fm2hf\" (UID: \"1aa3b494-7370-4251-afd5-01109e40cc85\") " pod="openstack/heat-cfnapi-77b878966d-fm2hf" Oct 14 09:05:57 crc kubenswrapper[4870]: I1014 09:05:57.153344 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-f4c547d9b-gjq77" Oct 14 09:05:57 crc kubenswrapper[4870]: I1014 09:05:57.162104 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-77b878966d-fm2hf" Oct 14 09:05:57 crc kubenswrapper[4870]: I1014 09:05:57.206777 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-6fb9fc4bf-jmsxf" Oct 14 09:05:57 crc kubenswrapper[4870]: I1014 09:05:57.533136 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-6bb7df6545-4x5wj"] Oct 14 09:05:57 crc kubenswrapper[4870]: I1014 09:05:57.710053 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-f4c547d9b-gjq77"] Oct 14 09:05:57 crc kubenswrapper[4870]: I1014 09:05:57.717045 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-77b878966d-fm2hf"] Oct 14 09:05:57 crc kubenswrapper[4870]: W1014 09:05:57.721799 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1aa3b494_7370_4251_afd5_01109e40cc85.slice/crio-42b888c347272a6673a7d4df3eb3811eb27441ecb809809cc5c376f178e5b26b WatchSource:0}: Error finding container 42b888c347272a6673a7d4df3eb3811eb27441ecb809809cc5c376f178e5b26b: Status 404 returned error can't find the container with id 42b888c347272a6673a7d4df3eb3811eb27441ecb809809cc5c376f178e5b26b Oct 14 09:05:58 crc kubenswrapper[4870]: I1014 09:05:58.347124 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-f4c547d9b-gjq77" event={"ID":"0e232c83-c81a-4c64-ad90-74b771732688","Type":"ContainerStarted","Data":"7c4d331001c91b48cb788a6077e350718e841ee81496c17d2cdab6092a42ddf8"} Oct 14 09:05:58 crc kubenswrapper[4870]: I1014 09:05:58.349351 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-6bb7df6545-4x5wj" event={"ID":"3ff05517-1df4-4a38-ad5c-c293546a56e3","Type":"ContainerStarted","Data":"8e64fadc9d5a7dada01d6954a6884e75b42143c15b8bc239b163fb77338e7a9d"} Oct 14 09:05:58 crc kubenswrapper[4870]: I1014 09:05:58.349400 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-6bb7df6545-4x5wj" event={"ID":"3ff05517-1df4-4a38-ad5c-c293546a56e3","Type":"ContainerStarted","Data":"9c0f79cef4dea469d15e6583b2240978cf58ad325c1b149e3706ad6306c802f3"} Oct 14 09:05:58 crc kubenswrapper[4870]: I1014 09:05:58.349435 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-engine-6bb7df6545-4x5wj" Oct 14 09:05:58 crc kubenswrapper[4870]: I1014 09:05:58.354933 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-77b878966d-fm2hf" event={"ID":"1aa3b494-7370-4251-afd5-01109e40cc85","Type":"ContainerStarted","Data":"42b888c347272a6673a7d4df3eb3811eb27441ecb809809cc5c376f178e5b26b"} Oct 14 09:05:58 crc kubenswrapper[4870]: I1014 09:05:58.369790 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-engine-6bb7df6545-4x5wj" podStartSLOduration=2.369771993 podStartE2EDuration="2.369771993s" podCreationTimestamp="2025-10-14 09:05:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 09:05:58.363061918 +0000 UTC m=+7494.060422289" watchObservedRunningTime="2025-10-14 09:05:58.369771993 +0000 UTC m=+7494.067132364" Oct 14 09:05:58 crc kubenswrapper[4870]: I1014 09:05:58.929213 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-6fb9fc4bf-jmsxf" Oct 14 09:05:58 crc kubenswrapper[4870]: I1014 09:05:58.989497 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7cff98d7df-2fx9m"] Oct 14 09:05:58 crc kubenswrapper[4870]: I1014 09:05:58.989729 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-7cff98d7df-2fx9m" podUID="7ad77931-c302-4ce8-81f0-78767bb80443" containerName="horizon-log" containerID="cri-o://4a68c4125fa0d27922a302bd38cd7ca956c788c5062a558fbf2fe6050cd6ae9a" gracePeriod=30 Oct 14 09:05:58 crc kubenswrapper[4870]: I1014 09:05:58.990114 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-7cff98d7df-2fx9m" podUID="7ad77931-c302-4ce8-81f0-78767bb80443" containerName="horizon" containerID="cri-o://d898a5c61cd9c202ebe91686a9d143d06fbb01cb5ba6341691ff053cbe036f0c" gracePeriod=30 Oct 14 09:05:59 crc kubenswrapper[4870]: I1014 09:05:59.056426 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8d51a0c5-822f-47d5-af81-b812ac44fe67" path="/var/lib/kubelet/pods/8d51a0c5-822f-47d5-af81-b812ac44fe67/volumes" Oct 14 09:06:00 crc kubenswrapper[4870]: I1014 09:06:00.033800 4870 scope.go:117] "RemoveContainer" containerID="8fcaeb3090f0c89cba1e66db524b4a246072daa35592012664aca450fa09ddc9" Oct 14 09:06:00 crc kubenswrapper[4870]: E1014 09:06:00.034377 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 09:06:00 crc kubenswrapper[4870]: I1014 09:06:00.371480 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-77b878966d-fm2hf" event={"ID":"1aa3b494-7370-4251-afd5-01109e40cc85","Type":"ContainerStarted","Data":"9d1a057b505c0b4c1b33ee89c245f6c2a750af18010ab179747346479cc35e9b"} Oct 14 09:06:00 crc kubenswrapper[4870]: I1014 09:06:00.372718 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-cfnapi-77b878966d-fm2hf" Oct 14 09:06:00 crc kubenswrapper[4870]: I1014 09:06:00.376944 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-f4c547d9b-gjq77" event={"ID":"0e232c83-c81a-4c64-ad90-74b771732688","Type":"ContainerStarted","Data":"bc70e0809b09a8c41ec2b93145496460a5e21aeb4f29001fd63169c9977b4c17"} Oct 14 09:06:00 crc kubenswrapper[4870]: I1014 09:06:00.377078 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-f4c547d9b-gjq77" Oct 14 09:06:00 crc kubenswrapper[4870]: I1014 09:06:00.390902 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-cfnapi-77b878966d-fm2hf" podStartSLOduration=2.750545255 podStartE2EDuration="4.390879988s" podCreationTimestamp="2025-10-14 09:05:56 +0000 UTC" firstStartedPulling="2025-10-14 09:05:57.726362746 +0000 UTC m=+7493.423723117" lastFinishedPulling="2025-10-14 09:05:59.366697479 +0000 UTC m=+7495.064057850" observedRunningTime="2025-10-14 09:06:00.388207172 +0000 UTC m=+7496.085567553" watchObservedRunningTime="2025-10-14 09:06:00.390879988 +0000 UTC m=+7496.088240359" Oct 14 09:06:02 crc kubenswrapper[4870]: I1014 09:06:02.399036 4870 generic.go:334] "Generic (PLEG): container finished" podID="7ad77931-c302-4ce8-81f0-78767bb80443" containerID="d898a5c61cd9c202ebe91686a9d143d06fbb01cb5ba6341691ff053cbe036f0c" exitCode=0 Oct 14 09:06:02 crc kubenswrapper[4870]: I1014 09:06:02.400087 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7cff98d7df-2fx9m" event={"ID":"7ad77931-c302-4ce8-81f0-78767bb80443","Type":"ContainerDied","Data":"d898a5c61cd9c202ebe91686a9d143d06fbb01cb5ba6341691ff053cbe036f0c"} Oct 14 09:06:07 crc kubenswrapper[4870]: I1014 09:06:07.086961 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-engine-6bb7df6545-4x5wj" Oct 14 09:06:07 crc kubenswrapper[4870]: I1014 09:06:07.104690 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-api-f4c547d9b-gjq77" podStartSLOduration=9.452463768 podStartE2EDuration="11.104670504s" podCreationTimestamp="2025-10-14 09:05:56 +0000 UTC" firstStartedPulling="2025-10-14 09:05:57.716882082 +0000 UTC m=+7493.414242453" lastFinishedPulling="2025-10-14 09:05:59.369088818 +0000 UTC m=+7495.066449189" observedRunningTime="2025-10-14 09:06:00.401182742 +0000 UTC m=+7496.098543113" watchObservedRunningTime="2025-10-14 09:06:07.104670504 +0000 UTC m=+7502.802030875" Oct 14 09:06:08 crc kubenswrapper[4870]: I1014 09:06:08.498992 4870 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-7cff98d7df-2fx9m" podUID="7ad77931-c302-4ce8-81f0-78767bb80443" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.106:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.106:8080: connect: connection refused" Oct 14 09:06:08 crc kubenswrapper[4870]: I1014 09:06:08.694922 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-api-f4c547d9b-gjq77" Oct 14 09:06:08 crc kubenswrapper[4870]: I1014 09:06:08.723129 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-cfnapi-77b878966d-fm2hf" Oct 14 09:06:13 crc kubenswrapper[4870]: I1014 09:06:13.036206 4870 scope.go:117] "RemoveContainer" containerID="8fcaeb3090f0c89cba1e66db524b4a246072daa35592012664aca450fa09ddc9" Oct 14 09:06:13 crc kubenswrapper[4870]: E1014 09:06:13.037837 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 09:06:18 crc kubenswrapper[4870]: I1014 09:06:18.475295 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dg96wq"] Oct 14 09:06:18 crc kubenswrapper[4870]: I1014 09:06:18.478582 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dg96wq" Oct 14 09:06:18 crc kubenswrapper[4870]: I1014 09:06:18.484833 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Oct 14 09:06:18 crc kubenswrapper[4870]: I1014 09:06:18.497605 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dg96wq"] Oct 14 09:06:18 crc kubenswrapper[4870]: I1014 09:06:18.498262 4870 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-7cff98d7df-2fx9m" podUID="7ad77931-c302-4ce8-81f0-78767bb80443" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.106:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.106:8080: connect: connection refused" Oct 14 09:06:18 crc kubenswrapper[4870]: I1014 09:06:18.578181 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b87acf8e-2f81-484d-99d4-eeb3e1278895-bundle\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dg96wq\" (UID: \"b87acf8e-2f81-484d-99d4-eeb3e1278895\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dg96wq" Oct 14 09:06:18 crc kubenswrapper[4870]: I1014 09:06:18.578322 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b87acf8e-2f81-484d-99d4-eeb3e1278895-util\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dg96wq\" (UID: \"b87acf8e-2f81-484d-99d4-eeb3e1278895\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dg96wq" Oct 14 09:06:18 crc kubenswrapper[4870]: I1014 09:06:18.578596 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s6xlw\" (UniqueName: \"kubernetes.io/projected/b87acf8e-2f81-484d-99d4-eeb3e1278895-kube-api-access-s6xlw\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dg96wq\" (UID: \"b87acf8e-2f81-484d-99d4-eeb3e1278895\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dg96wq" Oct 14 09:06:18 crc kubenswrapper[4870]: I1014 09:06:18.680422 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s6xlw\" (UniqueName: \"kubernetes.io/projected/b87acf8e-2f81-484d-99d4-eeb3e1278895-kube-api-access-s6xlw\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dg96wq\" (UID: \"b87acf8e-2f81-484d-99d4-eeb3e1278895\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dg96wq" Oct 14 09:06:18 crc kubenswrapper[4870]: I1014 09:06:18.680599 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b87acf8e-2f81-484d-99d4-eeb3e1278895-bundle\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dg96wq\" (UID: \"b87acf8e-2f81-484d-99d4-eeb3e1278895\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dg96wq" Oct 14 09:06:18 crc kubenswrapper[4870]: I1014 09:06:18.680691 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b87acf8e-2f81-484d-99d4-eeb3e1278895-util\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dg96wq\" (UID: \"b87acf8e-2f81-484d-99d4-eeb3e1278895\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dg96wq" Oct 14 09:06:18 crc kubenswrapper[4870]: I1014 09:06:18.681801 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b87acf8e-2f81-484d-99d4-eeb3e1278895-util\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dg96wq\" (UID: \"b87acf8e-2f81-484d-99d4-eeb3e1278895\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dg96wq" Oct 14 09:06:18 crc kubenswrapper[4870]: I1014 09:06:18.682078 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b87acf8e-2f81-484d-99d4-eeb3e1278895-bundle\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dg96wq\" (UID: \"b87acf8e-2f81-484d-99d4-eeb3e1278895\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dg96wq" Oct 14 09:06:18 crc kubenswrapper[4870]: I1014 09:06:18.719867 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s6xlw\" (UniqueName: \"kubernetes.io/projected/b87acf8e-2f81-484d-99d4-eeb3e1278895-kube-api-access-s6xlw\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dg96wq\" (UID: \"b87acf8e-2f81-484d-99d4-eeb3e1278895\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dg96wq" Oct 14 09:06:18 crc kubenswrapper[4870]: I1014 09:06:18.818113 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dg96wq" Oct 14 09:06:19 crc kubenswrapper[4870]: I1014 09:06:19.316375 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dg96wq"] Oct 14 09:06:19 crc kubenswrapper[4870]: I1014 09:06:19.614617 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dg96wq" event={"ID":"b87acf8e-2f81-484d-99d4-eeb3e1278895","Type":"ContainerStarted","Data":"6b5d51a3bd638ff60c0fab363204a896f0145db60bc733048d0840f483da302d"} Oct 14 09:06:19 crc kubenswrapper[4870]: I1014 09:06:19.614994 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dg96wq" event={"ID":"b87acf8e-2f81-484d-99d4-eeb3e1278895","Type":"ContainerStarted","Data":"f9f9f5c604abd50b6a5d62a84e77437870c831d24a5504f30bd15ffd4e340ef1"} Oct 14 09:06:20 crc kubenswrapper[4870]: I1014 09:06:20.629513 4870 generic.go:334] "Generic (PLEG): container finished" podID="b87acf8e-2f81-484d-99d4-eeb3e1278895" containerID="6b5d51a3bd638ff60c0fab363204a896f0145db60bc733048d0840f483da302d" exitCode=0 Oct 14 09:06:20 crc kubenswrapper[4870]: I1014 09:06:20.629578 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dg96wq" event={"ID":"b87acf8e-2f81-484d-99d4-eeb3e1278895","Type":"ContainerDied","Data":"6b5d51a3bd638ff60c0fab363204a896f0145db60bc733048d0840f483da302d"} Oct 14 09:06:24 crc kubenswrapper[4870]: I1014 09:06:24.034330 4870 scope.go:117] "RemoveContainer" containerID="8fcaeb3090f0c89cba1e66db524b4a246072daa35592012664aca450fa09ddc9" Oct 14 09:06:24 crc kubenswrapper[4870]: E1014 09:06:24.035146 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 09:06:28 crc kubenswrapper[4870]: I1014 09:06:28.498799 4870 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-7cff98d7df-2fx9m" podUID="7ad77931-c302-4ce8-81f0-78767bb80443" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.106:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.106:8080: connect: connection refused" Oct 14 09:06:28 crc kubenswrapper[4870]: I1014 09:06:28.499467 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-7cff98d7df-2fx9m" Oct 14 09:06:28 crc kubenswrapper[4870]: I1014 09:06:28.770210 4870 scope.go:117] "RemoveContainer" containerID="ed6f882b5aaf3efccc07a4239089b7fe033432102b1bbc861f8b1ab3577a69c4" Oct 14 09:06:28 crc kubenswrapper[4870]: I1014 09:06:28.815460 4870 scope.go:117] "RemoveContainer" containerID="580060d7584a6e8e9e9af94c1f92c8c6ad81baaab6f9d16525ef2919c61c2b31" Oct 14 09:06:28 crc kubenswrapper[4870]: I1014 09:06:28.965018 4870 scope.go:117] "RemoveContainer" containerID="6344e90ea632210985754aaf112afea23aaeab912ffe8717f5dd2f5ee1b46948" Oct 14 09:06:29 crc kubenswrapper[4870]: I1014 09:06:29.075482 4870 scope.go:117] "RemoveContainer" containerID="b8baaeb44c9d5a477c37b5e781b3be584d0acddcb67f90363b3985344c6c16a5" Oct 14 09:06:29 crc kubenswrapper[4870]: I1014 09:06:29.434627 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7cff98d7df-2fx9m" Oct 14 09:06:29 crc kubenswrapper[4870]: I1014 09:06:29.528179 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7ad77931-c302-4ce8-81f0-78767bb80443-logs\") pod \"7ad77931-c302-4ce8-81f0-78767bb80443\" (UID: \"7ad77931-c302-4ce8-81f0-78767bb80443\") " Oct 14 09:06:29 crc kubenswrapper[4870]: I1014 09:06:29.528546 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7ad77931-c302-4ce8-81f0-78767bb80443-logs" (OuterVolumeSpecName: "logs") pod "7ad77931-c302-4ce8-81f0-78767bb80443" (UID: "7ad77931-c302-4ce8-81f0-78767bb80443"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 09:06:29 crc kubenswrapper[4870]: I1014 09:06:29.528644 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7ad77931-c302-4ce8-81f0-78767bb80443-config-data\") pod \"7ad77931-c302-4ce8-81f0-78767bb80443\" (UID: \"7ad77931-c302-4ce8-81f0-78767bb80443\") " Oct 14 09:06:29 crc kubenswrapper[4870]: I1014 09:06:29.529592 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/7ad77931-c302-4ce8-81f0-78767bb80443-horizon-secret-key\") pod \"7ad77931-c302-4ce8-81f0-78767bb80443\" (UID: \"7ad77931-c302-4ce8-81f0-78767bb80443\") " Oct 14 09:06:29 crc kubenswrapper[4870]: I1014 09:06:29.529635 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7ad77931-c302-4ce8-81f0-78767bb80443-scripts\") pod \"7ad77931-c302-4ce8-81f0-78767bb80443\" (UID: \"7ad77931-c302-4ce8-81f0-78767bb80443\") " Oct 14 09:06:29 crc kubenswrapper[4870]: I1014 09:06:29.529772 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l2c8j\" (UniqueName: \"kubernetes.io/projected/7ad77931-c302-4ce8-81f0-78767bb80443-kube-api-access-l2c8j\") pod \"7ad77931-c302-4ce8-81f0-78767bb80443\" (UID: \"7ad77931-c302-4ce8-81f0-78767bb80443\") " Oct 14 09:06:29 crc kubenswrapper[4870]: I1014 09:06:29.530195 4870 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7ad77931-c302-4ce8-81f0-78767bb80443-logs\") on node \"crc\" DevicePath \"\"" Oct 14 09:06:29 crc kubenswrapper[4870]: I1014 09:06:29.534968 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7ad77931-c302-4ce8-81f0-78767bb80443-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "7ad77931-c302-4ce8-81f0-78767bb80443" (UID: "7ad77931-c302-4ce8-81f0-78767bb80443"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:06:29 crc kubenswrapper[4870]: I1014 09:06:29.535371 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7ad77931-c302-4ce8-81f0-78767bb80443-kube-api-access-l2c8j" (OuterVolumeSpecName: "kube-api-access-l2c8j") pod "7ad77931-c302-4ce8-81f0-78767bb80443" (UID: "7ad77931-c302-4ce8-81f0-78767bb80443"). InnerVolumeSpecName "kube-api-access-l2c8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:06:29 crc kubenswrapper[4870]: I1014 09:06:29.555525 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7ad77931-c302-4ce8-81f0-78767bb80443-scripts" (OuterVolumeSpecName: "scripts") pod "7ad77931-c302-4ce8-81f0-78767bb80443" (UID: "7ad77931-c302-4ce8-81f0-78767bb80443"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 09:06:29 crc kubenswrapper[4870]: I1014 09:06:29.557098 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7ad77931-c302-4ce8-81f0-78767bb80443-config-data" (OuterVolumeSpecName: "config-data") pod "7ad77931-c302-4ce8-81f0-78767bb80443" (UID: "7ad77931-c302-4ce8-81f0-78767bb80443"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 09:06:29 crc kubenswrapper[4870]: I1014 09:06:29.631876 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l2c8j\" (UniqueName: \"kubernetes.io/projected/7ad77931-c302-4ce8-81f0-78767bb80443-kube-api-access-l2c8j\") on node \"crc\" DevicePath \"\"" Oct 14 09:06:29 crc kubenswrapper[4870]: I1014 09:06:29.631928 4870 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7ad77931-c302-4ce8-81f0-78767bb80443-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 09:06:29 crc kubenswrapper[4870]: I1014 09:06:29.631950 4870 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/7ad77931-c302-4ce8-81f0-78767bb80443-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Oct 14 09:06:29 crc kubenswrapper[4870]: I1014 09:06:29.631971 4870 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7ad77931-c302-4ce8-81f0-78767bb80443-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 09:06:29 crc kubenswrapper[4870]: I1014 09:06:29.737486 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7cff98d7df-2fx9m" event={"ID":"7ad77931-c302-4ce8-81f0-78767bb80443","Type":"ContainerDied","Data":"4a68c4125fa0d27922a302bd38cd7ca956c788c5062a558fbf2fe6050cd6ae9a"} Oct 14 09:06:29 crc kubenswrapper[4870]: I1014 09:06:29.737510 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7cff98d7df-2fx9m" Oct 14 09:06:29 crc kubenswrapper[4870]: I1014 09:06:29.737570 4870 scope.go:117] "RemoveContainer" containerID="d898a5c61cd9c202ebe91686a9d143d06fbb01cb5ba6341691ff053cbe036f0c" Oct 14 09:06:29 crc kubenswrapper[4870]: I1014 09:06:29.737422 4870 generic.go:334] "Generic (PLEG): container finished" podID="7ad77931-c302-4ce8-81f0-78767bb80443" containerID="4a68c4125fa0d27922a302bd38cd7ca956c788c5062a558fbf2fe6050cd6ae9a" exitCode=137 Oct 14 09:06:29 crc kubenswrapper[4870]: I1014 09:06:29.737691 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7cff98d7df-2fx9m" event={"ID":"7ad77931-c302-4ce8-81f0-78767bb80443","Type":"ContainerDied","Data":"75811ed35ae717f13ca70040f547a8131fa29f5d02af80e4da49519a86594f06"} Oct 14 09:06:29 crc kubenswrapper[4870]: I1014 09:06:29.741232 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dg96wq" event={"ID":"b87acf8e-2f81-484d-99d4-eeb3e1278895","Type":"ContainerStarted","Data":"2e95111cea453ead1925f112088881ffcdd6d22c96f4103d1f55cde29518ec12"} Oct 14 09:06:29 crc kubenswrapper[4870]: I1014 09:06:29.796236 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7cff98d7df-2fx9m"] Oct 14 09:06:29 crc kubenswrapper[4870]: I1014 09:06:29.809805 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-7cff98d7df-2fx9m"] Oct 14 09:06:29 crc kubenswrapper[4870]: E1014 09:06:29.943841 4870 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7ad77931_c302_4ce8_81f0_78767bb80443.slice\": RecentStats: unable to find data in memory cache]" Oct 14 09:06:29 crc kubenswrapper[4870]: I1014 09:06:29.972550 4870 scope.go:117] "RemoveContainer" containerID="4a68c4125fa0d27922a302bd38cd7ca956c788c5062a558fbf2fe6050cd6ae9a" Oct 14 09:06:29 crc kubenswrapper[4870]: I1014 09:06:29.992929 4870 scope.go:117] "RemoveContainer" containerID="d898a5c61cd9c202ebe91686a9d143d06fbb01cb5ba6341691ff053cbe036f0c" Oct 14 09:06:29 crc kubenswrapper[4870]: E1014 09:06:29.993281 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d898a5c61cd9c202ebe91686a9d143d06fbb01cb5ba6341691ff053cbe036f0c\": container with ID starting with d898a5c61cd9c202ebe91686a9d143d06fbb01cb5ba6341691ff053cbe036f0c not found: ID does not exist" containerID="d898a5c61cd9c202ebe91686a9d143d06fbb01cb5ba6341691ff053cbe036f0c" Oct 14 09:06:29 crc kubenswrapper[4870]: I1014 09:06:29.993380 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d898a5c61cd9c202ebe91686a9d143d06fbb01cb5ba6341691ff053cbe036f0c"} err="failed to get container status \"d898a5c61cd9c202ebe91686a9d143d06fbb01cb5ba6341691ff053cbe036f0c\": rpc error: code = NotFound desc = could not find container \"d898a5c61cd9c202ebe91686a9d143d06fbb01cb5ba6341691ff053cbe036f0c\": container with ID starting with d898a5c61cd9c202ebe91686a9d143d06fbb01cb5ba6341691ff053cbe036f0c not found: ID does not exist" Oct 14 09:06:29 crc kubenswrapper[4870]: I1014 09:06:29.993486 4870 scope.go:117] "RemoveContainer" containerID="4a68c4125fa0d27922a302bd38cd7ca956c788c5062a558fbf2fe6050cd6ae9a" Oct 14 09:06:29 crc kubenswrapper[4870]: E1014 09:06:29.993850 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4a68c4125fa0d27922a302bd38cd7ca956c788c5062a558fbf2fe6050cd6ae9a\": container with ID starting with 4a68c4125fa0d27922a302bd38cd7ca956c788c5062a558fbf2fe6050cd6ae9a not found: ID does not exist" containerID="4a68c4125fa0d27922a302bd38cd7ca956c788c5062a558fbf2fe6050cd6ae9a" Oct 14 09:06:29 crc kubenswrapper[4870]: I1014 09:06:29.993894 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4a68c4125fa0d27922a302bd38cd7ca956c788c5062a558fbf2fe6050cd6ae9a"} err="failed to get container status \"4a68c4125fa0d27922a302bd38cd7ca956c788c5062a558fbf2fe6050cd6ae9a\": rpc error: code = NotFound desc = could not find container \"4a68c4125fa0d27922a302bd38cd7ca956c788c5062a558fbf2fe6050cd6ae9a\": container with ID starting with 4a68c4125fa0d27922a302bd38cd7ca956c788c5062a558fbf2fe6050cd6ae9a not found: ID does not exist" Oct 14 09:06:30 crc kubenswrapper[4870]: I1014 09:06:30.759683 4870 generic.go:334] "Generic (PLEG): container finished" podID="b87acf8e-2f81-484d-99d4-eeb3e1278895" containerID="2e95111cea453ead1925f112088881ffcdd6d22c96f4103d1f55cde29518ec12" exitCode=0 Oct 14 09:06:30 crc kubenswrapper[4870]: I1014 09:06:30.762396 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dg96wq" event={"ID":"b87acf8e-2f81-484d-99d4-eeb3e1278895","Type":"ContainerDied","Data":"2e95111cea453ead1925f112088881ffcdd6d22c96f4103d1f55cde29518ec12"} Oct 14 09:06:31 crc kubenswrapper[4870]: I1014 09:06:31.051553 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7ad77931-c302-4ce8-81f0-78767bb80443" path="/var/lib/kubelet/pods/7ad77931-c302-4ce8-81f0-78767bb80443/volumes" Oct 14 09:06:31 crc kubenswrapper[4870]: I1014 09:06:31.777280 4870 generic.go:334] "Generic (PLEG): container finished" podID="b87acf8e-2f81-484d-99d4-eeb3e1278895" containerID="735b446a6547174520fcfa3e42e2528d3af4fd7e1b175b23152b0660476d6d09" exitCode=0 Oct 14 09:06:31 crc kubenswrapper[4870]: I1014 09:06:31.777328 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dg96wq" event={"ID":"b87acf8e-2f81-484d-99d4-eeb3e1278895","Type":"ContainerDied","Data":"735b446a6547174520fcfa3e42e2528d3af4fd7e1b175b23152b0660476d6d09"} Oct 14 09:06:33 crc kubenswrapper[4870]: I1014 09:06:33.140701 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dg96wq" Oct 14 09:06:33 crc kubenswrapper[4870]: I1014 09:06:33.214275 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b87acf8e-2f81-484d-99d4-eeb3e1278895-util\") pod \"b87acf8e-2f81-484d-99d4-eeb3e1278895\" (UID: \"b87acf8e-2f81-484d-99d4-eeb3e1278895\") " Oct 14 09:06:33 crc kubenswrapper[4870]: I1014 09:06:33.214382 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s6xlw\" (UniqueName: \"kubernetes.io/projected/b87acf8e-2f81-484d-99d4-eeb3e1278895-kube-api-access-s6xlw\") pod \"b87acf8e-2f81-484d-99d4-eeb3e1278895\" (UID: \"b87acf8e-2f81-484d-99d4-eeb3e1278895\") " Oct 14 09:06:33 crc kubenswrapper[4870]: I1014 09:06:33.214772 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b87acf8e-2f81-484d-99d4-eeb3e1278895-bundle\") pod \"b87acf8e-2f81-484d-99d4-eeb3e1278895\" (UID: \"b87acf8e-2f81-484d-99d4-eeb3e1278895\") " Oct 14 09:06:33 crc kubenswrapper[4870]: I1014 09:06:33.216097 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b87acf8e-2f81-484d-99d4-eeb3e1278895-bundle" (OuterVolumeSpecName: "bundle") pod "b87acf8e-2f81-484d-99d4-eeb3e1278895" (UID: "b87acf8e-2f81-484d-99d4-eeb3e1278895"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 09:06:33 crc kubenswrapper[4870]: I1014 09:06:33.218635 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b87acf8e-2f81-484d-99d4-eeb3e1278895-util" (OuterVolumeSpecName: "util") pod "b87acf8e-2f81-484d-99d4-eeb3e1278895" (UID: "b87acf8e-2f81-484d-99d4-eeb3e1278895"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 09:06:33 crc kubenswrapper[4870]: I1014 09:06:33.230034 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b87acf8e-2f81-484d-99d4-eeb3e1278895-kube-api-access-s6xlw" (OuterVolumeSpecName: "kube-api-access-s6xlw") pod "b87acf8e-2f81-484d-99d4-eeb3e1278895" (UID: "b87acf8e-2f81-484d-99d4-eeb3e1278895"). InnerVolumeSpecName "kube-api-access-s6xlw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:06:33 crc kubenswrapper[4870]: I1014 09:06:33.318478 4870 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b87acf8e-2f81-484d-99d4-eeb3e1278895-util\") on node \"crc\" DevicePath \"\"" Oct 14 09:06:33 crc kubenswrapper[4870]: I1014 09:06:33.318522 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s6xlw\" (UniqueName: \"kubernetes.io/projected/b87acf8e-2f81-484d-99d4-eeb3e1278895-kube-api-access-s6xlw\") on node \"crc\" DevicePath \"\"" Oct 14 09:06:33 crc kubenswrapper[4870]: I1014 09:06:33.318536 4870 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b87acf8e-2f81-484d-99d4-eeb3e1278895-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 09:06:33 crc kubenswrapper[4870]: I1014 09:06:33.795997 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dg96wq" event={"ID":"b87acf8e-2f81-484d-99d4-eeb3e1278895","Type":"ContainerDied","Data":"f9f9f5c604abd50b6a5d62a84e77437870c831d24a5504f30bd15ffd4e340ef1"} Oct 14 09:06:33 crc kubenswrapper[4870]: I1014 09:06:33.796046 4870 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f9f9f5c604abd50b6a5d62a84e77437870c831d24a5504f30bd15ffd4e340ef1" Oct 14 09:06:33 crc kubenswrapper[4870]: I1014 09:06:33.796047 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dg96wq" Oct 14 09:06:35 crc kubenswrapper[4870]: I1014 09:06:35.041236 4870 scope.go:117] "RemoveContainer" containerID="8fcaeb3090f0c89cba1e66db524b4a246072daa35592012664aca450fa09ddc9" Oct 14 09:06:35 crc kubenswrapper[4870]: E1014 09:06:35.041711 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 09:06:40 crc kubenswrapper[4870]: I1014 09:06:40.061710 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-8h9wb"] Oct 14 09:06:40 crc kubenswrapper[4870]: I1014 09:06:40.078651 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-8h9wb"] Oct 14 09:06:41 crc kubenswrapper[4870]: I1014 09:06:41.044427 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="673ae6d4-e844-4bda-8be7-0cc398c456c9" path="/var/lib/kubelet/pods/673ae6d4-e844-4bda-8be7-0cc398c456c9/volumes" Oct 14 09:06:46 crc kubenswrapper[4870]: I1014 09:06:46.480662 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-7c8cf85677-fzq5k"] Oct 14 09:06:46 crc kubenswrapper[4870]: E1014 09:06:46.481541 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ad77931-c302-4ce8-81f0-78767bb80443" containerName="horizon" Oct 14 09:06:46 crc kubenswrapper[4870]: I1014 09:06:46.481557 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ad77931-c302-4ce8-81f0-78767bb80443" containerName="horizon" Oct 14 09:06:46 crc kubenswrapper[4870]: E1014 09:06:46.481591 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b87acf8e-2f81-484d-99d4-eeb3e1278895" containerName="util" Oct 14 09:06:46 crc kubenswrapper[4870]: I1014 09:06:46.481608 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="b87acf8e-2f81-484d-99d4-eeb3e1278895" containerName="util" Oct 14 09:06:46 crc kubenswrapper[4870]: E1014 09:06:46.481634 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b87acf8e-2f81-484d-99d4-eeb3e1278895" containerName="pull" Oct 14 09:06:46 crc kubenswrapper[4870]: I1014 09:06:46.481643 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="b87acf8e-2f81-484d-99d4-eeb3e1278895" containerName="pull" Oct 14 09:06:46 crc kubenswrapper[4870]: E1014 09:06:46.481653 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b87acf8e-2f81-484d-99d4-eeb3e1278895" containerName="extract" Oct 14 09:06:46 crc kubenswrapper[4870]: I1014 09:06:46.481660 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="b87acf8e-2f81-484d-99d4-eeb3e1278895" containerName="extract" Oct 14 09:06:46 crc kubenswrapper[4870]: E1014 09:06:46.481692 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ad77931-c302-4ce8-81f0-78767bb80443" containerName="horizon-log" Oct 14 09:06:46 crc kubenswrapper[4870]: I1014 09:06:46.481700 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ad77931-c302-4ce8-81f0-78767bb80443" containerName="horizon-log" Oct 14 09:06:46 crc kubenswrapper[4870]: I1014 09:06:46.481930 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="7ad77931-c302-4ce8-81f0-78767bb80443" containerName="horizon" Oct 14 09:06:46 crc kubenswrapper[4870]: I1014 09:06:46.481956 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="b87acf8e-2f81-484d-99d4-eeb3e1278895" containerName="extract" Oct 14 09:06:46 crc kubenswrapper[4870]: I1014 09:06:46.481978 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="7ad77931-c302-4ce8-81f0-78767bb80443" containerName="horizon-log" Oct 14 09:06:46 crc kubenswrapper[4870]: I1014 09:06:46.482914 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-7c8cf85677-fzq5k" Oct 14 09:06:46 crc kubenswrapper[4870]: I1014 09:06:46.484610 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"kube-root-ca.crt" Oct 14 09:06:46 crc kubenswrapper[4870]: I1014 09:06:46.486316 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-dockercfg-wrmlp" Oct 14 09:06:46 crc kubenswrapper[4870]: I1014 09:06:46.486633 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"openshift-service-ca.crt" Oct 14 09:06:46 crc kubenswrapper[4870]: I1014 09:06:46.490469 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-7c8cf85677-fzq5k"] Oct 14 09:06:46 crc kubenswrapper[4870]: I1014 09:06:46.575623 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gz7f7\" (UniqueName: \"kubernetes.io/projected/ea8c24a3-71f9-43d5-8524-d3d95dd3b9f6-kube-api-access-gz7f7\") pod \"obo-prometheus-operator-7c8cf85677-fzq5k\" (UID: \"ea8c24a3-71f9-43d5-8524-d3d95dd3b9f6\") " pod="openshift-operators/obo-prometheus-operator-7c8cf85677-fzq5k" Oct 14 09:06:46 crc kubenswrapper[4870]: I1014 09:06:46.614475 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-769c7ff77b-ksw25"] Oct 14 09:06:46 crc kubenswrapper[4870]: I1014 09:06:46.616045 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-769c7ff77b-ksw25" Oct 14 09:06:46 crc kubenswrapper[4870]: I1014 09:06:46.617839 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-service-cert" Oct 14 09:06:46 crc kubenswrapper[4870]: I1014 09:06:46.618192 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-dockercfg-v82pj" Oct 14 09:06:46 crc kubenswrapper[4870]: I1014 09:06:46.626775 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-769c7ff77b-prngx"] Oct 14 09:06:46 crc kubenswrapper[4870]: I1014 09:06:46.630053 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-769c7ff77b-prngx" Oct 14 09:06:46 crc kubenswrapper[4870]: I1014 09:06:46.642902 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-769c7ff77b-ksw25"] Oct 14 09:06:46 crc kubenswrapper[4870]: I1014 09:06:46.669520 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-769c7ff77b-prngx"] Oct 14 09:06:46 crc kubenswrapper[4870]: I1014 09:06:46.678794 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/203800e4-7772-492d-942b-1175c02174e8-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-769c7ff77b-prngx\" (UID: \"203800e4-7772-492d-942b-1175c02174e8\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-769c7ff77b-prngx" Oct 14 09:06:46 crc kubenswrapper[4870]: I1014 09:06:46.678856 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/203800e4-7772-492d-942b-1175c02174e8-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-769c7ff77b-prngx\" (UID: \"203800e4-7772-492d-942b-1175c02174e8\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-769c7ff77b-prngx" Oct 14 09:06:46 crc kubenswrapper[4870]: I1014 09:06:46.678904 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/b0c9563b-efee-4741-88f3-b09a08320257-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-769c7ff77b-ksw25\" (UID: \"b0c9563b-efee-4741-88f3-b09a08320257\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-769c7ff77b-ksw25" Oct 14 09:06:46 crc kubenswrapper[4870]: I1014 09:06:46.678946 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/b0c9563b-efee-4741-88f3-b09a08320257-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-769c7ff77b-ksw25\" (UID: \"b0c9563b-efee-4741-88f3-b09a08320257\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-769c7ff77b-ksw25" Oct 14 09:06:46 crc kubenswrapper[4870]: I1014 09:06:46.679003 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gz7f7\" (UniqueName: \"kubernetes.io/projected/ea8c24a3-71f9-43d5-8524-d3d95dd3b9f6-kube-api-access-gz7f7\") pod \"obo-prometheus-operator-7c8cf85677-fzq5k\" (UID: \"ea8c24a3-71f9-43d5-8524-d3d95dd3b9f6\") " pod="openshift-operators/obo-prometheus-operator-7c8cf85677-fzq5k" Oct 14 09:06:46 crc kubenswrapper[4870]: I1014 09:06:46.710920 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gz7f7\" (UniqueName: \"kubernetes.io/projected/ea8c24a3-71f9-43d5-8524-d3d95dd3b9f6-kube-api-access-gz7f7\") pod \"obo-prometheus-operator-7c8cf85677-fzq5k\" (UID: \"ea8c24a3-71f9-43d5-8524-d3d95dd3b9f6\") " pod="openshift-operators/obo-prometheus-operator-7c8cf85677-fzq5k" Oct 14 09:06:46 crc kubenswrapper[4870]: I1014 09:06:46.780779 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/b0c9563b-efee-4741-88f3-b09a08320257-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-769c7ff77b-ksw25\" (UID: \"b0c9563b-efee-4741-88f3-b09a08320257\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-769c7ff77b-ksw25" Oct 14 09:06:46 crc kubenswrapper[4870]: I1014 09:06:46.781260 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/203800e4-7772-492d-942b-1175c02174e8-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-769c7ff77b-prngx\" (UID: \"203800e4-7772-492d-942b-1175c02174e8\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-769c7ff77b-prngx" Oct 14 09:06:46 crc kubenswrapper[4870]: I1014 09:06:46.781317 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/203800e4-7772-492d-942b-1175c02174e8-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-769c7ff77b-prngx\" (UID: \"203800e4-7772-492d-942b-1175c02174e8\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-769c7ff77b-prngx" Oct 14 09:06:46 crc kubenswrapper[4870]: I1014 09:06:46.781394 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/b0c9563b-efee-4741-88f3-b09a08320257-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-769c7ff77b-ksw25\" (UID: \"b0c9563b-efee-4741-88f3-b09a08320257\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-769c7ff77b-ksw25" Oct 14 09:06:46 crc kubenswrapper[4870]: I1014 09:06:46.785881 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/203800e4-7772-492d-942b-1175c02174e8-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-769c7ff77b-prngx\" (UID: \"203800e4-7772-492d-942b-1175c02174e8\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-769c7ff77b-prngx" Oct 14 09:06:46 crc kubenswrapper[4870]: I1014 09:06:46.787100 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/b0c9563b-efee-4741-88f3-b09a08320257-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-769c7ff77b-ksw25\" (UID: \"b0c9563b-efee-4741-88f3-b09a08320257\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-769c7ff77b-ksw25" Oct 14 09:06:46 crc kubenswrapper[4870]: I1014 09:06:46.787491 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/b0c9563b-efee-4741-88f3-b09a08320257-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-769c7ff77b-ksw25\" (UID: \"b0c9563b-efee-4741-88f3-b09a08320257\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-769c7ff77b-ksw25" Oct 14 09:06:46 crc kubenswrapper[4870]: I1014 09:06:46.796971 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/observability-operator-cc5f78dfc-xmgpw"] Oct 14 09:06:46 crc kubenswrapper[4870]: I1014 09:06:46.798334 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-cc5f78dfc-xmgpw" Oct 14 09:06:46 crc kubenswrapper[4870]: I1014 09:06:46.801192 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-sa-dockercfg-k7h8t" Oct 14 09:06:46 crc kubenswrapper[4870]: I1014 09:06:46.810403 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-tls" Oct 14 09:06:46 crc kubenswrapper[4870]: I1014 09:06:46.811300 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/203800e4-7772-492d-942b-1175c02174e8-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-769c7ff77b-prngx\" (UID: \"203800e4-7772-492d-942b-1175c02174e8\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-769c7ff77b-prngx" Oct 14 09:06:46 crc kubenswrapper[4870]: I1014 09:06:46.817836 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-cc5f78dfc-xmgpw"] Oct 14 09:06:46 crc kubenswrapper[4870]: I1014 09:06:46.848082 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-7c8cf85677-fzq5k" Oct 14 09:06:46 crc kubenswrapper[4870]: I1014 09:06:46.951219 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-769c7ff77b-ksw25" Oct 14 09:06:46 crc kubenswrapper[4870]: I1014 09:06:46.965489 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-769c7ff77b-prngx" Oct 14 09:06:46 crc kubenswrapper[4870]: I1014 09:06:46.984507 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xtw6w\" (UniqueName: \"kubernetes.io/projected/8b4db008-9294-40b8-9b49-96273ac9627a-kube-api-access-xtw6w\") pod \"observability-operator-cc5f78dfc-xmgpw\" (UID: \"8b4db008-9294-40b8-9b49-96273ac9627a\") " pod="openshift-operators/observability-operator-cc5f78dfc-xmgpw" Oct 14 09:06:46 crc kubenswrapper[4870]: I1014 09:06:46.984648 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/8b4db008-9294-40b8-9b49-96273ac9627a-observability-operator-tls\") pod \"observability-operator-cc5f78dfc-xmgpw\" (UID: \"8b4db008-9294-40b8-9b49-96273ac9627a\") " pod="openshift-operators/observability-operator-cc5f78dfc-xmgpw" Oct 14 09:06:47 crc kubenswrapper[4870]: I1014 09:06:47.018515 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/perses-operator-54bc95c9fb-ghvk7"] Oct 14 09:06:47 crc kubenswrapper[4870]: I1014 09:06:47.020043 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-54bc95c9fb-ghvk7" Oct 14 09:06:47 crc kubenswrapper[4870]: I1014 09:06:47.028233 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"perses-operator-dockercfg-7rp9c" Oct 14 09:06:47 crc kubenswrapper[4870]: I1014 09:06:47.028245 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-54bc95c9fb-ghvk7"] Oct 14 09:06:47 crc kubenswrapper[4870]: I1014 09:06:47.089083 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xtw6w\" (UniqueName: \"kubernetes.io/projected/8b4db008-9294-40b8-9b49-96273ac9627a-kube-api-access-xtw6w\") pod \"observability-operator-cc5f78dfc-xmgpw\" (UID: \"8b4db008-9294-40b8-9b49-96273ac9627a\") " pod="openshift-operators/observability-operator-cc5f78dfc-xmgpw" Oct 14 09:06:47 crc kubenswrapper[4870]: I1014 09:06:47.089220 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/8b4db008-9294-40b8-9b49-96273ac9627a-observability-operator-tls\") pod \"observability-operator-cc5f78dfc-xmgpw\" (UID: \"8b4db008-9294-40b8-9b49-96273ac9627a\") " pod="openshift-operators/observability-operator-cc5f78dfc-xmgpw" Oct 14 09:06:47 crc kubenswrapper[4870]: I1014 09:06:47.094175 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/8b4db008-9294-40b8-9b49-96273ac9627a-observability-operator-tls\") pod \"observability-operator-cc5f78dfc-xmgpw\" (UID: \"8b4db008-9294-40b8-9b49-96273ac9627a\") " pod="openshift-operators/observability-operator-cc5f78dfc-xmgpw" Oct 14 09:06:47 crc kubenswrapper[4870]: I1014 09:06:47.113617 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xtw6w\" (UniqueName: \"kubernetes.io/projected/8b4db008-9294-40b8-9b49-96273ac9627a-kube-api-access-xtw6w\") pod \"observability-operator-cc5f78dfc-xmgpw\" (UID: \"8b4db008-9294-40b8-9b49-96273ac9627a\") " pod="openshift-operators/observability-operator-cc5f78dfc-xmgpw" Oct 14 09:06:47 crc kubenswrapper[4870]: I1014 09:06:47.196331 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/c8bc8628-7639-4a34-b1f3-5de19a0614a8-openshift-service-ca\") pod \"perses-operator-54bc95c9fb-ghvk7\" (UID: \"c8bc8628-7639-4a34-b1f3-5de19a0614a8\") " pod="openshift-operators/perses-operator-54bc95c9fb-ghvk7" Oct 14 09:06:47 crc kubenswrapper[4870]: I1014 09:06:47.196400 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-drv4h\" (UniqueName: \"kubernetes.io/projected/c8bc8628-7639-4a34-b1f3-5de19a0614a8-kube-api-access-drv4h\") pod \"perses-operator-54bc95c9fb-ghvk7\" (UID: \"c8bc8628-7639-4a34-b1f3-5de19a0614a8\") " pod="openshift-operators/perses-operator-54bc95c9fb-ghvk7" Oct 14 09:06:47 crc kubenswrapper[4870]: I1014 09:06:47.254978 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-cc5f78dfc-xmgpw" Oct 14 09:06:47 crc kubenswrapper[4870]: I1014 09:06:47.299670 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/c8bc8628-7639-4a34-b1f3-5de19a0614a8-openshift-service-ca\") pod \"perses-operator-54bc95c9fb-ghvk7\" (UID: \"c8bc8628-7639-4a34-b1f3-5de19a0614a8\") " pod="openshift-operators/perses-operator-54bc95c9fb-ghvk7" Oct 14 09:06:47 crc kubenswrapper[4870]: I1014 09:06:47.299746 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-drv4h\" (UniqueName: \"kubernetes.io/projected/c8bc8628-7639-4a34-b1f3-5de19a0614a8-kube-api-access-drv4h\") pod \"perses-operator-54bc95c9fb-ghvk7\" (UID: \"c8bc8628-7639-4a34-b1f3-5de19a0614a8\") " pod="openshift-operators/perses-operator-54bc95c9fb-ghvk7" Oct 14 09:06:47 crc kubenswrapper[4870]: I1014 09:06:47.300561 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/c8bc8628-7639-4a34-b1f3-5de19a0614a8-openshift-service-ca\") pod \"perses-operator-54bc95c9fb-ghvk7\" (UID: \"c8bc8628-7639-4a34-b1f3-5de19a0614a8\") " pod="openshift-operators/perses-operator-54bc95c9fb-ghvk7" Oct 14 09:06:47 crc kubenswrapper[4870]: I1014 09:06:47.331232 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-drv4h\" (UniqueName: \"kubernetes.io/projected/c8bc8628-7639-4a34-b1f3-5de19a0614a8-kube-api-access-drv4h\") pod \"perses-operator-54bc95c9fb-ghvk7\" (UID: \"c8bc8628-7639-4a34-b1f3-5de19a0614a8\") " pod="openshift-operators/perses-operator-54bc95c9fb-ghvk7" Oct 14 09:06:47 crc kubenswrapper[4870]: I1014 09:06:47.360003 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-54bc95c9fb-ghvk7" Oct 14 09:06:47 crc kubenswrapper[4870]: I1014 09:06:47.462707 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-7c8cf85677-fzq5k"] Oct 14 09:06:47 crc kubenswrapper[4870]: I1014 09:06:47.586869 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-769c7ff77b-ksw25"] Oct 14 09:06:47 crc kubenswrapper[4870]: W1014 09:06:47.638255 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb0c9563b_efee_4741_88f3_b09a08320257.slice/crio-57e9bdc112d68f2f96642a61e1ef9859a965c02a5c99eebc6f7ef8ada6808499 WatchSource:0}: Error finding container 57e9bdc112d68f2f96642a61e1ef9859a965c02a5c99eebc6f7ef8ada6808499: Status 404 returned error can't find the container with id 57e9bdc112d68f2f96642a61e1ef9859a965c02a5c99eebc6f7ef8ada6808499 Oct 14 09:06:47 crc kubenswrapper[4870]: I1014 09:06:47.708210 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-769c7ff77b-prngx"] Oct 14 09:06:47 crc kubenswrapper[4870]: I1014 09:06:47.785423 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-cc5f78dfc-xmgpw"] Oct 14 09:06:47 crc kubenswrapper[4870]: W1014 09:06:47.796426 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8b4db008_9294_40b8_9b49_96273ac9627a.slice/crio-6500e3298daeebd7302d7e7a793d1268660686e04bbe48e8127d5784c459030b WatchSource:0}: Error finding container 6500e3298daeebd7302d7e7a793d1268660686e04bbe48e8127d5784c459030b: Status 404 returned error can't find the container with id 6500e3298daeebd7302d7e7a793d1268660686e04bbe48e8127d5784c459030b Oct 14 09:06:47 crc kubenswrapper[4870]: I1014 09:06:47.935463 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-7c8cf85677-fzq5k" event={"ID":"ea8c24a3-71f9-43d5-8524-d3d95dd3b9f6","Type":"ContainerStarted","Data":"ce0dc8cffd900009de238542e124786c9e0faf3fcda3014a1f16098e2c7fb6aa"} Oct 14 09:06:47 crc kubenswrapper[4870]: I1014 09:06:47.937818 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-769c7ff77b-prngx" event={"ID":"203800e4-7772-492d-942b-1175c02174e8","Type":"ContainerStarted","Data":"e7f73a20fa90d5758257991d4193ea2e12a86b3c501d7f48f6b96d2f3f2474c0"} Oct 14 09:06:47 crc kubenswrapper[4870]: I1014 09:06:47.940888 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-cc5f78dfc-xmgpw" event={"ID":"8b4db008-9294-40b8-9b49-96273ac9627a","Type":"ContainerStarted","Data":"6500e3298daeebd7302d7e7a793d1268660686e04bbe48e8127d5784c459030b"} Oct 14 09:06:47 crc kubenswrapper[4870]: I1014 09:06:47.941797 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-769c7ff77b-ksw25" event={"ID":"b0c9563b-efee-4741-88f3-b09a08320257","Type":"ContainerStarted","Data":"57e9bdc112d68f2f96642a61e1ef9859a965c02a5c99eebc6f7ef8ada6808499"} Oct 14 09:06:48 crc kubenswrapper[4870]: I1014 09:06:48.076121 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-54bc95c9fb-ghvk7"] Oct 14 09:06:48 crc kubenswrapper[4870]: W1014 09:06:48.082615 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc8bc8628_7639_4a34_b1f3_5de19a0614a8.slice/crio-d647db1291e41b361f84dd2c8a37910bbbea9373b08d625d7b5687f58af43d05 WatchSource:0}: Error finding container d647db1291e41b361f84dd2c8a37910bbbea9373b08d625d7b5687f58af43d05: Status 404 returned error can't find the container with id d647db1291e41b361f84dd2c8a37910bbbea9373b08d625d7b5687f58af43d05 Oct 14 09:06:48 crc kubenswrapper[4870]: I1014 09:06:48.960849 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-54bc95c9fb-ghvk7" event={"ID":"c8bc8628-7639-4a34-b1f3-5de19a0614a8","Type":"ContainerStarted","Data":"d647db1291e41b361f84dd2c8a37910bbbea9373b08d625d7b5687f58af43d05"} Oct 14 09:06:50 crc kubenswrapper[4870]: I1014 09:06:50.026339 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-6d91-account-create-7mzvp"] Oct 14 09:06:50 crc kubenswrapper[4870]: I1014 09:06:50.035246 4870 scope.go:117] "RemoveContainer" containerID="8fcaeb3090f0c89cba1e66db524b4a246072daa35592012664aca450fa09ddc9" Oct 14 09:06:50 crc kubenswrapper[4870]: E1014 09:06:50.035576 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 09:06:50 crc kubenswrapper[4870]: I1014 09:06:50.040431 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-6d91-account-create-7mzvp"] Oct 14 09:06:51 crc kubenswrapper[4870]: I1014 09:06:51.047809 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4e4335da-b836-4ca3-b683-f658d7085175" path="/var/lib/kubelet/pods/4e4335da-b836-4ca3-b683-f658d7085175/volumes" Oct 14 09:06:56 crc kubenswrapper[4870]: I1014 09:06:56.029197 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-769c7ff77b-ksw25" event={"ID":"b0c9563b-efee-4741-88f3-b09a08320257","Type":"ContainerStarted","Data":"a781b233624ebc22e957383e07a6ffecd848dcd971af6922e663dfc138edb36c"} Oct 14 09:06:56 crc kubenswrapper[4870]: I1014 09:06:56.030518 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-7c8cf85677-fzq5k" event={"ID":"ea8c24a3-71f9-43d5-8524-d3d95dd3b9f6","Type":"ContainerStarted","Data":"513340d60fe50c0ce1cb6ea60710a9ffaf638244c4b570fe5a32aab6d21c53d6"} Oct 14 09:06:56 crc kubenswrapper[4870]: I1014 09:06:56.031971 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-54bc95c9fb-ghvk7" event={"ID":"c8bc8628-7639-4a34-b1f3-5de19a0614a8","Type":"ContainerStarted","Data":"288917799b20fed60d3c097ddf5b01b72b928ee42b467ce2459929957c9ca29a"} Oct 14 09:06:56 crc kubenswrapper[4870]: I1014 09:06:56.032079 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/perses-operator-54bc95c9fb-ghvk7" Oct 14 09:06:56 crc kubenswrapper[4870]: I1014 09:06:56.033403 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-769c7ff77b-prngx" event={"ID":"203800e4-7772-492d-942b-1175c02174e8","Type":"ContainerStarted","Data":"b1d16bc7ca56bc7ab4057c1c1807c753f969ab1fe7c86eca290d194df1b12228"} Oct 14 09:06:56 crc kubenswrapper[4870]: I1014 09:06:56.034714 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-cc5f78dfc-xmgpw" event={"ID":"8b4db008-9294-40b8-9b49-96273ac9627a","Type":"ContainerStarted","Data":"dc32d7dd7155c80df34a4bfabc0c9394651817b30b2d9e82c1973008184f6b93"} Oct 14 09:06:56 crc kubenswrapper[4870]: I1014 09:06:56.034911 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/observability-operator-cc5f78dfc-xmgpw" Oct 14 09:06:56 crc kubenswrapper[4870]: I1014 09:06:56.044785 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/observability-operator-cc5f78dfc-xmgpw" Oct 14 09:06:56 crc kubenswrapper[4870]: I1014 09:06:56.057950 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-769c7ff77b-ksw25" podStartSLOduration=2.681077496 podStartE2EDuration="10.057930465s" podCreationTimestamp="2025-10-14 09:06:46 +0000 UTC" firstStartedPulling="2025-10-14 09:06:47.663402868 +0000 UTC m=+7543.360763239" lastFinishedPulling="2025-10-14 09:06:55.040255827 +0000 UTC m=+7550.737616208" observedRunningTime="2025-10-14 09:06:56.046863382 +0000 UTC m=+7551.744223753" watchObservedRunningTime="2025-10-14 09:06:56.057930465 +0000 UTC m=+7551.755290836" Oct 14 09:06:56 crc kubenswrapper[4870]: I1014 09:06:56.080674 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/observability-operator-cc5f78dfc-xmgpw" podStartSLOduration=2.783427353 podStartE2EDuration="10.080657776s" podCreationTimestamp="2025-10-14 09:06:46 +0000 UTC" firstStartedPulling="2025-10-14 09:06:47.7995727 +0000 UTC m=+7543.496933071" lastFinishedPulling="2025-10-14 09:06:55.096803133 +0000 UTC m=+7550.794163494" observedRunningTime="2025-10-14 09:06:56.07595594 +0000 UTC m=+7551.773316311" watchObservedRunningTime="2025-10-14 09:06:56.080657776 +0000 UTC m=+7551.778018147" Oct 14 09:06:56 crc kubenswrapper[4870]: I1014 09:06:56.156778 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/perses-operator-54bc95c9fb-ghvk7" podStartSLOduration=3.215056442 podStartE2EDuration="10.156759026s" podCreationTimestamp="2025-10-14 09:06:46 +0000 UTC" firstStartedPulling="2025-10-14 09:06:48.084950057 +0000 UTC m=+7543.782310428" lastFinishedPulling="2025-10-14 09:06:55.026652631 +0000 UTC m=+7550.724013012" observedRunningTime="2025-10-14 09:06:56.130656941 +0000 UTC m=+7551.828017332" watchObservedRunningTime="2025-10-14 09:06:56.156759026 +0000 UTC m=+7551.854119397" Oct 14 09:06:56 crc kubenswrapper[4870]: I1014 09:06:56.219029 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-769c7ff77b-prngx" podStartSLOduration=2.942746919 podStartE2EDuration="10.219003843s" podCreationTimestamp="2025-10-14 09:06:46 +0000 UTC" firstStartedPulling="2025-10-14 09:06:47.747742421 +0000 UTC m=+7543.445102782" lastFinishedPulling="2025-10-14 09:06:55.023999325 +0000 UTC m=+7550.721359706" observedRunningTime="2025-10-14 09:06:56.15654643 +0000 UTC m=+7551.853906801" watchObservedRunningTime="2025-10-14 09:06:56.219003843 +0000 UTC m=+7551.916364214" Oct 14 09:06:56 crc kubenswrapper[4870]: I1014 09:06:56.232848 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-7c8cf85677-fzq5k" podStartSLOduration=2.703174873 podStartE2EDuration="10.232827244s" podCreationTimestamp="2025-10-14 09:06:46 +0000 UTC" firstStartedPulling="2025-10-14 09:06:47.497464151 +0000 UTC m=+7543.194824522" lastFinishedPulling="2025-10-14 09:06:55.027116512 +0000 UTC m=+7550.724476893" observedRunningTime="2025-10-14 09:06:56.176990325 +0000 UTC m=+7551.874350696" watchObservedRunningTime="2025-10-14 09:06:56.232827244 +0000 UTC m=+7551.930187615" Oct 14 09:07:04 crc kubenswrapper[4870]: I1014 09:07:04.035405 4870 scope.go:117] "RemoveContainer" containerID="8fcaeb3090f0c89cba1e66db524b4a246072daa35592012664aca450fa09ddc9" Oct 14 09:07:04 crc kubenswrapper[4870]: E1014 09:07:04.036476 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 09:07:07 crc kubenswrapper[4870]: I1014 09:07:07.364883 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/perses-operator-54bc95c9fb-ghvk7" Oct 14 09:07:11 crc kubenswrapper[4870]: I1014 09:07:11.274533 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Oct 14 09:07:11 crc kubenswrapper[4870]: I1014 09:07:11.275158 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstackclient" podUID="fff3c418-2e18-4393-99c2-f24ad07746bb" containerName="openstackclient" containerID="cri-o://5f2890b7d74b88dc7653d308e99317cec8d57d4714978e696efc0b53a6d9c3e9" gracePeriod=2 Oct 14 09:07:11 crc kubenswrapper[4870]: I1014 09:07:11.288903 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Oct 14 09:07:11 crc kubenswrapper[4870]: I1014 09:07:11.307817 4870 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="fff3c418-2e18-4393-99c2-f24ad07746bb" podUID="285cda3d-f2c3-4dc1-9a2b-a24ddf6942c4" Oct 14 09:07:11 crc kubenswrapper[4870]: I1014 09:07:11.334791 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Oct 14 09:07:11 crc kubenswrapper[4870]: E1014 09:07:11.335185 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fff3c418-2e18-4393-99c2-f24ad07746bb" containerName="openstackclient" Oct 14 09:07:11 crc kubenswrapper[4870]: I1014 09:07:11.335201 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="fff3c418-2e18-4393-99c2-f24ad07746bb" containerName="openstackclient" Oct 14 09:07:11 crc kubenswrapper[4870]: I1014 09:07:11.335376 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="fff3c418-2e18-4393-99c2-f24ad07746bb" containerName="openstackclient" Oct 14 09:07:11 crc kubenswrapper[4870]: I1014 09:07:11.336041 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 14 09:07:11 crc kubenswrapper[4870]: I1014 09:07:11.348194 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 14 09:07:11 crc kubenswrapper[4870]: I1014 09:07:11.429501 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Oct 14 09:07:11 crc kubenswrapper[4870]: E1014 09:07:11.430286 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[kube-api-access-fl2xp openstack-config openstack-config-secret], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/openstackclient" podUID="285cda3d-f2c3-4dc1-9a2b-a24ddf6942c4" Oct 14 09:07:11 crc kubenswrapper[4870]: I1014 09:07:11.439940 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Oct 14 09:07:11 crc kubenswrapper[4870]: I1014 09:07:11.460957 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fl2xp\" (UniqueName: \"kubernetes.io/projected/285cda3d-f2c3-4dc1-9a2b-a24ddf6942c4-kube-api-access-fl2xp\") pod \"openstackclient\" (UID: \"285cda3d-f2c3-4dc1-9a2b-a24ddf6942c4\") " pod="openstack/openstackclient" Oct 14 09:07:11 crc kubenswrapper[4870]: I1014 09:07:11.461058 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/285cda3d-f2c3-4dc1-9a2b-a24ddf6942c4-openstack-config-secret\") pod \"openstackclient\" (UID: \"285cda3d-f2c3-4dc1-9a2b-a24ddf6942c4\") " pod="openstack/openstackclient" Oct 14 09:07:11 crc kubenswrapper[4870]: I1014 09:07:11.461139 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/285cda3d-f2c3-4dc1-9a2b-a24ddf6942c4-openstack-config\") pod \"openstackclient\" (UID: \"285cda3d-f2c3-4dc1-9a2b-a24ddf6942c4\") " pod="openstack/openstackclient" Oct 14 09:07:11 crc kubenswrapper[4870]: I1014 09:07:11.464607 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Oct 14 09:07:11 crc kubenswrapper[4870]: I1014 09:07:11.466732 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 14 09:07:11 crc kubenswrapper[4870]: I1014 09:07:11.477967 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 14 09:07:11 crc kubenswrapper[4870]: I1014 09:07:11.563078 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fl2xp\" (UniqueName: \"kubernetes.io/projected/285cda3d-f2c3-4dc1-9a2b-a24ddf6942c4-kube-api-access-fl2xp\") pod \"openstackclient\" (UID: \"285cda3d-f2c3-4dc1-9a2b-a24ddf6942c4\") " pod="openstack/openstackclient" Oct 14 09:07:11 crc kubenswrapper[4870]: I1014 09:07:11.563127 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/285cda3d-f2c3-4dc1-9a2b-a24ddf6942c4-openstack-config-secret\") pod \"openstackclient\" (UID: \"285cda3d-f2c3-4dc1-9a2b-a24ddf6942c4\") " pod="openstack/openstackclient" Oct 14 09:07:11 crc kubenswrapper[4870]: I1014 09:07:11.563157 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/73cb6ac7-5103-401c-8d79-6d2d337b133c-openstack-config-secret\") pod \"openstackclient\" (UID: \"73cb6ac7-5103-401c-8d79-6d2d337b133c\") " pod="openstack/openstackclient" Oct 14 09:07:11 crc kubenswrapper[4870]: I1014 09:07:11.563207 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/285cda3d-f2c3-4dc1-9a2b-a24ddf6942c4-openstack-config\") pod \"openstackclient\" (UID: \"285cda3d-f2c3-4dc1-9a2b-a24ddf6942c4\") " pod="openstack/openstackclient" Oct 14 09:07:11 crc kubenswrapper[4870]: I1014 09:07:11.563239 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/73cb6ac7-5103-401c-8d79-6d2d337b133c-openstack-config\") pod \"openstackclient\" (UID: \"73cb6ac7-5103-401c-8d79-6d2d337b133c\") " pod="openstack/openstackclient" Oct 14 09:07:11 crc kubenswrapper[4870]: I1014 09:07:11.563302 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g492x\" (UniqueName: \"kubernetes.io/projected/73cb6ac7-5103-401c-8d79-6d2d337b133c-kube-api-access-g492x\") pod \"openstackclient\" (UID: \"73cb6ac7-5103-401c-8d79-6d2d337b133c\") " pod="openstack/openstackclient" Oct 14 09:07:11 crc kubenswrapper[4870]: I1014 09:07:11.564995 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/285cda3d-f2c3-4dc1-9a2b-a24ddf6942c4-openstack-config\") pod \"openstackclient\" (UID: \"285cda3d-f2c3-4dc1-9a2b-a24ddf6942c4\") " pod="openstack/openstackclient" Oct 14 09:07:11 crc kubenswrapper[4870]: E1014 09:07:11.566909 4870 projected.go:194] Error preparing data for projected volume kube-api-access-fl2xp for pod openstack/openstackclient: failed to fetch token: serviceaccounts "openstackclient-openstackclient" is forbidden: the UID in the bound object reference (285cda3d-f2c3-4dc1-9a2b-a24ddf6942c4) does not match the UID in record. The object might have been deleted and then recreated Oct 14 09:07:11 crc kubenswrapper[4870]: E1014 09:07:11.566978 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/285cda3d-f2c3-4dc1-9a2b-a24ddf6942c4-kube-api-access-fl2xp podName:285cda3d-f2c3-4dc1-9a2b-a24ddf6942c4 nodeName:}" failed. No retries permitted until 2025-10-14 09:07:12.066961213 +0000 UTC m=+7567.764321584 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-fl2xp" (UniqueName: "kubernetes.io/projected/285cda3d-f2c3-4dc1-9a2b-a24ddf6942c4-kube-api-access-fl2xp") pod "openstackclient" (UID: "285cda3d-f2c3-4dc1-9a2b-a24ddf6942c4") : failed to fetch token: serviceaccounts "openstackclient-openstackclient" is forbidden: the UID in the bound object reference (285cda3d-f2c3-4dc1-9a2b-a24ddf6942c4) does not match the UID in record. The object might have been deleted and then recreated Oct 14 09:07:11 crc kubenswrapper[4870]: I1014 09:07:11.571973 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/285cda3d-f2c3-4dc1-9a2b-a24ddf6942c4-openstack-config-secret\") pod \"openstackclient\" (UID: \"285cda3d-f2c3-4dc1-9a2b-a24ddf6942c4\") " pod="openstack/openstackclient" Oct 14 09:07:11 crc kubenswrapper[4870]: I1014 09:07:11.615019 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Oct 14 09:07:11 crc kubenswrapper[4870]: I1014 09:07:11.616321 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 14 09:07:11 crc kubenswrapper[4870]: I1014 09:07:11.618299 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-wt44z" Oct 14 09:07:11 crc kubenswrapper[4870]: I1014 09:07:11.633783 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 14 09:07:11 crc kubenswrapper[4870]: I1014 09:07:11.665309 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/73cb6ac7-5103-401c-8d79-6d2d337b133c-openstack-config-secret\") pod \"openstackclient\" (UID: \"73cb6ac7-5103-401c-8d79-6d2d337b133c\") " pod="openstack/openstackclient" Oct 14 09:07:11 crc kubenswrapper[4870]: I1014 09:07:11.665413 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/73cb6ac7-5103-401c-8d79-6d2d337b133c-openstack-config\") pod \"openstackclient\" (UID: \"73cb6ac7-5103-401c-8d79-6d2d337b133c\") " pod="openstack/openstackclient" Oct 14 09:07:11 crc kubenswrapper[4870]: I1014 09:07:11.665505 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g492x\" (UniqueName: \"kubernetes.io/projected/73cb6ac7-5103-401c-8d79-6d2d337b133c-kube-api-access-g492x\") pod \"openstackclient\" (UID: \"73cb6ac7-5103-401c-8d79-6d2d337b133c\") " pod="openstack/openstackclient" Oct 14 09:07:11 crc kubenswrapper[4870]: I1014 09:07:11.668171 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/73cb6ac7-5103-401c-8d79-6d2d337b133c-openstack-config\") pod \"openstackclient\" (UID: \"73cb6ac7-5103-401c-8d79-6d2d337b133c\") " pod="openstack/openstackclient" Oct 14 09:07:11 crc kubenswrapper[4870]: I1014 09:07:11.690006 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/73cb6ac7-5103-401c-8d79-6d2d337b133c-openstack-config-secret\") pod \"openstackclient\" (UID: \"73cb6ac7-5103-401c-8d79-6d2d337b133c\") " pod="openstack/openstackclient" Oct 14 09:07:11 crc kubenswrapper[4870]: I1014 09:07:11.698240 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g492x\" (UniqueName: \"kubernetes.io/projected/73cb6ac7-5103-401c-8d79-6d2d337b133c-kube-api-access-g492x\") pod \"openstackclient\" (UID: \"73cb6ac7-5103-401c-8d79-6d2d337b133c\") " pod="openstack/openstackclient" Oct 14 09:07:11 crc kubenswrapper[4870]: I1014 09:07:11.770923 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zb64w\" (UniqueName: \"kubernetes.io/projected/f1c15a86-a683-46ae-8ca4-431ebbf213c9-kube-api-access-zb64w\") pod \"kube-state-metrics-0\" (UID: \"f1c15a86-a683-46ae-8ca4-431ebbf213c9\") " pod="openstack/kube-state-metrics-0" Oct 14 09:07:11 crc kubenswrapper[4870]: I1014 09:07:11.783979 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 14 09:07:11 crc kubenswrapper[4870]: I1014 09:07:11.872768 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zb64w\" (UniqueName: \"kubernetes.io/projected/f1c15a86-a683-46ae-8ca4-431ebbf213c9-kube-api-access-zb64w\") pod \"kube-state-metrics-0\" (UID: \"f1c15a86-a683-46ae-8ca4-431ebbf213c9\") " pod="openstack/kube-state-metrics-0" Oct 14 09:07:11 crc kubenswrapper[4870]: I1014 09:07:11.897968 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zb64w\" (UniqueName: \"kubernetes.io/projected/f1c15a86-a683-46ae-8ca4-431ebbf213c9-kube-api-access-zb64w\") pod \"kube-state-metrics-0\" (UID: \"f1c15a86-a683-46ae-8ca4-431ebbf213c9\") " pod="openstack/kube-state-metrics-0" Oct 14 09:07:11 crc kubenswrapper[4870]: I1014 09:07:11.933890 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 14 09:07:12 crc kubenswrapper[4870]: I1014 09:07:12.082373 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fl2xp\" (UniqueName: \"kubernetes.io/projected/285cda3d-f2c3-4dc1-9a2b-a24ddf6942c4-kube-api-access-fl2xp\") pod \"openstackclient\" (UID: \"285cda3d-f2c3-4dc1-9a2b-a24ddf6942c4\") " pod="openstack/openstackclient" Oct 14 09:07:12 crc kubenswrapper[4870]: E1014 09:07:12.091915 4870 projected.go:194] Error preparing data for projected volume kube-api-access-fl2xp for pod openstack/openstackclient: failed to fetch token: serviceaccounts "openstackclient-openstackclient" is forbidden: the UID in the bound object reference (285cda3d-f2c3-4dc1-9a2b-a24ddf6942c4) does not match the UID in record. The object might have been deleted and then recreated Oct 14 09:07:12 crc kubenswrapper[4870]: E1014 09:07:12.091979 4870 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/285cda3d-f2c3-4dc1-9a2b-a24ddf6942c4-kube-api-access-fl2xp podName:285cda3d-f2c3-4dc1-9a2b-a24ddf6942c4 nodeName:}" failed. No retries permitted until 2025-10-14 09:07:13.091961766 +0000 UTC m=+7568.789322137 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-fl2xp" (UniqueName: "kubernetes.io/projected/285cda3d-f2c3-4dc1-9a2b-a24ddf6942c4-kube-api-access-fl2xp") pod "openstackclient" (UID: "285cda3d-f2c3-4dc1-9a2b-a24ddf6942c4") : failed to fetch token: serviceaccounts "openstackclient-openstackclient" is forbidden: the UID in the bound object reference (285cda3d-f2c3-4dc1-9a2b-a24ddf6942c4) does not match the UID in record. The object might have been deleted and then recreated Oct 14 09:07:12 crc kubenswrapper[4870]: I1014 09:07:12.225847 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 14 09:07:12 crc kubenswrapper[4870]: I1014 09:07:12.237543 4870 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="285cda3d-f2c3-4dc1-9a2b-a24ddf6942c4" podUID="73cb6ac7-5103-401c-8d79-6d2d337b133c" Oct 14 09:07:12 crc kubenswrapper[4870]: I1014 09:07:12.252505 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 14 09:07:12 crc kubenswrapper[4870]: I1014 09:07:12.297177 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/alertmanager-metric-storage-0"] Oct 14 09:07:12 crc kubenswrapper[4870]: I1014 09:07:12.307968 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/alertmanager-metric-storage-0" Oct 14 09:07:12 crc kubenswrapper[4870]: I1014 09:07:12.311279 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"alertmanager-metric-storage-tls-assets-0" Oct 14 09:07:12 crc kubenswrapper[4870]: I1014 09:07:12.311579 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-alertmanager-dockercfg-h478v" Oct 14 09:07:12 crc kubenswrapper[4870]: I1014 09:07:12.311821 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"alertmanager-metric-storage-web-config" Oct 14 09:07:12 crc kubenswrapper[4870]: I1014 09:07:12.319522 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"alertmanager-metric-storage-generated" Oct 14 09:07:12 crc kubenswrapper[4870]: I1014 09:07:12.326291 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/alertmanager-metric-storage-0"] Oct 14 09:07:12 crc kubenswrapper[4870]: I1014 09:07:12.395964 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/285cda3d-f2c3-4dc1-9a2b-a24ddf6942c4-openstack-config-secret\") pod \"285cda3d-f2c3-4dc1-9a2b-a24ddf6942c4\" (UID: \"285cda3d-f2c3-4dc1-9a2b-a24ddf6942c4\") " Oct 14 09:07:12 crc kubenswrapper[4870]: I1014 09:07:12.396068 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/285cda3d-f2c3-4dc1-9a2b-a24ddf6942c4-openstack-config\") pod \"285cda3d-f2c3-4dc1-9a2b-a24ddf6942c4\" (UID: \"285cda3d-f2c3-4dc1-9a2b-a24ddf6942c4\") " Oct 14 09:07:12 crc kubenswrapper[4870]: I1014 09:07:12.396375 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"alertmanager-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/ffbac7fa-fdbf-42d6-9941-472fa69c381b-alertmanager-metric-storage-db\") pod \"alertmanager-metric-storage-0\" (UID: \"ffbac7fa-fdbf-42d6-9941-472fa69c381b\") " pod="openstack/alertmanager-metric-storage-0" Oct 14 09:07:12 crc kubenswrapper[4870]: I1014 09:07:12.396414 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2l7qf\" (UniqueName: \"kubernetes.io/projected/ffbac7fa-fdbf-42d6-9941-472fa69c381b-kube-api-access-2l7qf\") pod \"alertmanager-metric-storage-0\" (UID: \"ffbac7fa-fdbf-42d6-9941-472fa69c381b\") " pod="openstack/alertmanager-metric-storage-0" Oct 14 09:07:12 crc kubenswrapper[4870]: I1014 09:07:12.396433 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/ffbac7fa-fdbf-42d6-9941-472fa69c381b-web-config\") pod \"alertmanager-metric-storage-0\" (UID: \"ffbac7fa-fdbf-42d6-9941-472fa69c381b\") " pod="openstack/alertmanager-metric-storage-0" Oct 14 09:07:12 crc kubenswrapper[4870]: I1014 09:07:12.396491 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/ffbac7fa-fdbf-42d6-9941-472fa69c381b-config-volume\") pod \"alertmanager-metric-storage-0\" (UID: \"ffbac7fa-fdbf-42d6-9941-472fa69c381b\") " pod="openstack/alertmanager-metric-storage-0" Oct 14 09:07:12 crc kubenswrapper[4870]: I1014 09:07:12.396514 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/ffbac7fa-fdbf-42d6-9941-472fa69c381b-tls-assets\") pod \"alertmanager-metric-storage-0\" (UID: \"ffbac7fa-fdbf-42d6-9941-472fa69c381b\") " pod="openstack/alertmanager-metric-storage-0" Oct 14 09:07:12 crc kubenswrapper[4870]: I1014 09:07:12.396587 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/ffbac7fa-fdbf-42d6-9941-472fa69c381b-config-out\") pod \"alertmanager-metric-storage-0\" (UID: \"ffbac7fa-fdbf-42d6-9941-472fa69c381b\") " pod="openstack/alertmanager-metric-storage-0" Oct 14 09:07:12 crc kubenswrapper[4870]: I1014 09:07:12.396636 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fl2xp\" (UniqueName: \"kubernetes.io/projected/285cda3d-f2c3-4dc1-9a2b-a24ddf6942c4-kube-api-access-fl2xp\") on node \"crc\" DevicePath \"\"" Oct 14 09:07:12 crc kubenswrapper[4870]: I1014 09:07:12.399517 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/285cda3d-f2c3-4dc1-9a2b-a24ddf6942c4-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "285cda3d-f2c3-4dc1-9a2b-a24ddf6942c4" (UID: "285cda3d-f2c3-4dc1-9a2b-a24ddf6942c4"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 09:07:12 crc kubenswrapper[4870]: I1014 09:07:12.416832 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/285cda3d-f2c3-4dc1-9a2b-a24ddf6942c4-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "285cda3d-f2c3-4dc1-9a2b-a24ddf6942c4" (UID: "285cda3d-f2c3-4dc1-9a2b-a24ddf6942c4"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:07:12 crc kubenswrapper[4870]: I1014 09:07:12.456488 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 14 09:07:12 crc kubenswrapper[4870]: I1014 09:07:12.498754 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2l7qf\" (UniqueName: \"kubernetes.io/projected/ffbac7fa-fdbf-42d6-9941-472fa69c381b-kube-api-access-2l7qf\") pod \"alertmanager-metric-storage-0\" (UID: \"ffbac7fa-fdbf-42d6-9941-472fa69c381b\") " pod="openstack/alertmanager-metric-storage-0" Oct 14 09:07:12 crc kubenswrapper[4870]: I1014 09:07:12.498862 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/ffbac7fa-fdbf-42d6-9941-472fa69c381b-web-config\") pod \"alertmanager-metric-storage-0\" (UID: \"ffbac7fa-fdbf-42d6-9941-472fa69c381b\") " pod="openstack/alertmanager-metric-storage-0" Oct 14 09:07:12 crc kubenswrapper[4870]: I1014 09:07:12.498987 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/ffbac7fa-fdbf-42d6-9941-472fa69c381b-config-volume\") pod \"alertmanager-metric-storage-0\" (UID: \"ffbac7fa-fdbf-42d6-9941-472fa69c381b\") " pod="openstack/alertmanager-metric-storage-0" Oct 14 09:07:12 crc kubenswrapper[4870]: I1014 09:07:12.499068 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/ffbac7fa-fdbf-42d6-9941-472fa69c381b-tls-assets\") pod \"alertmanager-metric-storage-0\" (UID: \"ffbac7fa-fdbf-42d6-9941-472fa69c381b\") " pod="openstack/alertmanager-metric-storage-0" Oct 14 09:07:12 crc kubenswrapper[4870]: I1014 09:07:12.499203 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/ffbac7fa-fdbf-42d6-9941-472fa69c381b-config-out\") pod \"alertmanager-metric-storage-0\" (UID: \"ffbac7fa-fdbf-42d6-9941-472fa69c381b\") " pod="openstack/alertmanager-metric-storage-0" Oct 14 09:07:12 crc kubenswrapper[4870]: I1014 09:07:12.499737 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"alertmanager-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/ffbac7fa-fdbf-42d6-9941-472fa69c381b-alertmanager-metric-storage-db\") pod \"alertmanager-metric-storage-0\" (UID: \"ffbac7fa-fdbf-42d6-9941-472fa69c381b\") " pod="openstack/alertmanager-metric-storage-0" Oct 14 09:07:12 crc kubenswrapper[4870]: I1014 09:07:12.500103 4870 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/285cda3d-f2c3-4dc1-9a2b-a24ddf6942c4-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Oct 14 09:07:12 crc kubenswrapper[4870]: I1014 09:07:12.500185 4870 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/285cda3d-f2c3-4dc1-9a2b-a24ddf6942c4-openstack-config\") on node \"crc\" DevicePath \"\"" Oct 14 09:07:12 crc kubenswrapper[4870]: I1014 09:07:12.502859 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"alertmanager-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/ffbac7fa-fdbf-42d6-9941-472fa69c381b-alertmanager-metric-storage-db\") pod \"alertmanager-metric-storage-0\" (UID: \"ffbac7fa-fdbf-42d6-9941-472fa69c381b\") " pod="openstack/alertmanager-metric-storage-0" Oct 14 09:07:12 crc kubenswrapper[4870]: I1014 09:07:12.504261 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/ffbac7fa-fdbf-42d6-9941-472fa69c381b-tls-assets\") pod \"alertmanager-metric-storage-0\" (UID: \"ffbac7fa-fdbf-42d6-9941-472fa69c381b\") " pod="openstack/alertmanager-metric-storage-0" Oct 14 09:07:12 crc kubenswrapper[4870]: I1014 09:07:12.505599 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/ffbac7fa-fdbf-42d6-9941-472fa69c381b-config-volume\") pod \"alertmanager-metric-storage-0\" (UID: \"ffbac7fa-fdbf-42d6-9941-472fa69c381b\") " pod="openstack/alertmanager-metric-storage-0" Oct 14 09:07:12 crc kubenswrapper[4870]: I1014 09:07:12.515252 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/ffbac7fa-fdbf-42d6-9941-472fa69c381b-web-config\") pod \"alertmanager-metric-storage-0\" (UID: \"ffbac7fa-fdbf-42d6-9941-472fa69c381b\") " pod="openstack/alertmanager-metric-storage-0" Oct 14 09:07:12 crc kubenswrapper[4870]: I1014 09:07:12.522807 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2l7qf\" (UniqueName: \"kubernetes.io/projected/ffbac7fa-fdbf-42d6-9941-472fa69c381b-kube-api-access-2l7qf\") pod \"alertmanager-metric-storage-0\" (UID: \"ffbac7fa-fdbf-42d6-9941-472fa69c381b\") " pod="openstack/alertmanager-metric-storage-0" Oct 14 09:07:12 crc kubenswrapper[4870]: I1014 09:07:12.523549 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/ffbac7fa-fdbf-42d6-9941-472fa69c381b-config-out\") pod \"alertmanager-metric-storage-0\" (UID: \"ffbac7fa-fdbf-42d6-9941-472fa69c381b\") " pod="openstack/alertmanager-metric-storage-0" Oct 14 09:07:12 crc kubenswrapper[4870]: I1014 09:07:12.526517 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 14 09:07:12 crc kubenswrapper[4870]: I1014 09:07:12.646936 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/alertmanager-metric-storage-0" Oct 14 09:07:12 crc kubenswrapper[4870]: I1014 09:07:12.932805 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/prometheus-metric-storage-0"] Oct 14 09:07:12 crc kubenswrapper[4870]: I1014 09:07:12.937553 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Oct 14 09:07:12 crc kubenswrapper[4870]: I1014 09:07:12.954018 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-web-config" Oct 14 09:07:12 crc kubenswrapper[4870]: I1014 09:07:12.954158 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage" Oct 14 09:07:12 crc kubenswrapper[4870]: I1014 09:07:12.954255 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-0" Oct 14 09:07:12 crc kubenswrapper[4870]: I1014 09:07:12.954530 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-prometheus-dockercfg-x7l48" Oct 14 09:07:12 crc kubenswrapper[4870]: I1014 09:07:12.958771 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-tls-assets-0" Oct 14 09:07:12 crc kubenswrapper[4870]: I1014 09:07:12.960257 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-thanos-prometheus-http-client-file" Oct 14 09:07:12 crc kubenswrapper[4870]: I1014 09:07:12.979002 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Oct 14 09:07:13 crc kubenswrapper[4870]: I1014 09:07:13.010049 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/1920aea5-2efe-4faf-9a54-d5bb6f83cc08-config\") pod \"prometheus-metric-storage-0\" (UID: \"1920aea5-2efe-4faf-9a54-d5bb6f83cc08\") " pod="openstack/prometheus-metric-storage-0" Oct 14 09:07:13 crc kubenswrapper[4870]: I1014 09:07:13.010290 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/1920aea5-2efe-4faf-9a54-d5bb6f83cc08-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"1920aea5-2efe-4faf-9a54-d5bb6f83cc08\") " pod="openstack/prometheus-metric-storage-0" Oct 14 09:07:13 crc kubenswrapper[4870]: I1014 09:07:13.010377 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-acd4a217-ece4-4ec0-9e29-288aef7e78b8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-acd4a217-ece4-4ec0-9e29-288aef7e78b8\") pod \"prometheus-metric-storage-0\" (UID: \"1920aea5-2efe-4faf-9a54-d5bb6f83cc08\") " pod="openstack/prometheus-metric-storage-0" Oct 14 09:07:13 crc kubenswrapper[4870]: I1014 09:07:13.010460 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/1920aea5-2efe-4faf-9a54-d5bb6f83cc08-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"1920aea5-2efe-4faf-9a54-d5bb6f83cc08\") " pod="openstack/prometheus-metric-storage-0" Oct 14 09:07:13 crc kubenswrapper[4870]: I1014 09:07:13.010573 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dzkjs\" (UniqueName: \"kubernetes.io/projected/1920aea5-2efe-4faf-9a54-d5bb6f83cc08-kube-api-access-dzkjs\") pod \"prometheus-metric-storage-0\" (UID: \"1920aea5-2efe-4faf-9a54-d5bb6f83cc08\") " pod="openstack/prometheus-metric-storage-0" Oct 14 09:07:13 crc kubenswrapper[4870]: I1014 09:07:13.010651 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/1920aea5-2efe-4faf-9a54-d5bb6f83cc08-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"1920aea5-2efe-4faf-9a54-d5bb6f83cc08\") " pod="openstack/prometheus-metric-storage-0" Oct 14 09:07:13 crc kubenswrapper[4870]: I1014 09:07:13.010728 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/1920aea5-2efe-4faf-9a54-d5bb6f83cc08-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"1920aea5-2efe-4faf-9a54-d5bb6f83cc08\") " pod="openstack/prometheus-metric-storage-0" Oct 14 09:07:13 crc kubenswrapper[4870]: I1014 09:07:13.010876 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/1920aea5-2efe-4faf-9a54-d5bb6f83cc08-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"1920aea5-2efe-4faf-9a54-d5bb6f83cc08\") " pod="openstack/prometheus-metric-storage-0" Oct 14 09:07:13 crc kubenswrapper[4870]: I1014 09:07:13.087682 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="285cda3d-f2c3-4dc1-9a2b-a24ddf6942c4" path="/var/lib/kubelet/pods/285cda3d-f2c3-4dc1-9a2b-a24ddf6942c4/volumes" Oct 14 09:07:13 crc kubenswrapper[4870]: I1014 09:07:13.124795 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/1920aea5-2efe-4faf-9a54-d5bb6f83cc08-config\") pod \"prometheus-metric-storage-0\" (UID: \"1920aea5-2efe-4faf-9a54-d5bb6f83cc08\") " pod="openstack/prometheus-metric-storage-0" Oct 14 09:07:13 crc kubenswrapper[4870]: I1014 09:07:13.124865 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/1920aea5-2efe-4faf-9a54-d5bb6f83cc08-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"1920aea5-2efe-4faf-9a54-d5bb6f83cc08\") " pod="openstack/prometheus-metric-storage-0" Oct 14 09:07:13 crc kubenswrapper[4870]: I1014 09:07:13.124908 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-acd4a217-ece4-4ec0-9e29-288aef7e78b8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-acd4a217-ece4-4ec0-9e29-288aef7e78b8\") pod \"prometheus-metric-storage-0\" (UID: \"1920aea5-2efe-4faf-9a54-d5bb6f83cc08\") " pod="openstack/prometheus-metric-storage-0" Oct 14 09:07:13 crc kubenswrapper[4870]: I1014 09:07:13.124933 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/1920aea5-2efe-4faf-9a54-d5bb6f83cc08-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"1920aea5-2efe-4faf-9a54-d5bb6f83cc08\") " pod="openstack/prometheus-metric-storage-0" Oct 14 09:07:13 crc kubenswrapper[4870]: I1014 09:07:13.124991 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dzkjs\" (UniqueName: \"kubernetes.io/projected/1920aea5-2efe-4faf-9a54-d5bb6f83cc08-kube-api-access-dzkjs\") pod \"prometheus-metric-storage-0\" (UID: \"1920aea5-2efe-4faf-9a54-d5bb6f83cc08\") " pod="openstack/prometheus-metric-storage-0" Oct 14 09:07:13 crc kubenswrapper[4870]: I1014 09:07:13.125014 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/1920aea5-2efe-4faf-9a54-d5bb6f83cc08-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"1920aea5-2efe-4faf-9a54-d5bb6f83cc08\") " pod="openstack/prometheus-metric-storage-0" Oct 14 09:07:13 crc kubenswrapper[4870]: I1014 09:07:13.125051 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/1920aea5-2efe-4faf-9a54-d5bb6f83cc08-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"1920aea5-2efe-4faf-9a54-d5bb6f83cc08\") " pod="openstack/prometheus-metric-storage-0" Oct 14 09:07:13 crc kubenswrapper[4870]: I1014 09:07:13.125221 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/1920aea5-2efe-4faf-9a54-d5bb6f83cc08-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"1920aea5-2efe-4faf-9a54-d5bb6f83cc08\") " pod="openstack/prometheus-metric-storage-0" Oct 14 09:07:13 crc kubenswrapper[4870]: I1014 09:07:13.125879 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/1920aea5-2efe-4faf-9a54-d5bb6f83cc08-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"1920aea5-2efe-4faf-9a54-d5bb6f83cc08\") " pod="openstack/prometheus-metric-storage-0" Oct 14 09:07:13 crc kubenswrapper[4870]: I1014 09:07:13.161246 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/1920aea5-2efe-4faf-9a54-d5bb6f83cc08-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"1920aea5-2efe-4faf-9a54-d5bb6f83cc08\") " pod="openstack/prometheus-metric-storage-0" Oct 14 09:07:13 crc kubenswrapper[4870]: I1014 09:07:13.167654 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/1920aea5-2efe-4faf-9a54-d5bb6f83cc08-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"1920aea5-2efe-4faf-9a54-d5bb6f83cc08\") " pod="openstack/prometheus-metric-storage-0" Oct 14 09:07:13 crc kubenswrapper[4870]: I1014 09:07:13.189933 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/1920aea5-2efe-4faf-9a54-d5bb6f83cc08-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"1920aea5-2efe-4faf-9a54-d5bb6f83cc08\") " pod="openstack/prometheus-metric-storage-0" Oct 14 09:07:13 crc kubenswrapper[4870]: I1014 09:07:13.191358 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/1920aea5-2efe-4faf-9a54-d5bb6f83cc08-config\") pod \"prometheus-metric-storage-0\" (UID: \"1920aea5-2efe-4faf-9a54-d5bb6f83cc08\") " pod="openstack/prometheus-metric-storage-0" Oct 14 09:07:13 crc kubenswrapper[4870]: I1014 09:07:13.211002 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/1920aea5-2efe-4faf-9a54-d5bb6f83cc08-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"1920aea5-2efe-4faf-9a54-d5bb6f83cc08\") " pod="openstack/prometheus-metric-storage-0" Oct 14 09:07:13 crc kubenswrapper[4870]: I1014 09:07:13.219980 4870 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 14 09:07:13 crc kubenswrapper[4870]: I1014 09:07:13.220210 4870 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-acd4a217-ece4-4ec0-9e29-288aef7e78b8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-acd4a217-ece4-4ec0-9e29-288aef7e78b8\") pod \"prometheus-metric-storage-0\" (UID: \"1920aea5-2efe-4faf-9a54-d5bb6f83cc08\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/c38e6abdac4aee513c8f21a833c64a21e4e50b9d1958ad34709a8cd1e49bdd4c/globalmount\"" pod="openstack/prometheus-metric-storage-0" Oct 14 09:07:13 crc kubenswrapper[4870]: I1014 09:07:13.224200 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dzkjs\" (UniqueName: \"kubernetes.io/projected/1920aea5-2efe-4faf-9a54-d5bb6f83cc08-kube-api-access-dzkjs\") pod \"prometheus-metric-storage-0\" (UID: \"1920aea5-2efe-4faf-9a54-d5bb6f83cc08\") " pod="openstack/prometheus-metric-storage-0" Oct 14 09:07:13 crc kubenswrapper[4870]: I1014 09:07:13.265085 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"73cb6ac7-5103-401c-8d79-6d2d337b133c","Type":"ContainerStarted","Data":"74ffdd0ce8597d351f6e2d9b8016149fcf06805aacc3b5917412bcf4586606c3"} Oct 14 09:07:13 crc kubenswrapper[4870]: I1014 09:07:13.266135 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"73cb6ac7-5103-401c-8d79-6d2d337b133c","Type":"ContainerStarted","Data":"52bb5a0bf7af71bf652397c6851aca5ce79ba9505c5bd261744f4dcf6e8e6082"} Oct 14 09:07:13 crc kubenswrapper[4870]: I1014 09:07:13.293724 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 14 09:07:13 crc kubenswrapper[4870]: I1014 09:07:13.293818 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"f1c15a86-a683-46ae-8ca4-431ebbf213c9","Type":"ContainerStarted","Data":"924d7b6b9a2564ad4d43a2d99d8e93c0103d467f6a718151b03be558d1fc7be6"} Oct 14 09:07:13 crc kubenswrapper[4870]: I1014 09:07:13.339327 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=2.339307675 podStartE2EDuration="2.339307675s" podCreationTimestamp="2025-10-14 09:07:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 09:07:13.327262868 +0000 UTC m=+7569.024623249" watchObservedRunningTime="2025-10-14 09:07:13.339307675 +0000 UTC m=+7569.036668046" Oct 14 09:07:13 crc kubenswrapper[4870]: I1014 09:07:13.340493 4870 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="285cda3d-f2c3-4dc1-9a2b-a24ddf6942c4" podUID="73cb6ac7-5103-401c-8d79-6d2d337b133c" Oct 14 09:07:13 crc kubenswrapper[4870]: I1014 09:07:13.407695 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/alertmanager-metric-storage-0"] Oct 14 09:07:13 crc kubenswrapper[4870]: I1014 09:07:13.458520 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-acd4a217-ece4-4ec0-9e29-288aef7e78b8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-acd4a217-ece4-4ec0-9e29-288aef7e78b8\") pod \"prometheus-metric-storage-0\" (UID: \"1920aea5-2efe-4faf-9a54-d5bb6f83cc08\") " pod="openstack/prometheus-metric-storage-0" Oct 14 09:07:13 crc kubenswrapper[4870]: I1014 09:07:13.599065 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Oct 14 09:07:13 crc kubenswrapper[4870]: I1014 09:07:13.900850 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 14 09:07:13 crc kubenswrapper[4870]: I1014 09:07:13.962569 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rgcgw\" (UniqueName: \"kubernetes.io/projected/fff3c418-2e18-4393-99c2-f24ad07746bb-kube-api-access-rgcgw\") pod \"fff3c418-2e18-4393-99c2-f24ad07746bb\" (UID: \"fff3c418-2e18-4393-99c2-f24ad07746bb\") " Oct 14 09:07:13 crc kubenswrapper[4870]: I1014 09:07:13.962661 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/fff3c418-2e18-4393-99c2-f24ad07746bb-openstack-config-secret\") pod \"fff3c418-2e18-4393-99c2-f24ad07746bb\" (UID: \"fff3c418-2e18-4393-99c2-f24ad07746bb\") " Oct 14 09:07:13 crc kubenswrapper[4870]: I1014 09:07:13.962803 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/fff3c418-2e18-4393-99c2-f24ad07746bb-openstack-config\") pod \"fff3c418-2e18-4393-99c2-f24ad07746bb\" (UID: \"fff3c418-2e18-4393-99c2-f24ad07746bb\") " Oct 14 09:07:13 crc kubenswrapper[4870]: I1014 09:07:13.971352 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fff3c418-2e18-4393-99c2-f24ad07746bb-kube-api-access-rgcgw" (OuterVolumeSpecName: "kube-api-access-rgcgw") pod "fff3c418-2e18-4393-99c2-f24ad07746bb" (UID: "fff3c418-2e18-4393-99c2-f24ad07746bb"). InnerVolumeSpecName "kube-api-access-rgcgw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:07:13 crc kubenswrapper[4870]: I1014 09:07:13.991630 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fff3c418-2e18-4393-99c2-f24ad07746bb-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "fff3c418-2e18-4393-99c2-f24ad07746bb" (UID: "fff3c418-2e18-4393-99c2-f24ad07746bb"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 09:07:14 crc kubenswrapper[4870]: I1014 09:07:14.029179 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-4p4kx"] Oct 14 09:07:14 crc kubenswrapper[4870]: I1014 09:07:14.040468 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-4p4kx"] Oct 14 09:07:14 crc kubenswrapper[4870]: I1014 09:07:14.042231 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fff3c418-2e18-4393-99c2-f24ad07746bb-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "fff3c418-2e18-4393-99c2-f24ad07746bb" (UID: "fff3c418-2e18-4393-99c2-f24ad07746bb"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:07:14 crc kubenswrapper[4870]: I1014 09:07:14.065129 4870 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/fff3c418-2e18-4393-99c2-f24ad07746bb-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Oct 14 09:07:14 crc kubenswrapper[4870]: I1014 09:07:14.065156 4870 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/fff3c418-2e18-4393-99c2-f24ad07746bb-openstack-config\") on node \"crc\" DevicePath \"\"" Oct 14 09:07:14 crc kubenswrapper[4870]: I1014 09:07:14.065165 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rgcgw\" (UniqueName: \"kubernetes.io/projected/fff3c418-2e18-4393-99c2-f24ad07746bb-kube-api-access-rgcgw\") on node \"crc\" DevicePath \"\"" Oct 14 09:07:14 crc kubenswrapper[4870]: I1014 09:07:14.270333 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Oct 14 09:07:14 crc kubenswrapper[4870]: I1014 09:07:14.307839 4870 generic.go:334] "Generic (PLEG): container finished" podID="fff3c418-2e18-4393-99c2-f24ad07746bb" containerID="5f2890b7d74b88dc7653d308e99317cec8d57d4714978e696efc0b53a6d9c3e9" exitCode=137 Oct 14 09:07:14 crc kubenswrapper[4870]: I1014 09:07:14.307956 4870 scope.go:117] "RemoveContainer" containerID="5f2890b7d74b88dc7653d308e99317cec8d57d4714978e696efc0b53a6d9c3e9" Oct 14 09:07:14 crc kubenswrapper[4870]: I1014 09:07:14.307957 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 14 09:07:14 crc kubenswrapper[4870]: I1014 09:07:14.309746 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"1920aea5-2efe-4faf-9a54-d5bb6f83cc08","Type":"ContainerStarted","Data":"99ae3756e9d32a78d68b157b7ee145418f48cca2f088358e2a73e4c7cd91e790"} Oct 14 09:07:14 crc kubenswrapper[4870]: I1014 09:07:14.317244 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"ffbac7fa-fdbf-42d6-9941-472fa69c381b","Type":"ContainerStarted","Data":"492949cdf70f22c27380ec7a6a9fb3c2bc52a63137bae55b9d93ca27b4c4d634"} Oct 14 09:07:14 crc kubenswrapper[4870]: I1014 09:07:14.327745 4870 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="fff3c418-2e18-4393-99c2-f24ad07746bb" podUID="73cb6ac7-5103-401c-8d79-6d2d337b133c" Oct 14 09:07:14 crc kubenswrapper[4870]: I1014 09:07:14.339176 4870 scope.go:117] "RemoveContainer" containerID="5f2890b7d74b88dc7653d308e99317cec8d57d4714978e696efc0b53a6d9c3e9" Oct 14 09:07:14 crc kubenswrapper[4870]: E1014 09:07:14.351686 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5f2890b7d74b88dc7653d308e99317cec8d57d4714978e696efc0b53a6d9c3e9\": container with ID starting with 5f2890b7d74b88dc7653d308e99317cec8d57d4714978e696efc0b53a6d9c3e9 not found: ID does not exist" containerID="5f2890b7d74b88dc7653d308e99317cec8d57d4714978e696efc0b53a6d9c3e9" Oct 14 09:07:14 crc kubenswrapper[4870]: I1014 09:07:14.351739 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5f2890b7d74b88dc7653d308e99317cec8d57d4714978e696efc0b53a6d9c3e9"} err="failed to get container status \"5f2890b7d74b88dc7653d308e99317cec8d57d4714978e696efc0b53a6d9c3e9\": rpc error: code = NotFound desc = could not find container \"5f2890b7d74b88dc7653d308e99317cec8d57d4714978e696efc0b53a6d9c3e9\": container with ID starting with 5f2890b7d74b88dc7653d308e99317cec8d57d4714978e696efc0b53a6d9c3e9 not found: ID does not exist" Oct 14 09:07:15 crc kubenswrapper[4870]: I1014 09:07:15.053339 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f7a69635-d246-4920-b96b-6b232fa4e813" path="/var/lib/kubelet/pods/f7a69635-d246-4920-b96b-6b232fa4e813/volumes" Oct 14 09:07:15 crc kubenswrapper[4870]: I1014 09:07:15.056998 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fff3c418-2e18-4393-99c2-f24ad07746bb" path="/var/lib/kubelet/pods/fff3c418-2e18-4393-99c2-f24ad07746bb/volumes" Oct 14 09:07:19 crc kubenswrapper[4870]: I1014 09:07:19.034533 4870 scope.go:117] "RemoveContainer" containerID="8fcaeb3090f0c89cba1e66db524b4a246072daa35592012664aca450fa09ddc9" Oct 14 09:07:19 crc kubenswrapper[4870]: E1014 09:07:19.035392 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 09:07:19 crc kubenswrapper[4870]: I1014 09:07:19.389143 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"f1c15a86-a683-46ae-8ca4-431ebbf213c9","Type":"ContainerStarted","Data":"53b9dbeb351dc0701c5d9ba60d028fdc0a849094ad90946f2cb61ae415396941"} Oct 14 09:07:19 crc kubenswrapper[4870]: I1014 09:07:19.389303 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Oct 14 09:07:19 crc kubenswrapper[4870]: I1014 09:07:19.403024 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=3.017966081 podStartE2EDuration="8.403001979s" podCreationTimestamp="2025-10-14 09:07:11 +0000 UTC" firstStartedPulling="2025-10-14 09:07:12.539084536 +0000 UTC m=+7568.236444907" lastFinishedPulling="2025-10-14 09:07:17.924120424 +0000 UTC m=+7573.621480805" observedRunningTime="2025-10-14 09:07:19.401581174 +0000 UTC m=+7575.098941545" watchObservedRunningTime="2025-10-14 09:07:19.403001979 +0000 UTC m=+7575.100362360" Oct 14 09:07:20 crc kubenswrapper[4870]: I1014 09:07:20.405389 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"1920aea5-2efe-4faf-9a54-d5bb6f83cc08","Type":"ContainerStarted","Data":"7a2bb501f2d9eeda89300f155faa3d703834f4264bd858b3ed14f4debee093b6"} Oct 14 09:07:20 crc kubenswrapper[4870]: I1014 09:07:20.407667 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"ffbac7fa-fdbf-42d6-9941-472fa69c381b","Type":"ContainerStarted","Data":"029045c7d1d2dba21ac4a764aa53af9b8592e054e43126112be18312a001ab78"} Oct 14 09:07:27 crc kubenswrapper[4870]: I1014 09:07:27.498245 4870 generic.go:334] "Generic (PLEG): container finished" podID="1920aea5-2efe-4faf-9a54-d5bb6f83cc08" containerID="7a2bb501f2d9eeda89300f155faa3d703834f4264bd858b3ed14f4debee093b6" exitCode=0 Oct 14 09:07:27 crc kubenswrapper[4870]: I1014 09:07:27.498362 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"1920aea5-2efe-4faf-9a54-d5bb6f83cc08","Type":"ContainerDied","Data":"7a2bb501f2d9eeda89300f155faa3d703834f4264bd858b3ed14f4debee093b6"} Oct 14 09:07:28 crc kubenswrapper[4870]: I1014 09:07:28.511128 4870 generic.go:334] "Generic (PLEG): container finished" podID="ffbac7fa-fdbf-42d6-9941-472fa69c381b" containerID="029045c7d1d2dba21ac4a764aa53af9b8592e054e43126112be18312a001ab78" exitCode=0 Oct 14 09:07:28 crc kubenswrapper[4870]: I1014 09:07:28.511257 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"ffbac7fa-fdbf-42d6-9941-472fa69c381b","Type":"ContainerDied","Data":"029045c7d1d2dba21ac4a764aa53af9b8592e054e43126112be18312a001ab78"} Oct 14 09:07:29 crc kubenswrapper[4870]: I1014 09:07:29.453211 4870 scope.go:117] "RemoveContainer" containerID="df6172ec65f91a6d51dc77c1a71e028a0ae2e669355403f5e0db3860d68e8a0c" Oct 14 09:07:29 crc kubenswrapper[4870]: I1014 09:07:29.474750 4870 scope.go:117] "RemoveContainer" containerID="f6a6d0a40aa24e08080ff7782a1374691a10405c3eb48f5a12139f0887eec7e5" Oct 14 09:07:29 crc kubenswrapper[4870]: I1014 09:07:29.549946 4870 scope.go:117] "RemoveContainer" containerID="dab6722d8bb2837aa5d068e5d4d12b75bd61194b8d7ab4c71e41dd75604954b7" Oct 14 09:07:29 crc kubenswrapper[4870]: I1014 09:07:29.576237 4870 scope.go:117] "RemoveContainer" containerID="8e6799c323a696d937ce6adae78594ef16d1caaf215d3b849a6c746b0dccc342" Oct 14 09:07:29 crc kubenswrapper[4870]: I1014 09:07:29.634360 4870 scope.go:117] "RemoveContainer" containerID="f2b6e0bf724555237b504d3bf97ca25246546d718a3200ab1c570224f37f8f59" Oct 14 09:07:31 crc kubenswrapper[4870]: I1014 09:07:31.544096 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"ffbac7fa-fdbf-42d6-9941-472fa69c381b","Type":"ContainerStarted","Data":"7c26397df5d5da136817274bb9fedca50bb48c35d77d4662059722f99d43edc2"} Oct 14 09:07:31 crc kubenswrapper[4870]: I1014 09:07:31.941592 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Oct 14 09:07:33 crc kubenswrapper[4870]: I1014 09:07:33.564909 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"1920aea5-2efe-4faf-9a54-d5bb6f83cc08","Type":"ContainerStarted","Data":"c571d5e4ebe20d86d90c3821e0eafa7a6174113b85d38d77339caeb87ef488ea"} Oct 14 09:07:34 crc kubenswrapper[4870]: I1014 09:07:34.034254 4870 scope.go:117] "RemoveContainer" containerID="8fcaeb3090f0c89cba1e66db524b4a246072daa35592012664aca450fa09ddc9" Oct 14 09:07:34 crc kubenswrapper[4870]: E1014 09:07:34.034566 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 09:07:36 crc kubenswrapper[4870]: I1014 09:07:36.601573 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"1920aea5-2efe-4faf-9a54-d5bb6f83cc08","Type":"ContainerStarted","Data":"918385f60ffd1e48c397f9fc1b4d252b7f193217f4763354cc019448166d4254"} Oct 14 09:07:37 crc kubenswrapper[4870]: I1014 09:07:37.616247 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"ffbac7fa-fdbf-42d6-9941-472fa69c381b","Type":"ContainerStarted","Data":"0e2388be553cbfaf0acc2c39b375a77ee449eb68c438302ed75776edfc8fe1e4"} Oct 14 09:07:37 crc kubenswrapper[4870]: I1014 09:07:37.616895 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/alertmanager-metric-storage-0" Oct 14 09:07:37 crc kubenswrapper[4870]: I1014 09:07:37.621181 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/alertmanager-metric-storage-0" Oct 14 09:07:37 crc kubenswrapper[4870]: I1014 09:07:37.655399 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/alertmanager-metric-storage-0" podStartSLOduration=8.334707946 podStartE2EDuration="25.655376526s" podCreationTimestamp="2025-10-14 09:07:12 +0000 UTC" firstStartedPulling="2025-10-14 09:07:13.43261902 +0000 UTC m=+7569.129979391" lastFinishedPulling="2025-10-14 09:07:30.7532876 +0000 UTC m=+7586.450647971" observedRunningTime="2025-10-14 09:07:37.645894611 +0000 UTC m=+7593.343254992" watchObservedRunningTime="2025-10-14 09:07:37.655376526 +0000 UTC m=+7593.352736907" Oct 14 09:07:41 crc kubenswrapper[4870]: I1014 09:07:41.685330 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"1920aea5-2efe-4faf-9a54-d5bb6f83cc08","Type":"ContainerStarted","Data":"a676b85ed510438b6172e0f452a49e457b0a6d87c26d6dd399795220e4d41c34"} Oct 14 09:07:41 crc kubenswrapper[4870]: I1014 09:07:41.721790 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/prometheus-metric-storage-0" podStartSLOduration=3.9384196019999997 podStartE2EDuration="30.721769753s" podCreationTimestamp="2025-10-14 09:07:11 +0000 UTC" firstStartedPulling="2025-10-14 09:07:14.279820649 +0000 UTC m=+7569.977181030" lastFinishedPulling="2025-10-14 09:07:41.06317078 +0000 UTC m=+7596.760531181" observedRunningTime="2025-10-14 09:07:41.720012199 +0000 UTC m=+7597.417372610" watchObservedRunningTime="2025-10-14 09:07:41.721769753 +0000 UTC m=+7597.419130144" Oct 14 09:07:43 crc kubenswrapper[4870]: I1014 09:07:43.600510 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/prometheus-metric-storage-0" Oct 14 09:07:43 crc kubenswrapper[4870]: I1014 09:07:43.601465 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/prometheus-metric-storage-0" Oct 14 09:07:43 crc kubenswrapper[4870]: I1014 09:07:43.605108 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/prometheus-metric-storage-0" Oct 14 09:07:43 crc kubenswrapper[4870]: I1014 09:07:43.711823 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/prometheus-metric-storage-0" Oct 14 09:07:44 crc kubenswrapper[4870]: I1014 09:07:44.047363 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-v5vqn"] Oct 14 09:07:44 crc kubenswrapper[4870]: I1014 09:07:44.066274 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-v5vqn"] Oct 14 09:07:45 crc kubenswrapper[4870]: I1014 09:07:45.043871 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6e1e7452-1d68-46e7-872f-973667e7d3d6" path="/var/lib/kubelet/pods/6e1e7452-1d68-46e7-872f-973667e7d3d6/volumes" Oct 14 09:07:46 crc kubenswrapper[4870]: I1014 09:07:46.236537 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 14 09:07:46 crc kubenswrapper[4870]: I1014 09:07:46.240190 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 14 09:07:46 crc kubenswrapper[4870]: I1014 09:07:46.249868 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 14 09:07:46 crc kubenswrapper[4870]: I1014 09:07:46.249968 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 14 09:07:46 crc kubenswrapper[4870]: I1014 09:07:46.250016 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 14 09:07:46 crc kubenswrapper[4870]: I1014 09:07:46.367268 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/620a8d3a-3b05-4ea5-bd1a-6f40433d11a9-run-httpd\") pod \"ceilometer-0\" (UID: \"620a8d3a-3b05-4ea5-bd1a-6f40433d11a9\") " pod="openstack/ceilometer-0" Oct 14 09:07:46 crc kubenswrapper[4870]: I1014 09:07:46.367361 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/620a8d3a-3b05-4ea5-bd1a-6f40433d11a9-scripts\") pod \"ceilometer-0\" (UID: \"620a8d3a-3b05-4ea5-bd1a-6f40433d11a9\") " pod="openstack/ceilometer-0" Oct 14 09:07:46 crc kubenswrapper[4870]: I1014 09:07:46.367455 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fzbkb\" (UniqueName: \"kubernetes.io/projected/620a8d3a-3b05-4ea5-bd1a-6f40433d11a9-kube-api-access-fzbkb\") pod \"ceilometer-0\" (UID: \"620a8d3a-3b05-4ea5-bd1a-6f40433d11a9\") " pod="openstack/ceilometer-0" Oct 14 09:07:46 crc kubenswrapper[4870]: I1014 09:07:46.367506 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/620a8d3a-3b05-4ea5-bd1a-6f40433d11a9-config-data\") pod \"ceilometer-0\" (UID: \"620a8d3a-3b05-4ea5-bd1a-6f40433d11a9\") " pod="openstack/ceilometer-0" Oct 14 09:07:46 crc kubenswrapper[4870]: I1014 09:07:46.367586 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/620a8d3a-3b05-4ea5-bd1a-6f40433d11a9-log-httpd\") pod \"ceilometer-0\" (UID: \"620a8d3a-3b05-4ea5-bd1a-6f40433d11a9\") " pod="openstack/ceilometer-0" Oct 14 09:07:46 crc kubenswrapper[4870]: I1014 09:07:46.367613 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/620a8d3a-3b05-4ea5-bd1a-6f40433d11a9-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"620a8d3a-3b05-4ea5-bd1a-6f40433d11a9\") " pod="openstack/ceilometer-0" Oct 14 09:07:46 crc kubenswrapper[4870]: I1014 09:07:46.367648 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/620a8d3a-3b05-4ea5-bd1a-6f40433d11a9-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"620a8d3a-3b05-4ea5-bd1a-6f40433d11a9\") " pod="openstack/ceilometer-0" Oct 14 09:07:46 crc kubenswrapper[4870]: I1014 09:07:46.468913 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/620a8d3a-3b05-4ea5-bd1a-6f40433d11a9-log-httpd\") pod \"ceilometer-0\" (UID: \"620a8d3a-3b05-4ea5-bd1a-6f40433d11a9\") " pod="openstack/ceilometer-0" Oct 14 09:07:46 crc kubenswrapper[4870]: I1014 09:07:46.468973 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/620a8d3a-3b05-4ea5-bd1a-6f40433d11a9-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"620a8d3a-3b05-4ea5-bd1a-6f40433d11a9\") " pod="openstack/ceilometer-0" Oct 14 09:07:46 crc kubenswrapper[4870]: I1014 09:07:46.469018 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/620a8d3a-3b05-4ea5-bd1a-6f40433d11a9-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"620a8d3a-3b05-4ea5-bd1a-6f40433d11a9\") " pod="openstack/ceilometer-0" Oct 14 09:07:46 crc kubenswrapper[4870]: I1014 09:07:46.469048 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/620a8d3a-3b05-4ea5-bd1a-6f40433d11a9-run-httpd\") pod \"ceilometer-0\" (UID: \"620a8d3a-3b05-4ea5-bd1a-6f40433d11a9\") " pod="openstack/ceilometer-0" Oct 14 09:07:46 crc kubenswrapper[4870]: I1014 09:07:46.469096 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/620a8d3a-3b05-4ea5-bd1a-6f40433d11a9-scripts\") pod \"ceilometer-0\" (UID: \"620a8d3a-3b05-4ea5-bd1a-6f40433d11a9\") " pod="openstack/ceilometer-0" Oct 14 09:07:46 crc kubenswrapper[4870]: I1014 09:07:46.469148 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fzbkb\" (UniqueName: \"kubernetes.io/projected/620a8d3a-3b05-4ea5-bd1a-6f40433d11a9-kube-api-access-fzbkb\") pod \"ceilometer-0\" (UID: \"620a8d3a-3b05-4ea5-bd1a-6f40433d11a9\") " pod="openstack/ceilometer-0" Oct 14 09:07:46 crc kubenswrapper[4870]: I1014 09:07:46.469185 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/620a8d3a-3b05-4ea5-bd1a-6f40433d11a9-config-data\") pod \"ceilometer-0\" (UID: \"620a8d3a-3b05-4ea5-bd1a-6f40433d11a9\") " pod="openstack/ceilometer-0" Oct 14 09:07:46 crc kubenswrapper[4870]: I1014 09:07:46.469698 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/620a8d3a-3b05-4ea5-bd1a-6f40433d11a9-run-httpd\") pod \"ceilometer-0\" (UID: \"620a8d3a-3b05-4ea5-bd1a-6f40433d11a9\") " pod="openstack/ceilometer-0" Oct 14 09:07:46 crc kubenswrapper[4870]: I1014 09:07:46.469982 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/620a8d3a-3b05-4ea5-bd1a-6f40433d11a9-log-httpd\") pod \"ceilometer-0\" (UID: \"620a8d3a-3b05-4ea5-bd1a-6f40433d11a9\") " pod="openstack/ceilometer-0" Oct 14 09:07:46 crc kubenswrapper[4870]: I1014 09:07:46.476096 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/620a8d3a-3b05-4ea5-bd1a-6f40433d11a9-config-data\") pod \"ceilometer-0\" (UID: \"620a8d3a-3b05-4ea5-bd1a-6f40433d11a9\") " pod="openstack/ceilometer-0" Oct 14 09:07:46 crc kubenswrapper[4870]: I1014 09:07:46.477217 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/620a8d3a-3b05-4ea5-bd1a-6f40433d11a9-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"620a8d3a-3b05-4ea5-bd1a-6f40433d11a9\") " pod="openstack/ceilometer-0" Oct 14 09:07:46 crc kubenswrapper[4870]: I1014 09:07:46.478892 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/620a8d3a-3b05-4ea5-bd1a-6f40433d11a9-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"620a8d3a-3b05-4ea5-bd1a-6f40433d11a9\") " pod="openstack/ceilometer-0" Oct 14 09:07:46 crc kubenswrapper[4870]: I1014 09:07:46.485608 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/620a8d3a-3b05-4ea5-bd1a-6f40433d11a9-scripts\") pod \"ceilometer-0\" (UID: \"620a8d3a-3b05-4ea5-bd1a-6f40433d11a9\") " pod="openstack/ceilometer-0" Oct 14 09:07:46 crc kubenswrapper[4870]: I1014 09:07:46.491402 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fzbkb\" (UniqueName: \"kubernetes.io/projected/620a8d3a-3b05-4ea5-bd1a-6f40433d11a9-kube-api-access-fzbkb\") pod \"ceilometer-0\" (UID: \"620a8d3a-3b05-4ea5-bd1a-6f40433d11a9\") " pod="openstack/ceilometer-0" Oct 14 09:07:46 crc kubenswrapper[4870]: I1014 09:07:46.572174 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 14 09:07:47 crc kubenswrapper[4870]: I1014 09:07:47.106345 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 14 09:07:47 crc kubenswrapper[4870]: W1014 09:07:47.114535 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod620a8d3a_3b05_4ea5_bd1a_6f40433d11a9.slice/crio-e5c205781105471151c48f465e495dc36465d062c20dc52e4cb0856ce1a319dc WatchSource:0}: Error finding container e5c205781105471151c48f465e495dc36465d062c20dc52e4cb0856ce1a319dc: Status 404 returned error can't find the container with id e5c205781105471151c48f465e495dc36465d062c20dc52e4cb0856ce1a319dc Oct 14 09:07:47 crc kubenswrapper[4870]: I1014 09:07:47.780269 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"620a8d3a-3b05-4ea5-bd1a-6f40433d11a9","Type":"ContainerStarted","Data":"e5c205781105471151c48f465e495dc36465d062c20dc52e4cb0856ce1a319dc"} Oct 14 09:07:48 crc kubenswrapper[4870]: I1014 09:07:48.034004 4870 scope.go:117] "RemoveContainer" containerID="8fcaeb3090f0c89cba1e66db524b4a246072daa35592012664aca450fa09ddc9" Oct 14 09:07:48 crc kubenswrapper[4870]: E1014 09:07:48.034313 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 09:07:51 crc kubenswrapper[4870]: I1014 09:07:51.827290 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"620a8d3a-3b05-4ea5-bd1a-6f40433d11a9","Type":"ContainerStarted","Data":"533ef6e22525e1381bb31273c227ba87cf77eceb2c8712d18f5e5f95a4c2892a"} Oct 14 09:07:52 crc kubenswrapper[4870]: I1014 09:07:52.851665 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"620a8d3a-3b05-4ea5-bd1a-6f40433d11a9","Type":"ContainerStarted","Data":"0cb67aa3076c37ce75e040a7fda40630104e3813c56d7660b2b9918616bd68ac"} Oct 14 09:07:54 crc kubenswrapper[4870]: I1014 09:07:54.041571 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-bfc7-account-create-9n2mf"] Oct 14 09:07:54 crc kubenswrapper[4870]: I1014 09:07:54.067573 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-bfc7-account-create-9n2mf"] Oct 14 09:07:54 crc kubenswrapper[4870]: I1014 09:07:54.874130 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"620a8d3a-3b05-4ea5-bd1a-6f40433d11a9","Type":"ContainerStarted","Data":"e91ba300647ffc4556c5aa3dd7a1c6268471cb72acd6d40f651b16b77b541548"} Oct 14 09:07:55 crc kubenswrapper[4870]: I1014 09:07:55.064950 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="21b0ad50-1397-43e9-97d1-2bd8a4aff236" path="/var/lib/kubelet/pods/21b0ad50-1397-43e9-97d1-2bd8a4aff236/volumes" Oct 14 09:07:55 crc kubenswrapper[4870]: I1014 09:07:55.894339 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"620a8d3a-3b05-4ea5-bd1a-6f40433d11a9","Type":"ContainerStarted","Data":"c55f79de5e91cc683b6fedc9c34be9ae3402ba92f16f1bd24bd57d2b0d1093d6"} Oct 14 09:07:55 crc kubenswrapper[4870]: I1014 09:07:55.894746 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 14 09:07:55 crc kubenswrapper[4870]: I1014 09:07:55.927650 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.636369466 podStartE2EDuration="9.927631723s" podCreationTimestamp="2025-10-14 09:07:46 +0000 UTC" firstStartedPulling="2025-10-14 09:07:47.116966541 +0000 UTC m=+7602.814326912" lastFinishedPulling="2025-10-14 09:07:55.408228788 +0000 UTC m=+7611.105589169" observedRunningTime="2025-10-14 09:07:55.925148402 +0000 UTC m=+7611.622508783" watchObservedRunningTime="2025-10-14 09:07:55.927631723 +0000 UTC m=+7611.624992114" Oct 14 09:08:00 crc kubenswrapper[4870]: I1014 09:08:00.035132 4870 scope.go:117] "RemoveContainer" containerID="8fcaeb3090f0c89cba1e66db524b4a246072daa35592012664aca450fa09ddc9" Oct 14 09:08:00 crc kubenswrapper[4870]: E1014 09:08:00.036536 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 09:08:02 crc kubenswrapper[4870]: I1014 09:08:02.864285 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-db-create-zhtzj"] Oct 14 09:08:02 crc kubenswrapper[4870]: I1014 09:08:02.866269 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-zhtzj" Oct 14 09:08:02 crc kubenswrapper[4870]: I1014 09:08:02.913745 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-create-zhtzj"] Oct 14 09:08:02 crc kubenswrapper[4870]: I1014 09:08:02.923545 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rzxnv\" (UniqueName: \"kubernetes.io/projected/89dc0446-9ebc-48d8-99e9-b77564ef3273-kube-api-access-rzxnv\") pod \"aodh-db-create-zhtzj\" (UID: \"89dc0446-9ebc-48d8-99e9-b77564ef3273\") " pod="openstack/aodh-db-create-zhtzj" Oct 14 09:08:03 crc kubenswrapper[4870]: I1014 09:08:03.025710 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rzxnv\" (UniqueName: \"kubernetes.io/projected/89dc0446-9ebc-48d8-99e9-b77564ef3273-kube-api-access-rzxnv\") pod \"aodh-db-create-zhtzj\" (UID: \"89dc0446-9ebc-48d8-99e9-b77564ef3273\") " pod="openstack/aodh-db-create-zhtzj" Oct 14 09:08:03 crc kubenswrapper[4870]: I1014 09:08:03.042386 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-kfqr4"] Oct 14 09:08:03 crc kubenswrapper[4870]: I1014 09:08:03.069205 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rzxnv\" (UniqueName: \"kubernetes.io/projected/89dc0446-9ebc-48d8-99e9-b77564ef3273-kube-api-access-rzxnv\") pod \"aodh-db-create-zhtzj\" (UID: \"89dc0446-9ebc-48d8-99e9-b77564ef3273\") " pod="openstack/aodh-db-create-zhtzj" Oct 14 09:08:03 crc kubenswrapper[4870]: I1014 09:08:03.073008 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-kfqr4"] Oct 14 09:08:03 crc kubenswrapper[4870]: I1014 09:08:03.192510 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-zhtzj" Oct 14 09:08:03 crc kubenswrapper[4870]: I1014 09:08:03.766379 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-create-zhtzj"] Oct 14 09:08:03 crc kubenswrapper[4870]: W1014 09:08:03.774697 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod89dc0446_9ebc_48d8_99e9_b77564ef3273.slice/crio-c68354fb1a1663c96e6392446b2ca8af7e1a17955eb4d07f2bac9d6af6decfd8 WatchSource:0}: Error finding container c68354fb1a1663c96e6392446b2ca8af7e1a17955eb4d07f2bac9d6af6decfd8: Status 404 returned error can't find the container with id c68354fb1a1663c96e6392446b2ca8af7e1a17955eb4d07f2bac9d6af6decfd8 Oct 14 09:08:03 crc kubenswrapper[4870]: I1014 09:08:03.990776 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-create-zhtzj" event={"ID":"89dc0446-9ebc-48d8-99e9-b77564ef3273","Type":"ContainerStarted","Data":"c68354fb1a1663c96e6392446b2ca8af7e1a17955eb4d07f2bac9d6af6decfd8"} Oct 14 09:08:05 crc kubenswrapper[4870]: I1014 09:08:05.013196 4870 generic.go:334] "Generic (PLEG): container finished" podID="89dc0446-9ebc-48d8-99e9-b77564ef3273" containerID="9db545bb783b2952b072caa15b2a293309e1ea6f049f61a0357c2dc02fd1dc15" exitCode=0 Oct 14 09:08:05 crc kubenswrapper[4870]: I1014 09:08:05.013258 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-create-zhtzj" event={"ID":"89dc0446-9ebc-48d8-99e9-b77564ef3273","Type":"ContainerDied","Data":"9db545bb783b2952b072caa15b2a293309e1ea6f049f61a0357c2dc02fd1dc15"} Oct 14 09:08:05 crc kubenswrapper[4870]: I1014 09:08:05.054676 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="32cde066-88bb-4e2b-b276-227282560585" path="/var/lib/kubelet/pods/32cde066-88bb-4e2b-b276-227282560585/volumes" Oct 14 09:08:06 crc kubenswrapper[4870]: I1014 09:08:06.441062 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-zhtzj" Oct 14 09:08:06 crc kubenswrapper[4870]: I1014 09:08:06.599719 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rzxnv\" (UniqueName: \"kubernetes.io/projected/89dc0446-9ebc-48d8-99e9-b77564ef3273-kube-api-access-rzxnv\") pod \"89dc0446-9ebc-48d8-99e9-b77564ef3273\" (UID: \"89dc0446-9ebc-48d8-99e9-b77564ef3273\") " Oct 14 09:08:06 crc kubenswrapper[4870]: I1014 09:08:06.607228 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/89dc0446-9ebc-48d8-99e9-b77564ef3273-kube-api-access-rzxnv" (OuterVolumeSpecName: "kube-api-access-rzxnv") pod "89dc0446-9ebc-48d8-99e9-b77564ef3273" (UID: "89dc0446-9ebc-48d8-99e9-b77564ef3273"). InnerVolumeSpecName "kube-api-access-rzxnv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:08:06 crc kubenswrapper[4870]: I1014 09:08:06.703324 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rzxnv\" (UniqueName: \"kubernetes.io/projected/89dc0446-9ebc-48d8-99e9-b77564ef3273-kube-api-access-rzxnv\") on node \"crc\" DevicePath \"\"" Oct 14 09:08:07 crc kubenswrapper[4870]: I1014 09:08:07.037612 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-zhtzj" Oct 14 09:08:07 crc kubenswrapper[4870]: I1014 09:08:07.058619 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-create-zhtzj" event={"ID":"89dc0446-9ebc-48d8-99e9-b77564ef3273","Type":"ContainerDied","Data":"c68354fb1a1663c96e6392446b2ca8af7e1a17955eb4d07f2bac9d6af6decfd8"} Oct 14 09:08:07 crc kubenswrapper[4870]: I1014 09:08:07.058689 4870 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c68354fb1a1663c96e6392446b2ca8af7e1a17955eb4d07f2bac9d6af6decfd8" Oct 14 09:08:12 crc kubenswrapper[4870]: I1014 09:08:12.913685 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-479c-account-create-nsqjn"] Oct 14 09:08:12 crc kubenswrapper[4870]: E1014 09:08:12.914612 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="89dc0446-9ebc-48d8-99e9-b77564ef3273" containerName="mariadb-database-create" Oct 14 09:08:12 crc kubenswrapper[4870]: I1014 09:08:12.914629 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="89dc0446-9ebc-48d8-99e9-b77564ef3273" containerName="mariadb-database-create" Oct 14 09:08:12 crc kubenswrapper[4870]: I1014 09:08:12.914914 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="89dc0446-9ebc-48d8-99e9-b77564ef3273" containerName="mariadb-database-create" Oct 14 09:08:12 crc kubenswrapper[4870]: I1014 09:08:12.915844 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-479c-account-create-nsqjn" Oct 14 09:08:12 crc kubenswrapper[4870]: I1014 09:08:12.918312 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-db-secret" Oct 14 09:08:12 crc kubenswrapper[4870]: I1014 09:08:12.928728 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-479c-account-create-nsqjn"] Oct 14 09:08:12 crc kubenswrapper[4870]: I1014 09:08:12.960029 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4qs4k\" (UniqueName: \"kubernetes.io/projected/b251834e-2ef7-4906-8452-4fa62834bf3a-kube-api-access-4qs4k\") pod \"aodh-479c-account-create-nsqjn\" (UID: \"b251834e-2ef7-4906-8452-4fa62834bf3a\") " pod="openstack/aodh-479c-account-create-nsqjn" Oct 14 09:08:13 crc kubenswrapper[4870]: I1014 09:08:13.062509 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4qs4k\" (UniqueName: \"kubernetes.io/projected/b251834e-2ef7-4906-8452-4fa62834bf3a-kube-api-access-4qs4k\") pod \"aodh-479c-account-create-nsqjn\" (UID: \"b251834e-2ef7-4906-8452-4fa62834bf3a\") " pod="openstack/aodh-479c-account-create-nsqjn" Oct 14 09:08:13 crc kubenswrapper[4870]: I1014 09:08:13.080332 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4qs4k\" (UniqueName: \"kubernetes.io/projected/b251834e-2ef7-4906-8452-4fa62834bf3a-kube-api-access-4qs4k\") pod \"aodh-479c-account-create-nsqjn\" (UID: \"b251834e-2ef7-4906-8452-4fa62834bf3a\") " pod="openstack/aodh-479c-account-create-nsqjn" Oct 14 09:08:13 crc kubenswrapper[4870]: I1014 09:08:13.242186 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-479c-account-create-nsqjn" Oct 14 09:08:13 crc kubenswrapper[4870]: W1014 09:08:13.798354 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb251834e_2ef7_4906_8452_4fa62834bf3a.slice/crio-2e995d45f35abda55713680eff759cb842e645fda9e4d412a4e0cd3d9b3856c2 WatchSource:0}: Error finding container 2e995d45f35abda55713680eff759cb842e645fda9e4d412a4e0cd3d9b3856c2: Status 404 returned error can't find the container with id 2e995d45f35abda55713680eff759cb842e645fda9e4d412a4e0cd3d9b3856c2 Oct 14 09:08:13 crc kubenswrapper[4870]: I1014 09:08:13.802047 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-479c-account-create-nsqjn"] Oct 14 09:08:14 crc kubenswrapper[4870]: I1014 09:08:14.120944 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-479c-account-create-nsqjn" event={"ID":"b251834e-2ef7-4906-8452-4fa62834bf3a","Type":"ContainerStarted","Data":"15be86f675fd9ba70a551bd735e24b0141a57be493b09079edb6bf957d6ee554"} Oct 14 09:08:14 crc kubenswrapper[4870]: I1014 09:08:14.121038 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-479c-account-create-nsqjn" event={"ID":"b251834e-2ef7-4906-8452-4fa62834bf3a","Type":"ContainerStarted","Data":"2e995d45f35abda55713680eff759cb842e645fda9e4d412a4e0cd3d9b3856c2"} Oct 14 09:08:14 crc kubenswrapper[4870]: I1014 09:08:14.150625 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-479c-account-create-nsqjn" podStartSLOduration=2.150599223 podStartE2EDuration="2.150599223s" podCreationTimestamp="2025-10-14 09:08:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 09:08:14.141806496 +0000 UTC m=+7629.839166887" watchObservedRunningTime="2025-10-14 09:08:14.150599223 +0000 UTC m=+7629.847959634" Oct 14 09:08:15 crc kubenswrapper[4870]: I1014 09:08:15.045695 4870 scope.go:117] "RemoveContainer" containerID="8fcaeb3090f0c89cba1e66db524b4a246072daa35592012664aca450fa09ddc9" Oct 14 09:08:15 crc kubenswrapper[4870]: E1014 09:08:15.045981 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 09:08:15 crc kubenswrapper[4870]: I1014 09:08:15.134347 4870 generic.go:334] "Generic (PLEG): container finished" podID="b251834e-2ef7-4906-8452-4fa62834bf3a" containerID="15be86f675fd9ba70a551bd735e24b0141a57be493b09079edb6bf957d6ee554" exitCode=0 Oct 14 09:08:15 crc kubenswrapper[4870]: I1014 09:08:15.134430 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-479c-account-create-nsqjn" event={"ID":"b251834e-2ef7-4906-8452-4fa62834bf3a","Type":"ContainerDied","Data":"15be86f675fd9ba70a551bd735e24b0141a57be493b09079edb6bf957d6ee554"} Oct 14 09:08:16 crc kubenswrapper[4870]: I1014 09:08:16.591769 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Oct 14 09:08:16 crc kubenswrapper[4870]: I1014 09:08:16.670330 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-479c-account-create-nsqjn" Oct 14 09:08:16 crc kubenswrapper[4870]: I1014 09:08:16.853229 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4qs4k\" (UniqueName: \"kubernetes.io/projected/b251834e-2ef7-4906-8452-4fa62834bf3a-kube-api-access-4qs4k\") pod \"b251834e-2ef7-4906-8452-4fa62834bf3a\" (UID: \"b251834e-2ef7-4906-8452-4fa62834bf3a\") " Oct 14 09:08:16 crc kubenswrapper[4870]: I1014 09:08:16.862623 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b251834e-2ef7-4906-8452-4fa62834bf3a-kube-api-access-4qs4k" (OuterVolumeSpecName: "kube-api-access-4qs4k") pod "b251834e-2ef7-4906-8452-4fa62834bf3a" (UID: "b251834e-2ef7-4906-8452-4fa62834bf3a"). InnerVolumeSpecName "kube-api-access-4qs4k". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:08:16 crc kubenswrapper[4870]: I1014 09:08:16.956084 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4qs4k\" (UniqueName: \"kubernetes.io/projected/b251834e-2ef7-4906-8452-4fa62834bf3a-kube-api-access-4qs4k\") on node \"crc\" DevicePath \"\"" Oct 14 09:08:17 crc kubenswrapper[4870]: I1014 09:08:17.165152 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-479c-account-create-nsqjn" event={"ID":"b251834e-2ef7-4906-8452-4fa62834bf3a","Type":"ContainerDied","Data":"2e995d45f35abda55713680eff759cb842e645fda9e4d412a4e0cd3d9b3856c2"} Oct 14 09:08:17 crc kubenswrapper[4870]: I1014 09:08:17.165207 4870 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2e995d45f35abda55713680eff759cb842e645fda9e4d412a4e0cd3d9b3856c2" Oct 14 09:08:17 crc kubenswrapper[4870]: I1014 09:08:17.165237 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-479c-account-create-nsqjn" Oct 14 09:08:18 crc kubenswrapper[4870]: I1014 09:08:18.453673 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-db-sync-zkj55"] Oct 14 09:08:18 crc kubenswrapper[4870]: E1014 09:08:18.454616 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b251834e-2ef7-4906-8452-4fa62834bf3a" containerName="mariadb-account-create" Oct 14 09:08:18 crc kubenswrapper[4870]: I1014 09:08:18.455092 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="b251834e-2ef7-4906-8452-4fa62834bf3a" containerName="mariadb-account-create" Oct 14 09:08:18 crc kubenswrapper[4870]: I1014 09:08:18.455740 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="b251834e-2ef7-4906-8452-4fa62834bf3a" containerName="mariadb-account-create" Oct 14 09:08:18 crc kubenswrapper[4870]: I1014 09:08:18.457387 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-zkj55" Oct 14 09:08:18 crc kubenswrapper[4870]: I1014 09:08:18.462083 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-scripts" Oct 14 09:08:18 crc kubenswrapper[4870]: I1014 09:08:18.462196 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-config-data" Oct 14 09:08:18 crc kubenswrapper[4870]: I1014 09:08:18.462216 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-autoscaling-dockercfg-wstph" Oct 14 09:08:18 crc kubenswrapper[4870]: I1014 09:08:18.471009 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-sync-zkj55"] Oct 14 09:08:18 crc kubenswrapper[4870]: I1014 09:08:18.589295 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mhgwl\" (UniqueName: \"kubernetes.io/projected/deb1d307-0ab5-4053-a68e-513f2d7b0d53-kube-api-access-mhgwl\") pod \"aodh-db-sync-zkj55\" (UID: \"deb1d307-0ab5-4053-a68e-513f2d7b0d53\") " pod="openstack/aodh-db-sync-zkj55" Oct 14 09:08:18 crc kubenswrapper[4870]: I1014 09:08:18.589389 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/deb1d307-0ab5-4053-a68e-513f2d7b0d53-scripts\") pod \"aodh-db-sync-zkj55\" (UID: \"deb1d307-0ab5-4053-a68e-513f2d7b0d53\") " pod="openstack/aodh-db-sync-zkj55" Oct 14 09:08:18 crc kubenswrapper[4870]: I1014 09:08:18.589900 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/deb1d307-0ab5-4053-a68e-513f2d7b0d53-config-data\") pod \"aodh-db-sync-zkj55\" (UID: \"deb1d307-0ab5-4053-a68e-513f2d7b0d53\") " pod="openstack/aodh-db-sync-zkj55" Oct 14 09:08:18 crc kubenswrapper[4870]: I1014 09:08:18.589986 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/deb1d307-0ab5-4053-a68e-513f2d7b0d53-combined-ca-bundle\") pod \"aodh-db-sync-zkj55\" (UID: \"deb1d307-0ab5-4053-a68e-513f2d7b0d53\") " pod="openstack/aodh-db-sync-zkj55" Oct 14 09:08:18 crc kubenswrapper[4870]: I1014 09:08:18.692686 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/deb1d307-0ab5-4053-a68e-513f2d7b0d53-config-data\") pod \"aodh-db-sync-zkj55\" (UID: \"deb1d307-0ab5-4053-a68e-513f2d7b0d53\") " pod="openstack/aodh-db-sync-zkj55" Oct 14 09:08:18 crc kubenswrapper[4870]: I1014 09:08:18.692804 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/deb1d307-0ab5-4053-a68e-513f2d7b0d53-combined-ca-bundle\") pod \"aodh-db-sync-zkj55\" (UID: \"deb1d307-0ab5-4053-a68e-513f2d7b0d53\") " pod="openstack/aodh-db-sync-zkj55" Oct 14 09:08:18 crc kubenswrapper[4870]: I1014 09:08:18.692991 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mhgwl\" (UniqueName: \"kubernetes.io/projected/deb1d307-0ab5-4053-a68e-513f2d7b0d53-kube-api-access-mhgwl\") pod \"aodh-db-sync-zkj55\" (UID: \"deb1d307-0ab5-4053-a68e-513f2d7b0d53\") " pod="openstack/aodh-db-sync-zkj55" Oct 14 09:08:18 crc kubenswrapper[4870]: I1014 09:08:18.693122 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/deb1d307-0ab5-4053-a68e-513f2d7b0d53-scripts\") pod \"aodh-db-sync-zkj55\" (UID: \"deb1d307-0ab5-4053-a68e-513f2d7b0d53\") " pod="openstack/aodh-db-sync-zkj55" Oct 14 09:08:18 crc kubenswrapper[4870]: I1014 09:08:18.697514 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/deb1d307-0ab5-4053-a68e-513f2d7b0d53-combined-ca-bundle\") pod \"aodh-db-sync-zkj55\" (UID: \"deb1d307-0ab5-4053-a68e-513f2d7b0d53\") " pod="openstack/aodh-db-sync-zkj55" Oct 14 09:08:18 crc kubenswrapper[4870]: I1014 09:08:18.698344 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/deb1d307-0ab5-4053-a68e-513f2d7b0d53-config-data\") pod \"aodh-db-sync-zkj55\" (UID: \"deb1d307-0ab5-4053-a68e-513f2d7b0d53\") " pod="openstack/aodh-db-sync-zkj55" Oct 14 09:08:18 crc kubenswrapper[4870]: I1014 09:08:18.699916 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/deb1d307-0ab5-4053-a68e-513f2d7b0d53-scripts\") pod \"aodh-db-sync-zkj55\" (UID: \"deb1d307-0ab5-4053-a68e-513f2d7b0d53\") " pod="openstack/aodh-db-sync-zkj55" Oct 14 09:08:18 crc kubenswrapper[4870]: I1014 09:08:18.721134 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mhgwl\" (UniqueName: \"kubernetes.io/projected/deb1d307-0ab5-4053-a68e-513f2d7b0d53-kube-api-access-mhgwl\") pod \"aodh-db-sync-zkj55\" (UID: \"deb1d307-0ab5-4053-a68e-513f2d7b0d53\") " pod="openstack/aodh-db-sync-zkj55" Oct 14 09:08:18 crc kubenswrapper[4870]: I1014 09:08:18.779374 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-zkj55" Oct 14 09:08:19 crc kubenswrapper[4870]: I1014 09:08:19.308410 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-sync-zkj55"] Oct 14 09:08:20 crc kubenswrapper[4870]: I1014 09:08:20.195370 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-zkj55" event={"ID":"deb1d307-0ab5-4053-a68e-513f2d7b0d53","Type":"ContainerStarted","Data":"7f28dcdf5395dfa9c108cd9af291f42355c0ccf01f49bcffced2da0b622979e2"} Oct 14 09:08:24 crc kubenswrapper[4870]: I1014 09:08:24.254543 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-zkj55" event={"ID":"deb1d307-0ab5-4053-a68e-513f2d7b0d53","Type":"ContainerStarted","Data":"fb1fa218005cc3197d0da536bf6930392c1110ab18fdfc98181646cb4f8bf54d"} Oct 14 09:08:24 crc kubenswrapper[4870]: I1014 09:08:24.279139 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-db-sync-zkj55" podStartSLOduration=1.990493141 podStartE2EDuration="6.279123626s" podCreationTimestamp="2025-10-14 09:08:18 +0000 UTC" firstStartedPulling="2025-10-14 09:08:19.356710651 +0000 UTC m=+7635.054071022" lastFinishedPulling="2025-10-14 09:08:23.645341126 +0000 UTC m=+7639.342701507" observedRunningTime="2025-10-14 09:08:24.274985914 +0000 UTC m=+7639.972346305" watchObservedRunningTime="2025-10-14 09:08:24.279123626 +0000 UTC m=+7639.976483997" Oct 14 09:08:26 crc kubenswrapper[4870]: I1014 09:08:26.278079 4870 generic.go:334] "Generic (PLEG): container finished" podID="deb1d307-0ab5-4053-a68e-513f2d7b0d53" containerID="fb1fa218005cc3197d0da536bf6930392c1110ab18fdfc98181646cb4f8bf54d" exitCode=0 Oct 14 09:08:26 crc kubenswrapper[4870]: I1014 09:08:26.278156 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-zkj55" event={"ID":"deb1d307-0ab5-4053-a68e-513f2d7b0d53","Type":"ContainerDied","Data":"fb1fa218005cc3197d0da536bf6930392c1110ab18fdfc98181646cb4f8bf54d"} Oct 14 09:08:27 crc kubenswrapper[4870]: I1014 09:08:27.034638 4870 scope.go:117] "RemoveContainer" containerID="8fcaeb3090f0c89cba1e66db524b4a246072daa35592012664aca450fa09ddc9" Oct 14 09:08:27 crc kubenswrapper[4870]: E1014 09:08:27.035178 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 09:08:27 crc kubenswrapper[4870]: I1014 09:08:27.711350 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-zkj55" Oct 14 09:08:27 crc kubenswrapper[4870]: I1014 09:08:27.810257 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/deb1d307-0ab5-4053-a68e-513f2d7b0d53-config-data\") pod \"deb1d307-0ab5-4053-a68e-513f2d7b0d53\" (UID: \"deb1d307-0ab5-4053-a68e-513f2d7b0d53\") " Oct 14 09:08:27 crc kubenswrapper[4870]: I1014 09:08:27.810534 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/deb1d307-0ab5-4053-a68e-513f2d7b0d53-combined-ca-bundle\") pod \"deb1d307-0ab5-4053-a68e-513f2d7b0d53\" (UID: \"deb1d307-0ab5-4053-a68e-513f2d7b0d53\") " Oct 14 09:08:27 crc kubenswrapper[4870]: I1014 09:08:27.810637 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mhgwl\" (UniqueName: \"kubernetes.io/projected/deb1d307-0ab5-4053-a68e-513f2d7b0d53-kube-api-access-mhgwl\") pod \"deb1d307-0ab5-4053-a68e-513f2d7b0d53\" (UID: \"deb1d307-0ab5-4053-a68e-513f2d7b0d53\") " Oct 14 09:08:27 crc kubenswrapper[4870]: I1014 09:08:27.810789 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/deb1d307-0ab5-4053-a68e-513f2d7b0d53-scripts\") pod \"deb1d307-0ab5-4053-a68e-513f2d7b0d53\" (UID: \"deb1d307-0ab5-4053-a68e-513f2d7b0d53\") " Oct 14 09:08:27 crc kubenswrapper[4870]: I1014 09:08:27.817220 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/deb1d307-0ab5-4053-a68e-513f2d7b0d53-kube-api-access-mhgwl" (OuterVolumeSpecName: "kube-api-access-mhgwl") pod "deb1d307-0ab5-4053-a68e-513f2d7b0d53" (UID: "deb1d307-0ab5-4053-a68e-513f2d7b0d53"). InnerVolumeSpecName "kube-api-access-mhgwl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:08:27 crc kubenswrapper[4870]: I1014 09:08:27.818502 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/deb1d307-0ab5-4053-a68e-513f2d7b0d53-scripts" (OuterVolumeSpecName: "scripts") pod "deb1d307-0ab5-4053-a68e-513f2d7b0d53" (UID: "deb1d307-0ab5-4053-a68e-513f2d7b0d53"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:08:27 crc kubenswrapper[4870]: I1014 09:08:27.851209 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/deb1d307-0ab5-4053-a68e-513f2d7b0d53-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "deb1d307-0ab5-4053-a68e-513f2d7b0d53" (UID: "deb1d307-0ab5-4053-a68e-513f2d7b0d53"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:08:27 crc kubenswrapper[4870]: I1014 09:08:27.859671 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/deb1d307-0ab5-4053-a68e-513f2d7b0d53-config-data" (OuterVolumeSpecName: "config-data") pod "deb1d307-0ab5-4053-a68e-513f2d7b0d53" (UID: "deb1d307-0ab5-4053-a68e-513f2d7b0d53"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:08:27 crc kubenswrapper[4870]: I1014 09:08:27.914747 4870 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/deb1d307-0ab5-4053-a68e-513f2d7b0d53-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 09:08:27 crc kubenswrapper[4870]: I1014 09:08:27.914787 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mhgwl\" (UniqueName: \"kubernetes.io/projected/deb1d307-0ab5-4053-a68e-513f2d7b0d53-kube-api-access-mhgwl\") on node \"crc\" DevicePath \"\"" Oct 14 09:08:27 crc kubenswrapper[4870]: I1014 09:08:27.914808 4870 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/deb1d307-0ab5-4053-a68e-513f2d7b0d53-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 09:08:27 crc kubenswrapper[4870]: I1014 09:08:27.914821 4870 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/deb1d307-0ab5-4053-a68e-513f2d7b0d53-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 09:08:28 crc kubenswrapper[4870]: I1014 09:08:28.301984 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-zkj55" event={"ID":"deb1d307-0ab5-4053-a68e-513f2d7b0d53","Type":"ContainerDied","Data":"7f28dcdf5395dfa9c108cd9af291f42355c0ccf01f49bcffced2da0b622979e2"} Oct 14 09:08:28 crc kubenswrapper[4870]: I1014 09:08:28.302041 4870 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7f28dcdf5395dfa9c108cd9af291f42355c0ccf01f49bcffced2da0b622979e2" Oct 14 09:08:28 crc kubenswrapper[4870]: I1014 09:08:28.302076 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-zkj55" Oct 14 09:08:29 crc kubenswrapper[4870]: I1014 09:08:29.823855 4870 scope.go:117] "RemoveContainer" containerID="adacb51ba823d60acf3fb6ca55a5842d901122fccb80169da2fe7d33d7baff7c" Oct 14 09:08:29 crc kubenswrapper[4870]: I1014 09:08:29.867873 4870 scope.go:117] "RemoveContainer" containerID="a26882018748ab7886af431748d8380dfd8592cd004d6110b333d865a49e851c" Oct 14 09:08:29 crc kubenswrapper[4870]: I1014 09:08:29.923519 4870 scope.go:117] "RemoveContainer" containerID="94d794a728ce92614be836c095267b61dda5f462fbda98a73e90746bc7ee4b6a" Oct 14 09:08:32 crc kubenswrapper[4870]: I1014 09:08:32.937683 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-0"] Oct 14 09:08:32 crc kubenswrapper[4870]: E1014 09:08:32.940875 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="deb1d307-0ab5-4053-a68e-513f2d7b0d53" containerName="aodh-db-sync" Oct 14 09:08:32 crc kubenswrapper[4870]: I1014 09:08:32.941209 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="deb1d307-0ab5-4053-a68e-513f2d7b0d53" containerName="aodh-db-sync" Oct 14 09:08:32 crc kubenswrapper[4870]: I1014 09:08:32.941635 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="deb1d307-0ab5-4053-a68e-513f2d7b0d53" containerName="aodh-db-sync" Oct 14 09:08:32 crc kubenswrapper[4870]: I1014 09:08:32.944880 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Oct 14 09:08:32 crc kubenswrapper[4870]: I1014 09:08:32.956055 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-config-data" Oct 14 09:08:32 crc kubenswrapper[4870]: I1014 09:08:32.956335 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-autoscaling-dockercfg-wstph" Oct 14 09:08:32 crc kubenswrapper[4870]: I1014 09:08:32.956568 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-scripts" Oct 14 09:08:32 crc kubenswrapper[4870]: I1014 09:08:32.959643 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Oct 14 09:08:33 crc kubenswrapper[4870]: I1014 09:08:33.036761 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/81cb33bd-91aa-4f54-a805-596957cb2b26-scripts\") pod \"aodh-0\" (UID: \"81cb33bd-91aa-4f54-a805-596957cb2b26\") " pod="openstack/aodh-0" Oct 14 09:08:33 crc kubenswrapper[4870]: I1014 09:08:33.036826 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h44rq\" (UniqueName: \"kubernetes.io/projected/81cb33bd-91aa-4f54-a805-596957cb2b26-kube-api-access-h44rq\") pod \"aodh-0\" (UID: \"81cb33bd-91aa-4f54-a805-596957cb2b26\") " pod="openstack/aodh-0" Oct 14 09:08:33 crc kubenswrapper[4870]: I1014 09:08:33.037441 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81cb33bd-91aa-4f54-a805-596957cb2b26-combined-ca-bundle\") pod \"aodh-0\" (UID: \"81cb33bd-91aa-4f54-a805-596957cb2b26\") " pod="openstack/aodh-0" Oct 14 09:08:33 crc kubenswrapper[4870]: I1014 09:08:33.037545 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/81cb33bd-91aa-4f54-a805-596957cb2b26-config-data\") pod \"aodh-0\" (UID: \"81cb33bd-91aa-4f54-a805-596957cb2b26\") " pod="openstack/aodh-0" Oct 14 09:08:33 crc kubenswrapper[4870]: I1014 09:08:33.140923 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81cb33bd-91aa-4f54-a805-596957cb2b26-combined-ca-bundle\") pod \"aodh-0\" (UID: \"81cb33bd-91aa-4f54-a805-596957cb2b26\") " pod="openstack/aodh-0" Oct 14 09:08:33 crc kubenswrapper[4870]: I1014 09:08:33.141001 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/81cb33bd-91aa-4f54-a805-596957cb2b26-config-data\") pod \"aodh-0\" (UID: \"81cb33bd-91aa-4f54-a805-596957cb2b26\") " pod="openstack/aodh-0" Oct 14 09:08:33 crc kubenswrapper[4870]: I1014 09:08:33.141124 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/81cb33bd-91aa-4f54-a805-596957cb2b26-scripts\") pod \"aodh-0\" (UID: \"81cb33bd-91aa-4f54-a805-596957cb2b26\") " pod="openstack/aodh-0" Oct 14 09:08:33 crc kubenswrapper[4870]: I1014 09:08:33.141158 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h44rq\" (UniqueName: \"kubernetes.io/projected/81cb33bd-91aa-4f54-a805-596957cb2b26-kube-api-access-h44rq\") pod \"aodh-0\" (UID: \"81cb33bd-91aa-4f54-a805-596957cb2b26\") " pod="openstack/aodh-0" Oct 14 09:08:33 crc kubenswrapper[4870]: I1014 09:08:33.158852 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81cb33bd-91aa-4f54-a805-596957cb2b26-combined-ca-bundle\") pod \"aodh-0\" (UID: \"81cb33bd-91aa-4f54-a805-596957cb2b26\") " pod="openstack/aodh-0" Oct 14 09:08:33 crc kubenswrapper[4870]: I1014 09:08:33.159211 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/81cb33bd-91aa-4f54-a805-596957cb2b26-config-data\") pod \"aodh-0\" (UID: \"81cb33bd-91aa-4f54-a805-596957cb2b26\") " pod="openstack/aodh-0" Oct 14 09:08:33 crc kubenswrapper[4870]: I1014 09:08:33.162198 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h44rq\" (UniqueName: \"kubernetes.io/projected/81cb33bd-91aa-4f54-a805-596957cb2b26-kube-api-access-h44rq\") pod \"aodh-0\" (UID: \"81cb33bd-91aa-4f54-a805-596957cb2b26\") " pod="openstack/aodh-0" Oct 14 09:08:33 crc kubenswrapper[4870]: I1014 09:08:33.166709 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/81cb33bd-91aa-4f54-a805-596957cb2b26-scripts\") pod \"aodh-0\" (UID: \"81cb33bd-91aa-4f54-a805-596957cb2b26\") " pod="openstack/aodh-0" Oct 14 09:08:33 crc kubenswrapper[4870]: I1014 09:08:33.315631 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Oct 14 09:08:33 crc kubenswrapper[4870]: I1014 09:08:33.785142 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Oct 14 09:08:34 crc kubenswrapper[4870]: I1014 09:08:34.379968 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"81cb33bd-91aa-4f54-a805-596957cb2b26","Type":"ContainerStarted","Data":"bd78a4e6d1a79a7c91f2b7b0079d02b0a848a03b5323dae7b149c6559a4a1af0"} Oct 14 09:08:34 crc kubenswrapper[4870]: I1014 09:08:34.380286 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"81cb33bd-91aa-4f54-a805-596957cb2b26","Type":"ContainerStarted","Data":"b5357f933aed557f0e7eea5421699c3262630bac139c6eb62632093a70dc7674"} Oct 14 09:08:35 crc kubenswrapper[4870]: I1014 09:08:35.366480 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 14 09:08:35 crc kubenswrapper[4870]: I1014 09:08:35.367133 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="620a8d3a-3b05-4ea5-bd1a-6f40433d11a9" containerName="ceilometer-central-agent" containerID="cri-o://533ef6e22525e1381bb31273c227ba87cf77eceb2c8712d18f5e5f95a4c2892a" gracePeriod=30 Oct 14 09:08:35 crc kubenswrapper[4870]: I1014 09:08:35.367252 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="620a8d3a-3b05-4ea5-bd1a-6f40433d11a9" containerName="proxy-httpd" containerID="cri-o://c55f79de5e91cc683b6fedc9c34be9ae3402ba92f16f1bd24bd57d2b0d1093d6" gracePeriod=30 Oct 14 09:08:35 crc kubenswrapper[4870]: I1014 09:08:35.367343 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="620a8d3a-3b05-4ea5-bd1a-6f40433d11a9" containerName="ceilometer-notification-agent" containerID="cri-o://0cb67aa3076c37ce75e040a7fda40630104e3813c56d7660b2b9918616bd68ac" gracePeriod=30 Oct 14 09:08:35 crc kubenswrapper[4870]: I1014 09:08:35.367476 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="620a8d3a-3b05-4ea5-bd1a-6f40433d11a9" containerName="sg-core" containerID="cri-o://e91ba300647ffc4556c5aa3dd7a1c6268471cb72acd6d40f651b16b77b541548" gracePeriod=30 Oct 14 09:08:35 crc kubenswrapper[4870]: I1014 09:08:35.395919 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"81cb33bd-91aa-4f54-a805-596957cb2b26","Type":"ContainerStarted","Data":"ac01a02e637834222ae587cbd0533b212182ba64d2d6324722bc0d0d4b2727de"} Oct 14 09:08:36 crc kubenswrapper[4870]: I1014 09:08:36.422535 4870 generic.go:334] "Generic (PLEG): container finished" podID="620a8d3a-3b05-4ea5-bd1a-6f40433d11a9" containerID="c55f79de5e91cc683b6fedc9c34be9ae3402ba92f16f1bd24bd57d2b0d1093d6" exitCode=0 Oct 14 09:08:36 crc kubenswrapper[4870]: I1014 09:08:36.423097 4870 generic.go:334] "Generic (PLEG): container finished" podID="620a8d3a-3b05-4ea5-bd1a-6f40433d11a9" containerID="e91ba300647ffc4556c5aa3dd7a1c6268471cb72acd6d40f651b16b77b541548" exitCode=2 Oct 14 09:08:36 crc kubenswrapper[4870]: I1014 09:08:36.423109 4870 generic.go:334] "Generic (PLEG): container finished" podID="620a8d3a-3b05-4ea5-bd1a-6f40433d11a9" containerID="533ef6e22525e1381bb31273c227ba87cf77eceb2c8712d18f5e5f95a4c2892a" exitCode=0 Oct 14 09:08:36 crc kubenswrapper[4870]: I1014 09:08:36.422614 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"620a8d3a-3b05-4ea5-bd1a-6f40433d11a9","Type":"ContainerDied","Data":"c55f79de5e91cc683b6fedc9c34be9ae3402ba92f16f1bd24bd57d2b0d1093d6"} Oct 14 09:08:36 crc kubenswrapper[4870]: I1014 09:08:36.423189 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"620a8d3a-3b05-4ea5-bd1a-6f40433d11a9","Type":"ContainerDied","Data":"e91ba300647ffc4556c5aa3dd7a1c6268471cb72acd6d40f651b16b77b541548"} Oct 14 09:08:36 crc kubenswrapper[4870]: I1014 09:08:36.423200 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"620a8d3a-3b05-4ea5-bd1a-6f40433d11a9","Type":"ContainerDied","Data":"533ef6e22525e1381bb31273c227ba87cf77eceb2c8712d18f5e5f95a4c2892a"} Oct 14 09:08:36 crc kubenswrapper[4870]: I1014 09:08:36.426616 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"81cb33bd-91aa-4f54-a805-596957cb2b26","Type":"ContainerStarted","Data":"994e2d14b092617d678dc73519237a771a375f800bf1a7dfe1ad7b88a59f7622"} Oct 14 09:08:38 crc kubenswrapper[4870]: I1014 09:08:38.446495 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"81cb33bd-91aa-4f54-a805-596957cb2b26","Type":"ContainerStarted","Data":"98938f704168a79df1e4d1eb88e4822a39408f69f398e5e8f704803fdcaddde8"} Oct 14 09:08:38 crc kubenswrapper[4870]: I1014 09:08:38.468628 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-0" podStartSLOduration=2.928795337 podStartE2EDuration="6.468607672s" podCreationTimestamp="2025-10-14 09:08:32 +0000 UTC" firstStartedPulling="2025-10-14 09:08:33.787969438 +0000 UTC m=+7649.485329809" lastFinishedPulling="2025-10-14 09:08:37.327781763 +0000 UTC m=+7653.025142144" observedRunningTime="2025-10-14 09:08:38.463917986 +0000 UTC m=+7654.161278357" watchObservedRunningTime="2025-10-14 09:08:38.468607672 +0000 UTC m=+7654.165968043" Oct 14 09:08:40 crc kubenswrapper[4870]: I1014 09:08:40.409770 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 14 09:08:40 crc kubenswrapper[4870]: I1014 09:08:40.480938 4870 generic.go:334] "Generic (PLEG): container finished" podID="620a8d3a-3b05-4ea5-bd1a-6f40433d11a9" containerID="0cb67aa3076c37ce75e040a7fda40630104e3813c56d7660b2b9918616bd68ac" exitCode=0 Oct 14 09:08:40 crc kubenswrapper[4870]: I1014 09:08:40.481009 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"620a8d3a-3b05-4ea5-bd1a-6f40433d11a9","Type":"ContainerDied","Data":"0cb67aa3076c37ce75e040a7fda40630104e3813c56d7660b2b9918616bd68ac"} Oct 14 09:08:40 crc kubenswrapper[4870]: I1014 09:08:40.481053 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 14 09:08:40 crc kubenswrapper[4870]: I1014 09:08:40.481091 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"620a8d3a-3b05-4ea5-bd1a-6f40433d11a9","Type":"ContainerDied","Data":"e5c205781105471151c48f465e495dc36465d062c20dc52e4cb0856ce1a319dc"} Oct 14 09:08:40 crc kubenswrapper[4870]: I1014 09:08:40.481129 4870 scope.go:117] "RemoveContainer" containerID="c55f79de5e91cc683b6fedc9c34be9ae3402ba92f16f1bd24bd57d2b0d1093d6" Oct 14 09:08:40 crc kubenswrapper[4870]: I1014 09:08:40.491016 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/620a8d3a-3b05-4ea5-bd1a-6f40433d11a9-combined-ca-bundle\") pod \"620a8d3a-3b05-4ea5-bd1a-6f40433d11a9\" (UID: \"620a8d3a-3b05-4ea5-bd1a-6f40433d11a9\") " Oct 14 09:08:40 crc kubenswrapper[4870]: I1014 09:08:40.491129 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fzbkb\" (UniqueName: \"kubernetes.io/projected/620a8d3a-3b05-4ea5-bd1a-6f40433d11a9-kube-api-access-fzbkb\") pod \"620a8d3a-3b05-4ea5-bd1a-6f40433d11a9\" (UID: \"620a8d3a-3b05-4ea5-bd1a-6f40433d11a9\") " Oct 14 09:08:40 crc kubenswrapper[4870]: I1014 09:08:40.491182 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/620a8d3a-3b05-4ea5-bd1a-6f40433d11a9-run-httpd\") pod \"620a8d3a-3b05-4ea5-bd1a-6f40433d11a9\" (UID: \"620a8d3a-3b05-4ea5-bd1a-6f40433d11a9\") " Oct 14 09:08:40 crc kubenswrapper[4870]: I1014 09:08:40.491322 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/620a8d3a-3b05-4ea5-bd1a-6f40433d11a9-log-httpd\") pod \"620a8d3a-3b05-4ea5-bd1a-6f40433d11a9\" (UID: \"620a8d3a-3b05-4ea5-bd1a-6f40433d11a9\") " Oct 14 09:08:40 crc kubenswrapper[4870]: I1014 09:08:40.491388 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/620a8d3a-3b05-4ea5-bd1a-6f40433d11a9-sg-core-conf-yaml\") pod \"620a8d3a-3b05-4ea5-bd1a-6f40433d11a9\" (UID: \"620a8d3a-3b05-4ea5-bd1a-6f40433d11a9\") " Oct 14 09:08:40 crc kubenswrapper[4870]: I1014 09:08:40.491464 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/620a8d3a-3b05-4ea5-bd1a-6f40433d11a9-scripts\") pod \"620a8d3a-3b05-4ea5-bd1a-6f40433d11a9\" (UID: \"620a8d3a-3b05-4ea5-bd1a-6f40433d11a9\") " Oct 14 09:08:40 crc kubenswrapper[4870]: I1014 09:08:40.491524 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/620a8d3a-3b05-4ea5-bd1a-6f40433d11a9-config-data\") pod \"620a8d3a-3b05-4ea5-bd1a-6f40433d11a9\" (UID: \"620a8d3a-3b05-4ea5-bd1a-6f40433d11a9\") " Oct 14 09:08:40 crc kubenswrapper[4870]: I1014 09:08:40.491970 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/620a8d3a-3b05-4ea5-bd1a-6f40433d11a9-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "620a8d3a-3b05-4ea5-bd1a-6f40433d11a9" (UID: "620a8d3a-3b05-4ea5-bd1a-6f40433d11a9"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 09:08:40 crc kubenswrapper[4870]: I1014 09:08:40.491987 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/620a8d3a-3b05-4ea5-bd1a-6f40433d11a9-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "620a8d3a-3b05-4ea5-bd1a-6f40433d11a9" (UID: "620a8d3a-3b05-4ea5-bd1a-6f40433d11a9"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 09:08:40 crc kubenswrapper[4870]: I1014 09:08:40.492533 4870 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/620a8d3a-3b05-4ea5-bd1a-6f40433d11a9-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 14 09:08:40 crc kubenswrapper[4870]: I1014 09:08:40.492578 4870 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/620a8d3a-3b05-4ea5-bd1a-6f40433d11a9-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 14 09:08:40 crc kubenswrapper[4870]: I1014 09:08:40.497457 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/620a8d3a-3b05-4ea5-bd1a-6f40433d11a9-scripts" (OuterVolumeSpecName: "scripts") pod "620a8d3a-3b05-4ea5-bd1a-6f40433d11a9" (UID: "620a8d3a-3b05-4ea5-bd1a-6f40433d11a9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:08:40 crc kubenswrapper[4870]: I1014 09:08:40.497501 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/620a8d3a-3b05-4ea5-bd1a-6f40433d11a9-kube-api-access-fzbkb" (OuterVolumeSpecName: "kube-api-access-fzbkb") pod "620a8d3a-3b05-4ea5-bd1a-6f40433d11a9" (UID: "620a8d3a-3b05-4ea5-bd1a-6f40433d11a9"). InnerVolumeSpecName "kube-api-access-fzbkb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:08:40 crc kubenswrapper[4870]: I1014 09:08:40.512559 4870 scope.go:117] "RemoveContainer" containerID="e91ba300647ffc4556c5aa3dd7a1c6268471cb72acd6d40f651b16b77b541548" Oct 14 09:08:40 crc kubenswrapper[4870]: I1014 09:08:40.526603 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/620a8d3a-3b05-4ea5-bd1a-6f40433d11a9-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "620a8d3a-3b05-4ea5-bd1a-6f40433d11a9" (UID: "620a8d3a-3b05-4ea5-bd1a-6f40433d11a9"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:08:40 crc kubenswrapper[4870]: I1014 09:08:40.585636 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/620a8d3a-3b05-4ea5-bd1a-6f40433d11a9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "620a8d3a-3b05-4ea5-bd1a-6f40433d11a9" (UID: "620a8d3a-3b05-4ea5-bd1a-6f40433d11a9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:08:40 crc kubenswrapper[4870]: I1014 09:08:40.594812 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fzbkb\" (UniqueName: \"kubernetes.io/projected/620a8d3a-3b05-4ea5-bd1a-6f40433d11a9-kube-api-access-fzbkb\") on node \"crc\" DevicePath \"\"" Oct 14 09:08:40 crc kubenswrapper[4870]: I1014 09:08:40.594841 4870 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/620a8d3a-3b05-4ea5-bd1a-6f40433d11a9-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 14 09:08:40 crc kubenswrapper[4870]: I1014 09:08:40.594851 4870 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/620a8d3a-3b05-4ea5-bd1a-6f40433d11a9-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 09:08:40 crc kubenswrapper[4870]: I1014 09:08:40.594861 4870 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/620a8d3a-3b05-4ea5-bd1a-6f40433d11a9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 09:08:40 crc kubenswrapper[4870]: I1014 09:08:40.596870 4870 scope.go:117] "RemoveContainer" containerID="0cb67aa3076c37ce75e040a7fda40630104e3813c56d7660b2b9918616bd68ac" Oct 14 09:08:40 crc kubenswrapper[4870]: I1014 09:08:40.614204 4870 scope.go:117] "RemoveContainer" containerID="533ef6e22525e1381bb31273c227ba87cf77eceb2c8712d18f5e5f95a4c2892a" Oct 14 09:08:40 crc kubenswrapper[4870]: I1014 09:08:40.616263 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/620a8d3a-3b05-4ea5-bd1a-6f40433d11a9-config-data" (OuterVolumeSpecName: "config-data") pod "620a8d3a-3b05-4ea5-bd1a-6f40433d11a9" (UID: "620a8d3a-3b05-4ea5-bd1a-6f40433d11a9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:08:40 crc kubenswrapper[4870]: I1014 09:08:40.640339 4870 scope.go:117] "RemoveContainer" containerID="c55f79de5e91cc683b6fedc9c34be9ae3402ba92f16f1bd24bd57d2b0d1093d6" Oct 14 09:08:40 crc kubenswrapper[4870]: E1014 09:08:40.640852 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c55f79de5e91cc683b6fedc9c34be9ae3402ba92f16f1bd24bd57d2b0d1093d6\": container with ID starting with c55f79de5e91cc683b6fedc9c34be9ae3402ba92f16f1bd24bd57d2b0d1093d6 not found: ID does not exist" containerID="c55f79de5e91cc683b6fedc9c34be9ae3402ba92f16f1bd24bd57d2b0d1093d6" Oct 14 09:08:40 crc kubenswrapper[4870]: I1014 09:08:40.640904 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c55f79de5e91cc683b6fedc9c34be9ae3402ba92f16f1bd24bd57d2b0d1093d6"} err="failed to get container status \"c55f79de5e91cc683b6fedc9c34be9ae3402ba92f16f1bd24bd57d2b0d1093d6\": rpc error: code = NotFound desc = could not find container \"c55f79de5e91cc683b6fedc9c34be9ae3402ba92f16f1bd24bd57d2b0d1093d6\": container with ID starting with c55f79de5e91cc683b6fedc9c34be9ae3402ba92f16f1bd24bd57d2b0d1093d6 not found: ID does not exist" Oct 14 09:08:40 crc kubenswrapper[4870]: I1014 09:08:40.640932 4870 scope.go:117] "RemoveContainer" containerID="e91ba300647ffc4556c5aa3dd7a1c6268471cb72acd6d40f651b16b77b541548" Oct 14 09:08:40 crc kubenswrapper[4870]: E1014 09:08:40.641356 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e91ba300647ffc4556c5aa3dd7a1c6268471cb72acd6d40f651b16b77b541548\": container with ID starting with e91ba300647ffc4556c5aa3dd7a1c6268471cb72acd6d40f651b16b77b541548 not found: ID does not exist" containerID="e91ba300647ffc4556c5aa3dd7a1c6268471cb72acd6d40f651b16b77b541548" Oct 14 09:08:40 crc kubenswrapper[4870]: I1014 09:08:40.641404 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e91ba300647ffc4556c5aa3dd7a1c6268471cb72acd6d40f651b16b77b541548"} err="failed to get container status \"e91ba300647ffc4556c5aa3dd7a1c6268471cb72acd6d40f651b16b77b541548\": rpc error: code = NotFound desc = could not find container \"e91ba300647ffc4556c5aa3dd7a1c6268471cb72acd6d40f651b16b77b541548\": container with ID starting with e91ba300647ffc4556c5aa3dd7a1c6268471cb72acd6d40f651b16b77b541548 not found: ID does not exist" Oct 14 09:08:40 crc kubenswrapper[4870]: I1014 09:08:40.641463 4870 scope.go:117] "RemoveContainer" containerID="0cb67aa3076c37ce75e040a7fda40630104e3813c56d7660b2b9918616bd68ac" Oct 14 09:08:40 crc kubenswrapper[4870]: E1014 09:08:40.641913 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0cb67aa3076c37ce75e040a7fda40630104e3813c56d7660b2b9918616bd68ac\": container with ID starting with 0cb67aa3076c37ce75e040a7fda40630104e3813c56d7660b2b9918616bd68ac not found: ID does not exist" containerID="0cb67aa3076c37ce75e040a7fda40630104e3813c56d7660b2b9918616bd68ac" Oct 14 09:08:40 crc kubenswrapper[4870]: I1014 09:08:40.641963 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0cb67aa3076c37ce75e040a7fda40630104e3813c56d7660b2b9918616bd68ac"} err="failed to get container status \"0cb67aa3076c37ce75e040a7fda40630104e3813c56d7660b2b9918616bd68ac\": rpc error: code = NotFound desc = could not find container \"0cb67aa3076c37ce75e040a7fda40630104e3813c56d7660b2b9918616bd68ac\": container with ID starting with 0cb67aa3076c37ce75e040a7fda40630104e3813c56d7660b2b9918616bd68ac not found: ID does not exist" Oct 14 09:08:40 crc kubenswrapper[4870]: I1014 09:08:40.641979 4870 scope.go:117] "RemoveContainer" containerID="533ef6e22525e1381bb31273c227ba87cf77eceb2c8712d18f5e5f95a4c2892a" Oct 14 09:08:40 crc kubenswrapper[4870]: E1014 09:08:40.642317 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"533ef6e22525e1381bb31273c227ba87cf77eceb2c8712d18f5e5f95a4c2892a\": container with ID starting with 533ef6e22525e1381bb31273c227ba87cf77eceb2c8712d18f5e5f95a4c2892a not found: ID does not exist" containerID="533ef6e22525e1381bb31273c227ba87cf77eceb2c8712d18f5e5f95a4c2892a" Oct 14 09:08:40 crc kubenswrapper[4870]: I1014 09:08:40.642355 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"533ef6e22525e1381bb31273c227ba87cf77eceb2c8712d18f5e5f95a4c2892a"} err="failed to get container status \"533ef6e22525e1381bb31273c227ba87cf77eceb2c8712d18f5e5f95a4c2892a\": rpc error: code = NotFound desc = could not find container \"533ef6e22525e1381bb31273c227ba87cf77eceb2c8712d18f5e5f95a4c2892a\": container with ID starting with 533ef6e22525e1381bb31273c227ba87cf77eceb2c8712d18f5e5f95a4c2892a not found: ID does not exist" Oct 14 09:08:40 crc kubenswrapper[4870]: I1014 09:08:40.697316 4870 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/620a8d3a-3b05-4ea5-bd1a-6f40433d11a9-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 09:08:40 crc kubenswrapper[4870]: I1014 09:08:40.829342 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 14 09:08:40 crc kubenswrapper[4870]: I1014 09:08:40.836738 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 14 09:08:40 crc kubenswrapper[4870]: I1014 09:08:40.856012 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 14 09:08:40 crc kubenswrapper[4870]: E1014 09:08:40.856449 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="620a8d3a-3b05-4ea5-bd1a-6f40433d11a9" containerName="sg-core" Oct 14 09:08:40 crc kubenswrapper[4870]: I1014 09:08:40.856465 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="620a8d3a-3b05-4ea5-bd1a-6f40433d11a9" containerName="sg-core" Oct 14 09:08:40 crc kubenswrapper[4870]: E1014 09:08:40.856477 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="620a8d3a-3b05-4ea5-bd1a-6f40433d11a9" containerName="ceilometer-notification-agent" Oct 14 09:08:40 crc kubenswrapper[4870]: I1014 09:08:40.856483 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="620a8d3a-3b05-4ea5-bd1a-6f40433d11a9" containerName="ceilometer-notification-agent" Oct 14 09:08:40 crc kubenswrapper[4870]: E1014 09:08:40.856500 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="620a8d3a-3b05-4ea5-bd1a-6f40433d11a9" containerName="ceilometer-central-agent" Oct 14 09:08:40 crc kubenswrapper[4870]: I1014 09:08:40.856507 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="620a8d3a-3b05-4ea5-bd1a-6f40433d11a9" containerName="ceilometer-central-agent" Oct 14 09:08:40 crc kubenswrapper[4870]: E1014 09:08:40.856513 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="620a8d3a-3b05-4ea5-bd1a-6f40433d11a9" containerName="proxy-httpd" Oct 14 09:08:40 crc kubenswrapper[4870]: I1014 09:08:40.856518 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="620a8d3a-3b05-4ea5-bd1a-6f40433d11a9" containerName="proxy-httpd" Oct 14 09:08:40 crc kubenswrapper[4870]: I1014 09:08:40.856720 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="620a8d3a-3b05-4ea5-bd1a-6f40433d11a9" containerName="sg-core" Oct 14 09:08:40 crc kubenswrapper[4870]: I1014 09:08:40.856736 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="620a8d3a-3b05-4ea5-bd1a-6f40433d11a9" containerName="ceilometer-central-agent" Oct 14 09:08:40 crc kubenswrapper[4870]: I1014 09:08:40.856749 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="620a8d3a-3b05-4ea5-bd1a-6f40433d11a9" containerName="proxy-httpd" Oct 14 09:08:40 crc kubenswrapper[4870]: I1014 09:08:40.856757 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="620a8d3a-3b05-4ea5-bd1a-6f40433d11a9" containerName="ceilometer-notification-agent" Oct 14 09:08:40 crc kubenswrapper[4870]: I1014 09:08:40.858550 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 14 09:08:40 crc kubenswrapper[4870]: I1014 09:08:40.861047 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 14 09:08:40 crc kubenswrapper[4870]: I1014 09:08:40.861922 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 14 09:08:40 crc kubenswrapper[4870]: I1014 09:08:40.878145 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 14 09:08:40 crc kubenswrapper[4870]: I1014 09:08:40.905064 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0c1abbc6-c3c5-4fae-8200-fa3633518f50-log-httpd\") pod \"ceilometer-0\" (UID: \"0c1abbc6-c3c5-4fae-8200-fa3633518f50\") " pod="openstack/ceilometer-0" Oct 14 09:08:40 crc kubenswrapper[4870]: I1014 09:08:40.905432 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c1abbc6-c3c5-4fae-8200-fa3633518f50-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0c1abbc6-c3c5-4fae-8200-fa3633518f50\") " pod="openstack/ceilometer-0" Oct 14 09:08:40 crc kubenswrapper[4870]: I1014 09:08:40.905639 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0c1abbc6-c3c5-4fae-8200-fa3633518f50-run-httpd\") pod \"ceilometer-0\" (UID: \"0c1abbc6-c3c5-4fae-8200-fa3633518f50\") " pod="openstack/ceilometer-0" Oct 14 09:08:40 crc kubenswrapper[4870]: I1014 09:08:40.905807 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c1abbc6-c3c5-4fae-8200-fa3633518f50-config-data\") pod \"ceilometer-0\" (UID: \"0c1abbc6-c3c5-4fae-8200-fa3633518f50\") " pod="openstack/ceilometer-0" Oct 14 09:08:40 crc kubenswrapper[4870]: I1014 09:08:40.905981 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4bm8w\" (UniqueName: \"kubernetes.io/projected/0c1abbc6-c3c5-4fae-8200-fa3633518f50-kube-api-access-4bm8w\") pod \"ceilometer-0\" (UID: \"0c1abbc6-c3c5-4fae-8200-fa3633518f50\") " pod="openstack/ceilometer-0" Oct 14 09:08:40 crc kubenswrapper[4870]: I1014 09:08:40.906177 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0c1abbc6-c3c5-4fae-8200-fa3633518f50-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0c1abbc6-c3c5-4fae-8200-fa3633518f50\") " pod="openstack/ceilometer-0" Oct 14 09:08:40 crc kubenswrapper[4870]: I1014 09:08:40.906398 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0c1abbc6-c3c5-4fae-8200-fa3633518f50-scripts\") pod \"ceilometer-0\" (UID: \"0c1abbc6-c3c5-4fae-8200-fa3633518f50\") " pod="openstack/ceilometer-0" Oct 14 09:08:41 crc kubenswrapper[4870]: I1014 09:08:41.009654 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0c1abbc6-c3c5-4fae-8200-fa3633518f50-log-httpd\") pod \"ceilometer-0\" (UID: \"0c1abbc6-c3c5-4fae-8200-fa3633518f50\") " pod="openstack/ceilometer-0" Oct 14 09:08:41 crc kubenswrapper[4870]: I1014 09:08:41.009985 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c1abbc6-c3c5-4fae-8200-fa3633518f50-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0c1abbc6-c3c5-4fae-8200-fa3633518f50\") " pod="openstack/ceilometer-0" Oct 14 09:08:41 crc kubenswrapper[4870]: I1014 09:08:41.010219 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0c1abbc6-c3c5-4fae-8200-fa3633518f50-run-httpd\") pod \"ceilometer-0\" (UID: \"0c1abbc6-c3c5-4fae-8200-fa3633518f50\") " pod="openstack/ceilometer-0" Oct 14 09:08:41 crc kubenswrapper[4870]: I1014 09:08:41.010458 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0c1abbc6-c3c5-4fae-8200-fa3633518f50-run-httpd\") pod \"ceilometer-0\" (UID: \"0c1abbc6-c3c5-4fae-8200-fa3633518f50\") " pod="openstack/ceilometer-0" Oct 14 09:08:41 crc kubenswrapper[4870]: I1014 09:08:41.010107 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0c1abbc6-c3c5-4fae-8200-fa3633518f50-log-httpd\") pod \"ceilometer-0\" (UID: \"0c1abbc6-c3c5-4fae-8200-fa3633518f50\") " pod="openstack/ceilometer-0" Oct 14 09:08:41 crc kubenswrapper[4870]: I1014 09:08:41.010466 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c1abbc6-c3c5-4fae-8200-fa3633518f50-config-data\") pod \"ceilometer-0\" (UID: \"0c1abbc6-c3c5-4fae-8200-fa3633518f50\") " pod="openstack/ceilometer-0" Oct 14 09:08:41 crc kubenswrapper[4870]: I1014 09:08:41.010781 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4bm8w\" (UniqueName: \"kubernetes.io/projected/0c1abbc6-c3c5-4fae-8200-fa3633518f50-kube-api-access-4bm8w\") pod \"ceilometer-0\" (UID: \"0c1abbc6-c3c5-4fae-8200-fa3633518f50\") " pod="openstack/ceilometer-0" Oct 14 09:08:41 crc kubenswrapper[4870]: I1014 09:08:41.010848 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0c1abbc6-c3c5-4fae-8200-fa3633518f50-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0c1abbc6-c3c5-4fae-8200-fa3633518f50\") " pod="openstack/ceilometer-0" Oct 14 09:08:41 crc kubenswrapper[4870]: I1014 09:08:41.010993 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0c1abbc6-c3c5-4fae-8200-fa3633518f50-scripts\") pod \"ceilometer-0\" (UID: \"0c1abbc6-c3c5-4fae-8200-fa3633518f50\") " pod="openstack/ceilometer-0" Oct 14 09:08:41 crc kubenswrapper[4870]: I1014 09:08:41.014470 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c1abbc6-c3c5-4fae-8200-fa3633518f50-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0c1abbc6-c3c5-4fae-8200-fa3633518f50\") " pod="openstack/ceilometer-0" Oct 14 09:08:41 crc kubenswrapper[4870]: I1014 09:08:41.015096 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0c1abbc6-c3c5-4fae-8200-fa3633518f50-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0c1abbc6-c3c5-4fae-8200-fa3633518f50\") " pod="openstack/ceilometer-0" Oct 14 09:08:41 crc kubenswrapper[4870]: I1014 09:08:41.015969 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c1abbc6-c3c5-4fae-8200-fa3633518f50-config-data\") pod \"ceilometer-0\" (UID: \"0c1abbc6-c3c5-4fae-8200-fa3633518f50\") " pod="openstack/ceilometer-0" Oct 14 09:08:41 crc kubenswrapper[4870]: I1014 09:08:41.017887 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0c1abbc6-c3c5-4fae-8200-fa3633518f50-scripts\") pod \"ceilometer-0\" (UID: \"0c1abbc6-c3c5-4fae-8200-fa3633518f50\") " pod="openstack/ceilometer-0" Oct 14 09:08:41 crc kubenswrapper[4870]: I1014 09:08:41.027002 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4bm8w\" (UniqueName: \"kubernetes.io/projected/0c1abbc6-c3c5-4fae-8200-fa3633518f50-kube-api-access-4bm8w\") pod \"ceilometer-0\" (UID: \"0c1abbc6-c3c5-4fae-8200-fa3633518f50\") " pod="openstack/ceilometer-0" Oct 14 09:08:41 crc kubenswrapper[4870]: I1014 09:08:41.033931 4870 scope.go:117] "RemoveContainer" containerID="8fcaeb3090f0c89cba1e66db524b4a246072daa35592012664aca450fa09ddc9" Oct 14 09:08:41 crc kubenswrapper[4870]: E1014 09:08:41.034329 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 09:08:41 crc kubenswrapper[4870]: I1014 09:08:41.051227 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="620a8d3a-3b05-4ea5-bd1a-6f40433d11a9" path="/var/lib/kubelet/pods/620a8d3a-3b05-4ea5-bd1a-6f40433d11a9/volumes" Oct 14 09:08:41 crc kubenswrapper[4870]: I1014 09:08:41.184781 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 14 09:08:41 crc kubenswrapper[4870]: I1014 09:08:41.744924 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 14 09:08:42 crc kubenswrapper[4870]: I1014 09:08:42.512929 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0c1abbc6-c3c5-4fae-8200-fa3633518f50","Type":"ContainerStarted","Data":"6c66907b3ba7cfdfa4c257675d38768ee4dfd691a99ca1aebf71de86c862b364"} Oct 14 09:08:42 crc kubenswrapper[4870]: I1014 09:08:42.513423 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0c1abbc6-c3c5-4fae-8200-fa3633518f50","Type":"ContainerStarted","Data":"e17032cae0cba82e58a163171c945ba5d0f8bae398d4814f6d06d61311984cb1"} Oct 14 09:08:43 crc kubenswrapper[4870]: I1014 09:08:43.525761 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0c1abbc6-c3c5-4fae-8200-fa3633518f50","Type":"ContainerStarted","Data":"81dc195c974a1f500fd17010993c6de1b08465d9468744524f534488e538b64f"} Oct 14 09:08:44 crc kubenswrapper[4870]: I1014 09:08:44.045712 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-db-create-9495v"] Oct 14 09:08:44 crc kubenswrapper[4870]: I1014 09:08:44.047837 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-9495v" Oct 14 09:08:44 crc kubenswrapper[4870]: I1014 09:08:44.060175 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-create-9495v"] Oct 14 09:08:44 crc kubenswrapper[4870]: I1014 09:08:44.093796 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tmjsh\" (UniqueName: \"kubernetes.io/projected/f51823d2-e754-4dce-8ad1-6d3a6752e5ed-kube-api-access-tmjsh\") pod \"manila-db-create-9495v\" (UID: \"f51823d2-e754-4dce-8ad1-6d3a6752e5ed\") " pod="openstack/manila-db-create-9495v" Oct 14 09:08:44 crc kubenswrapper[4870]: I1014 09:08:44.195149 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tmjsh\" (UniqueName: \"kubernetes.io/projected/f51823d2-e754-4dce-8ad1-6d3a6752e5ed-kube-api-access-tmjsh\") pod \"manila-db-create-9495v\" (UID: \"f51823d2-e754-4dce-8ad1-6d3a6752e5ed\") " pod="openstack/manila-db-create-9495v" Oct 14 09:08:44 crc kubenswrapper[4870]: I1014 09:08:44.217583 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tmjsh\" (UniqueName: \"kubernetes.io/projected/f51823d2-e754-4dce-8ad1-6d3a6752e5ed-kube-api-access-tmjsh\") pod \"manila-db-create-9495v\" (UID: \"f51823d2-e754-4dce-8ad1-6d3a6752e5ed\") " pod="openstack/manila-db-create-9495v" Oct 14 09:08:44 crc kubenswrapper[4870]: I1014 09:08:44.426503 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-9495v" Oct 14 09:08:44 crc kubenswrapper[4870]: I1014 09:08:44.554585 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0c1abbc6-c3c5-4fae-8200-fa3633518f50","Type":"ContainerStarted","Data":"11569cd009cbf79d170157a97f6048b3dc7a1902a13f0da7b0909f183b5fee3c"} Oct 14 09:08:44 crc kubenswrapper[4870]: I1014 09:08:44.927386 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-create-9495v"] Oct 14 09:08:45 crc kubenswrapper[4870]: I1014 09:08:45.576087 4870 generic.go:334] "Generic (PLEG): container finished" podID="f51823d2-e754-4dce-8ad1-6d3a6752e5ed" containerID="d794ce3cb85d49e9fc592ac73a2b3a39f4b6085a8da52d52c364d6d6048ba80f" exitCode=0 Oct 14 09:08:45 crc kubenswrapper[4870]: I1014 09:08:45.576258 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-create-9495v" event={"ID":"f51823d2-e754-4dce-8ad1-6d3a6752e5ed","Type":"ContainerDied","Data":"d794ce3cb85d49e9fc592ac73a2b3a39f4b6085a8da52d52c364d6d6048ba80f"} Oct 14 09:08:45 crc kubenswrapper[4870]: I1014 09:08:45.576336 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-create-9495v" event={"ID":"f51823d2-e754-4dce-8ad1-6d3a6752e5ed","Type":"ContainerStarted","Data":"0676d12b59962dacfa4ac6f98cd0d4c09021eb481e498ee938598bdd64b2d549"} Oct 14 09:08:46 crc kubenswrapper[4870]: I1014 09:08:46.593986 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0c1abbc6-c3c5-4fae-8200-fa3633518f50","Type":"ContainerStarted","Data":"fc6a3a9c4e15f823dafcbbfd5aa9442770632652275e8ec060b454c537254c5c"} Oct 14 09:08:46 crc kubenswrapper[4870]: I1014 09:08:46.594691 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 14 09:08:46 crc kubenswrapper[4870]: I1014 09:08:46.622128 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.634264359 podStartE2EDuration="6.622089877s" podCreationTimestamp="2025-10-14 09:08:40 +0000 UTC" firstStartedPulling="2025-10-14 09:08:41.758490645 +0000 UTC m=+7657.455851016" lastFinishedPulling="2025-10-14 09:08:45.746316163 +0000 UTC m=+7661.443676534" observedRunningTime="2025-10-14 09:08:46.614288055 +0000 UTC m=+7662.311648466" watchObservedRunningTime="2025-10-14 09:08:46.622089877 +0000 UTC m=+7662.319450258" Oct 14 09:08:47 crc kubenswrapper[4870]: I1014 09:08:47.073130 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-9495v" Oct 14 09:08:47 crc kubenswrapper[4870]: I1014 09:08:47.166578 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tmjsh\" (UniqueName: \"kubernetes.io/projected/f51823d2-e754-4dce-8ad1-6d3a6752e5ed-kube-api-access-tmjsh\") pod \"f51823d2-e754-4dce-8ad1-6d3a6752e5ed\" (UID: \"f51823d2-e754-4dce-8ad1-6d3a6752e5ed\") " Oct 14 09:08:47 crc kubenswrapper[4870]: I1014 09:08:47.175854 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f51823d2-e754-4dce-8ad1-6d3a6752e5ed-kube-api-access-tmjsh" (OuterVolumeSpecName: "kube-api-access-tmjsh") pod "f51823d2-e754-4dce-8ad1-6d3a6752e5ed" (UID: "f51823d2-e754-4dce-8ad1-6d3a6752e5ed"). InnerVolumeSpecName "kube-api-access-tmjsh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:08:47 crc kubenswrapper[4870]: I1014 09:08:47.271172 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tmjsh\" (UniqueName: \"kubernetes.io/projected/f51823d2-e754-4dce-8ad1-6d3a6752e5ed-kube-api-access-tmjsh\") on node \"crc\" DevicePath \"\"" Oct 14 09:08:47 crc kubenswrapper[4870]: I1014 09:08:47.610630 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-create-9495v" event={"ID":"f51823d2-e754-4dce-8ad1-6d3a6752e5ed","Type":"ContainerDied","Data":"0676d12b59962dacfa4ac6f98cd0d4c09021eb481e498ee938598bdd64b2d549"} Oct 14 09:08:47 crc kubenswrapper[4870]: I1014 09:08:47.610666 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-9495v" Oct 14 09:08:47 crc kubenswrapper[4870]: I1014 09:08:47.610699 4870 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0676d12b59962dacfa4ac6f98cd0d4c09021eb481e498ee938598bdd64b2d549" Oct 14 09:08:54 crc kubenswrapper[4870]: I1014 09:08:54.034248 4870 scope.go:117] "RemoveContainer" containerID="8fcaeb3090f0c89cba1e66db524b4a246072daa35592012664aca450fa09ddc9" Oct 14 09:08:54 crc kubenswrapper[4870]: E1014 09:08:54.034920 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 09:08:54 crc kubenswrapper[4870]: I1014 09:08:54.156849 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-cbde-account-create-vcwvz"] Oct 14 09:08:54 crc kubenswrapper[4870]: E1014 09:08:54.157395 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f51823d2-e754-4dce-8ad1-6d3a6752e5ed" containerName="mariadb-database-create" Oct 14 09:08:54 crc kubenswrapper[4870]: I1014 09:08:54.157417 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="f51823d2-e754-4dce-8ad1-6d3a6752e5ed" containerName="mariadb-database-create" Oct 14 09:08:54 crc kubenswrapper[4870]: I1014 09:08:54.157884 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="f51823d2-e754-4dce-8ad1-6d3a6752e5ed" containerName="mariadb-database-create" Oct 14 09:08:54 crc kubenswrapper[4870]: I1014 09:08:54.158730 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-cbde-account-create-vcwvz" Oct 14 09:08:54 crc kubenswrapper[4870]: I1014 09:08:54.162835 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-db-secret" Oct 14 09:08:54 crc kubenswrapper[4870]: I1014 09:08:54.176684 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-cbde-account-create-vcwvz"] Oct 14 09:08:54 crc kubenswrapper[4870]: I1014 09:08:54.250160 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qmksm\" (UniqueName: \"kubernetes.io/projected/43692e8e-fa17-4f76-9a5f-24a28270152f-kube-api-access-qmksm\") pod \"manila-cbde-account-create-vcwvz\" (UID: \"43692e8e-fa17-4f76-9a5f-24a28270152f\") " pod="openstack/manila-cbde-account-create-vcwvz" Oct 14 09:08:54 crc kubenswrapper[4870]: I1014 09:08:54.352799 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qmksm\" (UniqueName: \"kubernetes.io/projected/43692e8e-fa17-4f76-9a5f-24a28270152f-kube-api-access-qmksm\") pod \"manila-cbde-account-create-vcwvz\" (UID: \"43692e8e-fa17-4f76-9a5f-24a28270152f\") " pod="openstack/manila-cbde-account-create-vcwvz" Oct 14 09:08:54 crc kubenswrapper[4870]: I1014 09:08:54.386881 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qmksm\" (UniqueName: \"kubernetes.io/projected/43692e8e-fa17-4f76-9a5f-24a28270152f-kube-api-access-qmksm\") pod \"manila-cbde-account-create-vcwvz\" (UID: \"43692e8e-fa17-4f76-9a5f-24a28270152f\") " pod="openstack/manila-cbde-account-create-vcwvz" Oct 14 09:08:54 crc kubenswrapper[4870]: I1014 09:08:54.491923 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-cbde-account-create-vcwvz" Oct 14 09:08:55 crc kubenswrapper[4870]: I1014 09:08:55.090055 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-cbde-account-create-vcwvz"] Oct 14 09:08:55 crc kubenswrapper[4870]: I1014 09:08:55.709916 4870 generic.go:334] "Generic (PLEG): container finished" podID="43692e8e-fa17-4f76-9a5f-24a28270152f" containerID="c5729775861fa7384aca80299f35064424dffaeb877e688615688092998be592" exitCode=0 Oct 14 09:08:55 crc kubenswrapper[4870]: I1014 09:08:55.710014 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-cbde-account-create-vcwvz" event={"ID":"43692e8e-fa17-4f76-9a5f-24a28270152f","Type":"ContainerDied","Data":"c5729775861fa7384aca80299f35064424dffaeb877e688615688092998be592"} Oct 14 09:08:55 crc kubenswrapper[4870]: I1014 09:08:55.710366 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-cbde-account-create-vcwvz" event={"ID":"43692e8e-fa17-4f76-9a5f-24a28270152f","Type":"ContainerStarted","Data":"5b9929c540dc905a7521522f90585ec4bf32c9cd63c3a3bcf1ffc510e08a1a4c"} Oct 14 09:08:57 crc kubenswrapper[4870]: I1014 09:08:57.226490 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-cbde-account-create-vcwvz" Oct 14 09:08:57 crc kubenswrapper[4870]: I1014 09:08:57.325479 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qmksm\" (UniqueName: \"kubernetes.io/projected/43692e8e-fa17-4f76-9a5f-24a28270152f-kube-api-access-qmksm\") pod \"43692e8e-fa17-4f76-9a5f-24a28270152f\" (UID: \"43692e8e-fa17-4f76-9a5f-24a28270152f\") " Oct 14 09:08:57 crc kubenswrapper[4870]: I1014 09:08:57.331719 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43692e8e-fa17-4f76-9a5f-24a28270152f-kube-api-access-qmksm" (OuterVolumeSpecName: "kube-api-access-qmksm") pod "43692e8e-fa17-4f76-9a5f-24a28270152f" (UID: "43692e8e-fa17-4f76-9a5f-24a28270152f"). InnerVolumeSpecName "kube-api-access-qmksm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:08:57 crc kubenswrapper[4870]: I1014 09:08:57.428592 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qmksm\" (UniqueName: \"kubernetes.io/projected/43692e8e-fa17-4f76-9a5f-24a28270152f-kube-api-access-qmksm\") on node \"crc\" DevicePath \"\"" Oct 14 09:08:57 crc kubenswrapper[4870]: I1014 09:08:57.740343 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-cbde-account-create-vcwvz" event={"ID":"43692e8e-fa17-4f76-9a5f-24a28270152f","Type":"ContainerDied","Data":"5b9929c540dc905a7521522f90585ec4bf32c9cd63c3a3bcf1ffc510e08a1a4c"} Oct 14 09:08:57 crc kubenswrapper[4870]: I1014 09:08:57.740384 4870 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5b9929c540dc905a7521522f90585ec4bf32c9cd63c3a3bcf1ffc510e08a1a4c" Oct 14 09:08:57 crc kubenswrapper[4870]: I1014 09:08:57.740484 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-cbde-account-create-vcwvz" Oct 14 09:08:59 crc kubenswrapper[4870]: I1014 09:08:59.513474 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-db-sync-wq597"] Oct 14 09:08:59 crc kubenswrapper[4870]: E1014 09:08:59.514573 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="43692e8e-fa17-4f76-9a5f-24a28270152f" containerName="mariadb-account-create" Oct 14 09:08:59 crc kubenswrapper[4870]: I1014 09:08:59.514597 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="43692e8e-fa17-4f76-9a5f-24a28270152f" containerName="mariadb-account-create" Oct 14 09:08:59 crc kubenswrapper[4870]: I1014 09:08:59.516742 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="43692e8e-fa17-4f76-9a5f-24a28270152f" containerName="mariadb-account-create" Oct 14 09:08:59 crc kubenswrapper[4870]: I1014 09:08:59.519139 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-wq597" Oct 14 09:08:59 crc kubenswrapper[4870]: I1014 09:08:59.521019 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-config-data" Oct 14 09:08:59 crc kubenswrapper[4870]: I1014 09:08:59.521558 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-manila-dockercfg-97p94" Oct 14 09:08:59 crc kubenswrapper[4870]: I1014 09:08:59.548943 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-sync-wq597"] Oct 14 09:08:59 crc kubenswrapper[4870]: I1014 09:08:59.570194 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/ce5d88b9-7a1f-4aac-aa4c-e9d36c4a5a47-job-config-data\") pod \"manila-db-sync-wq597\" (UID: \"ce5d88b9-7a1f-4aac-aa4c-e9d36c4a5a47\") " pod="openstack/manila-db-sync-wq597" Oct 14 09:08:59 crc kubenswrapper[4870]: I1014 09:08:59.570290 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce5d88b9-7a1f-4aac-aa4c-e9d36c4a5a47-combined-ca-bundle\") pod \"manila-db-sync-wq597\" (UID: \"ce5d88b9-7a1f-4aac-aa4c-e9d36c4a5a47\") " pod="openstack/manila-db-sync-wq597" Oct 14 09:08:59 crc kubenswrapper[4870]: I1014 09:08:59.570340 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ce5d88b9-7a1f-4aac-aa4c-e9d36c4a5a47-config-data\") pod \"manila-db-sync-wq597\" (UID: \"ce5d88b9-7a1f-4aac-aa4c-e9d36c4a5a47\") " pod="openstack/manila-db-sync-wq597" Oct 14 09:08:59 crc kubenswrapper[4870]: I1014 09:08:59.570501 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cc4st\" (UniqueName: \"kubernetes.io/projected/ce5d88b9-7a1f-4aac-aa4c-e9d36c4a5a47-kube-api-access-cc4st\") pod \"manila-db-sync-wq597\" (UID: \"ce5d88b9-7a1f-4aac-aa4c-e9d36c4a5a47\") " pod="openstack/manila-db-sync-wq597" Oct 14 09:08:59 crc kubenswrapper[4870]: I1014 09:08:59.672230 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cc4st\" (UniqueName: \"kubernetes.io/projected/ce5d88b9-7a1f-4aac-aa4c-e9d36c4a5a47-kube-api-access-cc4st\") pod \"manila-db-sync-wq597\" (UID: \"ce5d88b9-7a1f-4aac-aa4c-e9d36c4a5a47\") " pod="openstack/manila-db-sync-wq597" Oct 14 09:08:59 crc kubenswrapper[4870]: I1014 09:08:59.672350 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/ce5d88b9-7a1f-4aac-aa4c-e9d36c4a5a47-job-config-data\") pod \"manila-db-sync-wq597\" (UID: \"ce5d88b9-7a1f-4aac-aa4c-e9d36c4a5a47\") " pod="openstack/manila-db-sync-wq597" Oct 14 09:08:59 crc kubenswrapper[4870]: I1014 09:08:59.672416 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce5d88b9-7a1f-4aac-aa4c-e9d36c4a5a47-combined-ca-bundle\") pod \"manila-db-sync-wq597\" (UID: \"ce5d88b9-7a1f-4aac-aa4c-e9d36c4a5a47\") " pod="openstack/manila-db-sync-wq597" Oct 14 09:08:59 crc kubenswrapper[4870]: I1014 09:08:59.672482 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ce5d88b9-7a1f-4aac-aa4c-e9d36c4a5a47-config-data\") pod \"manila-db-sync-wq597\" (UID: \"ce5d88b9-7a1f-4aac-aa4c-e9d36c4a5a47\") " pod="openstack/manila-db-sync-wq597" Oct 14 09:08:59 crc kubenswrapper[4870]: I1014 09:08:59.678376 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/ce5d88b9-7a1f-4aac-aa4c-e9d36c4a5a47-job-config-data\") pod \"manila-db-sync-wq597\" (UID: \"ce5d88b9-7a1f-4aac-aa4c-e9d36c4a5a47\") " pod="openstack/manila-db-sync-wq597" Oct 14 09:08:59 crc kubenswrapper[4870]: I1014 09:08:59.678759 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ce5d88b9-7a1f-4aac-aa4c-e9d36c4a5a47-config-data\") pod \"manila-db-sync-wq597\" (UID: \"ce5d88b9-7a1f-4aac-aa4c-e9d36c4a5a47\") " pod="openstack/manila-db-sync-wq597" Oct 14 09:08:59 crc kubenswrapper[4870]: I1014 09:08:59.681635 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce5d88b9-7a1f-4aac-aa4c-e9d36c4a5a47-combined-ca-bundle\") pod \"manila-db-sync-wq597\" (UID: \"ce5d88b9-7a1f-4aac-aa4c-e9d36c4a5a47\") " pod="openstack/manila-db-sync-wq597" Oct 14 09:08:59 crc kubenswrapper[4870]: I1014 09:08:59.698112 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cc4st\" (UniqueName: \"kubernetes.io/projected/ce5d88b9-7a1f-4aac-aa4c-e9d36c4a5a47-kube-api-access-cc4st\") pod \"manila-db-sync-wq597\" (UID: \"ce5d88b9-7a1f-4aac-aa4c-e9d36c4a5a47\") " pod="openstack/manila-db-sync-wq597" Oct 14 09:08:59 crc kubenswrapper[4870]: I1014 09:08:59.837260 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-wq597" Oct 14 09:09:00 crc kubenswrapper[4870]: I1014 09:09:00.708867 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-sync-wq597"] Oct 14 09:09:00 crc kubenswrapper[4870]: W1014 09:09:00.714566 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podce5d88b9_7a1f_4aac_aa4c_e9d36c4a5a47.slice/crio-c383c3be5a863bb491e5e29edaf03a1e1c1f8bcfa7dac675731e4268a360a101 WatchSource:0}: Error finding container c383c3be5a863bb491e5e29edaf03a1e1c1f8bcfa7dac675731e4268a360a101: Status 404 returned error can't find the container with id c383c3be5a863bb491e5e29edaf03a1e1c1f8bcfa7dac675731e4268a360a101 Oct 14 09:09:00 crc kubenswrapper[4870]: I1014 09:09:00.778683 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-wq597" event={"ID":"ce5d88b9-7a1f-4aac-aa4c-e9d36c4a5a47","Type":"ContainerStarted","Data":"c383c3be5a863bb491e5e29edaf03a1e1c1f8bcfa7dac675731e4268a360a101"} Oct 14 09:09:04 crc kubenswrapper[4870]: I1014 09:09:04.052869 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-scw7l"] Oct 14 09:09:04 crc kubenswrapper[4870]: I1014 09:09:04.066061 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-9htzx"] Oct 14 09:09:04 crc kubenswrapper[4870]: I1014 09:09:04.078917 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-scw7l"] Oct 14 09:09:04 crc kubenswrapper[4870]: I1014 09:09:04.087859 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-9htzx"] Oct 14 09:09:04 crc kubenswrapper[4870]: I1014 09:09:04.095348 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-p9h2b"] Oct 14 09:09:04 crc kubenswrapper[4870]: I1014 09:09:04.102985 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-p9h2b"] Oct 14 09:09:05 crc kubenswrapper[4870]: I1014 09:09:05.048906 4870 scope.go:117] "RemoveContainer" containerID="8fcaeb3090f0c89cba1e66db524b4a246072daa35592012664aca450fa09ddc9" Oct 14 09:09:05 crc kubenswrapper[4870]: E1014 09:09:05.049328 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 09:09:05 crc kubenswrapper[4870]: I1014 09:09:05.051451 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="16fe9c99-c775-4b99-8132-535f9daa9a9f" path="/var/lib/kubelet/pods/16fe9c99-c775-4b99-8132-535f9daa9a9f/volumes" Oct 14 09:09:05 crc kubenswrapper[4870]: I1014 09:09:05.052007 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2ea8030c-8337-467b-be99-b8eb2e48ea56" path="/var/lib/kubelet/pods/2ea8030c-8337-467b-be99-b8eb2e48ea56/volumes" Oct 14 09:09:05 crc kubenswrapper[4870]: I1014 09:09:05.052521 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8fba7d20-e48c-4e8e-b0a6-21ed0aea5ab1" path="/var/lib/kubelet/pods/8fba7d20-e48c-4e8e-b0a6-21ed0aea5ab1/volumes" Oct 14 09:09:06 crc kubenswrapper[4870]: I1014 09:09:06.849175 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-wq597" event={"ID":"ce5d88b9-7a1f-4aac-aa4c-e9d36c4a5a47","Type":"ContainerStarted","Data":"f21671f988885335521bbaced2823640aa1fdc272bab39ec2d2f3a1cb560addd"} Oct 14 09:09:06 crc kubenswrapper[4870]: I1014 09:09:06.879751 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-db-sync-wq597" podStartSLOduration=3.166478528 podStartE2EDuration="7.879728758s" podCreationTimestamp="2025-10-14 09:08:59 +0000 UTC" firstStartedPulling="2025-10-14 09:09:00.716504125 +0000 UTC m=+7676.413864496" lastFinishedPulling="2025-10-14 09:09:05.429754355 +0000 UTC m=+7681.127114726" observedRunningTime="2025-10-14 09:09:06.869979457 +0000 UTC m=+7682.567339868" watchObservedRunningTime="2025-10-14 09:09:06.879728758 +0000 UTC m=+7682.577089129" Oct 14 09:09:07 crc kubenswrapper[4870]: I1014 09:09:07.864686 4870 generic.go:334] "Generic (PLEG): container finished" podID="ce5d88b9-7a1f-4aac-aa4c-e9d36c4a5a47" containerID="f21671f988885335521bbaced2823640aa1fdc272bab39ec2d2f3a1cb560addd" exitCode=0 Oct 14 09:09:07 crc kubenswrapper[4870]: I1014 09:09:07.864784 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-wq597" event={"ID":"ce5d88b9-7a1f-4aac-aa4c-e9d36c4a5a47","Type":"ContainerDied","Data":"f21671f988885335521bbaced2823640aa1fdc272bab39ec2d2f3a1cb560addd"} Oct 14 09:09:09 crc kubenswrapper[4870]: I1014 09:09:09.534948 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-wq597" Oct 14 09:09:09 crc kubenswrapper[4870]: I1014 09:09:09.594146 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/ce5d88b9-7a1f-4aac-aa4c-e9d36c4a5a47-job-config-data\") pod \"ce5d88b9-7a1f-4aac-aa4c-e9d36c4a5a47\" (UID: \"ce5d88b9-7a1f-4aac-aa4c-e9d36c4a5a47\") " Oct 14 09:09:09 crc kubenswrapper[4870]: I1014 09:09:09.594454 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce5d88b9-7a1f-4aac-aa4c-e9d36c4a5a47-combined-ca-bundle\") pod \"ce5d88b9-7a1f-4aac-aa4c-e9d36c4a5a47\" (UID: \"ce5d88b9-7a1f-4aac-aa4c-e9d36c4a5a47\") " Oct 14 09:09:09 crc kubenswrapper[4870]: I1014 09:09:09.594618 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cc4st\" (UniqueName: \"kubernetes.io/projected/ce5d88b9-7a1f-4aac-aa4c-e9d36c4a5a47-kube-api-access-cc4st\") pod \"ce5d88b9-7a1f-4aac-aa4c-e9d36c4a5a47\" (UID: \"ce5d88b9-7a1f-4aac-aa4c-e9d36c4a5a47\") " Oct 14 09:09:09 crc kubenswrapper[4870]: I1014 09:09:09.594736 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ce5d88b9-7a1f-4aac-aa4c-e9d36c4a5a47-config-data\") pod \"ce5d88b9-7a1f-4aac-aa4c-e9d36c4a5a47\" (UID: \"ce5d88b9-7a1f-4aac-aa4c-e9d36c4a5a47\") " Oct 14 09:09:09 crc kubenswrapper[4870]: I1014 09:09:09.602386 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ce5d88b9-7a1f-4aac-aa4c-e9d36c4a5a47-config-data" (OuterVolumeSpecName: "config-data") pod "ce5d88b9-7a1f-4aac-aa4c-e9d36c4a5a47" (UID: "ce5d88b9-7a1f-4aac-aa4c-e9d36c4a5a47"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:09:09 crc kubenswrapper[4870]: I1014 09:09:09.606656 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ce5d88b9-7a1f-4aac-aa4c-e9d36c4a5a47-job-config-data" (OuterVolumeSpecName: "job-config-data") pod "ce5d88b9-7a1f-4aac-aa4c-e9d36c4a5a47" (UID: "ce5d88b9-7a1f-4aac-aa4c-e9d36c4a5a47"). InnerVolumeSpecName "job-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:09:09 crc kubenswrapper[4870]: I1014 09:09:09.608573 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ce5d88b9-7a1f-4aac-aa4c-e9d36c4a5a47-kube-api-access-cc4st" (OuterVolumeSpecName: "kube-api-access-cc4st") pod "ce5d88b9-7a1f-4aac-aa4c-e9d36c4a5a47" (UID: "ce5d88b9-7a1f-4aac-aa4c-e9d36c4a5a47"). InnerVolumeSpecName "kube-api-access-cc4st". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:09:09 crc kubenswrapper[4870]: I1014 09:09:09.637714 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ce5d88b9-7a1f-4aac-aa4c-e9d36c4a5a47-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ce5d88b9-7a1f-4aac-aa4c-e9d36c4a5a47" (UID: "ce5d88b9-7a1f-4aac-aa4c-e9d36c4a5a47"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:09:09 crc kubenswrapper[4870]: I1014 09:09:09.697655 4870 reconciler_common.go:293] "Volume detached for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/ce5d88b9-7a1f-4aac-aa4c-e9d36c4a5a47-job-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 09:09:09 crc kubenswrapper[4870]: I1014 09:09:09.697685 4870 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce5d88b9-7a1f-4aac-aa4c-e9d36c4a5a47-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 09:09:09 crc kubenswrapper[4870]: I1014 09:09:09.698088 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cc4st\" (UniqueName: \"kubernetes.io/projected/ce5d88b9-7a1f-4aac-aa4c-e9d36c4a5a47-kube-api-access-cc4st\") on node \"crc\" DevicePath \"\"" Oct 14 09:09:09 crc kubenswrapper[4870]: I1014 09:09:09.698099 4870 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ce5d88b9-7a1f-4aac-aa4c-e9d36c4a5a47-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 09:09:09 crc kubenswrapper[4870]: I1014 09:09:09.897319 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-wq597" event={"ID":"ce5d88b9-7a1f-4aac-aa4c-e9d36c4a5a47","Type":"ContainerDied","Data":"c383c3be5a863bb491e5e29edaf03a1e1c1f8bcfa7dac675731e4268a360a101"} Oct 14 09:09:09 crc kubenswrapper[4870]: I1014 09:09:09.897363 4870 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c383c3be5a863bb491e5e29edaf03a1e1c1f8bcfa7dac675731e4268a360a101" Oct 14 09:09:09 crc kubenswrapper[4870]: I1014 09:09:09.897761 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-wq597" Oct 14 09:09:10 crc kubenswrapper[4870]: I1014 09:09:10.215665 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-scheduler-0"] Oct 14 09:09:10 crc kubenswrapper[4870]: E1014 09:09:10.216150 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce5d88b9-7a1f-4aac-aa4c-e9d36c4a5a47" containerName="manila-db-sync" Oct 14 09:09:10 crc kubenswrapper[4870]: I1014 09:09:10.216179 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce5d88b9-7a1f-4aac-aa4c-e9d36c4a5a47" containerName="manila-db-sync" Oct 14 09:09:10 crc kubenswrapper[4870]: I1014 09:09:10.216474 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="ce5d88b9-7a1f-4aac-aa4c-e9d36c4a5a47" containerName="manila-db-sync" Oct 14 09:09:10 crc kubenswrapper[4870]: I1014 09:09:10.217863 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Oct 14 09:09:10 crc kubenswrapper[4870]: I1014 09:09:10.221842 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-manila-dockercfg-97p94" Oct 14 09:09:10 crc kubenswrapper[4870]: I1014 09:09:10.222026 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-scripts" Oct 14 09:09:10 crc kubenswrapper[4870]: I1014 09:09:10.223436 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-scheduler-config-data" Oct 14 09:09:10 crc kubenswrapper[4870]: I1014 09:09:10.223470 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-config-data" Oct 14 09:09:10 crc kubenswrapper[4870]: I1014 09:09:10.252981 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-share-share1-0"] Oct 14 09:09:10 crc kubenswrapper[4870]: I1014 09:09:10.255117 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Oct 14 09:09:10 crc kubenswrapper[4870]: I1014 09:09:10.260184 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-share-share1-config-data" Oct 14 09:09:10 crc kubenswrapper[4870]: I1014 09:09:10.269208 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-scheduler-0"] Oct 14 09:09:10 crc kubenswrapper[4870]: I1014 09:09:10.284502 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-share-share1-0"] Oct 14 09:09:10 crc kubenswrapper[4870]: I1014 09:09:10.312340 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/dea5cc89-8770-43ab-a5f3-4986cfac76af-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"dea5cc89-8770-43ab-a5f3-4986cfac76af\") " pod="openstack/manila-share-share1-0" Oct 14 09:09:10 crc kubenswrapper[4870]: I1014 09:09:10.312375 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-55dr8\" (UniqueName: \"kubernetes.io/projected/f1580185-56b7-4ef3-b148-e3934a5df6eb-kube-api-access-55dr8\") pod \"manila-scheduler-0\" (UID: \"f1580185-56b7-4ef3-b148-e3934a5df6eb\") " pod="openstack/manila-scheduler-0" Oct 14 09:09:10 crc kubenswrapper[4870]: I1014 09:09:10.312417 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dea5cc89-8770-43ab-a5f3-4986cfac76af-config-data\") pod \"manila-share-share1-0\" (UID: \"dea5cc89-8770-43ab-a5f3-4986cfac76af\") " pod="openstack/manila-share-share1-0" Oct 14 09:09:10 crc kubenswrapper[4870]: I1014 09:09:10.312452 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f1580185-56b7-4ef3-b148-e3934a5df6eb-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"f1580185-56b7-4ef3-b148-e3934a5df6eb\") " pod="openstack/manila-scheduler-0" Oct 14 09:09:10 crc kubenswrapper[4870]: I1014 09:09:10.312490 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dea5cc89-8770-43ab-a5f3-4986cfac76af-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"dea5cc89-8770-43ab-a5f3-4986cfac76af\") " pod="openstack/manila-share-share1-0" Oct 14 09:09:10 crc kubenswrapper[4870]: I1014 09:09:10.312542 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dea5cc89-8770-43ab-a5f3-4986cfac76af-scripts\") pod \"manila-share-share1-0\" (UID: \"dea5cc89-8770-43ab-a5f3-4986cfac76af\") " pod="openstack/manila-share-share1-0" Oct 14 09:09:10 crc kubenswrapper[4870]: I1014 09:09:10.312562 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f1580185-56b7-4ef3-b148-e3934a5df6eb-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"f1580185-56b7-4ef3-b148-e3934a5df6eb\") " pod="openstack/manila-scheduler-0" Oct 14 09:09:10 crc kubenswrapper[4870]: I1014 09:09:10.312581 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f1580185-56b7-4ef3-b148-e3934a5df6eb-config-data\") pod \"manila-scheduler-0\" (UID: \"f1580185-56b7-4ef3-b148-e3934a5df6eb\") " pod="openstack/manila-scheduler-0" Oct 14 09:09:10 crc kubenswrapper[4870]: I1014 09:09:10.312615 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f1580185-56b7-4ef3-b148-e3934a5df6eb-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"f1580185-56b7-4ef3-b148-e3934a5df6eb\") " pod="openstack/manila-scheduler-0" Oct 14 09:09:10 crc kubenswrapper[4870]: I1014 09:09:10.312810 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/dea5cc89-8770-43ab-a5f3-4986cfac76af-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"dea5cc89-8770-43ab-a5f3-4986cfac76af\") " pod="openstack/manila-share-share1-0" Oct 14 09:09:10 crc kubenswrapper[4870]: I1014 09:09:10.312931 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-df7mw\" (UniqueName: \"kubernetes.io/projected/dea5cc89-8770-43ab-a5f3-4986cfac76af-kube-api-access-df7mw\") pod \"manila-share-share1-0\" (UID: \"dea5cc89-8770-43ab-a5f3-4986cfac76af\") " pod="openstack/manila-share-share1-0" Oct 14 09:09:10 crc kubenswrapper[4870]: I1014 09:09:10.313059 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f1580185-56b7-4ef3-b148-e3934a5df6eb-scripts\") pod \"manila-scheduler-0\" (UID: \"f1580185-56b7-4ef3-b148-e3934a5df6eb\") " pod="openstack/manila-scheduler-0" Oct 14 09:09:10 crc kubenswrapper[4870]: I1014 09:09:10.313204 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/dea5cc89-8770-43ab-a5f3-4986cfac76af-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"dea5cc89-8770-43ab-a5f3-4986cfac76af\") " pod="openstack/manila-share-share1-0" Oct 14 09:09:10 crc kubenswrapper[4870]: I1014 09:09:10.313253 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/dea5cc89-8770-43ab-a5f3-4986cfac76af-ceph\") pod \"manila-share-share1-0\" (UID: \"dea5cc89-8770-43ab-a5f3-4986cfac76af\") " pod="openstack/manila-share-share1-0" Oct 14 09:09:10 crc kubenswrapper[4870]: I1014 09:09:10.416877 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/dea5cc89-8770-43ab-a5f3-4986cfac76af-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"dea5cc89-8770-43ab-a5f3-4986cfac76af\") " pod="openstack/manila-share-share1-0" Oct 14 09:09:10 crc kubenswrapper[4870]: I1014 09:09:10.416929 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/dea5cc89-8770-43ab-a5f3-4986cfac76af-ceph\") pod \"manila-share-share1-0\" (UID: \"dea5cc89-8770-43ab-a5f3-4986cfac76af\") " pod="openstack/manila-share-share1-0" Oct 14 09:09:10 crc kubenswrapper[4870]: I1014 09:09:10.416958 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/dea5cc89-8770-43ab-a5f3-4986cfac76af-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"dea5cc89-8770-43ab-a5f3-4986cfac76af\") " pod="openstack/manila-share-share1-0" Oct 14 09:09:10 crc kubenswrapper[4870]: I1014 09:09:10.416979 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-55dr8\" (UniqueName: \"kubernetes.io/projected/f1580185-56b7-4ef3-b148-e3934a5df6eb-kube-api-access-55dr8\") pod \"manila-scheduler-0\" (UID: \"f1580185-56b7-4ef3-b148-e3934a5df6eb\") " pod="openstack/manila-scheduler-0" Oct 14 09:09:10 crc kubenswrapper[4870]: I1014 09:09:10.417025 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dea5cc89-8770-43ab-a5f3-4986cfac76af-config-data\") pod \"manila-share-share1-0\" (UID: \"dea5cc89-8770-43ab-a5f3-4986cfac76af\") " pod="openstack/manila-share-share1-0" Oct 14 09:09:10 crc kubenswrapper[4870]: I1014 09:09:10.417066 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f1580185-56b7-4ef3-b148-e3934a5df6eb-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"f1580185-56b7-4ef3-b148-e3934a5df6eb\") " pod="openstack/manila-scheduler-0" Oct 14 09:09:10 crc kubenswrapper[4870]: I1014 09:09:10.417107 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dea5cc89-8770-43ab-a5f3-4986cfac76af-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"dea5cc89-8770-43ab-a5f3-4986cfac76af\") " pod="openstack/manila-share-share1-0" Oct 14 09:09:10 crc kubenswrapper[4870]: I1014 09:09:10.417177 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dea5cc89-8770-43ab-a5f3-4986cfac76af-scripts\") pod \"manila-share-share1-0\" (UID: \"dea5cc89-8770-43ab-a5f3-4986cfac76af\") " pod="openstack/manila-share-share1-0" Oct 14 09:09:10 crc kubenswrapper[4870]: I1014 09:09:10.417194 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f1580185-56b7-4ef3-b148-e3934a5df6eb-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"f1580185-56b7-4ef3-b148-e3934a5df6eb\") " pod="openstack/manila-scheduler-0" Oct 14 09:09:10 crc kubenswrapper[4870]: I1014 09:09:10.417218 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f1580185-56b7-4ef3-b148-e3934a5df6eb-config-data\") pod \"manila-scheduler-0\" (UID: \"f1580185-56b7-4ef3-b148-e3934a5df6eb\") " pod="openstack/manila-scheduler-0" Oct 14 09:09:10 crc kubenswrapper[4870]: I1014 09:09:10.417258 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f1580185-56b7-4ef3-b148-e3934a5df6eb-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"f1580185-56b7-4ef3-b148-e3934a5df6eb\") " pod="openstack/manila-scheduler-0" Oct 14 09:09:10 crc kubenswrapper[4870]: I1014 09:09:10.417287 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/dea5cc89-8770-43ab-a5f3-4986cfac76af-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"dea5cc89-8770-43ab-a5f3-4986cfac76af\") " pod="openstack/manila-share-share1-0" Oct 14 09:09:10 crc kubenswrapper[4870]: I1014 09:09:10.417312 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-df7mw\" (UniqueName: \"kubernetes.io/projected/dea5cc89-8770-43ab-a5f3-4986cfac76af-kube-api-access-df7mw\") pod \"manila-share-share1-0\" (UID: \"dea5cc89-8770-43ab-a5f3-4986cfac76af\") " pod="openstack/manila-share-share1-0" Oct 14 09:09:10 crc kubenswrapper[4870]: I1014 09:09:10.417359 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f1580185-56b7-4ef3-b148-e3934a5df6eb-scripts\") pod \"manila-scheduler-0\" (UID: \"f1580185-56b7-4ef3-b148-e3934a5df6eb\") " pod="openstack/manila-scheduler-0" Oct 14 09:09:10 crc kubenswrapper[4870]: I1014 09:09:10.420701 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/dea5cc89-8770-43ab-a5f3-4986cfac76af-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"dea5cc89-8770-43ab-a5f3-4986cfac76af\") " pod="openstack/manila-share-share1-0" Oct 14 09:09:10 crc kubenswrapper[4870]: I1014 09:09:10.420766 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/dea5cc89-8770-43ab-a5f3-4986cfac76af-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"dea5cc89-8770-43ab-a5f3-4986cfac76af\") " pod="openstack/manila-share-share1-0" Oct 14 09:09:10 crc kubenswrapper[4870]: I1014 09:09:10.422277 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f1580185-56b7-4ef3-b148-e3934a5df6eb-scripts\") pod \"manila-scheduler-0\" (UID: \"f1580185-56b7-4ef3-b148-e3934a5df6eb\") " pod="openstack/manila-scheduler-0" Oct 14 09:09:10 crc kubenswrapper[4870]: I1014 09:09:10.424506 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-849c9dd687-6z7z9"] Oct 14 09:09:10 crc kubenswrapper[4870]: I1014 09:09:10.426294 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-849c9dd687-6z7z9" Oct 14 09:09:10 crc kubenswrapper[4870]: I1014 09:09:10.427994 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dea5cc89-8770-43ab-a5f3-4986cfac76af-scripts\") pod \"manila-share-share1-0\" (UID: \"dea5cc89-8770-43ab-a5f3-4986cfac76af\") " pod="openstack/manila-share-share1-0" Oct 14 09:09:10 crc kubenswrapper[4870]: I1014 09:09:10.428069 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f1580185-56b7-4ef3-b148-e3934a5df6eb-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"f1580185-56b7-4ef3-b148-e3934a5df6eb\") " pod="openstack/manila-scheduler-0" Oct 14 09:09:10 crc kubenswrapper[4870]: I1014 09:09:10.430564 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/dea5cc89-8770-43ab-a5f3-4986cfac76af-ceph\") pod \"manila-share-share1-0\" (UID: \"dea5cc89-8770-43ab-a5f3-4986cfac76af\") " pod="openstack/manila-share-share1-0" Oct 14 09:09:10 crc kubenswrapper[4870]: I1014 09:09:10.450370 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dea5cc89-8770-43ab-a5f3-4986cfac76af-config-data\") pod \"manila-share-share1-0\" (UID: \"dea5cc89-8770-43ab-a5f3-4986cfac76af\") " pod="openstack/manila-share-share1-0" Oct 14 09:09:10 crc kubenswrapper[4870]: I1014 09:09:10.451776 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-849c9dd687-6z7z9"] Oct 14 09:09:10 crc kubenswrapper[4870]: I1014 09:09:10.453210 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f1580185-56b7-4ef3-b148-e3934a5df6eb-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"f1580185-56b7-4ef3-b148-e3934a5df6eb\") " pod="openstack/manila-scheduler-0" Oct 14 09:09:10 crc kubenswrapper[4870]: I1014 09:09:10.454308 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dea5cc89-8770-43ab-a5f3-4986cfac76af-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"dea5cc89-8770-43ab-a5f3-4986cfac76af\") " pod="openstack/manila-share-share1-0" Oct 14 09:09:10 crc kubenswrapper[4870]: I1014 09:09:10.456907 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f1580185-56b7-4ef3-b148-e3934a5df6eb-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"f1580185-56b7-4ef3-b148-e3934a5df6eb\") " pod="openstack/manila-scheduler-0" Oct 14 09:09:10 crc kubenswrapper[4870]: I1014 09:09:10.462194 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/dea5cc89-8770-43ab-a5f3-4986cfac76af-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"dea5cc89-8770-43ab-a5f3-4986cfac76af\") " pod="openstack/manila-share-share1-0" Oct 14 09:09:10 crc kubenswrapper[4870]: I1014 09:09:10.464917 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-55dr8\" (UniqueName: \"kubernetes.io/projected/f1580185-56b7-4ef3-b148-e3934a5df6eb-kube-api-access-55dr8\") pod \"manila-scheduler-0\" (UID: \"f1580185-56b7-4ef3-b148-e3934a5df6eb\") " pod="openstack/manila-scheduler-0" Oct 14 09:09:10 crc kubenswrapper[4870]: I1014 09:09:10.478179 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f1580185-56b7-4ef3-b148-e3934a5df6eb-config-data\") pod \"manila-scheduler-0\" (UID: \"f1580185-56b7-4ef3-b148-e3934a5df6eb\") " pod="openstack/manila-scheduler-0" Oct 14 09:09:10 crc kubenswrapper[4870]: I1014 09:09:10.480156 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-df7mw\" (UniqueName: \"kubernetes.io/projected/dea5cc89-8770-43ab-a5f3-4986cfac76af-kube-api-access-df7mw\") pod \"manila-share-share1-0\" (UID: \"dea5cc89-8770-43ab-a5f3-4986cfac76af\") " pod="openstack/manila-share-share1-0" Oct 14 09:09:10 crc kubenswrapper[4870]: I1014 09:09:10.519330 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b5697993-d512-4996-b3d0-dce02c8b8114-config\") pod \"dnsmasq-dns-849c9dd687-6z7z9\" (UID: \"b5697993-d512-4996-b3d0-dce02c8b8114\") " pod="openstack/dnsmasq-dns-849c9dd687-6z7z9" Oct 14 09:09:10 crc kubenswrapper[4870]: I1014 09:09:10.519392 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b5697993-d512-4996-b3d0-dce02c8b8114-ovsdbserver-nb\") pod \"dnsmasq-dns-849c9dd687-6z7z9\" (UID: \"b5697993-d512-4996-b3d0-dce02c8b8114\") " pod="openstack/dnsmasq-dns-849c9dd687-6z7z9" Oct 14 09:09:10 crc kubenswrapper[4870]: I1014 09:09:10.525269 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b5697993-d512-4996-b3d0-dce02c8b8114-ovsdbserver-sb\") pod \"dnsmasq-dns-849c9dd687-6z7z9\" (UID: \"b5697993-d512-4996-b3d0-dce02c8b8114\") " pod="openstack/dnsmasq-dns-849c9dd687-6z7z9" Oct 14 09:09:10 crc kubenswrapper[4870]: I1014 09:09:10.525352 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b5697993-d512-4996-b3d0-dce02c8b8114-dns-svc\") pod \"dnsmasq-dns-849c9dd687-6z7z9\" (UID: \"b5697993-d512-4996-b3d0-dce02c8b8114\") " pod="openstack/dnsmasq-dns-849c9dd687-6z7z9" Oct 14 09:09:10 crc kubenswrapper[4870]: I1014 09:09:10.525430 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xfzh9\" (UniqueName: \"kubernetes.io/projected/b5697993-d512-4996-b3d0-dce02c8b8114-kube-api-access-xfzh9\") pod \"dnsmasq-dns-849c9dd687-6z7z9\" (UID: \"b5697993-d512-4996-b3d0-dce02c8b8114\") " pod="openstack/dnsmasq-dns-849c9dd687-6z7z9" Oct 14 09:09:10 crc kubenswrapper[4870]: I1014 09:09:10.547479 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Oct 14 09:09:10 crc kubenswrapper[4870]: I1014 09:09:10.584801 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Oct 14 09:09:10 crc kubenswrapper[4870]: I1014 09:09:10.585367 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-api-0"] Oct 14 09:09:10 crc kubenswrapper[4870]: I1014 09:09:10.588124 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Oct 14 09:09:10 crc kubenswrapper[4870]: I1014 09:09:10.590540 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-api-config-data" Oct 14 09:09:10 crc kubenswrapper[4870]: I1014 09:09:10.603708 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-api-0"] Oct 14 09:09:10 crc kubenswrapper[4870]: I1014 09:09:10.627229 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0e863093-887f-414c-bfed-ffb1be0f0033-etc-machine-id\") pod \"manila-api-0\" (UID: \"0e863093-887f-414c-bfed-ffb1be0f0033\") " pod="openstack/manila-api-0" Oct 14 09:09:10 crc kubenswrapper[4870]: I1014 09:09:10.627276 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0e863093-887f-414c-bfed-ffb1be0f0033-scripts\") pod \"manila-api-0\" (UID: \"0e863093-887f-414c-bfed-ffb1be0f0033\") " pod="openstack/manila-api-0" Oct 14 09:09:10 crc kubenswrapper[4870]: I1014 09:09:10.627315 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b5697993-d512-4996-b3d0-dce02c8b8114-config\") pod \"dnsmasq-dns-849c9dd687-6z7z9\" (UID: \"b5697993-d512-4996-b3d0-dce02c8b8114\") " pod="openstack/dnsmasq-dns-849c9dd687-6z7z9" Oct 14 09:09:10 crc kubenswrapper[4870]: I1014 09:09:10.627346 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b5697993-d512-4996-b3d0-dce02c8b8114-ovsdbserver-nb\") pod \"dnsmasq-dns-849c9dd687-6z7z9\" (UID: \"b5697993-d512-4996-b3d0-dce02c8b8114\") " pod="openstack/dnsmasq-dns-849c9dd687-6z7z9" Oct 14 09:09:10 crc kubenswrapper[4870]: I1014 09:09:10.627378 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e863093-887f-414c-bfed-ffb1be0f0033-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"0e863093-887f-414c-bfed-ffb1be0f0033\") " pod="openstack/manila-api-0" Oct 14 09:09:10 crc kubenswrapper[4870]: I1014 09:09:10.627402 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0e863093-887f-414c-bfed-ffb1be0f0033-logs\") pod \"manila-api-0\" (UID: \"0e863093-887f-414c-bfed-ffb1be0f0033\") " pod="openstack/manila-api-0" Oct 14 09:09:10 crc kubenswrapper[4870]: I1014 09:09:10.627737 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j2f54\" (UniqueName: \"kubernetes.io/projected/0e863093-887f-414c-bfed-ffb1be0f0033-kube-api-access-j2f54\") pod \"manila-api-0\" (UID: \"0e863093-887f-414c-bfed-ffb1be0f0033\") " pod="openstack/manila-api-0" Oct 14 09:09:10 crc kubenswrapper[4870]: I1014 09:09:10.628432 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b5697993-d512-4996-b3d0-dce02c8b8114-config\") pod \"dnsmasq-dns-849c9dd687-6z7z9\" (UID: \"b5697993-d512-4996-b3d0-dce02c8b8114\") " pod="openstack/dnsmasq-dns-849c9dd687-6z7z9" Oct 14 09:09:10 crc kubenswrapper[4870]: I1014 09:09:10.628602 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b5697993-d512-4996-b3d0-dce02c8b8114-ovsdbserver-nb\") pod \"dnsmasq-dns-849c9dd687-6z7z9\" (UID: \"b5697993-d512-4996-b3d0-dce02c8b8114\") " pod="openstack/dnsmasq-dns-849c9dd687-6z7z9" Oct 14 09:09:10 crc kubenswrapper[4870]: I1014 09:09:10.628722 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b5697993-d512-4996-b3d0-dce02c8b8114-ovsdbserver-sb\") pod \"dnsmasq-dns-849c9dd687-6z7z9\" (UID: \"b5697993-d512-4996-b3d0-dce02c8b8114\") " pod="openstack/dnsmasq-dns-849c9dd687-6z7z9" Oct 14 09:09:10 crc kubenswrapper[4870]: I1014 09:09:10.628758 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b5697993-d512-4996-b3d0-dce02c8b8114-dns-svc\") pod \"dnsmasq-dns-849c9dd687-6z7z9\" (UID: \"b5697993-d512-4996-b3d0-dce02c8b8114\") " pod="openstack/dnsmasq-dns-849c9dd687-6z7z9" Oct 14 09:09:10 crc kubenswrapper[4870]: I1014 09:09:10.628780 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0e863093-887f-414c-bfed-ffb1be0f0033-config-data-custom\") pod \"manila-api-0\" (UID: \"0e863093-887f-414c-bfed-ffb1be0f0033\") " pod="openstack/manila-api-0" Oct 14 09:09:10 crc kubenswrapper[4870]: I1014 09:09:10.628816 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0e863093-887f-414c-bfed-ffb1be0f0033-config-data\") pod \"manila-api-0\" (UID: \"0e863093-887f-414c-bfed-ffb1be0f0033\") " pod="openstack/manila-api-0" Oct 14 09:09:10 crc kubenswrapper[4870]: I1014 09:09:10.628835 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xfzh9\" (UniqueName: \"kubernetes.io/projected/b5697993-d512-4996-b3d0-dce02c8b8114-kube-api-access-xfzh9\") pod \"dnsmasq-dns-849c9dd687-6z7z9\" (UID: \"b5697993-d512-4996-b3d0-dce02c8b8114\") " pod="openstack/dnsmasq-dns-849c9dd687-6z7z9" Oct 14 09:09:10 crc kubenswrapper[4870]: I1014 09:09:10.629646 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b5697993-d512-4996-b3d0-dce02c8b8114-ovsdbserver-sb\") pod \"dnsmasq-dns-849c9dd687-6z7z9\" (UID: \"b5697993-d512-4996-b3d0-dce02c8b8114\") " pod="openstack/dnsmasq-dns-849c9dd687-6z7z9" Oct 14 09:09:10 crc kubenswrapper[4870]: I1014 09:09:10.630128 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b5697993-d512-4996-b3d0-dce02c8b8114-dns-svc\") pod \"dnsmasq-dns-849c9dd687-6z7z9\" (UID: \"b5697993-d512-4996-b3d0-dce02c8b8114\") " pod="openstack/dnsmasq-dns-849c9dd687-6z7z9" Oct 14 09:09:10 crc kubenswrapper[4870]: I1014 09:09:10.649298 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xfzh9\" (UniqueName: \"kubernetes.io/projected/b5697993-d512-4996-b3d0-dce02c8b8114-kube-api-access-xfzh9\") pod \"dnsmasq-dns-849c9dd687-6z7z9\" (UID: \"b5697993-d512-4996-b3d0-dce02c8b8114\") " pod="openstack/dnsmasq-dns-849c9dd687-6z7z9" Oct 14 09:09:10 crc kubenswrapper[4870]: I1014 09:09:10.750072 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0e863093-887f-414c-bfed-ffb1be0f0033-etc-machine-id\") pod \"manila-api-0\" (UID: \"0e863093-887f-414c-bfed-ffb1be0f0033\") " pod="openstack/manila-api-0" Oct 14 09:09:10 crc kubenswrapper[4870]: I1014 09:09:10.750388 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0e863093-887f-414c-bfed-ffb1be0f0033-scripts\") pod \"manila-api-0\" (UID: \"0e863093-887f-414c-bfed-ffb1be0f0033\") " pod="openstack/manila-api-0" Oct 14 09:09:10 crc kubenswrapper[4870]: I1014 09:09:10.750459 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e863093-887f-414c-bfed-ffb1be0f0033-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"0e863093-887f-414c-bfed-ffb1be0f0033\") " pod="openstack/manila-api-0" Oct 14 09:09:10 crc kubenswrapper[4870]: I1014 09:09:10.750486 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0e863093-887f-414c-bfed-ffb1be0f0033-logs\") pod \"manila-api-0\" (UID: \"0e863093-887f-414c-bfed-ffb1be0f0033\") " pod="openstack/manila-api-0" Oct 14 09:09:10 crc kubenswrapper[4870]: I1014 09:09:10.750532 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j2f54\" (UniqueName: \"kubernetes.io/projected/0e863093-887f-414c-bfed-ffb1be0f0033-kube-api-access-j2f54\") pod \"manila-api-0\" (UID: \"0e863093-887f-414c-bfed-ffb1be0f0033\") " pod="openstack/manila-api-0" Oct 14 09:09:10 crc kubenswrapper[4870]: I1014 09:09:10.750574 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0e863093-887f-414c-bfed-ffb1be0f0033-config-data-custom\") pod \"manila-api-0\" (UID: \"0e863093-887f-414c-bfed-ffb1be0f0033\") " pod="openstack/manila-api-0" Oct 14 09:09:10 crc kubenswrapper[4870]: I1014 09:09:10.750601 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0e863093-887f-414c-bfed-ffb1be0f0033-config-data\") pod \"manila-api-0\" (UID: \"0e863093-887f-414c-bfed-ffb1be0f0033\") " pod="openstack/manila-api-0" Oct 14 09:09:10 crc kubenswrapper[4870]: I1014 09:09:10.750252 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0e863093-887f-414c-bfed-ffb1be0f0033-etc-machine-id\") pod \"manila-api-0\" (UID: \"0e863093-887f-414c-bfed-ffb1be0f0033\") " pod="openstack/manila-api-0" Oct 14 09:09:10 crc kubenswrapper[4870]: I1014 09:09:10.757139 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0e863093-887f-414c-bfed-ffb1be0f0033-logs\") pod \"manila-api-0\" (UID: \"0e863093-887f-414c-bfed-ffb1be0f0033\") " pod="openstack/manila-api-0" Oct 14 09:09:10 crc kubenswrapper[4870]: I1014 09:09:10.759895 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e863093-887f-414c-bfed-ffb1be0f0033-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"0e863093-887f-414c-bfed-ffb1be0f0033\") " pod="openstack/manila-api-0" Oct 14 09:09:10 crc kubenswrapper[4870]: I1014 09:09:10.760511 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0e863093-887f-414c-bfed-ffb1be0f0033-config-data-custom\") pod \"manila-api-0\" (UID: \"0e863093-887f-414c-bfed-ffb1be0f0033\") " pod="openstack/manila-api-0" Oct 14 09:09:10 crc kubenswrapper[4870]: I1014 09:09:10.762256 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0e863093-887f-414c-bfed-ffb1be0f0033-config-data\") pod \"manila-api-0\" (UID: \"0e863093-887f-414c-bfed-ffb1be0f0033\") " pod="openstack/manila-api-0" Oct 14 09:09:10 crc kubenswrapper[4870]: I1014 09:09:10.777797 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j2f54\" (UniqueName: \"kubernetes.io/projected/0e863093-887f-414c-bfed-ffb1be0f0033-kube-api-access-j2f54\") pod \"manila-api-0\" (UID: \"0e863093-887f-414c-bfed-ffb1be0f0033\") " pod="openstack/manila-api-0" Oct 14 09:09:10 crc kubenswrapper[4870]: I1014 09:09:10.778949 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0e863093-887f-414c-bfed-ffb1be0f0033-scripts\") pod \"manila-api-0\" (UID: \"0e863093-887f-414c-bfed-ffb1be0f0033\") " pod="openstack/manila-api-0" Oct 14 09:09:10 crc kubenswrapper[4870]: I1014 09:09:10.784537 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-849c9dd687-6z7z9" Oct 14 09:09:10 crc kubenswrapper[4870]: I1014 09:09:10.798179 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Oct 14 09:09:11 crc kubenswrapper[4870]: I1014 09:09:11.121486 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-scheduler-0"] Oct 14 09:09:11 crc kubenswrapper[4870]: I1014 09:09:11.200122 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Oct 14 09:09:11 crc kubenswrapper[4870]: I1014 09:09:11.381761 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-share-share1-0"] Oct 14 09:09:11 crc kubenswrapper[4870]: I1014 09:09:11.416334 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-849c9dd687-6z7z9"] Oct 14 09:09:11 crc kubenswrapper[4870]: I1014 09:09:11.494295 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-api-0"] Oct 14 09:09:11 crc kubenswrapper[4870]: I1014 09:09:11.930224 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-849c9dd687-6z7z9" event={"ID":"b5697993-d512-4996-b3d0-dce02c8b8114","Type":"ContainerStarted","Data":"f312035f067b128f5801edc80f77d1ab8cb42e4a382cb303068a5d2429226e96"} Oct 14 09:09:11 crc kubenswrapper[4870]: I1014 09:09:11.930414 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-849c9dd687-6z7z9" event={"ID":"b5697993-d512-4996-b3d0-dce02c8b8114","Type":"ContainerStarted","Data":"596b4b8116137f5df05039c3af7209fe77be3937cf83f26780fad788c94a72a4"} Oct 14 09:09:11 crc kubenswrapper[4870]: I1014 09:09:11.935744 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"dea5cc89-8770-43ab-a5f3-4986cfac76af","Type":"ContainerStarted","Data":"6af532beb5ef92c2be1066525a8b7d03eee1d1dc8eb271983950253aab72a459"} Oct 14 09:09:11 crc kubenswrapper[4870]: I1014 09:09:11.944685 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"0e863093-887f-414c-bfed-ffb1be0f0033","Type":"ContainerStarted","Data":"927d118ef8b4f8f1a56342ed09df334b4d2648ecc6793ccf7b157ca25e8acbdf"} Oct 14 09:09:11 crc kubenswrapper[4870]: I1014 09:09:11.960413 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"f1580185-56b7-4ef3-b148-e3934a5df6eb","Type":"ContainerStarted","Data":"887f82e375d3d7487238c82e9cc9319e6117612d1cc00d2615ca918fb06317ee"} Oct 14 09:09:12 crc kubenswrapper[4870]: I1014 09:09:12.977335 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"f1580185-56b7-4ef3-b148-e3934a5df6eb","Type":"ContainerStarted","Data":"efdaadebeffef069ddd13d03dbead314ec261cdaafe620c7c7401a49b44f5ec0"} Oct 14 09:09:12 crc kubenswrapper[4870]: I1014 09:09:12.977947 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"f1580185-56b7-4ef3-b148-e3934a5df6eb","Type":"ContainerStarted","Data":"cd0a9ff20960bc52f57d3321a2a4cc10760f0e676cc536061c3e3889777f7f02"} Oct 14 09:09:12 crc kubenswrapper[4870]: I1014 09:09:12.981406 4870 generic.go:334] "Generic (PLEG): container finished" podID="b5697993-d512-4996-b3d0-dce02c8b8114" containerID="f312035f067b128f5801edc80f77d1ab8cb42e4a382cb303068a5d2429226e96" exitCode=0 Oct 14 09:09:12 crc kubenswrapper[4870]: I1014 09:09:12.982108 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-849c9dd687-6z7z9" event={"ID":"b5697993-d512-4996-b3d0-dce02c8b8114","Type":"ContainerDied","Data":"f312035f067b128f5801edc80f77d1ab8cb42e4a382cb303068a5d2429226e96"} Oct 14 09:09:12 crc kubenswrapper[4870]: I1014 09:09:12.982133 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-849c9dd687-6z7z9" event={"ID":"b5697993-d512-4996-b3d0-dce02c8b8114","Type":"ContainerStarted","Data":"37f1920472504bf0e5fe2a0d3ddb02ba22238fccc5e86737d9be4e16e27ca98b"} Oct 14 09:09:12 crc kubenswrapper[4870]: I1014 09:09:12.982151 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-849c9dd687-6z7z9" Oct 14 09:09:12 crc kubenswrapper[4870]: I1014 09:09:12.988816 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"0e863093-887f-414c-bfed-ffb1be0f0033","Type":"ContainerStarted","Data":"7129226adfa80a6566c5d3d617e80a1e78a471055f5d8e778e1a96cf35478b43"} Oct 14 09:09:12 crc kubenswrapper[4870]: I1014 09:09:12.988866 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"0e863093-887f-414c-bfed-ffb1be0f0033","Type":"ContainerStarted","Data":"c0d954779dc2035661e7b2568c4926e26532780644b061caa7ad7d6f756b3391"} Oct 14 09:09:12 crc kubenswrapper[4870]: I1014 09:09:12.996337 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/manila-api-0" Oct 14 09:09:13 crc kubenswrapper[4870]: I1014 09:09:13.009734 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-scheduler-0" podStartSLOduration=2.653929744 podStartE2EDuration="3.009703058s" podCreationTimestamp="2025-10-14 09:09:10 +0000 UTC" firstStartedPulling="2025-10-14 09:09:11.112603076 +0000 UTC m=+7686.809963447" lastFinishedPulling="2025-10-14 09:09:11.46837639 +0000 UTC m=+7687.165736761" observedRunningTime="2025-10-14 09:09:13.006233013 +0000 UTC m=+7688.703593384" watchObservedRunningTime="2025-10-14 09:09:13.009703058 +0000 UTC m=+7688.707063429" Oct 14 09:09:13 crc kubenswrapper[4870]: I1014 09:09:13.040472 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-api-0" podStartSLOduration=3.040452407 podStartE2EDuration="3.040452407s" podCreationTimestamp="2025-10-14 09:09:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 09:09:13.030325197 +0000 UTC m=+7688.727685568" watchObservedRunningTime="2025-10-14 09:09:13.040452407 +0000 UTC m=+7688.737812778" Oct 14 09:09:13 crc kubenswrapper[4870]: I1014 09:09:13.068017 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-849c9dd687-6z7z9" podStartSLOduration=3.067996418 podStartE2EDuration="3.067996418s" podCreationTimestamp="2025-10-14 09:09:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 09:09:13.050888945 +0000 UTC m=+7688.748249316" watchObservedRunningTime="2025-10-14 09:09:13.067996418 +0000 UTC m=+7688.765356789" Oct 14 09:09:14 crc kubenswrapper[4870]: I1014 09:09:14.051722 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-f10d-account-create-pkpj9"] Oct 14 09:09:14 crc kubenswrapper[4870]: I1014 09:09:14.063393 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-e4ea-account-create-8t749"] Oct 14 09:09:14 crc kubenswrapper[4870]: I1014 09:09:14.069070 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-f10d-account-create-pkpj9"] Oct 14 09:09:14 crc kubenswrapper[4870]: I1014 09:09:14.076455 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-e4ea-account-create-8t749"] Oct 14 09:09:15 crc kubenswrapper[4870]: I1014 09:09:15.031995 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-66ee-account-create-wjnkl"] Oct 14 09:09:15 crc kubenswrapper[4870]: I1014 09:09:15.048896 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ee31f93f-00ce-45ea-8e8c-4c12036e3d11" path="/var/lib/kubelet/pods/ee31f93f-00ce-45ea-8e8c-4c12036e3d11/volumes" Oct 14 09:09:15 crc kubenswrapper[4870]: I1014 09:09:15.049594 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f022b544-37f1-4e1e-a1bc-614f4f36b8c3" path="/var/lib/kubelet/pods/f022b544-37f1-4e1e-a1bc-614f4f36b8c3/volumes" Oct 14 09:09:15 crc kubenswrapper[4870]: I1014 09:09:15.050264 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-66ee-account-create-wjnkl"] Oct 14 09:09:16 crc kubenswrapper[4870]: I1014 09:09:16.033480 4870 scope.go:117] "RemoveContainer" containerID="8fcaeb3090f0c89cba1e66db524b4a246072daa35592012664aca450fa09ddc9" Oct 14 09:09:16 crc kubenswrapper[4870]: E1014 09:09:16.034021 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 09:09:17 crc kubenswrapper[4870]: I1014 09:09:17.048132 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="984f8c9e-ebdd-4c8f-b093-8d30b38b5a2a" path="/var/lib/kubelet/pods/984f8c9e-ebdd-4c8f-b093-8d30b38b5a2a/volumes" Oct 14 09:09:20 crc kubenswrapper[4870]: I1014 09:09:20.079359 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"dea5cc89-8770-43ab-a5f3-4986cfac76af","Type":"ContainerStarted","Data":"dd46c21024eaeffa23981ccc06077e56568b96004e9e3340edeaebd3dbc9ba6c"} Oct 14 09:09:20 crc kubenswrapper[4870]: I1014 09:09:20.080000 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"dea5cc89-8770-43ab-a5f3-4986cfac76af","Type":"ContainerStarted","Data":"27cc803273feb0b61460ea245990a7bd8886270ff4bb5611b4c97551f00dac71"} Oct 14 09:09:20 crc kubenswrapper[4870]: I1014 09:09:20.104705 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-share-share1-0" podStartSLOduration=2.55577416 podStartE2EDuration="10.104684527s" podCreationTimestamp="2025-10-14 09:09:10 +0000 UTC" firstStartedPulling="2025-10-14 09:09:11.413563367 +0000 UTC m=+7687.110923738" lastFinishedPulling="2025-10-14 09:09:18.962473724 +0000 UTC m=+7694.659834105" observedRunningTime="2025-10-14 09:09:20.100852082 +0000 UTC m=+7695.798212443" watchObservedRunningTime="2025-10-14 09:09:20.104684527 +0000 UTC m=+7695.802044898" Oct 14 09:09:20 crc kubenswrapper[4870]: I1014 09:09:20.548616 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/manila-scheduler-0" Oct 14 09:09:20 crc kubenswrapper[4870]: I1014 09:09:20.586066 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/manila-share-share1-0" Oct 14 09:09:20 crc kubenswrapper[4870]: I1014 09:09:20.786656 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-849c9dd687-6z7z9" Oct 14 09:09:20 crc kubenswrapper[4870]: I1014 09:09:20.894508 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6b44c8b5f5-47x86"] Oct 14 09:09:20 crc kubenswrapper[4870]: I1014 09:09:20.894991 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6b44c8b5f5-47x86" podUID="d666d204-b974-4c9a-aaba-6d02302fc332" containerName="dnsmasq-dns" containerID="cri-o://654085a705d3044a54f19852baa77fbbc5a4c6cb8bff2383483e23651297a478" gracePeriod=10 Oct 14 09:09:21 crc kubenswrapper[4870]: I1014 09:09:21.098407 4870 generic.go:334] "Generic (PLEG): container finished" podID="d666d204-b974-4c9a-aaba-6d02302fc332" containerID="654085a705d3044a54f19852baa77fbbc5a4c6cb8bff2383483e23651297a478" exitCode=0 Oct 14 09:09:21 crc kubenswrapper[4870]: I1014 09:09:21.099349 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b44c8b5f5-47x86" event={"ID":"d666d204-b974-4c9a-aaba-6d02302fc332","Type":"ContainerDied","Data":"654085a705d3044a54f19852baa77fbbc5a4c6cb8bff2383483e23651297a478"} Oct 14 09:09:21 crc kubenswrapper[4870]: I1014 09:09:21.577700 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b44c8b5f5-47x86" Oct 14 09:09:21 crc kubenswrapper[4870]: I1014 09:09:21.714190 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d666d204-b974-4c9a-aaba-6d02302fc332-dns-svc\") pod \"d666d204-b974-4c9a-aaba-6d02302fc332\" (UID: \"d666d204-b974-4c9a-aaba-6d02302fc332\") " Oct 14 09:09:21 crc kubenswrapper[4870]: I1014 09:09:21.714274 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2qzvg\" (UniqueName: \"kubernetes.io/projected/d666d204-b974-4c9a-aaba-6d02302fc332-kube-api-access-2qzvg\") pod \"d666d204-b974-4c9a-aaba-6d02302fc332\" (UID: \"d666d204-b974-4c9a-aaba-6d02302fc332\") " Oct 14 09:09:21 crc kubenswrapper[4870]: I1014 09:09:21.714350 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d666d204-b974-4c9a-aaba-6d02302fc332-config\") pod \"d666d204-b974-4c9a-aaba-6d02302fc332\" (UID: \"d666d204-b974-4c9a-aaba-6d02302fc332\") " Oct 14 09:09:21 crc kubenswrapper[4870]: I1014 09:09:21.714430 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d666d204-b974-4c9a-aaba-6d02302fc332-ovsdbserver-nb\") pod \"d666d204-b974-4c9a-aaba-6d02302fc332\" (UID: \"d666d204-b974-4c9a-aaba-6d02302fc332\") " Oct 14 09:09:21 crc kubenswrapper[4870]: I1014 09:09:21.714818 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d666d204-b974-4c9a-aaba-6d02302fc332-ovsdbserver-sb\") pod \"d666d204-b974-4c9a-aaba-6d02302fc332\" (UID: \"d666d204-b974-4c9a-aaba-6d02302fc332\") " Oct 14 09:09:21 crc kubenswrapper[4870]: I1014 09:09:21.722592 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d666d204-b974-4c9a-aaba-6d02302fc332-kube-api-access-2qzvg" (OuterVolumeSpecName: "kube-api-access-2qzvg") pod "d666d204-b974-4c9a-aaba-6d02302fc332" (UID: "d666d204-b974-4c9a-aaba-6d02302fc332"). InnerVolumeSpecName "kube-api-access-2qzvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:09:21 crc kubenswrapper[4870]: I1014 09:09:21.766376 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d666d204-b974-4c9a-aaba-6d02302fc332-config" (OuterVolumeSpecName: "config") pod "d666d204-b974-4c9a-aaba-6d02302fc332" (UID: "d666d204-b974-4c9a-aaba-6d02302fc332"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 09:09:21 crc kubenswrapper[4870]: I1014 09:09:21.779601 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d666d204-b974-4c9a-aaba-6d02302fc332-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "d666d204-b974-4c9a-aaba-6d02302fc332" (UID: "d666d204-b974-4c9a-aaba-6d02302fc332"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 09:09:21 crc kubenswrapper[4870]: I1014 09:09:21.786979 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d666d204-b974-4c9a-aaba-6d02302fc332-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "d666d204-b974-4c9a-aaba-6d02302fc332" (UID: "d666d204-b974-4c9a-aaba-6d02302fc332"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 09:09:21 crc kubenswrapper[4870]: I1014 09:09:21.798429 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d666d204-b974-4c9a-aaba-6d02302fc332-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "d666d204-b974-4c9a-aaba-6d02302fc332" (UID: "d666d204-b974-4c9a-aaba-6d02302fc332"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 09:09:21 crc kubenswrapper[4870]: I1014 09:09:21.816920 4870 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d666d204-b974-4c9a-aaba-6d02302fc332-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 14 09:09:21 crc kubenswrapper[4870]: I1014 09:09:21.816955 4870 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d666d204-b974-4c9a-aaba-6d02302fc332-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 14 09:09:21 crc kubenswrapper[4870]: I1014 09:09:21.816971 4870 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d666d204-b974-4c9a-aaba-6d02302fc332-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 14 09:09:21 crc kubenswrapper[4870]: I1014 09:09:21.816989 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2qzvg\" (UniqueName: \"kubernetes.io/projected/d666d204-b974-4c9a-aaba-6d02302fc332-kube-api-access-2qzvg\") on node \"crc\" DevicePath \"\"" Oct 14 09:09:21 crc kubenswrapper[4870]: I1014 09:09:21.817003 4870 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d666d204-b974-4c9a-aaba-6d02302fc332-config\") on node \"crc\" DevicePath \"\"" Oct 14 09:09:22 crc kubenswrapper[4870]: I1014 09:09:22.109602 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b44c8b5f5-47x86" event={"ID":"d666d204-b974-4c9a-aaba-6d02302fc332","Type":"ContainerDied","Data":"2f090c359b4dfbaf97b1ac18905a081b8a4a20d8ac2fb08f4290e4ec192971e3"} Oct 14 09:09:22 crc kubenswrapper[4870]: I1014 09:09:22.109667 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b44c8b5f5-47x86" Oct 14 09:09:22 crc kubenswrapper[4870]: I1014 09:09:22.109682 4870 scope.go:117] "RemoveContainer" containerID="654085a705d3044a54f19852baa77fbbc5a4c6cb8bff2383483e23651297a478" Oct 14 09:09:22 crc kubenswrapper[4870]: I1014 09:09:22.141787 4870 scope.go:117] "RemoveContainer" containerID="99dca1d602683adf2005164f939b7d374fb1c110208d96e73dd1db9ad628d9d1" Oct 14 09:09:22 crc kubenswrapper[4870]: I1014 09:09:22.162274 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6b44c8b5f5-47x86"] Oct 14 09:09:22 crc kubenswrapper[4870]: I1014 09:09:22.173921 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6b44c8b5f5-47x86"] Oct 14 09:09:23 crc kubenswrapper[4870]: I1014 09:09:23.047023 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d666d204-b974-4c9a-aaba-6d02302fc332" path="/var/lib/kubelet/pods/d666d204-b974-4c9a-aaba-6d02302fc332/volumes" Oct 14 09:09:23 crc kubenswrapper[4870]: I1014 09:09:23.870281 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 14 09:09:23 crc kubenswrapper[4870]: I1014 09:09:23.871940 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0c1abbc6-c3c5-4fae-8200-fa3633518f50" containerName="ceilometer-central-agent" containerID="cri-o://6c66907b3ba7cfdfa4c257675d38768ee4dfd691a99ca1aebf71de86c862b364" gracePeriod=30 Oct 14 09:09:23 crc kubenswrapper[4870]: I1014 09:09:23.871971 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0c1abbc6-c3c5-4fae-8200-fa3633518f50" containerName="proxy-httpd" containerID="cri-o://fc6a3a9c4e15f823dafcbbfd5aa9442770632652275e8ec060b454c537254c5c" gracePeriod=30 Oct 14 09:09:23 crc kubenswrapper[4870]: I1014 09:09:23.871988 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0c1abbc6-c3c5-4fae-8200-fa3633518f50" containerName="sg-core" containerID="cri-o://11569cd009cbf79d170157a97f6048b3dc7a1902a13f0da7b0909f183b5fee3c" gracePeriod=30 Oct 14 09:09:23 crc kubenswrapper[4870]: I1014 09:09:23.872019 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0c1abbc6-c3c5-4fae-8200-fa3633518f50" containerName="ceilometer-notification-agent" containerID="cri-o://81dc195c974a1f500fd17010993c6de1b08465d9468744524f534488e538b64f" gracePeriod=30 Oct 14 09:09:24 crc kubenswrapper[4870]: I1014 09:09:24.135620 4870 generic.go:334] "Generic (PLEG): container finished" podID="0c1abbc6-c3c5-4fae-8200-fa3633518f50" containerID="fc6a3a9c4e15f823dafcbbfd5aa9442770632652275e8ec060b454c537254c5c" exitCode=0 Oct 14 09:09:24 crc kubenswrapper[4870]: I1014 09:09:24.135910 4870 generic.go:334] "Generic (PLEG): container finished" podID="0c1abbc6-c3c5-4fae-8200-fa3633518f50" containerID="11569cd009cbf79d170157a97f6048b3dc7a1902a13f0da7b0909f183b5fee3c" exitCode=2 Oct 14 09:09:24 crc kubenswrapper[4870]: I1014 09:09:24.135700 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0c1abbc6-c3c5-4fae-8200-fa3633518f50","Type":"ContainerDied","Data":"fc6a3a9c4e15f823dafcbbfd5aa9442770632652275e8ec060b454c537254c5c"} Oct 14 09:09:24 crc kubenswrapper[4870]: I1014 09:09:24.135947 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0c1abbc6-c3c5-4fae-8200-fa3633518f50","Type":"ContainerDied","Data":"11569cd009cbf79d170157a97f6048b3dc7a1902a13f0da7b0909f183b5fee3c"} Oct 14 09:09:25 crc kubenswrapper[4870]: I1014 09:09:25.146926 4870 generic.go:334] "Generic (PLEG): container finished" podID="0c1abbc6-c3c5-4fae-8200-fa3633518f50" containerID="6c66907b3ba7cfdfa4c257675d38768ee4dfd691a99ca1aebf71de86c862b364" exitCode=0 Oct 14 09:09:25 crc kubenswrapper[4870]: I1014 09:09:25.146996 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0c1abbc6-c3c5-4fae-8200-fa3633518f50","Type":"ContainerDied","Data":"6c66907b3ba7cfdfa4c257675d38768ee4dfd691a99ca1aebf71de86c862b364"} Oct 14 09:09:25 crc kubenswrapper[4870]: I1014 09:09:25.670078 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 14 09:09:25 crc kubenswrapper[4870]: I1014 09:09:25.809142 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0c1abbc6-c3c5-4fae-8200-fa3633518f50-run-httpd\") pod \"0c1abbc6-c3c5-4fae-8200-fa3633518f50\" (UID: \"0c1abbc6-c3c5-4fae-8200-fa3633518f50\") " Oct 14 09:09:25 crc kubenswrapper[4870]: I1014 09:09:25.809510 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c1abbc6-c3c5-4fae-8200-fa3633518f50-config-data\") pod \"0c1abbc6-c3c5-4fae-8200-fa3633518f50\" (UID: \"0c1abbc6-c3c5-4fae-8200-fa3633518f50\") " Oct 14 09:09:25 crc kubenswrapper[4870]: I1014 09:09:25.809602 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0c1abbc6-c3c5-4fae-8200-fa3633518f50-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "0c1abbc6-c3c5-4fae-8200-fa3633518f50" (UID: "0c1abbc6-c3c5-4fae-8200-fa3633518f50"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 09:09:25 crc kubenswrapper[4870]: I1014 09:09:25.809777 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0c1abbc6-c3c5-4fae-8200-fa3633518f50-sg-core-conf-yaml\") pod \"0c1abbc6-c3c5-4fae-8200-fa3633518f50\" (UID: \"0c1abbc6-c3c5-4fae-8200-fa3633518f50\") " Oct 14 09:09:25 crc kubenswrapper[4870]: I1014 09:09:25.809821 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c1abbc6-c3c5-4fae-8200-fa3633518f50-combined-ca-bundle\") pod \"0c1abbc6-c3c5-4fae-8200-fa3633518f50\" (UID: \"0c1abbc6-c3c5-4fae-8200-fa3633518f50\") " Oct 14 09:09:25 crc kubenswrapper[4870]: I1014 09:09:25.809885 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0c1abbc6-c3c5-4fae-8200-fa3633518f50-log-httpd\") pod \"0c1abbc6-c3c5-4fae-8200-fa3633518f50\" (UID: \"0c1abbc6-c3c5-4fae-8200-fa3633518f50\") " Oct 14 09:09:25 crc kubenswrapper[4870]: I1014 09:09:25.809965 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4bm8w\" (UniqueName: \"kubernetes.io/projected/0c1abbc6-c3c5-4fae-8200-fa3633518f50-kube-api-access-4bm8w\") pod \"0c1abbc6-c3c5-4fae-8200-fa3633518f50\" (UID: \"0c1abbc6-c3c5-4fae-8200-fa3633518f50\") " Oct 14 09:09:25 crc kubenswrapper[4870]: I1014 09:09:25.810368 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0c1abbc6-c3c5-4fae-8200-fa3633518f50-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "0c1abbc6-c3c5-4fae-8200-fa3633518f50" (UID: "0c1abbc6-c3c5-4fae-8200-fa3633518f50"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 09:09:25 crc kubenswrapper[4870]: I1014 09:09:25.810489 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0c1abbc6-c3c5-4fae-8200-fa3633518f50-scripts\") pod \"0c1abbc6-c3c5-4fae-8200-fa3633518f50\" (UID: \"0c1abbc6-c3c5-4fae-8200-fa3633518f50\") " Oct 14 09:09:25 crc kubenswrapper[4870]: I1014 09:09:25.811310 4870 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0c1abbc6-c3c5-4fae-8200-fa3633518f50-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 14 09:09:25 crc kubenswrapper[4870]: I1014 09:09:25.811334 4870 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0c1abbc6-c3c5-4fae-8200-fa3633518f50-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 14 09:09:25 crc kubenswrapper[4870]: I1014 09:09:25.820926 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0c1abbc6-c3c5-4fae-8200-fa3633518f50-kube-api-access-4bm8w" (OuterVolumeSpecName: "kube-api-access-4bm8w") pod "0c1abbc6-c3c5-4fae-8200-fa3633518f50" (UID: "0c1abbc6-c3c5-4fae-8200-fa3633518f50"). InnerVolumeSpecName "kube-api-access-4bm8w". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:09:25 crc kubenswrapper[4870]: I1014 09:09:25.821085 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c1abbc6-c3c5-4fae-8200-fa3633518f50-scripts" (OuterVolumeSpecName: "scripts") pod "0c1abbc6-c3c5-4fae-8200-fa3633518f50" (UID: "0c1abbc6-c3c5-4fae-8200-fa3633518f50"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:09:25 crc kubenswrapper[4870]: I1014 09:09:25.837433 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c1abbc6-c3c5-4fae-8200-fa3633518f50-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "0c1abbc6-c3c5-4fae-8200-fa3633518f50" (UID: "0c1abbc6-c3c5-4fae-8200-fa3633518f50"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:09:25 crc kubenswrapper[4870]: I1014 09:09:25.900746 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c1abbc6-c3c5-4fae-8200-fa3633518f50-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0c1abbc6-c3c5-4fae-8200-fa3633518f50" (UID: "0c1abbc6-c3c5-4fae-8200-fa3633518f50"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:09:25 crc kubenswrapper[4870]: I1014 09:09:25.913649 4870 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0c1abbc6-c3c5-4fae-8200-fa3633518f50-scripts\") on node \"crc\" DevicePath \"\"" Oct 14 09:09:25 crc kubenswrapper[4870]: I1014 09:09:25.914461 4870 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0c1abbc6-c3c5-4fae-8200-fa3633518f50-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 14 09:09:25 crc kubenswrapper[4870]: I1014 09:09:25.914570 4870 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c1abbc6-c3c5-4fae-8200-fa3633518f50-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 09:09:25 crc kubenswrapper[4870]: I1014 09:09:25.914654 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4bm8w\" (UniqueName: \"kubernetes.io/projected/0c1abbc6-c3c5-4fae-8200-fa3633518f50-kube-api-access-4bm8w\") on node \"crc\" DevicePath \"\"" Oct 14 09:09:25 crc kubenswrapper[4870]: I1014 09:09:25.919627 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c1abbc6-c3c5-4fae-8200-fa3633518f50-config-data" (OuterVolumeSpecName: "config-data") pod "0c1abbc6-c3c5-4fae-8200-fa3633518f50" (UID: "0c1abbc6-c3c5-4fae-8200-fa3633518f50"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:09:26 crc kubenswrapper[4870]: I1014 09:09:26.016314 4870 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c1abbc6-c3c5-4fae-8200-fa3633518f50-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 09:09:26 crc kubenswrapper[4870]: I1014 09:09:26.161179 4870 generic.go:334] "Generic (PLEG): container finished" podID="0c1abbc6-c3c5-4fae-8200-fa3633518f50" containerID="81dc195c974a1f500fd17010993c6de1b08465d9468744524f534488e538b64f" exitCode=0 Oct 14 09:09:26 crc kubenswrapper[4870]: I1014 09:09:26.161257 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 14 09:09:26 crc kubenswrapper[4870]: I1014 09:09:26.161252 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0c1abbc6-c3c5-4fae-8200-fa3633518f50","Type":"ContainerDied","Data":"81dc195c974a1f500fd17010993c6de1b08465d9468744524f534488e538b64f"} Oct 14 09:09:26 crc kubenswrapper[4870]: I1014 09:09:26.161385 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0c1abbc6-c3c5-4fae-8200-fa3633518f50","Type":"ContainerDied","Data":"e17032cae0cba82e58a163171c945ba5d0f8bae398d4814f6d06d61311984cb1"} Oct 14 09:09:26 crc kubenswrapper[4870]: I1014 09:09:26.161402 4870 scope.go:117] "RemoveContainer" containerID="fc6a3a9c4e15f823dafcbbfd5aa9442770632652275e8ec060b454c537254c5c" Oct 14 09:09:26 crc kubenswrapper[4870]: I1014 09:09:26.189933 4870 scope.go:117] "RemoveContainer" containerID="11569cd009cbf79d170157a97f6048b3dc7a1902a13f0da7b0909f183b5fee3c" Oct 14 09:09:26 crc kubenswrapper[4870]: I1014 09:09:26.264301 4870 scope.go:117] "RemoveContainer" containerID="81dc195c974a1f500fd17010993c6de1b08465d9468744524f534488e538b64f" Oct 14 09:09:26 crc kubenswrapper[4870]: I1014 09:09:26.287295 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 14 09:09:26 crc kubenswrapper[4870]: I1014 09:09:26.294411 4870 scope.go:117] "RemoveContainer" containerID="6c66907b3ba7cfdfa4c257675d38768ee4dfd691a99ca1aebf71de86c862b364" Oct 14 09:09:26 crc kubenswrapper[4870]: I1014 09:09:26.298664 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 14 09:09:26 crc kubenswrapper[4870]: I1014 09:09:26.317905 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 14 09:09:26 crc kubenswrapper[4870]: E1014 09:09:26.318305 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c1abbc6-c3c5-4fae-8200-fa3633518f50" containerName="ceilometer-central-agent" Oct 14 09:09:26 crc kubenswrapper[4870]: I1014 09:09:26.318322 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c1abbc6-c3c5-4fae-8200-fa3633518f50" containerName="ceilometer-central-agent" Oct 14 09:09:26 crc kubenswrapper[4870]: E1014 09:09:26.318341 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d666d204-b974-4c9a-aaba-6d02302fc332" containerName="init" Oct 14 09:09:26 crc kubenswrapper[4870]: I1014 09:09:26.318347 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="d666d204-b974-4c9a-aaba-6d02302fc332" containerName="init" Oct 14 09:09:26 crc kubenswrapper[4870]: E1014 09:09:26.318364 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c1abbc6-c3c5-4fae-8200-fa3633518f50" containerName="ceilometer-notification-agent" Oct 14 09:09:26 crc kubenswrapper[4870]: I1014 09:09:26.318371 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c1abbc6-c3c5-4fae-8200-fa3633518f50" containerName="ceilometer-notification-agent" Oct 14 09:09:26 crc kubenswrapper[4870]: E1014 09:09:26.318385 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d666d204-b974-4c9a-aaba-6d02302fc332" containerName="dnsmasq-dns" Oct 14 09:09:26 crc kubenswrapper[4870]: I1014 09:09:26.318391 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="d666d204-b974-4c9a-aaba-6d02302fc332" containerName="dnsmasq-dns" Oct 14 09:09:26 crc kubenswrapper[4870]: E1014 09:09:26.318403 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c1abbc6-c3c5-4fae-8200-fa3633518f50" containerName="sg-core" Oct 14 09:09:26 crc kubenswrapper[4870]: I1014 09:09:26.318409 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c1abbc6-c3c5-4fae-8200-fa3633518f50" containerName="sg-core" Oct 14 09:09:26 crc kubenswrapper[4870]: E1014 09:09:26.318427 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c1abbc6-c3c5-4fae-8200-fa3633518f50" containerName="proxy-httpd" Oct 14 09:09:26 crc kubenswrapper[4870]: I1014 09:09:26.318447 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c1abbc6-c3c5-4fae-8200-fa3633518f50" containerName="proxy-httpd" Oct 14 09:09:26 crc kubenswrapper[4870]: I1014 09:09:26.318629 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="d666d204-b974-4c9a-aaba-6d02302fc332" containerName="dnsmasq-dns" Oct 14 09:09:26 crc kubenswrapper[4870]: I1014 09:09:26.318645 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="0c1abbc6-c3c5-4fae-8200-fa3633518f50" containerName="sg-core" Oct 14 09:09:26 crc kubenswrapper[4870]: I1014 09:09:26.318669 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="0c1abbc6-c3c5-4fae-8200-fa3633518f50" containerName="proxy-httpd" Oct 14 09:09:26 crc kubenswrapper[4870]: I1014 09:09:26.318680 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="0c1abbc6-c3c5-4fae-8200-fa3633518f50" containerName="ceilometer-central-agent" Oct 14 09:09:26 crc kubenswrapper[4870]: I1014 09:09:26.318691 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="0c1abbc6-c3c5-4fae-8200-fa3633518f50" containerName="ceilometer-notification-agent" Oct 14 09:09:26 crc kubenswrapper[4870]: I1014 09:09:26.320718 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 14 09:09:26 crc kubenswrapper[4870]: I1014 09:09:26.322744 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a26548c9-166f-422a-ad41-24302cab4447-log-httpd\") pod \"ceilometer-0\" (UID: \"a26548c9-166f-422a-ad41-24302cab4447\") " pod="openstack/ceilometer-0" Oct 14 09:09:26 crc kubenswrapper[4870]: I1014 09:09:26.322781 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a26548c9-166f-422a-ad41-24302cab4447-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a26548c9-166f-422a-ad41-24302cab4447\") " pod="openstack/ceilometer-0" Oct 14 09:09:26 crc kubenswrapper[4870]: I1014 09:09:26.322810 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a26548c9-166f-422a-ad41-24302cab4447-config-data\") pod \"ceilometer-0\" (UID: \"a26548c9-166f-422a-ad41-24302cab4447\") " pod="openstack/ceilometer-0" Oct 14 09:09:26 crc kubenswrapper[4870]: I1014 09:09:26.322864 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wnvrw\" (UniqueName: \"kubernetes.io/projected/a26548c9-166f-422a-ad41-24302cab4447-kube-api-access-wnvrw\") pod \"ceilometer-0\" (UID: \"a26548c9-166f-422a-ad41-24302cab4447\") " pod="openstack/ceilometer-0" Oct 14 09:09:26 crc kubenswrapper[4870]: I1014 09:09:26.322980 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a26548c9-166f-422a-ad41-24302cab4447-scripts\") pod \"ceilometer-0\" (UID: \"a26548c9-166f-422a-ad41-24302cab4447\") " pod="openstack/ceilometer-0" Oct 14 09:09:26 crc kubenswrapper[4870]: I1014 09:09:26.323021 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a26548c9-166f-422a-ad41-24302cab4447-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a26548c9-166f-422a-ad41-24302cab4447\") " pod="openstack/ceilometer-0" Oct 14 09:09:26 crc kubenswrapper[4870]: I1014 09:09:26.323094 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a26548c9-166f-422a-ad41-24302cab4447-run-httpd\") pod \"ceilometer-0\" (UID: \"a26548c9-166f-422a-ad41-24302cab4447\") " pod="openstack/ceilometer-0" Oct 14 09:09:26 crc kubenswrapper[4870]: I1014 09:09:26.326753 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 14 09:09:26 crc kubenswrapper[4870]: I1014 09:09:26.327176 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 14 09:09:26 crc kubenswrapper[4870]: I1014 09:09:26.334638 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 14 09:09:26 crc kubenswrapper[4870]: I1014 09:09:26.336241 4870 scope.go:117] "RemoveContainer" containerID="fc6a3a9c4e15f823dafcbbfd5aa9442770632652275e8ec060b454c537254c5c" Oct 14 09:09:26 crc kubenswrapper[4870]: E1014 09:09:26.336684 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fc6a3a9c4e15f823dafcbbfd5aa9442770632652275e8ec060b454c537254c5c\": container with ID starting with fc6a3a9c4e15f823dafcbbfd5aa9442770632652275e8ec060b454c537254c5c not found: ID does not exist" containerID="fc6a3a9c4e15f823dafcbbfd5aa9442770632652275e8ec060b454c537254c5c" Oct 14 09:09:26 crc kubenswrapper[4870]: I1014 09:09:26.336725 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fc6a3a9c4e15f823dafcbbfd5aa9442770632652275e8ec060b454c537254c5c"} err="failed to get container status \"fc6a3a9c4e15f823dafcbbfd5aa9442770632652275e8ec060b454c537254c5c\": rpc error: code = NotFound desc = could not find container \"fc6a3a9c4e15f823dafcbbfd5aa9442770632652275e8ec060b454c537254c5c\": container with ID starting with fc6a3a9c4e15f823dafcbbfd5aa9442770632652275e8ec060b454c537254c5c not found: ID does not exist" Oct 14 09:09:26 crc kubenswrapper[4870]: I1014 09:09:26.336748 4870 scope.go:117] "RemoveContainer" containerID="11569cd009cbf79d170157a97f6048b3dc7a1902a13f0da7b0909f183b5fee3c" Oct 14 09:09:26 crc kubenswrapper[4870]: E1014 09:09:26.337146 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"11569cd009cbf79d170157a97f6048b3dc7a1902a13f0da7b0909f183b5fee3c\": container with ID starting with 11569cd009cbf79d170157a97f6048b3dc7a1902a13f0da7b0909f183b5fee3c not found: ID does not exist" containerID="11569cd009cbf79d170157a97f6048b3dc7a1902a13f0da7b0909f183b5fee3c" Oct 14 09:09:26 crc kubenswrapper[4870]: I1014 09:09:26.337168 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"11569cd009cbf79d170157a97f6048b3dc7a1902a13f0da7b0909f183b5fee3c"} err="failed to get container status \"11569cd009cbf79d170157a97f6048b3dc7a1902a13f0da7b0909f183b5fee3c\": rpc error: code = NotFound desc = could not find container \"11569cd009cbf79d170157a97f6048b3dc7a1902a13f0da7b0909f183b5fee3c\": container with ID starting with 11569cd009cbf79d170157a97f6048b3dc7a1902a13f0da7b0909f183b5fee3c not found: ID does not exist" Oct 14 09:09:26 crc kubenswrapper[4870]: I1014 09:09:26.337181 4870 scope.go:117] "RemoveContainer" containerID="81dc195c974a1f500fd17010993c6de1b08465d9468744524f534488e538b64f" Oct 14 09:09:26 crc kubenswrapper[4870]: E1014 09:09:26.337386 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"81dc195c974a1f500fd17010993c6de1b08465d9468744524f534488e538b64f\": container with ID starting with 81dc195c974a1f500fd17010993c6de1b08465d9468744524f534488e538b64f not found: ID does not exist" containerID="81dc195c974a1f500fd17010993c6de1b08465d9468744524f534488e538b64f" Oct 14 09:09:26 crc kubenswrapper[4870]: I1014 09:09:26.337406 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"81dc195c974a1f500fd17010993c6de1b08465d9468744524f534488e538b64f"} err="failed to get container status \"81dc195c974a1f500fd17010993c6de1b08465d9468744524f534488e538b64f\": rpc error: code = NotFound desc = could not find container \"81dc195c974a1f500fd17010993c6de1b08465d9468744524f534488e538b64f\": container with ID starting with 81dc195c974a1f500fd17010993c6de1b08465d9468744524f534488e538b64f not found: ID does not exist" Oct 14 09:09:26 crc kubenswrapper[4870]: I1014 09:09:26.337419 4870 scope.go:117] "RemoveContainer" containerID="6c66907b3ba7cfdfa4c257675d38768ee4dfd691a99ca1aebf71de86c862b364" Oct 14 09:09:26 crc kubenswrapper[4870]: E1014 09:09:26.337739 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6c66907b3ba7cfdfa4c257675d38768ee4dfd691a99ca1aebf71de86c862b364\": container with ID starting with 6c66907b3ba7cfdfa4c257675d38768ee4dfd691a99ca1aebf71de86c862b364 not found: ID does not exist" containerID="6c66907b3ba7cfdfa4c257675d38768ee4dfd691a99ca1aebf71de86c862b364" Oct 14 09:09:26 crc kubenswrapper[4870]: I1014 09:09:26.337763 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6c66907b3ba7cfdfa4c257675d38768ee4dfd691a99ca1aebf71de86c862b364"} err="failed to get container status \"6c66907b3ba7cfdfa4c257675d38768ee4dfd691a99ca1aebf71de86c862b364\": rpc error: code = NotFound desc = could not find container \"6c66907b3ba7cfdfa4c257675d38768ee4dfd691a99ca1aebf71de86c862b364\": container with ID starting with 6c66907b3ba7cfdfa4c257675d38768ee4dfd691a99ca1aebf71de86c862b364 not found: ID does not exist" Oct 14 09:09:26 crc kubenswrapper[4870]: I1014 09:09:26.424063 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a26548c9-166f-422a-ad41-24302cab4447-run-httpd\") pod \"ceilometer-0\" (UID: \"a26548c9-166f-422a-ad41-24302cab4447\") " pod="openstack/ceilometer-0" Oct 14 09:09:26 crc kubenswrapper[4870]: I1014 09:09:26.424160 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a26548c9-166f-422a-ad41-24302cab4447-log-httpd\") pod \"ceilometer-0\" (UID: \"a26548c9-166f-422a-ad41-24302cab4447\") " pod="openstack/ceilometer-0" Oct 14 09:09:26 crc kubenswrapper[4870]: I1014 09:09:26.424930 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a26548c9-166f-422a-ad41-24302cab4447-run-httpd\") pod \"ceilometer-0\" (UID: \"a26548c9-166f-422a-ad41-24302cab4447\") " pod="openstack/ceilometer-0" Oct 14 09:09:26 crc kubenswrapper[4870]: I1014 09:09:26.425012 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a26548c9-166f-422a-ad41-24302cab4447-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a26548c9-166f-422a-ad41-24302cab4447\") " pod="openstack/ceilometer-0" Oct 14 09:09:26 crc kubenswrapper[4870]: I1014 09:09:26.425039 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a26548c9-166f-422a-ad41-24302cab4447-config-data\") pod \"ceilometer-0\" (UID: \"a26548c9-166f-422a-ad41-24302cab4447\") " pod="openstack/ceilometer-0" Oct 14 09:09:26 crc kubenswrapper[4870]: I1014 09:09:26.425087 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wnvrw\" (UniqueName: \"kubernetes.io/projected/a26548c9-166f-422a-ad41-24302cab4447-kube-api-access-wnvrw\") pod \"ceilometer-0\" (UID: \"a26548c9-166f-422a-ad41-24302cab4447\") " pod="openstack/ceilometer-0" Oct 14 09:09:26 crc kubenswrapper[4870]: I1014 09:09:26.425156 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a26548c9-166f-422a-ad41-24302cab4447-scripts\") pod \"ceilometer-0\" (UID: \"a26548c9-166f-422a-ad41-24302cab4447\") " pod="openstack/ceilometer-0" Oct 14 09:09:26 crc kubenswrapper[4870]: I1014 09:09:26.425186 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a26548c9-166f-422a-ad41-24302cab4447-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a26548c9-166f-422a-ad41-24302cab4447\") " pod="openstack/ceilometer-0" Oct 14 09:09:26 crc kubenswrapper[4870]: I1014 09:09:26.426085 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a26548c9-166f-422a-ad41-24302cab4447-log-httpd\") pod \"ceilometer-0\" (UID: \"a26548c9-166f-422a-ad41-24302cab4447\") " pod="openstack/ceilometer-0" Oct 14 09:09:26 crc kubenswrapper[4870]: I1014 09:09:26.429034 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a26548c9-166f-422a-ad41-24302cab4447-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a26548c9-166f-422a-ad41-24302cab4447\") " pod="openstack/ceilometer-0" Oct 14 09:09:26 crc kubenswrapper[4870]: I1014 09:09:26.429248 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a26548c9-166f-422a-ad41-24302cab4447-scripts\") pod \"ceilometer-0\" (UID: \"a26548c9-166f-422a-ad41-24302cab4447\") " pod="openstack/ceilometer-0" Oct 14 09:09:26 crc kubenswrapper[4870]: I1014 09:09:26.429749 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a26548c9-166f-422a-ad41-24302cab4447-config-data\") pod \"ceilometer-0\" (UID: \"a26548c9-166f-422a-ad41-24302cab4447\") " pod="openstack/ceilometer-0" Oct 14 09:09:26 crc kubenswrapper[4870]: I1014 09:09:26.430149 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a26548c9-166f-422a-ad41-24302cab4447-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a26548c9-166f-422a-ad41-24302cab4447\") " pod="openstack/ceilometer-0" Oct 14 09:09:26 crc kubenswrapper[4870]: I1014 09:09:26.443036 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wnvrw\" (UniqueName: \"kubernetes.io/projected/a26548c9-166f-422a-ad41-24302cab4447-kube-api-access-wnvrw\") pod \"ceilometer-0\" (UID: \"a26548c9-166f-422a-ad41-24302cab4447\") " pod="openstack/ceilometer-0" Oct 14 09:09:26 crc kubenswrapper[4870]: I1014 09:09:26.657272 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 14 09:09:27 crc kubenswrapper[4870]: I1014 09:09:27.046510 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0c1abbc6-c3c5-4fae-8200-fa3633518f50" path="/var/lib/kubelet/pods/0c1abbc6-c3c5-4fae-8200-fa3633518f50/volumes" Oct 14 09:09:27 crc kubenswrapper[4870]: I1014 09:09:27.123407 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 14 09:09:27 crc kubenswrapper[4870]: W1014 09:09:27.131701 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda26548c9_166f_422a_ad41_24302cab4447.slice/crio-b11c18039b303eb585d9f57ae58534821ef7783b0d1f7a715b99c7915bc9d0aa WatchSource:0}: Error finding container b11c18039b303eb585d9f57ae58534821ef7783b0d1f7a715b99c7915bc9d0aa: Status 404 returned error can't find the container with id b11c18039b303eb585d9f57ae58534821ef7783b0d1f7a715b99c7915bc9d0aa Oct 14 09:09:27 crc kubenswrapper[4870]: I1014 09:09:27.133805 4870 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 14 09:09:27 crc kubenswrapper[4870]: I1014 09:09:27.179025 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a26548c9-166f-422a-ad41-24302cab4447","Type":"ContainerStarted","Data":"b11c18039b303eb585d9f57ae58534821ef7783b0d1f7a715b99c7915bc9d0aa"} Oct 14 09:09:28 crc kubenswrapper[4870]: I1014 09:09:28.200017 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a26548c9-166f-422a-ad41-24302cab4447","Type":"ContainerStarted","Data":"28fca38e6f1429faafc349a8a8fda92d5eacb9561cda24827441e20d9b7f5ac5"} Oct 14 09:09:28 crc kubenswrapper[4870]: I1014 09:09:28.200325 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a26548c9-166f-422a-ad41-24302cab4447","Type":"ContainerStarted","Data":"d9028a2d484522ec3bf176d1899c5e5600f5a49b56ee56ca991955467345f2b2"} Oct 14 09:09:29 crc kubenswrapper[4870]: I1014 09:09:29.212119 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a26548c9-166f-422a-ad41-24302cab4447","Type":"ContainerStarted","Data":"583663dc7f70a20c74d967da950eca45ef4a121a9fbfc836f32ba0cf4ae24711"} Oct 14 09:09:30 crc kubenswrapper[4870]: I1014 09:09:30.127374 4870 scope.go:117] "RemoveContainer" containerID="4e79e333d58ceed9e942c90eccb0fcc1d37daa5266b237942c698e6f1e9bb5b1" Oct 14 09:09:30 crc kubenswrapper[4870]: I1014 09:09:30.169417 4870 scope.go:117] "RemoveContainer" containerID="c44545800af6e2b89c3a53d3ede0edff2502d2bed187c89bf6e629d76d11e590" Oct 14 09:09:30 crc kubenswrapper[4870]: I1014 09:09:30.222544 4870 scope.go:117] "RemoveContainer" containerID="fc13475221487a46b31d10b88df54efde5603c4ed03902033b2e560f96f5278c" Oct 14 09:09:30 crc kubenswrapper[4870]: I1014 09:09:30.280576 4870 scope.go:117] "RemoveContainer" containerID="12da0dee2ef976c06be85dab716622747e95edc7d4700015b5f01575ba23bd3e" Oct 14 09:09:30 crc kubenswrapper[4870]: I1014 09:09:30.369498 4870 scope.go:117] "RemoveContainer" containerID="0e314a6abc2dab8dca15f3a5e845c1abbd2ca7e8beb7c42bd2a1e0e2ee0f0790" Oct 14 09:09:30 crc kubenswrapper[4870]: I1014 09:09:30.405276 4870 scope.go:117] "RemoveContainer" containerID="e5c32fb72855303655e866115cac1797680ed1edbe001700b324e72837a8647f" Oct 14 09:09:31 crc kubenswrapper[4870]: I1014 09:09:31.038676 4870 scope.go:117] "RemoveContainer" containerID="8fcaeb3090f0c89cba1e66db524b4a246072daa35592012664aca450fa09ddc9" Oct 14 09:09:31 crc kubenswrapper[4870]: E1014 09:09:31.039089 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 09:09:32 crc kubenswrapper[4870]: I1014 09:09:32.050672 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-kxsd4"] Oct 14 09:09:32 crc kubenswrapper[4870]: I1014 09:09:32.062641 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-kxsd4"] Oct 14 09:09:32 crc kubenswrapper[4870]: I1014 09:09:32.167911 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/manila-api-0" Oct 14 09:09:32 crc kubenswrapper[4870]: I1014 09:09:32.266283 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/manila-scheduler-0" Oct 14 09:09:32 crc kubenswrapper[4870]: I1014 09:09:32.268522 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/manila-share-share1-0" Oct 14 09:09:32 crc kubenswrapper[4870]: I1014 09:09:32.269583 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a26548c9-166f-422a-ad41-24302cab4447","Type":"ContainerStarted","Data":"ac08d22099bb072e240c72046f4c9cb59178af000c8208ab4a1e7c7deb68fd6a"} Oct 14 09:09:32 crc kubenswrapper[4870]: I1014 09:09:32.271236 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 14 09:09:32 crc kubenswrapper[4870]: I1014 09:09:32.360057 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.477645231 podStartE2EDuration="6.360036275s" podCreationTimestamp="2025-10-14 09:09:26 +0000 UTC" firstStartedPulling="2025-10-14 09:09:27.133488331 +0000 UTC m=+7702.830848702" lastFinishedPulling="2025-10-14 09:09:31.015879375 +0000 UTC m=+7706.713239746" observedRunningTime="2025-10-14 09:09:32.345033465 +0000 UTC m=+7708.042393836" watchObservedRunningTime="2025-10-14 09:09:32.360036275 +0000 UTC m=+7708.057396646" Oct 14 09:09:33 crc kubenswrapper[4870]: I1014 09:09:33.045940 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57647978-3492-4dcf-85ae-ad3109e97980" path="/var/lib/kubelet/pods/57647978-3492-4dcf-85ae-ad3109e97980/volumes" Oct 14 09:09:44 crc kubenswrapper[4870]: I1014 09:09:44.035845 4870 scope.go:117] "RemoveContainer" containerID="8fcaeb3090f0c89cba1e66db524b4a246072daa35592012664aca450fa09ddc9" Oct 14 09:09:44 crc kubenswrapper[4870]: E1014 09:09:44.036733 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 09:09:47 crc kubenswrapper[4870]: I1014 09:09:47.012524 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-qzpjh"] Oct 14 09:09:47 crc kubenswrapper[4870]: I1014 09:09:47.017193 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-qzpjh" Oct 14 09:09:47 crc kubenswrapper[4870]: I1014 09:09:47.068209 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-qzpjh"] Oct 14 09:09:47 crc kubenswrapper[4870]: I1014 09:09:47.118056 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c898dca6-0a58-4692-8bd0-d4bcde9094b5-catalog-content\") pod \"community-operators-qzpjh\" (UID: \"c898dca6-0a58-4692-8bd0-d4bcde9094b5\") " pod="openshift-marketplace/community-operators-qzpjh" Oct 14 09:09:47 crc kubenswrapper[4870]: I1014 09:09:47.118157 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fn84x\" (UniqueName: \"kubernetes.io/projected/c898dca6-0a58-4692-8bd0-d4bcde9094b5-kube-api-access-fn84x\") pod \"community-operators-qzpjh\" (UID: \"c898dca6-0a58-4692-8bd0-d4bcde9094b5\") " pod="openshift-marketplace/community-operators-qzpjh" Oct 14 09:09:47 crc kubenswrapper[4870]: I1014 09:09:47.118276 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c898dca6-0a58-4692-8bd0-d4bcde9094b5-utilities\") pod \"community-operators-qzpjh\" (UID: \"c898dca6-0a58-4692-8bd0-d4bcde9094b5\") " pod="openshift-marketplace/community-operators-qzpjh" Oct 14 09:09:47 crc kubenswrapper[4870]: I1014 09:09:47.221575 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c898dca6-0a58-4692-8bd0-d4bcde9094b5-catalog-content\") pod \"community-operators-qzpjh\" (UID: \"c898dca6-0a58-4692-8bd0-d4bcde9094b5\") " pod="openshift-marketplace/community-operators-qzpjh" Oct 14 09:09:47 crc kubenswrapper[4870]: I1014 09:09:47.222366 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fn84x\" (UniqueName: \"kubernetes.io/projected/c898dca6-0a58-4692-8bd0-d4bcde9094b5-kube-api-access-fn84x\") pod \"community-operators-qzpjh\" (UID: \"c898dca6-0a58-4692-8bd0-d4bcde9094b5\") " pod="openshift-marketplace/community-operators-qzpjh" Oct 14 09:09:47 crc kubenswrapper[4870]: I1014 09:09:47.222399 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c898dca6-0a58-4692-8bd0-d4bcde9094b5-utilities\") pod \"community-operators-qzpjh\" (UID: \"c898dca6-0a58-4692-8bd0-d4bcde9094b5\") " pod="openshift-marketplace/community-operators-qzpjh" Oct 14 09:09:47 crc kubenswrapper[4870]: I1014 09:09:47.222827 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c898dca6-0a58-4692-8bd0-d4bcde9094b5-utilities\") pod \"community-operators-qzpjh\" (UID: \"c898dca6-0a58-4692-8bd0-d4bcde9094b5\") " pod="openshift-marketplace/community-operators-qzpjh" Oct 14 09:09:47 crc kubenswrapper[4870]: I1014 09:09:47.222291 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c898dca6-0a58-4692-8bd0-d4bcde9094b5-catalog-content\") pod \"community-operators-qzpjh\" (UID: \"c898dca6-0a58-4692-8bd0-d4bcde9094b5\") " pod="openshift-marketplace/community-operators-qzpjh" Oct 14 09:09:47 crc kubenswrapper[4870]: I1014 09:09:47.260785 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fn84x\" (UniqueName: \"kubernetes.io/projected/c898dca6-0a58-4692-8bd0-d4bcde9094b5-kube-api-access-fn84x\") pod \"community-operators-qzpjh\" (UID: \"c898dca6-0a58-4692-8bd0-d4bcde9094b5\") " pod="openshift-marketplace/community-operators-qzpjh" Oct 14 09:09:47 crc kubenswrapper[4870]: I1014 09:09:47.366243 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-qzpjh" Oct 14 09:09:48 crc kubenswrapper[4870]: I1014 09:09:48.014014 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-qzpjh"] Oct 14 09:09:48 crc kubenswrapper[4870]: I1014 09:09:48.452297 4870 generic.go:334] "Generic (PLEG): container finished" podID="c898dca6-0a58-4692-8bd0-d4bcde9094b5" containerID="3eb2e960069690e10e272a4ad3cb58a13f459bc7a925b221eaaf5cc5402b16cc" exitCode=0 Oct 14 09:09:48 crc kubenswrapper[4870]: I1014 09:09:48.452407 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qzpjh" event={"ID":"c898dca6-0a58-4692-8bd0-d4bcde9094b5","Type":"ContainerDied","Data":"3eb2e960069690e10e272a4ad3cb58a13f459bc7a925b221eaaf5cc5402b16cc"} Oct 14 09:09:48 crc kubenswrapper[4870]: I1014 09:09:48.452749 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qzpjh" event={"ID":"c898dca6-0a58-4692-8bd0-d4bcde9094b5","Type":"ContainerStarted","Data":"756a47f16f5a8ce25a2e1f3aa772ca8f7015d2043c9261a3a62148b5f63dc14a"} Oct 14 09:09:49 crc kubenswrapper[4870]: I1014 09:09:49.465136 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qzpjh" event={"ID":"c898dca6-0a58-4692-8bd0-d4bcde9094b5","Type":"ContainerStarted","Data":"c25f3e0b5735cd8fe5f810861bc935fdad4ed0afffdeb5c9fdcf49bdfdb26ef1"} Oct 14 09:09:51 crc kubenswrapper[4870]: I1014 09:09:51.046544 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-mvlrf"] Oct 14 09:09:51 crc kubenswrapper[4870]: I1014 09:09:51.046903 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-mvlrf"] Oct 14 09:09:51 crc kubenswrapper[4870]: I1014 09:09:51.488255 4870 generic.go:334] "Generic (PLEG): container finished" podID="c898dca6-0a58-4692-8bd0-d4bcde9094b5" containerID="c25f3e0b5735cd8fe5f810861bc935fdad4ed0afffdeb5c9fdcf49bdfdb26ef1" exitCode=0 Oct 14 09:09:51 crc kubenswrapper[4870]: I1014 09:09:51.488301 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qzpjh" event={"ID":"c898dca6-0a58-4692-8bd0-d4bcde9094b5","Type":"ContainerDied","Data":"c25f3e0b5735cd8fe5f810861bc935fdad4ed0afffdeb5c9fdcf49bdfdb26ef1"} Oct 14 09:09:52 crc kubenswrapper[4870]: I1014 09:09:52.046124 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-pl2ff"] Oct 14 09:09:52 crc kubenswrapper[4870]: I1014 09:09:52.055911 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-pl2ff"] Oct 14 09:09:52 crc kubenswrapper[4870]: I1014 09:09:52.497864 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qzpjh" event={"ID":"c898dca6-0a58-4692-8bd0-d4bcde9094b5","Type":"ContainerStarted","Data":"e1978992aacb702163aa46306220b1b09611ef89b8b854b479533cad3254a737"} Oct 14 09:09:52 crc kubenswrapper[4870]: I1014 09:09:52.525942 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-qzpjh" podStartSLOduration=2.909420001 podStartE2EDuration="6.52592295s" podCreationTimestamp="2025-10-14 09:09:46 +0000 UTC" firstStartedPulling="2025-10-14 09:09:48.454369555 +0000 UTC m=+7724.151729966" lastFinishedPulling="2025-10-14 09:09:52.070872534 +0000 UTC m=+7727.768232915" observedRunningTime="2025-10-14 09:09:52.517731288 +0000 UTC m=+7728.215091659" watchObservedRunningTime="2025-10-14 09:09:52.52592295 +0000 UTC m=+7728.223283321" Oct 14 09:09:53 crc kubenswrapper[4870]: I1014 09:09:53.058757 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c0c2d021-038b-4127-bceb-bbe26c1f3070" path="/var/lib/kubelet/pods/c0c2d021-038b-4127-bceb-bbe26c1f3070/volumes" Oct 14 09:09:53 crc kubenswrapper[4870]: I1014 09:09:53.060661 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cbd7ca07-b13e-421e-8e09-1b96dc55132a" path="/var/lib/kubelet/pods/cbd7ca07-b13e-421e-8e09-1b96dc55132a/volumes" Oct 14 09:09:55 crc kubenswrapper[4870]: I1014 09:09:55.044170 4870 scope.go:117] "RemoveContainer" containerID="8fcaeb3090f0c89cba1e66db524b4a246072daa35592012664aca450fa09ddc9" Oct 14 09:09:55 crc kubenswrapper[4870]: E1014 09:09:55.045003 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 09:09:56 crc kubenswrapper[4870]: I1014 09:09:56.668694 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Oct 14 09:09:57 crc kubenswrapper[4870]: I1014 09:09:57.367475 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-qzpjh" Oct 14 09:09:57 crc kubenswrapper[4870]: I1014 09:09:57.367534 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-qzpjh" Oct 14 09:09:57 crc kubenswrapper[4870]: I1014 09:09:57.447039 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-qzpjh" Oct 14 09:09:57 crc kubenswrapper[4870]: I1014 09:09:57.619347 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-qzpjh" Oct 14 09:09:57 crc kubenswrapper[4870]: I1014 09:09:57.698164 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-qzpjh"] Oct 14 09:09:59 crc kubenswrapper[4870]: I1014 09:09:59.587286 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-qzpjh" podUID="c898dca6-0a58-4692-8bd0-d4bcde9094b5" containerName="registry-server" containerID="cri-o://e1978992aacb702163aa46306220b1b09611ef89b8b854b479533cad3254a737" gracePeriod=2 Oct 14 09:10:00 crc kubenswrapper[4870]: I1014 09:10:00.075112 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-qzpjh" Oct 14 09:10:00 crc kubenswrapper[4870]: I1014 09:10:00.156270 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c898dca6-0a58-4692-8bd0-d4bcde9094b5-utilities\") pod \"c898dca6-0a58-4692-8bd0-d4bcde9094b5\" (UID: \"c898dca6-0a58-4692-8bd0-d4bcde9094b5\") " Oct 14 09:10:00 crc kubenswrapper[4870]: I1014 09:10:00.156612 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c898dca6-0a58-4692-8bd0-d4bcde9094b5-catalog-content\") pod \"c898dca6-0a58-4692-8bd0-d4bcde9094b5\" (UID: \"c898dca6-0a58-4692-8bd0-d4bcde9094b5\") " Oct 14 09:10:00 crc kubenswrapper[4870]: I1014 09:10:00.156790 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fn84x\" (UniqueName: \"kubernetes.io/projected/c898dca6-0a58-4692-8bd0-d4bcde9094b5-kube-api-access-fn84x\") pod \"c898dca6-0a58-4692-8bd0-d4bcde9094b5\" (UID: \"c898dca6-0a58-4692-8bd0-d4bcde9094b5\") " Oct 14 09:10:00 crc kubenswrapper[4870]: I1014 09:10:00.157331 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c898dca6-0a58-4692-8bd0-d4bcde9094b5-utilities" (OuterVolumeSpecName: "utilities") pod "c898dca6-0a58-4692-8bd0-d4bcde9094b5" (UID: "c898dca6-0a58-4692-8bd0-d4bcde9094b5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 09:10:00 crc kubenswrapper[4870]: I1014 09:10:00.157829 4870 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c898dca6-0a58-4692-8bd0-d4bcde9094b5-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 09:10:00 crc kubenswrapper[4870]: I1014 09:10:00.163163 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c898dca6-0a58-4692-8bd0-d4bcde9094b5-kube-api-access-fn84x" (OuterVolumeSpecName: "kube-api-access-fn84x") pod "c898dca6-0a58-4692-8bd0-d4bcde9094b5" (UID: "c898dca6-0a58-4692-8bd0-d4bcde9094b5"). InnerVolumeSpecName "kube-api-access-fn84x". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:10:00 crc kubenswrapper[4870]: I1014 09:10:00.247408 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c898dca6-0a58-4692-8bd0-d4bcde9094b5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c898dca6-0a58-4692-8bd0-d4bcde9094b5" (UID: "c898dca6-0a58-4692-8bd0-d4bcde9094b5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 09:10:00 crc kubenswrapper[4870]: I1014 09:10:00.259306 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fn84x\" (UniqueName: \"kubernetes.io/projected/c898dca6-0a58-4692-8bd0-d4bcde9094b5-kube-api-access-fn84x\") on node \"crc\" DevicePath \"\"" Oct 14 09:10:00 crc kubenswrapper[4870]: I1014 09:10:00.259359 4870 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c898dca6-0a58-4692-8bd0-d4bcde9094b5-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 09:10:00 crc kubenswrapper[4870]: I1014 09:10:00.605307 4870 generic.go:334] "Generic (PLEG): container finished" podID="c898dca6-0a58-4692-8bd0-d4bcde9094b5" containerID="e1978992aacb702163aa46306220b1b09611ef89b8b854b479533cad3254a737" exitCode=0 Oct 14 09:10:00 crc kubenswrapper[4870]: I1014 09:10:00.605492 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qzpjh" event={"ID":"c898dca6-0a58-4692-8bd0-d4bcde9094b5","Type":"ContainerDied","Data":"e1978992aacb702163aa46306220b1b09611ef89b8b854b479533cad3254a737"} Oct 14 09:10:00 crc kubenswrapper[4870]: I1014 09:10:00.605589 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qzpjh" event={"ID":"c898dca6-0a58-4692-8bd0-d4bcde9094b5","Type":"ContainerDied","Data":"756a47f16f5a8ce25a2e1f3aa772ca8f7015d2043c9261a3a62148b5f63dc14a"} Oct 14 09:10:00 crc kubenswrapper[4870]: I1014 09:10:00.605591 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-qzpjh" Oct 14 09:10:00 crc kubenswrapper[4870]: I1014 09:10:00.605620 4870 scope.go:117] "RemoveContainer" containerID="e1978992aacb702163aa46306220b1b09611ef89b8b854b479533cad3254a737" Oct 14 09:10:00 crc kubenswrapper[4870]: I1014 09:10:00.654783 4870 scope.go:117] "RemoveContainer" containerID="c25f3e0b5735cd8fe5f810861bc935fdad4ed0afffdeb5c9fdcf49bdfdb26ef1" Oct 14 09:10:00 crc kubenswrapper[4870]: I1014 09:10:00.674638 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-qzpjh"] Oct 14 09:10:00 crc kubenswrapper[4870]: I1014 09:10:00.689253 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-qzpjh"] Oct 14 09:10:00 crc kubenswrapper[4870]: I1014 09:10:00.702365 4870 scope.go:117] "RemoveContainer" containerID="3eb2e960069690e10e272a4ad3cb58a13f459bc7a925b221eaaf5cc5402b16cc" Oct 14 09:10:00 crc kubenswrapper[4870]: I1014 09:10:00.778652 4870 scope.go:117] "RemoveContainer" containerID="e1978992aacb702163aa46306220b1b09611ef89b8b854b479533cad3254a737" Oct 14 09:10:00 crc kubenswrapper[4870]: E1014 09:10:00.779949 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e1978992aacb702163aa46306220b1b09611ef89b8b854b479533cad3254a737\": container with ID starting with e1978992aacb702163aa46306220b1b09611ef89b8b854b479533cad3254a737 not found: ID does not exist" containerID="e1978992aacb702163aa46306220b1b09611ef89b8b854b479533cad3254a737" Oct 14 09:10:00 crc kubenswrapper[4870]: I1014 09:10:00.780004 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e1978992aacb702163aa46306220b1b09611ef89b8b854b479533cad3254a737"} err="failed to get container status \"e1978992aacb702163aa46306220b1b09611ef89b8b854b479533cad3254a737\": rpc error: code = NotFound desc = could not find container \"e1978992aacb702163aa46306220b1b09611ef89b8b854b479533cad3254a737\": container with ID starting with e1978992aacb702163aa46306220b1b09611ef89b8b854b479533cad3254a737 not found: ID does not exist" Oct 14 09:10:00 crc kubenswrapper[4870]: I1014 09:10:00.780038 4870 scope.go:117] "RemoveContainer" containerID="c25f3e0b5735cd8fe5f810861bc935fdad4ed0afffdeb5c9fdcf49bdfdb26ef1" Oct 14 09:10:00 crc kubenswrapper[4870]: E1014 09:10:00.786160 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c25f3e0b5735cd8fe5f810861bc935fdad4ed0afffdeb5c9fdcf49bdfdb26ef1\": container with ID starting with c25f3e0b5735cd8fe5f810861bc935fdad4ed0afffdeb5c9fdcf49bdfdb26ef1 not found: ID does not exist" containerID="c25f3e0b5735cd8fe5f810861bc935fdad4ed0afffdeb5c9fdcf49bdfdb26ef1" Oct 14 09:10:00 crc kubenswrapper[4870]: I1014 09:10:00.786213 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c25f3e0b5735cd8fe5f810861bc935fdad4ed0afffdeb5c9fdcf49bdfdb26ef1"} err="failed to get container status \"c25f3e0b5735cd8fe5f810861bc935fdad4ed0afffdeb5c9fdcf49bdfdb26ef1\": rpc error: code = NotFound desc = could not find container \"c25f3e0b5735cd8fe5f810861bc935fdad4ed0afffdeb5c9fdcf49bdfdb26ef1\": container with ID starting with c25f3e0b5735cd8fe5f810861bc935fdad4ed0afffdeb5c9fdcf49bdfdb26ef1 not found: ID does not exist" Oct 14 09:10:00 crc kubenswrapper[4870]: I1014 09:10:00.786279 4870 scope.go:117] "RemoveContainer" containerID="3eb2e960069690e10e272a4ad3cb58a13f459bc7a925b221eaaf5cc5402b16cc" Oct 14 09:10:00 crc kubenswrapper[4870]: E1014 09:10:00.786865 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3eb2e960069690e10e272a4ad3cb58a13f459bc7a925b221eaaf5cc5402b16cc\": container with ID starting with 3eb2e960069690e10e272a4ad3cb58a13f459bc7a925b221eaaf5cc5402b16cc not found: ID does not exist" containerID="3eb2e960069690e10e272a4ad3cb58a13f459bc7a925b221eaaf5cc5402b16cc" Oct 14 09:10:00 crc kubenswrapper[4870]: I1014 09:10:00.787161 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3eb2e960069690e10e272a4ad3cb58a13f459bc7a925b221eaaf5cc5402b16cc"} err="failed to get container status \"3eb2e960069690e10e272a4ad3cb58a13f459bc7a925b221eaaf5cc5402b16cc\": rpc error: code = NotFound desc = could not find container \"3eb2e960069690e10e272a4ad3cb58a13f459bc7a925b221eaaf5cc5402b16cc\": container with ID starting with 3eb2e960069690e10e272a4ad3cb58a13f459bc7a925b221eaaf5cc5402b16cc not found: ID does not exist" Oct 14 09:10:01 crc kubenswrapper[4870]: I1014 09:10:01.054390 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c898dca6-0a58-4692-8bd0-d4bcde9094b5" path="/var/lib/kubelet/pods/c898dca6-0a58-4692-8bd0-d4bcde9094b5/volumes" Oct 14 09:10:08 crc kubenswrapper[4870]: I1014 09:10:08.033823 4870 scope.go:117] "RemoveContainer" containerID="8fcaeb3090f0c89cba1e66db524b4a246072daa35592012664aca450fa09ddc9" Oct 14 09:10:08 crc kubenswrapper[4870]: E1014 09:10:08.034860 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 09:10:12 crc kubenswrapper[4870]: I1014 09:10:12.061247 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-fbdnv"] Oct 14 09:10:12 crc kubenswrapper[4870]: I1014 09:10:12.071788 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-fbdnv"] Oct 14 09:10:13 crc kubenswrapper[4870]: I1014 09:10:13.069204 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2447a450-aee2-4e46-9bac-fa754b69ecaf" path="/var/lib/kubelet/pods/2447a450-aee2-4e46-9bac-fa754b69ecaf/volumes" Oct 14 09:10:16 crc kubenswrapper[4870]: I1014 09:10:16.253050 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5bd5f99ff7-r6rlk"] Oct 14 09:10:16 crc kubenswrapper[4870]: E1014 09:10:16.253779 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c898dca6-0a58-4692-8bd0-d4bcde9094b5" containerName="extract-content" Oct 14 09:10:16 crc kubenswrapper[4870]: I1014 09:10:16.253795 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="c898dca6-0a58-4692-8bd0-d4bcde9094b5" containerName="extract-content" Oct 14 09:10:16 crc kubenswrapper[4870]: E1014 09:10:16.253822 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c898dca6-0a58-4692-8bd0-d4bcde9094b5" containerName="registry-server" Oct 14 09:10:16 crc kubenswrapper[4870]: I1014 09:10:16.253830 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="c898dca6-0a58-4692-8bd0-d4bcde9094b5" containerName="registry-server" Oct 14 09:10:16 crc kubenswrapper[4870]: E1014 09:10:16.253862 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c898dca6-0a58-4692-8bd0-d4bcde9094b5" containerName="extract-utilities" Oct 14 09:10:16 crc kubenswrapper[4870]: I1014 09:10:16.253869 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="c898dca6-0a58-4692-8bd0-d4bcde9094b5" containerName="extract-utilities" Oct 14 09:10:16 crc kubenswrapper[4870]: I1014 09:10:16.254077 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="c898dca6-0a58-4692-8bd0-d4bcde9094b5" containerName="registry-server" Oct 14 09:10:16 crc kubenswrapper[4870]: I1014 09:10:16.263157 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5bd5f99ff7-r6rlk" Oct 14 09:10:16 crc kubenswrapper[4870]: I1014 09:10:16.272566 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d6e3771b-e2bb-47d0-8d71-4b8c34d0a786-config\") pod \"dnsmasq-dns-5bd5f99ff7-r6rlk\" (UID: \"d6e3771b-e2bb-47d0-8d71-4b8c34d0a786\") " pod="openstack/dnsmasq-dns-5bd5f99ff7-r6rlk" Oct 14 09:10:16 crc kubenswrapper[4870]: I1014 09:10:16.272642 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d6e3771b-e2bb-47d0-8d71-4b8c34d0a786-ovsdbserver-nb\") pod \"dnsmasq-dns-5bd5f99ff7-r6rlk\" (UID: \"d6e3771b-e2bb-47d0-8d71-4b8c34d0a786\") " pod="openstack/dnsmasq-dns-5bd5f99ff7-r6rlk" Oct 14 09:10:16 crc kubenswrapper[4870]: I1014 09:10:16.272748 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d6e3771b-e2bb-47d0-8d71-4b8c34d0a786-ovsdbserver-sb\") pod \"dnsmasq-dns-5bd5f99ff7-r6rlk\" (UID: \"d6e3771b-e2bb-47d0-8d71-4b8c34d0a786\") " pod="openstack/dnsmasq-dns-5bd5f99ff7-r6rlk" Oct 14 09:10:16 crc kubenswrapper[4870]: I1014 09:10:16.272808 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/d6e3771b-e2bb-47d0-8d71-4b8c34d0a786-openstack-cell1\") pod \"dnsmasq-dns-5bd5f99ff7-r6rlk\" (UID: \"d6e3771b-e2bb-47d0-8d71-4b8c34d0a786\") " pod="openstack/dnsmasq-dns-5bd5f99ff7-r6rlk" Oct 14 09:10:16 crc kubenswrapper[4870]: I1014 09:10:16.272958 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d6e3771b-e2bb-47d0-8d71-4b8c34d0a786-dns-svc\") pod \"dnsmasq-dns-5bd5f99ff7-r6rlk\" (UID: \"d6e3771b-e2bb-47d0-8d71-4b8c34d0a786\") " pod="openstack/dnsmasq-dns-5bd5f99ff7-r6rlk" Oct 14 09:10:16 crc kubenswrapper[4870]: I1014 09:10:16.273095 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pmnnh\" (UniqueName: \"kubernetes.io/projected/d6e3771b-e2bb-47d0-8d71-4b8c34d0a786-kube-api-access-pmnnh\") pod \"dnsmasq-dns-5bd5f99ff7-r6rlk\" (UID: \"d6e3771b-e2bb-47d0-8d71-4b8c34d0a786\") " pod="openstack/dnsmasq-dns-5bd5f99ff7-r6rlk" Oct 14 09:10:16 crc kubenswrapper[4870]: I1014 09:10:16.273428 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1" Oct 14 09:10:16 crc kubenswrapper[4870]: I1014 09:10:16.296483 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5bd5f99ff7-r6rlk"] Oct 14 09:10:16 crc kubenswrapper[4870]: I1014 09:10:16.377230 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d6e3771b-e2bb-47d0-8d71-4b8c34d0a786-ovsdbserver-nb\") pod \"dnsmasq-dns-5bd5f99ff7-r6rlk\" (UID: \"d6e3771b-e2bb-47d0-8d71-4b8c34d0a786\") " pod="openstack/dnsmasq-dns-5bd5f99ff7-r6rlk" Oct 14 09:10:16 crc kubenswrapper[4870]: I1014 09:10:16.377387 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d6e3771b-e2bb-47d0-8d71-4b8c34d0a786-ovsdbserver-sb\") pod \"dnsmasq-dns-5bd5f99ff7-r6rlk\" (UID: \"d6e3771b-e2bb-47d0-8d71-4b8c34d0a786\") " pod="openstack/dnsmasq-dns-5bd5f99ff7-r6rlk" Oct 14 09:10:16 crc kubenswrapper[4870]: I1014 09:10:16.377459 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/d6e3771b-e2bb-47d0-8d71-4b8c34d0a786-openstack-cell1\") pod \"dnsmasq-dns-5bd5f99ff7-r6rlk\" (UID: \"d6e3771b-e2bb-47d0-8d71-4b8c34d0a786\") " pod="openstack/dnsmasq-dns-5bd5f99ff7-r6rlk" Oct 14 09:10:16 crc kubenswrapper[4870]: I1014 09:10:16.377525 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d6e3771b-e2bb-47d0-8d71-4b8c34d0a786-dns-svc\") pod \"dnsmasq-dns-5bd5f99ff7-r6rlk\" (UID: \"d6e3771b-e2bb-47d0-8d71-4b8c34d0a786\") " pod="openstack/dnsmasq-dns-5bd5f99ff7-r6rlk" Oct 14 09:10:16 crc kubenswrapper[4870]: I1014 09:10:16.377615 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pmnnh\" (UniqueName: \"kubernetes.io/projected/d6e3771b-e2bb-47d0-8d71-4b8c34d0a786-kube-api-access-pmnnh\") pod \"dnsmasq-dns-5bd5f99ff7-r6rlk\" (UID: \"d6e3771b-e2bb-47d0-8d71-4b8c34d0a786\") " pod="openstack/dnsmasq-dns-5bd5f99ff7-r6rlk" Oct 14 09:10:16 crc kubenswrapper[4870]: I1014 09:10:16.377708 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d6e3771b-e2bb-47d0-8d71-4b8c34d0a786-config\") pod \"dnsmasq-dns-5bd5f99ff7-r6rlk\" (UID: \"d6e3771b-e2bb-47d0-8d71-4b8c34d0a786\") " pod="openstack/dnsmasq-dns-5bd5f99ff7-r6rlk" Oct 14 09:10:16 crc kubenswrapper[4870]: I1014 09:10:16.378163 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d6e3771b-e2bb-47d0-8d71-4b8c34d0a786-ovsdbserver-nb\") pod \"dnsmasq-dns-5bd5f99ff7-r6rlk\" (UID: \"d6e3771b-e2bb-47d0-8d71-4b8c34d0a786\") " pod="openstack/dnsmasq-dns-5bd5f99ff7-r6rlk" Oct 14 09:10:16 crc kubenswrapper[4870]: I1014 09:10:16.378221 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d6e3771b-e2bb-47d0-8d71-4b8c34d0a786-ovsdbserver-sb\") pod \"dnsmasq-dns-5bd5f99ff7-r6rlk\" (UID: \"d6e3771b-e2bb-47d0-8d71-4b8c34d0a786\") " pod="openstack/dnsmasq-dns-5bd5f99ff7-r6rlk" Oct 14 09:10:16 crc kubenswrapper[4870]: I1014 09:10:16.378791 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d6e3771b-e2bb-47d0-8d71-4b8c34d0a786-config\") pod \"dnsmasq-dns-5bd5f99ff7-r6rlk\" (UID: \"d6e3771b-e2bb-47d0-8d71-4b8c34d0a786\") " pod="openstack/dnsmasq-dns-5bd5f99ff7-r6rlk" Oct 14 09:10:16 crc kubenswrapper[4870]: I1014 09:10:16.378959 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d6e3771b-e2bb-47d0-8d71-4b8c34d0a786-dns-svc\") pod \"dnsmasq-dns-5bd5f99ff7-r6rlk\" (UID: \"d6e3771b-e2bb-47d0-8d71-4b8c34d0a786\") " pod="openstack/dnsmasq-dns-5bd5f99ff7-r6rlk" Oct 14 09:10:16 crc kubenswrapper[4870]: I1014 09:10:16.379014 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/d6e3771b-e2bb-47d0-8d71-4b8c34d0a786-openstack-cell1\") pod \"dnsmasq-dns-5bd5f99ff7-r6rlk\" (UID: \"d6e3771b-e2bb-47d0-8d71-4b8c34d0a786\") " pod="openstack/dnsmasq-dns-5bd5f99ff7-r6rlk" Oct 14 09:10:16 crc kubenswrapper[4870]: I1014 09:10:16.402519 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pmnnh\" (UniqueName: \"kubernetes.io/projected/d6e3771b-e2bb-47d0-8d71-4b8c34d0a786-kube-api-access-pmnnh\") pod \"dnsmasq-dns-5bd5f99ff7-r6rlk\" (UID: \"d6e3771b-e2bb-47d0-8d71-4b8c34d0a786\") " pod="openstack/dnsmasq-dns-5bd5f99ff7-r6rlk" Oct 14 09:10:16 crc kubenswrapper[4870]: I1014 09:10:16.617911 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5bd5f99ff7-r6rlk" Oct 14 09:10:17 crc kubenswrapper[4870]: I1014 09:10:17.125511 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5bd5f99ff7-r6rlk"] Oct 14 09:10:17 crc kubenswrapper[4870]: I1014 09:10:17.822981 4870 generic.go:334] "Generic (PLEG): container finished" podID="d6e3771b-e2bb-47d0-8d71-4b8c34d0a786" containerID="d006dde91c5f38a31c1d94418a694b2cdb84fb994a2cc859c55225e773b7161e" exitCode=0 Oct 14 09:10:17 crc kubenswrapper[4870]: I1014 09:10:17.823042 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5bd5f99ff7-r6rlk" event={"ID":"d6e3771b-e2bb-47d0-8d71-4b8c34d0a786","Type":"ContainerDied","Data":"d006dde91c5f38a31c1d94418a694b2cdb84fb994a2cc859c55225e773b7161e"} Oct 14 09:10:17 crc kubenswrapper[4870]: I1014 09:10:17.823333 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5bd5f99ff7-r6rlk" event={"ID":"d6e3771b-e2bb-47d0-8d71-4b8c34d0a786","Type":"ContainerStarted","Data":"e32b91142bb001541ea8cec638d106e7cf22861cf926b05828c047d3bd6cb2b9"} Oct 14 09:10:18 crc kubenswrapper[4870]: I1014 09:10:18.838980 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5bd5f99ff7-r6rlk" event={"ID":"d6e3771b-e2bb-47d0-8d71-4b8c34d0a786","Type":"ContainerStarted","Data":"bfd5da21e90abfe40ac98135c19880c182d190bd748b4c393938af18b6b28ed2"} Oct 14 09:10:18 crc kubenswrapper[4870]: I1014 09:10:18.839564 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5bd5f99ff7-r6rlk" Oct 14 09:10:23 crc kubenswrapper[4870]: I1014 09:10:23.036415 4870 scope.go:117] "RemoveContainer" containerID="8fcaeb3090f0c89cba1e66db524b4a246072daa35592012664aca450fa09ddc9" Oct 14 09:10:23 crc kubenswrapper[4870]: E1014 09:10:23.039279 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 09:10:26 crc kubenswrapper[4870]: I1014 09:10:26.619699 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5bd5f99ff7-r6rlk" Oct 14 09:10:26 crc kubenswrapper[4870]: I1014 09:10:26.657351 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5bd5f99ff7-r6rlk" podStartSLOduration=10.657089855 podStartE2EDuration="10.657089855s" podCreationTimestamp="2025-10-14 09:10:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 09:10:18.868190101 +0000 UTC m=+7754.565550512" watchObservedRunningTime="2025-10-14 09:10:26.657089855 +0000 UTC m=+7762.354450236" Oct 14 09:10:26 crc kubenswrapper[4870]: I1014 09:10:26.711148 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-849c9dd687-6z7z9"] Oct 14 09:10:26 crc kubenswrapper[4870]: I1014 09:10:26.711454 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-849c9dd687-6z7z9" podUID="b5697993-d512-4996-b3d0-dce02c8b8114" containerName="dnsmasq-dns" containerID="cri-o://37f1920472504bf0e5fe2a0d3ddb02ba22238fccc5e86737d9be4e16e27ca98b" gracePeriod=10 Oct 14 09:10:26 crc kubenswrapper[4870]: I1014 09:10:26.853472 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-684cfb547c-6h8np"] Oct 14 09:10:26 crc kubenswrapper[4870]: I1014 09:10:26.856371 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-684cfb547c-6h8np" Oct 14 09:10:26 crc kubenswrapper[4870]: I1014 09:10:26.897470 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-684cfb547c-6h8np"] Oct 14 09:10:26 crc kubenswrapper[4870]: I1014 09:10:26.931368 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/8657fe11-9fab-465a-bc32-a62447369e2c-openstack-cell1\") pod \"dnsmasq-dns-684cfb547c-6h8np\" (UID: \"8657fe11-9fab-465a-bc32-a62447369e2c\") " pod="openstack/dnsmasq-dns-684cfb547c-6h8np" Oct 14 09:10:26 crc kubenswrapper[4870]: I1014 09:10:26.931431 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8657fe11-9fab-465a-bc32-a62447369e2c-config\") pod \"dnsmasq-dns-684cfb547c-6h8np\" (UID: \"8657fe11-9fab-465a-bc32-a62447369e2c\") " pod="openstack/dnsmasq-dns-684cfb547c-6h8np" Oct 14 09:10:26 crc kubenswrapper[4870]: I1014 09:10:26.931478 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jjd4q\" (UniqueName: \"kubernetes.io/projected/8657fe11-9fab-465a-bc32-a62447369e2c-kube-api-access-jjd4q\") pod \"dnsmasq-dns-684cfb547c-6h8np\" (UID: \"8657fe11-9fab-465a-bc32-a62447369e2c\") " pod="openstack/dnsmasq-dns-684cfb547c-6h8np" Oct 14 09:10:26 crc kubenswrapper[4870]: I1014 09:10:26.931510 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8657fe11-9fab-465a-bc32-a62447369e2c-dns-svc\") pod \"dnsmasq-dns-684cfb547c-6h8np\" (UID: \"8657fe11-9fab-465a-bc32-a62447369e2c\") " pod="openstack/dnsmasq-dns-684cfb547c-6h8np" Oct 14 09:10:26 crc kubenswrapper[4870]: I1014 09:10:26.931614 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8657fe11-9fab-465a-bc32-a62447369e2c-ovsdbserver-sb\") pod \"dnsmasq-dns-684cfb547c-6h8np\" (UID: \"8657fe11-9fab-465a-bc32-a62447369e2c\") " pod="openstack/dnsmasq-dns-684cfb547c-6h8np" Oct 14 09:10:26 crc kubenswrapper[4870]: I1014 09:10:26.931739 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8657fe11-9fab-465a-bc32-a62447369e2c-ovsdbserver-nb\") pod \"dnsmasq-dns-684cfb547c-6h8np\" (UID: \"8657fe11-9fab-465a-bc32-a62447369e2c\") " pod="openstack/dnsmasq-dns-684cfb547c-6h8np" Oct 14 09:10:26 crc kubenswrapper[4870]: I1014 09:10:26.939111 4870 generic.go:334] "Generic (PLEG): container finished" podID="b5697993-d512-4996-b3d0-dce02c8b8114" containerID="37f1920472504bf0e5fe2a0d3ddb02ba22238fccc5e86737d9be4e16e27ca98b" exitCode=0 Oct 14 09:10:26 crc kubenswrapper[4870]: I1014 09:10:26.939174 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-849c9dd687-6z7z9" event={"ID":"b5697993-d512-4996-b3d0-dce02c8b8114","Type":"ContainerDied","Data":"37f1920472504bf0e5fe2a0d3ddb02ba22238fccc5e86737d9be4e16e27ca98b"} Oct 14 09:10:26 crc kubenswrapper[4870]: E1014 09:10:26.943069 4870 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb5697993_d512_4996_b3d0_dce02c8b8114.slice/crio-conmon-37f1920472504bf0e5fe2a0d3ddb02ba22238fccc5e86737d9be4e16e27ca98b.scope\": RecentStats: unable to find data in memory cache]" Oct 14 09:10:27 crc kubenswrapper[4870]: I1014 09:10:27.033171 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/8657fe11-9fab-465a-bc32-a62447369e2c-openstack-cell1\") pod \"dnsmasq-dns-684cfb547c-6h8np\" (UID: \"8657fe11-9fab-465a-bc32-a62447369e2c\") " pod="openstack/dnsmasq-dns-684cfb547c-6h8np" Oct 14 09:10:27 crc kubenswrapper[4870]: I1014 09:10:27.033236 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8657fe11-9fab-465a-bc32-a62447369e2c-config\") pod \"dnsmasq-dns-684cfb547c-6h8np\" (UID: \"8657fe11-9fab-465a-bc32-a62447369e2c\") " pod="openstack/dnsmasq-dns-684cfb547c-6h8np" Oct 14 09:10:27 crc kubenswrapper[4870]: I1014 09:10:27.033269 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jjd4q\" (UniqueName: \"kubernetes.io/projected/8657fe11-9fab-465a-bc32-a62447369e2c-kube-api-access-jjd4q\") pod \"dnsmasq-dns-684cfb547c-6h8np\" (UID: \"8657fe11-9fab-465a-bc32-a62447369e2c\") " pod="openstack/dnsmasq-dns-684cfb547c-6h8np" Oct 14 09:10:27 crc kubenswrapper[4870]: I1014 09:10:27.033301 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8657fe11-9fab-465a-bc32-a62447369e2c-dns-svc\") pod \"dnsmasq-dns-684cfb547c-6h8np\" (UID: \"8657fe11-9fab-465a-bc32-a62447369e2c\") " pod="openstack/dnsmasq-dns-684cfb547c-6h8np" Oct 14 09:10:27 crc kubenswrapper[4870]: I1014 09:10:27.033365 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8657fe11-9fab-465a-bc32-a62447369e2c-ovsdbserver-sb\") pod \"dnsmasq-dns-684cfb547c-6h8np\" (UID: \"8657fe11-9fab-465a-bc32-a62447369e2c\") " pod="openstack/dnsmasq-dns-684cfb547c-6h8np" Oct 14 09:10:27 crc kubenswrapper[4870]: I1014 09:10:27.033417 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8657fe11-9fab-465a-bc32-a62447369e2c-ovsdbserver-nb\") pod \"dnsmasq-dns-684cfb547c-6h8np\" (UID: \"8657fe11-9fab-465a-bc32-a62447369e2c\") " pod="openstack/dnsmasq-dns-684cfb547c-6h8np" Oct 14 09:10:27 crc kubenswrapper[4870]: I1014 09:10:27.034163 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8657fe11-9fab-465a-bc32-a62447369e2c-ovsdbserver-nb\") pod \"dnsmasq-dns-684cfb547c-6h8np\" (UID: \"8657fe11-9fab-465a-bc32-a62447369e2c\") " pod="openstack/dnsmasq-dns-684cfb547c-6h8np" Oct 14 09:10:27 crc kubenswrapper[4870]: I1014 09:10:27.034213 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/8657fe11-9fab-465a-bc32-a62447369e2c-openstack-cell1\") pod \"dnsmasq-dns-684cfb547c-6h8np\" (UID: \"8657fe11-9fab-465a-bc32-a62447369e2c\") " pod="openstack/dnsmasq-dns-684cfb547c-6h8np" Oct 14 09:10:27 crc kubenswrapper[4870]: I1014 09:10:27.034702 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8657fe11-9fab-465a-bc32-a62447369e2c-config\") pod \"dnsmasq-dns-684cfb547c-6h8np\" (UID: \"8657fe11-9fab-465a-bc32-a62447369e2c\") " pod="openstack/dnsmasq-dns-684cfb547c-6h8np" Oct 14 09:10:27 crc kubenswrapper[4870]: I1014 09:10:27.034879 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8657fe11-9fab-465a-bc32-a62447369e2c-dns-svc\") pod \"dnsmasq-dns-684cfb547c-6h8np\" (UID: \"8657fe11-9fab-465a-bc32-a62447369e2c\") " pod="openstack/dnsmasq-dns-684cfb547c-6h8np" Oct 14 09:10:27 crc kubenswrapper[4870]: I1014 09:10:27.040039 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8657fe11-9fab-465a-bc32-a62447369e2c-ovsdbserver-sb\") pod \"dnsmasq-dns-684cfb547c-6h8np\" (UID: \"8657fe11-9fab-465a-bc32-a62447369e2c\") " pod="openstack/dnsmasq-dns-684cfb547c-6h8np" Oct 14 09:10:27 crc kubenswrapper[4870]: I1014 09:10:27.060329 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jjd4q\" (UniqueName: \"kubernetes.io/projected/8657fe11-9fab-465a-bc32-a62447369e2c-kube-api-access-jjd4q\") pod \"dnsmasq-dns-684cfb547c-6h8np\" (UID: \"8657fe11-9fab-465a-bc32-a62447369e2c\") " pod="openstack/dnsmasq-dns-684cfb547c-6h8np" Oct 14 09:10:27 crc kubenswrapper[4870]: I1014 09:10:27.207427 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-684cfb547c-6h8np" Oct 14 09:10:27 crc kubenswrapper[4870]: I1014 09:10:27.301615 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-849c9dd687-6z7z9" Oct 14 09:10:27 crc kubenswrapper[4870]: I1014 09:10:27.341478 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b5697993-d512-4996-b3d0-dce02c8b8114-config\") pod \"b5697993-d512-4996-b3d0-dce02c8b8114\" (UID: \"b5697993-d512-4996-b3d0-dce02c8b8114\") " Oct 14 09:10:27 crc kubenswrapper[4870]: I1014 09:10:27.341583 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b5697993-d512-4996-b3d0-dce02c8b8114-dns-svc\") pod \"b5697993-d512-4996-b3d0-dce02c8b8114\" (UID: \"b5697993-d512-4996-b3d0-dce02c8b8114\") " Oct 14 09:10:27 crc kubenswrapper[4870]: I1014 09:10:27.341655 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xfzh9\" (UniqueName: \"kubernetes.io/projected/b5697993-d512-4996-b3d0-dce02c8b8114-kube-api-access-xfzh9\") pod \"b5697993-d512-4996-b3d0-dce02c8b8114\" (UID: \"b5697993-d512-4996-b3d0-dce02c8b8114\") " Oct 14 09:10:27 crc kubenswrapper[4870]: I1014 09:10:27.341735 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b5697993-d512-4996-b3d0-dce02c8b8114-ovsdbserver-nb\") pod \"b5697993-d512-4996-b3d0-dce02c8b8114\" (UID: \"b5697993-d512-4996-b3d0-dce02c8b8114\") " Oct 14 09:10:27 crc kubenswrapper[4870]: I1014 09:10:27.341795 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b5697993-d512-4996-b3d0-dce02c8b8114-ovsdbserver-sb\") pod \"b5697993-d512-4996-b3d0-dce02c8b8114\" (UID: \"b5697993-d512-4996-b3d0-dce02c8b8114\") " Oct 14 09:10:27 crc kubenswrapper[4870]: I1014 09:10:27.348286 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b5697993-d512-4996-b3d0-dce02c8b8114-kube-api-access-xfzh9" (OuterVolumeSpecName: "kube-api-access-xfzh9") pod "b5697993-d512-4996-b3d0-dce02c8b8114" (UID: "b5697993-d512-4996-b3d0-dce02c8b8114"). InnerVolumeSpecName "kube-api-access-xfzh9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:10:27 crc kubenswrapper[4870]: I1014 09:10:27.394345 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b5697993-d512-4996-b3d0-dce02c8b8114-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "b5697993-d512-4996-b3d0-dce02c8b8114" (UID: "b5697993-d512-4996-b3d0-dce02c8b8114"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 09:10:27 crc kubenswrapper[4870]: I1014 09:10:27.417942 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b5697993-d512-4996-b3d0-dce02c8b8114-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "b5697993-d512-4996-b3d0-dce02c8b8114" (UID: "b5697993-d512-4996-b3d0-dce02c8b8114"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 09:10:27 crc kubenswrapper[4870]: I1014 09:10:27.420715 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b5697993-d512-4996-b3d0-dce02c8b8114-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "b5697993-d512-4996-b3d0-dce02c8b8114" (UID: "b5697993-d512-4996-b3d0-dce02c8b8114"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 09:10:27 crc kubenswrapper[4870]: I1014 09:10:27.430573 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b5697993-d512-4996-b3d0-dce02c8b8114-config" (OuterVolumeSpecName: "config") pod "b5697993-d512-4996-b3d0-dce02c8b8114" (UID: "b5697993-d512-4996-b3d0-dce02c8b8114"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 09:10:27 crc kubenswrapper[4870]: I1014 09:10:27.445169 4870 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b5697993-d512-4996-b3d0-dce02c8b8114-config\") on node \"crc\" DevicePath \"\"" Oct 14 09:10:27 crc kubenswrapper[4870]: I1014 09:10:27.445212 4870 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b5697993-d512-4996-b3d0-dce02c8b8114-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 14 09:10:27 crc kubenswrapper[4870]: I1014 09:10:27.445246 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xfzh9\" (UniqueName: \"kubernetes.io/projected/b5697993-d512-4996-b3d0-dce02c8b8114-kube-api-access-xfzh9\") on node \"crc\" DevicePath \"\"" Oct 14 09:10:27 crc kubenswrapper[4870]: I1014 09:10:27.445262 4870 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b5697993-d512-4996-b3d0-dce02c8b8114-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 14 09:10:27 crc kubenswrapper[4870]: I1014 09:10:27.445274 4870 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b5697993-d512-4996-b3d0-dce02c8b8114-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 14 09:10:27 crc kubenswrapper[4870]: I1014 09:10:27.695354 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-684cfb547c-6h8np"] Oct 14 09:10:27 crc kubenswrapper[4870]: W1014 09:10:27.699335 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8657fe11_9fab_465a_bc32_a62447369e2c.slice/crio-0d0a0b8d62735684c49f161b44e7687d7843a16ecd04c85721db15a70ed54b75 WatchSource:0}: Error finding container 0d0a0b8d62735684c49f161b44e7687d7843a16ecd04c85721db15a70ed54b75: Status 404 returned error can't find the container with id 0d0a0b8d62735684c49f161b44e7687d7843a16ecd04c85721db15a70ed54b75 Oct 14 09:10:27 crc kubenswrapper[4870]: I1014 09:10:27.954381 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-684cfb547c-6h8np" event={"ID":"8657fe11-9fab-465a-bc32-a62447369e2c","Type":"ContainerStarted","Data":"c8978b1f8aab207381f5916eca4944077ab5b450b57259db580bfd4c248e3577"} Oct 14 09:10:27 crc kubenswrapper[4870]: I1014 09:10:27.954754 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-684cfb547c-6h8np" event={"ID":"8657fe11-9fab-465a-bc32-a62447369e2c","Type":"ContainerStarted","Data":"0d0a0b8d62735684c49f161b44e7687d7843a16ecd04c85721db15a70ed54b75"} Oct 14 09:10:27 crc kubenswrapper[4870]: I1014 09:10:27.956318 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-849c9dd687-6z7z9" event={"ID":"b5697993-d512-4996-b3d0-dce02c8b8114","Type":"ContainerDied","Data":"596b4b8116137f5df05039c3af7209fe77be3937cf83f26780fad788c94a72a4"} Oct 14 09:10:27 crc kubenswrapper[4870]: I1014 09:10:27.956372 4870 scope.go:117] "RemoveContainer" containerID="37f1920472504bf0e5fe2a0d3ddb02ba22238fccc5e86737d9be4e16e27ca98b" Oct 14 09:10:27 crc kubenswrapper[4870]: I1014 09:10:27.956503 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-849c9dd687-6z7z9" Oct 14 09:10:28 crc kubenswrapper[4870]: I1014 09:10:28.188967 4870 scope.go:117] "RemoveContainer" containerID="f312035f067b128f5801edc80f77d1ab8cb42e4a382cb303068a5d2429226e96" Oct 14 09:10:28 crc kubenswrapper[4870]: I1014 09:10:28.244230 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-849c9dd687-6z7z9"] Oct 14 09:10:28 crc kubenswrapper[4870]: I1014 09:10:28.276165 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-849c9dd687-6z7z9"] Oct 14 09:10:28 crc kubenswrapper[4870]: I1014 09:10:28.971891 4870 generic.go:334] "Generic (PLEG): container finished" podID="8657fe11-9fab-465a-bc32-a62447369e2c" containerID="c8978b1f8aab207381f5916eca4944077ab5b450b57259db580bfd4c248e3577" exitCode=0 Oct 14 09:10:28 crc kubenswrapper[4870]: I1014 09:10:28.973918 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-684cfb547c-6h8np" event={"ID":"8657fe11-9fab-465a-bc32-a62447369e2c","Type":"ContainerDied","Data":"c8978b1f8aab207381f5916eca4944077ab5b450b57259db580bfd4c248e3577"} Oct 14 09:10:28 crc kubenswrapper[4870]: I1014 09:10:28.973990 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-684cfb547c-6h8np" event={"ID":"8657fe11-9fab-465a-bc32-a62447369e2c","Type":"ContainerStarted","Data":"0746b0c80c90ad65201147e7953ca31dcb0355ac914276d6b459b1ab8ccd9734"} Oct 14 09:10:28 crc kubenswrapper[4870]: I1014 09:10:28.974051 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-684cfb547c-6h8np" Oct 14 09:10:29 crc kubenswrapper[4870]: I1014 09:10:29.010999 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-684cfb547c-6h8np" podStartSLOduration=3.010959286 podStartE2EDuration="3.010959286s" podCreationTimestamp="2025-10-14 09:10:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 09:10:28.996292654 +0000 UTC m=+7764.693653015" watchObservedRunningTime="2025-10-14 09:10:29.010959286 +0000 UTC m=+7764.708319687" Oct 14 09:10:29 crc kubenswrapper[4870]: I1014 09:10:29.047520 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b5697993-d512-4996-b3d0-dce02c8b8114" path="/var/lib/kubelet/pods/b5697993-d512-4996-b3d0-dce02c8b8114/volumes" Oct 14 09:10:30 crc kubenswrapper[4870]: I1014 09:10:30.656728 4870 scope.go:117] "RemoveContainer" containerID="fd26f0ddc6114ecb8fa4383933073a6afec3038dbefa14de68ea6c388abc85cc" Oct 14 09:10:30 crc kubenswrapper[4870]: I1014 09:10:30.704435 4870 scope.go:117] "RemoveContainer" containerID="ffcade20088eb5ff83f1f617deef2c3d30eced0aac5f3c80b87cd3a1131acbae" Oct 14 09:10:30 crc kubenswrapper[4870]: I1014 09:10:30.773582 4870 scope.go:117] "RemoveContainer" containerID="be40d5b0502c566a97b58ff8206fb2ff8e99acc212f7cc433e012aa49735f5cf" Oct 14 09:10:30 crc kubenswrapper[4870]: I1014 09:10:30.834459 4870 scope.go:117] "RemoveContainer" containerID="d96697d8d135d62e8516d08211e951ebbd5b03db0de02038b259e5f52e2aa8c3" Oct 14 09:10:31 crc kubenswrapper[4870]: I1014 09:10:31.055574 4870 scope.go:117] "RemoveContainer" containerID="65c074e8a549fa46c3006560067a6b54b95899f9654c6d1d2b04c1ed63aabd3f" Oct 14 09:10:31 crc kubenswrapper[4870]: I1014 09:10:31.116420 4870 scope.go:117] "RemoveContainer" containerID="56bbe467d4c1379fd0b140aec08cccaffb367f9bbe2482f5c9f68117f91d1ac7" Oct 14 09:10:34 crc kubenswrapper[4870]: I1014 09:10:34.034004 4870 scope.go:117] "RemoveContainer" containerID="8fcaeb3090f0c89cba1e66db524b4a246072daa35592012664aca450fa09ddc9" Oct 14 09:10:35 crc kubenswrapper[4870]: I1014 09:10:35.087493 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" event={"ID":"874a5e45-dffd-4d17-b609-b7d3ed2eab07","Type":"ContainerStarted","Data":"a9b3d8bfba3e51117bb9f2ff67d8d2b345878f886b38f76fc3c8cf6d7f91c6de"} Oct 14 09:10:37 crc kubenswrapper[4870]: I1014 09:10:37.211713 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-684cfb547c-6h8np" Oct 14 09:10:37 crc kubenswrapper[4870]: I1014 09:10:37.276017 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5bd5f99ff7-r6rlk"] Oct 14 09:10:37 crc kubenswrapper[4870]: I1014 09:10:37.276287 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5bd5f99ff7-r6rlk" podUID="d6e3771b-e2bb-47d0-8d71-4b8c34d0a786" containerName="dnsmasq-dns" containerID="cri-o://bfd5da21e90abfe40ac98135c19880c182d190bd748b4c393938af18b6b28ed2" gracePeriod=10 Oct 14 09:10:37 crc kubenswrapper[4870]: I1014 09:10:37.463533 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-67dbcf57c9-fcxrn"] Oct 14 09:10:37 crc kubenswrapper[4870]: E1014 09:10:37.465034 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5697993-d512-4996-b3d0-dce02c8b8114" containerName="init" Oct 14 09:10:37 crc kubenswrapper[4870]: I1014 09:10:37.465058 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5697993-d512-4996-b3d0-dce02c8b8114" containerName="init" Oct 14 09:10:37 crc kubenswrapper[4870]: E1014 09:10:37.465071 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5697993-d512-4996-b3d0-dce02c8b8114" containerName="dnsmasq-dns" Oct 14 09:10:37 crc kubenswrapper[4870]: I1014 09:10:37.465078 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5697993-d512-4996-b3d0-dce02c8b8114" containerName="dnsmasq-dns" Oct 14 09:10:37 crc kubenswrapper[4870]: I1014 09:10:37.465399 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="b5697993-d512-4996-b3d0-dce02c8b8114" containerName="dnsmasq-dns" Oct 14 09:10:37 crc kubenswrapper[4870]: I1014 09:10:37.473859 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67dbcf57c9-fcxrn" Oct 14 09:10:37 crc kubenswrapper[4870]: I1014 09:10:37.481534 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-67dbcf57c9-fcxrn"] Oct 14 09:10:37 crc kubenswrapper[4870]: I1014 09:10:37.483207 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-networker" Oct 14 09:10:37 crc kubenswrapper[4870]: I1014 09:10:37.524626 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/e5c27b78-9614-44c4-b96e-db0380508659-openstack-cell1\") pod \"dnsmasq-dns-67dbcf57c9-fcxrn\" (UID: \"e5c27b78-9614-44c4-b96e-db0380508659\") " pod="openstack/dnsmasq-dns-67dbcf57c9-fcxrn" Oct 14 09:10:37 crc kubenswrapper[4870]: I1014 09:10:37.524747 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dn2fq\" (UniqueName: \"kubernetes.io/projected/e5c27b78-9614-44c4-b96e-db0380508659-kube-api-access-dn2fq\") pod \"dnsmasq-dns-67dbcf57c9-fcxrn\" (UID: \"e5c27b78-9614-44c4-b96e-db0380508659\") " pod="openstack/dnsmasq-dns-67dbcf57c9-fcxrn" Oct 14 09:10:37 crc kubenswrapper[4870]: I1014 09:10:37.524899 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e5c27b78-9614-44c4-b96e-db0380508659-dns-svc\") pod \"dnsmasq-dns-67dbcf57c9-fcxrn\" (UID: \"e5c27b78-9614-44c4-b96e-db0380508659\") " pod="openstack/dnsmasq-dns-67dbcf57c9-fcxrn" Oct 14 09:10:37 crc kubenswrapper[4870]: I1014 09:10:37.524960 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e5c27b78-9614-44c4-b96e-db0380508659-ovsdbserver-nb\") pod \"dnsmasq-dns-67dbcf57c9-fcxrn\" (UID: \"e5c27b78-9614-44c4-b96e-db0380508659\") " pod="openstack/dnsmasq-dns-67dbcf57c9-fcxrn" Oct 14 09:10:37 crc kubenswrapper[4870]: I1014 09:10:37.524986 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-networker\" (UniqueName: \"kubernetes.io/configmap/e5c27b78-9614-44c4-b96e-db0380508659-openstack-networker\") pod \"dnsmasq-dns-67dbcf57c9-fcxrn\" (UID: \"e5c27b78-9614-44c4-b96e-db0380508659\") " pod="openstack/dnsmasq-dns-67dbcf57c9-fcxrn" Oct 14 09:10:37 crc kubenswrapper[4870]: I1014 09:10:37.525082 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e5c27b78-9614-44c4-b96e-db0380508659-ovsdbserver-sb\") pod \"dnsmasq-dns-67dbcf57c9-fcxrn\" (UID: \"e5c27b78-9614-44c4-b96e-db0380508659\") " pod="openstack/dnsmasq-dns-67dbcf57c9-fcxrn" Oct 14 09:10:37 crc kubenswrapper[4870]: I1014 09:10:37.525109 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e5c27b78-9614-44c4-b96e-db0380508659-config\") pod \"dnsmasq-dns-67dbcf57c9-fcxrn\" (UID: \"e5c27b78-9614-44c4-b96e-db0380508659\") " pod="openstack/dnsmasq-dns-67dbcf57c9-fcxrn" Oct 14 09:10:37 crc kubenswrapper[4870]: I1014 09:10:37.626703 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e5c27b78-9614-44c4-b96e-db0380508659-dns-svc\") pod \"dnsmasq-dns-67dbcf57c9-fcxrn\" (UID: \"e5c27b78-9614-44c4-b96e-db0380508659\") " pod="openstack/dnsmasq-dns-67dbcf57c9-fcxrn" Oct 14 09:10:37 crc kubenswrapper[4870]: I1014 09:10:37.626766 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e5c27b78-9614-44c4-b96e-db0380508659-ovsdbserver-nb\") pod \"dnsmasq-dns-67dbcf57c9-fcxrn\" (UID: \"e5c27b78-9614-44c4-b96e-db0380508659\") " pod="openstack/dnsmasq-dns-67dbcf57c9-fcxrn" Oct 14 09:10:37 crc kubenswrapper[4870]: I1014 09:10:37.626789 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-networker\" (UniqueName: \"kubernetes.io/configmap/e5c27b78-9614-44c4-b96e-db0380508659-openstack-networker\") pod \"dnsmasq-dns-67dbcf57c9-fcxrn\" (UID: \"e5c27b78-9614-44c4-b96e-db0380508659\") " pod="openstack/dnsmasq-dns-67dbcf57c9-fcxrn" Oct 14 09:10:37 crc kubenswrapper[4870]: I1014 09:10:37.626858 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e5c27b78-9614-44c4-b96e-db0380508659-ovsdbserver-sb\") pod \"dnsmasq-dns-67dbcf57c9-fcxrn\" (UID: \"e5c27b78-9614-44c4-b96e-db0380508659\") " pod="openstack/dnsmasq-dns-67dbcf57c9-fcxrn" Oct 14 09:10:37 crc kubenswrapper[4870]: I1014 09:10:37.626880 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e5c27b78-9614-44c4-b96e-db0380508659-config\") pod \"dnsmasq-dns-67dbcf57c9-fcxrn\" (UID: \"e5c27b78-9614-44c4-b96e-db0380508659\") " pod="openstack/dnsmasq-dns-67dbcf57c9-fcxrn" Oct 14 09:10:37 crc kubenswrapper[4870]: I1014 09:10:37.626929 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/e5c27b78-9614-44c4-b96e-db0380508659-openstack-cell1\") pod \"dnsmasq-dns-67dbcf57c9-fcxrn\" (UID: \"e5c27b78-9614-44c4-b96e-db0380508659\") " pod="openstack/dnsmasq-dns-67dbcf57c9-fcxrn" Oct 14 09:10:37 crc kubenswrapper[4870]: I1014 09:10:37.626969 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dn2fq\" (UniqueName: \"kubernetes.io/projected/e5c27b78-9614-44c4-b96e-db0380508659-kube-api-access-dn2fq\") pod \"dnsmasq-dns-67dbcf57c9-fcxrn\" (UID: \"e5c27b78-9614-44c4-b96e-db0380508659\") " pod="openstack/dnsmasq-dns-67dbcf57c9-fcxrn" Oct 14 09:10:37 crc kubenswrapper[4870]: I1014 09:10:37.628347 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-networker\" (UniqueName: \"kubernetes.io/configmap/e5c27b78-9614-44c4-b96e-db0380508659-openstack-networker\") pod \"dnsmasq-dns-67dbcf57c9-fcxrn\" (UID: \"e5c27b78-9614-44c4-b96e-db0380508659\") " pod="openstack/dnsmasq-dns-67dbcf57c9-fcxrn" Oct 14 09:10:37 crc kubenswrapper[4870]: I1014 09:10:37.628398 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e5c27b78-9614-44c4-b96e-db0380508659-dns-svc\") pod \"dnsmasq-dns-67dbcf57c9-fcxrn\" (UID: \"e5c27b78-9614-44c4-b96e-db0380508659\") " pod="openstack/dnsmasq-dns-67dbcf57c9-fcxrn" Oct 14 09:10:37 crc kubenswrapper[4870]: I1014 09:10:37.628485 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e5c27b78-9614-44c4-b96e-db0380508659-ovsdbserver-sb\") pod \"dnsmasq-dns-67dbcf57c9-fcxrn\" (UID: \"e5c27b78-9614-44c4-b96e-db0380508659\") " pod="openstack/dnsmasq-dns-67dbcf57c9-fcxrn" Oct 14 09:10:37 crc kubenswrapper[4870]: I1014 09:10:37.628719 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/e5c27b78-9614-44c4-b96e-db0380508659-openstack-cell1\") pod \"dnsmasq-dns-67dbcf57c9-fcxrn\" (UID: \"e5c27b78-9614-44c4-b96e-db0380508659\") " pod="openstack/dnsmasq-dns-67dbcf57c9-fcxrn" Oct 14 09:10:37 crc kubenswrapper[4870]: I1014 09:10:37.628940 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e5c27b78-9614-44c4-b96e-db0380508659-ovsdbserver-nb\") pod \"dnsmasq-dns-67dbcf57c9-fcxrn\" (UID: \"e5c27b78-9614-44c4-b96e-db0380508659\") " pod="openstack/dnsmasq-dns-67dbcf57c9-fcxrn" Oct 14 09:10:37 crc kubenswrapper[4870]: I1014 09:10:37.629231 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e5c27b78-9614-44c4-b96e-db0380508659-config\") pod \"dnsmasq-dns-67dbcf57c9-fcxrn\" (UID: \"e5c27b78-9614-44c4-b96e-db0380508659\") " pod="openstack/dnsmasq-dns-67dbcf57c9-fcxrn" Oct 14 09:10:37 crc kubenswrapper[4870]: I1014 09:10:37.658734 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dn2fq\" (UniqueName: \"kubernetes.io/projected/e5c27b78-9614-44c4-b96e-db0380508659-kube-api-access-dn2fq\") pod \"dnsmasq-dns-67dbcf57c9-fcxrn\" (UID: \"e5c27b78-9614-44c4-b96e-db0380508659\") " pod="openstack/dnsmasq-dns-67dbcf57c9-fcxrn" Oct 14 09:10:37 crc kubenswrapper[4870]: I1014 09:10:37.810519 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67dbcf57c9-fcxrn" Oct 14 09:10:37 crc kubenswrapper[4870]: I1014 09:10:37.970463 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5bd5f99ff7-r6rlk" Oct 14 09:10:38 crc kubenswrapper[4870]: I1014 09:10:38.038155 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pmnnh\" (UniqueName: \"kubernetes.io/projected/d6e3771b-e2bb-47d0-8d71-4b8c34d0a786-kube-api-access-pmnnh\") pod \"d6e3771b-e2bb-47d0-8d71-4b8c34d0a786\" (UID: \"d6e3771b-e2bb-47d0-8d71-4b8c34d0a786\") " Oct 14 09:10:38 crc kubenswrapper[4870]: I1014 09:10:38.038288 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d6e3771b-e2bb-47d0-8d71-4b8c34d0a786-config\") pod \"d6e3771b-e2bb-47d0-8d71-4b8c34d0a786\" (UID: \"d6e3771b-e2bb-47d0-8d71-4b8c34d0a786\") " Oct 14 09:10:38 crc kubenswrapper[4870]: I1014 09:10:38.038497 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d6e3771b-e2bb-47d0-8d71-4b8c34d0a786-ovsdbserver-sb\") pod \"d6e3771b-e2bb-47d0-8d71-4b8c34d0a786\" (UID: \"d6e3771b-e2bb-47d0-8d71-4b8c34d0a786\") " Oct 14 09:10:38 crc kubenswrapper[4870]: I1014 09:10:38.038558 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d6e3771b-e2bb-47d0-8d71-4b8c34d0a786-dns-svc\") pod \"d6e3771b-e2bb-47d0-8d71-4b8c34d0a786\" (UID: \"d6e3771b-e2bb-47d0-8d71-4b8c34d0a786\") " Oct 14 09:10:38 crc kubenswrapper[4870]: I1014 09:10:38.038640 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/d6e3771b-e2bb-47d0-8d71-4b8c34d0a786-openstack-cell1\") pod \"d6e3771b-e2bb-47d0-8d71-4b8c34d0a786\" (UID: \"d6e3771b-e2bb-47d0-8d71-4b8c34d0a786\") " Oct 14 09:10:38 crc kubenswrapper[4870]: I1014 09:10:38.038683 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d6e3771b-e2bb-47d0-8d71-4b8c34d0a786-ovsdbserver-nb\") pod \"d6e3771b-e2bb-47d0-8d71-4b8c34d0a786\" (UID: \"d6e3771b-e2bb-47d0-8d71-4b8c34d0a786\") " Oct 14 09:10:38 crc kubenswrapper[4870]: I1014 09:10:38.059807 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d6e3771b-e2bb-47d0-8d71-4b8c34d0a786-kube-api-access-pmnnh" (OuterVolumeSpecName: "kube-api-access-pmnnh") pod "d6e3771b-e2bb-47d0-8d71-4b8c34d0a786" (UID: "d6e3771b-e2bb-47d0-8d71-4b8c34d0a786"). InnerVolumeSpecName "kube-api-access-pmnnh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:10:38 crc kubenswrapper[4870]: I1014 09:10:38.098333 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d6e3771b-e2bb-47d0-8d71-4b8c34d0a786-openstack-cell1" (OuterVolumeSpecName: "openstack-cell1") pod "d6e3771b-e2bb-47d0-8d71-4b8c34d0a786" (UID: "d6e3771b-e2bb-47d0-8d71-4b8c34d0a786"). InnerVolumeSpecName "openstack-cell1". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 09:10:38 crc kubenswrapper[4870]: I1014 09:10:38.123905 4870 generic.go:334] "Generic (PLEG): container finished" podID="d6e3771b-e2bb-47d0-8d71-4b8c34d0a786" containerID="bfd5da21e90abfe40ac98135c19880c182d190bd748b4c393938af18b6b28ed2" exitCode=0 Oct 14 09:10:38 crc kubenswrapper[4870]: I1014 09:10:38.123953 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5bd5f99ff7-r6rlk" event={"ID":"d6e3771b-e2bb-47d0-8d71-4b8c34d0a786","Type":"ContainerDied","Data":"bfd5da21e90abfe40ac98135c19880c182d190bd748b4c393938af18b6b28ed2"} Oct 14 09:10:38 crc kubenswrapper[4870]: I1014 09:10:38.123980 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5bd5f99ff7-r6rlk" event={"ID":"d6e3771b-e2bb-47d0-8d71-4b8c34d0a786","Type":"ContainerDied","Data":"e32b91142bb001541ea8cec638d106e7cf22861cf926b05828c047d3bd6cb2b9"} Oct 14 09:10:38 crc kubenswrapper[4870]: I1014 09:10:38.123996 4870 scope.go:117] "RemoveContainer" containerID="bfd5da21e90abfe40ac98135c19880c182d190bd748b4c393938af18b6b28ed2" Oct 14 09:10:38 crc kubenswrapper[4870]: I1014 09:10:38.124115 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5bd5f99ff7-r6rlk" Oct 14 09:10:38 crc kubenswrapper[4870]: I1014 09:10:38.125474 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d6e3771b-e2bb-47d0-8d71-4b8c34d0a786-config" (OuterVolumeSpecName: "config") pod "d6e3771b-e2bb-47d0-8d71-4b8c34d0a786" (UID: "d6e3771b-e2bb-47d0-8d71-4b8c34d0a786"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 09:10:38 crc kubenswrapper[4870]: I1014 09:10:38.133122 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d6e3771b-e2bb-47d0-8d71-4b8c34d0a786-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "d6e3771b-e2bb-47d0-8d71-4b8c34d0a786" (UID: "d6e3771b-e2bb-47d0-8d71-4b8c34d0a786"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 09:10:38 crc kubenswrapper[4870]: I1014 09:10:38.134838 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d6e3771b-e2bb-47d0-8d71-4b8c34d0a786-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "d6e3771b-e2bb-47d0-8d71-4b8c34d0a786" (UID: "d6e3771b-e2bb-47d0-8d71-4b8c34d0a786"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 09:10:38 crc kubenswrapper[4870]: I1014 09:10:38.140946 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pmnnh\" (UniqueName: \"kubernetes.io/projected/d6e3771b-e2bb-47d0-8d71-4b8c34d0a786-kube-api-access-pmnnh\") on node \"crc\" DevicePath \"\"" Oct 14 09:10:38 crc kubenswrapper[4870]: I1014 09:10:38.140972 4870 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d6e3771b-e2bb-47d0-8d71-4b8c34d0a786-config\") on node \"crc\" DevicePath \"\"" Oct 14 09:10:38 crc kubenswrapper[4870]: I1014 09:10:38.140982 4870 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d6e3771b-e2bb-47d0-8d71-4b8c34d0a786-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 14 09:10:38 crc kubenswrapper[4870]: I1014 09:10:38.140991 4870 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d6e3771b-e2bb-47d0-8d71-4b8c34d0a786-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 14 09:10:38 crc kubenswrapper[4870]: I1014 09:10:38.141001 4870 reconciler_common.go:293] "Volume detached for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/d6e3771b-e2bb-47d0-8d71-4b8c34d0a786-openstack-cell1\") on node \"crc\" DevicePath \"\"" Oct 14 09:10:38 crc kubenswrapper[4870]: I1014 09:10:38.145621 4870 scope.go:117] "RemoveContainer" containerID="d006dde91c5f38a31c1d94418a694b2cdb84fb994a2cc859c55225e773b7161e" Oct 14 09:10:38 crc kubenswrapper[4870]: I1014 09:10:38.149412 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d6e3771b-e2bb-47d0-8d71-4b8c34d0a786-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "d6e3771b-e2bb-47d0-8d71-4b8c34d0a786" (UID: "d6e3771b-e2bb-47d0-8d71-4b8c34d0a786"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 09:10:38 crc kubenswrapper[4870]: I1014 09:10:38.163214 4870 scope.go:117] "RemoveContainer" containerID="bfd5da21e90abfe40ac98135c19880c182d190bd748b4c393938af18b6b28ed2" Oct 14 09:10:38 crc kubenswrapper[4870]: E1014 09:10:38.163692 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bfd5da21e90abfe40ac98135c19880c182d190bd748b4c393938af18b6b28ed2\": container with ID starting with bfd5da21e90abfe40ac98135c19880c182d190bd748b4c393938af18b6b28ed2 not found: ID does not exist" containerID="bfd5da21e90abfe40ac98135c19880c182d190bd748b4c393938af18b6b28ed2" Oct 14 09:10:38 crc kubenswrapper[4870]: I1014 09:10:38.163722 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bfd5da21e90abfe40ac98135c19880c182d190bd748b4c393938af18b6b28ed2"} err="failed to get container status \"bfd5da21e90abfe40ac98135c19880c182d190bd748b4c393938af18b6b28ed2\": rpc error: code = NotFound desc = could not find container \"bfd5da21e90abfe40ac98135c19880c182d190bd748b4c393938af18b6b28ed2\": container with ID starting with bfd5da21e90abfe40ac98135c19880c182d190bd748b4c393938af18b6b28ed2 not found: ID does not exist" Oct 14 09:10:38 crc kubenswrapper[4870]: I1014 09:10:38.163742 4870 scope.go:117] "RemoveContainer" containerID="d006dde91c5f38a31c1d94418a694b2cdb84fb994a2cc859c55225e773b7161e" Oct 14 09:10:38 crc kubenswrapper[4870]: E1014 09:10:38.163985 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d006dde91c5f38a31c1d94418a694b2cdb84fb994a2cc859c55225e773b7161e\": container with ID starting with d006dde91c5f38a31c1d94418a694b2cdb84fb994a2cc859c55225e773b7161e not found: ID does not exist" containerID="d006dde91c5f38a31c1d94418a694b2cdb84fb994a2cc859c55225e773b7161e" Oct 14 09:10:38 crc kubenswrapper[4870]: I1014 09:10:38.164031 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d006dde91c5f38a31c1d94418a694b2cdb84fb994a2cc859c55225e773b7161e"} err="failed to get container status \"d006dde91c5f38a31c1d94418a694b2cdb84fb994a2cc859c55225e773b7161e\": rpc error: code = NotFound desc = could not find container \"d006dde91c5f38a31c1d94418a694b2cdb84fb994a2cc859c55225e773b7161e\": container with ID starting with d006dde91c5f38a31c1d94418a694b2cdb84fb994a2cc859c55225e773b7161e not found: ID does not exist" Oct 14 09:10:38 crc kubenswrapper[4870]: I1014 09:10:38.243504 4870 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d6e3771b-e2bb-47d0-8d71-4b8c34d0a786-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 14 09:10:38 crc kubenswrapper[4870]: I1014 09:10:38.304471 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-67dbcf57c9-fcxrn"] Oct 14 09:10:38 crc kubenswrapper[4870]: W1014 09:10:38.325271 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode5c27b78_9614_44c4_b96e_db0380508659.slice/crio-d1d37e3096c9038f26ea99eab30cac84d7cadc60c2912c801119f87ff00cce8a WatchSource:0}: Error finding container d1d37e3096c9038f26ea99eab30cac84d7cadc60c2912c801119f87ff00cce8a: Status 404 returned error can't find the container with id d1d37e3096c9038f26ea99eab30cac84d7cadc60c2912c801119f87ff00cce8a Oct 14 09:10:38 crc kubenswrapper[4870]: I1014 09:10:38.530563 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5bd5f99ff7-r6rlk"] Oct 14 09:10:38 crc kubenswrapper[4870]: I1014 09:10:38.538922 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5bd5f99ff7-r6rlk"] Oct 14 09:10:39 crc kubenswrapper[4870]: I1014 09:10:39.055372 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d6e3771b-e2bb-47d0-8d71-4b8c34d0a786" path="/var/lib/kubelet/pods/d6e3771b-e2bb-47d0-8d71-4b8c34d0a786/volumes" Oct 14 09:10:39 crc kubenswrapper[4870]: I1014 09:10:39.139687 4870 generic.go:334] "Generic (PLEG): container finished" podID="e5c27b78-9614-44c4-b96e-db0380508659" containerID="efbb36a6c810ef3612b36cf8e84e610caf79f5b312f16be58ee6ef555369e3e1" exitCode=0 Oct 14 09:10:39 crc kubenswrapper[4870]: I1014 09:10:39.139772 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67dbcf57c9-fcxrn" event={"ID":"e5c27b78-9614-44c4-b96e-db0380508659","Type":"ContainerDied","Data":"efbb36a6c810ef3612b36cf8e84e610caf79f5b312f16be58ee6ef555369e3e1"} Oct 14 09:10:39 crc kubenswrapper[4870]: I1014 09:10:39.139853 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67dbcf57c9-fcxrn" event={"ID":"e5c27b78-9614-44c4-b96e-db0380508659","Type":"ContainerStarted","Data":"d1d37e3096c9038f26ea99eab30cac84d7cadc60c2912c801119f87ff00cce8a"} Oct 14 09:10:40 crc kubenswrapper[4870]: I1014 09:10:40.155638 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67dbcf57c9-fcxrn" event={"ID":"e5c27b78-9614-44c4-b96e-db0380508659","Type":"ContainerStarted","Data":"1574192c2fe8827792ae6b316917929bfed0d6936ba80ca02145feeb7a35ad93"} Oct 14 09:10:40 crc kubenswrapper[4870]: I1014 09:10:40.156105 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-67dbcf57c9-fcxrn" Oct 14 09:10:40 crc kubenswrapper[4870]: I1014 09:10:40.174223 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-67dbcf57c9-fcxrn" podStartSLOduration=3.174198618 podStartE2EDuration="3.174198618s" podCreationTimestamp="2025-10-14 09:10:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 09:10:40.172631549 +0000 UTC m=+7775.869991910" watchObservedRunningTime="2025-10-14 09:10:40.174198618 +0000 UTC m=+7775.871559019" Oct 14 09:10:47 crc kubenswrapper[4870]: I1014 09:10:47.813095 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-67dbcf57c9-fcxrn" Oct 14 09:10:47 crc kubenswrapper[4870]: I1014 09:10:47.932852 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-684cfb547c-6h8np"] Oct 14 09:10:47 crc kubenswrapper[4870]: I1014 09:10:47.933476 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-684cfb547c-6h8np" podUID="8657fe11-9fab-465a-bc32-a62447369e2c" containerName="dnsmasq-dns" containerID="cri-o://0746b0c80c90ad65201147e7953ca31dcb0355ac914276d6b459b1ab8ccd9734" gracePeriod=10 Oct 14 09:10:48 crc kubenswrapper[4870]: I1014 09:10:48.244831 4870 generic.go:334] "Generic (PLEG): container finished" podID="8657fe11-9fab-465a-bc32-a62447369e2c" containerID="0746b0c80c90ad65201147e7953ca31dcb0355ac914276d6b459b1ab8ccd9734" exitCode=0 Oct 14 09:10:48 crc kubenswrapper[4870]: I1014 09:10:48.244877 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-684cfb547c-6h8np" event={"ID":"8657fe11-9fab-465a-bc32-a62447369e2c","Type":"ContainerDied","Data":"0746b0c80c90ad65201147e7953ca31dcb0355ac914276d6b459b1ab8ccd9734"} Oct 14 09:10:48 crc kubenswrapper[4870]: I1014 09:10:48.490526 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-684cfb547c-6h8np" Oct 14 09:10:48 crc kubenswrapper[4870]: I1014 09:10:48.626299 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8657fe11-9fab-465a-bc32-a62447369e2c-dns-svc\") pod \"8657fe11-9fab-465a-bc32-a62447369e2c\" (UID: \"8657fe11-9fab-465a-bc32-a62447369e2c\") " Oct 14 09:10:48 crc kubenswrapper[4870]: I1014 09:10:48.626351 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8657fe11-9fab-465a-bc32-a62447369e2c-ovsdbserver-nb\") pod \"8657fe11-9fab-465a-bc32-a62447369e2c\" (UID: \"8657fe11-9fab-465a-bc32-a62447369e2c\") " Oct 14 09:10:48 crc kubenswrapper[4870]: I1014 09:10:48.626403 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/8657fe11-9fab-465a-bc32-a62447369e2c-openstack-cell1\") pod \"8657fe11-9fab-465a-bc32-a62447369e2c\" (UID: \"8657fe11-9fab-465a-bc32-a62447369e2c\") " Oct 14 09:10:48 crc kubenswrapper[4870]: I1014 09:10:48.626487 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8657fe11-9fab-465a-bc32-a62447369e2c-config\") pod \"8657fe11-9fab-465a-bc32-a62447369e2c\" (UID: \"8657fe11-9fab-465a-bc32-a62447369e2c\") " Oct 14 09:10:48 crc kubenswrapper[4870]: I1014 09:10:48.626505 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8657fe11-9fab-465a-bc32-a62447369e2c-ovsdbserver-sb\") pod \"8657fe11-9fab-465a-bc32-a62447369e2c\" (UID: \"8657fe11-9fab-465a-bc32-a62447369e2c\") " Oct 14 09:10:48 crc kubenswrapper[4870]: I1014 09:10:48.626746 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jjd4q\" (UniqueName: \"kubernetes.io/projected/8657fe11-9fab-465a-bc32-a62447369e2c-kube-api-access-jjd4q\") pod \"8657fe11-9fab-465a-bc32-a62447369e2c\" (UID: \"8657fe11-9fab-465a-bc32-a62447369e2c\") " Oct 14 09:10:48 crc kubenswrapper[4870]: I1014 09:10:48.643790 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8657fe11-9fab-465a-bc32-a62447369e2c-kube-api-access-jjd4q" (OuterVolumeSpecName: "kube-api-access-jjd4q") pod "8657fe11-9fab-465a-bc32-a62447369e2c" (UID: "8657fe11-9fab-465a-bc32-a62447369e2c"). InnerVolumeSpecName "kube-api-access-jjd4q". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:10:48 crc kubenswrapper[4870]: I1014 09:10:48.696116 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8657fe11-9fab-465a-bc32-a62447369e2c-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "8657fe11-9fab-465a-bc32-a62447369e2c" (UID: "8657fe11-9fab-465a-bc32-a62447369e2c"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 09:10:48 crc kubenswrapper[4870]: I1014 09:10:48.697953 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8657fe11-9fab-465a-bc32-a62447369e2c-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "8657fe11-9fab-465a-bc32-a62447369e2c" (UID: "8657fe11-9fab-465a-bc32-a62447369e2c"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 09:10:48 crc kubenswrapper[4870]: I1014 09:10:48.701145 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8657fe11-9fab-465a-bc32-a62447369e2c-openstack-cell1" (OuterVolumeSpecName: "openstack-cell1") pod "8657fe11-9fab-465a-bc32-a62447369e2c" (UID: "8657fe11-9fab-465a-bc32-a62447369e2c"). InnerVolumeSpecName "openstack-cell1". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 09:10:48 crc kubenswrapper[4870]: I1014 09:10:48.706101 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8657fe11-9fab-465a-bc32-a62447369e2c-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "8657fe11-9fab-465a-bc32-a62447369e2c" (UID: "8657fe11-9fab-465a-bc32-a62447369e2c"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 09:10:48 crc kubenswrapper[4870]: I1014 09:10:48.729877 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jjd4q\" (UniqueName: \"kubernetes.io/projected/8657fe11-9fab-465a-bc32-a62447369e2c-kube-api-access-jjd4q\") on node \"crc\" DevicePath \"\"" Oct 14 09:10:48 crc kubenswrapper[4870]: I1014 09:10:48.729929 4870 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8657fe11-9fab-465a-bc32-a62447369e2c-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 14 09:10:48 crc kubenswrapper[4870]: I1014 09:10:48.729939 4870 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8657fe11-9fab-465a-bc32-a62447369e2c-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 14 09:10:48 crc kubenswrapper[4870]: I1014 09:10:48.729950 4870 reconciler_common.go:293] "Volume detached for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/8657fe11-9fab-465a-bc32-a62447369e2c-openstack-cell1\") on node \"crc\" DevicePath \"\"" Oct 14 09:10:48 crc kubenswrapper[4870]: I1014 09:10:48.729959 4870 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8657fe11-9fab-465a-bc32-a62447369e2c-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 14 09:10:48 crc kubenswrapper[4870]: I1014 09:10:48.731733 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8657fe11-9fab-465a-bc32-a62447369e2c-config" (OuterVolumeSpecName: "config") pod "8657fe11-9fab-465a-bc32-a62447369e2c" (UID: "8657fe11-9fab-465a-bc32-a62447369e2c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 09:10:48 crc kubenswrapper[4870]: I1014 09:10:48.831172 4870 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8657fe11-9fab-465a-bc32-a62447369e2c-config\") on node \"crc\" DevicePath \"\"" Oct 14 09:10:49 crc kubenswrapper[4870]: I1014 09:10:49.270729 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-684cfb547c-6h8np" event={"ID":"8657fe11-9fab-465a-bc32-a62447369e2c","Type":"ContainerDied","Data":"0d0a0b8d62735684c49f161b44e7687d7843a16ecd04c85721db15a70ed54b75"} Oct 14 09:10:49 crc kubenswrapper[4870]: I1014 09:10:49.271076 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-684cfb547c-6h8np" Oct 14 09:10:49 crc kubenswrapper[4870]: I1014 09:10:49.271117 4870 scope.go:117] "RemoveContainer" containerID="0746b0c80c90ad65201147e7953ca31dcb0355ac914276d6b459b1ab8ccd9734" Oct 14 09:10:49 crc kubenswrapper[4870]: I1014 09:10:49.301700 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-684cfb547c-6h8np"] Oct 14 09:10:49 crc kubenswrapper[4870]: I1014 09:10:49.309131 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-684cfb547c-6h8np"] Oct 14 09:10:49 crc kubenswrapper[4870]: I1014 09:10:49.313946 4870 scope.go:117] "RemoveContainer" containerID="c8978b1f8aab207381f5916eca4944077ab5b450b57259db580bfd4c248e3577" Oct 14 09:10:51 crc kubenswrapper[4870]: I1014 09:10:51.045171 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8657fe11-9fab-465a-bc32-a62447369e2c" path="/var/lib/kubelet/pods/8657fe11-9fab-465a-bc32-a62447369e2c/volumes" Oct 14 09:10:55 crc kubenswrapper[4870]: I1014 09:10:55.061671 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-c7pv6"] Oct 14 09:10:55 crc kubenswrapper[4870]: I1014 09:10:55.071887 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-c7pv6"] Oct 14 09:10:57 crc kubenswrapper[4870]: I1014 09:10:57.055825 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="651aa59d-e6a6-48b2-806d-8a2e2531814f" path="/var/lib/kubelet/pods/651aa59d-e6a6-48b2-806d-8a2e2531814f/volumes" Oct 14 09:11:03 crc kubenswrapper[4870]: I1014 09:11:03.085231 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c6xncd"] Oct 14 09:11:03 crc kubenswrapper[4870]: E1014 09:11:03.086727 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6e3771b-e2bb-47d0-8d71-4b8c34d0a786" containerName="dnsmasq-dns" Oct 14 09:11:03 crc kubenswrapper[4870]: I1014 09:11:03.086752 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6e3771b-e2bb-47d0-8d71-4b8c34d0a786" containerName="dnsmasq-dns" Oct 14 09:11:03 crc kubenswrapper[4870]: E1014 09:11:03.086797 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6e3771b-e2bb-47d0-8d71-4b8c34d0a786" containerName="init" Oct 14 09:11:03 crc kubenswrapper[4870]: I1014 09:11:03.086810 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6e3771b-e2bb-47d0-8d71-4b8c34d0a786" containerName="init" Oct 14 09:11:03 crc kubenswrapper[4870]: E1014 09:11:03.086834 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8657fe11-9fab-465a-bc32-a62447369e2c" containerName="dnsmasq-dns" Oct 14 09:11:03 crc kubenswrapper[4870]: I1014 09:11:03.086846 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="8657fe11-9fab-465a-bc32-a62447369e2c" containerName="dnsmasq-dns" Oct 14 09:11:03 crc kubenswrapper[4870]: E1014 09:11:03.086869 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8657fe11-9fab-465a-bc32-a62447369e2c" containerName="init" Oct 14 09:11:03 crc kubenswrapper[4870]: I1014 09:11:03.086881 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="8657fe11-9fab-465a-bc32-a62447369e2c" containerName="init" Oct 14 09:11:03 crc kubenswrapper[4870]: I1014 09:11:03.087212 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="d6e3771b-e2bb-47d0-8d71-4b8c34d0a786" containerName="dnsmasq-dns" Oct 14 09:11:03 crc kubenswrapper[4870]: I1014 09:11:03.087245 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="8657fe11-9fab-465a-bc32-a62447369e2c" containerName="dnsmasq-dns" Oct 14 09:11:03 crc kubenswrapper[4870]: I1014 09:11:03.088674 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c6xncd" Oct 14 09:11:03 crc kubenswrapper[4870]: I1014 09:11:03.093854 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 14 09:11:03 crc kubenswrapper[4870]: I1014 09:11:03.094423 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 14 09:11:03 crc kubenswrapper[4870]: I1014 09:11:03.094552 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-wrz7w" Oct 14 09:11:03 crc kubenswrapper[4870]: I1014 09:11:03.097015 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 14 09:11:03 crc kubenswrapper[4870]: I1014 09:11:03.143739 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/pre-adoption-validation-openstack-pre-adoption-openstack-nwfcbm"] Oct 14 09:11:03 crc kubenswrapper[4870]: I1014 09:11:03.146758 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-nwfcbm" Oct 14 09:11:03 crc kubenswrapper[4870]: I1014 09:11:03.148752 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-networker" Oct 14 09:11:03 crc kubenswrapper[4870]: I1014 09:11:03.149498 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-networker-dockercfg-54hzr" Oct 14 09:11:03 crc kubenswrapper[4870]: I1014 09:11:03.159998 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c6xncd"] Oct 14 09:11:03 crc kubenswrapper[4870]: I1014 09:11:03.176832 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/pre-adoption-validation-openstack-pre-adoption-openstack-nwfcbm"] Oct 14 09:11:03 crc kubenswrapper[4870]: I1014 09:11:03.234897 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2233d692-f4ea-4390-adde-275dab829e73-inventory\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-c6xncd\" (UID: \"2233d692-f4ea-4390-adde-275dab829e73\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c6xncd" Oct 14 09:11:03 crc kubenswrapper[4870]: I1014 09:11:03.235011 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pre-adoption-validation-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2233d692-f4ea-4390-adde-275dab829e73-pre-adoption-validation-combined-ca-bundle\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-c6xncd\" (UID: \"2233d692-f4ea-4390-adde-275dab829e73\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c6xncd" Oct 14 09:11:03 crc kubenswrapper[4870]: I1014 09:11:03.235195 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/2233d692-f4ea-4390-adde-275dab829e73-ceph\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-c6xncd\" (UID: \"2233d692-f4ea-4390-adde-275dab829e73\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c6xncd" Oct 14 09:11:03 crc kubenswrapper[4870]: I1014 09:11:03.235247 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2233d692-f4ea-4390-adde-275dab829e73-ssh-key\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-c6xncd\" (UID: \"2233d692-f4ea-4390-adde-275dab829e73\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c6xncd" Oct 14 09:11:03 crc kubenswrapper[4870]: I1014 09:11:03.235329 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sbm65\" (UniqueName: \"kubernetes.io/projected/2233d692-f4ea-4390-adde-275dab829e73-kube-api-access-sbm65\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-c6xncd\" (UID: \"2233d692-f4ea-4390-adde-275dab829e73\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c6xncd" Oct 14 09:11:03 crc kubenswrapper[4870]: I1014 09:11:03.337041 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/364d6666-a8e2-48f6-b42d-3d676de7f45a-inventory\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-nwfcbm\" (UID: \"364d6666-a8e2-48f6-b42d-3d676de7f45a\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-nwfcbm" Oct 14 09:11:03 crc kubenswrapper[4870]: I1014 09:11:03.337984 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2233d692-f4ea-4390-adde-275dab829e73-inventory\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-c6xncd\" (UID: \"2233d692-f4ea-4390-adde-275dab829e73\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c6xncd" Oct 14 09:11:03 crc kubenswrapper[4870]: I1014 09:11:03.338065 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pre-adoption-validation-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2233d692-f4ea-4390-adde-275dab829e73-pre-adoption-validation-combined-ca-bundle\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-c6xncd\" (UID: \"2233d692-f4ea-4390-adde-275dab829e73\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c6xncd" Oct 14 09:11:03 crc kubenswrapper[4870]: I1014 09:11:03.338104 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pre-adoption-validation-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/364d6666-a8e2-48f6-b42d-3d676de7f45a-pre-adoption-validation-combined-ca-bundle\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-nwfcbm\" (UID: \"364d6666-a8e2-48f6-b42d-3d676de7f45a\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-nwfcbm" Oct 14 09:11:03 crc kubenswrapper[4870]: I1014 09:11:03.338160 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/364d6666-a8e2-48f6-b42d-3d676de7f45a-ssh-key\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-nwfcbm\" (UID: \"364d6666-a8e2-48f6-b42d-3d676de7f45a\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-nwfcbm" Oct 14 09:11:03 crc kubenswrapper[4870]: I1014 09:11:03.338189 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/2233d692-f4ea-4390-adde-275dab829e73-ceph\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-c6xncd\" (UID: \"2233d692-f4ea-4390-adde-275dab829e73\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c6xncd" Oct 14 09:11:03 crc kubenswrapper[4870]: I1014 09:11:03.338221 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2233d692-f4ea-4390-adde-275dab829e73-ssh-key\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-c6xncd\" (UID: \"2233d692-f4ea-4390-adde-275dab829e73\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c6xncd" Oct 14 09:11:03 crc kubenswrapper[4870]: I1014 09:11:03.338288 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xsnsw\" (UniqueName: \"kubernetes.io/projected/364d6666-a8e2-48f6-b42d-3d676de7f45a-kube-api-access-xsnsw\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-nwfcbm\" (UID: \"364d6666-a8e2-48f6-b42d-3d676de7f45a\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-nwfcbm" Oct 14 09:11:03 crc kubenswrapper[4870]: I1014 09:11:03.338321 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sbm65\" (UniqueName: \"kubernetes.io/projected/2233d692-f4ea-4390-adde-275dab829e73-kube-api-access-sbm65\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-c6xncd\" (UID: \"2233d692-f4ea-4390-adde-275dab829e73\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c6xncd" Oct 14 09:11:03 crc kubenswrapper[4870]: I1014 09:11:03.344820 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pre-adoption-validation-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2233d692-f4ea-4390-adde-275dab829e73-pre-adoption-validation-combined-ca-bundle\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-c6xncd\" (UID: \"2233d692-f4ea-4390-adde-275dab829e73\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c6xncd" Oct 14 09:11:03 crc kubenswrapper[4870]: I1014 09:11:03.344947 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2233d692-f4ea-4390-adde-275dab829e73-inventory\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-c6xncd\" (UID: \"2233d692-f4ea-4390-adde-275dab829e73\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c6xncd" Oct 14 09:11:03 crc kubenswrapper[4870]: I1014 09:11:03.348065 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2233d692-f4ea-4390-adde-275dab829e73-ssh-key\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-c6xncd\" (UID: \"2233d692-f4ea-4390-adde-275dab829e73\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c6xncd" Oct 14 09:11:03 crc kubenswrapper[4870]: I1014 09:11:03.348567 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/2233d692-f4ea-4390-adde-275dab829e73-ceph\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-c6xncd\" (UID: \"2233d692-f4ea-4390-adde-275dab829e73\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c6xncd" Oct 14 09:11:03 crc kubenswrapper[4870]: I1014 09:11:03.372322 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sbm65\" (UniqueName: \"kubernetes.io/projected/2233d692-f4ea-4390-adde-275dab829e73-kube-api-access-sbm65\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-c6xncd\" (UID: \"2233d692-f4ea-4390-adde-275dab829e73\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c6xncd" Oct 14 09:11:03 crc kubenswrapper[4870]: I1014 09:11:03.440946 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xsnsw\" (UniqueName: \"kubernetes.io/projected/364d6666-a8e2-48f6-b42d-3d676de7f45a-kube-api-access-xsnsw\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-nwfcbm\" (UID: \"364d6666-a8e2-48f6-b42d-3d676de7f45a\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-nwfcbm" Oct 14 09:11:03 crc kubenswrapper[4870]: I1014 09:11:03.441083 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/364d6666-a8e2-48f6-b42d-3d676de7f45a-inventory\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-nwfcbm\" (UID: \"364d6666-a8e2-48f6-b42d-3d676de7f45a\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-nwfcbm" Oct 14 09:11:03 crc kubenswrapper[4870]: I1014 09:11:03.441272 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pre-adoption-validation-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/364d6666-a8e2-48f6-b42d-3d676de7f45a-pre-adoption-validation-combined-ca-bundle\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-nwfcbm\" (UID: \"364d6666-a8e2-48f6-b42d-3d676de7f45a\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-nwfcbm" Oct 14 09:11:03 crc kubenswrapper[4870]: I1014 09:11:03.441350 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/364d6666-a8e2-48f6-b42d-3d676de7f45a-ssh-key\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-nwfcbm\" (UID: \"364d6666-a8e2-48f6-b42d-3d676de7f45a\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-nwfcbm" Oct 14 09:11:03 crc kubenswrapper[4870]: I1014 09:11:03.445290 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/364d6666-a8e2-48f6-b42d-3d676de7f45a-ssh-key\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-nwfcbm\" (UID: \"364d6666-a8e2-48f6-b42d-3d676de7f45a\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-nwfcbm" Oct 14 09:11:03 crc kubenswrapper[4870]: I1014 09:11:03.446006 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/364d6666-a8e2-48f6-b42d-3d676de7f45a-inventory\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-nwfcbm\" (UID: \"364d6666-a8e2-48f6-b42d-3d676de7f45a\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-nwfcbm" Oct 14 09:11:03 crc kubenswrapper[4870]: I1014 09:11:03.446238 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pre-adoption-validation-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/364d6666-a8e2-48f6-b42d-3d676de7f45a-pre-adoption-validation-combined-ca-bundle\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-nwfcbm\" (UID: \"364d6666-a8e2-48f6-b42d-3d676de7f45a\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-nwfcbm" Oct 14 09:11:03 crc kubenswrapper[4870]: I1014 09:11:03.468883 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xsnsw\" (UniqueName: \"kubernetes.io/projected/364d6666-a8e2-48f6-b42d-3d676de7f45a-kube-api-access-xsnsw\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-nwfcbm\" (UID: \"364d6666-a8e2-48f6-b42d-3d676de7f45a\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-nwfcbm" Oct 14 09:11:03 crc kubenswrapper[4870]: I1014 09:11:03.477331 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c6xncd" Oct 14 09:11:03 crc kubenswrapper[4870]: I1014 09:11:03.482751 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-nwfcbm" Oct 14 09:11:04 crc kubenswrapper[4870]: I1014 09:11:04.134728 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/pre-adoption-validation-openstack-pre-adoption-openstack-nwfcbm"] Oct 14 09:11:04 crc kubenswrapper[4870]: I1014 09:11:04.204950 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c6xncd"] Oct 14 09:11:04 crc kubenswrapper[4870]: W1014 09:11:04.206544 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2233d692_f4ea_4390_adde_275dab829e73.slice/crio-c955a61d8eb37b91098416d0ef75f9727d262564b2d7b04bdcf41efdf2864d86 WatchSource:0}: Error finding container c955a61d8eb37b91098416d0ef75f9727d262564b2d7b04bdcf41efdf2864d86: Status 404 returned error can't find the container with id c955a61d8eb37b91098416d0ef75f9727d262564b2d7b04bdcf41efdf2864d86 Oct 14 09:11:04 crc kubenswrapper[4870]: I1014 09:11:04.473911 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-nwfcbm" event={"ID":"364d6666-a8e2-48f6-b42d-3d676de7f45a","Type":"ContainerStarted","Data":"40a695d179cb845ff95daed8c6385110fc55e662fef20c35afa65cec0267c20d"} Oct 14 09:11:04 crc kubenswrapper[4870]: I1014 09:11:04.475385 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c6xncd" event={"ID":"2233d692-f4ea-4390-adde-275dab829e73","Type":"ContainerStarted","Data":"c955a61d8eb37b91098416d0ef75f9727d262564b2d7b04bdcf41efdf2864d86"} Oct 14 09:11:06 crc kubenswrapper[4870]: I1014 09:11:06.027577 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-fafa-account-create-758c7"] Oct 14 09:11:06 crc kubenswrapper[4870]: I1014 09:11:06.039425 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-fafa-account-create-758c7"] Oct 14 09:11:07 crc kubenswrapper[4870]: I1014 09:11:07.065287 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ad2fe341-5486-4fec-b744-e294b5d034bc" path="/var/lib/kubelet/pods/ad2fe341-5486-4fec-b744-e294b5d034bc/volumes" Oct 14 09:11:13 crc kubenswrapper[4870]: I1014 09:11:13.473551 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 14 09:11:14 crc kubenswrapper[4870]: I1014 09:11:14.594638 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-nwfcbm" event={"ID":"364d6666-a8e2-48f6-b42d-3d676de7f45a","Type":"ContainerStarted","Data":"a26fe1c83606e58cbaffcb847ca612e13c34121249b447fd44f885895044c40d"} Oct 14 09:11:14 crc kubenswrapper[4870]: I1014 09:11:14.597288 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c6xncd" event={"ID":"2233d692-f4ea-4390-adde-275dab829e73","Type":"ContainerStarted","Data":"79d960d37554864403d73d8334bb1c79016bf1ddfa20c7f3a801ff3b02bc3c35"} Oct 14 09:11:14 crc kubenswrapper[4870]: I1014 09:11:14.657598 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-nwfcbm" podStartSLOduration=2.3281728360000002 podStartE2EDuration="11.656593594s" podCreationTimestamp="2025-10-14 09:11:03 +0000 UTC" firstStartedPulling="2025-10-14 09:11:04.142515031 +0000 UTC m=+7799.839875412" lastFinishedPulling="2025-10-14 09:11:13.470935789 +0000 UTC m=+7809.168296170" observedRunningTime="2025-10-14 09:11:14.646655689 +0000 UTC m=+7810.344016060" watchObservedRunningTime="2025-10-14 09:11:14.656593594 +0000 UTC m=+7810.353954005" Oct 14 09:11:23 crc kubenswrapper[4870]: I1014 09:11:23.709305 4870 generic.go:334] "Generic (PLEG): container finished" podID="364d6666-a8e2-48f6-b42d-3d676de7f45a" containerID="a26fe1c83606e58cbaffcb847ca612e13c34121249b447fd44f885895044c40d" exitCode=0 Oct 14 09:11:23 crc kubenswrapper[4870]: I1014 09:11:23.709462 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-nwfcbm" event={"ID":"364d6666-a8e2-48f6-b42d-3d676de7f45a","Type":"ContainerDied","Data":"a26fe1c83606e58cbaffcb847ca612e13c34121249b447fd44f885895044c40d"} Oct 14 09:11:23 crc kubenswrapper[4870]: I1014 09:11:23.743314 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c6xncd" podStartSLOduration=11.465990388 podStartE2EDuration="20.743282492s" podCreationTimestamp="2025-10-14 09:11:03 +0000 UTC" firstStartedPulling="2025-10-14 09:11:04.208892041 +0000 UTC m=+7799.906252412" lastFinishedPulling="2025-10-14 09:11:13.486184145 +0000 UTC m=+7809.183544516" observedRunningTime="2025-10-14 09:11:14.692790838 +0000 UTC m=+7810.390151199" watchObservedRunningTime="2025-10-14 09:11:23.743282492 +0000 UTC m=+7819.440642903" Oct 14 09:11:25 crc kubenswrapper[4870]: I1014 09:11:25.278962 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-nwfcbm" Oct 14 09:11:25 crc kubenswrapper[4870]: I1014 09:11:25.431360 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/364d6666-a8e2-48f6-b42d-3d676de7f45a-inventory\") pod \"364d6666-a8e2-48f6-b42d-3d676de7f45a\" (UID: \"364d6666-a8e2-48f6-b42d-3d676de7f45a\") " Oct 14 09:11:25 crc kubenswrapper[4870]: I1014 09:11:25.431474 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/364d6666-a8e2-48f6-b42d-3d676de7f45a-ssh-key\") pod \"364d6666-a8e2-48f6-b42d-3d676de7f45a\" (UID: \"364d6666-a8e2-48f6-b42d-3d676de7f45a\") " Oct 14 09:11:25 crc kubenswrapper[4870]: I1014 09:11:25.431572 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pre-adoption-validation-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/364d6666-a8e2-48f6-b42d-3d676de7f45a-pre-adoption-validation-combined-ca-bundle\") pod \"364d6666-a8e2-48f6-b42d-3d676de7f45a\" (UID: \"364d6666-a8e2-48f6-b42d-3d676de7f45a\") " Oct 14 09:11:25 crc kubenswrapper[4870]: I1014 09:11:25.431731 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xsnsw\" (UniqueName: \"kubernetes.io/projected/364d6666-a8e2-48f6-b42d-3d676de7f45a-kube-api-access-xsnsw\") pod \"364d6666-a8e2-48f6-b42d-3d676de7f45a\" (UID: \"364d6666-a8e2-48f6-b42d-3d676de7f45a\") " Oct 14 09:11:25 crc kubenswrapper[4870]: I1014 09:11:25.439076 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/364d6666-a8e2-48f6-b42d-3d676de7f45a-pre-adoption-validation-combined-ca-bundle" (OuterVolumeSpecName: "pre-adoption-validation-combined-ca-bundle") pod "364d6666-a8e2-48f6-b42d-3d676de7f45a" (UID: "364d6666-a8e2-48f6-b42d-3d676de7f45a"). InnerVolumeSpecName "pre-adoption-validation-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:11:25 crc kubenswrapper[4870]: I1014 09:11:25.442722 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/364d6666-a8e2-48f6-b42d-3d676de7f45a-kube-api-access-xsnsw" (OuterVolumeSpecName: "kube-api-access-xsnsw") pod "364d6666-a8e2-48f6-b42d-3d676de7f45a" (UID: "364d6666-a8e2-48f6-b42d-3d676de7f45a"). InnerVolumeSpecName "kube-api-access-xsnsw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:11:25 crc kubenswrapper[4870]: I1014 09:11:25.465903 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/364d6666-a8e2-48f6-b42d-3d676de7f45a-inventory" (OuterVolumeSpecName: "inventory") pod "364d6666-a8e2-48f6-b42d-3d676de7f45a" (UID: "364d6666-a8e2-48f6-b42d-3d676de7f45a"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:11:25 crc kubenswrapper[4870]: I1014 09:11:25.480371 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/364d6666-a8e2-48f6-b42d-3d676de7f45a-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "364d6666-a8e2-48f6-b42d-3d676de7f45a" (UID: "364d6666-a8e2-48f6-b42d-3d676de7f45a"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:11:25 crc kubenswrapper[4870]: I1014 09:11:25.535827 4870 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/364d6666-a8e2-48f6-b42d-3d676de7f45a-inventory\") on node \"crc\" DevicePath \"\"" Oct 14 09:11:25 crc kubenswrapper[4870]: I1014 09:11:25.535895 4870 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/364d6666-a8e2-48f6-b42d-3d676de7f45a-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 14 09:11:25 crc kubenswrapper[4870]: I1014 09:11:25.535907 4870 reconciler_common.go:293] "Volume detached for volume \"pre-adoption-validation-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/364d6666-a8e2-48f6-b42d-3d676de7f45a-pre-adoption-validation-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 09:11:25 crc kubenswrapper[4870]: I1014 09:11:25.535926 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xsnsw\" (UniqueName: \"kubernetes.io/projected/364d6666-a8e2-48f6-b42d-3d676de7f45a-kube-api-access-xsnsw\") on node \"crc\" DevicePath \"\"" Oct 14 09:11:25 crc kubenswrapper[4870]: I1014 09:11:25.744556 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-nwfcbm" Oct 14 09:11:25 crc kubenswrapper[4870]: I1014 09:11:25.744573 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-nwfcbm" event={"ID":"364d6666-a8e2-48f6-b42d-3d676de7f45a","Type":"ContainerDied","Data":"40a695d179cb845ff95daed8c6385110fc55e662fef20c35afa65cec0267c20d"} Oct 14 09:11:25 crc kubenswrapper[4870]: I1014 09:11:25.744640 4870 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="40a695d179cb845ff95daed8c6385110fc55e662fef20c35afa65cec0267c20d" Oct 14 09:11:25 crc kubenswrapper[4870]: I1014 09:11:25.747992 4870 generic.go:334] "Generic (PLEG): container finished" podID="2233d692-f4ea-4390-adde-275dab829e73" containerID="79d960d37554864403d73d8334bb1c79016bf1ddfa20c7f3a801ff3b02bc3c35" exitCode=0 Oct 14 09:11:25 crc kubenswrapper[4870]: I1014 09:11:25.748074 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c6xncd" event={"ID":"2233d692-f4ea-4390-adde-275dab829e73","Type":"ContainerDied","Data":"79d960d37554864403d73d8334bb1c79016bf1ddfa20c7f3a801ff3b02bc3c35"} Oct 14 09:11:27 crc kubenswrapper[4870]: I1014 09:11:27.262942 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c6xncd" Oct 14 09:11:27 crc kubenswrapper[4870]: I1014 09:11:27.377199 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2233d692-f4ea-4390-adde-275dab829e73-ssh-key\") pod \"2233d692-f4ea-4390-adde-275dab829e73\" (UID: \"2233d692-f4ea-4390-adde-275dab829e73\") " Oct 14 09:11:27 crc kubenswrapper[4870]: I1014 09:11:27.377318 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pre-adoption-validation-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2233d692-f4ea-4390-adde-275dab829e73-pre-adoption-validation-combined-ca-bundle\") pod \"2233d692-f4ea-4390-adde-275dab829e73\" (UID: \"2233d692-f4ea-4390-adde-275dab829e73\") " Oct 14 09:11:27 crc kubenswrapper[4870]: I1014 09:11:27.377478 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sbm65\" (UniqueName: \"kubernetes.io/projected/2233d692-f4ea-4390-adde-275dab829e73-kube-api-access-sbm65\") pod \"2233d692-f4ea-4390-adde-275dab829e73\" (UID: \"2233d692-f4ea-4390-adde-275dab829e73\") " Oct 14 09:11:27 crc kubenswrapper[4870]: I1014 09:11:27.377550 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/2233d692-f4ea-4390-adde-275dab829e73-ceph\") pod \"2233d692-f4ea-4390-adde-275dab829e73\" (UID: \"2233d692-f4ea-4390-adde-275dab829e73\") " Oct 14 09:11:27 crc kubenswrapper[4870]: I1014 09:11:27.377639 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2233d692-f4ea-4390-adde-275dab829e73-inventory\") pod \"2233d692-f4ea-4390-adde-275dab829e73\" (UID: \"2233d692-f4ea-4390-adde-275dab829e73\") " Oct 14 09:11:27 crc kubenswrapper[4870]: I1014 09:11:27.383370 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2233d692-f4ea-4390-adde-275dab829e73-ceph" (OuterVolumeSpecName: "ceph") pod "2233d692-f4ea-4390-adde-275dab829e73" (UID: "2233d692-f4ea-4390-adde-275dab829e73"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:11:27 crc kubenswrapper[4870]: I1014 09:11:27.383659 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2233d692-f4ea-4390-adde-275dab829e73-pre-adoption-validation-combined-ca-bundle" (OuterVolumeSpecName: "pre-adoption-validation-combined-ca-bundle") pod "2233d692-f4ea-4390-adde-275dab829e73" (UID: "2233d692-f4ea-4390-adde-275dab829e73"). InnerVolumeSpecName "pre-adoption-validation-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:11:27 crc kubenswrapper[4870]: I1014 09:11:27.397716 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2233d692-f4ea-4390-adde-275dab829e73-kube-api-access-sbm65" (OuterVolumeSpecName: "kube-api-access-sbm65") pod "2233d692-f4ea-4390-adde-275dab829e73" (UID: "2233d692-f4ea-4390-adde-275dab829e73"). InnerVolumeSpecName "kube-api-access-sbm65". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:11:27 crc kubenswrapper[4870]: I1014 09:11:27.416840 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2233d692-f4ea-4390-adde-275dab829e73-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "2233d692-f4ea-4390-adde-275dab829e73" (UID: "2233d692-f4ea-4390-adde-275dab829e73"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:11:27 crc kubenswrapper[4870]: I1014 09:11:27.417346 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2233d692-f4ea-4390-adde-275dab829e73-inventory" (OuterVolumeSpecName: "inventory") pod "2233d692-f4ea-4390-adde-275dab829e73" (UID: "2233d692-f4ea-4390-adde-275dab829e73"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:11:27 crc kubenswrapper[4870]: I1014 09:11:27.479478 4870 reconciler_common.go:293] "Volume detached for volume \"pre-adoption-validation-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2233d692-f4ea-4390-adde-275dab829e73-pre-adoption-validation-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 09:11:27 crc kubenswrapper[4870]: I1014 09:11:27.479507 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sbm65\" (UniqueName: \"kubernetes.io/projected/2233d692-f4ea-4390-adde-275dab829e73-kube-api-access-sbm65\") on node \"crc\" DevicePath \"\"" Oct 14 09:11:27 crc kubenswrapper[4870]: I1014 09:11:27.479517 4870 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/2233d692-f4ea-4390-adde-275dab829e73-ceph\") on node \"crc\" DevicePath \"\"" Oct 14 09:11:27 crc kubenswrapper[4870]: I1014 09:11:27.479525 4870 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2233d692-f4ea-4390-adde-275dab829e73-inventory\") on node \"crc\" DevicePath \"\"" Oct 14 09:11:27 crc kubenswrapper[4870]: I1014 09:11:27.479536 4870 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2233d692-f4ea-4390-adde-275dab829e73-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 14 09:11:27 crc kubenswrapper[4870]: I1014 09:11:27.772327 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c6xncd" event={"ID":"2233d692-f4ea-4390-adde-275dab829e73","Type":"ContainerDied","Data":"c955a61d8eb37b91098416d0ef75f9727d262564b2d7b04bdcf41efdf2864d86"} Oct 14 09:11:27 crc kubenswrapper[4870]: I1014 09:11:27.772373 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c6xncd" Oct 14 09:11:27 crc kubenswrapper[4870]: I1014 09:11:27.772388 4870 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c955a61d8eb37b91098416d0ef75f9727d262564b2d7b04bdcf41efdf2864d86" Oct 14 09:11:31 crc kubenswrapper[4870]: I1014 09:11:31.291355 4870 scope.go:117] "RemoveContainer" containerID="7221c8a48ccda4afcf145ed2303bc545cf64e8407ec3f7f95bbf196913e791c2" Oct 14 09:11:31 crc kubenswrapper[4870]: I1014 09:11:31.353845 4870 scope.go:117] "RemoveContainer" containerID="5bff5d9b13861ce498c2037fffc163d2c8ac9a090f60d4678677965ee8fdfec9" Oct 14 09:11:31 crc kubenswrapper[4870]: I1014 09:11:31.353983 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-tz9gr"] Oct 14 09:11:31 crc kubenswrapper[4870]: E1014 09:11:31.354387 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="364d6666-a8e2-48f6-b42d-3d676de7f45a" containerName="pre-adoption-validation-openstack-pre-adoption-openstack-networ" Oct 14 09:11:31 crc kubenswrapper[4870]: I1014 09:11:31.354404 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="364d6666-a8e2-48f6-b42d-3d676de7f45a" containerName="pre-adoption-validation-openstack-pre-adoption-openstack-networ" Oct 14 09:11:31 crc kubenswrapper[4870]: E1014 09:11:31.354481 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2233d692-f4ea-4390-adde-275dab829e73" containerName="pre-adoption-validation-openstack-pre-adoption-openstack-cell1" Oct 14 09:11:31 crc kubenswrapper[4870]: I1014 09:11:31.354492 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="2233d692-f4ea-4390-adde-275dab829e73" containerName="pre-adoption-validation-openstack-pre-adoption-openstack-cell1" Oct 14 09:11:31 crc kubenswrapper[4870]: I1014 09:11:31.354690 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="364d6666-a8e2-48f6-b42d-3d676de7f45a" containerName="pre-adoption-validation-openstack-pre-adoption-openstack-networ" Oct 14 09:11:31 crc kubenswrapper[4870]: I1014 09:11:31.354720 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="2233d692-f4ea-4390-adde-275dab829e73" containerName="pre-adoption-validation-openstack-pre-adoption-openstack-cell1" Oct 14 09:11:31 crc kubenswrapper[4870]: I1014 09:11:31.355527 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-tz9gr" Oct 14 09:11:31 crc kubenswrapper[4870]: I1014 09:11:31.359371 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 14 09:11:31 crc kubenswrapper[4870]: I1014 09:11:31.359592 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 14 09:11:31 crc kubenswrapper[4870]: I1014 09:11:31.359786 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 14 09:11:31 crc kubenswrapper[4870]: I1014 09:11:31.360476 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-wrz7w" Oct 14 09:11:31 crc kubenswrapper[4870]: I1014 09:11:31.368490 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tripleo-cleanup-tripleo-cleanup-openstack-networker-ncc5q"] Oct 14 09:11:31 crc kubenswrapper[4870]: I1014 09:11:31.370240 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-networker-ncc5q" Oct 14 09:11:31 crc kubenswrapper[4870]: I1014 09:11:31.377391 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-networker" Oct 14 09:11:31 crc kubenswrapper[4870]: I1014 09:11:31.377735 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-networker-dockercfg-54hzr" Oct 14 09:11:31 crc kubenswrapper[4870]: I1014 09:11:31.382134 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-tz9gr"] Oct 14 09:11:31 crc kubenswrapper[4870]: I1014 09:11:31.460140 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tripleo-cleanup-tripleo-cleanup-openstack-networker-ncc5q"] Oct 14 09:11:31 crc kubenswrapper[4870]: I1014 09:11:31.481504 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k9q8g\" (UniqueName: \"kubernetes.io/projected/ea75b5d6-ee4b-4a8d-a3af-4995af7bafb9-kube-api-access-k9q8g\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-tz9gr\" (UID: \"ea75b5d6-ee4b-4a8d-a3af-4995af7bafb9\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-tz9gr" Oct 14 09:11:31 crc kubenswrapper[4870]: I1014 09:11:31.481583 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tripleo-cleanup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b96078e0-b889-4b47-9bec-a1411c1fec36-tripleo-cleanup-combined-ca-bundle\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-networker-ncc5q\" (UID: \"b96078e0-b889-4b47-9bec-a1411c1fec36\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-networker-ncc5q" Oct 14 09:11:31 crc kubenswrapper[4870]: I1014 09:11:31.481622 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ea75b5d6-ee4b-4a8d-a3af-4995af7bafb9-ceph\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-tz9gr\" (UID: \"ea75b5d6-ee4b-4a8d-a3af-4995af7bafb9\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-tz9gr" Oct 14 09:11:31 crc kubenswrapper[4870]: I1014 09:11:31.481658 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b96078e0-b889-4b47-9bec-a1411c1fec36-inventory\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-networker-ncc5q\" (UID: \"b96078e0-b889-4b47-9bec-a1411c1fec36\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-networker-ncc5q" Oct 14 09:11:31 crc kubenswrapper[4870]: I1014 09:11:31.481677 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ea75b5d6-ee4b-4a8d-a3af-4995af7bafb9-ssh-key\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-tz9gr\" (UID: \"ea75b5d6-ee4b-4a8d-a3af-4995af7bafb9\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-tz9gr" Oct 14 09:11:31 crc kubenswrapper[4870]: I1014 09:11:31.481708 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tripleo-cleanup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ea75b5d6-ee4b-4a8d-a3af-4995af7bafb9-tripleo-cleanup-combined-ca-bundle\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-tz9gr\" (UID: \"ea75b5d6-ee4b-4a8d-a3af-4995af7bafb9\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-tz9gr" Oct 14 09:11:31 crc kubenswrapper[4870]: I1014 09:11:31.481750 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ea75b5d6-ee4b-4a8d-a3af-4995af7bafb9-inventory\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-tz9gr\" (UID: \"ea75b5d6-ee4b-4a8d-a3af-4995af7bafb9\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-tz9gr" Oct 14 09:11:31 crc kubenswrapper[4870]: I1014 09:11:31.481788 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b96078e0-b889-4b47-9bec-a1411c1fec36-ssh-key\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-networker-ncc5q\" (UID: \"b96078e0-b889-4b47-9bec-a1411c1fec36\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-networker-ncc5q" Oct 14 09:11:31 crc kubenswrapper[4870]: I1014 09:11:31.481806 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hn9gp\" (UniqueName: \"kubernetes.io/projected/b96078e0-b889-4b47-9bec-a1411c1fec36-kube-api-access-hn9gp\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-networker-ncc5q\" (UID: \"b96078e0-b889-4b47-9bec-a1411c1fec36\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-networker-ncc5q" Oct 14 09:11:31 crc kubenswrapper[4870]: I1014 09:11:31.583548 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ea75b5d6-ee4b-4a8d-a3af-4995af7bafb9-inventory\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-tz9gr\" (UID: \"ea75b5d6-ee4b-4a8d-a3af-4995af7bafb9\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-tz9gr" Oct 14 09:11:31 crc kubenswrapper[4870]: I1014 09:11:31.583652 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b96078e0-b889-4b47-9bec-a1411c1fec36-ssh-key\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-networker-ncc5q\" (UID: \"b96078e0-b889-4b47-9bec-a1411c1fec36\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-networker-ncc5q" Oct 14 09:11:31 crc kubenswrapper[4870]: I1014 09:11:31.583676 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hn9gp\" (UniqueName: \"kubernetes.io/projected/b96078e0-b889-4b47-9bec-a1411c1fec36-kube-api-access-hn9gp\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-networker-ncc5q\" (UID: \"b96078e0-b889-4b47-9bec-a1411c1fec36\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-networker-ncc5q" Oct 14 09:11:31 crc kubenswrapper[4870]: I1014 09:11:31.583723 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k9q8g\" (UniqueName: \"kubernetes.io/projected/ea75b5d6-ee4b-4a8d-a3af-4995af7bafb9-kube-api-access-k9q8g\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-tz9gr\" (UID: \"ea75b5d6-ee4b-4a8d-a3af-4995af7bafb9\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-tz9gr" Oct 14 09:11:31 crc kubenswrapper[4870]: I1014 09:11:31.583775 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tripleo-cleanup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b96078e0-b889-4b47-9bec-a1411c1fec36-tripleo-cleanup-combined-ca-bundle\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-networker-ncc5q\" (UID: \"b96078e0-b889-4b47-9bec-a1411c1fec36\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-networker-ncc5q" Oct 14 09:11:31 crc kubenswrapper[4870]: I1014 09:11:31.583817 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ea75b5d6-ee4b-4a8d-a3af-4995af7bafb9-ceph\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-tz9gr\" (UID: \"ea75b5d6-ee4b-4a8d-a3af-4995af7bafb9\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-tz9gr" Oct 14 09:11:31 crc kubenswrapper[4870]: I1014 09:11:31.583851 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b96078e0-b889-4b47-9bec-a1411c1fec36-inventory\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-networker-ncc5q\" (UID: \"b96078e0-b889-4b47-9bec-a1411c1fec36\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-networker-ncc5q" Oct 14 09:11:31 crc kubenswrapper[4870]: I1014 09:11:31.583869 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ea75b5d6-ee4b-4a8d-a3af-4995af7bafb9-ssh-key\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-tz9gr\" (UID: \"ea75b5d6-ee4b-4a8d-a3af-4995af7bafb9\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-tz9gr" Oct 14 09:11:31 crc kubenswrapper[4870]: I1014 09:11:31.583904 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tripleo-cleanup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ea75b5d6-ee4b-4a8d-a3af-4995af7bafb9-tripleo-cleanup-combined-ca-bundle\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-tz9gr\" (UID: \"ea75b5d6-ee4b-4a8d-a3af-4995af7bafb9\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-tz9gr" Oct 14 09:11:31 crc kubenswrapper[4870]: I1014 09:11:31.589171 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b96078e0-b889-4b47-9bec-a1411c1fec36-ssh-key\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-networker-ncc5q\" (UID: \"b96078e0-b889-4b47-9bec-a1411c1fec36\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-networker-ncc5q" Oct 14 09:11:31 crc kubenswrapper[4870]: I1014 09:11:31.589219 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ea75b5d6-ee4b-4a8d-a3af-4995af7bafb9-ssh-key\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-tz9gr\" (UID: \"ea75b5d6-ee4b-4a8d-a3af-4995af7bafb9\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-tz9gr" Oct 14 09:11:31 crc kubenswrapper[4870]: I1014 09:11:31.601860 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tripleo-cleanup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ea75b5d6-ee4b-4a8d-a3af-4995af7bafb9-tripleo-cleanup-combined-ca-bundle\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-tz9gr\" (UID: \"ea75b5d6-ee4b-4a8d-a3af-4995af7bafb9\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-tz9gr" Oct 14 09:11:31 crc kubenswrapper[4870]: I1014 09:11:31.601938 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ea75b5d6-ee4b-4a8d-a3af-4995af7bafb9-ceph\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-tz9gr\" (UID: \"ea75b5d6-ee4b-4a8d-a3af-4995af7bafb9\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-tz9gr" Oct 14 09:11:31 crc kubenswrapper[4870]: I1014 09:11:31.601962 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b96078e0-b889-4b47-9bec-a1411c1fec36-inventory\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-networker-ncc5q\" (UID: \"b96078e0-b889-4b47-9bec-a1411c1fec36\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-networker-ncc5q" Oct 14 09:11:31 crc kubenswrapper[4870]: I1014 09:11:31.601974 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tripleo-cleanup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b96078e0-b889-4b47-9bec-a1411c1fec36-tripleo-cleanup-combined-ca-bundle\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-networker-ncc5q\" (UID: \"b96078e0-b889-4b47-9bec-a1411c1fec36\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-networker-ncc5q" Oct 14 09:11:31 crc kubenswrapper[4870]: I1014 09:11:31.602677 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ea75b5d6-ee4b-4a8d-a3af-4995af7bafb9-inventory\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-tz9gr\" (UID: \"ea75b5d6-ee4b-4a8d-a3af-4995af7bafb9\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-tz9gr" Oct 14 09:11:31 crc kubenswrapper[4870]: I1014 09:11:31.609186 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hn9gp\" (UniqueName: \"kubernetes.io/projected/b96078e0-b889-4b47-9bec-a1411c1fec36-kube-api-access-hn9gp\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-networker-ncc5q\" (UID: \"b96078e0-b889-4b47-9bec-a1411c1fec36\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-networker-ncc5q" Oct 14 09:11:31 crc kubenswrapper[4870]: I1014 09:11:31.624211 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k9q8g\" (UniqueName: \"kubernetes.io/projected/ea75b5d6-ee4b-4a8d-a3af-4995af7bafb9-kube-api-access-k9q8g\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-tz9gr\" (UID: \"ea75b5d6-ee4b-4a8d-a3af-4995af7bafb9\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-tz9gr" Oct 14 09:11:31 crc kubenswrapper[4870]: I1014 09:11:31.709186 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-tz9gr" Oct 14 09:11:31 crc kubenswrapper[4870]: I1014 09:11:31.757398 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-networker-ncc5q" Oct 14 09:11:32 crc kubenswrapper[4870]: W1014 09:11:32.264841 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podea75b5d6_ee4b_4a8d_a3af_4995af7bafb9.slice/crio-8d60c257e44302f2a6d45852ba3f8354e60efdf9992de5d7fefb780858fa094a WatchSource:0}: Error finding container 8d60c257e44302f2a6d45852ba3f8354e60efdf9992de5d7fefb780858fa094a: Status 404 returned error can't find the container with id 8d60c257e44302f2a6d45852ba3f8354e60efdf9992de5d7fefb780858fa094a Oct 14 09:11:32 crc kubenswrapper[4870]: I1014 09:11:32.272298 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-tz9gr"] Oct 14 09:11:32 crc kubenswrapper[4870]: I1014 09:11:32.381934 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tripleo-cleanup-tripleo-cleanup-openstack-networker-ncc5q"] Oct 14 09:11:32 crc kubenswrapper[4870]: I1014 09:11:32.880988 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-networker-ncc5q" event={"ID":"b96078e0-b889-4b47-9bec-a1411c1fec36","Type":"ContainerStarted","Data":"ad2cb4f30a22bda7e820971b083bebbfc50e68fe1e135cbebd8a3f42a824bee1"} Oct 14 09:11:32 crc kubenswrapper[4870]: I1014 09:11:32.883392 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-tz9gr" event={"ID":"ea75b5d6-ee4b-4a8d-a3af-4995af7bafb9","Type":"ContainerStarted","Data":"8d60c257e44302f2a6d45852ba3f8354e60efdf9992de5d7fefb780858fa094a"} Oct 14 09:11:33 crc kubenswrapper[4870]: I1014 09:11:33.055180 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-x7dzz"] Oct 14 09:11:33 crc kubenswrapper[4870]: I1014 09:11:33.072792 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-x7dzz"] Oct 14 09:11:33 crc kubenswrapper[4870]: I1014 09:11:33.896049 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-tz9gr" event={"ID":"ea75b5d6-ee4b-4a8d-a3af-4995af7bafb9","Type":"ContainerStarted","Data":"bbcde4b193c5c8d0fc536d73ce644073d25ca7bbc4e4cad9878a8b687c145771"} Oct 14 09:11:33 crc kubenswrapper[4870]: I1014 09:11:33.898969 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-networker-ncc5q" event={"ID":"b96078e0-b889-4b47-9bec-a1411c1fec36","Type":"ContainerStarted","Data":"e0ef696961602aad945c2246f047575f06c8ba9512e813739a1d582140fb2a72"} Oct 14 09:11:33 crc kubenswrapper[4870]: I1014 09:11:33.934472 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-tz9gr" podStartSLOduration=2.056968934 podStartE2EDuration="2.934428111s" podCreationTimestamp="2025-10-14 09:11:31 +0000 UTC" firstStartedPulling="2025-10-14 09:11:32.266885065 +0000 UTC m=+7827.964245436" lastFinishedPulling="2025-10-14 09:11:33.144344242 +0000 UTC m=+7828.841704613" observedRunningTime="2025-10-14 09:11:33.926656629 +0000 UTC m=+7829.624017040" watchObservedRunningTime="2025-10-14 09:11:33.934428111 +0000 UTC m=+7829.631788502" Oct 14 09:11:33 crc kubenswrapper[4870]: I1014 09:11:33.942617 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-networker-ncc5q" podStartSLOduration=2.4186735759999998 podStartE2EDuration="2.942601253s" podCreationTimestamp="2025-10-14 09:11:31 +0000 UTC" firstStartedPulling="2025-10-14 09:11:32.373790115 +0000 UTC m=+7828.071150486" lastFinishedPulling="2025-10-14 09:11:32.897717762 +0000 UTC m=+7828.595078163" observedRunningTime="2025-10-14 09:11:33.940091021 +0000 UTC m=+7829.637451462" watchObservedRunningTime="2025-10-14 09:11:33.942601253 +0000 UTC m=+7829.639961624" Oct 14 09:11:35 crc kubenswrapper[4870]: I1014 09:11:35.045569 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a5cfe53c-f2e3-4660-a84b-e33429acfc0f" path="/var/lib/kubelet/pods/a5cfe53c-f2e3-4660-a84b-e33429acfc0f/volumes" Oct 14 09:12:31 crc kubenswrapper[4870]: I1014 09:12:31.532017 4870 scope.go:117] "RemoveContainer" containerID="fcf2557eeea7af11b86a0d1b9f17fde60af84369c41c9200aed16412b714bcd1" Oct 14 09:12:53 crc kubenswrapper[4870]: I1014 09:12:53.951311 4870 patch_prober.go:28] interesting pod/machine-config-daemon-7tvc7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 09:12:53 crc kubenswrapper[4870]: I1014 09:12:53.952242 4870 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 09:13:23 crc kubenswrapper[4870]: I1014 09:13:23.951001 4870 patch_prober.go:28] interesting pod/machine-config-daemon-7tvc7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 09:13:23 crc kubenswrapper[4870]: I1014 09:13:23.952001 4870 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 09:13:53 crc kubenswrapper[4870]: I1014 09:13:53.951777 4870 patch_prober.go:28] interesting pod/machine-config-daemon-7tvc7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 09:13:53 crc kubenswrapper[4870]: I1014 09:13:53.954254 4870 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 09:13:53 crc kubenswrapper[4870]: I1014 09:13:53.954531 4870 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" Oct 14 09:13:53 crc kubenswrapper[4870]: I1014 09:13:53.955971 4870 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"a9b3d8bfba3e51117bb9f2ff67d8d2b345878f886b38f76fc3c8cf6d7f91c6de"} pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 14 09:13:53 crc kubenswrapper[4870]: I1014 09:13:53.956264 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerName="machine-config-daemon" containerID="cri-o://a9b3d8bfba3e51117bb9f2ff67d8d2b345878f886b38f76fc3c8cf6d7f91c6de" gracePeriod=600 Oct 14 09:13:54 crc kubenswrapper[4870]: I1014 09:13:54.545993 4870 generic.go:334] "Generic (PLEG): container finished" podID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerID="a9b3d8bfba3e51117bb9f2ff67d8d2b345878f886b38f76fc3c8cf6d7f91c6de" exitCode=0 Oct 14 09:13:54 crc kubenswrapper[4870]: I1014 09:13:54.546075 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" event={"ID":"874a5e45-dffd-4d17-b609-b7d3ed2eab07","Type":"ContainerDied","Data":"a9b3d8bfba3e51117bb9f2ff67d8d2b345878f886b38f76fc3c8cf6d7f91c6de"} Oct 14 09:13:54 crc kubenswrapper[4870]: I1014 09:13:54.546883 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" event={"ID":"874a5e45-dffd-4d17-b609-b7d3ed2eab07","Type":"ContainerStarted","Data":"5a354d0d25c9ddf421bfdb89e3b92b75f73097cc5398e19453b6e98125f9ecf5"} Oct 14 09:13:54 crc kubenswrapper[4870]: I1014 09:13:54.546963 4870 scope.go:117] "RemoveContainer" containerID="8fcaeb3090f0c89cba1e66db524b4a246072daa35592012664aca450fa09ddc9" Oct 14 09:14:36 crc kubenswrapper[4870]: I1014 09:14:36.146613 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-wcvdl"] Oct 14 09:14:36 crc kubenswrapper[4870]: I1014 09:14:36.158903 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wcvdl" Oct 14 09:14:36 crc kubenswrapper[4870]: I1014 09:14:36.180282 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-wcvdl"] Oct 14 09:14:36 crc kubenswrapper[4870]: I1014 09:14:36.240424 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ksbjf\" (UniqueName: \"kubernetes.io/projected/6d31e213-518a-4d7d-a9c6-938fa07316f9-kube-api-access-ksbjf\") pod \"redhat-marketplace-wcvdl\" (UID: \"6d31e213-518a-4d7d-a9c6-938fa07316f9\") " pod="openshift-marketplace/redhat-marketplace-wcvdl" Oct 14 09:14:36 crc kubenswrapper[4870]: I1014 09:14:36.240860 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6d31e213-518a-4d7d-a9c6-938fa07316f9-catalog-content\") pod \"redhat-marketplace-wcvdl\" (UID: \"6d31e213-518a-4d7d-a9c6-938fa07316f9\") " pod="openshift-marketplace/redhat-marketplace-wcvdl" Oct 14 09:14:36 crc kubenswrapper[4870]: I1014 09:14:36.240962 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6d31e213-518a-4d7d-a9c6-938fa07316f9-utilities\") pod \"redhat-marketplace-wcvdl\" (UID: \"6d31e213-518a-4d7d-a9c6-938fa07316f9\") " pod="openshift-marketplace/redhat-marketplace-wcvdl" Oct 14 09:14:36 crc kubenswrapper[4870]: I1014 09:14:36.342320 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6d31e213-518a-4d7d-a9c6-938fa07316f9-catalog-content\") pod \"redhat-marketplace-wcvdl\" (UID: \"6d31e213-518a-4d7d-a9c6-938fa07316f9\") " pod="openshift-marketplace/redhat-marketplace-wcvdl" Oct 14 09:14:36 crc kubenswrapper[4870]: I1014 09:14:36.342465 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6d31e213-518a-4d7d-a9c6-938fa07316f9-utilities\") pod \"redhat-marketplace-wcvdl\" (UID: \"6d31e213-518a-4d7d-a9c6-938fa07316f9\") " pod="openshift-marketplace/redhat-marketplace-wcvdl" Oct 14 09:14:36 crc kubenswrapper[4870]: I1014 09:14:36.342616 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ksbjf\" (UniqueName: \"kubernetes.io/projected/6d31e213-518a-4d7d-a9c6-938fa07316f9-kube-api-access-ksbjf\") pod \"redhat-marketplace-wcvdl\" (UID: \"6d31e213-518a-4d7d-a9c6-938fa07316f9\") " pod="openshift-marketplace/redhat-marketplace-wcvdl" Oct 14 09:14:36 crc kubenswrapper[4870]: I1014 09:14:36.342823 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6d31e213-518a-4d7d-a9c6-938fa07316f9-catalog-content\") pod \"redhat-marketplace-wcvdl\" (UID: \"6d31e213-518a-4d7d-a9c6-938fa07316f9\") " pod="openshift-marketplace/redhat-marketplace-wcvdl" Oct 14 09:14:36 crc kubenswrapper[4870]: I1014 09:14:36.343000 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6d31e213-518a-4d7d-a9c6-938fa07316f9-utilities\") pod \"redhat-marketplace-wcvdl\" (UID: \"6d31e213-518a-4d7d-a9c6-938fa07316f9\") " pod="openshift-marketplace/redhat-marketplace-wcvdl" Oct 14 09:14:36 crc kubenswrapper[4870]: I1014 09:14:36.362507 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ksbjf\" (UniqueName: \"kubernetes.io/projected/6d31e213-518a-4d7d-a9c6-938fa07316f9-kube-api-access-ksbjf\") pod \"redhat-marketplace-wcvdl\" (UID: \"6d31e213-518a-4d7d-a9c6-938fa07316f9\") " pod="openshift-marketplace/redhat-marketplace-wcvdl" Oct 14 09:14:36 crc kubenswrapper[4870]: I1014 09:14:36.489415 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wcvdl" Oct 14 09:14:36 crc kubenswrapper[4870]: W1014 09:14:36.932627 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6d31e213_518a_4d7d_a9c6_938fa07316f9.slice/crio-448776e509bc0ac0d4b4d64546a3562232819fee2e769e3892b9e40cdbe3203f WatchSource:0}: Error finding container 448776e509bc0ac0d4b4d64546a3562232819fee2e769e3892b9e40cdbe3203f: Status 404 returned error can't find the container with id 448776e509bc0ac0d4b4d64546a3562232819fee2e769e3892b9e40cdbe3203f Oct 14 09:14:36 crc kubenswrapper[4870]: I1014 09:14:36.938466 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-wcvdl"] Oct 14 09:14:37 crc kubenswrapper[4870]: I1014 09:14:37.048688 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wcvdl" event={"ID":"6d31e213-518a-4d7d-a9c6-938fa07316f9","Type":"ContainerStarted","Data":"448776e509bc0ac0d4b4d64546a3562232819fee2e769e3892b9e40cdbe3203f"} Oct 14 09:14:38 crc kubenswrapper[4870]: I1014 09:14:38.058124 4870 generic.go:334] "Generic (PLEG): container finished" podID="6d31e213-518a-4d7d-a9c6-938fa07316f9" containerID="606d7281267f3f2bd10b6b67360ac98cb18b141044dc5d78de26531c8d528c06" exitCode=0 Oct 14 09:14:38 crc kubenswrapper[4870]: I1014 09:14:38.058239 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wcvdl" event={"ID":"6d31e213-518a-4d7d-a9c6-938fa07316f9","Type":"ContainerDied","Data":"606d7281267f3f2bd10b6b67360ac98cb18b141044dc5d78de26531c8d528c06"} Oct 14 09:14:38 crc kubenswrapper[4870]: I1014 09:14:38.062628 4870 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 14 09:14:38 crc kubenswrapper[4870]: I1014 09:14:38.909043 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-lgrkk"] Oct 14 09:14:38 crc kubenswrapper[4870]: I1014 09:14:38.912417 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lgrkk" Oct 14 09:14:38 crc kubenswrapper[4870]: I1014 09:14:38.929133 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-lgrkk"] Oct 14 09:14:39 crc kubenswrapper[4870]: I1014 09:14:39.001735 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a618a42e-e43b-4512-b665-ad057cda60c5-utilities\") pod \"redhat-operators-lgrkk\" (UID: \"a618a42e-e43b-4512-b665-ad057cda60c5\") " pod="openshift-marketplace/redhat-operators-lgrkk" Oct 14 09:14:39 crc kubenswrapper[4870]: I1014 09:14:39.001822 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a618a42e-e43b-4512-b665-ad057cda60c5-catalog-content\") pod \"redhat-operators-lgrkk\" (UID: \"a618a42e-e43b-4512-b665-ad057cda60c5\") " pod="openshift-marketplace/redhat-operators-lgrkk" Oct 14 09:14:39 crc kubenswrapper[4870]: I1014 09:14:39.002059 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-69m7j\" (UniqueName: \"kubernetes.io/projected/a618a42e-e43b-4512-b665-ad057cda60c5-kube-api-access-69m7j\") pod \"redhat-operators-lgrkk\" (UID: \"a618a42e-e43b-4512-b665-ad057cda60c5\") " pod="openshift-marketplace/redhat-operators-lgrkk" Oct 14 09:14:39 crc kubenswrapper[4870]: I1014 09:14:39.103911 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a618a42e-e43b-4512-b665-ad057cda60c5-utilities\") pod \"redhat-operators-lgrkk\" (UID: \"a618a42e-e43b-4512-b665-ad057cda60c5\") " pod="openshift-marketplace/redhat-operators-lgrkk" Oct 14 09:14:39 crc kubenswrapper[4870]: I1014 09:14:39.104073 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a618a42e-e43b-4512-b665-ad057cda60c5-catalog-content\") pod \"redhat-operators-lgrkk\" (UID: \"a618a42e-e43b-4512-b665-ad057cda60c5\") " pod="openshift-marketplace/redhat-operators-lgrkk" Oct 14 09:14:39 crc kubenswrapper[4870]: I1014 09:14:39.105057 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-69m7j\" (UniqueName: \"kubernetes.io/projected/a618a42e-e43b-4512-b665-ad057cda60c5-kube-api-access-69m7j\") pod \"redhat-operators-lgrkk\" (UID: \"a618a42e-e43b-4512-b665-ad057cda60c5\") " pod="openshift-marketplace/redhat-operators-lgrkk" Oct 14 09:14:39 crc kubenswrapper[4870]: I1014 09:14:39.105284 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a618a42e-e43b-4512-b665-ad057cda60c5-catalog-content\") pod \"redhat-operators-lgrkk\" (UID: \"a618a42e-e43b-4512-b665-ad057cda60c5\") " pod="openshift-marketplace/redhat-operators-lgrkk" Oct 14 09:14:39 crc kubenswrapper[4870]: I1014 09:14:39.105329 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a618a42e-e43b-4512-b665-ad057cda60c5-utilities\") pod \"redhat-operators-lgrkk\" (UID: \"a618a42e-e43b-4512-b665-ad057cda60c5\") " pod="openshift-marketplace/redhat-operators-lgrkk" Oct 14 09:14:39 crc kubenswrapper[4870]: I1014 09:14:39.134801 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-69m7j\" (UniqueName: \"kubernetes.io/projected/a618a42e-e43b-4512-b665-ad057cda60c5-kube-api-access-69m7j\") pod \"redhat-operators-lgrkk\" (UID: \"a618a42e-e43b-4512-b665-ad057cda60c5\") " pod="openshift-marketplace/redhat-operators-lgrkk" Oct 14 09:14:39 crc kubenswrapper[4870]: I1014 09:14:39.235747 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lgrkk" Oct 14 09:14:39 crc kubenswrapper[4870]: I1014 09:14:39.694568 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-lgrkk"] Oct 14 09:14:39 crc kubenswrapper[4870]: W1014 09:14:39.696903 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda618a42e_e43b_4512_b665_ad057cda60c5.slice/crio-9383bc4345e0a66a943ff1ea9b22013658f3cb9ed8d48fbb232d41ea0a848185 WatchSource:0}: Error finding container 9383bc4345e0a66a943ff1ea9b22013658f3cb9ed8d48fbb232d41ea0a848185: Status 404 returned error can't find the container with id 9383bc4345e0a66a943ff1ea9b22013658f3cb9ed8d48fbb232d41ea0a848185 Oct 14 09:14:40 crc kubenswrapper[4870]: I1014 09:14:40.080529 4870 generic.go:334] "Generic (PLEG): container finished" podID="6d31e213-518a-4d7d-a9c6-938fa07316f9" containerID="11c5d944d9cf2fe393ac990b49229b7c508931af3a1d5e3fe251235563d90e0f" exitCode=0 Oct 14 09:14:40 crc kubenswrapper[4870]: I1014 09:14:40.080599 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wcvdl" event={"ID":"6d31e213-518a-4d7d-a9c6-938fa07316f9","Type":"ContainerDied","Data":"11c5d944d9cf2fe393ac990b49229b7c508931af3a1d5e3fe251235563d90e0f"} Oct 14 09:14:40 crc kubenswrapper[4870]: I1014 09:14:40.084178 4870 generic.go:334] "Generic (PLEG): container finished" podID="a618a42e-e43b-4512-b665-ad057cda60c5" containerID="dfd8908cb37b890e80810a459689152e4e104d1e2a4559d5967bdd6b9adcf347" exitCode=0 Oct 14 09:14:40 crc kubenswrapper[4870]: I1014 09:14:40.084221 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lgrkk" event={"ID":"a618a42e-e43b-4512-b665-ad057cda60c5","Type":"ContainerDied","Data":"dfd8908cb37b890e80810a459689152e4e104d1e2a4559d5967bdd6b9adcf347"} Oct 14 09:14:40 crc kubenswrapper[4870]: I1014 09:14:40.084247 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lgrkk" event={"ID":"a618a42e-e43b-4512-b665-ad057cda60c5","Type":"ContainerStarted","Data":"9383bc4345e0a66a943ff1ea9b22013658f3cb9ed8d48fbb232d41ea0a848185"} Oct 14 09:14:41 crc kubenswrapper[4870]: I1014 09:14:41.106368 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wcvdl" event={"ID":"6d31e213-518a-4d7d-a9c6-938fa07316f9","Type":"ContainerStarted","Data":"43328cfe718015a48fd1f54d92587b19bd9d6010616522290df5e21edfced31f"} Oct 14 09:14:41 crc kubenswrapper[4870]: I1014 09:14:41.109326 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lgrkk" event={"ID":"a618a42e-e43b-4512-b665-ad057cda60c5","Type":"ContainerStarted","Data":"256dadbb3e55e80e0681de36cd131023e6d18aa26d3c0e58489e1b079350264d"} Oct 14 09:14:41 crc kubenswrapper[4870]: I1014 09:14:41.139376 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-wcvdl" podStartSLOduration=2.515851576 podStartE2EDuration="5.139348035s" podCreationTimestamp="2025-10-14 09:14:36 +0000 UTC" firstStartedPulling="2025-10-14 09:14:38.062191404 +0000 UTC m=+8013.759551795" lastFinishedPulling="2025-10-14 09:14:40.685687883 +0000 UTC m=+8016.383048254" observedRunningTime="2025-10-14 09:14:41.129713567 +0000 UTC m=+8016.827073938" watchObservedRunningTime="2025-10-14 09:14:41.139348035 +0000 UTC m=+8016.836708396" Oct 14 09:14:45 crc kubenswrapper[4870]: I1014 09:14:45.150889 4870 generic.go:334] "Generic (PLEG): container finished" podID="a618a42e-e43b-4512-b665-ad057cda60c5" containerID="256dadbb3e55e80e0681de36cd131023e6d18aa26d3c0e58489e1b079350264d" exitCode=0 Oct 14 09:14:45 crc kubenswrapper[4870]: I1014 09:14:45.151006 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lgrkk" event={"ID":"a618a42e-e43b-4512-b665-ad057cda60c5","Type":"ContainerDied","Data":"256dadbb3e55e80e0681de36cd131023e6d18aa26d3c0e58489e1b079350264d"} Oct 14 09:14:46 crc kubenswrapper[4870]: I1014 09:14:46.490511 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-wcvdl" Oct 14 09:14:46 crc kubenswrapper[4870]: I1014 09:14:46.492246 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-wcvdl" Oct 14 09:14:46 crc kubenswrapper[4870]: I1014 09:14:46.539417 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-wcvdl" Oct 14 09:14:47 crc kubenswrapper[4870]: I1014 09:14:47.177419 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lgrkk" event={"ID":"a618a42e-e43b-4512-b665-ad057cda60c5","Type":"ContainerStarted","Data":"07736105c2af079678c2f1eb687c5f6d54b242eb662834bb71ccb8892cbb7c89"} Oct 14 09:14:47 crc kubenswrapper[4870]: I1014 09:14:47.203607 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-lgrkk" podStartSLOduration=3.252112979 podStartE2EDuration="9.203587072s" podCreationTimestamp="2025-10-14 09:14:38 +0000 UTC" firstStartedPulling="2025-10-14 09:14:40.08576202 +0000 UTC m=+8015.783122391" lastFinishedPulling="2025-10-14 09:14:46.037236113 +0000 UTC m=+8021.734596484" observedRunningTime="2025-10-14 09:14:47.196121368 +0000 UTC m=+8022.893481739" watchObservedRunningTime="2025-10-14 09:14:47.203587072 +0000 UTC m=+8022.900947443" Oct 14 09:14:47 crc kubenswrapper[4870]: I1014 09:14:47.233341 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-wcvdl" Oct 14 09:14:48 crc kubenswrapper[4870]: I1014 09:14:48.696694 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-wcvdl"] Oct 14 09:14:49 crc kubenswrapper[4870]: I1014 09:14:49.236582 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-lgrkk" Oct 14 09:14:49 crc kubenswrapper[4870]: I1014 09:14:49.236650 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-lgrkk" Oct 14 09:14:50 crc kubenswrapper[4870]: I1014 09:14:50.205804 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-wcvdl" podUID="6d31e213-518a-4d7d-a9c6-938fa07316f9" containerName="registry-server" containerID="cri-o://43328cfe718015a48fd1f54d92587b19bd9d6010616522290df5e21edfced31f" gracePeriod=2 Oct 14 09:14:50 crc kubenswrapper[4870]: I1014 09:14:50.290710 4870 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-lgrkk" podUID="a618a42e-e43b-4512-b665-ad057cda60c5" containerName="registry-server" probeResult="failure" output=< Oct 14 09:14:50 crc kubenswrapper[4870]: timeout: failed to connect service ":50051" within 1s Oct 14 09:14:50 crc kubenswrapper[4870]: > Oct 14 09:14:50 crc kubenswrapper[4870]: I1014 09:14:50.821578 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wcvdl" Oct 14 09:14:50 crc kubenswrapper[4870]: I1014 09:14:50.880488 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6d31e213-518a-4d7d-a9c6-938fa07316f9-utilities\") pod \"6d31e213-518a-4d7d-a9c6-938fa07316f9\" (UID: \"6d31e213-518a-4d7d-a9c6-938fa07316f9\") " Oct 14 09:14:50 crc kubenswrapper[4870]: I1014 09:14:50.880608 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ksbjf\" (UniqueName: \"kubernetes.io/projected/6d31e213-518a-4d7d-a9c6-938fa07316f9-kube-api-access-ksbjf\") pod \"6d31e213-518a-4d7d-a9c6-938fa07316f9\" (UID: \"6d31e213-518a-4d7d-a9c6-938fa07316f9\") " Oct 14 09:14:50 crc kubenswrapper[4870]: I1014 09:14:50.880839 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6d31e213-518a-4d7d-a9c6-938fa07316f9-catalog-content\") pod \"6d31e213-518a-4d7d-a9c6-938fa07316f9\" (UID: \"6d31e213-518a-4d7d-a9c6-938fa07316f9\") " Oct 14 09:14:50 crc kubenswrapper[4870]: I1014 09:14:50.881355 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6d31e213-518a-4d7d-a9c6-938fa07316f9-utilities" (OuterVolumeSpecName: "utilities") pod "6d31e213-518a-4d7d-a9c6-938fa07316f9" (UID: "6d31e213-518a-4d7d-a9c6-938fa07316f9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 09:14:50 crc kubenswrapper[4870]: I1014 09:14:50.881920 4870 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6d31e213-518a-4d7d-a9c6-938fa07316f9-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 09:14:50 crc kubenswrapper[4870]: I1014 09:14:50.899625 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6d31e213-518a-4d7d-a9c6-938fa07316f9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6d31e213-518a-4d7d-a9c6-938fa07316f9" (UID: "6d31e213-518a-4d7d-a9c6-938fa07316f9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 09:14:50 crc kubenswrapper[4870]: I1014 09:14:50.905264 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6d31e213-518a-4d7d-a9c6-938fa07316f9-kube-api-access-ksbjf" (OuterVolumeSpecName: "kube-api-access-ksbjf") pod "6d31e213-518a-4d7d-a9c6-938fa07316f9" (UID: "6d31e213-518a-4d7d-a9c6-938fa07316f9"). InnerVolumeSpecName "kube-api-access-ksbjf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:14:50 crc kubenswrapper[4870]: I1014 09:14:50.983963 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ksbjf\" (UniqueName: \"kubernetes.io/projected/6d31e213-518a-4d7d-a9c6-938fa07316f9-kube-api-access-ksbjf\") on node \"crc\" DevicePath \"\"" Oct 14 09:14:50 crc kubenswrapper[4870]: I1014 09:14:50.983992 4870 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6d31e213-518a-4d7d-a9c6-938fa07316f9-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 09:14:51 crc kubenswrapper[4870]: I1014 09:14:51.217675 4870 generic.go:334] "Generic (PLEG): container finished" podID="6d31e213-518a-4d7d-a9c6-938fa07316f9" containerID="43328cfe718015a48fd1f54d92587b19bd9d6010616522290df5e21edfced31f" exitCode=0 Oct 14 09:14:51 crc kubenswrapper[4870]: I1014 09:14:51.218053 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wcvdl" Oct 14 09:14:51 crc kubenswrapper[4870]: I1014 09:14:51.217908 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wcvdl" event={"ID":"6d31e213-518a-4d7d-a9c6-938fa07316f9","Type":"ContainerDied","Data":"43328cfe718015a48fd1f54d92587b19bd9d6010616522290df5e21edfced31f"} Oct 14 09:14:51 crc kubenswrapper[4870]: I1014 09:14:51.218252 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wcvdl" event={"ID":"6d31e213-518a-4d7d-a9c6-938fa07316f9","Type":"ContainerDied","Data":"448776e509bc0ac0d4b4d64546a3562232819fee2e769e3892b9e40cdbe3203f"} Oct 14 09:14:51 crc kubenswrapper[4870]: I1014 09:14:51.218295 4870 scope.go:117] "RemoveContainer" containerID="43328cfe718015a48fd1f54d92587b19bd9d6010616522290df5e21edfced31f" Oct 14 09:14:51 crc kubenswrapper[4870]: I1014 09:14:51.248965 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-wcvdl"] Oct 14 09:14:51 crc kubenswrapper[4870]: I1014 09:14:51.251619 4870 scope.go:117] "RemoveContainer" containerID="11c5d944d9cf2fe393ac990b49229b7c508931af3a1d5e3fe251235563d90e0f" Oct 14 09:14:51 crc kubenswrapper[4870]: I1014 09:14:51.257412 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-wcvdl"] Oct 14 09:14:51 crc kubenswrapper[4870]: I1014 09:14:51.280521 4870 scope.go:117] "RemoveContainer" containerID="606d7281267f3f2bd10b6b67360ac98cb18b141044dc5d78de26531c8d528c06" Oct 14 09:14:51 crc kubenswrapper[4870]: I1014 09:14:51.336839 4870 scope.go:117] "RemoveContainer" containerID="43328cfe718015a48fd1f54d92587b19bd9d6010616522290df5e21edfced31f" Oct 14 09:14:51 crc kubenswrapper[4870]: E1014 09:14:51.337567 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"43328cfe718015a48fd1f54d92587b19bd9d6010616522290df5e21edfced31f\": container with ID starting with 43328cfe718015a48fd1f54d92587b19bd9d6010616522290df5e21edfced31f not found: ID does not exist" containerID="43328cfe718015a48fd1f54d92587b19bd9d6010616522290df5e21edfced31f" Oct 14 09:14:51 crc kubenswrapper[4870]: I1014 09:14:51.337619 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"43328cfe718015a48fd1f54d92587b19bd9d6010616522290df5e21edfced31f"} err="failed to get container status \"43328cfe718015a48fd1f54d92587b19bd9d6010616522290df5e21edfced31f\": rpc error: code = NotFound desc = could not find container \"43328cfe718015a48fd1f54d92587b19bd9d6010616522290df5e21edfced31f\": container with ID starting with 43328cfe718015a48fd1f54d92587b19bd9d6010616522290df5e21edfced31f not found: ID does not exist" Oct 14 09:14:51 crc kubenswrapper[4870]: I1014 09:14:51.337647 4870 scope.go:117] "RemoveContainer" containerID="11c5d944d9cf2fe393ac990b49229b7c508931af3a1d5e3fe251235563d90e0f" Oct 14 09:14:51 crc kubenswrapper[4870]: E1014 09:14:51.338201 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"11c5d944d9cf2fe393ac990b49229b7c508931af3a1d5e3fe251235563d90e0f\": container with ID starting with 11c5d944d9cf2fe393ac990b49229b7c508931af3a1d5e3fe251235563d90e0f not found: ID does not exist" containerID="11c5d944d9cf2fe393ac990b49229b7c508931af3a1d5e3fe251235563d90e0f" Oct 14 09:14:51 crc kubenswrapper[4870]: I1014 09:14:51.338241 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"11c5d944d9cf2fe393ac990b49229b7c508931af3a1d5e3fe251235563d90e0f"} err="failed to get container status \"11c5d944d9cf2fe393ac990b49229b7c508931af3a1d5e3fe251235563d90e0f\": rpc error: code = NotFound desc = could not find container \"11c5d944d9cf2fe393ac990b49229b7c508931af3a1d5e3fe251235563d90e0f\": container with ID starting with 11c5d944d9cf2fe393ac990b49229b7c508931af3a1d5e3fe251235563d90e0f not found: ID does not exist" Oct 14 09:14:51 crc kubenswrapper[4870]: I1014 09:14:51.338266 4870 scope.go:117] "RemoveContainer" containerID="606d7281267f3f2bd10b6b67360ac98cb18b141044dc5d78de26531c8d528c06" Oct 14 09:14:51 crc kubenswrapper[4870]: E1014 09:14:51.338506 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"606d7281267f3f2bd10b6b67360ac98cb18b141044dc5d78de26531c8d528c06\": container with ID starting with 606d7281267f3f2bd10b6b67360ac98cb18b141044dc5d78de26531c8d528c06 not found: ID does not exist" containerID="606d7281267f3f2bd10b6b67360ac98cb18b141044dc5d78de26531c8d528c06" Oct 14 09:14:51 crc kubenswrapper[4870]: I1014 09:14:51.338534 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"606d7281267f3f2bd10b6b67360ac98cb18b141044dc5d78de26531c8d528c06"} err="failed to get container status \"606d7281267f3f2bd10b6b67360ac98cb18b141044dc5d78de26531c8d528c06\": rpc error: code = NotFound desc = could not find container \"606d7281267f3f2bd10b6b67360ac98cb18b141044dc5d78de26531c8d528c06\": container with ID starting with 606d7281267f3f2bd10b6b67360ac98cb18b141044dc5d78de26531c8d528c06 not found: ID does not exist" Oct 14 09:14:53 crc kubenswrapper[4870]: I1014 09:14:53.069030 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6d31e213-518a-4d7d-a9c6-938fa07316f9" path="/var/lib/kubelet/pods/6d31e213-518a-4d7d-a9c6-938fa07316f9/volumes" Oct 14 09:14:59 crc kubenswrapper[4870]: I1014 09:14:59.281320 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-lgrkk" Oct 14 09:14:59 crc kubenswrapper[4870]: I1014 09:14:59.330338 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-lgrkk" Oct 14 09:14:59 crc kubenswrapper[4870]: I1014 09:14:59.516283 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-lgrkk"] Oct 14 09:15:00 crc kubenswrapper[4870]: I1014 09:15:00.151260 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29340555-qnz96"] Oct 14 09:15:00 crc kubenswrapper[4870]: E1014 09:15:00.151726 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d31e213-518a-4d7d-a9c6-938fa07316f9" containerName="extract-utilities" Oct 14 09:15:00 crc kubenswrapper[4870]: I1014 09:15:00.151743 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d31e213-518a-4d7d-a9c6-938fa07316f9" containerName="extract-utilities" Oct 14 09:15:00 crc kubenswrapper[4870]: E1014 09:15:00.151756 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d31e213-518a-4d7d-a9c6-938fa07316f9" containerName="extract-content" Oct 14 09:15:00 crc kubenswrapper[4870]: I1014 09:15:00.151763 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d31e213-518a-4d7d-a9c6-938fa07316f9" containerName="extract-content" Oct 14 09:15:00 crc kubenswrapper[4870]: E1014 09:15:00.151793 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d31e213-518a-4d7d-a9c6-938fa07316f9" containerName="registry-server" Oct 14 09:15:00 crc kubenswrapper[4870]: I1014 09:15:00.151799 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d31e213-518a-4d7d-a9c6-938fa07316f9" containerName="registry-server" Oct 14 09:15:00 crc kubenswrapper[4870]: I1014 09:15:00.152013 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="6d31e213-518a-4d7d-a9c6-938fa07316f9" containerName="registry-server" Oct 14 09:15:00 crc kubenswrapper[4870]: I1014 09:15:00.152885 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29340555-qnz96" Oct 14 09:15:00 crc kubenswrapper[4870]: I1014 09:15:00.156316 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 14 09:15:00 crc kubenswrapper[4870]: I1014 09:15:00.161489 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 14 09:15:00 crc kubenswrapper[4870]: I1014 09:15:00.166633 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29340555-qnz96"] Oct 14 09:15:00 crc kubenswrapper[4870]: I1014 09:15:00.279025 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e7cd0dbb-80e4-4f4b-9942-5082214636d5-config-volume\") pod \"collect-profiles-29340555-qnz96\" (UID: \"e7cd0dbb-80e4-4f4b-9942-5082214636d5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340555-qnz96" Oct 14 09:15:00 crc kubenswrapper[4870]: I1014 09:15:00.279392 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e7cd0dbb-80e4-4f4b-9942-5082214636d5-secret-volume\") pod \"collect-profiles-29340555-qnz96\" (UID: \"e7cd0dbb-80e4-4f4b-9942-5082214636d5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340555-qnz96" Oct 14 09:15:00 crc kubenswrapper[4870]: I1014 09:15:00.279488 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cfjdf\" (UniqueName: \"kubernetes.io/projected/e7cd0dbb-80e4-4f4b-9942-5082214636d5-kube-api-access-cfjdf\") pod \"collect-profiles-29340555-qnz96\" (UID: \"e7cd0dbb-80e4-4f4b-9942-5082214636d5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340555-qnz96" Oct 14 09:15:00 crc kubenswrapper[4870]: I1014 09:15:00.319317 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-lgrkk" podUID="a618a42e-e43b-4512-b665-ad057cda60c5" containerName="registry-server" containerID="cri-o://07736105c2af079678c2f1eb687c5f6d54b242eb662834bb71ccb8892cbb7c89" gracePeriod=2 Oct 14 09:15:00 crc kubenswrapper[4870]: I1014 09:15:00.381097 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cfjdf\" (UniqueName: \"kubernetes.io/projected/e7cd0dbb-80e4-4f4b-9942-5082214636d5-kube-api-access-cfjdf\") pod \"collect-profiles-29340555-qnz96\" (UID: \"e7cd0dbb-80e4-4f4b-9942-5082214636d5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340555-qnz96" Oct 14 09:15:00 crc kubenswrapper[4870]: I1014 09:15:00.381265 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e7cd0dbb-80e4-4f4b-9942-5082214636d5-config-volume\") pod \"collect-profiles-29340555-qnz96\" (UID: \"e7cd0dbb-80e4-4f4b-9942-5082214636d5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340555-qnz96" Oct 14 09:15:00 crc kubenswrapper[4870]: I1014 09:15:00.382060 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e7cd0dbb-80e4-4f4b-9942-5082214636d5-config-volume\") pod \"collect-profiles-29340555-qnz96\" (UID: \"e7cd0dbb-80e4-4f4b-9942-5082214636d5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340555-qnz96" Oct 14 09:15:00 crc kubenswrapper[4870]: I1014 09:15:00.382165 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e7cd0dbb-80e4-4f4b-9942-5082214636d5-secret-volume\") pod \"collect-profiles-29340555-qnz96\" (UID: \"e7cd0dbb-80e4-4f4b-9942-5082214636d5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340555-qnz96" Oct 14 09:15:00 crc kubenswrapper[4870]: I1014 09:15:00.387983 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e7cd0dbb-80e4-4f4b-9942-5082214636d5-secret-volume\") pod \"collect-profiles-29340555-qnz96\" (UID: \"e7cd0dbb-80e4-4f4b-9942-5082214636d5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340555-qnz96" Oct 14 09:15:00 crc kubenswrapper[4870]: I1014 09:15:00.402309 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cfjdf\" (UniqueName: \"kubernetes.io/projected/e7cd0dbb-80e4-4f4b-9942-5082214636d5-kube-api-access-cfjdf\") pod \"collect-profiles-29340555-qnz96\" (UID: \"e7cd0dbb-80e4-4f4b-9942-5082214636d5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340555-qnz96" Oct 14 09:15:00 crc kubenswrapper[4870]: I1014 09:15:00.493882 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29340555-qnz96" Oct 14 09:15:00 crc kubenswrapper[4870]: I1014 09:15:00.810666 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lgrkk" Oct 14 09:15:00 crc kubenswrapper[4870]: I1014 09:15:00.995668 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a618a42e-e43b-4512-b665-ad057cda60c5-catalog-content\") pod \"a618a42e-e43b-4512-b665-ad057cda60c5\" (UID: \"a618a42e-e43b-4512-b665-ad057cda60c5\") " Oct 14 09:15:00 crc kubenswrapper[4870]: I1014 09:15:00.996032 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-69m7j\" (UniqueName: \"kubernetes.io/projected/a618a42e-e43b-4512-b665-ad057cda60c5-kube-api-access-69m7j\") pod \"a618a42e-e43b-4512-b665-ad057cda60c5\" (UID: \"a618a42e-e43b-4512-b665-ad057cda60c5\") " Oct 14 09:15:00 crc kubenswrapper[4870]: I1014 09:15:00.996170 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a618a42e-e43b-4512-b665-ad057cda60c5-utilities\") pod \"a618a42e-e43b-4512-b665-ad057cda60c5\" (UID: \"a618a42e-e43b-4512-b665-ad057cda60c5\") " Oct 14 09:15:00 crc kubenswrapper[4870]: I1014 09:15:00.997367 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a618a42e-e43b-4512-b665-ad057cda60c5-utilities" (OuterVolumeSpecName: "utilities") pod "a618a42e-e43b-4512-b665-ad057cda60c5" (UID: "a618a42e-e43b-4512-b665-ad057cda60c5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 09:15:01 crc kubenswrapper[4870]: I1014 09:15:01.003430 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a618a42e-e43b-4512-b665-ad057cda60c5-kube-api-access-69m7j" (OuterVolumeSpecName: "kube-api-access-69m7j") pod "a618a42e-e43b-4512-b665-ad057cda60c5" (UID: "a618a42e-e43b-4512-b665-ad057cda60c5"). InnerVolumeSpecName "kube-api-access-69m7j". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:15:01 crc kubenswrapper[4870]: I1014 09:15:01.033684 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29340555-qnz96"] Oct 14 09:15:01 crc kubenswrapper[4870]: I1014 09:15:01.090323 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a618a42e-e43b-4512-b665-ad057cda60c5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a618a42e-e43b-4512-b665-ad057cda60c5" (UID: "a618a42e-e43b-4512-b665-ad057cda60c5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 09:15:01 crc kubenswrapper[4870]: I1014 09:15:01.099255 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-69m7j\" (UniqueName: \"kubernetes.io/projected/a618a42e-e43b-4512-b665-ad057cda60c5-kube-api-access-69m7j\") on node \"crc\" DevicePath \"\"" Oct 14 09:15:01 crc kubenswrapper[4870]: I1014 09:15:01.099298 4870 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a618a42e-e43b-4512-b665-ad057cda60c5-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 09:15:01 crc kubenswrapper[4870]: I1014 09:15:01.099312 4870 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a618a42e-e43b-4512-b665-ad057cda60c5-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 09:15:01 crc kubenswrapper[4870]: I1014 09:15:01.328834 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29340555-qnz96" event={"ID":"e7cd0dbb-80e4-4f4b-9942-5082214636d5","Type":"ContainerStarted","Data":"2616e7acafcec59999241eb7ea10f3a3a4a6900c2e7f284928cb66c4e2f89048"} Oct 14 09:15:01 crc kubenswrapper[4870]: I1014 09:15:01.328885 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29340555-qnz96" event={"ID":"e7cd0dbb-80e4-4f4b-9942-5082214636d5","Type":"ContainerStarted","Data":"9accd9039e6b61e3094762c39f3215e4aea9e02ed4bf61457ebedc5a7de4a407"} Oct 14 09:15:01 crc kubenswrapper[4870]: I1014 09:15:01.332439 4870 generic.go:334] "Generic (PLEG): container finished" podID="a618a42e-e43b-4512-b665-ad057cda60c5" containerID="07736105c2af079678c2f1eb687c5f6d54b242eb662834bb71ccb8892cbb7c89" exitCode=0 Oct 14 09:15:01 crc kubenswrapper[4870]: I1014 09:15:01.332500 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lgrkk" event={"ID":"a618a42e-e43b-4512-b665-ad057cda60c5","Type":"ContainerDied","Data":"07736105c2af079678c2f1eb687c5f6d54b242eb662834bb71ccb8892cbb7c89"} Oct 14 09:15:01 crc kubenswrapper[4870]: I1014 09:15:01.332529 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lgrkk" event={"ID":"a618a42e-e43b-4512-b665-ad057cda60c5","Type":"ContainerDied","Data":"9383bc4345e0a66a943ff1ea9b22013658f3cb9ed8d48fbb232d41ea0a848185"} Oct 14 09:15:01 crc kubenswrapper[4870]: I1014 09:15:01.332545 4870 scope.go:117] "RemoveContainer" containerID="07736105c2af079678c2f1eb687c5f6d54b242eb662834bb71ccb8892cbb7c89" Oct 14 09:15:01 crc kubenswrapper[4870]: I1014 09:15:01.332694 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lgrkk" Oct 14 09:15:01 crc kubenswrapper[4870]: I1014 09:15:01.364017 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29340555-qnz96" podStartSLOduration=1.36400098 podStartE2EDuration="1.36400098s" podCreationTimestamp="2025-10-14 09:15:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 09:15:01.344816646 +0000 UTC m=+8037.042177017" watchObservedRunningTime="2025-10-14 09:15:01.36400098 +0000 UTC m=+8037.061361351" Oct 14 09:15:01 crc kubenswrapper[4870]: I1014 09:15:01.368694 4870 scope.go:117] "RemoveContainer" containerID="256dadbb3e55e80e0681de36cd131023e6d18aa26d3c0e58489e1b079350264d" Oct 14 09:15:01 crc kubenswrapper[4870]: I1014 09:15:01.372578 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-lgrkk"] Oct 14 09:15:01 crc kubenswrapper[4870]: I1014 09:15:01.380830 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-lgrkk"] Oct 14 09:15:01 crc kubenswrapper[4870]: I1014 09:15:01.405305 4870 scope.go:117] "RemoveContainer" containerID="dfd8908cb37b890e80810a459689152e4e104d1e2a4559d5967bdd6b9adcf347" Oct 14 09:15:01 crc kubenswrapper[4870]: I1014 09:15:01.430351 4870 scope.go:117] "RemoveContainer" containerID="07736105c2af079678c2f1eb687c5f6d54b242eb662834bb71ccb8892cbb7c89" Oct 14 09:15:01 crc kubenswrapper[4870]: E1014 09:15:01.430743 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"07736105c2af079678c2f1eb687c5f6d54b242eb662834bb71ccb8892cbb7c89\": container with ID starting with 07736105c2af079678c2f1eb687c5f6d54b242eb662834bb71ccb8892cbb7c89 not found: ID does not exist" containerID="07736105c2af079678c2f1eb687c5f6d54b242eb662834bb71ccb8892cbb7c89" Oct 14 09:15:01 crc kubenswrapper[4870]: I1014 09:15:01.430776 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"07736105c2af079678c2f1eb687c5f6d54b242eb662834bb71ccb8892cbb7c89"} err="failed to get container status \"07736105c2af079678c2f1eb687c5f6d54b242eb662834bb71ccb8892cbb7c89\": rpc error: code = NotFound desc = could not find container \"07736105c2af079678c2f1eb687c5f6d54b242eb662834bb71ccb8892cbb7c89\": container with ID starting with 07736105c2af079678c2f1eb687c5f6d54b242eb662834bb71ccb8892cbb7c89 not found: ID does not exist" Oct 14 09:15:01 crc kubenswrapper[4870]: I1014 09:15:01.430798 4870 scope.go:117] "RemoveContainer" containerID="256dadbb3e55e80e0681de36cd131023e6d18aa26d3c0e58489e1b079350264d" Oct 14 09:15:01 crc kubenswrapper[4870]: E1014 09:15:01.431171 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"256dadbb3e55e80e0681de36cd131023e6d18aa26d3c0e58489e1b079350264d\": container with ID starting with 256dadbb3e55e80e0681de36cd131023e6d18aa26d3c0e58489e1b079350264d not found: ID does not exist" containerID="256dadbb3e55e80e0681de36cd131023e6d18aa26d3c0e58489e1b079350264d" Oct 14 09:15:01 crc kubenswrapper[4870]: I1014 09:15:01.431198 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"256dadbb3e55e80e0681de36cd131023e6d18aa26d3c0e58489e1b079350264d"} err="failed to get container status \"256dadbb3e55e80e0681de36cd131023e6d18aa26d3c0e58489e1b079350264d\": rpc error: code = NotFound desc = could not find container \"256dadbb3e55e80e0681de36cd131023e6d18aa26d3c0e58489e1b079350264d\": container with ID starting with 256dadbb3e55e80e0681de36cd131023e6d18aa26d3c0e58489e1b079350264d not found: ID does not exist" Oct 14 09:15:01 crc kubenswrapper[4870]: I1014 09:15:01.431219 4870 scope.go:117] "RemoveContainer" containerID="dfd8908cb37b890e80810a459689152e4e104d1e2a4559d5967bdd6b9adcf347" Oct 14 09:15:01 crc kubenswrapper[4870]: E1014 09:15:01.431441 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dfd8908cb37b890e80810a459689152e4e104d1e2a4559d5967bdd6b9adcf347\": container with ID starting with dfd8908cb37b890e80810a459689152e4e104d1e2a4559d5967bdd6b9adcf347 not found: ID does not exist" containerID="dfd8908cb37b890e80810a459689152e4e104d1e2a4559d5967bdd6b9adcf347" Oct 14 09:15:01 crc kubenswrapper[4870]: I1014 09:15:01.431483 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dfd8908cb37b890e80810a459689152e4e104d1e2a4559d5967bdd6b9adcf347"} err="failed to get container status \"dfd8908cb37b890e80810a459689152e4e104d1e2a4559d5967bdd6b9adcf347\": rpc error: code = NotFound desc = could not find container \"dfd8908cb37b890e80810a459689152e4e104d1e2a4559d5967bdd6b9adcf347\": container with ID starting with dfd8908cb37b890e80810a459689152e4e104d1e2a4559d5967bdd6b9adcf347 not found: ID does not exist" Oct 14 09:15:02 crc kubenswrapper[4870]: I1014 09:15:02.343913 4870 generic.go:334] "Generic (PLEG): container finished" podID="e7cd0dbb-80e4-4f4b-9942-5082214636d5" containerID="2616e7acafcec59999241eb7ea10f3a3a4a6900c2e7f284928cb66c4e2f89048" exitCode=0 Oct 14 09:15:02 crc kubenswrapper[4870]: I1014 09:15:02.343964 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29340555-qnz96" event={"ID":"e7cd0dbb-80e4-4f4b-9942-5082214636d5","Type":"ContainerDied","Data":"2616e7acafcec59999241eb7ea10f3a3a4a6900c2e7f284928cb66c4e2f89048"} Oct 14 09:15:03 crc kubenswrapper[4870]: I1014 09:15:03.047206 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a618a42e-e43b-4512-b665-ad057cda60c5" path="/var/lib/kubelet/pods/a618a42e-e43b-4512-b665-ad057cda60c5/volumes" Oct 14 09:15:03 crc kubenswrapper[4870]: I1014 09:15:03.715067 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29340555-qnz96" Oct 14 09:15:03 crc kubenswrapper[4870]: I1014 09:15:03.786155 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e7cd0dbb-80e4-4f4b-9942-5082214636d5-secret-volume\") pod \"e7cd0dbb-80e4-4f4b-9942-5082214636d5\" (UID: \"e7cd0dbb-80e4-4f4b-9942-5082214636d5\") " Oct 14 09:15:03 crc kubenswrapper[4870]: I1014 09:15:03.786278 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e7cd0dbb-80e4-4f4b-9942-5082214636d5-config-volume\") pod \"e7cd0dbb-80e4-4f4b-9942-5082214636d5\" (UID: \"e7cd0dbb-80e4-4f4b-9942-5082214636d5\") " Oct 14 09:15:03 crc kubenswrapper[4870]: I1014 09:15:03.786420 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfjdf\" (UniqueName: \"kubernetes.io/projected/e7cd0dbb-80e4-4f4b-9942-5082214636d5-kube-api-access-cfjdf\") pod \"e7cd0dbb-80e4-4f4b-9942-5082214636d5\" (UID: \"e7cd0dbb-80e4-4f4b-9942-5082214636d5\") " Oct 14 09:15:03 crc kubenswrapper[4870]: I1014 09:15:03.788582 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7cd0dbb-80e4-4f4b-9942-5082214636d5-config-volume" (OuterVolumeSpecName: "config-volume") pod "e7cd0dbb-80e4-4f4b-9942-5082214636d5" (UID: "e7cd0dbb-80e4-4f4b-9942-5082214636d5"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 09:15:03 crc kubenswrapper[4870]: I1014 09:15:03.792525 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7cd0dbb-80e4-4f4b-9942-5082214636d5-kube-api-access-cfjdf" (OuterVolumeSpecName: "kube-api-access-cfjdf") pod "e7cd0dbb-80e4-4f4b-9942-5082214636d5" (UID: "e7cd0dbb-80e4-4f4b-9942-5082214636d5"). InnerVolumeSpecName "kube-api-access-cfjdf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:15:03 crc kubenswrapper[4870]: I1014 09:15:03.798670 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7cd0dbb-80e4-4f4b-9942-5082214636d5-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "e7cd0dbb-80e4-4f4b-9942-5082214636d5" (UID: "e7cd0dbb-80e4-4f4b-9942-5082214636d5"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:15:03 crc kubenswrapper[4870]: I1014 09:15:03.888430 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfjdf\" (UniqueName: \"kubernetes.io/projected/e7cd0dbb-80e4-4f4b-9942-5082214636d5-kube-api-access-cfjdf\") on node \"crc\" DevicePath \"\"" Oct 14 09:15:03 crc kubenswrapper[4870]: I1014 09:15:03.888513 4870 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e7cd0dbb-80e4-4f4b-9942-5082214636d5-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 14 09:15:03 crc kubenswrapper[4870]: I1014 09:15:03.888531 4870 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e7cd0dbb-80e4-4f4b-9942-5082214636d5-config-volume\") on node \"crc\" DevicePath \"\"" Oct 14 09:15:04 crc kubenswrapper[4870]: I1014 09:15:04.367906 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29340555-qnz96" event={"ID":"e7cd0dbb-80e4-4f4b-9942-5082214636d5","Type":"ContainerDied","Data":"9accd9039e6b61e3094762c39f3215e4aea9e02ed4bf61457ebedc5a7de4a407"} Oct 14 09:15:04 crc kubenswrapper[4870]: I1014 09:15:04.368300 4870 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9accd9039e6b61e3094762c39f3215e4aea9e02ed4bf61457ebedc5a7de4a407" Oct 14 09:15:04 crc kubenswrapper[4870]: I1014 09:15:04.368135 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29340555-qnz96" Oct 14 09:15:04 crc kubenswrapper[4870]: I1014 09:15:04.444035 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29340510-vb6cv"] Oct 14 09:15:04 crc kubenswrapper[4870]: I1014 09:15:04.451956 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29340510-vb6cv"] Oct 14 09:15:05 crc kubenswrapper[4870]: I1014 09:15:05.055991 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6702b606-3463-40c8-934c-baaaf9735e01" path="/var/lib/kubelet/pods/6702b606-3463-40c8-934c-baaaf9735e01/volumes" Oct 14 09:15:30 crc kubenswrapper[4870]: I1014 09:15:30.042730 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-db-create-njft6"] Oct 14 09:15:30 crc kubenswrapper[4870]: I1014 09:15:30.055747 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-db-create-njft6"] Oct 14 09:15:31 crc kubenswrapper[4870]: I1014 09:15:31.049930 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e10cd527-69c5-4c6c-bd52-774cdb2dc9e3" path="/var/lib/kubelet/pods/e10cd527-69c5-4c6c-bd52-774cdb2dc9e3/volumes" Oct 14 09:15:31 crc kubenswrapper[4870]: I1014 09:15:31.686466 4870 scope.go:117] "RemoveContainer" containerID="50ccc31eebb6c05607156d85a7862b4203834eae708689e65c496ea8a5560964" Oct 14 09:15:31 crc kubenswrapper[4870]: I1014 09:15:31.728574 4870 scope.go:117] "RemoveContainer" containerID="39257698f6cdcb9408754a91d15b0bc347d48f9bd53382fceab1014d43df1138" Oct 14 09:15:40 crc kubenswrapper[4870]: I1014 09:15:40.041840 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-c54c-account-create-vgr7d"] Oct 14 09:15:40 crc kubenswrapper[4870]: I1014 09:15:40.055900 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-c54c-account-create-vgr7d"] Oct 14 09:15:41 crc kubenswrapper[4870]: I1014 09:15:41.051642 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2c21a773-0454-49b9-8a63-5312e7a64826" path="/var/lib/kubelet/pods/2c21a773-0454-49b9-8a63-5312e7a64826/volumes" Oct 14 09:15:55 crc kubenswrapper[4870]: I1014 09:15:55.071206 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-db-sync-lrs97"] Oct 14 09:15:55 crc kubenswrapper[4870]: I1014 09:15:55.087056 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-db-sync-lrs97"] Oct 14 09:15:57 crc kubenswrapper[4870]: I1014 09:15:57.045842 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ddf1fd10-61d6-41a6-a261-5fa5388602bb" path="/var/lib/kubelet/pods/ddf1fd10-61d6-41a6-a261-5fa5388602bb/volumes" Oct 14 09:16:00 crc kubenswrapper[4870]: I1014 09:16:00.016628 4870 generic.go:334] "Generic (PLEG): container finished" podID="b96078e0-b889-4b47-9bec-a1411c1fec36" containerID="e0ef696961602aad945c2246f047575f06c8ba9512e813739a1d582140fb2a72" exitCode=0 Oct 14 09:16:00 crc kubenswrapper[4870]: I1014 09:16:00.016710 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-networker-ncc5q" event={"ID":"b96078e0-b889-4b47-9bec-a1411c1fec36","Type":"ContainerDied","Data":"e0ef696961602aad945c2246f047575f06c8ba9512e813739a1d582140fb2a72"} Oct 14 09:16:01 crc kubenswrapper[4870]: I1014 09:16:01.507122 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-networker-ncc5q" Oct 14 09:16:01 crc kubenswrapper[4870]: I1014 09:16:01.632588 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tripleo-cleanup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b96078e0-b889-4b47-9bec-a1411c1fec36-tripleo-cleanup-combined-ca-bundle\") pod \"b96078e0-b889-4b47-9bec-a1411c1fec36\" (UID: \"b96078e0-b889-4b47-9bec-a1411c1fec36\") " Oct 14 09:16:01 crc kubenswrapper[4870]: I1014 09:16:01.632781 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b96078e0-b889-4b47-9bec-a1411c1fec36-inventory\") pod \"b96078e0-b889-4b47-9bec-a1411c1fec36\" (UID: \"b96078e0-b889-4b47-9bec-a1411c1fec36\") " Oct 14 09:16:01 crc kubenswrapper[4870]: I1014 09:16:01.632879 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hn9gp\" (UniqueName: \"kubernetes.io/projected/b96078e0-b889-4b47-9bec-a1411c1fec36-kube-api-access-hn9gp\") pod \"b96078e0-b889-4b47-9bec-a1411c1fec36\" (UID: \"b96078e0-b889-4b47-9bec-a1411c1fec36\") " Oct 14 09:16:01 crc kubenswrapper[4870]: I1014 09:16:01.632989 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b96078e0-b889-4b47-9bec-a1411c1fec36-ssh-key\") pod \"b96078e0-b889-4b47-9bec-a1411c1fec36\" (UID: \"b96078e0-b889-4b47-9bec-a1411c1fec36\") " Oct 14 09:16:01 crc kubenswrapper[4870]: I1014 09:16:01.638284 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b96078e0-b889-4b47-9bec-a1411c1fec36-tripleo-cleanup-combined-ca-bundle" (OuterVolumeSpecName: "tripleo-cleanup-combined-ca-bundle") pod "b96078e0-b889-4b47-9bec-a1411c1fec36" (UID: "b96078e0-b889-4b47-9bec-a1411c1fec36"). InnerVolumeSpecName "tripleo-cleanup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:16:01 crc kubenswrapper[4870]: I1014 09:16:01.638323 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b96078e0-b889-4b47-9bec-a1411c1fec36-kube-api-access-hn9gp" (OuterVolumeSpecName: "kube-api-access-hn9gp") pod "b96078e0-b889-4b47-9bec-a1411c1fec36" (UID: "b96078e0-b889-4b47-9bec-a1411c1fec36"). InnerVolumeSpecName "kube-api-access-hn9gp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:16:01 crc kubenswrapper[4870]: I1014 09:16:01.662368 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b96078e0-b889-4b47-9bec-a1411c1fec36-inventory" (OuterVolumeSpecName: "inventory") pod "b96078e0-b889-4b47-9bec-a1411c1fec36" (UID: "b96078e0-b889-4b47-9bec-a1411c1fec36"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:16:01 crc kubenswrapper[4870]: I1014 09:16:01.690156 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b96078e0-b889-4b47-9bec-a1411c1fec36-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "b96078e0-b889-4b47-9bec-a1411c1fec36" (UID: "b96078e0-b889-4b47-9bec-a1411c1fec36"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:16:01 crc kubenswrapper[4870]: I1014 09:16:01.770043 4870 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b96078e0-b889-4b47-9bec-a1411c1fec36-inventory\") on node \"crc\" DevicePath \"\"" Oct 14 09:16:01 crc kubenswrapper[4870]: I1014 09:16:01.770110 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hn9gp\" (UniqueName: \"kubernetes.io/projected/b96078e0-b889-4b47-9bec-a1411c1fec36-kube-api-access-hn9gp\") on node \"crc\" DevicePath \"\"" Oct 14 09:16:01 crc kubenswrapper[4870]: I1014 09:16:01.770134 4870 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b96078e0-b889-4b47-9bec-a1411c1fec36-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 14 09:16:01 crc kubenswrapper[4870]: I1014 09:16:01.770164 4870 reconciler_common.go:293] "Volume detached for volume \"tripleo-cleanup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b96078e0-b889-4b47-9bec-a1411c1fec36-tripleo-cleanup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 09:16:02 crc kubenswrapper[4870]: I1014 09:16:02.042301 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-networker-ncc5q" event={"ID":"b96078e0-b889-4b47-9bec-a1411c1fec36","Type":"ContainerDied","Data":"ad2cb4f30a22bda7e820971b083bebbfc50e68fe1e135cbebd8a3f42a824bee1"} Oct 14 09:16:02 crc kubenswrapper[4870]: I1014 09:16:02.042337 4870 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ad2cb4f30a22bda7e820971b083bebbfc50e68fe1e135cbebd8a3f42a824bee1" Oct 14 09:16:02 crc kubenswrapper[4870]: I1014 09:16:02.042366 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-networker-ncc5q" Oct 14 09:16:23 crc kubenswrapper[4870]: I1014 09:16:23.950956 4870 patch_prober.go:28] interesting pod/machine-config-daemon-7tvc7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 09:16:23 crc kubenswrapper[4870]: I1014 09:16:23.951667 4870 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 09:16:31 crc kubenswrapper[4870]: I1014 09:16:31.928958 4870 scope.go:117] "RemoveContainer" containerID="6bfdccb12a2aa548d1f24fc373dd026d9c8de828f33bf089aa12dc1b00379fe4" Oct 14 09:16:31 crc kubenswrapper[4870]: I1014 09:16:31.975160 4870 scope.go:117] "RemoveContainer" containerID="1b75db4079b0f6952970dc908c691c7371d927a7b96d26ad3cdd316012caef08" Oct 14 09:16:53 crc kubenswrapper[4870]: I1014 09:16:53.951103 4870 patch_prober.go:28] interesting pod/machine-config-daemon-7tvc7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 09:16:53 crc kubenswrapper[4870]: I1014 09:16:53.951765 4870 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 09:17:23 crc kubenswrapper[4870]: I1014 09:17:23.951369 4870 patch_prober.go:28] interesting pod/machine-config-daemon-7tvc7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 09:17:23 crc kubenswrapper[4870]: I1014 09:17:23.951977 4870 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 09:17:23 crc kubenswrapper[4870]: I1014 09:17:23.952046 4870 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" Oct 14 09:17:23 crc kubenswrapper[4870]: I1014 09:17:23.952978 4870 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"5a354d0d25c9ddf421bfdb89e3b92b75f73097cc5398e19453b6e98125f9ecf5"} pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 14 09:17:23 crc kubenswrapper[4870]: I1014 09:17:23.953068 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerName="machine-config-daemon" containerID="cri-o://5a354d0d25c9ddf421bfdb89e3b92b75f73097cc5398e19453b6e98125f9ecf5" gracePeriod=600 Oct 14 09:17:24 crc kubenswrapper[4870]: E1014 09:17:24.086784 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 09:17:25 crc kubenswrapper[4870]: I1014 09:17:25.074143 4870 generic.go:334] "Generic (PLEG): container finished" podID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerID="5a354d0d25c9ddf421bfdb89e3b92b75f73097cc5398e19453b6e98125f9ecf5" exitCode=0 Oct 14 09:17:25 crc kubenswrapper[4870]: I1014 09:17:25.074611 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" event={"ID":"874a5e45-dffd-4d17-b609-b7d3ed2eab07","Type":"ContainerDied","Data":"5a354d0d25c9ddf421bfdb89e3b92b75f73097cc5398e19453b6e98125f9ecf5"} Oct 14 09:17:25 crc kubenswrapper[4870]: I1014 09:17:25.074667 4870 scope.go:117] "RemoveContainer" containerID="a9b3d8bfba3e51117bb9f2ff67d8d2b345878f886b38f76fc3c8cf6d7f91c6de" Oct 14 09:17:25 crc kubenswrapper[4870]: I1014 09:17:25.075779 4870 scope.go:117] "RemoveContainer" containerID="5a354d0d25c9ddf421bfdb89e3b92b75f73097cc5398e19453b6e98125f9ecf5" Oct 14 09:17:25 crc kubenswrapper[4870]: E1014 09:17:25.084553 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 09:17:40 crc kubenswrapper[4870]: I1014 09:17:40.033734 4870 scope.go:117] "RemoveContainer" containerID="5a354d0d25c9ddf421bfdb89e3b92b75f73097cc5398e19453b6e98125f9ecf5" Oct 14 09:17:40 crc kubenswrapper[4870]: E1014 09:17:40.034570 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 09:17:51 crc kubenswrapper[4870]: I1014 09:17:51.034897 4870 scope.go:117] "RemoveContainer" containerID="5a354d0d25c9ddf421bfdb89e3b92b75f73097cc5398e19453b6e98125f9ecf5" Oct 14 09:17:51 crc kubenswrapper[4870]: E1014 09:17:51.035743 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 09:18:04 crc kubenswrapper[4870]: I1014 09:18:04.035622 4870 scope.go:117] "RemoveContainer" containerID="5a354d0d25c9ddf421bfdb89e3b92b75f73097cc5398e19453b6e98125f9ecf5" Oct 14 09:18:04 crc kubenswrapper[4870]: E1014 09:18:04.036677 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 09:18:07 crc kubenswrapper[4870]: I1014 09:18:07.050205 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-db-create-zhtzj"] Oct 14 09:18:07 crc kubenswrapper[4870]: I1014 09:18:07.060025 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-db-create-zhtzj"] Oct 14 09:18:09 crc kubenswrapper[4870]: I1014 09:18:09.058079 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="89dc0446-9ebc-48d8-99e9-b77564ef3273" path="/var/lib/kubelet/pods/89dc0446-9ebc-48d8-99e9-b77564ef3273/volumes" Oct 14 09:18:17 crc kubenswrapper[4870]: I1014 09:18:17.058232 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-479c-account-create-nsqjn"] Oct 14 09:18:17 crc kubenswrapper[4870]: I1014 09:18:17.078748 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-479c-account-create-nsqjn"] Oct 14 09:18:19 crc kubenswrapper[4870]: I1014 09:18:19.035042 4870 scope.go:117] "RemoveContainer" containerID="5a354d0d25c9ddf421bfdb89e3b92b75f73097cc5398e19453b6e98125f9ecf5" Oct 14 09:18:19 crc kubenswrapper[4870]: E1014 09:18:19.036693 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 09:18:19 crc kubenswrapper[4870]: I1014 09:18:19.053033 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b251834e-2ef7-4906-8452-4fa62834bf3a" path="/var/lib/kubelet/pods/b251834e-2ef7-4906-8452-4fa62834bf3a/volumes" Oct 14 09:18:28 crc kubenswrapper[4870]: I1014 09:18:28.047369 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-db-sync-zkj55"] Oct 14 09:18:28 crc kubenswrapper[4870]: I1014 09:18:28.055792 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-db-sync-zkj55"] Oct 14 09:18:29 crc kubenswrapper[4870]: I1014 09:18:29.057030 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="deb1d307-0ab5-4053-a68e-513f2d7b0d53" path="/var/lib/kubelet/pods/deb1d307-0ab5-4053-a68e-513f2d7b0d53/volumes" Oct 14 09:18:32 crc kubenswrapper[4870]: I1014 09:18:32.135656 4870 scope.go:117] "RemoveContainer" containerID="fb1fa218005cc3197d0da536bf6930392c1110ab18fdfc98181646cb4f8bf54d" Oct 14 09:18:32 crc kubenswrapper[4870]: I1014 09:18:32.191995 4870 scope.go:117] "RemoveContainer" containerID="9db545bb783b2952b072caa15b2a293309e1ea6f049f61a0357c2dc02fd1dc15" Oct 14 09:18:32 crc kubenswrapper[4870]: I1014 09:18:32.245360 4870 scope.go:117] "RemoveContainer" containerID="15be86f675fd9ba70a551bd735e24b0141a57be493b09079edb6bf957d6ee554" Oct 14 09:18:33 crc kubenswrapper[4870]: I1014 09:18:33.036285 4870 scope.go:117] "RemoveContainer" containerID="5a354d0d25c9ddf421bfdb89e3b92b75f73097cc5398e19453b6e98125f9ecf5" Oct 14 09:18:33 crc kubenswrapper[4870]: E1014 09:18:33.037179 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 09:18:48 crc kubenswrapper[4870]: I1014 09:18:48.034767 4870 scope.go:117] "RemoveContainer" containerID="5a354d0d25c9ddf421bfdb89e3b92b75f73097cc5398e19453b6e98125f9ecf5" Oct 14 09:18:48 crc kubenswrapper[4870]: E1014 09:18:48.036100 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 09:18:48 crc kubenswrapper[4870]: I1014 09:18:48.044060 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-db-create-9495v"] Oct 14 09:18:48 crc kubenswrapper[4870]: I1014 09:18:48.057748 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-db-create-9495v"] Oct 14 09:18:49 crc kubenswrapper[4870]: I1014 09:18:49.056788 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f51823d2-e754-4dce-8ad1-6d3a6752e5ed" path="/var/lib/kubelet/pods/f51823d2-e754-4dce-8ad1-6d3a6752e5ed/volumes" Oct 14 09:18:58 crc kubenswrapper[4870]: I1014 09:18:58.068117 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-cbde-account-create-vcwvz"] Oct 14 09:18:58 crc kubenswrapper[4870]: I1014 09:18:58.081533 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-cbde-account-create-vcwvz"] Oct 14 09:18:59 crc kubenswrapper[4870]: I1014 09:18:59.059907 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43692e8e-fa17-4f76-9a5f-24a28270152f" path="/var/lib/kubelet/pods/43692e8e-fa17-4f76-9a5f-24a28270152f/volumes" Oct 14 09:19:01 crc kubenswrapper[4870]: I1014 09:19:01.034990 4870 scope.go:117] "RemoveContainer" containerID="5a354d0d25c9ddf421bfdb89e3b92b75f73097cc5398e19453b6e98125f9ecf5" Oct 14 09:19:01 crc kubenswrapper[4870]: E1014 09:19:01.035935 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 09:19:09 crc kubenswrapper[4870]: I1014 09:19:09.062605 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-db-sync-wq597"] Oct 14 09:19:09 crc kubenswrapper[4870]: I1014 09:19:09.063634 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-db-sync-wq597"] Oct 14 09:19:11 crc kubenswrapper[4870]: I1014 09:19:11.055103 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ce5d88b9-7a1f-4aac-aa4c-e9d36c4a5a47" path="/var/lib/kubelet/pods/ce5d88b9-7a1f-4aac-aa4c-e9d36c4a5a47/volumes" Oct 14 09:19:14 crc kubenswrapper[4870]: I1014 09:19:14.034600 4870 scope.go:117] "RemoveContainer" containerID="5a354d0d25c9ddf421bfdb89e3b92b75f73097cc5398e19453b6e98125f9ecf5" Oct 14 09:19:14 crc kubenswrapper[4870]: E1014 09:19:14.036479 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 09:19:14 crc kubenswrapper[4870]: I1014 09:19:14.807017 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-rtldx"] Oct 14 09:19:14 crc kubenswrapper[4870]: E1014 09:19:14.808296 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7cd0dbb-80e4-4f4b-9942-5082214636d5" containerName="collect-profiles" Oct 14 09:19:14 crc kubenswrapper[4870]: I1014 09:19:14.808323 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7cd0dbb-80e4-4f4b-9942-5082214636d5" containerName="collect-profiles" Oct 14 09:19:14 crc kubenswrapper[4870]: E1014 09:19:14.808338 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a618a42e-e43b-4512-b665-ad057cda60c5" containerName="extract-content" Oct 14 09:19:14 crc kubenswrapper[4870]: I1014 09:19:14.808348 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="a618a42e-e43b-4512-b665-ad057cda60c5" containerName="extract-content" Oct 14 09:19:14 crc kubenswrapper[4870]: E1014 09:19:14.808378 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b96078e0-b889-4b47-9bec-a1411c1fec36" containerName="tripleo-cleanup-tripleo-cleanup-openstack-networker" Oct 14 09:19:14 crc kubenswrapper[4870]: I1014 09:19:14.808390 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="b96078e0-b889-4b47-9bec-a1411c1fec36" containerName="tripleo-cleanup-tripleo-cleanup-openstack-networker" Oct 14 09:19:14 crc kubenswrapper[4870]: E1014 09:19:14.808407 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a618a42e-e43b-4512-b665-ad057cda60c5" containerName="extract-utilities" Oct 14 09:19:14 crc kubenswrapper[4870]: I1014 09:19:14.808415 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="a618a42e-e43b-4512-b665-ad057cda60c5" containerName="extract-utilities" Oct 14 09:19:14 crc kubenswrapper[4870]: E1014 09:19:14.808426 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a618a42e-e43b-4512-b665-ad057cda60c5" containerName="registry-server" Oct 14 09:19:14 crc kubenswrapper[4870]: I1014 09:19:14.808436 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="a618a42e-e43b-4512-b665-ad057cda60c5" containerName="registry-server" Oct 14 09:19:14 crc kubenswrapper[4870]: I1014 09:19:14.808806 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="e7cd0dbb-80e4-4f4b-9942-5082214636d5" containerName="collect-profiles" Oct 14 09:19:14 crc kubenswrapper[4870]: I1014 09:19:14.808839 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="b96078e0-b889-4b47-9bec-a1411c1fec36" containerName="tripleo-cleanup-tripleo-cleanup-openstack-networker" Oct 14 09:19:14 crc kubenswrapper[4870]: I1014 09:19:14.808886 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="a618a42e-e43b-4512-b665-ad057cda60c5" containerName="registry-server" Oct 14 09:19:14 crc kubenswrapper[4870]: I1014 09:19:14.811397 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rtldx" Oct 14 09:19:14 crc kubenswrapper[4870]: I1014 09:19:14.848087 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rtldx"] Oct 14 09:19:14 crc kubenswrapper[4870]: I1014 09:19:14.918013 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/62176685-e507-4577-bdd4-a2b87200f111-utilities\") pod \"certified-operators-rtldx\" (UID: \"62176685-e507-4577-bdd4-a2b87200f111\") " pod="openshift-marketplace/certified-operators-rtldx" Oct 14 09:19:14 crc kubenswrapper[4870]: I1014 09:19:14.918239 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xkdxv\" (UniqueName: \"kubernetes.io/projected/62176685-e507-4577-bdd4-a2b87200f111-kube-api-access-xkdxv\") pod \"certified-operators-rtldx\" (UID: \"62176685-e507-4577-bdd4-a2b87200f111\") " pod="openshift-marketplace/certified-operators-rtldx" Oct 14 09:19:14 crc kubenswrapper[4870]: I1014 09:19:14.918466 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/62176685-e507-4577-bdd4-a2b87200f111-catalog-content\") pod \"certified-operators-rtldx\" (UID: \"62176685-e507-4577-bdd4-a2b87200f111\") " pod="openshift-marketplace/certified-operators-rtldx" Oct 14 09:19:15 crc kubenswrapper[4870]: I1014 09:19:15.020559 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/62176685-e507-4577-bdd4-a2b87200f111-utilities\") pod \"certified-operators-rtldx\" (UID: \"62176685-e507-4577-bdd4-a2b87200f111\") " pod="openshift-marketplace/certified-operators-rtldx" Oct 14 09:19:15 crc kubenswrapper[4870]: I1014 09:19:15.020695 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xkdxv\" (UniqueName: \"kubernetes.io/projected/62176685-e507-4577-bdd4-a2b87200f111-kube-api-access-xkdxv\") pod \"certified-operators-rtldx\" (UID: \"62176685-e507-4577-bdd4-a2b87200f111\") " pod="openshift-marketplace/certified-operators-rtldx" Oct 14 09:19:15 crc kubenswrapper[4870]: I1014 09:19:15.020749 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/62176685-e507-4577-bdd4-a2b87200f111-catalog-content\") pod \"certified-operators-rtldx\" (UID: \"62176685-e507-4577-bdd4-a2b87200f111\") " pod="openshift-marketplace/certified-operators-rtldx" Oct 14 09:19:15 crc kubenswrapper[4870]: I1014 09:19:15.021217 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/62176685-e507-4577-bdd4-a2b87200f111-utilities\") pod \"certified-operators-rtldx\" (UID: \"62176685-e507-4577-bdd4-a2b87200f111\") " pod="openshift-marketplace/certified-operators-rtldx" Oct 14 09:19:15 crc kubenswrapper[4870]: I1014 09:19:15.021230 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/62176685-e507-4577-bdd4-a2b87200f111-catalog-content\") pod \"certified-operators-rtldx\" (UID: \"62176685-e507-4577-bdd4-a2b87200f111\") " pod="openshift-marketplace/certified-operators-rtldx" Oct 14 09:19:15 crc kubenswrapper[4870]: I1014 09:19:15.046774 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xkdxv\" (UniqueName: \"kubernetes.io/projected/62176685-e507-4577-bdd4-a2b87200f111-kube-api-access-xkdxv\") pod \"certified-operators-rtldx\" (UID: \"62176685-e507-4577-bdd4-a2b87200f111\") " pod="openshift-marketplace/certified-operators-rtldx" Oct 14 09:19:15 crc kubenswrapper[4870]: I1014 09:19:15.139535 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rtldx" Oct 14 09:19:15 crc kubenswrapper[4870]: I1014 09:19:15.664042 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rtldx"] Oct 14 09:19:16 crc kubenswrapper[4870]: I1014 09:19:16.388427 4870 generic.go:334] "Generic (PLEG): container finished" podID="62176685-e507-4577-bdd4-a2b87200f111" containerID="4a3712a1343df8a2ca891b7aff502635c8ddd1315b9deca700eb56419b0c202d" exitCode=0 Oct 14 09:19:16 crc kubenswrapper[4870]: I1014 09:19:16.388560 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rtldx" event={"ID":"62176685-e507-4577-bdd4-a2b87200f111","Type":"ContainerDied","Data":"4a3712a1343df8a2ca891b7aff502635c8ddd1315b9deca700eb56419b0c202d"} Oct 14 09:19:16 crc kubenswrapper[4870]: I1014 09:19:16.388787 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rtldx" event={"ID":"62176685-e507-4577-bdd4-a2b87200f111","Type":"ContainerStarted","Data":"6721e52d8a594e533436dfb2631c26455e3f3d1fb79898128b6011d60fb89837"} Oct 14 09:19:17 crc kubenswrapper[4870]: I1014 09:19:17.410578 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rtldx" event={"ID":"62176685-e507-4577-bdd4-a2b87200f111","Type":"ContainerStarted","Data":"776f20a994123e77e340f2a368032185d253de9651adba349a3e02f4caf4f059"} Oct 14 09:19:19 crc kubenswrapper[4870]: I1014 09:19:19.437107 4870 generic.go:334] "Generic (PLEG): container finished" podID="62176685-e507-4577-bdd4-a2b87200f111" containerID="776f20a994123e77e340f2a368032185d253de9651adba349a3e02f4caf4f059" exitCode=0 Oct 14 09:19:19 crc kubenswrapper[4870]: I1014 09:19:19.437228 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rtldx" event={"ID":"62176685-e507-4577-bdd4-a2b87200f111","Type":"ContainerDied","Data":"776f20a994123e77e340f2a368032185d253de9651adba349a3e02f4caf4f059"} Oct 14 09:19:20 crc kubenswrapper[4870]: I1014 09:19:20.452376 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rtldx" event={"ID":"62176685-e507-4577-bdd4-a2b87200f111","Type":"ContainerStarted","Data":"0c5ed389b8a36aeb6ded6aac78e2a3f80d36df61d18983cb0df804557ad89ed1"} Oct 14 09:19:20 crc kubenswrapper[4870]: I1014 09:19:20.485020 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-rtldx" podStartSLOduration=3.027833015 podStartE2EDuration="6.4849943s" podCreationTimestamp="2025-10-14 09:19:14 +0000 UTC" firstStartedPulling="2025-10-14 09:19:16.39063598 +0000 UTC m=+8292.087996351" lastFinishedPulling="2025-10-14 09:19:19.847797255 +0000 UTC m=+8295.545157636" observedRunningTime="2025-10-14 09:19:20.476609674 +0000 UTC m=+8296.173970085" watchObservedRunningTime="2025-10-14 09:19:20.4849943 +0000 UTC m=+8296.182354711" Oct 14 09:19:24 crc kubenswrapper[4870]: I1014 09:19:24.500125 4870 generic.go:334] "Generic (PLEG): container finished" podID="ea75b5d6-ee4b-4a8d-a3af-4995af7bafb9" containerID="bbcde4b193c5c8d0fc536d73ce644073d25ca7bbc4e4cad9878a8b687c145771" exitCode=0 Oct 14 09:19:24 crc kubenswrapper[4870]: I1014 09:19:24.500201 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-tz9gr" event={"ID":"ea75b5d6-ee4b-4a8d-a3af-4995af7bafb9","Type":"ContainerDied","Data":"bbcde4b193c5c8d0fc536d73ce644073d25ca7bbc4e4cad9878a8b687c145771"} Oct 14 09:19:25 crc kubenswrapper[4870]: I1014 09:19:25.140571 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-rtldx" Oct 14 09:19:25 crc kubenswrapper[4870]: I1014 09:19:25.141550 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-rtldx" Oct 14 09:19:25 crc kubenswrapper[4870]: I1014 09:19:25.199123 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-rtldx" Oct 14 09:19:25 crc kubenswrapper[4870]: I1014 09:19:25.576223 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-rtldx" Oct 14 09:19:25 crc kubenswrapper[4870]: I1014 09:19:25.955046 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-tz9gr" Oct 14 09:19:26 crc kubenswrapper[4870]: I1014 09:19:26.081784 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ea75b5d6-ee4b-4a8d-a3af-4995af7bafb9-ceph\") pod \"ea75b5d6-ee4b-4a8d-a3af-4995af7bafb9\" (UID: \"ea75b5d6-ee4b-4a8d-a3af-4995af7bafb9\") " Oct 14 09:19:26 crc kubenswrapper[4870]: I1014 09:19:26.081999 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tripleo-cleanup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ea75b5d6-ee4b-4a8d-a3af-4995af7bafb9-tripleo-cleanup-combined-ca-bundle\") pod \"ea75b5d6-ee4b-4a8d-a3af-4995af7bafb9\" (UID: \"ea75b5d6-ee4b-4a8d-a3af-4995af7bafb9\") " Oct 14 09:19:26 crc kubenswrapper[4870]: I1014 09:19:26.082041 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ea75b5d6-ee4b-4a8d-a3af-4995af7bafb9-ssh-key\") pod \"ea75b5d6-ee4b-4a8d-a3af-4995af7bafb9\" (UID: \"ea75b5d6-ee4b-4a8d-a3af-4995af7bafb9\") " Oct 14 09:19:26 crc kubenswrapper[4870]: I1014 09:19:26.082123 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k9q8g\" (UniqueName: \"kubernetes.io/projected/ea75b5d6-ee4b-4a8d-a3af-4995af7bafb9-kube-api-access-k9q8g\") pod \"ea75b5d6-ee4b-4a8d-a3af-4995af7bafb9\" (UID: \"ea75b5d6-ee4b-4a8d-a3af-4995af7bafb9\") " Oct 14 09:19:26 crc kubenswrapper[4870]: I1014 09:19:26.082188 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ea75b5d6-ee4b-4a8d-a3af-4995af7bafb9-inventory\") pod \"ea75b5d6-ee4b-4a8d-a3af-4995af7bafb9\" (UID: \"ea75b5d6-ee4b-4a8d-a3af-4995af7bafb9\") " Oct 14 09:19:26 crc kubenswrapper[4870]: I1014 09:19:26.092704 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ea75b5d6-ee4b-4a8d-a3af-4995af7bafb9-tripleo-cleanup-combined-ca-bundle" (OuterVolumeSpecName: "tripleo-cleanup-combined-ca-bundle") pod "ea75b5d6-ee4b-4a8d-a3af-4995af7bafb9" (UID: "ea75b5d6-ee4b-4a8d-a3af-4995af7bafb9"). InnerVolumeSpecName "tripleo-cleanup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:19:26 crc kubenswrapper[4870]: I1014 09:19:26.097123 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ea75b5d6-ee4b-4a8d-a3af-4995af7bafb9-kube-api-access-k9q8g" (OuterVolumeSpecName: "kube-api-access-k9q8g") pod "ea75b5d6-ee4b-4a8d-a3af-4995af7bafb9" (UID: "ea75b5d6-ee4b-4a8d-a3af-4995af7bafb9"). InnerVolumeSpecName "kube-api-access-k9q8g". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:19:26 crc kubenswrapper[4870]: I1014 09:19:26.101970 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ea75b5d6-ee4b-4a8d-a3af-4995af7bafb9-ceph" (OuterVolumeSpecName: "ceph") pod "ea75b5d6-ee4b-4a8d-a3af-4995af7bafb9" (UID: "ea75b5d6-ee4b-4a8d-a3af-4995af7bafb9"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:19:26 crc kubenswrapper[4870]: I1014 09:19:26.112297 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ea75b5d6-ee4b-4a8d-a3af-4995af7bafb9-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "ea75b5d6-ee4b-4a8d-a3af-4995af7bafb9" (UID: "ea75b5d6-ee4b-4a8d-a3af-4995af7bafb9"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:19:26 crc kubenswrapper[4870]: I1014 09:19:26.131393 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ea75b5d6-ee4b-4a8d-a3af-4995af7bafb9-inventory" (OuterVolumeSpecName: "inventory") pod "ea75b5d6-ee4b-4a8d-a3af-4995af7bafb9" (UID: "ea75b5d6-ee4b-4a8d-a3af-4995af7bafb9"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:19:26 crc kubenswrapper[4870]: I1014 09:19:26.184808 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k9q8g\" (UniqueName: \"kubernetes.io/projected/ea75b5d6-ee4b-4a8d-a3af-4995af7bafb9-kube-api-access-k9q8g\") on node \"crc\" DevicePath \"\"" Oct 14 09:19:26 crc kubenswrapper[4870]: I1014 09:19:26.184852 4870 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ea75b5d6-ee4b-4a8d-a3af-4995af7bafb9-inventory\") on node \"crc\" DevicePath \"\"" Oct 14 09:19:26 crc kubenswrapper[4870]: I1014 09:19:26.184865 4870 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ea75b5d6-ee4b-4a8d-a3af-4995af7bafb9-ceph\") on node \"crc\" DevicePath \"\"" Oct 14 09:19:26 crc kubenswrapper[4870]: I1014 09:19:26.184879 4870 reconciler_common.go:293] "Volume detached for volume \"tripleo-cleanup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ea75b5d6-ee4b-4a8d-a3af-4995af7bafb9-tripleo-cleanup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 09:19:26 crc kubenswrapper[4870]: I1014 09:19:26.184896 4870 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ea75b5d6-ee4b-4a8d-a3af-4995af7bafb9-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 14 09:19:26 crc kubenswrapper[4870]: I1014 09:19:26.524105 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-tz9gr" event={"ID":"ea75b5d6-ee4b-4a8d-a3af-4995af7bafb9","Type":"ContainerDied","Data":"8d60c257e44302f2a6d45852ba3f8354e60efdf9992de5d7fefb780858fa094a"} Oct 14 09:19:26 crc kubenswrapper[4870]: I1014 09:19:26.524179 4870 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8d60c257e44302f2a6d45852ba3f8354e60efdf9992de5d7fefb780858fa094a" Oct 14 09:19:26 crc kubenswrapper[4870]: I1014 09:19:26.524389 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-tz9gr" Oct 14 09:19:29 crc kubenswrapper[4870]: I1014 09:19:29.034874 4870 scope.go:117] "RemoveContainer" containerID="5a354d0d25c9ddf421bfdb89e3b92b75f73097cc5398e19453b6e98125f9ecf5" Oct 14 09:19:29 crc kubenswrapper[4870]: E1014 09:19:29.035395 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 09:19:29 crc kubenswrapper[4870]: I1014 09:19:29.584561 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-rtldx"] Oct 14 09:19:29 crc kubenswrapper[4870]: I1014 09:19:29.584833 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-rtldx" podUID="62176685-e507-4577-bdd4-a2b87200f111" containerName="registry-server" containerID="cri-o://0c5ed389b8a36aeb6ded6aac78e2a3f80d36df61d18983cb0df804557ad89ed1" gracePeriod=2 Oct 14 09:19:30 crc kubenswrapper[4870]: I1014 09:19:30.080774 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rtldx" Oct 14 09:19:30 crc kubenswrapper[4870]: I1014 09:19:30.193201 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/62176685-e507-4577-bdd4-a2b87200f111-utilities\") pod \"62176685-e507-4577-bdd4-a2b87200f111\" (UID: \"62176685-e507-4577-bdd4-a2b87200f111\") " Oct 14 09:19:30 crc kubenswrapper[4870]: I1014 09:19:30.193276 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xkdxv\" (UniqueName: \"kubernetes.io/projected/62176685-e507-4577-bdd4-a2b87200f111-kube-api-access-xkdxv\") pod \"62176685-e507-4577-bdd4-a2b87200f111\" (UID: \"62176685-e507-4577-bdd4-a2b87200f111\") " Oct 14 09:19:30 crc kubenswrapper[4870]: I1014 09:19:30.193324 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/62176685-e507-4577-bdd4-a2b87200f111-catalog-content\") pod \"62176685-e507-4577-bdd4-a2b87200f111\" (UID: \"62176685-e507-4577-bdd4-a2b87200f111\") " Oct 14 09:19:30 crc kubenswrapper[4870]: I1014 09:19:30.195058 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/62176685-e507-4577-bdd4-a2b87200f111-utilities" (OuterVolumeSpecName: "utilities") pod "62176685-e507-4577-bdd4-a2b87200f111" (UID: "62176685-e507-4577-bdd4-a2b87200f111"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 09:19:30 crc kubenswrapper[4870]: I1014 09:19:30.198399 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/62176685-e507-4577-bdd4-a2b87200f111-kube-api-access-xkdxv" (OuterVolumeSpecName: "kube-api-access-xkdxv") pod "62176685-e507-4577-bdd4-a2b87200f111" (UID: "62176685-e507-4577-bdd4-a2b87200f111"). InnerVolumeSpecName "kube-api-access-xkdxv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:19:30 crc kubenswrapper[4870]: I1014 09:19:30.258556 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/62176685-e507-4577-bdd4-a2b87200f111-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "62176685-e507-4577-bdd4-a2b87200f111" (UID: "62176685-e507-4577-bdd4-a2b87200f111"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 09:19:30 crc kubenswrapper[4870]: I1014 09:19:30.295498 4870 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/62176685-e507-4577-bdd4-a2b87200f111-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 09:19:30 crc kubenswrapper[4870]: I1014 09:19:30.295547 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xkdxv\" (UniqueName: \"kubernetes.io/projected/62176685-e507-4577-bdd4-a2b87200f111-kube-api-access-xkdxv\") on node \"crc\" DevicePath \"\"" Oct 14 09:19:30 crc kubenswrapper[4870]: I1014 09:19:30.295563 4870 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/62176685-e507-4577-bdd4-a2b87200f111-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 09:19:30 crc kubenswrapper[4870]: I1014 09:19:30.574914 4870 generic.go:334] "Generic (PLEG): container finished" podID="62176685-e507-4577-bdd4-a2b87200f111" containerID="0c5ed389b8a36aeb6ded6aac78e2a3f80d36df61d18983cb0df804557ad89ed1" exitCode=0 Oct 14 09:19:30 crc kubenswrapper[4870]: I1014 09:19:30.575037 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rtldx" Oct 14 09:19:30 crc kubenswrapper[4870]: I1014 09:19:30.575035 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rtldx" event={"ID":"62176685-e507-4577-bdd4-a2b87200f111","Type":"ContainerDied","Data":"0c5ed389b8a36aeb6ded6aac78e2a3f80d36df61d18983cb0df804557ad89ed1"} Oct 14 09:19:30 crc kubenswrapper[4870]: I1014 09:19:30.575495 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rtldx" event={"ID":"62176685-e507-4577-bdd4-a2b87200f111","Type":"ContainerDied","Data":"6721e52d8a594e533436dfb2631c26455e3f3d1fb79898128b6011d60fb89837"} Oct 14 09:19:30 crc kubenswrapper[4870]: I1014 09:19:30.575523 4870 scope.go:117] "RemoveContainer" containerID="0c5ed389b8a36aeb6ded6aac78e2a3f80d36df61d18983cb0df804557ad89ed1" Oct 14 09:19:30 crc kubenswrapper[4870]: I1014 09:19:30.599428 4870 scope.go:117] "RemoveContainer" containerID="776f20a994123e77e340f2a368032185d253de9651adba349a3e02f4caf4f059" Oct 14 09:19:30 crc kubenswrapper[4870]: I1014 09:19:30.640315 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-rtldx"] Oct 14 09:19:30 crc kubenswrapper[4870]: I1014 09:19:30.645348 4870 scope.go:117] "RemoveContainer" containerID="4a3712a1343df8a2ca891b7aff502635c8ddd1315b9deca700eb56419b0c202d" Oct 14 09:19:30 crc kubenswrapper[4870]: I1014 09:19:30.651539 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-rtldx"] Oct 14 09:19:30 crc kubenswrapper[4870]: I1014 09:19:30.708750 4870 scope.go:117] "RemoveContainer" containerID="0c5ed389b8a36aeb6ded6aac78e2a3f80d36df61d18983cb0df804557ad89ed1" Oct 14 09:19:30 crc kubenswrapper[4870]: E1014 09:19:30.709624 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0c5ed389b8a36aeb6ded6aac78e2a3f80d36df61d18983cb0df804557ad89ed1\": container with ID starting with 0c5ed389b8a36aeb6ded6aac78e2a3f80d36df61d18983cb0df804557ad89ed1 not found: ID does not exist" containerID="0c5ed389b8a36aeb6ded6aac78e2a3f80d36df61d18983cb0df804557ad89ed1" Oct 14 09:19:30 crc kubenswrapper[4870]: I1014 09:19:30.709708 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0c5ed389b8a36aeb6ded6aac78e2a3f80d36df61d18983cb0df804557ad89ed1"} err="failed to get container status \"0c5ed389b8a36aeb6ded6aac78e2a3f80d36df61d18983cb0df804557ad89ed1\": rpc error: code = NotFound desc = could not find container \"0c5ed389b8a36aeb6ded6aac78e2a3f80d36df61d18983cb0df804557ad89ed1\": container with ID starting with 0c5ed389b8a36aeb6ded6aac78e2a3f80d36df61d18983cb0df804557ad89ed1 not found: ID does not exist" Oct 14 09:19:30 crc kubenswrapper[4870]: I1014 09:19:30.709763 4870 scope.go:117] "RemoveContainer" containerID="776f20a994123e77e340f2a368032185d253de9651adba349a3e02f4caf4f059" Oct 14 09:19:30 crc kubenswrapper[4870]: E1014 09:19:30.710582 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"776f20a994123e77e340f2a368032185d253de9651adba349a3e02f4caf4f059\": container with ID starting with 776f20a994123e77e340f2a368032185d253de9651adba349a3e02f4caf4f059 not found: ID does not exist" containerID="776f20a994123e77e340f2a368032185d253de9651adba349a3e02f4caf4f059" Oct 14 09:19:30 crc kubenswrapper[4870]: I1014 09:19:30.710642 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"776f20a994123e77e340f2a368032185d253de9651adba349a3e02f4caf4f059"} err="failed to get container status \"776f20a994123e77e340f2a368032185d253de9651adba349a3e02f4caf4f059\": rpc error: code = NotFound desc = could not find container \"776f20a994123e77e340f2a368032185d253de9651adba349a3e02f4caf4f059\": container with ID starting with 776f20a994123e77e340f2a368032185d253de9651adba349a3e02f4caf4f059 not found: ID does not exist" Oct 14 09:19:30 crc kubenswrapper[4870]: I1014 09:19:30.710872 4870 scope.go:117] "RemoveContainer" containerID="4a3712a1343df8a2ca891b7aff502635c8ddd1315b9deca700eb56419b0c202d" Oct 14 09:19:30 crc kubenswrapper[4870]: E1014 09:19:30.711433 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4a3712a1343df8a2ca891b7aff502635c8ddd1315b9deca700eb56419b0c202d\": container with ID starting with 4a3712a1343df8a2ca891b7aff502635c8ddd1315b9deca700eb56419b0c202d not found: ID does not exist" containerID="4a3712a1343df8a2ca891b7aff502635c8ddd1315b9deca700eb56419b0c202d" Oct 14 09:19:30 crc kubenswrapper[4870]: I1014 09:19:30.711515 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4a3712a1343df8a2ca891b7aff502635c8ddd1315b9deca700eb56419b0c202d"} err="failed to get container status \"4a3712a1343df8a2ca891b7aff502635c8ddd1315b9deca700eb56419b0c202d\": rpc error: code = NotFound desc = could not find container \"4a3712a1343df8a2ca891b7aff502635c8ddd1315b9deca700eb56419b0c202d\": container with ID starting with 4a3712a1343df8a2ca891b7aff502635c8ddd1315b9deca700eb56419b0c202d not found: ID does not exist" Oct 14 09:19:31 crc kubenswrapper[4870]: I1014 09:19:31.047938 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="62176685-e507-4577-bdd4-a2b87200f111" path="/var/lib/kubelet/pods/62176685-e507-4577-bdd4-a2b87200f111/volumes" Oct 14 09:19:32 crc kubenswrapper[4870]: I1014 09:19:32.402598 4870 scope.go:117] "RemoveContainer" containerID="c5729775861fa7384aca80299f35064424dffaeb877e688615688092998be592" Oct 14 09:19:32 crc kubenswrapper[4870]: I1014 09:19:32.442809 4870 scope.go:117] "RemoveContainer" containerID="f21671f988885335521bbaced2823640aa1fdc272bab39ec2d2f3a1cb560addd" Oct 14 09:19:32 crc kubenswrapper[4870]: I1014 09:19:32.504647 4870 scope.go:117] "RemoveContainer" containerID="d794ce3cb85d49e9fc592ac73a2b3a39f4b6085a8da52d52c364d6d6048ba80f" Oct 14 09:19:35 crc kubenswrapper[4870]: I1014 09:19:35.589254 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-openstack-openstack-cell1-flgnw"] Oct 14 09:19:35 crc kubenswrapper[4870]: E1014 09:19:35.591526 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62176685-e507-4577-bdd4-a2b87200f111" containerName="extract-content" Oct 14 09:19:35 crc kubenswrapper[4870]: I1014 09:19:35.591658 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="62176685-e507-4577-bdd4-a2b87200f111" containerName="extract-content" Oct 14 09:19:35 crc kubenswrapper[4870]: E1014 09:19:35.591691 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62176685-e507-4577-bdd4-a2b87200f111" containerName="extract-utilities" Oct 14 09:19:35 crc kubenswrapper[4870]: I1014 09:19:35.591702 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="62176685-e507-4577-bdd4-a2b87200f111" containerName="extract-utilities" Oct 14 09:19:35 crc kubenswrapper[4870]: E1014 09:19:35.591733 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea75b5d6-ee4b-4a8d-a3af-4995af7bafb9" containerName="tripleo-cleanup-tripleo-cleanup-openstack-cell1" Oct 14 09:19:35 crc kubenswrapper[4870]: I1014 09:19:35.591773 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea75b5d6-ee4b-4a8d-a3af-4995af7bafb9" containerName="tripleo-cleanup-tripleo-cleanup-openstack-cell1" Oct 14 09:19:35 crc kubenswrapper[4870]: E1014 09:19:35.591810 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62176685-e507-4577-bdd4-a2b87200f111" containerName="registry-server" Oct 14 09:19:35 crc kubenswrapper[4870]: I1014 09:19:35.591818 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="62176685-e507-4577-bdd4-a2b87200f111" containerName="registry-server" Oct 14 09:19:35 crc kubenswrapper[4870]: I1014 09:19:35.592103 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="62176685-e507-4577-bdd4-a2b87200f111" containerName="registry-server" Oct 14 09:19:35 crc kubenswrapper[4870]: I1014 09:19:35.592126 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="ea75b5d6-ee4b-4a8d-a3af-4995af7bafb9" containerName="tripleo-cleanup-tripleo-cleanup-openstack-cell1" Oct 14 09:19:35 crc kubenswrapper[4870]: I1014 09:19:35.593239 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-openstack-openstack-cell1-flgnw" Oct 14 09:19:35 crc kubenswrapper[4870]: I1014 09:19:35.596518 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 14 09:19:35 crc kubenswrapper[4870]: I1014 09:19:35.596908 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 14 09:19:35 crc kubenswrapper[4870]: I1014 09:19:35.598270 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 14 09:19:35 crc kubenswrapper[4870]: I1014 09:19:35.598727 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-wrz7w" Oct 14 09:19:35 crc kubenswrapper[4870]: I1014 09:19:35.601296 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-openstack-openstack-networker-772fx"] Oct 14 09:19:35 crc kubenswrapper[4870]: I1014 09:19:35.603139 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-openstack-openstack-networker-772fx" Oct 14 09:19:35 crc kubenswrapper[4870]: I1014 09:19:35.605974 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-networker-dockercfg-54hzr" Oct 14 09:19:35 crc kubenswrapper[4870]: I1014 09:19:35.606537 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-networker" Oct 14 09:19:35 crc kubenswrapper[4870]: I1014 09:19:35.610255 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-openstack-openstack-cell1-flgnw"] Oct 14 09:19:35 crc kubenswrapper[4870]: I1014 09:19:35.619831 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-openstack-openstack-networker-772fx"] Oct 14 09:19:35 crc kubenswrapper[4870]: I1014 09:19:35.722575 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/139b40af-cdfc-417f-9807-a67634524bcb-inventory\") pod \"bootstrap-openstack-openstack-networker-772fx\" (UID: \"139b40af-cdfc-417f-9807-a67634524bcb\") " pod="openstack/bootstrap-openstack-openstack-networker-772fx" Oct 14 09:19:35 crc kubenswrapper[4870]: I1014 09:19:35.722960 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/295dcd08-48f4-484b-a2e5-5c03db6afd73-bootstrap-combined-ca-bundle\") pod \"bootstrap-openstack-openstack-cell1-flgnw\" (UID: \"295dcd08-48f4-484b-a2e5-5c03db6afd73\") " pod="openstack/bootstrap-openstack-openstack-cell1-flgnw" Oct 14 09:19:35 crc kubenswrapper[4870]: I1014 09:19:35.723212 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/295dcd08-48f4-484b-a2e5-5c03db6afd73-ssh-key\") pod \"bootstrap-openstack-openstack-cell1-flgnw\" (UID: \"295dcd08-48f4-484b-a2e5-5c03db6afd73\") " pod="openstack/bootstrap-openstack-openstack-cell1-flgnw" Oct 14 09:19:35 crc kubenswrapper[4870]: I1014 09:19:35.723369 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/139b40af-cdfc-417f-9807-a67634524bcb-bootstrap-combined-ca-bundle\") pod \"bootstrap-openstack-openstack-networker-772fx\" (UID: \"139b40af-cdfc-417f-9807-a67634524bcb\") " pod="openstack/bootstrap-openstack-openstack-networker-772fx" Oct 14 09:19:35 crc kubenswrapper[4870]: I1014 09:19:35.723571 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r5f89\" (UniqueName: \"kubernetes.io/projected/139b40af-cdfc-417f-9807-a67634524bcb-kube-api-access-r5f89\") pod \"bootstrap-openstack-openstack-networker-772fx\" (UID: \"139b40af-cdfc-417f-9807-a67634524bcb\") " pod="openstack/bootstrap-openstack-openstack-networker-772fx" Oct 14 09:19:35 crc kubenswrapper[4870]: I1014 09:19:35.723772 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/295dcd08-48f4-484b-a2e5-5c03db6afd73-inventory\") pod \"bootstrap-openstack-openstack-cell1-flgnw\" (UID: \"295dcd08-48f4-484b-a2e5-5c03db6afd73\") " pod="openstack/bootstrap-openstack-openstack-cell1-flgnw" Oct 14 09:19:35 crc kubenswrapper[4870]: I1014 09:19:35.723935 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/295dcd08-48f4-484b-a2e5-5c03db6afd73-ceph\") pod \"bootstrap-openstack-openstack-cell1-flgnw\" (UID: \"295dcd08-48f4-484b-a2e5-5c03db6afd73\") " pod="openstack/bootstrap-openstack-openstack-cell1-flgnw" Oct 14 09:19:35 crc kubenswrapper[4870]: I1014 09:19:35.724128 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/139b40af-cdfc-417f-9807-a67634524bcb-ssh-key\") pod \"bootstrap-openstack-openstack-networker-772fx\" (UID: \"139b40af-cdfc-417f-9807-a67634524bcb\") " pod="openstack/bootstrap-openstack-openstack-networker-772fx" Oct 14 09:19:35 crc kubenswrapper[4870]: I1014 09:19:35.724537 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w87ld\" (UniqueName: \"kubernetes.io/projected/295dcd08-48f4-484b-a2e5-5c03db6afd73-kube-api-access-w87ld\") pod \"bootstrap-openstack-openstack-cell1-flgnw\" (UID: \"295dcd08-48f4-484b-a2e5-5c03db6afd73\") " pod="openstack/bootstrap-openstack-openstack-cell1-flgnw" Oct 14 09:19:35 crc kubenswrapper[4870]: I1014 09:19:35.826856 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/139b40af-cdfc-417f-9807-a67634524bcb-bootstrap-combined-ca-bundle\") pod \"bootstrap-openstack-openstack-networker-772fx\" (UID: \"139b40af-cdfc-417f-9807-a67634524bcb\") " pod="openstack/bootstrap-openstack-openstack-networker-772fx" Oct 14 09:19:35 crc kubenswrapper[4870]: I1014 09:19:35.827198 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r5f89\" (UniqueName: \"kubernetes.io/projected/139b40af-cdfc-417f-9807-a67634524bcb-kube-api-access-r5f89\") pod \"bootstrap-openstack-openstack-networker-772fx\" (UID: \"139b40af-cdfc-417f-9807-a67634524bcb\") " pod="openstack/bootstrap-openstack-openstack-networker-772fx" Oct 14 09:19:35 crc kubenswrapper[4870]: I1014 09:19:35.827312 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/295dcd08-48f4-484b-a2e5-5c03db6afd73-inventory\") pod \"bootstrap-openstack-openstack-cell1-flgnw\" (UID: \"295dcd08-48f4-484b-a2e5-5c03db6afd73\") " pod="openstack/bootstrap-openstack-openstack-cell1-flgnw" Oct 14 09:19:35 crc kubenswrapper[4870]: I1014 09:19:35.827401 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/295dcd08-48f4-484b-a2e5-5c03db6afd73-ceph\") pod \"bootstrap-openstack-openstack-cell1-flgnw\" (UID: \"295dcd08-48f4-484b-a2e5-5c03db6afd73\") " pod="openstack/bootstrap-openstack-openstack-cell1-flgnw" Oct 14 09:19:35 crc kubenswrapper[4870]: I1014 09:19:35.827522 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/139b40af-cdfc-417f-9807-a67634524bcb-ssh-key\") pod \"bootstrap-openstack-openstack-networker-772fx\" (UID: \"139b40af-cdfc-417f-9807-a67634524bcb\") " pod="openstack/bootstrap-openstack-openstack-networker-772fx" Oct 14 09:19:35 crc kubenswrapper[4870]: I1014 09:19:35.827654 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w87ld\" (UniqueName: \"kubernetes.io/projected/295dcd08-48f4-484b-a2e5-5c03db6afd73-kube-api-access-w87ld\") pod \"bootstrap-openstack-openstack-cell1-flgnw\" (UID: \"295dcd08-48f4-484b-a2e5-5c03db6afd73\") " pod="openstack/bootstrap-openstack-openstack-cell1-flgnw" Oct 14 09:19:35 crc kubenswrapper[4870]: I1014 09:19:35.827765 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/139b40af-cdfc-417f-9807-a67634524bcb-inventory\") pod \"bootstrap-openstack-openstack-networker-772fx\" (UID: \"139b40af-cdfc-417f-9807-a67634524bcb\") " pod="openstack/bootstrap-openstack-openstack-networker-772fx" Oct 14 09:19:35 crc kubenswrapper[4870]: I1014 09:19:35.827858 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/295dcd08-48f4-484b-a2e5-5c03db6afd73-bootstrap-combined-ca-bundle\") pod \"bootstrap-openstack-openstack-cell1-flgnw\" (UID: \"295dcd08-48f4-484b-a2e5-5c03db6afd73\") " pod="openstack/bootstrap-openstack-openstack-cell1-flgnw" Oct 14 09:19:35 crc kubenswrapper[4870]: I1014 09:19:35.827986 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/295dcd08-48f4-484b-a2e5-5c03db6afd73-ssh-key\") pod \"bootstrap-openstack-openstack-cell1-flgnw\" (UID: \"295dcd08-48f4-484b-a2e5-5c03db6afd73\") " pod="openstack/bootstrap-openstack-openstack-cell1-flgnw" Oct 14 09:19:35 crc kubenswrapper[4870]: I1014 09:19:35.837753 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/295dcd08-48f4-484b-a2e5-5c03db6afd73-ceph\") pod \"bootstrap-openstack-openstack-cell1-flgnw\" (UID: \"295dcd08-48f4-484b-a2e5-5c03db6afd73\") " pod="openstack/bootstrap-openstack-openstack-cell1-flgnw" Oct 14 09:19:35 crc kubenswrapper[4870]: I1014 09:19:35.837786 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/139b40af-cdfc-417f-9807-a67634524bcb-inventory\") pod \"bootstrap-openstack-openstack-networker-772fx\" (UID: \"139b40af-cdfc-417f-9807-a67634524bcb\") " pod="openstack/bootstrap-openstack-openstack-networker-772fx" Oct 14 09:19:35 crc kubenswrapper[4870]: I1014 09:19:35.837758 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/295dcd08-48f4-484b-a2e5-5c03db6afd73-ssh-key\") pod \"bootstrap-openstack-openstack-cell1-flgnw\" (UID: \"295dcd08-48f4-484b-a2e5-5c03db6afd73\") " pod="openstack/bootstrap-openstack-openstack-cell1-flgnw" Oct 14 09:19:35 crc kubenswrapper[4870]: I1014 09:19:35.837825 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/139b40af-cdfc-417f-9807-a67634524bcb-bootstrap-combined-ca-bundle\") pod \"bootstrap-openstack-openstack-networker-772fx\" (UID: \"139b40af-cdfc-417f-9807-a67634524bcb\") " pod="openstack/bootstrap-openstack-openstack-networker-772fx" Oct 14 09:19:35 crc kubenswrapper[4870]: I1014 09:19:35.838286 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/295dcd08-48f4-484b-a2e5-5c03db6afd73-bootstrap-combined-ca-bundle\") pod \"bootstrap-openstack-openstack-cell1-flgnw\" (UID: \"295dcd08-48f4-484b-a2e5-5c03db6afd73\") " pod="openstack/bootstrap-openstack-openstack-cell1-flgnw" Oct 14 09:19:35 crc kubenswrapper[4870]: I1014 09:19:35.838331 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/139b40af-cdfc-417f-9807-a67634524bcb-ssh-key\") pod \"bootstrap-openstack-openstack-networker-772fx\" (UID: \"139b40af-cdfc-417f-9807-a67634524bcb\") " pod="openstack/bootstrap-openstack-openstack-networker-772fx" Oct 14 09:19:35 crc kubenswrapper[4870]: I1014 09:19:35.838518 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/295dcd08-48f4-484b-a2e5-5c03db6afd73-inventory\") pod \"bootstrap-openstack-openstack-cell1-flgnw\" (UID: \"295dcd08-48f4-484b-a2e5-5c03db6afd73\") " pod="openstack/bootstrap-openstack-openstack-cell1-flgnw" Oct 14 09:19:35 crc kubenswrapper[4870]: I1014 09:19:35.848032 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w87ld\" (UniqueName: \"kubernetes.io/projected/295dcd08-48f4-484b-a2e5-5c03db6afd73-kube-api-access-w87ld\") pod \"bootstrap-openstack-openstack-cell1-flgnw\" (UID: \"295dcd08-48f4-484b-a2e5-5c03db6afd73\") " pod="openstack/bootstrap-openstack-openstack-cell1-flgnw" Oct 14 09:19:35 crc kubenswrapper[4870]: I1014 09:19:35.852084 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r5f89\" (UniqueName: \"kubernetes.io/projected/139b40af-cdfc-417f-9807-a67634524bcb-kube-api-access-r5f89\") pod \"bootstrap-openstack-openstack-networker-772fx\" (UID: \"139b40af-cdfc-417f-9807-a67634524bcb\") " pod="openstack/bootstrap-openstack-openstack-networker-772fx" Oct 14 09:19:35 crc kubenswrapper[4870]: I1014 09:19:35.919192 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-openstack-openstack-cell1-flgnw" Oct 14 09:19:35 crc kubenswrapper[4870]: I1014 09:19:35.941968 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-openstack-openstack-networker-772fx" Oct 14 09:19:36 crc kubenswrapper[4870]: I1014 09:19:36.533424 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-openstack-openstack-cell1-flgnw"] Oct 14 09:19:36 crc kubenswrapper[4870]: I1014 09:19:36.662641 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-openstack-openstack-cell1-flgnw" event={"ID":"295dcd08-48f4-484b-a2e5-5c03db6afd73","Type":"ContainerStarted","Data":"d65b3abb557bf7348bf5d24bc8b3f521a7619339016ee57434691f84cf2e2733"} Oct 14 09:19:37 crc kubenswrapper[4870]: I1014 09:19:37.080728 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-openstack-openstack-networker-772fx"] Oct 14 09:19:37 crc kubenswrapper[4870]: W1014 09:19:37.090930 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod139b40af_cdfc_417f_9807_a67634524bcb.slice/crio-075d87b8165e4c7d32d7c4425ad1542a9e959f3b45e9b465d18dee13ae9908f4 WatchSource:0}: Error finding container 075d87b8165e4c7d32d7c4425ad1542a9e959f3b45e9b465d18dee13ae9908f4: Status 404 returned error can't find the container with id 075d87b8165e4c7d32d7c4425ad1542a9e959f3b45e9b465d18dee13ae9908f4 Oct 14 09:19:37 crc kubenswrapper[4870]: I1014 09:19:37.693224 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-openstack-openstack-cell1-flgnw" event={"ID":"295dcd08-48f4-484b-a2e5-5c03db6afd73","Type":"ContainerStarted","Data":"12d1d18223965d1ddf8f69f9bd1a095cb514e1f416ac1ed4446936db064e9a91"} Oct 14 09:19:37 crc kubenswrapper[4870]: I1014 09:19:37.701634 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-openstack-openstack-networker-772fx" event={"ID":"139b40af-cdfc-417f-9807-a67634524bcb","Type":"ContainerStarted","Data":"075d87b8165e4c7d32d7c4425ad1542a9e959f3b45e9b465d18dee13ae9908f4"} Oct 14 09:19:37 crc kubenswrapper[4870]: I1014 09:19:37.730522 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-openstack-openstack-cell1-flgnw" podStartSLOduration=1.91488103 podStartE2EDuration="2.730499264s" podCreationTimestamp="2025-10-14 09:19:35 +0000 UTC" firstStartedPulling="2025-10-14 09:19:36.536068461 +0000 UTC m=+8312.233428832" lastFinishedPulling="2025-10-14 09:19:37.351686685 +0000 UTC m=+8313.049047066" observedRunningTime="2025-10-14 09:19:37.718426377 +0000 UTC m=+8313.415786758" watchObservedRunningTime="2025-10-14 09:19:37.730499264 +0000 UTC m=+8313.427859635" Oct 14 09:19:38 crc kubenswrapper[4870]: I1014 09:19:38.728174 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-openstack-openstack-networker-772fx" event={"ID":"139b40af-cdfc-417f-9807-a67634524bcb","Type":"ContainerStarted","Data":"e5d646542c3c92c7f6289f8f633e27cc9722c4057fa3c79caf06ed6d43009653"} Oct 14 09:19:38 crc kubenswrapper[4870]: I1014 09:19:38.775882 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-openstack-openstack-networker-772fx" podStartSLOduration=3.085255441 podStartE2EDuration="3.77585533s" podCreationTimestamp="2025-10-14 09:19:35 +0000 UTC" firstStartedPulling="2025-10-14 09:19:37.095525444 +0000 UTC m=+8312.792885825" lastFinishedPulling="2025-10-14 09:19:37.786125333 +0000 UTC m=+8313.483485714" observedRunningTime="2025-10-14 09:19:38.753234253 +0000 UTC m=+8314.450594664" watchObservedRunningTime="2025-10-14 09:19:38.77585533 +0000 UTC m=+8314.473215711" Oct 14 09:19:40 crc kubenswrapper[4870]: I1014 09:19:40.033526 4870 scope.go:117] "RemoveContainer" containerID="5a354d0d25c9ddf421bfdb89e3b92b75f73097cc5398e19453b6e98125f9ecf5" Oct 14 09:19:40 crc kubenswrapper[4870]: E1014 09:19:40.034051 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 09:19:54 crc kubenswrapper[4870]: I1014 09:19:54.034436 4870 scope.go:117] "RemoveContainer" containerID="5a354d0d25c9ddf421bfdb89e3b92b75f73097cc5398e19453b6e98125f9ecf5" Oct 14 09:19:54 crc kubenswrapper[4870]: E1014 09:19:54.035837 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 09:20:08 crc kubenswrapper[4870]: I1014 09:20:08.035347 4870 scope.go:117] "RemoveContainer" containerID="5a354d0d25c9ddf421bfdb89e3b92b75f73097cc5398e19453b6e98125f9ecf5" Oct 14 09:20:08 crc kubenswrapper[4870]: E1014 09:20:08.038309 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 09:20:22 crc kubenswrapper[4870]: I1014 09:20:22.034205 4870 scope.go:117] "RemoveContainer" containerID="5a354d0d25c9ddf421bfdb89e3b92b75f73097cc5398e19453b6e98125f9ecf5" Oct 14 09:20:22 crc kubenswrapper[4870]: E1014 09:20:22.035904 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 09:20:34 crc kubenswrapper[4870]: I1014 09:20:34.036060 4870 scope.go:117] "RemoveContainer" containerID="5a354d0d25c9ddf421bfdb89e3b92b75f73097cc5398e19453b6e98125f9ecf5" Oct 14 09:20:34 crc kubenswrapper[4870]: E1014 09:20:34.037531 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 09:20:48 crc kubenswrapper[4870]: I1014 09:20:48.034476 4870 scope.go:117] "RemoveContainer" containerID="5a354d0d25c9ddf421bfdb89e3b92b75f73097cc5398e19453b6e98125f9ecf5" Oct 14 09:20:48 crc kubenswrapper[4870]: E1014 09:20:48.035706 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 09:21:00 crc kubenswrapper[4870]: I1014 09:21:00.035159 4870 scope.go:117] "RemoveContainer" containerID="5a354d0d25c9ddf421bfdb89e3b92b75f73097cc5398e19453b6e98125f9ecf5" Oct 14 09:21:00 crc kubenswrapper[4870]: E1014 09:21:00.036435 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 09:21:13 crc kubenswrapper[4870]: I1014 09:21:13.033893 4870 scope.go:117] "RemoveContainer" containerID="5a354d0d25c9ddf421bfdb89e3b92b75f73097cc5398e19453b6e98125f9ecf5" Oct 14 09:21:13 crc kubenswrapper[4870]: E1014 09:21:13.034871 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 09:21:25 crc kubenswrapper[4870]: I1014 09:21:25.078719 4870 scope.go:117] "RemoveContainer" containerID="5a354d0d25c9ddf421bfdb89e3b92b75f73097cc5398e19453b6e98125f9ecf5" Oct 14 09:21:25 crc kubenswrapper[4870]: E1014 09:21:25.079660 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 09:21:36 crc kubenswrapper[4870]: I1014 09:21:36.034500 4870 scope.go:117] "RemoveContainer" containerID="5a354d0d25c9ddf421bfdb89e3b92b75f73097cc5398e19453b6e98125f9ecf5" Oct 14 09:21:36 crc kubenswrapper[4870]: E1014 09:21:36.035184 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 09:21:47 crc kubenswrapper[4870]: I1014 09:21:47.034164 4870 scope.go:117] "RemoveContainer" containerID="5a354d0d25c9ddf421bfdb89e3b92b75f73097cc5398e19453b6e98125f9ecf5" Oct 14 09:21:47 crc kubenswrapper[4870]: E1014 09:21:47.035086 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 09:22:02 crc kubenswrapper[4870]: I1014 09:22:02.036358 4870 scope.go:117] "RemoveContainer" containerID="5a354d0d25c9ddf421bfdb89e3b92b75f73097cc5398e19453b6e98125f9ecf5" Oct 14 09:22:02 crc kubenswrapper[4870]: E1014 09:22:02.037426 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 09:22:17 crc kubenswrapper[4870]: I1014 09:22:17.036474 4870 scope.go:117] "RemoveContainer" containerID="5a354d0d25c9ddf421bfdb89e3b92b75f73097cc5398e19453b6e98125f9ecf5" Oct 14 09:22:17 crc kubenswrapper[4870]: E1014 09:22:17.040029 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 09:22:31 crc kubenswrapper[4870]: I1014 09:22:31.037949 4870 scope.go:117] "RemoveContainer" containerID="5a354d0d25c9ddf421bfdb89e3b92b75f73097cc5398e19453b6e98125f9ecf5" Oct 14 09:22:31 crc kubenswrapper[4870]: I1014 09:22:31.704835 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" event={"ID":"874a5e45-dffd-4d17-b609-b7d3ed2eab07","Type":"ContainerStarted","Data":"c92a8ddf378f6420fa6f685f5a67ccca0eb5b167c543057b9d31affa78843af8"} Oct 14 09:22:33 crc kubenswrapper[4870]: I1014 09:22:33.726070 4870 generic.go:334] "Generic (PLEG): container finished" podID="295dcd08-48f4-484b-a2e5-5c03db6afd73" containerID="12d1d18223965d1ddf8f69f9bd1a095cb514e1f416ac1ed4446936db064e9a91" exitCode=0 Oct 14 09:22:33 crc kubenswrapper[4870]: I1014 09:22:33.726176 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-openstack-openstack-cell1-flgnw" event={"ID":"295dcd08-48f4-484b-a2e5-5c03db6afd73","Type":"ContainerDied","Data":"12d1d18223965d1ddf8f69f9bd1a095cb514e1f416ac1ed4446936db064e9a91"} Oct 14 09:22:35 crc kubenswrapper[4870]: I1014 09:22:35.228929 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-openstack-openstack-cell1-flgnw" Oct 14 09:22:35 crc kubenswrapper[4870]: I1014 09:22:35.390540 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/295dcd08-48f4-484b-a2e5-5c03db6afd73-ceph\") pod \"295dcd08-48f4-484b-a2e5-5c03db6afd73\" (UID: \"295dcd08-48f4-484b-a2e5-5c03db6afd73\") " Oct 14 09:22:35 crc kubenswrapper[4870]: I1014 09:22:35.391015 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/295dcd08-48f4-484b-a2e5-5c03db6afd73-ssh-key\") pod \"295dcd08-48f4-484b-a2e5-5c03db6afd73\" (UID: \"295dcd08-48f4-484b-a2e5-5c03db6afd73\") " Oct 14 09:22:35 crc kubenswrapper[4870]: I1014 09:22:35.391126 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/295dcd08-48f4-484b-a2e5-5c03db6afd73-bootstrap-combined-ca-bundle\") pod \"295dcd08-48f4-484b-a2e5-5c03db6afd73\" (UID: \"295dcd08-48f4-484b-a2e5-5c03db6afd73\") " Oct 14 09:22:35 crc kubenswrapper[4870]: I1014 09:22:35.391267 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/295dcd08-48f4-484b-a2e5-5c03db6afd73-inventory\") pod \"295dcd08-48f4-484b-a2e5-5c03db6afd73\" (UID: \"295dcd08-48f4-484b-a2e5-5c03db6afd73\") " Oct 14 09:22:35 crc kubenswrapper[4870]: I1014 09:22:35.391456 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w87ld\" (UniqueName: \"kubernetes.io/projected/295dcd08-48f4-484b-a2e5-5c03db6afd73-kube-api-access-w87ld\") pod \"295dcd08-48f4-484b-a2e5-5c03db6afd73\" (UID: \"295dcd08-48f4-484b-a2e5-5c03db6afd73\") " Oct 14 09:22:35 crc kubenswrapper[4870]: I1014 09:22:35.397739 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/295dcd08-48f4-484b-a2e5-5c03db6afd73-ceph" (OuterVolumeSpecName: "ceph") pod "295dcd08-48f4-484b-a2e5-5c03db6afd73" (UID: "295dcd08-48f4-484b-a2e5-5c03db6afd73"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:22:35 crc kubenswrapper[4870]: I1014 09:22:35.400765 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/295dcd08-48f4-484b-a2e5-5c03db6afd73-kube-api-access-w87ld" (OuterVolumeSpecName: "kube-api-access-w87ld") pod "295dcd08-48f4-484b-a2e5-5c03db6afd73" (UID: "295dcd08-48f4-484b-a2e5-5c03db6afd73"). InnerVolumeSpecName "kube-api-access-w87ld". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:22:35 crc kubenswrapper[4870]: I1014 09:22:35.411176 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/295dcd08-48f4-484b-a2e5-5c03db6afd73-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "295dcd08-48f4-484b-a2e5-5c03db6afd73" (UID: "295dcd08-48f4-484b-a2e5-5c03db6afd73"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:22:35 crc kubenswrapper[4870]: I1014 09:22:35.427258 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/295dcd08-48f4-484b-a2e5-5c03db6afd73-inventory" (OuterVolumeSpecName: "inventory") pod "295dcd08-48f4-484b-a2e5-5c03db6afd73" (UID: "295dcd08-48f4-484b-a2e5-5c03db6afd73"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:22:35 crc kubenswrapper[4870]: I1014 09:22:35.428041 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/295dcd08-48f4-484b-a2e5-5c03db6afd73-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "295dcd08-48f4-484b-a2e5-5c03db6afd73" (UID: "295dcd08-48f4-484b-a2e5-5c03db6afd73"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:22:35 crc kubenswrapper[4870]: I1014 09:22:35.493952 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w87ld\" (UniqueName: \"kubernetes.io/projected/295dcd08-48f4-484b-a2e5-5c03db6afd73-kube-api-access-w87ld\") on node \"crc\" DevicePath \"\"" Oct 14 09:22:35 crc kubenswrapper[4870]: I1014 09:22:35.493986 4870 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/295dcd08-48f4-484b-a2e5-5c03db6afd73-ceph\") on node \"crc\" DevicePath \"\"" Oct 14 09:22:35 crc kubenswrapper[4870]: I1014 09:22:35.493995 4870 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/295dcd08-48f4-484b-a2e5-5c03db6afd73-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 14 09:22:35 crc kubenswrapper[4870]: I1014 09:22:35.494004 4870 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/295dcd08-48f4-484b-a2e5-5c03db6afd73-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 09:22:35 crc kubenswrapper[4870]: I1014 09:22:35.494013 4870 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/295dcd08-48f4-484b-a2e5-5c03db6afd73-inventory\") on node \"crc\" DevicePath \"\"" Oct 14 09:22:35 crc kubenswrapper[4870]: I1014 09:22:35.751682 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-openstack-openstack-cell1-flgnw" event={"ID":"295dcd08-48f4-484b-a2e5-5c03db6afd73","Type":"ContainerDied","Data":"d65b3abb557bf7348bf5d24bc8b3f521a7619339016ee57434691f84cf2e2733"} Oct 14 09:22:35 crc kubenswrapper[4870]: I1014 09:22:35.751737 4870 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d65b3abb557bf7348bf5d24bc8b3f521a7619339016ee57434691f84cf2e2733" Oct 14 09:22:35 crc kubenswrapper[4870]: I1014 09:22:35.751823 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-openstack-openstack-cell1-flgnw" Oct 14 09:22:35 crc kubenswrapper[4870]: I1014 09:22:35.850818 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/download-cache-openstack-openstack-cell1-cfvhm"] Oct 14 09:22:35 crc kubenswrapper[4870]: E1014 09:22:35.851301 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="295dcd08-48f4-484b-a2e5-5c03db6afd73" containerName="bootstrap-openstack-openstack-cell1" Oct 14 09:22:35 crc kubenswrapper[4870]: I1014 09:22:35.851319 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="295dcd08-48f4-484b-a2e5-5c03db6afd73" containerName="bootstrap-openstack-openstack-cell1" Oct 14 09:22:35 crc kubenswrapper[4870]: I1014 09:22:35.851647 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="295dcd08-48f4-484b-a2e5-5c03db6afd73" containerName="bootstrap-openstack-openstack-cell1" Oct 14 09:22:35 crc kubenswrapper[4870]: I1014 09:22:35.852547 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-openstack-openstack-cell1-cfvhm" Oct 14 09:22:35 crc kubenswrapper[4870]: I1014 09:22:35.857306 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 14 09:22:35 crc kubenswrapper[4870]: I1014 09:22:35.866078 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-openstack-openstack-cell1-cfvhm"] Oct 14 09:22:35 crc kubenswrapper[4870]: I1014 09:22:35.868321 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-wrz7w" Oct 14 09:22:35 crc kubenswrapper[4870]: I1014 09:22:35.902659 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e4ecb749-bbab-40eb-82be-bd39305cb787-ceph\") pod \"download-cache-openstack-openstack-cell1-cfvhm\" (UID: \"e4ecb749-bbab-40eb-82be-bd39305cb787\") " pod="openstack/download-cache-openstack-openstack-cell1-cfvhm" Oct 14 09:22:35 crc kubenswrapper[4870]: I1014 09:22:35.902741 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e4ecb749-bbab-40eb-82be-bd39305cb787-ssh-key\") pod \"download-cache-openstack-openstack-cell1-cfvhm\" (UID: \"e4ecb749-bbab-40eb-82be-bd39305cb787\") " pod="openstack/download-cache-openstack-openstack-cell1-cfvhm" Oct 14 09:22:35 crc kubenswrapper[4870]: I1014 09:22:35.902770 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xxlb6\" (UniqueName: \"kubernetes.io/projected/e4ecb749-bbab-40eb-82be-bd39305cb787-kube-api-access-xxlb6\") pod \"download-cache-openstack-openstack-cell1-cfvhm\" (UID: \"e4ecb749-bbab-40eb-82be-bd39305cb787\") " pod="openstack/download-cache-openstack-openstack-cell1-cfvhm" Oct 14 09:22:35 crc kubenswrapper[4870]: I1014 09:22:35.902809 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e4ecb749-bbab-40eb-82be-bd39305cb787-inventory\") pod \"download-cache-openstack-openstack-cell1-cfvhm\" (UID: \"e4ecb749-bbab-40eb-82be-bd39305cb787\") " pod="openstack/download-cache-openstack-openstack-cell1-cfvhm" Oct 14 09:22:36 crc kubenswrapper[4870]: I1014 09:22:36.004405 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e4ecb749-bbab-40eb-82be-bd39305cb787-ceph\") pod \"download-cache-openstack-openstack-cell1-cfvhm\" (UID: \"e4ecb749-bbab-40eb-82be-bd39305cb787\") " pod="openstack/download-cache-openstack-openstack-cell1-cfvhm" Oct 14 09:22:36 crc kubenswrapper[4870]: I1014 09:22:36.005196 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e4ecb749-bbab-40eb-82be-bd39305cb787-ssh-key\") pod \"download-cache-openstack-openstack-cell1-cfvhm\" (UID: \"e4ecb749-bbab-40eb-82be-bd39305cb787\") " pod="openstack/download-cache-openstack-openstack-cell1-cfvhm" Oct 14 09:22:36 crc kubenswrapper[4870]: I1014 09:22:36.005238 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xxlb6\" (UniqueName: \"kubernetes.io/projected/e4ecb749-bbab-40eb-82be-bd39305cb787-kube-api-access-xxlb6\") pod \"download-cache-openstack-openstack-cell1-cfvhm\" (UID: \"e4ecb749-bbab-40eb-82be-bd39305cb787\") " pod="openstack/download-cache-openstack-openstack-cell1-cfvhm" Oct 14 09:22:36 crc kubenswrapper[4870]: I1014 09:22:36.005294 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e4ecb749-bbab-40eb-82be-bd39305cb787-inventory\") pod \"download-cache-openstack-openstack-cell1-cfvhm\" (UID: \"e4ecb749-bbab-40eb-82be-bd39305cb787\") " pod="openstack/download-cache-openstack-openstack-cell1-cfvhm" Oct 14 09:22:36 crc kubenswrapper[4870]: I1014 09:22:36.009460 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e4ecb749-bbab-40eb-82be-bd39305cb787-inventory\") pod \"download-cache-openstack-openstack-cell1-cfvhm\" (UID: \"e4ecb749-bbab-40eb-82be-bd39305cb787\") " pod="openstack/download-cache-openstack-openstack-cell1-cfvhm" Oct 14 09:22:36 crc kubenswrapper[4870]: I1014 09:22:36.009533 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e4ecb749-bbab-40eb-82be-bd39305cb787-ssh-key\") pod \"download-cache-openstack-openstack-cell1-cfvhm\" (UID: \"e4ecb749-bbab-40eb-82be-bd39305cb787\") " pod="openstack/download-cache-openstack-openstack-cell1-cfvhm" Oct 14 09:22:36 crc kubenswrapper[4870]: I1014 09:22:36.009578 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e4ecb749-bbab-40eb-82be-bd39305cb787-ceph\") pod \"download-cache-openstack-openstack-cell1-cfvhm\" (UID: \"e4ecb749-bbab-40eb-82be-bd39305cb787\") " pod="openstack/download-cache-openstack-openstack-cell1-cfvhm" Oct 14 09:22:36 crc kubenswrapper[4870]: I1014 09:22:36.023156 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xxlb6\" (UniqueName: \"kubernetes.io/projected/e4ecb749-bbab-40eb-82be-bd39305cb787-kube-api-access-xxlb6\") pod \"download-cache-openstack-openstack-cell1-cfvhm\" (UID: \"e4ecb749-bbab-40eb-82be-bd39305cb787\") " pod="openstack/download-cache-openstack-openstack-cell1-cfvhm" Oct 14 09:22:36 crc kubenswrapper[4870]: I1014 09:22:36.174491 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-openstack-openstack-cell1-cfvhm" Oct 14 09:22:36 crc kubenswrapper[4870]: I1014 09:22:36.714717 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-openstack-openstack-cell1-cfvhm"] Oct 14 09:22:36 crc kubenswrapper[4870]: W1014 09:22:36.717712 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode4ecb749_bbab_40eb_82be_bd39305cb787.slice/crio-136029c2da65fcffe17de432c7c879033c3baa52072ba0dc6880c046f29a4f78 WatchSource:0}: Error finding container 136029c2da65fcffe17de432c7c879033c3baa52072ba0dc6880c046f29a4f78: Status 404 returned error can't find the container with id 136029c2da65fcffe17de432c7c879033c3baa52072ba0dc6880c046f29a4f78 Oct 14 09:22:36 crc kubenswrapper[4870]: I1014 09:22:36.719659 4870 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 14 09:22:36 crc kubenswrapper[4870]: I1014 09:22:36.788639 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-openstack-openstack-cell1-cfvhm" event={"ID":"e4ecb749-bbab-40eb-82be-bd39305cb787","Type":"ContainerStarted","Data":"136029c2da65fcffe17de432c7c879033c3baa52072ba0dc6880c046f29a4f78"} Oct 14 09:22:37 crc kubenswrapper[4870]: I1014 09:22:37.803010 4870 generic.go:334] "Generic (PLEG): container finished" podID="139b40af-cdfc-417f-9807-a67634524bcb" containerID="e5d646542c3c92c7f6289f8f633e27cc9722c4057fa3c79caf06ed6d43009653" exitCode=0 Oct 14 09:22:37 crc kubenswrapper[4870]: I1014 09:22:37.803133 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-openstack-openstack-networker-772fx" event={"ID":"139b40af-cdfc-417f-9807-a67634524bcb","Type":"ContainerDied","Data":"e5d646542c3c92c7f6289f8f633e27cc9722c4057fa3c79caf06ed6d43009653"} Oct 14 09:22:37 crc kubenswrapper[4870]: I1014 09:22:37.806231 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-openstack-openstack-cell1-cfvhm" event={"ID":"e4ecb749-bbab-40eb-82be-bd39305cb787","Type":"ContainerStarted","Data":"98cdfd2cd44f60b9a7f9b548102592c58d141e98b1b69f0c8b02c6800ec8f35c"} Oct 14 09:22:37 crc kubenswrapper[4870]: I1014 09:22:37.876862 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/download-cache-openstack-openstack-cell1-cfvhm" podStartSLOduration=2.078321428 podStartE2EDuration="2.876835492s" podCreationTimestamp="2025-10-14 09:22:35 +0000 UTC" firstStartedPulling="2025-10-14 09:22:36.719450071 +0000 UTC m=+8492.416810442" lastFinishedPulling="2025-10-14 09:22:37.517964135 +0000 UTC m=+8493.215324506" observedRunningTime="2025-10-14 09:22:37.862346746 +0000 UTC m=+8493.559707137" watchObservedRunningTime="2025-10-14 09:22:37.876835492 +0000 UTC m=+8493.574195863" Oct 14 09:22:39 crc kubenswrapper[4870]: I1014 09:22:39.289929 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-openstack-openstack-networker-772fx" Oct 14 09:22:39 crc kubenswrapper[4870]: I1014 09:22:39.480777 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/139b40af-cdfc-417f-9807-a67634524bcb-inventory\") pod \"139b40af-cdfc-417f-9807-a67634524bcb\" (UID: \"139b40af-cdfc-417f-9807-a67634524bcb\") " Oct 14 09:22:39 crc kubenswrapper[4870]: I1014 09:22:39.480870 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/139b40af-cdfc-417f-9807-a67634524bcb-ssh-key\") pod \"139b40af-cdfc-417f-9807-a67634524bcb\" (UID: \"139b40af-cdfc-417f-9807-a67634524bcb\") " Oct 14 09:22:39 crc kubenswrapper[4870]: I1014 09:22:39.480901 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/139b40af-cdfc-417f-9807-a67634524bcb-bootstrap-combined-ca-bundle\") pod \"139b40af-cdfc-417f-9807-a67634524bcb\" (UID: \"139b40af-cdfc-417f-9807-a67634524bcb\") " Oct 14 09:22:39 crc kubenswrapper[4870]: I1014 09:22:39.480937 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r5f89\" (UniqueName: \"kubernetes.io/projected/139b40af-cdfc-417f-9807-a67634524bcb-kube-api-access-r5f89\") pod \"139b40af-cdfc-417f-9807-a67634524bcb\" (UID: \"139b40af-cdfc-417f-9807-a67634524bcb\") " Oct 14 09:22:39 crc kubenswrapper[4870]: I1014 09:22:39.486293 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/139b40af-cdfc-417f-9807-a67634524bcb-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "139b40af-cdfc-417f-9807-a67634524bcb" (UID: "139b40af-cdfc-417f-9807-a67634524bcb"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:22:39 crc kubenswrapper[4870]: I1014 09:22:39.486540 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/139b40af-cdfc-417f-9807-a67634524bcb-kube-api-access-r5f89" (OuterVolumeSpecName: "kube-api-access-r5f89") pod "139b40af-cdfc-417f-9807-a67634524bcb" (UID: "139b40af-cdfc-417f-9807-a67634524bcb"). InnerVolumeSpecName "kube-api-access-r5f89". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:22:39 crc kubenswrapper[4870]: I1014 09:22:39.515558 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/139b40af-cdfc-417f-9807-a67634524bcb-inventory" (OuterVolumeSpecName: "inventory") pod "139b40af-cdfc-417f-9807-a67634524bcb" (UID: "139b40af-cdfc-417f-9807-a67634524bcb"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:22:39 crc kubenswrapper[4870]: I1014 09:22:39.523778 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/139b40af-cdfc-417f-9807-a67634524bcb-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "139b40af-cdfc-417f-9807-a67634524bcb" (UID: "139b40af-cdfc-417f-9807-a67634524bcb"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:22:39 crc kubenswrapper[4870]: I1014 09:22:39.583387 4870 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/139b40af-cdfc-417f-9807-a67634524bcb-inventory\") on node \"crc\" DevicePath \"\"" Oct 14 09:22:39 crc kubenswrapper[4870]: I1014 09:22:39.583627 4870 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/139b40af-cdfc-417f-9807-a67634524bcb-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 14 09:22:39 crc kubenswrapper[4870]: I1014 09:22:39.583705 4870 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/139b40af-cdfc-417f-9807-a67634524bcb-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 09:22:39 crc kubenswrapper[4870]: I1014 09:22:39.583778 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r5f89\" (UniqueName: \"kubernetes.io/projected/139b40af-cdfc-417f-9807-a67634524bcb-kube-api-access-r5f89\") on node \"crc\" DevicePath \"\"" Oct 14 09:22:39 crc kubenswrapper[4870]: I1014 09:22:39.829593 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-openstack-openstack-networker-772fx" event={"ID":"139b40af-cdfc-417f-9807-a67634524bcb","Type":"ContainerDied","Data":"075d87b8165e4c7d32d7c4425ad1542a9e959f3b45e9b465d18dee13ae9908f4"} Oct 14 09:22:39 crc kubenswrapper[4870]: I1014 09:22:39.829924 4870 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="075d87b8165e4c7d32d7c4425ad1542a9e959f3b45e9b465d18dee13ae9908f4" Oct 14 09:22:39 crc kubenswrapper[4870]: I1014 09:22:39.830042 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-openstack-openstack-networker-772fx" Oct 14 09:22:39 crc kubenswrapper[4870]: I1014 09:22:39.905596 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/download-cache-openstack-openstack-networker-pqffh"] Oct 14 09:22:39 crc kubenswrapper[4870]: E1014 09:22:39.906028 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="139b40af-cdfc-417f-9807-a67634524bcb" containerName="bootstrap-openstack-openstack-networker" Oct 14 09:22:39 crc kubenswrapper[4870]: I1014 09:22:39.906045 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="139b40af-cdfc-417f-9807-a67634524bcb" containerName="bootstrap-openstack-openstack-networker" Oct 14 09:22:39 crc kubenswrapper[4870]: I1014 09:22:39.906235 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="139b40af-cdfc-417f-9807-a67634524bcb" containerName="bootstrap-openstack-openstack-networker" Oct 14 09:22:39 crc kubenswrapper[4870]: I1014 09:22:39.906931 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-openstack-openstack-networker-pqffh" Oct 14 09:22:39 crc kubenswrapper[4870]: I1014 09:22:39.909022 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-networker-dockercfg-54hzr" Oct 14 09:22:39 crc kubenswrapper[4870]: I1014 09:22:39.909635 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-networker" Oct 14 09:22:39 crc kubenswrapper[4870]: I1014 09:22:39.922884 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-openstack-openstack-networker-pqffh"] Oct 14 09:22:40 crc kubenswrapper[4870]: I1014 09:22:40.093756 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-78nkr\" (UniqueName: \"kubernetes.io/projected/ea545890-c197-4826-8f15-380af2a44ad9-kube-api-access-78nkr\") pod \"download-cache-openstack-openstack-networker-pqffh\" (UID: \"ea545890-c197-4826-8f15-380af2a44ad9\") " pod="openstack/download-cache-openstack-openstack-networker-pqffh" Oct 14 09:22:40 crc kubenswrapper[4870]: I1014 09:22:40.094247 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ea545890-c197-4826-8f15-380af2a44ad9-ssh-key\") pod \"download-cache-openstack-openstack-networker-pqffh\" (UID: \"ea545890-c197-4826-8f15-380af2a44ad9\") " pod="openstack/download-cache-openstack-openstack-networker-pqffh" Oct 14 09:22:40 crc kubenswrapper[4870]: I1014 09:22:40.094627 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ea545890-c197-4826-8f15-380af2a44ad9-inventory\") pod \"download-cache-openstack-openstack-networker-pqffh\" (UID: \"ea545890-c197-4826-8f15-380af2a44ad9\") " pod="openstack/download-cache-openstack-openstack-networker-pqffh" Oct 14 09:22:40 crc kubenswrapper[4870]: I1014 09:22:40.196615 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ea545890-c197-4826-8f15-380af2a44ad9-inventory\") pod \"download-cache-openstack-openstack-networker-pqffh\" (UID: \"ea545890-c197-4826-8f15-380af2a44ad9\") " pod="openstack/download-cache-openstack-openstack-networker-pqffh" Oct 14 09:22:40 crc kubenswrapper[4870]: I1014 09:22:40.197675 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-78nkr\" (UniqueName: \"kubernetes.io/projected/ea545890-c197-4826-8f15-380af2a44ad9-kube-api-access-78nkr\") pod \"download-cache-openstack-openstack-networker-pqffh\" (UID: \"ea545890-c197-4826-8f15-380af2a44ad9\") " pod="openstack/download-cache-openstack-openstack-networker-pqffh" Oct 14 09:22:40 crc kubenswrapper[4870]: I1014 09:22:40.197739 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ea545890-c197-4826-8f15-380af2a44ad9-ssh-key\") pod \"download-cache-openstack-openstack-networker-pqffh\" (UID: \"ea545890-c197-4826-8f15-380af2a44ad9\") " pod="openstack/download-cache-openstack-openstack-networker-pqffh" Oct 14 09:22:40 crc kubenswrapper[4870]: I1014 09:22:40.217728 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ea545890-c197-4826-8f15-380af2a44ad9-ssh-key\") pod \"download-cache-openstack-openstack-networker-pqffh\" (UID: \"ea545890-c197-4826-8f15-380af2a44ad9\") " pod="openstack/download-cache-openstack-openstack-networker-pqffh" Oct 14 09:22:40 crc kubenswrapper[4870]: I1014 09:22:40.218330 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ea545890-c197-4826-8f15-380af2a44ad9-inventory\") pod \"download-cache-openstack-openstack-networker-pqffh\" (UID: \"ea545890-c197-4826-8f15-380af2a44ad9\") " pod="openstack/download-cache-openstack-openstack-networker-pqffh" Oct 14 09:22:40 crc kubenswrapper[4870]: I1014 09:22:40.232978 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-78nkr\" (UniqueName: \"kubernetes.io/projected/ea545890-c197-4826-8f15-380af2a44ad9-kube-api-access-78nkr\") pod \"download-cache-openstack-openstack-networker-pqffh\" (UID: \"ea545890-c197-4826-8f15-380af2a44ad9\") " pod="openstack/download-cache-openstack-openstack-networker-pqffh" Oct 14 09:22:40 crc kubenswrapper[4870]: I1014 09:22:40.529335 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-openstack-openstack-networker-pqffh" Oct 14 09:22:41 crc kubenswrapper[4870]: I1014 09:22:41.108188 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-openstack-openstack-networker-pqffh"] Oct 14 09:22:41 crc kubenswrapper[4870]: W1014 09:22:41.116429 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podea545890_c197_4826_8f15_380af2a44ad9.slice/crio-21f8212e9fb5ac22b9d9b822b4c11c95ccbf1ac1b56b8aee95d93eebc52608bf WatchSource:0}: Error finding container 21f8212e9fb5ac22b9d9b822b4c11c95ccbf1ac1b56b8aee95d93eebc52608bf: Status 404 returned error can't find the container with id 21f8212e9fb5ac22b9d9b822b4c11c95ccbf1ac1b56b8aee95d93eebc52608bf Oct 14 09:22:41 crc kubenswrapper[4870]: I1014 09:22:41.849601 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-openstack-openstack-networker-pqffh" event={"ID":"ea545890-c197-4826-8f15-380af2a44ad9","Type":"ContainerStarted","Data":"21f8212e9fb5ac22b9d9b822b4c11c95ccbf1ac1b56b8aee95d93eebc52608bf"} Oct 14 09:22:42 crc kubenswrapper[4870]: I1014 09:22:42.861169 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-openstack-openstack-networker-pqffh" event={"ID":"ea545890-c197-4826-8f15-380af2a44ad9","Type":"ContainerStarted","Data":"e17558a034462eca6c5adb2aa53ccb7499124e62a19e47e45f5e60e3a2783b8c"} Oct 14 09:22:42 crc kubenswrapper[4870]: I1014 09:22:42.882116 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/download-cache-openstack-openstack-networker-pqffh" podStartSLOduration=3.365103494 podStartE2EDuration="3.882097291s" podCreationTimestamp="2025-10-14 09:22:39 +0000 UTC" firstStartedPulling="2025-10-14 09:22:41.119178114 +0000 UTC m=+8496.816538485" lastFinishedPulling="2025-10-14 09:22:41.636171901 +0000 UTC m=+8497.333532282" observedRunningTime="2025-10-14 09:22:42.879941478 +0000 UTC m=+8498.577301859" watchObservedRunningTime="2025-10-14 09:22:42.882097291 +0000 UTC m=+8498.579457672" Oct 14 09:22:46 crc kubenswrapper[4870]: I1014 09:22:46.074843 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-67qhp"] Oct 14 09:22:46 crc kubenswrapper[4870]: I1014 09:22:46.077278 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-67qhp" Oct 14 09:22:46 crc kubenswrapper[4870]: I1014 09:22:46.098812 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-67qhp"] Oct 14 09:22:46 crc kubenswrapper[4870]: I1014 09:22:46.244798 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e284b3e4-c59d-4f0c-8fe3-d675893c3979-utilities\") pod \"community-operators-67qhp\" (UID: \"e284b3e4-c59d-4f0c-8fe3-d675893c3979\") " pod="openshift-marketplace/community-operators-67qhp" Oct 14 09:22:46 crc kubenswrapper[4870]: I1014 09:22:46.245009 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rbcdj\" (UniqueName: \"kubernetes.io/projected/e284b3e4-c59d-4f0c-8fe3-d675893c3979-kube-api-access-rbcdj\") pod \"community-operators-67qhp\" (UID: \"e284b3e4-c59d-4f0c-8fe3-d675893c3979\") " pod="openshift-marketplace/community-operators-67qhp" Oct 14 09:22:46 crc kubenswrapper[4870]: I1014 09:22:46.245123 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e284b3e4-c59d-4f0c-8fe3-d675893c3979-catalog-content\") pod \"community-operators-67qhp\" (UID: \"e284b3e4-c59d-4f0c-8fe3-d675893c3979\") " pod="openshift-marketplace/community-operators-67qhp" Oct 14 09:22:46 crc kubenswrapper[4870]: I1014 09:22:46.347192 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e284b3e4-c59d-4f0c-8fe3-d675893c3979-utilities\") pod \"community-operators-67qhp\" (UID: \"e284b3e4-c59d-4f0c-8fe3-d675893c3979\") " pod="openshift-marketplace/community-operators-67qhp" Oct 14 09:22:46 crc kubenswrapper[4870]: I1014 09:22:46.347354 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rbcdj\" (UniqueName: \"kubernetes.io/projected/e284b3e4-c59d-4f0c-8fe3-d675893c3979-kube-api-access-rbcdj\") pod \"community-operators-67qhp\" (UID: \"e284b3e4-c59d-4f0c-8fe3-d675893c3979\") " pod="openshift-marketplace/community-operators-67qhp" Oct 14 09:22:46 crc kubenswrapper[4870]: I1014 09:22:46.347454 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e284b3e4-c59d-4f0c-8fe3-d675893c3979-catalog-content\") pod \"community-operators-67qhp\" (UID: \"e284b3e4-c59d-4f0c-8fe3-d675893c3979\") " pod="openshift-marketplace/community-operators-67qhp" Oct 14 09:22:46 crc kubenswrapper[4870]: I1014 09:22:46.347810 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e284b3e4-c59d-4f0c-8fe3-d675893c3979-utilities\") pod \"community-operators-67qhp\" (UID: \"e284b3e4-c59d-4f0c-8fe3-d675893c3979\") " pod="openshift-marketplace/community-operators-67qhp" Oct 14 09:22:46 crc kubenswrapper[4870]: I1014 09:22:46.347921 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e284b3e4-c59d-4f0c-8fe3-d675893c3979-catalog-content\") pod \"community-operators-67qhp\" (UID: \"e284b3e4-c59d-4f0c-8fe3-d675893c3979\") " pod="openshift-marketplace/community-operators-67qhp" Oct 14 09:22:46 crc kubenswrapper[4870]: I1014 09:22:46.379484 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rbcdj\" (UniqueName: \"kubernetes.io/projected/e284b3e4-c59d-4f0c-8fe3-d675893c3979-kube-api-access-rbcdj\") pod \"community-operators-67qhp\" (UID: \"e284b3e4-c59d-4f0c-8fe3-d675893c3979\") " pod="openshift-marketplace/community-operators-67qhp" Oct 14 09:22:46 crc kubenswrapper[4870]: I1014 09:22:46.421039 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-67qhp" Oct 14 09:22:47 crc kubenswrapper[4870]: I1014 09:22:47.015857 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-67qhp"] Oct 14 09:22:47 crc kubenswrapper[4870]: I1014 09:22:47.911691 4870 generic.go:334] "Generic (PLEG): container finished" podID="e284b3e4-c59d-4f0c-8fe3-d675893c3979" containerID="65873e3572fa95eb55491d804c911350e8a682027db59433122bc36206ea2749" exitCode=0 Oct 14 09:22:47 crc kubenswrapper[4870]: I1014 09:22:47.911821 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-67qhp" event={"ID":"e284b3e4-c59d-4f0c-8fe3-d675893c3979","Type":"ContainerDied","Data":"65873e3572fa95eb55491d804c911350e8a682027db59433122bc36206ea2749"} Oct 14 09:22:47 crc kubenswrapper[4870]: I1014 09:22:47.912191 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-67qhp" event={"ID":"e284b3e4-c59d-4f0c-8fe3-d675893c3979","Type":"ContainerStarted","Data":"5c49b80067d9b753e46ace29805ee59f40d1d7fc1f6f32c8228834a002c59e07"} Oct 14 09:22:49 crc kubenswrapper[4870]: I1014 09:22:49.981392 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-67qhp" event={"ID":"e284b3e4-c59d-4f0c-8fe3-d675893c3979","Type":"ContainerStarted","Data":"20a5524e4c85c815641565c5c9a0e15d4e7246a234804bfd9cafe3d674a88b86"} Oct 14 09:22:50 crc kubenswrapper[4870]: I1014 09:22:50.996706 4870 generic.go:334] "Generic (PLEG): container finished" podID="e284b3e4-c59d-4f0c-8fe3-d675893c3979" containerID="20a5524e4c85c815641565c5c9a0e15d4e7246a234804bfd9cafe3d674a88b86" exitCode=0 Oct 14 09:22:50 crc kubenswrapper[4870]: I1014 09:22:50.996836 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-67qhp" event={"ID":"e284b3e4-c59d-4f0c-8fe3-d675893c3979","Type":"ContainerDied","Data":"20a5524e4c85c815641565c5c9a0e15d4e7246a234804bfd9cafe3d674a88b86"} Oct 14 09:22:52 crc kubenswrapper[4870]: I1014 09:22:52.009338 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-67qhp" event={"ID":"e284b3e4-c59d-4f0c-8fe3-d675893c3979","Type":"ContainerStarted","Data":"b1ad048ead9e51043f61d4a6a5795d4795263881de995b69555715d2602a7d0b"} Oct 14 09:22:52 crc kubenswrapper[4870]: I1014 09:22:52.033234 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-67qhp" podStartSLOduration=2.507208937 podStartE2EDuration="6.033218026s" podCreationTimestamp="2025-10-14 09:22:46 +0000 UTC" firstStartedPulling="2025-10-14 09:22:47.915998724 +0000 UTC m=+8503.613359095" lastFinishedPulling="2025-10-14 09:22:51.442007803 +0000 UTC m=+8507.139368184" observedRunningTime="2025-10-14 09:22:52.026841219 +0000 UTC m=+8507.724201590" watchObservedRunningTime="2025-10-14 09:22:52.033218026 +0000 UTC m=+8507.730578387" Oct 14 09:22:56 crc kubenswrapper[4870]: I1014 09:22:56.421866 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-67qhp" Oct 14 09:22:56 crc kubenswrapper[4870]: I1014 09:22:56.422595 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-67qhp" Oct 14 09:22:56 crc kubenswrapper[4870]: I1014 09:22:56.485221 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-67qhp" Oct 14 09:22:57 crc kubenswrapper[4870]: I1014 09:22:57.148510 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-67qhp" Oct 14 09:22:57 crc kubenswrapper[4870]: I1014 09:22:57.217133 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-67qhp"] Oct 14 09:22:59 crc kubenswrapper[4870]: I1014 09:22:59.094327 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-67qhp" podUID="e284b3e4-c59d-4f0c-8fe3-d675893c3979" containerName="registry-server" containerID="cri-o://b1ad048ead9e51043f61d4a6a5795d4795263881de995b69555715d2602a7d0b" gracePeriod=2 Oct 14 09:22:59 crc kubenswrapper[4870]: I1014 09:22:59.622566 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-67qhp" Oct 14 09:22:59 crc kubenswrapper[4870]: I1014 09:22:59.805939 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e284b3e4-c59d-4f0c-8fe3-d675893c3979-utilities\") pod \"e284b3e4-c59d-4f0c-8fe3-d675893c3979\" (UID: \"e284b3e4-c59d-4f0c-8fe3-d675893c3979\") " Oct 14 09:22:59 crc kubenswrapper[4870]: I1014 09:22:59.806213 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rbcdj\" (UniqueName: \"kubernetes.io/projected/e284b3e4-c59d-4f0c-8fe3-d675893c3979-kube-api-access-rbcdj\") pod \"e284b3e4-c59d-4f0c-8fe3-d675893c3979\" (UID: \"e284b3e4-c59d-4f0c-8fe3-d675893c3979\") " Oct 14 09:22:59 crc kubenswrapper[4870]: I1014 09:22:59.806274 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e284b3e4-c59d-4f0c-8fe3-d675893c3979-catalog-content\") pod \"e284b3e4-c59d-4f0c-8fe3-d675893c3979\" (UID: \"e284b3e4-c59d-4f0c-8fe3-d675893c3979\") " Oct 14 09:22:59 crc kubenswrapper[4870]: I1014 09:22:59.806956 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e284b3e4-c59d-4f0c-8fe3-d675893c3979-utilities" (OuterVolumeSpecName: "utilities") pod "e284b3e4-c59d-4f0c-8fe3-d675893c3979" (UID: "e284b3e4-c59d-4f0c-8fe3-d675893c3979"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 09:22:59 crc kubenswrapper[4870]: I1014 09:22:59.813585 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e284b3e4-c59d-4f0c-8fe3-d675893c3979-kube-api-access-rbcdj" (OuterVolumeSpecName: "kube-api-access-rbcdj") pod "e284b3e4-c59d-4f0c-8fe3-d675893c3979" (UID: "e284b3e4-c59d-4f0c-8fe3-d675893c3979"). InnerVolumeSpecName "kube-api-access-rbcdj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:22:59 crc kubenswrapper[4870]: I1014 09:22:59.873671 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e284b3e4-c59d-4f0c-8fe3-d675893c3979-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e284b3e4-c59d-4f0c-8fe3-d675893c3979" (UID: "e284b3e4-c59d-4f0c-8fe3-d675893c3979"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 09:22:59 crc kubenswrapper[4870]: I1014 09:22:59.909071 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rbcdj\" (UniqueName: \"kubernetes.io/projected/e284b3e4-c59d-4f0c-8fe3-d675893c3979-kube-api-access-rbcdj\") on node \"crc\" DevicePath \"\"" Oct 14 09:22:59 crc kubenswrapper[4870]: I1014 09:22:59.909117 4870 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e284b3e4-c59d-4f0c-8fe3-d675893c3979-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 09:22:59 crc kubenswrapper[4870]: I1014 09:22:59.909131 4870 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e284b3e4-c59d-4f0c-8fe3-d675893c3979-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 09:23:00 crc kubenswrapper[4870]: I1014 09:23:00.105474 4870 generic.go:334] "Generic (PLEG): container finished" podID="e284b3e4-c59d-4f0c-8fe3-d675893c3979" containerID="b1ad048ead9e51043f61d4a6a5795d4795263881de995b69555715d2602a7d0b" exitCode=0 Oct 14 09:23:00 crc kubenswrapper[4870]: I1014 09:23:00.105574 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-67qhp" Oct 14 09:23:00 crc kubenswrapper[4870]: I1014 09:23:00.105578 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-67qhp" event={"ID":"e284b3e4-c59d-4f0c-8fe3-d675893c3979","Type":"ContainerDied","Data":"b1ad048ead9e51043f61d4a6a5795d4795263881de995b69555715d2602a7d0b"} Oct 14 09:23:00 crc kubenswrapper[4870]: I1014 09:23:00.107045 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-67qhp" event={"ID":"e284b3e4-c59d-4f0c-8fe3-d675893c3979","Type":"ContainerDied","Data":"5c49b80067d9b753e46ace29805ee59f40d1d7fc1f6f32c8228834a002c59e07"} Oct 14 09:23:00 crc kubenswrapper[4870]: I1014 09:23:00.107123 4870 scope.go:117] "RemoveContainer" containerID="b1ad048ead9e51043f61d4a6a5795d4795263881de995b69555715d2602a7d0b" Oct 14 09:23:00 crc kubenswrapper[4870]: I1014 09:23:00.141881 4870 scope.go:117] "RemoveContainer" containerID="20a5524e4c85c815641565c5c9a0e15d4e7246a234804bfd9cafe3d674a88b86" Oct 14 09:23:00 crc kubenswrapper[4870]: I1014 09:23:00.148550 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-67qhp"] Oct 14 09:23:00 crc kubenswrapper[4870]: I1014 09:23:00.159994 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-67qhp"] Oct 14 09:23:00 crc kubenswrapper[4870]: I1014 09:23:00.171941 4870 scope.go:117] "RemoveContainer" containerID="65873e3572fa95eb55491d804c911350e8a682027db59433122bc36206ea2749" Oct 14 09:23:00 crc kubenswrapper[4870]: I1014 09:23:00.245425 4870 scope.go:117] "RemoveContainer" containerID="b1ad048ead9e51043f61d4a6a5795d4795263881de995b69555715d2602a7d0b" Oct 14 09:23:00 crc kubenswrapper[4870]: E1014 09:23:00.245942 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b1ad048ead9e51043f61d4a6a5795d4795263881de995b69555715d2602a7d0b\": container with ID starting with b1ad048ead9e51043f61d4a6a5795d4795263881de995b69555715d2602a7d0b not found: ID does not exist" containerID="b1ad048ead9e51043f61d4a6a5795d4795263881de995b69555715d2602a7d0b" Oct 14 09:23:00 crc kubenswrapper[4870]: I1014 09:23:00.245975 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b1ad048ead9e51043f61d4a6a5795d4795263881de995b69555715d2602a7d0b"} err="failed to get container status \"b1ad048ead9e51043f61d4a6a5795d4795263881de995b69555715d2602a7d0b\": rpc error: code = NotFound desc = could not find container \"b1ad048ead9e51043f61d4a6a5795d4795263881de995b69555715d2602a7d0b\": container with ID starting with b1ad048ead9e51043f61d4a6a5795d4795263881de995b69555715d2602a7d0b not found: ID does not exist" Oct 14 09:23:00 crc kubenswrapper[4870]: I1014 09:23:00.245997 4870 scope.go:117] "RemoveContainer" containerID="20a5524e4c85c815641565c5c9a0e15d4e7246a234804bfd9cafe3d674a88b86" Oct 14 09:23:00 crc kubenswrapper[4870]: E1014 09:23:00.246278 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"20a5524e4c85c815641565c5c9a0e15d4e7246a234804bfd9cafe3d674a88b86\": container with ID starting with 20a5524e4c85c815641565c5c9a0e15d4e7246a234804bfd9cafe3d674a88b86 not found: ID does not exist" containerID="20a5524e4c85c815641565c5c9a0e15d4e7246a234804bfd9cafe3d674a88b86" Oct 14 09:23:00 crc kubenswrapper[4870]: I1014 09:23:00.246299 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"20a5524e4c85c815641565c5c9a0e15d4e7246a234804bfd9cafe3d674a88b86"} err="failed to get container status \"20a5524e4c85c815641565c5c9a0e15d4e7246a234804bfd9cafe3d674a88b86\": rpc error: code = NotFound desc = could not find container \"20a5524e4c85c815641565c5c9a0e15d4e7246a234804bfd9cafe3d674a88b86\": container with ID starting with 20a5524e4c85c815641565c5c9a0e15d4e7246a234804bfd9cafe3d674a88b86 not found: ID does not exist" Oct 14 09:23:00 crc kubenswrapper[4870]: I1014 09:23:00.246312 4870 scope.go:117] "RemoveContainer" containerID="65873e3572fa95eb55491d804c911350e8a682027db59433122bc36206ea2749" Oct 14 09:23:00 crc kubenswrapper[4870]: E1014 09:23:00.246593 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"65873e3572fa95eb55491d804c911350e8a682027db59433122bc36206ea2749\": container with ID starting with 65873e3572fa95eb55491d804c911350e8a682027db59433122bc36206ea2749 not found: ID does not exist" containerID="65873e3572fa95eb55491d804c911350e8a682027db59433122bc36206ea2749" Oct 14 09:23:00 crc kubenswrapper[4870]: I1014 09:23:00.246613 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"65873e3572fa95eb55491d804c911350e8a682027db59433122bc36206ea2749"} err="failed to get container status \"65873e3572fa95eb55491d804c911350e8a682027db59433122bc36206ea2749\": rpc error: code = NotFound desc = could not find container \"65873e3572fa95eb55491d804c911350e8a682027db59433122bc36206ea2749\": container with ID starting with 65873e3572fa95eb55491d804c911350e8a682027db59433122bc36206ea2749 not found: ID does not exist" Oct 14 09:23:01 crc kubenswrapper[4870]: I1014 09:23:01.046574 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e284b3e4-c59d-4f0c-8fe3-d675893c3979" path="/var/lib/kubelet/pods/e284b3e4-c59d-4f0c-8fe3-d675893c3979/volumes" Oct 14 09:23:48 crc kubenswrapper[4870]: I1014 09:23:48.626422 4870 generic.go:334] "Generic (PLEG): container finished" podID="ea545890-c197-4826-8f15-380af2a44ad9" containerID="e17558a034462eca6c5adb2aa53ccb7499124e62a19e47e45f5e60e3a2783b8c" exitCode=0 Oct 14 09:23:48 crc kubenswrapper[4870]: I1014 09:23:48.626497 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-openstack-openstack-networker-pqffh" event={"ID":"ea545890-c197-4826-8f15-380af2a44ad9","Type":"ContainerDied","Data":"e17558a034462eca6c5adb2aa53ccb7499124e62a19e47e45f5e60e3a2783b8c"} Oct 14 09:23:50 crc kubenswrapper[4870]: I1014 09:23:50.143979 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-openstack-openstack-networker-pqffh" Oct 14 09:23:50 crc kubenswrapper[4870]: I1014 09:23:50.331753 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ea545890-c197-4826-8f15-380af2a44ad9-ssh-key\") pod \"ea545890-c197-4826-8f15-380af2a44ad9\" (UID: \"ea545890-c197-4826-8f15-380af2a44ad9\") " Oct 14 09:23:50 crc kubenswrapper[4870]: I1014 09:23:50.331975 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-78nkr\" (UniqueName: \"kubernetes.io/projected/ea545890-c197-4826-8f15-380af2a44ad9-kube-api-access-78nkr\") pod \"ea545890-c197-4826-8f15-380af2a44ad9\" (UID: \"ea545890-c197-4826-8f15-380af2a44ad9\") " Oct 14 09:23:50 crc kubenswrapper[4870]: I1014 09:23:50.332149 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ea545890-c197-4826-8f15-380af2a44ad9-inventory\") pod \"ea545890-c197-4826-8f15-380af2a44ad9\" (UID: \"ea545890-c197-4826-8f15-380af2a44ad9\") " Oct 14 09:23:50 crc kubenswrapper[4870]: I1014 09:23:50.341913 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ea545890-c197-4826-8f15-380af2a44ad9-kube-api-access-78nkr" (OuterVolumeSpecName: "kube-api-access-78nkr") pod "ea545890-c197-4826-8f15-380af2a44ad9" (UID: "ea545890-c197-4826-8f15-380af2a44ad9"). InnerVolumeSpecName "kube-api-access-78nkr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:23:50 crc kubenswrapper[4870]: I1014 09:23:50.363459 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ea545890-c197-4826-8f15-380af2a44ad9-inventory" (OuterVolumeSpecName: "inventory") pod "ea545890-c197-4826-8f15-380af2a44ad9" (UID: "ea545890-c197-4826-8f15-380af2a44ad9"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:23:50 crc kubenswrapper[4870]: I1014 09:23:50.368854 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ea545890-c197-4826-8f15-380af2a44ad9-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "ea545890-c197-4826-8f15-380af2a44ad9" (UID: "ea545890-c197-4826-8f15-380af2a44ad9"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:23:50 crc kubenswrapper[4870]: I1014 09:23:50.435072 4870 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ea545890-c197-4826-8f15-380af2a44ad9-inventory\") on node \"crc\" DevicePath \"\"" Oct 14 09:23:50 crc kubenswrapper[4870]: I1014 09:23:50.435122 4870 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ea545890-c197-4826-8f15-380af2a44ad9-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 14 09:23:50 crc kubenswrapper[4870]: I1014 09:23:50.435134 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-78nkr\" (UniqueName: \"kubernetes.io/projected/ea545890-c197-4826-8f15-380af2a44ad9-kube-api-access-78nkr\") on node \"crc\" DevicePath \"\"" Oct 14 09:23:50 crc kubenswrapper[4870]: I1014 09:23:50.650990 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-openstack-openstack-networker-pqffh" event={"ID":"ea545890-c197-4826-8f15-380af2a44ad9","Type":"ContainerDied","Data":"21f8212e9fb5ac22b9d9b822b4c11c95ccbf1ac1b56b8aee95d93eebc52608bf"} Oct 14 09:23:50 crc kubenswrapper[4870]: I1014 09:23:50.651038 4870 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="21f8212e9fb5ac22b9d9b822b4c11c95ccbf1ac1b56b8aee95d93eebc52608bf" Oct 14 09:23:50 crc kubenswrapper[4870]: I1014 09:23:50.651047 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-openstack-openstack-networker-pqffh" Oct 14 09:23:50 crc kubenswrapper[4870]: I1014 09:23:50.761153 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-openstack-openstack-networker-8chgx"] Oct 14 09:23:50 crc kubenswrapper[4870]: E1014 09:23:50.761765 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea545890-c197-4826-8f15-380af2a44ad9" containerName="download-cache-openstack-openstack-networker" Oct 14 09:23:50 crc kubenswrapper[4870]: I1014 09:23:50.761790 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea545890-c197-4826-8f15-380af2a44ad9" containerName="download-cache-openstack-openstack-networker" Oct 14 09:23:50 crc kubenswrapper[4870]: E1014 09:23:50.761843 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e284b3e4-c59d-4f0c-8fe3-d675893c3979" containerName="extract-utilities" Oct 14 09:23:50 crc kubenswrapper[4870]: I1014 09:23:50.761853 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="e284b3e4-c59d-4f0c-8fe3-d675893c3979" containerName="extract-utilities" Oct 14 09:23:50 crc kubenswrapper[4870]: E1014 09:23:50.761873 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e284b3e4-c59d-4f0c-8fe3-d675893c3979" containerName="registry-server" Oct 14 09:23:50 crc kubenswrapper[4870]: I1014 09:23:50.761881 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="e284b3e4-c59d-4f0c-8fe3-d675893c3979" containerName="registry-server" Oct 14 09:23:50 crc kubenswrapper[4870]: E1014 09:23:50.761906 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e284b3e4-c59d-4f0c-8fe3-d675893c3979" containerName="extract-content" Oct 14 09:23:50 crc kubenswrapper[4870]: I1014 09:23:50.761917 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="e284b3e4-c59d-4f0c-8fe3-d675893c3979" containerName="extract-content" Oct 14 09:23:50 crc kubenswrapper[4870]: I1014 09:23:50.762190 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="ea545890-c197-4826-8f15-380af2a44ad9" containerName="download-cache-openstack-openstack-networker" Oct 14 09:23:50 crc kubenswrapper[4870]: I1014 09:23:50.762241 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="e284b3e4-c59d-4f0c-8fe3-d675893c3979" containerName="registry-server" Oct 14 09:23:50 crc kubenswrapper[4870]: I1014 09:23:50.764422 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-openstack-openstack-networker-8chgx" Oct 14 09:23:50 crc kubenswrapper[4870]: I1014 09:23:50.767913 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-networker" Oct 14 09:23:50 crc kubenswrapper[4870]: I1014 09:23:50.768064 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-networker-dockercfg-54hzr" Oct 14 09:23:50 crc kubenswrapper[4870]: I1014 09:23:50.772412 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-openstack-openstack-networker-8chgx"] Oct 14 09:23:50 crc kubenswrapper[4870]: I1014 09:23:50.841801 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5a77a035-4684-43ed-b3a1-1e9dfa53dd63-ssh-key\") pod \"configure-network-openstack-openstack-networker-8chgx\" (UID: \"5a77a035-4684-43ed-b3a1-1e9dfa53dd63\") " pod="openstack/configure-network-openstack-openstack-networker-8chgx" Oct 14 09:23:50 crc kubenswrapper[4870]: I1014 09:23:50.842356 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5a77a035-4684-43ed-b3a1-1e9dfa53dd63-inventory\") pod \"configure-network-openstack-openstack-networker-8chgx\" (UID: \"5a77a035-4684-43ed-b3a1-1e9dfa53dd63\") " pod="openstack/configure-network-openstack-openstack-networker-8chgx" Oct 14 09:23:50 crc kubenswrapper[4870]: I1014 09:23:50.842501 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jrlsz\" (UniqueName: \"kubernetes.io/projected/5a77a035-4684-43ed-b3a1-1e9dfa53dd63-kube-api-access-jrlsz\") pod \"configure-network-openstack-openstack-networker-8chgx\" (UID: \"5a77a035-4684-43ed-b3a1-1e9dfa53dd63\") " pod="openstack/configure-network-openstack-openstack-networker-8chgx" Oct 14 09:23:50 crc kubenswrapper[4870]: I1014 09:23:50.944794 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5a77a035-4684-43ed-b3a1-1e9dfa53dd63-inventory\") pod \"configure-network-openstack-openstack-networker-8chgx\" (UID: \"5a77a035-4684-43ed-b3a1-1e9dfa53dd63\") " pod="openstack/configure-network-openstack-openstack-networker-8chgx" Oct 14 09:23:50 crc kubenswrapper[4870]: I1014 09:23:50.945202 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jrlsz\" (UniqueName: \"kubernetes.io/projected/5a77a035-4684-43ed-b3a1-1e9dfa53dd63-kube-api-access-jrlsz\") pod \"configure-network-openstack-openstack-networker-8chgx\" (UID: \"5a77a035-4684-43ed-b3a1-1e9dfa53dd63\") " pod="openstack/configure-network-openstack-openstack-networker-8chgx" Oct 14 09:23:50 crc kubenswrapper[4870]: I1014 09:23:50.945646 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5a77a035-4684-43ed-b3a1-1e9dfa53dd63-ssh-key\") pod \"configure-network-openstack-openstack-networker-8chgx\" (UID: \"5a77a035-4684-43ed-b3a1-1e9dfa53dd63\") " pod="openstack/configure-network-openstack-openstack-networker-8chgx" Oct 14 09:23:50 crc kubenswrapper[4870]: I1014 09:23:50.954082 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5a77a035-4684-43ed-b3a1-1e9dfa53dd63-ssh-key\") pod \"configure-network-openstack-openstack-networker-8chgx\" (UID: \"5a77a035-4684-43ed-b3a1-1e9dfa53dd63\") " pod="openstack/configure-network-openstack-openstack-networker-8chgx" Oct 14 09:23:50 crc kubenswrapper[4870]: I1014 09:23:50.955074 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5a77a035-4684-43ed-b3a1-1e9dfa53dd63-inventory\") pod \"configure-network-openstack-openstack-networker-8chgx\" (UID: \"5a77a035-4684-43ed-b3a1-1e9dfa53dd63\") " pod="openstack/configure-network-openstack-openstack-networker-8chgx" Oct 14 09:23:50 crc kubenswrapper[4870]: I1014 09:23:50.968980 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jrlsz\" (UniqueName: \"kubernetes.io/projected/5a77a035-4684-43ed-b3a1-1e9dfa53dd63-kube-api-access-jrlsz\") pod \"configure-network-openstack-openstack-networker-8chgx\" (UID: \"5a77a035-4684-43ed-b3a1-1e9dfa53dd63\") " pod="openstack/configure-network-openstack-openstack-networker-8chgx" Oct 14 09:23:51 crc kubenswrapper[4870]: I1014 09:23:51.091621 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-openstack-openstack-networker-8chgx" Oct 14 09:23:51 crc kubenswrapper[4870]: I1014 09:23:51.722147 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-openstack-openstack-networker-8chgx"] Oct 14 09:23:52 crc kubenswrapper[4870]: I1014 09:23:52.672429 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-openstack-openstack-networker-8chgx" event={"ID":"5a77a035-4684-43ed-b3a1-1e9dfa53dd63","Type":"ContainerStarted","Data":"103378df8d767f0720abf25d9975e8b54a425d29b961aec145adcbac9ddfdbe0"} Oct 14 09:23:52 crc kubenswrapper[4870]: I1014 09:23:52.673840 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-openstack-openstack-networker-8chgx" event={"ID":"5a77a035-4684-43ed-b3a1-1e9dfa53dd63","Type":"ContainerStarted","Data":"f81d72997eaad618f533cd3e5313321e0612fc6796d5fe528bddb12ed290b7e5"} Oct 14 09:23:52 crc kubenswrapper[4870]: I1014 09:23:52.695031 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-openstack-openstack-networker-8chgx" podStartSLOduration=2.05139359 podStartE2EDuration="2.695011223s" podCreationTimestamp="2025-10-14 09:23:50 +0000 UTC" firstStartedPulling="2025-10-14 09:23:51.736242567 +0000 UTC m=+8567.433602948" lastFinishedPulling="2025-10-14 09:23:52.37986019 +0000 UTC m=+8568.077220581" observedRunningTime="2025-10-14 09:23:52.68678525 +0000 UTC m=+8568.384145651" watchObservedRunningTime="2025-10-14 09:23:52.695011223 +0000 UTC m=+8568.392371594" Oct 14 09:24:11 crc kubenswrapper[4870]: I1014 09:24:11.913935 4870 generic.go:334] "Generic (PLEG): container finished" podID="e4ecb749-bbab-40eb-82be-bd39305cb787" containerID="98cdfd2cd44f60b9a7f9b548102592c58d141e98b1b69f0c8b02c6800ec8f35c" exitCode=0 Oct 14 09:24:11 crc kubenswrapper[4870]: I1014 09:24:11.914044 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-openstack-openstack-cell1-cfvhm" event={"ID":"e4ecb749-bbab-40eb-82be-bd39305cb787","Type":"ContainerDied","Data":"98cdfd2cd44f60b9a7f9b548102592c58d141e98b1b69f0c8b02c6800ec8f35c"} Oct 14 09:24:13 crc kubenswrapper[4870]: I1014 09:24:13.534490 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-openstack-openstack-cell1-cfvhm" Oct 14 09:24:13 crc kubenswrapper[4870]: I1014 09:24:13.708208 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e4ecb749-bbab-40eb-82be-bd39305cb787-ceph\") pod \"e4ecb749-bbab-40eb-82be-bd39305cb787\" (UID: \"e4ecb749-bbab-40eb-82be-bd39305cb787\") " Oct 14 09:24:13 crc kubenswrapper[4870]: I1014 09:24:13.708280 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e4ecb749-bbab-40eb-82be-bd39305cb787-ssh-key\") pod \"e4ecb749-bbab-40eb-82be-bd39305cb787\" (UID: \"e4ecb749-bbab-40eb-82be-bd39305cb787\") " Oct 14 09:24:13 crc kubenswrapper[4870]: I1014 09:24:13.708550 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xxlb6\" (UniqueName: \"kubernetes.io/projected/e4ecb749-bbab-40eb-82be-bd39305cb787-kube-api-access-xxlb6\") pod \"e4ecb749-bbab-40eb-82be-bd39305cb787\" (UID: \"e4ecb749-bbab-40eb-82be-bd39305cb787\") " Oct 14 09:24:13 crc kubenswrapper[4870]: I1014 09:24:13.708599 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e4ecb749-bbab-40eb-82be-bd39305cb787-inventory\") pod \"e4ecb749-bbab-40eb-82be-bd39305cb787\" (UID: \"e4ecb749-bbab-40eb-82be-bd39305cb787\") " Oct 14 09:24:13 crc kubenswrapper[4870]: I1014 09:24:13.716714 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e4ecb749-bbab-40eb-82be-bd39305cb787-ceph" (OuterVolumeSpecName: "ceph") pod "e4ecb749-bbab-40eb-82be-bd39305cb787" (UID: "e4ecb749-bbab-40eb-82be-bd39305cb787"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:24:13 crc kubenswrapper[4870]: I1014 09:24:13.717703 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e4ecb749-bbab-40eb-82be-bd39305cb787-kube-api-access-xxlb6" (OuterVolumeSpecName: "kube-api-access-xxlb6") pod "e4ecb749-bbab-40eb-82be-bd39305cb787" (UID: "e4ecb749-bbab-40eb-82be-bd39305cb787"). InnerVolumeSpecName "kube-api-access-xxlb6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:24:13 crc kubenswrapper[4870]: I1014 09:24:13.754693 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e4ecb749-bbab-40eb-82be-bd39305cb787-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "e4ecb749-bbab-40eb-82be-bd39305cb787" (UID: "e4ecb749-bbab-40eb-82be-bd39305cb787"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:24:13 crc kubenswrapper[4870]: I1014 09:24:13.761509 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e4ecb749-bbab-40eb-82be-bd39305cb787-inventory" (OuterVolumeSpecName: "inventory") pod "e4ecb749-bbab-40eb-82be-bd39305cb787" (UID: "e4ecb749-bbab-40eb-82be-bd39305cb787"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:24:13 crc kubenswrapper[4870]: I1014 09:24:13.812205 4870 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e4ecb749-bbab-40eb-82be-bd39305cb787-ceph\") on node \"crc\" DevicePath \"\"" Oct 14 09:24:13 crc kubenswrapper[4870]: I1014 09:24:13.812261 4870 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e4ecb749-bbab-40eb-82be-bd39305cb787-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 14 09:24:13 crc kubenswrapper[4870]: I1014 09:24:13.812284 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xxlb6\" (UniqueName: \"kubernetes.io/projected/e4ecb749-bbab-40eb-82be-bd39305cb787-kube-api-access-xxlb6\") on node \"crc\" DevicePath \"\"" Oct 14 09:24:13 crc kubenswrapper[4870]: I1014 09:24:13.812305 4870 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e4ecb749-bbab-40eb-82be-bd39305cb787-inventory\") on node \"crc\" DevicePath \"\"" Oct 14 09:24:13 crc kubenswrapper[4870]: I1014 09:24:13.937089 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-openstack-openstack-cell1-cfvhm" event={"ID":"e4ecb749-bbab-40eb-82be-bd39305cb787","Type":"ContainerDied","Data":"136029c2da65fcffe17de432c7c879033c3baa52072ba0dc6880c046f29a4f78"} Oct 14 09:24:13 crc kubenswrapper[4870]: I1014 09:24:13.937132 4870 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="136029c2da65fcffe17de432c7c879033c3baa52072ba0dc6880c046f29a4f78" Oct 14 09:24:13 crc kubenswrapper[4870]: I1014 09:24:13.937194 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-openstack-openstack-cell1-cfvhm" Oct 14 09:24:14 crc kubenswrapper[4870]: I1014 09:24:14.066490 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-openstack-openstack-cell1-jglgw"] Oct 14 09:24:14 crc kubenswrapper[4870]: E1014 09:24:14.067274 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e4ecb749-bbab-40eb-82be-bd39305cb787" containerName="download-cache-openstack-openstack-cell1" Oct 14 09:24:14 crc kubenswrapper[4870]: I1014 09:24:14.067310 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="e4ecb749-bbab-40eb-82be-bd39305cb787" containerName="download-cache-openstack-openstack-cell1" Oct 14 09:24:14 crc kubenswrapper[4870]: I1014 09:24:14.067785 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="e4ecb749-bbab-40eb-82be-bd39305cb787" containerName="download-cache-openstack-openstack-cell1" Oct 14 09:24:14 crc kubenswrapper[4870]: I1014 09:24:14.069127 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-openstack-openstack-cell1-jglgw" Oct 14 09:24:14 crc kubenswrapper[4870]: I1014 09:24:14.074274 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-wrz7w" Oct 14 09:24:14 crc kubenswrapper[4870]: I1014 09:24:14.074469 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 14 09:24:14 crc kubenswrapper[4870]: I1014 09:24:14.082185 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-openstack-openstack-cell1-jglgw"] Oct 14 09:24:14 crc kubenswrapper[4870]: I1014 09:24:14.222627 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f4qtx\" (UniqueName: \"kubernetes.io/projected/2e2a40a0-7698-4294-a107-639c5de9e5f7-kube-api-access-f4qtx\") pod \"configure-network-openstack-openstack-cell1-jglgw\" (UID: \"2e2a40a0-7698-4294-a107-639c5de9e5f7\") " pod="openstack/configure-network-openstack-openstack-cell1-jglgw" Oct 14 09:24:14 crc kubenswrapper[4870]: I1014 09:24:14.223051 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/2e2a40a0-7698-4294-a107-639c5de9e5f7-ceph\") pod \"configure-network-openstack-openstack-cell1-jglgw\" (UID: \"2e2a40a0-7698-4294-a107-639c5de9e5f7\") " pod="openstack/configure-network-openstack-openstack-cell1-jglgw" Oct 14 09:24:14 crc kubenswrapper[4870]: I1014 09:24:14.223105 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2e2a40a0-7698-4294-a107-639c5de9e5f7-inventory\") pod \"configure-network-openstack-openstack-cell1-jglgw\" (UID: \"2e2a40a0-7698-4294-a107-639c5de9e5f7\") " pod="openstack/configure-network-openstack-openstack-cell1-jglgw" Oct 14 09:24:14 crc kubenswrapper[4870]: I1014 09:24:14.223131 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2e2a40a0-7698-4294-a107-639c5de9e5f7-ssh-key\") pod \"configure-network-openstack-openstack-cell1-jglgw\" (UID: \"2e2a40a0-7698-4294-a107-639c5de9e5f7\") " pod="openstack/configure-network-openstack-openstack-cell1-jglgw" Oct 14 09:24:14 crc kubenswrapper[4870]: I1014 09:24:14.325015 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2e2a40a0-7698-4294-a107-639c5de9e5f7-inventory\") pod \"configure-network-openstack-openstack-cell1-jglgw\" (UID: \"2e2a40a0-7698-4294-a107-639c5de9e5f7\") " pod="openstack/configure-network-openstack-openstack-cell1-jglgw" Oct 14 09:24:14 crc kubenswrapper[4870]: I1014 09:24:14.325077 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2e2a40a0-7698-4294-a107-639c5de9e5f7-ssh-key\") pod \"configure-network-openstack-openstack-cell1-jglgw\" (UID: \"2e2a40a0-7698-4294-a107-639c5de9e5f7\") " pod="openstack/configure-network-openstack-openstack-cell1-jglgw" Oct 14 09:24:14 crc kubenswrapper[4870]: I1014 09:24:14.325203 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f4qtx\" (UniqueName: \"kubernetes.io/projected/2e2a40a0-7698-4294-a107-639c5de9e5f7-kube-api-access-f4qtx\") pod \"configure-network-openstack-openstack-cell1-jglgw\" (UID: \"2e2a40a0-7698-4294-a107-639c5de9e5f7\") " pod="openstack/configure-network-openstack-openstack-cell1-jglgw" Oct 14 09:24:14 crc kubenswrapper[4870]: I1014 09:24:14.325406 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/2e2a40a0-7698-4294-a107-639c5de9e5f7-ceph\") pod \"configure-network-openstack-openstack-cell1-jglgw\" (UID: \"2e2a40a0-7698-4294-a107-639c5de9e5f7\") " pod="openstack/configure-network-openstack-openstack-cell1-jglgw" Oct 14 09:24:14 crc kubenswrapper[4870]: I1014 09:24:14.330560 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2e2a40a0-7698-4294-a107-639c5de9e5f7-inventory\") pod \"configure-network-openstack-openstack-cell1-jglgw\" (UID: \"2e2a40a0-7698-4294-a107-639c5de9e5f7\") " pod="openstack/configure-network-openstack-openstack-cell1-jglgw" Oct 14 09:24:14 crc kubenswrapper[4870]: I1014 09:24:14.331148 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/2e2a40a0-7698-4294-a107-639c5de9e5f7-ceph\") pod \"configure-network-openstack-openstack-cell1-jglgw\" (UID: \"2e2a40a0-7698-4294-a107-639c5de9e5f7\") " pod="openstack/configure-network-openstack-openstack-cell1-jglgw" Oct 14 09:24:14 crc kubenswrapper[4870]: I1014 09:24:14.331830 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2e2a40a0-7698-4294-a107-639c5de9e5f7-ssh-key\") pod \"configure-network-openstack-openstack-cell1-jglgw\" (UID: \"2e2a40a0-7698-4294-a107-639c5de9e5f7\") " pod="openstack/configure-network-openstack-openstack-cell1-jglgw" Oct 14 09:24:14 crc kubenswrapper[4870]: I1014 09:24:14.342498 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f4qtx\" (UniqueName: \"kubernetes.io/projected/2e2a40a0-7698-4294-a107-639c5de9e5f7-kube-api-access-f4qtx\") pod \"configure-network-openstack-openstack-cell1-jglgw\" (UID: \"2e2a40a0-7698-4294-a107-639c5de9e5f7\") " pod="openstack/configure-network-openstack-openstack-cell1-jglgw" Oct 14 09:24:14 crc kubenswrapper[4870]: I1014 09:24:14.396458 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-openstack-openstack-cell1-jglgw" Oct 14 09:24:15 crc kubenswrapper[4870]: I1014 09:24:15.001295 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-openstack-openstack-cell1-jglgw"] Oct 14 09:24:15 crc kubenswrapper[4870]: I1014 09:24:15.960734 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-openstack-openstack-cell1-jglgw" event={"ID":"2e2a40a0-7698-4294-a107-639c5de9e5f7","Type":"ContainerStarted","Data":"3e2dad7b3aaf70ec9e109156fff022a87697812df024a3b55a9a42126a76af40"} Oct 14 09:24:16 crc kubenswrapper[4870]: I1014 09:24:16.979179 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-openstack-openstack-cell1-jglgw" event={"ID":"2e2a40a0-7698-4294-a107-639c5de9e5f7","Type":"ContainerStarted","Data":"9720f4cd5d8308c67eb620a27dfba942d1e4473b861629329e69434035610e9a"} Oct 14 09:24:17 crc kubenswrapper[4870]: I1014 09:24:17.011424 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-openstack-openstack-cell1-jglgw" podStartSLOduration=2.230816148 podStartE2EDuration="3.01140036s" podCreationTimestamp="2025-10-14 09:24:14 +0000 UTC" firstStartedPulling="2025-10-14 09:24:15.00369385 +0000 UTC m=+8590.701054241" lastFinishedPulling="2025-10-14 09:24:15.784278042 +0000 UTC m=+8591.481638453" observedRunningTime="2025-10-14 09:24:17.006219202 +0000 UTC m=+8592.703579633" watchObservedRunningTime="2025-10-14 09:24:17.01140036 +0000 UTC m=+8592.708760771" Oct 14 09:24:53 crc kubenswrapper[4870]: I1014 09:24:53.951005 4870 patch_prober.go:28] interesting pod/machine-config-daemon-7tvc7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 09:24:53 crc kubenswrapper[4870]: I1014 09:24:53.953032 4870 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 09:24:55 crc kubenswrapper[4870]: I1014 09:24:55.454223 4870 generic.go:334] "Generic (PLEG): container finished" podID="5a77a035-4684-43ed-b3a1-1e9dfa53dd63" containerID="103378df8d767f0720abf25d9975e8b54a425d29b961aec145adcbac9ddfdbe0" exitCode=0 Oct 14 09:24:55 crc kubenswrapper[4870]: I1014 09:24:55.454607 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-openstack-openstack-networker-8chgx" event={"ID":"5a77a035-4684-43ed-b3a1-1e9dfa53dd63","Type":"ContainerDied","Data":"103378df8d767f0720abf25d9975e8b54a425d29b961aec145adcbac9ddfdbe0"} Oct 14 09:24:57 crc kubenswrapper[4870]: I1014 09:24:57.107796 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-openstack-openstack-networker-8chgx" Oct 14 09:24:57 crc kubenswrapper[4870]: I1014 09:24:57.245464 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5a77a035-4684-43ed-b3a1-1e9dfa53dd63-inventory\") pod \"5a77a035-4684-43ed-b3a1-1e9dfa53dd63\" (UID: \"5a77a035-4684-43ed-b3a1-1e9dfa53dd63\") " Oct 14 09:24:57 crc kubenswrapper[4870]: I1014 09:24:57.245693 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jrlsz\" (UniqueName: \"kubernetes.io/projected/5a77a035-4684-43ed-b3a1-1e9dfa53dd63-kube-api-access-jrlsz\") pod \"5a77a035-4684-43ed-b3a1-1e9dfa53dd63\" (UID: \"5a77a035-4684-43ed-b3a1-1e9dfa53dd63\") " Oct 14 09:24:57 crc kubenswrapper[4870]: I1014 09:24:57.245854 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5a77a035-4684-43ed-b3a1-1e9dfa53dd63-ssh-key\") pod \"5a77a035-4684-43ed-b3a1-1e9dfa53dd63\" (UID: \"5a77a035-4684-43ed-b3a1-1e9dfa53dd63\") " Oct 14 09:24:57 crc kubenswrapper[4870]: I1014 09:24:57.254264 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5a77a035-4684-43ed-b3a1-1e9dfa53dd63-kube-api-access-jrlsz" (OuterVolumeSpecName: "kube-api-access-jrlsz") pod "5a77a035-4684-43ed-b3a1-1e9dfa53dd63" (UID: "5a77a035-4684-43ed-b3a1-1e9dfa53dd63"). InnerVolumeSpecName "kube-api-access-jrlsz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:24:57 crc kubenswrapper[4870]: I1014 09:24:57.282508 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5a77a035-4684-43ed-b3a1-1e9dfa53dd63-inventory" (OuterVolumeSpecName: "inventory") pod "5a77a035-4684-43ed-b3a1-1e9dfa53dd63" (UID: "5a77a035-4684-43ed-b3a1-1e9dfa53dd63"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:24:57 crc kubenswrapper[4870]: I1014 09:24:57.300257 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5a77a035-4684-43ed-b3a1-1e9dfa53dd63-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "5a77a035-4684-43ed-b3a1-1e9dfa53dd63" (UID: "5a77a035-4684-43ed-b3a1-1e9dfa53dd63"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:24:57 crc kubenswrapper[4870]: I1014 09:24:57.348826 4870 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5a77a035-4684-43ed-b3a1-1e9dfa53dd63-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 14 09:24:57 crc kubenswrapper[4870]: I1014 09:24:57.348875 4870 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5a77a035-4684-43ed-b3a1-1e9dfa53dd63-inventory\") on node \"crc\" DevicePath \"\"" Oct 14 09:24:57 crc kubenswrapper[4870]: I1014 09:24:57.348893 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jrlsz\" (UniqueName: \"kubernetes.io/projected/5a77a035-4684-43ed-b3a1-1e9dfa53dd63-kube-api-access-jrlsz\") on node \"crc\" DevicePath \"\"" Oct 14 09:24:57 crc kubenswrapper[4870]: I1014 09:24:57.533552 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-openstack-openstack-networker-8chgx" event={"ID":"5a77a035-4684-43ed-b3a1-1e9dfa53dd63","Type":"ContainerDied","Data":"f81d72997eaad618f533cd3e5313321e0612fc6796d5fe528bddb12ed290b7e5"} Oct 14 09:24:57 crc kubenswrapper[4870]: I1014 09:24:57.533590 4870 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f81d72997eaad618f533cd3e5313321e0612fc6796d5fe528bddb12ed290b7e5" Oct 14 09:24:57 crc kubenswrapper[4870]: I1014 09:24:57.533644 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-openstack-openstack-networker-8chgx" Oct 14 09:24:57 crc kubenswrapper[4870]: I1014 09:24:57.602074 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-openstack-openstack-networker-ps772"] Oct 14 09:24:57 crc kubenswrapper[4870]: E1014 09:24:57.602480 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5a77a035-4684-43ed-b3a1-1e9dfa53dd63" containerName="configure-network-openstack-openstack-networker" Oct 14 09:24:57 crc kubenswrapper[4870]: I1014 09:24:57.602498 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="5a77a035-4684-43ed-b3a1-1e9dfa53dd63" containerName="configure-network-openstack-openstack-networker" Oct 14 09:24:57 crc kubenswrapper[4870]: I1014 09:24:57.602722 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="5a77a035-4684-43ed-b3a1-1e9dfa53dd63" containerName="configure-network-openstack-openstack-networker" Oct 14 09:24:57 crc kubenswrapper[4870]: I1014 09:24:57.603719 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-openstack-openstack-networker-ps772" Oct 14 09:24:57 crc kubenswrapper[4870]: I1014 09:24:57.608492 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-networker" Oct 14 09:24:57 crc kubenswrapper[4870]: I1014 09:24:57.608695 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-networker-dockercfg-54hzr" Oct 14 09:24:57 crc kubenswrapper[4870]: I1014 09:24:57.636899 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-openstack-openstack-networker-ps772"] Oct 14 09:24:57 crc kubenswrapper[4870]: I1014 09:24:57.654660 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tzmqw\" (UniqueName: \"kubernetes.io/projected/1cd91851-2684-4634-8d0e-78857205d00a-kube-api-access-tzmqw\") pod \"validate-network-openstack-openstack-networker-ps772\" (UID: \"1cd91851-2684-4634-8d0e-78857205d00a\") " pod="openstack/validate-network-openstack-openstack-networker-ps772" Oct 14 09:24:57 crc kubenswrapper[4870]: I1014 09:24:57.654727 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1cd91851-2684-4634-8d0e-78857205d00a-inventory\") pod \"validate-network-openstack-openstack-networker-ps772\" (UID: \"1cd91851-2684-4634-8d0e-78857205d00a\") " pod="openstack/validate-network-openstack-openstack-networker-ps772" Oct 14 09:24:57 crc kubenswrapper[4870]: I1014 09:24:57.654783 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1cd91851-2684-4634-8d0e-78857205d00a-ssh-key\") pod \"validate-network-openstack-openstack-networker-ps772\" (UID: \"1cd91851-2684-4634-8d0e-78857205d00a\") " pod="openstack/validate-network-openstack-openstack-networker-ps772" Oct 14 09:24:57 crc kubenswrapper[4870]: I1014 09:24:57.758214 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tzmqw\" (UniqueName: \"kubernetes.io/projected/1cd91851-2684-4634-8d0e-78857205d00a-kube-api-access-tzmqw\") pod \"validate-network-openstack-openstack-networker-ps772\" (UID: \"1cd91851-2684-4634-8d0e-78857205d00a\") " pod="openstack/validate-network-openstack-openstack-networker-ps772" Oct 14 09:24:57 crc kubenswrapper[4870]: I1014 09:24:57.758369 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1cd91851-2684-4634-8d0e-78857205d00a-inventory\") pod \"validate-network-openstack-openstack-networker-ps772\" (UID: \"1cd91851-2684-4634-8d0e-78857205d00a\") " pod="openstack/validate-network-openstack-openstack-networker-ps772" Oct 14 09:24:57 crc kubenswrapper[4870]: I1014 09:24:57.758605 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1cd91851-2684-4634-8d0e-78857205d00a-ssh-key\") pod \"validate-network-openstack-openstack-networker-ps772\" (UID: \"1cd91851-2684-4634-8d0e-78857205d00a\") " pod="openstack/validate-network-openstack-openstack-networker-ps772" Oct 14 09:24:57 crc kubenswrapper[4870]: I1014 09:24:57.762145 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1cd91851-2684-4634-8d0e-78857205d00a-inventory\") pod \"validate-network-openstack-openstack-networker-ps772\" (UID: \"1cd91851-2684-4634-8d0e-78857205d00a\") " pod="openstack/validate-network-openstack-openstack-networker-ps772" Oct 14 09:24:57 crc kubenswrapper[4870]: I1014 09:24:57.763696 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1cd91851-2684-4634-8d0e-78857205d00a-ssh-key\") pod \"validate-network-openstack-openstack-networker-ps772\" (UID: \"1cd91851-2684-4634-8d0e-78857205d00a\") " pod="openstack/validate-network-openstack-openstack-networker-ps772" Oct 14 09:24:57 crc kubenswrapper[4870]: I1014 09:24:57.777551 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tzmqw\" (UniqueName: \"kubernetes.io/projected/1cd91851-2684-4634-8d0e-78857205d00a-kube-api-access-tzmqw\") pod \"validate-network-openstack-openstack-networker-ps772\" (UID: \"1cd91851-2684-4634-8d0e-78857205d00a\") " pod="openstack/validate-network-openstack-openstack-networker-ps772" Oct 14 09:24:57 crc kubenswrapper[4870]: I1014 09:24:57.932916 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-openstack-openstack-networker-ps772" Oct 14 09:24:58 crc kubenswrapper[4870]: I1014 09:24:58.582034 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-openstack-openstack-networker-ps772"] Oct 14 09:24:59 crc kubenswrapper[4870]: I1014 09:24:59.565826 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-openstack-openstack-networker-ps772" event={"ID":"1cd91851-2684-4634-8d0e-78857205d00a","Type":"ContainerStarted","Data":"01b69761d5ee006ab20ea29d4febe5f61c9c6f13dadc4b142825a74e27d99f69"} Oct 14 09:25:00 crc kubenswrapper[4870]: I1014 09:25:00.575626 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-openstack-openstack-networker-ps772" event={"ID":"1cd91851-2684-4634-8d0e-78857205d00a","Type":"ContainerStarted","Data":"9afdbaf85da6754e9014b5e8493b9a201d1d3205060f1a0998a9f5718b5a2339"} Oct 14 09:25:06 crc kubenswrapper[4870]: I1014 09:25:06.653522 4870 generic.go:334] "Generic (PLEG): container finished" podID="1cd91851-2684-4634-8d0e-78857205d00a" containerID="9afdbaf85da6754e9014b5e8493b9a201d1d3205060f1a0998a9f5718b5a2339" exitCode=0 Oct 14 09:25:06 crc kubenswrapper[4870]: I1014 09:25:06.653645 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-openstack-openstack-networker-ps772" event={"ID":"1cd91851-2684-4634-8d0e-78857205d00a","Type":"ContainerDied","Data":"9afdbaf85da6754e9014b5e8493b9a201d1d3205060f1a0998a9f5718b5a2339"} Oct 14 09:25:08 crc kubenswrapper[4870]: I1014 09:25:08.237555 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-openstack-openstack-networker-ps772" Oct 14 09:25:08 crc kubenswrapper[4870]: I1014 09:25:08.323612 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1cd91851-2684-4634-8d0e-78857205d00a-ssh-key\") pod \"1cd91851-2684-4634-8d0e-78857205d00a\" (UID: \"1cd91851-2684-4634-8d0e-78857205d00a\") " Oct 14 09:25:08 crc kubenswrapper[4870]: I1014 09:25:08.323743 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1cd91851-2684-4634-8d0e-78857205d00a-inventory\") pod \"1cd91851-2684-4634-8d0e-78857205d00a\" (UID: \"1cd91851-2684-4634-8d0e-78857205d00a\") " Oct 14 09:25:08 crc kubenswrapper[4870]: I1014 09:25:08.323783 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tzmqw\" (UniqueName: \"kubernetes.io/projected/1cd91851-2684-4634-8d0e-78857205d00a-kube-api-access-tzmqw\") pod \"1cd91851-2684-4634-8d0e-78857205d00a\" (UID: \"1cd91851-2684-4634-8d0e-78857205d00a\") " Oct 14 09:25:08 crc kubenswrapper[4870]: I1014 09:25:08.331973 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1cd91851-2684-4634-8d0e-78857205d00a-kube-api-access-tzmqw" (OuterVolumeSpecName: "kube-api-access-tzmqw") pod "1cd91851-2684-4634-8d0e-78857205d00a" (UID: "1cd91851-2684-4634-8d0e-78857205d00a"). InnerVolumeSpecName "kube-api-access-tzmqw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:25:08 crc kubenswrapper[4870]: I1014 09:25:08.365522 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1cd91851-2684-4634-8d0e-78857205d00a-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "1cd91851-2684-4634-8d0e-78857205d00a" (UID: "1cd91851-2684-4634-8d0e-78857205d00a"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:25:08 crc kubenswrapper[4870]: I1014 09:25:08.370086 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1cd91851-2684-4634-8d0e-78857205d00a-inventory" (OuterVolumeSpecName: "inventory") pod "1cd91851-2684-4634-8d0e-78857205d00a" (UID: "1cd91851-2684-4634-8d0e-78857205d00a"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:25:08 crc kubenswrapper[4870]: I1014 09:25:08.426653 4870 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1cd91851-2684-4634-8d0e-78857205d00a-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 14 09:25:08 crc kubenswrapper[4870]: I1014 09:25:08.426699 4870 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1cd91851-2684-4634-8d0e-78857205d00a-inventory\") on node \"crc\" DevicePath \"\"" Oct 14 09:25:08 crc kubenswrapper[4870]: I1014 09:25:08.426717 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tzmqw\" (UniqueName: \"kubernetes.io/projected/1cd91851-2684-4634-8d0e-78857205d00a-kube-api-access-tzmqw\") on node \"crc\" DevicePath \"\"" Oct 14 09:25:08 crc kubenswrapper[4870]: I1014 09:25:08.680572 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-openstack-openstack-networker-ps772" event={"ID":"1cd91851-2684-4634-8d0e-78857205d00a","Type":"ContainerDied","Data":"01b69761d5ee006ab20ea29d4febe5f61c9c6f13dadc4b142825a74e27d99f69"} Oct 14 09:25:08 crc kubenswrapper[4870]: I1014 09:25:08.680623 4870 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="01b69761d5ee006ab20ea29d4febe5f61c9c6f13dadc4b142825a74e27d99f69" Oct 14 09:25:08 crc kubenswrapper[4870]: I1014 09:25:08.680689 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-openstack-openstack-networker-ps772" Oct 14 09:25:08 crc kubenswrapper[4870]: I1014 09:25:08.767536 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-openstack-openstack-networker-wdjn4"] Oct 14 09:25:08 crc kubenswrapper[4870]: E1014 09:25:08.768328 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1cd91851-2684-4634-8d0e-78857205d00a" containerName="validate-network-openstack-openstack-networker" Oct 14 09:25:08 crc kubenswrapper[4870]: I1014 09:25:08.768426 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="1cd91851-2684-4634-8d0e-78857205d00a" containerName="validate-network-openstack-openstack-networker" Oct 14 09:25:08 crc kubenswrapper[4870]: I1014 09:25:08.768806 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="1cd91851-2684-4634-8d0e-78857205d00a" containerName="validate-network-openstack-openstack-networker" Oct 14 09:25:08 crc kubenswrapper[4870]: I1014 09:25:08.775810 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-openstack-openstack-networker-wdjn4" Oct 14 09:25:08 crc kubenswrapper[4870]: I1014 09:25:08.778066 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-networker" Oct 14 09:25:08 crc kubenswrapper[4870]: I1014 09:25:08.778736 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-networker-dockercfg-54hzr" Oct 14 09:25:08 crc kubenswrapper[4870]: I1014 09:25:08.785152 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-openstack-openstack-networker-wdjn4"] Oct 14 09:25:08 crc kubenswrapper[4870]: I1014 09:25:08.834558 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ed0d4c93-b579-4253-867f-d191d43c35c7-inventory\") pod \"install-os-openstack-openstack-networker-wdjn4\" (UID: \"ed0d4c93-b579-4253-867f-d191d43c35c7\") " pod="openstack/install-os-openstack-openstack-networker-wdjn4" Oct 14 09:25:08 crc kubenswrapper[4870]: I1014 09:25:08.834954 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ed0d4c93-b579-4253-867f-d191d43c35c7-ssh-key\") pod \"install-os-openstack-openstack-networker-wdjn4\" (UID: \"ed0d4c93-b579-4253-867f-d191d43c35c7\") " pod="openstack/install-os-openstack-openstack-networker-wdjn4" Oct 14 09:25:08 crc kubenswrapper[4870]: I1014 09:25:08.835097 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xrdbh\" (UniqueName: \"kubernetes.io/projected/ed0d4c93-b579-4253-867f-d191d43c35c7-kube-api-access-xrdbh\") pod \"install-os-openstack-openstack-networker-wdjn4\" (UID: \"ed0d4c93-b579-4253-867f-d191d43c35c7\") " pod="openstack/install-os-openstack-openstack-networker-wdjn4" Oct 14 09:25:08 crc kubenswrapper[4870]: I1014 09:25:08.936418 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ed0d4c93-b579-4253-867f-d191d43c35c7-ssh-key\") pod \"install-os-openstack-openstack-networker-wdjn4\" (UID: \"ed0d4c93-b579-4253-867f-d191d43c35c7\") " pod="openstack/install-os-openstack-openstack-networker-wdjn4" Oct 14 09:25:08 crc kubenswrapper[4870]: I1014 09:25:08.936537 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xrdbh\" (UniqueName: \"kubernetes.io/projected/ed0d4c93-b579-4253-867f-d191d43c35c7-kube-api-access-xrdbh\") pod \"install-os-openstack-openstack-networker-wdjn4\" (UID: \"ed0d4c93-b579-4253-867f-d191d43c35c7\") " pod="openstack/install-os-openstack-openstack-networker-wdjn4" Oct 14 09:25:08 crc kubenswrapper[4870]: I1014 09:25:08.936596 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ed0d4c93-b579-4253-867f-d191d43c35c7-inventory\") pod \"install-os-openstack-openstack-networker-wdjn4\" (UID: \"ed0d4c93-b579-4253-867f-d191d43c35c7\") " pod="openstack/install-os-openstack-openstack-networker-wdjn4" Oct 14 09:25:08 crc kubenswrapper[4870]: I1014 09:25:08.940802 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ed0d4c93-b579-4253-867f-d191d43c35c7-ssh-key\") pod \"install-os-openstack-openstack-networker-wdjn4\" (UID: \"ed0d4c93-b579-4253-867f-d191d43c35c7\") " pod="openstack/install-os-openstack-openstack-networker-wdjn4" Oct 14 09:25:08 crc kubenswrapper[4870]: I1014 09:25:08.953679 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ed0d4c93-b579-4253-867f-d191d43c35c7-inventory\") pod \"install-os-openstack-openstack-networker-wdjn4\" (UID: \"ed0d4c93-b579-4253-867f-d191d43c35c7\") " pod="openstack/install-os-openstack-openstack-networker-wdjn4" Oct 14 09:25:08 crc kubenswrapper[4870]: I1014 09:25:08.960654 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xrdbh\" (UniqueName: \"kubernetes.io/projected/ed0d4c93-b579-4253-867f-d191d43c35c7-kube-api-access-xrdbh\") pod \"install-os-openstack-openstack-networker-wdjn4\" (UID: \"ed0d4c93-b579-4253-867f-d191d43c35c7\") " pod="openstack/install-os-openstack-openstack-networker-wdjn4" Oct 14 09:25:09 crc kubenswrapper[4870]: I1014 09:25:09.099313 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-openstack-openstack-networker-wdjn4" Oct 14 09:25:09 crc kubenswrapper[4870]: W1014 09:25:09.714066 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poded0d4c93_b579_4253_867f_d191d43c35c7.slice/crio-375ec7325f35e1398712dca62c42da8c9b50484035fabfcb726d6a25cba89c73 WatchSource:0}: Error finding container 375ec7325f35e1398712dca62c42da8c9b50484035fabfcb726d6a25cba89c73: Status 404 returned error can't find the container with id 375ec7325f35e1398712dca62c42da8c9b50484035fabfcb726d6a25cba89c73 Oct 14 09:25:09 crc kubenswrapper[4870]: I1014 09:25:09.718544 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-openstack-openstack-networker-wdjn4"] Oct 14 09:25:10 crc kubenswrapper[4870]: I1014 09:25:10.706596 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-openstack-openstack-networker-wdjn4" event={"ID":"ed0d4c93-b579-4253-867f-d191d43c35c7","Type":"ContainerStarted","Data":"528bb0301df717845635b2eaea75b95255d1b6840d6443eeb25e2d0d8009ce31"} Oct 14 09:25:10 crc kubenswrapper[4870]: I1014 09:25:10.707095 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-openstack-openstack-networker-wdjn4" event={"ID":"ed0d4c93-b579-4253-867f-d191d43c35c7","Type":"ContainerStarted","Data":"375ec7325f35e1398712dca62c42da8c9b50484035fabfcb726d6a25cba89c73"} Oct 14 09:25:10 crc kubenswrapper[4870]: I1014 09:25:10.726370 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-openstack-openstack-networker-wdjn4" podStartSLOduration=2.133375908 podStartE2EDuration="2.726351735s" podCreationTimestamp="2025-10-14 09:25:08 +0000 UTC" firstStartedPulling="2025-10-14 09:25:09.71720773 +0000 UTC m=+8645.414568101" lastFinishedPulling="2025-10-14 09:25:10.310183527 +0000 UTC m=+8646.007543928" observedRunningTime="2025-10-14 09:25:10.72371796 +0000 UTC m=+8646.421078351" watchObservedRunningTime="2025-10-14 09:25:10.726351735 +0000 UTC m=+8646.423712106" Oct 14 09:25:18 crc kubenswrapper[4870]: I1014 09:25:18.822083 4870 generic.go:334] "Generic (PLEG): container finished" podID="2e2a40a0-7698-4294-a107-639c5de9e5f7" containerID="9720f4cd5d8308c67eb620a27dfba942d1e4473b861629329e69434035610e9a" exitCode=0 Oct 14 09:25:18 crc kubenswrapper[4870]: I1014 09:25:18.822152 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-openstack-openstack-cell1-jglgw" event={"ID":"2e2a40a0-7698-4294-a107-639c5de9e5f7","Type":"ContainerDied","Data":"9720f4cd5d8308c67eb620a27dfba942d1e4473b861629329e69434035610e9a"} Oct 14 09:25:20 crc kubenswrapper[4870]: I1014 09:25:20.370814 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-openstack-openstack-cell1-jglgw" Oct 14 09:25:20 crc kubenswrapper[4870]: I1014 09:25:20.540941 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/2e2a40a0-7698-4294-a107-639c5de9e5f7-ceph\") pod \"2e2a40a0-7698-4294-a107-639c5de9e5f7\" (UID: \"2e2a40a0-7698-4294-a107-639c5de9e5f7\") " Oct 14 09:25:20 crc kubenswrapper[4870]: I1014 09:25:20.541326 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f4qtx\" (UniqueName: \"kubernetes.io/projected/2e2a40a0-7698-4294-a107-639c5de9e5f7-kube-api-access-f4qtx\") pod \"2e2a40a0-7698-4294-a107-639c5de9e5f7\" (UID: \"2e2a40a0-7698-4294-a107-639c5de9e5f7\") " Oct 14 09:25:20 crc kubenswrapper[4870]: I1014 09:25:20.541352 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2e2a40a0-7698-4294-a107-639c5de9e5f7-ssh-key\") pod \"2e2a40a0-7698-4294-a107-639c5de9e5f7\" (UID: \"2e2a40a0-7698-4294-a107-639c5de9e5f7\") " Oct 14 09:25:20 crc kubenswrapper[4870]: I1014 09:25:20.541623 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2e2a40a0-7698-4294-a107-639c5de9e5f7-inventory\") pod \"2e2a40a0-7698-4294-a107-639c5de9e5f7\" (UID: \"2e2a40a0-7698-4294-a107-639c5de9e5f7\") " Oct 14 09:25:20 crc kubenswrapper[4870]: I1014 09:25:20.547008 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2e2a40a0-7698-4294-a107-639c5de9e5f7-kube-api-access-f4qtx" (OuterVolumeSpecName: "kube-api-access-f4qtx") pod "2e2a40a0-7698-4294-a107-639c5de9e5f7" (UID: "2e2a40a0-7698-4294-a107-639c5de9e5f7"). InnerVolumeSpecName "kube-api-access-f4qtx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:25:20 crc kubenswrapper[4870]: I1014 09:25:20.548001 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e2a40a0-7698-4294-a107-639c5de9e5f7-ceph" (OuterVolumeSpecName: "ceph") pod "2e2a40a0-7698-4294-a107-639c5de9e5f7" (UID: "2e2a40a0-7698-4294-a107-639c5de9e5f7"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:25:20 crc kubenswrapper[4870]: I1014 09:25:20.574219 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e2a40a0-7698-4294-a107-639c5de9e5f7-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "2e2a40a0-7698-4294-a107-639c5de9e5f7" (UID: "2e2a40a0-7698-4294-a107-639c5de9e5f7"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:25:20 crc kubenswrapper[4870]: I1014 09:25:20.580085 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e2a40a0-7698-4294-a107-639c5de9e5f7-inventory" (OuterVolumeSpecName: "inventory") pod "2e2a40a0-7698-4294-a107-639c5de9e5f7" (UID: "2e2a40a0-7698-4294-a107-639c5de9e5f7"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:25:20 crc kubenswrapper[4870]: I1014 09:25:20.644902 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f4qtx\" (UniqueName: \"kubernetes.io/projected/2e2a40a0-7698-4294-a107-639c5de9e5f7-kube-api-access-f4qtx\") on node \"crc\" DevicePath \"\"" Oct 14 09:25:20 crc kubenswrapper[4870]: I1014 09:25:20.644948 4870 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2e2a40a0-7698-4294-a107-639c5de9e5f7-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 14 09:25:20 crc kubenswrapper[4870]: I1014 09:25:20.644962 4870 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2e2a40a0-7698-4294-a107-639c5de9e5f7-inventory\") on node \"crc\" DevicePath \"\"" Oct 14 09:25:20 crc kubenswrapper[4870]: I1014 09:25:20.644975 4870 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/2e2a40a0-7698-4294-a107-639c5de9e5f7-ceph\") on node \"crc\" DevicePath \"\"" Oct 14 09:25:20 crc kubenswrapper[4870]: I1014 09:25:20.849525 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-openstack-openstack-cell1-jglgw" event={"ID":"2e2a40a0-7698-4294-a107-639c5de9e5f7","Type":"ContainerDied","Data":"3e2dad7b3aaf70ec9e109156fff022a87697812df024a3b55a9a42126a76af40"} Oct 14 09:25:20 crc kubenswrapper[4870]: I1014 09:25:20.849571 4870 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3e2dad7b3aaf70ec9e109156fff022a87697812df024a3b55a9a42126a76af40" Oct 14 09:25:20 crc kubenswrapper[4870]: I1014 09:25:20.849571 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-openstack-openstack-cell1-jglgw" Oct 14 09:25:20 crc kubenswrapper[4870]: I1014 09:25:20.933556 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-openstack-openstack-cell1-lrc5r"] Oct 14 09:25:20 crc kubenswrapper[4870]: E1014 09:25:20.934757 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e2a40a0-7698-4294-a107-639c5de9e5f7" containerName="configure-network-openstack-openstack-cell1" Oct 14 09:25:20 crc kubenswrapper[4870]: I1014 09:25:20.934914 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e2a40a0-7698-4294-a107-639c5de9e5f7" containerName="configure-network-openstack-openstack-cell1" Oct 14 09:25:20 crc kubenswrapper[4870]: I1014 09:25:20.935493 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e2a40a0-7698-4294-a107-639c5de9e5f7" containerName="configure-network-openstack-openstack-cell1" Oct 14 09:25:20 crc kubenswrapper[4870]: I1014 09:25:20.936957 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-openstack-openstack-cell1-lrc5r" Oct 14 09:25:20 crc kubenswrapper[4870]: I1014 09:25:20.939905 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 14 09:25:20 crc kubenswrapper[4870]: I1014 09:25:20.940221 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-wrz7w" Oct 14 09:25:20 crc kubenswrapper[4870]: I1014 09:25:20.941330 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-openstack-openstack-cell1-lrc5r"] Oct 14 09:25:21 crc kubenswrapper[4870]: I1014 09:25:21.055258 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/34d9c179-4fe6-4431-b495-6803732cb3ee-inventory\") pod \"validate-network-openstack-openstack-cell1-lrc5r\" (UID: \"34d9c179-4fe6-4431-b495-6803732cb3ee\") " pod="openstack/validate-network-openstack-openstack-cell1-lrc5r" Oct 14 09:25:21 crc kubenswrapper[4870]: I1014 09:25:21.056263 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/34d9c179-4fe6-4431-b495-6803732cb3ee-ceph\") pod \"validate-network-openstack-openstack-cell1-lrc5r\" (UID: \"34d9c179-4fe6-4431-b495-6803732cb3ee\") " pod="openstack/validate-network-openstack-openstack-cell1-lrc5r" Oct 14 09:25:21 crc kubenswrapper[4870]: I1014 09:25:21.056390 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qbgqh\" (UniqueName: \"kubernetes.io/projected/34d9c179-4fe6-4431-b495-6803732cb3ee-kube-api-access-qbgqh\") pod \"validate-network-openstack-openstack-cell1-lrc5r\" (UID: \"34d9c179-4fe6-4431-b495-6803732cb3ee\") " pod="openstack/validate-network-openstack-openstack-cell1-lrc5r" Oct 14 09:25:21 crc kubenswrapper[4870]: I1014 09:25:21.058113 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/34d9c179-4fe6-4431-b495-6803732cb3ee-ssh-key\") pod \"validate-network-openstack-openstack-cell1-lrc5r\" (UID: \"34d9c179-4fe6-4431-b495-6803732cb3ee\") " pod="openstack/validate-network-openstack-openstack-cell1-lrc5r" Oct 14 09:25:21 crc kubenswrapper[4870]: I1014 09:25:21.160445 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/34d9c179-4fe6-4431-b495-6803732cb3ee-ssh-key\") pod \"validate-network-openstack-openstack-cell1-lrc5r\" (UID: \"34d9c179-4fe6-4431-b495-6803732cb3ee\") " pod="openstack/validate-network-openstack-openstack-cell1-lrc5r" Oct 14 09:25:21 crc kubenswrapper[4870]: I1014 09:25:21.160864 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/34d9c179-4fe6-4431-b495-6803732cb3ee-inventory\") pod \"validate-network-openstack-openstack-cell1-lrc5r\" (UID: \"34d9c179-4fe6-4431-b495-6803732cb3ee\") " pod="openstack/validate-network-openstack-openstack-cell1-lrc5r" Oct 14 09:25:21 crc kubenswrapper[4870]: I1014 09:25:21.160990 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/34d9c179-4fe6-4431-b495-6803732cb3ee-ceph\") pod \"validate-network-openstack-openstack-cell1-lrc5r\" (UID: \"34d9c179-4fe6-4431-b495-6803732cb3ee\") " pod="openstack/validate-network-openstack-openstack-cell1-lrc5r" Oct 14 09:25:21 crc kubenswrapper[4870]: I1014 09:25:21.161119 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qbgqh\" (UniqueName: \"kubernetes.io/projected/34d9c179-4fe6-4431-b495-6803732cb3ee-kube-api-access-qbgqh\") pod \"validate-network-openstack-openstack-cell1-lrc5r\" (UID: \"34d9c179-4fe6-4431-b495-6803732cb3ee\") " pod="openstack/validate-network-openstack-openstack-cell1-lrc5r" Oct 14 09:25:21 crc kubenswrapper[4870]: I1014 09:25:21.165797 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/34d9c179-4fe6-4431-b495-6803732cb3ee-ceph\") pod \"validate-network-openstack-openstack-cell1-lrc5r\" (UID: \"34d9c179-4fe6-4431-b495-6803732cb3ee\") " pod="openstack/validate-network-openstack-openstack-cell1-lrc5r" Oct 14 09:25:21 crc kubenswrapper[4870]: I1014 09:25:21.165888 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/34d9c179-4fe6-4431-b495-6803732cb3ee-ssh-key\") pod \"validate-network-openstack-openstack-cell1-lrc5r\" (UID: \"34d9c179-4fe6-4431-b495-6803732cb3ee\") " pod="openstack/validate-network-openstack-openstack-cell1-lrc5r" Oct 14 09:25:21 crc kubenswrapper[4870]: I1014 09:25:21.168516 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/34d9c179-4fe6-4431-b495-6803732cb3ee-inventory\") pod \"validate-network-openstack-openstack-cell1-lrc5r\" (UID: \"34d9c179-4fe6-4431-b495-6803732cb3ee\") " pod="openstack/validate-network-openstack-openstack-cell1-lrc5r" Oct 14 09:25:21 crc kubenswrapper[4870]: I1014 09:25:21.188093 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qbgqh\" (UniqueName: \"kubernetes.io/projected/34d9c179-4fe6-4431-b495-6803732cb3ee-kube-api-access-qbgqh\") pod \"validate-network-openstack-openstack-cell1-lrc5r\" (UID: \"34d9c179-4fe6-4431-b495-6803732cb3ee\") " pod="openstack/validate-network-openstack-openstack-cell1-lrc5r" Oct 14 09:25:21 crc kubenswrapper[4870]: I1014 09:25:21.253997 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-openstack-openstack-cell1-lrc5r" Oct 14 09:25:21 crc kubenswrapper[4870]: I1014 09:25:21.668429 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-openstack-openstack-cell1-lrc5r"] Oct 14 09:25:21 crc kubenswrapper[4870]: W1014 09:25:21.672098 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod34d9c179_4fe6_4431_b495_6803732cb3ee.slice/crio-c6fef86b934abfea4c0eecb3f882d2eb6eaba384fe88d855ee5a31f3a9e3f0c5 WatchSource:0}: Error finding container c6fef86b934abfea4c0eecb3f882d2eb6eaba384fe88d855ee5a31f3a9e3f0c5: Status 404 returned error can't find the container with id c6fef86b934abfea4c0eecb3f882d2eb6eaba384fe88d855ee5a31f3a9e3f0c5 Oct 14 09:25:21 crc kubenswrapper[4870]: I1014 09:25:21.860857 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-openstack-openstack-cell1-lrc5r" event={"ID":"34d9c179-4fe6-4431-b495-6803732cb3ee","Type":"ContainerStarted","Data":"c6fef86b934abfea4c0eecb3f882d2eb6eaba384fe88d855ee5a31f3a9e3f0c5"} Oct 14 09:25:22 crc kubenswrapper[4870]: I1014 09:25:22.879964 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-openstack-openstack-cell1-lrc5r" event={"ID":"34d9c179-4fe6-4431-b495-6803732cb3ee","Type":"ContainerStarted","Data":"921d85cf2340c08f961437820630639e48c0639e08629c49b1f3777334575344"} Oct 14 09:25:22 crc kubenswrapper[4870]: I1014 09:25:22.914642 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-openstack-openstack-cell1-lrc5r" podStartSLOduration=2.496362474 podStartE2EDuration="2.914617213s" podCreationTimestamp="2025-10-14 09:25:20 +0000 UTC" firstStartedPulling="2025-10-14 09:25:21.674582539 +0000 UTC m=+8657.371942920" lastFinishedPulling="2025-10-14 09:25:22.092837288 +0000 UTC m=+8657.790197659" observedRunningTime="2025-10-14 09:25:22.901006028 +0000 UTC m=+8658.598366399" watchObservedRunningTime="2025-10-14 09:25:22.914617213 +0000 UTC m=+8658.611977624" Oct 14 09:25:23 crc kubenswrapper[4870]: I1014 09:25:23.950602 4870 patch_prober.go:28] interesting pod/machine-config-daemon-7tvc7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 09:25:23 crc kubenswrapper[4870]: I1014 09:25:23.950964 4870 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 09:25:27 crc kubenswrapper[4870]: I1014 09:25:27.939986 4870 generic.go:334] "Generic (PLEG): container finished" podID="34d9c179-4fe6-4431-b495-6803732cb3ee" containerID="921d85cf2340c08f961437820630639e48c0639e08629c49b1f3777334575344" exitCode=0 Oct 14 09:25:27 crc kubenswrapper[4870]: I1014 09:25:27.940077 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-openstack-openstack-cell1-lrc5r" event={"ID":"34d9c179-4fe6-4431-b495-6803732cb3ee","Type":"ContainerDied","Data":"921d85cf2340c08f961437820630639e48c0639e08629c49b1f3777334575344"} Oct 14 09:25:28 crc kubenswrapper[4870]: I1014 09:25:28.118119 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-2j55s"] Oct 14 09:25:28 crc kubenswrapper[4870]: I1014 09:25:28.120969 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2j55s" Oct 14 09:25:28 crc kubenswrapper[4870]: I1014 09:25:28.133958 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-2j55s"] Oct 14 09:25:28 crc kubenswrapper[4870]: I1014 09:25:28.211752 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c3ac7db2-1733-4569-ba26-2f9ba0c4a045-catalog-content\") pod \"redhat-marketplace-2j55s\" (UID: \"c3ac7db2-1733-4569-ba26-2f9ba0c4a045\") " pod="openshift-marketplace/redhat-marketplace-2j55s" Oct 14 09:25:28 crc kubenswrapper[4870]: I1014 09:25:28.211929 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c3ac7db2-1733-4569-ba26-2f9ba0c4a045-utilities\") pod \"redhat-marketplace-2j55s\" (UID: \"c3ac7db2-1733-4569-ba26-2f9ba0c4a045\") " pod="openshift-marketplace/redhat-marketplace-2j55s" Oct 14 09:25:28 crc kubenswrapper[4870]: I1014 09:25:28.211975 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s69c9\" (UniqueName: \"kubernetes.io/projected/c3ac7db2-1733-4569-ba26-2f9ba0c4a045-kube-api-access-s69c9\") pod \"redhat-marketplace-2j55s\" (UID: \"c3ac7db2-1733-4569-ba26-2f9ba0c4a045\") " pod="openshift-marketplace/redhat-marketplace-2j55s" Oct 14 09:25:28 crc kubenswrapper[4870]: I1014 09:25:28.313616 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c3ac7db2-1733-4569-ba26-2f9ba0c4a045-catalog-content\") pod \"redhat-marketplace-2j55s\" (UID: \"c3ac7db2-1733-4569-ba26-2f9ba0c4a045\") " pod="openshift-marketplace/redhat-marketplace-2j55s" Oct 14 09:25:28 crc kubenswrapper[4870]: I1014 09:25:28.313732 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c3ac7db2-1733-4569-ba26-2f9ba0c4a045-utilities\") pod \"redhat-marketplace-2j55s\" (UID: \"c3ac7db2-1733-4569-ba26-2f9ba0c4a045\") " pod="openshift-marketplace/redhat-marketplace-2j55s" Oct 14 09:25:28 crc kubenswrapper[4870]: I1014 09:25:28.313758 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s69c9\" (UniqueName: \"kubernetes.io/projected/c3ac7db2-1733-4569-ba26-2f9ba0c4a045-kube-api-access-s69c9\") pod \"redhat-marketplace-2j55s\" (UID: \"c3ac7db2-1733-4569-ba26-2f9ba0c4a045\") " pod="openshift-marketplace/redhat-marketplace-2j55s" Oct 14 09:25:28 crc kubenswrapper[4870]: I1014 09:25:28.314139 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c3ac7db2-1733-4569-ba26-2f9ba0c4a045-catalog-content\") pod \"redhat-marketplace-2j55s\" (UID: \"c3ac7db2-1733-4569-ba26-2f9ba0c4a045\") " pod="openshift-marketplace/redhat-marketplace-2j55s" Oct 14 09:25:28 crc kubenswrapper[4870]: I1014 09:25:28.314388 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c3ac7db2-1733-4569-ba26-2f9ba0c4a045-utilities\") pod \"redhat-marketplace-2j55s\" (UID: \"c3ac7db2-1733-4569-ba26-2f9ba0c4a045\") " pod="openshift-marketplace/redhat-marketplace-2j55s" Oct 14 09:25:28 crc kubenswrapper[4870]: I1014 09:25:28.337008 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s69c9\" (UniqueName: \"kubernetes.io/projected/c3ac7db2-1733-4569-ba26-2f9ba0c4a045-kube-api-access-s69c9\") pod \"redhat-marketplace-2j55s\" (UID: \"c3ac7db2-1733-4569-ba26-2f9ba0c4a045\") " pod="openshift-marketplace/redhat-marketplace-2j55s" Oct 14 09:25:28 crc kubenswrapper[4870]: I1014 09:25:28.464647 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2j55s" Oct 14 09:25:29 crc kubenswrapper[4870]: I1014 09:25:29.024372 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-2j55s"] Oct 14 09:25:29 crc kubenswrapper[4870]: W1014 09:25:29.041852 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc3ac7db2_1733_4569_ba26_2f9ba0c4a045.slice/crio-9c114e005211059fdaeb7da752d44dab6d9ab2ff3a6543ee247f7e7bc15dc8e3 WatchSource:0}: Error finding container 9c114e005211059fdaeb7da752d44dab6d9ab2ff3a6543ee247f7e7bc15dc8e3: Status 404 returned error can't find the container with id 9c114e005211059fdaeb7da752d44dab6d9ab2ff3a6543ee247f7e7bc15dc8e3 Oct 14 09:25:29 crc kubenswrapper[4870]: I1014 09:25:29.422472 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-openstack-openstack-cell1-lrc5r" Oct 14 09:25:29 crc kubenswrapper[4870]: I1014 09:25:29.544821 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/34d9c179-4fe6-4431-b495-6803732cb3ee-ssh-key\") pod \"34d9c179-4fe6-4431-b495-6803732cb3ee\" (UID: \"34d9c179-4fe6-4431-b495-6803732cb3ee\") " Oct 14 09:25:29 crc kubenswrapper[4870]: I1014 09:25:29.544886 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qbgqh\" (UniqueName: \"kubernetes.io/projected/34d9c179-4fe6-4431-b495-6803732cb3ee-kube-api-access-qbgqh\") pod \"34d9c179-4fe6-4431-b495-6803732cb3ee\" (UID: \"34d9c179-4fe6-4431-b495-6803732cb3ee\") " Oct 14 09:25:29 crc kubenswrapper[4870]: I1014 09:25:29.544947 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/34d9c179-4fe6-4431-b495-6803732cb3ee-ceph\") pod \"34d9c179-4fe6-4431-b495-6803732cb3ee\" (UID: \"34d9c179-4fe6-4431-b495-6803732cb3ee\") " Oct 14 09:25:29 crc kubenswrapper[4870]: I1014 09:25:29.545179 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/34d9c179-4fe6-4431-b495-6803732cb3ee-inventory\") pod \"34d9c179-4fe6-4431-b495-6803732cb3ee\" (UID: \"34d9c179-4fe6-4431-b495-6803732cb3ee\") " Oct 14 09:25:29 crc kubenswrapper[4870]: I1014 09:25:29.558562 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/34d9c179-4fe6-4431-b495-6803732cb3ee-kube-api-access-qbgqh" (OuterVolumeSpecName: "kube-api-access-qbgqh") pod "34d9c179-4fe6-4431-b495-6803732cb3ee" (UID: "34d9c179-4fe6-4431-b495-6803732cb3ee"). InnerVolumeSpecName "kube-api-access-qbgqh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:25:29 crc kubenswrapper[4870]: I1014 09:25:29.569321 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/34d9c179-4fe6-4431-b495-6803732cb3ee-ceph" (OuterVolumeSpecName: "ceph") pod "34d9c179-4fe6-4431-b495-6803732cb3ee" (UID: "34d9c179-4fe6-4431-b495-6803732cb3ee"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:25:29 crc kubenswrapper[4870]: I1014 09:25:29.578528 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/34d9c179-4fe6-4431-b495-6803732cb3ee-inventory" (OuterVolumeSpecName: "inventory") pod "34d9c179-4fe6-4431-b495-6803732cb3ee" (UID: "34d9c179-4fe6-4431-b495-6803732cb3ee"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:25:29 crc kubenswrapper[4870]: I1014 09:25:29.600687 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/34d9c179-4fe6-4431-b495-6803732cb3ee-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "34d9c179-4fe6-4431-b495-6803732cb3ee" (UID: "34d9c179-4fe6-4431-b495-6803732cb3ee"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:25:29 crc kubenswrapper[4870]: I1014 09:25:29.647427 4870 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/34d9c179-4fe6-4431-b495-6803732cb3ee-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 14 09:25:29 crc kubenswrapper[4870]: I1014 09:25:29.647491 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qbgqh\" (UniqueName: \"kubernetes.io/projected/34d9c179-4fe6-4431-b495-6803732cb3ee-kube-api-access-qbgqh\") on node \"crc\" DevicePath \"\"" Oct 14 09:25:29 crc kubenswrapper[4870]: I1014 09:25:29.647511 4870 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/34d9c179-4fe6-4431-b495-6803732cb3ee-ceph\") on node \"crc\" DevicePath \"\"" Oct 14 09:25:29 crc kubenswrapper[4870]: I1014 09:25:29.647524 4870 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/34d9c179-4fe6-4431-b495-6803732cb3ee-inventory\") on node \"crc\" DevicePath \"\"" Oct 14 09:25:29 crc kubenswrapper[4870]: I1014 09:25:29.986899 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-openstack-openstack-cell1-lrc5r" event={"ID":"34d9c179-4fe6-4431-b495-6803732cb3ee","Type":"ContainerDied","Data":"c6fef86b934abfea4c0eecb3f882d2eb6eaba384fe88d855ee5a31f3a9e3f0c5"} Oct 14 09:25:29 crc kubenswrapper[4870]: I1014 09:25:29.986987 4870 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c6fef86b934abfea4c0eecb3f882d2eb6eaba384fe88d855ee5a31f3a9e3f0c5" Oct 14 09:25:29 crc kubenswrapper[4870]: I1014 09:25:29.987106 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-openstack-openstack-cell1-lrc5r" Oct 14 09:25:29 crc kubenswrapper[4870]: I1014 09:25:29.994219 4870 generic.go:334] "Generic (PLEG): container finished" podID="c3ac7db2-1733-4569-ba26-2f9ba0c4a045" containerID="7ab065224e82ca82c73fe0df2a8fbcd30e1de92538a58261eba76fde5fc6d572" exitCode=0 Oct 14 09:25:29 crc kubenswrapper[4870]: I1014 09:25:29.994299 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2j55s" event={"ID":"c3ac7db2-1733-4569-ba26-2f9ba0c4a045","Type":"ContainerDied","Data":"7ab065224e82ca82c73fe0df2a8fbcd30e1de92538a58261eba76fde5fc6d572"} Oct 14 09:25:29 crc kubenswrapper[4870]: I1014 09:25:29.994346 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2j55s" event={"ID":"c3ac7db2-1733-4569-ba26-2f9ba0c4a045","Type":"ContainerStarted","Data":"9c114e005211059fdaeb7da752d44dab6d9ab2ff3a6543ee247f7e7bc15dc8e3"} Oct 14 09:25:30 crc kubenswrapper[4870]: I1014 09:25:30.086894 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-openstack-openstack-cell1-54frh"] Oct 14 09:25:30 crc kubenswrapper[4870]: E1014 09:25:30.087465 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34d9c179-4fe6-4431-b495-6803732cb3ee" containerName="validate-network-openstack-openstack-cell1" Oct 14 09:25:30 crc kubenswrapper[4870]: I1014 09:25:30.087482 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="34d9c179-4fe6-4431-b495-6803732cb3ee" containerName="validate-network-openstack-openstack-cell1" Oct 14 09:25:30 crc kubenswrapper[4870]: I1014 09:25:30.087756 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="34d9c179-4fe6-4431-b495-6803732cb3ee" containerName="validate-network-openstack-openstack-cell1" Oct 14 09:25:30 crc kubenswrapper[4870]: I1014 09:25:30.088704 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-openstack-openstack-cell1-54frh" Oct 14 09:25:30 crc kubenswrapper[4870]: I1014 09:25:30.092244 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-wrz7w" Oct 14 09:25:30 crc kubenswrapper[4870]: I1014 09:25:30.093345 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 14 09:25:30 crc kubenswrapper[4870]: I1014 09:25:30.100642 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-openstack-openstack-cell1-54frh"] Oct 14 09:25:30 crc kubenswrapper[4870]: I1014 09:25:30.161211 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qb74v\" (UniqueName: \"kubernetes.io/projected/faed3d6f-4637-44f8-8925-0e82407fcd44-kube-api-access-qb74v\") pod \"install-os-openstack-openstack-cell1-54frh\" (UID: \"faed3d6f-4637-44f8-8925-0e82407fcd44\") " pod="openstack/install-os-openstack-openstack-cell1-54frh" Oct 14 09:25:30 crc kubenswrapper[4870]: I1014 09:25:30.162231 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/faed3d6f-4637-44f8-8925-0e82407fcd44-ceph\") pod \"install-os-openstack-openstack-cell1-54frh\" (UID: \"faed3d6f-4637-44f8-8925-0e82407fcd44\") " pod="openstack/install-os-openstack-openstack-cell1-54frh" Oct 14 09:25:30 crc kubenswrapper[4870]: I1014 09:25:30.162342 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/faed3d6f-4637-44f8-8925-0e82407fcd44-ssh-key\") pod \"install-os-openstack-openstack-cell1-54frh\" (UID: \"faed3d6f-4637-44f8-8925-0e82407fcd44\") " pod="openstack/install-os-openstack-openstack-cell1-54frh" Oct 14 09:25:30 crc kubenswrapper[4870]: I1014 09:25:30.162665 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/faed3d6f-4637-44f8-8925-0e82407fcd44-inventory\") pod \"install-os-openstack-openstack-cell1-54frh\" (UID: \"faed3d6f-4637-44f8-8925-0e82407fcd44\") " pod="openstack/install-os-openstack-openstack-cell1-54frh" Oct 14 09:25:30 crc kubenswrapper[4870]: I1014 09:25:30.270193 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/faed3d6f-4637-44f8-8925-0e82407fcd44-inventory\") pod \"install-os-openstack-openstack-cell1-54frh\" (UID: \"faed3d6f-4637-44f8-8925-0e82407fcd44\") " pod="openstack/install-os-openstack-openstack-cell1-54frh" Oct 14 09:25:30 crc kubenswrapper[4870]: I1014 09:25:30.270479 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qb74v\" (UniqueName: \"kubernetes.io/projected/faed3d6f-4637-44f8-8925-0e82407fcd44-kube-api-access-qb74v\") pod \"install-os-openstack-openstack-cell1-54frh\" (UID: \"faed3d6f-4637-44f8-8925-0e82407fcd44\") " pod="openstack/install-os-openstack-openstack-cell1-54frh" Oct 14 09:25:30 crc kubenswrapper[4870]: I1014 09:25:30.271086 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/faed3d6f-4637-44f8-8925-0e82407fcd44-ceph\") pod \"install-os-openstack-openstack-cell1-54frh\" (UID: \"faed3d6f-4637-44f8-8925-0e82407fcd44\") " pod="openstack/install-os-openstack-openstack-cell1-54frh" Oct 14 09:25:30 crc kubenswrapper[4870]: I1014 09:25:30.271170 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/faed3d6f-4637-44f8-8925-0e82407fcd44-ssh-key\") pod \"install-os-openstack-openstack-cell1-54frh\" (UID: \"faed3d6f-4637-44f8-8925-0e82407fcd44\") " pod="openstack/install-os-openstack-openstack-cell1-54frh" Oct 14 09:25:30 crc kubenswrapper[4870]: I1014 09:25:30.278583 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/faed3d6f-4637-44f8-8925-0e82407fcd44-ceph\") pod \"install-os-openstack-openstack-cell1-54frh\" (UID: \"faed3d6f-4637-44f8-8925-0e82407fcd44\") " pod="openstack/install-os-openstack-openstack-cell1-54frh" Oct 14 09:25:30 crc kubenswrapper[4870]: I1014 09:25:30.278770 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/faed3d6f-4637-44f8-8925-0e82407fcd44-inventory\") pod \"install-os-openstack-openstack-cell1-54frh\" (UID: \"faed3d6f-4637-44f8-8925-0e82407fcd44\") " pod="openstack/install-os-openstack-openstack-cell1-54frh" Oct 14 09:25:30 crc kubenswrapper[4870]: I1014 09:25:30.281117 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/faed3d6f-4637-44f8-8925-0e82407fcd44-ssh-key\") pod \"install-os-openstack-openstack-cell1-54frh\" (UID: \"faed3d6f-4637-44f8-8925-0e82407fcd44\") " pod="openstack/install-os-openstack-openstack-cell1-54frh" Oct 14 09:25:30 crc kubenswrapper[4870]: I1014 09:25:30.292363 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qb74v\" (UniqueName: \"kubernetes.io/projected/faed3d6f-4637-44f8-8925-0e82407fcd44-kube-api-access-qb74v\") pod \"install-os-openstack-openstack-cell1-54frh\" (UID: \"faed3d6f-4637-44f8-8925-0e82407fcd44\") " pod="openstack/install-os-openstack-openstack-cell1-54frh" Oct 14 09:25:30 crc kubenswrapper[4870]: I1014 09:25:30.419383 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-openstack-openstack-cell1-54frh" Oct 14 09:25:31 crc kubenswrapper[4870]: I1014 09:25:31.069931 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-openstack-openstack-cell1-54frh"] Oct 14 09:25:32 crc kubenswrapper[4870]: I1014 09:25:32.024728 4870 generic.go:334] "Generic (PLEG): container finished" podID="c3ac7db2-1733-4569-ba26-2f9ba0c4a045" containerID="68a4cea724ab4e31a49a10093f9af51bed6f07b2b2be1a5ad2d2c48262cc06a3" exitCode=0 Oct 14 09:25:32 crc kubenswrapper[4870]: I1014 09:25:32.025343 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2j55s" event={"ID":"c3ac7db2-1733-4569-ba26-2f9ba0c4a045","Type":"ContainerDied","Data":"68a4cea724ab4e31a49a10093f9af51bed6f07b2b2be1a5ad2d2c48262cc06a3"} Oct 14 09:25:32 crc kubenswrapper[4870]: I1014 09:25:32.031032 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-openstack-openstack-cell1-54frh" event={"ID":"faed3d6f-4637-44f8-8925-0e82407fcd44","Type":"ContainerStarted","Data":"b014edbf89143aa5b8cccbe618d85eccd8cc29d9db3aa71566e0d135dec82cbf"} Oct 14 09:25:32 crc kubenswrapper[4870]: I1014 09:25:32.031080 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-openstack-openstack-cell1-54frh" event={"ID":"faed3d6f-4637-44f8-8925-0e82407fcd44","Type":"ContainerStarted","Data":"530b8b2bc995d83f9cd65e89dc3b23a1042fa5953596b8d882c9b1911832a584"} Oct 14 09:25:32 crc kubenswrapper[4870]: I1014 09:25:32.066045 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-openstack-openstack-cell1-54frh" podStartSLOduration=1.6367398560000002 podStartE2EDuration="2.066027756s" podCreationTimestamp="2025-10-14 09:25:30 +0000 UTC" firstStartedPulling="2025-10-14 09:25:31.083906456 +0000 UTC m=+8666.781266837" lastFinishedPulling="2025-10-14 09:25:31.513194366 +0000 UTC m=+8667.210554737" observedRunningTime="2025-10-14 09:25:32.062289104 +0000 UTC m=+8667.759649475" watchObservedRunningTime="2025-10-14 09:25:32.066027756 +0000 UTC m=+8667.763388127" Oct 14 09:25:34 crc kubenswrapper[4870]: I1014 09:25:34.063581 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2j55s" event={"ID":"c3ac7db2-1733-4569-ba26-2f9ba0c4a045","Type":"ContainerStarted","Data":"6dea80739bffdc0296d3ccd48fe2d68423a47ec2dfa6fc4f830243a45c18e277"} Oct 14 09:25:34 crc kubenswrapper[4870]: I1014 09:25:34.084888 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-2j55s" podStartSLOduration=3.137929094 podStartE2EDuration="6.084868959s" podCreationTimestamp="2025-10-14 09:25:28 +0000 UTC" firstStartedPulling="2025-10-14 09:25:29.998397222 +0000 UTC m=+8665.695757633" lastFinishedPulling="2025-10-14 09:25:32.945337117 +0000 UTC m=+8668.642697498" observedRunningTime="2025-10-14 09:25:34.078627675 +0000 UTC m=+8669.775988086" watchObservedRunningTime="2025-10-14 09:25:34.084868959 +0000 UTC m=+8669.782229330" Oct 14 09:25:38 crc kubenswrapper[4870]: I1014 09:25:38.465185 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-2j55s" Oct 14 09:25:38 crc kubenswrapper[4870]: I1014 09:25:38.465584 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-2j55s" Oct 14 09:25:38 crc kubenswrapper[4870]: I1014 09:25:38.523926 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-2j55s" Oct 14 09:25:39 crc kubenswrapper[4870]: I1014 09:25:39.201289 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-2j55s" Oct 14 09:25:39 crc kubenswrapper[4870]: I1014 09:25:39.253336 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-2j55s"] Oct 14 09:25:41 crc kubenswrapper[4870]: I1014 09:25:41.133782 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-2j55s" podUID="c3ac7db2-1733-4569-ba26-2f9ba0c4a045" containerName="registry-server" containerID="cri-o://6dea80739bffdc0296d3ccd48fe2d68423a47ec2dfa6fc4f830243a45c18e277" gracePeriod=2 Oct 14 09:25:41 crc kubenswrapper[4870]: I1014 09:25:41.686429 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2j55s" Oct 14 09:25:41 crc kubenswrapper[4870]: I1014 09:25:41.750261 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c3ac7db2-1733-4569-ba26-2f9ba0c4a045-utilities\") pod \"c3ac7db2-1733-4569-ba26-2f9ba0c4a045\" (UID: \"c3ac7db2-1733-4569-ba26-2f9ba0c4a045\") " Oct 14 09:25:41 crc kubenswrapper[4870]: I1014 09:25:41.750393 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c3ac7db2-1733-4569-ba26-2f9ba0c4a045-catalog-content\") pod \"c3ac7db2-1733-4569-ba26-2f9ba0c4a045\" (UID: \"c3ac7db2-1733-4569-ba26-2f9ba0c4a045\") " Oct 14 09:25:41 crc kubenswrapper[4870]: I1014 09:25:41.750630 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s69c9\" (UniqueName: \"kubernetes.io/projected/c3ac7db2-1733-4569-ba26-2f9ba0c4a045-kube-api-access-s69c9\") pod \"c3ac7db2-1733-4569-ba26-2f9ba0c4a045\" (UID: \"c3ac7db2-1733-4569-ba26-2f9ba0c4a045\") " Oct 14 09:25:41 crc kubenswrapper[4870]: I1014 09:25:41.754793 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c3ac7db2-1733-4569-ba26-2f9ba0c4a045-utilities" (OuterVolumeSpecName: "utilities") pod "c3ac7db2-1733-4569-ba26-2f9ba0c4a045" (UID: "c3ac7db2-1733-4569-ba26-2f9ba0c4a045"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 09:25:41 crc kubenswrapper[4870]: I1014 09:25:41.765659 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c3ac7db2-1733-4569-ba26-2f9ba0c4a045-kube-api-access-s69c9" (OuterVolumeSpecName: "kube-api-access-s69c9") pod "c3ac7db2-1733-4569-ba26-2f9ba0c4a045" (UID: "c3ac7db2-1733-4569-ba26-2f9ba0c4a045"). InnerVolumeSpecName "kube-api-access-s69c9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:25:41 crc kubenswrapper[4870]: I1014 09:25:41.811259 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c3ac7db2-1733-4569-ba26-2f9ba0c4a045-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c3ac7db2-1733-4569-ba26-2f9ba0c4a045" (UID: "c3ac7db2-1733-4569-ba26-2f9ba0c4a045"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 09:25:41 crc kubenswrapper[4870]: I1014 09:25:41.855809 4870 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c3ac7db2-1733-4569-ba26-2f9ba0c4a045-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 09:25:41 crc kubenswrapper[4870]: I1014 09:25:41.855859 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s69c9\" (UniqueName: \"kubernetes.io/projected/c3ac7db2-1733-4569-ba26-2f9ba0c4a045-kube-api-access-s69c9\") on node \"crc\" DevicePath \"\"" Oct 14 09:25:41 crc kubenswrapper[4870]: I1014 09:25:41.855876 4870 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c3ac7db2-1733-4569-ba26-2f9ba0c4a045-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 09:25:42 crc kubenswrapper[4870]: I1014 09:25:42.145832 4870 generic.go:334] "Generic (PLEG): container finished" podID="c3ac7db2-1733-4569-ba26-2f9ba0c4a045" containerID="6dea80739bffdc0296d3ccd48fe2d68423a47ec2dfa6fc4f830243a45c18e277" exitCode=0 Oct 14 09:25:42 crc kubenswrapper[4870]: I1014 09:25:42.145882 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2j55s" event={"ID":"c3ac7db2-1733-4569-ba26-2f9ba0c4a045","Type":"ContainerDied","Data":"6dea80739bffdc0296d3ccd48fe2d68423a47ec2dfa6fc4f830243a45c18e277"} Oct 14 09:25:42 crc kubenswrapper[4870]: I1014 09:25:42.145925 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2j55s" event={"ID":"c3ac7db2-1733-4569-ba26-2f9ba0c4a045","Type":"ContainerDied","Data":"9c114e005211059fdaeb7da752d44dab6d9ab2ff3a6543ee247f7e7bc15dc8e3"} Oct 14 09:25:42 crc kubenswrapper[4870]: I1014 09:25:42.145948 4870 scope.go:117] "RemoveContainer" containerID="6dea80739bffdc0296d3ccd48fe2d68423a47ec2dfa6fc4f830243a45c18e277" Oct 14 09:25:42 crc kubenswrapper[4870]: I1014 09:25:42.145949 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2j55s" Oct 14 09:25:42 crc kubenswrapper[4870]: I1014 09:25:42.188502 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-2j55s"] Oct 14 09:25:42 crc kubenswrapper[4870]: I1014 09:25:42.196340 4870 scope.go:117] "RemoveContainer" containerID="68a4cea724ab4e31a49a10093f9af51bed6f07b2b2be1a5ad2d2c48262cc06a3" Oct 14 09:25:42 crc kubenswrapper[4870]: I1014 09:25:42.199257 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-2j55s"] Oct 14 09:25:42 crc kubenswrapper[4870]: I1014 09:25:42.218781 4870 scope.go:117] "RemoveContainer" containerID="7ab065224e82ca82c73fe0df2a8fbcd30e1de92538a58261eba76fde5fc6d572" Oct 14 09:25:42 crc kubenswrapper[4870]: I1014 09:25:42.276007 4870 scope.go:117] "RemoveContainer" containerID="6dea80739bffdc0296d3ccd48fe2d68423a47ec2dfa6fc4f830243a45c18e277" Oct 14 09:25:42 crc kubenswrapper[4870]: E1014 09:25:42.276502 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6dea80739bffdc0296d3ccd48fe2d68423a47ec2dfa6fc4f830243a45c18e277\": container with ID starting with 6dea80739bffdc0296d3ccd48fe2d68423a47ec2dfa6fc4f830243a45c18e277 not found: ID does not exist" containerID="6dea80739bffdc0296d3ccd48fe2d68423a47ec2dfa6fc4f830243a45c18e277" Oct 14 09:25:42 crc kubenswrapper[4870]: I1014 09:25:42.276578 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6dea80739bffdc0296d3ccd48fe2d68423a47ec2dfa6fc4f830243a45c18e277"} err="failed to get container status \"6dea80739bffdc0296d3ccd48fe2d68423a47ec2dfa6fc4f830243a45c18e277\": rpc error: code = NotFound desc = could not find container \"6dea80739bffdc0296d3ccd48fe2d68423a47ec2dfa6fc4f830243a45c18e277\": container with ID starting with 6dea80739bffdc0296d3ccd48fe2d68423a47ec2dfa6fc4f830243a45c18e277 not found: ID does not exist" Oct 14 09:25:42 crc kubenswrapper[4870]: I1014 09:25:42.276606 4870 scope.go:117] "RemoveContainer" containerID="68a4cea724ab4e31a49a10093f9af51bed6f07b2b2be1a5ad2d2c48262cc06a3" Oct 14 09:25:42 crc kubenswrapper[4870]: E1014 09:25:42.277151 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"68a4cea724ab4e31a49a10093f9af51bed6f07b2b2be1a5ad2d2c48262cc06a3\": container with ID starting with 68a4cea724ab4e31a49a10093f9af51bed6f07b2b2be1a5ad2d2c48262cc06a3 not found: ID does not exist" containerID="68a4cea724ab4e31a49a10093f9af51bed6f07b2b2be1a5ad2d2c48262cc06a3" Oct 14 09:25:42 crc kubenswrapper[4870]: I1014 09:25:42.277170 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"68a4cea724ab4e31a49a10093f9af51bed6f07b2b2be1a5ad2d2c48262cc06a3"} err="failed to get container status \"68a4cea724ab4e31a49a10093f9af51bed6f07b2b2be1a5ad2d2c48262cc06a3\": rpc error: code = NotFound desc = could not find container \"68a4cea724ab4e31a49a10093f9af51bed6f07b2b2be1a5ad2d2c48262cc06a3\": container with ID starting with 68a4cea724ab4e31a49a10093f9af51bed6f07b2b2be1a5ad2d2c48262cc06a3 not found: ID does not exist" Oct 14 09:25:42 crc kubenswrapper[4870]: I1014 09:25:42.277182 4870 scope.go:117] "RemoveContainer" containerID="7ab065224e82ca82c73fe0df2a8fbcd30e1de92538a58261eba76fde5fc6d572" Oct 14 09:25:42 crc kubenswrapper[4870]: E1014 09:25:42.277581 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7ab065224e82ca82c73fe0df2a8fbcd30e1de92538a58261eba76fde5fc6d572\": container with ID starting with 7ab065224e82ca82c73fe0df2a8fbcd30e1de92538a58261eba76fde5fc6d572 not found: ID does not exist" containerID="7ab065224e82ca82c73fe0df2a8fbcd30e1de92538a58261eba76fde5fc6d572" Oct 14 09:25:42 crc kubenswrapper[4870]: I1014 09:25:42.277611 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7ab065224e82ca82c73fe0df2a8fbcd30e1de92538a58261eba76fde5fc6d572"} err="failed to get container status \"7ab065224e82ca82c73fe0df2a8fbcd30e1de92538a58261eba76fde5fc6d572\": rpc error: code = NotFound desc = could not find container \"7ab065224e82ca82c73fe0df2a8fbcd30e1de92538a58261eba76fde5fc6d572\": container with ID starting with 7ab065224e82ca82c73fe0df2a8fbcd30e1de92538a58261eba76fde5fc6d572 not found: ID does not exist" Oct 14 09:25:43 crc kubenswrapper[4870]: I1014 09:25:43.056777 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c3ac7db2-1733-4569-ba26-2f9ba0c4a045" path="/var/lib/kubelet/pods/c3ac7db2-1733-4569-ba26-2f9ba0c4a045/volumes" Oct 14 09:25:53 crc kubenswrapper[4870]: I1014 09:25:53.950680 4870 patch_prober.go:28] interesting pod/machine-config-daemon-7tvc7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 09:25:53 crc kubenswrapper[4870]: I1014 09:25:53.951224 4870 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 09:25:53 crc kubenswrapper[4870]: I1014 09:25:53.951286 4870 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" Oct 14 09:25:53 crc kubenswrapper[4870]: I1014 09:25:53.952317 4870 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c92a8ddf378f6420fa6f685f5a67ccca0eb5b167c543057b9d31affa78843af8"} pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 14 09:25:53 crc kubenswrapper[4870]: I1014 09:25:53.952404 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerName="machine-config-daemon" containerID="cri-o://c92a8ddf378f6420fa6f685f5a67ccca0eb5b167c543057b9d31affa78843af8" gracePeriod=600 Oct 14 09:25:54 crc kubenswrapper[4870]: I1014 09:25:54.304588 4870 generic.go:334] "Generic (PLEG): container finished" podID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerID="c92a8ddf378f6420fa6f685f5a67ccca0eb5b167c543057b9d31affa78843af8" exitCode=0 Oct 14 09:25:54 crc kubenswrapper[4870]: I1014 09:25:54.304727 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" event={"ID":"874a5e45-dffd-4d17-b609-b7d3ed2eab07","Type":"ContainerDied","Data":"c92a8ddf378f6420fa6f685f5a67ccca0eb5b167c543057b9d31affa78843af8"} Oct 14 09:25:54 crc kubenswrapper[4870]: I1014 09:25:54.305058 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" event={"ID":"874a5e45-dffd-4d17-b609-b7d3ed2eab07","Type":"ContainerStarted","Data":"3eea07ce0e4dde9363f92f5f44f530c26c10b42e335dd2bafb6764d9cbdc203c"} Oct 14 09:25:54 crc kubenswrapper[4870]: I1014 09:25:54.305075 4870 scope.go:117] "RemoveContainer" containerID="5a354d0d25c9ddf421bfdb89e3b92b75f73097cc5398e19453b6e98125f9ecf5" Oct 14 09:26:00 crc kubenswrapper[4870]: I1014 09:26:00.378306 4870 generic.go:334] "Generic (PLEG): container finished" podID="ed0d4c93-b579-4253-867f-d191d43c35c7" containerID="528bb0301df717845635b2eaea75b95255d1b6840d6443eeb25e2d0d8009ce31" exitCode=0 Oct 14 09:26:00 crc kubenswrapper[4870]: I1014 09:26:00.378354 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-openstack-openstack-networker-wdjn4" event={"ID":"ed0d4c93-b579-4253-867f-d191d43c35c7","Type":"ContainerDied","Data":"528bb0301df717845635b2eaea75b95255d1b6840d6443eeb25e2d0d8009ce31"} Oct 14 09:26:01 crc kubenswrapper[4870]: I1014 09:26:01.932522 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-openstack-openstack-networker-wdjn4" Oct 14 09:26:02 crc kubenswrapper[4870]: I1014 09:26:02.035605 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xrdbh\" (UniqueName: \"kubernetes.io/projected/ed0d4c93-b579-4253-867f-d191d43c35c7-kube-api-access-xrdbh\") pod \"ed0d4c93-b579-4253-867f-d191d43c35c7\" (UID: \"ed0d4c93-b579-4253-867f-d191d43c35c7\") " Oct 14 09:26:02 crc kubenswrapper[4870]: I1014 09:26:02.035696 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ed0d4c93-b579-4253-867f-d191d43c35c7-inventory\") pod \"ed0d4c93-b579-4253-867f-d191d43c35c7\" (UID: \"ed0d4c93-b579-4253-867f-d191d43c35c7\") " Oct 14 09:26:02 crc kubenswrapper[4870]: I1014 09:26:02.035877 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ed0d4c93-b579-4253-867f-d191d43c35c7-ssh-key\") pod \"ed0d4c93-b579-4253-867f-d191d43c35c7\" (UID: \"ed0d4c93-b579-4253-867f-d191d43c35c7\") " Oct 14 09:26:02 crc kubenswrapper[4870]: I1014 09:26:02.069770 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ed0d4c93-b579-4253-867f-d191d43c35c7-kube-api-access-xrdbh" (OuterVolumeSpecName: "kube-api-access-xrdbh") pod "ed0d4c93-b579-4253-867f-d191d43c35c7" (UID: "ed0d4c93-b579-4253-867f-d191d43c35c7"). InnerVolumeSpecName "kube-api-access-xrdbh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:26:02 crc kubenswrapper[4870]: I1014 09:26:02.091549 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ed0d4c93-b579-4253-867f-d191d43c35c7-inventory" (OuterVolumeSpecName: "inventory") pod "ed0d4c93-b579-4253-867f-d191d43c35c7" (UID: "ed0d4c93-b579-4253-867f-d191d43c35c7"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:26:02 crc kubenswrapper[4870]: I1014 09:26:02.128963 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ed0d4c93-b579-4253-867f-d191d43c35c7-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "ed0d4c93-b579-4253-867f-d191d43c35c7" (UID: "ed0d4c93-b579-4253-867f-d191d43c35c7"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:26:02 crc kubenswrapper[4870]: I1014 09:26:02.139835 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xrdbh\" (UniqueName: \"kubernetes.io/projected/ed0d4c93-b579-4253-867f-d191d43c35c7-kube-api-access-xrdbh\") on node \"crc\" DevicePath \"\"" Oct 14 09:26:02 crc kubenswrapper[4870]: I1014 09:26:02.139872 4870 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ed0d4c93-b579-4253-867f-d191d43c35c7-inventory\") on node \"crc\" DevicePath \"\"" Oct 14 09:26:02 crc kubenswrapper[4870]: I1014 09:26:02.139881 4870 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ed0d4c93-b579-4253-867f-d191d43c35c7-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 14 09:26:02 crc kubenswrapper[4870]: I1014 09:26:02.410627 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-openstack-openstack-networker-wdjn4" event={"ID":"ed0d4c93-b579-4253-867f-d191d43c35c7","Type":"ContainerDied","Data":"375ec7325f35e1398712dca62c42da8c9b50484035fabfcb726d6a25cba89c73"} Oct 14 09:26:02 crc kubenswrapper[4870]: I1014 09:26:02.410678 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-openstack-openstack-networker-wdjn4" Oct 14 09:26:02 crc kubenswrapper[4870]: I1014 09:26:02.410686 4870 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="375ec7325f35e1398712dca62c42da8c9b50484035fabfcb726d6a25cba89c73" Oct 14 09:26:02 crc kubenswrapper[4870]: I1014 09:26:02.489351 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-openstack-openstack-networker-nwsds"] Oct 14 09:26:02 crc kubenswrapper[4870]: E1014 09:26:02.489912 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c3ac7db2-1733-4569-ba26-2f9ba0c4a045" containerName="extract-utilities" Oct 14 09:26:02 crc kubenswrapper[4870]: I1014 09:26:02.490084 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="c3ac7db2-1733-4569-ba26-2f9ba0c4a045" containerName="extract-utilities" Oct 14 09:26:02 crc kubenswrapper[4870]: E1014 09:26:02.490137 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed0d4c93-b579-4253-867f-d191d43c35c7" containerName="install-os-openstack-openstack-networker" Oct 14 09:26:02 crc kubenswrapper[4870]: I1014 09:26:02.490148 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed0d4c93-b579-4253-867f-d191d43c35c7" containerName="install-os-openstack-openstack-networker" Oct 14 09:26:02 crc kubenswrapper[4870]: E1014 09:26:02.490162 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c3ac7db2-1733-4569-ba26-2f9ba0c4a045" containerName="extract-content" Oct 14 09:26:02 crc kubenswrapper[4870]: I1014 09:26:02.490170 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="c3ac7db2-1733-4569-ba26-2f9ba0c4a045" containerName="extract-content" Oct 14 09:26:02 crc kubenswrapper[4870]: E1014 09:26:02.490196 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c3ac7db2-1733-4569-ba26-2f9ba0c4a045" containerName="registry-server" Oct 14 09:26:02 crc kubenswrapper[4870]: I1014 09:26:02.490203 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="c3ac7db2-1733-4569-ba26-2f9ba0c4a045" containerName="registry-server" Oct 14 09:26:02 crc kubenswrapper[4870]: I1014 09:26:02.490540 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="ed0d4c93-b579-4253-867f-d191d43c35c7" containerName="install-os-openstack-openstack-networker" Oct 14 09:26:02 crc kubenswrapper[4870]: I1014 09:26:02.490590 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="c3ac7db2-1733-4569-ba26-2f9ba0c4a045" containerName="registry-server" Oct 14 09:26:02 crc kubenswrapper[4870]: I1014 09:26:02.491542 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-openstack-openstack-networker-nwsds" Oct 14 09:26:02 crc kubenswrapper[4870]: I1014 09:26:02.502180 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-networker-dockercfg-54hzr" Oct 14 09:26:02 crc kubenswrapper[4870]: I1014 09:26:02.502505 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-networker" Oct 14 09:26:02 crc kubenswrapper[4870]: I1014 09:26:02.508856 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-openstack-openstack-networker-nwsds"] Oct 14 09:26:02 crc kubenswrapper[4870]: I1014 09:26:02.548787 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h4c8k\" (UniqueName: \"kubernetes.io/projected/324ce306-ca93-447c-a19c-711764988cc9-kube-api-access-h4c8k\") pod \"configure-os-openstack-openstack-networker-nwsds\" (UID: \"324ce306-ca93-447c-a19c-711764988cc9\") " pod="openstack/configure-os-openstack-openstack-networker-nwsds" Oct 14 09:26:02 crc kubenswrapper[4870]: I1014 09:26:02.548849 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/324ce306-ca93-447c-a19c-711764988cc9-ssh-key\") pod \"configure-os-openstack-openstack-networker-nwsds\" (UID: \"324ce306-ca93-447c-a19c-711764988cc9\") " pod="openstack/configure-os-openstack-openstack-networker-nwsds" Oct 14 09:26:02 crc kubenswrapper[4870]: I1014 09:26:02.548896 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/324ce306-ca93-447c-a19c-711764988cc9-inventory\") pod \"configure-os-openstack-openstack-networker-nwsds\" (UID: \"324ce306-ca93-447c-a19c-711764988cc9\") " pod="openstack/configure-os-openstack-openstack-networker-nwsds" Oct 14 09:26:02 crc kubenswrapper[4870]: I1014 09:26:02.650956 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h4c8k\" (UniqueName: \"kubernetes.io/projected/324ce306-ca93-447c-a19c-711764988cc9-kube-api-access-h4c8k\") pod \"configure-os-openstack-openstack-networker-nwsds\" (UID: \"324ce306-ca93-447c-a19c-711764988cc9\") " pod="openstack/configure-os-openstack-openstack-networker-nwsds" Oct 14 09:26:02 crc kubenswrapper[4870]: I1014 09:26:02.651392 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/324ce306-ca93-447c-a19c-711764988cc9-ssh-key\") pod \"configure-os-openstack-openstack-networker-nwsds\" (UID: \"324ce306-ca93-447c-a19c-711764988cc9\") " pod="openstack/configure-os-openstack-openstack-networker-nwsds" Oct 14 09:26:02 crc kubenswrapper[4870]: I1014 09:26:02.651500 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/324ce306-ca93-447c-a19c-711764988cc9-inventory\") pod \"configure-os-openstack-openstack-networker-nwsds\" (UID: \"324ce306-ca93-447c-a19c-711764988cc9\") " pod="openstack/configure-os-openstack-openstack-networker-nwsds" Oct 14 09:26:02 crc kubenswrapper[4870]: I1014 09:26:02.655352 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/324ce306-ca93-447c-a19c-711764988cc9-inventory\") pod \"configure-os-openstack-openstack-networker-nwsds\" (UID: \"324ce306-ca93-447c-a19c-711764988cc9\") " pod="openstack/configure-os-openstack-openstack-networker-nwsds" Oct 14 09:26:02 crc kubenswrapper[4870]: I1014 09:26:02.657699 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/324ce306-ca93-447c-a19c-711764988cc9-ssh-key\") pod \"configure-os-openstack-openstack-networker-nwsds\" (UID: \"324ce306-ca93-447c-a19c-711764988cc9\") " pod="openstack/configure-os-openstack-openstack-networker-nwsds" Oct 14 09:26:02 crc kubenswrapper[4870]: I1014 09:26:02.672553 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h4c8k\" (UniqueName: \"kubernetes.io/projected/324ce306-ca93-447c-a19c-711764988cc9-kube-api-access-h4c8k\") pod \"configure-os-openstack-openstack-networker-nwsds\" (UID: \"324ce306-ca93-447c-a19c-711764988cc9\") " pod="openstack/configure-os-openstack-openstack-networker-nwsds" Oct 14 09:26:02 crc kubenswrapper[4870]: I1014 09:26:02.812137 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-openstack-openstack-networker-nwsds" Oct 14 09:26:03 crc kubenswrapper[4870]: I1014 09:26:03.412299 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-openstack-openstack-networker-nwsds"] Oct 14 09:26:04 crc kubenswrapper[4870]: I1014 09:26:04.445801 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-openstack-openstack-networker-nwsds" event={"ID":"324ce306-ca93-447c-a19c-711764988cc9","Type":"ContainerStarted","Data":"d49605b56b610bbb2f9cc84184f1a33ddfd2b5e03a3383c7f2768f1536af2a3a"} Oct 14 09:26:05 crc kubenswrapper[4870]: I1014 09:26:05.458553 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-openstack-openstack-networker-nwsds" event={"ID":"324ce306-ca93-447c-a19c-711764988cc9","Type":"ContainerStarted","Data":"2e954cbed599e0ecac3592fe40304f116e75d89a5fd72d38a3729306e4df75f5"} Oct 14 09:26:05 crc kubenswrapper[4870]: I1014 09:26:05.496142 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-openstack-openstack-networker-nwsds" podStartSLOduration=2.5636478609999998 podStartE2EDuration="3.49612065s" podCreationTimestamp="2025-10-14 09:26:02 +0000 UTC" firstStartedPulling="2025-10-14 09:26:03.422179882 +0000 UTC m=+8699.119540253" lastFinishedPulling="2025-10-14 09:26:04.354652661 +0000 UTC m=+8700.052013042" observedRunningTime="2025-10-14 09:26:05.494017338 +0000 UTC m=+8701.191377709" watchObservedRunningTime="2025-10-14 09:26:05.49612065 +0000 UTC m=+8701.193481031" Oct 14 09:26:19 crc kubenswrapper[4870]: I1014 09:26:19.620925 4870 generic.go:334] "Generic (PLEG): container finished" podID="faed3d6f-4637-44f8-8925-0e82407fcd44" containerID="b014edbf89143aa5b8cccbe618d85eccd8cc29d9db3aa71566e0d135dec82cbf" exitCode=0 Oct 14 09:26:19 crc kubenswrapper[4870]: I1014 09:26:19.621045 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-openstack-openstack-cell1-54frh" event={"ID":"faed3d6f-4637-44f8-8925-0e82407fcd44","Type":"ContainerDied","Data":"b014edbf89143aa5b8cccbe618d85eccd8cc29d9db3aa71566e0d135dec82cbf"} Oct 14 09:26:21 crc kubenswrapper[4870]: I1014 09:26:21.162616 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-openstack-openstack-cell1-54frh" Oct 14 09:26:21 crc kubenswrapper[4870]: I1014 09:26:21.241596 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qb74v\" (UniqueName: \"kubernetes.io/projected/faed3d6f-4637-44f8-8925-0e82407fcd44-kube-api-access-qb74v\") pod \"faed3d6f-4637-44f8-8925-0e82407fcd44\" (UID: \"faed3d6f-4637-44f8-8925-0e82407fcd44\") " Oct 14 09:26:21 crc kubenswrapper[4870]: I1014 09:26:21.241720 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/faed3d6f-4637-44f8-8925-0e82407fcd44-ssh-key\") pod \"faed3d6f-4637-44f8-8925-0e82407fcd44\" (UID: \"faed3d6f-4637-44f8-8925-0e82407fcd44\") " Oct 14 09:26:21 crc kubenswrapper[4870]: I1014 09:26:21.241888 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/faed3d6f-4637-44f8-8925-0e82407fcd44-ceph\") pod \"faed3d6f-4637-44f8-8925-0e82407fcd44\" (UID: \"faed3d6f-4637-44f8-8925-0e82407fcd44\") " Oct 14 09:26:21 crc kubenswrapper[4870]: I1014 09:26:21.241956 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/faed3d6f-4637-44f8-8925-0e82407fcd44-inventory\") pod \"faed3d6f-4637-44f8-8925-0e82407fcd44\" (UID: \"faed3d6f-4637-44f8-8925-0e82407fcd44\") " Oct 14 09:26:21 crc kubenswrapper[4870]: I1014 09:26:21.247795 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/faed3d6f-4637-44f8-8925-0e82407fcd44-ceph" (OuterVolumeSpecName: "ceph") pod "faed3d6f-4637-44f8-8925-0e82407fcd44" (UID: "faed3d6f-4637-44f8-8925-0e82407fcd44"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:26:21 crc kubenswrapper[4870]: I1014 09:26:21.249039 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/faed3d6f-4637-44f8-8925-0e82407fcd44-kube-api-access-qb74v" (OuterVolumeSpecName: "kube-api-access-qb74v") pod "faed3d6f-4637-44f8-8925-0e82407fcd44" (UID: "faed3d6f-4637-44f8-8925-0e82407fcd44"). InnerVolumeSpecName "kube-api-access-qb74v". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:26:21 crc kubenswrapper[4870]: I1014 09:26:21.271808 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/faed3d6f-4637-44f8-8925-0e82407fcd44-inventory" (OuterVolumeSpecName: "inventory") pod "faed3d6f-4637-44f8-8925-0e82407fcd44" (UID: "faed3d6f-4637-44f8-8925-0e82407fcd44"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:26:21 crc kubenswrapper[4870]: I1014 09:26:21.272561 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/faed3d6f-4637-44f8-8925-0e82407fcd44-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "faed3d6f-4637-44f8-8925-0e82407fcd44" (UID: "faed3d6f-4637-44f8-8925-0e82407fcd44"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:26:21 crc kubenswrapper[4870]: I1014 09:26:21.345966 4870 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/faed3d6f-4637-44f8-8925-0e82407fcd44-inventory\") on node \"crc\" DevicePath \"\"" Oct 14 09:26:21 crc kubenswrapper[4870]: I1014 09:26:21.346001 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qb74v\" (UniqueName: \"kubernetes.io/projected/faed3d6f-4637-44f8-8925-0e82407fcd44-kube-api-access-qb74v\") on node \"crc\" DevicePath \"\"" Oct 14 09:26:21 crc kubenswrapper[4870]: I1014 09:26:21.346011 4870 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/faed3d6f-4637-44f8-8925-0e82407fcd44-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 14 09:26:21 crc kubenswrapper[4870]: I1014 09:26:21.346019 4870 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/faed3d6f-4637-44f8-8925-0e82407fcd44-ceph\") on node \"crc\" DevicePath \"\"" Oct 14 09:26:21 crc kubenswrapper[4870]: I1014 09:26:21.645903 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-openstack-openstack-cell1-54frh" event={"ID":"faed3d6f-4637-44f8-8925-0e82407fcd44","Type":"ContainerDied","Data":"530b8b2bc995d83f9cd65e89dc3b23a1042fa5953596b8d882c9b1911832a584"} Oct 14 09:26:21 crc kubenswrapper[4870]: I1014 09:26:21.645941 4870 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="530b8b2bc995d83f9cd65e89dc3b23a1042fa5953596b8d882c9b1911832a584" Oct 14 09:26:21 crc kubenswrapper[4870]: I1014 09:26:21.645994 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-openstack-openstack-cell1-54frh" Oct 14 09:26:21 crc kubenswrapper[4870]: I1014 09:26:21.747109 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-openstack-openstack-cell1-p9242"] Oct 14 09:26:21 crc kubenswrapper[4870]: E1014 09:26:21.747743 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="faed3d6f-4637-44f8-8925-0e82407fcd44" containerName="install-os-openstack-openstack-cell1" Oct 14 09:26:21 crc kubenswrapper[4870]: I1014 09:26:21.747761 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="faed3d6f-4637-44f8-8925-0e82407fcd44" containerName="install-os-openstack-openstack-cell1" Oct 14 09:26:21 crc kubenswrapper[4870]: I1014 09:26:21.747995 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="faed3d6f-4637-44f8-8925-0e82407fcd44" containerName="install-os-openstack-openstack-cell1" Oct 14 09:26:21 crc kubenswrapper[4870]: I1014 09:26:21.751481 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-openstack-openstack-cell1-p9242" Oct 14 09:26:21 crc kubenswrapper[4870]: I1014 09:26:21.754934 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 14 09:26:21 crc kubenswrapper[4870]: I1014 09:26:21.755002 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-wrz7w" Oct 14 09:26:21 crc kubenswrapper[4870]: I1014 09:26:21.762270 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-openstack-openstack-cell1-p9242"] Oct 14 09:26:21 crc kubenswrapper[4870]: I1014 09:26:21.859732 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6815b76f-1435-4980-9b2f-9c36e02d6f63-ssh-key\") pod \"configure-os-openstack-openstack-cell1-p9242\" (UID: \"6815b76f-1435-4980-9b2f-9c36e02d6f63\") " pod="openstack/configure-os-openstack-openstack-cell1-p9242" Oct 14 09:26:21 crc kubenswrapper[4870]: I1014 09:26:21.859859 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6815b76f-1435-4980-9b2f-9c36e02d6f63-inventory\") pod \"configure-os-openstack-openstack-cell1-p9242\" (UID: \"6815b76f-1435-4980-9b2f-9c36e02d6f63\") " pod="openstack/configure-os-openstack-openstack-cell1-p9242" Oct 14 09:26:21 crc kubenswrapper[4870]: I1014 09:26:21.860178 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/6815b76f-1435-4980-9b2f-9c36e02d6f63-ceph\") pod \"configure-os-openstack-openstack-cell1-p9242\" (UID: \"6815b76f-1435-4980-9b2f-9c36e02d6f63\") " pod="openstack/configure-os-openstack-openstack-cell1-p9242" Oct 14 09:26:21 crc kubenswrapper[4870]: I1014 09:26:21.860242 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gkd79\" (UniqueName: \"kubernetes.io/projected/6815b76f-1435-4980-9b2f-9c36e02d6f63-kube-api-access-gkd79\") pod \"configure-os-openstack-openstack-cell1-p9242\" (UID: \"6815b76f-1435-4980-9b2f-9c36e02d6f63\") " pod="openstack/configure-os-openstack-openstack-cell1-p9242" Oct 14 09:26:21 crc kubenswrapper[4870]: I1014 09:26:21.964023 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/6815b76f-1435-4980-9b2f-9c36e02d6f63-ceph\") pod \"configure-os-openstack-openstack-cell1-p9242\" (UID: \"6815b76f-1435-4980-9b2f-9c36e02d6f63\") " pod="openstack/configure-os-openstack-openstack-cell1-p9242" Oct 14 09:26:21 crc kubenswrapper[4870]: I1014 09:26:21.964090 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gkd79\" (UniqueName: \"kubernetes.io/projected/6815b76f-1435-4980-9b2f-9c36e02d6f63-kube-api-access-gkd79\") pod \"configure-os-openstack-openstack-cell1-p9242\" (UID: \"6815b76f-1435-4980-9b2f-9c36e02d6f63\") " pod="openstack/configure-os-openstack-openstack-cell1-p9242" Oct 14 09:26:21 crc kubenswrapper[4870]: I1014 09:26:21.964269 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6815b76f-1435-4980-9b2f-9c36e02d6f63-ssh-key\") pod \"configure-os-openstack-openstack-cell1-p9242\" (UID: \"6815b76f-1435-4980-9b2f-9c36e02d6f63\") " pod="openstack/configure-os-openstack-openstack-cell1-p9242" Oct 14 09:26:21 crc kubenswrapper[4870]: I1014 09:26:21.964424 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6815b76f-1435-4980-9b2f-9c36e02d6f63-inventory\") pod \"configure-os-openstack-openstack-cell1-p9242\" (UID: \"6815b76f-1435-4980-9b2f-9c36e02d6f63\") " pod="openstack/configure-os-openstack-openstack-cell1-p9242" Oct 14 09:26:22 crc kubenswrapper[4870]: I1014 09:26:22.007851 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gkd79\" (UniqueName: \"kubernetes.io/projected/6815b76f-1435-4980-9b2f-9c36e02d6f63-kube-api-access-gkd79\") pod \"configure-os-openstack-openstack-cell1-p9242\" (UID: \"6815b76f-1435-4980-9b2f-9c36e02d6f63\") " pod="openstack/configure-os-openstack-openstack-cell1-p9242" Oct 14 09:26:22 crc kubenswrapper[4870]: I1014 09:26:22.026031 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6815b76f-1435-4980-9b2f-9c36e02d6f63-ssh-key\") pod \"configure-os-openstack-openstack-cell1-p9242\" (UID: \"6815b76f-1435-4980-9b2f-9c36e02d6f63\") " pod="openstack/configure-os-openstack-openstack-cell1-p9242" Oct 14 09:26:22 crc kubenswrapper[4870]: I1014 09:26:22.040130 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/6815b76f-1435-4980-9b2f-9c36e02d6f63-ceph\") pod \"configure-os-openstack-openstack-cell1-p9242\" (UID: \"6815b76f-1435-4980-9b2f-9c36e02d6f63\") " pod="openstack/configure-os-openstack-openstack-cell1-p9242" Oct 14 09:26:22 crc kubenswrapper[4870]: I1014 09:26:22.056983 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6815b76f-1435-4980-9b2f-9c36e02d6f63-inventory\") pod \"configure-os-openstack-openstack-cell1-p9242\" (UID: \"6815b76f-1435-4980-9b2f-9c36e02d6f63\") " pod="openstack/configure-os-openstack-openstack-cell1-p9242" Oct 14 09:26:22 crc kubenswrapper[4870]: I1014 09:26:22.079019 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-openstack-openstack-cell1-p9242" Oct 14 09:26:22 crc kubenswrapper[4870]: I1014 09:26:22.654289 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-openstack-openstack-cell1-p9242"] Oct 14 09:26:22 crc kubenswrapper[4870]: W1014 09:26:22.662653 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6815b76f_1435_4980_9b2f_9c36e02d6f63.slice/crio-2901991058b49fca004ce6999d69308a310503df61b3ae09a654d7d450e8d23b WatchSource:0}: Error finding container 2901991058b49fca004ce6999d69308a310503df61b3ae09a654d7d450e8d23b: Status 404 returned error can't find the container with id 2901991058b49fca004ce6999d69308a310503df61b3ae09a654d7d450e8d23b Oct 14 09:26:23 crc kubenswrapper[4870]: I1014 09:26:23.666645 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-openstack-openstack-cell1-p9242" event={"ID":"6815b76f-1435-4980-9b2f-9c36e02d6f63","Type":"ContainerStarted","Data":"ad6b483a02773539d1db7c37872626347dd76fdf0150922c8af78056f93d2e4b"} Oct 14 09:26:23 crc kubenswrapper[4870]: I1014 09:26:23.666928 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-openstack-openstack-cell1-p9242" event={"ID":"6815b76f-1435-4980-9b2f-9c36e02d6f63","Type":"ContainerStarted","Data":"2901991058b49fca004ce6999d69308a310503df61b3ae09a654d7d450e8d23b"} Oct 14 09:26:23 crc kubenswrapper[4870]: I1014 09:26:23.705117 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-openstack-openstack-cell1-p9242" podStartSLOduration=2.20813531 podStartE2EDuration="2.705087585s" podCreationTimestamp="2025-10-14 09:26:21 +0000 UTC" firstStartedPulling="2025-10-14 09:26:22.665522273 +0000 UTC m=+8718.362882644" lastFinishedPulling="2025-10-14 09:26:23.162474538 +0000 UTC m=+8718.859834919" observedRunningTime="2025-10-14 09:26:23.685632117 +0000 UTC m=+8719.382992508" watchObservedRunningTime="2025-10-14 09:26:23.705087585 +0000 UTC m=+8719.402447986" Oct 14 09:26:56 crc kubenswrapper[4870]: I1014 09:26:56.073164 4870 generic.go:334] "Generic (PLEG): container finished" podID="324ce306-ca93-447c-a19c-711764988cc9" containerID="2e954cbed599e0ecac3592fe40304f116e75d89a5fd72d38a3729306e4df75f5" exitCode=2 Oct 14 09:26:56 crc kubenswrapper[4870]: I1014 09:26:56.073288 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-openstack-openstack-networker-nwsds" event={"ID":"324ce306-ca93-447c-a19c-711764988cc9","Type":"ContainerDied","Data":"2e954cbed599e0ecac3592fe40304f116e75d89a5fd72d38a3729306e4df75f5"} Oct 14 09:26:57 crc kubenswrapper[4870]: I1014 09:26:57.623550 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-openstack-openstack-networker-nwsds" Oct 14 09:26:57 crc kubenswrapper[4870]: I1014 09:26:57.705561 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h4c8k\" (UniqueName: \"kubernetes.io/projected/324ce306-ca93-447c-a19c-711764988cc9-kube-api-access-h4c8k\") pod \"324ce306-ca93-447c-a19c-711764988cc9\" (UID: \"324ce306-ca93-447c-a19c-711764988cc9\") " Oct 14 09:26:57 crc kubenswrapper[4870]: I1014 09:26:57.705615 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/324ce306-ca93-447c-a19c-711764988cc9-inventory\") pod \"324ce306-ca93-447c-a19c-711764988cc9\" (UID: \"324ce306-ca93-447c-a19c-711764988cc9\") " Oct 14 09:26:57 crc kubenswrapper[4870]: I1014 09:26:57.705645 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/324ce306-ca93-447c-a19c-711764988cc9-ssh-key\") pod \"324ce306-ca93-447c-a19c-711764988cc9\" (UID: \"324ce306-ca93-447c-a19c-711764988cc9\") " Oct 14 09:26:57 crc kubenswrapper[4870]: I1014 09:26:57.728388 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/324ce306-ca93-447c-a19c-711764988cc9-kube-api-access-h4c8k" (OuterVolumeSpecName: "kube-api-access-h4c8k") pod "324ce306-ca93-447c-a19c-711764988cc9" (UID: "324ce306-ca93-447c-a19c-711764988cc9"). InnerVolumeSpecName "kube-api-access-h4c8k". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:26:57 crc kubenswrapper[4870]: I1014 09:26:57.740001 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/324ce306-ca93-447c-a19c-711764988cc9-inventory" (OuterVolumeSpecName: "inventory") pod "324ce306-ca93-447c-a19c-711764988cc9" (UID: "324ce306-ca93-447c-a19c-711764988cc9"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:26:57 crc kubenswrapper[4870]: I1014 09:26:57.740349 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/324ce306-ca93-447c-a19c-711764988cc9-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "324ce306-ca93-447c-a19c-711764988cc9" (UID: "324ce306-ca93-447c-a19c-711764988cc9"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:26:57 crc kubenswrapper[4870]: I1014 09:26:57.807996 4870 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/324ce306-ca93-447c-a19c-711764988cc9-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 14 09:26:57 crc kubenswrapper[4870]: I1014 09:26:57.808035 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h4c8k\" (UniqueName: \"kubernetes.io/projected/324ce306-ca93-447c-a19c-711764988cc9-kube-api-access-h4c8k\") on node \"crc\" DevicePath \"\"" Oct 14 09:26:57 crc kubenswrapper[4870]: I1014 09:26:57.808044 4870 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/324ce306-ca93-447c-a19c-711764988cc9-inventory\") on node \"crc\" DevicePath \"\"" Oct 14 09:26:58 crc kubenswrapper[4870]: I1014 09:26:58.095142 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-openstack-openstack-networker-nwsds" event={"ID":"324ce306-ca93-447c-a19c-711764988cc9","Type":"ContainerDied","Data":"d49605b56b610bbb2f9cc84184f1a33ddfd2b5e03a3383c7f2768f1536af2a3a"} Oct 14 09:26:58 crc kubenswrapper[4870]: I1014 09:26:58.095397 4870 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d49605b56b610bbb2f9cc84184f1a33ddfd2b5e03a3383c7f2768f1536af2a3a" Oct 14 09:26:58 crc kubenswrapper[4870]: I1014 09:26:58.095181 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-openstack-openstack-networker-nwsds" Oct 14 09:27:05 crc kubenswrapper[4870]: I1014 09:27:05.068147 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-openstack-openstack-networker-gbdf9"] Oct 14 09:27:05 crc kubenswrapper[4870]: E1014 09:27:05.069300 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="324ce306-ca93-447c-a19c-711764988cc9" containerName="configure-os-openstack-openstack-networker" Oct 14 09:27:05 crc kubenswrapper[4870]: I1014 09:27:05.069320 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="324ce306-ca93-447c-a19c-711764988cc9" containerName="configure-os-openstack-openstack-networker" Oct 14 09:27:05 crc kubenswrapper[4870]: I1014 09:27:05.069692 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="324ce306-ca93-447c-a19c-711764988cc9" containerName="configure-os-openstack-openstack-networker" Oct 14 09:27:05 crc kubenswrapper[4870]: I1014 09:27:05.070809 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-openstack-openstack-networker-gbdf9" Oct 14 09:27:05 crc kubenswrapper[4870]: I1014 09:27:05.075065 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-networker-dockercfg-54hzr" Oct 14 09:27:05 crc kubenswrapper[4870]: I1014 09:27:05.075870 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-networker" Oct 14 09:27:05 crc kubenswrapper[4870]: I1014 09:27:05.082705 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-openstack-openstack-networker-gbdf9"] Oct 14 09:27:05 crc kubenswrapper[4870]: I1014 09:27:05.174732 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0d62b70c-9f3a-4029-92ff-05143f46254e-inventory\") pod \"configure-os-openstack-openstack-networker-gbdf9\" (UID: \"0d62b70c-9f3a-4029-92ff-05143f46254e\") " pod="openstack/configure-os-openstack-openstack-networker-gbdf9" Oct 14 09:27:05 crc kubenswrapper[4870]: I1014 09:27:05.174994 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nr7c4\" (UniqueName: \"kubernetes.io/projected/0d62b70c-9f3a-4029-92ff-05143f46254e-kube-api-access-nr7c4\") pod \"configure-os-openstack-openstack-networker-gbdf9\" (UID: \"0d62b70c-9f3a-4029-92ff-05143f46254e\") " pod="openstack/configure-os-openstack-openstack-networker-gbdf9" Oct 14 09:27:05 crc kubenswrapper[4870]: I1014 09:27:05.175166 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0d62b70c-9f3a-4029-92ff-05143f46254e-ssh-key\") pod \"configure-os-openstack-openstack-networker-gbdf9\" (UID: \"0d62b70c-9f3a-4029-92ff-05143f46254e\") " pod="openstack/configure-os-openstack-openstack-networker-gbdf9" Oct 14 09:27:05 crc kubenswrapper[4870]: I1014 09:27:05.276866 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0d62b70c-9f3a-4029-92ff-05143f46254e-inventory\") pod \"configure-os-openstack-openstack-networker-gbdf9\" (UID: \"0d62b70c-9f3a-4029-92ff-05143f46254e\") " pod="openstack/configure-os-openstack-openstack-networker-gbdf9" Oct 14 09:27:05 crc kubenswrapper[4870]: I1014 09:27:05.277079 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nr7c4\" (UniqueName: \"kubernetes.io/projected/0d62b70c-9f3a-4029-92ff-05143f46254e-kube-api-access-nr7c4\") pod \"configure-os-openstack-openstack-networker-gbdf9\" (UID: \"0d62b70c-9f3a-4029-92ff-05143f46254e\") " pod="openstack/configure-os-openstack-openstack-networker-gbdf9" Oct 14 09:27:05 crc kubenswrapper[4870]: I1014 09:27:05.277209 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0d62b70c-9f3a-4029-92ff-05143f46254e-ssh-key\") pod \"configure-os-openstack-openstack-networker-gbdf9\" (UID: \"0d62b70c-9f3a-4029-92ff-05143f46254e\") " pod="openstack/configure-os-openstack-openstack-networker-gbdf9" Oct 14 09:27:05 crc kubenswrapper[4870]: I1014 09:27:05.285723 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0d62b70c-9f3a-4029-92ff-05143f46254e-inventory\") pod \"configure-os-openstack-openstack-networker-gbdf9\" (UID: \"0d62b70c-9f3a-4029-92ff-05143f46254e\") " pod="openstack/configure-os-openstack-openstack-networker-gbdf9" Oct 14 09:27:05 crc kubenswrapper[4870]: I1014 09:27:05.286014 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0d62b70c-9f3a-4029-92ff-05143f46254e-ssh-key\") pod \"configure-os-openstack-openstack-networker-gbdf9\" (UID: \"0d62b70c-9f3a-4029-92ff-05143f46254e\") " pod="openstack/configure-os-openstack-openstack-networker-gbdf9" Oct 14 09:27:05 crc kubenswrapper[4870]: I1014 09:27:05.299338 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nr7c4\" (UniqueName: \"kubernetes.io/projected/0d62b70c-9f3a-4029-92ff-05143f46254e-kube-api-access-nr7c4\") pod \"configure-os-openstack-openstack-networker-gbdf9\" (UID: \"0d62b70c-9f3a-4029-92ff-05143f46254e\") " pod="openstack/configure-os-openstack-openstack-networker-gbdf9" Oct 14 09:27:05 crc kubenswrapper[4870]: I1014 09:27:05.414757 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-openstack-openstack-networker-gbdf9" Oct 14 09:27:05 crc kubenswrapper[4870]: W1014 09:27:05.967601 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0d62b70c_9f3a_4029_92ff_05143f46254e.slice/crio-9c4f071e834ea009725e08074713ed465ccc379acc8d6d9f47f86402fc504b5f WatchSource:0}: Error finding container 9c4f071e834ea009725e08074713ed465ccc379acc8d6d9f47f86402fc504b5f: Status 404 returned error can't find the container with id 9c4f071e834ea009725e08074713ed465ccc379acc8d6d9f47f86402fc504b5f Oct 14 09:27:05 crc kubenswrapper[4870]: I1014 09:27:05.977128 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-openstack-openstack-networker-gbdf9"] Oct 14 09:27:06 crc kubenswrapper[4870]: I1014 09:27:06.199659 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-openstack-openstack-networker-gbdf9" event={"ID":"0d62b70c-9f3a-4029-92ff-05143f46254e","Type":"ContainerStarted","Data":"9c4f071e834ea009725e08074713ed465ccc379acc8d6d9f47f86402fc504b5f"} Oct 14 09:27:07 crc kubenswrapper[4870]: I1014 09:27:07.211786 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-openstack-openstack-networker-gbdf9" event={"ID":"0d62b70c-9f3a-4029-92ff-05143f46254e","Type":"ContainerStarted","Data":"d1ed073e24a95f53b23b27d31555df8cf06a5ad7d80b85bdc2725b5e80106170"} Oct 14 09:27:07 crc kubenswrapper[4870]: I1014 09:27:07.246000 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-openstack-openstack-networker-gbdf9" podStartSLOduration=1.7469358050000001 podStartE2EDuration="2.245977921s" podCreationTimestamp="2025-10-14 09:27:05 +0000 UTC" firstStartedPulling="2025-10-14 09:27:05.970258869 +0000 UTC m=+8761.667619250" lastFinishedPulling="2025-10-14 09:27:06.469300985 +0000 UTC m=+8762.166661366" observedRunningTime="2025-10-14 09:27:07.23048106 +0000 UTC m=+8762.927841431" watchObservedRunningTime="2025-10-14 09:27:07.245977921 +0000 UTC m=+8762.943338292" Oct 14 09:27:15 crc kubenswrapper[4870]: I1014 09:27:15.313235 4870 generic.go:334] "Generic (PLEG): container finished" podID="6815b76f-1435-4980-9b2f-9c36e02d6f63" containerID="ad6b483a02773539d1db7c37872626347dd76fdf0150922c8af78056f93d2e4b" exitCode=2 Oct 14 09:27:15 crc kubenswrapper[4870]: I1014 09:27:15.313430 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-openstack-openstack-cell1-p9242" event={"ID":"6815b76f-1435-4980-9b2f-9c36e02d6f63","Type":"ContainerDied","Data":"ad6b483a02773539d1db7c37872626347dd76fdf0150922c8af78056f93d2e4b"} Oct 14 09:27:16 crc kubenswrapper[4870]: I1014 09:27:16.761110 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-openstack-openstack-cell1-p9242" Oct 14 09:27:16 crc kubenswrapper[4870]: I1014 09:27:16.855045 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6815b76f-1435-4980-9b2f-9c36e02d6f63-inventory\") pod \"6815b76f-1435-4980-9b2f-9c36e02d6f63\" (UID: \"6815b76f-1435-4980-9b2f-9c36e02d6f63\") " Oct 14 09:27:16 crc kubenswrapper[4870]: I1014 09:27:16.855117 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/6815b76f-1435-4980-9b2f-9c36e02d6f63-ceph\") pod \"6815b76f-1435-4980-9b2f-9c36e02d6f63\" (UID: \"6815b76f-1435-4980-9b2f-9c36e02d6f63\") " Oct 14 09:27:16 crc kubenswrapper[4870]: I1014 09:27:16.855206 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gkd79\" (UniqueName: \"kubernetes.io/projected/6815b76f-1435-4980-9b2f-9c36e02d6f63-kube-api-access-gkd79\") pod \"6815b76f-1435-4980-9b2f-9c36e02d6f63\" (UID: \"6815b76f-1435-4980-9b2f-9c36e02d6f63\") " Oct 14 09:27:16 crc kubenswrapper[4870]: I1014 09:27:16.855337 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6815b76f-1435-4980-9b2f-9c36e02d6f63-ssh-key\") pod \"6815b76f-1435-4980-9b2f-9c36e02d6f63\" (UID: \"6815b76f-1435-4980-9b2f-9c36e02d6f63\") " Oct 14 09:27:16 crc kubenswrapper[4870]: I1014 09:27:16.861094 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6815b76f-1435-4980-9b2f-9c36e02d6f63-ceph" (OuterVolumeSpecName: "ceph") pod "6815b76f-1435-4980-9b2f-9c36e02d6f63" (UID: "6815b76f-1435-4980-9b2f-9c36e02d6f63"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:27:16 crc kubenswrapper[4870]: I1014 09:27:16.864073 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6815b76f-1435-4980-9b2f-9c36e02d6f63-kube-api-access-gkd79" (OuterVolumeSpecName: "kube-api-access-gkd79") pod "6815b76f-1435-4980-9b2f-9c36e02d6f63" (UID: "6815b76f-1435-4980-9b2f-9c36e02d6f63"). InnerVolumeSpecName "kube-api-access-gkd79". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:27:16 crc kubenswrapper[4870]: I1014 09:27:16.886209 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6815b76f-1435-4980-9b2f-9c36e02d6f63-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "6815b76f-1435-4980-9b2f-9c36e02d6f63" (UID: "6815b76f-1435-4980-9b2f-9c36e02d6f63"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:27:16 crc kubenswrapper[4870]: I1014 09:27:16.900436 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6815b76f-1435-4980-9b2f-9c36e02d6f63-inventory" (OuterVolumeSpecName: "inventory") pod "6815b76f-1435-4980-9b2f-9c36e02d6f63" (UID: "6815b76f-1435-4980-9b2f-9c36e02d6f63"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:27:16 crc kubenswrapper[4870]: I1014 09:27:16.956956 4870 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6815b76f-1435-4980-9b2f-9c36e02d6f63-inventory\") on node \"crc\" DevicePath \"\"" Oct 14 09:27:16 crc kubenswrapper[4870]: I1014 09:27:16.956981 4870 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/6815b76f-1435-4980-9b2f-9c36e02d6f63-ceph\") on node \"crc\" DevicePath \"\"" Oct 14 09:27:16 crc kubenswrapper[4870]: I1014 09:27:16.956991 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gkd79\" (UniqueName: \"kubernetes.io/projected/6815b76f-1435-4980-9b2f-9c36e02d6f63-kube-api-access-gkd79\") on node \"crc\" DevicePath \"\"" Oct 14 09:27:16 crc kubenswrapper[4870]: I1014 09:27:16.956999 4870 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6815b76f-1435-4980-9b2f-9c36e02d6f63-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 14 09:27:17 crc kubenswrapper[4870]: I1014 09:27:17.333554 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-openstack-openstack-cell1-p9242" event={"ID":"6815b76f-1435-4980-9b2f-9c36e02d6f63","Type":"ContainerDied","Data":"2901991058b49fca004ce6999d69308a310503df61b3ae09a654d7d450e8d23b"} Oct 14 09:27:17 crc kubenswrapper[4870]: I1014 09:27:17.333599 4870 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2901991058b49fca004ce6999d69308a310503df61b3ae09a654d7d450e8d23b" Oct 14 09:27:17 crc kubenswrapper[4870]: I1014 09:27:17.333655 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-openstack-openstack-cell1-p9242" Oct 14 09:27:24 crc kubenswrapper[4870]: I1014 09:27:24.046176 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-openstack-openstack-cell1-7928w"] Oct 14 09:27:24 crc kubenswrapper[4870]: E1014 09:27:24.050337 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6815b76f-1435-4980-9b2f-9c36e02d6f63" containerName="configure-os-openstack-openstack-cell1" Oct 14 09:27:24 crc kubenswrapper[4870]: I1014 09:27:24.050365 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="6815b76f-1435-4980-9b2f-9c36e02d6f63" containerName="configure-os-openstack-openstack-cell1" Oct 14 09:27:24 crc kubenswrapper[4870]: I1014 09:27:24.050732 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="6815b76f-1435-4980-9b2f-9c36e02d6f63" containerName="configure-os-openstack-openstack-cell1" Oct 14 09:27:24 crc kubenswrapper[4870]: I1014 09:27:24.051933 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-openstack-openstack-cell1-7928w" Oct 14 09:27:24 crc kubenswrapper[4870]: I1014 09:27:24.056006 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 14 09:27:24 crc kubenswrapper[4870]: I1014 09:27:24.056075 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-wrz7w" Oct 14 09:27:24 crc kubenswrapper[4870]: I1014 09:27:24.118216 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-openstack-openstack-cell1-7928w"] Oct 14 09:27:24 crc kubenswrapper[4870]: I1014 09:27:24.147265 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c2rgp\" (UniqueName: \"kubernetes.io/projected/e3f48d1d-384a-4b16-964a-5834de86fdb3-kube-api-access-c2rgp\") pod \"configure-os-openstack-openstack-cell1-7928w\" (UID: \"e3f48d1d-384a-4b16-964a-5834de86fdb3\") " pod="openstack/configure-os-openstack-openstack-cell1-7928w" Oct 14 09:27:24 crc kubenswrapper[4870]: I1014 09:27:24.147367 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e3f48d1d-384a-4b16-964a-5834de86fdb3-ceph\") pod \"configure-os-openstack-openstack-cell1-7928w\" (UID: \"e3f48d1d-384a-4b16-964a-5834de86fdb3\") " pod="openstack/configure-os-openstack-openstack-cell1-7928w" Oct 14 09:27:24 crc kubenswrapper[4870]: I1014 09:27:24.147475 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e3f48d1d-384a-4b16-964a-5834de86fdb3-inventory\") pod \"configure-os-openstack-openstack-cell1-7928w\" (UID: \"e3f48d1d-384a-4b16-964a-5834de86fdb3\") " pod="openstack/configure-os-openstack-openstack-cell1-7928w" Oct 14 09:27:24 crc kubenswrapper[4870]: I1014 09:27:24.147769 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e3f48d1d-384a-4b16-964a-5834de86fdb3-ssh-key\") pod \"configure-os-openstack-openstack-cell1-7928w\" (UID: \"e3f48d1d-384a-4b16-964a-5834de86fdb3\") " pod="openstack/configure-os-openstack-openstack-cell1-7928w" Oct 14 09:27:24 crc kubenswrapper[4870]: I1014 09:27:24.249520 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e3f48d1d-384a-4b16-964a-5834de86fdb3-inventory\") pod \"configure-os-openstack-openstack-cell1-7928w\" (UID: \"e3f48d1d-384a-4b16-964a-5834de86fdb3\") " pod="openstack/configure-os-openstack-openstack-cell1-7928w" Oct 14 09:27:24 crc kubenswrapper[4870]: I1014 09:27:24.250540 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e3f48d1d-384a-4b16-964a-5834de86fdb3-ssh-key\") pod \"configure-os-openstack-openstack-cell1-7928w\" (UID: \"e3f48d1d-384a-4b16-964a-5834de86fdb3\") " pod="openstack/configure-os-openstack-openstack-cell1-7928w" Oct 14 09:27:24 crc kubenswrapper[4870]: I1014 09:27:24.250940 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c2rgp\" (UniqueName: \"kubernetes.io/projected/e3f48d1d-384a-4b16-964a-5834de86fdb3-kube-api-access-c2rgp\") pod \"configure-os-openstack-openstack-cell1-7928w\" (UID: \"e3f48d1d-384a-4b16-964a-5834de86fdb3\") " pod="openstack/configure-os-openstack-openstack-cell1-7928w" Oct 14 09:27:24 crc kubenswrapper[4870]: I1014 09:27:24.250994 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e3f48d1d-384a-4b16-964a-5834de86fdb3-ceph\") pod \"configure-os-openstack-openstack-cell1-7928w\" (UID: \"e3f48d1d-384a-4b16-964a-5834de86fdb3\") " pod="openstack/configure-os-openstack-openstack-cell1-7928w" Oct 14 09:27:24 crc kubenswrapper[4870]: I1014 09:27:24.265991 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e3f48d1d-384a-4b16-964a-5834de86fdb3-inventory\") pod \"configure-os-openstack-openstack-cell1-7928w\" (UID: \"e3f48d1d-384a-4b16-964a-5834de86fdb3\") " pod="openstack/configure-os-openstack-openstack-cell1-7928w" Oct 14 09:27:24 crc kubenswrapper[4870]: I1014 09:27:24.274978 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e3f48d1d-384a-4b16-964a-5834de86fdb3-ssh-key\") pod \"configure-os-openstack-openstack-cell1-7928w\" (UID: \"e3f48d1d-384a-4b16-964a-5834de86fdb3\") " pod="openstack/configure-os-openstack-openstack-cell1-7928w" Oct 14 09:27:24 crc kubenswrapper[4870]: I1014 09:27:24.276151 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e3f48d1d-384a-4b16-964a-5834de86fdb3-ceph\") pod \"configure-os-openstack-openstack-cell1-7928w\" (UID: \"e3f48d1d-384a-4b16-964a-5834de86fdb3\") " pod="openstack/configure-os-openstack-openstack-cell1-7928w" Oct 14 09:27:24 crc kubenswrapper[4870]: I1014 09:27:24.287819 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c2rgp\" (UniqueName: \"kubernetes.io/projected/e3f48d1d-384a-4b16-964a-5834de86fdb3-kube-api-access-c2rgp\") pod \"configure-os-openstack-openstack-cell1-7928w\" (UID: \"e3f48d1d-384a-4b16-964a-5834de86fdb3\") " pod="openstack/configure-os-openstack-openstack-cell1-7928w" Oct 14 09:27:24 crc kubenswrapper[4870]: I1014 09:27:24.390878 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-openstack-openstack-cell1-7928w" Oct 14 09:27:24 crc kubenswrapper[4870]: I1014 09:27:24.909322 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-openstack-openstack-cell1-7928w"] Oct 14 09:27:25 crc kubenswrapper[4870]: I1014 09:27:25.438013 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-openstack-openstack-cell1-7928w" event={"ID":"e3f48d1d-384a-4b16-964a-5834de86fdb3","Type":"ContainerStarted","Data":"6009edbf25dfcd9c651f667f2ad5fa7e736919ccd3b7e88809df796f97ba7937"} Oct 14 09:27:26 crc kubenswrapper[4870]: I1014 09:27:26.449372 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-openstack-openstack-cell1-7928w" event={"ID":"e3f48d1d-384a-4b16-964a-5834de86fdb3","Type":"ContainerStarted","Data":"c0f81efd93cbeb30917b532fced7c2095edcaea800e6f40418be417e0b7ef7e2"} Oct 14 09:27:26 crc kubenswrapper[4870]: I1014 09:27:26.473917 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-openstack-openstack-cell1-7928w" podStartSLOduration=1.901339089 podStartE2EDuration="2.473898014s" podCreationTimestamp="2025-10-14 09:27:24 +0000 UTC" firstStartedPulling="2025-10-14 09:27:24.911278773 +0000 UTC m=+8780.608639144" lastFinishedPulling="2025-10-14 09:27:25.483837688 +0000 UTC m=+8781.181198069" observedRunningTime="2025-10-14 09:27:26.472362976 +0000 UTC m=+8782.169723347" watchObservedRunningTime="2025-10-14 09:27:26.473898014 +0000 UTC m=+8782.171258385" Oct 14 09:27:56 crc kubenswrapper[4870]: I1014 09:27:56.769345 4870 generic.go:334] "Generic (PLEG): container finished" podID="0d62b70c-9f3a-4029-92ff-05143f46254e" containerID="d1ed073e24a95f53b23b27d31555df8cf06a5ad7d80b85bdc2725b5e80106170" exitCode=0 Oct 14 09:27:56 crc kubenswrapper[4870]: I1014 09:27:56.770118 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-openstack-openstack-networker-gbdf9" event={"ID":"0d62b70c-9f3a-4029-92ff-05143f46254e","Type":"ContainerDied","Data":"d1ed073e24a95f53b23b27d31555df8cf06a5ad7d80b85bdc2725b5e80106170"} Oct 14 09:27:58 crc kubenswrapper[4870]: I1014 09:27:58.341698 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-openstack-openstack-networker-gbdf9" Oct 14 09:27:58 crc kubenswrapper[4870]: I1014 09:27:58.497129 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nr7c4\" (UniqueName: \"kubernetes.io/projected/0d62b70c-9f3a-4029-92ff-05143f46254e-kube-api-access-nr7c4\") pod \"0d62b70c-9f3a-4029-92ff-05143f46254e\" (UID: \"0d62b70c-9f3a-4029-92ff-05143f46254e\") " Oct 14 09:27:58 crc kubenswrapper[4870]: I1014 09:27:58.497637 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0d62b70c-9f3a-4029-92ff-05143f46254e-ssh-key\") pod \"0d62b70c-9f3a-4029-92ff-05143f46254e\" (UID: \"0d62b70c-9f3a-4029-92ff-05143f46254e\") " Oct 14 09:27:58 crc kubenswrapper[4870]: I1014 09:27:58.497698 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0d62b70c-9f3a-4029-92ff-05143f46254e-inventory\") pod \"0d62b70c-9f3a-4029-92ff-05143f46254e\" (UID: \"0d62b70c-9f3a-4029-92ff-05143f46254e\") " Oct 14 09:27:58 crc kubenswrapper[4870]: I1014 09:27:58.505525 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0d62b70c-9f3a-4029-92ff-05143f46254e-kube-api-access-nr7c4" (OuterVolumeSpecName: "kube-api-access-nr7c4") pod "0d62b70c-9f3a-4029-92ff-05143f46254e" (UID: "0d62b70c-9f3a-4029-92ff-05143f46254e"). InnerVolumeSpecName "kube-api-access-nr7c4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:27:58 crc kubenswrapper[4870]: I1014 09:27:58.532056 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0d62b70c-9f3a-4029-92ff-05143f46254e-inventory" (OuterVolumeSpecName: "inventory") pod "0d62b70c-9f3a-4029-92ff-05143f46254e" (UID: "0d62b70c-9f3a-4029-92ff-05143f46254e"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:27:58 crc kubenswrapper[4870]: I1014 09:27:58.533708 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0d62b70c-9f3a-4029-92ff-05143f46254e-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "0d62b70c-9f3a-4029-92ff-05143f46254e" (UID: "0d62b70c-9f3a-4029-92ff-05143f46254e"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:27:58 crc kubenswrapper[4870]: I1014 09:27:58.600027 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nr7c4\" (UniqueName: \"kubernetes.io/projected/0d62b70c-9f3a-4029-92ff-05143f46254e-kube-api-access-nr7c4\") on node \"crc\" DevicePath \"\"" Oct 14 09:27:58 crc kubenswrapper[4870]: I1014 09:27:58.600069 4870 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0d62b70c-9f3a-4029-92ff-05143f46254e-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 14 09:27:58 crc kubenswrapper[4870]: I1014 09:27:58.600081 4870 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0d62b70c-9f3a-4029-92ff-05143f46254e-inventory\") on node \"crc\" DevicePath \"\"" Oct 14 09:27:58 crc kubenswrapper[4870]: I1014 09:27:58.792164 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-openstack-openstack-networker-gbdf9" event={"ID":"0d62b70c-9f3a-4029-92ff-05143f46254e","Type":"ContainerDied","Data":"9c4f071e834ea009725e08074713ed465ccc379acc8d6d9f47f86402fc504b5f"} Oct 14 09:27:58 crc kubenswrapper[4870]: I1014 09:27:58.792208 4870 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9c4f071e834ea009725e08074713ed465ccc379acc8d6d9f47f86402fc504b5f" Oct 14 09:27:58 crc kubenswrapper[4870]: I1014 09:27:58.792276 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-openstack-openstack-networker-gbdf9" Oct 14 09:27:58 crc kubenswrapper[4870]: I1014 09:27:58.918317 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-openstack-openstack-networker-89k68"] Oct 14 09:27:58 crc kubenswrapper[4870]: E1014 09:27:58.919694 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d62b70c-9f3a-4029-92ff-05143f46254e" containerName="configure-os-openstack-openstack-networker" Oct 14 09:27:58 crc kubenswrapper[4870]: I1014 09:27:58.919734 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d62b70c-9f3a-4029-92ff-05143f46254e" containerName="configure-os-openstack-openstack-networker" Oct 14 09:27:58 crc kubenswrapper[4870]: I1014 09:27:58.920077 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="0d62b70c-9f3a-4029-92ff-05143f46254e" containerName="configure-os-openstack-openstack-networker" Oct 14 09:27:58 crc kubenswrapper[4870]: I1014 09:27:58.921313 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-openstack-openstack-networker-89k68" Oct 14 09:27:58 crc kubenswrapper[4870]: I1014 09:27:58.924609 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-networker-dockercfg-54hzr" Oct 14 09:27:58 crc kubenswrapper[4870]: I1014 09:27:58.924929 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-networker" Oct 14 09:27:58 crc kubenswrapper[4870]: I1014 09:27:58.939048 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-openstack-openstack-networker-89k68"] Oct 14 09:27:59 crc kubenswrapper[4870]: I1014 09:27:59.008131 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b3e9780b-d240-4d71-bb80-3e21346f7eb9-inventory\") pod \"run-os-openstack-openstack-networker-89k68\" (UID: \"b3e9780b-d240-4d71-bb80-3e21346f7eb9\") " pod="openstack/run-os-openstack-openstack-networker-89k68" Oct 14 09:27:59 crc kubenswrapper[4870]: I1014 09:27:59.008319 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b3e9780b-d240-4d71-bb80-3e21346f7eb9-ssh-key\") pod \"run-os-openstack-openstack-networker-89k68\" (UID: \"b3e9780b-d240-4d71-bb80-3e21346f7eb9\") " pod="openstack/run-os-openstack-openstack-networker-89k68" Oct 14 09:27:59 crc kubenswrapper[4870]: I1014 09:27:59.008632 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4tf8j\" (UniqueName: \"kubernetes.io/projected/b3e9780b-d240-4d71-bb80-3e21346f7eb9-kube-api-access-4tf8j\") pod \"run-os-openstack-openstack-networker-89k68\" (UID: \"b3e9780b-d240-4d71-bb80-3e21346f7eb9\") " pod="openstack/run-os-openstack-openstack-networker-89k68" Oct 14 09:27:59 crc kubenswrapper[4870]: I1014 09:27:59.110401 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b3e9780b-d240-4d71-bb80-3e21346f7eb9-inventory\") pod \"run-os-openstack-openstack-networker-89k68\" (UID: \"b3e9780b-d240-4d71-bb80-3e21346f7eb9\") " pod="openstack/run-os-openstack-openstack-networker-89k68" Oct 14 09:27:59 crc kubenswrapper[4870]: I1014 09:27:59.110548 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b3e9780b-d240-4d71-bb80-3e21346f7eb9-ssh-key\") pod \"run-os-openstack-openstack-networker-89k68\" (UID: \"b3e9780b-d240-4d71-bb80-3e21346f7eb9\") " pod="openstack/run-os-openstack-openstack-networker-89k68" Oct 14 09:27:59 crc kubenswrapper[4870]: I1014 09:27:59.110677 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4tf8j\" (UniqueName: \"kubernetes.io/projected/b3e9780b-d240-4d71-bb80-3e21346f7eb9-kube-api-access-4tf8j\") pod \"run-os-openstack-openstack-networker-89k68\" (UID: \"b3e9780b-d240-4d71-bb80-3e21346f7eb9\") " pod="openstack/run-os-openstack-openstack-networker-89k68" Oct 14 09:27:59 crc kubenswrapper[4870]: I1014 09:27:59.116635 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b3e9780b-d240-4d71-bb80-3e21346f7eb9-ssh-key\") pod \"run-os-openstack-openstack-networker-89k68\" (UID: \"b3e9780b-d240-4d71-bb80-3e21346f7eb9\") " pod="openstack/run-os-openstack-openstack-networker-89k68" Oct 14 09:27:59 crc kubenswrapper[4870]: I1014 09:27:59.124075 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b3e9780b-d240-4d71-bb80-3e21346f7eb9-inventory\") pod \"run-os-openstack-openstack-networker-89k68\" (UID: \"b3e9780b-d240-4d71-bb80-3e21346f7eb9\") " pod="openstack/run-os-openstack-openstack-networker-89k68" Oct 14 09:27:59 crc kubenswrapper[4870]: I1014 09:27:59.141840 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4tf8j\" (UniqueName: \"kubernetes.io/projected/b3e9780b-d240-4d71-bb80-3e21346f7eb9-kube-api-access-4tf8j\") pod \"run-os-openstack-openstack-networker-89k68\" (UID: \"b3e9780b-d240-4d71-bb80-3e21346f7eb9\") " pod="openstack/run-os-openstack-openstack-networker-89k68" Oct 14 09:27:59 crc kubenswrapper[4870]: I1014 09:27:59.239468 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-openstack-openstack-networker-89k68" Oct 14 09:28:00 crc kubenswrapper[4870]: I1014 09:28:00.008472 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-openstack-openstack-networker-89k68"] Oct 14 09:28:00 crc kubenswrapper[4870]: I1014 09:28:00.018176 4870 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 14 09:28:00 crc kubenswrapper[4870]: I1014 09:28:00.823107 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-openstack-openstack-networker-89k68" event={"ID":"b3e9780b-d240-4d71-bb80-3e21346f7eb9","Type":"ContainerStarted","Data":"0d1b1115d82dcededf84c336d9b5ee551d2a29b9ce95d30f31e198c4f59c682c"} Oct 14 09:28:01 crc kubenswrapper[4870]: I1014 09:28:01.843248 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-openstack-openstack-networker-89k68" event={"ID":"b3e9780b-d240-4d71-bb80-3e21346f7eb9","Type":"ContainerStarted","Data":"ccb6cf93bbe299554aa02e51385215484974cc8a883d1165834d19b82ca10cc6"} Oct 14 09:28:01 crc kubenswrapper[4870]: I1014 09:28:01.864552 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-openstack-openstack-networker-89k68" podStartSLOduration=3.269466598 podStartE2EDuration="3.864534896s" podCreationTimestamp="2025-10-14 09:27:58 +0000 UTC" firstStartedPulling="2025-10-14 09:28:00.017820737 +0000 UTC m=+8815.715181118" lastFinishedPulling="2025-10-14 09:28:00.612889055 +0000 UTC m=+8816.310249416" observedRunningTime="2025-10-14 09:28:01.858150479 +0000 UTC m=+8817.555510850" watchObservedRunningTime="2025-10-14 09:28:01.864534896 +0000 UTC m=+8817.561895267" Oct 14 09:28:09 crc kubenswrapper[4870]: I1014 09:28:09.978805 4870 generic.go:334] "Generic (PLEG): container finished" podID="b3e9780b-d240-4d71-bb80-3e21346f7eb9" containerID="ccb6cf93bbe299554aa02e51385215484974cc8a883d1165834d19b82ca10cc6" exitCode=0 Oct 14 09:28:09 crc kubenswrapper[4870]: I1014 09:28:09.979926 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-openstack-openstack-networker-89k68" event={"ID":"b3e9780b-d240-4d71-bb80-3e21346f7eb9","Type":"ContainerDied","Data":"ccb6cf93bbe299554aa02e51385215484974cc8a883d1165834d19b82ca10cc6"} Oct 14 09:28:11 crc kubenswrapper[4870]: I1014 09:28:11.459600 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-openstack-openstack-networker-89k68" Oct 14 09:28:11 crc kubenswrapper[4870]: I1014 09:28:11.522463 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b3e9780b-d240-4d71-bb80-3e21346f7eb9-inventory\") pod \"b3e9780b-d240-4d71-bb80-3e21346f7eb9\" (UID: \"b3e9780b-d240-4d71-bb80-3e21346f7eb9\") " Oct 14 09:28:11 crc kubenswrapper[4870]: I1014 09:28:11.522717 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4tf8j\" (UniqueName: \"kubernetes.io/projected/b3e9780b-d240-4d71-bb80-3e21346f7eb9-kube-api-access-4tf8j\") pod \"b3e9780b-d240-4d71-bb80-3e21346f7eb9\" (UID: \"b3e9780b-d240-4d71-bb80-3e21346f7eb9\") " Oct 14 09:28:11 crc kubenswrapper[4870]: I1014 09:28:11.522799 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b3e9780b-d240-4d71-bb80-3e21346f7eb9-ssh-key\") pod \"b3e9780b-d240-4d71-bb80-3e21346f7eb9\" (UID: \"b3e9780b-d240-4d71-bb80-3e21346f7eb9\") " Oct 14 09:28:11 crc kubenswrapper[4870]: I1014 09:28:11.535405 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b3e9780b-d240-4d71-bb80-3e21346f7eb9-kube-api-access-4tf8j" (OuterVolumeSpecName: "kube-api-access-4tf8j") pod "b3e9780b-d240-4d71-bb80-3e21346f7eb9" (UID: "b3e9780b-d240-4d71-bb80-3e21346f7eb9"). InnerVolumeSpecName "kube-api-access-4tf8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:28:11 crc kubenswrapper[4870]: I1014 09:28:11.553935 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b3e9780b-d240-4d71-bb80-3e21346f7eb9-inventory" (OuterVolumeSpecName: "inventory") pod "b3e9780b-d240-4d71-bb80-3e21346f7eb9" (UID: "b3e9780b-d240-4d71-bb80-3e21346f7eb9"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:28:11 crc kubenswrapper[4870]: I1014 09:28:11.555269 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b3e9780b-d240-4d71-bb80-3e21346f7eb9-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "b3e9780b-d240-4d71-bb80-3e21346f7eb9" (UID: "b3e9780b-d240-4d71-bb80-3e21346f7eb9"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:28:11 crc kubenswrapper[4870]: I1014 09:28:11.625543 4870 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b3e9780b-d240-4d71-bb80-3e21346f7eb9-inventory\") on node \"crc\" DevicePath \"\"" Oct 14 09:28:11 crc kubenswrapper[4870]: I1014 09:28:11.625583 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4tf8j\" (UniqueName: \"kubernetes.io/projected/b3e9780b-d240-4d71-bb80-3e21346f7eb9-kube-api-access-4tf8j\") on node \"crc\" DevicePath \"\"" Oct 14 09:28:11 crc kubenswrapper[4870]: I1014 09:28:11.625595 4870 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b3e9780b-d240-4d71-bb80-3e21346f7eb9-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 14 09:28:12 crc kubenswrapper[4870]: I1014 09:28:12.002904 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-openstack-openstack-networker-89k68" event={"ID":"b3e9780b-d240-4d71-bb80-3e21346f7eb9","Type":"ContainerDied","Data":"0d1b1115d82dcededf84c336d9b5ee551d2a29b9ce95d30f31e198c4f59c682c"} Oct 14 09:28:12 crc kubenswrapper[4870]: I1014 09:28:12.002952 4870 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0d1b1115d82dcededf84c336d9b5ee551d2a29b9ce95d30f31e198c4f59c682c" Oct 14 09:28:12 crc kubenswrapper[4870]: I1014 09:28:12.003006 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-openstack-openstack-networker-89k68" Oct 14 09:28:12 crc kubenswrapper[4870]: I1014 09:28:12.079645 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-openstack-openstack-networker-j54sj"] Oct 14 09:28:12 crc kubenswrapper[4870]: E1014 09:28:12.080158 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3e9780b-d240-4d71-bb80-3e21346f7eb9" containerName="run-os-openstack-openstack-networker" Oct 14 09:28:12 crc kubenswrapper[4870]: I1014 09:28:12.080179 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3e9780b-d240-4d71-bb80-3e21346f7eb9" containerName="run-os-openstack-openstack-networker" Oct 14 09:28:12 crc kubenswrapper[4870]: I1014 09:28:12.080507 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="b3e9780b-d240-4d71-bb80-3e21346f7eb9" containerName="run-os-openstack-openstack-networker" Oct 14 09:28:12 crc kubenswrapper[4870]: I1014 09:28:12.081407 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-openstack-openstack-networker-j54sj" Oct 14 09:28:12 crc kubenswrapper[4870]: I1014 09:28:12.088006 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-networker" Oct 14 09:28:12 crc kubenswrapper[4870]: I1014 09:28:12.088077 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-networker-dockercfg-54hzr" Oct 14 09:28:12 crc kubenswrapper[4870]: I1014 09:28:12.094980 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-openstack-openstack-networker-j54sj"] Oct 14 09:28:12 crc kubenswrapper[4870]: I1014 09:28:12.134829 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2df5305c-8a09-4db0-9fe3-28ea067728a5-ssh-key\") pod \"reboot-os-openstack-openstack-networker-j54sj\" (UID: \"2df5305c-8a09-4db0-9fe3-28ea067728a5\") " pod="openstack/reboot-os-openstack-openstack-networker-j54sj" Oct 14 09:28:12 crc kubenswrapper[4870]: I1014 09:28:12.134889 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2df5305c-8a09-4db0-9fe3-28ea067728a5-inventory\") pod \"reboot-os-openstack-openstack-networker-j54sj\" (UID: \"2df5305c-8a09-4db0-9fe3-28ea067728a5\") " pod="openstack/reboot-os-openstack-openstack-networker-j54sj" Oct 14 09:28:12 crc kubenswrapper[4870]: I1014 09:28:12.135065 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sl8gv\" (UniqueName: \"kubernetes.io/projected/2df5305c-8a09-4db0-9fe3-28ea067728a5-kube-api-access-sl8gv\") pod \"reboot-os-openstack-openstack-networker-j54sj\" (UID: \"2df5305c-8a09-4db0-9fe3-28ea067728a5\") " pod="openstack/reboot-os-openstack-openstack-networker-j54sj" Oct 14 09:28:12 crc kubenswrapper[4870]: I1014 09:28:12.236309 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sl8gv\" (UniqueName: \"kubernetes.io/projected/2df5305c-8a09-4db0-9fe3-28ea067728a5-kube-api-access-sl8gv\") pod \"reboot-os-openstack-openstack-networker-j54sj\" (UID: \"2df5305c-8a09-4db0-9fe3-28ea067728a5\") " pod="openstack/reboot-os-openstack-openstack-networker-j54sj" Oct 14 09:28:12 crc kubenswrapper[4870]: I1014 09:28:12.236838 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2df5305c-8a09-4db0-9fe3-28ea067728a5-ssh-key\") pod \"reboot-os-openstack-openstack-networker-j54sj\" (UID: \"2df5305c-8a09-4db0-9fe3-28ea067728a5\") " pod="openstack/reboot-os-openstack-openstack-networker-j54sj" Oct 14 09:28:12 crc kubenswrapper[4870]: I1014 09:28:12.236890 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2df5305c-8a09-4db0-9fe3-28ea067728a5-inventory\") pod \"reboot-os-openstack-openstack-networker-j54sj\" (UID: \"2df5305c-8a09-4db0-9fe3-28ea067728a5\") " pod="openstack/reboot-os-openstack-openstack-networker-j54sj" Oct 14 09:28:12 crc kubenswrapper[4870]: I1014 09:28:12.247915 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2df5305c-8a09-4db0-9fe3-28ea067728a5-ssh-key\") pod \"reboot-os-openstack-openstack-networker-j54sj\" (UID: \"2df5305c-8a09-4db0-9fe3-28ea067728a5\") " pod="openstack/reboot-os-openstack-openstack-networker-j54sj" Oct 14 09:28:12 crc kubenswrapper[4870]: I1014 09:28:12.247915 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2df5305c-8a09-4db0-9fe3-28ea067728a5-inventory\") pod \"reboot-os-openstack-openstack-networker-j54sj\" (UID: \"2df5305c-8a09-4db0-9fe3-28ea067728a5\") " pod="openstack/reboot-os-openstack-openstack-networker-j54sj" Oct 14 09:28:12 crc kubenswrapper[4870]: I1014 09:28:12.255239 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sl8gv\" (UniqueName: \"kubernetes.io/projected/2df5305c-8a09-4db0-9fe3-28ea067728a5-kube-api-access-sl8gv\") pod \"reboot-os-openstack-openstack-networker-j54sj\" (UID: \"2df5305c-8a09-4db0-9fe3-28ea067728a5\") " pod="openstack/reboot-os-openstack-openstack-networker-j54sj" Oct 14 09:28:12 crc kubenswrapper[4870]: I1014 09:28:12.409172 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-openstack-openstack-networker-j54sj" Oct 14 09:28:12 crc kubenswrapper[4870]: I1014 09:28:12.987887 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-openstack-openstack-networker-j54sj"] Oct 14 09:28:13 crc kubenswrapper[4870]: I1014 09:28:13.023067 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-openstack-openstack-networker-j54sj" event={"ID":"2df5305c-8a09-4db0-9fe3-28ea067728a5","Type":"ContainerStarted","Data":"774e74f91bd37b08697be4b49312f252aa6356615db980dae7e51b2c88c4968a"} Oct 14 09:28:14 crc kubenswrapper[4870]: I1014 09:28:14.052271 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-openstack-openstack-networker-j54sj" event={"ID":"2df5305c-8a09-4db0-9fe3-28ea067728a5","Type":"ContainerStarted","Data":"a1fc915b817ee94754fb805c592ba5d46d7873bc404132785d7b41e5a2dcc4be"} Oct 14 09:28:14 crc kubenswrapper[4870]: I1014 09:28:14.078306 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-openstack-openstack-networker-j54sj" podStartSLOduration=1.463147119 podStartE2EDuration="2.078286422s" podCreationTimestamp="2025-10-14 09:28:12 +0000 UTC" firstStartedPulling="2025-10-14 09:28:13.007751886 +0000 UTC m=+8828.705112287" lastFinishedPulling="2025-10-14 09:28:13.622891189 +0000 UTC m=+8829.320251590" observedRunningTime="2025-10-14 09:28:14.070560662 +0000 UTC m=+8829.767921053" watchObservedRunningTime="2025-10-14 09:28:14.078286422 +0000 UTC m=+8829.775646803" Oct 14 09:28:15 crc kubenswrapper[4870]: I1014 09:28:15.067837 4870 generic.go:334] "Generic (PLEG): container finished" podID="e3f48d1d-384a-4b16-964a-5834de86fdb3" containerID="c0f81efd93cbeb30917b532fced7c2095edcaea800e6f40418be417e0b7ef7e2" exitCode=0 Oct 14 09:28:15 crc kubenswrapper[4870]: I1014 09:28:15.067933 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-openstack-openstack-cell1-7928w" event={"ID":"e3f48d1d-384a-4b16-964a-5834de86fdb3","Type":"ContainerDied","Data":"c0f81efd93cbeb30917b532fced7c2095edcaea800e6f40418be417e0b7ef7e2"} Oct 14 09:28:16 crc kubenswrapper[4870]: I1014 09:28:16.556038 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-openstack-openstack-cell1-7928w" Oct 14 09:28:16 crc kubenswrapper[4870]: I1014 09:28:16.743116 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e3f48d1d-384a-4b16-964a-5834de86fdb3-ssh-key\") pod \"e3f48d1d-384a-4b16-964a-5834de86fdb3\" (UID: \"e3f48d1d-384a-4b16-964a-5834de86fdb3\") " Oct 14 09:28:16 crc kubenswrapper[4870]: I1014 09:28:16.743324 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c2rgp\" (UniqueName: \"kubernetes.io/projected/e3f48d1d-384a-4b16-964a-5834de86fdb3-kube-api-access-c2rgp\") pod \"e3f48d1d-384a-4b16-964a-5834de86fdb3\" (UID: \"e3f48d1d-384a-4b16-964a-5834de86fdb3\") " Oct 14 09:28:16 crc kubenswrapper[4870]: I1014 09:28:16.743395 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e3f48d1d-384a-4b16-964a-5834de86fdb3-ceph\") pod \"e3f48d1d-384a-4b16-964a-5834de86fdb3\" (UID: \"e3f48d1d-384a-4b16-964a-5834de86fdb3\") " Oct 14 09:28:16 crc kubenswrapper[4870]: I1014 09:28:16.743510 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e3f48d1d-384a-4b16-964a-5834de86fdb3-inventory\") pod \"e3f48d1d-384a-4b16-964a-5834de86fdb3\" (UID: \"e3f48d1d-384a-4b16-964a-5834de86fdb3\") " Oct 14 09:28:16 crc kubenswrapper[4870]: I1014 09:28:16.749761 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e3f48d1d-384a-4b16-964a-5834de86fdb3-ceph" (OuterVolumeSpecName: "ceph") pod "e3f48d1d-384a-4b16-964a-5834de86fdb3" (UID: "e3f48d1d-384a-4b16-964a-5834de86fdb3"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:28:16 crc kubenswrapper[4870]: I1014 09:28:16.752819 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e3f48d1d-384a-4b16-964a-5834de86fdb3-kube-api-access-c2rgp" (OuterVolumeSpecName: "kube-api-access-c2rgp") pod "e3f48d1d-384a-4b16-964a-5834de86fdb3" (UID: "e3f48d1d-384a-4b16-964a-5834de86fdb3"). InnerVolumeSpecName "kube-api-access-c2rgp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:28:16 crc kubenswrapper[4870]: I1014 09:28:16.776066 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e3f48d1d-384a-4b16-964a-5834de86fdb3-inventory" (OuterVolumeSpecName: "inventory") pod "e3f48d1d-384a-4b16-964a-5834de86fdb3" (UID: "e3f48d1d-384a-4b16-964a-5834de86fdb3"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:28:16 crc kubenswrapper[4870]: I1014 09:28:16.779615 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e3f48d1d-384a-4b16-964a-5834de86fdb3-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "e3f48d1d-384a-4b16-964a-5834de86fdb3" (UID: "e3f48d1d-384a-4b16-964a-5834de86fdb3"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:28:16 crc kubenswrapper[4870]: I1014 09:28:16.845795 4870 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e3f48d1d-384a-4b16-964a-5834de86fdb3-ceph\") on node \"crc\" DevicePath \"\"" Oct 14 09:28:16 crc kubenswrapper[4870]: I1014 09:28:16.845835 4870 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e3f48d1d-384a-4b16-964a-5834de86fdb3-inventory\") on node \"crc\" DevicePath \"\"" Oct 14 09:28:16 crc kubenswrapper[4870]: I1014 09:28:16.845848 4870 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e3f48d1d-384a-4b16-964a-5834de86fdb3-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 14 09:28:16 crc kubenswrapper[4870]: I1014 09:28:16.845861 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c2rgp\" (UniqueName: \"kubernetes.io/projected/e3f48d1d-384a-4b16-964a-5834de86fdb3-kube-api-access-c2rgp\") on node \"crc\" DevicePath \"\"" Oct 14 09:28:17 crc kubenswrapper[4870]: I1014 09:28:17.098846 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-openstack-openstack-cell1-7928w" event={"ID":"e3f48d1d-384a-4b16-964a-5834de86fdb3","Type":"ContainerDied","Data":"6009edbf25dfcd9c651f667f2ad5fa7e736919ccd3b7e88809df796f97ba7937"} Oct 14 09:28:17 crc kubenswrapper[4870]: I1014 09:28:17.098900 4870 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6009edbf25dfcd9c651f667f2ad5fa7e736919ccd3b7e88809df796f97ba7937" Oct 14 09:28:17 crc kubenswrapper[4870]: I1014 09:28:17.098910 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-openstack-openstack-cell1-7928w" Oct 14 09:28:17 crc kubenswrapper[4870]: I1014 09:28:17.192602 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-openstack-rmpkj"] Oct 14 09:28:17 crc kubenswrapper[4870]: E1014 09:28:17.193154 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3f48d1d-384a-4b16-964a-5834de86fdb3" containerName="configure-os-openstack-openstack-cell1" Oct 14 09:28:17 crc kubenswrapper[4870]: I1014 09:28:17.193176 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3f48d1d-384a-4b16-964a-5834de86fdb3" containerName="configure-os-openstack-openstack-cell1" Oct 14 09:28:17 crc kubenswrapper[4870]: I1014 09:28:17.193485 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="e3f48d1d-384a-4b16-964a-5834de86fdb3" containerName="configure-os-openstack-openstack-cell1" Oct 14 09:28:17 crc kubenswrapper[4870]: I1014 09:28:17.194245 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-openstack-rmpkj" Oct 14 09:28:17 crc kubenswrapper[4870]: I1014 09:28:17.199927 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 14 09:28:17 crc kubenswrapper[4870]: I1014 09:28:17.200936 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-wrz7w" Oct 14 09:28:17 crc kubenswrapper[4870]: I1014 09:28:17.215719 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-openstack-rmpkj"] Oct 14 09:28:17 crc kubenswrapper[4870]: I1014 09:28:17.357650 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/b48a6df4-d071-487d-b843-b60af6b5a141-ssh-key-openstack-cell1\") pod \"ssh-known-hosts-openstack-rmpkj\" (UID: \"b48a6df4-d071-487d-b843-b60af6b5a141\") " pod="openstack/ssh-known-hosts-openstack-rmpkj" Oct 14 09:28:17 crc kubenswrapper[4870]: I1014 09:28:17.357759 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/b48a6df4-d071-487d-b843-b60af6b5a141-inventory-0\") pod \"ssh-known-hosts-openstack-rmpkj\" (UID: \"b48a6df4-d071-487d-b843-b60af6b5a141\") " pod="openstack/ssh-known-hosts-openstack-rmpkj" Oct 14 09:28:17 crc kubenswrapper[4870]: I1014 09:28:17.357811 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-1\" (UniqueName: \"kubernetes.io/secret/b48a6df4-d071-487d-b843-b60af6b5a141-inventory-1\") pod \"ssh-known-hosts-openstack-rmpkj\" (UID: \"b48a6df4-d071-487d-b843-b60af6b5a141\") " pod="openstack/ssh-known-hosts-openstack-rmpkj" Oct 14 09:28:17 crc kubenswrapper[4870]: I1014 09:28:17.357875 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ldk5z\" (UniqueName: \"kubernetes.io/projected/b48a6df4-d071-487d-b843-b60af6b5a141-kube-api-access-ldk5z\") pod \"ssh-known-hosts-openstack-rmpkj\" (UID: \"b48a6df4-d071-487d-b843-b60af6b5a141\") " pod="openstack/ssh-known-hosts-openstack-rmpkj" Oct 14 09:28:17 crc kubenswrapper[4870]: I1014 09:28:17.357936 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-networker\" (UniqueName: \"kubernetes.io/secret/b48a6df4-d071-487d-b843-b60af6b5a141-ssh-key-openstack-networker\") pod \"ssh-known-hosts-openstack-rmpkj\" (UID: \"b48a6df4-d071-487d-b843-b60af6b5a141\") " pod="openstack/ssh-known-hosts-openstack-rmpkj" Oct 14 09:28:17 crc kubenswrapper[4870]: I1014 09:28:17.358008 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b48a6df4-d071-487d-b843-b60af6b5a141-ceph\") pod \"ssh-known-hosts-openstack-rmpkj\" (UID: \"b48a6df4-d071-487d-b843-b60af6b5a141\") " pod="openstack/ssh-known-hosts-openstack-rmpkj" Oct 14 09:28:17 crc kubenswrapper[4870]: I1014 09:28:17.460101 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/b48a6df4-d071-487d-b843-b60af6b5a141-ssh-key-openstack-cell1\") pod \"ssh-known-hosts-openstack-rmpkj\" (UID: \"b48a6df4-d071-487d-b843-b60af6b5a141\") " pod="openstack/ssh-known-hosts-openstack-rmpkj" Oct 14 09:28:17 crc kubenswrapper[4870]: I1014 09:28:17.460219 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/b48a6df4-d071-487d-b843-b60af6b5a141-inventory-0\") pod \"ssh-known-hosts-openstack-rmpkj\" (UID: \"b48a6df4-d071-487d-b843-b60af6b5a141\") " pod="openstack/ssh-known-hosts-openstack-rmpkj" Oct 14 09:28:17 crc kubenswrapper[4870]: I1014 09:28:17.460255 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-1\" (UniqueName: \"kubernetes.io/secret/b48a6df4-d071-487d-b843-b60af6b5a141-inventory-1\") pod \"ssh-known-hosts-openstack-rmpkj\" (UID: \"b48a6df4-d071-487d-b843-b60af6b5a141\") " pod="openstack/ssh-known-hosts-openstack-rmpkj" Oct 14 09:28:17 crc kubenswrapper[4870]: I1014 09:28:17.460318 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ldk5z\" (UniqueName: \"kubernetes.io/projected/b48a6df4-d071-487d-b843-b60af6b5a141-kube-api-access-ldk5z\") pod \"ssh-known-hosts-openstack-rmpkj\" (UID: \"b48a6df4-d071-487d-b843-b60af6b5a141\") " pod="openstack/ssh-known-hosts-openstack-rmpkj" Oct 14 09:28:17 crc kubenswrapper[4870]: I1014 09:28:17.460378 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-networker\" (UniqueName: \"kubernetes.io/secret/b48a6df4-d071-487d-b843-b60af6b5a141-ssh-key-openstack-networker\") pod \"ssh-known-hosts-openstack-rmpkj\" (UID: \"b48a6df4-d071-487d-b843-b60af6b5a141\") " pod="openstack/ssh-known-hosts-openstack-rmpkj" Oct 14 09:28:17 crc kubenswrapper[4870]: I1014 09:28:17.460461 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b48a6df4-d071-487d-b843-b60af6b5a141-ceph\") pod \"ssh-known-hosts-openstack-rmpkj\" (UID: \"b48a6df4-d071-487d-b843-b60af6b5a141\") " pod="openstack/ssh-known-hosts-openstack-rmpkj" Oct 14 09:28:17 crc kubenswrapper[4870]: I1014 09:28:17.464588 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/b48a6df4-d071-487d-b843-b60af6b5a141-inventory-0\") pod \"ssh-known-hosts-openstack-rmpkj\" (UID: \"b48a6df4-d071-487d-b843-b60af6b5a141\") " pod="openstack/ssh-known-hosts-openstack-rmpkj" Oct 14 09:28:17 crc kubenswrapper[4870]: I1014 09:28:17.464914 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-1\" (UniqueName: \"kubernetes.io/secret/b48a6df4-d071-487d-b843-b60af6b5a141-inventory-1\") pod \"ssh-known-hosts-openstack-rmpkj\" (UID: \"b48a6df4-d071-487d-b843-b60af6b5a141\") " pod="openstack/ssh-known-hosts-openstack-rmpkj" Oct 14 09:28:17 crc kubenswrapper[4870]: I1014 09:28:17.465181 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b48a6df4-d071-487d-b843-b60af6b5a141-ceph\") pod \"ssh-known-hosts-openstack-rmpkj\" (UID: \"b48a6df4-d071-487d-b843-b60af6b5a141\") " pod="openstack/ssh-known-hosts-openstack-rmpkj" Oct 14 09:28:17 crc kubenswrapper[4870]: I1014 09:28:17.466541 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-networker\" (UniqueName: \"kubernetes.io/secret/b48a6df4-d071-487d-b843-b60af6b5a141-ssh-key-openstack-networker\") pod \"ssh-known-hosts-openstack-rmpkj\" (UID: \"b48a6df4-d071-487d-b843-b60af6b5a141\") " pod="openstack/ssh-known-hosts-openstack-rmpkj" Oct 14 09:28:17 crc kubenswrapper[4870]: I1014 09:28:17.466837 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/b48a6df4-d071-487d-b843-b60af6b5a141-ssh-key-openstack-cell1\") pod \"ssh-known-hosts-openstack-rmpkj\" (UID: \"b48a6df4-d071-487d-b843-b60af6b5a141\") " pod="openstack/ssh-known-hosts-openstack-rmpkj" Oct 14 09:28:17 crc kubenswrapper[4870]: I1014 09:28:17.487221 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ldk5z\" (UniqueName: \"kubernetes.io/projected/b48a6df4-d071-487d-b843-b60af6b5a141-kube-api-access-ldk5z\") pod \"ssh-known-hosts-openstack-rmpkj\" (UID: \"b48a6df4-d071-487d-b843-b60af6b5a141\") " pod="openstack/ssh-known-hosts-openstack-rmpkj" Oct 14 09:28:17 crc kubenswrapper[4870]: I1014 09:28:17.513589 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-openstack-rmpkj" Oct 14 09:28:18 crc kubenswrapper[4870]: I1014 09:28:18.125819 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-openstack-rmpkj"] Oct 14 09:28:18 crc kubenswrapper[4870]: W1014 09:28:18.133425 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb48a6df4_d071_487d_b843_b60af6b5a141.slice/crio-1a4d0eccb9800495d357a5d830dd321eba0fe886c6646858aeeb5f0d1141ffc6 WatchSource:0}: Error finding container 1a4d0eccb9800495d357a5d830dd321eba0fe886c6646858aeeb5f0d1141ffc6: Status 404 returned error can't find the container with id 1a4d0eccb9800495d357a5d830dd321eba0fe886c6646858aeeb5f0d1141ffc6 Oct 14 09:28:19 crc kubenswrapper[4870]: I1014 09:28:19.123346 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-openstack-rmpkj" event={"ID":"b48a6df4-d071-487d-b843-b60af6b5a141","Type":"ContainerStarted","Data":"20377a2200f1f3cf330053a616d125dfb303de3f29b9577f898e55bc2be34ce8"} Oct 14 09:28:19 crc kubenswrapper[4870]: I1014 09:28:19.123702 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-openstack-rmpkj" event={"ID":"b48a6df4-d071-487d-b843-b60af6b5a141","Type":"ContainerStarted","Data":"1a4d0eccb9800495d357a5d830dd321eba0fe886c6646858aeeb5f0d1141ffc6"} Oct 14 09:28:19 crc kubenswrapper[4870]: I1014 09:28:19.147930 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-openstack-rmpkj" podStartSLOduration=1.5806199269999999 podStartE2EDuration="2.147883402s" podCreationTimestamp="2025-10-14 09:28:17 +0000 UTC" firstStartedPulling="2025-10-14 09:28:18.135666381 +0000 UTC m=+8833.833026762" lastFinishedPulling="2025-10-14 09:28:18.702929856 +0000 UTC m=+8834.400290237" observedRunningTime="2025-10-14 09:28:19.142899709 +0000 UTC m=+8834.840260120" watchObservedRunningTime="2025-10-14 09:28:19.147883402 +0000 UTC m=+8834.845243783" Oct 14 09:28:23 crc kubenswrapper[4870]: I1014 09:28:23.951525 4870 patch_prober.go:28] interesting pod/machine-config-daemon-7tvc7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 09:28:23 crc kubenswrapper[4870]: I1014 09:28:23.952267 4870 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 09:28:30 crc kubenswrapper[4870]: I1014 09:28:30.255035 4870 generic.go:334] "Generic (PLEG): container finished" podID="2df5305c-8a09-4db0-9fe3-28ea067728a5" containerID="a1fc915b817ee94754fb805c592ba5d46d7873bc404132785d7b41e5a2dcc4be" exitCode=0 Oct 14 09:28:30 crc kubenswrapper[4870]: I1014 09:28:30.255159 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-openstack-openstack-networker-j54sj" event={"ID":"2df5305c-8a09-4db0-9fe3-28ea067728a5","Type":"ContainerDied","Data":"a1fc915b817ee94754fb805c592ba5d46d7873bc404132785d7b41e5a2dcc4be"} Oct 14 09:28:31 crc kubenswrapper[4870]: I1014 09:28:31.818863 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-openstack-openstack-networker-j54sj" Oct 14 09:28:31 crc kubenswrapper[4870]: I1014 09:28:31.839393 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2df5305c-8a09-4db0-9fe3-28ea067728a5-inventory\") pod \"2df5305c-8a09-4db0-9fe3-28ea067728a5\" (UID: \"2df5305c-8a09-4db0-9fe3-28ea067728a5\") " Oct 14 09:28:31 crc kubenswrapper[4870]: I1014 09:28:31.839713 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sl8gv\" (UniqueName: \"kubernetes.io/projected/2df5305c-8a09-4db0-9fe3-28ea067728a5-kube-api-access-sl8gv\") pod \"2df5305c-8a09-4db0-9fe3-28ea067728a5\" (UID: \"2df5305c-8a09-4db0-9fe3-28ea067728a5\") " Oct 14 09:28:31 crc kubenswrapper[4870]: I1014 09:28:31.839789 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2df5305c-8a09-4db0-9fe3-28ea067728a5-ssh-key\") pod \"2df5305c-8a09-4db0-9fe3-28ea067728a5\" (UID: \"2df5305c-8a09-4db0-9fe3-28ea067728a5\") " Oct 14 09:28:31 crc kubenswrapper[4870]: I1014 09:28:31.855390 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2df5305c-8a09-4db0-9fe3-28ea067728a5-kube-api-access-sl8gv" (OuterVolumeSpecName: "kube-api-access-sl8gv") pod "2df5305c-8a09-4db0-9fe3-28ea067728a5" (UID: "2df5305c-8a09-4db0-9fe3-28ea067728a5"). InnerVolumeSpecName "kube-api-access-sl8gv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:28:31 crc kubenswrapper[4870]: I1014 09:28:31.888675 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2df5305c-8a09-4db0-9fe3-28ea067728a5-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "2df5305c-8a09-4db0-9fe3-28ea067728a5" (UID: "2df5305c-8a09-4db0-9fe3-28ea067728a5"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:28:31 crc kubenswrapper[4870]: I1014 09:28:31.908183 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2df5305c-8a09-4db0-9fe3-28ea067728a5-inventory" (OuterVolumeSpecName: "inventory") pod "2df5305c-8a09-4db0-9fe3-28ea067728a5" (UID: "2df5305c-8a09-4db0-9fe3-28ea067728a5"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:28:31 crc kubenswrapper[4870]: I1014 09:28:31.942174 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sl8gv\" (UniqueName: \"kubernetes.io/projected/2df5305c-8a09-4db0-9fe3-28ea067728a5-kube-api-access-sl8gv\") on node \"crc\" DevicePath \"\"" Oct 14 09:28:31 crc kubenswrapper[4870]: I1014 09:28:31.942207 4870 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2df5305c-8a09-4db0-9fe3-28ea067728a5-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 14 09:28:31 crc kubenswrapper[4870]: I1014 09:28:31.942219 4870 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2df5305c-8a09-4db0-9fe3-28ea067728a5-inventory\") on node \"crc\" DevicePath \"\"" Oct 14 09:28:32 crc kubenswrapper[4870]: I1014 09:28:32.294818 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-openstack-openstack-networker-j54sj" event={"ID":"2df5305c-8a09-4db0-9fe3-28ea067728a5","Type":"ContainerDied","Data":"774e74f91bd37b08697be4b49312f252aa6356615db980dae7e51b2c88c4968a"} Oct 14 09:28:32 crc kubenswrapper[4870]: I1014 09:28:32.295352 4870 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="774e74f91bd37b08697be4b49312f252aa6356615db980dae7e51b2c88c4968a" Oct 14 09:28:32 crc kubenswrapper[4870]: I1014 09:28:32.295161 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-openstack-openstack-networker-j54sj" Oct 14 09:28:32 crc kubenswrapper[4870]: I1014 09:28:32.412738 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-openstack-openstack-networker-lqg8j"] Oct 14 09:28:32 crc kubenswrapper[4870]: E1014 09:28:32.413514 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2df5305c-8a09-4db0-9fe3-28ea067728a5" containerName="reboot-os-openstack-openstack-networker" Oct 14 09:28:32 crc kubenswrapper[4870]: I1014 09:28:32.413532 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="2df5305c-8a09-4db0-9fe3-28ea067728a5" containerName="reboot-os-openstack-openstack-networker" Oct 14 09:28:32 crc kubenswrapper[4870]: I1014 09:28:32.413955 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="2df5305c-8a09-4db0-9fe3-28ea067728a5" containerName="reboot-os-openstack-openstack-networker" Oct 14 09:28:32 crc kubenswrapper[4870]: I1014 09:28:32.414957 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-openstack-openstack-networker-lqg8j" Oct 14 09:28:32 crc kubenswrapper[4870]: I1014 09:28:32.419185 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-networker-dockercfg-54hzr" Oct 14 09:28:32 crc kubenswrapper[4870]: I1014 09:28:32.437180 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-openstack-openstack-networker-lqg8j"] Oct 14 09:28:32 crc kubenswrapper[4870]: I1014 09:28:32.451121 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5570af78-f5d2-414d-8437-1e9b8210f0de-ovn-combined-ca-bundle\") pod \"install-certs-openstack-openstack-networker-lqg8j\" (UID: \"5570af78-f5d2-414d-8437-1e9b8210f0de\") " pod="openstack/install-certs-openstack-openstack-networker-lqg8j" Oct 14 09:28:32 crc kubenswrapper[4870]: I1014 09:28:32.451170 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hq494\" (UniqueName: \"kubernetes.io/projected/5570af78-f5d2-414d-8437-1e9b8210f0de-kube-api-access-hq494\") pod \"install-certs-openstack-openstack-networker-lqg8j\" (UID: \"5570af78-f5d2-414d-8437-1e9b8210f0de\") " pod="openstack/install-certs-openstack-openstack-networker-lqg8j" Oct 14 09:28:32 crc kubenswrapper[4870]: I1014 09:28:32.451250 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5570af78-f5d2-414d-8437-1e9b8210f0de-inventory\") pod \"install-certs-openstack-openstack-networker-lqg8j\" (UID: \"5570af78-f5d2-414d-8437-1e9b8210f0de\") " pod="openstack/install-certs-openstack-openstack-networker-lqg8j" Oct 14 09:28:32 crc kubenswrapper[4870]: I1014 09:28:32.451377 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5570af78-f5d2-414d-8437-1e9b8210f0de-neutron-metadata-combined-ca-bundle\") pod \"install-certs-openstack-openstack-networker-lqg8j\" (UID: \"5570af78-f5d2-414d-8437-1e9b8210f0de\") " pod="openstack/install-certs-openstack-openstack-networker-lqg8j" Oct 14 09:28:32 crc kubenswrapper[4870]: I1014 09:28:32.451397 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5570af78-f5d2-414d-8437-1e9b8210f0de-bootstrap-combined-ca-bundle\") pod \"install-certs-openstack-openstack-networker-lqg8j\" (UID: \"5570af78-f5d2-414d-8437-1e9b8210f0de\") " pod="openstack/install-certs-openstack-openstack-networker-lqg8j" Oct 14 09:28:32 crc kubenswrapper[4870]: I1014 09:28:32.451469 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5570af78-f5d2-414d-8437-1e9b8210f0de-ssh-key\") pod \"install-certs-openstack-openstack-networker-lqg8j\" (UID: \"5570af78-f5d2-414d-8437-1e9b8210f0de\") " pod="openstack/install-certs-openstack-openstack-networker-lqg8j" Oct 14 09:28:32 crc kubenswrapper[4870]: I1014 09:28:32.553265 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5570af78-f5d2-414d-8437-1e9b8210f0de-ssh-key\") pod \"install-certs-openstack-openstack-networker-lqg8j\" (UID: \"5570af78-f5d2-414d-8437-1e9b8210f0de\") " pod="openstack/install-certs-openstack-openstack-networker-lqg8j" Oct 14 09:28:32 crc kubenswrapper[4870]: I1014 09:28:32.553358 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5570af78-f5d2-414d-8437-1e9b8210f0de-ovn-combined-ca-bundle\") pod \"install-certs-openstack-openstack-networker-lqg8j\" (UID: \"5570af78-f5d2-414d-8437-1e9b8210f0de\") " pod="openstack/install-certs-openstack-openstack-networker-lqg8j" Oct 14 09:28:32 crc kubenswrapper[4870]: I1014 09:28:32.553387 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hq494\" (UniqueName: \"kubernetes.io/projected/5570af78-f5d2-414d-8437-1e9b8210f0de-kube-api-access-hq494\") pod \"install-certs-openstack-openstack-networker-lqg8j\" (UID: \"5570af78-f5d2-414d-8437-1e9b8210f0de\") " pod="openstack/install-certs-openstack-openstack-networker-lqg8j" Oct 14 09:28:32 crc kubenswrapper[4870]: I1014 09:28:32.553498 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5570af78-f5d2-414d-8437-1e9b8210f0de-inventory\") pod \"install-certs-openstack-openstack-networker-lqg8j\" (UID: \"5570af78-f5d2-414d-8437-1e9b8210f0de\") " pod="openstack/install-certs-openstack-openstack-networker-lqg8j" Oct 14 09:28:32 crc kubenswrapper[4870]: I1014 09:28:32.553654 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5570af78-f5d2-414d-8437-1e9b8210f0de-neutron-metadata-combined-ca-bundle\") pod \"install-certs-openstack-openstack-networker-lqg8j\" (UID: \"5570af78-f5d2-414d-8437-1e9b8210f0de\") " pod="openstack/install-certs-openstack-openstack-networker-lqg8j" Oct 14 09:28:32 crc kubenswrapper[4870]: I1014 09:28:32.553695 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5570af78-f5d2-414d-8437-1e9b8210f0de-bootstrap-combined-ca-bundle\") pod \"install-certs-openstack-openstack-networker-lqg8j\" (UID: \"5570af78-f5d2-414d-8437-1e9b8210f0de\") " pod="openstack/install-certs-openstack-openstack-networker-lqg8j" Oct 14 09:28:32 crc kubenswrapper[4870]: I1014 09:28:32.558595 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5570af78-f5d2-414d-8437-1e9b8210f0de-neutron-metadata-combined-ca-bundle\") pod \"install-certs-openstack-openstack-networker-lqg8j\" (UID: \"5570af78-f5d2-414d-8437-1e9b8210f0de\") " pod="openstack/install-certs-openstack-openstack-networker-lqg8j" Oct 14 09:28:32 crc kubenswrapper[4870]: I1014 09:28:32.559249 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5570af78-f5d2-414d-8437-1e9b8210f0de-inventory\") pod \"install-certs-openstack-openstack-networker-lqg8j\" (UID: \"5570af78-f5d2-414d-8437-1e9b8210f0de\") " pod="openstack/install-certs-openstack-openstack-networker-lqg8j" Oct 14 09:28:32 crc kubenswrapper[4870]: I1014 09:28:32.560485 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5570af78-f5d2-414d-8437-1e9b8210f0de-ovn-combined-ca-bundle\") pod \"install-certs-openstack-openstack-networker-lqg8j\" (UID: \"5570af78-f5d2-414d-8437-1e9b8210f0de\") " pod="openstack/install-certs-openstack-openstack-networker-lqg8j" Oct 14 09:28:32 crc kubenswrapper[4870]: I1014 09:28:32.561480 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5570af78-f5d2-414d-8437-1e9b8210f0de-ssh-key\") pod \"install-certs-openstack-openstack-networker-lqg8j\" (UID: \"5570af78-f5d2-414d-8437-1e9b8210f0de\") " pod="openstack/install-certs-openstack-openstack-networker-lqg8j" Oct 14 09:28:32 crc kubenswrapper[4870]: I1014 09:28:32.561782 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5570af78-f5d2-414d-8437-1e9b8210f0de-bootstrap-combined-ca-bundle\") pod \"install-certs-openstack-openstack-networker-lqg8j\" (UID: \"5570af78-f5d2-414d-8437-1e9b8210f0de\") " pod="openstack/install-certs-openstack-openstack-networker-lqg8j" Oct 14 09:28:32 crc kubenswrapper[4870]: I1014 09:28:32.572244 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hq494\" (UniqueName: \"kubernetes.io/projected/5570af78-f5d2-414d-8437-1e9b8210f0de-kube-api-access-hq494\") pod \"install-certs-openstack-openstack-networker-lqg8j\" (UID: \"5570af78-f5d2-414d-8437-1e9b8210f0de\") " pod="openstack/install-certs-openstack-openstack-networker-lqg8j" Oct 14 09:28:32 crc kubenswrapper[4870]: I1014 09:28:32.739009 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-openstack-openstack-networker-lqg8j" Oct 14 09:28:33 crc kubenswrapper[4870]: I1014 09:28:33.339499 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-openstack-openstack-networker-lqg8j"] Oct 14 09:28:33 crc kubenswrapper[4870]: W1014 09:28:33.343762 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5570af78_f5d2_414d_8437_1e9b8210f0de.slice/crio-67f26bec13d24b1474b7b57fe4234a40cd3d12746c3a4a53143c775654130d72 WatchSource:0}: Error finding container 67f26bec13d24b1474b7b57fe4234a40cd3d12746c3a4a53143c775654130d72: Status 404 returned error can't find the container with id 67f26bec13d24b1474b7b57fe4234a40cd3d12746c3a4a53143c775654130d72 Oct 14 09:28:34 crc kubenswrapper[4870]: I1014 09:28:34.321734 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-openstack-openstack-networker-lqg8j" event={"ID":"5570af78-f5d2-414d-8437-1e9b8210f0de","Type":"ContainerStarted","Data":"67f26bec13d24b1474b7b57fe4234a40cd3d12746c3a4a53143c775654130d72"} Oct 14 09:28:34 crc kubenswrapper[4870]: I1014 09:28:34.332977 4870 generic.go:334] "Generic (PLEG): container finished" podID="b48a6df4-d071-487d-b843-b60af6b5a141" containerID="20377a2200f1f3cf330053a616d125dfb303de3f29b9577f898e55bc2be34ce8" exitCode=0 Oct 14 09:28:34 crc kubenswrapper[4870]: I1014 09:28:34.333261 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-openstack-rmpkj" event={"ID":"b48a6df4-d071-487d-b843-b60af6b5a141","Type":"ContainerDied","Data":"20377a2200f1f3cf330053a616d125dfb303de3f29b9577f898e55bc2be34ce8"} Oct 14 09:28:35 crc kubenswrapper[4870]: I1014 09:28:35.350040 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-openstack-openstack-networker-lqg8j" event={"ID":"5570af78-f5d2-414d-8437-1e9b8210f0de","Type":"ContainerStarted","Data":"5fea7c16f23957fbfb2510f92eb80f4c27c0decd8bf744f373e64a83107b3379"} Oct 14 09:28:35 crc kubenswrapper[4870]: I1014 09:28:35.390989 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-openstack-openstack-networker-lqg8j" podStartSLOduration=2.225079878 podStartE2EDuration="3.390960518s" podCreationTimestamp="2025-10-14 09:28:32 +0000 UTC" firstStartedPulling="2025-10-14 09:28:33.347687864 +0000 UTC m=+8849.045048245" lastFinishedPulling="2025-10-14 09:28:34.513568494 +0000 UTC m=+8850.210928885" observedRunningTime="2025-10-14 09:28:35.380669365 +0000 UTC m=+8851.078029756" watchObservedRunningTime="2025-10-14 09:28:35.390960518 +0000 UTC m=+8851.088320919" Oct 14 09:28:35 crc kubenswrapper[4870]: I1014 09:28:35.855634 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-openstack-rmpkj" Oct 14 09:28:35 crc kubenswrapper[4870]: I1014 09:28:35.931223 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-1\" (UniqueName: \"kubernetes.io/secret/b48a6df4-d071-487d-b843-b60af6b5a141-inventory-1\") pod \"b48a6df4-d071-487d-b843-b60af6b5a141\" (UID: \"b48a6df4-d071-487d-b843-b60af6b5a141\") " Oct 14 09:28:35 crc kubenswrapper[4870]: I1014 09:28:35.931324 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-networker\" (UniqueName: \"kubernetes.io/secret/b48a6df4-d071-487d-b843-b60af6b5a141-ssh-key-openstack-networker\") pod \"b48a6df4-d071-487d-b843-b60af6b5a141\" (UID: \"b48a6df4-d071-487d-b843-b60af6b5a141\") " Oct 14 09:28:35 crc kubenswrapper[4870]: I1014 09:28:35.931413 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ldk5z\" (UniqueName: \"kubernetes.io/projected/b48a6df4-d071-487d-b843-b60af6b5a141-kube-api-access-ldk5z\") pod \"b48a6df4-d071-487d-b843-b60af6b5a141\" (UID: \"b48a6df4-d071-487d-b843-b60af6b5a141\") " Oct 14 09:28:35 crc kubenswrapper[4870]: I1014 09:28:35.931563 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b48a6df4-d071-487d-b843-b60af6b5a141-ceph\") pod \"b48a6df4-d071-487d-b843-b60af6b5a141\" (UID: \"b48a6df4-d071-487d-b843-b60af6b5a141\") " Oct 14 09:28:35 crc kubenswrapper[4870]: I1014 09:28:35.931640 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/b48a6df4-d071-487d-b843-b60af6b5a141-inventory-0\") pod \"b48a6df4-d071-487d-b843-b60af6b5a141\" (UID: \"b48a6df4-d071-487d-b843-b60af6b5a141\") " Oct 14 09:28:35 crc kubenswrapper[4870]: I1014 09:28:35.931679 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/b48a6df4-d071-487d-b843-b60af6b5a141-ssh-key-openstack-cell1\") pod \"b48a6df4-d071-487d-b843-b60af6b5a141\" (UID: \"b48a6df4-d071-487d-b843-b60af6b5a141\") " Oct 14 09:28:35 crc kubenswrapper[4870]: I1014 09:28:35.937663 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b48a6df4-d071-487d-b843-b60af6b5a141-ceph" (OuterVolumeSpecName: "ceph") pod "b48a6df4-d071-487d-b843-b60af6b5a141" (UID: "b48a6df4-d071-487d-b843-b60af6b5a141"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:28:35 crc kubenswrapper[4870]: I1014 09:28:35.937837 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b48a6df4-d071-487d-b843-b60af6b5a141-kube-api-access-ldk5z" (OuterVolumeSpecName: "kube-api-access-ldk5z") pod "b48a6df4-d071-487d-b843-b60af6b5a141" (UID: "b48a6df4-d071-487d-b843-b60af6b5a141"). InnerVolumeSpecName "kube-api-access-ldk5z". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:28:35 crc kubenswrapper[4870]: I1014 09:28:35.959983 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b48a6df4-d071-487d-b843-b60af6b5a141-inventory-1" (OuterVolumeSpecName: "inventory-1") pod "b48a6df4-d071-487d-b843-b60af6b5a141" (UID: "b48a6df4-d071-487d-b843-b60af6b5a141"). InnerVolumeSpecName "inventory-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:28:35 crc kubenswrapper[4870]: I1014 09:28:35.964878 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b48a6df4-d071-487d-b843-b60af6b5a141-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "b48a6df4-d071-487d-b843-b60af6b5a141" (UID: "b48a6df4-d071-487d-b843-b60af6b5a141"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:28:35 crc kubenswrapper[4870]: I1014 09:28:35.965992 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b48a6df4-d071-487d-b843-b60af6b5a141-ssh-key-openstack-cell1" (OuterVolumeSpecName: "ssh-key-openstack-cell1") pod "b48a6df4-d071-487d-b843-b60af6b5a141" (UID: "b48a6df4-d071-487d-b843-b60af6b5a141"). InnerVolumeSpecName "ssh-key-openstack-cell1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:28:35 crc kubenswrapper[4870]: I1014 09:28:35.967137 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b48a6df4-d071-487d-b843-b60af6b5a141-ssh-key-openstack-networker" (OuterVolumeSpecName: "ssh-key-openstack-networker") pod "b48a6df4-d071-487d-b843-b60af6b5a141" (UID: "b48a6df4-d071-487d-b843-b60af6b5a141"). InnerVolumeSpecName "ssh-key-openstack-networker". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:28:36 crc kubenswrapper[4870]: I1014 09:28:36.035099 4870 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-networker\" (UniqueName: \"kubernetes.io/secret/b48a6df4-d071-487d-b843-b60af6b5a141-ssh-key-openstack-networker\") on node \"crc\" DevicePath \"\"" Oct 14 09:28:36 crc kubenswrapper[4870]: I1014 09:28:36.035127 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ldk5z\" (UniqueName: \"kubernetes.io/projected/b48a6df4-d071-487d-b843-b60af6b5a141-kube-api-access-ldk5z\") on node \"crc\" DevicePath \"\"" Oct 14 09:28:36 crc kubenswrapper[4870]: I1014 09:28:36.035140 4870 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b48a6df4-d071-487d-b843-b60af6b5a141-ceph\") on node \"crc\" DevicePath \"\"" Oct 14 09:28:36 crc kubenswrapper[4870]: I1014 09:28:36.035150 4870 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/b48a6df4-d071-487d-b843-b60af6b5a141-inventory-0\") on node \"crc\" DevicePath \"\"" Oct 14 09:28:36 crc kubenswrapper[4870]: I1014 09:28:36.035161 4870 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/b48a6df4-d071-487d-b843-b60af6b5a141-ssh-key-openstack-cell1\") on node \"crc\" DevicePath \"\"" Oct 14 09:28:36 crc kubenswrapper[4870]: I1014 09:28:36.035172 4870 reconciler_common.go:293] "Volume detached for volume \"inventory-1\" (UniqueName: \"kubernetes.io/secret/b48a6df4-d071-487d-b843-b60af6b5a141-inventory-1\") on node \"crc\" DevicePath \"\"" Oct 14 09:28:36 crc kubenswrapper[4870]: I1014 09:28:36.364425 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-openstack-rmpkj" event={"ID":"b48a6df4-d071-487d-b843-b60af6b5a141","Type":"ContainerDied","Data":"1a4d0eccb9800495d357a5d830dd321eba0fe886c6646858aeeb5f0d1141ffc6"} Oct 14 09:28:36 crc kubenswrapper[4870]: I1014 09:28:36.365414 4870 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1a4d0eccb9800495d357a5d830dd321eba0fe886c6646858aeeb5f0d1141ffc6" Oct 14 09:28:36 crc kubenswrapper[4870]: I1014 09:28:36.364483 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-openstack-rmpkj" Oct 14 09:28:36 crc kubenswrapper[4870]: I1014 09:28:36.461567 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-openstack-openstack-cell1-m4fnw"] Oct 14 09:28:36 crc kubenswrapper[4870]: E1014 09:28:36.464278 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b48a6df4-d071-487d-b843-b60af6b5a141" containerName="ssh-known-hosts-openstack" Oct 14 09:28:36 crc kubenswrapper[4870]: I1014 09:28:36.464332 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="b48a6df4-d071-487d-b843-b60af6b5a141" containerName="ssh-known-hosts-openstack" Oct 14 09:28:36 crc kubenswrapper[4870]: I1014 09:28:36.464936 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="b48a6df4-d071-487d-b843-b60af6b5a141" containerName="ssh-known-hosts-openstack" Oct 14 09:28:36 crc kubenswrapper[4870]: I1014 09:28:36.466235 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-openstack-openstack-cell1-m4fnw" Oct 14 09:28:36 crc kubenswrapper[4870]: I1014 09:28:36.468169 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 14 09:28:36 crc kubenswrapper[4870]: I1014 09:28:36.471135 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-wrz7w" Oct 14 09:28:36 crc kubenswrapper[4870]: I1014 09:28:36.484739 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-openstack-openstack-cell1-m4fnw"] Oct 14 09:28:36 crc kubenswrapper[4870]: I1014 09:28:36.548343 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fe4c4a5a-274a-4adf-929a-46a1057c51ee-ssh-key\") pod \"run-os-openstack-openstack-cell1-m4fnw\" (UID: \"fe4c4a5a-274a-4adf-929a-46a1057c51ee\") " pod="openstack/run-os-openstack-openstack-cell1-m4fnw" Oct 14 09:28:36 crc kubenswrapper[4870]: I1014 09:28:36.548391 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/fe4c4a5a-274a-4adf-929a-46a1057c51ee-ceph\") pod \"run-os-openstack-openstack-cell1-m4fnw\" (UID: \"fe4c4a5a-274a-4adf-929a-46a1057c51ee\") " pod="openstack/run-os-openstack-openstack-cell1-m4fnw" Oct 14 09:28:36 crc kubenswrapper[4870]: I1014 09:28:36.548429 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-thtwm\" (UniqueName: \"kubernetes.io/projected/fe4c4a5a-274a-4adf-929a-46a1057c51ee-kube-api-access-thtwm\") pod \"run-os-openstack-openstack-cell1-m4fnw\" (UID: \"fe4c4a5a-274a-4adf-929a-46a1057c51ee\") " pod="openstack/run-os-openstack-openstack-cell1-m4fnw" Oct 14 09:28:36 crc kubenswrapper[4870]: I1014 09:28:36.548554 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fe4c4a5a-274a-4adf-929a-46a1057c51ee-inventory\") pod \"run-os-openstack-openstack-cell1-m4fnw\" (UID: \"fe4c4a5a-274a-4adf-929a-46a1057c51ee\") " pod="openstack/run-os-openstack-openstack-cell1-m4fnw" Oct 14 09:28:36 crc kubenswrapper[4870]: I1014 09:28:36.649955 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fe4c4a5a-274a-4adf-929a-46a1057c51ee-inventory\") pod \"run-os-openstack-openstack-cell1-m4fnw\" (UID: \"fe4c4a5a-274a-4adf-929a-46a1057c51ee\") " pod="openstack/run-os-openstack-openstack-cell1-m4fnw" Oct 14 09:28:36 crc kubenswrapper[4870]: I1014 09:28:36.650020 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fe4c4a5a-274a-4adf-929a-46a1057c51ee-ssh-key\") pod \"run-os-openstack-openstack-cell1-m4fnw\" (UID: \"fe4c4a5a-274a-4adf-929a-46a1057c51ee\") " pod="openstack/run-os-openstack-openstack-cell1-m4fnw" Oct 14 09:28:36 crc kubenswrapper[4870]: I1014 09:28:36.650059 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/fe4c4a5a-274a-4adf-929a-46a1057c51ee-ceph\") pod \"run-os-openstack-openstack-cell1-m4fnw\" (UID: \"fe4c4a5a-274a-4adf-929a-46a1057c51ee\") " pod="openstack/run-os-openstack-openstack-cell1-m4fnw" Oct 14 09:28:36 crc kubenswrapper[4870]: I1014 09:28:36.650113 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-thtwm\" (UniqueName: \"kubernetes.io/projected/fe4c4a5a-274a-4adf-929a-46a1057c51ee-kube-api-access-thtwm\") pod \"run-os-openstack-openstack-cell1-m4fnw\" (UID: \"fe4c4a5a-274a-4adf-929a-46a1057c51ee\") " pod="openstack/run-os-openstack-openstack-cell1-m4fnw" Oct 14 09:28:36 crc kubenswrapper[4870]: I1014 09:28:36.653701 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fe4c4a5a-274a-4adf-929a-46a1057c51ee-inventory\") pod \"run-os-openstack-openstack-cell1-m4fnw\" (UID: \"fe4c4a5a-274a-4adf-929a-46a1057c51ee\") " pod="openstack/run-os-openstack-openstack-cell1-m4fnw" Oct 14 09:28:36 crc kubenswrapper[4870]: I1014 09:28:36.653730 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fe4c4a5a-274a-4adf-929a-46a1057c51ee-ssh-key\") pod \"run-os-openstack-openstack-cell1-m4fnw\" (UID: \"fe4c4a5a-274a-4adf-929a-46a1057c51ee\") " pod="openstack/run-os-openstack-openstack-cell1-m4fnw" Oct 14 09:28:36 crc kubenswrapper[4870]: I1014 09:28:36.654366 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/fe4c4a5a-274a-4adf-929a-46a1057c51ee-ceph\") pod \"run-os-openstack-openstack-cell1-m4fnw\" (UID: \"fe4c4a5a-274a-4adf-929a-46a1057c51ee\") " pod="openstack/run-os-openstack-openstack-cell1-m4fnw" Oct 14 09:28:36 crc kubenswrapper[4870]: I1014 09:28:36.667638 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-thtwm\" (UniqueName: \"kubernetes.io/projected/fe4c4a5a-274a-4adf-929a-46a1057c51ee-kube-api-access-thtwm\") pod \"run-os-openstack-openstack-cell1-m4fnw\" (UID: \"fe4c4a5a-274a-4adf-929a-46a1057c51ee\") " pod="openstack/run-os-openstack-openstack-cell1-m4fnw" Oct 14 09:28:36 crc kubenswrapper[4870]: I1014 09:28:36.795108 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-openstack-openstack-cell1-m4fnw" Oct 14 09:28:37 crc kubenswrapper[4870]: I1014 09:28:37.463725 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-openstack-openstack-cell1-m4fnw"] Oct 14 09:28:38 crc kubenswrapper[4870]: I1014 09:28:38.388810 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-openstack-openstack-cell1-m4fnw" event={"ID":"fe4c4a5a-274a-4adf-929a-46a1057c51ee","Type":"ContainerStarted","Data":"111597cf1db9f1f7eef2f1cf5eda3d068513f0f73b1b7064ad339add7778074d"} Oct 14 09:28:38 crc kubenswrapper[4870]: I1014 09:28:38.389175 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-openstack-openstack-cell1-m4fnw" event={"ID":"fe4c4a5a-274a-4adf-929a-46a1057c51ee","Type":"ContainerStarted","Data":"cc58b0932a7c3d17b32926ff7ad4fe8f992ff5500d03438604e3b945d5aeb343"} Oct 14 09:28:38 crc kubenswrapper[4870]: I1014 09:28:38.419748 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-openstack-openstack-cell1-m4fnw" podStartSLOduration=1.848109013 podStartE2EDuration="2.419716024s" podCreationTimestamp="2025-10-14 09:28:36 +0000 UTC" firstStartedPulling="2025-10-14 09:28:37.478746167 +0000 UTC m=+8853.176106538" lastFinishedPulling="2025-10-14 09:28:38.050353168 +0000 UTC m=+8853.747713549" observedRunningTime="2025-10-14 09:28:38.404674744 +0000 UTC m=+8854.102035155" watchObservedRunningTime="2025-10-14 09:28:38.419716024 +0000 UTC m=+8854.117076435" Oct 14 09:28:46 crc kubenswrapper[4870]: I1014 09:28:46.495333 4870 generic.go:334] "Generic (PLEG): container finished" podID="5570af78-f5d2-414d-8437-1e9b8210f0de" containerID="5fea7c16f23957fbfb2510f92eb80f4c27c0decd8bf744f373e64a83107b3379" exitCode=0 Oct 14 09:28:46 crc kubenswrapper[4870]: I1014 09:28:46.495455 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-openstack-openstack-networker-lqg8j" event={"ID":"5570af78-f5d2-414d-8437-1e9b8210f0de","Type":"ContainerDied","Data":"5fea7c16f23957fbfb2510f92eb80f4c27c0decd8bf744f373e64a83107b3379"} Oct 14 09:28:47 crc kubenswrapper[4870]: I1014 09:28:47.508675 4870 generic.go:334] "Generic (PLEG): container finished" podID="fe4c4a5a-274a-4adf-929a-46a1057c51ee" containerID="111597cf1db9f1f7eef2f1cf5eda3d068513f0f73b1b7064ad339add7778074d" exitCode=0 Oct 14 09:28:47 crc kubenswrapper[4870]: I1014 09:28:47.508878 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-openstack-openstack-cell1-m4fnw" event={"ID":"fe4c4a5a-274a-4adf-929a-46a1057c51ee","Type":"ContainerDied","Data":"111597cf1db9f1f7eef2f1cf5eda3d068513f0f73b1b7064ad339add7778074d"} Oct 14 09:28:48 crc kubenswrapper[4870]: I1014 09:28:48.034967 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-openstack-openstack-networker-lqg8j" Oct 14 09:28:48 crc kubenswrapper[4870]: I1014 09:28:48.210830 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hq494\" (UniqueName: \"kubernetes.io/projected/5570af78-f5d2-414d-8437-1e9b8210f0de-kube-api-access-hq494\") pod \"5570af78-f5d2-414d-8437-1e9b8210f0de\" (UID: \"5570af78-f5d2-414d-8437-1e9b8210f0de\") " Oct 14 09:28:48 crc kubenswrapper[4870]: I1014 09:28:48.211385 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5570af78-f5d2-414d-8437-1e9b8210f0de-inventory\") pod \"5570af78-f5d2-414d-8437-1e9b8210f0de\" (UID: \"5570af78-f5d2-414d-8437-1e9b8210f0de\") " Oct 14 09:28:48 crc kubenswrapper[4870]: I1014 09:28:48.211995 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5570af78-f5d2-414d-8437-1e9b8210f0de-bootstrap-combined-ca-bundle\") pod \"5570af78-f5d2-414d-8437-1e9b8210f0de\" (UID: \"5570af78-f5d2-414d-8437-1e9b8210f0de\") " Oct 14 09:28:48 crc kubenswrapper[4870]: I1014 09:28:48.212755 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5570af78-f5d2-414d-8437-1e9b8210f0de-ovn-combined-ca-bundle\") pod \"5570af78-f5d2-414d-8437-1e9b8210f0de\" (UID: \"5570af78-f5d2-414d-8437-1e9b8210f0de\") " Oct 14 09:28:48 crc kubenswrapper[4870]: I1014 09:28:48.212804 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5570af78-f5d2-414d-8437-1e9b8210f0de-neutron-metadata-combined-ca-bundle\") pod \"5570af78-f5d2-414d-8437-1e9b8210f0de\" (UID: \"5570af78-f5d2-414d-8437-1e9b8210f0de\") " Oct 14 09:28:48 crc kubenswrapper[4870]: I1014 09:28:48.212922 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5570af78-f5d2-414d-8437-1e9b8210f0de-ssh-key\") pod \"5570af78-f5d2-414d-8437-1e9b8210f0de\" (UID: \"5570af78-f5d2-414d-8437-1e9b8210f0de\") " Oct 14 09:28:48 crc kubenswrapper[4870]: I1014 09:28:48.218117 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5570af78-f5d2-414d-8437-1e9b8210f0de-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "5570af78-f5d2-414d-8437-1e9b8210f0de" (UID: "5570af78-f5d2-414d-8437-1e9b8210f0de"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:28:48 crc kubenswrapper[4870]: I1014 09:28:48.218800 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5570af78-f5d2-414d-8437-1e9b8210f0de-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "5570af78-f5d2-414d-8437-1e9b8210f0de" (UID: "5570af78-f5d2-414d-8437-1e9b8210f0de"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:28:48 crc kubenswrapper[4870]: I1014 09:28:48.219348 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5570af78-f5d2-414d-8437-1e9b8210f0de-kube-api-access-hq494" (OuterVolumeSpecName: "kube-api-access-hq494") pod "5570af78-f5d2-414d-8437-1e9b8210f0de" (UID: "5570af78-f5d2-414d-8437-1e9b8210f0de"). InnerVolumeSpecName "kube-api-access-hq494". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:28:48 crc kubenswrapper[4870]: I1014 09:28:48.220942 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5570af78-f5d2-414d-8437-1e9b8210f0de-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "5570af78-f5d2-414d-8437-1e9b8210f0de" (UID: "5570af78-f5d2-414d-8437-1e9b8210f0de"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:28:48 crc kubenswrapper[4870]: I1014 09:28:48.268502 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5570af78-f5d2-414d-8437-1e9b8210f0de-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "5570af78-f5d2-414d-8437-1e9b8210f0de" (UID: "5570af78-f5d2-414d-8437-1e9b8210f0de"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:28:48 crc kubenswrapper[4870]: I1014 09:28:48.269940 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5570af78-f5d2-414d-8437-1e9b8210f0de-inventory" (OuterVolumeSpecName: "inventory") pod "5570af78-f5d2-414d-8437-1e9b8210f0de" (UID: "5570af78-f5d2-414d-8437-1e9b8210f0de"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:28:48 crc kubenswrapper[4870]: I1014 09:28:48.316816 4870 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5570af78-f5d2-414d-8437-1e9b8210f0de-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 09:28:48 crc kubenswrapper[4870]: I1014 09:28:48.316865 4870 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5570af78-f5d2-414d-8437-1e9b8210f0de-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 09:28:48 crc kubenswrapper[4870]: I1014 09:28:48.316885 4870 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5570af78-f5d2-414d-8437-1e9b8210f0de-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 09:28:48 crc kubenswrapper[4870]: I1014 09:28:48.316899 4870 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5570af78-f5d2-414d-8437-1e9b8210f0de-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 14 09:28:48 crc kubenswrapper[4870]: I1014 09:28:48.316911 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hq494\" (UniqueName: \"kubernetes.io/projected/5570af78-f5d2-414d-8437-1e9b8210f0de-kube-api-access-hq494\") on node \"crc\" DevicePath \"\"" Oct 14 09:28:48 crc kubenswrapper[4870]: I1014 09:28:48.316925 4870 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5570af78-f5d2-414d-8437-1e9b8210f0de-inventory\") on node \"crc\" DevicePath \"\"" Oct 14 09:28:48 crc kubenswrapper[4870]: I1014 09:28:48.525593 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-openstack-openstack-networker-lqg8j" event={"ID":"5570af78-f5d2-414d-8437-1e9b8210f0de","Type":"ContainerDied","Data":"67f26bec13d24b1474b7b57fe4234a40cd3d12746c3a4a53143c775654130d72"} Oct 14 09:28:48 crc kubenswrapper[4870]: I1014 09:28:48.525720 4870 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="67f26bec13d24b1474b7b57fe4234a40cd3d12746c3a4a53143c775654130d72" Oct 14 09:28:48 crc kubenswrapper[4870]: I1014 09:28:48.525642 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-openstack-openstack-networker-lqg8j" Oct 14 09:28:48 crc kubenswrapper[4870]: I1014 09:28:48.612810 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-openstack-openstack-networker-hw7gw"] Oct 14 09:28:48 crc kubenswrapper[4870]: E1014 09:28:48.613614 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5570af78-f5d2-414d-8437-1e9b8210f0de" containerName="install-certs-openstack-openstack-networker" Oct 14 09:28:48 crc kubenswrapper[4870]: I1014 09:28:48.613637 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="5570af78-f5d2-414d-8437-1e9b8210f0de" containerName="install-certs-openstack-openstack-networker" Oct 14 09:28:48 crc kubenswrapper[4870]: I1014 09:28:48.613898 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="5570af78-f5d2-414d-8437-1e9b8210f0de" containerName="install-certs-openstack-openstack-networker" Oct 14 09:28:48 crc kubenswrapper[4870]: I1014 09:28:48.614771 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-openstack-openstack-networker-hw7gw" Oct 14 09:28:48 crc kubenswrapper[4870]: I1014 09:28:48.618507 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-networker-dockercfg-54hzr" Oct 14 09:28:48 crc kubenswrapper[4870]: I1014 09:28:48.619164 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-networker" Oct 14 09:28:48 crc kubenswrapper[4870]: I1014 09:28:48.619301 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Oct 14 09:28:48 crc kubenswrapper[4870]: I1014 09:28:48.635149 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-openstack-openstack-networker-hw7gw"] Oct 14 09:28:48 crc kubenswrapper[4870]: I1014 09:28:48.728658 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/f27b6002-352b-4191-8b00-347f35a563bf-ovncontroller-config-0\") pod \"ovn-openstack-openstack-networker-hw7gw\" (UID: \"f27b6002-352b-4191-8b00-347f35a563bf\") " pod="openstack/ovn-openstack-openstack-networker-hw7gw" Oct 14 09:28:48 crc kubenswrapper[4870]: I1014 09:28:48.728751 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c67mm\" (UniqueName: \"kubernetes.io/projected/f27b6002-352b-4191-8b00-347f35a563bf-kube-api-access-c67mm\") pod \"ovn-openstack-openstack-networker-hw7gw\" (UID: \"f27b6002-352b-4191-8b00-347f35a563bf\") " pod="openstack/ovn-openstack-openstack-networker-hw7gw" Oct 14 09:28:48 crc kubenswrapper[4870]: I1014 09:28:48.729430 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f27b6002-352b-4191-8b00-347f35a563bf-ovn-combined-ca-bundle\") pod \"ovn-openstack-openstack-networker-hw7gw\" (UID: \"f27b6002-352b-4191-8b00-347f35a563bf\") " pod="openstack/ovn-openstack-openstack-networker-hw7gw" Oct 14 09:28:48 crc kubenswrapper[4870]: I1014 09:28:48.729536 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f27b6002-352b-4191-8b00-347f35a563bf-ssh-key\") pod \"ovn-openstack-openstack-networker-hw7gw\" (UID: \"f27b6002-352b-4191-8b00-347f35a563bf\") " pod="openstack/ovn-openstack-openstack-networker-hw7gw" Oct 14 09:28:48 crc kubenswrapper[4870]: I1014 09:28:48.729564 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f27b6002-352b-4191-8b00-347f35a563bf-inventory\") pod \"ovn-openstack-openstack-networker-hw7gw\" (UID: \"f27b6002-352b-4191-8b00-347f35a563bf\") " pod="openstack/ovn-openstack-openstack-networker-hw7gw" Oct 14 09:28:48 crc kubenswrapper[4870]: I1014 09:28:48.832237 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f27b6002-352b-4191-8b00-347f35a563bf-ovn-combined-ca-bundle\") pod \"ovn-openstack-openstack-networker-hw7gw\" (UID: \"f27b6002-352b-4191-8b00-347f35a563bf\") " pod="openstack/ovn-openstack-openstack-networker-hw7gw" Oct 14 09:28:48 crc kubenswrapper[4870]: I1014 09:28:48.832302 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f27b6002-352b-4191-8b00-347f35a563bf-ssh-key\") pod \"ovn-openstack-openstack-networker-hw7gw\" (UID: \"f27b6002-352b-4191-8b00-347f35a563bf\") " pod="openstack/ovn-openstack-openstack-networker-hw7gw" Oct 14 09:28:48 crc kubenswrapper[4870]: I1014 09:28:48.832325 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f27b6002-352b-4191-8b00-347f35a563bf-inventory\") pod \"ovn-openstack-openstack-networker-hw7gw\" (UID: \"f27b6002-352b-4191-8b00-347f35a563bf\") " pod="openstack/ovn-openstack-openstack-networker-hw7gw" Oct 14 09:28:48 crc kubenswrapper[4870]: I1014 09:28:48.832380 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/f27b6002-352b-4191-8b00-347f35a563bf-ovncontroller-config-0\") pod \"ovn-openstack-openstack-networker-hw7gw\" (UID: \"f27b6002-352b-4191-8b00-347f35a563bf\") " pod="openstack/ovn-openstack-openstack-networker-hw7gw" Oct 14 09:28:48 crc kubenswrapper[4870]: I1014 09:28:48.832454 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c67mm\" (UniqueName: \"kubernetes.io/projected/f27b6002-352b-4191-8b00-347f35a563bf-kube-api-access-c67mm\") pod \"ovn-openstack-openstack-networker-hw7gw\" (UID: \"f27b6002-352b-4191-8b00-347f35a563bf\") " pod="openstack/ovn-openstack-openstack-networker-hw7gw" Oct 14 09:28:48 crc kubenswrapper[4870]: I1014 09:28:48.833799 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/f27b6002-352b-4191-8b00-347f35a563bf-ovncontroller-config-0\") pod \"ovn-openstack-openstack-networker-hw7gw\" (UID: \"f27b6002-352b-4191-8b00-347f35a563bf\") " pod="openstack/ovn-openstack-openstack-networker-hw7gw" Oct 14 09:28:48 crc kubenswrapper[4870]: I1014 09:28:48.837744 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f27b6002-352b-4191-8b00-347f35a563bf-inventory\") pod \"ovn-openstack-openstack-networker-hw7gw\" (UID: \"f27b6002-352b-4191-8b00-347f35a563bf\") " pod="openstack/ovn-openstack-openstack-networker-hw7gw" Oct 14 09:28:48 crc kubenswrapper[4870]: I1014 09:28:48.838774 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f27b6002-352b-4191-8b00-347f35a563bf-ovn-combined-ca-bundle\") pod \"ovn-openstack-openstack-networker-hw7gw\" (UID: \"f27b6002-352b-4191-8b00-347f35a563bf\") " pod="openstack/ovn-openstack-openstack-networker-hw7gw" Oct 14 09:28:48 crc kubenswrapper[4870]: I1014 09:28:48.848223 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f27b6002-352b-4191-8b00-347f35a563bf-ssh-key\") pod \"ovn-openstack-openstack-networker-hw7gw\" (UID: \"f27b6002-352b-4191-8b00-347f35a563bf\") " pod="openstack/ovn-openstack-openstack-networker-hw7gw" Oct 14 09:28:48 crc kubenswrapper[4870]: I1014 09:28:48.853953 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c67mm\" (UniqueName: \"kubernetes.io/projected/f27b6002-352b-4191-8b00-347f35a563bf-kube-api-access-c67mm\") pod \"ovn-openstack-openstack-networker-hw7gw\" (UID: \"f27b6002-352b-4191-8b00-347f35a563bf\") " pod="openstack/ovn-openstack-openstack-networker-hw7gw" Oct 14 09:28:48 crc kubenswrapper[4870]: I1014 09:28:48.934622 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-openstack-openstack-networker-hw7gw" Oct 14 09:28:49 crc kubenswrapper[4870]: I1014 09:28:49.069763 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-openstack-openstack-cell1-m4fnw" Oct 14 09:28:49 crc kubenswrapper[4870]: I1014 09:28:49.239284 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-thtwm\" (UniqueName: \"kubernetes.io/projected/fe4c4a5a-274a-4adf-929a-46a1057c51ee-kube-api-access-thtwm\") pod \"fe4c4a5a-274a-4adf-929a-46a1057c51ee\" (UID: \"fe4c4a5a-274a-4adf-929a-46a1057c51ee\") " Oct 14 09:28:49 crc kubenswrapper[4870]: I1014 09:28:49.239379 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/fe4c4a5a-274a-4adf-929a-46a1057c51ee-ceph\") pod \"fe4c4a5a-274a-4adf-929a-46a1057c51ee\" (UID: \"fe4c4a5a-274a-4adf-929a-46a1057c51ee\") " Oct 14 09:28:49 crc kubenswrapper[4870]: I1014 09:28:49.239567 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fe4c4a5a-274a-4adf-929a-46a1057c51ee-inventory\") pod \"fe4c4a5a-274a-4adf-929a-46a1057c51ee\" (UID: \"fe4c4a5a-274a-4adf-929a-46a1057c51ee\") " Oct 14 09:28:49 crc kubenswrapper[4870]: I1014 09:28:49.239677 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fe4c4a5a-274a-4adf-929a-46a1057c51ee-ssh-key\") pod \"fe4c4a5a-274a-4adf-929a-46a1057c51ee\" (UID: \"fe4c4a5a-274a-4adf-929a-46a1057c51ee\") " Oct 14 09:28:49 crc kubenswrapper[4870]: I1014 09:28:49.251789 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fe4c4a5a-274a-4adf-929a-46a1057c51ee-kube-api-access-thtwm" (OuterVolumeSpecName: "kube-api-access-thtwm") pod "fe4c4a5a-274a-4adf-929a-46a1057c51ee" (UID: "fe4c4a5a-274a-4adf-929a-46a1057c51ee"). InnerVolumeSpecName "kube-api-access-thtwm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:28:49 crc kubenswrapper[4870]: I1014 09:28:49.259604 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fe4c4a5a-274a-4adf-929a-46a1057c51ee-ceph" (OuterVolumeSpecName: "ceph") pod "fe4c4a5a-274a-4adf-929a-46a1057c51ee" (UID: "fe4c4a5a-274a-4adf-929a-46a1057c51ee"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:28:49 crc kubenswrapper[4870]: I1014 09:28:49.275501 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fe4c4a5a-274a-4adf-929a-46a1057c51ee-inventory" (OuterVolumeSpecName: "inventory") pod "fe4c4a5a-274a-4adf-929a-46a1057c51ee" (UID: "fe4c4a5a-274a-4adf-929a-46a1057c51ee"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:28:49 crc kubenswrapper[4870]: I1014 09:28:49.284578 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fe4c4a5a-274a-4adf-929a-46a1057c51ee-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "fe4c4a5a-274a-4adf-929a-46a1057c51ee" (UID: "fe4c4a5a-274a-4adf-929a-46a1057c51ee"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:28:49 crc kubenswrapper[4870]: I1014 09:28:49.329376 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-openstack-openstack-networker-hw7gw"] Oct 14 09:28:49 crc kubenswrapper[4870]: I1014 09:28:49.343058 4870 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fe4c4a5a-274a-4adf-929a-46a1057c51ee-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 14 09:28:49 crc kubenswrapper[4870]: I1014 09:28:49.343424 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-thtwm\" (UniqueName: \"kubernetes.io/projected/fe4c4a5a-274a-4adf-929a-46a1057c51ee-kube-api-access-thtwm\") on node \"crc\" DevicePath \"\"" Oct 14 09:28:49 crc kubenswrapper[4870]: I1014 09:28:49.343473 4870 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/fe4c4a5a-274a-4adf-929a-46a1057c51ee-ceph\") on node \"crc\" DevicePath \"\"" Oct 14 09:28:49 crc kubenswrapper[4870]: I1014 09:28:49.343491 4870 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fe4c4a5a-274a-4adf-929a-46a1057c51ee-inventory\") on node \"crc\" DevicePath \"\"" Oct 14 09:28:49 crc kubenswrapper[4870]: I1014 09:28:49.552571 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-openstack-openstack-cell1-m4fnw" Oct 14 09:28:49 crc kubenswrapper[4870]: I1014 09:28:49.552563 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-openstack-openstack-cell1-m4fnw" event={"ID":"fe4c4a5a-274a-4adf-929a-46a1057c51ee","Type":"ContainerDied","Data":"cc58b0932a7c3d17b32926ff7ad4fe8f992ff5500d03438604e3b945d5aeb343"} Oct 14 09:28:49 crc kubenswrapper[4870]: I1014 09:28:49.552733 4870 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cc58b0932a7c3d17b32926ff7ad4fe8f992ff5500d03438604e3b945d5aeb343" Oct 14 09:28:49 crc kubenswrapper[4870]: I1014 09:28:49.556403 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-openstack-openstack-networker-hw7gw" event={"ID":"f27b6002-352b-4191-8b00-347f35a563bf","Type":"ContainerStarted","Data":"bc5d5fdd064099fbf8aec679928d23bbbb18bb242af736e665ef383ea8b68275"} Oct 14 09:28:49 crc kubenswrapper[4870]: I1014 09:28:49.631245 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-openstack-openstack-cell1-cx8gp"] Oct 14 09:28:49 crc kubenswrapper[4870]: E1014 09:28:49.631873 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fe4c4a5a-274a-4adf-929a-46a1057c51ee" containerName="run-os-openstack-openstack-cell1" Oct 14 09:28:49 crc kubenswrapper[4870]: I1014 09:28:49.631895 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="fe4c4a5a-274a-4adf-929a-46a1057c51ee" containerName="run-os-openstack-openstack-cell1" Oct 14 09:28:49 crc kubenswrapper[4870]: I1014 09:28:49.632185 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="fe4c4a5a-274a-4adf-929a-46a1057c51ee" containerName="run-os-openstack-openstack-cell1" Oct 14 09:28:49 crc kubenswrapper[4870]: I1014 09:28:49.633131 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-openstack-openstack-cell1-cx8gp" Oct 14 09:28:49 crc kubenswrapper[4870]: I1014 09:28:49.635309 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 14 09:28:49 crc kubenswrapper[4870]: I1014 09:28:49.635315 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-wrz7w" Oct 14 09:28:49 crc kubenswrapper[4870]: I1014 09:28:49.657990 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-openstack-openstack-cell1-cx8gp"] Oct 14 09:28:49 crc kubenswrapper[4870]: I1014 09:28:49.756541 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f7861d1b-2199-407a-8abb-594c42ec2544-inventory\") pod \"reboot-os-openstack-openstack-cell1-cx8gp\" (UID: \"f7861d1b-2199-407a-8abb-594c42ec2544\") " pod="openstack/reboot-os-openstack-openstack-cell1-cx8gp" Oct 14 09:28:49 crc kubenswrapper[4870]: I1014 09:28:49.757057 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f7861d1b-2199-407a-8abb-594c42ec2544-ceph\") pod \"reboot-os-openstack-openstack-cell1-cx8gp\" (UID: \"f7861d1b-2199-407a-8abb-594c42ec2544\") " pod="openstack/reboot-os-openstack-openstack-cell1-cx8gp" Oct 14 09:28:49 crc kubenswrapper[4870]: I1014 09:28:49.757209 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f7861d1b-2199-407a-8abb-594c42ec2544-ssh-key\") pod \"reboot-os-openstack-openstack-cell1-cx8gp\" (UID: \"f7861d1b-2199-407a-8abb-594c42ec2544\") " pod="openstack/reboot-os-openstack-openstack-cell1-cx8gp" Oct 14 09:28:49 crc kubenswrapper[4870]: I1014 09:28:49.757862 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2vfnj\" (UniqueName: \"kubernetes.io/projected/f7861d1b-2199-407a-8abb-594c42ec2544-kube-api-access-2vfnj\") pod \"reboot-os-openstack-openstack-cell1-cx8gp\" (UID: \"f7861d1b-2199-407a-8abb-594c42ec2544\") " pod="openstack/reboot-os-openstack-openstack-cell1-cx8gp" Oct 14 09:28:49 crc kubenswrapper[4870]: E1014 09:28:49.807149 4870 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfe4c4a5a_274a_4adf_929a_46a1057c51ee.slice/crio-cc58b0932a7c3d17b32926ff7ad4fe8f992ff5500d03438604e3b945d5aeb343\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfe4c4a5a_274a_4adf_929a_46a1057c51ee.slice\": RecentStats: unable to find data in memory cache]" Oct 14 09:28:49 crc kubenswrapper[4870]: I1014 09:28:49.859451 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2vfnj\" (UniqueName: \"kubernetes.io/projected/f7861d1b-2199-407a-8abb-594c42ec2544-kube-api-access-2vfnj\") pod \"reboot-os-openstack-openstack-cell1-cx8gp\" (UID: \"f7861d1b-2199-407a-8abb-594c42ec2544\") " pod="openstack/reboot-os-openstack-openstack-cell1-cx8gp" Oct 14 09:28:49 crc kubenswrapper[4870]: I1014 09:28:49.859719 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f7861d1b-2199-407a-8abb-594c42ec2544-inventory\") pod \"reboot-os-openstack-openstack-cell1-cx8gp\" (UID: \"f7861d1b-2199-407a-8abb-594c42ec2544\") " pod="openstack/reboot-os-openstack-openstack-cell1-cx8gp" Oct 14 09:28:49 crc kubenswrapper[4870]: I1014 09:28:49.859804 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f7861d1b-2199-407a-8abb-594c42ec2544-ceph\") pod \"reboot-os-openstack-openstack-cell1-cx8gp\" (UID: \"f7861d1b-2199-407a-8abb-594c42ec2544\") " pod="openstack/reboot-os-openstack-openstack-cell1-cx8gp" Oct 14 09:28:49 crc kubenswrapper[4870]: I1014 09:28:49.859834 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f7861d1b-2199-407a-8abb-594c42ec2544-ssh-key\") pod \"reboot-os-openstack-openstack-cell1-cx8gp\" (UID: \"f7861d1b-2199-407a-8abb-594c42ec2544\") " pod="openstack/reboot-os-openstack-openstack-cell1-cx8gp" Oct 14 09:28:49 crc kubenswrapper[4870]: I1014 09:28:49.866216 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f7861d1b-2199-407a-8abb-594c42ec2544-inventory\") pod \"reboot-os-openstack-openstack-cell1-cx8gp\" (UID: \"f7861d1b-2199-407a-8abb-594c42ec2544\") " pod="openstack/reboot-os-openstack-openstack-cell1-cx8gp" Oct 14 09:28:49 crc kubenswrapper[4870]: I1014 09:28:49.867925 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f7861d1b-2199-407a-8abb-594c42ec2544-ceph\") pod \"reboot-os-openstack-openstack-cell1-cx8gp\" (UID: \"f7861d1b-2199-407a-8abb-594c42ec2544\") " pod="openstack/reboot-os-openstack-openstack-cell1-cx8gp" Oct 14 09:28:49 crc kubenswrapper[4870]: I1014 09:28:49.869589 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f7861d1b-2199-407a-8abb-594c42ec2544-ssh-key\") pod \"reboot-os-openstack-openstack-cell1-cx8gp\" (UID: \"f7861d1b-2199-407a-8abb-594c42ec2544\") " pod="openstack/reboot-os-openstack-openstack-cell1-cx8gp" Oct 14 09:28:49 crc kubenswrapper[4870]: I1014 09:28:49.880203 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2vfnj\" (UniqueName: \"kubernetes.io/projected/f7861d1b-2199-407a-8abb-594c42ec2544-kube-api-access-2vfnj\") pod \"reboot-os-openstack-openstack-cell1-cx8gp\" (UID: \"f7861d1b-2199-407a-8abb-594c42ec2544\") " pod="openstack/reboot-os-openstack-openstack-cell1-cx8gp" Oct 14 09:28:49 crc kubenswrapper[4870]: I1014 09:28:49.958115 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-openstack-openstack-cell1-cx8gp" Oct 14 09:28:50 crc kubenswrapper[4870]: I1014 09:28:50.536415 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-openstack-openstack-cell1-cx8gp"] Oct 14 09:28:50 crc kubenswrapper[4870]: W1014 09:28:50.544502 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf7861d1b_2199_407a_8abb_594c42ec2544.slice/crio-b03c12589257bebff4b39419a58b94c128089020e8b0f78fdaeff1f9dd29a42f WatchSource:0}: Error finding container b03c12589257bebff4b39419a58b94c128089020e8b0f78fdaeff1f9dd29a42f: Status 404 returned error can't find the container with id b03c12589257bebff4b39419a58b94c128089020e8b0f78fdaeff1f9dd29a42f Oct 14 09:28:50 crc kubenswrapper[4870]: I1014 09:28:50.566946 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-openstack-openstack-networker-hw7gw" event={"ID":"f27b6002-352b-4191-8b00-347f35a563bf","Type":"ContainerStarted","Data":"719983fe94249f44a37d90c5dc7f1043c46541671e7d7b4f0c197a89bdcb8c70"} Oct 14 09:28:50 crc kubenswrapper[4870]: I1014 09:28:50.568957 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-openstack-openstack-cell1-cx8gp" event={"ID":"f7861d1b-2199-407a-8abb-594c42ec2544","Type":"ContainerStarted","Data":"b03c12589257bebff4b39419a58b94c128089020e8b0f78fdaeff1f9dd29a42f"} Oct 14 09:28:50 crc kubenswrapper[4870]: I1014 09:28:50.600501 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-openstack-openstack-networker-hw7gw" podStartSLOduration=2.141894346 podStartE2EDuration="2.600473127s" podCreationTimestamp="2025-10-14 09:28:48 +0000 UTC" firstStartedPulling="2025-10-14 09:28:49.334838423 +0000 UTC m=+8865.032198804" lastFinishedPulling="2025-10-14 09:28:49.793417184 +0000 UTC m=+8865.490777585" observedRunningTime="2025-10-14 09:28:50.58313183 +0000 UTC m=+8866.280492201" watchObservedRunningTime="2025-10-14 09:28:50.600473127 +0000 UTC m=+8866.297833498" Oct 14 09:28:51 crc kubenswrapper[4870]: I1014 09:28:51.586584 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-openstack-openstack-cell1-cx8gp" event={"ID":"f7861d1b-2199-407a-8abb-594c42ec2544","Type":"ContainerStarted","Data":"f2686be2c39bcb3605c96b4c2815165b29a8b2804448adf0ca2d1020e685ea45"} Oct 14 09:28:51 crc kubenswrapper[4870]: I1014 09:28:51.624386 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-openstack-openstack-cell1-cx8gp" podStartSLOduration=2.163813185 podStartE2EDuration="2.624360424s" podCreationTimestamp="2025-10-14 09:28:49 +0000 UTC" firstStartedPulling="2025-10-14 09:28:50.547300009 +0000 UTC m=+8866.244660380" lastFinishedPulling="2025-10-14 09:28:51.007847228 +0000 UTC m=+8866.705207619" observedRunningTime="2025-10-14 09:28:51.617772272 +0000 UTC m=+8867.315132663" watchObservedRunningTime="2025-10-14 09:28:51.624360424 +0000 UTC m=+8867.321720835" Oct 14 09:28:53 crc kubenswrapper[4870]: I1014 09:28:53.951035 4870 patch_prober.go:28] interesting pod/machine-config-daemon-7tvc7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 09:28:53 crc kubenswrapper[4870]: I1014 09:28:53.952832 4870 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 09:29:06 crc kubenswrapper[4870]: I1014 09:29:06.804608 4870 generic.go:334] "Generic (PLEG): container finished" podID="f7861d1b-2199-407a-8abb-594c42ec2544" containerID="f2686be2c39bcb3605c96b4c2815165b29a8b2804448adf0ca2d1020e685ea45" exitCode=0 Oct 14 09:29:06 crc kubenswrapper[4870]: I1014 09:29:06.804728 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-openstack-openstack-cell1-cx8gp" event={"ID":"f7861d1b-2199-407a-8abb-594c42ec2544","Type":"ContainerDied","Data":"f2686be2c39bcb3605c96b4c2815165b29a8b2804448adf0ca2d1020e685ea45"} Oct 14 09:29:08 crc kubenswrapper[4870]: I1014 09:29:08.421990 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-openstack-openstack-cell1-cx8gp" Oct 14 09:29:08 crc kubenswrapper[4870]: I1014 09:29:08.527226 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f7861d1b-2199-407a-8abb-594c42ec2544-ssh-key\") pod \"f7861d1b-2199-407a-8abb-594c42ec2544\" (UID: \"f7861d1b-2199-407a-8abb-594c42ec2544\") " Oct 14 09:29:08 crc kubenswrapper[4870]: I1014 09:29:08.527326 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f7861d1b-2199-407a-8abb-594c42ec2544-inventory\") pod \"f7861d1b-2199-407a-8abb-594c42ec2544\" (UID: \"f7861d1b-2199-407a-8abb-594c42ec2544\") " Oct 14 09:29:08 crc kubenswrapper[4870]: I1014 09:29:08.527373 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2vfnj\" (UniqueName: \"kubernetes.io/projected/f7861d1b-2199-407a-8abb-594c42ec2544-kube-api-access-2vfnj\") pod \"f7861d1b-2199-407a-8abb-594c42ec2544\" (UID: \"f7861d1b-2199-407a-8abb-594c42ec2544\") " Oct 14 09:29:08 crc kubenswrapper[4870]: I1014 09:29:08.527686 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f7861d1b-2199-407a-8abb-594c42ec2544-ceph\") pod \"f7861d1b-2199-407a-8abb-594c42ec2544\" (UID: \"f7861d1b-2199-407a-8abb-594c42ec2544\") " Oct 14 09:29:08 crc kubenswrapper[4870]: I1014 09:29:08.534678 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7861d1b-2199-407a-8abb-594c42ec2544-ceph" (OuterVolumeSpecName: "ceph") pod "f7861d1b-2199-407a-8abb-594c42ec2544" (UID: "f7861d1b-2199-407a-8abb-594c42ec2544"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:29:08 crc kubenswrapper[4870]: I1014 09:29:08.534964 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f7861d1b-2199-407a-8abb-594c42ec2544-kube-api-access-2vfnj" (OuterVolumeSpecName: "kube-api-access-2vfnj") pod "f7861d1b-2199-407a-8abb-594c42ec2544" (UID: "f7861d1b-2199-407a-8abb-594c42ec2544"). InnerVolumeSpecName "kube-api-access-2vfnj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:29:08 crc kubenswrapper[4870]: I1014 09:29:08.564326 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7861d1b-2199-407a-8abb-594c42ec2544-inventory" (OuterVolumeSpecName: "inventory") pod "f7861d1b-2199-407a-8abb-594c42ec2544" (UID: "f7861d1b-2199-407a-8abb-594c42ec2544"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:29:08 crc kubenswrapper[4870]: I1014 09:29:08.568391 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7861d1b-2199-407a-8abb-594c42ec2544-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "f7861d1b-2199-407a-8abb-594c42ec2544" (UID: "f7861d1b-2199-407a-8abb-594c42ec2544"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:29:08 crc kubenswrapper[4870]: I1014 09:29:08.631746 4870 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f7861d1b-2199-407a-8abb-594c42ec2544-ceph\") on node \"crc\" DevicePath \"\"" Oct 14 09:29:08 crc kubenswrapper[4870]: I1014 09:29:08.631806 4870 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f7861d1b-2199-407a-8abb-594c42ec2544-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 14 09:29:08 crc kubenswrapper[4870]: I1014 09:29:08.631827 4870 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f7861d1b-2199-407a-8abb-594c42ec2544-inventory\") on node \"crc\" DevicePath \"\"" Oct 14 09:29:08 crc kubenswrapper[4870]: I1014 09:29:08.631845 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2vfnj\" (UniqueName: \"kubernetes.io/projected/f7861d1b-2199-407a-8abb-594c42ec2544-kube-api-access-2vfnj\") on node \"crc\" DevicePath \"\"" Oct 14 09:29:08 crc kubenswrapper[4870]: I1014 09:29:08.835309 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-openstack-openstack-cell1-cx8gp" event={"ID":"f7861d1b-2199-407a-8abb-594c42ec2544","Type":"ContainerDied","Data":"b03c12589257bebff4b39419a58b94c128089020e8b0f78fdaeff1f9dd29a42f"} Oct 14 09:29:08 crc kubenswrapper[4870]: I1014 09:29:08.835732 4870 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b03c12589257bebff4b39419a58b94c128089020e8b0f78fdaeff1f9dd29a42f" Oct 14 09:29:08 crc kubenswrapper[4870]: I1014 09:29:08.835504 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-openstack-openstack-cell1-cx8gp" Oct 14 09:29:08 crc kubenswrapper[4870]: I1014 09:29:08.957429 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-openstack-openstack-cell1-q4zrw"] Oct 14 09:29:08 crc kubenswrapper[4870]: E1014 09:29:08.957889 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7861d1b-2199-407a-8abb-594c42ec2544" containerName="reboot-os-openstack-openstack-cell1" Oct 14 09:29:08 crc kubenswrapper[4870]: I1014 09:29:08.957905 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7861d1b-2199-407a-8abb-594c42ec2544" containerName="reboot-os-openstack-openstack-cell1" Oct 14 09:29:08 crc kubenswrapper[4870]: I1014 09:29:08.958125 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="f7861d1b-2199-407a-8abb-594c42ec2544" containerName="reboot-os-openstack-openstack-cell1" Oct 14 09:29:08 crc kubenswrapper[4870]: I1014 09:29:08.958846 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-openstack-openstack-cell1-q4zrw" Oct 14 09:29:08 crc kubenswrapper[4870]: I1014 09:29:08.964765 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-wrz7w" Oct 14 09:29:08 crc kubenswrapper[4870]: I1014 09:29:08.966154 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 14 09:29:08 crc kubenswrapper[4870]: I1014 09:29:08.981028 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-openstack-openstack-cell1-q4zrw"] Oct 14 09:29:09 crc kubenswrapper[4870]: I1014 09:29:09.040838 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a90e5cfd-b6bb-45a3-8b31-cfa84c3ebf93-libvirt-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-q4zrw\" (UID: \"a90e5cfd-b6bb-45a3-8b31-cfa84c3ebf93\") " pod="openstack/install-certs-openstack-openstack-cell1-q4zrw" Oct 14 09:29:09 crc kubenswrapper[4870]: I1014 09:29:09.040978 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a90e5cfd-b6bb-45a3-8b31-cfa84c3ebf93-neutron-dhcp-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-q4zrw\" (UID: \"a90e5cfd-b6bb-45a3-8b31-cfa84c3ebf93\") " pod="openstack/install-certs-openstack-openstack-cell1-q4zrw" Oct 14 09:29:09 crc kubenswrapper[4870]: I1014 09:29:09.041016 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a90e5cfd-b6bb-45a3-8b31-cfa84c3ebf93-ssh-key\") pod \"install-certs-openstack-openstack-cell1-q4zrw\" (UID: \"a90e5cfd-b6bb-45a3-8b31-cfa84c3ebf93\") " pod="openstack/install-certs-openstack-openstack-cell1-q4zrw" Oct 14 09:29:09 crc kubenswrapper[4870]: I1014 09:29:09.041050 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a90e5cfd-b6bb-45a3-8b31-cfa84c3ebf93-bootstrap-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-q4zrw\" (UID: \"a90e5cfd-b6bb-45a3-8b31-cfa84c3ebf93\") " pod="openstack/install-certs-openstack-openstack-cell1-q4zrw" Oct 14 09:29:09 crc kubenswrapper[4870]: I1014 09:29:09.041130 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a90e5cfd-b6bb-45a3-8b31-cfa84c3ebf93-inventory\") pod \"install-certs-openstack-openstack-cell1-q4zrw\" (UID: \"a90e5cfd-b6bb-45a3-8b31-cfa84c3ebf93\") " pod="openstack/install-certs-openstack-openstack-cell1-q4zrw" Oct 14 09:29:09 crc kubenswrapper[4870]: I1014 09:29:09.041194 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/a90e5cfd-b6bb-45a3-8b31-cfa84c3ebf93-ceph\") pod \"install-certs-openstack-openstack-cell1-q4zrw\" (UID: \"a90e5cfd-b6bb-45a3-8b31-cfa84c3ebf93\") " pod="openstack/install-certs-openstack-openstack-cell1-q4zrw" Oct 14 09:29:09 crc kubenswrapper[4870]: I1014 09:29:09.041297 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a90e5cfd-b6bb-45a3-8b31-cfa84c3ebf93-neutron-sriov-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-q4zrw\" (UID: \"a90e5cfd-b6bb-45a3-8b31-cfa84c3ebf93\") " pod="openstack/install-certs-openstack-openstack-cell1-q4zrw" Oct 14 09:29:09 crc kubenswrapper[4870]: I1014 09:29:09.041324 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a90e5cfd-b6bb-45a3-8b31-cfa84c3ebf93-ovn-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-q4zrw\" (UID: \"a90e5cfd-b6bb-45a3-8b31-cfa84c3ebf93\") " pod="openstack/install-certs-openstack-openstack-cell1-q4zrw" Oct 14 09:29:09 crc kubenswrapper[4870]: I1014 09:29:09.041361 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a90e5cfd-b6bb-45a3-8b31-cfa84c3ebf93-nova-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-q4zrw\" (UID: \"a90e5cfd-b6bb-45a3-8b31-cfa84c3ebf93\") " pod="openstack/install-certs-openstack-openstack-cell1-q4zrw" Oct 14 09:29:09 crc kubenswrapper[4870]: I1014 09:29:09.041388 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a90e5cfd-b6bb-45a3-8b31-cfa84c3ebf93-neutron-metadata-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-q4zrw\" (UID: \"a90e5cfd-b6bb-45a3-8b31-cfa84c3ebf93\") " pod="openstack/install-certs-openstack-openstack-cell1-q4zrw" Oct 14 09:29:09 crc kubenswrapper[4870]: I1014 09:29:09.041433 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a90e5cfd-b6bb-45a3-8b31-cfa84c3ebf93-telemetry-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-q4zrw\" (UID: \"a90e5cfd-b6bb-45a3-8b31-cfa84c3ebf93\") " pod="openstack/install-certs-openstack-openstack-cell1-q4zrw" Oct 14 09:29:09 crc kubenswrapper[4870]: I1014 09:29:09.041484 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-894vw\" (UniqueName: \"kubernetes.io/projected/a90e5cfd-b6bb-45a3-8b31-cfa84c3ebf93-kube-api-access-894vw\") pod \"install-certs-openstack-openstack-cell1-q4zrw\" (UID: \"a90e5cfd-b6bb-45a3-8b31-cfa84c3ebf93\") " pod="openstack/install-certs-openstack-openstack-cell1-q4zrw" Oct 14 09:29:09 crc kubenswrapper[4870]: I1014 09:29:09.143614 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a90e5cfd-b6bb-45a3-8b31-cfa84c3ebf93-libvirt-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-q4zrw\" (UID: \"a90e5cfd-b6bb-45a3-8b31-cfa84c3ebf93\") " pod="openstack/install-certs-openstack-openstack-cell1-q4zrw" Oct 14 09:29:09 crc kubenswrapper[4870]: I1014 09:29:09.144727 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a90e5cfd-b6bb-45a3-8b31-cfa84c3ebf93-neutron-dhcp-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-q4zrw\" (UID: \"a90e5cfd-b6bb-45a3-8b31-cfa84c3ebf93\") " pod="openstack/install-certs-openstack-openstack-cell1-q4zrw" Oct 14 09:29:09 crc kubenswrapper[4870]: I1014 09:29:09.145263 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a90e5cfd-b6bb-45a3-8b31-cfa84c3ebf93-ssh-key\") pod \"install-certs-openstack-openstack-cell1-q4zrw\" (UID: \"a90e5cfd-b6bb-45a3-8b31-cfa84c3ebf93\") " pod="openstack/install-certs-openstack-openstack-cell1-q4zrw" Oct 14 09:29:09 crc kubenswrapper[4870]: I1014 09:29:09.145322 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a90e5cfd-b6bb-45a3-8b31-cfa84c3ebf93-bootstrap-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-q4zrw\" (UID: \"a90e5cfd-b6bb-45a3-8b31-cfa84c3ebf93\") " pod="openstack/install-certs-openstack-openstack-cell1-q4zrw" Oct 14 09:29:09 crc kubenswrapper[4870]: I1014 09:29:09.145387 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a90e5cfd-b6bb-45a3-8b31-cfa84c3ebf93-inventory\") pod \"install-certs-openstack-openstack-cell1-q4zrw\" (UID: \"a90e5cfd-b6bb-45a3-8b31-cfa84c3ebf93\") " pod="openstack/install-certs-openstack-openstack-cell1-q4zrw" Oct 14 09:29:09 crc kubenswrapper[4870]: I1014 09:29:09.145546 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/a90e5cfd-b6bb-45a3-8b31-cfa84c3ebf93-ceph\") pod \"install-certs-openstack-openstack-cell1-q4zrw\" (UID: \"a90e5cfd-b6bb-45a3-8b31-cfa84c3ebf93\") " pod="openstack/install-certs-openstack-openstack-cell1-q4zrw" Oct 14 09:29:09 crc kubenswrapper[4870]: I1014 09:29:09.145631 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a90e5cfd-b6bb-45a3-8b31-cfa84c3ebf93-neutron-sriov-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-q4zrw\" (UID: \"a90e5cfd-b6bb-45a3-8b31-cfa84c3ebf93\") " pod="openstack/install-certs-openstack-openstack-cell1-q4zrw" Oct 14 09:29:09 crc kubenswrapper[4870]: I1014 09:29:09.145681 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a90e5cfd-b6bb-45a3-8b31-cfa84c3ebf93-ovn-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-q4zrw\" (UID: \"a90e5cfd-b6bb-45a3-8b31-cfa84c3ebf93\") " pod="openstack/install-certs-openstack-openstack-cell1-q4zrw" Oct 14 09:29:09 crc kubenswrapper[4870]: I1014 09:29:09.145735 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a90e5cfd-b6bb-45a3-8b31-cfa84c3ebf93-nova-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-q4zrw\" (UID: \"a90e5cfd-b6bb-45a3-8b31-cfa84c3ebf93\") " pod="openstack/install-certs-openstack-openstack-cell1-q4zrw" Oct 14 09:29:09 crc kubenswrapper[4870]: I1014 09:29:09.145770 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a90e5cfd-b6bb-45a3-8b31-cfa84c3ebf93-neutron-metadata-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-q4zrw\" (UID: \"a90e5cfd-b6bb-45a3-8b31-cfa84c3ebf93\") " pod="openstack/install-certs-openstack-openstack-cell1-q4zrw" Oct 14 09:29:09 crc kubenswrapper[4870]: I1014 09:29:09.145806 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a90e5cfd-b6bb-45a3-8b31-cfa84c3ebf93-telemetry-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-q4zrw\" (UID: \"a90e5cfd-b6bb-45a3-8b31-cfa84c3ebf93\") " pod="openstack/install-certs-openstack-openstack-cell1-q4zrw" Oct 14 09:29:09 crc kubenswrapper[4870]: I1014 09:29:09.145844 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-894vw\" (UniqueName: \"kubernetes.io/projected/a90e5cfd-b6bb-45a3-8b31-cfa84c3ebf93-kube-api-access-894vw\") pod \"install-certs-openstack-openstack-cell1-q4zrw\" (UID: \"a90e5cfd-b6bb-45a3-8b31-cfa84c3ebf93\") " pod="openstack/install-certs-openstack-openstack-cell1-q4zrw" Oct 14 09:29:09 crc kubenswrapper[4870]: I1014 09:29:09.150241 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a90e5cfd-b6bb-45a3-8b31-cfa84c3ebf93-neutron-metadata-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-q4zrw\" (UID: \"a90e5cfd-b6bb-45a3-8b31-cfa84c3ebf93\") " pod="openstack/install-certs-openstack-openstack-cell1-q4zrw" Oct 14 09:29:09 crc kubenswrapper[4870]: I1014 09:29:09.150761 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a90e5cfd-b6bb-45a3-8b31-cfa84c3ebf93-libvirt-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-q4zrw\" (UID: \"a90e5cfd-b6bb-45a3-8b31-cfa84c3ebf93\") " pod="openstack/install-certs-openstack-openstack-cell1-q4zrw" Oct 14 09:29:09 crc kubenswrapper[4870]: I1014 09:29:09.151329 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a90e5cfd-b6bb-45a3-8b31-cfa84c3ebf93-telemetry-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-q4zrw\" (UID: \"a90e5cfd-b6bb-45a3-8b31-cfa84c3ebf93\") " pod="openstack/install-certs-openstack-openstack-cell1-q4zrw" Oct 14 09:29:09 crc kubenswrapper[4870]: I1014 09:29:09.151507 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a90e5cfd-b6bb-45a3-8b31-cfa84c3ebf93-ssh-key\") pod \"install-certs-openstack-openstack-cell1-q4zrw\" (UID: \"a90e5cfd-b6bb-45a3-8b31-cfa84c3ebf93\") " pod="openstack/install-certs-openstack-openstack-cell1-q4zrw" Oct 14 09:29:09 crc kubenswrapper[4870]: I1014 09:29:09.151553 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a90e5cfd-b6bb-45a3-8b31-cfa84c3ebf93-nova-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-q4zrw\" (UID: \"a90e5cfd-b6bb-45a3-8b31-cfa84c3ebf93\") " pod="openstack/install-certs-openstack-openstack-cell1-q4zrw" Oct 14 09:29:09 crc kubenswrapper[4870]: I1014 09:29:09.152416 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a90e5cfd-b6bb-45a3-8b31-cfa84c3ebf93-neutron-sriov-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-q4zrw\" (UID: \"a90e5cfd-b6bb-45a3-8b31-cfa84c3ebf93\") " pod="openstack/install-certs-openstack-openstack-cell1-q4zrw" Oct 14 09:29:09 crc kubenswrapper[4870]: I1014 09:29:09.154062 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a90e5cfd-b6bb-45a3-8b31-cfa84c3ebf93-ovn-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-q4zrw\" (UID: \"a90e5cfd-b6bb-45a3-8b31-cfa84c3ebf93\") " pod="openstack/install-certs-openstack-openstack-cell1-q4zrw" Oct 14 09:29:09 crc kubenswrapper[4870]: I1014 09:29:09.159423 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a90e5cfd-b6bb-45a3-8b31-cfa84c3ebf93-bootstrap-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-q4zrw\" (UID: \"a90e5cfd-b6bb-45a3-8b31-cfa84c3ebf93\") " pod="openstack/install-certs-openstack-openstack-cell1-q4zrw" Oct 14 09:29:09 crc kubenswrapper[4870]: I1014 09:29:09.162690 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a90e5cfd-b6bb-45a3-8b31-cfa84c3ebf93-inventory\") pod \"install-certs-openstack-openstack-cell1-q4zrw\" (UID: \"a90e5cfd-b6bb-45a3-8b31-cfa84c3ebf93\") " pod="openstack/install-certs-openstack-openstack-cell1-q4zrw" Oct 14 09:29:09 crc kubenswrapper[4870]: I1014 09:29:09.165050 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a90e5cfd-b6bb-45a3-8b31-cfa84c3ebf93-neutron-dhcp-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-q4zrw\" (UID: \"a90e5cfd-b6bb-45a3-8b31-cfa84c3ebf93\") " pod="openstack/install-certs-openstack-openstack-cell1-q4zrw" Oct 14 09:29:09 crc kubenswrapper[4870]: I1014 09:29:09.152737 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/a90e5cfd-b6bb-45a3-8b31-cfa84c3ebf93-ceph\") pod \"install-certs-openstack-openstack-cell1-q4zrw\" (UID: \"a90e5cfd-b6bb-45a3-8b31-cfa84c3ebf93\") " pod="openstack/install-certs-openstack-openstack-cell1-q4zrw" Oct 14 09:29:09 crc kubenswrapper[4870]: I1014 09:29:09.173336 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-894vw\" (UniqueName: \"kubernetes.io/projected/a90e5cfd-b6bb-45a3-8b31-cfa84c3ebf93-kube-api-access-894vw\") pod \"install-certs-openstack-openstack-cell1-q4zrw\" (UID: \"a90e5cfd-b6bb-45a3-8b31-cfa84c3ebf93\") " pod="openstack/install-certs-openstack-openstack-cell1-q4zrw" Oct 14 09:29:09 crc kubenswrapper[4870]: I1014 09:29:09.276725 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-openstack-openstack-cell1-q4zrw" Oct 14 09:29:09 crc kubenswrapper[4870]: I1014 09:29:09.839554 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-openstack-openstack-cell1-q4zrw"] Oct 14 09:29:09 crc kubenswrapper[4870]: W1014 09:29:09.846446 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda90e5cfd_b6bb_45a3_8b31_cfa84c3ebf93.slice/crio-b3fa518a94ee3abcbb6e29998f431a33f28aa32033dc987e657a1accb73b1415 WatchSource:0}: Error finding container b3fa518a94ee3abcbb6e29998f431a33f28aa32033dc987e657a1accb73b1415: Status 404 returned error can't find the container with id b3fa518a94ee3abcbb6e29998f431a33f28aa32033dc987e657a1accb73b1415 Oct 14 09:29:10 crc kubenswrapper[4870]: I1014 09:29:10.865646 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-openstack-openstack-cell1-q4zrw" event={"ID":"a90e5cfd-b6bb-45a3-8b31-cfa84c3ebf93","Type":"ContainerStarted","Data":"5cf83ec245634e137d8e0c7599a907899702bd4bfe58d5b28a23b596ad714933"} Oct 14 09:29:10 crc kubenswrapper[4870]: I1014 09:29:10.866268 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-openstack-openstack-cell1-q4zrw" event={"ID":"a90e5cfd-b6bb-45a3-8b31-cfa84c3ebf93","Type":"ContainerStarted","Data":"b3fa518a94ee3abcbb6e29998f431a33f28aa32033dc987e657a1accb73b1415"} Oct 14 09:29:10 crc kubenswrapper[4870]: I1014 09:29:10.902182 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-openstack-openstack-cell1-q4zrw" podStartSLOduration=2.455864204 podStartE2EDuration="2.902155083s" podCreationTimestamp="2025-10-14 09:29:08 +0000 UTC" firstStartedPulling="2025-10-14 09:29:09.851068476 +0000 UTC m=+8885.548428847" lastFinishedPulling="2025-10-14 09:29:10.297359335 +0000 UTC m=+8885.994719726" observedRunningTime="2025-10-14 09:29:10.891259205 +0000 UTC m=+8886.588619606" watchObservedRunningTime="2025-10-14 09:29:10.902155083 +0000 UTC m=+8886.599515494" Oct 14 09:29:23 crc kubenswrapper[4870]: I1014 09:29:23.951275 4870 patch_prober.go:28] interesting pod/machine-config-daemon-7tvc7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 09:29:23 crc kubenswrapper[4870]: I1014 09:29:23.951945 4870 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 09:29:23 crc kubenswrapper[4870]: I1014 09:29:23.952022 4870 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" Oct 14 09:29:23 crc kubenswrapper[4870]: I1014 09:29:23.953410 4870 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"3eea07ce0e4dde9363f92f5f44f530c26c10b42e335dd2bafb6764d9cbdc203c"} pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 14 09:29:23 crc kubenswrapper[4870]: I1014 09:29:23.953549 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerName="machine-config-daemon" containerID="cri-o://3eea07ce0e4dde9363f92f5f44f530c26c10b42e335dd2bafb6764d9cbdc203c" gracePeriod=600 Oct 14 09:29:24 crc kubenswrapper[4870]: E1014 09:29:24.087065 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 09:29:25 crc kubenswrapper[4870]: I1014 09:29:25.083054 4870 generic.go:334] "Generic (PLEG): container finished" podID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerID="3eea07ce0e4dde9363f92f5f44f530c26c10b42e335dd2bafb6764d9cbdc203c" exitCode=0 Oct 14 09:29:25 crc kubenswrapper[4870]: I1014 09:29:25.083096 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" event={"ID":"874a5e45-dffd-4d17-b609-b7d3ed2eab07","Type":"ContainerDied","Data":"3eea07ce0e4dde9363f92f5f44f530c26c10b42e335dd2bafb6764d9cbdc203c"} Oct 14 09:29:25 crc kubenswrapper[4870]: I1014 09:29:25.083127 4870 scope.go:117] "RemoveContainer" containerID="c92a8ddf378f6420fa6f685f5a67ccca0eb5b167c543057b9d31affa78843af8" Oct 14 09:29:25 crc kubenswrapper[4870]: I1014 09:29:25.084074 4870 scope.go:117] "RemoveContainer" containerID="3eea07ce0e4dde9363f92f5f44f530c26c10b42e335dd2bafb6764d9cbdc203c" Oct 14 09:29:25 crc kubenswrapper[4870]: E1014 09:29:25.084383 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 09:29:31 crc kubenswrapper[4870]: I1014 09:29:31.153420 4870 generic.go:334] "Generic (PLEG): container finished" podID="a90e5cfd-b6bb-45a3-8b31-cfa84c3ebf93" containerID="5cf83ec245634e137d8e0c7599a907899702bd4bfe58d5b28a23b596ad714933" exitCode=0 Oct 14 09:29:31 crc kubenswrapper[4870]: I1014 09:29:31.153466 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-openstack-openstack-cell1-q4zrw" event={"ID":"a90e5cfd-b6bb-45a3-8b31-cfa84c3ebf93","Type":"ContainerDied","Data":"5cf83ec245634e137d8e0c7599a907899702bd4bfe58d5b28a23b596ad714933"} Oct 14 09:29:32 crc kubenswrapper[4870]: I1014 09:29:32.671850 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-openstack-openstack-cell1-q4zrw" Oct 14 09:29:32 crc kubenswrapper[4870]: I1014 09:29:32.735284 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a90e5cfd-b6bb-45a3-8b31-cfa84c3ebf93-neutron-metadata-combined-ca-bundle\") pod \"a90e5cfd-b6bb-45a3-8b31-cfa84c3ebf93\" (UID: \"a90e5cfd-b6bb-45a3-8b31-cfa84c3ebf93\") " Oct 14 09:29:32 crc kubenswrapper[4870]: I1014 09:29:32.735339 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-894vw\" (UniqueName: \"kubernetes.io/projected/a90e5cfd-b6bb-45a3-8b31-cfa84c3ebf93-kube-api-access-894vw\") pod \"a90e5cfd-b6bb-45a3-8b31-cfa84c3ebf93\" (UID: \"a90e5cfd-b6bb-45a3-8b31-cfa84c3ebf93\") " Oct 14 09:29:32 crc kubenswrapper[4870]: I1014 09:29:32.735472 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a90e5cfd-b6bb-45a3-8b31-cfa84c3ebf93-bootstrap-combined-ca-bundle\") pod \"a90e5cfd-b6bb-45a3-8b31-cfa84c3ebf93\" (UID: \"a90e5cfd-b6bb-45a3-8b31-cfa84c3ebf93\") " Oct 14 09:29:32 crc kubenswrapper[4870]: I1014 09:29:32.735508 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a90e5cfd-b6bb-45a3-8b31-cfa84c3ebf93-neutron-sriov-combined-ca-bundle\") pod \"a90e5cfd-b6bb-45a3-8b31-cfa84c3ebf93\" (UID: \"a90e5cfd-b6bb-45a3-8b31-cfa84c3ebf93\") " Oct 14 09:29:32 crc kubenswrapper[4870]: I1014 09:29:32.735557 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a90e5cfd-b6bb-45a3-8b31-cfa84c3ebf93-nova-combined-ca-bundle\") pod \"a90e5cfd-b6bb-45a3-8b31-cfa84c3ebf93\" (UID: \"a90e5cfd-b6bb-45a3-8b31-cfa84c3ebf93\") " Oct 14 09:29:32 crc kubenswrapper[4870]: I1014 09:29:32.735586 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a90e5cfd-b6bb-45a3-8b31-cfa84c3ebf93-neutron-dhcp-combined-ca-bundle\") pod \"a90e5cfd-b6bb-45a3-8b31-cfa84c3ebf93\" (UID: \"a90e5cfd-b6bb-45a3-8b31-cfa84c3ebf93\") " Oct 14 09:29:32 crc kubenswrapper[4870]: I1014 09:29:32.735742 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a90e5cfd-b6bb-45a3-8b31-cfa84c3ebf93-inventory\") pod \"a90e5cfd-b6bb-45a3-8b31-cfa84c3ebf93\" (UID: \"a90e5cfd-b6bb-45a3-8b31-cfa84c3ebf93\") " Oct 14 09:29:32 crc kubenswrapper[4870]: I1014 09:29:32.735802 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a90e5cfd-b6bb-45a3-8b31-cfa84c3ebf93-ssh-key\") pod \"a90e5cfd-b6bb-45a3-8b31-cfa84c3ebf93\" (UID: \"a90e5cfd-b6bb-45a3-8b31-cfa84c3ebf93\") " Oct 14 09:29:32 crc kubenswrapper[4870]: I1014 09:29:32.736665 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/a90e5cfd-b6bb-45a3-8b31-cfa84c3ebf93-ceph\") pod \"a90e5cfd-b6bb-45a3-8b31-cfa84c3ebf93\" (UID: \"a90e5cfd-b6bb-45a3-8b31-cfa84c3ebf93\") " Oct 14 09:29:32 crc kubenswrapper[4870]: I1014 09:29:32.736742 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a90e5cfd-b6bb-45a3-8b31-cfa84c3ebf93-libvirt-combined-ca-bundle\") pod \"a90e5cfd-b6bb-45a3-8b31-cfa84c3ebf93\" (UID: \"a90e5cfd-b6bb-45a3-8b31-cfa84c3ebf93\") " Oct 14 09:29:32 crc kubenswrapper[4870]: I1014 09:29:32.736770 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a90e5cfd-b6bb-45a3-8b31-cfa84c3ebf93-ovn-combined-ca-bundle\") pod \"a90e5cfd-b6bb-45a3-8b31-cfa84c3ebf93\" (UID: \"a90e5cfd-b6bb-45a3-8b31-cfa84c3ebf93\") " Oct 14 09:29:32 crc kubenswrapper[4870]: I1014 09:29:32.736968 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a90e5cfd-b6bb-45a3-8b31-cfa84c3ebf93-telemetry-combined-ca-bundle\") pod \"a90e5cfd-b6bb-45a3-8b31-cfa84c3ebf93\" (UID: \"a90e5cfd-b6bb-45a3-8b31-cfa84c3ebf93\") " Oct 14 09:29:32 crc kubenswrapper[4870]: I1014 09:29:32.743245 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a90e5cfd-b6bb-45a3-8b31-cfa84c3ebf93-ceph" (OuterVolumeSpecName: "ceph") pod "a90e5cfd-b6bb-45a3-8b31-cfa84c3ebf93" (UID: "a90e5cfd-b6bb-45a3-8b31-cfa84c3ebf93"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:29:32 crc kubenswrapper[4870]: I1014 09:29:32.744412 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a90e5cfd-b6bb-45a3-8b31-cfa84c3ebf93-neutron-sriov-combined-ca-bundle" (OuterVolumeSpecName: "neutron-sriov-combined-ca-bundle") pod "a90e5cfd-b6bb-45a3-8b31-cfa84c3ebf93" (UID: "a90e5cfd-b6bb-45a3-8b31-cfa84c3ebf93"). InnerVolumeSpecName "neutron-sriov-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:29:32 crc kubenswrapper[4870]: I1014 09:29:32.745040 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a90e5cfd-b6bb-45a3-8b31-cfa84c3ebf93-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "a90e5cfd-b6bb-45a3-8b31-cfa84c3ebf93" (UID: "a90e5cfd-b6bb-45a3-8b31-cfa84c3ebf93"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:29:32 crc kubenswrapper[4870]: I1014 09:29:32.745519 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a90e5cfd-b6bb-45a3-8b31-cfa84c3ebf93-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "a90e5cfd-b6bb-45a3-8b31-cfa84c3ebf93" (UID: "a90e5cfd-b6bb-45a3-8b31-cfa84c3ebf93"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:29:32 crc kubenswrapper[4870]: I1014 09:29:32.745586 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a90e5cfd-b6bb-45a3-8b31-cfa84c3ebf93-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "a90e5cfd-b6bb-45a3-8b31-cfa84c3ebf93" (UID: "a90e5cfd-b6bb-45a3-8b31-cfa84c3ebf93"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:29:32 crc kubenswrapper[4870]: I1014 09:29:32.745939 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a90e5cfd-b6bb-45a3-8b31-cfa84c3ebf93-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "a90e5cfd-b6bb-45a3-8b31-cfa84c3ebf93" (UID: "a90e5cfd-b6bb-45a3-8b31-cfa84c3ebf93"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:29:32 crc kubenswrapper[4870]: I1014 09:29:32.746077 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a90e5cfd-b6bb-45a3-8b31-cfa84c3ebf93-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "a90e5cfd-b6bb-45a3-8b31-cfa84c3ebf93" (UID: "a90e5cfd-b6bb-45a3-8b31-cfa84c3ebf93"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:29:32 crc kubenswrapper[4870]: I1014 09:29:32.748594 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a90e5cfd-b6bb-45a3-8b31-cfa84c3ebf93-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "a90e5cfd-b6bb-45a3-8b31-cfa84c3ebf93" (UID: "a90e5cfd-b6bb-45a3-8b31-cfa84c3ebf93"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:29:32 crc kubenswrapper[4870]: I1014 09:29:32.748873 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a90e5cfd-b6bb-45a3-8b31-cfa84c3ebf93-neutron-dhcp-combined-ca-bundle" (OuterVolumeSpecName: "neutron-dhcp-combined-ca-bundle") pod "a90e5cfd-b6bb-45a3-8b31-cfa84c3ebf93" (UID: "a90e5cfd-b6bb-45a3-8b31-cfa84c3ebf93"). InnerVolumeSpecName "neutron-dhcp-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:29:32 crc kubenswrapper[4870]: I1014 09:29:32.753763 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a90e5cfd-b6bb-45a3-8b31-cfa84c3ebf93-kube-api-access-894vw" (OuterVolumeSpecName: "kube-api-access-894vw") pod "a90e5cfd-b6bb-45a3-8b31-cfa84c3ebf93" (UID: "a90e5cfd-b6bb-45a3-8b31-cfa84c3ebf93"). InnerVolumeSpecName "kube-api-access-894vw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:29:32 crc kubenswrapper[4870]: I1014 09:29:32.769749 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a90e5cfd-b6bb-45a3-8b31-cfa84c3ebf93-inventory" (OuterVolumeSpecName: "inventory") pod "a90e5cfd-b6bb-45a3-8b31-cfa84c3ebf93" (UID: "a90e5cfd-b6bb-45a3-8b31-cfa84c3ebf93"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:29:32 crc kubenswrapper[4870]: I1014 09:29:32.773290 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a90e5cfd-b6bb-45a3-8b31-cfa84c3ebf93-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "a90e5cfd-b6bb-45a3-8b31-cfa84c3ebf93" (UID: "a90e5cfd-b6bb-45a3-8b31-cfa84c3ebf93"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:29:32 crc kubenswrapper[4870]: I1014 09:29:32.839765 4870 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a90e5cfd-b6bb-45a3-8b31-cfa84c3ebf93-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 09:29:32 crc kubenswrapper[4870]: I1014 09:29:32.839808 4870 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a90e5cfd-b6bb-45a3-8b31-cfa84c3ebf93-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 09:29:32 crc kubenswrapper[4870]: I1014 09:29:32.839824 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-894vw\" (UniqueName: \"kubernetes.io/projected/a90e5cfd-b6bb-45a3-8b31-cfa84c3ebf93-kube-api-access-894vw\") on node \"crc\" DevicePath \"\"" Oct 14 09:29:32 crc kubenswrapper[4870]: I1014 09:29:32.839838 4870 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a90e5cfd-b6bb-45a3-8b31-cfa84c3ebf93-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 09:29:32 crc kubenswrapper[4870]: I1014 09:29:32.839851 4870 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a90e5cfd-b6bb-45a3-8b31-cfa84c3ebf93-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 09:29:32 crc kubenswrapper[4870]: I1014 09:29:32.839862 4870 reconciler_common.go:293] "Volume detached for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a90e5cfd-b6bb-45a3-8b31-cfa84c3ebf93-neutron-sriov-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 09:29:32 crc kubenswrapper[4870]: I1014 09:29:32.839874 4870 reconciler_common.go:293] "Volume detached for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a90e5cfd-b6bb-45a3-8b31-cfa84c3ebf93-neutron-dhcp-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 09:29:32 crc kubenswrapper[4870]: I1014 09:29:32.839889 4870 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a90e5cfd-b6bb-45a3-8b31-cfa84c3ebf93-inventory\") on node \"crc\" DevicePath \"\"" Oct 14 09:29:32 crc kubenswrapper[4870]: I1014 09:29:32.839904 4870 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a90e5cfd-b6bb-45a3-8b31-cfa84c3ebf93-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 14 09:29:32 crc kubenswrapper[4870]: I1014 09:29:32.839917 4870 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/a90e5cfd-b6bb-45a3-8b31-cfa84c3ebf93-ceph\") on node \"crc\" DevicePath \"\"" Oct 14 09:29:32 crc kubenswrapper[4870]: I1014 09:29:32.839928 4870 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a90e5cfd-b6bb-45a3-8b31-cfa84c3ebf93-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 09:29:32 crc kubenswrapper[4870]: I1014 09:29:32.839943 4870 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a90e5cfd-b6bb-45a3-8b31-cfa84c3ebf93-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 09:29:33 crc kubenswrapper[4870]: I1014 09:29:33.177874 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-openstack-openstack-cell1-q4zrw" event={"ID":"a90e5cfd-b6bb-45a3-8b31-cfa84c3ebf93","Type":"ContainerDied","Data":"b3fa518a94ee3abcbb6e29998f431a33f28aa32033dc987e657a1accb73b1415"} Oct 14 09:29:33 crc kubenswrapper[4870]: I1014 09:29:33.178222 4870 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b3fa518a94ee3abcbb6e29998f431a33f28aa32033dc987e657a1accb73b1415" Oct 14 09:29:33 crc kubenswrapper[4870]: I1014 09:29:33.177998 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-openstack-openstack-cell1-q4zrw" Oct 14 09:29:33 crc kubenswrapper[4870]: I1014 09:29:33.304880 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceph-client-openstack-openstack-cell1-f7d6q"] Oct 14 09:29:33 crc kubenswrapper[4870]: E1014 09:29:33.305838 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a90e5cfd-b6bb-45a3-8b31-cfa84c3ebf93" containerName="install-certs-openstack-openstack-cell1" Oct 14 09:29:33 crc kubenswrapper[4870]: I1014 09:29:33.305871 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="a90e5cfd-b6bb-45a3-8b31-cfa84c3ebf93" containerName="install-certs-openstack-openstack-cell1" Oct 14 09:29:33 crc kubenswrapper[4870]: I1014 09:29:33.306295 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="a90e5cfd-b6bb-45a3-8b31-cfa84c3ebf93" containerName="install-certs-openstack-openstack-cell1" Oct 14 09:29:33 crc kubenswrapper[4870]: I1014 09:29:33.307734 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-openstack-openstack-cell1-f7d6q" Oct 14 09:29:33 crc kubenswrapper[4870]: I1014 09:29:33.310246 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-wrz7w" Oct 14 09:29:33 crc kubenswrapper[4870]: I1014 09:29:33.311133 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 14 09:29:33 crc kubenswrapper[4870]: I1014 09:29:33.324422 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-client-openstack-openstack-cell1-f7d6q"] Oct 14 09:29:33 crc kubenswrapper[4870]: I1014 09:29:33.359203 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/aee008c6-821e-44b3-a557-11dda24d715b-ceph\") pod \"ceph-client-openstack-openstack-cell1-f7d6q\" (UID: \"aee008c6-821e-44b3-a557-11dda24d715b\") " pod="openstack/ceph-client-openstack-openstack-cell1-f7d6q" Oct 14 09:29:33 crc kubenswrapper[4870]: I1014 09:29:33.359275 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/aee008c6-821e-44b3-a557-11dda24d715b-ssh-key\") pod \"ceph-client-openstack-openstack-cell1-f7d6q\" (UID: \"aee008c6-821e-44b3-a557-11dda24d715b\") " pod="openstack/ceph-client-openstack-openstack-cell1-f7d6q" Oct 14 09:29:33 crc kubenswrapper[4870]: I1014 09:29:33.359318 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/aee008c6-821e-44b3-a557-11dda24d715b-inventory\") pod \"ceph-client-openstack-openstack-cell1-f7d6q\" (UID: \"aee008c6-821e-44b3-a557-11dda24d715b\") " pod="openstack/ceph-client-openstack-openstack-cell1-f7d6q" Oct 14 09:29:33 crc kubenswrapper[4870]: I1014 09:29:33.359505 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nrmc7\" (UniqueName: \"kubernetes.io/projected/aee008c6-821e-44b3-a557-11dda24d715b-kube-api-access-nrmc7\") pod \"ceph-client-openstack-openstack-cell1-f7d6q\" (UID: \"aee008c6-821e-44b3-a557-11dda24d715b\") " pod="openstack/ceph-client-openstack-openstack-cell1-f7d6q" Oct 14 09:29:33 crc kubenswrapper[4870]: I1014 09:29:33.461808 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nrmc7\" (UniqueName: \"kubernetes.io/projected/aee008c6-821e-44b3-a557-11dda24d715b-kube-api-access-nrmc7\") pod \"ceph-client-openstack-openstack-cell1-f7d6q\" (UID: \"aee008c6-821e-44b3-a557-11dda24d715b\") " pod="openstack/ceph-client-openstack-openstack-cell1-f7d6q" Oct 14 09:29:33 crc kubenswrapper[4870]: I1014 09:29:33.461984 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/aee008c6-821e-44b3-a557-11dda24d715b-ceph\") pod \"ceph-client-openstack-openstack-cell1-f7d6q\" (UID: \"aee008c6-821e-44b3-a557-11dda24d715b\") " pod="openstack/ceph-client-openstack-openstack-cell1-f7d6q" Oct 14 09:29:33 crc kubenswrapper[4870]: I1014 09:29:33.462004 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/aee008c6-821e-44b3-a557-11dda24d715b-ssh-key\") pod \"ceph-client-openstack-openstack-cell1-f7d6q\" (UID: \"aee008c6-821e-44b3-a557-11dda24d715b\") " pod="openstack/ceph-client-openstack-openstack-cell1-f7d6q" Oct 14 09:29:33 crc kubenswrapper[4870]: I1014 09:29:33.462023 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/aee008c6-821e-44b3-a557-11dda24d715b-inventory\") pod \"ceph-client-openstack-openstack-cell1-f7d6q\" (UID: \"aee008c6-821e-44b3-a557-11dda24d715b\") " pod="openstack/ceph-client-openstack-openstack-cell1-f7d6q" Oct 14 09:29:33 crc kubenswrapper[4870]: I1014 09:29:33.467076 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/aee008c6-821e-44b3-a557-11dda24d715b-ceph\") pod \"ceph-client-openstack-openstack-cell1-f7d6q\" (UID: \"aee008c6-821e-44b3-a557-11dda24d715b\") " pod="openstack/ceph-client-openstack-openstack-cell1-f7d6q" Oct 14 09:29:33 crc kubenswrapper[4870]: I1014 09:29:33.467209 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/aee008c6-821e-44b3-a557-11dda24d715b-inventory\") pod \"ceph-client-openstack-openstack-cell1-f7d6q\" (UID: \"aee008c6-821e-44b3-a557-11dda24d715b\") " pod="openstack/ceph-client-openstack-openstack-cell1-f7d6q" Oct 14 09:29:33 crc kubenswrapper[4870]: I1014 09:29:33.467419 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/aee008c6-821e-44b3-a557-11dda24d715b-ssh-key\") pod \"ceph-client-openstack-openstack-cell1-f7d6q\" (UID: \"aee008c6-821e-44b3-a557-11dda24d715b\") " pod="openstack/ceph-client-openstack-openstack-cell1-f7d6q" Oct 14 09:29:33 crc kubenswrapper[4870]: I1014 09:29:33.480080 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nrmc7\" (UniqueName: \"kubernetes.io/projected/aee008c6-821e-44b3-a557-11dda24d715b-kube-api-access-nrmc7\") pod \"ceph-client-openstack-openstack-cell1-f7d6q\" (UID: \"aee008c6-821e-44b3-a557-11dda24d715b\") " pod="openstack/ceph-client-openstack-openstack-cell1-f7d6q" Oct 14 09:29:33 crc kubenswrapper[4870]: I1014 09:29:33.673775 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-openstack-openstack-cell1-f7d6q" Oct 14 09:29:34 crc kubenswrapper[4870]: I1014 09:29:34.240013 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-client-openstack-openstack-cell1-f7d6q"] Oct 14 09:29:35 crc kubenswrapper[4870]: I1014 09:29:35.196288 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-openstack-openstack-cell1-f7d6q" event={"ID":"aee008c6-821e-44b3-a557-11dda24d715b","Type":"ContainerStarted","Data":"e95218498e17a43a941a5f0f110c72054f0dc8992213ae3c474342db8e6afcd1"} Oct 14 09:29:36 crc kubenswrapper[4870]: I1014 09:29:36.217660 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-openstack-openstack-cell1-f7d6q" event={"ID":"aee008c6-821e-44b3-a557-11dda24d715b","Type":"ContainerStarted","Data":"972e4f10477252a8985d1e5efd04af9c7765c6f2254e2631e0d577de3de000c8"} Oct 14 09:29:36 crc kubenswrapper[4870]: I1014 09:29:36.259796 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceph-client-openstack-openstack-cell1-f7d6q" podStartSLOduration=2.5722019510000003 podStartE2EDuration="3.259773216s" podCreationTimestamp="2025-10-14 09:29:33 +0000 UTC" firstStartedPulling="2025-10-14 09:29:34.245917665 +0000 UTC m=+8909.943278036" lastFinishedPulling="2025-10-14 09:29:34.9334889 +0000 UTC m=+8910.630849301" observedRunningTime="2025-10-14 09:29:36.247196837 +0000 UTC m=+8911.944557228" watchObservedRunningTime="2025-10-14 09:29:36.259773216 +0000 UTC m=+8911.957133597" Oct 14 09:29:38 crc kubenswrapper[4870]: I1014 09:29:38.033848 4870 scope.go:117] "RemoveContainer" containerID="3eea07ce0e4dde9363f92f5f44f530c26c10b42e335dd2bafb6764d9cbdc203c" Oct 14 09:29:38 crc kubenswrapper[4870]: E1014 09:29:38.034170 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 09:29:40 crc kubenswrapper[4870]: I1014 09:29:40.891682 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-49fcc"] Oct 14 09:29:40 crc kubenswrapper[4870]: I1014 09:29:40.896089 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-49fcc" Oct 14 09:29:40 crc kubenswrapper[4870]: I1014 09:29:40.917609 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-49fcc"] Oct 14 09:29:40 crc kubenswrapper[4870]: I1014 09:29:40.983082 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4bf79468-ba38-4ab2-ac48-0e00d24134fa-catalog-content\") pod \"certified-operators-49fcc\" (UID: \"4bf79468-ba38-4ab2-ac48-0e00d24134fa\") " pod="openshift-marketplace/certified-operators-49fcc" Oct 14 09:29:40 crc kubenswrapper[4870]: I1014 09:29:40.983315 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c878t\" (UniqueName: \"kubernetes.io/projected/4bf79468-ba38-4ab2-ac48-0e00d24134fa-kube-api-access-c878t\") pod \"certified-operators-49fcc\" (UID: \"4bf79468-ba38-4ab2-ac48-0e00d24134fa\") " pod="openshift-marketplace/certified-operators-49fcc" Oct 14 09:29:40 crc kubenswrapper[4870]: I1014 09:29:40.983482 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4bf79468-ba38-4ab2-ac48-0e00d24134fa-utilities\") pod \"certified-operators-49fcc\" (UID: \"4bf79468-ba38-4ab2-ac48-0e00d24134fa\") " pod="openshift-marketplace/certified-operators-49fcc" Oct 14 09:29:41 crc kubenswrapper[4870]: I1014 09:29:41.090937 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4bf79468-ba38-4ab2-ac48-0e00d24134fa-catalog-content\") pod \"certified-operators-49fcc\" (UID: \"4bf79468-ba38-4ab2-ac48-0e00d24134fa\") " pod="openshift-marketplace/certified-operators-49fcc" Oct 14 09:29:41 crc kubenswrapper[4870]: I1014 09:29:41.091088 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c878t\" (UniqueName: \"kubernetes.io/projected/4bf79468-ba38-4ab2-ac48-0e00d24134fa-kube-api-access-c878t\") pod \"certified-operators-49fcc\" (UID: \"4bf79468-ba38-4ab2-ac48-0e00d24134fa\") " pod="openshift-marketplace/certified-operators-49fcc" Oct 14 09:29:41 crc kubenswrapper[4870]: I1014 09:29:41.091138 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4bf79468-ba38-4ab2-ac48-0e00d24134fa-utilities\") pod \"certified-operators-49fcc\" (UID: \"4bf79468-ba38-4ab2-ac48-0e00d24134fa\") " pod="openshift-marketplace/certified-operators-49fcc" Oct 14 09:29:41 crc kubenswrapper[4870]: I1014 09:29:41.091787 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4bf79468-ba38-4ab2-ac48-0e00d24134fa-utilities\") pod \"certified-operators-49fcc\" (UID: \"4bf79468-ba38-4ab2-ac48-0e00d24134fa\") " pod="openshift-marketplace/certified-operators-49fcc" Oct 14 09:29:41 crc kubenswrapper[4870]: I1014 09:29:41.098809 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4bf79468-ba38-4ab2-ac48-0e00d24134fa-catalog-content\") pod \"certified-operators-49fcc\" (UID: \"4bf79468-ba38-4ab2-ac48-0e00d24134fa\") " pod="openshift-marketplace/certified-operators-49fcc" Oct 14 09:29:41 crc kubenswrapper[4870]: I1014 09:29:41.131499 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c878t\" (UniqueName: \"kubernetes.io/projected/4bf79468-ba38-4ab2-ac48-0e00d24134fa-kube-api-access-c878t\") pod \"certified-operators-49fcc\" (UID: \"4bf79468-ba38-4ab2-ac48-0e00d24134fa\") " pod="openshift-marketplace/certified-operators-49fcc" Oct 14 09:29:41 crc kubenswrapper[4870]: I1014 09:29:41.231312 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-49fcc" Oct 14 09:29:41 crc kubenswrapper[4870]: I1014 09:29:41.305883 4870 generic.go:334] "Generic (PLEG): container finished" podID="aee008c6-821e-44b3-a557-11dda24d715b" containerID="972e4f10477252a8985d1e5efd04af9c7765c6f2254e2631e0d577de3de000c8" exitCode=0 Oct 14 09:29:41 crc kubenswrapper[4870]: I1014 09:29:41.306649 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-openstack-openstack-cell1-f7d6q" event={"ID":"aee008c6-821e-44b3-a557-11dda24d715b","Type":"ContainerDied","Data":"972e4f10477252a8985d1e5efd04af9c7765c6f2254e2631e0d577de3de000c8"} Oct 14 09:29:41 crc kubenswrapper[4870]: I1014 09:29:41.852102 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-49fcc"] Oct 14 09:29:42 crc kubenswrapper[4870]: I1014 09:29:42.319645 4870 generic.go:334] "Generic (PLEG): container finished" podID="4bf79468-ba38-4ab2-ac48-0e00d24134fa" containerID="e444aeefff188f6d9ad8b3351735b4a6148e5eb1def902fe200dcc1ce070c43e" exitCode=0 Oct 14 09:29:42 crc kubenswrapper[4870]: I1014 09:29:42.319763 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-49fcc" event={"ID":"4bf79468-ba38-4ab2-ac48-0e00d24134fa","Type":"ContainerDied","Data":"e444aeefff188f6d9ad8b3351735b4a6148e5eb1def902fe200dcc1ce070c43e"} Oct 14 09:29:42 crc kubenswrapper[4870]: I1014 09:29:42.319974 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-49fcc" event={"ID":"4bf79468-ba38-4ab2-ac48-0e00d24134fa","Type":"ContainerStarted","Data":"0a411a4899cba520c5dfd9a97bbf8f7f1cb6534c844ac0ff60932de6d3fc91d5"} Oct 14 09:29:42 crc kubenswrapper[4870]: I1014 09:29:42.766274 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-openstack-openstack-cell1-f7d6q" Oct 14 09:29:42 crc kubenswrapper[4870]: I1014 09:29:42.937141 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/aee008c6-821e-44b3-a557-11dda24d715b-ceph\") pod \"aee008c6-821e-44b3-a557-11dda24d715b\" (UID: \"aee008c6-821e-44b3-a557-11dda24d715b\") " Oct 14 09:29:42 crc kubenswrapper[4870]: I1014 09:29:42.937404 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/aee008c6-821e-44b3-a557-11dda24d715b-ssh-key\") pod \"aee008c6-821e-44b3-a557-11dda24d715b\" (UID: \"aee008c6-821e-44b3-a557-11dda24d715b\") " Oct 14 09:29:42 crc kubenswrapper[4870]: I1014 09:29:42.937471 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nrmc7\" (UniqueName: \"kubernetes.io/projected/aee008c6-821e-44b3-a557-11dda24d715b-kube-api-access-nrmc7\") pod \"aee008c6-821e-44b3-a557-11dda24d715b\" (UID: \"aee008c6-821e-44b3-a557-11dda24d715b\") " Oct 14 09:29:42 crc kubenswrapper[4870]: I1014 09:29:42.937577 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/aee008c6-821e-44b3-a557-11dda24d715b-inventory\") pod \"aee008c6-821e-44b3-a557-11dda24d715b\" (UID: \"aee008c6-821e-44b3-a557-11dda24d715b\") " Oct 14 09:29:42 crc kubenswrapper[4870]: I1014 09:29:42.945669 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aee008c6-821e-44b3-a557-11dda24d715b-ceph" (OuterVolumeSpecName: "ceph") pod "aee008c6-821e-44b3-a557-11dda24d715b" (UID: "aee008c6-821e-44b3-a557-11dda24d715b"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:29:42 crc kubenswrapper[4870]: I1014 09:29:42.946169 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aee008c6-821e-44b3-a557-11dda24d715b-kube-api-access-nrmc7" (OuterVolumeSpecName: "kube-api-access-nrmc7") pod "aee008c6-821e-44b3-a557-11dda24d715b" (UID: "aee008c6-821e-44b3-a557-11dda24d715b"). InnerVolumeSpecName "kube-api-access-nrmc7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:29:42 crc kubenswrapper[4870]: I1014 09:29:42.981125 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aee008c6-821e-44b3-a557-11dda24d715b-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "aee008c6-821e-44b3-a557-11dda24d715b" (UID: "aee008c6-821e-44b3-a557-11dda24d715b"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:29:42 crc kubenswrapper[4870]: I1014 09:29:42.997178 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aee008c6-821e-44b3-a557-11dda24d715b-inventory" (OuterVolumeSpecName: "inventory") pod "aee008c6-821e-44b3-a557-11dda24d715b" (UID: "aee008c6-821e-44b3-a557-11dda24d715b"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:29:43 crc kubenswrapper[4870]: I1014 09:29:43.040989 4870 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/aee008c6-821e-44b3-a557-11dda24d715b-inventory\") on node \"crc\" DevicePath \"\"" Oct 14 09:29:43 crc kubenswrapper[4870]: I1014 09:29:43.041040 4870 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/aee008c6-821e-44b3-a557-11dda24d715b-ceph\") on node \"crc\" DevicePath \"\"" Oct 14 09:29:43 crc kubenswrapper[4870]: I1014 09:29:43.041054 4870 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/aee008c6-821e-44b3-a557-11dda24d715b-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 14 09:29:43 crc kubenswrapper[4870]: I1014 09:29:43.041072 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nrmc7\" (UniqueName: \"kubernetes.io/projected/aee008c6-821e-44b3-a557-11dda24d715b-kube-api-access-nrmc7\") on node \"crc\" DevicePath \"\"" Oct 14 09:29:43 crc kubenswrapper[4870]: I1014 09:29:43.332766 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-openstack-openstack-cell1-f7d6q" event={"ID":"aee008c6-821e-44b3-a557-11dda24d715b","Type":"ContainerDied","Data":"e95218498e17a43a941a5f0f110c72054f0dc8992213ae3c474342db8e6afcd1"} Oct 14 09:29:43 crc kubenswrapper[4870]: I1014 09:29:43.332803 4870 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e95218498e17a43a941a5f0f110c72054f0dc8992213ae3c474342db8e6afcd1" Oct 14 09:29:43 crc kubenswrapper[4870]: I1014 09:29:43.332859 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-openstack-openstack-cell1-f7d6q" Oct 14 09:29:43 crc kubenswrapper[4870]: I1014 09:29:43.426018 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-openstack-openstack-cell1-6ggqp"] Oct 14 09:29:43 crc kubenswrapper[4870]: E1014 09:29:43.426919 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aee008c6-821e-44b3-a557-11dda24d715b" containerName="ceph-client-openstack-openstack-cell1" Oct 14 09:29:43 crc kubenswrapper[4870]: I1014 09:29:43.426946 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="aee008c6-821e-44b3-a557-11dda24d715b" containerName="ceph-client-openstack-openstack-cell1" Oct 14 09:29:43 crc kubenswrapper[4870]: I1014 09:29:43.427250 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="aee008c6-821e-44b3-a557-11dda24d715b" containerName="ceph-client-openstack-openstack-cell1" Oct 14 09:29:43 crc kubenswrapper[4870]: I1014 09:29:43.428235 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-openstack-openstack-cell1-6ggqp" Oct 14 09:29:43 crc kubenswrapper[4870]: I1014 09:29:43.430903 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-wrz7w" Oct 14 09:29:43 crc kubenswrapper[4870]: I1014 09:29:43.431325 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 14 09:29:43 crc kubenswrapper[4870]: I1014 09:29:43.440247 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-openstack-openstack-cell1-6ggqp"] Oct 14 09:29:43 crc kubenswrapper[4870]: I1014 09:29:43.449118 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6735cf95-3968-4805-b08c-e3d9231e0d6f-inventory\") pod \"ovn-openstack-openstack-cell1-6ggqp\" (UID: \"6735cf95-3968-4805-b08c-e3d9231e0d6f\") " pod="openstack/ovn-openstack-openstack-cell1-6ggqp" Oct 14 09:29:43 crc kubenswrapper[4870]: I1014 09:29:43.449213 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6735cf95-3968-4805-b08c-e3d9231e0d6f-ssh-key\") pod \"ovn-openstack-openstack-cell1-6ggqp\" (UID: \"6735cf95-3968-4805-b08c-e3d9231e0d6f\") " pod="openstack/ovn-openstack-openstack-cell1-6ggqp" Oct 14 09:29:43 crc kubenswrapper[4870]: I1014 09:29:43.449261 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/6735cf95-3968-4805-b08c-e3d9231e0d6f-ceph\") pod \"ovn-openstack-openstack-cell1-6ggqp\" (UID: \"6735cf95-3968-4805-b08c-e3d9231e0d6f\") " pod="openstack/ovn-openstack-openstack-cell1-6ggqp" Oct 14 09:29:43 crc kubenswrapper[4870]: I1014 09:29:43.449292 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/6735cf95-3968-4805-b08c-e3d9231e0d6f-ovncontroller-config-0\") pod \"ovn-openstack-openstack-cell1-6ggqp\" (UID: \"6735cf95-3968-4805-b08c-e3d9231e0d6f\") " pod="openstack/ovn-openstack-openstack-cell1-6ggqp" Oct 14 09:29:43 crc kubenswrapper[4870]: I1014 09:29:43.449358 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6735cf95-3968-4805-b08c-e3d9231e0d6f-ovn-combined-ca-bundle\") pod \"ovn-openstack-openstack-cell1-6ggqp\" (UID: \"6735cf95-3968-4805-b08c-e3d9231e0d6f\") " pod="openstack/ovn-openstack-openstack-cell1-6ggqp" Oct 14 09:29:43 crc kubenswrapper[4870]: I1014 09:29:43.449416 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2xfvk\" (UniqueName: \"kubernetes.io/projected/6735cf95-3968-4805-b08c-e3d9231e0d6f-kube-api-access-2xfvk\") pod \"ovn-openstack-openstack-cell1-6ggqp\" (UID: \"6735cf95-3968-4805-b08c-e3d9231e0d6f\") " pod="openstack/ovn-openstack-openstack-cell1-6ggqp" Oct 14 09:29:43 crc kubenswrapper[4870]: I1014 09:29:43.551458 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6735cf95-3968-4805-b08c-e3d9231e0d6f-ssh-key\") pod \"ovn-openstack-openstack-cell1-6ggqp\" (UID: \"6735cf95-3968-4805-b08c-e3d9231e0d6f\") " pod="openstack/ovn-openstack-openstack-cell1-6ggqp" Oct 14 09:29:43 crc kubenswrapper[4870]: I1014 09:29:43.551547 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/6735cf95-3968-4805-b08c-e3d9231e0d6f-ceph\") pod \"ovn-openstack-openstack-cell1-6ggqp\" (UID: \"6735cf95-3968-4805-b08c-e3d9231e0d6f\") " pod="openstack/ovn-openstack-openstack-cell1-6ggqp" Oct 14 09:29:43 crc kubenswrapper[4870]: I1014 09:29:43.551578 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/6735cf95-3968-4805-b08c-e3d9231e0d6f-ovncontroller-config-0\") pod \"ovn-openstack-openstack-cell1-6ggqp\" (UID: \"6735cf95-3968-4805-b08c-e3d9231e0d6f\") " pod="openstack/ovn-openstack-openstack-cell1-6ggqp" Oct 14 09:29:43 crc kubenswrapper[4870]: I1014 09:29:43.551615 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6735cf95-3968-4805-b08c-e3d9231e0d6f-ovn-combined-ca-bundle\") pod \"ovn-openstack-openstack-cell1-6ggqp\" (UID: \"6735cf95-3968-4805-b08c-e3d9231e0d6f\") " pod="openstack/ovn-openstack-openstack-cell1-6ggqp" Oct 14 09:29:43 crc kubenswrapper[4870]: I1014 09:29:43.551690 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2xfvk\" (UniqueName: \"kubernetes.io/projected/6735cf95-3968-4805-b08c-e3d9231e0d6f-kube-api-access-2xfvk\") pod \"ovn-openstack-openstack-cell1-6ggqp\" (UID: \"6735cf95-3968-4805-b08c-e3d9231e0d6f\") " pod="openstack/ovn-openstack-openstack-cell1-6ggqp" Oct 14 09:29:43 crc kubenswrapper[4870]: I1014 09:29:43.551840 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6735cf95-3968-4805-b08c-e3d9231e0d6f-inventory\") pod \"ovn-openstack-openstack-cell1-6ggqp\" (UID: \"6735cf95-3968-4805-b08c-e3d9231e0d6f\") " pod="openstack/ovn-openstack-openstack-cell1-6ggqp" Oct 14 09:29:43 crc kubenswrapper[4870]: I1014 09:29:43.552864 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/6735cf95-3968-4805-b08c-e3d9231e0d6f-ovncontroller-config-0\") pod \"ovn-openstack-openstack-cell1-6ggqp\" (UID: \"6735cf95-3968-4805-b08c-e3d9231e0d6f\") " pod="openstack/ovn-openstack-openstack-cell1-6ggqp" Oct 14 09:29:43 crc kubenswrapper[4870]: I1014 09:29:43.555178 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/6735cf95-3968-4805-b08c-e3d9231e0d6f-ceph\") pod \"ovn-openstack-openstack-cell1-6ggqp\" (UID: \"6735cf95-3968-4805-b08c-e3d9231e0d6f\") " pod="openstack/ovn-openstack-openstack-cell1-6ggqp" Oct 14 09:29:43 crc kubenswrapper[4870]: I1014 09:29:43.555289 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6735cf95-3968-4805-b08c-e3d9231e0d6f-ssh-key\") pod \"ovn-openstack-openstack-cell1-6ggqp\" (UID: \"6735cf95-3968-4805-b08c-e3d9231e0d6f\") " pod="openstack/ovn-openstack-openstack-cell1-6ggqp" Oct 14 09:29:43 crc kubenswrapper[4870]: I1014 09:29:43.555429 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6735cf95-3968-4805-b08c-e3d9231e0d6f-inventory\") pod \"ovn-openstack-openstack-cell1-6ggqp\" (UID: \"6735cf95-3968-4805-b08c-e3d9231e0d6f\") " pod="openstack/ovn-openstack-openstack-cell1-6ggqp" Oct 14 09:29:43 crc kubenswrapper[4870]: I1014 09:29:43.556067 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6735cf95-3968-4805-b08c-e3d9231e0d6f-ovn-combined-ca-bundle\") pod \"ovn-openstack-openstack-cell1-6ggqp\" (UID: \"6735cf95-3968-4805-b08c-e3d9231e0d6f\") " pod="openstack/ovn-openstack-openstack-cell1-6ggqp" Oct 14 09:29:43 crc kubenswrapper[4870]: I1014 09:29:43.571850 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2xfvk\" (UniqueName: \"kubernetes.io/projected/6735cf95-3968-4805-b08c-e3d9231e0d6f-kube-api-access-2xfvk\") pod \"ovn-openstack-openstack-cell1-6ggqp\" (UID: \"6735cf95-3968-4805-b08c-e3d9231e0d6f\") " pod="openstack/ovn-openstack-openstack-cell1-6ggqp" Oct 14 09:29:43 crc kubenswrapper[4870]: I1014 09:29:43.753077 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-openstack-openstack-cell1-6ggqp" Oct 14 09:29:44 crc kubenswrapper[4870]: I1014 09:29:44.368135 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-49fcc" event={"ID":"4bf79468-ba38-4ab2-ac48-0e00d24134fa","Type":"ContainerStarted","Data":"f1fccb6e258c69e72be028ace087a689c84d706f8e96a73646c53dea8b75c8f0"} Oct 14 09:29:44 crc kubenswrapper[4870]: I1014 09:29:44.445761 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-openstack-openstack-cell1-6ggqp"] Oct 14 09:29:45 crc kubenswrapper[4870]: I1014 09:29:45.385133 4870 generic.go:334] "Generic (PLEG): container finished" podID="4bf79468-ba38-4ab2-ac48-0e00d24134fa" containerID="f1fccb6e258c69e72be028ace087a689c84d706f8e96a73646c53dea8b75c8f0" exitCode=0 Oct 14 09:29:45 crc kubenswrapper[4870]: I1014 09:29:45.385250 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-49fcc" event={"ID":"4bf79468-ba38-4ab2-ac48-0e00d24134fa","Type":"ContainerDied","Data":"f1fccb6e258c69e72be028ace087a689c84d706f8e96a73646c53dea8b75c8f0"} Oct 14 09:29:45 crc kubenswrapper[4870]: I1014 09:29:45.390200 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-openstack-openstack-cell1-6ggqp" event={"ID":"6735cf95-3968-4805-b08c-e3d9231e0d6f","Type":"ContainerStarted","Data":"a1c19f3dc8f555d04df68e05aabfa9283768f5146699265678c61314e66d84c9"} Oct 14 09:29:45 crc kubenswrapper[4870]: I1014 09:29:45.390233 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-openstack-openstack-cell1-6ggqp" event={"ID":"6735cf95-3968-4805-b08c-e3d9231e0d6f","Type":"ContainerStarted","Data":"3fbba42e989e82f46d6bb8b6ce23f99480d1740012f6054c21f9aa47077be420"} Oct 14 09:29:45 crc kubenswrapper[4870]: I1014 09:29:45.474921 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-openstack-openstack-cell1-6ggqp" podStartSLOduration=1.9949861100000001 podStartE2EDuration="2.474899015s" podCreationTimestamp="2025-10-14 09:29:43 +0000 UTC" firstStartedPulling="2025-10-14 09:29:44.451815028 +0000 UTC m=+8920.149175399" lastFinishedPulling="2025-10-14 09:29:44.931727933 +0000 UTC m=+8920.629088304" observedRunningTime="2025-10-14 09:29:45.436193593 +0000 UTC m=+8921.133554004" watchObservedRunningTime="2025-10-14 09:29:45.474899015 +0000 UTC m=+8921.172259396" Oct 14 09:29:46 crc kubenswrapper[4870]: I1014 09:29:46.408542 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-49fcc" event={"ID":"4bf79468-ba38-4ab2-ac48-0e00d24134fa","Type":"ContainerStarted","Data":"c76e02d9698954c73ee1eab9db0e0cd5664c5c93303095f9c2935cae0b433457"} Oct 14 09:29:46 crc kubenswrapper[4870]: I1014 09:29:46.437142 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-49fcc" podStartSLOduration=2.8047415300000003 podStartE2EDuration="6.437126716s" podCreationTimestamp="2025-10-14 09:29:40 +0000 UTC" firstStartedPulling="2025-10-14 09:29:42.323352918 +0000 UTC m=+8918.020713289" lastFinishedPulling="2025-10-14 09:29:45.955738094 +0000 UTC m=+8921.653098475" observedRunningTime="2025-10-14 09:29:46.433687402 +0000 UTC m=+8922.131047773" watchObservedRunningTime="2025-10-14 09:29:46.437126716 +0000 UTC m=+8922.134487087" Oct 14 09:29:51 crc kubenswrapper[4870]: I1014 09:29:51.232418 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-49fcc" Oct 14 09:29:51 crc kubenswrapper[4870]: I1014 09:29:51.233072 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-49fcc" Oct 14 09:29:51 crc kubenswrapper[4870]: I1014 09:29:51.321362 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-49fcc" Oct 14 09:29:51 crc kubenswrapper[4870]: I1014 09:29:51.557104 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-49fcc" Oct 14 09:29:51 crc kubenswrapper[4870]: I1014 09:29:51.612549 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-49fcc"] Oct 14 09:29:53 crc kubenswrapper[4870]: I1014 09:29:53.037232 4870 scope.go:117] "RemoveContainer" containerID="3eea07ce0e4dde9363f92f5f44f530c26c10b42e335dd2bafb6764d9cbdc203c" Oct 14 09:29:53 crc kubenswrapper[4870]: E1014 09:29:53.038167 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 09:29:53 crc kubenswrapper[4870]: I1014 09:29:53.515710 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-49fcc" podUID="4bf79468-ba38-4ab2-ac48-0e00d24134fa" containerName="registry-server" containerID="cri-o://c76e02d9698954c73ee1eab9db0e0cd5664c5c93303095f9c2935cae0b433457" gracePeriod=2 Oct 14 09:29:54 crc kubenswrapper[4870]: I1014 09:29:54.016935 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-49fcc" Oct 14 09:29:54 crc kubenswrapper[4870]: I1014 09:29:54.081819 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4bf79468-ba38-4ab2-ac48-0e00d24134fa-utilities\") pod \"4bf79468-ba38-4ab2-ac48-0e00d24134fa\" (UID: \"4bf79468-ba38-4ab2-ac48-0e00d24134fa\") " Oct 14 09:29:54 crc kubenswrapper[4870]: I1014 09:29:54.082143 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4bf79468-ba38-4ab2-ac48-0e00d24134fa-catalog-content\") pod \"4bf79468-ba38-4ab2-ac48-0e00d24134fa\" (UID: \"4bf79468-ba38-4ab2-ac48-0e00d24134fa\") " Oct 14 09:29:54 crc kubenswrapper[4870]: I1014 09:29:54.082417 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c878t\" (UniqueName: \"kubernetes.io/projected/4bf79468-ba38-4ab2-ac48-0e00d24134fa-kube-api-access-c878t\") pod \"4bf79468-ba38-4ab2-ac48-0e00d24134fa\" (UID: \"4bf79468-ba38-4ab2-ac48-0e00d24134fa\") " Oct 14 09:29:54 crc kubenswrapper[4870]: I1014 09:29:54.082860 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4bf79468-ba38-4ab2-ac48-0e00d24134fa-utilities" (OuterVolumeSpecName: "utilities") pod "4bf79468-ba38-4ab2-ac48-0e00d24134fa" (UID: "4bf79468-ba38-4ab2-ac48-0e00d24134fa"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 09:29:54 crc kubenswrapper[4870]: I1014 09:29:54.088275 4870 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4bf79468-ba38-4ab2-ac48-0e00d24134fa-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 09:29:54 crc kubenswrapper[4870]: I1014 09:29:54.093020 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bf79468-ba38-4ab2-ac48-0e00d24134fa-kube-api-access-c878t" (OuterVolumeSpecName: "kube-api-access-c878t") pod "4bf79468-ba38-4ab2-ac48-0e00d24134fa" (UID: "4bf79468-ba38-4ab2-ac48-0e00d24134fa"). InnerVolumeSpecName "kube-api-access-c878t". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:29:54 crc kubenswrapper[4870]: I1014 09:29:54.153381 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4bf79468-ba38-4ab2-ac48-0e00d24134fa-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4bf79468-ba38-4ab2-ac48-0e00d24134fa" (UID: "4bf79468-ba38-4ab2-ac48-0e00d24134fa"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 09:29:54 crc kubenswrapper[4870]: I1014 09:29:54.190697 4870 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4bf79468-ba38-4ab2-ac48-0e00d24134fa-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 09:29:54 crc kubenswrapper[4870]: I1014 09:29:54.190727 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c878t\" (UniqueName: \"kubernetes.io/projected/4bf79468-ba38-4ab2-ac48-0e00d24134fa-kube-api-access-c878t\") on node \"crc\" DevicePath \"\"" Oct 14 09:29:54 crc kubenswrapper[4870]: I1014 09:29:54.532396 4870 generic.go:334] "Generic (PLEG): container finished" podID="4bf79468-ba38-4ab2-ac48-0e00d24134fa" containerID="c76e02d9698954c73ee1eab9db0e0cd5664c5c93303095f9c2935cae0b433457" exitCode=0 Oct 14 09:29:54 crc kubenswrapper[4870]: I1014 09:29:54.532561 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-49fcc" Oct 14 09:29:54 crc kubenswrapper[4870]: I1014 09:29:54.532791 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-49fcc" event={"ID":"4bf79468-ba38-4ab2-ac48-0e00d24134fa","Type":"ContainerDied","Data":"c76e02d9698954c73ee1eab9db0e0cd5664c5c93303095f9c2935cae0b433457"} Oct 14 09:29:54 crc kubenswrapper[4870]: I1014 09:29:54.532829 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-49fcc" event={"ID":"4bf79468-ba38-4ab2-ac48-0e00d24134fa","Type":"ContainerDied","Data":"0a411a4899cba520c5dfd9a97bbf8f7f1cb6534c844ac0ff60932de6d3fc91d5"} Oct 14 09:29:54 crc kubenswrapper[4870]: I1014 09:29:54.532855 4870 scope.go:117] "RemoveContainer" containerID="c76e02d9698954c73ee1eab9db0e0cd5664c5c93303095f9c2935cae0b433457" Oct 14 09:29:54 crc kubenswrapper[4870]: I1014 09:29:54.570568 4870 scope.go:117] "RemoveContainer" containerID="f1fccb6e258c69e72be028ace087a689c84d706f8e96a73646c53dea8b75c8f0" Oct 14 09:29:54 crc kubenswrapper[4870]: I1014 09:29:54.600590 4870 scope.go:117] "RemoveContainer" containerID="e444aeefff188f6d9ad8b3351735b4a6148e5eb1def902fe200dcc1ce070c43e" Oct 14 09:29:54 crc kubenswrapper[4870]: I1014 09:29:54.602208 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-49fcc"] Oct 14 09:29:54 crc kubenswrapper[4870]: I1014 09:29:54.618272 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-49fcc"] Oct 14 09:29:54 crc kubenswrapper[4870]: I1014 09:29:54.665215 4870 scope.go:117] "RemoveContainer" containerID="c76e02d9698954c73ee1eab9db0e0cd5664c5c93303095f9c2935cae0b433457" Oct 14 09:29:54 crc kubenswrapper[4870]: E1014 09:29:54.666182 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c76e02d9698954c73ee1eab9db0e0cd5664c5c93303095f9c2935cae0b433457\": container with ID starting with c76e02d9698954c73ee1eab9db0e0cd5664c5c93303095f9c2935cae0b433457 not found: ID does not exist" containerID="c76e02d9698954c73ee1eab9db0e0cd5664c5c93303095f9c2935cae0b433457" Oct 14 09:29:54 crc kubenswrapper[4870]: I1014 09:29:54.666301 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c76e02d9698954c73ee1eab9db0e0cd5664c5c93303095f9c2935cae0b433457"} err="failed to get container status \"c76e02d9698954c73ee1eab9db0e0cd5664c5c93303095f9c2935cae0b433457\": rpc error: code = NotFound desc = could not find container \"c76e02d9698954c73ee1eab9db0e0cd5664c5c93303095f9c2935cae0b433457\": container with ID starting with c76e02d9698954c73ee1eab9db0e0cd5664c5c93303095f9c2935cae0b433457 not found: ID does not exist" Oct 14 09:29:54 crc kubenswrapper[4870]: I1014 09:29:54.666401 4870 scope.go:117] "RemoveContainer" containerID="f1fccb6e258c69e72be028ace087a689c84d706f8e96a73646c53dea8b75c8f0" Oct 14 09:29:54 crc kubenswrapper[4870]: E1014 09:29:54.666802 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f1fccb6e258c69e72be028ace087a689c84d706f8e96a73646c53dea8b75c8f0\": container with ID starting with f1fccb6e258c69e72be028ace087a689c84d706f8e96a73646c53dea8b75c8f0 not found: ID does not exist" containerID="f1fccb6e258c69e72be028ace087a689c84d706f8e96a73646c53dea8b75c8f0" Oct 14 09:29:54 crc kubenswrapper[4870]: I1014 09:29:54.666824 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f1fccb6e258c69e72be028ace087a689c84d706f8e96a73646c53dea8b75c8f0"} err="failed to get container status \"f1fccb6e258c69e72be028ace087a689c84d706f8e96a73646c53dea8b75c8f0\": rpc error: code = NotFound desc = could not find container \"f1fccb6e258c69e72be028ace087a689c84d706f8e96a73646c53dea8b75c8f0\": container with ID starting with f1fccb6e258c69e72be028ace087a689c84d706f8e96a73646c53dea8b75c8f0 not found: ID does not exist" Oct 14 09:29:54 crc kubenswrapper[4870]: I1014 09:29:54.666838 4870 scope.go:117] "RemoveContainer" containerID="e444aeefff188f6d9ad8b3351735b4a6148e5eb1def902fe200dcc1ce070c43e" Oct 14 09:29:54 crc kubenswrapper[4870]: E1014 09:29:54.667240 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e444aeefff188f6d9ad8b3351735b4a6148e5eb1def902fe200dcc1ce070c43e\": container with ID starting with e444aeefff188f6d9ad8b3351735b4a6148e5eb1def902fe200dcc1ce070c43e not found: ID does not exist" containerID="e444aeefff188f6d9ad8b3351735b4a6148e5eb1def902fe200dcc1ce070c43e" Oct 14 09:29:54 crc kubenswrapper[4870]: I1014 09:29:54.667260 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e444aeefff188f6d9ad8b3351735b4a6148e5eb1def902fe200dcc1ce070c43e"} err="failed to get container status \"e444aeefff188f6d9ad8b3351735b4a6148e5eb1def902fe200dcc1ce070c43e\": rpc error: code = NotFound desc = could not find container \"e444aeefff188f6d9ad8b3351735b4a6148e5eb1def902fe200dcc1ce070c43e\": container with ID starting with e444aeefff188f6d9ad8b3351735b4a6148e5eb1def902fe200dcc1ce070c43e not found: ID does not exist" Oct 14 09:29:55 crc kubenswrapper[4870]: I1014 09:29:55.079869 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bf79468-ba38-4ab2-ac48-0e00d24134fa" path="/var/lib/kubelet/pods/4bf79468-ba38-4ab2-ac48-0e00d24134fa/volumes" Oct 14 09:30:00 crc kubenswrapper[4870]: I1014 09:30:00.162049 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29340570-kh9fl"] Oct 14 09:30:00 crc kubenswrapper[4870]: E1014 09:30:00.163152 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4bf79468-ba38-4ab2-ac48-0e00d24134fa" containerName="extract-utilities" Oct 14 09:30:00 crc kubenswrapper[4870]: I1014 09:30:00.163264 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="4bf79468-ba38-4ab2-ac48-0e00d24134fa" containerName="extract-utilities" Oct 14 09:30:00 crc kubenswrapper[4870]: E1014 09:30:00.163292 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4bf79468-ba38-4ab2-ac48-0e00d24134fa" containerName="extract-content" Oct 14 09:30:00 crc kubenswrapper[4870]: I1014 09:30:00.163300 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="4bf79468-ba38-4ab2-ac48-0e00d24134fa" containerName="extract-content" Oct 14 09:30:00 crc kubenswrapper[4870]: E1014 09:30:00.163332 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4bf79468-ba38-4ab2-ac48-0e00d24134fa" containerName="registry-server" Oct 14 09:30:00 crc kubenswrapper[4870]: I1014 09:30:00.163339 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="4bf79468-ba38-4ab2-ac48-0e00d24134fa" containerName="registry-server" Oct 14 09:30:00 crc kubenswrapper[4870]: I1014 09:30:00.163757 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="4bf79468-ba38-4ab2-ac48-0e00d24134fa" containerName="registry-server" Oct 14 09:30:00 crc kubenswrapper[4870]: I1014 09:30:00.164745 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29340570-kh9fl" Oct 14 09:30:00 crc kubenswrapper[4870]: I1014 09:30:00.167555 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 14 09:30:00 crc kubenswrapper[4870]: I1014 09:30:00.168199 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 14 09:30:00 crc kubenswrapper[4870]: I1014 09:30:00.184553 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29340570-kh9fl"] Oct 14 09:30:00 crc kubenswrapper[4870]: I1014 09:30:00.339019 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/557da624-75cc-4387-8771-241da876c655-config-volume\") pod \"collect-profiles-29340570-kh9fl\" (UID: \"557da624-75cc-4387-8771-241da876c655\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340570-kh9fl" Oct 14 09:30:00 crc kubenswrapper[4870]: I1014 09:30:00.339252 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/557da624-75cc-4387-8771-241da876c655-secret-volume\") pod \"collect-profiles-29340570-kh9fl\" (UID: \"557da624-75cc-4387-8771-241da876c655\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340570-kh9fl" Oct 14 09:30:00 crc kubenswrapper[4870]: I1014 09:30:00.339395 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9445q\" (UniqueName: \"kubernetes.io/projected/557da624-75cc-4387-8771-241da876c655-kube-api-access-9445q\") pod \"collect-profiles-29340570-kh9fl\" (UID: \"557da624-75cc-4387-8771-241da876c655\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340570-kh9fl" Oct 14 09:30:00 crc kubenswrapper[4870]: I1014 09:30:00.440917 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/557da624-75cc-4387-8771-241da876c655-secret-volume\") pod \"collect-profiles-29340570-kh9fl\" (UID: \"557da624-75cc-4387-8771-241da876c655\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340570-kh9fl" Oct 14 09:30:00 crc kubenswrapper[4870]: I1014 09:30:00.441009 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9445q\" (UniqueName: \"kubernetes.io/projected/557da624-75cc-4387-8771-241da876c655-kube-api-access-9445q\") pod \"collect-profiles-29340570-kh9fl\" (UID: \"557da624-75cc-4387-8771-241da876c655\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340570-kh9fl" Oct 14 09:30:00 crc kubenswrapper[4870]: I1014 09:30:00.441117 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/557da624-75cc-4387-8771-241da876c655-config-volume\") pod \"collect-profiles-29340570-kh9fl\" (UID: \"557da624-75cc-4387-8771-241da876c655\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340570-kh9fl" Oct 14 09:30:00 crc kubenswrapper[4870]: I1014 09:30:00.442483 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/557da624-75cc-4387-8771-241da876c655-config-volume\") pod \"collect-profiles-29340570-kh9fl\" (UID: \"557da624-75cc-4387-8771-241da876c655\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340570-kh9fl" Oct 14 09:30:00 crc kubenswrapper[4870]: I1014 09:30:00.451316 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/557da624-75cc-4387-8771-241da876c655-secret-volume\") pod \"collect-profiles-29340570-kh9fl\" (UID: \"557da624-75cc-4387-8771-241da876c655\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340570-kh9fl" Oct 14 09:30:00 crc kubenswrapper[4870]: I1014 09:30:00.456487 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9445q\" (UniqueName: \"kubernetes.io/projected/557da624-75cc-4387-8771-241da876c655-kube-api-access-9445q\") pod \"collect-profiles-29340570-kh9fl\" (UID: \"557da624-75cc-4387-8771-241da876c655\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340570-kh9fl" Oct 14 09:30:00 crc kubenswrapper[4870]: I1014 09:30:00.503938 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29340570-kh9fl" Oct 14 09:30:00 crc kubenswrapper[4870]: I1014 09:30:00.965788 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29340570-kh9fl"] Oct 14 09:30:01 crc kubenswrapper[4870]: I1014 09:30:01.613876 4870 generic.go:334] "Generic (PLEG): container finished" podID="557da624-75cc-4387-8771-241da876c655" containerID="be81d519402e34c4d49c058e6868c5a6cb81f4fe590402a93b44a4fe61a99a55" exitCode=0 Oct 14 09:30:01 crc kubenswrapper[4870]: I1014 09:30:01.614191 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29340570-kh9fl" event={"ID":"557da624-75cc-4387-8771-241da876c655","Type":"ContainerDied","Data":"be81d519402e34c4d49c058e6868c5a6cb81f4fe590402a93b44a4fe61a99a55"} Oct 14 09:30:01 crc kubenswrapper[4870]: I1014 09:30:01.614224 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29340570-kh9fl" event={"ID":"557da624-75cc-4387-8771-241da876c655","Type":"ContainerStarted","Data":"89d092b93cf2235cb886a793803c1563b755cdc7dc8b9e78b80e784ac63097a0"} Oct 14 09:30:02 crc kubenswrapper[4870]: I1014 09:30:02.988736 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29340570-kh9fl" Oct 14 09:30:03 crc kubenswrapper[4870]: I1014 09:30:03.091559 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9445q\" (UniqueName: \"kubernetes.io/projected/557da624-75cc-4387-8771-241da876c655-kube-api-access-9445q\") pod \"557da624-75cc-4387-8771-241da876c655\" (UID: \"557da624-75cc-4387-8771-241da876c655\") " Oct 14 09:30:03 crc kubenswrapper[4870]: I1014 09:30:03.091600 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/557da624-75cc-4387-8771-241da876c655-config-volume\") pod \"557da624-75cc-4387-8771-241da876c655\" (UID: \"557da624-75cc-4387-8771-241da876c655\") " Oct 14 09:30:03 crc kubenswrapper[4870]: I1014 09:30:03.091710 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/557da624-75cc-4387-8771-241da876c655-secret-volume\") pod \"557da624-75cc-4387-8771-241da876c655\" (UID: \"557da624-75cc-4387-8771-241da876c655\") " Oct 14 09:30:03 crc kubenswrapper[4870]: I1014 09:30:03.092666 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/557da624-75cc-4387-8771-241da876c655-config-volume" (OuterVolumeSpecName: "config-volume") pod "557da624-75cc-4387-8771-241da876c655" (UID: "557da624-75cc-4387-8771-241da876c655"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 09:30:03 crc kubenswrapper[4870]: I1014 09:30:03.098833 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/557da624-75cc-4387-8771-241da876c655-kube-api-access-9445q" (OuterVolumeSpecName: "kube-api-access-9445q") pod "557da624-75cc-4387-8771-241da876c655" (UID: "557da624-75cc-4387-8771-241da876c655"). InnerVolumeSpecName "kube-api-access-9445q". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:30:03 crc kubenswrapper[4870]: I1014 09:30:03.104627 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/557da624-75cc-4387-8771-241da876c655-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "557da624-75cc-4387-8771-241da876c655" (UID: "557da624-75cc-4387-8771-241da876c655"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:30:03 crc kubenswrapper[4870]: I1014 09:30:03.195013 4870 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/557da624-75cc-4387-8771-241da876c655-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 14 09:30:03 crc kubenswrapper[4870]: I1014 09:30:03.195073 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9445q\" (UniqueName: \"kubernetes.io/projected/557da624-75cc-4387-8771-241da876c655-kube-api-access-9445q\") on node \"crc\" DevicePath \"\"" Oct 14 09:30:03 crc kubenswrapper[4870]: I1014 09:30:03.195095 4870 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/557da624-75cc-4387-8771-241da876c655-config-volume\") on node \"crc\" DevicePath \"\"" Oct 14 09:30:03 crc kubenswrapper[4870]: I1014 09:30:03.637642 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29340570-kh9fl" event={"ID":"557da624-75cc-4387-8771-241da876c655","Type":"ContainerDied","Data":"89d092b93cf2235cb886a793803c1563b755cdc7dc8b9e78b80e784ac63097a0"} Oct 14 09:30:03 crc kubenswrapper[4870]: I1014 09:30:03.638029 4870 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="89d092b93cf2235cb886a793803c1563b755cdc7dc8b9e78b80e784ac63097a0" Oct 14 09:30:03 crc kubenswrapper[4870]: I1014 09:30:03.637726 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29340570-kh9fl" Oct 14 09:30:04 crc kubenswrapper[4870]: I1014 09:30:04.082392 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29340525-c7ww6"] Oct 14 09:30:04 crc kubenswrapper[4870]: I1014 09:30:04.093055 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29340525-c7ww6"] Oct 14 09:30:05 crc kubenswrapper[4870]: I1014 09:30:05.061550 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c0c03bf7-dad0-4518-bf17-08e81691880c" path="/var/lib/kubelet/pods/c0c03bf7-dad0-4518-bf17-08e81691880c/volumes" Oct 14 09:30:07 crc kubenswrapper[4870]: I1014 09:30:07.035531 4870 scope.go:117] "RemoveContainer" containerID="3eea07ce0e4dde9363f92f5f44f530c26c10b42e335dd2bafb6764d9cbdc203c" Oct 14 09:30:07 crc kubenswrapper[4870]: E1014 09:30:07.036181 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 09:30:10 crc kubenswrapper[4870]: I1014 09:30:10.734093 4870 generic.go:334] "Generic (PLEG): container finished" podID="f27b6002-352b-4191-8b00-347f35a563bf" containerID="719983fe94249f44a37d90c5dc7f1043c46541671e7d7b4f0c197a89bdcb8c70" exitCode=0 Oct 14 09:30:10 crc kubenswrapper[4870]: I1014 09:30:10.734202 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-openstack-openstack-networker-hw7gw" event={"ID":"f27b6002-352b-4191-8b00-347f35a563bf","Type":"ContainerDied","Data":"719983fe94249f44a37d90c5dc7f1043c46541671e7d7b4f0c197a89bdcb8c70"} Oct 14 09:30:12 crc kubenswrapper[4870]: I1014 09:30:12.258981 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-openstack-openstack-networker-hw7gw" Oct 14 09:30:12 crc kubenswrapper[4870]: I1014 09:30:12.400680 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c67mm\" (UniqueName: \"kubernetes.io/projected/f27b6002-352b-4191-8b00-347f35a563bf-kube-api-access-c67mm\") pod \"f27b6002-352b-4191-8b00-347f35a563bf\" (UID: \"f27b6002-352b-4191-8b00-347f35a563bf\") " Oct 14 09:30:12 crc kubenswrapper[4870]: I1014 09:30:12.400779 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f27b6002-352b-4191-8b00-347f35a563bf-ssh-key\") pod \"f27b6002-352b-4191-8b00-347f35a563bf\" (UID: \"f27b6002-352b-4191-8b00-347f35a563bf\") " Oct 14 09:30:12 crc kubenswrapper[4870]: I1014 09:30:12.400821 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/f27b6002-352b-4191-8b00-347f35a563bf-ovncontroller-config-0\") pod \"f27b6002-352b-4191-8b00-347f35a563bf\" (UID: \"f27b6002-352b-4191-8b00-347f35a563bf\") " Oct 14 09:30:12 crc kubenswrapper[4870]: I1014 09:30:12.400843 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f27b6002-352b-4191-8b00-347f35a563bf-ovn-combined-ca-bundle\") pod \"f27b6002-352b-4191-8b00-347f35a563bf\" (UID: \"f27b6002-352b-4191-8b00-347f35a563bf\") " Oct 14 09:30:12 crc kubenswrapper[4870]: I1014 09:30:12.400886 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f27b6002-352b-4191-8b00-347f35a563bf-inventory\") pod \"f27b6002-352b-4191-8b00-347f35a563bf\" (UID: \"f27b6002-352b-4191-8b00-347f35a563bf\") " Oct 14 09:30:12 crc kubenswrapper[4870]: I1014 09:30:12.410216 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f27b6002-352b-4191-8b00-347f35a563bf-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "f27b6002-352b-4191-8b00-347f35a563bf" (UID: "f27b6002-352b-4191-8b00-347f35a563bf"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:30:12 crc kubenswrapper[4870]: I1014 09:30:12.421572 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f27b6002-352b-4191-8b00-347f35a563bf-kube-api-access-c67mm" (OuterVolumeSpecName: "kube-api-access-c67mm") pod "f27b6002-352b-4191-8b00-347f35a563bf" (UID: "f27b6002-352b-4191-8b00-347f35a563bf"). InnerVolumeSpecName "kube-api-access-c67mm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:30:12 crc kubenswrapper[4870]: I1014 09:30:12.432978 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f27b6002-352b-4191-8b00-347f35a563bf-inventory" (OuterVolumeSpecName: "inventory") pod "f27b6002-352b-4191-8b00-347f35a563bf" (UID: "f27b6002-352b-4191-8b00-347f35a563bf"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:30:12 crc kubenswrapper[4870]: I1014 09:30:12.437237 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f27b6002-352b-4191-8b00-347f35a563bf-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "f27b6002-352b-4191-8b00-347f35a563bf" (UID: "f27b6002-352b-4191-8b00-347f35a563bf"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 09:30:12 crc kubenswrapper[4870]: I1014 09:30:12.437418 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f27b6002-352b-4191-8b00-347f35a563bf-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "f27b6002-352b-4191-8b00-347f35a563bf" (UID: "f27b6002-352b-4191-8b00-347f35a563bf"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:30:12 crc kubenswrapper[4870]: I1014 09:30:12.503008 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c67mm\" (UniqueName: \"kubernetes.io/projected/f27b6002-352b-4191-8b00-347f35a563bf-kube-api-access-c67mm\") on node \"crc\" DevicePath \"\"" Oct 14 09:30:12 crc kubenswrapper[4870]: I1014 09:30:12.503049 4870 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f27b6002-352b-4191-8b00-347f35a563bf-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 14 09:30:12 crc kubenswrapper[4870]: I1014 09:30:12.503060 4870 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/f27b6002-352b-4191-8b00-347f35a563bf-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Oct 14 09:30:12 crc kubenswrapper[4870]: I1014 09:30:12.503070 4870 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f27b6002-352b-4191-8b00-347f35a563bf-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 09:30:12 crc kubenswrapper[4870]: I1014 09:30:12.503081 4870 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f27b6002-352b-4191-8b00-347f35a563bf-inventory\") on node \"crc\" DevicePath \"\"" Oct 14 09:30:12 crc kubenswrapper[4870]: I1014 09:30:12.765694 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-openstack-openstack-networker-hw7gw" event={"ID":"f27b6002-352b-4191-8b00-347f35a563bf","Type":"ContainerDied","Data":"bc5d5fdd064099fbf8aec679928d23bbbb18bb242af736e665ef383ea8b68275"} Oct 14 09:30:12 crc kubenswrapper[4870]: I1014 09:30:12.766208 4870 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bc5d5fdd064099fbf8aec679928d23bbbb18bb242af736e665ef383ea8b68275" Oct 14 09:30:12 crc kubenswrapper[4870]: I1014 09:30:12.766301 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-openstack-openstack-networker-hw7gw" Oct 14 09:30:12 crc kubenswrapper[4870]: I1014 09:30:12.863945 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-openstack-openstack-networker-7fknh"] Oct 14 09:30:12 crc kubenswrapper[4870]: E1014 09:30:12.864463 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f27b6002-352b-4191-8b00-347f35a563bf" containerName="ovn-openstack-openstack-networker" Oct 14 09:30:12 crc kubenswrapper[4870]: I1014 09:30:12.864483 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="f27b6002-352b-4191-8b00-347f35a563bf" containerName="ovn-openstack-openstack-networker" Oct 14 09:30:12 crc kubenswrapper[4870]: E1014 09:30:12.864526 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="557da624-75cc-4387-8771-241da876c655" containerName="collect-profiles" Oct 14 09:30:12 crc kubenswrapper[4870]: I1014 09:30:12.864535 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="557da624-75cc-4387-8771-241da876c655" containerName="collect-profiles" Oct 14 09:30:12 crc kubenswrapper[4870]: I1014 09:30:12.864797 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="f27b6002-352b-4191-8b00-347f35a563bf" containerName="ovn-openstack-openstack-networker" Oct 14 09:30:12 crc kubenswrapper[4870]: I1014 09:30:12.864838 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="557da624-75cc-4387-8771-241da876c655" containerName="collect-profiles" Oct 14 09:30:12 crc kubenswrapper[4870]: I1014 09:30:12.865694 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-openstack-openstack-networker-7fknh" Oct 14 09:30:12 crc kubenswrapper[4870]: I1014 09:30:12.868633 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Oct 14 09:30:12 crc kubenswrapper[4870]: I1014 09:30:12.869108 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-networker" Oct 14 09:30:12 crc kubenswrapper[4870]: I1014 09:30:12.869288 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Oct 14 09:30:12 crc kubenswrapper[4870]: I1014 09:30:12.869291 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-networker-dockercfg-54hzr" Oct 14 09:30:12 crc kubenswrapper[4870]: I1014 09:30:12.885501 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-openstack-openstack-networker-7fknh"] Oct 14 09:30:13 crc kubenswrapper[4870]: I1014 09:30:13.013206 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-csc74\" (UniqueName: \"kubernetes.io/projected/6b3e50e6-d053-4f51-9a47-60dc660562a7-kube-api-access-csc74\") pod \"neutron-metadata-openstack-openstack-networker-7fknh\" (UID: \"6b3e50e6-d053-4f51-9a47-60dc660562a7\") " pod="openstack/neutron-metadata-openstack-openstack-networker-7fknh" Oct 14 09:30:13 crc kubenswrapper[4870]: I1014 09:30:13.013362 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6b3e50e6-d053-4f51-9a47-60dc660562a7-inventory\") pod \"neutron-metadata-openstack-openstack-networker-7fknh\" (UID: \"6b3e50e6-d053-4f51-9a47-60dc660562a7\") " pod="openstack/neutron-metadata-openstack-openstack-networker-7fknh" Oct 14 09:30:13 crc kubenswrapper[4870]: I1014 09:30:13.013838 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6b3e50e6-d053-4f51-9a47-60dc660562a7-ssh-key\") pod \"neutron-metadata-openstack-openstack-networker-7fknh\" (UID: \"6b3e50e6-d053-4f51-9a47-60dc660562a7\") " pod="openstack/neutron-metadata-openstack-openstack-networker-7fknh" Oct 14 09:30:13 crc kubenswrapper[4870]: I1014 09:30:13.014335 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b3e50e6-d053-4f51-9a47-60dc660562a7-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-openstack-openstack-networker-7fknh\" (UID: \"6b3e50e6-d053-4f51-9a47-60dc660562a7\") " pod="openstack/neutron-metadata-openstack-openstack-networker-7fknh" Oct 14 09:30:13 crc kubenswrapper[4870]: I1014 09:30:13.014531 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/6b3e50e6-d053-4f51-9a47-60dc660562a7-nova-metadata-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-networker-7fknh\" (UID: \"6b3e50e6-d053-4f51-9a47-60dc660562a7\") " pod="openstack/neutron-metadata-openstack-openstack-networker-7fknh" Oct 14 09:30:13 crc kubenswrapper[4870]: I1014 09:30:13.014593 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/6b3e50e6-d053-4f51-9a47-60dc660562a7-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-networker-7fknh\" (UID: \"6b3e50e6-d053-4f51-9a47-60dc660562a7\") " pod="openstack/neutron-metadata-openstack-openstack-networker-7fknh" Oct 14 09:30:13 crc kubenswrapper[4870]: I1014 09:30:13.117264 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6b3e50e6-d053-4f51-9a47-60dc660562a7-ssh-key\") pod \"neutron-metadata-openstack-openstack-networker-7fknh\" (UID: \"6b3e50e6-d053-4f51-9a47-60dc660562a7\") " pod="openstack/neutron-metadata-openstack-openstack-networker-7fknh" Oct 14 09:30:13 crc kubenswrapper[4870]: I1014 09:30:13.117427 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b3e50e6-d053-4f51-9a47-60dc660562a7-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-openstack-openstack-networker-7fknh\" (UID: \"6b3e50e6-d053-4f51-9a47-60dc660562a7\") " pod="openstack/neutron-metadata-openstack-openstack-networker-7fknh" Oct 14 09:30:13 crc kubenswrapper[4870]: I1014 09:30:13.117516 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/6b3e50e6-d053-4f51-9a47-60dc660562a7-nova-metadata-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-networker-7fknh\" (UID: \"6b3e50e6-d053-4f51-9a47-60dc660562a7\") " pod="openstack/neutron-metadata-openstack-openstack-networker-7fknh" Oct 14 09:30:13 crc kubenswrapper[4870]: I1014 09:30:13.117555 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/6b3e50e6-d053-4f51-9a47-60dc660562a7-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-networker-7fknh\" (UID: \"6b3e50e6-d053-4f51-9a47-60dc660562a7\") " pod="openstack/neutron-metadata-openstack-openstack-networker-7fknh" Oct 14 09:30:13 crc kubenswrapper[4870]: I1014 09:30:13.117636 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-csc74\" (UniqueName: \"kubernetes.io/projected/6b3e50e6-d053-4f51-9a47-60dc660562a7-kube-api-access-csc74\") pod \"neutron-metadata-openstack-openstack-networker-7fknh\" (UID: \"6b3e50e6-d053-4f51-9a47-60dc660562a7\") " pod="openstack/neutron-metadata-openstack-openstack-networker-7fknh" Oct 14 09:30:13 crc kubenswrapper[4870]: I1014 09:30:13.117738 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6b3e50e6-d053-4f51-9a47-60dc660562a7-inventory\") pod \"neutron-metadata-openstack-openstack-networker-7fknh\" (UID: \"6b3e50e6-d053-4f51-9a47-60dc660562a7\") " pod="openstack/neutron-metadata-openstack-openstack-networker-7fknh" Oct 14 09:30:13 crc kubenswrapper[4870]: I1014 09:30:13.122885 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6b3e50e6-d053-4f51-9a47-60dc660562a7-inventory\") pod \"neutron-metadata-openstack-openstack-networker-7fknh\" (UID: \"6b3e50e6-d053-4f51-9a47-60dc660562a7\") " pod="openstack/neutron-metadata-openstack-openstack-networker-7fknh" Oct 14 09:30:13 crc kubenswrapper[4870]: I1014 09:30:13.122911 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b3e50e6-d053-4f51-9a47-60dc660562a7-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-openstack-openstack-networker-7fknh\" (UID: \"6b3e50e6-d053-4f51-9a47-60dc660562a7\") " pod="openstack/neutron-metadata-openstack-openstack-networker-7fknh" Oct 14 09:30:13 crc kubenswrapper[4870]: I1014 09:30:13.123058 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/6b3e50e6-d053-4f51-9a47-60dc660562a7-nova-metadata-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-networker-7fknh\" (UID: \"6b3e50e6-d053-4f51-9a47-60dc660562a7\") " pod="openstack/neutron-metadata-openstack-openstack-networker-7fknh" Oct 14 09:30:13 crc kubenswrapper[4870]: I1014 09:30:13.127069 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/6b3e50e6-d053-4f51-9a47-60dc660562a7-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-networker-7fknh\" (UID: \"6b3e50e6-d053-4f51-9a47-60dc660562a7\") " pod="openstack/neutron-metadata-openstack-openstack-networker-7fknh" Oct 14 09:30:13 crc kubenswrapper[4870]: I1014 09:30:13.137664 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6b3e50e6-d053-4f51-9a47-60dc660562a7-ssh-key\") pod \"neutron-metadata-openstack-openstack-networker-7fknh\" (UID: \"6b3e50e6-d053-4f51-9a47-60dc660562a7\") " pod="openstack/neutron-metadata-openstack-openstack-networker-7fknh" Oct 14 09:30:13 crc kubenswrapper[4870]: I1014 09:30:13.141564 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-csc74\" (UniqueName: \"kubernetes.io/projected/6b3e50e6-d053-4f51-9a47-60dc660562a7-kube-api-access-csc74\") pod \"neutron-metadata-openstack-openstack-networker-7fknh\" (UID: \"6b3e50e6-d053-4f51-9a47-60dc660562a7\") " pod="openstack/neutron-metadata-openstack-openstack-networker-7fknh" Oct 14 09:30:13 crc kubenswrapper[4870]: I1014 09:30:13.193746 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-openstack-openstack-networker-7fknh" Oct 14 09:30:13 crc kubenswrapper[4870]: I1014 09:30:13.827768 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-openstack-openstack-networker-7fknh"] Oct 14 09:30:14 crc kubenswrapper[4870]: I1014 09:30:14.792507 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-openstack-openstack-networker-7fknh" event={"ID":"6b3e50e6-d053-4f51-9a47-60dc660562a7","Type":"ContainerStarted","Data":"f9414b88652437c981e66f991f11d5fd2cb1abe36f09f62ae45b0a5d9e671231"} Oct 14 09:30:15 crc kubenswrapper[4870]: I1014 09:30:15.805194 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-openstack-openstack-networker-7fknh" event={"ID":"6b3e50e6-d053-4f51-9a47-60dc660562a7","Type":"ContainerStarted","Data":"21f95ee9e1f0958e2d157709d1a833909552561be0575b01102ebad9a1542ee8"} Oct 14 09:30:19 crc kubenswrapper[4870]: I1014 09:30:19.040725 4870 scope.go:117] "RemoveContainer" containerID="3eea07ce0e4dde9363f92f5f44f530c26c10b42e335dd2bafb6764d9cbdc203c" Oct 14 09:30:19 crc kubenswrapper[4870]: E1014 09:30:19.041722 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 09:30:30 crc kubenswrapper[4870]: I1014 09:30:30.034118 4870 scope.go:117] "RemoveContainer" containerID="3eea07ce0e4dde9363f92f5f44f530c26c10b42e335dd2bafb6764d9cbdc203c" Oct 14 09:30:30 crc kubenswrapper[4870]: E1014 09:30:30.034920 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 09:30:32 crc kubenswrapper[4870]: I1014 09:30:32.891199 4870 scope.go:117] "RemoveContainer" containerID="3d6a38840140499b75f90504af412a2dbff358a946941c4e51eab87c2bf193e0" Oct 14 09:30:42 crc kubenswrapper[4870]: I1014 09:30:42.034073 4870 scope.go:117] "RemoveContainer" containerID="3eea07ce0e4dde9363f92f5f44f530c26c10b42e335dd2bafb6764d9cbdc203c" Oct 14 09:30:42 crc kubenswrapper[4870]: E1014 09:30:42.034939 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 09:30:56 crc kubenswrapper[4870]: I1014 09:30:56.303023 4870 generic.go:334] "Generic (PLEG): container finished" podID="6735cf95-3968-4805-b08c-e3d9231e0d6f" containerID="a1c19f3dc8f555d04df68e05aabfa9283768f5146699265678c61314e66d84c9" exitCode=0 Oct 14 09:30:56 crc kubenswrapper[4870]: I1014 09:30:56.303128 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-openstack-openstack-cell1-6ggqp" event={"ID":"6735cf95-3968-4805-b08c-e3d9231e0d6f","Type":"ContainerDied","Data":"a1c19f3dc8f555d04df68e05aabfa9283768f5146699265678c61314e66d84c9"} Oct 14 09:30:56 crc kubenswrapper[4870]: I1014 09:30:56.342398 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-metadata-openstack-openstack-networker-7fknh" podStartSLOduration=43.354525129 podStartE2EDuration="44.34237139s" podCreationTimestamp="2025-10-14 09:30:12 +0000 UTC" firstStartedPulling="2025-10-14 09:30:13.835570192 +0000 UTC m=+8949.532930583" lastFinishedPulling="2025-10-14 09:30:14.823416463 +0000 UTC m=+8950.520776844" observedRunningTime="2025-10-14 09:30:15.832693911 +0000 UTC m=+8951.530054302" watchObservedRunningTime="2025-10-14 09:30:56.34237139 +0000 UTC m=+8992.039731811" Oct 14 09:30:57 crc kubenswrapper[4870]: I1014 09:30:57.034746 4870 scope.go:117] "RemoveContainer" containerID="3eea07ce0e4dde9363f92f5f44f530c26c10b42e335dd2bafb6764d9cbdc203c" Oct 14 09:30:57 crc kubenswrapper[4870]: E1014 09:30:57.035487 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 09:30:57 crc kubenswrapper[4870]: I1014 09:30:57.863011 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-openstack-openstack-cell1-6ggqp" Oct 14 09:30:58 crc kubenswrapper[4870]: I1014 09:30:58.025570 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/6735cf95-3968-4805-b08c-e3d9231e0d6f-ceph\") pod \"6735cf95-3968-4805-b08c-e3d9231e0d6f\" (UID: \"6735cf95-3968-4805-b08c-e3d9231e0d6f\") " Oct 14 09:30:58 crc kubenswrapper[4870]: I1014 09:30:58.025649 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6735cf95-3968-4805-b08c-e3d9231e0d6f-inventory\") pod \"6735cf95-3968-4805-b08c-e3d9231e0d6f\" (UID: \"6735cf95-3968-4805-b08c-e3d9231e0d6f\") " Oct 14 09:30:58 crc kubenswrapper[4870]: I1014 09:30:58.025898 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6735cf95-3968-4805-b08c-e3d9231e0d6f-ssh-key\") pod \"6735cf95-3968-4805-b08c-e3d9231e0d6f\" (UID: \"6735cf95-3968-4805-b08c-e3d9231e0d6f\") " Oct 14 09:30:58 crc kubenswrapper[4870]: I1014 09:30:58.026031 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/6735cf95-3968-4805-b08c-e3d9231e0d6f-ovncontroller-config-0\") pod \"6735cf95-3968-4805-b08c-e3d9231e0d6f\" (UID: \"6735cf95-3968-4805-b08c-e3d9231e0d6f\") " Oct 14 09:30:58 crc kubenswrapper[4870]: I1014 09:30:58.026130 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6735cf95-3968-4805-b08c-e3d9231e0d6f-ovn-combined-ca-bundle\") pod \"6735cf95-3968-4805-b08c-e3d9231e0d6f\" (UID: \"6735cf95-3968-4805-b08c-e3d9231e0d6f\") " Oct 14 09:30:58 crc kubenswrapper[4870]: I1014 09:30:58.026253 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2xfvk\" (UniqueName: \"kubernetes.io/projected/6735cf95-3968-4805-b08c-e3d9231e0d6f-kube-api-access-2xfvk\") pod \"6735cf95-3968-4805-b08c-e3d9231e0d6f\" (UID: \"6735cf95-3968-4805-b08c-e3d9231e0d6f\") " Oct 14 09:30:58 crc kubenswrapper[4870]: I1014 09:30:58.032526 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6735cf95-3968-4805-b08c-e3d9231e0d6f-ceph" (OuterVolumeSpecName: "ceph") pod "6735cf95-3968-4805-b08c-e3d9231e0d6f" (UID: "6735cf95-3968-4805-b08c-e3d9231e0d6f"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:30:58 crc kubenswrapper[4870]: I1014 09:30:58.032626 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6735cf95-3968-4805-b08c-e3d9231e0d6f-kube-api-access-2xfvk" (OuterVolumeSpecName: "kube-api-access-2xfvk") pod "6735cf95-3968-4805-b08c-e3d9231e0d6f" (UID: "6735cf95-3968-4805-b08c-e3d9231e0d6f"). InnerVolumeSpecName "kube-api-access-2xfvk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:30:58 crc kubenswrapper[4870]: I1014 09:30:58.034031 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6735cf95-3968-4805-b08c-e3d9231e0d6f-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "6735cf95-3968-4805-b08c-e3d9231e0d6f" (UID: "6735cf95-3968-4805-b08c-e3d9231e0d6f"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:30:58 crc kubenswrapper[4870]: I1014 09:30:58.055652 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6735cf95-3968-4805-b08c-e3d9231e0d6f-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "6735cf95-3968-4805-b08c-e3d9231e0d6f" (UID: "6735cf95-3968-4805-b08c-e3d9231e0d6f"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 09:30:58 crc kubenswrapper[4870]: I1014 09:30:58.057738 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6735cf95-3968-4805-b08c-e3d9231e0d6f-inventory" (OuterVolumeSpecName: "inventory") pod "6735cf95-3968-4805-b08c-e3d9231e0d6f" (UID: "6735cf95-3968-4805-b08c-e3d9231e0d6f"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:30:58 crc kubenswrapper[4870]: I1014 09:30:58.082674 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6735cf95-3968-4805-b08c-e3d9231e0d6f-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "6735cf95-3968-4805-b08c-e3d9231e0d6f" (UID: "6735cf95-3968-4805-b08c-e3d9231e0d6f"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:30:58 crc kubenswrapper[4870]: I1014 09:30:58.130015 4870 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/6735cf95-3968-4805-b08c-e3d9231e0d6f-ceph\") on node \"crc\" DevicePath \"\"" Oct 14 09:30:58 crc kubenswrapper[4870]: I1014 09:30:58.130078 4870 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6735cf95-3968-4805-b08c-e3d9231e0d6f-inventory\") on node \"crc\" DevicePath \"\"" Oct 14 09:30:58 crc kubenswrapper[4870]: I1014 09:30:58.130098 4870 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6735cf95-3968-4805-b08c-e3d9231e0d6f-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 14 09:30:58 crc kubenswrapper[4870]: I1014 09:30:58.130118 4870 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/6735cf95-3968-4805-b08c-e3d9231e0d6f-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Oct 14 09:30:58 crc kubenswrapper[4870]: I1014 09:30:58.130137 4870 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6735cf95-3968-4805-b08c-e3d9231e0d6f-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 09:30:58 crc kubenswrapper[4870]: I1014 09:30:58.130155 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2xfvk\" (UniqueName: \"kubernetes.io/projected/6735cf95-3968-4805-b08c-e3d9231e0d6f-kube-api-access-2xfvk\") on node \"crc\" DevicePath \"\"" Oct 14 09:30:58 crc kubenswrapper[4870]: I1014 09:30:58.324246 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-openstack-openstack-cell1-6ggqp" event={"ID":"6735cf95-3968-4805-b08c-e3d9231e0d6f","Type":"ContainerDied","Data":"3fbba42e989e82f46d6bb8b6ce23f99480d1740012f6054c21f9aa47077be420"} Oct 14 09:30:58 crc kubenswrapper[4870]: I1014 09:30:58.324305 4870 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3fbba42e989e82f46d6bb8b6ce23f99480d1740012f6054c21f9aa47077be420" Oct 14 09:30:58 crc kubenswrapper[4870]: I1014 09:30:58.324343 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-openstack-openstack-cell1-6ggqp" Oct 14 09:30:58 crc kubenswrapper[4870]: I1014 09:30:58.445736 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-openstack-openstack-cell1-kh4dj"] Oct 14 09:30:58 crc kubenswrapper[4870]: E1014 09:30:58.447534 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6735cf95-3968-4805-b08c-e3d9231e0d6f" containerName="ovn-openstack-openstack-cell1" Oct 14 09:30:58 crc kubenswrapper[4870]: I1014 09:30:58.447594 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="6735cf95-3968-4805-b08c-e3d9231e0d6f" containerName="ovn-openstack-openstack-cell1" Oct 14 09:30:58 crc kubenswrapper[4870]: I1014 09:30:58.448057 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="6735cf95-3968-4805-b08c-e3d9231e0d6f" containerName="ovn-openstack-openstack-cell1" Oct 14 09:30:58 crc kubenswrapper[4870]: I1014 09:30:58.449553 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-openstack-openstack-cell1-kh4dj" Oct 14 09:30:58 crc kubenswrapper[4870]: I1014 09:30:58.457406 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-wrz7w" Oct 14 09:30:58 crc kubenswrapper[4870]: I1014 09:30:58.457702 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 14 09:30:58 crc kubenswrapper[4870]: I1014 09:30:58.472506 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-openstack-openstack-cell1-kh4dj"] Oct 14 09:30:58 crc kubenswrapper[4870]: I1014 09:30:58.546106 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/793a3762-db45-4a91-a194-7710517917e8-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-openstack-openstack-cell1-kh4dj\" (UID: \"793a3762-db45-4a91-a194-7710517917e8\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-kh4dj" Oct 14 09:30:58 crc kubenswrapper[4870]: I1014 09:30:58.546276 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nz4jf\" (UniqueName: \"kubernetes.io/projected/793a3762-db45-4a91-a194-7710517917e8-kube-api-access-nz4jf\") pod \"neutron-metadata-openstack-openstack-cell1-kh4dj\" (UID: \"793a3762-db45-4a91-a194-7710517917e8\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-kh4dj" Oct 14 09:30:58 crc kubenswrapper[4870]: I1014 09:30:58.546325 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/793a3762-db45-4a91-a194-7710517917e8-ceph\") pod \"neutron-metadata-openstack-openstack-cell1-kh4dj\" (UID: \"793a3762-db45-4a91-a194-7710517917e8\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-kh4dj" Oct 14 09:30:58 crc kubenswrapper[4870]: I1014 09:30:58.546492 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/793a3762-db45-4a91-a194-7710517917e8-ssh-key\") pod \"neutron-metadata-openstack-openstack-cell1-kh4dj\" (UID: \"793a3762-db45-4a91-a194-7710517917e8\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-kh4dj" Oct 14 09:30:58 crc kubenswrapper[4870]: I1014 09:30:58.546872 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/793a3762-db45-4a91-a194-7710517917e8-inventory\") pod \"neutron-metadata-openstack-openstack-cell1-kh4dj\" (UID: \"793a3762-db45-4a91-a194-7710517917e8\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-kh4dj" Oct 14 09:30:58 crc kubenswrapper[4870]: I1014 09:30:58.547000 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/793a3762-db45-4a91-a194-7710517917e8-nova-metadata-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-kh4dj\" (UID: \"793a3762-db45-4a91-a194-7710517917e8\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-kh4dj" Oct 14 09:30:58 crc kubenswrapper[4870]: I1014 09:30:58.547089 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/793a3762-db45-4a91-a194-7710517917e8-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-kh4dj\" (UID: \"793a3762-db45-4a91-a194-7710517917e8\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-kh4dj" Oct 14 09:30:58 crc kubenswrapper[4870]: I1014 09:30:58.650323 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/793a3762-db45-4a91-a194-7710517917e8-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-kh4dj\" (UID: \"793a3762-db45-4a91-a194-7710517917e8\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-kh4dj" Oct 14 09:30:58 crc kubenswrapper[4870]: I1014 09:30:58.650473 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/793a3762-db45-4a91-a194-7710517917e8-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-openstack-openstack-cell1-kh4dj\" (UID: \"793a3762-db45-4a91-a194-7710517917e8\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-kh4dj" Oct 14 09:30:58 crc kubenswrapper[4870]: I1014 09:30:58.650537 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nz4jf\" (UniqueName: \"kubernetes.io/projected/793a3762-db45-4a91-a194-7710517917e8-kube-api-access-nz4jf\") pod \"neutron-metadata-openstack-openstack-cell1-kh4dj\" (UID: \"793a3762-db45-4a91-a194-7710517917e8\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-kh4dj" Oct 14 09:30:58 crc kubenswrapper[4870]: I1014 09:30:58.650568 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/793a3762-db45-4a91-a194-7710517917e8-ceph\") pod \"neutron-metadata-openstack-openstack-cell1-kh4dj\" (UID: \"793a3762-db45-4a91-a194-7710517917e8\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-kh4dj" Oct 14 09:30:58 crc kubenswrapper[4870]: I1014 09:30:58.650631 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/793a3762-db45-4a91-a194-7710517917e8-ssh-key\") pod \"neutron-metadata-openstack-openstack-cell1-kh4dj\" (UID: \"793a3762-db45-4a91-a194-7710517917e8\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-kh4dj" Oct 14 09:30:58 crc kubenswrapper[4870]: I1014 09:30:58.650686 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/793a3762-db45-4a91-a194-7710517917e8-inventory\") pod \"neutron-metadata-openstack-openstack-cell1-kh4dj\" (UID: \"793a3762-db45-4a91-a194-7710517917e8\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-kh4dj" Oct 14 09:30:58 crc kubenswrapper[4870]: I1014 09:30:58.650713 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/793a3762-db45-4a91-a194-7710517917e8-nova-metadata-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-kh4dj\" (UID: \"793a3762-db45-4a91-a194-7710517917e8\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-kh4dj" Oct 14 09:30:58 crc kubenswrapper[4870]: I1014 09:30:58.656641 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/793a3762-db45-4a91-a194-7710517917e8-ssh-key\") pod \"neutron-metadata-openstack-openstack-cell1-kh4dj\" (UID: \"793a3762-db45-4a91-a194-7710517917e8\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-kh4dj" Oct 14 09:30:58 crc kubenswrapper[4870]: I1014 09:30:58.656932 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/793a3762-db45-4a91-a194-7710517917e8-inventory\") pod \"neutron-metadata-openstack-openstack-cell1-kh4dj\" (UID: \"793a3762-db45-4a91-a194-7710517917e8\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-kh4dj" Oct 14 09:30:58 crc kubenswrapper[4870]: I1014 09:30:58.657101 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/793a3762-db45-4a91-a194-7710517917e8-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-kh4dj\" (UID: \"793a3762-db45-4a91-a194-7710517917e8\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-kh4dj" Oct 14 09:30:58 crc kubenswrapper[4870]: I1014 09:30:58.657132 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/793a3762-db45-4a91-a194-7710517917e8-ceph\") pod \"neutron-metadata-openstack-openstack-cell1-kh4dj\" (UID: \"793a3762-db45-4a91-a194-7710517917e8\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-kh4dj" Oct 14 09:30:58 crc kubenswrapper[4870]: I1014 09:30:58.657816 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/793a3762-db45-4a91-a194-7710517917e8-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-openstack-openstack-cell1-kh4dj\" (UID: \"793a3762-db45-4a91-a194-7710517917e8\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-kh4dj" Oct 14 09:30:58 crc kubenswrapper[4870]: I1014 09:30:58.661089 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/793a3762-db45-4a91-a194-7710517917e8-nova-metadata-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-kh4dj\" (UID: \"793a3762-db45-4a91-a194-7710517917e8\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-kh4dj" Oct 14 09:30:58 crc kubenswrapper[4870]: I1014 09:30:58.671398 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nz4jf\" (UniqueName: \"kubernetes.io/projected/793a3762-db45-4a91-a194-7710517917e8-kube-api-access-nz4jf\") pod \"neutron-metadata-openstack-openstack-cell1-kh4dj\" (UID: \"793a3762-db45-4a91-a194-7710517917e8\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-kh4dj" Oct 14 09:30:58 crc kubenswrapper[4870]: I1014 09:30:58.771552 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-openstack-openstack-cell1-kh4dj" Oct 14 09:30:59 crc kubenswrapper[4870]: I1014 09:30:59.352496 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-openstack-openstack-cell1-kh4dj"] Oct 14 09:31:00 crc kubenswrapper[4870]: I1014 09:31:00.349485 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-openstack-openstack-cell1-kh4dj" event={"ID":"793a3762-db45-4a91-a194-7710517917e8","Type":"ContainerStarted","Data":"b261d5bbfd861d532463daecf537d44483c7e95c13fa0089829271795ee5cf49"} Oct 14 09:31:01 crc kubenswrapper[4870]: I1014 09:31:01.359781 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-openstack-openstack-cell1-kh4dj" event={"ID":"793a3762-db45-4a91-a194-7710517917e8","Type":"ContainerStarted","Data":"ba1e52eaefd626d7e03def49c4bb954201894aefd1cf9e2e83a07ba3dbe4faa7"} Oct 14 09:31:01 crc kubenswrapper[4870]: I1014 09:31:01.386519 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-metadata-openstack-openstack-cell1-kh4dj" podStartSLOduration=2.200589542 podStartE2EDuration="3.386499235s" podCreationTimestamp="2025-10-14 09:30:58 +0000 UTC" firstStartedPulling="2025-10-14 09:30:59.366468992 +0000 UTC m=+8995.063829363" lastFinishedPulling="2025-10-14 09:31:00.552378675 +0000 UTC m=+8996.249739056" observedRunningTime="2025-10-14 09:31:01.383623764 +0000 UTC m=+8997.080984145" watchObservedRunningTime="2025-10-14 09:31:01.386499235 +0000 UTC m=+8997.083859606" Oct 14 09:31:08 crc kubenswrapper[4870]: I1014 09:31:08.034479 4870 scope.go:117] "RemoveContainer" containerID="3eea07ce0e4dde9363f92f5f44f530c26c10b42e335dd2bafb6764d9cbdc203c" Oct 14 09:31:08 crc kubenswrapper[4870]: E1014 09:31:08.035708 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 09:31:16 crc kubenswrapper[4870]: I1014 09:31:16.556524 4870 generic.go:334] "Generic (PLEG): container finished" podID="6b3e50e6-d053-4f51-9a47-60dc660562a7" containerID="21f95ee9e1f0958e2d157709d1a833909552561be0575b01102ebad9a1542ee8" exitCode=0 Oct 14 09:31:16 crc kubenswrapper[4870]: I1014 09:31:16.556632 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-openstack-openstack-networker-7fknh" event={"ID":"6b3e50e6-d053-4f51-9a47-60dc660562a7","Type":"ContainerDied","Data":"21f95ee9e1f0958e2d157709d1a833909552561be0575b01102ebad9a1542ee8"} Oct 14 09:31:18 crc kubenswrapper[4870]: I1014 09:31:18.039719 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-openstack-openstack-networker-7fknh" Oct 14 09:31:18 crc kubenswrapper[4870]: I1014 09:31:18.125817 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6b3e50e6-d053-4f51-9a47-60dc660562a7-ssh-key\") pod \"6b3e50e6-d053-4f51-9a47-60dc660562a7\" (UID: \"6b3e50e6-d053-4f51-9a47-60dc660562a7\") " Oct 14 09:31:18 crc kubenswrapper[4870]: I1014 09:31:18.125933 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/6b3e50e6-d053-4f51-9a47-60dc660562a7-neutron-ovn-metadata-agent-neutron-config-0\") pod \"6b3e50e6-d053-4f51-9a47-60dc660562a7\" (UID: \"6b3e50e6-d053-4f51-9a47-60dc660562a7\") " Oct 14 09:31:18 crc kubenswrapper[4870]: I1014 09:31:18.125995 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-csc74\" (UniqueName: \"kubernetes.io/projected/6b3e50e6-d053-4f51-9a47-60dc660562a7-kube-api-access-csc74\") pod \"6b3e50e6-d053-4f51-9a47-60dc660562a7\" (UID: \"6b3e50e6-d053-4f51-9a47-60dc660562a7\") " Oct 14 09:31:18 crc kubenswrapper[4870]: I1014 09:31:18.126048 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b3e50e6-d053-4f51-9a47-60dc660562a7-neutron-metadata-combined-ca-bundle\") pod \"6b3e50e6-d053-4f51-9a47-60dc660562a7\" (UID: \"6b3e50e6-d053-4f51-9a47-60dc660562a7\") " Oct 14 09:31:18 crc kubenswrapper[4870]: I1014 09:31:18.126118 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/6b3e50e6-d053-4f51-9a47-60dc660562a7-nova-metadata-neutron-config-0\") pod \"6b3e50e6-d053-4f51-9a47-60dc660562a7\" (UID: \"6b3e50e6-d053-4f51-9a47-60dc660562a7\") " Oct 14 09:31:18 crc kubenswrapper[4870]: I1014 09:31:18.126172 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6b3e50e6-d053-4f51-9a47-60dc660562a7-inventory\") pod \"6b3e50e6-d053-4f51-9a47-60dc660562a7\" (UID: \"6b3e50e6-d053-4f51-9a47-60dc660562a7\") " Oct 14 09:31:18 crc kubenswrapper[4870]: I1014 09:31:18.132758 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b3e50e6-d053-4f51-9a47-60dc660562a7-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "6b3e50e6-d053-4f51-9a47-60dc660562a7" (UID: "6b3e50e6-d053-4f51-9a47-60dc660562a7"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:31:18 crc kubenswrapper[4870]: I1014 09:31:18.134304 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6b3e50e6-d053-4f51-9a47-60dc660562a7-kube-api-access-csc74" (OuterVolumeSpecName: "kube-api-access-csc74") pod "6b3e50e6-d053-4f51-9a47-60dc660562a7" (UID: "6b3e50e6-d053-4f51-9a47-60dc660562a7"). InnerVolumeSpecName "kube-api-access-csc74". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:31:18 crc kubenswrapper[4870]: I1014 09:31:18.164409 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b3e50e6-d053-4f51-9a47-60dc660562a7-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "6b3e50e6-d053-4f51-9a47-60dc660562a7" (UID: "6b3e50e6-d053-4f51-9a47-60dc660562a7"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:31:18 crc kubenswrapper[4870]: I1014 09:31:18.167162 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b3e50e6-d053-4f51-9a47-60dc660562a7-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "6b3e50e6-d053-4f51-9a47-60dc660562a7" (UID: "6b3e50e6-d053-4f51-9a47-60dc660562a7"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:31:18 crc kubenswrapper[4870]: I1014 09:31:18.169063 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b3e50e6-d053-4f51-9a47-60dc660562a7-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "6b3e50e6-d053-4f51-9a47-60dc660562a7" (UID: "6b3e50e6-d053-4f51-9a47-60dc660562a7"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:31:18 crc kubenswrapper[4870]: I1014 09:31:18.179811 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b3e50e6-d053-4f51-9a47-60dc660562a7-inventory" (OuterVolumeSpecName: "inventory") pod "6b3e50e6-d053-4f51-9a47-60dc660562a7" (UID: "6b3e50e6-d053-4f51-9a47-60dc660562a7"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:31:18 crc kubenswrapper[4870]: I1014 09:31:18.229107 4870 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6b3e50e6-d053-4f51-9a47-60dc660562a7-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 14 09:31:18 crc kubenswrapper[4870]: I1014 09:31:18.229142 4870 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/6b3e50e6-d053-4f51-9a47-60dc660562a7-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Oct 14 09:31:18 crc kubenswrapper[4870]: I1014 09:31:18.229157 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-csc74\" (UniqueName: \"kubernetes.io/projected/6b3e50e6-d053-4f51-9a47-60dc660562a7-kube-api-access-csc74\") on node \"crc\" DevicePath \"\"" Oct 14 09:31:18 crc kubenswrapper[4870]: I1014 09:31:18.229172 4870 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b3e50e6-d053-4f51-9a47-60dc660562a7-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 09:31:18 crc kubenswrapper[4870]: I1014 09:31:18.229186 4870 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/6b3e50e6-d053-4f51-9a47-60dc660562a7-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Oct 14 09:31:18 crc kubenswrapper[4870]: I1014 09:31:18.229200 4870 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6b3e50e6-d053-4f51-9a47-60dc660562a7-inventory\") on node \"crc\" DevicePath \"\"" Oct 14 09:31:18 crc kubenswrapper[4870]: I1014 09:31:18.583164 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-openstack-openstack-networker-7fknh" event={"ID":"6b3e50e6-d053-4f51-9a47-60dc660562a7","Type":"ContainerDied","Data":"f9414b88652437c981e66f991f11d5fd2cb1abe36f09f62ae45b0a5d9e671231"} Oct 14 09:31:18 crc kubenswrapper[4870]: I1014 09:31:18.583208 4870 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f9414b88652437c981e66f991f11d5fd2cb1abe36f09f62ae45b0a5d9e671231" Oct 14 09:31:18 crc kubenswrapper[4870]: I1014 09:31:18.583263 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-openstack-openstack-networker-7fknh" Oct 14 09:31:22 crc kubenswrapper[4870]: I1014 09:31:22.034950 4870 scope.go:117] "RemoveContainer" containerID="3eea07ce0e4dde9363f92f5f44f530c26c10b42e335dd2bafb6764d9cbdc203c" Oct 14 09:31:22 crc kubenswrapper[4870]: E1014 09:31:22.035737 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 09:31:36 crc kubenswrapper[4870]: I1014 09:31:36.035893 4870 scope.go:117] "RemoveContainer" containerID="3eea07ce0e4dde9363f92f5f44f530c26c10b42e335dd2bafb6764d9cbdc203c" Oct 14 09:31:36 crc kubenswrapper[4870]: E1014 09:31:36.037372 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 09:31:48 crc kubenswrapper[4870]: I1014 09:31:48.034413 4870 scope.go:117] "RemoveContainer" containerID="3eea07ce0e4dde9363f92f5f44f530c26c10b42e335dd2bafb6764d9cbdc203c" Oct 14 09:31:48 crc kubenswrapper[4870]: E1014 09:31:48.035346 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 09:31:57 crc kubenswrapper[4870]: I1014 09:31:57.131239 4870 generic.go:334] "Generic (PLEG): container finished" podID="793a3762-db45-4a91-a194-7710517917e8" containerID="ba1e52eaefd626d7e03def49c4bb954201894aefd1cf9e2e83a07ba3dbe4faa7" exitCode=0 Oct 14 09:31:57 crc kubenswrapper[4870]: I1014 09:31:57.131336 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-openstack-openstack-cell1-kh4dj" event={"ID":"793a3762-db45-4a91-a194-7710517917e8","Type":"ContainerDied","Data":"ba1e52eaefd626d7e03def49c4bb954201894aefd1cf9e2e83a07ba3dbe4faa7"} Oct 14 09:31:58 crc kubenswrapper[4870]: I1014 09:31:58.603167 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-openstack-openstack-cell1-kh4dj" Oct 14 09:31:58 crc kubenswrapper[4870]: I1014 09:31:58.650416 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/793a3762-db45-4a91-a194-7710517917e8-neutron-ovn-metadata-agent-neutron-config-0\") pod \"793a3762-db45-4a91-a194-7710517917e8\" (UID: \"793a3762-db45-4a91-a194-7710517917e8\") " Oct 14 09:31:58 crc kubenswrapper[4870]: I1014 09:31:58.650539 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/793a3762-db45-4a91-a194-7710517917e8-inventory\") pod \"793a3762-db45-4a91-a194-7710517917e8\" (UID: \"793a3762-db45-4a91-a194-7710517917e8\") " Oct 14 09:31:58 crc kubenswrapper[4870]: I1014 09:31:58.650640 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nz4jf\" (UniqueName: \"kubernetes.io/projected/793a3762-db45-4a91-a194-7710517917e8-kube-api-access-nz4jf\") pod \"793a3762-db45-4a91-a194-7710517917e8\" (UID: \"793a3762-db45-4a91-a194-7710517917e8\") " Oct 14 09:31:58 crc kubenswrapper[4870]: I1014 09:31:58.650825 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/793a3762-db45-4a91-a194-7710517917e8-ceph\") pod \"793a3762-db45-4a91-a194-7710517917e8\" (UID: \"793a3762-db45-4a91-a194-7710517917e8\") " Oct 14 09:31:58 crc kubenswrapper[4870]: I1014 09:31:58.650873 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/793a3762-db45-4a91-a194-7710517917e8-neutron-metadata-combined-ca-bundle\") pod \"793a3762-db45-4a91-a194-7710517917e8\" (UID: \"793a3762-db45-4a91-a194-7710517917e8\") " Oct 14 09:31:58 crc kubenswrapper[4870]: I1014 09:31:58.651057 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/793a3762-db45-4a91-a194-7710517917e8-nova-metadata-neutron-config-0\") pod \"793a3762-db45-4a91-a194-7710517917e8\" (UID: \"793a3762-db45-4a91-a194-7710517917e8\") " Oct 14 09:31:58 crc kubenswrapper[4870]: I1014 09:31:58.651109 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/793a3762-db45-4a91-a194-7710517917e8-ssh-key\") pod \"793a3762-db45-4a91-a194-7710517917e8\" (UID: \"793a3762-db45-4a91-a194-7710517917e8\") " Oct 14 09:31:58 crc kubenswrapper[4870]: I1014 09:31:58.657022 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/793a3762-db45-4a91-a194-7710517917e8-ceph" (OuterVolumeSpecName: "ceph") pod "793a3762-db45-4a91-a194-7710517917e8" (UID: "793a3762-db45-4a91-a194-7710517917e8"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:31:58 crc kubenswrapper[4870]: I1014 09:31:58.657231 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/793a3762-db45-4a91-a194-7710517917e8-kube-api-access-nz4jf" (OuterVolumeSpecName: "kube-api-access-nz4jf") pod "793a3762-db45-4a91-a194-7710517917e8" (UID: "793a3762-db45-4a91-a194-7710517917e8"). InnerVolumeSpecName "kube-api-access-nz4jf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:31:58 crc kubenswrapper[4870]: I1014 09:31:58.658049 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/793a3762-db45-4a91-a194-7710517917e8-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "793a3762-db45-4a91-a194-7710517917e8" (UID: "793a3762-db45-4a91-a194-7710517917e8"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:31:58 crc kubenswrapper[4870]: I1014 09:31:58.683748 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/793a3762-db45-4a91-a194-7710517917e8-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "793a3762-db45-4a91-a194-7710517917e8" (UID: "793a3762-db45-4a91-a194-7710517917e8"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:31:58 crc kubenswrapper[4870]: I1014 09:31:58.700569 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/793a3762-db45-4a91-a194-7710517917e8-inventory" (OuterVolumeSpecName: "inventory") pod "793a3762-db45-4a91-a194-7710517917e8" (UID: "793a3762-db45-4a91-a194-7710517917e8"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:31:58 crc kubenswrapper[4870]: I1014 09:31:58.709600 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/793a3762-db45-4a91-a194-7710517917e8-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "793a3762-db45-4a91-a194-7710517917e8" (UID: "793a3762-db45-4a91-a194-7710517917e8"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:31:58 crc kubenswrapper[4870]: I1014 09:31:58.709694 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/793a3762-db45-4a91-a194-7710517917e8-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "793a3762-db45-4a91-a194-7710517917e8" (UID: "793a3762-db45-4a91-a194-7710517917e8"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:31:58 crc kubenswrapper[4870]: I1014 09:31:58.753089 4870 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/793a3762-db45-4a91-a194-7710517917e8-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Oct 14 09:31:58 crc kubenswrapper[4870]: I1014 09:31:58.753129 4870 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/793a3762-db45-4a91-a194-7710517917e8-inventory\") on node \"crc\" DevicePath \"\"" Oct 14 09:31:58 crc kubenswrapper[4870]: I1014 09:31:58.753140 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nz4jf\" (UniqueName: \"kubernetes.io/projected/793a3762-db45-4a91-a194-7710517917e8-kube-api-access-nz4jf\") on node \"crc\" DevicePath \"\"" Oct 14 09:31:58 crc kubenswrapper[4870]: I1014 09:31:58.753150 4870 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/793a3762-db45-4a91-a194-7710517917e8-ceph\") on node \"crc\" DevicePath \"\"" Oct 14 09:31:58 crc kubenswrapper[4870]: I1014 09:31:58.753161 4870 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/793a3762-db45-4a91-a194-7710517917e8-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 09:31:58 crc kubenswrapper[4870]: I1014 09:31:58.753171 4870 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/793a3762-db45-4a91-a194-7710517917e8-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Oct 14 09:31:58 crc kubenswrapper[4870]: I1014 09:31:58.753180 4870 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/793a3762-db45-4a91-a194-7710517917e8-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 14 09:31:59 crc kubenswrapper[4870]: I1014 09:31:59.158809 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-openstack-openstack-cell1-kh4dj" event={"ID":"793a3762-db45-4a91-a194-7710517917e8","Type":"ContainerDied","Data":"b261d5bbfd861d532463daecf537d44483c7e95c13fa0089829271795ee5cf49"} Oct 14 09:31:59 crc kubenswrapper[4870]: I1014 09:31:59.159140 4870 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b261d5bbfd861d532463daecf537d44483c7e95c13fa0089829271795ee5cf49" Oct 14 09:31:59 crc kubenswrapper[4870]: I1014 09:31:59.158872 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-openstack-openstack-cell1-kh4dj" Oct 14 09:31:59 crc kubenswrapper[4870]: I1014 09:31:59.301714 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-openstack-openstack-cell1-r6f6j"] Oct 14 09:31:59 crc kubenswrapper[4870]: E1014 09:31:59.302324 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b3e50e6-d053-4f51-9a47-60dc660562a7" containerName="neutron-metadata-openstack-openstack-networker" Oct 14 09:31:59 crc kubenswrapper[4870]: I1014 09:31:59.302352 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b3e50e6-d053-4f51-9a47-60dc660562a7" containerName="neutron-metadata-openstack-openstack-networker" Oct 14 09:31:59 crc kubenswrapper[4870]: E1014 09:31:59.302396 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="793a3762-db45-4a91-a194-7710517917e8" containerName="neutron-metadata-openstack-openstack-cell1" Oct 14 09:31:59 crc kubenswrapper[4870]: I1014 09:31:59.302404 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="793a3762-db45-4a91-a194-7710517917e8" containerName="neutron-metadata-openstack-openstack-cell1" Oct 14 09:31:59 crc kubenswrapper[4870]: I1014 09:31:59.302666 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="793a3762-db45-4a91-a194-7710517917e8" containerName="neutron-metadata-openstack-openstack-cell1" Oct 14 09:31:59 crc kubenswrapper[4870]: I1014 09:31:59.302698 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="6b3e50e6-d053-4f51-9a47-60dc660562a7" containerName="neutron-metadata-openstack-openstack-networker" Oct 14 09:31:59 crc kubenswrapper[4870]: I1014 09:31:59.303782 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-openstack-openstack-cell1-r6f6j" Oct 14 09:31:59 crc kubenswrapper[4870]: I1014 09:31:59.348471 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Oct 14 09:31:59 crc kubenswrapper[4870]: I1014 09:31:59.348561 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 14 09:31:59 crc kubenswrapper[4870]: I1014 09:31:59.348688 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 14 09:31:59 crc kubenswrapper[4870]: I1014 09:31:59.348744 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 14 09:31:59 crc kubenswrapper[4870]: I1014 09:31:59.348831 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-wrz7w" Oct 14 09:31:59 crc kubenswrapper[4870]: I1014 09:31:59.362727 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-openstack-openstack-cell1-r6f6j"] Oct 14 09:31:59 crc kubenswrapper[4870]: I1014 09:31:59.465683 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6795067a-45b7-42d4-abbb-44a3f094c02e-libvirt-combined-ca-bundle\") pod \"libvirt-openstack-openstack-cell1-r6f6j\" (UID: \"6795067a-45b7-42d4-abbb-44a3f094c02e\") " pod="openstack/libvirt-openstack-openstack-cell1-r6f6j" Oct 14 09:31:59 crc kubenswrapper[4870]: I1014 09:31:59.465755 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/6795067a-45b7-42d4-abbb-44a3f094c02e-libvirt-secret-0\") pod \"libvirt-openstack-openstack-cell1-r6f6j\" (UID: \"6795067a-45b7-42d4-abbb-44a3f094c02e\") " pod="openstack/libvirt-openstack-openstack-cell1-r6f6j" Oct 14 09:31:59 crc kubenswrapper[4870]: I1014 09:31:59.465846 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6795067a-45b7-42d4-abbb-44a3f094c02e-ssh-key\") pod \"libvirt-openstack-openstack-cell1-r6f6j\" (UID: \"6795067a-45b7-42d4-abbb-44a3f094c02e\") " pod="openstack/libvirt-openstack-openstack-cell1-r6f6j" Oct 14 09:31:59 crc kubenswrapper[4870]: I1014 09:31:59.466055 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6795067a-45b7-42d4-abbb-44a3f094c02e-inventory\") pod \"libvirt-openstack-openstack-cell1-r6f6j\" (UID: \"6795067a-45b7-42d4-abbb-44a3f094c02e\") " pod="openstack/libvirt-openstack-openstack-cell1-r6f6j" Oct 14 09:31:59 crc kubenswrapper[4870]: I1014 09:31:59.466159 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w5v4x\" (UniqueName: \"kubernetes.io/projected/6795067a-45b7-42d4-abbb-44a3f094c02e-kube-api-access-w5v4x\") pod \"libvirt-openstack-openstack-cell1-r6f6j\" (UID: \"6795067a-45b7-42d4-abbb-44a3f094c02e\") " pod="openstack/libvirt-openstack-openstack-cell1-r6f6j" Oct 14 09:31:59 crc kubenswrapper[4870]: I1014 09:31:59.466244 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/6795067a-45b7-42d4-abbb-44a3f094c02e-ceph\") pod \"libvirt-openstack-openstack-cell1-r6f6j\" (UID: \"6795067a-45b7-42d4-abbb-44a3f094c02e\") " pod="openstack/libvirt-openstack-openstack-cell1-r6f6j" Oct 14 09:31:59 crc kubenswrapper[4870]: I1014 09:31:59.567788 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6795067a-45b7-42d4-abbb-44a3f094c02e-inventory\") pod \"libvirt-openstack-openstack-cell1-r6f6j\" (UID: \"6795067a-45b7-42d4-abbb-44a3f094c02e\") " pod="openstack/libvirt-openstack-openstack-cell1-r6f6j" Oct 14 09:31:59 crc kubenswrapper[4870]: I1014 09:31:59.567853 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w5v4x\" (UniqueName: \"kubernetes.io/projected/6795067a-45b7-42d4-abbb-44a3f094c02e-kube-api-access-w5v4x\") pod \"libvirt-openstack-openstack-cell1-r6f6j\" (UID: \"6795067a-45b7-42d4-abbb-44a3f094c02e\") " pod="openstack/libvirt-openstack-openstack-cell1-r6f6j" Oct 14 09:31:59 crc kubenswrapper[4870]: I1014 09:31:59.567900 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/6795067a-45b7-42d4-abbb-44a3f094c02e-ceph\") pod \"libvirt-openstack-openstack-cell1-r6f6j\" (UID: \"6795067a-45b7-42d4-abbb-44a3f094c02e\") " pod="openstack/libvirt-openstack-openstack-cell1-r6f6j" Oct 14 09:31:59 crc kubenswrapper[4870]: I1014 09:31:59.567936 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6795067a-45b7-42d4-abbb-44a3f094c02e-libvirt-combined-ca-bundle\") pod \"libvirt-openstack-openstack-cell1-r6f6j\" (UID: \"6795067a-45b7-42d4-abbb-44a3f094c02e\") " pod="openstack/libvirt-openstack-openstack-cell1-r6f6j" Oct 14 09:31:59 crc kubenswrapper[4870]: I1014 09:31:59.567964 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/6795067a-45b7-42d4-abbb-44a3f094c02e-libvirt-secret-0\") pod \"libvirt-openstack-openstack-cell1-r6f6j\" (UID: \"6795067a-45b7-42d4-abbb-44a3f094c02e\") " pod="openstack/libvirt-openstack-openstack-cell1-r6f6j" Oct 14 09:31:59 crc kubenswrapper[4870]: I1014 09:31:59.568001 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6795067a-45b7-42d4-abbb-44a3f094c02e-ssh-key\") pod \"libvirt-openstack-openstack-cell1-r6f6j\" (UID: \"6795067a-45b7-42d4-abbb-44a3f094c02e\") " pod="openstack/libvirt-openstack-openstack-cell1-r6f6j" Oct 14 09:31:59 crc kubenswrapper[4870]: I1014 09:31:59.572844 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6795067a-45b7-42d4-abbb-44a3f094c02e-inventory\") pod \"libvirt-openstack-openstack-cell1-r6f6j\" (UID: \"6795067a-45b7-42d4-abbb-44a3f094c02e\") " pod="openstack/libvirt-openstack-openstack-cell1-r6f6j" Oct 14 09:31:59 crc kubenswrapper[4870]: I1014 09:31:59.573385 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/6795067a-45b7-42d4-abbb-44a3f094c02e-ceph\") pod \"libvirt-openstack-openstack-cell1-r6f6j\" (UID: \"6795067a-45b7-42d4-abbb-44a3f094c02e\") " pod="openstack/libvirt-openstack-openstack-cell1-r6f6j" Oct 14 09:31:59 crc kubenswrapper[4870]: I1014 09:31:59.573669 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6795067a-45b7-42d4-abbb-44a3f094c02e-ssh-key\") pod \"libvirt-openstack-openstack-cell1-r6f6j\" (UID: \"6795067a-45b7-42d4-abbb-44a3f094c02e\") " pod="openstack/libvirt-openstack-openstack-cell1-r6f6j" Oct 14 09:31:59 crc kubenswrapper[4870]: I1014 09:31:59.575845 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/6795067a-45b7-42d4-abbb-44a3f094c02e-libvirt-secret-0\") pod \"libvirt-openstack-openstack-cell1-r6f6j\" (UID: \"6795067a-45b7-42d4-abbb-44a3f094c02e\") " pod="openstack/libvirt-openstack-openstack-cell1-r6f6j" Oct 14 09:31:59 crc kubenswrapper[4870]: I1014 09:31:59.580194 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6795067a-45b7-42d4-abbb-44a3f094c02e-libvirt-combined-ca-bundle\") pod \"libvirt-openstack-openstack-cell1-r6f6j\" (UID: \"6795067a-45b7-42d4-abbb-44a3f094c02e\") " pod="openstack/libvirt-openstack-openstack-cell1-r6f6j" Oct 14 09:31:59 crc kubenswrapper[4870]: I1014 09:31:59.600724 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w5v4x\" (UniqueName: \"kubernetes.io/projected/6795067a-45b7-42d4-abbb-44a3f094c02e-kube-api-access-w5v4x\") pod \"libvirt-openstack-openstack-cell1-r6f6j\" (UID: \"6795067a-45b7-42d4-abbb-44a3f094c02e\") " pod="openstack/libvirt-openstack-openstack-cell1-r6f6j" Oct 14 09:31:59 crc kubenswrapper[4870]: I1014 09:31:59.667039 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-openstack-openstack-cell1-r6f6j" Oct 14 09:32:00 crc kubenswrapper[4870]: I1014 09:32:00.221214 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-openstack-openstack-cell1-r6f6j"] Oct 14 09:32:01 crc kubenswrapper[4870]: I1014 09:32:01.183971 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-openstack-openstack-cell1-r6f6j" event={"ID":"6795067a-45b7-42d4-abbb-44a3f094c02e","Type":"ContainerStarted","Data":"2937f3d884a82899f4de7c8f1eddebfe52d69b5a4cf7af7e9f33164106594de3"} Oct 14 09:32:03 crc kubenswrapper[4870]: I1014 09:32:03.034982 4870 scope.go:117] "RemoveContainer" containerID="3eea07ce0e4dde9363f92f5f44f530c26c10b42e335dd2bafb6764d9cbdc203c" Oct 14 09:32:03 crc kubenswrapper[4870]: E1014 09:32:03.035921 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 09:32:05 crc kubenswrapper[4870]: I1014 09:32:05.872139 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 14 09:32:06 crc kubenswrapper[4870]: I1014 09:32:06.252607 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-openstack-openstack-cell1-r6f6j" event={"ID":"6795067a-45b7-42d4-abbb-44a3f094c02e","Type":"ContainerStarted","Data":"0adf87fcdabfd384b02d3cd45a0f41efe7d18a4c6e35e121cc1c4228790d7ce6"} Oct 14 09:32:06 crc kubenswrapper[4870]: I1014 09:32:06.272589 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-openstack-openstack-cell1-r6f6j" podStartSLOduration=1.621530671 podStartE2EDuration="7.272571517s" podCreationTimestamp="2025-10-14 09:31:59 +0000 UTC" firstStartedPulling="2025-10-14 09:32:00.218453665 +0000 UTC m=+9055.915814036" lastFinishedPulling="2025-10-14 09:32:05.869494521 +0000 UTC m=+9061.566854882" observedRunningTime="2025-10-14 09:32:06.271241004 +0000 UTC m=+9061.968601385" watchObservedRunningTime="2025-10-14 09:32:06.272571517 +0000 UTC m=+9061.969931888" Oct 14 09:32:18 crc kubenswrapper[4870]: I1014 09:32:18.034885 4870 scope.go:117] "RemoveContainer" containerID="3eea07ce0e4dde9363f92f5f44f530c26c10b42e335dd2bafb6764d9cbdc203c" Oct 14 09:32:18 crc kubenswrapper[4870]: E1014 09:32:18.036224 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 09:32:30 crc kubenswrapper[4870]: I1014 09:32:30.033801 4870 scope.go:117] "RemoveContainer" containerID="3eea07ce0e4dde9363f92f5f44f530c26c10b42e335dd2bafb6764d9cbdc203c" Oct 14 09:32:30 crc kubenswrapper[4870]: E1014 09:32:30.034801 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 09:32:41 crc kubenswrapper[4870]: I1014 09:32:41.034188 4870 scope.go:117] "RemoveContainer" containerID="3eea07ce0e4dde9363f92f5f44f530c26c10b42e335dd2bafb6764d9cbdc203c" Oct 14 09:32:41 crc kubenswrapper[4870]: E1014 09:32:41.035100 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 09:32:54 crc kubenswrapper[4870]: I1014 09:32:54.035411 4870 scope.go:117] "RemoveContainer" containerID="3eea07ce0e4dde9363f92f5f44f530c26c10b42e335dd2bafb6764d9cbdc203c" Oct 14 09:32:54 crc kubenswrapper[4870]: E1014 09:32:54.036567 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 09:33:06 crc kubenswrapper[4870]: I1014 09:33:06.035157 4870 scope.go:117] "RemoveContainer" containerID="3eea07ce0e4dde9363f92f5f44f530c26c10b42e335dd2bafb6764d9cbdc203c" Oct 14 09:33:06 crc kubenswrapper[4870]: E1014 09:33:06.036400 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 09:33:14 crc kubenswrapper[4870]: I1014 09:33:14.431514 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-cwhjl"] Oct 14 09:33:14 crc kubenswrapper[4870]: I1014 09:33:14.435167 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-cwhjl" Oct 14 09:33:14 crc kubenswrapper[4870]: I1014 09:33:14.464697 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-cwhjl"] Oct 14 09:33:14 crc kubenswrapper[4870]: I1014 09:33:14.531827 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/957f8609-659c-4a09-b47b-93fa10cb89e4-utilities\") pod \"redhat-operators-cwhjl\" (UID: \"957f8609-659c-4a09-b47b-93fa10cb89e4\") " pod="openshift-marketplace/redhat-operators-cwhjl" Oct 14 09:33:14 crc kubenswrapper[4870]: I1014 09:33:14.531891 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9f2cr\" (UniqueName: \"kubernetes.io/projected/957f8609-659c-4a09-b47b-93fa10cb89e4-kube-api-access-9f2cr\") pod \"redhat-operators-cwhjl\" (UID: \"957f8609-659c-4a09-b47b-93fa10cb89e4\") " pod="openshift-marketplace/redhat-operators-cwhjl" Oct 14 09:33:14 crc kubenswrapper[4870]: I1014 09:33:14.531980 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/957f8609-659c-4a09-b47b-93fa10cb89e4-catalog-content\") pod \"redhat-operators-cwhjl\" (UID: \"957f8609-659c-4a09-b47b-93fa10cb89e4\") " pod="openshift-marketplace/redhat-operators-cwhjl" Oct 14 09:33:14 crc kubenswrapper[4870]: I1014 09:33:14.633664 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/957f8609-659c-4a09-b47b-93fa10cb89e4-catalog-content\") pod \"redhat-operators-cwhjl\" (UID: \"957f8609-659c-4a09-b47b-93fa10cb89e4\") " pod="openshift-marketplace/redhat-operators-cwhjl" Oct 14 09:33:14 crc kubenswrapper[4870]: I1014 09:33:14.634112 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/957f8609-659c-4a09-b47b-93fa10cb89e4-utilities\") pod \"redhat-operators-cwhjl\" (UID: \"957f8609-659c-4a09-b47b-93fa10cb89e4\") " pod="openshift-marketplace/redhat-operators-cwhjl" Oct 14 09:33:14 crc kubenswrapper[4870]: I1014 09:33:14.634227 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9f2cr\" (UniqueName: \"kubernetes.io/projected/957f8609-659c-4a09-b47b-93fa10cb89e4-kube-api-access-9f2cr\") pod \"redhat-operators-cwhjl\" (UID: \"957f8609-659c-4a09-b47b-93fa10cb89e4\") " pod="openshift-marketplace/redhat-operators-cwhjl" Oct 14 09:33:14 crc kubenswrapper[4870]: I1014 09:33:14.634329 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/957f8609-659c-4a09-b47b-93fa10cb89e4-catalog-content\") pod \"redhat-operators-cwhjl\" (UID: \"957f8609-659c-4a09-b47b-93fa10cb89e4\") " pod="openshift-marketplace/redhat-operators-cwhjl" Oct 14 09:33:14 crc kubenswrapper[4870]: I1014 09:33:14.634925 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/957f8609-659c-4a09-b47b-93fa10cb89e4-utilities\") pod \"redhat-operators-cwhjl\" (UID: \"957f8609-659c-4a09-b47b-93fa10cb89e4\") " pod="openshift-marketplace/redhat-operators-cwhjl" Oct 14 09:33:14 crc kubenswrapper[4870]: I1014 09:33:14.667153 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9f2cr\" (UniqueName: \"kubernetes.io/projected/957f8609-659c-4a09-b47b-93fa10cb89e4-kube-api-access-9f2cr\") pod \"redhat-operators-cwhjl\" (UID: \"957f8609-659c-4a09-b47b-93fa10cb89e4\") " pod="openshift-marketplace/redhat-operators-cwhjl" Oct 14 09:33:14 crc kubenswrapper[4870]: I1014 09:33:14.770829 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-cwhjl" Oct 14 09:33:15 crc kubenswrapper[4870]: I1014 09:33:15.285252 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-cwhjl"] Oct 14 09:33:16 crc kubenswrapper[4870]: I1014 09:33:16.160386 4870 generic.go:334] "Generic (PLEG): container finished" podID="957f8609-659c-4a09-b47b-93fa10cb89e4" containerID="e137771eead78521779f19232a8f00341b9a25f42a124d3b198d646a6b5ab831" exitCode=0 Oct 14 09:33:16 crc kubenswrapper[4870]: I1014 09:33:16.160479 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cwhjl" event={"ID":"957f8609-659c-4a09-b47b-93fa10cb89e4","Type":"ContainerDied","Data":"e137771eead78521779f19232a8f00341b9a25f42a124d3b198d646a6b5ab831"} Oct 14 09:33:16 crc kubenswrapper[4870]: I1014 09:33:16.160751 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cwhjl" event={"ID":"957f8609-659c-4a09-b47b-93fa10cb89e4","Type":"ContainerStarted","Data":"70c20db0269ad9feabb2ff09f13d620808861ecac89a71812acf0a127e996eab"} Oct 14 09:33:16 crc kubenswrapper[4870]: I1014 09:33:16.162640 4870 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 14 09:33:18 crc kubenswrapper[4870]: I1014 09:33:18.212018 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cwhjl" event={"ID":"957f8609-659c-4a09-b47b-93fa10cb89e4","Type":"ContainerStarted","Data":"958b5bef08164abf85d4f0453a2c2bb0a742ad40cc785ceb0e9f1803edeca858"} Oct 14 09:33:21 crc kubenswrapper[4870]: I1014 09:33:21.037367 4870 scope.go:117] "RemoveContainer" containerID="3eea07ce0e4dde9363f92f5f44f530c26c10b42e335dd2bafb6764d9cbdc203c" Oct 14 09:33:21 crc kubenswrapper[4870]: E1014 09:33:21.040570 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 09:33:21 crc kubenswrapper[4870]: I1014 09:33:21.253709 4870 generic.go:334] "Generic (PLEG): container finished" podID="957f8609-659c-4a09-b47b-93fa10cb89e4" containerID="958b5bef08164abf85d4f0453a2c2bb0a742ad40cc785ceb0e9f1803edeca858" exitCode=0 Oct 14 09:33:21 crc kubenswrapper[4870]: I1014 09:33:21.253753 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cwhjl" event={"ID":"957f8609-659c-4a09-b47b-93fa10cb89e4","Type":"ContainerDied","Data":"958b5bef08164abf85d4f0453a2c2bb0a742ad40cc785ceb0e9f1803edeca858"} Oct 14 09:33:22 crc kubenswrapper[4870]: I1014 09:33:22.268396 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cwhjl" event={"ID":"957f8609-659c-4a09-b47b-93fa10cb89e4","Type":"ContainerStarted","Data":"c6cd624c8b6d690699eac5f7d0197b7703a59de59fd3d948a049456f5d902fa4"} Oct 14 09:33:22 crc kubenswrapper[4870]: I1014 09:33:22.293647 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-cwhjl" podStartSLOduration=2.754833435 podStartE2EDuration="8.29363057s" podCreationTimestamp="2025-10-14 09:33:14 +0000 UTC" firstStartedPulling="2025-10-14 09:33:16.162418853 +0000 UTC m=+9131.859779224" lastFinishedPulling="2025-10-14 09:33:21.701215948 +0000 UTC m=+9137.398576359" observedRunningTime="2025-10-14 09:33:22.28466268 +0000 UTC m=+9137.982023071" watchObservedRunningTime="2025-10-14 09:33:22.29363057 +0000 UTC m=+9137.990990941" Oct 14 09:33:24 crc kubenswrapper[4870]: I1014 09:33:24.771948 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-cwhjl" Oct 14 09:33:24 crc kubenswrapper[4870]: I1014 09:33:24.772289 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-cwhjl" Oct 14 09:33:25 crc kubenswrapper[4870]: I1014 09:33:25.823267 4870 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-cwhjl" podUID="957f8609-659c-4a09-b47b-93fa10cb89e4" containerName="registry-server" probeResult="failure" output=< Oct 14 09:33:25 crc kubenswrapper[4870]: timeout: failed to connect service ":50051" within 1s Oct 14 09:33:25 crc kubenswrapper[4870]: > Oct 14 09:33:33 crc kubenswrapper[4870]: I1014 09:33:33.034243 4870 scope.go:117] "RemoveContainer" containerID="3eea07ce0e4dde9363f92f5f44f530c26c10b42e335dd2bafb6764d9cbdc203c" Oct 14 09:33:33 crc kubenswrapper[4870]: E1014 09:33:33.035177 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 09:33:34 crc kubenswrapper[4870]: I1014 09:33:34.907014 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-dgl2k"] Oct 14 09:33:34 crc kubenswrapper[4870]: I1014 09:33:34.910290 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dgl2k" Oct 14 09:33:34 crc kubenswrapper[4870]: I1014 09:33:34.918521 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-dgl2k"] Oct 14 09:33:35 crc kubenswrapper[4870]: I1014 09:33:35.035288 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/19d716ff-67a8-4872-8487-14ec65033f8b-utilities\") pod \"community-operators-dgl2k\" (UID: \"19d716ff-67a8-4872-8487-14ec65033f8b\") " pod="openshift-marketplace/community-operators-dgl2k" Oct 14 09:33:35 crc kubenswrapper[4870]: I1014 09:33:35.035353 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rz485\" (UniqueName: \"kubernetes.io/projected/19d716ff-67a8-4872-8487-14ec65033f8b-kube-api-access-rz485\") pod \"community-operators-dgl2k\" (UID: \"19d716ff-67a8-4872-8487-14ec65033f8b\") " pod="openshift-marketplace/community-operators-dgl2k" Oct 14 09:33:35 crc kubenswrapper[4870]: I1014 09:33:35.035496 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/19d716ff-67a8-4872-8487-14ec65033f8b-catalog-content\") pod \"community-operators-dgl2k\" (UID: \"19d716ff-67a8-4872-8487-14ec65033f8b\") " pod="openshift-marketplace/community-operators-dgl2k" Oct 14 09:33:35 crc kubenswrapper[4870]: I1014 09:33:35.137270 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/19d716ff-67a8-4872-8487-14ec65033f8b-catalog-content\") pod \"community-operators-dgl2k\" (UID: \"19d716ff-67a8-4872-8487-14ec65033f8b\") " pod="openshift-marketplace/community-operators-dgl2k" Oct 14 09:33:35 crc kubenswrapper[4870]: I1014 09:33:35.137380 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/19d716ff-67a8-4872-8487-14ec65033f8b-utilities\") pod \"community-operators-dgl2k\" (UID: \"19d716ff-67a8-4872-8487-14ec65033f8b\") " pod="openshift-marketplace/community-operators-dgl2k" Oct 14 09:33:35 crc kubenswrapper[4870]: I1014 09:33:35.137419 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rz485\" (UniqueName: \"kubernetes.io/projected/19d716ff-67a8-4872-8487-14ec65033f8b-kube-api-access-rz485\") pod \"community-operators-dgl2k\" (UID: \"19d716ff-67a8-4872-8487-14ec65033f8b\") " pod="openshift-marketplace/community-operators-dgl2k" Oct 14 09:33:35 crc kubenswrapper[4870]: I1014 09:33:35.137822 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/19d716ff-67a8-4872-8487-14ec65033f8b-catalog-content\") pod \"community-operators-dgl2k\" (UID: \"19d716ff-67a8-4872-8487-14ec65033f8b\") " pod="openshift-marketplace/community-operators-dgl2k" Oct 14 09:33:35 crc kubenswrapper[4870]: I1014 09:33:35.137953 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/19d716ff-67a8-4872-8487-14ec65033f8b-utilities\") pod \"community-operators-dgl2k\" (UID: \"19d716ff-67a8-4872-8487-14ec65033f8b\") " pod="openshift-marketplace/community-operators-dgl2k" Oct 14 09:33:35 crc kubenswrapper[4870]: I1014 09:33:35.154314 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rz485\" (UniqueName: \"kubernetes.io/projected/19d716ff-67a8-4872-8487-14ec65033f8b-kube-api-access-rz485\") pod \"community-operators-dgl2k\" (UID: \"19d716ff-67a8-4872-8487-14ec65033f8b\") " pod="openshift-marketplace/community-operators-dgl2k" Oct 14 09:33:35 crc kubenswrapper[4870]: I1014 09:33:35.235618 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dgl2k" Oct 14 09:33:35 crc kubenswrapper[4870]: I1014 09:33:35.834327 4870 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-cwhjl" podUID="957f8609-659c-4a09-b47b-93fa10cb89e4" containerName="registry-server" probeResult="failure" output=< Oct 14 09:33:35 crc kubenswrapper[4870]: timeout: failed to connect service ":50051" within 1s Oct 14 09:33:35 crc kubenswrapper[4870]: > Oct 14 09:33:35 crc kubenswrapper[4870]: I1014 09:33:35.862345 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-dgl2k"] Oct 14 09:33:36 crc kubenswrapper[4870]: I1014 09:33:36.448215 4870 generic.go:334] "Generic (PLEG): container finished" podID="19d716ff-67a8-4872-8487-14ec65033f8b" containerID="50afdb9f65a786df201ea9c98707cc2cd063fe7459a359bbe41b131a844d1d98" exitCode=0 Oct 14 09:33:36 crc kubenswrapper[4870]: I1014 09:33:36.448333 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dgl2k" event={"ID":"19d716ff-67a8-4872-8487-14ec65033f8b","Type":"ContainerDied","Data":"50afdb9f65a786df201ea9c98707cc2cd063fe7459a359bbe41b131a844d1d98"} Oct 14 09:33:36 crc kubenswrapper[4870]: I1014 09:33:36.448612 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dgl2k" event={"ID":"19d716ff-67a8-4872-8487-14ec65033f8b","Type":"ContainerStarted","Data":"49e13e7c0689c432b1f0f5b43b2e0dce0e93a079d48892eaa0eaa9e33f6b3fb6"} Oct 14 09:33:37 crc kubenswrapper[4870]: I1014 09:33:37.461169 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dgl2k" event={"ID":"19d716ff-67a8-4872-8487-14ec65033f8b","Type":"ContainerStarted","Data":"2c3dab145e8b438e528c45ef40018c6d4f57ce215542e4e6e7b56ad89c6a756f"} Oct 14 09:33:39 crc kubenswrapper[4870]: I1014 09:33:39.488055 4870 generic.go:334] "Generic (PLEG): container finished" podID="19d716ff-67a8-4872-8487-14ec65033f8b" containerID="2c3dab145e8b438e528c45ef40018c6d4f57ce215542e4e6e7b56ad89c6a756f" exitCode=0 Oct 14 09:33:39 crc kubenswrapper[4870]: I1014 09:33:39.488113 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dgl2k" event={"ID":"19d716ff-67a8-4872-8487-14ec65033f8b","Type":"ContainerDied","Data":"2c3dab145e8b438e528c45ef40018c6d4f57ce215542e4e6e7b56ad89c6a756f"} Oct 14 09:33:40 crc kubenswrapper[4870]: I1014 09:33:40.499579 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dgl2k" event={"ID":"19d716ff-67a8-4872-8487-14ec65033f8b","Type":"ContainerStarted","Data":"e30ff4ac70e7c7db2a20e6400aa3d01e3119e9bf5209951eeb64b6387c73dc75"} Oct 14 09:33:44 crc kubenswrapper[4870]: I1014 09:33:44.835053 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-cwhjl" Oct 14 09:33:44 crc kubenswrapper[4870]: I1014 09:33:44.860906 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-dgl2k" podStartSLOduration=7.180254952 podStartE2EDuration="10.860884079s" podCreationTimestamp="2025-10-14 09:33:34 +0000 UTC" firstStartedPulling="2025-10-14 09:33:36.451053906 +0000 UTC m=+9152.148414297" lastFinishedPulling="2025-10-14 09:33:40.131683053 +0000 UTC m=+9155.829043424" observedRunningTime="2025-10-14 09:33:40.523364965 +0000 UTC m=+9156.220725336" watchObservedRunningTime="2025-10-14 09:33:44.860884079 +0000 UTC m=+9160.558244460" Oct 14 09:33:44 crc kubenswrapper[4870]: I1014 09:33:44.893176 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-cwhjl" Oct 14 09:33:45 crc kubenswrapper[4870]: I1014 09:33:45.236260 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-dgl2k" Oct 14 09:33:45 crc kubenswrapper[4870]: I1014 09:33:45.236326 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-dgl2k" Oct 14 09:33:45 crc kubenswrapper[4870]: I1014 09:33:45.629686 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-cwhjl"] Oct 14 09:33:46 crc kubenswrapper[4870]: I1014 09:33:46.301396 4870 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-dgl2k" podUID="19d716ff-67a8-4872-8487-14ec65033f8b" containerName="registry-server" probeResult="failure" output=< Oct 14 09:33:46 crc kubenswrapper[4870]: timeout: failed to connect service ":50051" within 1s Oct 14 09:33:46 crc kubenswrapper[4870]: > Oct 14 09:33:46 crc kubenswrapper[4870]: I1014 09:33:46.577622 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-cwhjl" podUID="957f8609-659c-4a09-b47b-93fa10cb89e4" containerName="registry-server" containerID="cri-o://c6cd624c8b6d690699eac5f7d0197b7703a59de59fd3d948a049456f5d902fa4" gracePeriod=2 Oct 14 09:33:47 crc kubenswrapper[4870]: I1014 09:33:47.035407 4870 scope.go:117] "RemoveContainer" containerID="3eea07ce0e4dde9363f92f5f44f530c26c10b42e335dd2bafb6764d9cbdc203c" Oct 14 09:33:47 crc kubenswrapper[4870]: E1014 09:33:47.035945 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 09:33:47 crc kubenswrapper[4870]: I1014 09:33:47.194645 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-cwhjl" Oct 14 09:33:47 crc kubenswrapper[4870]: I1014 09:33:47.341588 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9f2cr\" (UniqueName: \"kubernetes.io/projected/957f8609-659c-4a09-b47b-93fa10cb89e4-kube-api-access-9f2cr\") pod \"957f8609-659c-4a09-b47b-93fa10cb89e4\" (UID: \"957f8609-659c-4a09-b47b-93fa10cb89e4\") " Oct 14 09:33:47 crc kubenswrapper[4870]: I1014 09:33:47.342984 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/957f8609-659c-4a09-b47b-93fa10cb89e4-utilities\") pod \"957f8609-659c-4a09-b47b-93fa10cb89e4\" (UID: \"957f8609-659c-4a09-b47b-93fa10cb89e4\") " Oct 14 09:33:47 crc kubenswrapper[4870]: I1014 09:33:47.343103 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/957f8609-659c-4a09-b47b-93fa10cb89e4-catalog-content\") pod \"957f8609-659c-4a09-b47b-93fa10cb89e4\" (UID: \"957f8609-659c-4a09-b47b-93fa10cb89e4\") " Oct 14 09:33:47 crc kubenswrapper[4870]: I1014 09:33:47.344866 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/957f8609-659c-4a09-b47b-93fa10cb89e4-utilities" (OuterVolumeSpecName: "utilities") pod "957f8609-659c-4a09-b47b-93fa10cb89e4" (UID: "957f8609-659c-4a09-b47b-93fa10cb89e4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 09:33:47 crc kubenswrapper[4870]: I1014 09:33:47.361338 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/957f8609-659c-4a09-b47b-93fa10cb89e4-kube-api-access-9f2cr" (OuterVolumeSpecName: "kube-api-access-9f2cr") pod "957f8609-659c-4a09-b47b-93fa10cb89e4" (UID: "957f8609-659c-4a09-b47b-93fa10cb89e4"). InnerVolumeSpecName "kube-api-access-9f2cr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:33:47 crc kubenswrapper[4870]: I1014 09:33:47.435791 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/957f8609-659c-4a09-b47b-93fa10cb89e4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "957f8609-659c-4a09-b47b-93fa10cb89e4" (UID: "957f8609-659c-4a09-b47b-93fa10cb89e4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 09:33:47 crc kubenswrapper[4870]: I1014 09:33:47.446897 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9f2cr\" (UniqueName: \"kubernetes.io/projected/957f8609-659c-4a09-b47b-93fa10cb89e4-kube-api-access-9f2cr\") on node \"crc\" DevicePath \"\"" Oct 14 09:33:47 crc kubenswrapper[4870]: I1014 09:33:47.446946 4870 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/957f8609-659c-4a09-b47b-93fa10cb89e4-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 09:33:47 crc kubenswrapper[4870]: I1014 09:33:47.446966 4870 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/957f8609-659c-4a09-b47b-93fa10cb89e4-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 09:33:47 crc kubenswrapper[4870]: I1014 09:33:47.595003 4870 generic.go:334] "Generic (PLEG): container finished" podID="957f8609-659c-4a09-b47b-93fa10cb89e4" containerID="c6cd624c8b6d690699eac5f7d0197b7703a59de59fd3d948a049456f5d902fa4" exitCode=0 Oct 14 09:33:47 crc kubenswrapper[4870]: I1014 09:33:47.595217 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cwhjl" event={"ID":"957f8609-659c-4a09-b47b-93fa10cb89e4","Type":"ContainerDied","Data":"c6cd624c8b6d690699eac5f7d0197b7703a59de59fd3d948a049456f5d902fa4"} Oct 14 09:33:47 crc kubenswrapper[4870]: I1014 09:33:47.595385 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-cwhjl" Oct 14 09:33:47 crc kubenswrapper[4870]: I1014 09:33:47.595689 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cwhjl" event={"ID":"957f8609-659c-4a09-b47b-93fa10cb89e4","Type":"ContainerDied","Data":"70c20db0269ad9feabb2ff09f13d620808861ecac89a71812acf0a127e996eab"} Oct 14 09:33:47 crc kubenswrapper[4870]: I1014 09:33:47.595733 4870 scope.go:117] "RemoveContainer" containerID="c6cd624c8b6d690699eac5f7d0197b7703a59de59fd3d948a049456f5d902fa4" Oct 14 09:33:47 crc kubenswrapper[4870]: I1014 09:33:47.639201 4870 scope.go:117] "RemoveContainer" containerID="958b5bef08164abf85d4f0453a2c2bb0a742ad40cc785ceb0e9f1803edeca858" Oct 14 09:33:47 crc kubenswrapper[4870]: I1014 09:33:47.648508 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-cwhjl"] Oct 14 09:33:47 crc kubenswrapper[4870]: I1014 09:33:47.661498 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-cwhjl"] Oct 14 09:33:47 crc kubenswrapper[4870]: I1014 09:33:47.676611 4870 scope.go:117] "RemoveContainer" containerID="e137771eead78521779f19232a8f00341b9a25f42a124d3b198d646a6b5ab831" Oct 14 09:33:47 crc kubenswrapper[4870]: I1014 09:33:47.733620 4870 scope.go:117] "RemoveContainer" containerID="c6cd624c8b6d690699eac5f7d0197b7703a59de59fd3d948a049456f5d902fa4" Oct 14 09:33:47 crc kubenswrapper[4870]: E1014 09:33:47.734341 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c6cd624c8b6d690699eac5f7d0197b7703a59de59fd3d948a049456f5d902fa4\": container with ID starting with c6cd624c8b6d690699eac5f7d0197b7703a59de59fd3d948a049456f5d902fa4 not found: ID does not exist" containerID="c6cd624c8b6d690699eac5f7d0197b7703a59de59fd3d948a049456f5d902fa4" Oct 14 09:33:47 crc kubenswrapper[4870]: I1014 09:33:47.734392 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c6cd624c8b6d690699eac5f7d0197b7703a59de59fd3d948a049456f5d902fa4"} err="failed to get container status \"c6cd624c8b6d690699eac5f7d0197b7703a59de59fd3d948a049456f5d902fa4\": rpc error: code = NotFound desc = could not find container \"c6cd624c8b6d690699eac5f7d0197b7703a59de59fd3d948a049456f5d902fa4\": container with ID starting with c6cd624c8b6d690699eac5f7d0197b7703a59de59fd3d948a049456f5d902fa4 not found: ID does not exist" Oct 14 09:33:47 crc kubenswrapper[4870]: I1014 09:33:47.734422 4870 scope.go:117] "RemoveContainer" containerID="958b5bef08164abf85d4f0453a2c2bb0a742ad40cc785ceb0e9f1803edeca858" Oct 14 09:33:47 crc kubenswrapper[4870]: E1014 09:33:47.734989 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"958b5bef08164abf85d4f0453a2c2bb0a742ad40cc785ceb0e9f1803edeca858\": container with ID starting with 958b5bef08164abf85d4f0453a2c2bb0a742ad40cc785ceb0e9f1803edeca858 not found: ID does not exist" containerID="958b5bef08164abf85d4f0453a2c2bb0a742ad40cc785ceb0e9f1803edeca858" Oct 14 09:33:47 crc kubenswrapper[4870]: I1014 09:33:47.735017 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"958b5bef08164abf85d4f0453a2c2bb0a742ad40cc785ceb0e9f1803edeca858"} err="failed to get container status \"958b5bef08164abf85d4f0453a2c2bb0a742ad40cc785ceb0e9f1803edeca858\": rpc error: code = NotFound desc = could not find container \"958b5bef08164abf85d4f0453a2c2bb0a742ad40cc785ceb0e9f1803edeca858\": container with ID starting with 958b5bef08164abf85d4f0453a2c2bb0a742ad40cc785ceb0e9f1803edeca858 not found: ID does not exist" Oct 14 09:33:47 crc kubenswrapper[4870]: I1014 09:33:47.735037 4870 scope.go:117] "RemoveContainer" containerID="e137771eead78521779f19232a8f00341b9a25f42a124d3b198d646a6b5ab831" Oct 14 09:33:47 crc kubenswrapper[4870]: E1014 09:33:47.735408 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e137771eead78521779f19232a8f00341b9a25f42a124d3b198d646a6b5ab831\": container with ID starting with e137771eead78521779f19232a8f00341b9a25f42a124d3b198d646a6b5ab831 not found: ID does not exist" containerID="e137771eead78521779f19232a8f00341b9a25f42a124d3b198d646a6b5ab831" Oct 14 09:33:47 crc kubenswrapper[4870]: I1014 09:33:47.735429 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e137771eead78521779f19232a8f00341b9a25f42a124d3b198d646a6b5ab831"} err="failed to get container status \"e137771eead78521779f19232a8f00341b9a25f42a124d3b198d646a6b5ab831\": rpc error: code = NotFound desc = could not find container \"e137771eead78521779f19232a8f00341b9a25f42a124d3b198d646a6b5ab831\": container with ID starting with e137771eead78521779f19232a8f00341b9a25f42a124d3b198d646a6b5ab831 not found: ID does not exist" Oct 14 09:33:49 crc kubenswrapper[4870]: I1014 09:33:49.044796 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="957f8609-659c-4a09-b47b-93fa10cb89e4" path="/var/lib/kubelet/pods/957f8609-659c-4a09-b47b-93fa10cb89e4/volumes" Oct 14 09:33:55 crc kubenswrapper[4870]: I1014 09:33:55.304555 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-dgl2k" Oct 14 09:33:55 crc kubenswrapper[4870]: I1014 09:33:55.373414 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-dgl2k" Oct 14 09:33:55 crc kubenswrapper[4870]: I1014 09:33:55.560318 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-dgl2k"] Oct 14 09:33:56 crc kubenswrapper[4870]: I1014 09:33:56.707740 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-dgl2k" podUID="19d716ff-67a8-4872-8487-14ec65033f8b" containerName="registry-server" containerID="cri-o://e30ff4ac70e7c7db2a20e6400aa3d01e3119e9bf5209951eeb64b6387c73dc75" gracePeriod=2 Oct 14 09:33:57 crc kubenswrapper[4870]: I1014 09:33:57.297426 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dgl2k" Oct 14 09:33:57 crc kubenswrapper[4870]: I1014 09:33:57.482191 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/19d716ff-67a8-4872-8487-14ec65033f8b-utilities\") pod \"19d716ff-67a8-4872-8487-14ec65033f8b\" (UID: \"19d716ff-67a8-4872-8487-14ec65033f8b\") " Oct 14 09:33:57 crc kubenswrapper[4870]: I1014 09:33:57.482403 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/19d716ff-67a8-4872-8487-14ec65033f8b-catalog-content\") pod \"19d716ff-67a8-4872-8487-14ec65033f8b\" (UID: \"19d716ff-67a8-4872-8487-14ec65033f8b\") " Oct 14 09:33:57 crc kubenswrapper[4870]: I1014 09:33:57.482657 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rz485\" (UniqueName: \"kubernetes.io/projected/19d716ff-67a8-4872-8487-14ec65033f8b-kube-api-access-rz485\") pod \"19d716ff-67a8-4872-8487-14ec65033f8b\" (UID: \"19d716ff-67a8-4872-8487-14ec65033f8b\") " Oct 14 09:33:57 crc kubenswrapper[4870]: I1014 09:33:57.483332 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/19d716ff-67a8-4872-8487-14ec65033f8b-utilities" (OuterVolumeSpecName: "utilities") pod "19d716ff-67a8-4872-8487-14ec65033f8b" (UID: "19d716ff-67a8-4872-8487-14ec65033f8b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 09:33:57 crc kubenswrapper[4870]: I1014 09:33:57.483817 4870 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/19d716ff-67a8-4872-8487-14ec65033f8b-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 09:33:57 crc kubenswrapper[4870]: I1014 09:33:57.488995 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/19d716ff-67a8-4872-8487-14ec65033f8b-kube-api-access-rz485" (OuterVolumeSpecName: "kube-api-access-rz485") pod "19d716ff-67a8-4872-8487-14ec65033f8b" (UID: "19d716ff-67a8-4872-8487-14ec65033f8b"). InnerVolumeSpecName "kube-api-access-rz485". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:33:57 crc kubenswrapper[4870]: I1014 09:33:57.564327 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/19d716ff-67a8-4872-8487-14ec65033f8b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "19d716ff-67a8-4872-8487-14ec65033f8b" (UID: "19d716ff-67a8-4872-8487-14ec65033f8b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 09:33:57 crc kubenswrapper[4870]: I1014 09:33:57.587670 4870 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/19d716ff-67a8-4872-8487-14ec65033f8b-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 09:33:57 crc kubenswrapper[4870]: I1014 09:33:57.587748 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rz485\" (UniqueName: \"kubernetes.io/projected/19d716ff-67a8-4872-8487-14ec65033f8b-kube-api-access-rz485\") on node \"crc\" DevicePath \"\"" Oct 14 09:33:57 crc kubenswrapper[4870]: I1014 09:33:57.727323 4870 generic.go:334] "Generic (PLEG): container finished" podID="19d716ff-67a8-4872-8487-14ec65033f8b" containerID="e30ff4ac70e7c7db2a20e6400aa3d01e3119e9bf5209951eeb64b6387c73dc75" exitCode=0 Oct 14 09:33:57 crc kubenswrapper[4870]: I1014 09:33:57.727390 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dgl2k" event={"ID":"19d716ff-67a8-4872-8487-14ec65033f8b","Type":"ContainerDied","Data":"e30ff4ac70e7c7db2a20e6400aa3d01e3119e9bf5209951eeb64b6387c73dc75"} Oct 14 09:33:57 crc kubenswrapper[4870]: I1014 09:33:57.727466 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dgl2k" event={"ID":"19d716ff-67a8-4872-8487-14ec65033f8b","Type":"ContainerDied","Data":"49e13e7c0689c432b1f0f5b43b2e0dce0e93a079d48892eaa0eaa9e33f6b3fb6"} Oct 14 09:33:57 crc kubenswrapper[4870]: I1014 09:33:57.727498 4870 scope.go:117] "RemoveContainer" containerID="e30ff4ac70e7c7db2a20e6400aa3d01e3119e9bf5209951eeb64b6387c73dc75" Oct 14 09:33:57 crc kubenswrapper[4870]: I1014 09:33:57.727540 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dgl2k" Oct 14 09:33:57 crc kubenswrapper[4870]: I1014 09:33:57.766113 4870 scope.go:117] "RemoveContainer" containerID="2c3dab145e8b438e528c45ef40018c6d4f57ce215542e4e6e7b56ad89c6a756f" Oct 14 09:33:57 crc kubenswrapper[4870]: I1014 09:33:57.798165 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-dgl2k"] Oct 14 09:33:57 crc kubenswrapper[4870]: I1014 09:33:57.813278 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-dgl2k"] Oct 14 09:33:57 crc kubenswrapper[4870]: I1014 09:33:57.815503 4870 scope.go:117] "RemoveContainer" containerID="50afdb9f65a786df201ea9c98707cc2cd063fe7459a359bbe41b131a844d1d98" Oct 14 09:33:57 crc kubenswrapper[4870]: I1014 09:33:57.837836 4870 scope.go:117] "RemoveContainer" containerID="e30ff4ac70e7c7db2a20e6400aa3d01e3119e9bf5209951eeb64b6387c73dc75" Oct 14 09:33:57 crc kubenswrapper[4870]: E1014 09:33:57.838265 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e30ff4ac70e7c7db2a20e6400aa3d01e3119e9bf5209951eeb64b6387c73dc75\": container with ID starting with e30ff4ac70e7c7db2a20e6400aa3d01e3119e9bf5209951eeb64b6387c73dc75 not found: ID does not exist" containerID="e30ff4ac70e7c7db2a20e6400aa3d01e3119e9bf5209951eeb64b6387c73dc75" Oct 14 09:33:57 crc kubenswrapper[4870]: I1014 09:33:57.838325 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e30ff4ac70e7c7db2a20e6400aa3d01e3119e9bf5209951eeb64b6387c73dc75"} err="failed to get container status \"e30ff4ac70e7c7db2a20e6400aa3d01e3119e9bf5209951eeb64b6387c73dc75\": rpc error: code = NotFound desc = could not find container \"e30ff4ac70e7c7db2a20e6400aa3d01e3119e9bf5209951eeb64b6387c73dc75\": container with ID starting with e30ff4ac70e7c7db2a20e6400aa3d01e3119e9bf5209951eeb64b6387c73dc75 not found: ID does not exist" Oct 14 09:33:57 crc kubenswrapper[4870]: I1014 09:33:57.838361 4870 scope.go:117] "RemoveContainer" containerID="2c3dab145e8b438e528c45ef40018c6d4f57ce215542e4e6e7b56ad89c6a756f" Oct 14 09:33:57 crc kubenswrapper[4870]: E1014 09:33:57.838708 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2c3dab145e8b438e528c45ef40018c6d4f57ce215542e4e6e7b56ad89c6a756f\": container with ID starting with 2c3dab145e8b438e528c45ef40018c6d4f57ce215542e4e6e7b56ad89c6a756f not found: ID does not exist" containerID="2c3dab145e8b438e528c45ef40018c6d4f57ce215542e4e6e7b56ad89c6a756f" Oct 14 09:33:57 crc kubenswrapper[4870]: I1014 09:33:57.838755 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2c3dab145e8b438e528c45ef40018c6d4f57ce215542e4e6e7b56ad89c6a756f"} err="failed to get container status \"2c3dab145e8b438e528c45ef40018c6d4f57ce215542e4e6e7b56ad89c6a756f\": rpc error: code = NotFound desc = could not find container \"2c3dab145e8b438e528c45ef40018c6d4f57ce215542e4e6e7b56ad89c6a756f\": container with ID starting with 2c3dab145e8b438e528c45ef40018c6d4f57ce215542e4e6e7b56ad89c6a756f not found: ID does not exist" Oct 14 09:33:57 crc kubenswrapper[4870]: I1014 09:33:57.838789 4870 scope.go:117] "RemoveContainer" containerID="50afdb9f65a786df201ea9c98707cc2cd063fe7459a359bbe41b131a844d1d98" Oct 14 09:33:57 crc kubenswrapper[4870]: E1014 09:33:57.839130 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"50afdb9f65a786df201ea9c98707cc2cd063fe7459a359bbe41b131a844d1d98\": container with ID starting with 50afdb9f65a786df201ea9c98707cc2cd063fe7459a359bbe41b131a844d1d98 not found: ID does not exist" containerID="50afdb9f65a786df201ea9c98707cc2cd063fe7459a359bbe41b131a844d1d98" Oct 14 09:33:57 crc kubenswrapper[4870]: I1014 09:33:57.839161 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"50afdb9f65a786df201ea9c98707cc2cd063fe7459a359bbe41b131a844d1d98"} err="failed to get container status \"50afdb9f65a786df201ea9c98707cc2cd063fe7459a359bbe41b131a844d1d98\": rpc error: code = NotFound desc = could not find container \"50afdb9f65a786df201ea9c98707cc2cd063fe7459a359bbe41b131a844d1d98\": container with ID starting with 50afdb9f65a786df201ea9c98707cc2cd063fe7459a359bbe41b131a844d1d98 not found: ID does not exist" Oct 14 09:33:58 crc kubenswrapper[4870]: I1014 09:33:58.033867 4870 scope.go:117] "RemoveContainer" containerID="3eea07ce0e4dde9363f92f5f44f530c26c10b42e335dd2bafb6764d9cbdc203c" Oct 14 09:33:58 crc kubenswrapper[4870]: E1014 09:33:58.034365 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 09:33:59 crc kubenswrapper[4870]: I1014 09:33:59.047574 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="19d716ff-67a8-4872-8487-14ec65033f8b" path="/var/lib/kubelet/pods/19d716ff-67a8-4872-8487-14ec65033f8b/volumes" Oct 14 09:34:13 crc kubenswrapper[4870]: I1014 09:34:13.034183 4870 scope.go:117] "RemoveContainer" containerID="3eea07ce0e4dde9363f92f5f44f530c26c10b42e335dd2bafb6764d9cbdc203c" Oct 14 09:34:13 crc kubenswrapper[4870]: E1014 09:34:13.035148 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 09:34:24 crc kubenswrapper[4870]: I1014 09:34:24.035087 4870 scope.go:117] "RemoveContainer" containerID="3eea07ce0e4dde9363f92f5f44f530c26c10b42e335dd2bafb6764d9cbdc203c" Oct 14 09:34:25 crc kubenswrapper[4870]: I1014 09:34:25.107513 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" event={"ID":"874a5e45-dffd-4d17-b609-b7d3ed2eab07","Type":"ContainerStarted","Data":"da933e0d9bd5082c26564c190825d7838bd95c38ef50de4f4eef55960b16e294"} Oct 14 09:36:53 crc kubenswrapper[4870]: I1014 09:36:53.951408 4870 patch_prober.go:28] interesting pod/machine-config-daemon-7tvc7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 09:36:53 crc kubenswrapper[4870]: I1014 09:36:53.953721 4870 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 09:37:17 crc kubenswrapper[4870]: I1014 09:37:17.346126 4870 generic.go:334] "Generic (PLEG): container finished" podID="6795067a-45b7-42d4-abbb-44a3f094c02e" containerID="0adf87fcdabfd384b02d3cd45a0f41efe7d18a4c6e35e121cc1c4228790d7ce6" exitCode=0 Oct 14 09:37:17 crc kubenswrapper[4870]: I1014 09:37:17.346261 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-openstack-openstack-cell1-r6f6j" event={"ID":"6795067a-45b7-42d4-abbb-44a3f094c02e","Type":"ContainerDied","Data":"0adf87fcdabfd384b02d3cd45a0f41efe7d18a4c6e35e121cc1c4228790d7ce6"} Oct 14 09:37:19 crc kubenswrapper[4870]: I1014 09:37:19.583297 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-openstack-openstack-cell1-r6f6j" Oct 14 09:37:19 crc kubenswrapper[4870]: I1014 09:37:19.749372 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/6795067a-45b7-42d4-abbb-44a3f094c02e-ceph\") pod \"6795067a-45b7-42d4-abbb-44a3f094c02e\" (UID: \"6795067a-45b7-42d4-abbb-44a3f094c02e\") " Oct 14 09:37:19 crc kubenswrapper[4870]: I1014 09:37:19.749553 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6795067a-45b7-42d4-abbb-44a3f094c02e-ssh-key\") pod \"6795067a-45b7-42d4-abbb-44a3f094c02e\" (UID: \"6795067a-45b7-42d4-abbb-44a3f094c02e\") " Oct 14 09:37:19 crc kubenswrapper[4870]: I1014 09:37:19.749662 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w5v4x\" (UniqueName: \"kubernetes.io/projected/6795067a-45b7-42d4-abbb-44a3f094c02e-kube-api-access-w5v4x\") pod \"6795067a-45b7-42d4-abbb-44a3f094c02e\" (UID: \"6795067a-45b7-42d4-abbb-44a3f094c02e\") " Oct 14 09:37:19 crc kubenswrapper[4870]: I1014 09:37:19.749786 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6795067a-45b7-42d4-abbb-44a3f094c02e-inventory\") pod \"6795067a-45b7-42d4-abbb-44a3f094c02e\" (UID: \"6795067a-45b7-42d4-abbb-44a3f094c02e\") " Oct 14 09:37:19 crc kubenswrapper[4870]: I1014 09:37:19.749841 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6795067a-45b7-42d4-abbb-44a3f094c02e-libvirt-combined-ca-bundle\") pod \"6795067a-45b7-42d4-abbb-44a3f094c02e\" (UID: \"6795067a-45b7-42d4-abbb-44a3f094c02e\") " Oct 14 09:37:19 crc kubenswrapper[4870]: I1014 09:37:19.749926 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/6795067a-45b7-42d4-abbb-44a3f094c02e-libvirt-secret-0\") pod \"6795067a-45b7-42d4-abbb-44a3f094c02e\" (UID: \"6795067a-45b7-42d4-abbb-44a3f094c02e\") " Oct 14 09:37:19 crc kubenswrapper[4870]: I1014 09:37:19.755931 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6795067a-45b7-42d4-abbb-44a3f094c02e-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "6795067a-45b7-42d4-abbb-44a3f094c02e" (UID: "6795067a-45b7-42d4-abbb-44a3f094c02e"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:37:19 crc kubenswrapper[4870]: I1014 09:37:19.756034 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6795067a-45b7-42d4-abbb-44a3f094c02e-ceph" (OuterVolumeSpecName: "ceph") pod "6795067a-45b7-42d4-abbb-44a3f094c02e" (UID: "6795067a-45b7-42d4-abbb-44a3f094c02e"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:37:19 crc kubenswrapper[4870]: I1014 09:37:19.769892 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6795067a-45b7-42d4-abbb-44a3f094c02e-kube-api-access-w5v4x" (OuterVolumeSpecName: "kube-api-access-w5v4x") pod "6795067a-45b7-42d4-abbb-44a3f094c02e" (UID: "6795067a-45b7-42d4-abbb-44a3f094c02e"). InnerVolumeSpecName "kube-api-access-w5v4x". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:37:19 crc kubenswrapper[4870]: I1014 09:37:19.783383 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6795067a-45b7-42d4-abbb-44a3f094c02e-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "6795067a-45b7-42d4-abbb-44a3f094c02e" (UID: "6795067a-45b7-42d4-abbb-44a3f094c02e"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:37:19 crc kubenswrapper[4870]: I1014 09:37:19.803290 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6795067a-45b7-42d4-abbb-44a3f094c02e-inventory" (OuterVolumeSpecName: "inventory") pod "6795067a-45b7-42d4-abbb-44a3f094c02e" (UID: "6795067a-45b7-42d4-abbb-44a3f094c02e"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:37:19 crc kubenswrapper[4870]: I1014 09:37:19.805953 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6795067a-45b7-42d4-abbb-44a3f094c02e-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "6795067a-45b7-42d4-abbb-44a3f094c02e" (UID: "6795067a-45b7-42d4-abbb-44a3f094c02e"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:37:19 crc kubenswrapper[4870]: I1014 09:37:19.853763 4870 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/6795067a-45b7-42d4-abbb-44a3f094c02e-ceph\") on node \"crc\" DevicePath \"\"" Oct 14 09:37:19 crc kubenswrapper[4870]: I1014 09:37:19.853818 4870 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6795067a-45b7-42d4-abbb-44a3f094c02e-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 14 09:37:19 crc kubenswrapper[4870]: I1014 09:37:19.853839 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w5v4x\" (UniqueName: \"kubernetes.io/projected/6795067a-45b7-42d4-abbb-44a3f094c02e-kube-api-access-w5v4x\") on node \"crc\" DevicePath \"\"" Oct 14 09:37:19 crc kubenswrapper[4870]: I1014 09:37:19.853859 4870 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6795067a-45b7-42d4-abbb-44a3f094c02e-inventory\") on node \"crc\" DevicePath \"\"" Oct 14 09:37:19 crc kubenswrapper[4870]: I1014 09:37:19.853877 4870 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6795067a-45b7-42d4-abbb-44a3f094c02e-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 09:37:19 crc kubenswrapper[4870]: I1014 09:37:19.853895 4870 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/6795067a-45b7-42d4-abbb-44a3f094c02e-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Oct 14 09:37:20 crc kubenswrapper[4870]: I1014 09:37:20.391574 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-openstack-openstack-cell1-r6f6j" Oct 14 09:37:20 crc kubenswrapper[4870]: I1014 09:37:20.391588 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-openstack-openstack-cell1-r6f6j" event={"ID":"6795067a-45b7-42d4-abbb-44a3f094c02e","Type":"ContainerDied","Data":"2937f3d884a82899f4de7c8f1eddebfe52d69b5a4cf7af7e9f33164106594de3"} Oct 14 09:37:20 crc kubenswrapper[4870]: I1014 09:37:20.392039 4870 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2937f3d884a82899f4de7c8f1eddebfe52d69b5a4cf7af7e9f33164106594de3" Oct 14 09:37:20 crc kubenswrapper[4870]: I1014 09:37:20.816771 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-openstack-openstack-cell1-z74pq"] Oct 14 09:37:20 crc kubenswrapper[4870]: E1014 09:37:20.817255 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="957f8609-659c-4a09-b47b-93fa10cb89e4" containerName="extract-content" Oct 14 09:37:20 crc kubenswrapper[4870]: I1014 09:37:20.817266 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="957f8609-659c-4a09-b47b-93fa10cb89e4" containerName="extract-content" Oct 14 09:37:20 crc kubenswrapper[4870]: E1014 09:37:20.817288 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6795067a-45b7-42d4-abbb-44a3f094c02e" containerName="libvirt-openstack-openstack-cell1" Oct 14 09:37:20 crc kubenswrapper[4870]: I1014 09:37:20.817294 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="6795067a-45b7-42d4-abbb-44a3f094c02e" containerName="libvirt-openstack-openstack-cell1" Oct 14 09:37:20 crc kubenswrapper[4870]: E1014 09:37:20.817306 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="957f8609-659c-4a09-b47b-93fa10cb89e4" containerName="extract-utilities" Oct 14 09:37:20 crc kubenswrapper[4870]: I1014 09:37:20.817313 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="957f8609-659c-4a09-b47b-93fa10cb89e4" containerName="extract-utilities" Oct 14 09:37:20 crc kubenswrapper[4870]: E1014 09:37:20.817325 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="957f8609-659c-4a09-b47b-93fa10cb89e4" containerName="registry-server" Oct 14 09:37:20 crc kubenswrapper[4870]: I1014 09:37:20.817331 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="957f8609-659c-4a09-b47b-93fa10cb89e4" containerName="registry-server" Oct 14 09:37:20 crc kubenswrapper[4870]: E1014 09:37:20.817355 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19d716ff-67a8-4872-8487-14ec65033f8b" containerName="registry-server" Oct 14 09:37:20 crc kubenswrapper[4870]: I1014 09:37:20.817360 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="19d716ff-67a8-4872-8487-14ec65033f8b" containerName="registry-server" Oct 14 09:37:20 crc kubenswrapper[4870]: E1014 09:37:20.817371 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19d716ff-67a8-4872-8487-14ec65033f8b" containerName="extract-content" Oct 14 09:37:20 crc kubenswrapper[4870]: I1014 09:37:20.817377 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="19d716ff-67a8-4872-8487-14ec65033f8b" containerName="extract-content" Oct 14 09:37:20 crc kubenswrapper[4870]: E1014 09:37:20.817391 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19d716ff-67a8-4872-8487-14ec65033f8b" containerName="extract-utilities" Oct 14 09:37:20 crc kubenswrapper[4870]: I1014 09:37:20.817396 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="19d716ff-67a8-4872-8487-14ec65033f8b" containerName="extract-utilities" Oct 14 09:37:20 crc kubenswrapper[4870]: I1014 09:37:20.817608 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="6795067a-45b7-42d4-abbb-44a3f094c02e" containerName="libvirt-openstack-openstack-cell1" Oct 14 09:37:20 crc kubenswrapper[4870]: I1014 09:37:20.817622 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="957f8609-659c-4a09-b47b-93fa10cb89e4" containerName="registry-server" Oct 14 09:37:20 crc kubenswrapper[4870]: I1014 09:37:20.817645 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="19d716ff-67a8-4872-8487-14ec65033f8b" containerName="registry-server" Oct 14 09:37:20 crc kubenswrapper[4870]: I1014 09:37:20.818526 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-openstack-cell1-z74pq" Oct 14 09:37:20 crc kubenswrapper[4870]: I1014 09:37:20.820994 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 14 09:37:20 crc kubenswrapper[4870]: I1014 09:37:20.821301 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 14 09:37:20 crc kubenswrapper[4870]: I1014 09:37:20.821523 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Oct 14 09:37:20 crc kubenswrapper[4870]: I1014 09:37:20.821581 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-cells-global-config" Oct 14 09:37:20 crc kubenswrapper[4870]: I1014 09:37:20.825795 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 14 09:37:20 crc kubenswrapper[4870]: I1014 09:37:20.826547 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Oct 14 09:37:20 crc kubenswrapper[4870]: I1014 09:37:20.826773 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-wrz7w" Oct 14 09:37:20 crc kubenswrapper[4870]: I1014 09:37:20.833898 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-openstack-openstack-cell1-z74pq"] Oct 14 09:37:20 crc kubenswrapper[4870]: I1014 09:37:20.977882 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/7f05f9ee-e89a-45a1-9f10-e68addbef873-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-openstack-cell1-z74pq\" (UID: \"7f05f9ee-e89a-45a1-9f10-e68addbef873\") " pod="openstack/nova-cell1-openstack-openstack-cell1-z74pq" Oct 14 09:37:20 crc kubenswrapper[4870]: I1014 09:37:20.977988 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/7f05f9ee-e89a-45a1-9f10-e68addbef873-nova-cells-global-config-1\") pod \"nova-cell1-openstack-openstack-cell1-z74pq\" (UID: \"7f05f9ee-e89a-45a1-9f10-e68addbef873\") " pod="openstack/nova-cell1-openstack-openstack-cell1-z74pq" Oct 14 09:37:20 crc kubenswrapper[4870]: I1014 09:37:20.978024 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dtrcm\" (UniqueName: \"kubernetes.io/projected/7f05f9ee-e89a-45a1-9f10-e68addbef873-kube-api-access-dtrcm\") pod \"nova-cell1-openstack-openstack-cell1-z74pq\" (UID: \"7f05f9ee-e89a-45a1-9f10-e68addbef873\") " pod="openstack/nova-cell1-openstack-openstack-cell1-z74pq" Oct 14 09:37:20 crc kubenswrapper[4870]: I1014 09:37:20.978089 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7f05f9ee-e89a-45a1-9f10-e68addbef873-inventory\") pod \"nova-cell1-openstack-openstack-cell1-z74pq\" (UID: \"7f05f9ee-e89a-45a1-9f10-e68addbef873\") " pod="openstack/nova-cell1-openstack-openstack-cell1-z74pq" Oct 14 09:37:20 crc kubenswrapper[4870]: I1014 09:37:20.978120 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/7f05f9ee-e89a-45a1-9f10-e68addbef873-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-openstack-cell1-z74pq\" (UID: \"7f05f9ee-e89a-45a1-9f10-e68addbef873\") " pod="openstack/nova-cell1-openstack-openstack-cell1-z74pq" Oct 14 09:37:20 crc kubenswrapper[4870]: I1014 09:37:20.978149 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/7f05f9ee-e89a-45a1-9f10-e68addbef873-nova-cells-global-config-0\") pod \"nova-cell1-openstack-openstack-cell1-z74pq\" (UID: \"7f05f9ee-e89a-45a1-9f10-e68addbef873\") " pod="openstack/nova-cell1-openstack-openstack-cell1-z74pq" Oct 14 09:37:20 crc kubenswrapper[4870]: I1014 09:37:20.978189 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7f05f9ee-e89a-45a1-9f10-e68addbef873-ssh-key\") pod \"nova-cell1-openstack-openstack-cell1-z74pq\" (UID: \"7f05f9ee-e89a-45a1-9f10-e68addbef873\") " pod="openstack/nova-cell1-openstack-openstack-cell1-z74pq" Oct 14 09:37:20 crc kubenswrapper[4870]: I1014 09:37:20.978234 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/7f05f9ee-e89a-45a1-9f10-e68addbef873-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-openstack-cell1-z74pq\" (UID: \"7f05f9ee-e89a-45a1-9f10-e68addbef873\") " pod="openstack/nova-cell1-openstack-openstack-cell1-z74pq" Oct 14 09:37:20 crc kubenswrapper[4870]: I1014 09:37:20.978264 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/7f05f9ee-e89a-45a1-9f10-e68addbef873-ceph\") pod \"nova-cell1-openstack-openstack-cell1-z74pq\" (UID: \"7f05f9ee-e89a-45a1-9f10-e68addbef873\") " pod="openstack/nova-cell1-openstack-openstack-cell1-z74pq" Oct 14 09:37:20 crc kubenswrapper[4870]: I1014 09:37:20.978287 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/7f05f9ee-e89a-45a1-9f10-e68addbef873-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-openstack-cell1-z74pq\" (UID: \"7f05f9ee-e89a-45a1-9f10-e68addbef873\") " pod="openstack/nova-cell1-openstack-openstack-cell1-z74pq" Oct 14 09:37:20 crc kubenswrapper[4870]: I1014 09:37:20.978317 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f05f9ee-e89a-45a1-9f10-e68addbef873-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-openstack-cell1-z74pq\" (UID: \"7f05f9ee-e89a-45a1-9f10-e68addbef873\") " pod="openstack/nova-cell1-openstack-openstack-cell1-z74pq" Oct 14 09:37:21 crc kubenswrapper[4870]: I1014 09:37:21.080029 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/7f05f9ee-e89a-45a1-9f10-e68addbef873-ceph\") pod \"nova-cell1-openstack-openstack-cell1-z74pq\" (UID: \"7f05f9ee-e89a-45a1-9f10-e68addbef873\") " pod="openstack/nova-cell1-openstack-openstack-cell1-z74pq" Oct 14 09:37:21 crc kubenswrapper[4870]: I1014 09:37:21.080080 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/7f05f9ee-e89a-45a1-9f10-e68addbef873-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-openstack-cell1-z74pq\" (UID: \"7f05f9ee-e89a-45a1-9f10-e68addbef873\") " pod="openstack/nova-cell1-openstack-openstack-cell1-z74pq" Oct 14 09:37:21 crc kubenswrapper[4870]: I1014 09:37:21.080124 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f05f9ee-e89a-45a1-9f10-e68addbef873-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-openstack-cell1-z74pq\" (UID: \"7f05f9ee-e89a-45a1-9f10-e68addbef873\") " pod="openstack/nova-cell1-openstack-openstack-cell1-z74pq" Oct 14 09:37:21 crc kubenswrapper[4870]: I1014 09:37:21.080172 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/7f05f9ee-e89a-45a1-9f10-e68addbef873-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-openstack-cell1-z74pq\" (UID: \"7f05f9ee-e89a-45a1-9f10-e68addbef873\") " pod="openstack/nova-cell1-openstack-openstack-cell1-z74pq" Oct 14 09:37:21 crc kubenswrapper[4870]: I1014 09:37:21.080224 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/7f05f9ee-e89a-45a1-9f10-e68addbef873-nova-cells-global-config-1\") pod \"nova-cell1-openstack-openstack-cell1-z74pq\" (UID: \"7f05f9ee-e89a-45a1-9f10-e68addbef873\") " pod="openstack/nova-cell1-openstack-openstack-cell1-z74pq" Oct 14 09:37:21 crc kubenswrapper[4870]: I1014 09:37:21.080254 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dtrcm\" (UniqueName: \"kubernetes.io/projected/7f05f9ee-e89a-45a1-9f10-e68addbef873-kube-api-access-dtrcm\") pod \"nova-cell1-openstack-openstack-cell1-z74pq\" (UID: \"7f05f9ee-e89a-45a1-9f10-e68addbef873\") " pod="openstack/nova-cell1-openstack-openstack-cell1-z74pq" Oct 14 09:37:21 crc kubenswrapper[4870]: I1014 09:37:21.080299 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7f05f9ee-e89a-45a1-9f10-e68addbef873-inventory\") pod \"nova-cell1-openstack-openstack-cell1-z74pq\" (UID: \"7f05f9ee-e89a-45a1-9f10-e68addbef873\") " pod="openstack/nova-cell1-openstack-openstack-cell1-z74pq" Oct 14 09:37:21 crc kubenswrapper[4870]: I1014 09:37:21.080325 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/7f05f9ee-e89a-45a1-9f10-e68addbef873-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-openstack-cell1-z74pq\" (UID: \"7f05f9ee-e89a-45a1-9f10-e68addbef873\") " pod="openstack/nova-cell1-openstack-openstack-cell1-z74pq" Oct 14 09:37:21 crc kubenswrapper[4870]: I1014 09:37:21.080349 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/7f05f9ee-e89a-45a1-9f10-e68addbef873-nova-cells-global-config-0\") pod \"nova-cell1-openstack-openstack-cell1-z74pq\" (UID: \"7f05f9ee-e89a-45a1-9f10-e68addbef873\") " pod="openstack/nova-cell1-openstack-openstack-cell1-z74pq" Oct 14 09:37:21 crc kubenswrapper[4870]: I1014 09:37:21.080377 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7f05f9ee-e89a-45a1-9f10-e68addbef873-ssh-key\") pod \"nova-cell1-openstack-openstack-cell1-z74pq\" (UID: \"7f05f9ee-e89a-45a1-9f10-e68addbef873\") " pod="openstack/nova-cell1-openstack-openstack-cell1-z74pq" Oct 14 09:37:21 crc kubenswrapper[4870]: I1014 09:37:21.080413 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/7f05f9ee-e89a-45a1-9f10-e68addbef873-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-openstack-cell1-z74pq\" (UID: \"7f05f9ee-e89a-45a1-9f10-e68addbef873\") " pod="openstack/nova-cell1-openstack-openstack-cell1-z74pq" Oct 14 09:37:21 crc kubenswrapper[4870]: I1014 09:37:21.082405 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/7f05f9ee-e89a-45a1-9f10-e68addbef873-nova-cells-global-config-0\") pod \"nova-cell1-openstack-openstack-cell1-z74pq\" (UID: \"7f05f9ee-e89a-45a1-9f10-e68addbef873\") " pod="openstack/nova-cell1-openstack-openstack-cell1-z74pq" Oct 14 09:37:21 crc kubenswrapper[4870]: I1014 09:37:21.083803 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/7f05f9ee-e89a-45a1-9f10-e68addbef873-nova-cells-global-config-1\") pod \"nova-cell1-openstack-openstack-cell1-z74pq\" (UID: \"7f05f9ee-e89a-45a1-9f10-e68addbef873\") " pod="openstack/nova-cell1-openstack-openstack-cell1-z74pq" Oct 14 09:37:21 crc kubenswrapper[4870]: I1014 09:37:21.084820 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/7f05f9ee-e89a-45a1-9f10-e68addbef873-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-openstack-cell1-z74pq\" (UID: \"7f05f9ee-e89a-45a1-9f10-e68addbef873\") " pod="openstack/nova-cell1-openstack-openstack-cell1-z74pq" Oct 14 09:37:21 crc kubenswrapper[4870]: I1014 09:37:21.084989 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/7f05f9ee-e89a-45a1-9f10-e68addbef873-ceph\") pod \"nova-cell1-openstack-openstack-cell1-z74pq\" (UID: \"7f05f9ee-e89a-45a1-9f10-e68addbef873\") " pod="openstack/nova-cell1-openstack-openstack-cell1-z74pq" Oct 14 09:37:21 crc kubenswrapper[4870]: I1014 09:37:21.086518 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/7f05f9ee-e89a-45a1-9f10-e68addbef873-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-openstack-cell1-z74pq\" (UID: \"7f05f9ee-e89a-45a1-9f10-e68addbef873\") " pod="openstack/nova-cell1-openstack-openstack-cell1-z74pq" Oct 14 09:37:21 crc kubenswrapper[4870]: I1014 09:37:21.086826 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/7f05f9ee-e89a-45a1-9f10-e68addbef873-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-openstack-cell1-z74pq\" (UID: \"7f05f9ee-e89a-45a1-9f10-e68addbef873\") " pod="openstack/nova-cell1-openstack-openstack-cell1-z74pq" Oct 14 09:37:21 crc kubenswrapper[4870]: I1014 09:37:21.087338 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7f05f9ee-e89a-45a1-9f10-e68addbef873-ssh-key\") pod \"nova-cell1-openstack-openstack-cell1-z74pq\" (UID: \"7f05f9ee-e89a-45a1-9f10-e68addbef873\") " pod="openstack/nova-cell1-openstack-openstack-cell1-z74pq" Oct 14 09:37:21 crc kubenswrapper[4870]: I1014 09:37:21.088616 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/7f05f9ee-e89a-45a1-9f10-e68addbef873-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-openstack-cell1-z74pq\" (UID: \"7f05f9ee-e89a-45a1-9f10-e68addbef873\") " pod="openstack/nova-cell1-openstack-openstack-cell1-z74pq" Oct 14 09:37:21 crc kubenswrapper[4870]: I1014 09:37:21.088900 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f05f9ee-e89a-45a1-9f10-e68addbef873-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-openstack-cell1-z74pq\" (UID: \"7f05f9ee-e89a-45a1-9f10-e68addbef873\") " pod="openstack/nova-cell1-openstack-openstack-cell1-z74pq" Oct 14 09:37:21 crc kubenswrapper[4870]: I1014 09:37:21.091192 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7f05f9ee-e89a-45a1-9f10-e68addbef873-inventory\") pod \"nova-cell1-openstack-openstack-cell1-z74pq\" (UID: \"7f05f9ee-e89a-45a1-9f10-e68addbef873\") " pod="openstack/nova-cell1-openstack-openstack-cell1-z74pq" Oct 14 09:37:21 crc kubenswrapper[4870]: I1014 09:37:21.105498 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dtrcm\" (UniqueName: \"kubernetes.io/projected/7f05f9ee-e89a-45a1-9f10-e68addbef873-kube-api-access-dtrcm\") pod \"nova-cell1-openstack-openstack-cell1-z74pq\" (UID: \"7f05f9ee-e89a-45a1-9f10-e68addbef873\") " pod="openstack/nova-cell1-openstack-openstack-cell1-z74pq" Oct 14 09:37:21 crc kubenswrapper[4870]: I1014 09:37:21.149517 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-openstack-cell1-z74pq" Oct 14 09:37:21 crc kubenswrapper[4870]: I1014 09:37:21.802885 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-openstack-openstack-cell1-z74pq"] Oct 14 09:37:21 crc kubenswrapper[4870]: W1014 09:37:21.953485 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7f05f9ee_e89a_45a1_9f10_e68addbef873.slice/crio-1ee51b2bc67e54ff74bad07c9dcc53b7b5d650eca97c4deb6adaca787d7ff659 WatchSource:0}: Error finding container 1ee51b2bc67e54ff74bad07c9dcc53b7b5d650eca97c4deb6adaca787d7ff659: Status 404 returned error can't find the container with id 1ee51b2bc67e54ff74bad07c9dcc53b7b5d650eca97c4deb6adaca787d7ff659 Oct 14 09:37:22 crc kubenswrapper[4870]: I1014 09:37:22.415112 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-openstack-cell1-z74pq" event={"ID":"7f05f9ee-e89a-45a1-9f10-e68addbef873","Type":"ContainerStarted","Data":"1ee51b2bc67e54ff74bad07c9dcc53b7b5d650eca97c4deb6adaca787d7ff659"} Oct 14 09:37:23 crc kubenswrapper[4870]: I1014 09:37:23.431968 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-openstack-cell1-z74pq" event={"ID":"7f05f9ee-e89a-45a1-9f10-e68addbef873","Type":"ContainerStarted","Data":"79e5f5b43ed20f499f20ea656227a81276ec7adf1e8e2547aff2a69066f85e71"} Oct 14 09:37:23 crc kubenswrapper[4870]: I1014 09:37:23.466631 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-openstack-openstack-cell1-z74pq" podStartSLOduration=2.952835576 podStartE2EDuration="3.466598407s" podCreationTimestamp="2025-10-14 09:37:20 +0000 UTC" firstStartedPulling="2025-10-14 09:37:21.958767726 +0000 UTC m=+9377.656128127" lastFinishedPulling="2025-10-14 09:37:22.472530577 +0000 UTC m=+9378.169890958" observedRunningTime="2025-10-14 09:37:23.462079986 +0000 UTC m=+9379.159440387" watchObservedRunningTime="2025-10-14 09:37:23.466598407 +0000 UTC m=+9379.163958818" Oct 14 09:37:23 crc kubenswrapper[4870]: I1014 09:37:23.951171 4870 patch_prober.go:28] interesting pod/machine-config-daemon-7tvc7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 09:37:23 crc kubenswrapper[4870]: I1014 09:37:23.951582 4870 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 09:37:53 crc kubenswrapper[4870]: I1014 09:37:53.950895 4870 patch_prober.go:28] interesting pod/machine-config-daemon-7tvc7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 09:37:53 crc kubenswrapper[4870]: I1014 09:37:53.951551 4870 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 09:37:53 crc kubenswrapper[4870]: I1014 09:37:53.951615 4870 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" Oct 14 09:37:53 crc kubenswrapper[4870]: I1014 09:37:53.952731 4870 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"da933e0d9bd5082c26564c190825d7838bd95c38ef50de4f4eef55960b16e294"} pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 14 09:37:53 crc kubenswrapper[4870]: I1014 09:37:53.952798 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerName="machine-config-daemon" containerID="cri-o://da933e0d9bd5082c26564c190825d7838bd95c38ef50de4f4eef55960b16e294" gracePeriod=600 Oct 14 09:37:54 crc kubenswrapper[4870]: I1014 09:37:54.817008 4870 generic.go:334] "Generic (PLEG): container finished" podID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerID="da933e0d9bd5082c26564c190825d7838bd95c38ef50de4f4eef55960b16e294" exitCode=0 Oct 14 09:37:54 crc kubenswrapper[4870]: I1014 09:37:54.817090 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" event={"ID":"874a5e45-dffd-4d17-b609-b7d3ed2eab07","Type":"ContainerDied","Data":"da933e0d9bd5082c26564c190825d7838bd95c38ef50de4f4eef55960b16e294"} Oct 14 09:37:54 crc kubenswrapper[4870]: I1014 09:37:54.817676 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" event={"ID":"874a5e45-dffd-4d17-b609-b7d3ed2eab07","Type":"ContainerStarted","Data":"6d5792c105b7bc7694bce51cbf8b085eda0318f921dffcecf5167175f90aa527"} Oct 14 09:37:54 crc kubenswrapper[4870]: I1014 09:37:54.817710 4870 scope.go:117] "RemoveContainer" containerID="3eea07ce0e4dde9363f92f5f44f530c26c10b42e335dd2bafb6764d9cbdc203c" Oct 14 09:39:36 crc kubenswrapper[4870]: I1014 09:39:36.180431 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-58xgr"] Oct 14 09:39:36 crc kubenswrapper[4870]: I1014 09:39:36.183628 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-58xgr" Oct 14 09:39:36 crc kubenswrapper[4870]: I1014 09:39:36.196097 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/78366820-fd1b-4c74-8214-1c2bc8b05f1c-utilities\") pod \"redhat-marketplace-58xgr\" (UID: \"78366820-fd1b-4c74-8214-1c2bc8b05f1c\") " pod="openshift-marketplace/redhat-marketplace-58xgr" Oct 14 09:39:36 crc kubenswrapper[4870]: I1014 09:39:36.196144 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2b8qq\" (UniqueName: \"kubernetes.io/projected/78366820-fd1b-4c74-8214-1c2bc8b05f1c-kube-api-access-2b8qq\") pod \"redhat-marketplace-58xgr\" (UID: \"78366820-fd1b-4c74-8214-1c2bc8b05f1c\") " pod="openshift-marketplace/redhat-marketplace-58xgr" Oct 14 09:39:36 crc kubenswrapper[4870]: I1014 09:39:36.196331 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/78366820-fd1b-4c74-8214-1c2bc8b05f1c-catalog-content\") pod \"redhat-marketplace-58xgr\" (UID: \"78366820-fd1b-4c74-8214-1c2bc8b05f1c\") " pod="openshift-marketplace/redhat-marketplace-58xgr" Oct 14 09:39:36 crc kubenswrapper[4870]: I1014 09:39:36.217378 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-58xgr"] Oct 14 09:39:36 crc kubenswrapper[4870]: I1014 09:39:36.297842 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/78366820-fd1b-4c74-8214-1c2bc8b05f1c-utilities\") pod \"redhat-marketplace-58xgr\" (UID: \"78366820-fd1b-4c74-8214-1c2bc8b05f1c\") " pod="openshift-marketplace/redhat-marketplace-58xgr" Oct 14 09:39:36 crc kubenswrapper[4870]: I1014 09:39:36.297899 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2b8qq\" (UniqueName: \"kubernetes.io/projected/78366820-fd1b-4c74-8214-1c2bc8b05f1c-kube-api-access-2b8qq\") pod \"redhat-marketplace-58xgr\" (UID: \"78366820-fd1b-4c74-8214-1c2bc8b05f1c\") " pod="openshift-marketplace/redhat-marketplace-58xgr" Oct 14 09:39:36 crc kubenswrapper[4870]: I1014 09:39:36.297948 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/78366820-fd1b-4c74-8214-1c2bc8b05f1c-catalog-content\") pod \"redhat-marketplace-58xgr\" (UID: \"78366820-fd1b-4c74-8214-1c2bc8b05f1c\") " pod="openshift-marketplace/redhat-marketplace-58xgr" Oct 14 09:39:36 crc kubenswrapper[4870]: I1014 09:39:36.298325 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/78366820-fd1b-4c74-8214-1c2bc8b05f1c-utilities\") pod \"redhat-marketplace-58xgr\" (UID: \"78366820-fd1b-4c74-8214-1c2bc8b05f1c\") " pod="openshift-marketplace/redhat-marketplace-58xgr" Oct 14 09:39:36 crc kubenswrapper[4870]: I1014 09:39:36.298406 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/78366820-fd1b-4c74-8214-1c2bc8b05f1c-catalog-content\") pod \"redhat-marketplace-58xgr\" (UID: \"78366820-fd1b-4c74-8214-1c2bc8b05f1c\") " pod="openshift-marketplace/redhat-marketplace-58xgr" Oct 14 09:39:36 crc kubenswrapper[4870]: I1014 09:39:36.318959 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2b8qq\" (UniqueName: \"kubernetes.io/projected/78366820-fd1b-4c74-8214-1c2bc8b05f1c-kube-api-access-2b8qq\") pod \"redhat-marketplace-58xgr\" (UID: \"78366820-fd1b-4c74-8214-1c2bc8b05f1c\") " pod="openshift-marketplace/redhat-marketplace-58xgr" Oct 14 09:39:36 crc kubenswrapper[4870]: I1014 09:39:36.522950 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-58xgr" Oct 14 09:39:36 crc kubenswrapper[4870]: I1014 09:39:36.989593 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-58xgr"] Oct 14 09:39:37 crc kubenswrapper[4870]: W1014 09:39:37.007463 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod78366820_fd1b_4c74_8214_1c2bc8b05f1c.slice/crio-23cef815cd3401b7af086072f7679fe6406e7ed4d239098dabd4668e54a287a8 WatchSource:0}: Error finding container 23cef815cd3401b7af086072f7679fe6406e7ed4d239098dabd4668e54a287a8: Status 404 returned error can't find the container with id 23cef815cd3401b7af086072f7679fe6406e7ed4d239098dabd4668e54a287a8 Oct 14 09:39:37 crc kubenswrapper[4870]: I1014 09:39:37.100329 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-58xgr" event={"ID":"78366820-fd1b-4c74-8214-1c2bc8b05f1c","Type":"ContainerStarted","Data":"23cef815cd3401b7af086072f7679fe6406e7ed4d239098dabd4668e54a287a8"} Oct 14 09:39:38 crc kubenswrapper[4870]: I1014 09:39:38.132288 4870 generic.go:334] "Generic (PLEG): container finished" podID="78366820-fd1b-4c74-8214-1c2bc8b05f1c" containerID="e509f7324efc246094acd4b839200f4fd5efcd851ad582af0b25df3dc22e27fb" exitCode=0 Oct 14 09:39:38 crc kubenswrapper[4870]: I1014 09:39:38.132496 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-58xgr" event={"ID":"78366820-fd1b-4c74-8214-1c2bc8b05f1c","Type":"ContainerDied","Data":"e509f7324efc246094acd4b839200f4fd5efcd851ad582af0b25df3dc22e27fb"} Oct 14 09:39:38 crc kubenswrapper[4870]: I1014 09:39:38.135412 4870 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 14 09:39:39 crc kubenswrapper[4870]: I1014 09:39:39.146508 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-58xgr" event={"ID":"78366820-fd1b-4c74-8214-1c2bc8b05f1c","Type":"ContainerStarted","Data":"65159c4a1640f76eaecbaf0cdf0a4ded366512e3c6cf429cd21cebf2f970458b"} Oct 14 09:39:40 crc kubenswrapper[4870]: I1014 09:39:40.160531 4870 generic.go:334] "Generic (PLEG): container finished" podID="78366820-fd1b-4c74-8214-1c2bc8b05f1c" containerID="65159c4a1640f76eaecbaf0cdf0a4ded366512e3c6cf429cd21cebf2f970458b" exitCode=0 Oct 14 09:39:40 crc kubenswrapper[4870]: I1014 09:39:40.160612 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-58xgr" event={"ID":"78366820-fd1b-4c74-8214-1c2bc8b05f1c","Type":"ContainerDied","Data":"65159c4a1640f76eaecbaf0cdf0a4ded366512e3c6cf429cd21cebf2f970458b"} Oct 14 09:39:41 crc kubenswrapper[4870]: I1014 09:39:41.174711 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-58xgr" event={"ID":"78366820-fd1b-4c74-8214-1c2bc8b05f1c","Type":"ContainerStarted","Data":"12d3574484090915b0c5b2d1e8e4fb3c3d123b7baa08b4efebddc82adc01c902"} Oct 14 09:39:41 crc kubenswrapper[4870]: I1014 09:39:41.198344 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-58xgr" podStartSLOduration=2.692406546 podStartE2EDuration="5.198324776s" podCreationTimestamp="2025-10-14 09:39:36 +0000 UTC" firstStartedPulling="2025-10-14 09:39:38.134821468 +0000 UTC m=+9513.832181869" lastFinishedPulling="2025-10-14 09:39:40.640739728 +0000 UTC m=+9516.338100099" observedRunningTime="2025-10-14 09:39:41.191767885 +0000 UTC m=+9516.889128286" watchObservedRunningTime="2025-10-14 09:39:41.198324776 +0000 UTC m=+9516.895685147" Oct 14 09:39:46 crc kubenswrapper[4870]: I1014 09:39:46.523554 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-58xgr" Oct 14 09:39:46 crc kubenswrapper[4870]: I1014 09:39:46.524200 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-58xgr" Oct 14 09:39:46 crc kubenswrapper[4870]: I1014 09:39:46.585048 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-58xgr" Oct 14 09:39:47 crc kubenswrapper[4870]: I1014 09:39:47.314973 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-58xgr" Oct 14 09:39:47 crc kubenswrapper[4870]: I1014 09:39:47.368214 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-58xgr"] Oct 14 09:39:49 crc kubenswrapper[4870]: I1014 09:39:49.284401 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-58xgr" podUID="78366820-fd1b-4c74-8214-1c2bc8b05f1c" containerName="registry-server" containerID="cri-o://12d3574484090915b0c5b2d1e8e4fb3c3d123b7baa08b4efebddc82adc01c902" gracePeriod=2 Oct 14 09:39:49 crc kubenswrapper[4870]: I1014 09:39:49.815475 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-58xgr" Oct 14 09:39:49 crc kubenswrapper[4870]: I1014 09:39:49.919252 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/78366820-fd1b-4c74-8214-1c2bc8b05f1c-utilities\") pod \"78366820-fd1b-4c74-8214-1c2bc8b05f1c\" (UID: \"78366820-fd1b-4c74-8214-1c2bc8b05f1c\") " Oct 14 09:39:49 crc kubenswrapper[4870]: I1014 09:39:49.919313 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2b8qq\" (UniqueName: \"kubernetes.io/projected/78366820-fd1b-4c74-8214-1c2bc8b05f1c-kube-api-access-2b8qq\") pod \"78366820-fd1b-4c74-8214-1c2bc8b05f1c\" (UID: \"78366820-fd1b-4c74-8214-1c2bc8b05f1c\") " Oct 14 09:39:49 crc kubenswrapper[4870]: I1014 09:39:49.919347 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/78366820-fd1b-4c74-8214-1c2bc8b05f1c-catalog-content\") pod \"78366820-fd1b-4c74-8214-1c2bc8b05f1c\" (UID: \"78366820-fd1b-4c74-8214-1c2bc8b05f1c\") " Oct 14 09:39:49 crc kubenswrapper[4870]: I1014 09:39:49.920208 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/78366820-fd1b-4c74-8214-1c2bc8b05f1c-utilities" (OuterVolumeSpecName: "utilities") pod "78366820-fd1b-4c74-8214-1c2bc8b05f1c" (UID: "78366820-fd1b-4c74-8214-1c2bc8b05f1c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 09:39:49 crc kubenswrapper[4870]: I1014 09:39:49.931729 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/78366820-fd1b-4c74-8214-1c2bc8b05f1c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "78366820-fd1b-4c74-8214-1c2bc8b05f1c" (UID: "78366820-fd1b-4c74-8214-1c2bc8b05f1c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 09:39:49 crc kubenswrapper[4870]: I1014 09:39:49.932084 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/78366820-fd1b-4c74-8214-1c2bc8b05f1c-kube-api-access-2b8qq" (OuterVolumeSpecName: "kube-api-access-2b8qq") pod "78366820-fd1b-4c74-8214-1c2bc8b05f1c" (UID: "78366820-fd1b-4c74-8214-1c2bc8b05f1c"). InnerVolumeSpecName "kube-api-access-2b8qq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:39:50 crc kubenswrapper[4870]: I1014 09:39:50.022016 4870 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/78366820-fd1b-4c74-8214-1c2bc8b05f1c-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 09:39:50 crc kubenswrapper[4870]: I1014 09:39:50.022047 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2b8qq\" (UniqueName: \"kubernetes.io/projected/78366820-fd1b-4c74-8214-1c2bc8b05f1c-kube-api-access-2b8qq\") on node \"crc\" DevicePath \"\"" Oct 14 09:39:50 crc kubenswrapper[4870]: I1014 09:39:50.022059 4870 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/78366820-fd1b-4c74-8214-1c2bc8b05f1c-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 09:39:50 crc kubenswrapper[4870]: I1014 09:39:50.295936 4870 generic.go:334] "Generic (PLEG): container finished" podID="78366820-fd1b-4c74-8214-1c2bc8b05f1c" containerID="12d3574484090915b0c5b2d1e8e4fb3c3d123b7baa08b4efebddc82adc01c902" exitCode=0 Oct 14 09:39:50 crc kubenswrapper[4870]: I1014 09:39:50.296164 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-58xgr" event={"ID":"78366820-fd1b-4c74-8214-1c2bc8b05f1c","Type":"ContainerDied","Data":"12d3574484090915b0c5b2d1e8e4fb3c3d123b7baa08b4efebddc82adc01c902"} Oct 14 09:39:50 crc kubenswrapper[4870]: I1014 09:39:50.296221 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-58xgr" Oct 14 09:39:50 crc kubenswrapper[4870]: I1014 09:39:50.296242 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-58xgr" event={"ID":"78366820-fd1b-4c74-8214-1c2bc8b05f1c","Type":"ContainerDied","Data":"23cef815cd3401b7af086072f7679fe6406e7ed4d239098dabd4668e54a287a8"} Oct 14 09:39:50 crc kubenswrapper[4870]: I1014 09:39:50.296268 4870 scope.go:117] "RemoveContainer" containerID="12d3574484090915b0c5b2d1e8e4fb3c3d123b7baa08b4efebddc82adc01c902" Oct 14 09:39:50 crc kubenswrapper[4870]: I1014 09:39:50.322010 4870 scope.go:117] "RemoveContainer" containerID="65159c4a1640f76eaecbaf0cdf0a4ded366512e3c6cf429cd21cebf2f970458b" Oct 14 09:39:50 crc kubenswrapper[4870]: I1014 09:39:50.340580 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-58xgr"] Oct 14 09:39:50 crc kubenswrapper[4870]: I1014 09:39:50.355509 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-58xgr"] Oct 14 09:39:50 crc kubenswrapper[4870]: I1014 09:39:50.357261 4870 scope.go:117] "RemoveContainer" containerID="e509f7324efc246094acd4b839200f4fd5efcd851ad582af0b25df3dc22e27fb" Oct 14 09:39:50 crc kubenswrapper[4870]: I1014 09:39:50.401988 4870 scope.go:117] "RemoveContainer" containerID="12d3574484090915b0c5b2d1e8e4fb3c3d123b7baa08b4efebddc82adc01c902" Oct 14 09:39:50 crc kubenswrapper[4870]: E1014 09:39:50.402480 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"12d3574484090915b0c5b2d1e8e4fb3c3d123b7baa08b4efebddc82adc01c902\": container with ID starting with 12d3574484090915b0c5b2d1e8e4fb3c3d123b7baa08b4efebddc82adc01c902 not found: ID does not exist" containerID="12d3574484090915b0c5b2d1e8e4fb3c3d123b7baa08b4efebddc82adc01c902" Oct 14 09:39:50 crc kubenswrapper[4870]: I1014 09:39:50.402540 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"12d3574484090915b0c5b2d1e8e4fb3c3d123b7baa08b4efebddc82adc01c902"} err="failed to get container status \"12d3574484090915b0c5b2d1e8e4fb3c3d123b7baa08b4efebddc82adc01c902\": rpc error: code = NotFound desc = could not find container \"12d3574484090915b0c5b2d1e8e4fb3c3d123b7baa08b4efebddc82adc01c902\": container with ID starting with 12d3574484090915b0c5b2d1e8e4fb3c3d123b7baa08b4efebddc82adc01c902 not found: ID does not exist" Oct 14 09:39:50 crc kubenswrapper[4870]: I1014 09:39:50.402587 4870 scope.go:117] "RemoveContainer" containerID="65159c4a1640f76eaecbaf0cdf0a4ded366512e3c6cf429cd21cebf2f970458b" Oct 14 09:39:50 crc kubenswrapper[4870]: E1014 09:39:50.403030 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"65159c4a1640f76eaecbaf0cdf0a4ded366512e3c6cf429cd21cebf2f970458b\": container with ID starting with 65159c4a1640f76eaecbaf0cdf0a4ded366512e3c6cf429cd21cebf2f970458b not found: ID does not exist" containerID="65159c4a1640f76eaecbaf0cdf0a4ded366512e3c6cf429cd21cebf2f970458b" Oct 14 09:39:50 crc kubenswrapper[4870]: I1014 09:39:50.403067 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"65159c4a1640f76eaecbaf0cdf0a4ded366512e3c6cf429cd21cebf2f970458b"} err="failed to get container status \"65159c4a1640f76eaecbaf0cdf0a4ded366512e3c6cf429cd21cebf2f970458b\": rpc error: code = NotFound desc = could not find container \"65159c4a1640f76eaecbaf0cdf0a4ded366512e3c6cf429cd21cebf2f970458b\": container with ID starting with 65159c4a1640f76eaecbaf0cdf0a4ded366512e3c6cf429cd21cebf2f970458b not found: ID does not exist" Oct 14 09:39:50 crc kubenswrapper[4870]: I1014 09:39:50.403093 4870 scope.go:117] "RemoveContainer" containerID="e509f7324efc246094acd4b839200f4fd5efcd851ad582af0b25df3dc22e27fb" Oct 14 09:39:50 crc kubenswrapper[4870]: E1014 09:39:50.403369 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e509f7324efc246094acd4b839200f4fd5efcd851ad582af0b25df3dc22e27fb\": container with ID starting with e509f7324efc246094acd4b839200f4fd5efcd851ad582af0b25df3dc22e27fb not found: ID does not exist" containerID="e509f7324efc246094acd4b839200f4fd5efcd851ad582af0b25df3dc22e27fb" Oct 14 09:39:50 crc kubenswrapper[4870]: I1014 09:39:50.403416 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e509f7324efc246094acd4b839200f4fd5efcd851ad582af0b25df3dc22e27fb"} err="failed to get container status \"e509f7324efc246094acd4b839200f4fd5efcd851ad582af0b25df3dc22e27fb\": rpc error: code = NotFound desc = could not find container \"e509f7324efc246094acd4b839200f4fd5efcd851ad582af0b25df3dc22e27fb\": container with ID starting with e509f7324efc246094acd4b839200f4fd5efcd851ad582af0b25df3dc22e27fb not found: ID does not exist" Oct 14 09:39:51 crc kubenswrapper[4870]: I1014 09:39:51.049333 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="78366820-fd1b-4c74-8214-1c2bc8b05f1c" path="/var/lib/kubelet/pods/78366820-fd1b-4c74-8214-1c2bc8b05f1c/volumes" Oct 14 09:40:12 crc kubenswrapper[4870]: I1014 09:40:12.532256 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-66rm7"] Oct 14 09:40:12 crc kubenswrapper[4870]: E1014 09:40:12.533428 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="78366820-fd1b-4c74-8214-1c2bc8b05f1c" containerName="registry-server" Oct 14 09:40:12 crc kubenswrapper[4870]: I1014 09:40:12.533519 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="78366820-fd1b-4c74-8214-1c2bc8b05f1c" containerName="registry-server" Oct 14 09:40:12 crc kubenswrapper[4870]: E1014 09:40:12.533546 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="78366820-fd1b-4c74-8214-1c2bc8b05f1c" containerName="extract-content" Oct 14 09:40:12 crc kubenswrapper[4870]: I1014 09:40:12.533554 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="78366820-fd1b-4c74-8214-1c2bc8b05f1c" containerName="extract-content" Oct 14 09:40:12 crc kubenswrapper[4870]: E1014 09:40:12.533610 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="78366820-fd1b-4c74-8214-1c2bc8b05f1c" containerName="extract-utilities" Oct 14 09:40:12 crc kubenswrapper[4870]: I1014 09:40:12.533619 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="78366820-fd1b-4c74-8214-1c2bc8b05f1c" containerName="extract-utilities" Oct 14 09:40:12 crc kubenswrapper[4870]: I1014 09:40:12.533872 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="78366820-fd1b-4c74-8214-1c2bc8b05f1c" containerName="registry-server" Oct 14 09:40:12 crc kubenswrapper[4870]: I1014 09:40:12.535863 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-66rm7" Oct 14 09:40:12 crc kubenswrapper[4870]: I1014 09:40:12.557426 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-66rm7"] Oct 14 09:40:12 crc kubenswrapper[4870]: I1014 09:40:12.671554 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/591dfcef-72ec-4445-9900-c679a1f73a5a-utilities\") pod \"certified-operators-66rm7\" (UID: \"591dfcef-72ec-4445-9900-c679a1f73a5a\") " pod="openshift-marketplace/certified-operators-66rm7" Oct 14 09:40:12 crc kubenswrapper[4870]: I1014 09:40:12.671703 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/591dfcef-72ec-4445-9900-c679a1f73a5a-catalog-content\") pod \"certified-operators-66rm7\" (UID: \"591dfcef-72ec-4445-9900-c679a1f73a5a\") " pod="openshift-marketplace/certified-operators-66rm7" Oct 14 09:40:12 crc kubenswrapper[4870]: I1014 09:40:12.671737 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tx9ht\" (UniqueName: \"kubernetes.io/projected/591dfcef-72ec-4445-9900-c679a1f73a5a-kube-api-access-tx9ht\") pod \"certified-operators-66rm7\" (UID: \"591dfcef-72ec-4445-9900-c679a1f73a5a\") " pod="openshift-marketplace/certified-operators-66rm7" Oct 14 09:40:12 crc kubenswrapper[4870]: I1014 09:40:12.774496 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/591dfcef-72ec-4445-9900-c679a1f73a5a-utilities\") pod \"certified-operators-66rm7\" (UID: \"591dfcef-72ec-4445-9900-c679a1f73a5a\") " pod="openshift-marketplace/certified-operators-66rm7" Oct 14 09:40:12 crc kubenswrapper[4870]: I1014 09:40:12.774661 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/591dfcef-72ec-4445-9900-c679a1f73a5a-catalog-content\") pod \"certified-operators-66rm7\" (UID: \"591dfcef-72ec-4445-9900-c679a1f73a5a\") " pod="openshift-marketplace/certified-operators-66rm7" Oct 14 09:40:12 crc kubenswrapper[4870]: I1014 09:40:12.774699 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tx9ht\" (UniqueName: \"kubernetes.io/projected/591dfcef-72ec-4445-9900-c679a1f73a5a-kube-api-access-tx9ht\") pod \"certified-operators-66rm7\" (UID: \"591dfcef-72ec-4445-9900-c679a1f73a5a\") " pod="openshift-marketplace/certified-operators-66rm7" Oct 14 09:40:12 crc kubenswrapper[4870]: I1014 09:40:12.775653 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/591dfcef-72ec-4445-9900-c679a1f73a5a-utilities\") pod \"certified-operators-66rm7\" (UID: \"591dfcef-72ec-4445-9900-c679a1f73a5a\") " pod="openshift-marketplace/certified-operators-66rm7" Oct 14 09:40:12 crc kubenswrapper[4870]: I1014 09:40:12.775935 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/591dfcef-72ec-4445-9900-c679a1f73a5a-catalog-content\") pod \"certified-operators-66rm7\" (UID: \"591dfcef-72ec-4445-9900-c679a1f73a5a\") " pod="openshift-marketplace/certified-operators-66rm7" Oct 14 09:40:12 crc kubenswrapper[4870]: I1014 09:40:12.795651 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tx9ht\" (UniqueName: \"kubernetes.io/projected/591dfcef-72ec-4445-9900-c679a1f73a5a-kube-api-access-tx9ht\") pod \"certified-operators-66rm7\" (UID: \"591dfcef-72ec-4445-9900-c679a1f73a5a\") " pod="openshift-marketplace/certified-operators-66rm7" Oct 14 09:40:12 crc kubenswrapper[4870]: I1014 09:40:12.898954 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-66rm7" Oct 14 09:40:13 crc kubenswrapper[4870]: I1014 09:40:13.417239 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-66rm7"] Oct 14 09:40:13 crc kubenswrapper[4870]: I1014 09:40:13.617200 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-66rm7" event={"ID":"591dfcef-72ec-4445-9900-c679a1f73a5a","Type":"ContainerStarted","Data":"2ed15bbb40c429868f485cea2184fa93e5ce131f2eebe6137fbd548cef655049"} Oct 14 09:40:14 crc kubenswrapper[4870]: I1014 09:40:14.632696 4870 generic.go:334] "Generic (PLEG): container finished" podID="591dfcef-72ec-4445-9900-c679a1f73a5a" containerID="c6bb00c25a1754dd60f845f7dcb78f2294bff46f1aa0c278741197ce68a9e21e" exitCode=0 Oct 14 09:40:14 crc kubenswrapper[4870]: I1014 09:40:14.632748 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-66rm7" event={"ID":"591dfcef-72ec-4445-9900-c679a1f73a5a","Type":"ContainerDied","Data":"c6bb00c25a1754dd60f845f7dcb78f2294bff46f1aa0c278741197ce68a9e21e"} Oct 14 09:40:15 crc kubenswrapper[4870]: I1014 09:40:15.647947 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-66rm7" event={"ID":"591dfcef-72ec-4445-9900-c679a1f73a5a","Type":"ContainerStarted","Data":"654fab7a99ef3161dd7fdeca3069b2648c0a9718930483dc7317e3cb3df5665d"} Oct 14 09:40:16 crc kubenswrapper[4870]: I1014 09:40:16.675136 4870 generic.go:334] "Generic (PLEG): container finished" podID="591dfcef-72ec-4445-9900-c679a1f73a5a" containerID="654fab7a99ef3161dd7fdeca3069b2648c0a9718930483dc7317e3cb3df5665d" exitCode=0 Oct 14 09:40:16 crc kubenswrapper[4870]: I1014 09:40:16.675218 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-66rm7" event={"ID":"591dfcef-72ec-4445-9900-c679a1f73a5a","Type":"ContainerDied","Data":"654fab7a99ef3161dd7fdeca3069b2648c0a9718930483dc7317e3cb3df5665d"} Oct 14 09:40:17 crc kubenswrapper[4870]: I1014 09:40:17.690130 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-66rm7" event={"ID":"591dfcef-72ec-4445-9900-c679a1f73a5a","Type":"ContainerStarted","Data":"091a40917baf7e3ca48c776c2619b3d52bc941edd5b05d706046ede926ef6293"} Oct 14 09:40:17 crc kubenswrapper[4870]: I1014 09:40:17.713172 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-66rm7" podStartSLOduration=3.22458697 podStartE2EDuration="5.713146394s" podCreationTimestamp="2025-10-14 09:40:12 +0000 UTC" firstStartedPulling="2025-10-14 09:40:14.634769362 +0000 UTC m=+9550.332129743" lastFinishedPulling="2025-10-14 09:40:17.123328786 +0000 UTC m=+9552.820689167" observedRunningTime="2025-10-14 09:40:17.707139076 +0000 UTC m=+9553.404499507" watchObservedRunningTime="2025-10-14 09:40:17.713146394 +0000 UTC m=+9553.410506805" Oct 14 09:40:22 crc kubenswrapper[4870]: I1014 09:40:22.899374 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-66rm7" Oct 14 09:40:22 crc kubenswrapper[4870]: I1014 09:40:22.899940 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-66rm7" Oct 14 09:40:23 crc kubenswrapper[4870]: I1014 09:40:23.699586 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-66rm7" Oct 14 09:40:23 crc kubenswrapper[4870]: I1014 09:40:23.950910 4870 patch_prober.go:28] interesting pod/machine-config-daemon-7tvc7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 09:40:23 crc kubenswrapper[4870]: I1014 09:40:23.950964 4870 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 09:40:23 crc kubenswrapper[4870]: I1014 09:40:23.962045 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-66rm7" Oct 14 09:40:24 crc kubenswrapper[4870]: I1014 09:40:24.021513 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-66rm7"] Oct 14 09:40:25 crc kubenswrapper[4870]: I1014 09:40:25.853060 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-66rm7" podUID="591dfcef-72ec-4445-9900-c679a1f73a5a" containerName="registry-server" containerID="cri-o://091a40917baf7e3ca48c776c2619b3d52bc941edd5b05d706046ede926ef6293" gracePeriod=2 Oct 14 09:40:26 crc kubenswrapper[4870]: I1014 09:40:26.369667 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-66rm7" Oct 14 09:40:26 crc kubenswrapper[4870]: I1014 09:40:26.533717 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/591dfcef-72ec-4445-9900-c679a1f73a5a-catalog-content\") pod \"591dfcef-72ec-4445-9900-c679a1f73a5a\" (UID: \"591dfcef-72ec-4445-9900-c679a1f73a5a\") " Oct 14 09:40:26 crc kubenswrapper[4870]: I1014 09:40:26.533819 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tx9ht\" (UniqueName: \"kubernetes.io/projected/591dfcef-72ec-4445-9900-c679a1f73a5a-kube-api-access-tx9ht\") pod \"591dfcef-72ec-4445-9900-c679a1f73a5a\" (UID: \"591dfcef-72ec-4445-9900-c679a1f73a5a\") " Oct 14 09:40:26 crc kubenswrapper[4870]: I1014 09:40:26.533940 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/591dfcef-72ec-4445-9900-c679a1f73a5a-utilities\") pod \"591dfcef-72ec-4445-9900-c679a1f73a5a\" (UID: \"591dfcef-72ec-4445-9900-c679a1f73a5a\") " Oct 14 09:40:26 crc kubenswrapper[4870]: I1014 09:40:26.535510 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/591dfcef-72ec-4445-9900-c679a1f73a5a-utilities" (OuterVolumeSpecName: "utilities") pod "591dfcef-72ec-4445-9900-c679a1f73a5a" (UID: "591dfcef-72ec-4445-9900-c679a1f73a5a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 09:40:26 crc kubenswrapper[4870]: I1014 09:40:26.540622 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/591dfcef-72ec-4445-9900-c679a1f73a5a-kube-api-access-tx9ht" (OuterVolumeSpecName: "kube-api-access-tx9ht") pod "591dfcef-72ec-4445-9900-c679a1f73a5a" (UID: "591dfcef-72ec-4445-9900-c679a1f73a5a"). InnerVolumeSpecName "kube-api-access-tx9ht". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:40:26 crc kubenswrapper[4870]: I1014 09:40:26.603895 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/591dfcef-72ec-4445-9900-c679a1f73a5a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "591dfcef-72ec-4445-9900-c679a1f73a5a" (UID: "591dfcef-72ec-4445-9900-c679a1f73a5a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 09:40:26 crc kubenswrapper[4870]: I1014 09:40:26.636300 4870 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/591dfcef-72ec-4445-9900-c679a1f73a5a-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 09:40:26 crc kubenswrapper[4870]: I1014 09:40:26.636355 4870 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/591dfcef-72ec-4445-9900-c679a1f73a5a-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 09:40:26 crc kubenswrapper[4870]: I1014 09:40:26.636379 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tx9ht\" (UniqueName: \"kubernetes.io/projected/591dfcef-72ec-4445-9900-c679a1f73a5a-kube-api-access-tx9ht\") on node \"crc\" DevicePath \"\"" Oct 14 09:40:26 crc kubenswrapper[4870]: I1014 09:40:26.867802 4870 generic.go:334] "Generic (PLEG): container finished" podID="591dfcef-72ec-4445-9900-c679a1f73a5a" containerID="091a40917baf7e3ca48c776c2619b3d52bc941edd5b05d706046ede926ef6293" exitCode=0 Oct 14 09:40:26 crc kubenswrapper[4870]: I1014 09:40:26.867845 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-66rm7" event={"ID":"591dfcef-72ec-4445-9900-c679a1f73a5a","Type":"ContainerDied","Data":"091a40917baf7e3ca48c776c2619b3d52bc941edd5b05d706046ede926ef6293"} Oct 14 09:40:26 crc kubenswrapper[4870]: I1014 09:40:26.867870 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-66rm7" event={"ID":"591dfcef-72ec-4445-9900-c679a1f73a5a","Type":"ContainerDied","Data":"2ed15bbb40c429868f485cea2184fa93e5ce131f2eebe6137fbd548cef655049"} Oct 14 09:40:26 crc kubenswrapper[4870]: I1014 09:40:26.867896 4870 scope.go:117] "RemoveContainer" containerID="091a40917baf7e3ca48c776c2619b3d52bc941edd5b05d706046ede926ef6293" Oct 14 09:40:26 crc kubenswrapper[4870]: I1014 09:40:26.867933 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-66rm7" Oct 14 09:40:26 crc kubenswrapper[4870]: I1014 09:40:26.894987 4870 scope.go:117] "RemoveContainer" containerID="654fab7a99ef3161dd7fdeca3069b2648c0a9718930483dc7317e3cb3df5665d" Oct 14 09:40:26 crc kubenswrapper[4870]: I1014 09:40:26.921253 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-66rm7"] Oct 14 09:40:26 crc kubenswrapper[4870]: I1014 09:40:26.931592 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-66rm7"] Oct 14 09:40:26 crc kubenswrapper[4870]: I1014 09:40:26.936194 4870 scope.go:117] "RemoveContainer" containerID="c6bb00c25a1754dd60f845f7dcb78f2294bff46f1aa0c278741197ce68a9e21e" Oct 14 09:40:27 crc kubenswrapper[4870]: I1014 09:40:27.015850 4870 scope.go:117] "RemoveContainer" containerID="091a40917baf7e3ca48c776c2619b3d52bc941edd5b05d706046ede926ef6293" Oct 14 09:40:27 crc kubenswrapper[4870]: E1014 09:40:27.016266 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"091a40917baf7e3ca48c776c2619b3d52bc941edd5b05d706046ede926ef6293\": container with ID starting with 091a40917baf7e3ca48c776c2619b3d52bc941edd5b05d706046ede926ef6293 not found: ID does not exist" containerID="091a40917baf7e3ca48c776c2619b3d52bc941edd5b05d706046ede926ef6293" Oct 14 09:40:27 crc kubenswrapper[4870]: I1014 09:40:27.016289 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"091a40917baf7e3ca48c776c2619b3d52bc941edd5b05d706046ede926ef6293"} err="failed to get container status \"091a40917baf7e3ca48c776c2619b3d52bc941edd5b05d706046ede926ef6293\": rpc error: code = NotFound desc = could not find container \"091a40917baf7e3ca48c776c2619b3d52bc941edd5b05d706046ede926ef6293\": container with ID starting with 091a40917baf7e3ca48c776c2619b3d52bc941edd5b05d706046ede926ef6293 not found: ID does not exist" Oct 14 09:40:27 crc kubenswrapper[4870]: I1014 09:40:27.016311 4870 scope.go:117] "RemoveContainer" containerID="654fab7a99ef3161dd7fdeca3069b2648c0a9718930483dc7317e3cb3df5665d" Oct 14 09:40:27 crc kubenswrapper[4870]: E1014 09:40:27.016703 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"654fab7a99ef3161dd7fdeca3069b2648c0a9718930483dc7317e3cb3df5665d\": container with ID starting with 654fab7a99ef3161dd7fdeca3069b2648c0a9718930483dc7317e3cb3df5665d not found: ID does not exist" containerID="654fab7a99ef3161dd7fdeca3069b2648c0a9718930483dc7317e3cb3df5665d" Oct 14 09:40:27 crc kubenswrapper[4870]: I1014 09:40:27.016718 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"654fab7a99ef3161dd7fdeca3069b2648c0a9718930483dc7317e3cb3df5665d"} err="failed to get container status \"654fab7a99ef3161dd7fdeca3069b2648c0a9718930483dc7317e3cb3df5665d\": rpc error: code = NotFound desc = could not find container \"654fab7a99ef3161dd7fdeca3069b2648c0a9718930483dc7317e3cb3df5665d\": container with ID starting with 654fab7a99ef3161dd7fdeca3069b2648c0a9718930483dc7317e3cb3df5665d not found: ID does not exist" Oct 14 09:40:27 crc kubenswrapper[4870]: I1014 09:40:27.016734 4870 scope.go:117] "RemoveContainer" containerID="c6bb00c25a1754dd60f845f7dcb78f2294bff46f1aa0c278741197ce68a9e21e" Oct 14 09:40:27 crc kubenswrapper[4870]: E1014 09:40:27.016968 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c6bb00c25a1754dd60f845f7dcb78f2294bff46f1aa0c278741197ce68a9e21e\": container with ID starting with c6bb00c25a1754dd60f845f7dcb78f2294bff46f1aa0c278741197ce68a9e21e not found: ID does not exist" containerID="c6bb00c25a1754dd60f845f7dcb78f2294bff46f1aa0c278741197ce68a9e21e" Oct 14 09:40:27 crc kubenswrapper[4870]: I1014 09:40:27.016984 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c6bb00c25a1754dd60f845f7dcb78f2294bff46f1aa0c278741197ce68a9e21e"} err="failed to get container status \"c6bb00c25a1754dd60f845f7dcb78f2294bff46f1aa0c278741197ce68a9e21e\": rpc error: code = NotFound desc = could not find container \"c6bb00c25a1754dd60f845f7dcb78f2294bff46f1aa0c278741197ce68a9e21e\": container with ID starting with c6bb00c25a1754dd60f845f7dcb78f2294bff46f1aa0c278741197ce68a9e21e not found: ID does not exist" Oct 14 09:40:27 crc kubenswrapper[4870]: I1014 09:40:27.046310 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="591dfcef-72ec-4445-9900-c679a1f73a5a" path="/var/lib/kubelet/pods/591dfcef-72ec-4445-9900-c679a1f73a5a/volumes" Oct 14 09:40:53 crc kubenswrapper[4870]: I1014 09:40:53.951645 4870 patch_prober.go:28] interesting pod/machine-config-daemon-7tvc7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 09:40:53 crc kubenswrapper[4870]: I1014 09:40:53.952693 4870 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 09:41:23 crc kubenswrapper[4870]: I1014 09:41:23.950783 4870 patch_prober.go:28] interesting pod/machine-config-daemon-7tvc7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 09:41:23 crc kubenswrapper[4870]: I1014 09:41:23.951562 4870 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 09:41:23 crc kubenswrapper[4870]: I1014 09:41:23.951644 4870 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" Oct 14 09:41:23 crc kubenswrapper[4870]: I1014 09:41:23.952885 4870 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"6d5792c105b7bc7694bce51cbf8b085eda0318f921dffcecf5167175f90aa527"} pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 14 09:41:23 crc kubenswrapper[4870]: I1014 09:41:23.952995 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerName="machine-config-daemon" containerID="cri-o://6d5792c105b7bc7694bce51cbf8b085eda0318f921dffcecf5167175f90aa527" gracePeriod=600 Oct 14 09:41:24 crc kubenswrapper[4870]: E1014 09:41:24.092576 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 09:41:24 crc kubenswrapper[4870]: I1014 09:41:24.702253 4870 generic.go:334] "Generic (PLEG): container finished" podID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerID="6d5792c105b7bc7694bce51cbf8b085eda0318f921dffcecf5167175f90aa527" exitCode=0 Oct 14 09:41:24 crc kubenswrapper[4870]: I1014 09:41:24.702340 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" event={"ID":"874a5e45-dffd-4d17-b609-b7d3ed2eab07","Type":"ContainerDied","Data":"6d5792c105b7bc7694bce51cbf8b085eda0318f921dffcecf5167175f90aa527"} Oct 14 09:41:24 crc kubenswrapper[4870]: I1014 09:41:24.702768 4870 scope.go:117] "RemoveContainer" containerID="da933e0d9bd5082c26564c190825d7838bd95c38ef50de4f4eef55960b16e294" Oct 14 09:41:24 crc kubenswrapper[4870]: I1014 09:41:24.703525 4870 scope.go:117] "RemoveContainer" containerID="6d5792c105b7bc7694bce51cbf8b085eda0318f921dffcecf5167175f90aa527" Oct 14 09:41:24 crc kubenswrapper[4870]: E1014 09:41:24.704046 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 09:41:37 crc kubenswrapper[4870]: I1014 09:41:37.034409 4870 scope.go:117] "RemoveContainer" containerID="6d5792c105b7bc7694bce51cbf8b085eda0318f921dffcecf5167175f90aa527" Oct 14 09:41:37 crc kubenswrapper[4870]: E1014 09:41:37.038200 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 09:41:44 crc kubenswrapper[4870]: I1014 09:41:44.976845 4870 generic.go:334] "Generic (PLEG): container finished" podID="7f05f9ee-e89a-45a1-9f10-e68addbef873" containerID="79e5f5b43ed20f499f20ea656227a81276ec7adf1e8e2547aff2a69066f85e71" exitCode=0 Oct 14 09:41:44 crc kubenswrapper[4870]: I1014 09:41:44.976966 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-openstack-cell1-z74pq" event={"ID":"7f05f9ee-e89a-45a1-9f10-e68addbef873","Type":"ContainerDied","Data":"79e5f5b43ed20f499f20ea656227a81276ec7adf1e8e2547aff2a69066f85e71"} Oct 14 09:41:46 crc kubenswrapper[4870]: I1014 09:41:46.517914 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-openstack-cell1-z74pq" Oct 14 09:41:46 crc kubenswrapper[4870]: I1014 09:41:46.622006 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/7f05f9ee-e89a-45a1-9f10-e68addbef873-nova-migration-ssh-key-1\") pod \"7f05f9ee-e89a-45a1-9f10-e68addbef873\" (UID: \"7f05f9ee-e89a-45a1-9f10-e68addbef873\") " Oct 14 09:41:46 crc kubenswrapper[4870]: I1014 09:41:46.622122 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/7f05f9ee-e89a-45a1-9f10-e68addbef873-nova-cell1-compute-config-0\") pod \"7f05f9ee-e89a-45a1-9f10-e68addbef873\" (UID: \"7f05f9ee-e89a-45a1-9f10-e68addbef873\") " Oct 14 09:41:46 crc kubenswrapper[4870]: I1014 09:41:46.622170 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7f05f9ee-e89a-45a1-9f10-e68addbef873-inventory\") pod \"7f05f9ee-e89a-45a1-9f10-e68addbef873\" (UID: \"7f05f9ee-e89a-45a1-9f10-e68addbef873\") " Oct 14 09:41:46 crc kubenswrapper[4870]: I1014 09:41:46.622353 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7f05f9ee-e89a-45a1-9f10-e68addbef873-ssh-key\") pod \"7f05f9ee-e89a-45a1-9f10-e68addbef873\" (UID: \"7f05f9ee-e89a-45a1-9f10-e68addbef873\") " Oct 14 09:41:46 crc kubenswrapper[4870]: I1014 09:41:46.622395 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/7f05f9ee-e89a-45a1-9f10-e68addbef873-nova-migration-ssh-key-0\") pod \"7f05f9ee-e89a-45a1-9f10-e68addbef873\" (UID: \"7f05f9ee-e89a-45a1-9f10-e68addbef873\") " Oct 14 09:41:46 crc kubenswrapper[4870]: I1014 09:41:46.622496 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dtrcm\" (UniqueName: \"kubernetes.io/projected/7f05f9ee-e89a-45a1-9f10-e68addbef873-kube-api-access-dtrcm\") pod \"7f05f9ee-e89a-45a1-9f10-e68addbef873\" (UID: \"7f05f9ee-e89a-45a1-9f10-e68addbef873\") " Oct 14 09:41:46 crc kubenswrapper[4870]: I1014 09:41:46.622590 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/7f05f9ee-e89a-45a1-9f10-e68addbef873-nova-cell1-compute-config-1\") pod \"7f05f9ee-e89a-45a1-9f10-e68addbef873\" (UID: \"7f05f9ee-e89a-45a1-9f10-e68addbef873\") " Oct 14 09:41:46 crc kubenswrapper[4870]: I1014 09:41:46.622680 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/7f05f9ee-e89a-45a1-9f10-e68addbef873-ceph\") pod \"7f05f9ee-e89a-45a1-9f10-e68addbef873\" (UID: \"7f05f9ee-e89a-45a1-9f10-e68addbef873\") " Oct 14 09:41:46 crc kubenswrapper[4870]: I1014 09:41:46.622757 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f05f9ee-e89a-45a1-9f10-e68addbef873-nova-cell1-combined-ca-bundle\") pod \"7f05f9ee-e89a-45a1-9f10-e68addbef873\" (UID: \"7f05f9ee-e89a-45a1-9f10-e68addbef873\") " Oct 14 09:41:46 crc kubenswrapper[4870]: I1014 09:41:46.622848 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/7f05f9ee-e89a-45a1-9f10-e68addbef873-nova-cells-global-config-1\") pod \"7f05f9ee-e89a-45a1-9f10-e68addbef873\" (UID: \"7f05f9ee-e89a-45a1-9f10-e68addbef873\") " Oct 14 09:41:46 crc kubenswrapper[4870]: I1014 09:41:46.622897 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/7f05f9ee-e89a-45a1-9f10-e68addbef873-nova-cells-global-config-0\") pod \"7f05f9ee-e89a-45a1-9f10-e68addbef873\" (UID: \"7f05f9ee-e89a-45a1-9f10-e68addbef873\") " Oct 14 09:41:46 crc kubenswrapper[4870]: I1014 09:41:46.634148 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7f05f9ee-e89a-45a1-9f10-e68addbef873-kube-api-access-dtrcm" (OuterVolumeSpecName: "kube-api-access-dtrcm") pod "7f05f9ee-e89a-45a1-9f10-e68addbef873" (UID: "7f05f9ee-e89a-45a1-9f10-e68addbef873"). InnerVolumeSpecName "kube-api-access-dtrcm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:41:46 crc kubenswrapper[4870]: I1014 09:41:46.647401 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7f05f9ee-e89a-45a1-9f10-e68addbef873-ceph" (OuterVolumeSpecName: "ceph") pod "7f05f9ee-e89a-45a1-9f10-e68addbef873" (UID: "7f05f9ee-e89a-45a1-9f10-e68addbef873"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:41:46 crc kubenswrapper[4870]: I1014 09:41:46.647723 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7f05f9ee-e89a-45a1-9f10-e68addbef873-nova-cell1-combined-ca-bundle" (OuterVolumeSpecName: "nova-cell1-combined-ca-bundle") pod "7f05f9ee-e89a-45a1-9f10-e68addbef873" (UID: "7f05f9ee-e89a-45a1-9f10-e68addbef873"). InnerVolumeSpecName "nova-cell1-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:41:46 crc kubenswrapper[4870]: I1014 09:41:46.666233 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7f05f9ee-e89a-45a1-9f10-e68addbef873-nova-cells-global-config-1" (OuterVolumeSpecName: "nova-cells-global-config-1") pod "7f05f9ee-e89a-45a1-9f10-e68addbef873" (UID: "7f05f9ee-e89a-45a1-9f10-e68addbef873"). InnerVolumeSpecName "nova-cells-global-config-1". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 09:41:46 crc kubenswrapper[4870]: I1014 09:41:46.675993 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7f05f9ee-e89a-45a1-9f10-e68addbef873-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "7f05f9ee-e89a-45a1-9f10-e68addbef873" (UID: "7f05f9ee-e89a-45a1-9f10-e68addbef873"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:41:46 crc kubenswrapper[4870]: I1014 09:41:46.680623 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7f05f9ee-e89a-45a1-9f10-e68addbef873-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "7f05f9ee-e89a-45a1-9f10-e68addbef873" (UID: "7f05f9ee-e89a-45a1-9f10-e68addbef873"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:41:46 crc kubenswrapper[4870]: I1014 09:41:46.693004 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7f05f9ee-e89a-45a1-9f10-e68addbef873-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "7f05f9ee-e89a-45a1-9f10-e68addbef873" (UID: "7f05f9ee-e89a-45a1-9f10-e68addbef873"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:41:46 crc kubenswrapper[4870]: I1014 09:41:46.699033 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7f05f9ee-e89a-45a1-9f10-e68addbef873-nova-cells-global-config-0" (OuterVolumeSpecName: "nova-cells-global-config-0") pod "7f05f9ee-e89a-45a1-9f10-e68addbef873" (UID: "7f05f9ee-e89a-45a1-9f10-e68addbef873"). InnerVolumeSpecName "nova-cells-global-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 09:41:46 crc kubenswrapper[4870]: I1014 09:41:46.703686 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7f05f9ee-e89a-45a1-9f10-e68addbef873-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "7f05f9ee-e89a-45a1-9f10-e68addbef873" (UID: "7f05f9ee-e89a-45a1-9f10-e68addbef873"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:41:46 crc kubenswrapper[4870]: I1014 09:41:46.709702 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7f05f9ee-e89a-45a1-9f10-e68addbef873-inventory" (OuterVolumeSpecName: "inventory") pod "7f05f9ee-e89a-45a1-9f10-e68addbef873" (UID: "7f05f9ee-e89a-45a1-9f10-e68addbef873"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:41:46 crc kubenswrapper[4870]: I1014 09:41:46.710687 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7f05f9ee-e89a-45a1-9f10-e68addbef873-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "7f05f9ee-e89a-45a1-9f10-e68addbef873" (UID: "7f05f9ee-e89a-45a1-9f10-e68addbef873"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:41:46 crc kubenswrapper[4870]: I1014 09:41:46.725054 4870 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/7f05f9ee-e89a-45a1-9f10-e68addbef873-ceph\") on node \"crc\" DevicePath \"\"" Oct 14 09:41:46 crc kubenswrapper[4870]: I1014 09:41:46.725090 4870 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f05f9ee-e89a-45a1-9f10-e68addbef873-nova-cell1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 09:41:46 crc kubenswrapper[4870]: I1014 09:41:46.725128 4870 reconciler_common.go:293] "Volume detached for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/7f05f9ee-e89a-45a1-9f10-e68addbef873-nova-cells-global-config-1\") on node \"crc\" DevicePath \"\"" Oct 14 09:41:46 crc kubenswrapper[4870]: I1014 09:41:46.725140 4870 reconciler_common.go:293] "Volume detached for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/7f05f9ee-e89a-45a1-9f10-e68addbef873-nova-cells-global-config-0\") on node \"crc\" DevicePath \"\"" Oct 14 09:41:46 crc kubenswrapper[4870]: I1014 09:41:46.725153 4870 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/7f05f9ee-e89a-45a1-9f10-e68addbef873-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Oct 14 09:41:46 crc kubenswrapper[4870]: I1014 09:41:46.725163 4870 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/7f05f9ee-e89a-45a1-9f10-e68addbef873-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Oct 14 09:41:46 crc kubenswrapper[4870]: I1014 09:41:46.725174 4870 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7f05f9ee-e89a-45a1-9f10-e68addbef873-inventory\") on node \"crc\" DevicePath \"\"" Oct 14 09:41:46 crc kubenswrapper[4870]: I1014 09:41:46.725184 4870 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7f05f9ee-e89a-45a1-9f10-e68addbef873-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 14 09:41:46 crc kubenswrapper[4870]: I1014 09:41:46.725193 4870 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/7f05f9ee-e89a-45a1-9f10-e68addbef873-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Oct 14 09:41:46 crc kubenswrapper[4870]: I1014 09:41:46.725204 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dtrcm\" (UniqueName: \"kubernetes.io/projected/7f05f9ee-e89a-45a1-9f10-e68addbef873-kube-api-access-dtrcm\") on node \"crc\" DevicePath \"\"" Oct 14 09:41:46 crc kubenswrapper[4870]: I1014 09:41:46.725215 4870 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/7f05f9ee-e89a-45a1-9f10-e68addbef873-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Oct 14 09:41:47 crc kubenswrapper[4870]: I1014 09:41:47.030572 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-openstack-cell1-z74pq" event={"ID":"7f05f9ee-e89a-45a1-9f10-e68addbef873","Type":"ContainerDied","Data":"1ee51b2bc67e54ff74bad07c9dcc53b7b5d650eca97c4deb6adaca787d7ff659"} Oct 14 09:41:47 crc kubenswrapper[4870]: I1014 09:41:47.031223 4870 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1ee51b2bc67e54ff74bad07c9dcc53b7b5d650eca97c4deb6adaca787d7ff659" Oct 14 09:41:47 crc kubenswrapper[4870]: I1014 09:41:47.030751 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-openstack-cell1-z74pq" Oct 14 09:41:47 crc kubenswrapper[4870]: I1014 09:41:47.221200 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-openstack-openstack-cell1-q2bvq"] Oct 14 09:41:47 crc kubenswrapper[4870]: E1014 09:41:47.221981 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7f05f9ee-e89a-45a1-9f10-e68addbef873" containerName="nova-cell1-openstack-openstack-cell1" Oct 14 09:41:47 crc kubenswrapper[4870]: I1014 09:41:47.222027 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="7f05f9ee-e89a-45a1-9f10-e68addbef873" containerName="nova-cell1-openstack-openstack-cell1" Oct 14 09:41:47 crc kubenswrapper[4870]: E1014 09:41:47.222066 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="591dfcef-72ec-4445-9900-c679a1f73a5a" containerName="extract-utilities" Oct 14 09:41:47 crc kubenswrapper[4870]: I1014 09:41:47.222082 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="591dfcef-72ec-4445-9900-c679a1f73a5a" containerName="extract-utilities" Oct 14 09:41:47 crc kubenswrapper[4870]: E1014 09:41:47.222132 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="591dfcef-72ec-4445-9900-c679a1f73a5a" containerName="extract-content" Oct 14 09:41:47 crc kubenswrapper[4870]: I1014 09:41:47.222148 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="591dfcef-72ec-4445-9900-c679a1f73a5a" containerName="extract-content" Oct 14 09:41:47 crc kubenswrapper[4870]: E1014 09:41:47.222187 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="591dfcef-72ec-4445-9900-c679a1f73a5a" containerName="registry-server" Oct 14 09:41:47 crc kubenswrapper[4870]: I1014 09:41:47.222202 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="591dfcef-72ec-4445-9900-c679a1f73a5a" containerName="registry-server" Oct 14 09:41:47 crc kubenswrapper[4870]: I1014 09:41:47.222650 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="591dfcef-72ec-4445-9900-c679a1f73a5a" containerName="registry-server" Oct 14 09:41:47 crc kubenswrapper[4870]: I1014 09:41:47.222699 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="7f05f9ee-e89a-45a1-9f10-e68addbef873" containerName="nova-cell1-openstack-openstack-cell1" Oct 14 09:41:47 crc kubenswrapper[4870]: I1014 09:41:47.224701 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-openstack-openstack-cell1-q2bvq" Oct 14 09:41:47 crc kubenswrapper[4870]: I1014 09:41:47.229049 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 14 09:41:47 crc kubenswrapper[4870]: I1014 09:41:47.229182 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-wrz7w" Oct 14 09:41:47 crc kubenswrapper[4870]: I1014 09:41:47.229228 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 14 09:41:47 crc kubenswrapper[4870]: I1014 09:41:47.229308 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-compute-config-data" Oct 14 09:41:47 crc kubenswrapper[4870]: I1014 09:41:47.229100 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 14 09:41:47 crc kubenswrapper[4870]: I1014 09:41:47.233721 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-openstack-openstack-cell1-q2bvq"] Oct 14 09:41:47 crc kubenswrapper[4870]: I1014 09:41:47.338095 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7nkrn\" (UniqueName: \"kubernetes.io/projected/1358f0f9-0ff8-415f-9861-359ba368c8dd-kube-api-access-7nkrn\") pod \"telemetry-openstack-openstack-cell1-q2bvq\" (UID: \"1358f0f9-0ff8-415f-9861-359ba368c8dd\") " pod="openstack/telemetry-openstack-openstack-cell1-q2bvq" Oct 14 09:41:47 crc kubenswrapper[4870]: I1014 09:41:47.338187 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/1358f0f9-0ff8-415f-9861-359ba368c8dd-ceilometer-compute-config-data-0\") pod \"telemetry-openstack-openstack-cell1-q2bvq\" (UID: \"1358f0f9-0ff8-415f-9861-359ba368c8dd\") " pod="openstack/telemetry-openstack-openstack-cell1-q2bvq" Oct 14 09:41:47 crc kubenswrapper[4870]: I1014 09:41:47.338239 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1358f0f9-0ff8-415f-9861-359ba368c8dd-telemetry-combined-ca-bundle\") pod \"telemetry-openstack-openstack-cell1-q2bvq\" (UID: \"1358f0f9-0ff8-415f-9861-359ba368c8dd\") " pod="openstack/telemetry-openstack-openstack-cell1-q2bvq" Oct 14 09:41:47 crc kubenswrapper[4870]: I1014 09:41:47.338389 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/1358f0f9-0ff8-415f-9861-359ba368c8dd-ceilometer-compute-config-data-1\") pod \"telemetry-openstack-openstack-cell1-q2bvq\" (UID: \"1358f0f9-0ff8-415f-9861-359ba368c8dd\") " pod="openstack/telemetry-openstack-openstack-cell1-q2bvq" Oct 14 09:41:47 crc kubenswrapper[4870]: I1014 09:41:47.338429 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/1358f0f9-0ff8-415f-9861-359ba368c8dd-ceph\") pod \"telemetry-openstack-openstack-cell1-q2bvq\" (UID: \"1358f0f9-0ff8-415f-9861-359ba368c8dd\") " pod="openstack/telemetry-openstack-openstack-cell1-q2bvq" Oct 14 09:41:47 crc kubenswrapper[4870]: I1014 09:41:47.338485 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/1358f0f9-0ff8-415f-9861-359ba368c8dd-ceilometer-compute-config-data-2\") pod \"telemetry-openstack-openstack-cell1-q2bvq\" (UID: \"1358f0f9-0ff8-415f-9861-359ba368c8dd\") " pod="openstack/telemetry-openstack-openstack-cell1-q2bvq" Oct 14 09:41:47 crc kubenswrapper[4870]: I1014 09:41:47.338513 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1358f0f9-0ff8-415f-9861-359ba368c8dd-ssh-key\") pod \"telemetry-openstack-openstack-cell1-q2bvq\" (UID: \"1358f0f9-0ff8-415f-9861-359ba368c8dd\") " pod="openstack/telemetry-openstack-openstack-cell1-q2bvq" Oct 14 09:41:47 crc kubenswrapper[4870]: I1014 09:41:47.338565 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1358f0f9-0ff8-415f-9861-359ba368c8dd-inventory\") pod \"telemetry-openstack-openstack-cell1-q2bvq\" (UID: \"1358f0f9-0ff8-415f-9861-359ba368c8dd\") " pod="openstack/telemetry-openstack-openstack-cell1-q2bvq" Oct 14 09:41:47 crc kubenswrapper[4870]: I1014 09:41:47.441349 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/1358f0f9-0ff8-415f-9861-359ba368c8dd-ceilometer-compute-config-data-0\") pod \"telemetry-openstack-openstack-cell1-q2bvq\" (UID: \"1358f0f9-0ff8-415f-9861-359ba368c8dd\") " pod="openstack/telemetry-openstack-openstack-cell1-q2bvq" Oct 14 09:41:47 crc kubenswrapper[4870]: I1014 09:41:47.441452 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1358f0f9-0ff8-415f-9861-359ba368c8dd-telemetry-combined-ca-bundle\") pod \"telemetry-openstack-openstack-cell1-q2bvq\" (UID: \"1358f0f9-0ff8-415f-9861-359ba368c8dd\") " pod="openstack/telemetry-openstack-openstack-cell1-q2bvq" Oct 14 09:41:47 crc kubenswrapper[4870]: I1014 09:41:47.441564 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/1358f0f9-0ff8-415f-9861-359ba368c8dd-ceilometer-compute-config-data-1\") pod \"telemetry-openstack-openstack-cell1-q2bvq\" (UID: \"1358f0f9-0ff8-415f-9861-359ba368c8dd\") " pod="openstack/telemetry-openstack-openstack-cell1-q2bvq" Oct 14 09:41:47 crc kubenswrapper[4870]: I1014 09:41:47.441596 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/1358f0f9-0ff8-415f-9861-359ba368c8dd-ceph\") pod \"telemetry-openstack-openstack-cell1-q2bvq\" (UID: \"1358f0f9-0ff8-415f-9861-359ba368c8dd\") " pod="openstack/telemetry-openstack-openstack-cell1-q2bvq" Oct 14 09:41:47 crc kubenswrapper[4870]: I1014 09:41:47.441631 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/1358f0f9-0ff8-415f-9861-359ba368c8dd-ceilometer-compute-config-data-2\") pod \"telemetry-openstack-openstack-cell1-q2bvq\" (UID: \"1358f0f9-0ff8-415f-9861-359ba368c8dd\") " pod="openstack/telemetry-openstack-openstack-cell1-q2bvq" Oct 14 09:41:47 crc kubenswrapper[4870]: I1014 09:41:47.441661 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1358f0f9-0ff8-415f-9861-359ba368c8dd-ssh-key\") pod \"telemetry-openstack-openstack-cell1-q2bvq\" (UID: \"1358f0f9-0ff8-415f-9861-359ba368c8dd\") " pod="openstack/telemetry-openstack-openstack-cell1-q2bvq" Oct 14 09:41:47 crc kubenswrapper[4870]: I1014 09:41:47.441725 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1358f0f9-0ff8-415f-9861-359ba368c8dd-inventory\") pod \"telemetry-openstack-openstack-cell1-q2bvq\" (UID: \"1358f0f9-0ff8-415f-9861-359ba368c8dd\") " pod="openstack/telemetry-openstack-openstack-cell1-q2bvq" Oct 14 09:41:47 crc kubenswrapper[4870]: I1014 09:41:47.441818 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7nkrn\" (UniqueName: \"kubernetes.io/projected/1358f0f9-0ff8-415f-9861-359ba368c8dd-kube-api-access-7nkrn\") pod \"telemetry-openstack-openstack-cell1-q2bvq\" (UID: \"1358f0f9-0ff8-415f-9861-359ba368c8dd\") " pod="openstack/telemetry-openstack-openstack-cell1-q2bvq" Oct 14 09:41:47 crc kubenswrapper[4870]: I1014 09:41:47.646818 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/1358f0f9-0ff8-415f-9861-359ba368c8dd-ceilometer-compute-config-data-0\") pod \"telemetry-openstack-openstack-cell1-q2bvq\" (UID: \"1358f0f9-0ff8-415f-9861-359ba368c8dd\") " pod="openstack/telemetry-openstack-openstack-cell1-q2bvq" Oct 14 09:41:47 crc kubenswrapper[4870]: I1014 09:41:47.646820 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1358f0f9-0ff8-415f-9861-359ba368c8dd-inventory\") pod \"telemetry-openstack-openstack-cell1-q2bvq\" (UID: \"1358f0f9-0ff8-415f-9861-359ba368c8dd\") " pod="openstack/telemetry-openstack-openstack-cell1-q2bvq" Oct 14 09:41:47 crc kubenswrapper[4870]: I1014 09:41:47.647092 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/1358f0f9-0ff8-415f-9861-359ba368c8dd-ceilometer-compute-config-data-1\") pod \"telemetry-openstack-openstack-cell1-q2bvq\" (UID: \"1358f0f9-0ff8-415f-9861-359ba368c8dd\") " pod="openstack/telemetry-openstack-openstack-cell1-q2bvq" Oct 14 09:41:47 crc kubenswrapper[4870]: I1014 09:41:47.647315 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/1358f0f9-0ff8-415f-9861-359ba368c8dd-ceilometer-compute-config-data-2\") pod \"telemetry-openstack-openstack-cell1-q2bvq\" (UID: \"1358f0f9-0ff8-415f-9861-359ba368c8dd\") " pod="openstack/telemetry-openstack-openstack-cell1-q2bvq" Oct 14 09:41:47 crc kubenswrapper[4870]: I1014 09:41:47.647764 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1358f0f9-0ff8-415f-9861-359ba368c8dd-ssh-key\") pod \"telemetry-openstack-openstack-cell1-q2bvq\" (UID: \"1358f0f9-0ff8-415f-9861-359ba368c8dd\") " pod="openstack/telemetry-openstack-openstack-cell1-q2bvq" Oct 14 09:41:47 crc kubenswrapper[4870]: I1014 09:41:47.648087 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/1358f0f9-0ff8-415f-9861-359ba368c8dd-ceph\") pod \"telemetry-openstack-openstack-cell1-q2bvq\" (UID: \"1358f0f9-0ff8-415f-9861-359ba368c8dd\") " pod="openstack/telemetry-openstack-openstack-cell1-q2bvq" Oct 14 09:41:47 crc kubenswrapper[4870]: I1014 09:41:47.650091 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1358f0f9-0ff8-415f-9861-359ba368c8dd-telemetry-combined-ca-bundle\") pod \"telemetry-openstack-openstack-cell1-q2bvq\" (UID: \"1358f0f9-0ff8-415f-9861-359ba368c8dd\") " pod="openstack/telemetry-openstack-openstack-cell1-q2bvq" Oct 14 09:41:47 crc kubenswrapper[4870]: I1014 09:41:47.658337 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7nkrn\" (UniqueName: \"kubernetes.io/projected/1358f0f9-0ff8-415f-9861-359ba368c8dd-kube-api-access-7nkrn\") pod \"telemetry-openstack-openstack-cell1-q2bvq\" (UID: \"1358f0f9-0ff8-415f-9861-359ba368c8dd\") " pod="openstack/telemetry-openstack-openstack-cell1-q2bvq" Oct 14 09:41:47 crc kubenswrapper[4870]: I1014 09:41:47.863940 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-openstack-openstack-cell1-q2bvq" Oct 14 09:41:48 crc kubenswrapper[4870]: I1014 09:41:48.035977 4870 scope.go:117] "RemoveContainer" containerID="6d5792c105b7bc7694bce51cbf8b085eda0318f921dffcecf5167175f90aa527" Oct 14 09:41:48 crc kubenswrapper[4870]: E1014 09:41:48.037436 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 09:41:48 crc kubenswrapper[4870]: I1014 09:41:48.552710 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-openstack-openstack-cell1-q2bvq"] Oct 14 09:41:49 crc kubenswrapper[4870]: I1014 09:41:49.072850 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-openstack-openstack-cell1-q2bvq" event={"ID":"1358f0f9-0ff8-415f-9861-359ba368c8dd","Type":"ContainerStarted","Data":"f47173a6a1ff9d180b133ec1d2ed3f160abb7e522937a39b0b7715734e20f8a1"} Oct 14 09:41:50 crc kubenswrapper[4870]: I1014 09:41:50.080088 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-openstack-openstack-cell1-q2bvq" event={"ID":"1358f0f9-0ff8-415f-9861-359ba368c8dd","Type":"ContainerStarted","Data":"4626063e8ca122ab584e37783a27fd4546253d894e17055595a5404d18ce4acd"} Oct 14 09:41:50 crc kubenswrapper[4870]: I1014 09:41:50.109754 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-openstack-openstack-cell1-q2bvq" podStartSLOduration=2.326398402 podStartE2EDuration="3.109737276s" podCreationTimestamp="2025-10-14 09:41:47 +0000 UTC" firstStartedPulling="2025-10-14 09:41:48.544773791 +0000 UTC m=+9644.242134192" lastFinishedPulling="2025-10-14 09:41:49.328112665 +0000 UTC m=+9645.025473066" observedRunningTime="2025-10-14 09:41:50.104979559 +0000 UTC m=+9645.802339930" watchObservedRunningTime="2025-10-14 09:41:50.109737276 +0000 UTC m=+9645.807097647" Oct 14 09:42:00 crc kubenswrapper[4870]: I1014 09:42:00.034343 4870 scope.go:117] "RemoveContainer" containerID="6d5792c105b7bc7694bce51cbf8b085eda0318f921dffcecf5167175f90aa527" Oct 14 09:42:00 crc kubenswrapper[4870]: E1014 09:42:00.035621 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 09:42:13 crc kubenswrapper[4870]: I1014 09:42:13.035748 4870 scope.go:117] "RemoveContainer" containerID="6d5792c105b7bc7694bce51cbf8b085eda0318f921dffcecf5167175f90aa527" Oct 14 09:42:13 crc kubenswrapper[4870]: E1014 09:42:13.037150 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 09:42:25 crc kubenswrapper[4870]: I1014 09:42:25.064296 4870 scope.go:117] "RemoveContainer" containerID="6d5792c105b7bc7694bce51cbf8b085eda0318f921dffcecf5167175f90aa527" Oct 14 09:42:25 crc kubenswrapper[4870]: E1014 09:42:25.065620 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 09:42:37 crc kubenswrapper[4870]: I1014 09:42:37.035693 4870 scope.go:117] "RemoveContainer" containerID="6d5792c105b7bc7694bce51cbf8b085eda0318f921dffcecf5167175f90aa527" Oct 14 09:42:37 crc kubenswrapper[4870]: E1014 09:42:37.037066 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 09:42:49 crc kubenswrapper[4870]: I1014 09:42:49.034775 4870 scope.go:117] "RemoveContainer" containerID="6d5792c105b7bc7694bce51cbf8b085eda0318f921dffcecf5167175f90aa527" Oct 14 09:42:49 crc kubenswrapper[4870]: E1014 09:42:49.035969 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 09:43:04 crc kubenswrapper[4870]: I1014 09:43:04.034821 4870 scope.go:117] "RemoveContainer" containerID="6d5792c105b7bc7694bce51cbf8b085eda0318f921dffcecf5167175f90aa527" Oct 14 09:43:04 crc kubenswrapper[4870]: E1014 09:43:04.035971 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 09:43:15 crc kubenswrapper[4870]: I1014 09:43:15.050990 4870 scope.go:117] "RemoveContainer" containerID="6d5792c105b7bc7694bce51cbf8b085eda0318f921dffcecf5167175f90aa527" Oct 14 09:43:15 crc kubenswrapper[4870]: E1014 09:43:15.052029 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 09:43:28 crc kubenswrapper[4870]: I1014 09:43:28.035381 4870 scope.go:117] "RemoveContainer" containerID="6d5792c105b7bc7694bce51cbf8b085eda0318f921dffcecf5167175f90aa527" Oct 14 09:43:28 crc kubenswrapper[4870]: E1014 09:43:28.036299 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 09:43:42 crc kubenswrapper[4870]: I1014 09:43:42.034798 4870 scope.go:117] "RemoveContainer" containerID="6d5792c105b7bc7694bce51cbf8b085eda0318f921dffcecf5167175f90aa527" Oct 14 09:43:42 crc kubenswrapper[4870]: E1014 09:43:42.036073 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 09:43:46 crc kubenswrapper[4870]: I1014 09:43:46.985965 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-x4pql"] Oct 14 09:43:46 crc kubenswrapper[4870]: I1014 09:43:46.991503 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-x4pql" Oct 14 09:43:47 crc kubenswrapper[4870]: I1014 09:43:47.000379 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-x4pql"] Oct 14 09:43:47 crc kubenswrapper[4870]: I1014 09:43:47.127711 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/34134b95-2953-41b0-9108-571578e8f291-catalog-content\") pod \"redhat-operators-x4pql\" (UID: \"34134b95-2953-41b0-9108-571578e8f291\") " pod="openshift-marketplace/redhat-operators-x4pql" Oct 14 09:43:47 crc kubenswrapper[4870]: I1014 09:43:47.127815 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/34134b95-2953-41b0-9108-571578e8f291-utilities\") pod \"redhat-operators-x4pql\" (UID: \"34134b95-2953-41b0-9108-571578e8f291\") " pod="openshift-marketplace/redhat-operators-x4pql" Oct 14 09:43:47 crc kubenswrapper[4870]: I1014 09:43:47.128051 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-94s8m\" (UniqueName: \"kubernetes.io/projected/34134b95-2953-41b0-9108-571578e8f291-kube-api-access-94s8m\") pod \"redhat-operators-x4pql\" (UID: \"34134b95-2953-41b0-9108-571578e8f291\") " pod="openshift-marketplace/redhat-operators-x4pql" Oct 14 09:43:47 crc kubenswrapper[4870]: I1014 09:43:47.230161 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-94s8m\" (UniqueName: \"kubernetes.io/projected/34134b95-2953-41b0-9108-571578e8f291-kube-api-access-94s8m\") pod \"redhat-operators-x4pql\" (UID: \"34134b95-2953-41b0-9108-571578e8f291\") " pod="openshift-marketplace/redhat-operators-x4pql" Oct 14 09:43:47 crc kubenswrapper[4870]: I1014 09:43:47.230206 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/34134b95-2953-41b0-9108-571578e8f291-catalog-content\") pod \"redhat-operators-x4pql\" (UID: \"34134b95-2953-41b0-9108-571578e8f291\") " pod="openshift-marketplace/redhat-operators-x4pql" Oct 14 09:43:47 crc kubenswrapper[4870]: I1014 09:43:47.230258 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/34134b95-2953-41b0-9108-571578e8f291-utilities\") pod \"redhat-operators-x4pql\" (UID: \"34134b95-2953-41b0-9108-571578e8f291\") " pod="openshift-marketplace/redhat-operators-x4pql" Oct 14 09:43:47 crc kubenswrapper[4870]: I1014 09:43:47.230784 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/34134b95-2953-41b0-9108-571578e8f291-catalog-content\") pod \"redhat-operators-x4pql\" (UID: \"34134b95-2953-41b0-9108-571578e8f291\") " pod="openshift-marketplace/redhat-operators-x4pql" Oct 14 09:43:47 crc kubenswrapper[4870]: I1014 09:43:47.230825 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/34134b95-2953-41b0-9108-571578e8f291-utilities\") pod \"redhat-operators-x4pql\" (UID: \"34134b95-2953-41b0-9108-571578e8f291\") " pod="openshift-marketplace/redhat-operators-x4pql" Oct 14 09:43:47 crc kubenswrapper[4870]: I1014 09:43:47.250314 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-94s8m\" (UniqueName: \"kubernetes.io/projected/34134b95-2953-41b0-9108-571578e8f291-kube-api-access-94s8m\") pod \"redhat-operators-x4pql\" (UID: \"34134b95-2953-41b0-9108-571578e8f291\") " pod="openshift-marketplace/redhat-operators-x4pql" Oct 14 09:43:47 crc kubenswrapper[4870]: I1014 09:43:47.317340 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-x4pql" Oct 14 09:43:47 crc kubenswrapper[4870]: I1014 09:43:47.786033 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-x4pql"] Oct 14 09:43:48 crc kubenswrapper[4870]: I1014 09:43:48.596563 4870 generic.go:334] "Generic (PLEG): container finished" podID="34134b95-2953-41b0-9108-571578e8f291" containerID="f47306cf237946e5ea13bbfc13ccfcc64d14c8bfbf5765a1961eb665e176635f" exitCode=0 Oct 14 09:43:48 crc kubenswrapper[4870]: I1014 09:43:48.596719 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x4pql" event={"ID":"34134b95-2953-41b0-9108-571578e8f291","Type":"ContainerDied","Data":"f47306cf237946e5ea13bbfc13ccfcc64d14c8bfbf5765a1961eb665e176635f"} Oct 14 09:43:48 crc kubenswrapper[4870]: I1014 09:43:48.596940 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x4pql" event={"ID":"34134b95-2953-41b0-9108-571578e8f291","Type":"ContainerStarted","Data":"d64eebd33b65470d5ab2aa0af3be515441ddca2a8347ed8934eb1ef55f6ed6c8"} Oct 14 09:43:49 crc kubenswrapper[4870]: I1014 09:43:49.611792 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x4pql" event={"ID":"34134b95-2953-41b0-9108-571578e8f291","Type":"ContainerStarted","Data":"9f020610380da4cc009d56c712c9195ebafff1e27e1e08f885ff06a6ddd0bd82"} Oct 14 09:43:53 crc kubenswrapper[4870]: I1014 09:43:53.035276 4870 scope.go:117] "RemoveContainer" containerID="6d5792c105b7bc7694bce51cbf8b085eda0318f921dffcecf5167175f90aa527" Oct 14 09:43:53 crc kubenswrapper[4870]: E1014 09:43:53.038342 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 09:43:53 crc kubenswrapper[4870]: I1014 09:43:53.657249 4870 generic.go:334] "Generic (PLEG): container finished" podID="34134b95-2953-41b0-9108-571578e8f291" containerID="9f020610380da4cc009d56c712c9195ebafff1e27e1e08f885ff06a6ddd0bd82" exitCode=0 Oct 14 09:43:53 crc kubenswrapper[4870]: I1014 09:43:53.657302 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x4pql" event={"ID":"34134b95-2953-41b0-9108-571578e8f291","Type":"ContainerDied","Data":"9f020610380da4cc009d56c712c9195ebafff1e27e1e08f885ff06a6ddd0bd82"} Oct 14 09:43:54 crc kubenswrapper[4870]: I1014 09:43:54.671587 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x4pql" event={"ID":"34134b95-2953-41b0-9108-571578e8f291","Type":"ContainerStarted","Data":"2faba40e44783eb52f06b36eb985d2f5ec65cf2800da4a199cbcc5bce4674ebd"} Oct 14 09:43:54 crc kubenswrapper[4870]: I1014 09:43:54.698316 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-x4pql" podStartSLOduration=3.138319676 podStartE2EDuration="8.698295741s" podCreationTimestamp="2025-10-14 09:43:46 +0000 UTC" firstStartedPulling="2025-10-14 09:43:48.598922055 +0000 UTC m=+9764.296282436" lastFinishedPulling="2025-10-14 09:43:54.15889811 +0000 UTC m=+9769.856258501" observedRunningTime="2025-10-14 09:43:54.693914623 +0000 UTC m=+9770.391275004" watchObservedRunningTime="2025-10-14 09:43:54.698295741 +0000 UTC m=+9770.395656122" Oct 14 09:43:55 crc kubenswrapper[4870]: I1014 09:43:55.660885 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-whg22"] Oct 14 09:43:55 crc kubenswrapper[4870]: I1014 09:43:55.663241 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-whg22" Oct 14 09:43:55 crc kubenswrapper[4870]: I1014 09:43:55.687883 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-whg22"] Oct 14 09:43:55 crc kubenswrapper[4870]: I1014 09:43:55.717372 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mfhrk\" (UniqueName: \"kubernetes.io/projected/32dd5f3d-46c8-4c01-9316-cfc0be533b15-kube-api-access-mfhrk\") pod \"community-operators-whg22\" (UID: \"32dd5f3d-46c8-4c01-9316-cfc0be533b15\") " pod="openshift-marketplace/community-operators-whg22" Oct 14 09:43:55 crc kubenswrapper[4870]: I1014 09:43:55.717427 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/32dd5f3d-46c8-4c01-9316-cfc0be533b15-utilities\") pod \"community-operators-whg22\" (UID: \"32dd5f3d-46c8-4c01-9316-cfc0be533b15\") " pod="openshift-marketplace/community-operators-whg22" Oct 14 09:43:55 crc kubenswrapper[4870]: I1014 09:43:55.717480 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/32dd5f3d-46c8-4c01-9316-cfc0be533b15-catalog-content\") pod \"community-operators-whg22\" (UID: \"32dd5f3d-46c8-4c01-9316-cfc0be533b15\") " pod="openshift-marketplace/community-operators-whg22" Oct 14 09:43:55 crc kubenswrapper[4870]: I1014 09:43:55.819996 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mfhrk\" (UniqueName: \"kubernetes.io/projected/32dd5f3d-46c8-4c01-9316-cfc0be533b15-kube-api-access-mfhrk\") pod \"community-operators-whg22\" (UID: \"32dd5f3d-46c8-4c01-9316-cfc0be533b15\") " pod="openshift-marketplace/community-operators-whg22" Oct 14 09:43:55 crc kubenswrapper[4870]: I1014 09:43:55.820080 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/32dd5f3d-46c8-4c01-9316-cfc0be533b15-utilities\") pod \"community-operators-whg22\" (UID: \"32dd5f3d-46c8-4c01-9316-cfc0be533b15\") " pod="openshift-marketplace/community-operators-whg22" Oct 14 09:43:55 crc kubenswrapper[4870]: I1014 09:43:55.820114 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/32dd5f3d-46c8-4c01-9316-cfc0be533b15-catalog-content\") pod \"community-operators-whg22\" (UID: \"32dd5f3d-46c8-4c01-9316-cfc0be533b15\") " pod="openshift-marketplace/community-operators-whg22" Oct 14 09:43:55 crc kubenswrapper[4870]: I1014 09:43:55.820806 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/32dd5f3d-46c8-4c01-9316-cfc0be533b15-catalog-content\") pod \"community-operators-whg22\" (UID: \"32dd5f3d-46c8-4c01-9316-cfc0be533b15\") " pod="openshift-marketplace/community-operators-whg22" Oct 14 09:43:55 crc kubenswrapper[4870]: I1014 09:43:55.821344 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/32dd5f3d-46c8-4c01-9316-cfc0be533b15-utilities\") pod \"community-operators-whg22\" (UID: \"32dd5f3d-46c8-4c01-9316-cfc0be533b15\") " pod="openshift-marketplace/community-operators-whg22" Oct 14 09:43:56 crc kubenswrapper[4870]: I1014 09:43:56.057684 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mfhrk\" (UniqueName: \"kubernetes.io/projected/32dd5f3d-46c8-4c01-9316-cfc0be533b15-kube-api-access-mfhrk\") pod \"community-operators-whg22\" (UID: \"32dd5f3d-46c8-4c01-9316-cfc0be533b15\") " pod="openshift-marketplace/community-operators-whg22" Oct 14 09:43:56 crc kubenswrapper[4870]: I1014 09:43:56.340936 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-whg22" Oct 14 09:43:56 crc kubenswrapper[4870]: I1014 09:43:56.867509 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-whg22"] Oct 14 09:43:57 crc kubenswrapper[4870]: I1014 09:43:57.318961 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-x4pql" Oct 14 09:43:57 crc kubenswrapper[4870]: I1014 09:43:57.319379 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-x4pql" Oct 14 09:43:57 crc kubenswrapper[4870]: I1014 09:43:57.716191 4870 generic.go:334] "Generic (PLEG): container finished" podID="32dd5f3d-46c8-4c01-9316-cfc0be533b15" containerID="1fe13df2a8e818718696f389f8554e4e42c5e391059fe9add8519317e571a9da" exitCode=0 Oct 14 09:43:57 crc kubenswrapper[4870]: I1014 09:43:57.716242 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-whg22" event={"ID":"32dd5f3d-46c8-4c01-9316-cfc0be533b15","Type":"ContainerDied","Data":"1fe13df2a8e818718696f389f8554e4e42c5e391059fe9add8519317e571a9da"} Oct 14 09:43:57 crc kubenswrapper[4870]: I1014 09:43:57.716297 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-whg22" event={"ID":"32dd5f3d-46c8-4c01-9316-cfc0be533b15","Type":"ContainerStarted","Data":"3fdec7c7f74b26a674c9b9b979ab799ab076a41951fcc601f459e71ba7501a08"} Oct 14 09:43:58 crc kubenswrapper[4870]: I1014 09:43:58.374446 4870 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-x4pql" podUID="34134b95-2953-41b0-9108-571578e8f291" containerName="registry-server" probeResult="failure" output=< Oct 14 09:43:58 crc kubenswrapper[4870]: timeout: failed to connect service ":50051" within 1s Oct 14 09:43:58 crc kubenswrapper[4870]: > Oct 14 09:43:58 crc kubenswrapper[4870]: I1014 09:43:58.733769 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-whg22" event={"ID":"32dd5f3d-46c8-4c01-9316-cfc0be533b15","Type":"ContainerStarted","Data":"abeb0402a3c5a3a05e725fdf4dec5ea9f7f9ffe34c508183e3fd1fd56bd07626"} Oct 14 09:44:00 crc kubenswrapper[4870]: I1014 09:44:00.757320 4870 generic.go:334] "Generic (PLEG): container finished" podID="32dd5f3d-46c8-4c01-9316-cfc0be533b15" containerID="abeb0402a3c5a3a05e725fdf4dec5ea9f7f9ffe34c508183e3fd1fd56bd07626" exitCode=0 Oct 14 09:44:00 crc kubenswrapper[4870]: I1014 09:44:00.757501 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-whg22" event={"ID":"32dd5f3d-46c8-4c01-9316-cfc0be533b15","Type":"ContainerDied","Data":"abeb0402a3c5a3a05e725fdf4dec5ea9f7f9ffe34c508183e3fd1fd56bd07626"} Oct 14 09:44:01 crc kubenswrapper[4870]: I1014 09:44:01.768885 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-whg22" event={"ID":"32dd5f3d-46c8-4c01-9316-cfc0be533b15","Type":"ContainerStarted","Data":"e25538b09bd54376d95fbce6182de00ee0f1e17bd73fd221f910280e21528847"} Oct 14 09:44:01 crc kubenswrapper[4870]: I1014 09:44:01.795321 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-whg22" podStartSLOduration=3.324010069 podStartE2EDuration="6.795298033s" podCreationTimestamp="2025-10-14 09:43:55 +0000 UTC" firstStartedPulling="2025-10-14 09:43:57.71872628 +0000 UTC m=+9773.416086651" lastFinishedPulling="2025-10-14 09:44:01.190014244 +0000 UTC m=+9776.887374615" observedRunningTime="2025-10-14 09:44:01.78418669 +0000 UTC m=+9777.481547061" watchObservedRunningTime="2025-10-14 09:44:01.795298033 +0000 UTC m=+9777.492658404" Oct 14 09:44:06 crc kubenswrapper[4870]: I1014 09:44:06.341535 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-whg22" Oct 14 09:44:06 crc kubenswrapper[4870]: I1014 09:44:06.342201 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-whg22" Oct 14 09:44:07 crc kubenswrapper[4870]: I1014 09:44:07.408398 4870 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-whg22" podUID="32dd5f3d-46c8-4c01-9316-cfc0be533b15" containerName="registry-server" probeResult="failure" output=< Oct 14 09:44:07 crc kubenswrapper[4870]: timeout: failed to connect service ":50051" within 1s Oct 14 09:44:07 crc kubenswrapper[4870]: > Oct 14 09:44:08 crc kubenswrapper[4870]: I1014 09:44:08.035648 4870 scope.go:117] "RemoveContainer" containerID="6d5792c105b7bc7694bce51cbf8b085eda0318f921dffcecf5167175f90aa527" Oct 14 09:44:08 crc kubenswrapper[4870]: E1014 09:44:08.037078 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 09:44:08 crc kubenswrapper[4870]: I1014 09:44:08.366854 4870 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-x4pql" podUID="34134b95-2953-41b0-9108-571578e8f291" containerName="registry-server" probeResult="failure" output=< Oct 14 09:44:08 crc kubenswrapper[4870]: timeout: failed to connect service ":50051" within 1s Oct 14 09:44:08 crc kubenswrapper[4870]: > Oct 14 09:44:16 crc kubenswrapper[4870]: I1014 09:44:16.405620 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-whg22" Oct 14 09:44:16 crc kubenswrapper[4870]: I1014 09:44:16.510558 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-whg22" Oct 14 09:44:16 crc kubenswrapper[4870]: I1014 09:44:16.649214 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-whg22"] Oct 14 09:44:17 crc kubenswrapper[4870]: I1014 09:44:17.949572 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-whg22" podUID="32dd5f3d-46c8-4c01-9316-cfc0be533b15" containerName="registry-server" containerID="cri-o://e25538b09bd54376d95fbce6182de00ee0f1e17bd73fd221f910280e21528847" gracePeriod=2 Oct 14 09:44:18 crc kubenswrapper[4870]: I1014 09:44:18.370983 4870 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-x4pql" podUID="34134b95-2953-41b0-9108-571578e8f291" containerName="registry-server" probeResult="failure" output=< Oct 14 09:44:18 crc kubenswrapper[4870]: timeout: failed to connect service ":50051" within 1s Oct 14 09:44:18 crc kubenswrapper[4870]: > Oct 14 09:44:18 crc kubenswrapper[4870]: I1014 09:44:18.519578 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-whg22" Oct 14 09:44:18 crc kubenswrapper[4870]: I1014 09:44:18.539100 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/32dd5f3d-46c8-4c01-9316-cfc0be533b15-utilities\") pod \"32dd5f3d-46c8-4c01-9316-cfc0be533b15\" (UID: \"32dd5f3d-46c8-4c01-9316-cfc0be533b15\") " Oct 14 09:44:18 crc kubenswrapper[4870]: I1014 09:44:18.539373 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mfhrk\" (UniqueName: \"kubernetes.io/projected/32dd5f3d-46c8-4c01-9316-cfc0be533b15-kube-api-access-mfhrk\") pod \"32dd5f3d-46c8-4c01-9316-cfc0be533b15\" (UID: \"32dd5f3d-46c8-4c01-9316-cfc0be533b15\") " Oct 14 09:44:18 crc kubenswrapper[4870]: I1014 09:44:18.539529 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/32dd5f3d-46c8-4c01-9316-cfc0be533b15-catalog-content\") pod \"32dd5f3d-46c8-4c01-9316-cfc0be533b15\" (UID: \"32dd5f3d-46c8-4c01-9316-cfc0be533b15\") " Oct 14 09:44:18 crc kubenswrapper[4870]: I1014 09:44:18.539883 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/32dd5f3d-46c8-4c01-9316-cfc0be533b15-utilities" (OuterVolumeSpecName: "utilities") pod "32dd5f3d-46c8-4c01-9316-cfc0be533b15" (UID: "32dd5f3d-46c8-4c01-9316-cfc0be533b15"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 09:44:18 crc kubenswrapper[4870]: I1014 09:44:18.548753 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/32dd5f3d-46c8-4c01-9316-cfc0be533b15-kube-api-access-mfhrk" (OuterVolumeSpecName: "kube-api-access-mfhrk") pod "32dd5f3d-46c8-4c01-9316-cfc0be533b15" (UID: "32dd5f3d-46c8-4c01-9316-cfc0be533b15"). InnerVolumeSpecName "kube-api-access-mfhrk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:44:18 crc kubenswrapper[4870]: I1014 09:44:18.580814 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/32dd5f3d-46c8-4c01-9316-cfc0be533b15-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "32dd5f3d-46c8-4c01-9316-cfc0be533b15" (UID: "32dd5f3d-46c8-4c01-9316-cfc0be533b15"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 09:44:18 crc kubenswrapper[4870]: I1014 09:44:18.640873 4870 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/32dd5f3d-46c8-4c01-9316-cfc0be533b15-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 09:44:18 crc kubenswrapper[4870]: I1014 09:44:18.640908 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mfhrk\" (UniqueName: \"kubernetes.io/projected/32dd5f3d-46c8-4c01-9316-cfc0be533b15-kube-api-access-mfhrk\") on node \"crc\" DevicePath \"\"" Oct 14 09:44:18 crc kubenswrapper[4870]: I1014 09:44:18.640917 4870 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/32dd5f3d-46c8-4c01-9316-cfc0be533b15-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 09:44:18 crc kubenswrapper[4870]: I1014 09:44:18.962244 4870 generic.go:334] "Generic (PLEG): container finished" podID="32dd5f3d-46c8-4c01-9316-cfc0be533b15" containerID="e25538b09bd54376d95fbce6182de00ee0f1e17bd73fd221f910280e21528847" exitCode=0 Oct 14 09:44:18 crc kubenswrapper[4870]: I1014 09:44:18.962312 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-whg22" event={"ID":"32dd5f3d-46c8-4c01-9316-cfc0be533b15","Type":"ContainerDied","Data":"e25538b09bd54376d95fbce6182de00ee0f1e17bd73fd221f910280e21528847"} Oct 14 09:44:18 crc kubenswrapper[4870]: I1014 09:44:18.962371 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-whg22" event={"ID":"32dd5f3d-46c8-4c01-9316-cfc0be533b15","Type":"ContainerDied","Data":"3fdec7c7f74b26a674c9b9b979ab799ab076a41951fcc601f459e71ba7501a08"} Oct 14 09:44:18 crc kubenswrapper[4870]: I1014 09:44:18.962368 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-whg22" Oct 14 09:44:18 crc kubenswrapper[4870]: I1014 09:44:18.962396 4870 scope.go:117] "RemoveContainer" containerID="e25538b09bd54376d95fbce6182de00ee0f1e17bd73fd221f910280e21528847" Oct 14 09:44:18 crc kubenswrapper[4870]: I1014 09:44:18.991525 4870 scope.go:117] "RemoveContainer" containerID="abeb0402a3c5a3a05e725fdf4dec5ea9f7f9ffe34c508183e3fd1fd56bd07626" Oct 14 09:44:19 crc kubenswrapper[4870]: I1014 09:44:19.014980 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-whg22"] Oct 14 09:44:19 crc kubenswrapper[4870]: I1014 09:44:19.045407 4870 scope.go:117] "RemoveContainer" containerID="6d5792c105b7bc7694bce51cbf8b085eda0318f921dffcecf5167175f90aa527" Oct 14 09:44:19 crc kubenswrapper[4870]: E1014 09:44:19.045958 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 09:44:19 crc kubenswrapper[4870]: I1014 09:44:19.070328 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-whg22"] Oct 14 09:44:19 crc kubenswrapper[4870]: I1014 09:44:19.893973 4870 scope.go:117] "RemoveContainer" containerID="1fe13df2a8e818718696f389f8554e4e42c5e391059fe9add8519317e571a9da" Oct 14 09:44:19 crc kubenswrapper[4870]: I1014 09:44:19.962455 4870 scope.go:117] "RemoveContainer" containerID="e25538b09bd54376d95fbce6182de00ee0f1e17bd73fd221f910280e21528847" Oct 14 09:44:19 crc kubenswrapper[4870]: E1014 09:44:19.963064 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e25538b09bd54376d95fbce6182de00ee0f1e17bd73fd221f910280e21528847\": container with ID starting with e25538b09bd54376d95fbce6182de00ee0f1e17bd73fd221f910280e21528847 not found: ID does not exist" containerID="e25538b09bd54376d95fbce6182de00ee0f1e17bd73fd221f910280e21528847" Oct 14 09:44:19 crc kubenswrapper[4870]: I1014 09:44:19.963119 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e25538b09bd54376d95fbce6182de00ee0f1e17bd73fd221f910280e21528847"} err="failed to get container status \"e25538b09bd54376d95fbce6182de00ee0f1e17bd73fd221f910280e21528847\": rpc error: code = NotFound desc = could not find container \"e25538b09bd54376d95fbce6182de00ee0f1e17bd73fd221f910280e21528847\": container with ID starting with e25538b09bd54376d95fbce6182de00ee0f1e17bd73fd221f910280e21528847 not found: ID does not exist" Oct 14 09:44:19 crc kubenswrapper[4870]: I1014 09:44:19.963149 4870 scope.go:117] "RemoveContainer" containerID="abeb0402a3c5a3a05e725fdf4dec5ea9f7f9ffe34c508183e3fd1fd56bd07626" Oct 14 09:44:19 crc kubenswrapper[4870]: E1014 09:44:19.963619 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"abeb0402a3c5a3a05e725fdf4dec5ea9f7f9ffe34c508183e3fd1fd56bd07626\": container with ID starting with abeb0402a3c5a3a05e725fdf4dec5ea9f7f9ffe34c508183e3fd1fd56bd07626 not found: ID does not exist" containerID="abeb0402a3c5a3a05e725fdf4dec5ea9f7f9ffe34c508183e3fd1fd56bd07626" Oct 14 09:44:19 crc kubenswrapper[4870]: I1014 09:44:19.963646 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"abeb0402a3c5a3a05e725fdf4dec5ea9f7f9ffe34c508183e3fd1fd56bd07626"} err="failed to get container status \"abeb0402a3c5a3a05e725fdf4dec5ea9f7f9ffe34c508183e3fd1fd56bd07626\": rpc error: code = NotFound desc = could not find container \"abeb0402a3c5a3a05e725fdf4dec5ea9f7f9ffe34c508183e3fd1fd56bd07626\": container with ID starting with abeb0402a3c5a3a05e725fdf4dec5ea9f7f9ffe34c508183e3fd1fd56bd07626 not found: ID does not exist" Oct 14 09:44:19 crc kubenswrapper[4870]: I1014 09:44:19.963668 4870 scope.go:117] "RemoveContainer" containerID="1fe13df2a8e818718696f389f8554e4e42c5e391059fe9add8519317e571a9da" Oct 14 09:44:19 crc kubenswrapper[4870]: E1014 09:44:19.963893 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1fe13df2a8e818718696f389f8554e4e42c5e391059fe9add8519317e571a9da\": container with ID starting with 1fe13df2a8e818718696f389f8554e4e42c5e391059fe9add8519317e571a9da not found: ID does not exist" containerID="1fe13df2a8e818718696f389f8554e4e42c5e391059fe9add8519317e571a9da" Oct 14 09:44:19 crc kubenswrapper[4870]: I1014 09:44:19.963912 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1fe13df2a8e818718696f389f8554e4e42c5e391059fe9add8519317e571a9da"} err="failed to get container status \"1fe13df2a8e818718696f389f8554e4e42c5e391059fe9add8519317e571a9da\": rpc error: code = NotFound desc = could not find container \"1fe13df2a8e818718696f389f8554e4e42c5e391059fe9add8519317e571a9da\": container with ID starting with 1fe13df2a8e818718696f389f8554e4e42c5e391059fe9add8519317e571a9da not found: ID does not exist" Oct 14 09:44:21 crc kubenswrapper[4870]: I1014 09:44:21.059337 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="32dd5f3d-46c8-4c01-9316-cfc0be533b15" path="/var/lib/kubelet/pods/32dd5f3d-46c8-4c01-9316-cfc0be533b15/volumes" Oct 14 09:44:27 crc kubenswrapper[4870]: I1014 09:44:27.361995 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-x4pql" Oct 14 09:44:27 crc kubenswrapper[4870]: I1014 09:44:27.440663 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-x4pql" Oct 14 09:44:27 crc kubenswrapper[4870]: I1014 09:44:27.612996 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-x4pql"] Oct 14 09:44:29 crc kubenswrapper[4870]: I1014 09:44:29.106977 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-x4pql" podUID="34134b95-2953-41b0-9108-571578e8f291" containerName="registry-server" containerID="cri-o://2faba40e44783eb52f06b36eb985d2f5ec65cf2800da4a199cbcc5bce4674ebd" gracePeriod=2 Oct 14 09:44:29 crc kubenswrapper[4870]: I1014 09:44:29.701403 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-x4pql" Oct 14 09:44:29 crc kubenswrapper[4870]: I1014 09:44:29.814245 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/34134b95-2953-41b0-9108-571578e8f291-utilities\") pod \"34134b95-2953-41b0-9108-571578e8f291\" (UID: \"34134b95-2953-41b0-9108-571578e8f291\") " Oct 14 09:44:29 crc kubenswrapper[4870]: I1014 09:44:29.814799 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-94s8m\" (UniqueName: \"kubernetes.io/projected/34134b95-2953-41b0-9108-571578e8f291-kube-api-access-94s8m\") pod \"34134b95-2953-41b0-9108-571578e8f291\" (UID: \"34134b95-2953-41b0-9108-571578e8f291\") " Oct 14 09:44:29 crc kubenswrapper[4870]: I1014 09:44:29.814925 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/34134b95-2953-41b0-9108-571578e8f291-catalog-content\") pod \"34134b95-2953-41b0-9108-571578e8f291\" (UID: \"34134b95-2953-41b0-9108-571578e8f291\") " Oct 14 09:44:29 crc kubenswrapper[4870]: I1014 09:44:29.815169 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/34134b95-2953-41b0-9108-571578e8f291-utilities" (OuterVolumeSpecName: "utilities") pod "34134b95-2953-41b0-9108-571578e8f291" (UID: "34134b95-2953-41b0-9108-571578e8f291"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 09:44:29 crc kubenswrapper[4870]: I1014 09:44:29.815577 4870 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/34134b95-2953-41b0-9108-571578e8f291-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 09:44:29 crc kubenswrapper[4870]: I1014 09:44:29.820587 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/34134b95-2953-41b0-9108-571578e8f291-kube-api-access-94s8m" (OuterVolumeSpecName: "kube-api-access-94s8m") pod "34134b95-2953-41b0-9108-571578e8f291" (UID: "34134b95-2953-41b0-9108-571578e8f291"). InnerVolumeSpecName "kube-api-access-94s8m". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:44:29 crc kubenswrapper[4870]: I1014 09:44:29.886868 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/34134b95-2953-41b0-9108-571578e8f291-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "34134b95-2953-41b0-9108-571578e8f291" (UID: "34134b95-2953-41b0-9108-571578e8f291"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 09:44:29 crc kubenswrapper[4870]: I1014 09:44:29.917894 4870 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/34134b95-2953-41b0-9108-571578e8f291-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 09:44:29 crc kubenswrapper[4870]: I1014 09:44:29.917947 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-94s8m\" (UniqueName: \"kubernetes.io/projected/34134b95-2953-41b0-9108-571578e8f291-kube-api-access-94s8m\") on node \"crc\" DevicePath \"\"" Oct 14 09:44:30 crc kubenswrapper[4870]: I1014 09:44:30.119496 4870 generic.go:334] "Generic (PLEG): container finished" podID="34134b95-2953-41b0-9108-571578e8f291" containerID="2faba40e44783eb52f06b36eb985d2f5ec65cf2800da4a199cbcc5bce4674ebd" exitCode=0 Oct 14 09:44:30 crc kubenswrapper[4870]: I1014 09:44:30.119548 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-x4pql" Oct 14 09:44:30 crc kubenswrapper[4870]: I1014 09:44:30.119576 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x4pql" event={"ID":"34134b95-2953-41b0-9108-571578e8f291","Type":"ContainerDied","Data":"2faba40e44783eb52f06b36eb985d2f5ec65cf2800da4a199cbcc5bce4674ebd"} Oct 14 09:44:30 crc kubenswrapper[4870]: I1014 09:44:30.119635 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x4pql" event={"ID":"34134b95-2953-41b0-9108-571578e8f291","Type":"ContainerDied","Data":"d64eebd33b65470d5ab2aa0af3be515441ddca2a8347ed8934eb1ef55f6ed6c8"} Oct 14 09:44:30 crc kubenswrapper[4870]: I1014 09:44:30.119656 4870 scope.go:117] "RemoveContainer" containerID="2faba40e44783eb52f06b36eb985d2f5ec65cf2800da4a199cbcc5bce4674ebd" Oct 14 09:44:30 crc kubenswrapper[4870]: I1014 09:44:30.150846 4870 scope.go:117] "RemoveContainer" containerID="9f020610380da4cc009d56c712c9195ebafff1e27e1e08f885ff06a6ddd0bd82" Oct 14 09:44:30 crc kubenswrapper[4870]: I1014 09:44:30.155300 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-x4pql"] Oct 14 09:44:30 crc kubenswrapper[4870]: I1014 09:44:30.164265 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-x4pql"] Oct 14 09:44:30 crc kubenswrapper[4870]: I1014 09:44:30.181352 4870 scope.go:117] "RemoveContainer" containerID="f47306cf237946e5ea13bbfc13ccfcc64d14c8bfbf5765a1961eb665e176635f" Oct 14 09:44:30 crc kubenswrapper[4870]: I1014 09:44:30.236728 4870 scope.go:117] "RemoveContainer" containerID="2faba40e44783eb52f06b36eb985d2f5ec65cf2800da4a199cbcc5bce4674ebd" Oct 14 09:44:30 crc kubenswrapper[4870]: E1014 09:44:30.237477 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2faba40e44783eb52f06b36eb985d2f5ec65cf2800da4a199cbcc5bce4674ebd\": container with ID starting with 2faba40e44783eb52f06b36eb985d2f5ec65cf2800da4a199cbcc5bce4674ebd not found: ID does not exist" containerID="2faba40e44783eb52f06b36eb985d2f5ec65cf2800da4a199cbcc5bce4674ebd" Oct 14 09:44:30 crc kubenswrapper[4870]: I1014 09:44:30.237526 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2faba40e44783eb52f06b36eb985d2f5ec65cf2800da4a199cbcc5bce4674ebd"} err="failed to get container status \"2faba40e44783eb52f06b36eb985d2f5ec65cf2800da4a199cbcc5bce4674ebd\": rpc error: code = NotFound desc = could not find container \"2faba40e44783eb52f06b36eb985d2f5ec65cf2800da4a199cbcc5bce4674ebd\": container with ID starting with 2faba40e44783eb52f06b36eb985d2f5ec65cf2800da4a199cbcc5bce4674ebd not found: ID does not exist" Oct 14 09:44:30 crc kubenswrapper[4870]: I1014 09:44:30.237553 4870 scope.go:117] "RemoveContainer" containerID="9f020610380da4cc009d56c712c9195ebafff1e27e1e08f885ff06a6ddd0bd82" Oct 14 09:44:30 crc kubenswrapper[4870]: E1014 09:44:30.237988 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9f020610380da4cc009d56c712c9195ebafff1e27e1e08f885ff06a6ddd0bd82\": container with ID starting with 9f020610380da4cc009d56c712c9195ebafff1e27e1e08f885ff06a6ddd0bd82 not found: ID does not exist" containerID="9f020610380da4cc009d56c712c9195ebafff1e27e1e08f885ff06a6ddd0bd82" Oct 14 09:44:30 crc kubenswrapper[4870]: I1014 09:44:30.238028 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9f020610380da4cc009d56c712c9195ebafff1e27e1e08f885ff06a6ddd0bd82"} err="failed to get container status \"9f020610380da4cc009d56c712c9195ebafff1e27e1e08f885ff06a6ddd0bd82\": rpc error: code = NotFound desc = could not find container \"9f020610380da4cc009d56c712c9195ebafff1e27e1e08f885ff06a6ddd0bd82\": container with ID starting with 9f020610380da4cc009d56c712c9195ebafff1e27e1e08f885ff06a6ddd0bd82 not found: ID does not exist" Oct 14 09:44:30 crc kubenswrapper[4870]: I1014 09:44:30.238056 4870 scope.go:117] "RemoveContainer" containerID="f47306cf237946e5ea13bbfc13ccfcc64d14c8bfbf5765a1961eb665e176635f" Oct 14 09:44:30 crc kubenswrapper[4870]: E1014 09:44:30.238344 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f47306cf237946e5ea13bbfc13ccfcc64d14c8bfbf5765a1961eb665e176635f\": container with ID starting with f47306cf237946e5ea13bbfc13ccfcc64d14c8bfbf5765a1961eb665e176635f not found: ID does not exist" containerID="f47306cf237946e5ea13bbfc13ccfcc64d14c8bfbf5765a1961eb665e176635f" Oct 14 09:44:30 crc kubenswrapper[4870]: I1014 09:44:30.238379 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f47306cf237946e5ea13bbfc13ccfcc64d14c8bfbf5765a1961eb665e176635f"} err="failed to get container status \"f47306cf237946e5ea13bbfc13ccfcc64d14c8bfbf5765a1961eb665e176635f\": rpc error: code = NotFound desc = could not find container \"f47306cf237946e5ea13bbfc13ccfcc64d14c8bfbf5765a1961eb665e176635f\": container with ID starting with f47306cf237946e5ea13bbfc13ccfcc64d14c8bfbf5765a1961eb665e176635f not found: ID does not exist" Oct 14 09:44:31 crc kubenswrapper[4870]: I1014 09:44:31.058408 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="34134b95-2953-41b0-9108-571578e8f291" path="/var/lib/kubelet/pods/34134b95-2953-41b0-9108-571578e8f291/volumes" Oct 14 09:44:33 crc kubenswrapper[4870]: I1014 09:44:33.034472 4870 scope.go:117] "RemoveContainer" containerID="6d5792c105b7bc7694bce51cbf8b085eda0318f921dffcecf5167175f90aa527" Oct 14 09:44:33 crc kubenswrapper[4870]: E1014 09:44:33.035238 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 09:44:47 crc kubenswrapper[4870]: I1014 09:44:47.034791 4870 scope.go:117] "RemoveContainer" containerID="6d5792c105b7bc7694bce51cbf8b085eda0318f921dffcecf5167175f90aa527" Oct 14 09:44:47 crc kubenswrapper[4870]: E1014 09:44:47.035748 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 09:45:00 crc kubenswrapper[4870]: I1014 09:45:00.034812 4870 scope.go:117] "RemoveContainer" containerID="6d5792c105b7bc7694bce51cbf8b085eda0318f921dffcecf5167175f90aa527" Oct 14 09:45:00 crc kubenswrapper[4870]: E1014 09:45:00.035779 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 09:45:00 crc kubenswrapper[4870]: I1014 09:45:00.189651 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29340585-q9shp"] Oct 14 09:45:00 crc kubenswrapper[4870]: E1014 09:45:00.191359 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34134b95-2953-41b0-9108-571578e8f291" containerName="extract-content" Oct 14 09:45:00 crc kubenswrapper[4870]: I1014 09:45:00.191414 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="34134b95-2953-41b0-9108-571578e8f291" containerName="extract-content" Oct 14 09:45:00 crc kubenswrapper[4870]: E1014 09:45:00.191600 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="32dd5f3d-46c8-4c01-9316-cfc0be533b15" containerName="extract-content" Oct 14 09:45:00 crc kubenswrapper[4870]: I1014 09:45:00.191636 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="32dd5f3d-46c8-4c01-9316-cfc0be533b15" containerName="extract-content" Oct 14 09:45:00 crc kubenswrapper[4870]: E1014 09:45:00.191684 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="32dd5f3d-46c8-4c01-9316-cfc0be533b15" containerName="registry-server" Oct 14 09:45:00 crc kubenswrapper[4870]: I1014 09:45:00.191705 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="32dd5f3d-46c8-4c01-9316-cfc0be533b15" containerName="registry-server" Oct 14 09:45:00 crc kubenswrapper[4870]: E1014 09:45:00.191745 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="32dd5f3d-46c8-4c01-9316-cfc0be533b15" containerName="extract-utilities" Oct 14 09:45:00 crc kubenswrapper[4870]: I1014 09:45:00.191767 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="32dd5f3d-46c8-4c01-9316-cfc0be533b15" containerName="extract-utilities" Oct 14 09:45:00 crc kubenswrapper[4870]: E1014 09:45:00.191829 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34134b95-2953-41b0-9108-571578e8f291" containerName="extract-utilities" Oct 14 09:45:00 crc kubenswrapper[4870]: I1014 09:45:00.191849 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="34134b95-2953-41b0-9108-571578e8f291" containerName="extract-utilities" Oct 14 09:45:00 crc kubenswrapper[4870]: E1014 09:45:00.191891 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34134b95-2953-41b0-9108-571578e8f291" containerName="registry-server" Oct 14 09:45:00 crc kubenswrapper[4870]: I1014 09:45:00.191911 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="34134b95-2953-41b0-9108-571578e8f291" containerName="registry-server" Oct 14 09:45:00 crc kubenswrapper[4870]: I1014 09:45:00.193028 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="34134b95-2953-41b0-9108-571578e8f291" containerName="registry-server" Oct 14 09:45:00 crc kubenswrapper[4870]: I1014 09:45:00.193128 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="32dd5f3d-46c8-4c01-9316-cfc0be533b15" containerName="registry-server" Oct 14 09:45:00 crc kubenswrapper[4870]: I1014 09:45:00.196414 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29340585-q9shp" Oct 14 09:45:00 crc kubenswrapper[4870]: I1014 09:45:00.199501 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29340585-q9shp"] Oct 14 09:45:00 crc kubenswrapper[4870]: I1014 09:45:00.201705 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 14 09:45:00 crc kubenswrapper[4870]: I1014 09:45:00.201994 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 14 09:45:00 crc kubenswrapper[4870]: I1014 09:45:00.328789 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8ps4v\" (UniqueName: \"kubernetes.io/projected/0ffb284a-79dd-4b65-b824-c1c0211d4a0a-kube-api-access-8ps4v\") pod \"collect-profiles-29340585-q9shp\" (UID: \"0ffb284a-79dd-4b65-b824-c1c0211d4a0a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340585-q9shp" Oct 14 09:45:00 crc kubenswrapper[4870]: I1014 09:45:00.328933 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0ffb284a-79dd-4b65-b824-c1c0211d4a0a-secret-volume\") pod \"collect-profiles-29340585-q9shp\" (UID: \"0ffb284a-79dd-4b65-b824-c1c0211d4a0a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340585-q9shp" Oct 14 09:45:00 crc kubenswrapper[4870]: I1014 09:45:00.329002 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0ffb284a-79dd-4b65-b824-c1c0211d4a0a-config-volume\") pod \"collect-profiles-29340585-q9shp\" (UID: \"0ffb284a-79dd-4b65-b824-c1c0211d4a0a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340585-q9shp" Oct 14 09:45:00 crc kubenswrapper[4870]: I1014 09:45:00.432078 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8ps4v\" (UniqueName: \"kubernetes.io/projected/0ffb284a-79dd-4b65-b824-c1c0211d4a0a-kube-api-access-8ps4v\") pod \"collect-profiles-29340585-q9shp\" (UID: \"0ffb284a-79dd-4b65-b824-c1c0211d4a0a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340585-q9shp" Oct 14 09:45:00 crc kubenswrapper[4870]: I1014 09:45:00.432222 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0ffb284a-79dd-4b65-b824-c1c0211d4a0a-secret-volume\") pod \"collect-profiles-29340585-q9shp\" (UID: \"0ffb284a-79dd-4b65-b824-c1c0211d4a0a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340585-q9shp" Oct 14 09:45:00 crc kubenswrapper[4870]: I1014 09:45:00.432306 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0ffb284a-79dd-4b65-b824-c1c0211d4a0a-config-volume\") pod \"collect-profiles-29340585-q9shp\" (UID: \"0ffb284a-79dd-4b65-b824-c1c0211d4a0a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340585-q9shp" Oct 14 09:45:00 crc kubenswrapper[4870]: I1014 09:45:00.433926 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0ffb284a-79dd-4b65-b824-c1c0211d4a0a-config-volume\") pod \"collect-profiles-29340585-q9shp\" (UID: \"0ffb284a-79dd-4b65-b824-c1c0211d4a0a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340585-q9shp" Oct 14 09:45:00 crc kubenswrapper[4870]: I1014 09:45:00.441850 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0ffb284a-79dd-4b65-b824-c1c0211d4a0a-secret-volume\") pod \"collect-profiles-29340585-q9shp\" (UID: \"0ffb284a-79dd-4b65-b824-c1c0211d4a0a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340585-q9shp" Oct 14 09:45:00 crc kubenswrapper[4870]: I1014 09:45:00.450589 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8ps4v\" (UniqueName: \"kubernetes.io/projected/0ffb284a-79dd-4b65-b824-c1c0211d4a0a-kube-api-access-8ps4v\") pod \"collect-profiles-29340585-q9shp\" (UID: \"0ffb284a-79dd-4b65-b824-c1c0211d4a0a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340585-q9shp" Oct 14 09:45:00 crc kubenswrapper[4870]: I1014 09:45:00.538251 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29340585-q9shp" Oct 14 09:45:01 crc kubenswrapper[4870]: I1014 09:45:01.064684 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29340585-q9shp"] Oct 14 09:45:01 crc kubenswrapper[4870]: I1014 09:45:01.539773 4870 generic.go:334] "Generic (PLEG): container finished" podID="0ffb284a-79dd-4b65-b824-c1c0211d4a0a" containerID="d019c9eaf21cb0cbc817f0291648727457f79b6bd70d6e62f41211f6d66a537a" exitCode=0 Oct 14 09:45:01 crc kubenswrapper[4870]: I1014 09:45:01.539831 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29340585-q9shp" event={"ID":"0ffb284a-79dd-4b65-b824-c1c0211d4a0a","Type":"ContainerDied","Data":"d019c9eaf21cb0cbc817f0291648727457f79b6bd70d6e62f41211f6d66a537a"} Oct 14 09:45:01 crc kubenswrapper[4870]: I1014 09:45:01.539874 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29340585-q9shp" event={"ID":"0ffb284a-79dd-4b65-b824-c1c0211d4a0a","Type":"ContainerStarted","Data":"3977776d048925b15f778356477a1ba1cc7f4953c2c5c6d126637070342d0d95"} Oct 14 09:45:03 crc kubenswrapper[4870]: I1014 09:45:03.066949 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29340585-q9shp" Oct 14 09:45:03 crc kubenswrapper[4870]: I1014 09:45:03.200702 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0ffb284a-79dd-4b65-b824-c1c0211d4a0a-secret-volume\") pod \"0ffb284a-79dd-4b65-b824-c1c0211d4a0a\" (UID: \"0ffb284a-79dd-4b65-b824-c1c0211d4a0a\") " Oct 14 09:45:03 crc kubenswrapper[4870]: I1014 09:45:03.201353 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8ps4v\" (UniqueName: \"kubernetes.io/projected/0ffb284a-79dd-4b65-b824-c1c0211d4a0a-kube-api-access-8ps4v\") pod \"0ffb284a-79dd-4b65-b824-c1c0211d4a0a\" (UID: \"0ffb284a-79dd-4b65-b824-c1c0211d4a0a\") " Oct 14 09:45:03 crc kubenswrapper[4870]: I1014 09:45:03.201398 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0ffb284a-79dd-4b65-b824-c1c0211d4a0a-config-volume\") pod \"0ffb284a-79dd-4b65-b824-c1c0211d4a0a\" (UID: \"0ffb284a-79dd-4b65-b824-c1c0211d4a0a\") " Oct 14 09:45:03 crc kubenswrapper[4870]: I1014 09:45:03.202140 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0ffb284a-79dd-4b65-b824-c1c0211d4a0a-config-volume" (OuterVolumeSpecName: "config-volume") pod "0ffb284a-79dd-4b65-b824-c1c0211d4a0a" (UID: "0ffb284a-79dd-4b65-b824-c1c0211d4a0a"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 09:45:03 crc kubenswrapper[4870]: I1014 09:45:03.202381 4870 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0ffb284a-79dd-4b65-b824-c1c0211d4a0a-config-volume\") on node \"crc\" DevicePath \"\"" Oct 14 09:45:03 crc kubenswrapper[4870]: I1014 09:45:03.213847 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0ffb284a-79dd-4b65-b824-c1c0211d4a0a-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "0ffb284a-79dd-4b65-b824-c1c0211d4a0a" (UID: "0ffb284a-79dd-4b65-b824-c1c0211d4a0a"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:45:03 crc kubenswrapper[4870]: I1014 09:45:03.217164 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0ffb284a-79dd-4b65-b824-c1c0211d4a0a-kube-api-access-8ps4v" (OuterVolumeSpecName: "kube-api-access-8ps4v") pod "0ffb284a-79dd-4b65-b824-c1c0211d4a0a" (UID: "0ffb284a-79dd-4b65-b824-c1c0211d4a0a"). InnerVolumeSpecName "kube-api-access-8ps4v". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:45:03 crc kubenswrapper[4870]: I1014 09:45:03.304766 4870 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0ffb284a-79dd-4b65-b824-c1c0211d4a0a-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 14 09:45:03 crc kubenswrapper[4870]: I1014 09:45:03.304841 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8ps4v\" (UniqueName: \"kubernetes.io/projected/0ffb284a-79dd-4b65-b824-c1c0211d4a0a-kube-api-access-8ps4v\") on node \"crc\" DevicePath \"\"" Oct 14 09:45:03 crc kubenswrapper[4870]: I1014 09:45:03.586974 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29340585-q9shp" event={"ID":"0ffb284a-79dd-4b65-b824-c1c0211d4a0a","Type":"ContainerDied","Data":"3977776d048925b15f778356477a1ba1cc7f4953c2c5c6d126637070342d0d95"} Oct 14 09:45:03 crc kubenswrapper[4870]: I1014 09:45:03.587035 4870 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3977776d048925b15f778356477a1ba1cc7f4953c2c5c6d126637070342d0d95" Oct 14 09:45:03 crc kubenswrapper[4870]: I1014 09:45:03.587102 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29340585-q9shp" Oct 14 09:45:04 crc kubenswrapper[4870]: I1014 09:45:04.165718 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29340540-jw8p5"] Oct 14 09:45:04 crc kubenswrapper[4870]: I1014 09:45:04.177423 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29340540-jw8p5"] Oct 14 09:45:05 crc kubenswrapper[4870]: I1014 09:45:05.053959 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4c777f9d-8c71-4d17-ac52-ae341bd066a2" path="/var/lib/kubelet/pods/4c777f9d-8c71-4d17-ac52-ae341bd066a2/volumes" Oct 14 09:45:14 crc kubenswrapper[4870]: I1014 09:45:14.035107 4870 scope.go:117] "RemoveContainer" containerID="6d5792c105b7bc7694bce51cbf8b085eda0318f921dffcecf5167175f90aa527" Oct 14 09:45:14 crc kubenswrapper[4870]: E1014 09:45:14.036425 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 09:45:29 crc kubenswrapper[4870]: I1014 09:45:29.034885 4870 scope.go:117] "RemoveContainer" containerID="6d5792c105b7bc7694bce51cbf8b085eda0318f921dffcecf5167175f90aa527" Oct 14 09:45:29 crc kubenswrapper[4870]: E1014 09:45:29.035923 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 09:45:33 crc kubenswrapper[4870]: I1014 09:45:33.521768 4870 scope.go:117] "RemoveContainer" containerID="6d6e738bef449215a9a338332d1ae8ffd723ed0b9e5443e70020c01178dcd790" Oct 14 09:45:44 crc kubenswrapper[4870]: I1014 09:45:44.035477 4870 scope.go:117] "RemoveContainer" containerID="6d5792c105b7bc7694bce51cbf8b085eda0318f921dffcecf5167175f90aa527" Oct 14 09:45:44 crc kubenswrapper[4870]: E1014 09:45:44.036685 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 09:45:55 crc kubenswrapper[4870]: I1014 09:45:55.041461 4870 scope.go:117] "RemoveContainer" containerID="6d5792c105b7bc7694bce51cbf8b085eda0318f921dffcecf5167175f90aa527" Oct 14 09:45:55 crc kubenswrapper[4870]: E1014 09:45:55.042698 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 09:46:09 crc kubenswrapper[4870]: I1014 09:46:09.035882 4870 scope.go:117] "RemoveContainer" containerID="6d5792c105b7bc7694bce51cbf8b085eda0318f921dffcecf5167175f90aa527" Oct 14 09:46:09 crc kubenswrapper[4870]: E1014 09:46:09.037616 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 09:46:24 crc kubenswrapper[4870]: I1014 09:46:24.035165 4870 scope.go:117] "RemoveContainer" containerID="6d5792c105b7bc7694bce51cbf8b085eda0318f921dffcecf5167175f90aa527" Oct 14 09:46:24 crc kubenswrapper[4870]: I1014 09:46:24.754419 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" event={"ID":"874a5e45-dffd-4d17-b609-b7d3ed2eab07","Type":"ContainerStarted","Data":"4d8d8702edb8183ff39b29852bc295d1ac61316a22904ed8a1cfe721472e2347"} Oct 14 09:48:53 crc kubenswrapper[4870]: I1014 09:48:53.950967 4870 patch_prober.go:28] interesting pod/machine-config-daemon-7tvc7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 09:48:53 crc kubenswrapper[4870]: I1014 09:48:53.951946 4870 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 09:49:23 crc kubenswrapper[4870]: I1014 09:49:23.951731 4870 patch_prober.go:28] interesting pod/machine-config-daemon-7tvc7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 09:49:23 crc kubenswrapper[4870]: I1014 09:49:23.952509 4870 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 09:49:37 crc kubenswrapper[4870]: I1014 09:49:37.521255 4870 generic.go:334] "Generic (PLEG): container finished" podID="1358f0f9-0ff8-415f-9861-359ba368c8dd" containerID="4626063e8ca122ab584e37783a27fd4546253d894e17055595a5404d18ce4acd" exitCode=0 Oct 14 09:49:37 crc kubenswrapper[4870]: I1014 09:49:37.521490 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-openstack-openstack-cell1-q2bvq" event={"ID":"1358f0f9-0ff8-415f-9861-359ba368c8dd","Type":"ContainerDied","Data":"4626063e8ca122ab584e37783a27fd4546253d894e17055595a5404d18ce4acd"} Oct 14 09:49:39 crc kubenswrapper[4870]: I1014 09:49:39.124411 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-openstack-openstack-cell1-q2bvq" Oct 14 09:49:39 crc kubenswrapper[4870]: I1014 09:49:39.212969 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/1358f0f9-0ff8-415f-9861-359ba368c8dd-ceilometer-compute-config-data-1\") pod \"1358f0f9-0ff8-415f-9861-359ba368c8dd\" (UID: \"1358f0f9-0ff8-415f-9861-359ba368c8dd\") " Oct 14 09:49:39 crc kubenswrapper[4870]: I1014 09:49:39.213013 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/1358f0f9-0ff8-415f-9861-359ba368c8dd-ceilometer-compute-config-data-2\") pod \"1358f0f9-0ff8-415f-9861-359ba368c8dd\" (UID: \"1358f0f9-0ff8-415f-9861-359ba368c8dd\") " Oct 14 09:49:39 crc kubenswrapper[4870]: I1014 09:49:39.213059 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7nkrn\" (UniqueName: \"kubernetes.io/projected/1358f0f9-0ff8-415f-9861-359ba368c8dd-kube-api-access-7nkrn\") pod \"1358f0f9-0ff8-415f-9861-359ba368c8dd\" (UID: \"1358f0f9-0ff8-415f-9861-359ba368c8dd\") " Oct 14 09:49:39 crc kubenswrapper[4870]: I1014 09:49:39.213097 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/1358f0f9-0ff8-415f-9861-359ba368c8dd-ceph\") pod \"1358f0f9-0ff8-415f-9861-359ba368c8dd\" (UID: \"1358f0f9-0ff8-415f-9861-359ba368c8dd\") " Oct 14 09:49:39 crc kubenswrapper[4870]: I1014 09:49:39.213151 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/1358f0f9-0ff8-415f-9861-359ba368c8dd-ceilometer-compute-config-data-0\") pod \"1358f0f9-0ff8-415f-9861-359ba368c8dd\" (UID: \"1358f0f9-0ff8-415f-9861-359ba368c8dd\") " Oct 14 09:49:39 crc kubenswrapper[4870]: I1014 09:49:39.213196 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1358f0f9-0ff8-415f-9861-359ba368c8dd-telemetry-combined-ca-bundle\") pod \"1358f0f9-0ff8-415f-9861-359ba368c8dd\" (UID: \"1358f0f9-0ff8-415f-9861-359ba368c8dd\") " Oct 14 09:49:39 crc kubenswrapper[4870]: I1014 09:49:39.213215 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1358f0f9-0ff8-415f-9861-359ba368c8dd-ssh-key\") pod \"1358f0f9-0ff8-415f-9861-359ba368c8dd\" (UID: \"1358f0f9-0ff8-415f-9861-359ba368c8dd\") " Oct 14 09:49:39 crc kubenswrapper[4870]: I1014 09:49:39.213264 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1358f0f9-0ff8-415f-9861-359ba368c8dd-inventory\") pod \"1358f0f9-0ff8-415f-9861-359ba368c8dd\" (UID: \"1358f0f9-0ff8-415f-9861-359ba368c8dd\") " Oct 14 09:49:39 crc kubenswrapper[4870]: I1014 09:49:39.221747 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1358f0f9-0ff8-415f-9861-359ba368c8dd-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "1358f0f9-0ff8-415f-9861-359ba368c8dd" (UID: "1358f0f9-0ff8-415f-9861-359ba368c8dd"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:49:39 crc kubenswrapper[4870]: I1014 09:49:39.224724 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1358f0f9-0ff8-415f-9861-359ba368c8dd-kube-api-access-7nkrn" (OuterVolumeSpecName: "kube-api-access-7nkrn") pod "1358f0f9-0ff8-415f-9861-359ba368c8dd" (UID: "1358f0f9-0ff8-415f-9861-359ba368c8dd"). InnerVolumeSpecName "kube-api-access-7nkrn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:49:39 crc kubenswrapper[4870]: I1014 09:49:39.226120 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1358f0f9-0ff8-415f-9861-359ba368c8dd-ceph" (OuterVolumeSpecName: "ceph") pod "1358f0f9-0ff8-415f-9861-359ba368c8dd" (UID: "1358f0f9-0ff8-415f-9861-359ba368c8dd"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:49:39 crc kubenswrapper[4870]: I1014 09:49:39.244606 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1358f0f9-0ff8-415f-9861-359ba368c8dd-ceilometer-compute-config-data-0" (OuterVolumeSpecName: "ceilometer-compute-config-data-0") pod "1358f0f9-0ff8-415f-9861-359ba368c8dd" (UID: "1358f0f9-0ff8-415f-9861-359ba368c8dd"). InnerVolumeSpecName "ceilometer-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:49:39 crc kubenswrapper[4870]: I1014 09:49:39.248839 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1358f0f9-0ff8-415f-9861-359ba368c8dd-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "1358f0f9-0ff8-415f-9861-359ba368c8dd" (UID: "1358f0f9-0ff8-415f-9861-359ba368c8dd"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:49:39 crc kubenswrapper[4870]: I1014 09:49:39.251070 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1358f0f9-0ff8-415f-9861-359ba368c8dd-ceilometer-compute-config-data-2" (OuterVolumeSpecName: "ceilometer-compute-config-data-2") pod "1358f0f9-0ff8-415f-9861-359ba368c8dd" (UID: "1358f0f9-0ff8-415f-9861-359ba368c8dd"). InnerVolumeSpecName "ceilometer-compute-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:49:39 crc kubenswrapper[4870]: I1014 09:49:39.258181 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1358f0f9-0ff8-415f-9861-359ba368c8dd-ceilometer-compute-config-data-1" (OuterVolumeSpecName: "ceilometer-compute-config-data-1") pod "1358f0f9-0ff8-415f-9861-359ba368c8dd" (UID: "1358f0f9-0ff8-415f-9861-359ba368c8dd"). InnerVolumeSpecName "ceilometer-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:49:39 crc kubenswrapper[4870]: I1014 09:49:39.269635 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1358f0f9-0ff8-415f-9861-359ba368c8dd-inventory" (OuterVolumeSpecName: "inventory") pod "1358f0f9-0ff8-415f-9861-359ba368c8dd" (UID: "1358f0f9-0ff8-415f-9861-359ba368c8dd"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:49:39 crc kubenswrapper[4870]: I1014 09:49:39.315583 4870 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/1358f0f9-0ff8-415f-9861-359ba368c8dd-ceilometer-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Oct 14 09:49:39 crc kubenswrapper[4870]: I1014 09:49:39.315622 4870 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/1358f0f9-0ff8-415f-9861-359ba368c8dd-ceilometer-compute-config-data-2\") on node \"crc\" DevicePath \"\"" Oct 14 09:49:39 crc kubenswrapper[4870]: I1014 09:49:39.315634 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7nkrn\" (UniqueName: \"kubernetes.io/projected/1358f0f9-0ff8-415f-9861-359ba368c8dd-kube-api-access-7nkrn\") on node \"crc\" DevicePath \"\"" Oct 14 09:49:39 crc kubenswrapper[4870]: I1014 09:49:39.315646 4870 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/1358f0f9-0ff8-415f-9861-359ba368c8dd-ceph\") on node \"crc\" DevicePath \"\"" Oct 14 09:49:39 crc kubenswrapper[4870]: I1014 09:49:39.315657 4870 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/1358f0f9-0ff8-415f-9861-359ba368c8dd-ceilometer-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Oct 14 09:49:39 crc kubenswrapper[4870]: I1014 09:49:39.315666 4870 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1358f0f9-0ff8-415f-9861-359ba368c8dd-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 09:49:39 crc kubenswrapper[4870]: I1014 09:49:39.315676 4870 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1358f0f9-0ff8-415f-9861-359ba368c8dd-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 14 09:49:39 crc kubenswrapper[4870]: I1014 09:49:39.315686 4870 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1358f0f9-0ff8-415f-9861-359ba368c8dd-inventory\") on node \"crc\" DevicePath \"\"" Oct 14 09:49:39 crc kubenswrapper[4870]: I1014 09:49:39.547283 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-openstack-openstack-cell1-q2bvq" event={"ID":"1358f0f9-0ff8-415f-9861-359ba368c8dd","Type":"ContainerDied","Data":"f47173a6a1ff9d180b133ec1d2ed3f160abb7e522937a39b0b7715734e20f8a1"} Oct 14 09:49:39 crc kubenswrapper[4870]: I1014 09:49:39.547430 4870 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f47173a6a1ff9d180b133ec1d2ed3f160abb7e522937a39b0b7715734e20f8a1" Oct 14 09:49:39 crc kubenswrapper[4870]: I1014 09:49:39.547472 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-openstack-openstack-cell1-q2bvq" Oct 14 09:49:39 crc kubenswrapper[4870]: I1014 09:49:39.727235 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-sriov-openstack-openstack-cell1-7zwk7"] Oct 14 09:49:39 crc kubenswrapper[4870]: E1014 09:49:39.728203 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1358f0f9-0ff8-415f-9861-359ba368c8dd" containerName="telemetry-openstack-openstack-cell1" Oct 14 09:49:39 crc kubenswrapper[4870]: I1014 09:49:39.728229 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="1358f0f9-0ff8-415f-9861-359ba368c8dd" containerName="telemetry-openstack-openstack-cell1" Oct 14 09:49:39 crc kubenswrapper[4870]: E1014 09:49:39.728277 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ffb284a-79dd-4b65-b824-c1c0211d4a0a" containerName="collect-profiles" Oct 14 09:49:39 crc kubenswrapper[4870]: I1014 09:49:39.728290 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ffb284a-79dd-4b65-b824-c1c0211d4a0a" containerName="collect-profiles" Oct 14 09:49:39 crc kubenswrapper[4870]: I1014 09:49:39.728728 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="0ffb284a-79dd-4b65-b824-c1c0211d4a0a" containerName="collect-profiles" Oct 14 09:49:39 crc kubenswrapper[4870]: I1014 09:49:39.728787 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="1358f0f9-0ff8-415f-9861-359ba368c8dd" containerName="telemetry-openstack-openstack-cell1" Oct 14 09:49:39 crc kubenswrapper[4870]: I1014 09:49:39.729976 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-sriov-openstack-openstack-cell1-7zwk7" Oct 14 09:49:39 crc kubenswrapper[4870]: I1014 09:49:39.733004 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-sriov-agent-neutron-config" Oct 14 09:49:39 crc kubenswrapper[4870]: I1014 09:49:39.733212 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-wrz7w" Oct 14 09:49:39 crc kubenswrapper[4870]: I1014 09:49:39.733303 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 14 09:49:39 crc kubenswrapper[4870]: I1014 09:49:39.733472 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 14 09:49:39 crc kubenswrapper[4870]: I1014 09:49:39.733603 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 14 09:49:39 crc kubenswrapper[4870]: I1014 09:49:39.736273 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-sriov-openstack-openstack-cell1-7zwk7"] Oct 14 09:49:39 crc kubenswrapper[4870]: I1014 09:49:39.927422 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/59c26d47-f3a1-4103-a00d-5ee15b387c4a-ssh-key\") pod \"neutron-sriov-openstack-openstack-cell1-7zwk7\" (UID: \"59c26d47-f3a1-4103-a00d-5ee15b387c4a\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-7zwk7" Oct 14 09:49:39 crc kubenswrapper[4870]: I1014 09:49:39.927576 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/59c26d47-f3a1-4103-a00d-5ee15b387c4a-inventory\") pod \"neutron-sriov-openstack-openstack-cell1-7zwk7\" (UID: \"59c26d47-f3a1-4103-a00d-5ee15b387c4a\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-7zwk7" Oct 14 09:49:39 crc kubenswrapper[4870]: I1014 09:49:39.927702 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/59c26d47-f3a1-4103-a00d-5ee15b387c4a-neutron-sriov-combined-ca-bundle\") pod \"neutron-sriov-openstack-openstack-cell1-7zwk7\" (UID: \"59c26d47-f3a1-4103-a00d-5ee15b387c4a\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-7zwk7" Oct 14 09:49:39 crc kubenswrapper[4870]: I1014 09:49:39.927736 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q8skt\" (UniqueName: \"kubernetes.io/projected/59c26d47-f3a1-4103-a00d-5ee15b387c4a-kube-api-access-q8skt\") pod \"neutron-sriov-openstack-openstack-cell1-7zwk7\" (UID: \"59c26d47-f3a1-4103-a00d-5ee15b387c4a\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-7zwk7" Oct 14 09:49:39 crc kubenswrapper[4870]: I1014 09:49:39.927777 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/59c26d47-f3a1-4103-a00d-5ee15b387c4a-ceph\") pod \"neutron-sriov-openstack-openstack-cell1-7zwk7\" (UID: \"59c26d47-f3a1-4103-a00d-5ee15b387c4a\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-7zwk7" Oct 14 09:49:39 crc kubenswrapper[4870]: I1014 09:49:39.927849 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-sriov-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/59c26d47-f3a1-4103-a00d-5ee15b387c4a-neutron-sriov-agent-neutron-config-0\") pod \"neutron-sriov-openstack-openstack-cell1-7zwk7\" (UID: \"59c26d47-f3a1-4103-a00d-5ee15b387c4a\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-7zwk7" Oct 14 09:49:40 crc kubenswrapper[4870]: I1014 09:49:40.029203 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/59c26d47-f3a1-4103-a00d-5ee15b387c4a-inventory\") pod \"neutron-sriov-openstack-openstack-cell1-7zwk7\" (UID: \"59c26d47-f3a1-4103-a00d-5ee15b387c4a\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-7zwk7" Oct 14 09:49:40 crc kubenswrapper[4870]: I1014 09:49:40.029341 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/59c26d47-f3a1-4103-a00d-5ee15b387c4a-neutron-sriov-combined-ca-bundle\") pod \"neutron-sriov-openstack-openstack-cell1-7zwk7\" (UID: \"59c26d47-f3a1-4103-a00d-5ee15b387c4a\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-7zwk7" Oct 14 09:49:40 crc kubenswrapper[4870]: I1014 09:49:40.029378 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q8skt\" (UniqueName: \"kubernetes.io/projected/59c26d47-f3a1-4103-a00d-5ee15b387c4a-kube-api-access-q8skt\") pod \"neutron-sriov-openstack-openstack-cell1-7zwk7\" (UID: \"59c26d47-f3a1-4103-a00d-5ee15b387c4a\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-7zwk7" Oct 14 09:49:40 crc kubenswrapper[4870]: I1014 09:49:40.029416 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/59c26d47-f3a1-4103-a00d-5ee15b387c4a-ceph\") pod \"neutron-sriov-openstack-openstack-cell1-7zwk7\" (UID: \"59c26d47-f3a1-4103-a00d-5ee15b387c4a\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-7zwk7" Oct 14 09:49:40 crc kubenswrapper[4870]: I1014 09:49:40.029491 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-sriov-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/59c26d47-f3a1-4103-a00d-5ee15b387c4a-neutron-sriov-agent-neutron-config-0\") pod \"neutron-sriov-openstack-openstack-cell1-7zwk7\" (UID: \"59c26d47-f3a1-4103-a00d-5ee15b387c4a\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-7zwk7" Oct 14 09:49:40 crc kubenswrapper[4870]: I1014 09:49:40.029538 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/59c26d47-f3a1-4103-a00d-5ee15b387c4a-ssh-key\") pod \"neutron-sriov-openstack-openstack-cell1-7zwk7\" (UID: \"59c26d47-f3a1-4103-a00d-5ee15b387c4a\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-7zwk7" Oct 14 09:49:40 crc kubenswrapper[4870]: I1014 09:49:40.037007 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/59c26d47-f3a1-4103-a00d-5ee15b387c4a-neutron-sriov-combined-ca-bundle\") pod \"neutron-sriov-openstack-openstack-cell1-7zwk7\" (UID: \"59c26d47-f3a1-4103-a00d-5ee15b387c4a\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-7zwk7" Oct 14 09:49:40 crc kubenswrapper[4870]: I1014 09:49:40.037108 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/59c26d47-f3a1-4103-a00d-5ee15b387c4a-ceph\") pod \"neutron-sriov-openstack-openstack-cell1-7zwk7\" (UID: \"59c26d47-f3a1-4103-a00d-5ee15b387c4a\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-7zwk7" Oct 14 09:49:40 crc kubenswrapper[4870]: I1014 09:49:40.037656 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/59c26d47-f3a1-4103-a00d-5ee15b387c4a-inventory\") pod \"neutron-sriov-openstack-openstack-cell1-7zwk7\" (UID: \"59c26d47-f3a1-4103-a00d-5ee15b387c4a\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-7zwk7" Oct 14 09:49:40 crc kubenswrapper[4870]: I1014 09:49:40.038980 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-sriov-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/59c26d47-f3a1-4103-a00d-5ee15b387c4a-neutron-sriov-agent-neutron-config-0\") pod \"neutron-sriov-openstack-openstack-cell1-7zwk7\" (UID: \"59c26d47-f3a1-4103-a00d-5ee15b387c4a\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-7zwk7" Oct 14 09:49:40 crc kubenswrapper[4870]: I1014 09:49:40.040979 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/59c26d47-f3a1-4103-a00d-5ee15b387c4a-ssh-key\") pod \"neutron-sriov-openstack-openstack-cell1-7zwk7\" (UID: \"59c26d47-f3a1-4103-a00d-5ee15b387c4a\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-7zwk7" Oct 14 09:49:40 crc kubenswrapper[4870]: I1014 09:49:40.054303 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q8skt\" (UniqueName: \"kubernetes.io/projected/59c26d47-f3a1-4103-a00d-5ee15b387c4a-kube-api-access-q8skt\") pod \"neutron-sriov-openstack-openstack-cell1-7zwk7\" (UID: \"59c26d47-f3a1-4103-a00d-5ee15b387c4a\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-7zwk7" Oct 14 09:49:40 crc kubenswrapper[4870]: I1014 09:49:40.057086 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-sriov-openstack-openstack-cell1-7zwk7" Oct 14 09:49:40 crc kubenswrapper[4870]: I1014 09:49:40.652004 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-sriov-openstack-openstack-cell1-7zwk7"] Oct 14 09:49:41 crc kubenswrapper[4870]: W1014 09:49:41.159648 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod59c26d47_f3a1_4103_a00d_5ee15b387c4a.slice/crio-fa6c0fd057686830232b71f9bab869abfb61f61e5c8965f9bba7dba2d5b54f4a WatchSource:0}: Error finding container fa6c0fd057686830232b71f9bab869abfb61f61e5c8965f9bba7dba2d5b54f4a: Status 404 returned error can't find the container with id fa6c0fd057686830232b71f9bab869abfb61f61e5c8965f9bba7dba2d5b54f4a Oct 14 09:49:41 crc kubenswrapper[4870]: I1014 09:49:41.163919 4870 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 14 09:49:41 crc kubenswrapper[4870]: I1014 09:49:41.590205 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-sriov-openstack-openstack-cell1-7zwk7" event={"ID":"59c26d47-f3a1-4103-a00d-5ee15b387c4a","Type":"ContainerStarted","Data":"fa6c0fd057686830232b71f9bab869abfb61f61e5c8965f9bba7dba2d5b54f4a"} Oct 14 09:49:42 crc kubenswrapper[4870]: I1014 09:49:42.608755 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-sriov-openstack-openstack-cell1-7zwk7" event={"ID":"59c26d47-f3a1-4103-a00d-5ee15b387c4a","Type":"ContainerStarted","Data":"352511aa4e9ed04b460e136058a6ea68dd0aa6675e879931506cbd3d6e534f76"} Oct 14 09:49:42 crc kubenswrapper[4870]: I1014 09:49:42.651265 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-sriov-openstack-openstack-cell1-7zwk7" podStartSLOduration=3.112357018 podStartE2EDuration="3.651237565s" podCreationTimestamp="2025-10-14 09:49:39 +0000 UTC" firstStartedPulling="2025-10-14 09:49:41.163644922 +0000 UTC m=+10116.861005303" lastFinishedPulling="2025-10-14 09:49:41.702525439 +0000 UTC m=+10117.399885850" observedRunningTime="2025-10-14 09:49:42.638113553 +0000 UTC m=+10118.335473954" watchObservedRunningTime="2025-10-14 09:49:42.651237565 +0000 UTC m=+10118.348597966" Oct 14 09:49:53 crc kubenswrapper[4870]: I1014 09:49:53.951308 4870 patch_prober.go:28] interesting pod/machine-config-daemon-7tvc7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 09:49:53 crc kubenswrapper[4870]: I1014 09:49:53.952294 4870 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 09:49:53 crc kubenswrapper[4870]: I1014 09:49:53.952379 4870 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" Oct 14 09:49:53 crc kubenswrapper[4870]: I1014 09:49:53.954032 4870 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"4d8d8702edb8183ff39b29852bc295d1ac61316a22904ed8a1cfe721472e2347"} pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 14 09:49:53 crc kubenswrapper[4870]: I1014 09:49:53.954167 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerName="machine-config-daemon" containerID="cri-o://4d8d8702edb8183ff39b29852bc295d1ac61316a22904ed8a1cfe721472e2347" gracePeriod=600 Oct 14 09:49:54 crc kubenswrapper[4870]: I1014 09:49:54.786525 4870 generic.go:334] "Generic (PLEG): container finished" podID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerID="4d8d8702edb8183ff39b29852bc295d1ac61316a22904ed8a1cfe721472e2347" exitCode=0 Oct 14 09:49:54 crc kubenswrapper[4870]: I1014 09:49:54.786632 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" event={"ID":"874a5e45-dffd-4d17-b609-b7d3ed2eab07","Type":"ContainerDied","Data":"4d8d8702edb8183ff39b29852bc295d1ac61316a22904ed8a1cfe721472e2347"} Oct 14 09:49:54 crc kubenswrapper[4870]: I1014 09:49:54.787262 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" event={"ID":"874a5e45-dffd-4d17-b609-b7d3ed2eab07","Type":"ContainerStarted","Data":"c52e0ba5edaea8b07f60b9f76084bc5c51b52575d871193fc95ac2cefd996fea"} Oct 14 09:49:54 crc kubenswrapper[4870]: I1014 09:49:54.787298 4870 scope.go:117] "RemoveContainer" containerID="6d5792c105b7bc7694bce51cbf8b085eda0318f921dffcecf5167175f90aa527" Oct 14 09:49:56 crc kubenswrapper[4870]: I1014 09:49:56.870750 4870 trace.go:236] Trace[1437816399]: "Calculate volume metrics of ovndbcluster-nb-etc-ovn for pod openstack/ovsdbserver-nb-0" (14-Oct-2025 09:49:55.870) (total time: 1000ms): Oct 14 09:49:56 crc kubenswrapper[4870]: Trace[1437816399]: [1.000669602s] [1.000669602s] END Oct 14 09:50:07 crc kubenswrapper[4870]: I1014 09:50:07.861861 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-w2wh9"] Oct 14 09:50:07 crc kubenswrapper[4870]: I1014 09:50:07.868076 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-w2wh9" Oct 14 09:50:07 crc kubenswrapper[4870]: I1014 09:50:07.879769 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-w2wh9"] Oct 14 09:50:07 crc kubenswrapper[4870]: I1014 09:50:07.970635 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2e5f7f0c-129a-4b34-9276-82dacd66ca3d-catalog-content\") pod \"redhat-marketplace-w2wh9\" (UID: \"2e5f7f0c-129a-4b34-9276-82dacd66ca3d\") " pod="openshift-marketplace/redhat-marketplace-w2wh9" Oct 14 09:50:07 crc kubenswrapper[4870]: I1014 09:50:07.970858 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zxk2q\" (UniqueName: \"kubernetes.io/projected/2e5f7f0c-129a-4b34-9276-82dacd66ca3d-kube-api-access-zxk2q\") pod \"redhat-marketplace-w2wh9\" (UID: \"2e5f7f0c-129a-4b34-9276-82dacd66ca3d\") " pod="openshift-marketplace/redhat-marketplace-w2wh9" Oct 14 09:50:07 crc kubenswrapper[4870]: I1014 09:50:07.970892 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2e5f7f0c-129a-4b34-9276-82dacd66ca3d-utilities\") pod \"redhat-marketplace-w2wh9\" (UID: \"2e5f7f0c-129a-4b34-9276-82dacd66ca3d\") " pod="openshift-marketplace/redhat-marketplace-w2wh9" Oct 14 09:50:08 crc kubenswrapper[4870]: I1014 09:50:08.073530 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zxk2q\" (UniqueName: \"kubernetes.io/projected/2e5f7f0c-129a-4b34-9276-82dacd66ca3d-kube-api-access-zxk2q\") pod \"redhat-marketplace-w2wh9\" (UID: \"2e5f7f0c-129a-4b34-9276-82dacd66ca3d\") " pod="openshift-marketplace/redhat-marketplace-w2wh9" Oct 14 09:50:08 crc kubenswrapper[4870]: I1014 09:50:08.073584 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2e5f7f0c-129a-4b34-9276-82dacd66ca3d-utilities\") pod \"redhat-marketplace-w2wh9\" (UID: \"2e5f7f0c-129a-4b34-9276-82dacd66ca3d\") " pod="openshift-marketplace/redhat-marketplace-w2wh9" Oct 14 09:50:08 crc kubenswrapper[4870]: I1014 09:50:08.073669 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2e5f7f0c-129a-4b34-9276-82dacd66ca3d-catalog-content\") pod \"redhat-marketplace-w2wh9\" (UID: \"2e5f7f0c-129a-4b34-9276-82dacd66ca3d\") " pod="openshift-marketplace/redhat-marketplace-w2wh9" Oct 14 09:50:08 crc kubenswrapper[4870]: I1014 09:50:08.074140 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2e5f7f0c-129a-4b34-9276-82dacd66ca3d-catalog-content\") pod \"redhat-marketplace-w2wh9\" (UID: \"2e5f7f0c-129a-4b34-9276-82dacd66ca3d\") " pod="openshift-marketplace/redhat-marketplace-w2wh9" Oct 14 09:50:08 crc kubenswrapper[4870]: I1014 09:50:08.074139 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2e5f7f0c-129a-4b34-9276-82dacd66ca3d-utilities\") pod \"redhat-marketplace-w2wh9\" (UID: \"2e5f7f0c-129a-4b34-9276-82dacd66ca3d\") " pod="openshift-marketplace/redhat-marketplace-w2wh9" Oct 14 09:50:08 crc kubenswrapper[4870]: I1014 09:50:08.095276 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zxk2q\" (UniqueName: \"kubernetes.io/projected/2e5f7f0c-129a-4b34-9276-82dacd66ca3d-kube-api-access-zxk2q\") pod \"redhat-marketplace-w2wh9\" (UID: \"2e5f7f0c-129a-4b34-9276-82dacd66ca3d\") " pod="openshift-marketplace/redhat-marketplace-w2wh9" Oct 14 09:50:08 crc kubenswrapper[4870]: I1014 09:50:08.193244 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-w2wh9" Oct 14 09:50:08 crc kubenswrapper[4870]: I1014 09:50:08.709161 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-w2wh9"] Oct 14 09:50:08 crc kubenswrapper[4870]: I1014 09:50:08.992149 4870 generic.go:334] "Generic (PLEG): container finished" podID="2e5f7f0c-129a-4b34-9276-82dacd66ca3d" containerID="9c0ba9e501df0399b31b3a222ed3f8030c15f9a9070c49bae2794bfe7f1bd259" exitCode=0 Oct 14 09:50:08 crc kubenswrapper[4870]: I1014 09:50:08.992267 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-w2wh9" event={"ID":"2e5f7f0c-129a-4b34-9276-82dacd66ca3d","Type":"ContainerDied","Data":"9c0ba9e501df0399b31b3a222ed3f8030c15f9a9070c49bae2794bfe7f1bd259"} Oct 14 09:50:08 crc kubenswrapper[4870]: I1014 09:50:08.992570 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-w2wh9" event={"ID":"2e5f7f0c-129a-4b34-9276-82dacd66ca3d","Type":"ContainerStarted","Data":"57a092ac65b815bb4b21fcb66001792e7fe0c500066f3edad46760a8bec6c1d4"} Oct 14 09:50:10 crc kubenswrapper[4870]: I1014 09:50:10.002715 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-w2wh9" event={"ID":"2e5f7f0c-129a-4b34-9276-82dacd66ca3d","Type":"ContainerStarted","Data":"7539542b0f5bb3d2928fa3eee2ce422d821037144ba4673b8dcd37b2c7b26eab"} Oct 14 09:50:11 crc kubenswrapper[4870]: I1014 09:50:11.015995 4870 generic.go:334] "Generic (PLEG): container finished" podID="2e5f7f0c-129a-4b34-9276-82dacd66ca3d" containerID="7539542b0f5bb3d2928fa3eee2ce422d821037144ba4673b8dcd37b2c7b26eab" exitCode=0 Oct 14 09:50:11 crc kubenswrapper[4870]: I1014 09:50:11.016051 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-w2wh9" event={"ID":"2e5f7f0c-129a-4b34-9276-82dacd66ca3d","Type":"ContainerDied","Data":"7539542b0f5bb3d2928fa3eee2ce422d821037144ba4673b8dcd37b2c7b26eab"} Oct 14 09:50:12 crc kubenswrapper[4870]: I1014 09:50:12.068567 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-w2wh9" event={"ID":"2e5f7f0c-129a-4b34-9276-82dacd66ca3d","Type":"ContainerStarted","Data":"b5fac135a718bafd9ffa10566c987e3a972fdc13d86ce6b026b7245f8d34f495"} Oct 14 09:50:12 crc kubenswrapper[4870]: I1014 09:50:12.100342 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-w2wh9" podStartSLOduration=2.692019825 podStartE2EDuration="5.100312447s" podCreationTimestamp="2025-10-14 09:50:07 +0000 UTC" firstStartedPulling="2025-10-14 09:50:08.995638651 +0000 UTC m=+10144.692999022" lastFinishedPulling="2025-10-14 09:50:11.403931283 +0000 UTC m=+10147.101291644" observedRunningTime="2025-10-14 09:50:12.087303107 +0000 UTC m=+10147.784663498" watchObservedRunningTime="2025-10-14 09:50:12.100312447 +0000 UTC m=+10147.797672858" Oct 14 09:50:18 crc kubenswrapper[4870]: I1014 09:50:18.194147 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-w2wh9" Oct 14 09:50:18 crc kubenswrapper[4870]: I1014 09:50:18.194979 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-w2wh9" Oct 14 09:50:18 crc kubenswrapper[4870]: I1014 09:50:18.261805 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-w2wh9" Oct 14 09:50:19 crc kubenswrapper[4870]: I1014 09:50:19.200103 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-w2wh9" Oct 14 09:50:19 crc kubenswrapper[4870]: I1014 09:50:19.253629 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-w2wh9"] Oct 14 09:50:21 crc kubenswrapper[4870]: I1014 09:50:21.171461 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-w2wh9" podUID="2e5f7f0c-129a-4b34-9276-82dacd66ca3d" containerName="registry-server" containerID="cri-o://b5fac135a718bafd9ffa10566c987e3a972fdc13d86ce6b026b7245f8d34f495" gracePeriod=2 Oct 14 09:50:21 crc kubenswrapper[4870]: I1014 09:50:21.680651 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-w2wh9" Oct 14 09:50:21 crc kubenswrapper[4870]: I1014 09:50:21.797250 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zxk2q\" (UniqueName: \"kubernetes.io/projected/2e5f7f0c-129a-4b34-9276-82dacd66ca3d-kube-api-access-zxk2q\") pod \"2e5f7f0c-129a-4b34-9276-82dacd66ca3d\" (UID: \"2e5f7f0c-129a-4b34-9276-82dacd66ca3d\") " Oct 14 09:50:21 crc kubenswrapper[4870]: I1014 09:50:21.797367 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2e5f7f0c-129a-4b34-9276-82dacd66ca3d-catalog-content\") pod \"2e5f7f0c-129a-4b34-9276-82dacd66ca3d\" (UID: \"2e5f7f0c-129a-4b34-9276-82dacd66ca3d\") " Oct 14 09:50:21 crc kubenswrapper[4870]: I1014 09:50:21.797473 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2e5f7f0c-129a-4b34-9276-82dacd66ca3d-utilities\") pod \"2e5f7f0c-129a-4b34-9276-82dacd66ca3d\" (UID: \"2e5f7f0c-129a-4b34-9276-82dacd66ca3d\") " Oct 14 09:50:21 crc kubenswrapper[4870]: I1014 09:50:21.799027 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2e5f7f0c-129a-4b34-9276-82dacd66ca3d-utilities" (OuterVolumeSpecName: "utilities") pod "2e5f7f0c-129a-4b34-9276-82dacd66ca3d" (UID: "2e5f7f0c-129a-4b34-9276-82dacd66ca3d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 09:50:21 crc kubenswrapper[4870]: I1014 09:50:21.803723 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2e5f7f0c-129a-4b34-9276-82dacd66ca3d-kube-api-access-zxk2q" (OuterVolumeSpecName: "kube-api-access-zxk2q") pod "2e5f7f0c-129a-4b34-9276-82dacd66ca3d" (UID: "2e5f7f0c-129a-4b34-9276-82dacd66ca3d"). InnerVolumeSpecName "kube-api-access-zxk2q". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:50:21 crc kubenswrapper[4870]: I1014 09:50:21.812894 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2e5f7f0c-129a-4b34-9276-82dacd66ca3d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2e5f7f0c-129a-4b34-9276-82dacd66ca3d" (UID: "2e5f7f0c-129a-4b34-9276-82dacd66ca3d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 09:50:21 crc kubenswrapper[4870]: I1014 09:50:21.901039 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zxk2q\" (UniqueName: \"kubernetes.io/projected/2e5f7f0c-129a-4b34-9276-82dacd66ca3d-kube-api-access-zxk2q\") on node \"crc\" DevicePath \"\"" Oct 14 09:50:21 crc kubenswrapper[4870]: I1014 09:50:21.901112 4870 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2e5f7f0c-129a-4b34-9276-82dacd66ca3d-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 09:50:21 crc kubenswrapper[4870]: I1014 09:50:21.901130 4870 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2e5f7f0c-129a-4b34-9276-82dacd66ca3d-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 09:50:22 crc kubenswrapper[4870]: I1014 09:50:22.187128 4870 generic.go:334] "Generic (PLEG): container finished" podID="2e5f7f0c-129a-4b34-9276-82dacd66ca3d" containerID="b5fac135a718bafd9ffa10566c987e3a972fdc13d86ce6b026b7245f8d34f495" exitCode=0 Oct 14 09:50:22 crc kubenswrapper[4870]: I1014 09:50:22.187178 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-w2wh9" event={"ID":"2e5f7f0c-129a-4b34-9276-82dacd66ca3d","Type":"ContainerDied","Data":"b5fac135a718bafd9ffa10566c987e3a972fdc13d86ce6b026b7245f8d34f495"} Oct 14 09:50:22 crc kubenswrapper[4870]: I1014 09:50:22.187216 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-w2wh9" event={"ID":"2e5f7f0c-129a-4b34-9276-82dacd66ca3d","Type":"ContainerDied","Data":"57a092ac65b815bb4b21fcb66001792e7fe0c500066f3edad46760a8bec6c1d4"} Oct 14 09:50:22 crc kubenswrapper[4870]: I1014 09:50:22.187236 4870 scope.go:117] "RemoveContainer" containerID="b5fac135a718bafd9ffa10566c987e3a972fdc13d86ce6b026b7245f8d34f495" Oct 14 09:50:22 crc kubenswrapper[4870]: I1014 09:50:22.187243 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-w2wh9" Oct 14 09:50:22 crc kubenswrapper[4870]: I1014 09:50:22.217007 4870 scope.go:117] "RemoveContainer" containerID="7539542b0f5bb3d2928fa3eee2ce422d821037144ba4673b8dcd37b2c7b26eab" Oct 14 09:50:22 crc kubenswrapper[4870]: I1014 09:50:22.252068 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-w2wh9"] Oct 14 09:50:22 crc kubenswrapper[4870]: I1014 09:50:22.260324 4870 scope.go:117] "RemoveContainer" containerID="9c0ba9e501df0399b31b3a222ed3f8030c15f9a9070c49bae2794bfe7f1bd259" Oct 14 09:50:22 crc kubenswrapper[4870]: I1014 09:50:22.266838 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-w2wh9"] Oct 14 09:50:22 crc kubenswrapper[4870]: I1014 09:50:22.314553 4870 scope.go:117] "RemoveContainer" containerID="b5fac135a718bafd9ffa10566c987e3a972fdc13d86ce6b026b7245f8d34f495" Oct 14 09:50:22 crc kubenswrapper[4870]: E1014 09:50:22.314968 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b5fac135a718bafd9ffa10566c987e3a972fdc13d86ce6b026b7245f8d34f495\": container with ID starting with b5fac135a718bafd9ffa10566c987e3a972fdc13d86ce6b026b7245f8d34f495 not found: ID does not exist" containerID="b5fac135a718bafd9ffa10566c987e3a972fdc13d86ce6b026b7245f8d34f495" Oct 14 09:50:22 crc kubenswrapper[4870]: I1014 09:50:22.315000 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b5fac135a718bafd9ffa10566c987e3a972fdc13d86ce6b026b7245f8d34f495"} err="failed to get container status \"b5fac135a718bafd9ffa10566c987e3a972fdc13d86ce6b026b7245f8d34f495\": rpc error: code = NotFound desc = could not find container \"b5fac135a718bafd9ffa10566c987e3a972fdc13d86ce6b026b7245f8d34f495\": container with ID starting with b5fac135a718bafd9ffa10566c987e3a972fdc13d86ce6b026b7245f8d34f495 not found: ID does not exist" Oct 14 09:50:22 crc kubenswrapper[4870]: I1014 09:50:22.315024 4870 scope.go:117] "RemoveContainer" containerID="7539542b0f5bb3d2928fa3eee2ce422d821037144ba4673b8dcd37b2c7b26eab" Oct 14 09:50:22 crc kubenswrapper[4870]: E1014 09:50:22.315353 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7539542b0f5bb3d2928fa3eee2ce422d821037144ba4673b8dcd37b2c7b26eab\": container with ID starting with 7539542b0f5bb3d2928fa3eee2ce422d821037144ba4673b8dcd37b2c7b26eab not found: ID does not exist" containerID="7539542b0f5bb3d2928fa3eee2ce422d821037144ba4673b8dcd37b2c7b26eab" Oct 14 09:50:22 crc kubenswrapper[4870]: I1014 09:50:22.315433 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7539542b0f5bb3d2928fa3eee2ce422d821037144ba4673b8dcd37b2c7b26eab"} err="failed to get container status \"7539542b0f5bb3d2928fa3eee2ce422d821037144ba4673b8dcd37b2c7b26eab\": rpc error: code = NotFound desc = could not find container \"7539542b0f5bb3d2928fa3eee2ce422d821037144ba4673b8dcd37b2c7b26eab\": container with ID starting with 7539542b0f5bb3d2928fa3eee2ce422d821037144ba4673b8dcd37b2c7b26eab not found: ID does not exist" Oct 14 09:50:22 crc kubenswrapper[4870]: I1014 09:50:22.315462 4870 scope.go:117] "RemoveContainer" containerID="9c0ba9e501df0399b31b3a222ed3f8030c15f9a9070c49bae2794bfe7f1bd259" Oct 14 09:50:22 crc kubenswrapper[4870]: E1014 09:50:22.315665 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9c0ba9e501df0399b31b3a222ed3f8030c15f9a9070c49bae2794bfe7f1bd259\": container with ID starting with 9c0ba9e501df0399b31b3a222ed3f8030c15f9a9070c49bae2794bfe7f1bd259 not found: ID does not exist" containerID="9c0ba9e501df0399b31b3a222ed3f8030c15f9a9070c49bae2794bfe7f1bd259" Oct 14 09:50:22 crc kubenswrapper[4870]: I1014 09:50:22.315690 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9c0ba9e501df0399b31b3a222ed3f8030c15f9a9070c49bae2794bfe7f1bd259"} err="failed to get container status \"9c0ba9e501df0399b31b3a222ed3f8030c15f9a9070c49bae2794bfe7f1bd259\": rpc error: code = NotFound desc = could not find container \"9c0ba9e501df0399b31b3a222ed3f8030c15f9a9070c49bae2794bfe7f1bd259\": container with ID starting with 9c0ba9e501df0399b31b3a222ed3f8030c15f9a9070c49bae2794bfe7f1bd259 not found: ID does not exist" Oct 14 09:50:23 crc kubenswrapper[4870]: I1014 09:50:23.050250 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2e5f7f0c-129a-4b34-9276-82dacd66ca3d" path="/var/lib/kubelet/pods/2e5f7f0c-129a-4b34-9276-82dacd66ca3d/volumes" Oct 14 09:50:31 crc kubenswrapper[4870]: I1014 09:50:31.311265 4870 generic.go:334] "Generic (PLEG): container finished" podID="59c26d47-f3a1-4103-a00d-5ee15b387c4a" containerID="352511aa4e9ed04b460e136058a6ea68dd0aa6675e879931506cbd3d6e534f76" exitCode=0 Oct 14 09:50:31 crc kubenswrapper[4870]: I1014 09:50:31.311337 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-sriov-openstack-openstack-cell1-7zwk7" event={"ID":"59c26d47-f3a1-4103-a00d-5ee15b387c4a","Type":"ContainerDied","Data":"352511aa4e9ed04b460e136058a6ea68dd0aa6675e879931506cbd3d6e534f76"} Oct 14 09:50:32 crc kubenswrapper[4870]: I1014 09:50:32.510203 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-f77jx"] Oct 14 09:50:32 crc kubenswrapper[4870]: E1014 09:50:32.511299 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e5f7f0c-129a-4b34-9276-82dacd66ca3d" containerName="extract-utilities" Oct 14 09:50:32 crc kubenswrapper[4870]: I1014 09:50:32.511320 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e5f7f0c-129a-4b34-9276-82dacd66ca3d" containerName="extract-utilities" Oct 14 09:50:32 crc kubenswrapper[4870]: E1014 09:50:32.511359 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e5f7f0c-129a-4b34-9276-82dacd66ca3d" containerName="extract-content" Oct 14 09:50:32 crc kubenswrapper[4870]: I1014 09:50:32.511371 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e5f7f0c-129a-4b34-9276-82dacd66ca3d" containerName="extract-content" Oct 14 09:50:32 crc kubenswrapper[4870]: E1014 09:50:32.511384 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e5f7f0c-129a-4b34-9276-82dacd66ca3d" containerName="registry-server" Oct 14 09:50:32 crc kubenswrapper[4870]: I1014 09:50:32.511396 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e5f7f0c-129a-4b34-9276-82dacd66ca3d" containerName="registry-server" Oct 14 09:50:32 crc kubenswrapper[4870]: I1014 09:50:32.511793 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e5f7f0c-129a-4b34-9276-82dacd66ca3d" containerName="registry-server" Oct 14 09:50:32 crc kubenswrapper[4870]: I1014 09:50:32.514571 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-f77jx" Oct 14 09:50:32 crc kubenswrapper[4870]: I1014 09:50:32.524262 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-f77jx"] Oct 14 09:50:32 crc kubenswrapper[4870]: I1014 09:50:32.572850 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mwjrg\" (UniqueName: \"kubernetes.io/projected/6c5fc210-e910-4b2a-b121-04ac3dfb28f3-kube-api-access-mwjrg\") pod \"certified-operators-f77jx\" (UID: \"6c5fc210-e910-4b2a-b121-04ac3dfb28f3\") " pod="openshift-marketplace/certified-operators-f77jx" Oct 14 09:50:32 crc kubenswrapper[4870]: I1014 09:50:32.572922 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6c5fc210-e910-4b2a-b121-04ac3dfb28f3-catalog-content\") pod \"certified-operators-f77jx\" (UID: \"6c5fc210-e910-4b2a-b121-04ac3dfb28f3\") " pod="openshift-marketplace/certified-operators-f77jx" Oct 14 09:50:32 crc kubenswrapper[4870]: I1014 09:50:32.573038 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6c5fc210-e910-4b2a-b121-04ac3dfb28f3-utilities\") pod \"certified-operators-f77jx\" (UID: \"6c5fc210-e910-4b2a-b121-04ac3dfb28f3\") " pod="openshift-marketplace/certified-operators-f77jx" Oct 14 09:50:32 crc kubenswrapper[4870]: I1014 09:50:32.675201 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mwjrg\" (UniqueName: \"kubernetes.io/projected/6c5fc210-e910-4b2a-b121-04ac3dfb28f3-kube-api-access-mwjrg\") pod \"certified-operators-f77jx\" (UID: \"6c5fc210-e910-4b2a-b121-04ac3dfb28f3\") " pod="openshift-marketplace/certified-operators-f77jx" Oct 14 09:50:32 crc kubenswrapper[4870]: I1014 09:50:32.675267 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6c5fc210-e910-4b2a-b121-04ac3dfb28f3-catalog-content\") pod \"certified-operators-f77jx\" (UID: \"6c5fc210-e910-4b2a-b121-04ac3dfb28f3\") " pod="openshift-marketplace/certified-operators-f77jx" Oct 14 09:50:32 crc kubenswrapper[4870]: I1014 09:50:32.675380 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6c5fc210-e910-4b2a-b121-04ac3dfb28f3-utilities\") pod \"certified-operators-f77jx\" (UID: \"6c5fc210-e910-4b2a-b121-04ac3dfb28f3\") " pod="openshift-marketplace/certified-operators-f77jx" Oct 14 09:50:32 crc kubenswrapper[4870]: I1014 09:50:32.676131 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6c5fc210-e910-4b2a-b121-04ac3dfb28f3-utilities\") pod \"certified-operators-f77jx\" (UID: \"6c5fc210-e910-4b2a-b121-04ac3dfb28f3\") " pod="openshift-marketplace/certified-operators-f77jx" Oct 14 09:50:32 crc kubenswrapper[4870]: I1014 09:50:32.676245 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6c5fc210-e910-4b2a-b121-04ac3dfb28f3-catalog-content\") pod \"certified-operators-f77jx\" (UID: \"6c5fc210-e910-4b2a-b121-04ac3dfb28f3\") " pod="openshift-marketplace/certified-operators-f77jx" Oct 14 09:50:32 crc kubenswrapper[4870]: I1014 09:50:32.695525 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mwjrg\" (UniqueName: \"kubernetes.io/projected/6c5fc210-e910-4b2a-b121-04ac3dfb28f3-kube-api-access-mwjrg\") pod \"certified-operators-f77jx\" (UID: \"6c5fc210-e910-4b2a-b121-04ac3dfb28f3\") " pod="openshift-marketplace/certified-operators-f77jx" Oct 14 09:50:32 crc kubenswrapper[4870]: I1014 09:50:32.852281 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-sriov-openstack-openstack-cell1-7zwk7" Oct 14 09:50:32 crc kubenswrapper[4870]: I1014 09:50:32.855416 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-f77jx" Oct 14 09:50:32 crc kubenswrapper[4870]: I1014 09:50:32.984762 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/59c26d47-f3a1-4103-a00d-5ee15b387c4a-ssh-key\") pod \"59c26d47-f3a1-4103-a00d-5ee15b387c4a\" (UID: \"59c26d47-f3a1-4103-a00d-5ee15b387c4a\") " Oct 14 09:50:32 crc kubenswrapper[4870]: I1014 09:50:32.984872 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q8skt\" (UniqueName: \"kubernetes.io/projected/59c26d47-f3a1-4103-a00d-5ee15b387c4a-kube-api-access-q8skt\") pod \"59c26d47-f3a1-4103-a00d-5ee15b387c4a\" (UID: \"59c26d47-f3a1-4103-a00d-5ee15b387c4a\") " Oct 14 09:50:32 crc kubenswrapper[4870]: I1014 09:50:32.984903 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/59c26d47-f3a1-4103-a00d-5ee15b387c4a-neutron-sriov-combined-ca-bundle\") pod \"59c26d47-f3a1-4103-a00d-5ee15b387c4a\" (UID: \"59c26d47-f3a1-4103-a00d-5ee15b387c4a\") " Oct 14 09:50:32 crc kubenswrapper[4870]: I1014 09:50:32.985096 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/59c26d47-f3a1-4103-a00d-5ee15b387c4a-inventory\") pod \"59c26d47-f3a1-4103-a00d-5ee15b387c4a\" (UID: \"59c26d47-f3a1-4103-a00d-5ee15b387c4a\") " Oct 14 09:50:32 crc kubenswrapper[4870]: I1014 09:50:32.985159 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-sriov-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/59c26d47-f3a1-4103-a00d-5ee15b387c4a-neutron-sriov-agent-neutron-config-0\") pod \"59c26d47-f3a1-4103-a00d-5ee15b387c4a\" (UID: \"59c26d47-f3a1-4103-a00d-5ee15b387c4a\") " Oct 14 09:50:32 crc kubenswrapper[4870]: I1014 09:50:32.985228 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/59c26d47-f3a1-4103-a00d-5ee15b387c4a-ceph\") pod \"59c26d47-f3a1-4103-a00d-5ee15b387c4a\" (UID: \"59c26d47-f3a1-4103-a00d-5ee15b387c4a\") " Oct 14 09:50:32 crc kubenswrapper[4870]: I1014 09:50:32.990509 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/59c26d47-f3a1-4103-a00d-5ee15b387c4a-kube-api-access-q8skt" (OuterVolumeSpecName: "kube-api-access-q8skt") pod "59c26d47-f3a1-4103-a00d-5ee15b387c4a" (UID: "59c26d47-f3a1-4103-a00d-5ee15b387c4a"). InnerVolumeSpecName "kube-api-access-q8skt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:50:32 crc kubenswrapper[4870]: I1014 09:50:32.998120 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/59c26d47-f3a1-4103-a00d-5ee15b387c4a-neutron-sriov-combined-ca-bundle" (OuterVolumeSpecName: "neutron-sriov-combined-ca-bundle") pod "59c26d47-f3a1-4103-a00d-5ee15b387c4a" (UID: "59c26d47-f3a1-4103-a00d-5ee15b387c4a"). InnerVolumeSpecName "neutron-sriov-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:50:32 crc kubenswrapper[4870]: I1014 09:50:32.998616 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/59c26d47-f3a1-4103-a00d-5ee15b387c4a-ceph" (OuterVolumeSpecName: "ceph") pod "59c26d47-f3a1-4103-a00d-5ee15b387c4a" (UID: "59c26d47-f3a1-4103-a00d-5ee15b387c4a"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:50:33 crc kubenswrapper[4870]: I1014 09:50:33.047563 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/59c26d47-f3a1-4103-a00d-5ee15b387c4a-inventory" (OuterVolumeSpecName: "inventory") pod "59c26d47-f3a1-4103-a00d-5ee15b387c4a" (UID: "59c26d47-f3a1-4103-a00d-5ee15b387c4a"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:50:33 crc kubenswrapper[4870]: I1014 09:50:33.061477 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/59c26d47-f3a1-4103-a00d-5ee15b387c4a-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "59c26d47-f3a1-4103-a00d-5ee15b387c4a" (UID: "59c26d47-f3a1-4103-a00d-5ee15b387c4a"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:50:33 crc kubenswrapper[4870]: I1014 09:50:33.076852 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/59c26d47-f3a1-4103-a00d-5ee15b387c4a-neutron-sriov-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-sriov-agent-neutron-config-0") pod "59c26d47-f3a1-4103-a00d-5ee15b387c4a" (UID: "59c26d47-f3a1-4103-a00d-5ee15b387c4a"). InnerVolumeSpecName "neutron-sriov-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:50:33 crc kubenswrapper[4870]: I1014 09:50:33.091627 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q8skt\" (UniqueName: \"kubernetes.io/projected/59c26d47-f3a1-4103-a00d-5ee15b387c4a-kube-api-access-q8skt\") on node \"crc\" DevicePath \"\"" Oct 14 09:50:33 crc kubenswrapper[4870]: I1014 09:50:33.091661 4870 reconciler_common.go:293] "Volume detached for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/59c26d47-f3a1-4103-a00d-5ee15b387c4a-neutron-sriov-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 09:50:33 crc kubenswrapper[4870]: I1014 09:50:33.091674 4870 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/59c26d47-f3a1-4103-a00d-5ee15b387c4a-inventory\") on node \"crc\" DevicePath \"\"" Oct 14 09:50:33 crc kubenswrapper[4870]: I1014 09:50:33.091689 4870 reconciler_common.go:293] "Volume detached for volume \"neutron-sriov-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/59c26d47-f3a1-4103-a00d-5ee15b387c4a-neutron-sriov-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Oct 14 09:50:33 crc kubenswrapper[4870]: I1014 09:50:33.091755 4870 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/59c26d47-f3a1-4103-a00d-5ee15b387c4a-ceph\") on node \"crc\" DevicePath \"\"" Oct 14 09:50:33 crc kubenswrapper[4870]: I1014 09:50:33.091765 4870 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/59c26d47-f3a1-4103-a00d-5ee15b387c4a-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 14 09:50:33 crc kubenswrapper[4870]: I1014 09:50:33.331400 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-sriov-openstack-openstack-cell1-7zwk7" event={"ID":"59c26d47-f3a1-4103-a00d-5ee15b387c4a","Type":"ContainerDied","Data":"fa6c0fd057686830232b71f9bab869abfb61f61e5c8965f9bba7dba2d5b54f4a"} Oct 14 09:50:33 crc kubenswrapper[4870]: I1014 09:50:33.331443 4870 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fa6c0fd057686830232b71f9bab869abfb61f61e5c8965f9bba7dba2d5b54f4a" Oct 14 09:50:33 crc kubenswrapper[4870]: I1014 09:50:33.331491 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-sriov-openstack-openstack-cell1-7zwk7" Oct 14 09:50:33 crc kubenswrapper[4870]: I1014 09:50:33.423827 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-f77jx"] Oct 14 09:50:33 crc kubenswrapper[4870]: I1014 09:50:33.455620 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-dhcp-openstack-openstack-cell1-tmmz7"] Oct 14 09:50:33 crc kubenswrapper[4870]: E1014 09:50:33.456159 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="59c26d47-f3a1-4103-a00d-5ee15b387c4a" containerName="neutron-sriov-openstack-openstack-cell1" Oct 14 09:50:33 crc kubenswrapper[4870]: I1014 09:50:33.456180 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="59c26d47-f3a1-4103-a00d-5ee15b387c4a" containerName="neutron-sriov-openstack-openstack-cell1" Oct 14 09:50:33 crc kubenswrapper[4870]: I1014 09:50:33.456404 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="59c26d47-f3a1-4103-a00d-5ee15b387c4a" containerName="neutron-sriov-openstack-openstack-cell1" Oct 14 09:50:33 crc kubenswrapper[4870]: I1014 09:50:33.457228 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-dhcp-openstack-openstack-cell1-tmmz7" Oct 14 09:50:33 crc kubenswrapper[4870]: I1014 09:50:33.465051 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-dhcp-agent-neutron-config" Oct 14 09:50:33 crc kubenswrapper[4870]: I1014 09:50:33.465946 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 14 09:50:33 crc kubenswrapper[4870]: I1014 09:50:33.466039 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 14 09:50:33 crc kubenswrapper[4870]: I1014 09:50:33.466180 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-wrz7w" Oct 14 09:50:33 crc kubenswrapper[4870]: I1014 09:50:33.466180 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 14 09:50:33 crc kubenswrapper[4870]: I1014 09:50:33.491564 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-dhcp-openstack-openstack-cell1-tmmz7"] Oct 14 09:50:33 crc kubenswrapper[4870]: I1014 09:50:33.625438 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/0417d7ee-a386-402e-bfa8-72b8dc5cf38e-ceph\") pod \"neutron-dhcp-openstack-openstack-cell1-tmmz7\" (UID: \"0417d7ee-a386-402e-bfa8-72b8dc5cf38e\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-tmmz7" Oct 14 09:50:33 crc kubenswrapper[4870]: I1014 09:50:33.625511 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0417d7ee-a386-402e-bfa8-72b8dc5cf38e-inventory\") pod \"neutron-dhcp-openstack-openstack-cell1-tmmz7\" (UID: \"0417d7ee-a386-402e-bfa8-72b8dc5cf38e\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-tmmz7" Oct 14 09:50:33 crc kubenswrapper[4870]: I1014 09:50:33.625577 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0417d7ee-a386-402e-bfa8-72b8dc5cf38e-neutron-dhcp-combined-ca-bundle\") pod \"neutron-dhcp-openstack-openstack-cell1-tmmz7\" (UID: \"0417d7ee-a386-402e-bfa8-72b8dc5cf38e\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-tmmz7" Oct 14 09:50:33 crc kubenswrapper[4870]: I1014 09:50:33.625631 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hxcmg\" (UniqueName: \"kubernetes.io/projected/0417d7ee-a386-402e-bfa8-72b8dc5cf38e-kube-api-access-hxcmg\") pod \"neutron-dhcp-openstack-openstack-cell1-tmmz7\" (UID: \"0417d7ee-a386-402e-bfa8-72b8dc5cf38e\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-tmmz7" Oct 14 09:50:33 crc kubenswrapper[4870]: I1014 09:50:33.625665 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-dhcp-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/0417d7ee-a386-402e-bfa8-72b8dc5cf38e-neutron-dhcp-agent-neutron-config-0\") pod \"neutron-dhcp-openstack-openstack-cell1-tmmz7\" (UID: \"0417d7ee-a386-402e-bfa8-72b8dc5cf38e\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-tmmz7" Oct 14 09:50:33 crc kubenswrapper[4870]: I1014 09:50:33.625717 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0417d7ee-a386-402e-bfa8-72b8dc5cf38e-ssh-key\") pod \"neutron-dhcp-openstack-openstack-cell1-tmmz7\" (UID: \"0417d7ee-a386-402e-bfa8-72b8dc5cf38e\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-tmmz7" Oct 14 09:50:33 crc kubenswrapper[4870]: I1014 09:50:33.727428 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/0417d7ee-a386-402e-bfa8-72b8dc5cf38e-ceph\") pod \"neutron-dhcp-openstack-openstack-cell1-tmmz7\" (UID: \"0417d7ee-a386-402e-bfa8-72b8dc5cf38e\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-tmmz7" Oct 14 09:50:33 crc kubenswrapper[4870]: I1014 09:50:33.727795 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0417d7ee-a386-402e-bfa8-72b8dc5cf38e-inventory\") pod \"neutron-dhcp-openstack-openstack-cell1-tmmz7\" (UID: \"0417d7ee-a386-402e-bfa8-72b8dc5cf38e\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-tmmz7" Oct 14 09:50:33 crc kubenswrapper[4870]: I1014 09:50:33.727840 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0417d7ee-a386-402e-bfa8-72b8dc5cf38e-neutron-dhcp-combined-ca-bundle\") pod \"neutron-dhcp-openstack-openstack-cell1-tmmz7\" (UID: \"0417d7ee-a386-402e-bfa8-72b8dc5cf38e\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-tmmz7" Oct 14 09:50:33 crc kubenswrapper[4870]: I1014 09:50:33.727915 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hxcmg\" (UniqueName: \"kubernetes.io/projected/0417d7ee-a386-402e-bfa8-72b8dc5cf38e-kube-api-access-hxcmg\") pod \"neutron-dhcp-openstack-openstack-cell1-tmmz7\" (UID: \"0417d7ee-a386-402e-bfa8-72b8dc5cf38e\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-tmmz7" Oct 14 09:50:33 crc kubenswrapper[4870]: I1014 09:50:33.727984 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-dhcp-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/0417d7ee-a386-402e-bfa8-72b8dc5cf38e-neutron-dhcp-agent-neutron-config-0\") pod \"neutron-dhcp-openstack-openstack-cell1-tmmz7\" (UID: \"0417d7ee-a386-402e-bfa8-72b8dc5cf38e\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-tmmz7" Oct 14 09:50:33 crc kubenswrapper[4870]: I1014 09:50:33.728070 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0417d7ee-a386-402e-bfa8-72b8dc5cf38e-ssh-key\") pod \"neutron-dhcp-openstack-openstack-cell1-tmmz7\" (UID: \"0417d7ee-a386-402e-bfa8-72b8dc5cf38e\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-tmmz7" Oct 14 09:50:33 crc kubenswrapper[4870]: I1014 09:50:33.732725 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0417d7ee-a386-402e-bfa8-72b8dc5cf38e-inventory\") pod \"neutron-dhcp-openstack-openstack-cell1-tmmz7\" (UID: \"0417d7ee-a386-402e-bfa8-72b8dc5cf38e\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-tmmz7" Oct 14 09:50:33 crc kubenswrapper[4870]: I1014 09:50:33.733038 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-dhcp-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/0417d7ee-a386-402e-bfa8-72b8dc5cf38e-neutron-dhcp-agent-neutron-config-0\") pod \"neutron-dhcp-openstack-openstack-cell1-tmmz7\" (UID: \"0417d7ee-a386-402e-bfa8-72b8dc5cf38e\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-tmmz7" Oct 14 09:50:33 crc kubenswrapper[4870]: I1014 09:50:33.733270 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/0417d7ee-a386-402e-bfa8-72b8dc5cf38e-ceph\") pod \"neutron-dhcp-openstack-openstack-cell1-tmmz7\" (UID: \"0417d7ee-a386-402e-bfa8-72b8dc5cf38e\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-tmmz7" Oct 14 09:50:33 crc kubenswrapper[4870]: I1014 09:50:33.733814 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0417d7ee-a386-402e-bfa8-72b8dc5cf38e-ssh-key\") pod \"neutron-dhcp-openstack-openstack-cell1-tmmz7\" (UID: \"0417d7ee-a386-402e-bfa8-72b8dc5cf38e\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-tmmz7" Oct 14 09:50:33 crc kubenswrapper[4870]: I1014 09:50:33.734724 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0417d7ee-a386-402e-bfa8-72b8dc5cf38e-neutron-dhcp-combined-ca-bundle\") pod \"neutron-dhcp-openstack-openstack-cell1-tmmz7\" (UID: \"0417d7ee-a386-402e-bfa8-72b8dc5cf38e\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-tmmz7" Oct 14 09:50:33 crc kubenswrapper[4870]: I1014 09:50:33.746704 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hxcmg\" (UniqueName: \"kubernetes.io/projected/0417d7ee-a386-402e-bfa8-72b8dc5cf38e-kube-api-access-hxcmg\") pod \"neutron-dhcp-openstack-openstack-cell1-tmmz7\" (UID: \"0417d7ee-a386-402e-bfa8-72b8dc5cf38e\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-tmmz7" Oct 14 09:50:33 crc kubenswrapper[4870]: I1014 09:50:33.783979 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-dhcp-openstack-openstack-cell1-tmmz7" Oct 14 09:50:34 crc kubenswrapper[4870]: I1014 09:50:34.344530 4870 generic.go:334] "Generic (PLEG): container finished" podID="6c5fc210-e910-4b2a-b121-04ac3dfb28f3" containerID="01161a62ff3281b8faccb261844b26b3e31474e366985e62a27a56387ea04aa1" exitCode=0 Oct 14 09:50:34 crc kubenswrapper[4870]: I1014 09:50:34.344626 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-f77jx" event={"ID":"6c5fc210-e910-4b2a-b121-04ac3dfb28f3","Type":"ContainerDied","Data":"01161a62ff3281b8faccb261844b26b3e31474e366985e62a27a56387ea04aa1"} Oct 14 09:50:34 crc kubenswrapper[4870]: I1014 09:50:34.344989 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-f77jx" event={"ID":"6c5fc210-e910-4b2a-b121-04ac3dfb28f3","Type":"ContainerStarted","Data":"18ddaf6f66b52d9810f5bd6943ba59b8148dbbe4a471bd11c7579c6771204d97"} Oct 14 09:50:34 crc kubenswrapper[4870]: I1014 09:50:34.408734 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-dhcp-openstack-openstack-cell1-tmmz7"] Oct 14 09:50:35 crc kubenswrapper[4870]: I1014 09:50:35.361112 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-dhcp-openstack-openstack-cell1-tmmz7" event={"ID":"0417d7ee-a386-402e-bfa8-72b8dc5cf38e","Type":"ContainerStarted","Data":"840d8690d8af62974f2331b8947cbb887b9cef98c744e184ad314417873a2b55"} Oct 14 09:50:36 crc kubenswrapper[4870]: I1014 09:50:36.375935 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-dhcp-openstack-openstack-cell1-tmmz7" event={"ID":"0417d7ee-a386-402e-bfa8-72b8dc5cf38e","Type":"ContainerStarted","Data":"558de7880c05ee405588eb453add30ff96a1ddba7d6d1f84420ac389c041c76d"} Oct 14 09:50:36 crc kubenswrapper[4870]: I1014 09:50:36.378471 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-f77jx" event={"ID":"6c5fc210-e910-4b2a-b121-04ac3dfb28f3","Type":"ContainerStarted","Data":"a7a1586659e52de67f678b41e3780acfb4a3c882332181c9a068d4e13b2a981d"} Oct 14 09:50:36 crc kubenswrapper[4870]: I1014 09:50:36.407900 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-dhcp-openstack-openstack-cell1-tmmz7" podStartSLOduration=2.553585817 podStartE2EDuration="3.407872056s" podCreationTimestamp="2025-10-14 09:50:33 +0000 UTC" firstStartedPulling="2025-10-14 09:50:34.41840181 +0000 UTC m=+10170.115762181" lastFinishedPulling="2025-10-14 09:50:35.272688049 +0000 UTC m=+10170.970048420" observedRunningTime="2025-10-14 09:50:36.405631481 +0000 UTC m=+10172.102991882" watchObservedRunningTime="2025-10-14 09:50:36.407872056 +0000 UTC m=+10172.105232457" Oct 14 09:50:37 crc kubenswrapper[4870]: I1014 09:50:37.392684 4870 generic.go:334] "Generic (PLEG): container finished" podID="6c5fc210-e910-4b2a-b121-04ac3dfb28f3" containerID="a7a1586659e52de67f678b41e3780acfb4a3c882332181c9a068d4e13b2a981d" exitCode=0 Oct 14 09:50:37 crc kubenswrapper[4870]: I1014 09:50:37.392787 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-f77jx" event={"ID":"6c5fc210-e910-4b2a-b121-04ac3dfb28f3","Type":"ContainerDied","Data":"a7a1586659e52de67f678b41e3780acfb4a3c882332181c9a068d4e13b2a981d"} Oct 14 09:50:38 crc kubenswrapper[4870]: I1014 09:50:38.409311 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-f77jx" event={"ID":"6c5fc210-e910-4b2a-b121-04ac3dfb28f3","Type":"ContainerStarted","Data":"6936a03fa8461a074e41cc7cb15faa7a79c77b9b159fdf2d18bf5aa1f232ba91"} Oct 14 09:50:42 crc kubenswrapper[4870]: I1014 09:50:42.856506 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-f77jx" Oct 14 09:50:42 crc kubenswrapper[4870]: I1014 09:50:42.857337 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-f77jx" Oct 14 09:50:43 crc kubenswrapper[4870]: I1014 09:50:43.939766 4870 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-f77jx" podUID="6c5fc210-e910-4b2a-b121-04ac3dfb28f3" containerName="registry-server" probeResult="failure" output=< Oct 14 09:50:43 crc kubenswrapper[4870]: timeout: failed to connect service ":50051" within 1s Oct 14 09:50:43 crc kubenswrapper[4870]: > Oct 14 09:50:52 crc kubenswrapper[4870]: I1014 09:50:52.929422 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-f77jx" Oct 14 09:50:52 crc kubenswrapper[4870]: I1014 09:50:52.966946 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-f77jx" podStartSLOduration=17.503611928 podStartE2EDuration="20.966925229s" podCreationTimestamp="2025-10-14 09:50:32 +0000 UTC" firstStartedPulling="2025-10-14 09:50:34.347066075 +0000 UTC m=+10170.044426456" lastFinishedPulling="2025-10-14 09:50:37.810379376 +0000 UTC m=+10173.507739757" observedRunningTime="2025-10-14 09:50:38.432944796 +0000 UTC m=+10174.130305167" watchObservedRunningTime="2025-10-14 09:50:52.966925229 +0000 UTC m=+10188.664285600" Oct 14 09:50:53 crc kubenswrapper[4870]: I1014 09:50:53.019406 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-f77jx" Oct 14 09:50:53 crc kubenswrapper[4870]: I1014 09:50:53.179587 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-f77jx"] Oct 14 09:50:54 crc kubenswrapper[4870]: I1014 09:50:54.628919 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-f77jx" podUID="6c5fc210-e910-4b2a-b121-04ac3dfb28f3" containerName="registry-server" containerID="cri-o://6936a03fa8461a074e41cc7cb15faa7a79c77b9b159fdf2d18bf5aa1f232ba91" gracePeriod=2 Oct 14 09:50:55 crc kubenswrapper[4870]: I1014 09:50:55.294498 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-f77jx" Oct 14 09:50:55 crc kubenswrapper[4870]: I1014 09:50:55.386894 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mwjrg\" (UniqueName: \"kubernetes.io/projected/6c5fc210-e910-4b2a-b121-04ac3dfb28f3-kube-api-access-mwjrg\") pod \"6c5fc210-e910-4b2a-b121-04ac3dfb28f3\" (UID: \"6c5fc210-e910-4b2a-b121-04ac3dfb28f3\") " Oct 14 09:50:55 crc kubenswrapper[4870]: I1014 09:50:55.387176 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6c5fc210-e910-4b2a-b121-04ac3dfb28f3-catalog-content\") pod \"6c5fc210-e910-4b2a-b121-04ac3dfb28f3\" (UID: \"6c5fc210-e910-4b2a-b121-04ac3dfb28f3\") " Oct 14 09:50:55 crc kubenswrapper[4870]: I1014 09:50:55.387240 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6c5fc210-e910-4b2a-b121-04ac3dfb28f3-utilities\") pod \"6c5fc210-e910-4b2a-b121-04ac3dfb28f3\" (UID: \"6c5fc210-e910-4b2a-b121-04ac3dfb28f3\") " Oct 14 09:50:55 crc kubenswrapper[4870]: I1014 09:50:55.387850 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6c5fc210-e910-4b2a-b121-04ac3dfb28f3-utilities" (OuterVolumeSpecName: "utilities") pod "6c5fc210-e910-4b2a-b121-04ac3dfb28f3" (UID: "6c5fc210-e910-4b2a-b121-04ac3dfb28f3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 09:50:55 crc kubenswrapper[4870]: I1014 09:50:55.394329 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6c5fc210-e910-4b2a-b121-04ac3dfb28f3-kube-api-access-mwjrg" (OuterVolumeSpecName: "kube-api-access-mwjrg") pod "6c5fc210-e910-4b2a-b121-04ac3dfb28f3" (UID: "6c5fc210-e910-4b2a-b121-04ac3dfb28f3"). InnerVolumeSpecName "kube-api-access-mwjrg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:50:55 crc kubenswrapper[4870]: I1014 09:50:55.441357 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6c5fc210-e910-4b2a-b121-04ac3dfb28f3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6c5fc210-e910-4b2a-b121-04ac3dfb28f3" (UID: "6c5fc210-e910-4b2a-b121-04ac3dfb28f3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 09:50:55 crc kubenswrapper[4870]: I1014 09:50:55.490281 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mwjrg\" (UniqueName: \"kubernetes.io/projected/6c5fc210-e910-4b2a-b121-04ac3dfb28f3-kube-api-access-mwjrg\") on node \"crc\" DevicePath \"\"" Oct 14 09:50:55 crc kubenswrapper[4870]: I1014 09:50:55.490323 4870 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6c5fc210-e910-4b2a-b121-04ac3dfb28f3-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 09:50:55 crc kubenswrapper[4870]: I1014 09:50:55.490337 4870 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6c5fc210-e910-4b2a-b121-04ac3dfb28f3-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 09:50:55 crc kubenswrapper[4870]: I1014 09:50:55.640940 4870 generic.go:334] "Generic (PLEG): container finished" podID="6c5fc210-e910-4b2a-b121-04ac3dfb28f3" containerID="6936a03fa8461a074e41cc7cb15faa7a79c77b9b159fdf2d18bf5aa1f232ba91" exitCode=0 Oct 14 09:50:55 crc kubenswrapper[4870]: I1014 09:50:55.641001 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-f77jx" Oct 14 09:50:55 crc kubenswrapper[4870]: I1014 09:50:55.641021 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-f77jx" event={"ID":"6c5fc210-e910-4b2a-b121-04ac3dfb28f3","Type":"ContainerDied","Data":"6936a03fa8461a074e41cc7cb15faa7a79c77b9b159fdf2d18bf5aa1f232ba91"} Oct 14 09:50:55 crc kubenswrapper[4870]: I1014 09:50:55.641769 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-f77jx" event={"ID":"6c5fc210-e910-4b2a-b121-04ac3dfb28f3","Type":"ContainerDied","Data":"18ddaf6f66b52d9810f5bd6943ba59b8148dbbe4a471bd11c7579c6771204d97"} Oct 14 09:50:55 crc kubenswrapper[4870]: I1014 09:50:55.641788 4870 scope.go:117] "RemoveContainer" containerID="6936a03fa8461a074e41cc7cb15faa7a79c77b9b159fdf2d18bf5aa1f232ba91" Oct 14 09:50:55 crc kubenswrapper[4870]: I1014 09:50:55.682277 4870 scope.go:117] "RemoveContainer" containerID="a7a1586659e52de67f678b41e3780acfb4a3c882332181c9a068d4e13b2a981d" Oct 14 09:50:55 crc kubenswrapper[4870]: I1014 09:50:55.690497 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-f77jx"] Oct 14 09:50:55 crc kubenswrapper[4870]: I1014 09:50:55.731031 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-f77jx"] Oct 14 09:50:55 crc kubenswrapper[4870]: I1014 09:50:55.736185 4870 scope.go:117] "RemoveContainer" containerID="01161a62ff3281b8faccb261844b26b3e31474e366985e62a27a56387ea04aa1" Oct 14 09:50:55 crc kubenswrapper[4870]: I1014 09:50:55.762937 4870 scope.go:117] "RemoveContainer" containerID="6936a03fa8461a074e41cc7cb15faa7a79c77b9b159fdf2d18bf5aa1f232ba91" Oct 14 09:50:55 crc kubenswrapper[4870]: E1014 09:50:55.765585 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6936a03fa8461a074e41cc7cb15faa7a79c77b9b159fdf2d18bf5aa1f232ba91\": container with ID starting with 6936a03fa8461a074e41cc7cb15faa7a79c77b9b159fdf2d18bf5aa1f232ba91 not found: ID does not exist" containerID="6936a03fa8461a074e41cc7cb15faa7a79c77b9b159fdf2d18bf5aa1f232ba91" Oct 14 09:50:55 crc kubenswrapper[4870]: I1014 09:50:55.765668 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6936a03fa8461a074e41cc7cb15faa7a79c77b9b159fdf2d18bf5aa1f232ba91"} err="failed to get container status \"6936a03fa8461a074e41cc7cb15faa7a79c77b9b159fdf2d18bf5aa1f232ba91\": rpc error: code = NotFound desc = could not find container \"6936a03fa8461a074e41cc7cb15faa7a79c77b9b159fdf2d18bf5aa1f232ba91\": container with ID starting with 6936a03fa8461a074e41cc7cb15faa7a79c77b9b159fdf2d18bf5aa1f232ba91 not found: ID does not exist" Oct 14 09:50:55 crc kubenswrapper[4870]: I1014 09:50:55.765711 4870 scope.go:117] "RemoveContainer" containerID="a7a1586659e52de67f678b41e3780acfb4a3c882332181c9a068d4e13b2a981d" Oct 14 09:50:55 crc kubenswrapper[4870]: E1014 09:50:55.766353 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a7a1586659e52de67f678b41e3780acfb4a3c882332181c9a068d4e13b2a981d\": container with ID starting with a7a1586659e52de67f678b41e3780acfb4a3c882332181c9a068d4e13b2a981d not found: ID does not exist" containerID="a7a1586659e52de67f678b41e3780acfb4a3c882332181c9a068d4e13b2a981d" Oct 14 09:50:55 crc kubenswrapper[4870]: I1014 09:50:55.766407 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a7a1586659e52de67f678b41e3780acfb4a3c882332181c9a068d4e13b2a981d"} err="failed to get container status \"a7a1586659e52de67f678b41e3780acfb4a3c882332181c9a068d4e13b2a981d\": rpc error: code = NotFound desc = could not find container \"a7a1586659e52de67f678b41e3780acfb4a3c882332181c9a068d4e13b2a981d\": container with ID starting with a7a1586659e52de67f678b41e3780acfb4a3c882332181c9a068d4e13b2a981d not found: ID does not exist" Oct 14 09:50:55 crc kubenswrapper[4870]: I1014 09:50:55.766465 4870 scope.go:117] "RemoveContainer" containerID="01161a62ff3281b8faccb261844b26b3e31474e366985e62a27a56387ea04aa1" Oct 14 09:50:55 crc kubenswrapper[4870]: E1014 09:50:55.767376 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"01161a62ff3281b8faccb261844b26b3e31474e366985e62a27a56387ea04aa1\": container with ID starting with 01161a62ff3281b8faccb261844b26b3e31474e366985e62a27a56387ea04aa1 not found: ID does not exist" containerID="01161a62ff3281b8faccb261844b26b3e31474e366985e62a27a56387ea04aa1" Oct 14 09:50:55 crc kubenswrapper[4870]: I1014 09:50:55.767402 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"01161a62ff3281b8faccb261844b26b3e31474e366985e62a27a56387ea04aa1"} err="failed to get container status \"01161a62ff3281b8faccb261844b26b3e31474e366985e62a27a56387ea04aa1\": rpc error: code = NotFound desc = could not find container \"01161a62ff3281b8faccb261844b26b3e31474e366985e62a27a56387ea04aa1\": container with ID starting with 01161a62ff3281b8faccb261844b26b3e31474e366985e62a27a56387ea04aa1 not found: ID does not exist" Oct 14 09:50:57 crc kubenswrapper[4870]: I1014 09:50:57.056404 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6c5fc210-e910-4b2a-b121-04ac3dfb28f3" path="/var/lib/kubelet/pods/6c5fc210-e910-4b2a-b121-04ac3dfb28f3/volumes" Oct 14 09:51:44 crc kubenswrapper[4870]: I1014 09:51:44.322800 4870 generic.go:334] "Generic (PLEG): container finished" podID="0417d7ee-a386-402e-bfa8-72b8dc5cf38e" containerID="558de7880c05ee405588eb453add30ff96a1ddba7d6d1f84420ac389c041c76d" exitCode=0 Oct 14 09:51:44 crc kubenswrapper[4870]: I1014 09:51:44.322916 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-dhcp-openstack-openstack-cell1-tmmz7" event={"ID":"0417d7ee-a386-402e-bfa8-72b8dc5cf38e","Type":"ContainerDied","Data":"558de7880c05ee405588eb453add30ff96a1ddba7d6d1f84420ac389c041c76d"} Oct 14 09:51:45 crc kubenswrapper[4870]: I1014 09:51:45.895398 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-dhcp-openstack-openstack-cell1-tmmz7" Oct 14 09:51:45 crc kubenswrapper[4870]: I1014 09:51:45.965484 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hxcmg\" (UniqueName: \"kubernetes.io/projected/0417d7ee-a386-402e-bfa8-72b8dc5cf38e-kube-api-access-hxcmg\") pod \"0417d7ee-a386-402e-bfa8-72b8dc5cf38e\" (UID: \"0417d7ee-a386-402e-bfa8-72b8dc5cf38e\") " Oct 14 09:51:45 crc kubenswrapper[4870]: I1014 09:51:45.965621 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0417d7ee-a386-402e-bfa8-72b8dc5cf38e-ssh-key\") pod \"0417d7ee-a386-402e-bfa8-72b8dc5cf38e\" (UID: \"0417d7ee-a386-402e-bfa8-72b8dc5cf38e\") " Oct 14 09:51:45 crc kubenswrapper[4870]: I1014 09:51:45.965800 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/0417d7ee-a386-402e-bfa8-72b8dc5cf38e-ceph\") pod \"0417d7ee-a386-402e-bfa8-72b8dc5cf38e\" (UID: \"0417d7ee-a386-402e-bfa8-72b8dc5cf38e\") " Oct 14 09:51:45 crc kubenswrapper[4870]: I1014 09:51:45.965904 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0417d7ee-a386-402e-bfa8-72b8dc5cf38e-inventory\") pod \"0417d7ee-a386-402e-bfa8-72b8dc5cf38e\" (UID: \"0417d7ee-a386-402e-bfa8-72b8dc5cf38e\") " Oct 14 09:51:45 crc kubenswrapper[4870]: I1014 09:51:45.966018 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-dhcp-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/0417d7ee-a386-402e-bfa8-72b8dc5cf38e-neutron-dhcp-agent-neutron-config-0\") pod \"0417d7ee-a386-402e-bfa8-72b8dc5cf38e\" (UID: \"0417d7ee-a386-402e-bfa8-72b8dc5cf38e\") " Oct 14 09:51:45 crc kubenswrapper[4870]: I1014 09:51:45.966189 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0417d7ee-a386-402e-bfa8-72b8dc5cf38e-neutron-dhcp-combined-ca-bundle\") pod \"0417d7ee-a386-402e-bfa8-72b8dc5cf38e\" (UID: \"0417d7ee-a386-402e-bfa8-72b8dc5cf38e\") " Oct 14 09:51:45 crc kubenswrapper[4870]: I1014 09:51:45.981713 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0417d7ee-a386-402e-bfa8-72b8dc5cf38e-neutron-dhcp-combined-ca-bundle" (OuterVolumeSpecName: "neutron-dhcp-combined-ca-bundle") pod "0417d7ee-a386-402e-bfa8-72b8dc5cf38e" (UID: "0417d7ee-a386-402e-bfa8-72b8dc5cf38e"). InnerVolumeSpecName "neutron-dhcp-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:51:45 crc kubenswrapper[4870]: I1014 09:51:45.981791 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0417d7ee-a386-402e-bfa8-72b8dc5cf38e-ceph" (OuterVolumeSpecName: "ceph") pod "0417d7ee-a386-402e-bfa8-72b8dc5cf38e" (UID: "0417d7ee-a386-402e-bfa8-72b8dc5cf38e"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:51:45 crc kubenswrapper[4870]: I1014 09:51:45.984757 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0417d7ee-a386-402e-bfa8-72b8dc5cf38e-kube-api-access-hxcmg" (OuterVolumeSpecName: "kube-api-access-hxcmg") pod "0417d7ee-a386-402e-bfa8-72b8dc5cf38e" (UID: "0417d7ee-a386-402e-bfa8-72b8dc5cf38e"). InnerVolumeSpecName "kube-api-access-hxcmg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:51:46 crc kubenswrapper[4870]: I1014 09:51:46.004765 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0417d7ee-a386-402e-bfa8-72b8dc5cf38e-neutron-dhcp-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-dhcp-agent-neutron-config-0") pod "0417d7ee-a386-402e-bfa8-72b8dc5cf38e" (UID: "0417d7ee-a386-402e-bfa8-72b8dc5cf38e"). InnerVolumeSpecName "neutron-dhcp-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:51:46 crc kubenswrapper[4870]: I1014 09:51:46.006746 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0417d7ee-a386-402e-bfa8-72b8dc5cf38e-inventory" (OuterVolumeSpecName: "inventory") pod "0417d7ee-a386-402e-bfa8-72b8dc5cf38e" (UID: "0417d7ee-a386-402e-bfa8-72b8dc5cf38e"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:51:46 crc kubenswrapper[4870]: I1014 09:51:46.035081 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0417d7ee-a386-402e-bfa8-72b8dc5cf38e-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "0417d7ee-a386-402e-bfa8-72b8dc5cf38e" (UID: "0417d7ee-a386-402e-bfa8-72b8dc5cf38e"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:51:46 crc kubenswrapper[4870]: I1014 09:51:46.069869 4870 reconciler_common.go:293] "Volume detached for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0417d7ee-a386-402e-bfa8-72b8dc5cf38e-neutron-dhcp-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 09:51:46 crc kubenswrapper[4870]: I1014 09:51:46.069911 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hxcmg\" (UniqueName: \"kubernetes.io/projected/0417d7ee-a386-402e-bfa8-72b8dc5cf38e-kube-api-access-hxcmg\") on node \"crc\" DevicePath \"\"" Oct 14 09:51:46 crc kubenswrapper[4870]: I1014 09:51:46.069928 4870 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0417d7ee-a386-402e-bfa8-72b8dc5cf38e-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 14 09:51:46 crc kubenswrapper[4870]: I1014 09:51:46.069945 4870 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/0417d7ee-a386-402e-bfa8-72b8dc5cf38e-ceph\") on node \"crc\" DevicePath \"\"" Oct 14 09:51:46 crc kubenswrapper[4870]: I1014 09:51:46.069957 4870 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0417d7ee-a386-402e-bfa8-72b8dc5cf38e-inventory\") on node \"crc\" DevicePath \"\"" Oct 14 09:51:46 crc kubenswrapper[4870]: I1014 09:51:46.069969 4870 reconciler_common.go:293] "Volume detached for volume \"neutron-dhcp-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/0417d7ee-a386-402e-bfa8-72b8dc5cf38e-neutron-dhcp-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Oct 14 09:51:46 crc kubenswrapper[4870]: I1014 09:51:46.358746 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-dhcp-openstack-openstack-cell1-tmmz7" event={"ID":"0417d7ee-a386-402e-bfa8-72b8dc5cf38e","Type":"ContainerDied","Data":"840d8690d8af62974f2331b8947cbb887b9cef98c744e184ad314417873a2b55"} Oct 14 09:51:46 crc kubenswrapper[4870]: I1014 09:51:46.358805 4870 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="840d8690d8af62974f2331b8947cbb887b9cef98c744e184ad314417873a2b55" Oct 14 09:51:46 crc kubenswrapper[4870]: I1014 09:51:46.358882 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-dhcp-openstack-openstack-cell1-tmmz7" Oct 14 09:52:02 crc kubenswrapper[4870]: I1014 09:52:02.028612 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 14 09:52:02 crc kubenswrapper[4870]: I1014 09:52:02.030367 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell0-conductor-0" podUID="ee5dcbc8-b98e-48b9-a089-a9f5614249d3" containerName="nova-cell0-conductor-conductor" containerID="cri-o://910ed7bad34d9a7163b841990638ec4c809ae09c71d936618c95878703298b7e" gracePeriod=30 Oct 14 09:52:02 crc kubenswrapper[4870]: I1014 09:52:02.074631 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 14 09:52:02 crc kubenswrapper[4870]: I1014 09:52:02.074848 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-conductor-0" podUID="02ebb80d-7ee3-48e5-99e4-1d9e061d6e06" containerName="nova-cell1-conductor-conductor" containerID="cri-o://652700e216f5422f63850496eceac88c82acc847b709d8a19b42bc4103863ad9" gracePeriod=30 Oct 14 09:52:02 crc kubenswrapper[4870]: E1014 09:52:02.699495 4870 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="910ed7bad34d9a7163b841990638ec4c809ae09c71d936618c95878703298b7e" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 14 09:52:02 crc kubenswrapper[4870]: E1014 09:52:02.703102 4870 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="910ed7bad34d9a7163b841990638ec4c809ae09c71d936618c95878703298b7e" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 14 09:52:02 crc kubenswrapper[4870]: E1014 09:52:02.704424 4870 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="910ed7bad34d9a7163b841990638ec4c809ae09c71d936618c95878703298b7e" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 14 09:52:02 crc kubenswrapper[4870]: E1014 09:52:02.704561 4870 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-cell0-conductor-0" podUID="ee5dcbc8-b98e-48b9-a089-a9f5614249d3" containerName="nova-cell0-conductor-conductor" Oct 14 09:52:02 crc kubenswrapper[4870]: I1014 09:52:02.990176 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 14 09:52:02 crc kubenswrapper[4870]: I1014 09:52:02.990475 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="73281705-4198-40a4-b523-c0bfc322c833" containerName="nova-scheduler-scheduler" containerID="cri-o://c0020e7320b47c052dc0efb0f61570b5a39b13c7d39e5b5d0bc9ebec23c7b744" gracePeriod=30 Oct 14 09:52:03 crc kubenswrapper[4870]: I1014 09:52:03.008289 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 14 09:52:03 crc kubenswrapper[4870]: I1014 09:52:03.008786 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="e1055bc1-24f3-4893-ad77-c520a36428bb" containerName="nova-api-log" containerID="cri-o://f61fa490fab929fb5e534efb5255c5f168fdc3b95da9665beae2c6c2d3ba5803" gracePeriod=30 Oct 14 09:52:03 crc kubenswrapper[4870]: I1014 09:52:03.009282 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="e1055bc1-24f3-4893-ad77-c520a36428bb" containerName="nova-api-api" containerID="cri-o://36057428133bd01d5a6c0e75037e7dff3a30c1c00aa184489e5fee2e48ca9c66" gracePeriod=30 Oct 14 09:52:03 crc kubenswrapper[4870]: I1014 09:52:03.050372 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 14 09:52:03 crc kubenswrapper[4870]: I1014 09:52:03.050954 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="17f1a688-1bd1-4b88-ad00-51caf1af0ce0" containerName="nova-metadata-log" containerID="cri-o://990d26c8691a9651642a973ab6b25f404d460d767a6e4471a24b9aaf9d964606" gracePeriod=30 Oct 14 09:52:03 crc kubenswrapper[4870]: I1014 09:52:03.051087 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="17f1a688-1bd1-4b88-ad00-51caf1af0ce0" containerName="nova-metadata-metadata" containerID="cri-o://1bc76a1a314ba9d308a310397f670dc886227d98b6f2d40ae039086e08114b2d" gracePeriod=30 Oct 14 09:52:03 crc kubenswrapper[4870]: I1014 09:52:03.555202 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 14 09:52:03 crc kubenswrapper[4870]: I1014 09:52:03.624094 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/02ebb80d-7ee3-48e5-99e4-1d9e061d6e06-config-data\") pod \"02ebb80d-7ee3-48e5-99e4-1d9e061d6e06\" (UID: \"02ebb80d-7ee3-48e5-99e4-1d9e061d6e06\") " Oct 14 09:52:03 crc kubenswrapper[4870]: I1014 09:52:03.624307 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02ebb80d-7ee3-48e5-99e4-1d9e061d6e06-combined-ca-bundle\") pod \"02ebb80d-7ee3-48e5-99e4-1d9e061d6e06\" (UID: \"02ebb80d-7ee3-48e5-99e4-1d9e061d6e06\") " Oct 14 09:52:03 crc kubenswrapper[4870]: I1014 09:52:03.624341 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p9bkx\" (UniqueName: \"kubernetes.io/projected/02ebb80d-7ee3-48e5-99e4-1d9e061d6e06-kube-api-access-p9bkx\") pod \"02ebb80d-7ee3-48e5-99e4-1d9e061d6e06\" (UID: \"02ebb80d-7ee3-48e5-99e4-1d9e061d6e06\") " Oct 14 09:52:03 crc kubenswrapper[4870]: I1014 09:52:03.629267 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/02ebb80d-7ee3-48e5-99e4-1d9e061d6e06-kube-api-access-p9bkx" (OuterVolumeSpecName: "kube-api-access-p9bkx") pod "02ebb80d-7ee3-48e5-99e4-1d9e061d6e06" (UID: "02ebb80d-7ee3-48e5-99e4-1d9e061d6e06"). InnerVolumeSpecName "kube-api-access-p9bkx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:52:03 crc kubenswrapper[4870]: I1014 09:52:03.649510 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/02ebb80d-7ee3-48e5-99e4-1d9e061d6e06-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "02ebb80d-7ee3-48e5-99e4-1d9e061d6e06" (UID: "02ebb80d-7ee3-48e5-99e4-1d9e061d6e06"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:52:03 crc kubenswrapper[4870]: I1014 09:52:03.658558 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/02ebb80d-7ee3-48e5-99e4-1d9e061d6e06-config-data" (OuterVolumeSpecName: "config-data") pod "02ebb80d-7ee3-48e5-99e4-1d9e061d6e06" (UID: "02ebb80d-7ee3-48e5-99e4-1d9e061d6e06"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:52:03 crc kubenswrapper[4870]: I1014 09:52:03.684914 4870 generic.go:334] "Generic (PLEG): container finished" podID="02ebb80d-7ee3-48e5-99e4-1d9e061d6e06" containerID="652700e216f5422f63850496eceac88c82acc847b709d8a19b42bc4103863ad9" exitCode=0 Oct 14 09:52:03 crc kubenswrapper[4870]: I1014 09:52:03.684973 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 14 09:52:03 crc kubenswrapper[4870]: I1014 09:52:03.684989 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"02ebb80d-7ee3-48e5-99e4-1d9e061d6e06","Type":"ContainerDied","Data":"652700e216f5422f63850496eceac88c82acc847b709d8a19b42bc4103863ad9"} Oct 14 09:52:03 crc kubenswrapper[4870]: I1014 09:52:03.685024 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"02ebb80d-7ee3-48e5-99e4-1d9e061d6e06","Type":"ContainerDied","Data":"7de2685da46eb5150b97e8f6b62c03e1e3311ee185474328abef90e2a3d2e359"} Oct 14 09:52:03 crc kubenswrapper[4870]: I1014 09:52:03.685039 4870 scope.go:117] "RemoveContainer" containerID="652700e216f5422f63850496eceac88c82acc847b709d8a19b42bc4103863ad9" Oct 14 09:52:03 crc kubenswrapper[4870]: I1014 09:52:03.687645 4870 generic.go:334] "Generic (PLEG): container finished" podID="17f1a688-1bd1-4b88-ad00-51caf1af0ce0" containerID="990d26c8691a9651642a973ab6b25f404d460d767a6e4471a24b9aaf9d964606" exitCode=143 Oct 14 09:52:03 crc kubenswrapper[4870]: I1014 09:52:03.687700 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"17f1a688-1bd1-4b88-ad00-51caf1af0ce0","Type":"ContainerDied","Data":"990d26c8691a9651642a973ab6b25f404d460d767a6e4471a24b9aaf9d964606"} Oct 14 09:52:03 crc kubenswrapper[4870]: I1014 09:52:03.689820 4870 generic.go:334] "Generic (PLEG): container finished" podID="e1055bc1-24f3-4893-ad77-c520a36428bb" containerID="f61fa490fab929fb5e534efb5255c5f168fdc3b95da9665beae2c6c2d3ba5803" exitCode=143 Oct 14 09:52:03 crc kubenswrapper[4870]: I1014 09:52:03.689844 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e1055bc1-24f3-4893-ad77-c520a36428bb","Type":"ContainerDied","Data":"f61fa490fab929fb5e534efb5255c5f168fdc3b95da9665beae2c6c2d3ba5803"} Oct 14 09:52:03 crc kubenswrapper[4870]: I1014 09:52:03.725063 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 14 09:52:03 crc kubenswrapper[4870]: I1014 09:52:03.726362 4870 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/02ebb80d-7ee3-48e5-99e4-1d9e061d6e06-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 09:52:03 crc kubenswrapper[4870]: I1014 09:52:03.726401 4870 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02ebb80d-7ee3-48e5-99e4-1d9e061d6e06-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 09:52:03 crc kubenswrapper[4870]: I1014 09:52:03.726416 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p9bkx\" (UniqueName: \"kubernetes.io/projected/02ebb80d-7ee3-48e5-99e4-1d9e061d6e06-kube-api-access-p9bkx\") on node \"crc\" DevicePath \"\"" Oct 14 09:52:03 crc kubenswrapper[4870]: I1014 09:52:03.728810 4870 scope.go:117] "RemoveContainer" containerID="652700e216f5422f63850496eceac88c82acc847b709d8a19b42bc4103863ad9" Oct 14 09:52:03 crc kubenswrapper[4870]: E1014 09:52:03.729370 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"652700e216f5422f63850496eceac88c82acc847b709d8a19b42bc4103863ad9\": container with ID starting with 652700e216f5422f63850496eceac88c82acc847b709d8a19b42bc4103863ad9 not found: ID does not exist" containerID="652700e216f5422f63850496eceac88c82acc847b709d8a19b42bc4103863ad9" Oct 14 09:52:03 crc kubenswrapper[4870]: I1014 09:52:03.729414 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"652700e216f5422f63850496eceac88c82acc847b709d8a19b42bc4103863ad9"} err="failed to get container status \"652700e216f5422f63850496eceac88c82acc847b709d8a19b42bc4103863ad9\": rpc error: code = NotFound desc = could not find container \"652700e216f5422f63850496eceac88c82acc847b709d8a19b42bc4103863ad9\": container with ID starting with 652700e216f5422f63850496eceac88c82acc847b709d8a19b42bc4103863ad9 not found: ID does not exist" Oct 14 09:52:03 crc kubenswrapper[4870]: I1014 09:52:03.742833 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 14 09:52:03 crc kubenswrapper[4870]: I1014 09:52:03.755546 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 14 09:52:03 crc kubenswrapper[4870]: E1014 09:52:03.756046 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0417d7ee-a386-402e-bfa8-72b8dc5cf38e" containerName="neutron-dhcp-openstack-openstack-cell1" Oct 14 09:52:03 crc kubenswrapper[4870]: I1014 09:52:03.756074 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="0417d7ee-a386-402e-bfa8-72b8dc5cf38e" containerName="neutron-dhcp-openstack-openstack-cell1" Oct 14 09:52:03 crc kubenswrapper[4870]: E1014 09:52:03.756085 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6c5fc210-e910-4b2a-b121-04ac3dfb28f3" containerName="registry-server" Oct 14 09:52:03 crc kubenswrapper[4870]: I1014 09:52:03.756091 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="6c5fc210-e910-4b2a-b121-04ac3dfb28f3" containerName="registry-server" Oct 14 09:52:03 crc kubenswrapper[4870]: E1014 09:52:03.756099 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6c5fc210-e910-4b2a-b121-04ac3dfb28f3" containerName="extract-content" Oct 14 09:52:03 crc kubenswrapper[4870]: I1014 09:52:03.756105 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="6c5fc210-e910-4b2a-b121-04ac3dfb28f3" containerName="extract-content" Oct 14 09:52:03 crc kubenswrapper[4870]: E1014 09:52:03.756119 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6c5fc210-e910-4b2a-b121-04ac3dfb28f3" containerName="extract-utilities" Oct 14 09:52:03 crc kubenswrapper[4870]: I1014 09:52:03.756126 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="6c5fc210-e910-4b2a-b121-04ac3dfb28f3" containerName="extract-utilities" Oct 14 09:52:03 crc kubenswrapper[4870]: E1014 09:52:03.756150 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="02ebb80d-7ee3-48e5-99e4-1d9e061d6e06" containerName="nova-cell1-conductor-conductor" Oct 14 09:52:03 crc kubenswrapper[4870]: I1014 09:52:03.756156 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="02ebb80d-7ee3-48e5-99e4-1d9e061d6e06" containerName="nova-cell1-conductor-conductor" Oct 14 09:52:03 crc kubenswrapper[4870]: I1014 09:52:03.756356 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="6c5fc210-e910-4b2a-b121-04ac3dfb28f3" containerName="registry-server" Oct 14 09:52:03 crc kubenswrapper[4870]: I1014 09:52:03.756366 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="02ebb80d-7ee3-48e5-99e4-1d9e061d6e06" containerName="nova-cell1-conductor-conductor" Oct 14 09:52:03 crc kubenswrapper[4870]: I1014 09:52:03.756381 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="0417d7ee-a386-402e-bfa8-72b8dc5cf38e" containerName="neutron-dhcp-openstack-openstack-cell1" Oct 14 09:52:03 crc kubenswrapper[4870]: I1014 09:52:03.757223 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 14 09:52:03 crc kubenswrapper[4870]: I1014 09:52:03.758811 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Oct 14 09:52:03 crc kubenswrapper[4870]: I1014 09:52:03.765500 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 14 09:52:03 crc kubenswrapper[4870]: I1014 09:52:03.827935 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bgbqk\" (UniqueName: \"kubernetes.io/projected/60d06fa5-6fbe-4958-8a2e-a1781c615a46-kube-api-access-bgbqk\") pod \"nova-cell1-conductor-0\" (UID: \"60d06fa5-6fbe-4958-8a2e-a1781c615a46\") " pod="openstack/nova-cell1-conductor-0" Oct 14 09:52:03 crc kubenswrapper[4870]: I1014 09:52:03.828021 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60d06fa5-6fbe-4958-8a2e-a1781c615a46-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"60d06fa5-6fbe-4958-8a2e-a1781c615a46\") " pod="openstack/nova-cell1-conductor-0" Oct 14 09:52:03 crc kubenswrapper[4870]: I1014 09:52:03.828068 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/60d06fa5-6fbe-4958-8a2e-a1781c615a46-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"60d06fa5-6fbe-4958-8a2e-a1781c615a46\") " pod="openstack/nova-cell1-conductor-0" Oct 14 09:52:03 crc kubenswrapper[4870]: I1014 09:52:03.930184 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60d06fa5-6fbe-4958-8a2e-a1781c615a46-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"60d06fa5-6fbe-4958-8a2e-a1781c615a46\") " pod="openstack/nova-cell1-conductor-0" Oct 14 09:52:03 crc kubenswrapper[4870]: I1014 09:52:03.930257 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/60d06fa5-6fbe-4958-8a2e-a1781c615a46-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"60d06fa5-6fbe-4958-8a2e-a1781c615a46\") " pod="openstack/nova-cell1-conductor-0" Oct 14 09:52:03 crc kubenswrapper[4870]: I1014 09:52:03.930402 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bgbqk\" (UniqueName: \"kubernetes.io/projected/60d06fa5-6fbe-4958-8a2e-a1781c615a46-kube-api-access-bgbqk\") pod \"nova-cell1-conductor-0\" (UID: \"60d06fa5-6fbe-4958-8a2e-a1781c615a46\") " pod="openstack/nova-cell1-conductor-0" Oct 14 09:52:03 crc kubenswrapper[4870]: I1014 09:52:03.936046 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60d06fa5-6fbe-4958-8a2e-a1781c615a46-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"60d06fa5-6fbe-4958-8a2e-a1781c615a46\") " pod="openstack/nova-cell1-conductor-0" Oct 14 09:52:03 crc kubenswrapper[4870]: I1014 09:52:03.946424 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/60d06fa5-6fbe-4958-8a2e-a1781c615a46-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"60d06fa5-6fbe-4958-8a2e-a1781c615a46\") " pod="openstack/nova-cell1-conductor-0" Oct 14 09:52:03 crc kubenswrapper[4870]: I1014 09:52:03.951539 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bgbqk\" (UniqueName: \"kubernetes.io/projected/60d06fa5-6fbe-4958-8a2e-a1781c615a46-kube-api-access-bgbqk\") pod \"nova-cell1-conductor-0\" (UID: \"60d06fa5-6fbe-4958-8a2e-a1781c615a46\") " pod="openstack/nova-cell1-conductor-0" Oct 14 09:52:04 crc kubenswrapper[4870]: I1014 09:52:04.146311 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 14 09:52:04 crc kubenswrapper[4870]: I1014 09:52:04.702143 4870 generic.go:334] "Generic (PLEG): container finished" podID="73281705-4198-40a4-b523-c0bfc322c833" containerID="c0020e7320b47c052dc0efb0f61570b5a39b13c7d39e5b5d0bc9ebec23c7b744" exitCode=0 Oct 14 09:52:04 crc kubenswrapper[4870]: I1014 09:52:04.702198 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"73281705-4198-40a4-b523-c0bfc322c833","Type":"ContainerDied","Data":"c0020e7320b47c052dc0efb0f61570b5a39b13c7d39e5b5d0bc9ebec23c7b744"} Oct 14 09:52:04 crc kubenswrapper[4870]: I1014 09:52:04.702231 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"73281705-4198-40a4-b523-c0bfc322c833","Type":"ContainerDied","Data":"fef838a7c02ea978f46e98ee10f6903ba186b375270b82222309e4108eb295ce"} Oct 14 09:52:04 crc kubenswrapper[4870]: I1014 09:52:04.702247 4870 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fef838a7c02ea978f46e98ee10f6903ba186b375270b82222309e4108eb295ce" Oct 14 09:52:04 crc kubenswrapper[4870]: E1014 09:52:04.745147 4870 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of c0020e7320b47c052dc0efb0f61570b5a39b13c7d39e5b5d0bc9ebec23c7b744 is running failed: container process not found" containerID="c0020e7320b47c052dc0efb0f61570b5a39b13c7d39e5b5d0bc9ebec23c7b744" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 14 09:52:04 crc kubenswrapper[4870]: E1014 09:52:04.745608 4870 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of c0020e7320b47c052dc0efb0f61570b5a39b13c7d39e5b5d0bc9ebec23c7b744 is running failed: container process not found" containerID="c0020e7320b47c052dc0efb0f61570b5a39b13c7d39e5b5d0bc9ebec23c7b744" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 14 09:52:04 crc kubenswrapper[4870]: E1014 09:52:04.745974 4870 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of c0020e7320b47c052dc0efb0f61570b5a39b13c7d39e5b5d0bc9ebec23c7b744 is running failed: container process not found" containerID="c0020e7320b47c052dc0efb0f61570b5a39b13c7d39e5b5d0bc9ebec23c7b744" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 14 09:52:04 crc kubenswrapper[4870]: E1014 09:52:04.746042 4870 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of c0020e7320b47c052dc0efb0f61570b5a39b13c7d39e5b5d0bc9ebec23c7b744 is running failed: container process not found" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="73281705-4198-40a4-b523-c0bfc322c833" containerName="nova-scheduler-scheduler" Oct 14 09:52:05 crc kubenswrapper[4870]: I1014 09:52:05.012078 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 14 09:52:05 crc kubenswrapper[4870]: I1014 09:52:05.055473 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/73281705-4198-40a4-b523-c0bfc322c833-combined-ca-bundle\") pod \"73281705-4198-40a4-b523-c0bfc322c833\" (UID: \"73281705-4198-40a4-b523-c0bfc322c833\") " Oct 14 09:52:05 crc kubenswrapper[4870]: I1014 09:52:05.055667 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/73281705-4198-40a4-b523-c0bfc322c833-config-data\") pod \"73281705-4198-40a4-b523-c0bfc322c833\" (UID: \"73281705-4198-40a4-b523-c0bfc322c833\") " Oct 14 09:52:05 crc kubenswrapper[4870]: I1014 09:52:05.055708 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vscgf\" (UniqueName: \"kubernetes.io/projected/73281705-4198-40a4-b523-c0bfc322c833-kube-api-access-vscgf\") pod \"73281705-4198-40a4-b523-c0bfc322c833\" (UID: \"73281705-4198-40a4-b523-c0bfc322c833\") " Oct 14 09:52:05 crc kubenswrapper[4870]: I1014 09:52:05.064858 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/73281705-4198-40a4-b523-c0bfc322c833-kube-api-access-vscgf" (OuterVolumeSpecName: "kube-api-access-vscgf") pod "73281705-4198-40a4-b523-c0bfc322c833" (UID: "73281705-4198-40a4-b523-c0bfc322c833"). InnerVolumeSpecName "kube-api-access-vscgf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:52:05 crc kubenswrapper[4870]: I1014 09:52:05.064961 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="02ebb80d-7ee3-48e5-99e4-1d9e061d6e06" path="/var/lib/kubelet/pods/02ebb80d-7ee3-48e5-99e4-1d9e061d6e06/volumes" Oct 14 09:52:05 crc kubenswrapper[4870]: I1014 09:52:05.085943 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/73281705-4198-40a4-b523-c0bfc322c833-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "73281705-4198-40a4-b523-c0bfc322c833" (UID: "73281705-4198-40a4-b523-c0bfc322c833"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:52:05 crc kubenswrapper[4870]: I1014 09:52:05.086619 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/73281705-4198-40a4-b523-c0bfc322c833-config-data" (OuterVolumeSpecName: "config-data") pod "73281705-4198-40a4-b523-c0bfc322c833" (UID: "73281705-4198-40a4-b523-c0bfc322c833"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:52:05 crc kubenswrapper[4870]: I1014 09:52:05.157987 4870 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/73281705-4198-40a4-b523-c0bfc322c833-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 09:52:05 crc kubenswrapper[4870]: I1014 09:52:05.159578 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vscgf\" (UniqueName: \"kubernetes.io/projected/73281705-4198-40a4-b523-c0bfc322c833-kube-api-access-vscgf\") on node \"crc\" DevicePath \"\"" Oct 14 09:52:05 crc kubenswrapper[4870]: I1014 09:52:05.159647 4870 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/73281705-4198-40a4-b523-c0bfc322c833-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 09:52:05 crc kubenswrapper[4870]: I1014 09:52:05.367242 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 14 09:52:05 crc kubenswrapper[4870]: I1014 09:52:05.713358 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 14 09:52:05 crc kubenswrapper[4870]: I1014 09:52:05.713367 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"60d06fa5-6fbe-4958-8a2e-a1781c615a46","Type":"ContainerStarted","Data":"b75c9549208a197cbd5f2663a2ced537e1903db44f3e34af693d1de343186c36"} Oct 14 09:52:05 crc kubenswrapper[4870]: I1014 09:52:05.713416 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"60d06fa5-6fbe-4958-8a2e-a1781c615a46","Type":"ContainerStarted","Data":"67c3e636b66d31c25e259db675161c8ce7b495e332478f0c36f5360b1664526f"} Oct 14 09:52:05 crc kubenswrapper[4870]: I1014 09:52:05.754913 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.754896134 podStartE2EDuration="2.754896134s" podCreationTimestamp="2025-10-14 09:52:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 09:52:05.735033585 +0000 UTC m=+10261.432393966" watchObservedRunningTime="2025-10-14 09:52:05.754896134 +0000 UTC m=+10261.452256505" Oct 14 09:52:05 crc kubenswrapper[4870]: I1014 09:52:05.788524 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 14 09:52:05 crc kubenswrapper[4870]: I1014 09:52:05.798696 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Oct 14 09:52:05 crc kubenswrapper[4870]: I1014 09:52:05.811344 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 14 09:52:05 crc kubenswrapper[4870]: E1014 09:52:05.811793 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73281705-4198-40a4-b523-c0bfc322c833" containerName="nova-scheduler-scheduler" Oct 14 09:52:05 crc kubenswrapper[4870]: I1014 09:52:05.811810 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="73281705-4198-40a4-b523-c0bfc322c833" containerName="nova-scheduler-scheduler" Oct 14 09:52:05 crc kubenswrapper[4870]: I1014 09:52:05.812042 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="73281705-4198-40a4-b523-c0bfc322c833" containerName="nova-scheduler-scheduler" Oct 14 09:52:05 crc kubenswrapper[4870]: I1014 09:52:05.812772 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 14 09:52:05 crc kubenswrapper[4870]: I1014 09:52:05.815504 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 14 09:52:05 crc kubenswrapper[4870]: I1014 09:52:05.824026 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 14 09:52:05 crc kubenswrapper[4870]: I1014 09:52:05.882950 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dee94958-2c48-4715-964b-ebd6efe1f95e-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"dee94958-2c48-4715-964b-ebd6efe1f95e\") " pod="openstack/nova-scheduler-0" Oct 14 09:52:05 crc kubenswrapper[4870]: I1014 09:52:05.883016 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mhvtx\" (UniqueName: \"kubernetes.io/projected/dee94958-2c48-4715-964b-ebd6efe1f95e-kube-api-access-mhvtx\") pod \"nova-scheduler-0\" (UID: \"dee94958-2c48-4715-964b-ebd6efe1f95e\") " pod="openstack/nova-scheduler-0" Oct 14 09:52:05 crc kubenswrapper[4870]: I1014 09:52:05.883178 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dee94958-2c48-4715-964b-ebd6efe1f95e-config-data\") pod \"nova-scheduler-0\" (UID: \"dee94958-2c48-4715-964b-ebd6efe1f95e\") " pod="openstack/nova-scheduler-0" Oct 14 09:52:05 crc kubenswrapper[4870]: I1014 09:52:05.985425 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dee94958-2c48-4715-964b-ebd6efe1f95e-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"dee94958-2c48-4715-964b-ebd6efe1f95e\") " pod="openstack/nova-scheduler-0" Oct 14 09:52:05 crc kubenswrapper[4870]: I1014 09:52:05.985485 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mhvtx\" (UniqueName: \"kubernetes.io/projected/dee94958-2c48-4715-964b-ebd6efe1f95e-kube-api-access-mhvtx\") pod \"nova-scheduler-0\" (UID: \"dee94958-2c48-4715-964b-ebd6efe1f95e\") " pod="openstack/nova-scheduler-0" Oct 14 09:52:05 crc kubenswrapper[4870]: I1014 09:52:05.985567 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dee94958-2c48-4715-964b-ebd6efe1f95e-config-data\") pod \"nova-scheduler-0\" (UID: \"dee94958-2c48-4715-964b-ebd6efe1f95e\") " pod="openstack/nova-scheduler-0" Oct 14 09:52:05 crc kubenswrapper[4870]: I1014 09:52:05.989560 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dee94958-2c48-4715-964b-ebd6efe1f95e-config-data\") pod \"nova-scheduler-0\" (UID: \"dee94958-2c48-4715-964b-ebd6efe1f95e\") " pod="openstack/nova-scheduler-0" Oct 14 09:52:05 crc kubenswrapper[4870]: I1014 09:52:05.989951 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dee94958-2c48-4715-964b-ebd6efe1f95e-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"dee94958-2c48-4715-964b-ebd6efe1f95e\") " pod="openstack/nova-scheduler-0" Oct 14 09:52:06 crc kubenswrapper[4870]: I1014 09:52:06.001728 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mhvtx\" (UniqueName: \"kubernetes.io/projected/dee94958-2c48-4715-964b-ebd6efe1f95e-kube-api-access-mhvtx\") pod \"nova-scheduler-0\" (UID: \"dee94958-2c48-4715-964b-ebd6efe1f95e\") " pod="openstack/nova-scheduler-0" Oct 14 09:52:06 crc kubenswrapper[4870]: I1014 09:52:06.130293 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 14 09:52:06 crc kubenswrapper[4870]: I1014 09:52:06.455384 4870 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="17f1a688-1bd1-4b88-ad00-51caf1af0ce0" containerName="nova-metadata-log" probeResult="failure" output="Get \"http://10.217.1.94:8775/\": read tcp 10.217.0.2:60896->10.217.1.94:8775: read: connection reset by peer" Oct 14 09:52:06 crc kubenswrapper[4870]: I1014 09:52:06.455449 4870 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="17f1a688-1bd1-4b88-ad00-51caf1af0ce0" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"http://10.217.1.94:8775/\": read tcp 10.217.0.2:60912->10.217.1.94:8775: read: connection reset by peer" Oct 14 09:52:06 crc kubenswrapper[4870]: I1014 09:52:06.716924 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 14 09:52:06 crc kubenswrapper[4870]: I1014 09:52:06.724858 4870 generic.go:334] "Generic (PLEG): container finished" podID="e1055bc1-24f3-4893-ad77-c520a36428bb" containerID="36057428133bd01d5a6c0e75037e7dff3a30c1c00aa184489e5fee2e48ca9c66" exitCode=0 Oct 14 09:52:06 crc kubenswrapper[4870]: I1014 09:52:06.725528 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e1055bc1-24f3-4893-ad77-c520a36428bb","Type":"ContainerDied","Data":"36057428133bd01d5a6c0e75037e7dff3a30c1c00aa184489e5fee2e48ca9c66"} Oct 14 09:52:06 crc kubenswrapper[4870]: I1014 09:52:06.725701 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Oct 14 09:52:06 crc kubenswrapper[4870]: W1014 09:52:06.853542 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddee94958_2c48_4715_964b_ebd6efe1f95e.slice/crio-8a66311642f9b6ed85949fcd964bc4f1e252c49c4d6215865120d950ca4cd4e1 WatchSource:0}: Error finding container 8a66311642f9b6ed85949fcd964bc4f1e252c49c4d6215865120d950ca4cd4e1: Status 404 returned error can't find the container with id 8a66311642f9b6ed85949fcd964bc4f1e252c49c4d6215865120d950ca4cd4e1 Oct 14 09:52:07 crc kubenswrapper[4870]: I1014 09:52:07.049199 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="73281705-4198-40a4-b523-c0bfc322c833" path="/var/lib/kubelet/pods/73281705-4198-40a4-b523-c0bfc322c833/volumes" Oct 14 09:52:07 crc kubenswrapper[4870]: I1014 09:52:07.181846 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 14 09:52:07 crc kubenswrapper[4870]: I1014 09:52:07.227542 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e1055bc1-24f3-4893-ad77-c520a36428bb-combined-ca-bundle\") pod \"e1055bc1-24f3-4893-ad77-c520a36428bb\" (UID: \"e1055bc1-24f3-4893-ad77-c520a36428bb\") " Oct 14 09:52:07 crc kubenswrapper[4870]: I1014 09:52:07.227858 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tzv26\" (UniqueName: \"kubernetes.io/projected/e1055bc1-24f3-4893-ad77-c520a36428bb-kube-api-access-tzv26\") pod \"e1055bc1-24f3-4893-ad77-c520a36428bb\" (UID: \"e1055bc1-24f3-4893-ad77-c520a36428bb\") " Oct 14 09:52:07 crc kubenswrapper[4870]: I1014 09:52:07.227906 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e1055bc1-24f3-4893-ad77-c520a36428bb-logs\") pod \"e1055bc1-24f3-4893-ad77-c520a36428bb\" (UID: \"e1055bc1-24f3-4893-ad77-c520a36428bb\") " Oct 14 09:52:07 crc kubenswrapper[4870]: I1014 09:52:07.228037 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e1055bc1-24f3-4893-ad77-c520a36428bb-config-data\") pod \"e1055bc1-24f3-4893-ad77-c520a36428bb\" (UID: \"e1055bc1-24f3-4893-ad77-c520a36428bb\") " Oct 14 09:52:07 crc kubenswrapper[4870]: I1014 09:52:07.228613 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e1055bc1-24f3-4893-ad77-c520a36428bb-logs" (OuterVolumeSpecName: "logs") pod "e1055bc1-24f3-4893-ad77-c520a36428bb" (UID: "e1055bc1-24f3-4893-ad77-c520a36428bb"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 09:52:07 crc kubenswrapper[4870]: I1014 09:52:07.245224 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e1055bc1-24f3-4893-ad77-c520a36428bb-kube-api-access-tzv26" (OuterVolumeSpecName: "kube-api-access-tzv26") pod "e1055bc1-24f3-4893-ad77-c520a36428bb" (UID: "e1055bc1-24f3-4893-ad77-c520a36428bb"). InnerVolumeSpecName "kube-api-access-tzv26". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:52:07 crc kubenswrapper[4870]: I1014 09:52:07.292886 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e1055bc1-24f3-4893-ad77-c520a36428bb-config-data" (OuterVolumeSpecName: "config-data") pod "e1055bc1-24f3-4893-ad77-c520a36428bb" (UID: "e1055bc1-24f3-4893-ad77-c520a36428bb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:52:07 crc kubenswrapper[4870]: I1014 09:52:07.300687 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e1055bc1-24f3-4893-ad77-c520a36428bb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e1055bc1-24f3-4893-ad77-c520a36428bb" (UID: "e1055bc1-24f3-4893-ad77-c520a36428bb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:52:07 crc kubenswrapper[4870]: I1014 09:52:07.330367 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tzv26\" (UniqueName: \"kubernetes.io/projected/e1055bc1-24f3-4893-ad77-c520a36428bb-kube-api-access-tzv26\") on node \"crc\" DevicePath \"\"" Oct 14 09:52:07 crc kubenswrapper[4870]: I1014 09:52:07.330413 4870 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e1055bc1-24f3-4893-ad77-c520a36428bb-logs\") on node \"crc\" DevicePath \"\"" Oct 14 09:52:07 crc kubenswrapper[4870]: I1014 09:52:07.330428 4870 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e1055bc1-24f3-4893-ad77-c520a36428bb-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 09:52:07 crc kubenswrapper[4870]: I1014 09:52:07.330442 4870 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e1055bc1-24f3-4893-ad77-c520a36428bb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 09:52:07 crc kubenswrapper[4870]: I1014 09:52:07.407490 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 14 09:52:07 crc kubenswrapper[4870]: I1014 09:52:07.488523 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 14 09:52:07 crc kubenswrapper[4870]: I1014 09:52:07.533792 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b6p95\" (UniqueName: \"kubernetes.io/projected/17f1a688-1bd1-4b88-ad00-51caf1af0ce0-kube-api-access-b6p95\") pod \"17f1a688-1bd1-4b88-ad00-51caf1af0ce0\" (UID: \"17f1a688-1bd1-4b88-ad00-51caf1af0ce0\") " Oct 14 09:52:07 crc kubenswrapper[4870]: I1014 09:52:07.534125 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/17f1a688-1bd1-4b88-ad00-51caf1af0ce0-config-data\") pod \"17f1a688-1bd1-4b88-ad00-51caf1af0ce0\" (UID: \"17f1a688-1bd1-4b88-ad00-51caf1af0ce0\") " Oct 14 09:52:07 crc kubenswrapper[4870]: I1014 09:52:07.534223 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2lf8\" (UniqueName: \"kubernetes.io/projected/ee5dcbc8-b98e-48b9-a089-a9f5614249d3-kube-api-access-x2lf8\") pod \"ee5dcbc8-b98e-48b9-a089-a9f5614249d3\" (UID: \"ee5dcbc8-b98e-48b9-a089-a9f5614249d3\") " Oct 14 09:52:07 crc kubenswrapper[4870]: I1014 09:52:07.534361 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/17f1a688-1bd1-4b88-ad00-51caf1af0ce0-logs\") pod \"17f1a688-1bd1-4b88-ad00-51caf1af0ce0\" (UID: \"17f1a688-1bd1-4b88-ad00-51caf1af0ce0\") " Oct 14 09:52:07 crc kubenswrapper[4870]: I1014 09:52:07.534542 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee5dcbc8-b98e-48b9-a089-a9f5614249d3-combined-ca-bundle\") pod \"ee5dcbc8-b98e-48b9-a089-a9f5614249d3\" (UID: \"ee5dcbc8-b98e-48b9-a089-a9f5614249d3\") " Oct 14 09:52:07 crc kubenswrapper[4870]: I1014 09:52:07.534641 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17f1a688-1bd1-4b88-ad00-51caf1af0ce0-combined-ca-bundle\") pod \"17f1a688-1bd1-4b88-ad00-51caf1af0ce0\" (UID: \"17f1a688-1bd1-4b88-ad00-51caf1af0ce0\") " Oct 14 09:52:07 crc kubenswrapper[4870]: I1014 09:52:07.534798 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ee5dcbc8-b98e-48b9-a089-a9f5614249d3-config-data\") pod \"ee5dcbc8-b98e-48b9-a089-a9f5614249d3\" (UID: \"ee5dcbc8-b98e-48b9-a089-a9f5614249d3\") " Oct 14 09:52:07 crc kubenswrapper[4870]: I1014 09:52:07.536564 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/17f1a688-1bd1-4b88-ad00-51caf1af0ce0-kube-api-access-b6p95" (OuterVolumeSpecName: "kube-api-access-b6p95") pod "17f1a688-1bd1-4b88-ad00-51caf1af0ce0" (UID: "17f1a688-1bd1-4b88-ad00-51caf1af0ce0"). InnerVolumeSpecName "kube-api-access-b6p95". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:52:07 crc kubenswrapper[4870]: I1014 09:52:07.536904 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/17f1a688-1bd1-4b88-ad00-51caf1af0ce0-logs" (OuterVolumeSpecName: "logs") pod "17f1a688-1bd1-4b88-ad00-51caf1af0ce0" (UID: "17f1a688-1bd1-4b88-ad00-51caf1af0ce0"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 09:52:07 crc kubenswrapper[4870]: I1014 09:52:07.553618 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ee5dcbc8-b98e-48b9-a089-a9f5614249d3-kube-api-access-x2lf8" (OuterVolumeSpecName: "kube-api-access-x2lf8") pod "ee5dcbc8-b98e-48b9-a089-a9f5614249d3" (UID: "ee5dcbc8-b98e-48b9-a089-a9f5614249d3"). InnerVolumeSpecName "kube-api-access-x2lf8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:52:07 crc kubenswrapper[4870]: I1014 09:52:07.573610 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/17f1a688-1bd1-4b88-ad00-51caf1af0ce0-config-data" (OuterVolumeSpecName: "config-data") pod "17f1a688-1bd1-4b88-ad00-51caf1af0ce0" (UID: "17f1a688-1bd1-4b88-ad00-51caf1af0ce0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:52:07 crc kubenswrapper[4870]: I1014 09:52:07.577282 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee5dcbc8-b98e-48b9-a089-a9f5614249d3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ee5dcbc8-b98e-48b9-a089-a9f5614249d3" (UID: "ee5dcbc8-b98e-48b9-a089-a9f5614249d3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:52:07 crc kubenswrapper[4870]: I1014 09:52:07.577468 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee5dcbc8-b98e-48b9-a089-a9f5614249d3-config-data" (OuterVolumeSpecName: "config-data") pod "ee5dcbc8-b98e-48b9-a089-a9f5614249d3" (UID: "ee5dcbc8-b98e-48b9-a089-a9f5614249d3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:52:07 crc kubenswrapper[4870]: I1014 09:52:07.619640 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/17f1a688-1bd1-4b88-ad00-51caf1af0ce0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "17f1a688-1bd1-4b88-ad00-51caf1af0ce0" (UID: "17f1a688-1bd1-4b88-ad00-51caf1af0ce0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:52:07 crc kubenswrapper[4870]: I1014 09:52:07.637152 4870 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee5dcbc8-b98e-48b9-a089-a9f5614249d3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 09:52:07 crc kubenswrapper[4870]: I1014 09:52:07.637420 4870 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/17f1a688-1bd1-4b88-ad00-51caf1af0ce0-logs\") on node \"crc\" DevicePath \"\"" Oct 14 09:52:07 crc kubenswrapper[4870]: I1014 09:52:07.637500 4870 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17f1a688-1bd1-4b88-ad00-51caf1af0ce0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 09:52:07 crc kubenswrapper[4870]: I1014 09:52:07.637575 4870 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ee5dcbc8-b98e-48b9-a089-a9f5614249d3-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 09:52:07 crc kubenswrapper[4870]: I1014 09:52:07.637651 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b6p95\" (UniqueName: \"kubernetes.io/projected/17f1a688-1bd1-4b88-ad00-51caf1af0ce0-kube-api-access-b6p95\") on node \"crc\" DevicePath \"\"" Oct 14 09:52:07 crc kubenswrapper[4870]: I1014 09:52:07.637753 4870 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/17f1a688-1bd1-4b88-ad00-51caf1af0ce0-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 09:52:07 crc kubenswrapper[4870]: I1014 09:52:07.637815 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2lf8\" (UniqueName: \"kubernetes.io/projected/ee5dcbc8-b98e-48b9-a089-a9f5614249d3-kube-api-access-x2lf8\") on node \"crc\" DevicePath \"\"" Oct 14 09:52:07 crc kubenswrapper[4870]: I1014 09:52:07.761420 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e1055bc1-24f3-4893-ad77-c520a36428bb","Type":"ContainerDied","Data":"2e6295cdaa37c572684a9ab5e125e12fef5dbe0be1f1d9f3c30c40b8c4a44e04"} Oct 14 09:52:07 crc kubenswrapper[4870]: I1014 09:52:07.761696 4870 scope.go:117] "RemoveContainer" containerID="36057428133bd01d5a6c0e75037e7dff3a30c1c00aa184489e5fee2e48ca9c66" Oct 14 09:52:07 crc kubenswrapper[4870]: I1014 09:52:07.761899 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 14 09:52:07 crc kubenswrapper[4870]: I1014 09:52:07.773652 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"dee94958-2c48-4715-964b-ebd6efe1f95e","Type":"ContainerStarted","Data":"7a5c0c079c81a76643e6a50372e2aa3c93f662d50364fb9c25db8d54891644b6"} Oct 14 09:52:07 crc kubenswrapper[4870]: I1014 09:52:07.773702 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"dee94958-2c48-4715-964b-ebd6efe1f95e","Type":"ContainerStarted","Data":"8a66311642f9b6ed85949fcd964bc4f1e252c49c4d6215865120d950ca4cd4e1"} Oct 14 09:52:07 crc kubenswrapper[4870]: I1014 09:52:07.795742 4870 generic.go:334] "Generic (PLEG): container finished" podID="17f1a688-1bd1-4b88-ad00-51caf1af0ce0" containerID="1bc76a1a314ba9d308a310397f670dc886227d98b6f2d40ae039086e08114b2d" exitCode=0 Oct 14 09:52:07 crc kubenswrapper[4870]: I1014 09:52:07.795867 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"17f1a688-1bd1-4b88-ad00-51caf1af0ce0","Type":"ContainerDied","Data":"1bc76a1a314ba9d308a310397f670dc886227d98b6f2d40ae039086e08114b2d"} Oct 14 09:52:07 crc kubenswrapper[4870]: I1014 09:52:07.795907 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"17f1a688-1bd1-4b88-ad00-51caf1af0ce0","Type":"ContainerDied","Data":"4445c716ef7f432ea4d5783ad7075d6d63bcdc2920320f42fcc9bb29df12107a"} Oct 14 09:52:07 crc kubenswrapper[4870]: I1014 09:52:07.795997 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 14 09:52:07 crc kubenswrapper[4870]: I1014 09:52:07.802088 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.802069398 podStartE2EDuration="2.802069398s" podCreationTimestamp="2025-10-14 09:52:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 09:52:07.801196227 +0000 UTC m=+10263.498556598" watchObservedRunningTime="2025-10-14 09:52:07.802069398 +0000 UTC m=+10263.499429769" Oct 14 09:52:07 crc kubenswrapper[4870]: I1014 09:52:07.810497 4870 generic.go:334] "Generic (PLEG): container finished" podID="ee5dcbc8-b98e-48b9-a089-a9f5614249d3" containerID="910ed7bad34d9a7163b841990638ec4c809ae09c71d936618c95878703298b7e" exitCode=0 Oct 14 09:52:07 crc kubenswrapper[4870]: I1014 09:52:07.811336 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 14 09:52:07 crc kubenswrapper[4870]: I1014 09:52:07.817811 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"ee5dcbc8-b98e-48b9-a089-a9f5614249d3","Type":"ContainerDied","Data":"910ed7bad34d9a7163b841990638ec4c809ae09c71d936618c95878703298b7e"} Oct 14 09:52:07 crc kubenswrapper[4870]: I1014 09:52:07.817860 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"ee5dcbc8-b98e-48b9-a089-a9f5614249d3","Type":"ContainerDied","Data":"a7fc468608113d657aebbc527900ae007bc57e5d1082a93586d12cda783ca7a6"} Oct 14 09:52:07 crc kubenswrapper[4870]: I1014 09:52:07.842620 4870 scope.go:117] "RemoveContainer" containerID="f61fa490fab929fb5e534efb5255c5f168fdc3b95da9665beae2c6c2d3ba5803" Oct 14 09:52:07 crc kubenswrapper[4870]: I1014 09:52:07.883616 4870 scope.go:117] "RemoveContainer" containerID="1bc76a1a314ba9d308a310397f670dc886227d98b6f2d40ae039086e08114b2d" Oct 14 09:52:07 crc kubenswrapper[4870]: I1014 09:52:07.884159 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 14 09:52:07 crc kubenswrapper[4870]: I1014 09:52:07.907953 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 14 09:52:07 crc kubenswrapper[4870]: I1014 09:52:07.920658 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 14 09:52:07 crc kubenswrapper[4870]: E1014 09:52:07.921330 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1055bc1-24f3-4893-ad77-c520a36428bb" containerName="nova-api-api" Oct 14 09:52:07 crc kubenswrapper[4870]: I1014 09:52:07.921346 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1055bc1-24f3-4893-ad77-c520a36428bb" containerName="nova-api-api" Oct 14 09:52:07 crc kubenswrapper[4870]: E1014 09:52:07.921370 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee5dcbc8-b98e-48b9-a089-a9f5614249d3" containerName="nova-cell0-conductor-conductor" Oct 14 09:52:07 crc kubenswrapper[4870]: I1014 09:52:07.921379 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee5dcbc8-b98e-48b9-a089-a9f5614249d3" containerName="nova-cell0-conductor-conductor" Oct 14 09:52:07 crc kubenswrapper[4870]: E1014 09:52:07.921392 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17f1a688-1bd1-4b88-ad00-51caf1af0ce0" containerName="nova-metadata-log" Oct 14 09:52:07 crc kubenswrapper[4870]: I1014 09:52:07.921399 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="17f1a688-1bd1-4b88-ad00-51caf1af0ce0" containerName="nova-metadata-log" Oct 14 09:52:07 crc kubenswrapper[4870]: E1014 09:52:07.921409 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17f1a688-1bd1-4b88-ad00-51caf1af0ce0" containerName="nova-metadata-metadata" Oct 14 09:52:07 crc kubenswrapper[4870]: I1014 09:52:07.921414 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="17f1a688-1bd1-4b88-ad00-51caf1af0ce0" containerName="nova-metadata-metadata" Oct 14 09:52:07 crc kubenswrapper[4870]: E1014 09:52:07.921447 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1055bc1-24f3-4893-ad77-c520a36428bb" containerName="nova-api-log" Oct 14 09:52:07 crc kubenswrapper[4870]: I1014 09:52:07.921650 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1055bc1-24f3-4893-ad77-c520a36428bb" containerName="nova-api-log" Oct 14 09:52:07 crc kubenswrapper[4870]: I1014 09:52:07.921845 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="17f1a688-1bd1-4b88-ad00-51caf1af0ce0" containerName="nova-metadata-metadata" Oct 14 09:52:07 crc kubenswrapper[4870]: I1014 09:52:07.921868 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee5dcbc8-b98e-48b9-a089-a9f5614249d3" containerName="nova-cell0-conductor-conductor" Oct 14 09:52:07 crc kubenswrapper[4870]: I1014 09:52:07.921880 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="e1055bc1-24f3-4893-ad77-c520a36428bb" containerName="nova-api-log" Oct 14 09:52:07 crc kubenswrapper[4870]: I1014 09:52:07.921891 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="e1055bc1-24f3-4893-ad77-c520a36428bb" containerName="nova-api-api" Oct 14 09:52:07 crc kubenswrapper[4870]: I1014 09:52:07.921901 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="17f1a688-1bd1-4b88-ad00-51caf1af0ce0" containerName="nova-metadata-log" Oct 14 09:52:07 crc kubenswrapper[4870]: I1014 09:52:07.923193 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 14 09:52:07 crc kubenswrapper[4870]: I1014 09:52:07.926492 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 14 09:52:07 crc kubenswrapper[4870]: I1014 09:52:07.928997 4870 scope.go:117] "RemoveContainer" containerID="990d26c8691a9651642a973ab6b25f404d460d767a6e4471a24b9aaf9d964606" Oct 14 09:52:07 crc kubenswrapper[4870]: I1014 09:52:07.933031 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 14 09:52:07 crc kubenswrapper[4870]: I1014 09:52:07.947253 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 14 09:52:07 crc kubenswrapper[4870]: I1014 09:52:07.973519 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 14 09:52:07 crc kubenswrapper[4870]: I1014 09:52:07.980776 4870 scope.go:117] "RemoveContainer" containerID="1bc76a1a314ba9d308a310397f670dc886227d98b6f2d40ae039086e08114b2d" Oct 14 09:52:07 crc kubenswrapper[4870]: E1014 09:52:07.983357 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1bc76a1a314ba9d308a310397f670dc886227d98b6f2d40ae039086e08114b2d\": container with ID starting with 1bc76a1a314ba9d308a310397f670dc886227d98b6f2d40ae039086e08114b2d not found: ID does not exist" containerID="1bc76a1a314ba9d308a310397f670dc886227d98b6f2d40ae039086e08114b2d" Oct 14 09:52:07 crc kubenswrapper[4870]: I1014 09:52:07.983396 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1bc76a1a314ba9d308a310397f670dc886227d98b6f2d40ae039086e08114b2d"} err="failed to get container status \"1bc76a1a314ba9d308a310397f670dc886227d98b6f2d40ae039086e08114b2d\": rpc error: code = NotFound desc = could not find container \"1bc76a1a314ba9d308a310397f670dc886227d98b6f2d40ae039086e08114b2d\": container with ID starting with 1bc76a1a314ba9d308a310397f670dc886227d98b6f2d40ae039086e08114b2d not found: ID does not exist" Oct 14 09:52:07 crc kubenswrapper[4870]: I1014 09:52:07.983422 4870 scope.go:117] "RemoveContainer" containerID="990d26c8691a9651642a973ab6b25f404d460d767a6e4471a24b9aaf9d964606" Oct 14 09:52:07 crc kubenswrapper[4870]: E1014 09:52:07.983768 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"990d26c8691a9651642a973ab6b25f404d460d767a6e4471a24b9aaf9d964606\": container with ID starting with 990d26c8691a9651642a973ab6b25f404d460d767a6e4471a24b9aaf9d964606 not found: ID does not exist" containerID="990d26c8691a9651642a973ab6b25f404d460d767a6e4471a24b9aaf9d964606" Oct 14 09:52:07 crc kubenswrapper[4870]: I1014 09:52:07.983801 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"990d26c8691a9651642a973ab6b25f404d460d767a6e4471a24b9aaf9d964606"} err="failed to get container status \"990d26c8691a9651642a973ab6b25f404d460d767a6e4471a24b9aaf9d964606\": rpc error: code = NotFound desc = could not find container \"990d26c8691a9651642a973ab6b25f404d460d767a6e4471a24b9aaf9d964606\": container with ID starting with 990d26c8691a9651642a973ab6b25f404d460d767a6e4471a24b9aaf9d964606 not found: ID does not exist" Oct 14 09:52:07 crc kubenswrapper[4870]: I1014 09:52:07.983815 4870 scope.go:117] "RemoveContainer" containerID="910ed7bad34d9a7163b841990638ec4c809ae09c71d936618c95878703298b7e" Oct 14 09:52:07 crc kubenswrapper[4870]: I1014 09:52:07.984748 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 14 09:52:07 crc kubenswrapper[4870]: I1014 09:52:07.996283 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 14 09:52:08 crc kubenswrapper[4870]: I1014 09:52:08.003953 4870 scope.go:117] "RemoveContainer" containerID="910ed7bad34d9a7163b841990638ec4c809ae09c71d936618c95878703298b7e" Oct 14 09:52:08 crc kubenswrapper[4870]: E1014 09:52:08.004859 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"910ed7bad34d9a7163b841990638ec4c809ae09c71d936618c95878703298b7e\": container with ID starting with 910ed7bad34d9a7163b841990638ec4c809ae09c71d936618c95878703298b7e not found: ID does not exist" containerID="910ed7bad34d9a7163b841990638ec4c809ae09c71d936618c95878703298b7e" Oct 14 09:52:08 crc kubenswrapper[4870]: I1014 09:52:08.004890 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"910ed7bad34d9a7163b841990638ec4c809ae09c71d936618c95878703298b7e"} err="failed to get container status \"910ed7bad34d9a7163b841990638ec4c809ae09c71d936618c95878703298b7e\": rpc error: code = NotFound desc = could not find container \"910ed7bad34d9a7163b841990638ec4c809ae09c71d936618c95878703298b7e\": container with ID starting with 910ed7bad34d9a7163b841990638ec4c809ae09c71d936618c95878703298b7e not found: ID does not exist" Oct 14 09:52:08 crc kubenswrapper[4870]: I1014 09:52:08.005245 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 14 09:52:08 crc kubenswrapper[4870]: I1014 09:52:08.007020 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 14 09:52:08 crc kubenswrapper[4870]: I1014 09:52:08.008825 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 14 09:52:08 crc kubenswrapper[4870]: I1014 09:52:08.019073 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 14 09:52:08 crc kubenswrapper[4870]: I1014 09:52:08.025179 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 14 09:52:08 crc kubenswrapper[4870]: I1014 09:52:08.027230 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Oct 14 09:52:08 crc kubenswrapper[4870]: I1014 09:52:08.034737 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 14 09:52:08 crc kubenswrapper[4870]: I1014 09:52:08.053043 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 14 09:52:08 crc kubenswrapper[4870]: I1014 09:52:08.055939 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fce04f20-89b3-463d-b0eb-f32adcf5c8de-config-data\") pod \"nova-api-0\" (UID: \"fce04f20-89b3-463d-b0eb-f32adcf5c8de\") " pod="openstack/nova-api-0" Oct 14 09:52:08 crc kubenswrapper[4870]: I1014 09:52:08.055986 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xzkgw\" (UniqueName: \"kubernetes.io/projected/fce04f20-89b3-463d-b0eb-f32adcf5c8de-kube-api-access-xzkgw\") pod \"nova-api-0\" (UID: \"fce04f20-89b3-463d-b0eb-f32adcf5c8de\") " pod="openstack/nova-api-0" Oct 14 09:52:08 crc kubenswrapper[4870]: I1014 09:52:08.056049 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fce04f20-89b3-463d-b0eb-f32adcf5c8de-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"fce04f20-89b3-463d-b0eb-f32adcf5c8de\") " pod="openstack/nova-api-0" Oct 14 09:52:08 crc kubenswrapper[4870]: I1014 09:52:08.056095 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fce04f20-89b3-463d-b0eb-f32adcf5c8de-logs\") pod \"nova-api-0\" (UID: \"fce04f20-89b3-463d-b0eb-f32adcf5c8de\") " pod="openstack/nova-api-0" Oct 14 09:52:08 crc kubenswrapper[4870]: I1014 09:52:08.158212 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fce04f20-89b3-463d-b0eb-f32adcf5c8de-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"fce04f20-89b3-463d-b0eb-f32adcf5c8de\") " pod="openstack/nova-api-0" Oct 14 09:52:08 crc kubenswrapper[4870]: I1014 09:52:08.158274 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/52dbdd9e-e634-404b-8881-1a898ea568f0-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"52dbdd9e-e634-404b-8881-1a898ea568f0\") " pod="openstack/nova-cell0-conductor-0" Oct 14 09:52:08 crc kubenswrapper[4870]: I1014 09:52:08.159043 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fce04f20-89b3-463d-b0eb-f32adcf5c8de-logs\") pod \"nova-api-0\" (UID: \"fce04f20-89b3-463d-b0eb-f32adcf5c8de\") " pod="openstack/nova-api-0" Oct 14 09:52:08 crc kubenswrapper[4870]: I1014 09:52:08.159336 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fce04f20-89b3-463d-b0eb-f32adcf5c8de-logs\") pod \"nova-api-0\" (UID: \"fce04f20-89b3-463d-b0eb-f32adcf5c8de\") " pod="openstack/nova-api-0" Oct 14 09:52:08 crc kubenswrapper[4870]: I1014 09:52:08.159391 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9827ed43-9350-43db-8503-f4b9b0bd9c25-logs\") pod \"nova-metadata-0\" (UID: \"9827ed43-9350-43db-8503-f4b9b0bd9c25\") " pod="openstack/nova-metadata-0" Oct 14 09:52:08 crc kubenswrapper[4870]: I1014 09:52:08.159424 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/52dbdd9e-e634-404b-8881-1a898ea568f0-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"52dbdd9e-e634-404b-8881-1a898ea568f0\") " pod="openstack/nova-cell0-conductor-0" Oct 14 09:52:08 crc kubenswrapper[4870]: I1014 09:52:08.159509 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fce04f20-89b3-463d-b0eb-f32adcf5c8de-config-data\") pod \"nova-api-0\" (UID: \"fce04f20-89b3-463d-b0eb-f32adcf5c8de\") " pod="openstack/nova-api-0" Oct 14 09:52:08 crc kubenswrapper[4870]: I1014 09:52:08.159545 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xzkgw\" (UniqueName: \"kubernetes.io/projected/fce04f20-89b3-463d-b0eb-f32adcf5c8de-kube-api-access-xzkgw\") pod \"nova-api-0\" (UID: \"fce04f20-89b3-463d-b0eb-f32adcf5c8de\") " pod="openstack/nova-api-0" Oct 14 09:52:08 crc kubenswrapper[4870]: I1014 09:52:08.159567 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9827ed43-9350-43db-8503-f4b9b0bd9c25-config-data\") pod \"nova-metadata-0\" (UID: \"9827ed43-9350-43db-8503-f4b9b0bd9c25\") " pod="openstack/nova-metadata-0" Oct 14 09:52:08 crc kubenswrapper[4870]: I1014 09:52:08.159590 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vrrsx\" (UniqueName: \"kubernetes.io/projected/52dbdd9e-e634-404b-8881-1a898ea568f0-kube-api-access-vrrsx\") pod \"nova-cell0-conductor-0\" (UID: \"52dbdd9e-e634-404b-8881-1a898ea568f0\") " pod="openstack/nova-cell0-conductor-0" Oct 14 09:52:08 crc kubenswrapper[4870]: I1014 09:52:08.159612 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dmtqz\" (UniqueName: \"kubernetes.io/projected/9827ed43-9350-43db-8503-f4b9b0bd9c25-kube-api-access-dmtqz\") pod \"nova-metadata-0\" (UID: \"9827ed43-9350-43db-8503-f4b9b0bd9c25\") " pod="openstack/nova-metadata-0" Oct 14 09:52:08 crc kubenswrapper[4870]: I1014 09:52:08.159641 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9827ed43-9350-43db-8503-f4b9b0bd9c25-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"9827ed43-9350-43db-8503-f4b9b0bd9c25\") " pod="openstack/nova-metadata-0" Oct 14 09:52:08 crc kubenswrapper[4870]: I1014 09:52:08.162952 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fce04f20-89b3-463d-b0eb-f32adcf5c8de-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"fce04f20-89b3-463d-b0eb-f32adcf5c8de\") " pod="openstack/nova-api-0" Oct 14 09:52:08 crc kubenswrapper[4870]: I1014 09:52:08.163479 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fce04f20-89b3-463d-b0eb-f32adcf5c8de-config-data\") pod \"nova-api-0\" (UID: \"fce04f20-89b3-463d-b0eb-f32adcf5c8de\") " pod="openstack/nova-api-0" Oct 14 09:52:08 crc kubenswrapper[4870]: I1014 09:52:08.178091 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xzkgw\" (UniqueName: \"kubernetes.io/projected/fce04f20-89b3-463d-b0eb-f32adcf5c8de-kube-api-access-xzkgw\") pod \"nova-api-0\" (UID: \"fce04f20-89b3-463d-b0eb-f32adcf5c8de\") " pod="openstack/nova-api-0" Oct 14 09:52:08 crc kubenswrapper[4870]: I1014 09:52:08.261644 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/52dbdd9e-e634-404b-8881-1a898ea568f0-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"52dbdd9e-e634-404b-8881-1a898ea568f0\") " pod="openstack/nova-cell0-conductor-0" Oct 14 09:52:08 crc kubenswrapper[4870]: I1014 09:52:08.261741 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9827ed43-9350-43db-8503-f4b9b0bd9c25-logs\") pod \"nova-metadata-0\" (UID: \"9827ed43-9350-43db-8503-f4b9b0bd9c25\") " pod="openstack/nova-metadata-0" Oct 14 09:52:08 crc kubenswrapper[4870]: I1014 09:52:08.261768 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/52dbdd9e-e634-404b-8881-1a898ea568f0-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"52dbdd9e-e634-404b-8881-1a898ea568f0\") " pod="openstack/nova-cell0-conductor-0" Oct 14 09:52:08 crc kubenswrapper[4870]: I1014 09:52:08.261849 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9827ed43-9350-43db-8503-f4b9b0bd9c25-config-data\") pod \"nova-metadata-0\" (UID: \"9827ed43-9350-43db-8503-f4b9b0bd9c25\") " pod="openstack/nova-metadata-0" Oct 14 09:52:08 crc kubenswrapper[4870]: I1014 09:52:08.261872 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vrrsx\" (UniqueName: \"kubernetes.io/projected/52dbdd9e-e634-404b-8881-1a898ea568f0-kube-api-access-vrrsx\") pod \"nova-cell0-conductor-0\" (UID: \"52dbdd9e-e634-404b-8881-1a898ea568f0\") " pod="openstack/nova-cell0-conductor-0" Oct 14 09:52:08 crc kubenswrapper[4870]: I1014 09:52:08.261894 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dmtqz\" (UniqueName: \"kubernetes.io/projected/9827ed43-9350-43db-8503-f4b9b0bd9c25-kube-api-access-dmtqz\") pod \"nova-metadata-0\" (UID: \"9827ed43-9350-43db-8503-f4b9b0bd9c25\") " pod="openstack/nova-metadata-0" Oct 14 09:52:08 crc kubenswrapper[4870]: I1014 09:52:08.261920 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9827ed43-9350-43db-8503-f4b9b0bd9c25-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"9827ed43-9350-43db-8503-f4b9b0bd9c25\") " pod="openstack/nova-metadata-0" Oct 14 09:52:08 crc kubenswrapper[4870]: I1014 09:52:08.262637 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9827ed43-9350-43db-8503-f4b9b0bd9c25-logs\") pod \"nova-metadata-0\" (UID: \"9827ed43-9350-43db-8503-f4b9b0bd9c25\") " pod="openstack/nova-metadata-0" Oct 14 09:52:08 crc kubenswrapper[4870]: I1014 09:52:08.266032 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9827ed43-9350-43db-8503-f4b9b0bd9c25-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"9827ed43-9350-43db-8503-f4b9b0bd9c25\") " pod="openstack/nova-metadata-0" Oct 14 09:52:08 crc kubenswrapper[4870]: I1014 09:52:08.267427 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/52dbdd9e-e634-404b-8881-1a898ea568f0-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"52dbdd9e-e634-404b-8881-1a898ea568f0\") " pod="openstack/nova-cell0-conductor-0" Oct 14 09:52:08 crc kubenswrapper[4870]: I1014 09:52:08.268381 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9827ed43-9350-43db-8503-f4b9b0bd9c25-config-data\") pod \"nova-metadata-0\" (UID: \"9827ed43-9350-43db-8503-f4b9b0bd9c25\") " pod="openstack/nova-metadata-0" Oct 14 09:52:08 crc kubenswrapper[4870]: I1014 09:52:08.268839 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/52dbdd9e-e634-404b-8881-1a898ea568f0-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"52dbdd9e-e634-404b-8881-1a898ea568f0\") " pod="openstack/nova-cell0-conductor-0" Oct 14 09:52:08 crc kubenswrapper[4870]: I1014 09:52:08.270377 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 14 09:52:08 crc kubenswrapper[4870]: I1014 09:52:08.291393 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dmtqz\" (UniqueName: \"kubernetes.io/projected/9827ed43-9350-43db-8503-f4b9b0bd9c25-kube-api-access-dmtqz\") pod \"nova-metadata-0\" (UID: \"9827ed43-9350-43db-8503-f4b9b0bd9c25\") " pod="openstack/nova-metadata-0" Oct 14 09:52:08 crc kubenswrapper[4870]: I1014 09:52:08.292855 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vrrsx\" (UniqueName: \"kubernetes.io/projected/52dbdd9e-e634-404b-8881-1a898ea568f0-kube-api-access-vrrsx\") pod \"nova-cell0-conductor-0\" (UID: \"52dbdd9e-e634-404b-8881-1a898ea568f0\") " pod="openstack/nova-cell0-conductor-0" Oct 14 09:52:08 crc kubenswrapper[4870]: I1014 09:52:08.327926 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 14 09:52:08 crc kubenswrapper[4870]: I1014 09:52:08.345974 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 14 09:52:08 crc kubenswrapper[4870]: I1014 09:52:08.841597 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 14 09:52:08 crc kubenswrapper[4870]: I1014 09:52:08.921468 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 14 09:52:08 crc kubenswrapper[4870]: W1014 09:52:08.922988 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9827ed43_9350_43db_8503_f4b9b0bd9c25.slice/crio-a0700968e7f6bb8c370d9242c89160677dfec1f32b7d5e02b0fe1fe9b60c242b WatchSource:0}: Error finding container a0700968e7f6bb8c370d9242c89160677dfec1f32b7d5e02b0fe1fe9b60c242b: Status 404 returned error can't find the container with id a0700968e7f6bb8c370d9242c89160677dfec1f32b7d5e02b0fe1fe9b60c242b Oct 14 09:52:09 crc kubenswrapper[4870]: I1014 09:52:09.005250 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 14 09:52:09 crc kubenswrapper[4870]: I1014 09:52:09.052550 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="17f1a688-1bd1-4b88-ad00-51caf1af0ce0" path="/var/lib/kubelet/pods/17f1a688-1bd1-4b88-ad00-51caf1af0ce0/volumes" Oct 14 09:52:09 crc kubenswrapper[4870]: I1014 09:52:09.053232 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e1055bc1-24f3-4893-ad77-c520a36428bb" path="/var/lib/kubelet/pods/e1055bc1-24f3-4893-ad77-c520a36428bb/volumes" Oct 14 09:52:09 crc kubenswrapper[4870]: I1014 09:52:09.053964 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ee5dcbc8-b98e-48b9-a089-a9f5614249d3" path="/var/lib/kubelet/pods/ee5dcbc8-b98e-48b9-a089-a9f5614249d3/volumes" Oct 14 09:52:09 crc kubenswrapper[4870]: I1014 09:52:09.874551 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"52dbdd9e-e634-404b-8881-1a898ea568f0","Type":"ContainerStarted","Data":"b44d5e9286ec07a7aee6efd2bc58a481aca2f8cad12a4e8d9366580b6c28d79f"} Oct 14 09:52:09 crc kubenswrapper[4870]: I1014 09:52:09.875841 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"52dbdd9e-e634-404b-8881-1a898ea568f0","Type":"ContainerStarted","Data":"31f6e9ad3cc80ea067edec4a714f5a3bc201ca41f5fbd92aa1a6b1697d2dd774"} Oct 14 09:52:09 crc kubenswrapper[4870]: I1014 09:52:09.876086 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Oct 14 09:52:09 crc kubenswrapper[4870]: I1014 09:52:09.878132 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"9827ed43-9350-43db-8503-f4b9b0bd9c25","Type":"ContainerStarted","Data":"47e604fa5e159d579d2c2e4e2da23c524fc4add768354a4abbee93c6e6310dd1"} Oct 14 09:52:09 crc kubenswrapper[4870]: I1014 09:52:09.878225 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"9827ed43-9350-43db-8503-f4b9b0bd9c25","Type":"ContainerStarted","Data":"849b7b6e8e79c46e4e2c3b6536d50b89bbcd277e9704d65b5d349f45a01e967d"} Oct 14 09:52:09 crc kubenswrapper[4870]: I1014 09:52:09.878253 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"9827ed43-9350-43db-8503-f4b9b0bd9c25","Type":"ContainerStarted","Data":"a0700968e7f6bb8c370d9242c89160677dfec1f32b7d5e02b0fe1fe9b60c242b"} Oct 14 09:52:09 crc kubenswrapper[4870]: I1014 09:52:09.880746 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"fce04f20-89b3-463d-b0eb-f32adcf5c8de","Type":"ContainerStarted","Data":"7715b987cd8c4f7e41f7cbd7bf372079de63ec8820dbd64d081d833a0c6b2c88"} Oct 14 09:52:09 crc kubenswrapper[4870]: I1014 09:52:09.880794 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"fce04f20-89b3-463d-b0eb-f32adcf5c8de","Type":"ContainerStarted","Data":"5f894968b36c8f14a329b8d355292612f030594fb5d422790ef125cc040a9e2b"} Oct 14 09:52:09 crc kubenswrapper[4870]: I1014 09:52:09.880806 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"fce04f20-89b3-463d-b0eb-f32adcf5c8de","Type":"ContainerStarted","Data":"8fd0a0b9d55fa25c3e0e148788cc4fe4d8ce62bee60014311b89cfa0e922fec3"} Oct 14 09:52:09 crc kubenswrapper[4870]: I1014 09:52:09.900230 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.9001986349999997 podStartE2EDuration="2.900198635s" podCreationTimestamp="2025-10-14 09:52:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 09:52:09.892816164 +0000 UTC m=+10265.590176535" watchObservedRunningTime="2025-10-14 09:52:09.900198635 +0000 UTC m=+10265.597559016" Oct 14 09:52:09 crc kubenswrapper[4870]: I1014 09:52:09.934313 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.934294734 podStartE2EDuration="2.934294734s" podCreationTimestamp="2025-10-14 09:52:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 09:52:09.923417606 +0000 UTC m=+10265.620777997" watchObservedRunningTime="2025-10-14 09:52:09.934294734 +0000 UTC m=+10265.631655105" Oct 14 09:52:09 crc kubenswrapper[4870]: I1014 09:52:09.949872 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.949855607 podStartE2EDuration="2.949855607s" podCreationTimestamp="2025-10-14 09:52:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 09:52:09.943578792 +0000 UTC m=+10265.640939163" watchObservedRunningTime="2025-10-14 09:52:09.949855607 +0000 UTC m=+10265.647215978" Oct 14 09:52:11 crc kubenswrapper[4870]: I1014 09:52:11.130776 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 14 09:52:13 crc kubenswrapper[4870]: I1014 09:52:13.328699 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 14 09:52:13 crc kubenswrapper[4870]: I1014 09:52:13.331159 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 14 09:52:14 crc kubenswrapper[4870]: I1014 09:52:14.204651 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Oct 14 09:52:16 crc kubenswrapper[4870]: I1014 09:52:16.130950 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 14 09:52:16 crc kubenswrapper[4870]: I1014 09:52:16.176478 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 14 09:52:17 crc kubenswrapper[4870]: I1014 09:52:17.179656 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 14 09:52:18 crc kubenswrapper[4870]: I1014 09:52:18.270613 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 14 09:52:18 crc kubenswrapper[4870]: I1014 09:52:18.270764 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 14 09:52:18 crc kubenswrapper[4870]: I1014 09:52:18.329097 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 14 09:52:18 crc kubenswrapper[4870]: I1014 09:52:18.329163 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 14 09:52:18 crc kubenswrapper[4870]: I1014 09:52:18.382116 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Oct 14 09:52:19 crc kubenswrapper[4870]: I1014 09:52:19.353624 4870 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="fce04f20-89b3-463d-b0eb-f32adcf5c8de" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.1.201:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 14 09:52:19 crc kubenswrapper[4870]: I1014 09:52:19.353631 4870 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="fce04f20-89b3-463d-b0eb-f32adcf5c8de" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.1.201:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 14 09:52:19 crc kubenswrapper[4870]: I1014 09:52:19.435658 4870 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="9827ed43-9350-43db-8503-f4b9b0bd9c25" containerName="nova-metadata-log" probeResult="failure" output="Get \"http://10.217.1.202:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 14 09:52:19 crc kubenswrapper[4870]: I1014 09:52:19.435838 4870 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="9827ed43-9350-43db-8503-f4b9b0bd9c25" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"http://10.217.1.202:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 14 09:52:23 crc kubenswrapper[4870]: I1014 09:52:23.950804 4870 patch_prober.go:28] interesting pod/machine-config-daemon-7tvc7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 09:52:23 crc kubenswrapper[4870]: I1014 09:52:23.951428 4870 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 09:52:28 crc kubenswrapper[4870]: I1014 09:52:28.277520 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 14 09:52:28 crc kubenswrapper[4870]: I1014 09:52:28.278584 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 14 09:52:28 crc kubenswrapper[4870]: I1014 09:52:28.281601 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 14 09:52:28 crc kubenswrapper[4870]: I1014 09:52:28.283084 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 14 09:52:28 crc kubenswrapper[4870]: I1014 09:52:28.330430 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 14 09:52:28 crc kubenswrapper[4870]: I1014 09:52:28.330683 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 14 09:52:28 crc kubenswrapper[4870]: I1014 09:52:28.332122 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 14 09:52:29 crc kubenswrapper[4870]: I1014 09:52:29.170230 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 14 09:52:29 crc kubenswrapper[4870]: I1014 09:52:29.174411 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 14 09:52:29 crc kubenswrapper[4870]: I1014 09:52:29.175894 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 14 09:52:30 crc kubenswrapper[4870]: I1014 09:52:30.223204 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell7hdbp"] Oct 14 09:52:30 crc kubenswrapper[4870]: I1014 09:52:30.224975 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell7hdbp" Oct 14 09:52:30 crc kubenswrapper[4870]: I1014 09:52:30.227560 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Oct 14 09:52:30 crc kubenswrapper[4870]: I1014 09:52:30.228587 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 14 09:52:30 crc kubenswrapper[4870]: I1014 09:52:30.228589 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-wrz7w" Oct 14 09:52:30 crc kubenswrapper[4870]: I1014 09:52:30.228918 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 14 09:52:30 crc kubenswrapper[4870]: I1014 09:52:30.230695 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 14 09:52:30 crc kubenswrapper[4870]: I1014 09:52:30.230963 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-cells-global-config" Oct 14 09:52:30 crc kubenswrapper[4870]: I1014 09:52:30.232877 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Oct 14 09:52:30 crc kubenswrapper[4870]: I1014 09:52:30.246732 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell7hdbp"] Oct 14 09:52:30 crc kubenswrapper[4870]: I1014 09:52:30.384572 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ff165ad-7fb3-4c5a-918b-94950ba97a0a-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell7hdbp\" (UID: \"2ff165ad-7fb3-4c5a-918b-94950ba97a0a\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell7hdbp" Oct 14 09:52:30 crc kubenswrapper[4870]: I1014 09:52:30.384946 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2ff165ad-7fb3-4c5a-918b-94950ba97a0a-inventory\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell7hdbp\" (UID: \"2ff165ad-7fb3-4c5a-918b-94950ba97a0a\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell7hdbp" Oct 14 09:52:30 crc kubenswrapper[4870]: I1014 09:52:30.384983 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/2ff165ad-7fb3-4c5a-918b-94950ba97a0a-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell7hdbp\" (UID: \"2ff165ad-7fb3-4c5a-918b-94950ba97a0a\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell7hdbp" Oct 14 09:52:30 crc kubenswrapper[4870]: I1014 09:52:30.385021 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/2ff165ad-7fb3-4c5a-918b-94950ba97a0a-ceph\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell7hdbp\" (UID: \"2ff165ad-7fb3-4c5a-918b-94950ba97a0a\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell7hdbp" Oct 14 09:52:30 crc kubenswrapper[4870]: I1014 09:52:30.385054 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/2ff165ad-7fb3-4c5a-918b-94950ba97a0a-nova-cells-global-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell7hdbp\" (UID: \"2ff165ad-7fb3-4c5a-918b-94950ba97a0a\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell7hdbp" Oct 14 09:52:30 crc kubenswrapper[4870]: I1014 09:52:30.385078 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/2ff165ad-7fb3-4c5a-918b-94950ba97a0a-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell7hdbp\" (UID: \"2ff165ad-7fb3-4c5a-918b-94950ba97a0a\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell7hdbp" Oct 14 09:52:30 crc kubenswrapper[4870]: I1014 09:52:30.385133 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/2ff165ad-7fb3-4c5a-918b-94950ba97a0a-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell7hdbp\" (UID: \"2ff165ad-7fb3-4c5a-918b-94950ba97a0a\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell7hdbp" Oct 14 09:52:30 crc kubenswrapper[4870]: I1014 09:52:30.385209 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/2ff165ad-7fb3-4c5a-918b-94950ba97a0a-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell7hdbp\" (UID: \"2ff165ad-7fb3-4c5a-918b-94950ba97a0a\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell7hdbp" Oct 14 09:52:30 crc kubenswrapper[4870]: I1014 09:52:30.385246 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b8fh4\" (UniqueName: \"kubernetes.io/projected/2ff165ad-7fb3-4c5a-918b-94950ba97a0a-kube-api-access-b8fh4\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell7hdbp\" (UID: \"2ff165ad-7fb3-4c5a-918b-94950ba97a0a\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell7hdbp" Oct 14 09:52:30 crc kubenswrapper[4870]: I1014 09:52:30.385268 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2ff165ad-7fb3-4c5a-918b-94950ba97a0a-ssh-key\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell7hdbp\" (UID: \"2ff165ad-7fb3-4c5a-918b-94950ba97a0a\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell7hdbp" Oct 14 09:52:30 crc kubenswrapper[4870]: I1014 09:52:30.385289 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/2ff165ad-7fb3-4c5a-918b-94950ba97a0a-nova-cells-global-config-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell7hdbp\" (UID: \"2ff165ad-7fb3-4c5a-918b-94950ba97a0a\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell7hdbp" Oct 14 09:52:30 crc kubenswrapper[4870]: I1014 09:52:30.487152 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2ff165ad-7fb3-4c5a-918b-94950ba97a0a-inventory\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell7hdbp\" (UID: \"2ff165ad-7fb3-4c5a-918b-94950ba97a0a\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell7hdbp" Oct 14 09:52:30 crc kubenswrapper[4870]: I1014 09:52:30.487214 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/2ff165ad-7fb3-4c5a-918b-94950ba97a0a-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell7hdbp\" (UID: \"2ff165ad-7fb3-4c5a-918b-94950ba97a0a\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell7hdbp" Oct 14 09:52:30 crc kubenswrapper[4870]: I1014 09:52:30.487245 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/2ff165ad-7fb3-4c5a-918b-94950ba97a0a-ceph\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell7hdbp\" (UID: \"2ff165ad-7fb3-4c5a-918b-94950ba97a0a\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell7hdbp" Oct 14 09:52:30 crc kubenswrapper[4870]: I1014 09:52:30.487284 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/2ff165ad-7fb3-4c5a-918b-94950ba97a0a-nova-cells-global-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell7hdbp\" (UID: \"2ff165ad-7fb3-4c5a-918b-94950ba97a0a\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell7hdbp" Oct 14 09:52:30 crc kubenswrapper[4870]: I1014 09:52:30.487309 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/2ff165ad-7fb3-4c5a-918b-94950ba97a0a-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell7hdbp\" (UID: \"2ff165ad-7fb3-4c5a-918b-94950ba97a0a\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell7hdbp" Oct 14 09:52:30 crc kubenswrapper[4870]: I1014 09:52:30.487351 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/2ff165ad-7fb3-4c5a-918b-94950ba97a0a-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell7hdbp\" (UID: \"2ff165ad-7fb3-4c5a-918b-94950ba97a0a\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell7hdbp" Oct 14 09:52:30 crc kubenswrapper[4870]: I1014 09:52:30.487392 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/2ff165ad-7fb3-4c5a-918b-94950ba97a0a-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell7hdbp\" (UID: \"2ff165ad-7fb3-4c5a-918b-94950ba97a0a\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell7hdbp" Oct 14 09:52:30 crc kubenswrapper[4870]: I1014 09:52:30.487423 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b8fh4\" (UniqueName: \"kubernetes.io/projected/2ff165ad-7fb3-4c5a-918b-94950ba97a0a-kube-api-access-b8fh4\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell7hdbp\" (UID: \"2ff165ad-7fb3-4c5a-918b-94950ba97a0a\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell7hdbp" Oct 14 09:52:30 crc kubenswrapper[4870]: I1014 09:52:30.487458 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2ff165ad-7fb3-4c5a-918b-94950ba97a0a-ssh-key\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell7hdbp\" (UID: \"2ff165ad-7fb3-4c5a-918b-94950ba97a0a\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell7hdbp" Oct 14 09:52:30 crc kubenswrapper[4870]: I1014 09:52:30.487481 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/2ff165ad-7fb3-4c5a-918b-94950ba97a0a-nova-cells-global-config-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell7hdbp\" (UID: \"2ff165ad-7fb3-4c5a-918b-94950ba97a0a\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell7hdbp" Oct 14 09:52:30 crc kubenswrapper[4870]: I1014 09:52:30.487529 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ff165ad-7fb3-4c5a-918b-94950ba97a0a-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell7hdbp\" (UID: \"2ff165ad-7fb3-4c5a-918b-94950ba97a0a\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell7hdbp" Oct 14 09:52:30 crc kubenswrapper[4870]: I1014 09:52:30.493710 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/2ff165ad-7fb3-4c5a-918b-94950ba97a0a-nova-cells-global-config-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell7hdbp\" (UID: \"2ff165ad-7fb3-4c5a-918b-94950ba97a0a\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell7hdbp" Oct 14 09:52:30 crc kubenswrapper[4870]: I1014 09:52:30.494021 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/2ff165ad-7fb3-4c5a-918b-94950ba97a0a-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell7hdbp\" (UID: \"2ff165ad-7fb3-4c5a-918b-94950ba97a0a\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell7hdbp" Oct 14 09:52:30 crc kubenswrapper[4870]: I1014 09:52:30.494593 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2ff165ad-7fb3-4c5a-918b-94950ba97a0a-inventory\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell7hdbp\" (UID: \"2ff165ad-7fb3-4c5a-918b-94950ba97a0a\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell7hdbp" Oct 14 09:52:30 crc kubenswrapper[4870]: I1014 09:52:30.495021 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ff165ad-7fb3-4c5a-918b-94950ba97a0a-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell7hdbp\" (UID: \"2ff165ad-7fb3-4c5a-918b-94950ba97a0a\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell7hdbp" Oct 14 09:52:30 crc kubenswrapper[4870]: I1014 09:52:30.497067 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/2ff165ad-7fb3-4c5a-918b-94950ba97a0a-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell7hdbp\" (UID: \"2ff165ad-7fb3-4c5a-918b-94950ba97a0a\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell7hdbp" Oct 14 09:52:30 crc kubenswrapper[4870]: I1014 09:52:30.497809 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/2ff165ad-7fb3-4c5a-918b-94950ba97a0a-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell7hdbp\" (UID: \"2ff165ad-7fb3-4c5a-918b-94950ba97a0a\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell7hdbp" Oct 14 09:52:30 crc kubenswrapper[4870]: I1014 09:52:30.497999 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/2ff165ad-7fb3-4c5a-918b-94950ba97a0a-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell7hdbp\" (UID: \"2ff165ad-7fb3-4c5a-918b-94950ba97a0a\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell7hdbp" Oct 14 09:52:30 crc kubenswrapper[4870]: I1014 09:52:30.498186 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/2ff165ad-7fb3-4c5a-918b-94950ba97a0a-nova-cells-global-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell7hdbp\" (UID: \"2ff165ad-7fb3-4c5a-918b-94950ba97a0a\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell7hdbp" Oct 14 09:52:30 crc kubenswrapper[4870]: I1014 09:52:30.498639 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/2ff165ad-7fb3-4c5a-918b-94950ba97a0a-ceph\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell7hdbp\" (UID: \"2ff165ad-7fb3-4c5a-918b-94950ba97a0a\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell7hdbp" Oct 14 09:52:30 crc kubenswrapper[4870]: I1014 09:52:30.500599 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2ff165ad-7fb3-4c5a-918b-94950ba97a0a-ssh-key\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell7hdbp\" (UID: \"2ff165ad-7fb3-4c5a-918b-94950ba97a0a\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell7hdbp" Oct 14 09:52:30 crc kubenswrapper[4870]: I1014 09:52:30.508514 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b8fh4\" (UniqueName: \"kubernetes.io/projected/2ff165ad-7fb3-4c5a-918b-94950ba97a0a-kube-api-access-b8fh4\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell7hdbp\" (UID: \"2ff165ad-7fb3-4c5a-918b-94950ba97a0a\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell7hdbp" Oct 14 09:52:30 crc kubenswrapper[4870]: I1014 09:52:30.581939 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell7hdbp" Oct 14 09:52:31 crc kubenswrapper[4870]: I1014 09:52:31.163328 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell7hdbp"] Oct 14 09:52:31 crc kubenswrapper[4870]: W1014 09:52:31.655945 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2ff165ad_7fb3_4c5a_918b_94950ba97a0a.slice/crio-f499a75bc2f775ad0bc150b8be3cdd70b75177989a8ca75bff2034e70c7d1434 WatchSource:0}: Error finding container f499a75bc2f775ad0bc150b8be3cdd70b75177989a8ca75bff2034e70c7d1434: Status 404 returned error can't find the container with id f499a75bc2f775ad0bc150b8be3cdd70b75177989a8ca75bff2034e70c7d1434 Oct 14 09:52:32 crc kubenswrapper[4870]: I1014 09:52:32.211419 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell7hdbp" event={"ID":"2ff165ad-7fb3-4c5a-918b-94950ba97a0a","Type":"ContainerStarted","Data":"f499a75bc2f775ad0bc150b8be3cdd70b75177989a8ca75bff2034e70c7d1434"} Oct 14 09:52:33 crc kubenswrapper[4870]: I1014 09:52:33.231811 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell7hdbp" event={"ID":"2ff165ad-7fb3-4c5a-918b-94950ba97a0a","Type":"ContainerStarted","Data":"a6cd11bd728b708b6e281bb65a06fa0509f20de40fc02399a905b0cc430b8ce5"} Oct 14 09:52:33 crc kubenswrapper[4870]: I1014 09:52:33.277045 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell7hdbp" podStartSLOduration=2.692285237 podStartE2EDuration="3.277018718s" podCreationTimestamp="2025-10-14 09:52:30 +0000 UTC" firstStartedPulling="2025-10-14 09:52:31.664472071 +0000 UTC m=+10287.361832472" lastFinishedPulling="2025-10-14 09:52:32.249205582 +0000 UTC m=+10287.946565953" observedRunningTime="2025-10-14 09:52:33.261092116 +0000 UTC m=+10288.958452497" watchObservedRunningTime="2025-10-14 09:52:33.277018718 +0000 UTC m=+10288.974379099" Oct 14 09:52:33 crc kubenswrapper[4870]: I1014 09:52:33.816640 4870 scope.go:117] "RemoveContainer" containerID="c0020e7320b47c052dc0efb0f61570b5a39b13c7d39e5b5d0bc9ebec23c7b744" Oct 14 09:52:53 crc kubenswrapper[4870]: I1014 09:52:53.950730 4870 patch_prober.go:28] interesting pod/machine-config-daemon-7tvc7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 09:52:53 crc kubenswrapper[4870]: I1014 09:52:53.951330 4870 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 09:53:23 crc kubenswrapper[4870]: I1014 09:53:23.951530 4870 patch_prober.go:28] interesting pod/machine-config-daemon-7tvc7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 09:53:23 crc kubenswrapper[4870]: I1014 09:53:23.952130 4870 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 09:53:23 crc kubenswrapper[4870]: I1014 09:53:23.952198 4870 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" Oct 14 09:53:23 crc kubenswrapper[4870]: I1014 09:53:23.953399 4870 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c52e0ba5edaea8b07f60b9f76084bc5c51b52575d871193fc95ac2cefd996fea"} pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 14 09:53:23 crc kubenswrapper[4870]: I1014 09:53:23.953525 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerName="machine-config-daemon" containerID="cri-o://c52e0ba5edaea8b07f60b9f76084bc5c51b52575d871193fc95ac2cefd996fea" gracePeriod=600 Oct 14 09:53:24 crc kubenswrapper[4870]: E1014 09:53:24.091478 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 09:53:24 crc kubenswrapper[4870]: I1014 09:53:24.874558 4870 generic.go:334] "Generic (PLEG): container finished" podID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerID="c52e0ba5edaea8b07f60b9f76084bc5c51b52575d871193fc95ac2cefd996fea" exitCode=0 Oct 14 09:53:24 crc kubenswrapper[4870]: I1014 09:53:24.874682 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" event={"ID":"874a5e45-dffd-4d17-b609-b7d3ed2eab07","Type":"ContainerDied","Data":"c52e0ba5edaea8b07f60b9f76084bc5c51b52575d871193fc95ac2cefd996fea"} Oct 14 09:53:24 crc kubenswrapper[4870]: I1014 09:53:24.875096 4870 scope.go:117] "RemoveContainer" containerID="4d8d8702edb8183ff39b29852bc295d1ac61316a22904ed8a1cfe721472e2347" Oct 14 09:53:24 crc kubenswrapper[4870]: I1014 09:53:24.880841 4870 scope.go:117] "RemoveContainer" containerID="c52e0ba5edaea8b07f60b9f76084bc5c51b52575d871193fc95ac2cefd996fea" Oct 14 09:53:24 crc kubenswrapper[4870]: E1014 09:53:24.883712 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 09:53:40 crc kubenswrapper[4870]: I1014 09:53:40.034316 4870 scope.go:117] "RemoveContainer" containerID="c52e0ba5edaea8b07f60b9f76084bc5c51b52575d871193fc95ac2cefd996fea" Oct 14 09:53:40 crc kubenswrapper[4870]: E1014 09:53:40.035289 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 09:53:51 crc kubenswrapper[4870]: I1014 09:53:51.035143 4870 scope.go:117] "RemoveContainer" containerID="c52e0ba5edaea8b07f60b9f76084bc5c51b52575d871193fc95ac2cefd996fea" Oct 14 09:53:51 crc kubenswrapper[4870]: E1014 09:53:51.036402 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 09:54:06 crc kubenswrapper[4870]: I1014 09:54:06.034540 4870 scope.go:117] "RemoveContainer" containerID="c52e0ba5edaea8b07f60b9f76084bc5c51b52575d871193fc95ac2cefd996fea" Oct 14 09:54:06 crc kubenswrapper[4870]: E1014 09:54:06.035303 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 09:54:17 crc kubenswrapper[4870]: I1014 09:54:17.034892 4870 scope.go:117] "RemoveContainer" containerID="c52e0ba5edaea8b07f60b9f76084bc5c51b52575d871193fc95ac2cefd996fea" Oct 14 09:54:17 crc kubenswrapper[4870]: E1014 09:54:17.036037 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 09:54:30 crc kubenswrapper[4870]: I1014 09:54:30.034414 4870 scope.go:117] "RemoveContainer" containerID="c52e0ba5edaea8b07f60b9f76084bc5c51b52575d871193fc95ac2cefd996fea" Oct 14 09:54:30 crc kubenswrapper[4870]: E1014 09:54:30.035649 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 09:54:41 crc kubenswrapper[4870]: I1014 09:54:41.034511 4870 scope.go:117] "RemoveContainer" containerID="c52e0ba5edaea8b07f60b9f76084bc5c51b52575d871193fc95ac2cefd996fea" Oct 14 09:54:41 crc kubenswrapper[4870]: E1014 09:54:41.035366 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 09:54:55 crc kubenswrapper[4870]: I1014 09:54:55.051214 4870 scope.go:117] "RemoveContainer" containerID="c52e0ba5edaea8b07f60b9f76084bc5c51b52575d871193fc95ac2cefd996fea" Oct 14 09:54:55 crc kubenswrapper[4870]: E1014 09:54:55.052696 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 09:54:55 crc kubenswrapper[4870]: I1014 09:54:55.977259 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-lf25r"] Oct 14 09:54:55 crc kubenswrapper[4870]: I1014 09:54:55.980588 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lf25r" Oct 14 09:54:55 crc kubenswrapper[4870]: I1014 09:54:55.990142 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-lf25r"] Oct 14 09:54:56 crc kubenswrapper[4870]: I1014 09:54:56.089286 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jc9z5\" (UniqueName: \"kubernetes.io/projected/3f4a1094-671e-4c21-ba5c-c7f54bd9d2ae-kube-api-access-jc9z5\") pod \"redhat-operators-lf25r\" (UID: \"3f4a1094-671e-4c21-ba5c-c7f54bd9d2ae\") " pod="openshift-marketplace/redhat-operators-lf25r" Oct 14 09:54:56 crc kubenswrapper[4870]: I1014 09:54:56.089459 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3f4a1094-671e-4c21-ba5c-c7f54bd9d2ae-catalog-content\") pod \"redhat-operators-lf25r\" (UID: \"3f4a1094-671e-4c21-ba5c-c7f54bd9d2ae\") " pod="openshift-marketplace/redhat-operators-lf25r" Oct 14 09:54:56 crc kubenswrapper[4870]: I1014 09:54:56.089659 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3f4a1094-671e-4c21-ba5c-c7f54bd9d2ae-utilities\") pod \"redhat-operators-lf25r\" (UID: \"3f4a1094-671e-4c21-ba5c-c7f54bd9d2ae\") " pod="openshift-marketplace/redhat-operators-lf25r" Oct 14 09:54:56 crc kubenswrapper[4870]: I1014 09:54:56.191717 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jc9z5\" (UniqueName: \"kubernetes.io/projected/3f4a1094-671e-4c21-ba5c-c7f54bd9d2ae-kube-api-access-jc9z5\") pod \"redhat-operators-lf25r\" (UID: \"3f4a1094-671e-4c21-ba5c-c7f54bd9d2ae\") " pod="openshift-marketplace/redhat-operators-lf25r" Oct 14 09:54:56 crc kubenswrapper[4870]: I1014 09:54:56.191811 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3f4a1094-671e-4c21-ba5c-c7f54bd9d2ae-catalog-content\") pod \"redhat-operators-lf25r\" (UID: \"3f4a1094-671e-4c21-ba5c-c7f54bd9d2ae\") " pod="openshift-marketplace/redhat-operators-lf25r" Oct 14 09:54:56 crc kubenswrapper[4870]: I1014 09:54:56.191897 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3f4a1094-671e-4c21-ba5c-c7f54bd9d2ae-utilities\") pod \"redhat-operators-lf25r\" (UID: \"3f4a1094-671e-4c21-ba5c-c7f54bd9d2ae\") " pod="openshift-marketplace/redhat-operators-lf25r" Oct 14 09:54:56 crc kubenswrapper[4870]: I1014 09:54:56.192708 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3f4a1094-671e-4c21-ba5c-c7f54bd9d2ae-utilities\") pod \"redhat-operators-lf25r\" (UID: \"3f4a1094-671e-4c21-ba5c-c7f54bd9d2ae\") " pod="openshift-marketplace/redhat-operators-lf25r" Oct 14 09:54:56 crc kubenswrapper[4870]: I1014 09:54:56.193036 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3f4a1094-671e-4c21-ba5c-c7f54bd9d2ae-catalog-content\") pod \"redhat-operators-lf25r\" (UID: \"3f4a1094-671e-4c21-ba5c-c7f54bd9d2ae\") " pod="openshift-marketplace/redhat-operators-lf25r" Oct 14 09:54:56 crc kubenswrapper[4870]: I1014 09:54:56.225090 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jc9z5\" (UniqueName: \"kubernetes.io/projected/3f4a1094-671e-4c21-ba5c-c7f54bd9d2ae-kube-api-access-jc9z5\") pod \"redhat-operators-lf25r\" (UID: \"3f4a1094-671e-4c21-ba5c-c7f54bd9d2ae\") " pod="openshift-marketplace/redhat-operators-lf25r" Oct 14 09:54:56 crc kubenswrapper[4870]: I1014 09:54:56.304723 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lf25r" Oct 14 09:54:56 crc kubenswrapper[4870]: I1014 09:54:56.763354 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-lf25r"] Oct 14 09:54:57 crc kubenswrapper[4870]: I1014 09:54:57.117476 4870 generic.go:334] "Generic (PLEG): container finished" podID="3f4a1094-671e-4c21-ba5c-c7f54bd9d2ae" containerID="ce09cfb9f0da777d1155b192e80cd894692bf024e31adc1fe1d1af62557886c0" exitCode=0 Oct 14 09:54:57 crc kubenswrapper[4870]: I1014 09:54:57.117544 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lf25r" event={"ID":"3f4a1094-671e-4c21-ba5c-c7f54bd9d2ae","Type":"ContainerDied","Data":"ce09cfb9f0da777d1155b192e80cd894692bf024e31adc1fe1d1af62557886c0"} Oct 14 09:54:57 crc kubenswrapper[4870]: I1014 09:54:57.117786 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lf25r" event={"ID":"3f4a1094-671e-4c21-ba5c-c7f54bd9d2ae","Type":"ContainerStarted","Data":"fe83ad651c83626346b8521ae8b53f7e0a53225d57c2405c1cf80a47e2acd42f"} Oct 14 09:54:57 crc kubenswrapper[4870]: I1014 09:54:57.120348 4870 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 14 09:54:57 crc kubenswrapper[4870]: E1014 09:54:57.158990 4870 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3f4a1094_671e_4c21_ba5c_c7f54bd9d2ae.slice/crio-conmon-ce09cfb9f0da777d1155b192e80cd894692bf024e31adc1fe1d1af62557886c0.scope\": RecentStats: unable to find data in memory cache]" Oct 14 09:54:59 crc kubenswrapper[4870]: I1014 09:54:59.139752 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lf25r" event={"ID":"3f4a1094-671e-4c21-ba5c-c7f54bd9d2ae","Type":"ContainerStarted","Data":"4bf732048ba577734be5da74779b84f9ce5941c407632ec6bf0b8486af5640ab"} Oct 14 09:55:02 crc kubenswrapper[4870]: I1014 09:55:02.178492 4870 generic.go:334] "Generic (PLEG): container finished" podID="3f4a1094-671e-4c21-ba5c-c7f54bd9d2ae" containerID="4bf732048ba577734be5da74779b84f9ce5941c407632ec6bf0b8486af5640ab" exitCode=0 Oct 14 09:55:02 crc kubenswrapper[4870]: I1014 09:55:02.178595 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lf25r" event={"ID":"3f4a1094-671e-4c21-ba5c-c7f54bd9d2ae","Type":"ContainerDied","Data":"4bf732048ba577734be5da74779b84f9ce5941c407632ec6bf0b8486af5640ab"} Oct 14 09:55:03 crc kubenswrapper[4870]: I1014 09:55:03.194204 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lf25r" event={"ID":"3f4a1094-671e-4c21-ba5c-c7f54bd9d2ae","Type":"ContainerStarted","Data":"1c44ea96e8dc3987890a82acd056945ca0ecaab764fe17835133ee8915abde58"} Oct 14 09:55:03 crc kubenswrapper[4870]: I1014 09:55:03.216097 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-lf25r" podStartSLOduration=2.54768949 podStartE2EDuration="8.216077058s" podCreationTimestamp="2025-10-14 09:54:55 +0000 UTC" firstStartedPulling="2025-10-14 09:54:57.120132555 +0000 UTC m=+10432.817492926" lastFinishedPulling="2025-10-14 09:55:02.788520103 +0000 UTC m=+10438.485880494" observedRunningTime="2025-10-14 09:55:03.215595836 +0000 UTC m=+10438.912956207" watchObservedRunningTime="2025-10-14 09:55:03.216077058 +0000 UTC m=+10438.913437439" Oct 14 09:55:06 crc kubenswrapper[4870]: I1014 09:55:06.305907 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-lf25r" Oct 14 09:55:06 crc kubenswrapper[4870]: I1014 09:55:06.306546 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-lf25r" Oct 14 09:55:07 crc kubenswrapper[4870]: I1014 09:55:07.530120 4870 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-lf25r" podUID="3f4a1094-671e-4c21-ba5c-c7f54bd9d2ae" containerName="registry-server" probeResult="failure" output=< Oct 14 09:55:07 crc kubenswrapper[4870]: timeout: failed to connect service ":50051" within 1s Oct 14 09:55:07 crc kubenswrapper[4870]: > Oct 14 09:55:09 crc kubenswrapper[4870]: I1014 09:55:09.034684 4870 scope.go:117] "RemoveContainer" containerID="c52e0ba5edaea8b07f60b9f76084bc5c51b52575d871193fc95ac2cefd996fea" Oct 14 09:55:09 crc kubenswrapper[4870]: E1014 09:55:09.035600 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 09:55:12 crc kubenswrapper[4870]: I1014 09:55:12.951040 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-gwtg4"] Oct 14 09:55:12 crc kubenswrapper[4870]: I1014 09:55:12.956416 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gwtg4" Oct 14 09:55:12 crc kubenswrapper[4870]: I1014 09:55:12.962929 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-gwtg4"] Oct 14 09:55:13 crc kubenswrapper[4870]: I1014 09:55:13.095203 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/52b3807b-ba2a-4bb7-9b0c-07fb73fbb41d-utilities\") pod \"community-operators-gwtg4\" (UID: \"52b3807b-ba2a-4bb7-9b0c-07fb73fbb41d\") " pod="openshift-marketplace/community-operators-gwtg4" Oct 14 09:55:13 crc kubenswrapper[4870]: I1014 09:55:13.095336 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6rkdt\" (UniqueName: \"kubernetes.io/projected/52b3807b-ba2a-4bb7-9b0c-07fb73fbb41d-kube-api-access-6rkdt\") pod \"community-operators-gwtg4\" (UID: \"52b3807b-ba2a-4bb7-9b0c-07fb73fbb41d\") " pod="openshift-marketplace/community-operators-gwtg4" Oct 14 09:55:13 crc kubenswrapper[4870]: I1014 09:55:13.095420 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/52b3807b-ba2a-4bb7-9b0c-07fb73fbb41d-catalog-content\") pod \"community-operators-gwtg4\" (UID: \"52b3807b-ba2a-4bb7-9b0c-07fb73fbb41d\") " pod="openshift-marketplace/community-operators-gwtg4" Oct 14 09:55:13 crc kubenswrapper[4870]: I1014 09:55:13.197552 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6rkdt\" (UniqueName: \"kubernetes.io/projected/52b3807b-ba2a-4bb7-9b0c-07fb73fbb41d-kube-api-access-6rkdt\") pod \"community-operators-gwtg4\" (UID: \"52b3807b-ba2a-4bb7-9b0c-07fb73fbb41d\") " pod="openshift-marketplace/community-operators-gwtg4" Oct 14 09:55:13 crc kubenswrapper[4870]: I1014 09:55:13.197613 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/52b3807b-ba2a-4bb7-9b0c-07fb73fbb41d-catalog-content\") pod \"community-operators-gwtg4\" (UID: \"52b3807b-ba2a-4bb7-9b0c-07fb73fbb41d\") " pod="openshift-marketplace/community-operators-gwtg4" Oct 14 09:55:13 crc kubenswrapper[4870]: I1014 09:55:13.197741 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/52b3807b-ba2a-4bb7-9b0c-07fb73fbb41d-utilities\") pod \"community-operators-gwtg4\" (UID: \"52b3807b-ba2a-4bb7-9b0c-07fb73fbb41d\") " pod="openshift-marketplace/community-operators-gwtg4" Oct 14 09:55:13 crc kubenswrapper[4870]: I1014 09:55:13.198242 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/52b3807b-ba2a-4bb7-9b0c-07fb73fbb41d-utilities\") pod \"community-operators-gwtg4\" (UID: \"52b3807b-ba2a-4bb7-9b0c-07fb73fbb41d\") " pod="openshift-marketplace/community-operators-gwtg4" Oct 14 09:55:13 crc kubenswrapper[4870]: I1014 09:55:13.198421 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/52b3807b-ba2a-4bb7-9b0c-07fb73fbb41d-catalog-content\") pod \"community-operators-gwtg4\" (UID: \"52b3807b-ba2a-4bb7-9b0c-07fb73fbb41d\") " pod="openshift-marketplace/community-operators-gwtg4" Oct 14 09:55:13 crc kubenswrapper[4870]: I1014 09:55:13.227398 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6rkdt\" (UniqueName: \"kubernetes.io/projected/52b3807b-ba2a-4bb7-9b0c-07fb73fbb41d-kube-api-access-6rkdt\") pod \"community-operators-gwtg4\" (UID: \"52b3807b-ba2a-4bb7-9b0c-07fb73fbb41d\") " pod="openshift-marketplace/community-operators-gwtg4" Oct 14 09:55:13 crc kubenswrapper[4870]: I1014 09:55:13.288966 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gwtg4" Oct 14 09:55:13 crc kubenswrapper[4870]: W1014 09:55:13.925599 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod52b3807b_ba2a_4bb7_9b0c_07fb73fbb41d.slice/crio-0ee2dd76d8bb55b9d5e9028ab91f781ab2902c6322f61b31c4339fc0fc31e4f6 WatchSource:0}: Error finding container 0ee2dd76d8bb55b9d5e9028ab91f781ab2902c6322f61b31c4339fc0fc31e4f6: Status 404 returned error can't find the container with id 0ee2dd76d8bb55b9d5e9028ab91f781ab2902c6322f61b31c4339fc0fc31e4f6 Oct 14 09:55:13 crc kubenswrapper[4870]: I1014 09:55:13.926371 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-gwtg4"] Oct 14 09:55:14 crc kubenswrapper[4870]: I1014 09:55:14.318808 4870 generic.go:334] "Generic (PLEG): container finished" podID="52b3807b-ba2a-4bb7-9b0c-07fb73fbb41d" containerID="24002b237d7cc2bdeff1852042fe37310e22e47f3f47daae5260f73ac13c2c47" exitCode=0 Oct 14 09:55:14 crc kubenswrapper[4870]: I1014 09:55:14.318851 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gwtg4" event={"ID":"52b3807b-ba2a-4bb7-9b0c-07fb73fbb41d","Type":"ContainerDied","Data":"24002b237d7cc2bdeff1852042fe37310e22e47f3f47daae5260f73ac13c2c47"} Oct 14 09:55:14 crc kubenswrapper[4870]: I1014 09:55:14.318878 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gwtg4" event={"ID":"52b3807b-ba2a-4bb7-9b0c-07fb73fbb41d","Type":"ContainerStarted","Data":"0ee2dd76d8bb55b9d5e9028ab91f781ab2902c6322f61b31c4339fc0fc31e4f6"} Oct 14 09:55:16 crc kubenswrapper[4870]: I1014 09:55:16.346077 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gwtg4" event={"ID":"52b3807b-ba2a-4bb7-9b0c-07fb73fbb41d","Type":"ContainerStarted","Data":"f19d25bd04e805b73e5dfbd87b8225f23700ef892c5fe2da67b78a0bf902eff3"} Oct 14 09:55:16 crc kubenswrapper[4870]: I1014 09:55:16.386580 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-lf25r" Oct 14 09:55:16 crc kubenswrapper[4870]: I1014 09:55:16.468589 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-lf25r" Oct 14 09:55:18 crc kubenswrapper[4870]: I1014 09:55:18.304081 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-lf25r"] Oct 14 09:55:18 crc kubenswrapper[4870]: I1014 09:55:18.376699 4870 generic.go:334] "Generic (PLEG): container finished" podID="52b3807b-ba2a-4bb7-9b0c-07fb73fbb41d" containerID="f19d25bd04e805b73e5dfbd87b8225f23700ef892c5fe2da67b78a0bf902eff3" exitCode=0 Oct 14 09:55:18 crc kubenswrapper[4870]: I1014 09:55:18.376846 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gwtg4" event={"ID":"52b3807b-ba2a-4bb7-9b0c-07fb73fbb41d","Type":"ContainerDied","Data":"f19d25bd04e805b73e5dfbd87b8225f23700ef892c5fe2da67b78a0bf902eff3"} Oct 14 09:55:18 crc kubenswrapper[4870]: I1014 09:55:18.377548 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-lf25r" podUID="3f4a1094-671e-4c21-ba5c-c7f54bd9d2ae" containerName="registry-server" containerID="cri-o://1c44ea96e8dc3987890a82acd056945ca0ecaab764fe17835133ee8915abde58" gracePeriod=2 Oct 14 09:55:18 crc kubenswrapper[4870]: I1014 09:55:18.976454 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lf25r" Oct 14 09:55:19 crc kubenswrapper[4870]: I1014 09:55:19.042725 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3f4a1094-671e-4c21-ba5c-c7f54bd9d2ae-utilities\") pod \"3f4a1094-671e-4c21-ba5c-c7f54bd9d2ae\" (UID: \"3f4a1094-671e-4c21-ba5c-c7f54bd9d2ae\") " Oct 14 09:55:19 crc kubenswrapper[4870]: I1014 09:55:19.042916 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3f4a1094-671e-4c21-ba5c-c7f54bd9d2ae-catalog-content\") pod \"3f4a1094-671e-4c21-ba5c-c7f54bd9d2ae\" (UID: \"3f4a1094-671e-4c21-ba5c-c7f54bd9d2ae\") " Oct 14 09:55:19 crc kubenswrapper[4870]: I1014 09:55:19.043127 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jc9z5\" (UniqueName: \"kubernetes.io/projected/3f4a1094-671e-4c21-ba5c-c7f54bd9d2ae-kube-api-access-jc9z5\") pod \"3f4a1094-671e-4c21-ba5c-c7f54bd9d2ae\" (UID: \"3f4a1094-671e-4c21-ba5c-c7f54bd9d2ae\") " Oct 14 09:55:19 crc kubenswrapper[4870]: I1014 09:55:19.044278 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3f4a1094-671e-4c21-ba5c-c7f54bd9d2ae-utilities" (OuterVolumeSpecName: "utilities") pod "3f4a1094-671e-4c21-ba5c-c7f54bd9d2ae" (UID: "3f4a1094-671e-4c21-ba5c-c7f54bd9d2ae"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 09:55:19 crc kubenswrapper[4870]: I1014 09:55:19.053280 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3f4a1094-671e-4c21-ba5c-c7f54bd9d2ae-kube-api-access-jc9z5" (OuterVolumeSpecName: "kube-api-access-jc9z5") pod "3f4a1094-671e-4c21-ba5c-c7f54bd9d2ae" (UID: "3f4a1094-671e-4c21-ba5c-c7f54bd9d2ae"). InnerVolumeSpecName "kube-api-access-jc9z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:55:19 crc kubenswrapper[4870]: I1014 09:55:19.141962 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3f4a1094-671e-4c21-ba5c-c7f54bd9d2ae-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3f4a1094-671e-4c21-ba5c-c7f54bd9d2ae" (UID: "3f4a1094-671e-4c21-ba5c-c7f54bd9d2ae"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 09:55:19 crc kubenswrapper[4870]: I1014 09:55:19.147143 4870 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3f4a1094-671e-4c21-ba5c-c7f54bd9d2ae-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 09:55:19 crc kubenswrapper[4870]: I1014 09:55:19.147190 4870 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3f4a1094-671e-4c21-ba5c-c7f54bd9d2ae-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 09:55:19 crc kubenswrapper[4870]: I1014 09:55:19.147208 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jc9z5\" (UniqueName: \"kubernetes.io/projected/3f4a1094-671e-4c21-ba5c-c7f54bd9d2ae-kube-api-access-jc9z5\") on node \"crc\" DevicePath \"\"" Oct 14 09:55:19 crc kubenswrapper[4870]: I1014 09:55:19.392134 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gwtg4" event={"ID":"52b3807b-ba2a-4bb7-9b0c-07fb73fbb41d","Type":"ContainerStarted","Data":"07f5aa211089a65625149f4fd175c74550a2427826036a4061c53be5ac39d90b"} Oct 14 09:55:19 crc kubenswrapper[4870]: I1014 09:55:19.398284 4870 generic.go:334] "Generic (PLEG): container finished" podID="3f4a1094-671e-4c21-ba5c-c7f54bd9d2ae" containerID="1c44ea96e8dc3987890a82acd056945ca0ecaab764fe17835133ee8915abde58" exitCode=0 Oct 14 09:55:19 crc kubenswrapper[4870]: I1014 09:55:19.398316 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lf25r" event={"ID":"3f4a1094-671e-4c21-ba5c-c7f54bd9d2ae","Type":"ContainerDied","Data":"1c44ea96e8dc3987890a82acd056945ca0ecaab764fe17835133ee8915abde58"} Oct 14 09:55:19 crc kubenswrapper[4870]: I1014 09:55:19.398369 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lf25r" event={"ID":"3f4a1094-671e-4c21-ba5c-c7f54bd9d2ae","Type":"ContainerDied","Data":"fe83ad651c83626346b8521ae8b53f7e0a53225d57c2405c1cf80a47e2acd42f"} Oct 14 09:55:19 crc kubenswrapper[4870]: I1014 09:55:19.398391 4870 scope.go:117] "RemoveContainer" containerID="1c44ea96e8dc3987890a82acd056945ca0ecaab764fe17835133ee8915abde58" Oct 14 09:55:19 crc kubenswrapper[4870]: I1014 09:55:19.398422 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lf25r" Oct 14 09:55:19 crc kubenswrapper[4870]: I1014 09:55:19.418173 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-gwtg4" podStartSLOduration=2.891051762 podStartE2EDuration="7.418154424s" podCreationTimestamp="2025-10-14 09:55:12 +0000 UTC" firstStartedPulling="2025-10-14 09:55:14.320958793 +0000 UTC m=+10450.018319164" lastFinishedPulling="2025-10-14 09:55:18.848061455 +0000 UTC m=+10454.545421826" observedRunningTime="2025-10-14 09:55:19.414140265 +0000 UTC m=+10455.111500636" watchObservedRunningTime="2025-10-14 09:55:19.418154424 +0000 UTC m=+10455.115514795" Oct 14 09:55:19 crc kubenswrapper[4870]: I1014 09:55:19.431506 4870 scope.go:117] "RemoveContainer" containerID="4bf732048ba577734be5da74779b84f9ce5941c407632ec6bf0b8486af5640ab" Oct 14 09:55:19 crc kubenswrapper[4870]: I1014 09:55:19.456168 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-lf25r"] Oct 14 09:55:19 crc kubenswrapper[4870]: I1014 09:55:19.466222 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-lf25r"] Oct 14 09:55:20 crc kubenswrapper[4870]: I1014 09:55:20.072865 4870 scope.go:117] "RemoveContainer" containerID="ce09cfb9f0da777d1155b192e80cd894692bf024e31adc1fe1d1af62557886c0" Oct 14 09:55:20 crc kubenswrapper[4870]: I1014 09:55:20.134491 4870 scope.go:117] "RemoveContainer" containerID="1c44ea96e8dc3987890a82acd056945ca0ecaab764fe17835133ee8915abde58" Oct 14 09:55:20 crc kubenswrapper[4870]: E1014 09:55:20.134977 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1c44ea96e8dc3987890a82acd056945ca0ecaab764fe17835133ee8915abde58\": container with ID starting with 1c44ea96e8dc3987890a82acd056945ca0ecaab764fe17835133ee8915abde58 not found: ID does not exist" containerID="1c44ea96e8dc3987890a82acd056945ca0ecaab764fe17835133ee8915abde58" Oct 14 09:55:20 crc kubenswrapper[4870]: I1014 09:55:20.135023 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1c44ea96e8dc3987890a82acd056945ca0ecaab764fe17835133ee8915abde58"} err="failed to get container status \"1c44ea96e8dc3987890a82acd056945ca0ecaab764fe17835133ee8915abde58\": rpc error: code = NotFound desc = could not find container \"1c44ea96e8dc3987890a82acd056945ca0ecaab764fe17835133ee8915abde58\": container with ID starting with 1c44ea96e8dc3987890a82acd056945ca0ecaab764fe17835133ee8915abde58 not found: ID does not exist" Oct 14 09:55:20 crc kubenswrapper[4870]: I1014 09:55:20.135051 4870 scope.go:117] "RemoveContainer" containerID="4bf732048ba577734be5da74779b84f9ce5941c407632ec6bf0b8486af5640ab" Oct 14 09:55:20 crc kubenswrapper[4870]: E1014 09:55:20.135549 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4bf732048ba577734be5da74779b84f9ce5941c407632ec6bf0b8486af5640ab\": container with ID starting with 4bf732048ba577734be5da74779b84f9ce5941c407632ec6bf0b8486af5640ab not found: ID does not exist" containerID="4bf732048ba577734be5da74779b84f9ce5941c407632ec6bf0b8486af5640ab" Oct 14 09:55:20 crc kubenswrapper[4870]: I1014 09:55:20.135586 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4bf732048ba577734be5da74779b84f9ce5941c407632ec6bf0b8486af5640ab"} err="failed to get container status \"4bf732048ba577734be5da74779b84f9ce5941c407632ec6bf0b8486af5640ab\": rpc error: code = NotFound desc = could not find container \"4bf732048ba577734be5da74779b84f9ce5941c407632ec6bf0b8486af5640ab\": container with ID starting with 4bf732048ba577734be5da74779b84f9ce5941c407632ec6bf0b8486af5640ab not found: ID does not exist" Oct 14 09:55:20 crc kubenswrapper[4870]: I1014 09:55:20.135608 4870 scope.go:117] "RemoveContainer" containerID="ce09cfb9f0da777d1155b192e80cd894692bf024e31adc1fe1d1af62557886c0" Oct 14 09:55:20 crc kubenswrapper[4870]: E1014 09:55:20.135904 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ce09cfb9f0da777d1155b192e80cd894692bf024e31adc1fe1d1af62557886c0\": container with ID starting with ce09cfb9f0da777d1155b192e80cd894692bf024e31adc1fe1d1af62557886c0 not found: ID does not exist" containerID="ce09cfb9f0da777d1155b192e80cd894692bf024e31adc1fe1d1af62557886c0" Oct 14 09:55:20 crc kubenswrapper[4870]: I1014 09:55:20.135950 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ce09cfb9f0da777d1155b192e80cd894692bf024e31adc1fe1d1af62557886c0"} err="failed to get container status \"ce09cfb9f0da777d1155b192e80cd894692bf024e31adc1fe1d1af62557886c0\": rpc error: code = NotFound desc = could not find container \"ce09cfb9f0da777d1155b192e80cd894692bf024e31adc1fe1d1af62557886c0\": container with ID starting with ce09cfb9f0da777d1155b192e80cd894692bf024e31adc1fe1d1af62557886c0 not found: ID does not exist" Oct 14 09:55:21 crc kubenswrapper[4870]: I1014 09:55:21.057498 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3f4a1094-671e-4c21-ba5c-c7f54bd9d2ae" path="/var/lib/kubelet/pods/3f4a1094-671e-4c21-ba5c-c7f54bd9d2ae/volumes" Oct 14 09:55:22 crc kubenswrapper[4870]: I1014 09:55:22.034427 4870 scope.go:117] "RemoveContainer" containerID="c52e0ba5edaea8b07f60b9f76084bc5c51b52575d871193fc95ac2cefd996fea" Oct 14 09:55:22 crc kubenswrapper[4870]: E1014 09:55:22.035737 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 09:55:23 crc kubenswrapper[4870]: I1014 09:55:23.290390 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-gwtg4" Oct 14 09:55:23 crc kubenswrapper[4870]: I1014 09:55:23.290822 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-gwtg4" Oct 14 09:55:23 crc kubenswrapper[4870]: I1014 09:55:23.366426 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-gwtg4" Oct 14 09:55:33 crc kubenswrapper[4870]: I1014 09:55:33.344542 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-gwtg4" Oct 14 09:55:33 crc kubenswrapper[4870]: I1014 09:55:33.415646 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-gwtg4"] Oct 14 09:55:33 crc kubenswrapper[4870]: I1014 09:55:33.571569 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-gwtg4" podUID="52b3807b-ba2a-4bb7-9b0c-07fb73fbb41d" containerName="registry-server" containerID="cri-o://07f5aa211089a65625149f4fd175c74550a2427826036a4061c53be5ac39d90b" gracePeriod=2 Oct 14 09:55:34 crc kubenswrapper[4870]: I1014 09:55:34.082323 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gwtg4" Oct 14 09:55:34 crc kubenswrapper[4870]: I1014 09:55:34.098039 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/52b3807b-ba2a-4bb7-9b0c-07fb73fbb41d-catalog-content\") pod \"52b3807b-ba2a-4bb7-9b0c-07fb73fbb41d\" (UID: \"52b3807b-ba2a-4bb7-9b0c-07fb73fbb41d\") " Oct 14 09:55:34 crc kubenswrapper[4870]: I1014 09:55:34.098123 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/52b3807b-ba2a-4bb7-9b0c-07fb73fbb41d-utilities\") pod \"52b3807b-ba2a-4bb7-9b0c-07fb73fbb41d\" (UID: \"52b3807b-ba2a-4bb7-9b0c-07fb73fbb41d\") " Oct 14 09:55:34 crc kubenswrapper[4870]: I1014 09:55:34.098336 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6rkdt\" (UniqueName: \"kubernetes.io/projected/52b3807b-ba2a-4bb7-9b0c-07fb73fbb41d-kube-api-access-6rkdt\") pod \"52b3807b-ba2a-4bb7-9b0c-07fb73fbb41d\" (UID: \"52b3807b-ba2a-4bb7-9b0c-07fb73fbb41d\") " Oct 14 09:55:34 crc kubenswrapper[4870]: I1014 09:55:34.100584 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/52b3807b-ba2a-4bb7-9b0c-07fb73fbb41d-utilities" (OuterVolumeSpecName: "utilities") pod "52b3807b-ba2a-4bb7-9b0c-07fb73fbb41d" (UID: "52b3807b-ba2a-4bb7-9b0c-07fb73fbb41d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 09:55:34 crc kubenswrapper[4870]: I1014 09:55:34.104121 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/52b3807b-ba2a-4bb7-9b0c-07fb73fbb41d-kube-api-access-6rkdt" (OuterVolumeSpecName: "kube-api-access-6rkdt") pod "52b3807b-ba2a-4bb7-9b0c-07fb73fbb41d" (UID: "52b3807b-ba2a-4bb7-9b0c-07fb73fbb41d"). InnerVolumeSpecName "kube-api-access-6rkdt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:55:34 crc kubenswrapper[4870]: I1014 09:55:34.177552 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/52b3807b-ba2a-4bb7-9b0c-07fb73fbb41d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "52b3807b-ba2a-4bb7-9b0c-07fb73fbb41d" (UID: "52b3807b-ba2a-4bb7-9b0c-07fb73fbb41d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 09:55:34 crc kubenswrapper[4870]: I1014 09:55:34.200060 4870 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/52b3807b-ba2a-4bb7-9b0c-07fb73fbb41d-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 09:55:34 crc kubenswrapper[4870]: I1014 09:55:34.200098 4870 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/52b3807b-ba2a-4bb7-9b0c-07fb73fbb41d-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 09:55:34 crc kubenswrapper[4870]: I1014 09:55:34.200109 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6rkdt\" (UniqueName: \"kubernetes.io/projected/52b3807b-ba2a-4bb7-9b0c-07fb73fbb41d-kube-api-access-6rkdt\") on node \"crc\" DevicePath \"\"" Oct 14 09:55:34 crc kubenswrapper[4870]: I1014 09:55:34.583394 4870 generic.go:334] "Generic (PLEG): container finished" podID="52b3807b-ba2a-4bb7-9b0c-07fb73fbb41d" containerID="07f5aa211089a65625149f4fd175c74550a2427826036a4061c53be5ac39d90b" exitCode=0 Oct 14 09:55:34 crc kubenswrapper[4870]: I1014 09:55:34.583464 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gwtg4" event={"ID":"52b3807b-ba2a-4bb7-9b0c-07fb73fbb41d","Type":"ContainerDied","Data":"07f5aa211089a65625149f4fd175c74550a2427826036a4061c53be5ac39d90b"} Oct 14 09:55:34 crc kubenswrapper[4870]: I1014 09:55:34.583497 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gwtg4" event={"ID":"52b3807b-ba2a-4bb7-9b0c-07fb73fbb41d","Type":"ContainerDied","Data":"0ee2dd76d8bb55b9d5e9028ab91f781ab2902c6322f61b31c4339fc0fc31e4f6"} Oct 14 09:55:34 crc kubenswrapper[4870]: I1014 09:55:34.583522 4870 scope.go:117] "RemoveContainer" containerID="07f5aa211089a65625149f4fd175c74550a2427826036a4061c53be5ac39d90b" Oct 14 09:55:34 crc kubenswrapper[4870]: I1014 09:55:34.583685 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gwtg4" Oct 14 09:55:34 crc kubenswrapper[4870]: I1014 09:55:34.614618 4870 scope.go:117] "RemoveContainer" containerID="f19d25bd04e805b73e5dfbd87b8225f23700ef892c5fe2da67b78a0bf902eff3" Oct 14 09:55:34 crc kubenswrapper[4870]: I1014 09:55:34.651643 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-gwtg4"] Oct 14 09:55:34 crc kubenswrapper[4870]: I1014 09:55:34.670278 4870 scope.go:117] "RemoveContainer" containerID="24002b237d7cc2bdeff1852042fe37310e22e47f3f47daae5260f73ac13c2c47" Oct 14 09:55:34 crc kubenswrapper[4870]: I1014 09:55:34.677694 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-gwtg4"] Oct 14 09:55:34 crc kubenswrapper[4870]: I1014 09:55:34.706869 4870 scope.go:117] "RemoveContainer" containerID="07f5aa211089a65625149f4fd175c74550a2427826036a4061c53be5ac39d90b" Oct 14 09:55:34 crc kubenswrapper[4870]: E1014 09:55:34.707316 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"07f5aa211089a65625149f4fd175c74550a2427826036a4061c53be5ac39d90b\": container with ID starting with 07f5aa211089a65625149f4fd175c74550a2427826036a4061c53be5ac39d90b not found: ID does not exist" containerID="07f5aa211089a65625149f4fd175c74550a2427826036a4061c53be5ac39d90b" Oct 14 09:55:34 crc kubenswrapper[4870]: I1014 09:55:34.707345 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"07f5aa211089a65625149f4fd175c74550a2427826036a4061c53be5ac39d90b"} err="failed to get container status \"07f5aa211089a65625149f4fd175c74550a2427826036a4061c53be5ac39d90b\": rpc error: code = NotFound desc = could not find container \"07f5aa211089a65625149f4fd175c74550a2427826036a4061c53be5ac39d90b\": container with ID starting with 07f5aa211089a65625149f4fd175c74550a2427826036a4061c53be5ac39d90b not found: ID does not exist" Oct 14 09:55:34 crc kubenswrapper[4870]: I1014 09:55:34.707369 4870 scope.go:117] "RemoveContainer" containerID="f19d25bd04e805b73e5dfbd87b8225f23700ef892c5fe2da67b78a0bf902eff3" Oct 14 09:55:34 crc kubenswrapper[4870]: E1014 09:55:34.707684 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f19d25bd04e805b73e5dfbd87b8225f23700ef892c5fe2da67b78a0bf902eff3\": container with ID starting with f19d25bd04e805b73e5dfbd87b8225f23700ef892c5fe2da67b78a0bf902eff3 not found: ID does not exist" containerID="f19d25bd04e805b73e5dfbd87b8225f23700ef892c5fe2da67b78a0bf902eff3" Oct 14 09:55:34 crc kubenswrapper[4870]: I1014 09:55:34.707705 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f19d25bd04e805b73e5dfbd87b8225f23700ef892c5fe2da67b78a0bf902eff3"} err="failed to get container status \"f19d25bd04e805b73e5dfbd87b8225f23700ef892c5fe2da67b78a0bf902eff3\": rpc error: code = NotFound desc = could not find container \"f19d25bd04e805b73e5dfbd87b8225f23700ef892c5fe2da67b78a0bf902eff3\": container with ID starting with f19d25bd04e805b73e5dfbd87b8225f23700ef892c5fe2da67b78a0bf902eff3 not found: ID does not exist" Oct 14 09:55:34 crc kubenswrapper[4870]: I1014 09:55:34.707721 4870 scope.go:117] "RemoveContainer" containerID="24002b237d7cc2bdeff1852042fe37310e22e47f3f47daae5260f73ac13c2c47" Oct 14 09:55:34 crc kubenswrapper[4870]: E1014 09:55:34.707995 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"24002b237d7cc2bdeff1852042fe37310e22e47f3f47daae5260f73ac13c2c47\": container with ID starting with 24002b237d7cc2bdeff1852042fe37310e22e47f3f47daae5260f73ac13c2c47 not found: ID does not exist" containerID="24002b237d7cc2bdeff1852042fe37310e22e47f3f47daae5260f73ac13c2c47" Oct 14 09:55:34 crc kubenswrapper[4870]: I1014 09:55:34.708016 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"24002b237d7cc2bdeff1852042fe37310e22e47f3f47daae5260f73ac13c2c47"} err="failed to get container status \"24002b237d7cc2bdeff1852042fe37310e22e47f3f47daae5260f73ac13c2c47\": rpc error: code = NotFound desc = could not find container \"24002b237d7cc2bdeff1852042fe37310e22e47f3f47daae5260f73ac13c2c47\": container with ID starting with 24002b237d7cc2bdeff1852042fe37310e22e47f3f47daae5260f73ac13c2c47 not found: ID does not exist" Oct 14 09:55:35 crc kubenswrapper[4870]: I1014 09:55:35.058255 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="52b3807b-ba2a-4bb7-9b0c-07fb73fbb41d" path="/var/lib/kubelet/pods/52b3807b-ba2a-4bb7-9b0c-07fb73fbb41d/volumes" Oct 14 09:55:37 crc kubenswrapper[4870]: I1014 09:55:37.033935 4870 scope.go:117] "RemoveContainer" containerID="c52e0ba5edaea8b07f60b9f76084bc5c51b52575d871193fc95ac2cefd996fea" Oct 14 09:55:37 crc kubenswrapper[4870]: E1014 09:55:37.034641 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 09:55:49 crc kubenswrapper[4870]: I1014 09:55:49.035995 4870 scope.go:117] "RemoveContainer" containerID="c52e0ba5edaea8b07f60b9f76084bc5c51b52575d871193fc95ac2cefd996fea" Oct 14 09:55:49 crc kubenswrapper[4870]: E1014 09:55:49.037008 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 09:56:00 crc kubenswrapper[4870]: I1014 09:56:00.036034 4870 scope.go:117] "RemoveContainer" containerID="c52e0ba5edaea8b07f60b9f76084bc5c51b52575d871193fc95ac2cefd996fea" Oct 14 09:56:00 crc kubenswrapper[4870]: E1014 09:56:00.037288 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 09:56:11 crc kubenswrapper[4870]: I1014 09:56:11.043587 4870 scope.go:117] "RemoveContainer" containerID="c52e0ba5edaea8b07f60b9f76084bc5c51b52575d871193fc95ac2cefd996fea" Oct 14 09:56:11 crc kubenswrapper[4870]: E1014 09:56:11.044797 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 09:56:16 crc kubenswrapper[4870]: I1014 09:56:16.158482 4870 generic.go:334] "Generic (PLEG): container finished" podID="2ff165ad-7fb3-4c5a-918b-94950ba97a0a" containerID="a6cd11bd728b708b6e281bb65a06fa0509f20de40fc02399a905b0cc430b8ce5" exitCode=0 Oct 14 09:56:16 crc kubenswrapper[4870]: I1014 09:56:16.158566 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell7hdbp" event={"ID":"2ff165ad-7fb3-4c5a-918b-94950ba97a0a","Type":"ContainerDied","Data":"a6cd11bd728b708b6e281bb65a06fa0509f20de40fc02399a905b0cc430b8ce5"} Oct 14 09:56:17 crc kubenswrapper[4870]: I1014 09:56:17.720163 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell7hdbp" Oct 14 09:56:17 crc kubenswrapper[4870]: I1014 09:56:17.745285 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/2ff165ad-7fb3-4c5a-918b-94950ba97a0a-nova-cell1-compute-config-1\") pod \"2ff165ad-7fb3-4c5a-918b-94950ba97a0a\" (UID: \"2ff165ad-7fb3-4c5a-918b-94950ba97a0a\") " Oct 14 09:56:17 crc kubenswrapper[4870]: I1014 09:56:17.745382 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b8fh4\" (UniqueName: \"kubernetes.io/projected/2ff165ad-7fb3-4c5a-918b-94950ba97a0a-kube-api-access-b8fh4\") pod \"2ff165ad-7fb3-4c5a-918b-94950ba97a0a\" (UID: \"2ff165ad-7fb3-4c5a-918b-94950ba97a0a\") " Oct 14 09:56:17 crc kubenswrapper[4870]: I1014 09:56:17.745455 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/2ff165ad-7fb3-4c5a-918b-94950ba97a0a-nova-cells-global-config-0\") pod \"2ff165ad-7fb3-4c5a-918b-94950ba97a0a\" (UID: \"2ff165ad-7fb3-4c5a-918b-94950ba97a0a\") " Oct 14 09:56:17 crc kubenswrapper[4870]: I1014 09:56:17.745484 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/2ff165ad-7fb3-4c5a-918b-94950ba97a0a-nova-migration-ssh-key-1\") pod \"2ff165ad-7fb3-4c5a-918b-94950ba97a0a\" (UID: \"2ff165ad-7fb3-4c5a-918b-94950ba97a0a\") " Oct 14 09:56:17 crc kubenswrapper[4870]: I1014 09:56:17.745512 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/2ff165ad-7fb3-4c5a-918b-94950ba97a0a-nova-cells-global-config-1\") pod \"2ff165ad-7fb3-4c5a-918b-94950ba97a0a\" (UID: \"2ff165ad-7fb3-4c5a-918b-94950ba97a0a\") " Oct 14 09:56:17 crc kubenswrapper[4870]: I1014 09:56:17.745643 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/2ff165ad-7fb3-4c5a-918b-94950ba97a0a-nova-cell1-compute-config-0\") pod \"2ff165ad-7fb3-4c5a-918b-94950ba97a0a\" (UID: \"2ff165ad-7fb3-4c5a-918b-94950ba97a0a\") " Oct 14 09:56:17 crc kubenswrapper[4870]: I1014 09:56:17.745674 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/2ff165ad-7fb3-4c5a-918b-94950ba97a0a-nova-migration-ssh-key-0\") pod \"2ff165ad-7fb3-4c5a-918b-94950ba97a0a\" (UID: \"2ff165ad-7fb3-4c5a-918b-94950ba97a0a\") " Oct 14 09:56:17 crc kubenswrapper[4870]: I1014 09:56:17.745722 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2ff165ad-7fb3-4c5a-918b-94950ba97a0a-ssh-key\") pod \"2ff165ad-7fb3-4c5a-918b-94950ba97a0a\" (UID: \"2ff165ad-7fb3-4c5a-918b-94950ba97a0a\") " Oct 14 09:56:17 crc kubenswrapper[4870]: I1014 09:56:17.745813 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ff165ad-7fb3-4c5a-918b-94950ba97a0a-nova-cell1-combined-ca-bundle\") pod \"2ff165ad-7fb3-4c5a-918b-94950ba97a0a\" (UID: \"2ff165ad-7fb3-4c5a-918b-94950ba97a0a\") " Oct 14 09:56:17 crc kubenswrapper[4870]: I1014 09:56:17.745850 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/2ff165ad-7fb3-4c5a-918b-94950ba97a0a-ceph\") pod \"2ff165ad-7fb3-4c5a-918b-94950ba97a0a\" (UID: \"2ff165ad-7fb3-4c5a-918b-94950ba97a0a\") " Oct 14 09:56:17 crc kubenswrapper[4870]: I1014 09:56:17.745884 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2ff165ad-7fb3-4c5a-918b-94950ba97a0a-inventory\") pod \"2ff165ad-7fb3-4c5a-918b-94950ba97a0a\" (UID: \"2ff165ad-7fb3-4c5a-918b-94950ba97a0a\") " Oct 14 09:56:17 crc kubenswrapper[4870]: I1014 09:56:17.752643 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2ff165ad-7fb3-4c5a-918b-94950ba97a0a-ceph" (OuterVolumeSpecName: "ceph") pod "2ff165ad-7fb3-4c5a-918b-94950ba97a0a" (UID: "2ff165ad-7fb3-4c5a-918b-94950ba97a0a"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:56:17 crc kubenswrapper[4870]: I1014 09:56:17.759263 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2ff165ad-7fb3-4c5a-918b-94950ba97a0a-nova-cell1-combined-ca-bundle" (OuterVolumeSpecName: "nova-cell1-combined-ca-bundle") pod "2ff165ad-7fb3-4c5a-918b-94950ba97a0a" (UID: "2ff165ad-7fb3-4c5a-918b-94950ba97a0a"). InnerVolumeSpecName "nova-cell1-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:56:17 crc kubenswrapper[4870]: I1014 09:56:17.759294 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2ff165ad-7fb3-4c5a-918b-94950ba97a0a-kube-api-access-b8fh4" (OuterVolumeSpecName: "kube-api-access-b8fh4") pod "2ff165ad-7fb3-4c5a-918b-94950ba97a0a" (UID: "2ff165ad-7fb3-4c5a-918b-94950ba97a0a"). InnerVolumeSpecName "kube-api-access-b8fh4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:56:17 crc kubenswrapper[4870]: I1014 09:56:17.779091 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2ff165ad-7fb3-4c5a-918b-94950ba97a0a-nova-cells-global-config-1" (OuterVolumeSpecName: "nova-cells-global-config-1") pod "2ff165ad-7fb3-4c5a-918b-94950ba97a0a" (UID: "2ff165ad-7fb3-4c5a-918b-94950ba97a0a"). InnerVolumeSpecName "nova-cells-global-config-1". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 09:56:17 crc kubenswrapper[4870]: I1014 09:56:17.784687 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2ff165ad-7fb3-4c5a-918b-94950ba97a0a-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "2ff165ad-7fb3-4c5a-918b-94950ba97a0a" (UID: "2ff165ad-7fb3-4c5a-918b-94950ba97a0a"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:56:17 crc kubenswrapper[4870]: I1014 09:56:17.784710 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2ff165ad-7fb3-4c5a-918b-94950ba97a0a-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "2ff165ad-7fb3-4c5a-918b-94950ba97a0a" (UID: "2ff165ad-7fb3-4c5a-918b-94950ba97a0a"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:56:17 crc kubenswrapper[4870]: I1014 09:56:17.801405 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2ff165ad-7fb3-4c5a-918b-94950ba97a0a-nova-cells-global-config-0" (OuterVolumeSpecName: "nova-cells-global-config-0") pod "2ff165ad-7fb3-4c5a-918b-94950ba97a0a" (UID: "2ff165ad-7fb3-4c5a-918b-94950ba97a0a"). InnerVolumeSpecName "nova-cells-global-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 09:56:17 crc kubenswrapper[4870]: I1014 09:56:17.808226 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2ff165ad-7fb3-4c5a-918b-94950ba97a0a-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "2ff165ad-7fb3-4c5a-918b-94950ba97a0a" (UID: "2ff165ad-7fb3-4c5a-918b-94950ba97a0a"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:56:17 crc kubenswrapper[4870]: I1014 09:56:17.810228 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2ff165ad-7fb3-4c5a-918b-94950ba97a0a-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "2ff165ad-7fb3-4c5a-918b-94950ba97a0a" (UID: "2ff165ad-7fb3-4c5a-918b-94950ba97a0a"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:56:17 crc kubenswrapper[4870]: I1014 09:56:17.810366 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2ff165ad-7fb3-4c5a-918b-94950ba97a0a-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "2ff165ad-7fb3-4c5a-918b-94950ba97a0a" (UID: "2ff165ad-7fb3-4c5a-918b-94950ba97a0a"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:56:17 crc kubenswrapper[4870]: I1014 09:56:17.811650 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2ff165ad-7fb3-4c5a-918b-94950ba97a0a-inventory" (OuterVolumeSpecName: "inventory") pod "2ff165ad-7fb3-4c5a-918b-94950ba97a0a" (UID: "2ff165ad-7fb3-4c5a-918b-94950ba97a0a"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:56:17 crc kubenswrapper[4870]: I1014 09:56:17.848612 4870 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ff165ad-7fb3-4c5a-918b-94950ba97a0a-nova-cell1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 09:56:17 crc kubenswrapper[4870]: I1014 09:56:17.848648 4870 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/2ff165ad-7fb3-4c5a-918b-94950ba97a0a-ceph\") on node \"crc\" DevicePath \"\"" Oct 14 09:56:17 crc kubenswrapper[4870]: I1014 09:56:17.848658 4870 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2ff165ad-7fb3-4c5a-918b-94950ba97a0a-inventory\") on node \"crc\" DevicePath \"\"" Oct 14 09:56:17 crc kubenswrapper[4870]: I1014 09:56:17.848669 4870 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/2ff165ad-7fb3-4c5a-918b-94950ba97a0a-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Oct 14 09:56:17 crc kubenswrapper[4870]: I1014 09:56:17.848679 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b8fh4\" (UniqueName: \"kubernetes.io/projected/2ff165ad-7fb3-4c5a-918b-94950ba97a0a-kube-api-access-b8fh4\") on node \"crc\" DevicePath \"\"" Oct 14 09:56:17 crc kubenswrapper[4870]: I1014 09:56:17.848688 4870 reconciler_common.go:293] "Volume detached for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/2ff165ad-7fb3-4c5a-918b-94950ba97a0a-nova-cells-global-config-0\") on node \"crc\" DevicePath \"\"" Oct 14 09:56:17 crc kubenswrapper[4870]: I1014 09:56:17.848697 4870 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/2ff165ad-7fb3-4c5a-918b-94950ba97a0a-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Oct 14 09:56:17 crc kubenswrapper[4870]: I1014 09:56:17.848706 4870 reconciler_common.go:293] "Volume detached for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/2ff165ad-7fb3-4c5a-918b-94950ba97a0a-nova-cells-global-config-1\") on node \"crc\" DevicePath \"\"" Oct 14 09:56:17 crc kubenswrapper[4870]: I1014 09:56:17.848714 4870 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/2ff165ad-7fb3-4c5a-918b-94950ba97a0a-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Oct 14 09:56:17 crc kubenswrapper[4870]: I1014 09:56:17.848722 4870 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/2ff165ad-7fb3-4c5a-918b-94950ba97a0a-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Oct 14 09:56:17 crc kubenswrapper[4870]: I1014 09:56:17.848730 4870 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2ff165ad-7fb3-4c5a-918b-94950ba97a0a-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 14 09:56:18 crc kubenswrapper[4870]: I1014 09:56:18.190684 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell7hdbp" event={"ID":"2ff165ad-7fb3-4c5a-918b-94950ba97a0a","Type":"ContainerDied","Data":"f499a75bc2f775ad0bc150b8be3cdd70b75177989a8ca75bff2034e70c7d1434"} Oct 14 09:56:18 crc kubenswrapper[4870]: I1014 09:56:18.190752 4870 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f499a75bc2f775ad0bc150b8be3cdd70b75177989a8ca75bff2034e70c7d1434" Oct 14 09:56:18 crc kubenswrapper[4870]: I1014 09:56:18.190833 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell7hdbp" Oct 14 09:56:23 crc kubenswrapper[4870]: I1014 09:56:23.034795 4870 scope.go:117] "RemoveContainer" containerID="c52e0ba5edaea8b07f60b9f76084bc5c51b52575d871193fc95ac2cefd996fea" Oct 14 09:56:23 crc kubenswrapper[4870]: E1014 09:56:23.035875 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 09:56:35 crc kubenswrapper[4870]: I1014 09:56:35.054858 4870 scope.go:117] "RemoveContainer" containerID="c52e0ba5edaea8b07f60b9f76084bc5c51b52575d871193fc95ac2cefd996fea" Oct 14 09:56:35 crc kubenswrapper[4870]: E1014 09:56:35.056810 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 09:56:46 crc kubenswrapper[4870]: I1014 09:56:46.034724 4870 scope.go:117] "RemoveContainer" containerID="c52e0ba5edaea8b07f60b9f76084bc5c51b52575d871193fc95ac2cefd996fea" Oct 14 09:56:46 crc kubenswrapper[4870]: E1014 09:56:46.035732 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 09:56:59 crc kubenswrapper[4870]: I1014 09:56:59.035690 4870 scope.go:117] "RemoveContainer" containerID="c52e0ba5edaea8b07f60b9f76084bc5c51b52575d871193fc95ac2cefd996fea" Oct 14 09:56:59 crc kubenswrapper[4870]: E1014 09:56:59.037169 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 09:57:14 crc kubenswrapper[4870]: I1014 09:57:14.035152 4870 scope.go:117] "RemoveContainer" containerID="c52e0ba5edaea8b07f60b9f76084bc5c51b52575d871193fc95ac2cefd996fea" Oct 14 09:57:14 crc kubenswrapper[4870]: E1014 09:57:14.036650 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 09:57:29 crc kubenswrapper[4870]: I1014 09:57:29.035370 4870 scope.go:117] "RemoveContainer" containerID="c52e0ba5edaea8b07f60b9f76084bc5c51b52575d871193fc95ac2cefd996fea" Oct 14 09:57:29 crc kubenswrapper[4870]: E1014 09:57:29.036563 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 09:57:41 crc kubenswrapper[4870]: I1014 09:57:41.035396 4870 scope.go:117] "RemoveContainer" containerID="c52e0ba5edaea8b07f60b9f76084bc5c51b52575d871193fc95ac2cefd996fea" Oct 14 09:57:41 crc kubenswrapper[4870]: E1014 09:57:41.036643 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 09:57:54 crc kubenswrapper[4870]: I1014 09:57:54.035304 4870 scope.go:117] "RemoveContainer" containerID="c52e0ba5edaea8b07f60b9f76084bc5c51b52575d871193fc95ac2cefd996fea" Oct 14 09:57:54 crc kubenswrapper[4870]: E1014 09:57:54.036290 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 09:58:00 crc kubenswrapper[4870]: I1014 09:58:00.723759 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-copy-data"] Oct 14 09:58:00 crc kubenswrapper[4870]: I1014 09:58:00.724717 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/mariadb-copy-data" podUID="592ad2e9-ae45-4985-92f8-da4f3a76b5cb" containerName="adoption" containerID="cri-o://3dd253c3e0fa1fe373307af5bdeb4a9a5e12f8ee948dd1bc28a8e0c134af8acd" gracePeriod=30 Oct 14 09:58:07 crc kubenswrapper[4870]: I1014 09:58:07.040047 4870 scope.go:117] "RemoveContainer" containerID="c52e0ba5edaea8b07f60b9f76084bc5c51b52575d871193fc95ac2cefd996fea" Oct 14 09:58:07 crc kubenswrapper[4870]: E1014 09:58:07.041204 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 09:58:22 crc kubenswrapper[4870]: I1014 09:58:22.033928 4870 scope.go:117] "RemoveContainer" containerID="c52e0ba5edaea8b07f60b9f76084bc5c51b52575d871193fc95ac2cefd996fea" Oct 14 09:58:22 crc kubenswrapper[4870]: E1014 09:58:22.035010 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 09:58:30 crc kubenswrapper[4870]: I1014 09:58:30.861705 4870 generic.go:334] "Generic (PLEG): container finished" podID="592ad2e9-ae45-4985-92f8-da4f3a76b5cb" containerID="3dd253c3e0fa1fe373307af5bdeb4a9a5e12f8ee948dd1bc28a8e0c134af8acd" exitCode=137 Oct 14 09:58:30 crc kubenswrapper[4870]: I1014 09:58:30.861787 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-copy-data" event={"ID":"592ad2e9-ae45-4985-92f8-da4f3a76b5cb","Type":"ContainerDied","Data":"3dd253c3e0fa1fe373307af5bdeb4a9a5e12f8ee948dd1bc28a8e0c134af8acd"} Oct 14 09:58:31 crc kubenswrapper[4870]: I1014 09:58:31.321048 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-copy-data" Oct 14 09:58:31 crc kubenswrapper[4870]: I1014 09:58:31.464229 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mariadb-data\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2cf7c6e8-938d-4c3b-9b5b-7b262e943b94\") pod \"592ad2e9-ae45-4985-92f8-da4f3a76b5cb\" (UID: \"592ad2e9-ae45-4985-92f8-da4f3a76b5cb\") " Oct 14 09:58:31 crc kubenswrapper[4870]: I1014 09:58:31.464464 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qlct8\" (UniqueName: \"kubernetes.io/projected/592ad2e9-ae45-4985-92f8-da4f3a76b5cb-kube-api-access-qlct8\") pod \"592ad2e9-ae45-4985-92f8-da4f3a76b5cb\" (UID: \"592ad2e9-ae45-4985-92f8-da4f3a76b5cb\") " Oct 14 09:58:31 crc kubenswrapper[4870]: I1014 09:58:31.495903 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/592ad2e9-ae45-4985-92f8-da4f3a76b5cb-kube-api-access-qlct8" (OuterVolumeSpecName: "kube-api-access-qlct8") pod "592ad2e9-ae45-4985-92f8-da4f3a76b5cb" (UID: "592ad2e9-ae45-4985-92f8-da4f3a76b5cb"). InnerVolumeSpecName "kube-api-access-qlct8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:58:31 crc kubenswrapper[4870]: I1014 09:58:31.504271 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2cf7c6e8-938d-4c3b-9b5b-7b262e943b94" (OuterVolumeSpecName: "mariadb-data") pod "592ad2e9-ae45-4985-92f8-da4f3a76b5cb" (UID: "592ad2e9-ae45-4985-92f8-da4f3a76b5cb"). InnerVolumeSpecName "pvc-2cf7c6e8-938d-4c3b-9b5b-7b262e943b94". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 14 09:58:31 crc kubenswrapper[4870]: I1014 09:58:31.567482 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qlct8\" (UniqueName: \"kubernetes.io/projected/592ad2e9-ae45-4985-92f8-da4f3a76b5cb-kube-api-access-qlct8\") on node \"crc\" DevicePath \"\"" Oct 14 09:58:31 crc kubenswrapper[4870]: I1014 09:58:31.567540 4870 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-2cf7c6e8-938d-4c3b-9b5b-7b262e943b94\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2cf7c6e8-938d-4c3b-9b5b-7b262e943b94\") on node \"crc\" " Oct 14 09:58:31 crc kubenswrapper[4870]: I1014 09:58:31.593811 4870 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Oct 14 09:58:31 crc kubenswrapper[4870]: I1014 09:58:31.593971 4870 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-2cf7c6e8-938d-4c3b-9b5b-7b262e943b94" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2cf7c6e8-938d-4c3b-9b5b-7b262e943b94") on node "crc" Oct 14 09:58:31 crc kubenswrapper[4870]: I1014 09:58:31.669199 4870 reconciler_common.go:293] "Volume detached for volume \"pvc-2cf7c6e8-938d-4c3b-9b5b-7b262e943b94\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2cf7c6e8-938d-4c3b-9b5b-7b262e943b94\") on node \"crc\" DevicePath \"\"" Oct 14 09:58:31 crc kubenswrapper[4870]: I1014 09:58:31.880585 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-copy-data" event={"ID":"592ad2e9-ae45-4985-92f8-da4f3a76b5cb","Type":"ContainerDied","Data":"87cf07d511e054223dd4255a5b936e615fb5a362a3619fb51149bd78ff00a135"} Oct 14 09:58:31 crc kubenswrapper[4870]: I1014 09:58:31.880665 4870 scope.go:117] "RemoveContainer" containerID="3dd253c3e0fa1fe373307af5bdeb4a9a5e12f8ee948dd1bc28a8e0c134af8acd" Oct 14 09:58:31 crc kubenswrapper[4870]: I1014 09:58:31.880731 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-copy-data" Oct 14 09:58:31 crc kubenswrapper[4870]: I1014 09:58:31.938842 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-copy-data"] Oct 14 09:58:31 crc kubenswrapper[4870]: I1014 09:58:31.946762 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-copy-data"] Oct 14 09:58:32 crc kubenswrapper[4870]: I1014 09:58:32.570313 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-copy-data"] Oct 14 09:58:32 crc kubenswrapper[4870]: I1014 09:58:32.571100 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-copy-data" podUID="3cfdfc3c-8b85-4145-98ad-64c9f91c6860" containerName="adoption" containerID="cri-o://e4e6ab17398f4be1ba400d1713cb3dbe0d2c2b8ecdebc9c067675e1465fa9fff" gracePeriod=30 Oct 14 09:58:33 crc kubenswrapper[4870]: I1014 09:58:33.048603 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="592ad2e9-ae45-4985-92f8-da4f3a76b5cb" path="/var/lib/kubelet/pods/592ad2e9-ae45-4985-92f8-da4f3a76b5cb/volumes" Oct 14 09:58:34 crc kubenswrapper[4870]: I1014 09:58:34.034862 4870 scope.go:117] "RemoveContainer" containerID="c52e0ba5edaea8b07f60b9f76084bc5c51b52575d871193fc95ac2cefd996fea" Oct 14 09:58:34 crc kubenswrapper[4870]: I1014 09:58:34.919122 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" event={"ID":"874a5e45-dffd-4d17-b609-b7d3ed2eab07","Type":"ContainerStarted","Data":"fa43fd954e8c6e0aade481bc7b524b46d5559be95fdfa58bfe7f0b926272a5a5"} Oct 14 09:59:03 crc kubenswrapper[4870]: I1014 09:59:03.172123 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-copy-data" Oct 14 09:59:03 crc kubenswrapper[4870]: I1014 09:59:03.260033 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-data\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-64243ea5-9586-429f-ab5b-4806ef5e2614\") pod \"3cfdfc3c-8b85-4145-98ad-64c9f91c6860\" (UID: \"3cfdfc3c-8b85-4145-98ad-64c9f91c6860\") " Oct 14 09:59:03 crc kubenswrapper[4870]: I1014 09:59:03.260186 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-data-cert\" (UniqueName: \"kubernetes.io/secret/3cfdfc3c-8b85-4145-98ad-64c9f91c6860-ovn-data-cert\") pod \"3cfdfc3c-8b85-4145-98ad-64c9f91c6860\" (UID: \"3cfdfc3c-8b85-4145-98ad-64c9f91c6860\") " Oct 14 09:59:03 crc kubenswrapper[4870]: I1014 09:59:03.260237 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f5zx4\" (UniqueName: \"kubernetes.io/projected/3cfdfc3c-8b85-4145-98ad-64c9f91c6860-kube-api-access-f5zx4\") pod \"3cfdfc3c-8b85-4145-98ad-64c9f91c6860\" (UID: \"3cfdfc3c-8b85-4145-98ad-64c9f91c6860\") " Oct 14 09:59:03 crc kubenswrapper[4870]: I1014 09:59:03.265804 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cfdfc3c-8b85-4145-98ad-64c9f91c6860-kube-api-access-f5zx4" (OuterVolumeSpecName: "kube-api-access-f5zx4") pod "3cfdfc3c-8b85-4145-98ad-64c9f91c6860" (UID: "3cfdfc3c-8b85-4145-98ad-64c9f91c6860"). InnerVolumeSpecName "kube-api-access-f5zx4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 09:59:03 crc kubenswrapper[4870]: I1014 09:59:03.266099 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3cfdfc3c-8b85-4145-98ad-64c9f91c6860-ovn-data-cert" (OuterVolumeSpecName: "ovn-data-cert") pod "3cfdfc3c-8b85-4145-98ad-64c9f91c6860" (UID: "3cfdfc3c-8b85-4145-98ad-64c9f91c6860"). InnerVolumeSpecName "ovn-data-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 09:59:03 crc kubenswrapper[4870]: I1014 09:59:03.281312 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-64243ea5-9586-429f-ab5b-4806ef5e2614" (OuterVolumeSpecName: "ovn-data") pod "3cfdfc3c-8b85-4145-98ad-64c9f91c6860" (UID: "3cfdfc3c-8b85-4145-98ad-64c9f91c6860"). InnerVolumeSpecName "pvc-64243ea5-9586-429f-ab5b-4806ef5e2614". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 14 09:59:03 crc kubenswrapper[4870]: I1014 09:59:03.312986 4870 generic.go:334] "Generic (PLEG): container finished" podID="3cfdfc3c-8b85-4145-98ad-64c9f91c6860" containerID="e4e6ab17398f4be1ba400d1713cb3dbe0d2c2b8ecdebc9c067675e1465fa9fff" exitCode=137 Oct 14 09:59:03 crc kubenswrapper[4870]: I1014 09:59:03.313105 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-copy-data" Oct 14 09:59:03 crc kubenswrapper[4870]: I1014 09:59:03.313095 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-copy-data" event={"ID":"3cfdfc3c-8b85-4145-98ad-64c9f91c6860","Type":"ContainerDied","Data":"e4e6ab17398f4be1ba400d1713cb3dbe0d2c2b8ecdebc9c067675e1465fa9fff"} Oct 14 09:59:03 crc kubenswrapper[4870]: I1014 09:59:03.313399 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-copy-data" event={"ID":"3cfdfc3c-8b85-4145-98ad-64c9f91c6860","Type":"ContainerDied","Data":"e650e748983644c43cb483fe100167fe33ff6c31e22306a1be4c4a77e889b141"} Oct 14 09:59:03 crc kubenswrapper[4870]: I1014 09:59:03.313423 4870 scope.go:117] "RemoveContainer" containerID="e4e6ab17398f4be1ba400d1713cb3dbe0d2c2b8ecdebc9c067675e1465fa9fff" Oct 14 09:59:03 crc kubenswrapper[4870]: I1014 09:59:03.365239 4870 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-64243ea5-9586-429f-ab5b-4806ef5e2614\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-64243ea5-9586-429f-ab5b-4806ef5e2614\") on node \"crc\" " Oct 14 09:59:03 crc kubenswrapper[4870]: I1014 09:59:03.365279 4870 reconciler_common.go:293] "Volume detached for volume \"ovn-data-cert\" (UniqueName: \"kubernetes.io/secret/3cfdfc3c-8b85-4145-98ad-64c9f91c6860-ovn-data-cert\") on node \"crc\" DevicePath \"\"" Oct 14 09:59:03 crc kubenswrapper[4870]: I1014 09:59:03.365334 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f5zx4\" (UniqueName: \"kubernetes.io/projected/3cfdfc3c-8b85-4145-98ad-64c9f91c6860-kube-api-access-f5zx4\") on node \"crc\" DevicePath \"\"" Oct 14 09:59:03 crc kubenswrapper[4870]: I1014 09:59:03.373191 4870 scope.go:117] "RemoveContainer" containerID="e4e6ab17398f4be1ba400d1713cb3dbe0d2c2b8ecdebc9c067675e1465fa9fff" Oct 14 09:59:03 crc kubenswrapper[4870]: E1014 09:59:03.373875 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e4e6ab17398f4be1ba400d1713cb3dbe0d2c2b8ecdebc9c067675e1465fa9fff\": container with ID starting with e4e6ab17398f4be1ba400d1713cb3dbe0d2c2b8ecdebc9c067675e1465fa9fff not found: ID does not exist" containerID="e4e6ab17398f4be1ba400d1713cb3dbe0d2c2b8ecdebc9c067675e1465fa9fff" Oct 14 09:59:03 crc kubenswrapper[4870]: I1014 09:59:03.373931 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e4e6ab17398f4be1ba400d1713cb3dbe0d2c2b8ecdebc9c067675e1465fa9fff"} err="failed to get container status \"e4e6ab17398f4be1ba400d1713cb3dbe0d2c2b8ecdebc9c067675e1465fa9fff\": rpc error: code = NotFound desc = could not find container \"e4e6ab17398f4be1ba400d1713cb3dbe0d2c2b8ecdebc9c067675e1465fa9fff\": container with ID starting with e4e6ab17398f4be1ba400d1713cb3dbe0d2c2b8ecdebc9c067675e1465fa9fff not found: ID does not exist" Oct 14 09:59:03 crc kubenswrapper[4870]: I1014 09:59:03.374378 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-copy-data"] Oct 14 09:59:03 crc kubenswrapper[4870]: I1014 09:59:03.383386 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-copy-data"] Oct 14 09:59:03 crc kubenswrapper[4870]: I1014 09:59:03.410771 4870 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Oct 14 09:59:03 crc kubenswrapper[4870]: I1014 09:59:03.411147 4870 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-64243ea5-9586-429f-ab5b-4806ef5e2614" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-64243ea5-9586-429f-ab5b-4806ef5e2614") on node "crc" Oct 14 09:59:03 crc kubenswrapper[4870]: I1014 09:59:03.466414 4870 reconciler_common.go:293] "Volume detached for volume \"pvc-64243ea5-9586-429f-ab5b-4806ef5e2614\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-64243ea5-9586-429f-ab5b-4806ef5e2614\") on node \"crc\" DevicePath \"\"" Oct 14 09:59:05 crc kubenswrapper[4870]: I1014 09:59:05.049752 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cfdfc3c-8b85-4145-98ad-64c9f91c6860" path="/var/lib/kubelet/pods/3cfdfc3c-8b85-4145-98ad-64c9f91c6860/volumes" Oct 14 10:00:00 crc kubenswrapper[4870]: I1014 10:00:00.202530 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29340600-8bfb4"] Oct 14 10:00:00 crc kubenswrapper[4870]: E1014 10:00:00.203784 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="592ad2e9-ae45-4985-92f8-da4f3a76b5cb" containerName="adoption" Oct 14 10:00:00 crc kubenswrapper[4870]: I1014 10:00:00.203807 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="592ad2e9-ae45-4985-92f8-da4f3a76b5cb" containerName="adoption" Oct 14 10:00:00 crc kubenswrapper[4870]: E1014 10:00:00.203844 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f4a1094-671e-4c21-ba5c-c7f54bd9d2ae" containerName="registry-server" Oct 14 10:00:00 crc kubenswrapper[4870]: I1014 10:00:00.203858 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f4a1094-671e-4c21-ba5c-c7f54bd9d2ae" containerName="registry-server" Oct 14 10:00:00 crc kubenswrapper[4870]: E1014 10:00:00.203879 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="52b3807b-ba2a-4bb7-9b0c-07fb73fbb41d" containerName="extract-content" Oct 14 10:00:00 crc kubenswrapper[4870]: I1014 10:00:00.203892 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="52b3807b-ba2a-4bb7-9b0c-07fb73fbb41d" containerName="extract-content" Oct 14 10:00:00 crc kubenswrapper[4870]: E1014 10:00:00.203925 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ff165ad-7fb3-4c5a-918b-94950ba97a0a" containerName="nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell1" Oct 14 10:00:00 crc kubenswrapper[4870]: I1014 10:00:00.203942 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ff165ad-7fb3-4c5a-918b-94950ba97a0a" containerName="nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell1" Oct 14 10:00:00 crc kubenswrapper[4870]: E1014 10:00:00.203960 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="52b3807b-ba2a-4bb7-9b0c-07fb73fbb41d" containerName="registry-server" Oct 14 10:00:00 crc kubenswrapper[4870]: I1014 10:00:00.203973 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="52b3807b-ba2a-4bb7-9b0c-07fb73fbb41d" containerName="registry-server" Oct 14 10:00:00 crc kubenswrapper[4870]: E1014 10:00:00.204023 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3cfdfc3c-8b85-4145-98ad-64c9f91c6860" containerName="adoption" Oct 14 10:00:00 crc kubenswrapper[4870]: I1014 10:00:00.204037 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="3cfdfc3c-8b85-4145-98ad-64c9f91c6860" containerName="adoption" Oct 14 10:00:00 crc kubenswrapper[4870]: E1014 10:00:00.204071 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f4a1094-671e-4c21-ba5c-c7f54bd9d2ae" containerName="extract-utilities" Oct 14 10:00:00 crc kubenswrapper[4870]: I1014 10:00:00.204085 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f4a1094-671e-4c21-ba5c-c7f54bd9d2ae" containerName="extract-utilities" Oct 14 10:00:00 crc kubenswrapper[4870]: E1014 10:00:00.204110 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f4a1094-671e-4c21-ba5c-c7f54bd9d2ae" containerName="extract-content" Oct 14 10:00:00 crc kubenswrapper[4870]: I1014 10:00:00.204123 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f4a1094-671e-4c21-ba5c-c7f54bd9d2ae" containerName="extract-content" Oct 14 10:00:00 crc kubenswrapper[4870]: E1014 10:00:00.204146 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="52b3807b-ba2a-4bb7-9b0c-07fb73fbb41d" containerName="extract-utilities" Oct 14 10:00:00 crc kubenswrapper[4870]: I1014 10:00:00.204159 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="52b3807b-ba2a-4bb7-9b0c-07fb73fbb41d" containerName="extract-utilities" Oct 14 10:00:00 crc kubenswrapper[4870]: I1014 10:00:00.204584 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="52b3807b-ba2a-4bb7-9b0c-07fb73fbb41d" containerName="registry-server" Oct 14 10:00:00 crc kubenswrapper[4870]: I1014 10:00:00.204618 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="3f4a1094-671e-4c21-ba5c-c7f54bd9d2ae" containerName="registry-server" Oct 14 10:00:00 crc kubenswrapper[4870]: I1014 10:00:00.204671 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="2ff165ad-7fb3-4c5a-918b-94950ba97a0a" containerName="nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell1" Oct 14 10:00:00 crc kubenswrapper[4870]: I1014 10:00:00.204693 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="3cfdfc3c-8b85-4145-98ad-64c9f91c6860" containerName="adoption" Oct 14 10:00:00 crc kubenswrapper[4870]: I1014 10:00:00.204715 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="592ad2e9-ae45-4985-92f8-da4f3a76b5cb" containerName="adoption" Oct 14 10:00:00 crc kubenswrapper[4870]: I1014 10:00:00.205992 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29340600-8bfb4" Oct 14 10:00:00 crc kubenswrapper[4870]: I1014 10:00:00.210295 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 14 10:00:00 crc kubenswrapper[4870]: I1014 10:00:00.210357 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 14 10:00:00 crc kubenswrapper[4870]: I1014 10:00:00.221516 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29340600-8bfb4"] Oct 14 10:00:00 crc kubenswrapper[4870]: I1014 10:00:00.320663 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g7p7p\" (UniqueName: \"kubernetes.io/projected/3ab047ce-5401-4b0d-9820-fd0c8a944d65-kube-api-access-g7p7p\") pod \"collect-profiles-29340600-8bfb4\" (UID: \"3ab047ce-5401-4b0d-9820-fd0c8a944d65\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340600-8bfb4" Oct 14 10:00:00 crc kubenswrapper[4870]: I1014 10:00:00.320783 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3ab047ce-5401-4b0d-9820-fd0c8a944d65-secret-volume\") pod \"collect-profiles-29340600-8bfb4\" (UID: \"3ab047ce-5401-4b0d-9820-fd0c8a944d65\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340600-8bfb4" Oct 14 10:00:00 crc kubenswrapper[4870]: I1014 10:00:00.320977 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3ab047ce-5401-4b0d-9820-fd0c8a944d65-config-volume\") pod \"collect-profiles-29340600-8bfb4\" (UID: \"3ab047ce-5401-4b0d-9820-fd0c8a944d65\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340600-8bfb4" Oct 14 10:00:00 crc kubenswrapper[4870]: I1014 10:00:00.423207 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3ab047ce-5401-4b0d-9820-fd0c8a944d65-config-volume\") pod \"collect-profiles-29340600-8bfb4\" (UID: \"3ab047ce-5401-4b0d-9820-fd0c8a944d65\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340600-8bfb4" Oct 14 10:00:00 crc kubenswrapper[4870]: I1014 10:00:00.423342 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g7p7p\" (UniqueName: \"kubernetes.io/projected/3ab047ce-5401-4b0d-9820-fd0c8a944d65-kube-api-access-g7p7p\") pod \"collect-profiles-29340600-8bfb4\" (UID: \"3ab047ce-5401-4b0d-9820-fd0c8a944d65\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340600-8bfb4" Oct 14 10:00:00 crc kubenswrapper[4870]: I1014 10:00:00.423417 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3ab047ce-5401-4b0d-9820-fd0c8a944d65-secret-volume\") pod \"collect-profiles-29340600-8bfb4\" (UID: \"3ab047ce-5401-4b0d-9820-fd0c8a944d65\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340600-8bfb4" Oct 14 10:00:00 crc kubenswrapper[4870]: I1014 10:00:00.423972 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3ab047ce-5401-4b0d-9820-fd0c8a944d65-config-volume\") pod \"collect-profiles-29340600-8bfb4\" (UID: \"3ab047ce-5401-4b0d-9820-fd0c8a944d65\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340600-8bfb4" Oct 14 10:00:00 crc kubenswrapper[4870]: I1014 10:00:00.431745 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3ab047ce-5401-4b0d-9820-fd0c8a944d65-secret-volume\") pod \"collect-profiles-29340600-8bfb4\" (UID: \"3ab047ce-5401-4b0d-9820-fd0c8a944d65\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340600-8bfb4" Oct 14 10:00:00 crc kubenswrapper[4870]: I1014 10:00:00.438418 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g7p7p\" (UniqueName: \"kubernetes.io/projected/3ab047ce-5401-4b0d-9820-fd0c8a944d65-kube-api-access-g7p7p\") pod \"collect-profiles-29340600-8bfb4\" (UID: \"3ab047ce-5401-4b0d-9820-fd0c8a944d65\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340600-8bfb4" Oct 14 10:00:00 crc kubenswrapper[4870]: I1014 10:00:00.547875 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29340600-8bfb4" Oct 14 10:00:01 crc kubenswrapper[4870]: I1014 10:00:01.242145 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29340600-8bfb4"] Oct 14 10:00:02 crc kubenswrapper[4870]: I1014 10:00:02.131052 4870 generic.go:334] "Generic (PLEG): container finished" podID="3ab047ce-5401-4b0d-9820-fd0c8a944d65" containerID="9b86c65e5e03d31f848937e860117e041573a844755ffe3ea478b9f53bc0f5c2" exitCode=0 Oct 14 10:00:02 crc kubenswrapper[4870]: I1014 10:00:02.131111 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29340600-8bfb4" event={"ID":"3ab047ce-5401-4b0d-9820-fd0c8a944d65","Type":"ContainerDied","Data":"9b86c65e5e03d31f848937e860117e041573a844755ffe3ea478b9f53bc0f5c2"} Oct 14 10:00:02 crc kubenswrapper[4870]: I1014 10:00:02.131409 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29340600-8bfb4" event={"ID":"3ab047ce-5401-4b0d-9820-fd0c8a944d65","Type":"ContainerStarted","Data":"ec52ff8917960901178dd71ba3ec2b7c10a7fbeef2c52411aca70cca68254bea"} Oct 14 10:00:03 crc kubenswrapper[4870]: I1014 10:00:03.618594 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29340600-8bfb4" Oct 14 10:00:03 crc kubenswrapper[4870]: I1014 10:00:03.698127 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3ab047ce-5401-4b0d-9820-fd0c8a944d65-config-volume\") pod \"3ab047ce-5401-4b0d-9820-fd0c8a944d65\" (UID: \"3ab047ce-5401-4b0d-9820-fd0c8a944d65\") " Oct 14 10:00:03 crc kubenswrapper[4870]: I1014 10:00:03.698202 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3ab047ce-5401-4b0d-9820-fd0c8a944d65-secret-volume\") pod \"3ab047ce-5401-4b0d-9820-fd0c8a944d65\" (UID: \"3ab047ce-5401-4b0d-9820-fd0c8a944d65\") " Oct 14 10:00:03 crc kubenswrapper[4870]: I1014 10:00:03.698260 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g7p7p\" (UniqueName: \"kubernetes.io/projected/3ab047ce-5401-4b0d-9820-fd0c8a944d65-kube-api-access-g7p7p\") pod \"3ab047ce-5401-4b0d-9820-fd0c8a944d65\" (UID: \"3ab047ce-5401-4b0d-9820-fd0c8a944d65\") " Oct 14 10:00:03 crc kubenswrapper[4870]: I1014 10:00:03.699190 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3ab047ce-5401-4b0d-9820-fd0c8a944d65-config-volume" (OuterVolumeSpecName: "config-volume") pod "3ab047ce-5401-4b0d-9820-fd0c8a944d65" (UID: "3ab047ce-5401-4b0d-9820-fd0c8a944d65"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 10:00:03 crc kubenswrapper[4870]: I1014 10:00:03.704092 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab047ce-5401-4b0d-9820-fd0c8a944d65-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "3ab047ce-5401-4b0d-9820-fd0c8a944d65" (UID: "3ab047ce-5401-4b0d-9820-fd0c8a944d65"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 10:00:03 crc kubenswrapper[4870]: I1014 10:00:03.704312 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab047ce-5401-4b0d-9820-fd0c8a944d65-kube-api-access-g7p7p" (OuterVolumeSpecName: "kube-api-access-g7p7p") pod "3ab047ce-5401-4b0d-9820-fd0c8a944d65" (UID: "3ab047ce-5401-4b0d-9820-fd0c8a944d65"). InnerVolumeSpecName "kube-api-access-g7p7p". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 10:00:03 crc kubenswrapper[4870]: I1014 10:00:03.801218 4870 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3ab047ce-5401-4b0d-9820-fd0c8a944d65-config-volume\") on node \"crc\" DevicePath \"\"" Oct 14 10:00:03 crc kubenswrapper[4870]: I1014 10:00:03.801259 4870 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3ab047ce-5401-4b0d-9820-fd0c8a944d65-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 14 10:00:03 crc kubenswrapper[4870]: I1014 10:00:03.801274 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g7p7p\" (UniqueName: \"kubernetes.io/projected/3ab047ce-5401-4b0d-9820-fd0c8a944d65-kube-api-access-g7p7p\") on node \"crc\" DevicePath \"\"" Oct 14 10:00:04 crc kubenswrapper[4870]: I1014 10:00:04.154958 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29340600-8bfb4" event={"ID":"3ab047ce-5401-4b0d-9820-fd0c8a944d65","Type":"ContainerDied","Data":"ec52ff8917960901178dd71ba3ec2b7c10a7fbeef2c52411aca70cca68254bea"} Oct 14 10:00:04 crc kubenswrapper[4870]: I1014 10:00:04.155001 4870 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ec52ff8917960901178dd71ba3ec2b7c10a7fbeef2c52411aca70cca68254bea" Oct 14 10:00:04 crc kubenswrapper[4870]: I1014 10:00:04.155060 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29340600-8bfb4" Oct 14 10:00:04 crc kubenswrapper[4870]: I1014 10:00:04.709772 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29340555-qnz96"] Oct 14 10:00:04 crc kubenswrapper[4870]: I1014 10:00:04.720833 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29340555-qnz96"] Oct 14 10:00:05 crc kubenswrapper[4870]: I1014 10:00:05.060906 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7cd0dbb-80e4-4f4b-9942-5082214636d5" path="/var/lib/kubelet/pods/e7cd0dbb-80e4-4f4b-9942-5082214636d5/volumes" Oct 14 10:00:35 crc kubenswrapper[4870]: I1014 10:00:35.506500 4870 scope.go:117] "RemoveContainer" containerID="2616e7acafcec59999241eb7ea10f3a3a4a6900c2e7f284928cb66c4e2f89048" Oct 14 10:00:53 crc kubenswrapper[4870]: I1014 10:00:53.951069 4870 patch_prober.go:28] interesting pod/machine-config-daemon-7tvc7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 10:00:53 crc kubenswrapper[4870]: I1014 10:00:53.951879 4870 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 10:01:00 crc kubenswrapper[4870]: I1014 10:01:00.175940 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29340601-cnz9t"] Oct 14 10:01:00 crc kubenswrapper[4870]: E1014 10:01:00.177181 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3ab047ce-5401-4b0d-9820-fd0c8a944d65" containerName="collect-profiles" Oct 14 10:01:00 crc kubenswrapper[4870]: I1014 10:01:00.177207 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="3ab047ce-5401-4b0d-9820-fd0c8a944d65" containerName="collect-profiles" Oct 14 10:01:00 crc kubenswrapper[4870]: I1014 10:01:00.177685 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="3ab047ce-5401-4b0d-9820-fd0c8a944d65" containerName="collect-profiles" Oct 14 10:01:00 crc kubenswrapper[4870]: I1014 10:01:00.178999 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29340601-cnz9t" Oct 14 10:01:00 crc kubenswrapper[4870]: I1014 10:01:00.241520 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29340601-cnz9t"] Oct 14 10:01:00 crc kubenswrapper[4870]: I1014 10:01:00.242993 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ad4b83a-f2a8-439f-a02e-66424aa17de1-combined-ca-bundle\") pod \"keystone-cron-29340601-cnz9t\" (UID: \"4ad4b83a-f2a8-439f-a02e-66424aa17de1\") " pod="openstack/keystone-cron-29340601-cnz9t" Oct 14 10:01:00 crc kubenswrapper[4870]: I1014 10:01:00.244922 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ft6sx\" (UniqueName: \"kubernetes.io/projected/4ad4b83a-f2a8-439f-a02e-66424aa17de1-kube-api-access-ft6sx\") pod \"keystone-cron-29340601-cnz9t\" (UID: \"4ad4b83a-f2a8-439f-a02e-66424aa17de1\") " pod="openstack/keystone-cron-29340601-cnz9t" Oct 14 10:01:00 crc kubenswrapper[4870]: I1014 10:01:00.245102 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4ad4b83a-f2a8-439f-a02e-66424aa17de1-config-data\") pod \"keystone-cron-29340601-cnz9t\" (UID: \"4ad4b83a-f2a8-439f-a02e-66424aa17de1\") " pod="openstack/keystone-cron-29340601-cnz9t" Oct 14 10:01:00 crc kubenswrapper[4870]: I1014 10:01:00.245207 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/4ad4b83a-f2a8-439f-a02e-66424aa17de1-fernet-keys\") pod \"keystone-cron-29340601-cnz9t\" (UID: \"4ad4b83a-f2a8-439f-a02e-66424aa17de1\") " pod="openstack/keystone-cron-29340601-cnz9t" Oct 14 10:01:00 crc kubenswrapper[4870]: I1014 10:01:00.348597 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ad4b83a-f2a8-439f-a02e-66424aa17de1-combined-ca-bundle\") pod \"keystone-cron-29340601-cnz9t\" (UID: \"4ad4b83a-f2a8-439f-a02e-66424aa17de1\") " pod="openstack/keystone-cron-29340601-cnz9t" Oct 14 10:01:00 crc kubenswrapper[4870]: I1014 10:01:00.348746 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ft6sx\" (UniqueName: \"kubernetes.io/projected/4ad4b83a-f2a8-439f-a02e-66424aa17de1-kube-api-access-ft6sx\") pod \"keystone-cron-29340601-cnz9t\" (UID: \"4ad4b83a-f2a8-439f-a02e-66424aa17de1\") " pod="openstack/keystone-cron-29340601-cnz9t" Oct 14 10:01:00 crc kubenswrapper[4870]: I1014 10:01:00.348793 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4ad4b83a-f2a8-439f-a02e-66424aa17de1-config-data\") pod \"keystone-cron-29340601-cnz9t\" (UID: \"4ad4b83a-f2a8-439f-a02e-66424aa17de1\") " pod="openstack/keystone-cron-29340601-cnz9t" Oct 14 10:01:00 crc kubenswrapper[4870]: I1014 10:01:00.348823 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/4ad4b83a-f2a8-439f-a02e-66424aa17de1-fernet-keys\") pod \"keystone-cron-29340601-cnz9t\" (UID: \"4ad4b83a-f2a8-439f-a02e-66424aa17de1\") " pod="openstack/keystone-cron-29340601-cnz9t" Oct 14 10:01:00 crc kubenswrapper[4870]: I1014 10:01:00.359722 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ad4b83a-f2a8-439f-a02e-66424aa17de1-combined-ca-bundle\") pod \"keystone-cron-29340601-cnz9t\" (UID: \"4ad4b83a-f2a8-439f-a02e-66424aa17de1\") " pod="openstack/keystone-cron-29340601-cnz9t" Oct 14 10:01:00 crc kubenswrapper[4870]: I1014 10:01:00.359838 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/4ad4b83a-f2a8-439f-a02e-66424aa17de1-fernet-keys\") pod \"keystone-cron-29340601-cnz9t\" (UID: \"4ad4b83a-f2a8-439f-a02e-66424aa17de1\") " pod="openstack/keystone-cron-29340601-cnz9t" Oct 14 10:01:00 crc kubenswrapper[4870]: I1014 10:01:00.359945 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4ad4b83a-f2a8-439f-a02e-66424aa17de1-config-data\") pod \"keystone-cron-29340601-cnz9t\" (UID: \"4ad4b83a-f2a8-439f-a02e-66424aa17de1\") " pod="openstack/keystone-cron-29340601-cnz9t" Oct 14 10:01:00 crc kubenswrapper[4870]: I1014 10:01:00.367825 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ft6sx\" (UniqueName: \"kubernetes.io/projected/4ad4b83a-f2a8-439f-a02e-66424aa17de1-kube-api-access-ft6sx\") pod \"keystone-cron-29340601-cnz9t\" (UID: \"4ad4b83a-f2a8-439f-a02e-66424aa17de1\") " pod="openstack/keystone-cron-29340601-cnz9t" Oct 14 10:01:00 crc kubenswrapper[4870]: I1014 10:01:00.526009 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29340601-cnz9t" Oct 14 10:01:01 crc kubenswrapper[4870]: I1014 10:01:01.129364 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29340601-cnz9t"] Oct 14 10:01:01 crc kubenswrapper[4870]: I1014 10:01:01.957962 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29340601-cnz9t" event={"ID":"4ad4b83a-f2a8-439f-a02e-66424aa17de1","Type":"ContainerStarted","Data":"aa6741aed532728c05b1cf171356e822a5a7535cc8fdedb85d5729ade9100446"} Oct 14 10:01:01 crc kubenswrapper[4870]: I1014 10:01:01.958414 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29340601-cnz9t" event={"ID":"4ad4b83a-f2a8-439f-a02e-66424aa17de1","Type":"ContainerStarted","Data":"1633d9ca852a95de16157b4b4863054e0f222cdfe85bca3c66908bff960438a9"} Oct 14 10:01:01 crc kubenswrapper[4870]: I1014 10:01:01.980660 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29340601-cnz9t" podStartSLOduration=1.980608793 podStartE2EDuration="1.980608793s" podCreationTimestamp="2025-10-14 10:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-14 10:01:01.976255596 +0000 UTC m=+10797.673615987" watchObservedRunningTime="2025-10-14 10:01:01.980608793 +0000 UTC m=+10797.677969184" Oct 14 10:01:04 crc kubenswrapper[4870]: I1014 10:01:04.997060 4870 generic.go:334] "Generic (PLEG): container finished" podID="4ad4b83a-f2a8-439f-a02e-66424aa17de1" containerID="aa6741aed532728c05b1cf171356e822a5a7535cc8fdedb85d5729ade9100446" exitCode=0 Oct 14 10:01:04 crc kubenswrapper[4870]: I1014 10:01:04.997177 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29340601-cnz9t" event={"ID":"4ad4b83a-f2a8-439f-a02e-66424aa17de1","Type":"ContainerDied","Data":"aa6741aed532728c05b1cf171356e822a5a7535cc8fdedb85d5729ade9100446"} Oct 14 10:01:06 crc kubenswrapper[4870]: I1014 10:01:06.433285 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29340601-cnz9t" Oct 14 10:01:06 crc kubenswrapper[4870]: I1014 10:01:06.522289 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ad4b83a-f2a8-439f-a02e-66424aa17de1-combined-ca-bundle\") pod \"4ad4b83a-f2a8-439f-a02e-66424aa17de1\" (UID: \"4ad4b83a-f2a8-439f-a02e-66424aa17de1\") " Oct 14 10:01:06 crc kubenswrapper[4870]: I1014 10:01:06.523708 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4ad4b83a-f2a8-439f-a02e-66424aa17de1-config-data\") pod \"4ad4b83a-f2a8-439f-a02e-66424aa17de1\" (UID: \"4ad4b83a-f2a8-439f-a02e-66424aa17de1\") " Oct 14 10:01:06 crc kubenswrapper[4870]: I1014 10:01:06.524138 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ft6sx\" (UniqueName: \"kubernetes.io/projected/4ad4b83a-f2a8-439f-a02e-66424aa17de1-kube-api-access-ft6sx\") pod \"4ad4b83a-f2a8-439f-a02e-66424aa17de1\" (UID: \"4ad4b83a-f2a8-439f-a02e-66424aa17de1\") " Oct 14 10:01:06 crc kubenswrapper[4870]: I1014 10:01:06.524258 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/4ad4b83a-f2a8-439f-a02e-66424aa17de1-fernet-keys\") pod \"4ad4b83a-f2a8-439f-a02e-66424aa17de1\" (UID: \"4ad4b83a-f2a8-439f-a02e-66424aa17de1\") " Oct 14 10:01:06 crc kubenswrapper[4870]: I1014 10:01:06.529662 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4ad4b83a-f2a8-439f-a02e-66424aa17de1-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "4ad4b83a-f2a8-439f-a02e-66424aa17de1" (UID: "4ad4b83a-f2a8-439f-a02e-66424aa17de1"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 10:01:06 crc kubenswrapper[4870]: I1014 10:01:06.530618 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4ad4b83a-f2a8-439f-a02e-66424aa17de1-kube-api-access-ft6sx" (OuterVolumeSpecName: "kube-api-access-ft6sx") pod "4ad4b83a-f2a8-439f-a02e-66424aa17de1" (UID: "4ad4b83a-f2a8-439f-a02e-66424aa17de1"). InnerVolumeSpecName "kube-api-access-ft6sx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 10:01:06 crc kubenswrapper[4870]: I1014 10:01:06.590486 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4ad4b83a-f2a8-439f-a02e-66424aa17de1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4ad4b83a-f2a8-439f-a02e-66424aa17de1" (UID: "4ad4b83a-f2a8-439f-a02e-66424aa17de1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 10:01:06 crc kubenswrapper[4870]: I1014 10:01:06.610653 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4ad4b83a-f2a8-439f-a02e-66424aa17de1-config-data" (OuterVolumeSpecName: "config-data") pod "4ad4b83a-f2a8-439f-a02e-66424aa17de1" (UID: "4ad4b83a-f2a8-439f-a02e-66424aa17de1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 10:01:06 crc kubenswrapper[4870]: I1014 10:01:06.627052 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ft6sx\" (UniqueName: \"kubernetes.io/projected/4ad4b83a-f2a8-439f-a02e-66424aa17de1-kube-api-access-ft6sx\") on node \"crc\" DevicePath \"\"" Oct 14 10:01:06 crc kubenswrapper[4870]: I1014 10:01:06.627087 4870 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/4ad4b83a-f2a8-439f-a02e-66424aa17de1-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 14 10:01:06 crc kubenswrapper[4870]: I1014 10:01:06.627103 4870 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ad4b83a-f2a8-439f-a02e-66424aa17de1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 14 10:01:06 crc kubenswrapper[4870]: I1014 10:01:06.627116 4870 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4ad4b83a-f2a8-439f-a02e-66424aa17de1-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 10:01:07 crc kubenswrapper[4870]: I1014 10:01:07.028570 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29340601-cnz9t" event={"ID":"4ad4b83a-f2a8-439f-a02e-66424aa17de1","Type":"ContainerDied","Data":"1633d9ca852a95de16157b4b4863054e0f222cdfe85bca3c66908bff960438a9"} Oct 14 10:01:07 crc kubenswrapper[4870]: I1014 10:01:07.028623 4870 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1633d9ca852a95de16157b4b4863054e0f222cdfe85bca3c66908bff960438a9" Oct 14 10:01:07 crc kubenswrapper[4870]: I1014 10:01:07.028702 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29340601-cnz9t" Oct 14 10:01:23 crc kubenswrapper[4870]: I1014 10:01:23.951236 4870 patch_prober.go:28] interesting pod/machine-config-daemon-7tvc7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 10:01:23 crc kubenswrapper[4870]: I1014 10:01:23.954304 4870 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 10:01:28 crc kubenswrapper[4870]: I1014 10:01:28.513768 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-5mzbc"] Oct 14 10:01:28 crc kubenswrapper[4870]: E1014 10:01:28.515068 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ad4b83a-f2a8-439f-a02e-66424aa17de1" containerName="keystone-cron" Oct 14 10:01:28 crc kubenswrapper[4870]: I1014 10:01:28.515093 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ad4b83a-f2a8-439f-a02e-66424aa17de1" containerName="keystone-cron" Oct 14 10:01:28 crc kubenswrapper[4870]: I1014 10:01:28.515502 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="4ad4b83a-f2a8-439f-a02e-66424aa17de1" containerName="keystone-cron" Oct 14 10:01:28 crc kubenswrapper[4870]: I1014 10:01:28.520595 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5mzbc" Oct 14 10:01:28 crc kubenswrapper[4870]: I1014 10:01:28.548748 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-5mzbc"] Oct 14 10:01:28 crc kubenswrapper[4870]: I1014 10:01:28.665576 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vp7dn\" (UniqueName: \"kubernetes.io/projected/3e8d7e0e-015c-4105-bce9-e762ec9ce34c-kube-api-access-vp7dn\") pod \"certified-operators-5mzbc\" (UID: \"3e8d7e0e-015c-4105-bce9-e762ec9ce34c\") " pod="openshift-marketplace/certified-operators-5mzbc" Oct 14 10:01:28 crc kubenswrapper[4870]: I1014 10:01:28.665749 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3e8d7e0e-015c-4105-bce9-e762ec9ce34c-catalog-content\") pod \"certified-operators-5mzbc\" (UID: \"3e8d7e0e-015c-4105-bce9-e762ec9ce34c\") " pod="openshift-marketplace/certified-operators-5mzbc" Oct 14 10:01:28 crc kubenswrapper[4870]: I1014 10:01:28.665880 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3e8d7e0e-015c-4105-bce9-e762ec9ce34c-utilities\") pod \"certified-operators-5mzbc\" (UID: \"3e8d7e0e-015c-4105-bce9-e762ec9ce34c\") " pod="openshift-marketplace/certified-operators-5mzbc" Oct 14 10:01:28 crc kubenswrapper[4870]: I1014 10:01:28.768613 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3e8d7e0e-015c-4105-bce9-e762ec9ce34c-utilities\") pod \"certified-operators-5mzbc\" (UID: \"3e8d7e0e-015c-4105-bce9-e762ec9ce34c\") " pod="openshift-marketplace/certified-operators-5mzbc" Oct 14 10:01:28 crc kubenswrapper[4870]: I1014 10:01:28.768711 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vp7dn\" (UniqueName: \"kubernetes.io/projected/3e8d7e0e-015c-4105-bce9-e762ec9ce34c-kube-api-access-vp7dn\") pod \"certified-operators-5mzbc\" (UID: \"3e8d7e0e-015c-4105-bce9-e762ec9ce34c\") " pod="openshift-marketplace/certified-operators-5mzbc" Oct 14 10:01:28 crc kubenswrapper[4870]: I1014 10:01:28.768865 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3e8d7e0e-015c-4105-bce9-e762ec9ce34c-catalog-content\") pod \"certified-operators-5mzbc\" (UID: \"3e8d7e0e-015c-4105-bce9-e762ec9ce34c\") " pod="openshift-marketplace/certified-operators-5mzbc" Oct 14 10:01:28 crc kubenswrapper[4870]: I1014 10:01:28.769180 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3e8d7e0e-015c-4105-bce9-e762ec9ce34c-utilities\") pod \"certified-operators-5mzbc\" (UID: \"3e8d7e0e-015c-4105-bce9-e762ec9ce34c\") " pod="openshift-marketplace/certified-operators-5mzbc" Oct 14 10:01:28 crc kubenswrapper[4870]: I1014 10:01:28.769309 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3e8d7e0e-015c-4105-bce9-e762ec9ce34c-catalog-content\") pod \"certified-operators-5mzbc\" (UID: \"3e8d7e0e-015c-4105-bce9-e762ec9ce34c\") " pod="openshift-marketplace/certified-operators-5mzbc" Oct 14 10:01:28 crc kubenswrapper[4870]: I1014 10:01:28.811007 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vp7dn\" (UniqueName: \"kubernetes.io/projected/3e8d7e0e-015c-4105-bce9-e762ec9ce34c-kube-api-access-vp7dn\") pod \"certified-operators-5mzbc\" (UID: \"3e8d7e0e-015c-4105-bce9-e762ec9ce34c\") " pod="openshift-marketplace/certified-operators-5mzbc" Oct 14 10:01:28 crc kubenswrapper[4870]: I1014 10:01:28.846651 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5mzbc" Oct 14 10:01:29 crc kubenswrapper[4870]: I1014 10:01:29.382267 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-5mzbc"] Oct 14 10:01:29 crc kubenswrapper[4870]: W1014 10:01:29.390802 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3e8d7e0e_015c_4105_bce9_e762ec9ce34c.slice/crio-a576ae24d5e7567eae0154787c09c9e68750a6049bd630ba1d989a18ec35b030 WatchSource:0}: Error finding container a576ae24d5e7567eae0154787c09c9e68750a6049bd630ba1d989a18ec35b030: Status 404 returned error can't find the container with id a576ae24d5e7567eae0154787c09c9e68750a6049bd630ba1d989a18ec35b030 Oct 14 10:01:30 crc kubenswrapper[4870]: I1014 10:01:30.329537 4870 generic.go:334] "Generic (PLEG): container finished" podID="3e8d7e0e-015c-4105-bce9-e762ec9ce34c" containerID="8e36806d799675ca0fe382b716d4d469e97453ee72e2431144a0598875f192d9" exitCode=0 Oct 14 10:01:30 crc kubenswrapper[4870]: I1014 10:01:30.329608 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5mzbc" event={"ID":"3e8d7e0e-015c-4105-bce9-e762ec9ce34c","Type":"ContainerDied","Data":"8e36806d799675ca0fe382b716d4d469e97453ee72e2431144a0598875f192d9"} Oct 14 10:01:30 crc kubenswrapper[4870]: I1014 10:01:30.329915 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5mzbc" event={"ID":"3e8d7e0e-015c-4105-bce9-e762ec9ce34c","Type":"ContainerStarted","Data":"a576ae24d5e7567eae0154787c09c9e68750a6049bd630ba1d989a18ec35b030"} Oct 14 10:01:30 crc kubenswrapper[4870]: I1014 10:01:30.333101 4870 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 14 10:01:31 crc kubenswrapper[4870]: I1014 10:01:31.348010 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5mzbc" event={"ID":"3e8d7e0e-015c-4105-bce9-e762ec9ce34c","Type":"ContainerStarted","Data":"2b0a9c1fbb95283a8d2e5fc86c78add10f691231a77c327751daff89eccd84b9"} Oct 14 10:01:33 crc kubenswrapper[4870]: I1014 10:01:33.377097 4870 generic.go:334] "Generic (PLEG): container finished" podID="3e8d7e0e-015c-4105-bce9-e762ec9ce34c" containerID="2b0a9c1fbb95283a8d2e5fc86c78add10f691231a77c327751daff89eccd84b9" exitCode=0 Oct 14 10:01:33 crc kubenswrapper[4870]: I1014 10:01:33.377525 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5mzbc" event={"ID":"3e8d7e0e-015c-4105-bce9-e762ec9ce34c","Type":"ContainerDied","Data":"2b0a9c1fbb95283a8d2e5fc86c78add10f691231a77c327751daff89eccd84b9"} Oct 14 10:01:35 crc kubenswrapper[4870]: I1014 10:01:35.402725 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5mzbc" event={"ID":"3e8d7e0e-015c-4105-bce9-e762ec9ce34c","Type":"ContainerStarted","Data":"0fdf1c0818d14e95fd207fc705a97d6477be297147cbf254da43604c961f7f8c"} Oct 14 10:01:35 crc kubenswrapper[4870]: I1014 10:01:35.431762 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-5mzbc" podStartSLOduration=3.531738583 podStartE2EDuration="7.431738263s" podCreationTimestamp="2025-10-14 10:01:28 +0000 UTC" firstStartedPulling="2025-10-14 10:01:30.332622064 +0000 UTC m=+10826.029982475" lastFinishedPulling="2025-10-14 10:01:34.232621774 +0000 UTC m=+10829.929982155" observedRunningTime="2025-10-14 10:01:35.425833577 +0000 UTC m=+10831.123193988" watchObservedRunningTime="2025-10-14 10:01:35.431738263 +0000 UTC m=+10831.129098674" Oct 14 10:01:38 crc kubenswrapper[4870]: I1014 10:01:38.847814 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-5mzbc" Oct 14 10:01:38 crc kubenswrapper[4870]: I1014 10:01:38.848508 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-5mzbc" Oct 14 10:01:38 crc kubenswrapper[4870]: I1014 10:01:38.934391 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-5mzbc" Oct 14 10:01:39 crc kubenswrapper[4870]: I1014 10:01:39.546648 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-5mzbc" Oct 14 10:01:39 crc kubenswrapper[4870]: I1014 10:01:39.625403 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-5mzbc"] Oct 14 10:01:41 crc kubenswrapper[4870]: I1014 10:01:41.480682 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-5mzbc" podUID="3e8d7e0e-015c-4105-bce9-e762ec9ce34c" containerName="registry-server" containerID="cri-o://0fdf1c0818d14e95fd207fc705a97d6477be297147cbf254da43604c961f7f8c" gracePeriod=2 Oct 14 10:01:42 crc kubenswrapper[4870]: I1014 10:01:42.051942 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5mzbc" Oct 14 10:01:42 crc kubenswrapper[4870]: I1014 10:01:42.069344 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3e8d7e0e-015c-4105-bce9-e762ec9ce34c-catalog-content\") pod \"3e8d7e0e-015c-4105-bce9-e762ec9ce34c\" (UID: \"3e8d7e0e-015c-4105-bce9-e762ec9ce34c\") " Oct 14 10:01:42 crc kubenswrapper[4870]: I1014 10:01:42.069486 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3e8d7e0e-015c-4105-bce9-e762ec9ce34c-utilities\") pod \"3e8d7e0e-015c-4105-bce9-e762ec9ce34c\" (UID: \"3e8d7e0e-015c-4105-bce9-e762ec9ce34c\") " Oct 14 10:01:42 crc kubenswrapper[4870]: I1014 10:01:42.069545 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vp7dn\" (UniqueName: \"kubernetes.io/projected/3e8d7e0e-015c-4105-bce9-e762ec9ce34c-kube-api-access-vp7dn\") pod \"3e8d7e0e-015c-4105-bce9-e762ec9ce34c\" (UID: \"3e8d7e0e-015c-4105-bce9-e762ec9ce34c\") " Oct 14 10:01:42 crc kubenswrapper[4870]: I1014 10:01:42.070646 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3e8d7e0e-015c-4105-bce9-e762ec9ce34c-utilities" (OuterVolumeSpecName: "utilities") pod "3e8d7e0e-015c-4105-bce9-e762ec9ce34c" (UID: "3e8d7e0e-015c-4105-bce9-e762ec9ce34c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 10:01:42 crc kubenswrapper[4870]: I1014 10:01:42.071916 4870 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3e8d7e0e-015c-4105-bce9-e762ec9ce34c-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 10:01:42 crc kubenswrapper[4870]: I1014 10:01:42.076509 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3e8d7e0e-015c-4105-bce9-e762ec9ce34c-kube-api-access-vp7dn" (OuterVolumeSpecName: "kube-api-access-vp7dn") pod "3e8d7e0e-015c-4105-bce9-e762ec9ce34c" (UID: "3e8d7e0e-015c-4105-bce9-e762ec9ce34c"). InnerVolumeSpecName "kube-api-access-vp7dn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 10:01:42 crc kubenswrapper[4870]: I1014 10:01:42.148590 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3e8d7e0e-015c-4105-bce9-e762ec9ce34c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3e8d7e0e-015c-4105-bce9-e762ec9ce34c" (UID: "3e8d7e0e-015c-4105-bce9-e762ec9ce34c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 10:01:42 crc kubenswrapper[4870]: I1014 10:01:42.174137 4870 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3e8d7e0e-015c-4105-bce9-e762ec9ce34c-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 10:01:42 crc kubenswrapper[4870]: I1014 10:01:42.174183 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vp7dn\" (UniqueName: \"kubernetes.io/projected/3e8d7e0e-015c-4105-bce9-e762ec9ce34c-kube-api-access-vp7dn\") on node \"crc\" DevicePath \"\"" Oct 14 10:01:42 crc kubenswrapper[4870]: I1014 10:01:42.501170 4870 generic.go:334] "Generic (PLEG): container finished" podID="3e8d7e0e-015c-4105-bce9-e762ec9ce34c" containerID="0fdf1c0818d14e95fd207fc705a97d6477be297147cbf254da43604c961f7f8c" exitCode=0 Oct 14 10:01:42 crc kubenswrapper[4870]: I1014 10:01:42.501272 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5mzbc" Oct 14 10:01:42 crc kubenswrapper[4870]: I1014 10:01:42.501255 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5mzbc" event={"ID":"3e8d7e0e-015c-4105-bce9-e762ec9ce34c","Type":"ContainerDied","Data":"0fdf1c0818d14e95fd207fc705a97d6477be297147cbf254da43604c961f7f8c"} Oct 14 10:01:42 crc kubenswrapper[4870]: I1014 10:01:42.501558 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5mzbc" event={"ID":"3e8d7e0e-015c-4105-bce9-e762ec9ce34c","Type":"ContainerDied","Data":"a576ae24d5e7567eae0154787c09c9e68750a6049bd630ba1d989a18ec35b030"} Oct 14 10:01:42 crc kubenswrapper[4870]: I1014 10:01:42.501609 4870 scope.go:117] "RemoveContainer" containerID="0fdf1c0818d14e95fd207fc705a97d6477be297147cbf254da43604c961f7f8c" Oct 14 10:01:42 crc kubenswrapper[4870]: I1014 10:01:42.563054 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-5mzbc"] Oct 14 10:01:42 crc kubenswrapper[4870]: I1014 10:01:42.574218 4870 scope.go:117] "RemoveContainer" containerID="2b0a9c1fbb95283a8d2e5fc86c78add10f691231a77c327751daff89eccd84b9" Oct 14 10:01:42 crc kubenswrapper[4870]: I1014 10:01:42.578014 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-5mzbc"] Oct 14 10:01:42 crc kubenswrapper[4870]: I1014 10:01:42.617870 4870 scope.go:117] "RemoveContainer" containerID="8e36806d799675ca0fe382b716d4d469e97453ee72e2431144a0598875f192d9" Oct 14 10:01:42 crc kubenswrapper[4870]: I1014 10:01:42.670477 4870 scope.go:117] "RemoveContainer" containerID="0fdf1c0818d14e95fd207fc705a97d6477be297147cbf254da43604c961f7f8c" Oct 14 10:01:42 crc kubenswrapper[4870]: E1014 10:01:42.671180 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0fdf1c0818d14e95fd207fc705a97d6477be297147cbf254da43604c961f7f8c\": container with ID starting with 0fdf1c0818d14e95fd207fc705a97d6477be297147cbf254da43604c961f7f8c not found: ID does not exist" containerID="0fdf1c0818d14e95fd207fc705a97d6477be297147cbf254da43604c961f7f8c" Oct 14 10:01:42 crc kubenswrapper[4870]: I1014 10:01:42.671217 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0fdf1c0818d14e95fd207fc705a97d6477be297147cbf254da43604c961f7f8c"} err="failed to get container status \"0fdf1c0818d14e95fd207fc705a97d6477be297147cbf254da43604c961f7f8c\": rpc error: code = NotFound desc = could not find container \"0fdf1c0818d14e95fd207fc705a97d6477be297147cbf254da43604c961f7f8c\": container with ID starting with 0fdf1c0818d14e95fd207fc705a97d6477be297147cbf254da43604c961f7f8c not found: ID does not exist" Oct 14 10:01:42 crc kubenswrapper[4870]: I1014 10:01:42.671241 4870 scope.go:117] "RemoveContainer" containerID="2b0a9c1fbb95283a8d2e5fc86c78add10f691231a77c327751daff89eccd84b9" Oct 14 10:01:42 crc kubenswrapper[4870]: E1014 10:01:42.673680 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2b0a9c1fbb95283a8d2e5fc86c78add10f691231a77c327751daff89eccd84b9\": container with ID starting with 2b0a9c1fbb95283a8d2e5fc86c78add10f691231a77c327751daff89eccd84b9 not found: ID does not exist" containerID="2b0a9c1fbb95283a8d2e5fc86c78add10f691231a77c327751daff89eccd84b9" Oct 14 10:01:42 crc kubenswrapper[4870]: I1014 10:01:42.673710 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2b0a9c1fbb95283a8d2e5fc86c78add10f691231a77c327751daff89eccd84b9"} err="failed to get container status \"2b0a9c1fbb95283a8d2e5fc86c78add10f691231a77c327751daff89eccd84b9\": rpc error: code = NotFound desc = could not find container \"2b0a9c1fbb95283a8d2e5fc86c78add10f691231a77c327751daff89eccd84b9\": container with ID starting with 2b0a9c1fbb95283a8d2e5fc86c78add10f691231a77c327751daff89eccd84b9 not found: ID does not exist" Oct 14 10:01:42 crc kubenswrapper[4870]: I1014 10:01:42.673725 4870 scope.go:117] "RemoveContainer" containerID="8e36806d799675ca0fe382b716d4d469e97453ee72e2431144a0598875f192d9" Oct 14 10:01:42 crc kubenswrapper[4870]: E1014 10:01:42.674339 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8e36806d799675ca0fe382b716d4d469e97453ee72e2431144a0598875f192d9\": container with ID starting with 8e36806d799675ca0fe382b716d4d469e97453ee72e2431144a0598875f192d9 not found: ID does not exist" containerID="8e36806d799675ca0fe382b716d4d469e97453ee72e2431144a0598875f192d9" Oct 14 10:01:42 crc kubenswrapper[4870]: I1014 10:01:42.674415 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8e36806d799675ca0fe382b716d4d469e97453ee72e2431144a0598875f192d9"} err="failed to get container status \"8e36806d799675ca0fe382b716d4d469e97453ee72e2431144a0598875f192d9\": rpc error: code = NotFound desc = could not find container \"8e36806d799675ca0fe382b716d4d469e97453ee72e2431144a0598875f192d9\": container with ID starting with 8e36806d799675ca0fe382b716d4d469e97453ee72e2431144a0598875f192d9 not found: ID does not exist" Oct 14 10:01:43 crc kubenswrapper[4870]: I1014 10:01:43.055933 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3e8d7e0e-015c-4105-bce9-e762ec9ce34c" path="/var/lib/kubelet/pods/3e8d7e0e-015c-4105-bce9-e762ec9ce34c/volumes" Oct 14 10:01:53 crc kubenswrapper[4870]: I1014 10:01:53.950900 4870 patch_prober.go:28] interesting pod/machine-config-daemon-7tvc7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 10:01:53 crc kubenswrapper[4870]: I1014 10:01:53.951531 4870 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 10:01:53 crc kubenswrapper[4870]: I1014 10:01:53.951594 4870 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" Oct 14 10:01:53 crc kubenswrapper[4870]: I1014 10:01:53.952503 4870 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"fa43fd954e8c6e0aade481bc7b524b46d5559be95fdfa58bfe7f0b926272a5a5"} pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 14 10:01:53 crc kubenswrapper[4870]: I1014 10:01:53.952601 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerName="machine-config-daemon" containerID="cri-o://fa43fd954e8c6e0aade481bc7b524b46d5559be95fdfa58bfe7f0b926272a5a5" gracePeriod=600 Oct 14 10:01:54 crc kubenswrapper[4870]: I1014 10:01:54.669767 4870 generic.go:334] "Generic (PLEG): container finished" podID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerID="fa43fd954e8c6e0aade481bc7b524b46d5559be95fdfa58bfe7f0b926272a5a5" exitCode=0 Oct 14 10:01:54 crc kubenswrapper[4870]: I1014 10:01:54.669846 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" event={"ID":"874a5e45-dffd-4d17-b609-b7d3ed2eab07","Type":"ContainerDied","Data":"fa43fd954e8c6e0aade481bc7b524b46d5559be95fdfa58bfe7f0b926272a5a5"} Oct 14 10:01:54 crc kubenswrapper[4870]: I1014 10:01:54.670728 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" event={"ID":"874a5e45-dffd-4d17-b609-b7d3ed2eab07","Type":"ContainerStarted","Data":"972233f23b285fd01cec82ddb478cfd8b54d11806ccbfd7ce23b7ef84f4bee5c"} Oct 14 10:01:54 crc kubenswrapper[4870]: I1014 10:01:54.670771 4870 scope.go:117] "RemoveContainer" containerID="c52e0ba5edaea8b07f60b9f76084bc5c51b52575d871193fc95ac2cefd996fea" Oct 14 10:03:38 crc kubenswrapper[4870]: I1014 10:03:38.564482 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-web-config" Oct 14 10:04:23 crc kubenswrapper[4870]: I1014 10:04:23.950739 4870 patch_prober.go:28] interesting pod/machine-config-daemon-7tvc7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 10:04:23 crc kubenswrapper[4870]: I1014 10:04:23.951450 4870 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 10:04:32 crc kubenswrapper[4870]: I1014 10:04:32.885276 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-xxskh"] Oct 14 10:04:32 crc kubenswrapper[4870]: E1014 10:04:32.887012 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e8d7e0e-015c-4105-bce9-e762ec9ce34c" containerName="extract-utilities" Oct 14 10:04:32 crc kubenswrapper[4870]: I1014 10:04:32.887046 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e8d7e0e-015c-4105-bce9-e762ec9ce34c" containerName="extract-utilities" Oct 14 10:04:32 crc kubenswrapper[4870]: E1014 10:04:32.887095 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e8d7e0e-015c-4105-bce9-e762ec9ce34c" containerName="extract-content" Oct 14 10:04:32 crc kubenswrapper[4870]: I1014 10:04:32.887113 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e8d7e0e-015c-4105-bce9-e762ec9ce34c" containerName="extract-content" Oct 14 10:04:32 crc kubenswrapper[4870]: E1014 10:04:32.887203 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e8d7e0e-015c-4105-bce9-e762ec9ce34c" containerName="registry-server" Oct 14 10:04:32 crc kubenswrapper[4870]: I1014 10:04:32.887224 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e8d7e0e-015c-4105-bce9-e762ec9ce34c" containerName="registry-server" Oct 14 10:04:32 crc kubenswrapper[4870]: I1014 10:04:32.887779 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="3e8d7e0e-015c-4105-bce9-e762ec9ce34c" containerName="registry-server" Oct 14 10:04:32 crc kubenswrapper[4870]: I1014 10:04:32.891517 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xxskh" Oct 14 10:04:32 crc kubenswrapper[4870]: I1014 10:04:32.937786 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-xxskh"] Oct 14 10:04:33 crc kubenswrapper[4870]: I1014 10:04:33.022801 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rm28g\" (UniqueName: \"kubernetes.io/projected/fe1ae3d9-bdad-4f42-920a-4861e0ba6267-kube-api-access-rm28g\") pod \"redhat-marketplace-xxskh\" (UID: \"fe1ae3d9-bdad-4f42-920a-4861e0ba6267\") " pod="openshift-marketplace/redhat-marketplace-xxskh" Oct 14 10:04:33 crc kubenswrapper[4870]: I1014 10:04:33.022877 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fe1ae3d9-bdad-4f42-920a-4861e0ba6267-catalog-content\") pod \"redhat-marketplace-xxskh\" (UID: \"fe1ae3d9-bdad-4f42-920a-4861e0ba6267\") " pod="openshift-marketplace/redhat-marketplace-xxskh" Oct 14 10:04:33 crc kubenswrapper[4870]: I1014 10:04:33.023512 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fe1ae3d9-bdad-4f42-920a-4861e0ba6267-utilities\") pod \"redhat-marketplace-xxskh\" (UID: \"fe1ae3d9-bdad-4f42-920a-4861e0ba6267\") " pod="openshift-marketplace/redhat-marketplace-xxskh" Oct 14 10:04:33 crc kubenswrapper[4870]: I1014 10:04:33.125472 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rm28g\" (UniqueName: \"kubernetes.io/projected/fe1ae3d9-bdad-4f42-920a-4861e0ba6267-kube-api-access-rm28g\") pod \"redhat-marketplace-xxskh\" (UID: \"fe1ae3d9-bdad-4f42-920a-4861e0ba6267\") " pod="openshift-marketplace/redhat-marketplace-xxskh" Oct 14 10:04:33 crc kubenswrapper[4870]: I1014 10:04:33.125591 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fe1ae3d9-bdad-4f42-920a-4861e0ba6267-catalog-content\") pod \"redhat-marketplace-xxskh\" (UID: \"fe1ae3d9-bdad-4f42-920a-4861e0ba6267\") " pod="openshift-marketplace/redhat-marketplace-xxskh" Oct 14 10:04:33 crc kubenswrapper[4870]: I1014 10:04:33.125816 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fe1ae3d9-bdad-4f42-920a-4861e0ba6267-utilities\") pod \"redhat-marketplace-xxskh\" (UID: \"fe1ae3d9-bdad-4f42-920a-4861e0ba6267\") " pod="openshift-marketplace/redhat-marketplace-xxskh" Oct 14 10:04:33 crc kubenswrapper[4870]: I1014 10:04:33.127065 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fe1ae3d9-bdad-4f42-920a-4861e0ba6267-utilities\") pod \"redhat-marketplace-xxskh\" (UID: \"fe1ae3d9-bdad-4f42-920a-4861e0ba6267\") " pod="openshift-marketplace/redhat-marketplace-xxskh" Oct 14 10:04:33 crc kubenswrapper[4870]: I1014 10:04:33.127183 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fe1ae3d9-bdad-4f42-920a-4861e0ba6267-catalog-content\") pod \"redhat-marketplace-xxskh\" (UID: \"fe1ae3d9-bdad-4f42-920a-4861e0ba6267\") " pod="openshift-marketplace/redhat-marketplace-xxskh" Oct 14 10:04:33 crc kubenswrapper[4870]: I1014 10:04:33.151251 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rm28g\" (UniqueName: \"kubernetes.io/projected/fe1ae3d9-bdad-4f42-920a-4861e0ba6267-kube-api-access-rm28g\") pod \"redhat-marketplace-xxskh\" (UID: \"fe1ae3d9-bdad-4f42-920a-4861e0ba6267\") " pod="openshift-marketplace/redhat-marketplace-xxskh" Oct 14 10:04:33 crc kubenswrapper[4870]: I1014 10:04:33.252178 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xxskh" Oct 14 10:04:33 crc kubenswrapper[4870]: I1014 10:04:33.703579 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-xxskh"] Oct 14 10:04:34 crc kubenswrapper[4870]: I1014 10:04:34.934573 4870 generic.go:334] "Generic (PLEG): container finished" podID="fe1ae3d9-bdad-4f42-920a-4861e0ba6267" containerID="ce0714edbf8ff471ccce6feea781e247fc972c604d2da0695e6559165a14d3fa" exitCode=0 Oct 14 10:04:34 crc kubenswrapper[4870]: I1014 10:04:34.934709 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xxskh" event={"ID":"fe1ae3d9-bdad-4f42-920a-4861e0ba6267","Type":"ContainerDied","Data":"ce0714edbf8ff471ccce6feea781e247fc972c604d2da0695e6559165a14d3fa"} Oct 14 10:04:34 crc kubenswrapper[4870]: I1014 10:04:34.935291 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xxskh" event={"ID":"fe1ae3d9-bdad-4f42-920a-4861e0ba6267","Type":"ContainerStarted","Data":"c4418a8b7c3ba2af2ed67985861579fa024b133005176f63e700ba8db6ea0115"} Oct 14 10:04:35 crc kubenswrapper[4870]: I1014 10:04:35.948080 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xxskh" event={"ID":"fe1ae3d9-bdad-4f42-920a-4861e0ba6267","Type":"ContainerStarted","Data":"a244f28033f8540140e58237c3f4bf4dbfd2462f00786662bc8415d19ad95617"} Oct 14 10:04:36 crc kubenswrapper[4870]: I1014 10:04:36.961872 4870 generic.go:334] "Generic (PLEG): container finished" podID="fe1ae3d9-bdad-4f42-920a-4861e0ba6267" containerID="a244f28033f8540140e58237c3f4bf4dbfd2462f00786662bc8415d19ad95617" exitCode=0 Oct 14 10:04:36 crc kubenswrapper[4870]: I1014 10:04:36.961947 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xxskh" event={"ID":"fe1ae3d9-bdad-4f42-920a-4861e0ba6267","Type":"ContainerDied","Data":"a244f28033f8540140e58237c3f4bf4dbfd2462f00786662bc8415d19ad95617"} Oct 14 10:04:37 crc kubenswrapper[4870]: I1014 10:04:37.982293 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xxskh" event={"ID":"fe1ae3d9-bdad-4f42-920a-4861e0ba6267","Type":"ContainerStarted","Data":"580180cb4da143953befa8b8ef822282582ad6cee6a75ab784b6f3e0e2fa68d6"} Oct 14 10:04:38 crc kubenswrapper[4870]: I1014 10:04:38.014017 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-xxskh" podStartSLOduration=3.344504585 podStartE2EDuration="6.013994935s" podCreationTimestamp="2025-10-14 10:04:32 +0000 UTC" firstStartedPulling="2025-10-14 10:04:34.937853264 +0000 UTC m=+11010.635213675" lastFinishedPulling="2025-10-14 10:04:37.607343614 +0000 UTC m=+11013.304704025" observedRunningTime="2025-10-14 10:04:38.012310024 +0000 UTC m=+11013.709670405" watchObservedRunningTime="2025-10-14 10:04:38.013994935 +0000 UTC m=+11013.711355316" Oct 14 10:04:43 crc kubenswrapper[4870]: I1014 10:04:43.252664 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-xxskh" Oct 14 10:04:43 crc kubenswrapper[4870]: I1014 10:04:43.253582 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-xxskh" Oct 14 10:04:43 crc kubenswrapper[4870]: I1014 10:04:43.333035 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-xxskh" Oct 14 10:04:44 crc kubenswrapper[4870]: I1014 10:04:44.178133 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-xxskh" Oct 14 10:04:44 crc kubenswrapper[4870]: I1014 10:04:44.263221 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-xxskh"] Oct 14 10:04:46 crc kubenswrapper[4870]: I1014 10:04:46.109852 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-xxskh" podUID="fe1ae3d9-bdad-4f42-920a-4861e0ba6267" containerName="registry-server" containerID="cri-o://580180cb4da143953befa8b8ef822282582ad6cee6a75ab784b6f3e0e2fa68d6" gracePeriod=2 Oct 14 10:04:46 crc kubenswrapper[4870]: I1014 10:04:46.679327 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xxskh" Oct 14 10:04:46 crc kubenswrapper[4870]: I1014 10:04:46.760615 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fe1ae3d9-bdad-4f42-920a-4861e0ba6267-utilities\") pod \"fe1ae3d9-bdad-4f42-920a-4861e0ba6267\" (UID: \"fe1ae3d9-bdad-4f42-920a-4861e0ba6267\") " Oct 14 10:04:46 crc kubenswrapper[4870]: I1014 10:04:46.761060 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rm28g\" (UniqueName: \"kubernetes.io/projected/fe1ae3d9-bdad-4f42-920a-4861e0ba6267-kube-api-access-rm28g\") pod \"fe1ae3d9-bdad-4f42-920a-4861e0ba6267\" (UID: \"fe1ae3d9-bdad-4f42-920a-4861e0ba6267\") " Oct 14 10:04:46 crc kubenswrapper[4870]: I1014 10:04:46.761295 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fe1ae3d9-bdad-4f42-920a-4861e0ba6267-catalog-content\") pod \"fe1ae3d9-bdad-4f42-920a-4861e0ba6267\" (UID: \"fe1ae3d9-bdad-4f42-920a-4861e0ba6267\") " Oct 14 10:04:46 crc kubenswrapper[4870]: I1014 10:04:46.761772 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fe1ae3d9-bdad-4f42-920a-4861e0ba6267-utilities" (OuterVolumeSpecName: "utilities") pod "fe1ae3d9-bdad-4f42-920a-4861e0ba6267" (UID: "fe1ae3d9-bdad-4f42-920a-4861e0ba6267"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 10:04:46 crc kubenswrapper[4870]: I1014 10:04:46.767586 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fe1ae3d9-bdad-4f42-920a-4861e0ba6267-kube-api-access-rm28g" (OuterVolumeSpecName: "kube-api-access-rm28g") pod "fe1ae3d9-bdad-4f42-920a-4861e0ba6267" (UID: "fe1ae3d9-bdad-4f42-920a-4861e0ba6267"). InnerVolumeSpecName "kube-api-access-rm28g". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 10:04:46 crc kubenswrapper[4870]: I1014 10:04:46.779472 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fe1ae3d9-bdad-4f42-920a-4861e0ba6267-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "fe1ae3d9-bdad-4f42-920a-4861e0ba6267" (UID: "fe1ae3d9-bdad-4f42-920a-4861e0ba6267"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 10:04:46 crc kubenswrapper[4870]: I1014 10:04:46.864116 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rm28g\" (UniqueName: \"kubernetes.io/projected/fe1ae3d9-bdad-4f42-920a-4861e0ba6267-kube-api-access-rm28g\") on node \"crc\" DevicePath \"\"" Oct 14 10:04:46 crc kubenswrapper[4870]: I1014 10:04:46.864167 4870 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fe1ae3d9-bdad-4f42-920a-4861e0ba6267-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 10:04:46 crc kubenswrapper[4870]: I1014 10:04:46.864186 4870 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fe1ae3d9-bdad-4f42-920a-4861e0ba6267-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 10:04:47 crc kubenswrapper[4870]: I1014 10:04:47.128911 4870 generic.go:334] "Generic (PLEG): container finished" podID="fe1ae3d9-bdad-4f42-920a-4861e0ba6267" containerID="580180cb4da143953befa8b8ef822282582ad6cee6a75ab784b6f3e0e2fa68d6" exitCode=0 Oct 14 10:04:47 crc kubenswrapper[4870]: I1014 10:04:47.128982 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xxskh" event={"ID":"fe1ae3d9-bdad-4f42-920a-4861e0ba6267","Type":"ContainerDied","Data":"580180cb4da143953befa8b8ef822282582ad6cee6a75ab784b6f3e0e2fa68d6"} Oct 14 10:04:47 crc kubenswrapper[4870]: I1014 10:04:47.129062 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xxskh" Oct 14 10:04:47 crc kubenswrapper[4870]: I1014 10:04:47.129290 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xxskh" event={"ID":"fe1ae3d9-bdad-4f42-920a-4861e0ba6267","Type":"ContainerDied","Data":"c4418a8b7c3ba2af2ed67985861579fa024b133005176f63e700ba8db6ea0115"} Oct 14 10:04:47 crc kubenswrapper[4870]: I1014 10:04:47.129309 4870 scope.go:117] "RemoveContainer" containerID="580180cb4da143953befa8b8ef822282582ad6cee6a75ab784b6f3e0e2fa68d6" Oct 14 10:04:47 crc kubenswrapper[4870]: I1014 10:04:47.178411 4870 scope.go:117] "RemoveContainer" containerID="a244f28033f8540140e58237c3f4bf4dbfd2462f00786662bc8415d19ad95617" Oct 14 10:04:47 crc kubenswrapper[4870]: I1014 10:04:47.178589 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-xxskh"] Oct 14 10:04:47 crc kubenswrapper[4870]: I1014 10:04:47.198303 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-xxskh"] Oct 14 10:04:47 crc kubenswrapper[4870]: I1014 10:04:47.206317 4870 scope.go:117] "RemoveContainer" containerID="ce0714edbf8ff471ccce6feea781e247fc972c604d2da0695e6559165a14d3fa" Oct 14 10:04:47 crc kubenswrapper[4870]: I1014 10:04:47.864462 4870 scope.go:117] "RemoveContainer" containerID="580180cb4da143953befa8b8ef822282582ad6cee6a75ab784b6f3e0e2fa68d6" Oct 14 10:04:47 crc kubenswrapper[4870]: E1014 10:04:47.864965 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"580180cb4da143953befa8b8ef822282582ad6cee6a75ab784b6f3e0e2fa68d6\": container with ID starting with 580180cb4da143953befa8b8ef822282582ad6cee6a75ab784b6f3e0e2fa68d6 not found: ID does not exist" containerID="580180cb4da143953befa8b8ef822282582ad6cee6a75ab784b6f3e0e2fa68d6" Oct 14 10:04:47 crc kubenswrapper[4870]: I1014 10:04:47.865024 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"580180cb4da143953befa8b8ef822282582ad6cee6a75ab784b6f3e0e2fa68d6"} err="failed to get container status \"580180cb4da143953befa8b8ef822282582ad6cee6a75ab784b6f3e0e2fa68d6\": rpc error: code = NotFound desc = could not find container \"580180cb4da143953befa8b8ef822282582ad6cee6a75ab784b6f3e0e2fa68d6\": container with ID starting with 580180cb4da143953befa8b8ef822282582ad6cee6a75ab784b6f3e0e2fa68d6 not found: ID does not exist" Oct 14 10:04:47 crc kubenswrapper[4870]: I1014 10:04:47.865060 4870 scope.go:117] "RemoveContainer" containerID="a244f28033f8540140e58237c3f4bf4dbfd2462f00786662bc8415d19ad95617" Oct 14 10:04:47 crc kubenswrapper[4870]: E1014 10:04:47.865571 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a244f28033f8540140e58237c3f4bf4dbfd2462f00786662bc8415d19ad95617\": container with ID starting with a244f28033f8540140e58237c3f4bf4dbfd2462f00786662bc8415d19ad95617 not found: ID does not exist" containerID="a244f28033f8540140e58237c3f4bf4dbfd2462f00786662bc8415d19ad95617" Oct 14 10:04:47 crc kubenswrapper[4870]: I1014 10:04:47.865637 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a244f28033f8540140e58237c3f4bf4dbfd2462f00786662bc8415d19ad95617"} err="failed to get container status \"a244f28033f8540140e58237c3f4bf4dbfd2462f00786662bc8415d19ad95617\": rpc error: code = NotFound desc = could not find container \"a244f28033f8540140e58237c3f4bf4dbfd2462f00786662bc8415d19ad95617\": container with ID starting with a244f28033f8540140e58237c3f4bf4dbfd2462f00786662bc8415d19ad95617 not found: ID does not exist" Oct 14 10:04:47 crc kubenswrapper[4870]: I1014 10:04:47.865681 4870 scope.go:117] "RemoveContainer" containerID="ce0714edbf8ff471ccce6feea781e247fc972c604d2da0695e6559165a14d3fa" Oct 14 10:04:47 crc kubenswrapper[4870]: E1014 10:04:47.866120 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ce0714edbf8ff471ccce6feea781e247fc972c604d2da0695e6559165a14d3fa\": container with ID starting with ce0714edbf8ff471ccce6feea781e247fc972c604d2da0695e6559165a14d3fa not found: ID does not exist" containerID="ce0714edbf8ff471ccce6feea781e247fc972c604d2da0695e6559165a14d3fa" Oct 14 10:04:47 crc kubenswrapper[4870]: I1014 10:04:47.866162 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ce0714edbf8ff471ccce6feea781e247fc972c604d2da0695e6559165a14d3fa"} err="failed to get container status \"ce0714edbf8ff471ccce6feea781e247fc972c604d2da0695e6559165a14d3fa\": rpc error: code = NotFound desc = could not find container \"ce0714edbf8ff471ccce6feea781e247fc972c604d2da0695e6559165a14d3fa\": container with ID starting with ce0714edbf8ff471ccce6feea781e247fc972c604d2da0695e6559165a14d3fa not found: ID does not exist" Oct 14 10:04:49 crc kubenswrapper[4870]: I1014 10:04:49.057622 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fe1ae3d9-bdad-4f42-920a-4861e0ba6267" path="/var/lib/kubelet/pods/fe1ae3d9-bdad-4f42-920a-4861e0ba6267/volumes" Oct 14 10:04:53 crc kubenswrapper[4870]: I1014 10:04:53.950710 4870 patch_prober.go:28] interesting pod/machine-config-daemon-7tvc7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 10:04:53 crc kubenswrapper[4870]: I1014 10:04:53.951482 4870 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 10:05:11 crc kubenswrapper[4870]: I1014 10:05:11.292096 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-wgw82"] Oct 14 10:05:11 crc kubenswrapper[4870]: E1014 10:05:11.293337 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fe1ae3d9-bdad-4f42-920a-4861e0ba6267" containerName="extract-utilities" Oct 14 10:05:11 crc kubenswrapper[4870]: I1014 10:05:11.293360 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="fe1ae3d9-bdad-4f42-920a-4861e0ba6267" containerName="extract-utilities" Oct 14 10:05:11 crc kubenswrapper[4870]: E1014 10:05:11.293403 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fe1ae3d9-bdad-4f42-920a-4861e0ba6267" containerName="registry-server" Oct 14 10:05:11 crc kubenswrapper[4870]: I1014 10:05:11.293416 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="fe1ae3d9-bdad-4f42-920a-4861e0ba6267" containerName="registry-server" Oct 14 10:05:11 crc kubenswrapper[4870]: E1014 10:05:11.293481 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fe1ae3d9-bdad-4f42-920a-4861e0ba6267" containerName="extract-content" Oct 14 10:05:11 crc kubenswrapper[4870]: I1014 10:05:11.293492 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="fe1ae3d9-bdad-4f42-920a-4861e0ba6267" containerName="extract-content" Oct 14 10:05:11 crc kubenswrapper[4870]: I1014 10:05:11.293863 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="fe1ae3d9-bdad-4f42-920a-4861e0ba6267" containerName="registry-server" Oct 14 10:05:11 crc kubenswrapper[4870]: I1014 10:05:11.301648 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wgw82" Oct 14 10:05:11 crc kubenswrapper[4870]: I1014 10:05:11.328033 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-wgw82"] Oct 14 10:05:11 crc kubenswrapper[4870]: I1014 10:05:11.421300 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b0a42940-87de-4a91-b138-20f3bdee6e96-utilities\") pod \"redhat-operators-wgw82\" (UID: \"b0a42940-87de-4a91-b138-20f3bdee6e96\") " pod="openshift-marketplace/redhat-operators-wgw82" Oct 14 10:05:11 crc kubenswrapper[4870]: I1014 10:05:11.421380 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w5rs8\" (UniqueName: \"kubernetes.io/projected/b0a42940-87de-4a91-b138-20f3bdee6e96-kube-api-access-w5rs8\") pod \"redhat-operators-wgw82\" (UID: \"b0a42940-87de-4a91-b138-20f3bdee6e96\") " pod="openshift-marketplace/redhat-operators-wgw82" Oct 14 10:05:11 crc kubenswrapper[4870]: I1014 10:05:11.421503 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b0a42940-87de-4a91-b138-20f3bdee6e96-catalog-content\") pod \"redhat-operators-wgw82\" (UID: \"b0a42940-87de-4a91-b138-20f3bdee6e96\") " pod="openshift-marketplace/redhat-operators-wgw82" Oct 14 10:05:11 crc kubenswrapper[4870]: I1014 10:05:11.523098 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b0a42940-87de-4a91-b138-20f3bdee6e96-utilities\") pod \"redhat-operators-wgw82\" (UID: \"b0a42940-87de-4a91-b138-20f3bdee6e96\") " pod="openshift-marketplace/redhat-operators-wgw82" Oct 14 10:05:11 crc kubenswrapper[4870]: I1014 10:05:11.523178 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w5rs8\" (UniqueName: \"kubernetes.io/projected/b0a42940-87de-4a91-b138-20f3bdee6e96-kube-api-access-w5rs8\") pod \"redhat-operators-wgw82\" (UID: \"b0a42940-87de-4a91-b138-20f3bdee6e96\") " pod="openshift-marketplace/redhat-operators-wgw82" Oct 14 10:05:11 crc kubenswrapper[4870]: I1014 10:05:11.523297 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b0a42940-87de-4a91-b138-20f3bdee6e96-catalog-content\") pod \"redhat-operators-wgw82\" (UID: \"b0a42940-87de-4a91-b138-20f3bdee6e96\") " pod="openshift-marketplace/redhat-operators-wgw82" Oct 14 10:05:11 crc kubenswrapper[4870]: I1014 10:05:11.525035 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b0a42940-87de-4a91-b138-20f3bdee6e96-utilities\") pod \"redhat-operators-wgw82\" (UID: \"b0a42940-87de-4a91-b138-20f3bdee6e96\") " pod="openshift-marketplace/redhat-operators-wgw82" Oct 14 10:05:11 crc kubenswrapper[4870]: I1014 10:05:11.525791 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b0a42940-87de-4a91-b138-20f3bdee6e96-catalog-content\") pod \"redhat-operators-wgw82\" (UID: \"b0a42940-87de-4a91-b138-20f3bdee6e96\") " pod="openshift-marketplace/redhat-operators-wgw82" Oct 14 10:05:11 crc kubenswrapper[4870]: I1014 10:05:11.563258 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w5rs8\" (UniqueName: \"kubernetes.io/projected/b0a42940-87de-4a91-b138-20f3bdee6e96-kube-api-access-w5rs8\") pod \"redhat-operators-wgw82\" (UID: \"b0a42940-87de-4a91-b138-20f3bdee6e96\") " pod="openshift-marketplace/redhat-operators-wgw82" Oct 14 10:05:11 crc kubenswrapper[4870]: I1014 10:05:11.660820 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wgw82" Oct 14 10:05:12 crc kubenswrapper[4870]: I1014 10:05:12.196161 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-wgw82"] Oct 14 10:05:12 crc kubenswrapper[4870]: W1014 10:05:12.204584 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb0a42940_87de_4a91_b138_20f3bdee6e96.slice/crio-00f6215cab3e99e1a1cb59495b748db231c4095b3b99bcb020755f2693c9d3a1 WatchSource:0}: Error finding container 00f6215cab3e99e1a1cb59495b748db231c4095b3b99bcb020755f2693c9d3a1: Status 404 returned error can't find the container with id 00f6215cab3e99e1a1cb59495b748db231c4095b3b99bcb020755f2693c9d3a1 Oct 14 10:05:12 crc kubenswrapper[4870]: I1014 10:05:12.481105 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wgw82" event={"ID":"b0a42940-87de-4a91-b138-20f3bdee6e96","Type":"ContainerStarted","Data":"a2207def43dbd68999ce1608ef65946280c2d57c491c1dff2dab87bf61481544"} Oct 14 10:05:12 crc kubenswrapper[4870]: I1014 10:05:12.481159 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wgw82" event={"ID":"b0a42940-87de-4a91-b138-20f3bdee6e96","Type":"ContainerStarted","Data":"00f6215cab3e99e1a1cb59495b748db231c4095b3b99bcb020755f2693c9d3a1"} Oct 14 10:05:13 crc kubenswrapper[4870]: I1014 10:05:13.502410 4870 generic.go:334] "Generic (PLEG): container finished" podID="b0a42940-87de-4a91-b138-20f3bdee6e96" containerID="a2207def43dbd68999ce1608ef65946280c2d57c491c1dff2dab87bf61481544" exitCode=0 Oct 14 10:05:13 crc kubenswrapper[4870]: I1014 10:05:13.502804 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wgw82" event={"ID":"b0a42940-87de-4a91-b138-20f3bdee6e96","Type":"ContainerDied","Data":"a2207def43dbd68999ce1608ef65946280c2d57c491c1dff2dab87bf61481544"} Oct 14 10:05:15 crc kubenswrapper[4870]: I1014 10:05:15.525479 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wgw82" event={"ID":"b0a42940-87de-4a91-b138-20f3bdee6e96","Type":"ContainerStarted","Data":"35a470cba78f4609342b21a00609c4d4f36ba2cf0f83ff14af3b15aeb2aed92a"} Oct 14 10:05:20 crc kubenswrapper[4870]: I1014 10:05:20.600940 4870 generic.go:334] "Generic (PLEG): container finished" podID="b0a42940-87de-4a91-b138-20f3bdee6e96" containerID="35a470cba78f4609342b21a00609c4d4f36ba2cf0f83ff14af3b15aeb2aed92a" exitCode=0 Oct 14 10:05:20 crc kubenswrapper[4870]: I1014 10:05:20.601061 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wgw82" event={"ID":"b0a42940-87de-4a91-b138-20f3bdee6e96","Type":"ContainerDied","Data":"35a470cba78f4609342b21a00609c4d4f36ba2cf0f83ff14af3b15aeb2aed92a"} Oct 14 10:05:21 crc kubenswrapper[4870]: I1014 10:05:21.615027 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wgw82" event={"ID":"b0a42940-87de-4a91-b138-20f3bdee6e96","Type":"ContainerStarted","Data":"eea082c0c25e5d434ba0790d4062f17b2f38b85338f15e7fa51df5bd9ca931a8"} Oct 14 10:05:21 crc kubenswrapper[4870]: I1014 10:05:21.654880 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-wgw82" podStartSLOduration=3.054884763 podStartE2EDuration="10.654857714s" podCreationTimestamp="2025-10-14 10:05:11 +0000 UTC" firstStartedPulling="2025-10-14 10:05:13.504389895 +0000 UTC m=+11049.201750276" lastFinishedPulling="2025-10-14 10:05:21.104362856 +0000 UTC m=+11056.801723227" observedRunningTime="2025-10-14 10:05:21.638269046 +0000 UTC m=+11057.335629427" watchObservedRunningTime="2025-10-14 10:05:21.654857714 +0000 UTC m=+11057.352218105" Oct 14 10:05:21 crc kubenswrapper[4870]: I1014 10:05:21.662268 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-wgw82" Oct 14 10:05:21 crc kubenswrapper[4870]: I1014 10:05:21.662368 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-wgw82" Oct 14 10:05:22 crc kubenswrapper[4870]: I1014 10:05:22.715869 4870 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-wgw82" podUID="b0a42940-87de-4a91-b138-20f3bdee6e96" containerName="registry-server" probeResult="failure" output=< Oct 14 10:05:22 crc kubenswrapper[4870]: timeout: failed to connect service ":50051" within 1s Oct 14 10:05:22 crc kubenswrapper[4870]: > Oct 14 10:05:23 crc kubenswrapper[4870]: I1014 10:05:23.951663 4870 patch_prober.go:28] interesting pod/machine-config-daemon-7tvc7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 10:05:23 crc kubenswrapper[4870]: I1014 10:05:23.951755 4870 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 10:05:23 crc kubenswrapper[4870]: I1014 10:05:23.951823 4870 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" Oct 14 10:05:23 crc kubenswrapper[4870]: I1014 10:05:23.953007 4870 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"972233f23b285fd01cec82ddb478cfd8b54d11806ccbfd7ce23b7ef84f4bee5c"} pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 14 10:05:23 crc kubenswrapper[4870]: I1014 10:05:23.953146 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerName="machine-config-daemon" containerID="cri-o://972233f23b285fd01cec82ddb478cfd8b54d11806ccbfd7ce23b7ef84f4bee5c" gracePeriod=600 Oct 14 10:05:24 crc kubenswrapper[4870]: E1014 10:05:24.096372 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 10:05:24 crc kubenswrapper[4870]: I1014 10:05:24.652517 4870 generic.go:334] "Generic (PLEG): container finished" podID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerID="972233f23b285fd01cec82ddb478cfd8b54d11806ccbfd7ce23b7ef84f4bee5c" exitCode=0 Oct 14 10:05:24 crc kubenswrapper[4870]: I1014 10:05:24.652601 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" event={"ID":"874a5e45-dffd-4d17-b609-b7d3ed2eab07","Type":"ContainerDied","Data":"972233f23b285fd01cec82ddb478cfd8b54d11806ccbfd7ce23b7ef84f4bee5c"} Oct 14 10:05:24 crc kubenswrapper[4870]: I1014 10:05:24.652879 4870 scope.go:117] "RemoveContainer" containerID="fa43fd954e8c6e0aade481bc7b524b46d5559be95fdfa58bfe7f0b926272a5a5" Oct 14 10:05:24 crc kubenswrapper[4870]: I1014 10:05:24.653655 4870 scope.go:117] "RemoveContainer" containerID="972233f23b285fd01cec82ddb478cfd8b54d11806ccbfd7ce23b7ef84f4bee5c" Oct 14 10:05:24 crc kubenswrapper[4870]: E1014 10:05:24.653945 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 10:05:31 crc kubenswrapper[4870]: I1014 10:05:31.731149 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-wgw82" Oct 14 10:05:31 crc kubenswrapper[4870]: I1014 10:05:31.806202 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-wgw82" Oct 14 10:05:31 crc kubenswrapper[4870]: I1014 10:05:31.979713 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-wgw82"] Oct 14 10:05:33 crc kubenswrapper[4870]: I1014 10:05:33.757713 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-wgw82" podUID="b0a42940-87de-4a91-b138-20f3bdee6e96" containerName="registry-server" containerID="cri-o://eea082c0c25e5d434ba0790d4062f17b2f38b85338f15e7fa51df5bd9ca931a8" gracePeriod=2 Oct 14 10:05:34 crc kubenswrapper[4870]: I1014 10:05:34.345106 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wgw82" Oct 14 10:05:34 crc kubenswrapper[4870]: I1014 10:05:34.451138 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b0a42940-87de-4a91-b138-20f3bdee6e96-catalog-content\") pod \"b0a42940-87de-4a91-b138-20f3bdee6e96\" (UID: \"b0a42940-87de-4a91-b138-20f3bdee6e96\") " Oct 14 10:05:34 crc kubenswrapper[4870]: I1014 10:05:34.451227 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b0a42940-87de-4a91-b138-20f3bdee6e96-utilities\") pod \"b0a42940-87de-4a91-b138-20f3bdee6e96\" (UID: \"b0a42940-87de-4a91-b138-20f3bdee6e96\") " Oct 14 10:05:34 crc kubenswrapper[4870]: I1014 10:05:34.451365 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w5rs8\" (UniqueName: \"kubernetes.io/projected/b0a42940-87de-4a91-b138-20f3bdee6e96-kube-api-access-w5rs8\") pod \"b0a42940-87de-4a91-b138-20f3bdee6e96\" (UID: \"b0a42940-87de-4a91-b138-20f3bdee6e96\") " Oct 14 10:05:34 crc kubenswrapper[4870]: I1014 10:05:34.451989 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b0a42940-87de-4a91-b138-20f3bdee6e96-utilities" (OuterVolumeSpecName: "utilities") pod "b0a42940-87de-4a91-b138-20f3bdee6e96" (UID: "b0a42940-87de-4a91-b138-20f3bdee6e96"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 10:05:34 crc kubenswrapper[4870]: I1014 10:05:34.452531 4870 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b0a42940-87de-4a91-b138-20f3bdee6e96-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 10:05:34 crc kubenswrapper[4870]: I1014 10:05:34.456975 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b0a42940-87de-4a91-b138-20f3bdee6e96-kube-api-access-w5rs8" (OuterVolumeSpecName: "kube-api-access-w5rs8") pod "b0a42940-87de-4a91-b138-20f3bdee6e96" (UID: "b0a42940-87de-4a91-b138-20f3bdee6e96"). InnerVolumeSpecName "kube-api-access-w5rs8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 10:05:34 crc kubenswrapper[4870]: I1014 10:05:34.532784 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b0a42940-87de-4a91-b138-20f3bdee6e96-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b0a42940-87de-4a91-b138-20f3bdee6e96" (UID: "b0a42940-87de-4a91-b138-20f3bdee6e96"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 10:05:34 crc kubenswrapper[4870]: I1014 10:05:34.554471 4870 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b0a42940-87de-4a91-b138-20f3bdee6e96-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 10:05:34 crc kubenswrapper[4870]: I1014 10:05:34.554499 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w5rs8\" (UniqueName: \"kubernetes.io/projected/b0a42940-87de-4a91-b138-20f3bdee6e96-kube-api-access-w5rs8\") on node \"crc\" DevicePath \"\"" Oct 14 10:05:34 crc kubenswrapper[4870]: I1014 10:05:34.771580 4870 generic.go:334] "Generic (PLEG): container finished" podID="b0a42940-87de-4a91-b138-20f3bdee6e96" containerID="eea082c0c25e5d434ba0790d4062f17b2f38b85338f15e7fa51df5bd9ca931a8" exitCode=0 Oct 14 10:05:34 crc kubenswrapper[4870]: I1014 10:05:34.771665 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wgw82" event={"ID":"b0a42940-87de-4a91-b138-20f3bdee6e96","Type":"ContainerDied","Data":"eea082c0c25e5d434ba0790d4062f17b2f38b85338f15e7fa51df5bd9ca931a8"} Oct 14 10:05:34 crc kubenswrapper[4870]: I1014 10:05:34.771684 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wgw82" Oct 14 10:05:34 crc kubenswrapper[4870]: I1014 10:05:34.771732 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wgw82" event={"ID":"b0a42940-87de-4a91-b138-20f3bdee6e96","Type":"ContainerDied","Data":"00f6215cab3e99e1a1cb59495b748db231c4095b3b99bcb020755f2693c9d3a1"} Oct 14 10:05:34 crc kubenswrapper[4870]: I1014 10:05:34.771753 4870 scope.go:117] "RemoveContainer" containerID="eea082c0c25e5d434ba0790d4062f17b2f38b85338f15e7fa51df5bd9ca931a8" Oct 14 10:05:34 crc kubenswrapper[4870]: I1014 10:05:34.804197 4870 scope.go:117] "RemoveContainer" containerID="35a470cba78f4609342b21a00609c4d4f36ba2cf0f83ff14af3b15aeb2aed92a" Oct 14 10:05:34 crc kubenswrapper[4870]: I1014 10:05:34.834239 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-wgw82"] Oct 14 10:05:34 crc kubenswrapper[4870]: I1014 10:05:34.854492 4870 scope.go:117] "RemoveContainer" containerID="a2207def43dbd68999ce1608ef65946280c2d57c491c1dff2dab87bf61481544" Oct 14 10:05:34 crc kubenswrapper[4870]: I1014 10:05:34.861200 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-wgw82"] Oct 14 10:05:34 crc kubenswrapper[4870]: I1014 10:05:34.923372 4870 scope.go:117] "RemoveContainer" containerID="eea082c0c25e5d434ba0790d4062f17b2f38b85338f15e7fa51df5bd9ca931a8" Oct 14 10:05:34 crc kubenswrapper[4870]: E1014 10:05:34.924285 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eea082c0c25e5d434ba0790d4062f17b2f38b85338f15e7fa51df5bd9ca931a8\": container with ID starting with eea082c0c25e5d434ba0790d4062f17b2f38b85338f15e7fa51df5bd9ca931a8 not found: ID does not exist" containerID="eea082c0c25e5d434ba0790d4062f17b2f38b85338f15e7fa51df5bd9ca931a8" Oct 14 10:05:34 crc kubenswrapper[4870]: I1014 10:05:34.924419 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eea082c0c25e5d434ba0790d4062f17b2f38b85338f15e7fa51df5bd9ca931a8"} err="failed to get container status \"eea082c0c25e5d434ba0790d4062f17b2f38b85338f15e7fa51df5bd9ca931a8\": rpc error: code = NotFound desc = could not find container \"eea082c0c25e5d434ba0790d4062f17b2f38b85338f15e7fa51df5bd9ca931a8\": container with ID starting with eea082c0c25e5d434ba0790d4062f17b2f38b85338f15e7fa51df5bd9ca931a8 not found: ID does not exist" Oct 14 10:05:34 crc kubenswrapper[4870]: I1014 10:05:34.924515 4870 scope.go:117] "RemoveContainer" containerID="35a470cba78f4609342b21a00609c4d4f36ba2cf0f83ff14af3b15aeb2aed92a" Oct 14 10:05:34 crc kubenswrapper[4870]: E1014 10:05:34.925004 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"35a470cba78f4609342b21a00609c4d4f36ba2cf0f83ff14af3b15aeb2aed92a\": container with ID starting with 35a470cba78f4609342b21a00609c4d4f36ba2cf0f83ff14af3b15aeb2aed92a not found: ID does not exist" containerID="35a470cba78f4609342b21a00609c4d4f36ba2cf0f83ff14af3b15aeb2aed92a" Oct 14 10:05:34 crc kubenswrapper[4870]: I1014 10:05:34.925048 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"35a470cba78f4609342b21a00609c4d4f36ba2cf0f83ff14af3b15aeb2aed92a"} err="failed to get container status \"35a470cba78f4609342b21a00609c4d4f36ba2cf0f83ff14af3b15aeb2aed92a\": rpc error: code = NotFound desc = could not find container \"35a470cba78f4609342b21a00609c4d4f36ba2cf0f83ff14af3b15aeb2aed92a\": container with ID starting with 35a470cba78f4609342b21a00609c4d4f36ba2cf0f83ff14af3b15aeb2aed92a not found: ID does not exist" Oct 14 10:05:34 crc kubenswrapper[4870]: I1014 10:05:34.925074 4870 scope.go:117] "RemoveContainer" containerID="a2207def43dbd68999ce1608ef65946280c2d57c491c1dff2dab87bf61481544" Oct 14 10:05:34 crc kubenswrapper[4870]: E1014 10:05:34.925532 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a2207def43dbd68999ce1608ef65946280c2d57c491c1dff2dab87bf61481544\": container with ID starting with a2207def43dbd68999ce1608ef65946280c2d57c491c1dff2dab87bf61481544 not found: ID does not exist" containerID="a2207def43dbd68999ce1608ef65946280c2d57c491c1dff2dab87bf61481544" Oct 14 10:05:34 crc kubenswrapper[4870]: I1014 10:05:34.925589 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a2207def43dbd68999ce1608ef65946280c2d57c491c1dff2dab87bf61481544"} err="failed to get container status \"a2207def43dbd68999ce1608ef65946280c2d57c491c1dff2dab87bf61481544\": rpc error: code = NotFound desc = could not find container \"a2207def43dbd68999ce1608ef65946280c2d57c491c1dff2dab87bf61481544\": container with ID starting with a2207def43dbd68999ce1608ef65946280c2d57c491c1dff2dab87bf61481544 not found: ID does not exist" Oct 14 10:05:35 crc kubenswrapper[4870]: I1014 10:05:35.048994 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b0a42940-87de-4a91-b138-20f3bdee6e96" path="/var/lib/kubelet/pods/b0a42940-87de-4a91-b138-20f3bdee6e96/volumes" Oct 14 10:05:39 crc kubenswrapper[4870]: I1014 10:05:39.034519 4870 scope.go:117] "RemoveContainer" containerID="972233f23b285fd01cec82ddb478cfd8b54d11806ccbfd7ce23b7ef84f4bee5c" Oct 14 10:05:39 crc kubenswrapper[4870]: E1014 10:05:39.035770 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 10:05:50 crc kubenswrapper[4870]: I1014 10:05:50.034731 4870 scope.go:117] "RemoveContainer" containerID="972233f23b285fd01cec82ddb478cfd8b54d11806ccbfd7ce23b7ef84f4bee5c" Oct 14 10:05:50 crc kubenswrapper[4870]: E1014 10:05:50.035739 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 10:06:04 crc kubenswrapper[4870]: I1014 10:06:04.034549 4870 scope.go:117] "RemoveContainer" containerID="972233f23b285fd01cec82ddb478cfd8b54d11806ccbfd7ce23b7ef84f4bee5c" Oct 14 10:06:04 crc kubenswrapper[4870]: E1014 10:06:04.035920 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 10:06:17 crc kubenswrapper[4870]: I1014 10:06:17.033999 4870 scope.go:117] "RemoveContainer" containerID="972233f23b285fd01cec82ddb478cfd8b54d11806ccbfd7ce23b7ef84f4bee5c" Oct 14 10:06:17 crc kubenswrapper[4870]: E1014 10:06:17.034684 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 10:06:31 crc kubenswrapper[4870]: I1014 10:06:31.036003 4870 scope.go:117] "RemoveContainer" containerID="972233f23b285fd01cec82ddb478cfd8b54d11806ccbfd7ce23b7ef84f4bee5c" Oct 14 10:06:31 crc kubenswrapper[4870]: E1014 10:06:31.037389 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 10:06:42 crc kubenswrapper[4870]: I1014 10:06:42.034799 4870 scope.go:117] "RemoveContainer" containerID="972233f23b285fd01cec82ddb478cfd8b54d11806ccbfd7ce23b7ef84f4bee5c" Oct 14 10:06:42 crc kubenswrapper[4870]: E1014 10:06:42.036133 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 10:06:58 crc kubenswrapper[4870]: I1014 10:06:58.034995 4870 scope.go:117] "RemoveContainer" containerID="972233f23b285fd01cec82ddb478cfd8b54d11806ccbfd7ce23b7ef84f4bee5c" Oct 14 10:06:58 crc kubenswrapper[4870]: E1014 10:06:58.035987 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 10:07:10 crc kubenswrapper[4870]: I1014 10:07:10.034804 4870 scope.go:117] "RemoveContainer" containerID="972233f23b285fd01cec82ddb478cfd8b54d11806ccbfd7ce23b7ef84f4bee5c" Oct 14 10:07:10 crc kubenswrapper[4870]: E1014 10:07:10.038147 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 10:07:25 crc kubenswrapper[4870]: I1014 10:07:25.033989 4870 scope.go:117] "RemoveContainer" containerID="972233f23b285fd01cec82ddb478cfd8b54d11806ccbfd7ce23b7ef84f4bee5c" Oct 14 10:07:25 crc kubenswrapper[4870]: E1014 10:07:25.034865 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 10:07:38 crc kubenswrapper[4870]: I1014 10:07:38.034602 4870 scope.go:117] "RemoveContainer" containerID="972233f23b285fd01cec82ddb478cfd8b54d11806ccbfd7ce23b7ef84f4bee5c" Oct 14 10:07:38 crc kubenswrapper[4870]: E1014 10:07:38.035744 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 10:07:53 crc kubenswrapper[4870]: I1014 10:07:53.036217 4870 scope.go:117] "RemoveContainer" containerID="972233f23b285fd01cec82ddb478cfd8b54d11806ccbfd7ce23b7ef84f4bee5c" Oct 14 10:07:53 crc kubenswrapper[4870]: E1014 10:07:53.037980 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 10:08:07 crc kubenswrapper[4870]: I1014 10:08:07.034620 4870 scope.go:117] "RemoveContainer" containerID="972233f23b285fd01cec82ddb478cfd8b54d11806ccbfd7ce23b7ef84f4bee5c" Oct 14 10:08:07 crc kubenswrapper[4870]: E1014 10:08:07.037985 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 10:08:20 crc kubenswrapper[4870]: I1014 10:08:20.033886 4870 scope.go:117] "RemoveContainer" containerID="972233f23b285fd01cec82ddb478cfd8b54d11806ccbfd7ce23b7ef84f4bee5c" Oct 14 10:08:20 crc kubenswrapper[4870]: E1014 10:08:20.034746 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 10:08:23 crc kubenswrapper[4870]: I1014 10:08:23.855509 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tempest-tests-tempest"] Oct 14 10:08:23 crc kubenswrapper[4870]: E1014 10:08:23.856940 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b0a42940-87de-4a91-b138-20f3bdee6e96" containerName="extract-utilities" Oct 14 10:08:23 crc kubenswrapper[4870]: I1014 10:08:23.856974 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="b0a42940-87de-4a91-b138-20f3bdee6e96" containerName="extract-utilities" Oct 14 10:08:23 crc kubenswrapper[4870]: E1014 10:08:23.857036 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b0a42940-87de-4a91-b138-20f3bdee6e96" containerName="extract-content" Oct 14 10:08:23 crc kubenswrapper[4870]: I1014 10:08:23.857050 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="b0a42940-87de-4a91-b138-20f3bdee6e96" containerName="extract-content" Oct 14 10:08:23 crc kubenswrapper[4870]: E1014 10:08:23.857099 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b0a42940-87de-4a91-b138-20f3bdee6e96" containerName="registry-server" Oct 14 10:08:23 crc kubenswrapper[4870]: I1014 10:08:23.857113 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="b0a42940-87de-4a91-b138-20f3bdee6e96" containerName="registry-server" Oct 14 10:08:23 crc kubenswrapper[4870]: I1014 10:08:23.857565 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="b0a42940-87de-4a91-b138-20f3bdee6e96" containerName="registry-server" Oct 14 10:08:23 crc kubenswrapper[4870]: I1014 10:08:23.858944 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Oct 14 10:08:23 crc kubenswrapper[4870]: I1014 10:08:23.862782 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-custom-data-s0" Oct 14 10:08:23 crc kubenswrapper[4870]: I1014 10:08:23.863083 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"test-operator-controller-priv-key" Oct 14 10:08:23 crc kubenswrapper[4870]: I1014 10:08:23.863639 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Oct 14 10:08:23 crc kubenswrapper[4870]: I1014 10:08:23.864168 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-vnjjw" Oct 14 10:08:23 crc kubenswrapper[4870]: I1014 10:08:23.867570 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Oct 14 10:08:23 crc kubenswrapper[4870]: I1014 10:08:23.958525 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/2929a25a-a5e9-4a00-8b77-6939e3bf7d66-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"2929a25a-a5e9-4a00-8b77-6939e3bf7d66\") " pod="openstack/tempest-tests-tempest" Oct 14 10:08:23 crc kubenswrapper[4870]: I1014 10:08:23.958608 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2929a25a-a5e9-4a00-8b77-6939e3bf7d66-config-data\") pod \"tempest-tests-tempest\" (UID: \"2929a25a-a5e9-4a00-8b77-6939e3bf7d66\") " pod="openstack/tempest-tests-tempest" Oct 14 10:08:23 crc kubenswrapper[4870]: I1014 10:08:23.958644 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/2929a25a-a5e9-4a00-8b77-6939e3bf7d66-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"2929a25a-a5e9-4a00-8b77-6939e3bf7d66\") " pod="openstack/tempest-tests-tempest" Oct 14 10:08:24 crc kubenswrapper[4870]: I1014 10:08:24.060595 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/2929a25a-a5e9-4a00-8b77-6939e3bf7d66-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"2929a25a-a5e9-4a00-8b77-6939e3bf7d66\") " pod="openstack/tempest-tests-tempest" Oct 14 10:08:24 crc kubenswrapper[4870]: I1014 10:08:24.060683 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2929a25a-a5e9-4a00-8b77-6939e3bf7d66-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"2929a25a-a5e9-4a00-8b77-6939e3bf7d66\") " pod="openstack/tempest-tests-tempest" Oct 14 10:08:24 crc kubenswrapper[4870]: I1014 10:08:24.060736 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/2929a25a-a5e9-4a00-8b77-6939e3bf7d66-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"2929a25a-a5e9-4a00-8b77-6939e3bf7d66\") " pod="openstack/tempest-tests-tempest" Oct 14 10:08:24 crc kubenswrapper[4870]: I1014 10:08:24.060789 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/2929a25a-a5e9-4a00-8b77-6939e3bf7d66-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"2929a25a-a5e9-4a00-8b77-6939e3bf7d66\") " pod="openstack/tempest-tests-tempest" Oct 14 10:08:24 crc kubenswrapper[4870]: I1014 10:08:24.060816 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vbxrd\" (UniqueName: \"kubernetes.io/projected/2929a25a-a5e9-4a00-8b77-6939e3bf7d66-kube-api-access-vbxrd\") pod \"tempest-tests-tempest\" (UID: \"2929a25a-a5e9-4a00-8b77-6939e3bf7d66\") " pod="openstack/tempest-tests-tempest" Oct 14 10:08:24 crc kubenswrapper[4870]: I1014 10:08:24.060863 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2929a25a-a5e9-4a00-8b77-6939e3bf7d66-config-data\") pod \"tempest-tests-tempest\" (UID: \"2929a25a-a5e9-4a00-8b77-6939e3bf7d66\") " pod="openstack/tempest-tests-tempest" Oct 14 10:08:24 crc kubenswrapper[4870]: I1014 10:08:24.060899 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/2929a25a-a5e9-4a00-8b77-6939e3bf7d66-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"2929a25a-a5e9-4a00-8b77-6939e3bf7d66\") " pod="openstack/tempest-tests-tempest" Oct 14 10:08:24 crc kubenswrapper[4870]: I1014 10:08:24.060981 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"tempest-tests-tempest\" (UID: \"2929a25a-a5e9-4a00-8b77-6939e3bf7d66\") " pod="openstack/tempest-tests-tempest" Oct 14 10:08:24 crc kubenswrapper[4870]: I1014 10:08:24.061074 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/2929a25a-a5e9-4a00-8b77-6939e3bf7d66-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"2929a25a-a5e9-4a00-8b77-6939e3bf7d66\") " pod="openstack/tempest-tests-tempest" Oct 14 10:08:24 crc kubenswrapper[4870]: I1014 10:08:24.061833 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/2929a25a-a5e9-4a00-8b77-6939e3bf7d66-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"2929a25a-a5e9-4a00-8b77-6939e3bf7d66\") " pod="openstack/tempest-tests-tempest" Oct 14 10:08:24 crc kubenswrapper[4870]: I1014 10:08:24.062066 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2929a25a-a5e9-4a00-8b77-6939e3bf7d66-config-data\") pod \"tempest-tests-tempest\" (UID: \"2929a25a-a5e9-4a00-8b77-6939e3bf7d66\") " pod="openstack/tempest-tests-tempest" Oct 14 10:08:24 crc kubenswrapper[4870]: I1014 10:08:24.068866 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/2929a25a-a5e9-4a00-8b77-6939e3bf7d66-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"2929a25a-a5e9-4a00-8b77-6939e3bf7d66\") " pod="openstack/tempest-tests-tempest" Oct 14 10:08:24 crc kubenswrapper[4870]: I1014 10:08:24.163023 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"tempest-tests-tempest\" (UID: \"2929a25a-a5e9-4a00-8b77-6939e3bf7d66\") " pod="openstack/tempest-tests-tempest" Oct 14 10:08:24 crc kubenswrapper[4870]: I1014 10:08:24.163203 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/2929a25a-a5e9-4a00-8b77-6939e3bf7d66-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"2929a25a-a5e9-4a00-8b77-6939e3bf7d66\") " pod="openstack/tempest-tests-tempest" Oct 14 10:08:24 crc kubenswrapper[4870]: I1014 10:08:24.163324 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/2929a25a-a5e9-4a00-8b77-6939e3bf7d66-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"2929a25a-a5e9-4a00-8b77-6939e3bf7d66\") " pod="openstack/tempest-tests-tempest" Oct 14 10:08:24 crc kubenswrapper[4870]: I1014 10:08:24.163356 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2929a25a-a5e9-4a00-8b77-6939e3bf7d66-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"2929a25a-a5e9-4a00-8b77-6939e3bf7d66\") " pod="openstack/tempest-tests-tempest" Oct 14 10:08:24 crc kubenswrapper[4870]: I1014 10:08:24.163465 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/2929a25a-a5e9-4a00-8b77-6939e3bf7d66-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"2929a25a-a5e9-4a00-8b77-6939e3bf7d66\") " pod="openstack/tempest-tests-tempest" Oct 14 10:08:24 crc kubenswrapper[4870]: I1014 10:08:24.163506 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vbxrd\" (UniqueName: \"kubernetes.io/projected/2929a25a-a5e9-4a00-8b77-6939e3bf7d66-kube-api-access-vbxrd\") pod \"tempest-tests-tempest\" (UID: \"2929a25a-a5e9-4a00-8b77-6939e3bf7d66\") " pod="openstack/tempest-tests-tempest" Oct 14 10:08:24 crc kubenswrapper[4870]: I1014 10:08:24.163788 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/2929a25a-a5e9-4a00-8b77-6939e3bf7d66-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"2929a25a-a5e9-4a00-8b77-6939e3bf7d66\") " pod="openstack/tempest-tests-tempest" Oct 14 10:08:24 crc kubenswrapper[4870]: I1014 10:08:24.164730 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/2929a25a-a5e9-4a00-8b77-6939e3bf7d66-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"2929a25a-a5e9-4a00-8b77-6939e3bf7d66\") " pod="openstack/tempest-tests-tempest" Oct 14 10:08:24 crc kubenswrapper[4870]: I1014 10:08:24.165262 4870 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"tempest-tests-tempest\" (UID: \"2929a25a-a5e9-4a00-8b77-6939e3bf7d66\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/tempest-tests-tempest" Oct 14 10:08:24 crc kubenswrapper[4870]: I1014 10:08:24.170699 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2929a25a-a5e9-4a00-8b77-6939e3bf7d66-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"2929a25a-a5e9-4a00-8b77-6939e3bf7d66\") " pod="openstack/tempest-tests-tempest" Oct 14 10:08:24 crc kubenswrapper[4870]: I1014 10:08:24.172315 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/2929a25a-a5e9-4a00-8b77-6939e3bf7d66-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"2929a25a-a5e9-4a00-8b77-6939e3bf7d66\") " pod="openstack/tempest-tests-tempest" Oct 14 10:08:24 crc kubenswrapper[4870]: I1014 10:08:24.184751 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vbxrd\" (UniqueName: \"kubernetes.io/projected/2929a25a-a5e9-4a00-8b77-6939e3bf7d66-kube-api-access-vbxrd\") pod \"tempest-tests-tempest\" (UID: \"2929a25a-a5e9-4a00-8b77-6939e3bf7d66\") " pod="openstack/tempest-tests-tempest" Oct 14 10:08:24 crc kubenswrapper[4870]: I1014 10:08:24.218902 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"tempest-tests-tempest\" (UID: \"2929a25a-a5e9-4a00-8b77-6939e3bf7d66\") " pod="openstack/tempest-tests-tempest" Oct 14 10:08:24 crc kubenswrapper[4870]: I1014 10:08:24.234172 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Oct 14 10:08:24 crc kubenswrapper[4870]: I1014 10:08:24.715742 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Oct 14 10:08:24 crc kubenswrapper[4870]: I1014 10:08:24.731701 4870 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 14 10:08:25 crc kubenswrapper[4870]: I1014 10:08:25.135159 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"2929a25a-a5e9-4a00-8b77-6939e3bf7d66","Type":"ContainerStarted","Data":"48b2bf1ed0fca4be2d9e44f216574a00cff664e4ae7a7e02de24a3197a1a84a0"} Oct 14 10:08:35 crc kubenswrapper[4870]: I1014 10:08:35.042041 4870 scope.go:117] "RemoveContainer" containerID="972233f23b285fd01cec82ddb478cfd8b54d11806ccbfd7ce23b7ef84f4bee5c" Oct 14 10:08:35 crc kubenswrapper[4870]: E1014 10:08:35.043243 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 10:08:47 crc kubenswrapper[4870]: I1014 10:08:47.033862 4870 scope.go:117] "RemoveContainer" containerID="972233f23b285fd01cec82ddb478cfd8b54d11806ccbfd7ce23b7ef84f4bee5c" Oct 14 10:08:47 crc kubenswrapper[4870]: E1014 10:08:47.034672 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 10:09:02 crc kubenswrapper[4870]: I1014 10:09:02.034278 4870 scope.go:117] "RemoveContainer" containerID="972233f23b285fd01cec82ddb478cfd8b54d11806ccbfd7ce23b7ef84f4bee5c" Oct 14 10:09:02 crc kubenswrapper[4870]: E1014 10:09:02.036614 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 10:09:09 crc kubenswrapper[4870]: E1014 10:09:09.014674 4870 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-antelope-centos9/openstack-tempest-all:0468cb21803d466b2abfe00835cf1d2d" Oct 14 10:09:09 crc kubenswrapper[4870]: E1014 10:09:09.015401 4870 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-antelope-centos9/openstack-tempest-all:0468cb21803d466b2abfe00835cf1d2d" Oct 14 10:09:09 crc kubenswrapper[4870]: E1014 10:09:09.015640 4870 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:tempest-tests-tempest-tests-runner,Image:quay.rdoproject.org/podified-antelope-centos9/openstack-tempest-all:0468cb21803d466b2abfe00835cf1d2d,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:test-operator-ephemeral-workdir,ReadOnly:false,MountPath:/var/lib/tempest,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-temporary,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/test_operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-logs,ReadOnly:false,MountPath:/var/lib/tempest/external_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/etc/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/var/lib/tempest/.config/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/etc/openstack/secure.yaml,SubPath:secure.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ca-certs,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ssh-key,ReadOnly:false,MountPath:/var/lib/tempest/id_ecdsa,SubPath:ssh_key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-vbxrd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42480,RunAsNonRoot:*false,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:*true,RunAsGroup:*42480,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-custom-data-s0,},Optional:nil,},SecretRef:nil,},EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-env-vars-s0,},Optional:nil,},SecretRef:nil,},},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod tempest-tests-tempest_openstack(2929a25a-a5e9-4a00-8b77-6939e3bf7d66): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 14 10:09:09 crc kubenswrapper[4870]: E1014 10:09:09.017587 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/tempest-tests-tempest" podUID="2929a25a-a5e9-4a00-8b77-6939e3bf7d66" Oct 14 10:09:09 crc kubenswrapper[4870]: E1014 10:09:09.664994 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-antelope-centos9/openstack-tempest-all:0468cb21803d466b2abfe00835cf1d2d\\\"\"" pod="openstack/tempest-tests-tempest" podUID="2929a25a-a5e9-4a00-8b77-6939e3bf7d66" Oct 14 10:09:13 crc kubenswrapper[4870]: I1014 10:09:13.034337 4870 scope.go:117] "RemoveContainer" containerID="972233f23b285fd01cec82ddb478cfd8b54d11806ccbfd7ce23b7ef84f4bee5c" Oct 14 10:09:13 crc kubenswrapper[4870]: E1014 10:09:13.035132 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 10:09:24 crc kubenswrapper[4870]: I1014 10:09:24.232185 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Oct 14 10:09:25 crc kubenswrapper[4870]: I1014 10:09:25.870103 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"2929a25a-a5e9-4a00-8b77-6939e3bf7d66","Type":"ContainerStarted","Data":"264b4c7a0a082d24909cc6d5594818be1076b75d69cff6b79e25a89988f46238"} Oct 14 10:09:25 crc kubenswrapper[4870]: I1014 10:09:25.908381 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tempest-tests-tempest" podStartSLOduration=4.411130509 podStartE2EDuration="1m3.908350506s" podCreationTimestamp="2025-10-14 10:08:22 +0000 UTC" firstStartedPulling="2025-10-14 10:08:24.731375806 +0000 UTC m=+11240.428736197" lastFinishedPulling="2025-10-14 10:09:24.228595743 +0000 UTC m=+11299.925956194" observedRunningTime="2025-10-14 10:09:25.896681999 +0000 UTC m=+11301.594042370" watchObservedRunningTime="2025-10-14 10:09:25.908350506 +0000 UTC m=+11301.605710877" Oct 14 10:09:26 crc kubenswrapper[4870]: I1014 10:09:26.034299 4870 scope.go:117] "RemoveContainer" containerID="972233f23b285fd01cec82ddb478cfd8b54d11806ccbfd7ce23b7ef84f4bee5c" Oct 14 10:09:26 crc kubenswrapper[4870]: E1014 10:09:26.034711 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 10:09:40 crc kubenswrapper[4870]: I1014 10:09:40.033640 4870 scope.go:117] "RemoveContainer" containerID="972233f23b285fd01cec82ddb478cfd8b54d11806ccbfd7ce23b7ef84f4bee5c" Oct 14 10:09:40 crc kubenswrapper[4870]: E1014 10:09:40.034409 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 10:09:55 crc kubenswrapper[4870]: I1014 10:09:55.037959 4870 scope.go:117] "RemoveContainer" containerID="972233f23b285fd01cec82ddb478cfd8b54d11806ccbfd7ce23b7ef84f4bee5c" Oct 14 10:09:55 crc kubenswrapper[4870]: E1014 10:09:55.039294 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 10:10:06 crc kubenswrapper[4870]: I1014 10:10:06.036565 4870 scope.go:117] "RemoveContainer" containerID="972233f23b285fd01cec82ddb478cfd8b54d11806ccbfd7ce23b7ef84f4bee5c" Oct 14 10:10:06 crc kubenswrapper[4870]: E1014 10:10:06.038038 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 10:10:20 crc kubenswrapper[4870]: I1014 10:10:20.033740 4870 scope.go:117] "RemoveContainer" containerID="972233f23b285fd01cec82ddb478cfd8b54d11806ccbfd7ce23b7ef84f4bee5c" Oct 14 10:10:20 crc kubenswrapper[4870]: E1014 10:10:20.034411 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 10:10:35 crc kubenswrapper[4870]: I1014 10:10:35.041899 4870 scope.go:117] "RemoveContainer" containerID="972233f23b285fd01cec82ddb478cfd8b54d11806ccbfd7ce23b7ef84f4bee5c" Oct 14 10:10:35 crc kubenswrapper[4870]: I1014 10:10:35.822006 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" event={"ID":"874a5e45-dffd-4d17-b609-b7d3ed2eab07","Type":"ContainerStarted","Data":"2a49fc200e8e34001c67b2fb36a1c2bb51c1940419a6526be0496325545372e9"} Oct 14 10:10:49 crc kubenswrapper[4870]: I1014 10:10:49.074174 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-t7cs8"] Oct 14 10:10:49 crc kubenswrapper[4870]: I1014 10:10:49.077070 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-t7cs8" Oct 14 10:10:49 crc kubenswrapper[4870]: I1014 10:10:49.091085 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-t7cs8"] Oct 14 10:10:49 crc kubenswrapper[4870]: I1014 10:10:49.164640 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8r4pg\" (UniqueName: \"kubernetes.io/projected/c26243b2-d68a-4eea-9756-2170fa0607d9-kube-api-access-8r4pg\") pod \"community-operators-t7cs8\" (UID: \"c26243b2-d68a-4eea-9756-2170fa0607d9\") " pod="openshift-marketplace/community-operators-t7cs8" Oct 14 10:10:49 crc kubenswrapper[4870]: I1014 10:10:49.164705 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c26243b2-d68a-4eea-9756-2170fa0607d9-catalog-content\") pod \"community-operators-t7cs8\" (UID: \"c26243b2-d68a-4eea-9756-2170fa0607d9\") " pod="openshift-marketplace/community-operators-t7cs8" Oct 14 10:10:49 crc kubenswrapper[4870]: I1014 10:10:49.164877 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c26243b2-d68a-4eea-9756-2170fa0607d9-utilities\") pod \"community-operators-t7cs8\" (UID: \"c26243b2-d68a-4eea-9756-2170fa0607d9\") " pod="openshift-marketplace/community-operators-t7cs8" Oct 14 10:10:49 crc kubenswrapper[4870]: I1014 10:10:49.266360 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c26243b2-d68a-4eea-9756-2170fa0607d9-catalog-content\") pod \"community-operators-t7cs8\" (UID: \"c26243b2-d68a-4eea-9756-2170fa0607d9\") " pod="openshift-marketplace/community-operators-t7cs8" Oct 14 10:10:49 crc kubenswrapper[4870]: I1014 10:10:49.266513 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c26243b2-d68a-4eea-9756-2170fa0607d9-utilities\") pod \"community-operators-t7cs8\" (UID: \"c26243b2-d68a-4eea-9756-2170fa0607d9\") " pod="openshift-marketplace/community-operators-t7cs8" Oct 14 10:10:49 crc kubenswrapper[4870]: I1014 10:10:49.266592 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8r4pg\" (UniqueName: \"kubernetes.io/projected/c26243b2-d68a-4eea-9756-2170fa0607d9-kube-api-access-8r4pg\") pod \"community-operators-t7cs8\" (UID: \"c26243b2-d68a-4eea-9756-2170fa0607d9\") " pod="openshift-marketplace/community-operators-t7cs8" Oct 14 10:10:49 crc kubenswrapper[4870]: I1014 10:10:49.266965 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c26243b2-d68a-4eea-9756-2170fa0607d9-catalog-content\") pod \"community-operators-t7cs8\" (UID: \"c26243b2-d68a-4eea-9756-2170fa0607d9\") " pod="openshift-marketplace/community-operators-t7cs8" Oct 14 10:10:49 crc kubenswrapper[4870]: I1014 10:10:49.266988 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c26243b2-d68a-4eea-9756-2170fa0607d9-utilities\") pod \"community-operators-t7cs8\" (UID: \"c26243b2-d68a-4eea-9756-2170fa0607d9\") " pod="openshift-marketplace/community-operators-t7cs8" Oct 14 10:10:49 crc kubenswrapper[4870]: I1014 10:10:49.287354 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8r4pg\" (UniqueName: \"kubernetes.io/projected/c26243b2-d68a-4eea-9756-2170fa0607d9-kube-api-access-8r4pg\") pod \"community-operators-t7cs8\" (UID: \"c26243b2-d68a-4eea-9756-2170fa0607d9\") " pod="openshift-marketplace/community-operators-t7cs8" Oct 14 10:10:49 crc kubenswrapper[4870]: I1014 10:10:49.401557 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-t7cs8" Oct 14 10:10:50 crc kubenswrapper[4870]: W1014 10:10:50.018989 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc26243b2_d68a_4eea_9756_2170fa0607d9.slice/crio-4eb7210c25ea565b2d91d0a3900f60489f1aa7781424931d7e2743b13386dfd7 WatchSource:0}: Error finding container 4eb7210c25ea565b2d91d0a3900f60489f1aa7781424931d7e2743b13386dfd7: Status 404 returned error can't find the container with id 4eb7210c25ea565b2d91d0a3900f60489f1aa7781424931d7e2743b13386dfd7 Oct 14 10:10:50 crc kubenswrapper[4870]: I1014 10:10:50.021770 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-t7cs8"] Oct 14 10:10:50 crc kubenswrapper[4870]: I1014 10:10:50.975061 4870 generic.go:334] "Generic (PLEG): container finished" podID="c26243b2-d68a-4eea-9756-2170fa0607d9" containerID="b78aeefb00813ea811477b3587ff917c0c852a85d0fac9d61787f855abf9b7b4" exitCode=0 Oct 14 10:10:50 crc kubenswrapper[4870]: I1014 10:10:50.975172 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t7cs8" event={"ID":"c26243b2-d68a-4eea-9756-2170fa0607d9","Type":"ContainerDied","Data":"b78aeefb00813ea811477b3587ff917c0c852a85d0fac9d61787f855abf9b7b4"} Oct 14 10:10:50 crc kubenswrapper[4870]: I1014 10:10:50.975593 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t7cs8" event={"ID":"c26243b2-d68a-4eea-9756-2170fa0607d9","Type":"ContainerStarted","Data":"4eb7210c25ea565b2d91d0a3900f60489f1aa7781424931d7e2743b13386dfd7"} Oct 14 10:10:53 crc kubenswrapper[4870]: I1014 10:10:53.024499 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t7cs8" event={"ID":"c26243b2-d68a-4eea-9756-2170fa0607d9","Type":"ContainerStarted","Data":"f5eaac041cc64b90ddcc4b285cf5a55c4bafd1297838ddef21786b98ceabb80e"} Oct 14 10:10:54 crc kubenswrapper[4870]: I1014 10:10:54.035894 4870 generic.go:334] "Generic (PLEG): container finished" podID="c26243b2-d68a-4eea-9756-2170fa0607d9" containerID="f5eaac041cc64b90ddcc4b285cf5a55c4bafd1297838ddef21786b98ceabb80e" exitCode=0 Oct 14 10:10:54 crc kubenswrapper[4870]: I1014 10:10:54.036000 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t7cs8" event={"ID":"c26243b2-d68a-4eea-9756-2170fa0607d9","Type":"ContainerDied","Data":"f5eaac041cc64b90ddcc4b285cf5a55c4bafd1297838ddef21786b98ceabb80e"} Oct 14 10:10:56 crc kubenswrapper[4870]: I1014 10:10:56.058145 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t7cs8" event={"ID":"c26243b2-d68a-4eea-9756-2170fa0607d9","Type":"ContainerStarted","Data":"fc96f45df6ae8f2e602b9dfe585f5ba811e705217fdba0ebb31e1ab343566ee0"} Oct 14 10:10:56 crc kubenswrapper[4870]: I1014 10:10:56.083319 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-t7cs8" podStartSLOduration=3.07073162 podStartE2EDuration="7.083300738s" podCreationTimestamp="2025-10-14 10:10:49 +0000 UTC" firstStartedPulling="2025-10-14 10:10:50.977772182 +0000 UTC m=+11386.675132563" lastFinishedPulling="2025-10-14 10:10:54.99034131 +0000 UTC m=+11390.687701681" observedRunningTime="2025-10-14 10:10:56.075117226 +0000 UTC m=+11391.772477597" watchObservedRunningTime="2025-10-14 10:10:56.083300738 +0000 UTC m=+11391.780661109" Oct 14 10:10:59 crc kubenswrapper[4870]: I1014 10:10:59.402521 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-t7cs8" Oct 14 10:10:59 crc kubenswrapper[4870]: I1014 10:10:59.403066 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-t7cs8" Oct 14 10:10:59 crc kubenswrapper[4870]: I1014 10:10:59.450403 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-t7cs8" Oct 14 10:11:00 crc kubenswrapper[4870]: I1014 10:11:00.155764 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-t7cs8" Oct 14 10:11:00 crc kubenswrapper[4870]: I1014 10:11:00.263370 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-t7cs8"] Oct 14 10:11:02 crc kubenswrapper[4870]: I1014 10:11:02.108178 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-t7cs8" podUID="c26243b2-d68a-4eea-9756-2170fa0607d9" containerName="registry-server" containerID="cri-o://fc96f45df6ae8f2e602b9dfe585f5ba811e705217fdba0ebb31e1ab343566ee0" gracePeriod=2 Oct 14 10:11:02 crc kubenswrapper[4870]: I1014 10:11:02.697920 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-t7cs8" Oct 14 10:11:02 crc kubenswrapper[4870]: I1014 10:11:02.867619 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c26243b2-d68a-4eea-9756-2170fa0607d9-utilities\") pod \"c26243b2-d68a-4eea-9756-2170fa0607d9\" (UID: \"c26243b2-d68a-4eea-9756-2170fa0607d9\") " Oct 14 10:11:02 crc kubenswrapper[4870]: I1014 10:11:02.868043 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8r4pg\" (UniqueName: \"kubernetes.io/projected/c26243b2-d68a-4eea-9756-2170fa0607d9-kube-api-access-8r4pg\") pod \"c26243b2-d68a-4eea-9756-2170fa0607d9\" (UID: \"c26243b2-d68a-4eea-9756-2170fa0607d9\") " Oct 14 10:11:02 crc kubenswrapper[4870]: I1014 10:11:02.868077 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c26243b2-d68a-4eea-9756-2170fa0607d9-catalog-content\") pod \"c26243b2-d68a-4eea-9756-2170fa0607d9\" (UID: \"c26243b2-d68a-4eea-9756-2170fa0607d9\") " Oct 14 10:11:02 crc kubenswrapper[4870]: I1014 10:11:02.868884 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c26243b2-d68a-4eea-9756-2170fa0607d9-utilities" (OuterVolumeSpecName: "utilities") pod "c26243b2-d68a-4eea-9756-2170fa0607d9" (UID: "c26243b2-d68a-4eea-9756-2170fa0607d9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 10:11:02 crc kubenswrapper[4870]: I1014 10:11:02.877690 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c26243b2-d68a-4eea-9756-2170fa0607d9-kube-api-access-8r4pg" (OuterVolumeSpecName: "kube-api-access-8r4pg") pod "c26243b2-d68a-4eea-9756-2170fa0607d9" (UID: "c26243b2-d68a-4eea-9756-2170fa0607d9"). InnerVolumeSpecName "kube-api-access-8r4pg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 10:11:02 crc kubenswrapper[4870]: I1014 10:11:02.916075 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c26243b2-d68a-4eea-9756-2170fa0607d9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c26243b2-d68a-4eea-9756-2170fa0607d9" (UID: "c26243b2-d68a-4eea-9756-2170fa0607d9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 10:11:02 crc kubenswrapper[4870]: I1014 10:11:02.970346 4870 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c26243b2-d68a-4eea-9756-2170fa0607d9-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 10:11:02 crc kubenswrapper[4870]: I1014 10:11:02.970379 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8r4pg\" (UniqueName: \"kubernetes.io/projected/c26243b2-d68a-4eea-9756-2170fa0607d9-kube-api-access-8r4pg\") on node \"crc\" DevicePath \"\"" Oct 14 10:11:02 crc kubenswrapper[4870]: I1014 10:11:02.970389 4870 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c26243b2-d68a-4eea-9756-2170fa0607d9-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 10:11:03 crc kubenswrapper[4870]: I1014 10:11:03.120600 4870 generic.go:334] "Generic (PLEG): container finished" podID="c26243b2-d68a-4eea-9756-2170fa0607d9" containerID="fc96f45df6ae8f2e602b9dfe585f5ba811e705217fdba0ebb31e1ab343566ee0" exitCode=0 Oct 14 10:11:03 crc kubenswrapper[4870]: I1014 10:11:03.120668 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t7cs8" event={"ID":"c26243b2-d68a-4eea-9756-2170fa0607d9","Type":"ContainerDied","Data":"fc96f45df6ae8f2e602b9dfe585f5ba811e705217fdba0ebb31e1ab343566ee0"} Oct 14 10:11:03 crc kubenswrapper[4870]: I1014 10:11:03.120754 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t7cs8" event={"ID":"c26243b2-d68a-4eea-9756-2170fa0607d9","Type":"ContainerDied","Data":"4eb7210c25ea565b2d91d0a3900f60489f1aa7781424931d7e2743b13386dfd7"} Oct 14 10:11:03 crc kubenswrapper[4870]: I1014 10:11:03.120789 4870 scope.go:117] "RemoveContainer" containerID="fc96f45df6ae8f2e602b9dfe585f5ba811e705217fdba0ebb31e1ab343566ee0" Oct 14 10:11:03 crc kubenswrapper[4870]: I1014 10:11:03.120699 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-t7cs8" Oct 14 10:11:03 crc kubenswrapper[4870]: I1014 10:11:03.154780 4870 scope.go:117] "RemoveContainer" containerID="f5eaac041cc64b90ddcc4b285cf5a55c4bafd1297838ddef21786b98ceabb80e" Oct 14 10:11:03 crc kubenswrapper[4870]: I1014 10:11:03.154827 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-t7cs8"] Oct 14 10:11:03 crc kubenswrapper[4870]: I1014 10:11:03.172183 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-t7cs8"] Oct 14 10:11:03 crc kubenswrapper[4870]: I1014 10:11:03.178305 4870 scope.go:117] "RemoveContainer" containerID="b78aeefb00813ea811477b3587ff917c0c852a85d0fac9d61787f855abf9b7b4" Oct 14 10:11:03 crc kubenswrapper[4870]: I1014 10:11:03.259549 4870 scope.go:117] "RemoveContainer" containerID="fc96f45df6ae8f2e602b9dfe585f5ba811e705217fdba0ebb31e1ab343566ee0" Oct 14 10:11:03 crc kubenswrapper[4870]: E1014 10:11:03.260056 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fc96f45df6ae8f2e602b9dfe585f5ba811e705217fdba0ebb31e1ab343566ee0\": container with ID starting with fc96f45df6ae8f2e602b9dfe585f5ba811e705217fdba0ebb31e1ab343566ee0 not found: ID does not exist" containerID="fc96f45df6ae8f2e602b9dfe585f5ba811e705217fdba0ebb31e1ab343566ee0" Oct 14 10:11:03 crc kubenswrapper[4870]: I1014 10:11:03.260168 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fc96f45df6ae8f2e602b9dfe585f5ba811e705217fdba0ebb31e1ab343566ee0"} err="failed to get container status \"fc96f45df6ae8f2e602b9dfe585f5ba811e705217fdba0ebb31e1ab343566ee0\": rpc error: code = NotFound desc = could not find container \"fc96f45df6ae8f2e602b9dfe585f5ba811e705217fdba0ebb31e1ab343566ee0\": container with ID starting with fc96f45df6ae8f2e602b9dfe585f5ba811e705217fdba0ebb31e1ab343566ee0 not found: ID does not exist" Oct 14 10:11:03 crc kubenswrapper[4870]: I1014 10:11:03.260195 4870 scope.go:117] "RemoveContainer" containerID="f5eaac041cc64b90ddcc4b285cf5a55c4bafd1297838ddef21786b98ceabb80e" Oct 14 10:11:03 crc kubenswrapper[4870]: E1014 10:11:03.260700 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f5eaac041cc64b90ddcc4b285cf5a55c4bafd1297838ddef21786b98ceabb80e\": container with ID starting with f5eaac041cc64b90ddcc4b285cf5a55c4bafd1297838ddef21786b98ceabb80e not found: ID does not exist" containerID="f5eaac041cc64b90ddcc4b285cf5a55c4bafd1297838ddef21786b98ceabb80e" Oct 14 10:11:03 crc kubenswrapper[4870]: I1014 10:11:03.260748 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f5eaac041cc64b90ddcc4b285cf5a55c4bafd1297838ddef21786b98ceabb80e"} err="failed to get container status \"f5eaac041cc64b90ddcc4b285cf5a55c4bafd1297838ddef21786b98ceabb80e\": rpc error: code = NotFound desc = could not find container \"f5eaac041cc64b90ddcc4b285cf5a55c4bafd1297838ddef21786b98ceabb80e\": container with ID starting with f5eaac041cc64b90ddcc4b285cf5a55c4bafd1297838ddef21786b98ceabb80e not found: ID does not exist" Oct 14 10:11:03 crc kubenswrapper[4870]: I1014 10:11:03.260765 4870 scope.go:117] "RemoveContainer" containerID="b78aeefb00813ea811477b3587ff917c0c852a85d0fac9d61787f855abf9b7b4" Oct 14 10:11:03 crc kubenswrapper[4870]: E1014 10:11:03.261223 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b78aeefb00813ea811477b3587ff917c0c852a85d0fac9d61787f855abf9b7b4\": container with ID starting with b78aeefb00813ea811477b3587ff917c0c852a85d0fac9d61787f855abf9b7b4 not found: ID does not exist" containerID="b78aeefb00813ea811477b3587ff917c0c852a85d0fac9d61787f855abf9b7b4" Oct 14 10:11:03 crc kubenswrapper[4870]: I1014 10:11:03.261257 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b78aeefb00813ea811477b3587ff917c0c852a85d0fac9d61787f855abf9b7b4"} err="failed to get container status \"b78aeefb00813ea811477b3587ff917c0c852a85d0fac9d61787f855abf9b7b4\": rpc error: code = NotFound desc = could not find container \"b78aeefb00813ea811477b3587ff917c0c852a85d0fac9d61787f855abf9b7b4\": container with ID starting with b78aeefb00813ea811477b3587ff917c0c852a85d0fac9d61787f855abf9b7b4 not found: ID does not exist" Oct 14 10:11:05 crc kubenswrapper[4870]: I1014 10:11:05.048051 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c26243b2-d68a-4eea-9756-2170fa0607d9" path="/var/lib/kubelet/pods/c26243b2-d68a-4eea-9756-2170fa0607d9/volumes" Oct 14 10:11:39 crc kubenswrapper[4870]: I1014 10:11:39.120079 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-hkjx5"] Oct 14 10:11:39 crc kubenswrapper[4870]: E1014 10:11:39.121216 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c26243b2-d68a-4eea-9756-2170fa0607d9" containerName="extract-content" Oct 14 10:11:39 crc kubenswrapper[4870]: I1014 10:11:39.121235 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="c26243b2-d68a-4eea-9756-2170fa0607d9" containerName="extract-content" Oct 14 10:11:39 crc kubenswrapper[4870]: E1014 10:11:39.121299 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c26243b2-d68a-4eea-9756-2170fa0607d9" containerName="registry-server" Oct 14 10:11:39 crc kubenswrapper[4870]: I1014 10:11:39.121308 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="c26243b2-d68a-4eea-9756-2170fa0607d9" containerName="registry-server" Oct 14 10:11:39 crc kubenswrapper[4870]: E1014 10:11:39.121324 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c26243b2-d68a-4eea-9756-2170fa0607d9" containerName="extract-utilities" Oct 14 10:11:39 crc kubenswrapper[4870]: I1014 10:11:39.121332 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="c26243b2-d68a-4eea-9756-2170fa0607d9" containerName="extract-utilities" Oct 14 10:11:39 crc kubenswrapper[4870]: I1014 10:11:39.121619 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="c26243b2-d68a-4eea-9756-2170fa0607d9" containerName="registry-server" Oct 14 10:11:39 crc kubenswrapper[4870]: I1014 10:11:39.123717 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hkjx5" Oct 14 10:11:39 crc kubenswrapper[4870]: I1014 10:11:39.132431 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-hkjx5"] Oct 14 10:11:39 crc kubenswrapper[4870]: I1014 10:11:39.154186 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b6752061-06dc-4e8b-8144-2e6b6271b8c0-catalog-content\") pod \"certified-operators-hkjx5\" (UID: \"b6752061-06dc-4e8b-8144-2e6b6271b8c0\") " pod="openshift-marketplace/certified-operators-hkjx5" Oct 14 10:11:39 crc kubenswrapper[4870]: I1014 10:11:39.154311 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5km48\" (UniqueName: \"kubernetes.io/projected/b6752061-06dc-4e8b-8144-2e6b6271b8c0-kube-api-access-5km48\") pod \"certified-operators-hkjx5\" (UID: \"b6752061-06dc-4e8b-8144-2e6b6271b8c0\") " pod="openshift-marketplace/certified-operators-hkjx5" Oct 14 10:11:39 crc kubenswrapper[4870]: I1014 10:11:39.154401 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b6752061-06dc-4e8b-8144-2e6b6271b8c0-utilities\") pod \"certified-operators-hkjx5\" (UID: \"b6752061-06dc-4e8b-8144-2e6b6271b8c0\") " pod="openshift-marketplace/certified-operators-hkjx5" Oct 14 10:11:39 crc kubenswrapper[4870]: I1014 10:11:39.256169 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b6752061-06dc-4e8b-8144-2e6b6271b8c0-catalog-content\") pod \"certified-operators-hkjx5\" (UID: \"b6752061-06dc-4e8b-8144-2e6b6271b8c0\") " pod="openshift-marketplace/certified-operators-hkjx5" Oct 14 10:11:39 crc kubenswrapper[4870]: I1014 10:11:39.256273 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5km48\" (UniqueName: \"kubernetes.io/projected/b6752061-06dc-4e8b-8144-2e6b6271b8c0-kube-api-access-5km48\") pod \"certified-operators-hkjx5\" (UID: \"b6752061-06dc-4e8b-8144-2e6b6271b8c0\") " pod="openshift-marketplace/certified-operators-hkjx5" Oct 14 10:11:39 crc kubenswrapper[4870]: I1014 10:11:39.256354 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b6752061-06dc-4e8b-8144-2e6b6271b8c0-utilities\") pod \"certified-operators-hkjx5\" (UID: \"b6752061-06dc-4e8b-8144-2e6b6271b8c0\") " pod="openshift-marketplace/certified-operators-hkjx5" Oct 14 10:11:39 crc kubenswrapper[4870]: I1014 10:11:39.256949 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b6752061-06dc-4e8b-8144-2e6b6271b8c0-catalog-content\") pod \"certified-operators-hkjx5\" (UID: \"b6752061-06dc-4e8b-8144-2e6b6271b8c0\") " pod="openshift-marketplace/certified-operators-hkjx5" Oct 14 10:11:39 crc kubenswrapper[4870]: I1014 10:11:39.256988 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b6752061-06dc-4e8b-8144-2e6b6271b8c0-utilities\") pod \"certified-operators-hkjx5\" (UID: \"b6752061-06dc-4e8b-8144-2e6b6271b8c0\") " pod="openshift-marketplace/certified-operators-hkjx5" Oct 14 10:11:39 crc kubenswrapper[4870]: I1014 10:11:39.287506 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5km48\" (UniqueName: \"kubernetes.io/projected/b6752061-06dc-4e8b-8144-2e6b6271b8c0-kube-api-access-5km48\") pod \"certified-operators-hkjx5\" (UID: \"b6752061-06dc-4e8b-8144-2e6b6271b8c0\") " pod="openshift-marketplace/certified-operators-hkjx5" Oct 14 10:11:39 crc kubenswrapper[4870]: I1014 10:11:39.446200 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hkjx5" Oct 14 10:11:40 crc kubenswrapper[4870]: I1014 10:11:40.022982 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-hkjx5"] Oct 14 10:11:40 crc kubenswrapper[4870]: I1014 10:11:40.538859 4870 generic.go:334] "Generic (PLEG): container finished" podID="b6752061-06dc-4e8b-8144-2e6b6271b8c0" containerID="eabd2aff7c2295fd9d402b837f255b1a9ea554e08ba80215b20548e7090d8b7a" exitCode=0 Oct 14 10:11:40 crc kubenswrapper[4870]: I1014 10:11:40.538965 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hkjx5" event={"ID":"b6752061-06dc-4e8b-8144-2e6b6271b8c0","Type":"ContainerDied","Data":"eabd2aff7c2295fd9d402b837f255b1a9ea554e08ba80215b20548e7090d8b7a"} Oct 14 10:11:40 crc kubenswrapper[4870]: I1014 10:11:40.539321 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hkjx5" event={"ID":"b6752061-06dc-4e8b-8144-2e6b6271b8c0","Type":"ContainerStarted","Data":"bbdee949625673f6f3f4091d06103b08b6d5f1bb623267210760720a97b0dff3"} Oct 14 10:11:42 crc kubenswrapper[4870]: I1014 10:11:42.562964 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hkjx5" event={"ID":"b6752061-06dc-4e8b-8144-2e6b6271b8c0","Type":"ContainerStarted","Data":"af26eae17bca4d68f38f86af476ba61b0672b09373e21da20111fc1cc991eed0"} Oct 14 10:11:44 crc kubenswrapper[4870]: I1014 10:11:44.584860 4870 generic.go:334] "Generic (PLEG): container finished" podID="b6752061-06dc-4e8b-8144-2e6b6271b8c0" containerID="af26eae17bca4d68f38f86af476ba61b0672b09373e21da20111fc1cc991eed0" exitCode=0 Oct 14 10:11:44 crc kubenswrapper[4870]: I1014 10:11:44.584997 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hkjx5" event={"ID":"b6752061-06dc-4e8b-8144-2e6b6271b8c0","Type":"ContainerDied","Data":"af26eae17bca4d68f38f86af476ba61b0672b09373e21da20111fc1cc991eed0"} Oct 14 10:11:45 crc kubenswrapper[4870]: I1014 10:11:45.597561 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hkjx5" event={"ID":"b6752061-06dc-4e8b-8144-2e6b6271b8c0","Type":"ContainerStarted","Data":"e42bb2cba723f5cbbb395e3cd81808d426c1ee349c114399cc4f4e6eeeff8102"} Oct 14 10:11:45 crc kubenswrapper[4870]: I1014 10:11:45.618598 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-hkjx5" podStartSLOduration=2.162479803 podStartE2EDuration="6.618582494s" podCreationTimestamp="2025-10-14 10:11:39 +0000 UTC" firstStartedPulling="2025-10-14 10:11:40.540936344 +0000 UTC m=+11436.238296735" lastFinishedPulling="2025-10-14 10:11:44.997039055 +0000 UTC m=+11440.694399426" observedRunningTime="2025-10-14 10:11:45.614608677 +0000 UTC m=+11441.311969068" watchObservedRunningTime="2025-10-14 10:11:45.618582494 +0000 UTC m=+11441.315942855" Oct 14 10:11:49 crc kubenswrapper[4870]: I1014 10:11:49.446557 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-hkjx5" Oct 14 10:11:49 crc kubenswrapper[4870]: I1014 10:11:49.447990 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-hkjx5" Oct 14 10:11:49 crc kubenswrapper[4870]: I1014 10:11:49.502746 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-hkjx5" Oct 14 10:11:50 crc kubenswrapper[4870]: I1014 10:11:50.724833 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-hkjx5" Oct 14 10:11:50 crc kubenswrapper[4870]: I1014 10:11:50.780014 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-hkjx5"] Oct 14 10:11:52 crc kubenswrapper[4870]: I1014 10:11:52.668049 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-hkjx5" podUID="b6752061-06dc-4e8b-8144-2e6b6271b8c0" containerName="registry-server" containerID="cri-o://e42bb2cba723f5cbbb395e3cd81808d426c1ee349c114399cc4f4e6eeeff8102" gracePeriod=2 Oct 14 10:11:53 crc kubenswrapper[4870]: I1014 10:11:53.379302 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hkjx5" Oct 14 10:11:53 crc kubenswrapper[4870]: I1014 10:11:53.507212 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b6752061-06dc-4e8b-8144-2e6b6271b8c0-catalog-content\") pod \"b6752061-06dc-4e8b-8144-2e6b6271b8c0\" (UID: \"b6752061-06dc-4e8b-8144-2e6b6271b8c0\") " Oct 14 10:11:53 crc kubenswrapper[4870]: I1014 10:11:53.507470 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b6752061-06dc-4e8b-8144-2e6b6271b8c0-utilities\") pod \"b6752061-06dc-4e8b-8144-2e6b6271b8c0\" (UID: \"b6752061-06dc-4e8b-8144-2e6b6271b8c0\") " Oct 14 10:11:53 crc kubenswrapper[4870]: I1014 10:11:53.507539 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5km48\" (UniqueName: \"kubernetes.io/projected/b6752061-06dc-4e8b-8144-2e6b6271b8c0-kube-api-access-5km48\") pod \"b6752061-06dc-4e8b-8144-2e6b6271b8c0\" (UID: \"b6752061-06dc-4e8b-8144-2e6b6271b8c0\") " Oct 14 10:11:53 crc kubenswrapper[4870]: I1014 10:11:53.508520 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b6752061-06dc-4e8b-8144-2e6b6271b8c0-utilities" (OuterVolumeSpecName: "utilities") pod "b6752061-06dc-4e8b-8144-2e6b6271b8c0" (UID: "b6752061-06dc-4e8b-8144-2e6b6271b8c0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 10:11:53 crc kubenswrapper[4870]: I1014 10:11:53.521780 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6752061-06dc-4e8b-8144-2e6b6271b8c0-kube-api-access-5km48" (OuterVolumeSpecName: "kube-api-access-5km48") pod "b6752061-06dc-4e8b-8144-2e6b6271b8c0" (UID: "b6752061-06dc-4e8b-8144-2e6b6271b8c0"). InnerVolumeSpecName "kube-api-access-5km48". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 10:11:53 crc kubenswrapper[4870]: I1014 10:11:53.560793 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b6752061-06dc-4e8b-8144-2e6b6271b8c0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b6752061-06dc-4e8b-8144-2e6b6271b8c0" (UID: "b6752061-06dc-4e8b-8144-2e6b6271b8c0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 10:11:53 crc kubenswrapper[4870]: I1014 10:11:53.611033 4870 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b6752061-06dc-4e8b-8144-2e6b6271b8c0-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 10:11:53 crc kubenswrapper[4870]: I1014 10:11:53.611171 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5km48\" (UniqueName: \"kubernetes.io/projected/b6752061-06dc-4e8b-8144-2e6b6271b8c0-kube-api-access-5km48\") on node \"crc\" DevicePath \"\"" Oct 14 10:11:53 crc kubenswrapper[4870]: I1014 10:11:53.611190 4870 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b6752061-06dc-4e8b-8144-2e6b6271b8c0-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 10:11:53 crc kubenswrapper[4870]: I1014 10:11:53.709095 4870 generic.go:334] "Generic (PLEG): container finished" podID="b6752061-06dc-4e8b-8144-2e6b6271b8c0" containerID="e42bb2cba723f5cbbb395e3cd81808d426c1ee349c114399cc4f4e6eeeff8102" exitCode=0 Oct 14 10:11:53 crc kubenswrapper[4870]: I1014 10:11:53.709174 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hkjx5" event={"ID":"b6752061-06dc-4e8b-8144-2e6b6271b8c0","Type":"ContainerDied","Data":"e42bb2cba723f5cbbb395e3cd81808d426c1ee349c114399cc4f4e6eeeff8102"} Oct 14 10:11:53 crc kubenswrapper[4870]: I1014 10:11:53.709216 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hkjx5" event={"ID":"b6752061-06dc-4e8b-8144-2e6b6271b8c0","Type":"ContainerDied","Data":"bbdee949625673f6f3f4091d06103b08b6d5f1bb623267210760720a97b0dff3"} Oct 14 10:11:53 crc kubenswrapper[4870]: I1014 10:11:53.709241 4870 scope.go:117] "RemoveContainer" containerID="e42bb2cba723f5cbbb395e3cd81808d426c1ee349c114399cc4f4e6eeeff8102" Oct 14 10:11:53 crc kubenswrapper[4870]: I1014 10:11:53.710132 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hkjx5" Oct 14 10:11:53 crc kubenswrapper[4870]: I1014 10:11:53.754254 4870 scope.go:117] "RemoveContainer" containerID="af26eae17bca4d68f38f86af476ba61b0672b09373e21da20111fc1cc991eed0" Oct 14 10:11:53 crc kubenswrapper[4870]: I1014 10:11:53.774879 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-hkjx5"] Oct 14 10:11:53 crc kubenswrapper[4870]: I1014 10:11:53.789809 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-hkjx5"] Oct 14 10:11:53 crc kubenswrapper[4870]: I1014 10:11:53.793387 4870 scope.go:117] "RemoveContainer" containerID="eabd2aff7c2295fd9d402b837f255b1a9ea554e08ba80215b20548e7090d8b7a" Oct 14 10:11:53 crc kubenswrapper[4870]: I1014 10:11:53.839660 4870 scope.go:117] "RemoveContainer" containerID="e42bb2cba723f5cbbb395e3cd81808d426c1ee349c114399cc4f4e6eeeff8102" Oct 14 10:11:53 crc kubenswrapper[4870]: E1014 10:11:53.840199 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e42bb2cba723f5cbbb395e3cd81808d426c1ee349c114399cc4f4e6eeeff8102\": container with ID starting with e42bb2cba723f5cbbb395e3cd81808d426c1ee349c114399cc4f4e6eeeff8102 not found: ID does not exist" containerID="e42bb2cba723f5cbbb395e3cd81808d426c1ee349c114399cc4f4e6eeeff8102" Oct 14 10:11:53 crc kubenswrapper[4870]: I1014 10:11:53.840238 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e42bb2cba723f5cbbb395e3cd81808d426c1ee349c114399cc4f4e6eeeff8102"} err="failed to get container status \"e42bb2cba723f5cbbb395e3cd81808d426c1ee349c114399cc4f4e6eeeff8102\": rpc error: code = NotFound desc = could not find container \"e42bb2cba723f5cbbb395e3cd81808d426c1ee349c114399cc4f4e6eeeff8102\": container with ID starting with e42bb2cba723f5cbbb395e3cd81808d426c1ee349c114399cc4f4e6eeeff8102 not found: ID does not exist" Oct 14 10:11:53 crc kubenswrapper[4870]: I1014 10:11:53.840270 4870 scope.go:117] "RemoveContainer" containerID="af26eae17bca4d68f38f86af476ba61b0672b09373e21da20111fc1cc991eed0" Oct 14 10:11:53 crc kubenswrapper[4870]: E1014 10:11:53.841188 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"af26eae17bca4d68f38f86af476ba61b0672b09373e21da20111fc1cc991eed0\": container with ID starting with af26eae17bca4d68f38f86af476ba61b0672b09373e21da20111fc1cc991eed0 not found: ID does not exist" containerID="af26eae17bca4d68f38f86af476ba61b0672b09373e21da20111fc1cc991eed0" Oct 14 10:11:53 crc kubenswrapper[4870]: I1014 10:11:53.841258 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"af26eae17bca4d68f38f86af476ba61b0672b09373e21da20111fc1cc991eed0"} err="failed to get container status \"af26eae17bca4d68f38f86af476ba61b0672b09373e21da20111fc1cc991eed0\": rpc error: code = NotFound desc = could not find container \"af26eae17bca4d68f38f86af476ba61b0672b09373e21da20111fc1cc991eed0\": container with ID starting with af26eae17bca4d68f38f86af476ba61b0672b09373e21da20111fc1cc991eed0 not found: ID does not exist" Oct 14 10:11:53 crc kubenswrapper[4870]: I1014 10:11:53.841298 4870 scope.go:117] "RemoveContainer" containerID="eabd2aff7c2295fd9d402b837f255b1a9ea554e08ba80215b20548e7090d8b7a" Oct 14 10:11:53 crc kubenswrapper[4870]: E1014 10:11:53.841690 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eabd2aff7c2295fd9d402b837f255b1a9ea554e08ba80215b20548e7090d8b7a\": container with ID starting with eabd2aff7c2295fd9d402b837f255b1a9ea554e08ba80215b20548e7090d8b7a not found: ID does not exist" containerID="eabd2aff7c2295fd9d402b837f255b1a9ea554e08ba80215b20548e7090d8b7a" Oct 14 10:11:53 crc kubenswrapper[4870]: I1014 10:11:53.841722 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eabd2aff7c2295fd9d402b837f255b1a9ea554e08ba80215b20548e7090d8b7a"} err="failed to get container status \"eabd2aff7c2295fd9d402b837f255b1a9ea554e08ba80215b20548e7090d8b7a\": rpc error: code = NotFound desc = could not find container \"eabd2aff7c2295fd9d402b837f255b1a9ea554e08ba80215b20548e7090d8b7a\": container with ID starting with eabd2aff7c2295fd9d402b837f255b1a9ea554e08ba80215b20548e7090d8b7a not found: ID does not exist" Oct 14 10:11:55 crc kubenswrapper[4870]: I1014 10:11:55.052613 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6752061-06dc-4e8b-8144-2e6b6271b8c0" path="/var/lib/kubelet/pods/b6752061-06dc-4e8b-8144-2e6b6271b8c0/volumes" Oct 14 10:12:53 crc kubenswrapper[4870]: I1014 10:12:53.950870 4870 patch_prober.go:28] interesting pod/machine-config-daemon-7tvc7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 10:12:53 crc kubenswrapper[4870]: I1014 10:12:53.951986 4870 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 10:13:23 crc kubenswrapper[4870]: I1014 10:13:23.950747 4870 patch_prober.go:28] interesting pod/machine-config-daemon-7tvc7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 10:13:23 crc kubenswrapper[4870]: I1014 10:13:23.951319 4870 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 10:13:53 crc kubenswrapper[4870]: I1014 10:13:53.950526 4870 patch_prober.go:28] interesting pod/machine-config-daemon-7tvc7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 10:13:53 crc kubenswrapper[4870]: I1014 10:13:53.951098 4870 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 10:13:53 crc kubenswrapper[4870]: I1014 10:13:53.951156 4870 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" Oct 14 10:13:53 crc kubenswrapper[4870]: I1014 10:13:53.952039 4870 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"2a49fc200e8e34001c67b2fb36a1c2bb51c1940419a6526be0496325545372e9"} pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 14 10:13:53 crc kubenswrapper[4870]: I1014 10:13:53.952108 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerName="machine-config-daemon" containerID="cri-o://2a49fc200e8e34001c67b2fb36a1c2bb51c1940419a6526be0496325545372e9" gracePeriod=600 Oct 14 10:13:54 crc kubenswrapper[4870]: I1014 10:13:54.135523 4870 generic.go:334] "Generic (PLEG): container finished" podID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerID="2a49fc200e8e34001c67b2fb36a1c2bb51c1940419a6526be0496325545372e9" exitCode=0 Oct 14 10:13:54 crc kubenswrapper[4870]: I1014 10:13:54.135631 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" event={"ID":"874a5e45-dffd-4d17-b609-b7d3ed2eab07","Type":"ContainerDied","Data":"2a49fc200e8e34001c67b2fb36a1c2bb51c1940419a6526be0496325545372e9"} Oct 14 10:13:54 crc kubenswrapper[4870]: I1014 10:13:54.135883 4870 scope.go:117] "RemoveContainer" containerID="972233f23b285fd01cec82ddb478cfd8b54d11806ccbfd7ce23b7ef84f4bee5c" Oct 14 10:13:55 crc kubenswrapper[4870]: I1014 10:13:55.148353 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" event={"ID":"874a5e45-dffd-4d17-b609-b7d3ed2eab07","Type":"ContainerStarted","Data":"7468e2739d25e53085cbe91cf249f36e5760f983ae0240cf58bcc9bcb1b91355"} Oct 14 10:14:57 crc kubenswrapper[4870]: I1014 10:14:57.853369 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-ghnkh"] Oct 14 10:14:57 crc kubenswrapper[4870]: E1014 10:14:57.854792 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6752061-06dc-4e8b-8144-2e6b6271b8c0" containerName="registry-server" Oct 14 10:14:57 crc kubenswrapper[4870]: I1014 10:14:57.854811 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6752061-06dc-4e8b-8144-2e6b6271b8c0" containerName="registry-server" Oct 14 10:14:57 crc kubenswrapper[4870]: E1014 10:14:57.854856 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6752061-06dc-4e8b-8144-2e6b6271b8c0" containerName="extract-utilities" Oct 14 10:14:57 crc kubenswrapper[4870]: I1014 10:14:57.854864 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6752061-06dc-4e8b-8144-2e6b6271b8c0" containerName="extract-utilities" Oct 14 10:14:57 crc kubenswrapper[4870]: E1014 10:14:57.854890 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6752061-06dc-4e8b-8144-2e6b6271b8c0" containerName="extract-content" Oct 14 10:14:57 crc kubenswrapper[4870]: I1014 10:14:57.854898 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6752061-06dc-4e8b-8144-2e6b6271b8c0" containerName="extract-content" Oct 14 10:14:57 crc kubenswrapper[4870]: I1014 10:14:57.855138 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="b6752061-06dc-4e8b-8144-2e6b6271b8c0" containerName="registry-server" Oct 14 10:14:57 crc kubenswrapper[4870]: I1014 10:14:57.856788 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ghnkh" Oct 14 10:14:57 crc kubenswrapper[4870]: I1014 10:14:57.887244 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-ghnkh"] Oct 14 10:14:57 crc kubenswrapper[4870]: I1014 10:14:57.947334 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c274b0a9-ce86-4ad1-82e1-509592932675-catalog-content\") pod \"redhat-marketplace-ghnkh\" (UID: \"c274b0a9-ce86-4ad1-82e1-509592932675\") " pod="openshift-marketplace/redhat-marketplace-ghnkh" Oct 14 10:14:57 crc kubenswrapper[4870]: I1014 10:14:57.947727 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b6c6p\" (UniqueName: \"kubernetes.io/projected/c274b0a9-ce86-4ad1-82e1-509592932675-kube-api-access-b6c6p\") pod \"redhat-marketplace-ghnkh\" (UID: \"c274b0a9-ce86-4ad1-82e1-509592932675\") " pod="openshift-marketplace/redhat-marketplace-ghnkh" Oct 14 10:14:57 crc kubenswrapper[4870]: I1014 10:14:57.948250 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c274b0a9-ce86-4ad1-82e1-509592932675-utilities\") pod \"redhat-marketplace-ghnkh\" (UID: \"c274b0a9-ce86-4ad1-82e1-509592932675\") " pod="openshift-marketplace/redhat-marketplace-ghnkh" Oct 14 10:14:58 crc kubenswrapper[4870]: I1014 10:14:58.050231 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c274b0a9-ce86-4ad1-82e1-509592932675-utilities\") pod \"redhat-marketplace-ghnkh\" (UID: \"c274b0a9-ce86-4ad1-82e1-509592932675\") " pod="openshift-marketplace/redhat-marketplace-ghnkh" Oct 14 10:14:58 crc kubenswrapper[4870]: I1014 10:14:58.050586 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c274b0a9-ce86-4ad1-82e1-509592932675-catalog-content\") pod \"redhat-marketplace-ghnkh\" (UID: \"c274b0a9-ce86-4ad1-82e1-509592932675\") " pod="openshift-marketplace/redhat-marketplace-ghnkh" Oct 14 10:14:58 crc kubenswrapper[4870]: I1014 10:14:58.050732 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b6c6p\" (UniqueName: \"kubernetes.io/projected/c274b0a9-ce86-4ad1-82e1-509592932675-kube-api-access-b6c6p\") pod \"redhat-marketplace-ghnkh\" (UID: \"c274b0a9-ce86-4ad1-82e1-509592932675\") " pod="openshift-marketplace/redhat-marketplace-ghnkh" Oct 14 10:14:58 crc kubenswrapper[4870]: I1014 10:14:58.052068 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c274b0a9-ce86-4ad1-82e1-509592932675-catalog-content\") pod \"redhat-marketplace-ghnkh\" (UID: \"c274b0a9-ce86-4ad1-82e1-509592932675\") " pod="openshift-marketplace/redhat-marketplace-ghnkh" Oct 14 10:14:58 crc kubenswrapper[4870]: I1014 10:14:58.053159 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c274b0a9-ce86-4ad1-82e1-509592932675-utilities\") pod \"redhat-marketplace-ghnkh\" (UID: \"c274b0a9-ce86-4ad1-82e1-509592932675\") " pod="openshift-marketplace/redhat-marketplace-ghnkh" Oct 14 10:14:58 crc kubenswrapper[4870]: I1014 10:14:58.077500 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b6c6p\" (UniqueName: \"kubernetes.io/projected/c274b0a9-ce86-4ad1-82e1-509592932675-kube-api-access-b6c6p\") pod \"redhat-marketplace-ghnkh\" (UID: \"c274b0a9-ce86-4ad1-82e1-509592932675\") " pod="openshift-marketplace/redhat-marketplace-ghnkh" Oct 14 10:14:58 crc kubenswrapper[4870]: I1014 10:14:58.176221 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ghnkh" Oct 14 10:14:58 crc kubenswrapper[4870]: I1014 10:14:58.790751 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-ghnkh"] Oct 14 10:14:58 crc kubenswrapper[4870]: I1014 10:14:58.902789 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ghnkh" event={"ID":"c274b0a9-ce86-4ad1-82e1-509592932675","Type":"ContainerStarted","Data":"60b6a5a27e569f514345ba15da71fcb087a17b288682109c2a021f258ccc2bd5"} Oct 14 10:14:59 crc kubenswrapper[4870]: I1014 10:14:59.915643 4870 generic.go:334] "Generic (PLEG): container finished" podID="c274b0a9-ce86-4ad1-82e1-509592932675" containerID="c4cd8760991ef75744cbca3febabfa3fd4340af9d0b2598198dd4f21efa2281e" exitCode=0 Oct 14 10:14:59 crc kubenswrapper[4870]: I1014 10:14:59.915728 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ghnkh" event={"ID":"c274b0a9-ce86-4ad1-82e1-509592932675","Type":"ContainerDied","Data":"c4cd8760991ef75744cbca3febabfa3fd4340af9d0b2598198dd4f21efa2281e"} Oct 14 10:14:59 crc kubenswrapper[4870]: I1014 10:14:59.919209 4870 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 14 10:15:00 crc kubenswrapper[4870]: I1014 10:15:00.150680 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29340615-2kzb6"] Oct 14 10:15:00 crc kubenswrapper[4870]: I1014 10:15:00.152151 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29340615-2kzb6" Oct 14 10:15:00 crc kubenswrapper[4870]: I1014 10:15:00.155078 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 14 10:15:00 crc kubenswrapper[4870]: I1014 10:15:00.155789 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 14 10:15:00 crc kubenswrapper[4870]: I1014 10:15:00.204054 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29340615-2kzb6"] Oct 14 10:15:00 crc kubenswrapper[4870]: I1014 10:15:00.300945 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8fe77a0f-ea5d-48f3-9206-f4a4813f0818-config-volume\") pod \"collect-profiles-29340615-2kzb6\" (UID: \"8fe77a0f-ea5d-48f3-9206-f4a4813f0818\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340615-2kzb6" Oct 14 10:15:00 crc kubenswrapper[4870]: I1014 10:15:00.301057 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l7m8j\" (UniqueName: \"kubernetes.io/projected/8fe77a0f-ea5d-48f3-9206-f4a4813f0818-kube-api-access-l7m8j\") pod \"collect-profiles-29340615-2kzb6\" (UID: \"8fe77a0f-ea5d-48f3-9206-f4a4813f0818\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340615-2kzb6" Oct 14 10:15:00 crc kubenswrapper[4870]: I1014 10:15:00.301207 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8fe77a0f-ea5d-48f3-9206-f4a4813f0818-secret-volume\") pod \"collect-profiles-29340615-2kzb6\" (UID: \"8fe77a0f-ea5d-48f3-9206-f4a4813f0818\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340615-2kzb6" Oct 14 10:15:00 crc kubenswrapper[4870]: I1014 10:15:00.403116 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8fe77a0f-ea5d-48f3-9206-f4a4813f0818-config-volume\") pod \"collect-profiles-29340615-2kzb6\" (UID: \"8fe77a0f-ea5d-48f3-9206-f4a4813f0818\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340615-2kzb6" Oct 14 10:15:00 crc kubenswrapper[4870]: I1014 10:15:00.403246 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l7m8j\" (UniqueName: \"kubernetes.io/projected/8fe77a0f-ea5d-48f3-9206-f4a4813f0818-kube-api-access-l7m8j\") pod \"collect-profiles-29340615-2kzb6\" (UID: \"8fe77a0f-ea5d-48f3-9206-f4a4813f0818\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340615-2kzb6" Oct 14 10:15:00 crc kubenswrapper[4870]: I1014 10:15:00.403338 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8fe77a0f-ea5d-48f3-9206-f4a4813f0818-secret-volume\") pod \"collect-profiles-29340615-2kzb6\" (UID: \"8fe77a0f-ea5d-48f3-9206-f4a4813f0818\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340615-2kzb6" Oct 14 10:15:00 crc kubenswrapper[4870]: I1014 10:15:00.404213 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8fe77a0f-ea5d-48f3-9206-f4a4813f0818-config-volume\") pod \"collect-profiles-29340615-2kzb6\" (UID: \"8fe77a0f-ea5d-48f3-9206-f4a4813f0818\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340615-2kzb6" Oct 14 10:15:00 crc kubenswrapper[4870]: I1014 10:15:00.417154 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8fe77a0f-ea5d-48f3-9206-f4a4813f0818-secret-volume\") pod \"collect-profiles-29340615-2kzb6\" (UID: \"8fe77a0f-ea5d-48f3-9206-f4a4813f0818\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340615-2kzb6" Oct 14 10:15:00 crc kubenswrapper[4870]: I1014 10:15:00.428271 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l7m8j\" (UniqueName: \"kubernetes.io/projected/8fe77a0f-ea5d-48f3-9206-f4a4813f0818-kube-api-access-l7m8j\") pod \"collect-profiles-29340615-2kzb6\" (UID: \"8fe77a0f-ea5d-48f3-9206-f4a4813f0818\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340615-2kzb6" Oct 14 10:15:00 crc kubenswrapper[4870]: I1014 10:15:00.480520 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29340615-2kzb6" Oct 14 10:15:00 crc kubenswrapper[4870]: I1014 10:15:00.928886 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ghnkh" event={"ID":"c274b0a9-ce86-4ad1-82e1-509592932675","Type":"ContainerStarted","Data":"b1eb9272a96698fcbc0f50215144913e566b61e2460f8b3034ce54949c2e4188"} Oct 14 10:15:00 crc kubenswrapper[4870]: W1014 10:15:00.984012 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8fe77a0f_ea5d_48f3_9206_f4a4813f0818.slice/crio-e39cd57f202b463527bca6349ead1c384bdfe6531f1a08ad6cef4b23325edc86 WatchSource:0}: Error finding container e39cd57f202b463527bca6349ead1c384bdfe6531f1a08ad6cef4b23325edc86: Status 404 returned error can't find the container with id e39cd57f202b463527bca6349ead1c384bdfe6531f1a08ad6cef4b23325edc86 Oct 14 10:15:00 crc kubenswrapper[4870]: I1014 10:15:00.985846 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29340615-2kzb6"] Oct 14 10:15:01 crc kubenswrapper[4870]: I1014 10:15:01.947711 4870 generic.go:334] "Generic (PLEG): container finished" podID="8fe77a0f-ea5d-48f3-9206-f4a4813f0818" containerID="e49b85bc6295bf0eab0051f0f4162705e882da763a5539e6c0e65993261112f5" exitCode=0 Oct 14 10:15:01 crc kubenswrapper[4870]: I1014 10:15:01.947797 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29340615-2kzb6" event={"ID":"8fe77a0f-ea5d-48f3-9206-f4a4813f0818","Type":"ContainerDied","Data":"e49b85bc6295bf0eab0051f0f4162705e882da763a5539e6c0e65993261112f5"} Oct 14 10:15:01 crc kubenswrapper[4870]: I1014 10:15:01.948034 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29340615-2kzb6" event={"ID":"8fe77a0f-ea5d-48f3-9206-f4a4813f0818","Type":"ContainerStarted","Data":"e39cd57f202b463527bca6349ead1c384bdfe6531f1a08ad6cef4b23325edc86"} Oct 14 10:15:01 crc kubenswrapper[4870]: I1014 10:15:01.952056 4870 generic.go:334] "Generic (PLEG): container finished" podID="c274b0a9-ce86-4ad1-82e1-509592932675" containerID="b1eb9272a96698fcbc0f50215144913e566b61e2460f8b3034ce54949c2e4188" exitCode=0 Oct 14 10:15:01 crc kubenswrapper[4870]: I1014 10:15:01.952126 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ghnkh" event={"ID":"c274b0a9-ce86-4ad1-82e1-509592932675","Type":"ContainerDied","Data":"b1eb9272a96698fcbc0f50215144913e566b61e2460f8b3034ce54949c2e4188"} Oct 14 10:15:02 crc kubenswrapper[4870]: I1014 10:15:02.966520 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ghnkh" event={"ID":"c274b0a9-ce86-4ad1-82e1-509592932675","Type":"ContainerStarted","Data":"cf110bc597b7ae7d07aff32a30836bc98e772dea49b3ca0511e631d633872f22"} Oct 14 10:15:02 crc kubenswrapper[4870]: I1014 10:15:02.992536 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-ghnkh" podStartSLOduration=3.479481551 podStartE2EDuration="5.992510897s" podCreationTimestamp="2025-10-14 10:14:57 +0000 UTC" firstStartedPulling="2025-10-14 10:14:59.918940422 +0000 UTC m=+11635.616300803" lastFinishedPulling="2025-10-14 10:15:02.431969778 +0000 UTC m=+11638.129330149" observedRunningTime="2025-10-14 10:15:02.985188447 +0000 UTC m=+11638.682548838" watchObservedRunningTime="2025-10-14 10:15:02.992510897 +0000 UTC m=+11638.689871288" Oct 14 10:15:03 crc kubenswrapper[4870]: I1014 10:15:03.616871 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29340615-2kzb6" Oct 14 10:15:03 crc kubenswrapper[4870]: I1014 10:15:03.681339 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8fe77a0f-ea5d-48f3-9206-f4a4813f0818-config-volume\") pod \"8fe77a0f-ea5d-48f3-9206-f4a4813f0818\" (UID: \"8fe77a0f-ea5d-48f3-9206-f4a4813f0818\") " Oct 14 10:15:03 crc kubenswrapper[4870]: I1014 10:15:03.681581 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8fe77a0f-ea5d-48f3-9206-f4a4813f0818-secret-volume\") pod \"8fe77a0f-ea5d-48f3-9206-f4a4813f0818\" (UID: \"8fe77a0f-ea5d-48f3-9206-f4a4813f0818\") " Oct 14 10:15:03 crc kubenswrapper[4870]: I1014 10:15:03.681739 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l7m8j\" (UniqueName: \"kubernetes.io/projected/8fe77a0f-ea5d-48f3-9206-f4a4813f0818-kube-api-access-l7m8j\") pod \"8fe77a0f-ea5d-48f3-9206-f4a4813f0818\" (UID: \"8fe77a0f-ea5d-48f3-9206-f4a4813f0818\") " Oct 14 10:15:03 crc kubenswrapper[4870]: I1014 10:15:03.698778 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8fe77a0f-ea5d-48f3-9206-f4a4813f0818-kube-api-access-l7m8j" (OuterVolumeSpecName: "kube-api-access-l7m8j") pod "8fe77a0f-ea5d-48f3-9206-f4a4813f0818" (UID: "8fe77a0f-ea5d-48f3-9206-f4a4813f0818"). InnerVolumeSpecName "kube-api-access-l7m8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 10:15:03 crc kubenswrapper[4870]: I1014 10:15:03.698942 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8fe77a0f-ea5d-48f3-9206-f4a4813f0818-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "8fe77a0f-ea5d-48f3-9206-f4a4813f0818" (UID: "8fe77a0f-ea5d-48f3-9206-f4a4813f0818"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 10:15:03 crc kubenswrapper[4870]: I1014 10:15:03.713394 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8fe77a0f-ea5d-48f3-9206-f4a4813f0818-config-volume" (OuterVolumeSpecName: "config-volume") pod "8fe77a0f-ea5d-48f3-9206-f4a4813f0818" (UID: "8fe77a0f-ea5d-48f3-9206-f4a4813f0818"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 10:15:03 crc kubenswrapper[4870]: I1014 10:15:03.785780 4870 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8fe77a0f-ea5d-48f3-9206-f4a4813f0818-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 14 10:15:03 crc kubenswrapper[4870]: I1014 10:15:03.785832 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l7m8j\" (UniqueName: \"kubernetes.io/projected/8fe77a0f-ea5d-48f3-9206-f4a4813f0818-kube-api-access-l7m8j\") on node \"crc\" DevicePath \"\"" Oct 14 10:15:03 crc kubenswrapper[4870]: I1014 10:15:03.785846 4870 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8fe77a0f-ea5d-48f3-9206-f4a4813f0818-config-volume\") on node \"crc\" DevicePath \"\"" Oct 14 10:15:03 crc kubenswrapper[4870]: I1014 10:15:03.976757 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29340615-2kzb6" Oct 14 10:15:03 crc kubenswrapper[4870]: I1014 10:15:03.976945 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29340615-2kzb6" event={"ID":"8fe77a0f-ea5d-48f3-9206-f4a4813f0818","Type":"ContainerDied","Data":"e39cd57f202b463527bca6349ead1c384bdfe6531f1a08ad6cef4b23325edc86"} Oct 14 10:15:03 crc kubenswrapper[4870]: I1014 10:15:03.977790 4870 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e39cd57f202b463527bca6349ead1c384bdfe6531f1a08ad6cef4b23325edc86" Oct 14 10:15:04 crc kubenswrapper[4870]: I1014 10:15:04.690676 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29340570-kh9fl"] Oct 14 10:15:04 crc kubenswrapper[4870]: I1014 10:15:04.700590 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29340570-kh9fl"] Oct 14 10:15:05 crc kubenswrapper[4870]: I1014 10:15:05.053348 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="557da624-75cc-4387-8771-241da876c655" path="/var/lib/kubelet/pods/557da624-75cc-4387-8771-241da876c655/volumes" Oct 14 10:15:08 crc kubenswrapper[4870]: I1014 10:15:08.177491 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-ghnkh" Oct 14 10:15:08 crc kubenswrapper[4870]: I1014 10:15:08.177909 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-ghnkh" Oct 14 10:15:08 crc kubenswrapper[4870]: I1014 10:15:08.240059 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-ghnkh" Oct 14 10:15:09 crc kubenswrapper[4870]: I1014 10:15:09.096434 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-ghnkh" Oct 14 10:15:09 crc kubenswrapper[4870]: I1014 10:15:09.148273 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-ghnkh"] Oct 14 10:15:11 crc kubenswrapper[4870]: I1014 10:15:11.074226 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-ghnkh" podUID="c274b0a9-ce86-4ad1-82e1-509592932675" containerName="registry-server" containerID="cri-o://cf110bc597b7ae7d07aff32a30836bc98e772dea49b3ca0511e631d633872f22" gracePeriod=2 Oct 14 10:15:11 crc kubenswrapper[4870]: I1014 10:15:11.843842 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ghnkh" Oct 14 10:15:11 crc kubenswrapper[4870]: I1014 10:15:11.998091 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c274b0a9-ce86-4ad1-82e1-509592932675-utilities\") pod \"c274b0a9-ce86-4ad1-82e1-509592932675\" (UID: \"c274b0a9-ce86-4ad1-82e1-509592932675\") " Oct 14 10:15:11 crc kubenswrapper[4870]: I1014 10:15:11.998528 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b6c6p\" (UniqueName: \"kubernetes.io/projected/c274b0a9-ce86-4ad1-82e1-509592932675-kube-api-access-b6c6p\") pod \"c274b0a9-ce86-4ad1-82e1-509592932675\" (UID: \"c274b0a9-ce86-4ad1-82e1-509592932675\") " Oct 14 10:15:11 crc kubenswrapper[4870]: I1014 10:15:11.998608 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c274b0a9-ce86-4ad1-82e1-509592932675-catalog-content\") pod \"c274b0a9-ce86-4ad1-82e1-509592932675\" (UID: \"c274b0a9-ce86-4ad1-82e1-509592932675\") " Oct 14 10:15:12 crc kubenswrapper[4870]: I1014 10:15:12.001486 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c274b0a9-ce86-4ad1-82e1-509592932675-utilities" (OuterVolumeSpecName: "utilities") pod "c274b0a9-ce86-4ad1-82e1-509592932675" (UID: "c274b0a9-ce86-4ad1-82e1-509592932675"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 10:15:12 crc kubenswrapper[4870]: I1014 10:15:12.011436 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c274b0a9-ce86-4ad1-82e1-509592932675-kube-api-access-b6c6p" (OuterVolumeSpecName: "kube-api-access-b6c6p") pod "c274b0a9-ce86-4ad1-82e1-509592932675" (UID: "c274b0a9-ce86-4ad1-82e1-509592932675"). InnerVolumeSpecName "kube-api-access-b6c6p". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 10:15:12 crc kubenswrapper[4870]: I1014 10:15:12.011479 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c274b0a9-ce86-4ad1-82e1-509592932675-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c274b0a9-ce86-4ad1-82e1-509592932675" (UID: "c274b0a9-ce86-4ad1-82e1-509592932675"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 10:15:12 crc kubenswrapper[4870]: I1014 10:15:12.096907 4870 generic.go:334] "Generic (PLEG): container finished" podID="c274b0a9-ce86-4ad1-82e1-509592932675" containerID="cf110bc597b7ae7d07aff32a30836bc98e772dea49b3ca0511e631d633872f22" exitCode=0 Oct 14 10:15:12 crc kubenswrapper[4870]: I1014 10:15:12.096969 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ghnkh" event={"ID":"c274b0a9-ce86-4ad1-82e1-509592932675","Type":"ContainerDied","Data":"cf110bc597b7ae7d07aff32a30836bc98e772dea49b3ca0511e631d633872f22"} Oct 14 10:15:12 crc kubenswrapper[4870]: I1014 10:15:12.097002 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ghnkh" event={"ID":"c274b0a9-ce86-4ad1-82e1-509592932675","Type":"ContainerDied","Data":"60b6a5a27e569f514345ba15da71fcb087a17b288682109c2a021f258ccc2bd5"} Oct 14 10:15:12 crc kubenswrapper[4870]: I1014 10:15:12.097021 4870 scope.go:117] "RemoveContainer" containerID="cf110bc597b7ae7d07aff32a30836bc98e772dea49b3ca0511e631d633872f22" Oct 14 10:15:12 crc kubenswrapper[4870]: I1014 10:15:12.097212 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ghnkh" Oct 14 10:15:12 crc kubenswrapper[4870]: I1014 10:15:12.102117 4870 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c274b0a9-ce86-4ad1-82e1-509592932675-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 10:15:12 crc kubenswrapper[4870]: I1014 10:15:12.102156 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b6c6p\" (UniqueName: \"kubernetes.io/projected/c274b0a9-ce86-4ad1-82e1-509592932675-kube-api-access-b6c6p\") on node \"crc\" DevicePath \"\"" Oct 14 10:15:12 crc kubenswrapper[4870]: I1014 10:15:12.102169 4870 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c274b0a9-ce86-4ad1-82e1-509592932675-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 10:15:12 crc kubenswrapper[4870]: I1014 10:15:12.126620 4870 scope.go:117] "RemoveContainer" containerID="b1eb9272a96698fcbc0f50215144913e566b61e2460f8b3034ce54949c2e4188" Oct 14 10:15:12 crc kubenswrapper[4870]: I1014 10:15:12.152352 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-ghnkh"] Oct 14 10:15:12 crc kubenswrapper[4870]: I1014 10:15:12.164958 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-ghnkh"] Oct 14 10:15:12 crc kubenswrapper[4870]: I1014 10:15:12.165895 4870 scope.go:117] "RemoveContainer" containerID="c4cd8760991ef75744cbca3febabfa3fd4340af9d0b2598198dd4f21efa2281e" Oct 14 10:15:12 crc kubenswrapper[4870]: I1014 10:15:12.235470 4870 scope.go:117] "RemoveContainer" containerID="cf110bc597b7ae7d07aff32a30836bc98e772dea49b3ca0511e631d633872f22" Oct 14 10:15:12 crc kubenswrapper[4870]: E1014 10:15:12.235907 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cf110bc597b7ae7d07aff32a30836bc98e772dea49b3ca0511e631d633872f22\": container with ID starting with cf110bc597b7ae7d07aff32a30836bc98e772dea49b3ca0511e631d633872f22 not found: ID does not exist" containerID="cf110bc597b7ae7d07aff32a30836bc98e772dea49b3ca0511e631d633872f22" Oct 14 10:15:12 crc kubenswrapper[4870]: I1014 10:15:12.235956 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cf110bc597b7ae7d07aff32a30836bc98e772dea49b3ca0511e631d633872f22"} err="failed to get container status \"cf110bc597b7ae7d07aff32a30836bc98e772dea49b3ca0511e631d633872f22\": rpc error: code = NotFound desc = could not find container \"cf110bc597b7ae7d07aff32a30836bc98e772dea49b3ca0511e631d633872f22\": container with ID starting with cf110bc597b7ae7d07aff32a30836bc98e772dea49b3ca0511e631d633872f22 not found: ID does not exist" Oct 14 10:15:12 crc kubenswrapper[4870]: I1014 10:15:12.235978 4870 scope.go:117] "RemoveContainer" containerID="b1eb9272a96698fcbc0f50215144913e566b61e2460f8b3034ce54949c2e4188" Oct 14 10:15:12 crc kubenswrapper[4870]: E1014 10:15:12.236172 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b1eb9272a96698fcbc0f50215144913e566b61e2460f8b3034ce54949c2e4188\": container with ID starting with b1eb9272a96698fcbc0f50215144913e566b61e2460f8b3034ce54949c2e4188 not found: ID does not exist" containerID="b1eb9272a96698fcbc0f50215144913e566b61e2460f8b3034ce54949c2e4188" Oct 14 10:15:12 crc kubenswrapper[4870]: I1014 10:15:12.236192 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b1eb9272a96698fcbc0f50215144913e566b61e2460f8b3034ce54949c2e4188"} err="failed to get container status \"b1eb9272a96698fcbc0f50215144913e566b61e2460f8b3034ce54949c2e4188\": rpc error: code = NotFound desc = could not find container \"b1eb9272a96698fcbc0f50215144913e566b61e2460f8b3034ce54949c2e4188\": container with ID starting with b1eb9272a96698fcbc0f50215144913e566b61e2460f8b3034ce54949c2e4188 not found: ID does not exist" Oct 14 10:15:12 crc kubenswrapper[4870]: I1014 10:15:12.236207 4870 scope.go:117] "RemoveContainer" containerID="c4cd8760991ef75744cbca3febabfa3fd4340af9d0b2598198dd4f21efa2281e" Oct 14 10:15:12 crc kubenswrapper[4870]: E1014 10:15:12.236447 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c4cd8760991ef75744cbca3febabfa3fd4340af9d0b2598198dd4f21efa2281e\": container with ID starting with c4cd8760991ef75744cbca3febabfa3fd4340af9d0b2598198dd4f21efa2281e not found: ID does not exist" containerID="c4cd8760991ef75744cbca3febabfa3fd4340af9d0b2598198dd4f21efa2281e" Oct 14 10:15:12 crc kubenswrapper[4870]: I1014 10:15:12.236468 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c4cd8760991ef75744cbca3febabfa3fd4340af9d0b2598198dd4f21efa2281e"} err="failed to get container status \"c4cd8760991ef75744cbca3febabfa3fd4340af9d0b2598198dd4f21efa2281e\": rpc error: code = NotFound desc = could not find container \"c4cd8760991ef75744cbca3febabfa3fd4340af9d0b2598198dd4f21efa2281e\": container with ID starting with c4cd8760991ef75744cbca3febabfa3fd4340af9d0b2598198dd4f21efa2281e not found: ID does not exist" Oct 14 10:15:13 crc kubenswrapper[4870]: I1014 10:15:13.055782 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c274b0a9-ce86-4ad1-82e1-509592932675" path="/var/lib/kubelet/pods/c274b0a9-ce86-4ad1-82e1-509592932675/volumes" Oct 14 10:15:36 crc kubenswrapper[4870]: I1014 10:15:36.044374 4870 scope.go:117] "RemoveContainer" containerID="be81d519402e34c4d49c058e6868c5a6cb81f4fe590402a93b44a4fe61a99a55" Oct 14 10:16:19 crc kubenswrapper[4870]: I1014 10:16:19.227207 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-gdw8c"] Oct 14 10:16:19 crc kubenswrapper[4870]: E1014 10:16:19.228152 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c274b0a9-ce86-4ad1-82e1-509592932675" containerName="registry-server" Oct 14 10:16:19 crc kubenswrapper[4870]: I1014 10:16:19.228165 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="c274b0a9-ce86-4ad1-82e1-509592932675" containerName="registry-server" Oct 14 10:16:19 crc kubenswrapper[4870]: E1014 10:16:19.228180 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8fe77a0f-ea5d-48f3-9206-f4a4813f0818" containerName="collect-profiles" Oct 14 10:16:19 crc kubenswrapper[4870]: I1014 10:16:19.228187 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="8fe77a0f-ea5d-48f3-9206-f4a4813f0818" containerName="collect-profiles" Oct 14 10:16:19 crc kubenswrapper[4870]: E1014 10:16:19.228210 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c274b0a9-ce86-4ad1-82e1-509592932675" containerName="extract-utilities" Oct 14 10:16:19 crc kubenswrapper[4870]: I1014 10:16:19.228217 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="c274b0a9-ce86-4ad1-82e1-509592932675" containerName="extract-utilities" Oct 14 10:16:19 crc kubenswrapper[4870]: E1014 10:16:19.228235 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c274b0a9-ce86-4ad1-82e1-509592932675" containerName="extract-content" Oct 14 10:16:19 crc kubenswrapper[4870]: I1014 10:16:19.228241 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="c274b0a9-ce86-4ad1-82e1-509592932675" containerName="extract-content" Oct 14 10:16:19 crc kubenswrapper[4870]: I1014 10:16:19.228624 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="c274b0a9-ce86-4ad1-82e1-509592932675" containerName="registry-server" Oct 14 10:16:19 crc kubenswrapper[4870]: I1014 10:16:19.228646 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="8fe77a0f-ea5d-48f3-9206-f4a4813f0818" containerName="collect-profiles" Oct 14 10:16:19 crc kubenswrapper[4870]: I1014 10:16:19.230284 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gdw8c" Oct 14 10:16:19 crc kubenswrapper[4870]: I1014 10:16:19.247257 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-gdw8c"] Oct 14 10:16:19 crc kubenswrapper[4870]: I1014 10:16:19.397739 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6csrg\" (UniqueName: \"kubernetes.io/projected/3f73653c-d771-4da5-a66d-aa5c974f2898-kube-api-access-6csrg\") pod \"redhat-operators-gdw8c\" (UID: \"3f73653c-d771-4da5-a66d-aa5c974f2898\") " pod="openshift-marketplace/redhat-operators-gdw8c" Oct 14 10:16:19 crc kubenswrapper[4870]: I1014 10:16:19.397832 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3f73653c-d771-4da5-a66d-aa5c974f2898-utilities\") pod \"redhat-operators-gdw8c\" (UID: \"3f73653c-d771-4da5-a66d-aa5c974f2898\") " pod="openshift-marketplace/redhat-operators-gdw8c" Oct 14 10:16:19 crc kubenswrapper[4870]: I1014 10:16:19.397857 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3f73653c-d771-4da5-a66d-aa5c974f2898-catalog-content\") pod \"redhat-operators-gdw8c\" (UID: \"3f73653c-d771-4da5-a66d-aa5c974f2898\") " pod="openshift-marketplace/redhat-operators-gdw8c" Oct 14 10:16:19 crc kubenswrapper[4870]: I1014 10:16:19.499834 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6csrg\" (UniqueName: \"kubernetes.io/projected/3f73653c-d771-4da5-a66d-aa5c974f2898-kube-api-access-6csrg\") pod \"redhat-operators-gdw8c\" (UID: \"3f73653c-d771-4da5-a66d-aa5c974f2898\") " pod="openshift-marketplace/redhat-operators-gdw8c" Oct 14 10:16:19 crc kubenswrapper[4870]: I1014 10:16:19.500185 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3f73653c-d771-4da5-a66d-aa5c974f2898-utilities\") pod \"redhat-operators-gdw8c\" (UID: \"3f73653c-d771-4da5-a66d-aa5c974f2898\") " pod="openshift-marketplace/redhat-operators-gdw8c" Oct 14 10:16:19 crc kubenswrapper[4870]: I1014 10:16:19.500213 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3f73653c-d771-4da5-a66d-aa5c974f2898-catalog-content\") pod \"redhat-operators-gdw8c\" (UID: \"3f73653c-d771-4da5-a66d-aa5c974f2898\") " pod="openshift-marketplace/redhat-operators-gdw8c" Oct 14 10:16:19 crc kubenswrapper[4870]: I1014 10:16:19.500717 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3f73653c-d771-4da5-a66d-aa5c974f2898-catalog-content\") pod \"redhat-operators-gdw8c\" (UID: \"3f73653c-d771-4da5-a66d-aa5c974f2898\") " pod="openshift-marketplace/redhat-operators-gdw8c" Oct 14 10:16:19 crc kubenswrapper[4870]: I1014 10:16:19.500866 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3f73653c-d771-4da5-a66d-aa5c974f2898-utilities\") pod \"redhat-operators-gdw8c\" (UID: \"3f73653c-d771-4da5-a66d-aa5c974f2898\") " pod="openshift-marketplace/redhat-operators-gdw8c" Oct 14 10:16:19 crc kubenswrapper[4870]: I1014 10:16:19.519782 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6csrg\" (UniqueName: \"kubernetes.io/projected/3f73653c-d771-4da5-a66d-aa5c974f2898-kube-api-access-6csrg\") pod \"redhat-operators-gdw8c\" (UID: \"3f73653c-d771-4da5-a66d-aa5c974f2898\") " pod="openshift-marketplace/redhat-operators-gdw8c" Oct 14 10:16:19 crc kubenswrapper[4870]: I1014 10:16:19.555073 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gdw8c" Oct 14 10:16:20 crc kubenswrapper[4870]: I1014 10:16:20.048307 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-gdw8c"] Oct 14 10:16:20 crc kubenswrapper[4870]: I1014 10:16:20.912735 4870 generic.go:334] "Generic (PLEG): container finished" podID="3f73653c-d771-4da5-a66d-aa5c974f2898" containerID="0524087adc32d4a9fc953ea15c830c2bc2cd2ff1745659d6715b9eb107771433" exitCode=0 Oct 14 10:16:20 crc kubenswrapper[4870]: I1014 10:16:20.913183 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gdw8c" event={"ID":"3f73653c-d771-4da5-a66d-aa5c974f2898","Type":"ContainerDied","Data":"0524087adc32d4a9fc953ea15c830c2bc2cd2ff1745659d6715b9eb107771433"} Oct 14 10:16:20 crc kubenswrapper[4870]: I1014 10:16:20.913241 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gdw8c" event={"ID":"3f73653c-d771-4da5-a66d-aa5c974f2898","Type":"ContainerStarted","Data":"4ea5ed46c566e14ff8c8718e29745c082aad534b9b782badcb3ce71de77ec355"} Oct 14 10:16:23 crc kubenswrapper[4870]: I1014 10:16:23.951077 4870 patch_prober.go:28] interesting pod/machine-config-daemon-7tvc7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 10:16:23 crc kubenswrapper[4870]: I1014 10:16:23.951737 4870 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 10:16:32 crc kubenswrapper[4870]: I1014 10:16:32.098735 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gdw8c" event={"ID":"3f73653c-d771-4da5-a66d-aa5c974f2898","Type":"ContainerStarted","Data":"5bcca650bb353416c01c77b039f8213bf050b7edeecb503e3a169d53d60c0621"} Oct 14 10:16:33 crc kubenswrapper[4870]: I1014 10:16:33.115186 4870 generic.go:334] "Generic (PLEG): container finished" podID="3f73653c-d771-4da5-a66d-aa5c974f2898" containerID="5bcca650bb353416c01c77b039f8213bf050b7edeecb503e3a169d53d60c0621" exitCode=0 Oct 14 10:16:33 crc kubenswrapper[4870]: I1014 10:16:33.115313 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gdw8c" event={"ID":"3f73653c-d771-4da5-a66d-aa5c974f2898","Type":"ContainerDied","Data":"5bcca650bb353416c01c77b039f8213bf050b7edeecb503e3a169d53d60c0621"} Oct 14 10:16:35 crc kubenswrapper[4870]: I1014 10:16:35.140540 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gdw8c" event={"ID":"3f73653c-d771-4da5-a66d-aa5c974f2898","Type":"ContainerStarted","Data":"480bc7fbc8a159eb241723a54108dbab5868db60f27ab49f4298d966e10c7a9c"} Oct 14 10:16:35 crc kubenswrapper[4870]: I1014 10:16:35.167032 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-gdw8c" podStartSLOduration=3.334448459 podStartE2EDuration="16.167013391s" podCreationTimestamp="2025-10-14 10:16:19 +0000 UTC" firstStartedPulling="2025-10-14 10:16:20.915471207 +0000 UTC m=+11716.612831578" lastFinishedPulling="2025-10-14 10:16:33.748036139 +0000 UTC m=+11729.445396510" observedRunningTime="2025-10-14 10:16:35.158616145 +0000 UTC m=+11730.855976506" watchObservedRunningTime="2025-10-14 10:16:35.167013391 +0000 UTC m=+11730.864373762" Oct 14 10:16:39 crc kubenswrapper[4870]: I1014 10:16:39.555730 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-gdw8c" Oct 14 10:16:39 crc kubenswrapper[4870]: I1014 10:16:39.556253 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-gdw8c" Oct 14 10:16:40 crc kubenswrapper[4870]: I1014 10:16:40.609797 4870 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-gdw8c" podUID="3f73653c-d771-4da5-a66d-aa5c974f2898" containerName="registry-server" probeResult="failure" output=< Oct 14 10:16:40 crc kubenswrapper[4870]: timeout: failed to connect service ":50051" within 1s Oct 14 10:16:40 crc kubenswrapper[4870]: > Oct 14 10:16:49 crc kubenswrapper[4870]: I1014 10:16:49.630463 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-gdw8c" Oct 14 10:16:49 crc kubenswrapper[4870]: I1014 10:16:49.704430 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-gdw8c" Oct 14 10:16:50 crc kubenswrapper[4870]: I1014 10:16:50.270139 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-gdw8c"] Oct 14 10:16:50 crc kubenswrapper[4870]: I1014 10:16:50.429696 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-xwpt9"] Oct 14 10:16:50 crc kubenswrapper[4870]: I1014 10:16:50.429947 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-xwpt9" podUID="aeca7d4f-c3c4-404c-9bdf-a1ed0951ada9" containerName="registry-server" containerID="cri-o://f203f942444d205a40086ae2434be7198113f9d3ff89f9893f0a349715fba6b1" gracePeriod=2 Oct 14 10:16:51 crc kubenswrapper[4870]: I1014 10:16:51.347269 4870 generic.go:334] "Generic (PLEG): container finished" podID="aeca7d4f-c3c4-404c-9bdf-a1ed0951ada9" containerID="f203f942444d205a40086ae2434be7198113f9d3ff89f9893f0a349715fba6b1" exitCode=0 Oct 14 10:16:51 crc kubenswrapper[4870]: I1014 10:16:51.347538 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xwpt9" event={"ID":"aeca7d4f-c3c4-404c-9bdf-a1ed0951ada9","Type":"ContainerDied","Data":"f203f942444d205a40086ae2434be7198113f9d3ff89f9893f0a349715fba6b1"} Oct 14 10:16:51 crc kubenswrapper[4870]: I1014 10:16:51.349620 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xwpt9" event={"ID":"aeca7d4f-c3c4-404c-9bdf-a1ed0951ada9","Type":"ContainerDied","Data":"58b0cd7c623d8eeebb8aba3cfe7b5ffddd5adaa6cea4645a51617fd6556b774c"} Oct 14 10:16:51 crc kubenswrapper[4870]: I1014 10:16:51.349640 4870 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="58b0cd7c623d8eeebb8aba3cfe7b5ffddd5adaa6cea4645a51617fd6556b774c" Oct 14 10:16:51 crc kubenswrapper[4870]: I1014 10:16:51.425174 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xwpt9" Oct 14 10:16:51 crc kubenswrapper[4870]: I1014 10:16:51.547357 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aeca7d4f-c3c4-404c-9bdf-a1ed0951ada9-utilities\") pod \"aeca7d4f-c3c4-404c-9bdf-a1ed0951ada9\" (UID: \"aeca7d4f-c3c4-404c-9bdf-a1ed0951ada9\") " Oct 14 10:16:51 crc kubenswrapper[4870]: I1014 10:16:51.548293 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aeca7d4f-c3c4-404c-9bdf-a1ed0951ada9-catalog-content\") pod \"aeca7d4f-c3c4-404c-9bdf-a1ed0951ada9\" (UID: \"aeca7d4f-c3c4-404c-9bdf-a1ed0951ada9\") " Oct 14 10:16:51 crc kubenswrapper[4870]: I1014 10:16:51.548472 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ljbtr\" (UniqueName: \"kubernetes.io/projected/aeca7d4f-c3c4-404c-9bdf-a1ed0951ada9-kube-api-access-ljbtr\") pod \"aeca7d4f-c3c4-404c-9bdf-a1ed0951ada9\" (UID: \"aeca7d4f-c3c4-404c-9bdf-a1ed0951ada9\") " Oct 14 10:16:51 crc kubenswrapper[4870]: I1014 10:16:51.550722 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aeca7d4f-c3c4-404c-9bdf-a1ed0951ada9-utilities" (OuterVolumeSpecName: "utilities") pod "aeca7d4f-c3c4-404c-9bdf-a1ed0951ada9" (UID: "aeca7d4f-c3c4-404c-9bdf-a1ed0951ada9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 10:16:51 crc kubenswrapper[4870]: I1014 10:16:51.560187 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aeca7d4f-c3c4-404c-9bdf-a1ed0951ada9-kube-api-access-ljbtr" (OuterVolumeSpecName: "kube-api-access-ljbtr") pod "aeca7d4f-c3c4-404c-9bdf-a1ed0951ada9" (UID: "aeca7d4f-c3c4-404c-9bdf-a1ed0951ada9"). InnerVolumeSpecName "kube-api-access-ljbtr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 10:16:51 crc kubenswrapper[4870]: I1014 10:16:51.650778 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ljbtr\" (UniqueName: \"kubernetes.io/projected/aeca7d4f-c3c4-404c-9bdf-a1ed0951ada9-kube-api-access-ljbtr\") on node \"crc\" DevicePath \"\"" Oct 14 10:16:51 crc kubenswrapper[4870]: I1014 10:16:51.650814 4870 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aeca7d4f-c3c4-404c-9bdf-a1ed0951ada9-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 10:16:51 crc kubenswrapper[4870]: I1014 10:16:51.671692 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aeca7d4f-c3c4-404c-9bdf-a1ed0951ada9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "aeca7d4f-c3c4-404c-9bdf-a1ed0951ada9" (UID: "aeca7d4f-c3c4-404c-9bdf-a1ed0951ada9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 10:16:51 crc kubenswrapper[4870]: I1014 10:16:51.752642 4870 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aeca7d4f-c3c4-404c-9bdf-a1ed0951ada9-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 10:16:52 crc kubenswrapper[4870]: I1014 10:16:52.358249 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xwpt9" Oct 14 10:16:52 crc kubenswrapper[4870]: I1014 10:16:52.405213 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-xwpt9"] Oct 14 10:16:52 crc kubenswrapper[4870]: I1014 10:16:52.425045 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-xwpt9"] Oct 14 10:16:53 crc kubenswrapper[4870]: I1014 10:16:53.050687 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aeca7d4f-c3c4-404c-9bdf-a1ed0951ada9" path="/var/lib/kubelet/pods/aeca7d4f-c3c4-404c-9bdf-a1ed0951ada9/volumes" Oct 14 10:16:53 crc kubenswrapper[4870]: I1014 10:16:53.950810 4870 patch_prober.go:28] interesting pod/machine-config-daemon-7tvc7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 10:16:53 crc kubenswrapper[4870]: I1014 10:16:53.951187 4870 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 10:17:23 crc kubenswrapper[4870]: I1014 10:17:23.951424 4870 patch_prober.go:28] interesting pod/machine-config-daemon-7tvc7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 10:17:23 crc kubenswrapper[4870]: I1014 10:17:23.952628 4870 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 10:17:23 crc kubenswrapper[4870]: I1014 10:17:23.952723 4870 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" Oct 14 10:17:23 crc kubenswrapper[4870]: I1014 10:17:23.954311 4870 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"7468e2739d25e53085cbe91cf249f36e5760f983ae0240cf58bcc9bcb1b91355"} pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 14 10:17:23 crc kubenswrapper[4870]: I1014 10:17:23.954408 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerName="machine-config-daemon" containerID="cri-o://7468e2739d25e53085cbe91cf249f36e5760f983ae0240cf58bcc9bcb1b91355" gracePeriod=600 Oct 14 10:17:24 crc kubenswrapper[4870]: E1014 10:17:24.077596 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 10:17:24 crc kubenswrapper[4870]: I1014 10:17:24.786248 4870 generic.go:334] "Generic (PLEG): container finished" podID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerID="7468e2739d25e53085cbe91cf249f36e5760f983ae0240cf58bcc9bcb1b91355" exitCode=0 Oct 14 10:17:24 crc kubenswrapper[4870]: I1014 10:17:24.786588 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" event={"ID":"874a5e45-dffd-4d17-b609-b7d3ed2eab07","Type":"ContainerDied","Data":"7468e2739d25e53085cbe91cf249f36e5760f983ae0240cf58bcc9bcb1b91355"} Oct 14 10:17:24 crc kubenswrapper[4870]: I1014 10:17:24.786686 4870 scope.go:117] "RemoveContainer" containerID="2a49fc200e8e34001c67b2fb36a1c2bb51c1940419a6526be0496325545372e9" Oct 14 10:17:24 crc kubenswrapper[4870]: I1014 10:17:24.787792 4870 scope.go:117] "RemoveContainer" containerID="7468e2739d25e53085cbe91cf249f36e5760f983ae0240cf58bcc9bcb1b91355" Oct 14 10:17:24 crc kubenswrapper[4870]: E1014 10:17:24.788891 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 10:17:36 crc kubenswrapper[4870]: I1014 10:17:36.228590 4870 scope.go:117] "RemoveContainer" containerID="899754855bb11ba4691fc9e2d644607368e1d1d256d678f6715ec75365cb503f" Oct 14 10:17:36 crc kubenswrapper[4870]: I1014 10:17:36.269814 4870 scope.go:117] "RemoveContainer" containerID="f203f942444d205a40086ae2434be7198113f9d3ff89f9893f0a349715fba6b1" Oct 14 10:17:36 crc kubenswrapper[4870]: I1014 10:17:36.316027 4870 scope.go:117] "RemoveContainer" containerID="df1e991ca91d6a83ec52845be3ea04155de224c50912bb7dcbb13d8d2c334147" Oct 14 10:17:39 crc kubenswrapper[4870]: I1014 10:17:39.039414 4870 scope.go:117] "RemoveContainer" containerID="7468e2739d25e53085cbe91cf249f36e5760f983ae0240cf58bcc9bcb1b91355" Oct 14 10:17:39 crc kubenswrapper[4870]: E1014 10:17:39.040411 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 10:17:51 crc kubenswrapper[4870]: I1014 10:17:51.035984 4870 scope.go:117] "RemoveContainer" containerID="7468e2739d25e53085cbe91cf249f36e5760f983ae0240cf58bcc9bcb1b91355" Oct 14 10:17:51 crc kubenswrapper[4870]: E1014 10:17:51.036799 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 10:18:02 crc kubenswrapper[4870]: I1014 10:18:02.035942 4870 scope.go:117] "RemoveContainer" containerID="7468e2739d25e53085cbe91cf249f36e5760f983ae0240cf58bcc9bcb1b91355" Oct 14 10:18:02 crc kubenswrapper[4870]: E1014 10:18:02.036841 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 10:18:17 crc kubenswrapper[4870]: I1014 10:18:17.035523 4870 scope.go:117] "RemoveContainer" containerID="7468e2739d25e53085cbe91cf249f36e5760f983ae0240cf58bcc9bcb1b91355" Oct 14 10:18:17 crc kubenswrapper[4870]: E1014 10:18:17.036886 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 10:18:30 crc kubenswrapper[4870]: I1014 10:18:30.034317 4870 scope.go:117] "RemoveContainer" containerID="7468e2739d25e53085cbe91cf249f36e5760f983ae0240cf58bcc9bcb1b91355" Oct 14 10:18:30 crc kubenswrapper[4870]: E1014 10:18:30.035054 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 10:18:41 crc kubenswrapper[4870]: I1014 10:18:41.034430 4870 scope.go:117] "RemoveContainer" containerID="7468e2739d25e53085cbe91cf249f36e5760f983ae0240cf58bcc9bcb1b91355" Oct 14 10:18:41 crc kubenswrapper[4870]: E1014 10:18:41.035333 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 10:18:56 crc kubenswrapper[4870]: I1014 10:18:56.034010 4870 scope.go:117] "RemoveContainer" containerID="7468e2739d25e53085cbe91cf249f36e5760f983ae0240cf58bcc9bcb1b91355" Oct 14 10:18:56 crc kubenswrapper[4870]: E1014 10:18:56.034988 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 10:19:11 crc kubenswrapper[4870]: I1014 10:19:11.034309 4870 scope.go:117] "RemoveContainer" containerID="7468e2739d25e53085cbe91cf249f36e5760f983ae0240cf58bcc9bcb1b91355" Oct 14 10:19:11 crc kubenswrapper[4870]: E1014 10:19:11.035058 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 10:19:26 crc kubenswrapper[4870]: I1014 10:19:26.034331 4870 scope.go:117] "RemoveContainer" containerID="7468e2739d25e53085cbe91cf249f36e5760f983ae0240cf58bcc9bcb1b91355" Oct 14 10:19:26 crc kubenswrapper[4870]: E1014 10:19:26.035359 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 10:19:41 crc kubenswrapper[4870]: I1014 10:19:41.034933 4870 scope.go:117] "RemoveContainer" containerID="7468e2739d25e53085cbe91cf249f36e5760f983ae0240cf58bcc9bcb1b91355" Oct 14 10:19:41 crc kubenswrapper[4870]: E1014 10:19:41.035930 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 10:19:56 crc kubenswrapper[4870]: I1014 10:19:56.034555 4870 scope.go:117] "RemoveContainer" containerID="7468e2739d25e53085cbe91cf249f36e5760f983ae0240cf58bcc9bcb1b91355" Oct 14 10:19:56 crc kubenswrapper[4870]: E1014 10:19:56.035366 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 10:20:09 crc kubenswrapper[4870]: I1014 10:20:09.034852 4870 scope.go:117] "RemoveContainer" containerID="7468e2739d25e53085cbe91cf249f36e5760f983ae0240cf58bcc9bcb1b91355" Oct 14 10:20:09 crc kubenswrapper[4870]: E1014 10:20:09.036032 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 10:20:20 crc kubenswrapper[4870]: I1014 10:20:20.035032 4870 scope.go:117] "RemoveContainer" containerID="7468e2739d25e53085cbe91cf249f36e5760f983ae0240cf58bcc9bcb1b91355" Oct 14 10:20:20 crc kubenswrapper[4870]: E1014 10:20:20.036135 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 10:20:34 crc kubenswrapper[4870]: I1014 10:20:34.035264 4870 scope.go:117] "RemoveContainer" containerID="7468e2739d25e53085cbe91cf249f36e5760f983ae0240cf58bcc9bcb1b91355" Oct 14 10:20:34 crc kubenswrapper[4870]: E1014 10:20:34.036801 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 10:20:45 crc kubenswrapper[4870]: I1014 10:20:45.042392 4870 scope.go:117] "RemoveContainer" containerID="7468e2739d25e53085cbe91cf249f36e5760f983ae0240cf58bcc9bcb1b91355" Oct 14 10:20:45 crc kubenswrapper[4870]: E1014 10:20:45.043527 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 10:20:57 crc kubenswrapper[4870]: I1014 10:20:57.035781 4870 scope.go:117] "RemoveContainer" containerID="7468e2739d25e53085cbe91cf249f36e5760f983ae0240cf58bcc9bcb1b91355" Oct 14 10:20:57 crc kubenswrapper[4870]: E1014 10:20:57.036997 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 10:21:10 crc kubenswrapper[4870]: I1014 10:21:10.036799 4870 scope.go:117] "RemoveContainer" containerID="7468e2739d25e53085cbe91cf249f36e5760f983ae0240cf58bcc9bcb1b91355" Oct 14 10:21:10 crc kubenswrapper[4870]: E1014 10:21:10.038182 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 10:21:26 crc kubenswrapper[4870]: I1014 10:21:26.034516 4870 scope.go:117] "RemoveContainer" containerID="7468e2739d25e53085cbe91cf249f36e5760f983ae0240cf58bcc9bcb1b91355" Oct 14 10:21:26 crc kubenswrapper[4870]: E1014 10:21:26.036105 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 10:21:37 crc kubenswrapper[4870]: I1014 10:21:37.034873 4870 scope.go:117] "RemoveContainer" containerID="7468e2739d25e53085cbe91cf249f36e5760f983ae0240cf58bcc9bcb1b91355" Oct 14 10:21:37 crc kubenswrapper[4870]: E1014 10:21:37.040150 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 10:21:50 crc kubenswrapper[4870]: I1014 10:21:50.034254 4870 scope.go:117] "RemoveContainer" containerID="7468e2739d25e53085cbe91cf249f36e5760f983ae0240cf58bcc9bcb1b91355" Oct 14 10:21:50 crc kubenswrapper[4870]: E1014 10:21:50.035061 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 10:22:03 crc kubenswrapper[4870]: I1014 10:22:03.035917 4870 scope.go:117] "RemoveContainer" containerID="7468e2739d25e53085cbe91cf249f36e5760f983ae0240cf58bcc9bcb1b91355" Oct 14 10:22:03 crc kubenswrapper[4870]: E1014 10:22:03.042127 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 10:22:16 crc kubenswrapper[4870]: I1014 10:22:16.034275 4870 scope.go:117] "RemoveContainer" containerID="7468e2739d25e53085cbe91cf249f36e5760f983ae0240cf58bcc9bcb1b91355" Oct 14 10:22:16 crc kubenswrapper[4870]: E1014 10:22:16.035047 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 10:22:27 crc kubenswrapper[4870]: I1014 10:22:27.035383 4870 scope.go:117] "RemoveContainer" containerID="7468e2739d25e53085cbe91cf249f36e5760f983ae0240cf58bcc9bcb1b91355" Oct 14 10:22:27 crc kubenswrapper[4870]: I1014 10:22:27.376547 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" event={"ID":"874a5e45-dffd-4d17-b609-b7d3ed2eab07","Type":"ContainerStarted","Data":"4e0c842e7d29f428c954b866c1f3929be48b4bb5118abceba1fbf69beed50855"} Oct 14 10:23:05 crc kubenswrapper[4870]: I1014 10:23:05.986775 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-dgp7r"] Oct 14 10:23:05 crc kubenswrapper[4870]: E1014 10:23:05.988189 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aeca7d4f-c3c4-404c-9bdf-a1ed0951ada9" containerName="extract-utilities" Oct 14 10:23:05 crc kubenswrapper[4870]: I1014 10:23:05.988203 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="aeca7d4f-c3c4-404c-9bdf-a1ed0951ada9" containerName="extract-utilities" Oct 14 10:23:05 crc kubenswrapper[4870]: E1014 10:23:05.988237 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aeca7d4f-c3c4-404c-9bdf-a1ed0951ada9" containerName="registry-server" Oct 14 10:23:05 crc kubenswrapper[4870]: I1014 10:23:05.988243 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="aeca7d4f-c3c4-404c-9bdf-a1ed0951ada9" containerName="registry-server" Oct 14 10:23:05 crc kubenswrapper[4870]: E1014 10:23:05.988262 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aeca7d4f-c3c4-404c-9bdf-a1ed0951ada9" containerName="extract-content" Oct 14 10:23:05 crc kubenswrapper[4870]: I1014 10:23:05.988268 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="aeca7d4f-c3c4-404c-9bdf-a1ed0951ada9" containerName="extract-content" Oct 14 10:23:05 crc kubenswrapper[4870]: I1014 10:23:05.988496 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="aeca7d4f-c3c4-404c-9bdf-a1ed0951ada9" containerName="registry-server" Oct 14 10:23:05 crc kubenswrapper[4870]: I1014 10:23:05.990672 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dgp7r" Oct 14 10:23:06 crc kubenswrapper[4870]: I1014 10:23:06.026005 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-dgp7r"] Oct 14 10:23:06 crc kubenswrapper[4870]: I1014 10:23:06.160987 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t64dc\" (UniqueName: \"kubernetes.io/projected/b66523d6-2bc0-430c-ad75-e67ee9eee39d-kube-api-access-t64dc\") pod \"certified-operators-dgp7r\" (UID: \"b66523d6-2bc0-430c-ad75-e67ee9eee39d\") " pod="openshift-marketplace/certified-operators-dgp7r" Oct 14 10:23:06 crc kubenswrapper[4870]: I1014 10:23:06.161053 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b66523d6-2bc0-430c-ad75-e67ee9eee39d-catalog-content\") pod \"certified-operators-dgp7r\" (UID: \"b66523d6-2bc0-430c-ad75-e67ee9eee39d\") " pod="openshift-marketplace/certified-operators-dgp7r" Oct 14 10:23:06 crc kubenswrapper[4870]: I1014 10:23:06.161144 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b66523d6-2bc0-430c-ad75-e67ee9eee39d-utilities\") pod \"certified-operators-dgp7r\" (UID: \"b66523d6-2bc0-430c-ad75-e67ee9eee39d\") " pod="openshift-marketplace/certified-operators-dgp7r" Oct 14 10:23:06 crc kubenswrapper[4870]: I1014 10:23:06.262381 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t64dc\" (UniqueName: \"kubernetes.io/projected/b66523d6-2bc0-430c-ad75-e67ee9eee39d-kube-api-access-t64dc\") pod \"certified-operators-dgp7r\" (UID: \"b66523d6-2bc0-430c-ad75-e67ee9eee39d\") " pod="openshift-marketplace/certified-operators-dgp7r" Oct 14 10:23:06 crc kubenswrapper[4870]: I1014 10:23:06.262469 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b66523d6-2bc0-430c-ad75-e67ee9eee39d-catalog-content\") pod \"certified-operators-dgp7r\" (UID: \"b66523d6-2bc0-430c-ad75-e67ee9eee39d\") " pod="openshift-marketplace/certified-operators-dgp7r" Oct 14 10:23:06 crc kubenswrapper[4870]: I1014 10:23:06.262539 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b66523d6-2bc0-430c-ad75-e67ee9eee39d-utilities\") pod \"certified-operators-dgp7r\" (UID: \"b66523d6-2bc0-430c-ad75-e67ee9eee39d\") " pod="openshift-marketplace/certified-operators-dgp7r" Oct 14 10:23:06 crc kubenswrapper[4870]: I1014 10:23:06.263060 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b66523d6-2bc0-430c-ad75-e67ee9eee39d-utilities\") pod \"certified-operators-dgp7r\" (UID: \"b66523d6-2bc0-430c-ad75-e67ee9eee39d\") " pod="openshift-marketplace/certified-operators-dgp7r" Oct 14 10:23:06 crc kubenswrapper[4870]: I1014 10:23:06.263221 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b66523d6-2bc0-430c-ad75-e67ee9eee39d-catalog-content\") pod \"certified-operators-dgp7r\" (UID: \"b66523d6-2bc0-430c-ad75-e67ee9eee39d\") " pod="openshift-marketplace/certified-operators-dgp7r" Oct 14 10:23:06 crc kubenswrapper[4870]: I1014 10:23:06.299647 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t64dc\" (UniqueName: \"kubernetes.io/projected/b66523d6-2bc0-430c-ad75-e67ee9eee39d-kube-api-access-t64dc\") pod \"certified-operators-dgp7r\" (UID: \"b66523d6-2bc0-430c-ad75-e67ee9eee39d\") " pod="openshift-marketplace/certified-operators-dgp7r" Oct 14 10:23:06 crc kubenswrapper[4870]: I1014 10:23:06.313905 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dgp7r" Oct 14 10:23:06 crc kubenswrapper[4870]: I1014 10:23:06.855077 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-dgp7r"] Oct 14 10:23:07 crc kubenswrapper[4870]: I1014 10:23:07.857284 4870 generic.go:334] "Generic (PLEG): container finished" podID="b66523d6-2bc0-430c-ad75-e67ee9eee39d" containerID="36a8021d709c8177703591c13607d42f823c796b41ea6ef076fc5bf4e61b372a" exitCode=0 Oct 14 10:23:07 crc kubenswrapper[4870]: I1014 10:23:07.857594 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dgp7r" event={"ID":"b66523d6-2bc0-430c-ad75-e67ee9eee39d","Type":"ContainerDied","Data":"36a8021d709c8177703591c13607d42f823c796b41ea6ef076fc5bf4e61b372a"} Oct 14 10:23:07 crc kubenswrapper[4870]: I1014 10:23:07.857738 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dgp7r" event={"ID":"b66523d6-2bc0-430c-ad75-e67ee9eee39d","Type":"ContainerStarted","Data":"e3031f2dfc1896f5983f7bb13c3162c5b89618c050232bc16616147f52e85a12"} Oct 14 10:23:07 crc kubenswrapper[4870]: I1014 10:23:07.862176 4870 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 14 10:23:09 crc kubenswrapper[4870]: I1014 10:23:09.896699 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dgp7r" event={"ID":"b66523d6-2bc0-430c-ad75-e67ee9eee39d","Type":"ContainerStarted","Data":"e977448a9c15e70548e23973e2c5ad111f0a9bc6aa61763a298974a049395e33"} Oct 14 10:23:10 crc kubenswrapper[4870]: I1014 10:23:10.913605 4870 generic.go:334] "Generic (PLEG): container finished" podID="b66523d6-2bc0-430c-ad75-e67ee9eee39d" containerID="e977448a9c15e70548e23973e2c5ad111f0a9bc6aa61763a298974a049395e33" exitCode=0 Oct 14 10:23:10 crc kubenswrapper[4870]: I1014 10:23:10.913668 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dgp7r" event={"ID":"b66523d6-2bc0-430c-ad75-e67ee9eee39d","Type":"ContainerDied","Data":"e977448a9c15e70548e23973e2c5ad111f0a9bc6aa61763a298974a049395e33"} Oct 14 10:23:11 crc kubenswrapper[4870]: I1014 10:23:11.945885 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dgp7r" event={"ID":"b66523d6-2bc0-430c-ad75-e67ee9eee39d","Type":"ContainerStarted","Data":"7b7d59e54b23ba650fc959a79a06d1f024ffdbd1ba2bc6ccbda40873310090ff"} Oct 14 10:23:11 crc kubenswrapper[4870]: I1014 10:23:11.970424 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-dgp7r" podStartSLOduration=3.245679553 podStartE2EDuration="6.970404785s" podCreationTimestamp="2025-10-14 10:23:05 +0000 UTC" firstStartedPulling="2025-10-14 10:23:07.861661063 +0000 UTC m=+12123.559021464" lastFinishedPulling="2025-10-14 10:23:11.586386285 +0000 UTC m=+12127.283746696" observedRunningTime="2025-10-14 10:23:11.967869333 +0000 UTC m=+12127.665229704" watchObservedRunningTime="2025-10-14 10:23:11.970404785 +0000 UTC m=+12127.667765156" Oct 14 10:23:16 crc kubenswrapper[4870]: I1014 10:23:16.314584 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-dgp7r" Oct 14 10:23:16 crc kubenswrapper[4870]: I1014 10:23:16.314971 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-dgp7r" Oct 14 10:23:16 crc kubenswrapper[4870]: I1014 10:23:16.409904 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-dgp7r" Oct 14 10:23:17 crc kubenswrapper[4870]: I1014 10:23:17.085002 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-dgp7r" Oct 14 10:23:17 crc kubenswrapper[4870]: I1014 10:23:17.138555 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-dgp7r"] Oct 14 10:23:19 crc kubenswrapper[4870]: I1014 10:23:19.091246 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-dgp7r" podUID="b66523d6-2bc0-430c-ad75-e67ee9eee39d" containerName="registry-server" containerID="cri-o://7b7d59e54b23ba650fc959a79a06d1f024ffdbd1ba2bc6ccbda40873310090ff" gracePeriod=2 Oct 14 10:23:19 crc kubenswrapper[4870]: I1014 10:23:19.812779 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dgp7r" Oct 14 10:23:19 crc kubenswrapper[4870]: I1014 10:23:19.908099 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b66523d6-2bc0-430c-ad75-e67ee9eee39d-utilities\") pod \"b66523d6-2bc0-430c-ad75-e67ee9eee39d\" (UID: \"b66523d6-2bc0-430c-ad75-e67ee9eee39d\") " Oct 14 10:23:19 crc kubenswrapper[4870]: I1014 10:23:19.908188 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b66523d6-2bc0-430c-ad75-e67ee9eee39d-catalog-content\") pod \"b66523d6-2bc0-430c-ad75-e67ee9eee39d\" (UID: \"b66523d6-2bc0-430c-ad75-e67ee9eee39d\") " Oct 14 10:23:19 crc kubenswrapper[4870]: I1014 10:23:19.908274 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t64dc\" (UniqueName: \"kubernetes.io/projected/b66523d6-2bc0-430c-ad75-e67ee9eee39d-kube-api-access-t64dc\") pod \"b66523d6-2bc0-430c-ad75-e67ee9eee39d\" (UID: \"b66523d6-2bc0-430c-ad75-e67ee9eee39d\") " Oct 14 10:23:19 crc kubenswrapper[4870]: I1014 10:23:19.910763 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b66523d6-2bc0-430c-ad75-e67ee9eee39d-utilities" (OuterVolumeSpecName: "utilities") pod "b66523d6-2bc0-430c-ad75-e67ee9eee39d" (UID: "b66523d6-2bc0-430c-ad75-e67ee9eee39d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 10:23:19 crc kubenswrapper[4870]: I1014 10:23:19.915192 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b66523d6-2bc0-430c-ad75-e67ee9eee39d-kube-api-access-t64dc" (OuterVolumeSpecName: "kube-api-access-t64dc") pod "b66523d6-2bc0-430c-ad75-e67ee9eee39d" (UID: "b66523d6-2bc0-430c-ad75-e67ee9eee39d"). InnerVolumeSpecName "kube-api-access-t64dc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 10:23:19 crc kubenswrapper[4870]: I1014 10:23:19.969163 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b66523d6-2bc0-430c-ad75-e67ee9eee39d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b66523d6-2bc0-430c-ad75-e67ee9eee39d" (UID: "b66523d6-2bc0-430c-ad75-e67ee9eee39d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 10:23:20 crc kubenswrapper[4870]: I1014 10:23:20.011517 4870 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b66523d6-2bc0-430c-ad75-e67ee9eee39d-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 10:23:20 crc kubenswrapper[4870]: I1014 10:23:20.011563 4870 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b66523d6-2bc0-430c-ad75-e67ee9eee39d-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 10:23:20 crc kubenswrapper[4870]: I1014 10:23:20.011578 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t64dc\" (UniqueName: \"kubernetes.io/projected/b66523d6-2bc0-430c-ad75-e67ee9eee39d-kube-api-access-t64dc\") on node \"crc\" DevicePath \"\"" Oct 14 10:23:20 crc kubenswrapper[4870]: I1014 10:23:20.104436 4870 generic.go:334] "Generic (PLEG): container finished" podID="b66523d6-2bc0-430c-ad75-e67ee9eee39d" containerID="7b7d59e54b23ba650fc959a79a06d1f024ffdbd1ba2bc6ccbda40873310090ff" exitCode=0 Oct 14 10:23:20 crc kubenswrapper[4870]: I1014 10:23:20.104868 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dgp7r" event={"ID":"b66523d6-2bc0-430c-ad75-e67ee9eee39d","Type":"ContainerDied","Data":"7b7d59e54b23ba650fc959a79a06d1f024ffdbd1ba2bc6ccbda40873310090ff"} Oct 14 10:23:20 crc kubenswrapper[4870]: I1014 10:23:20.104911 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dgp7r" event={"ID":"b66523d6-2bc0-430c-ad75-e67ee9eee39d","Type":"ContainerDied","Data":"e3031f2dfc1896f5983f7bb13c3162c5b89618c050232bc16616147f52e85a12"} Oct 14 10:23:20 crc kubenswrapper[4870]: I1014 10:23:20.104941 4870 scope.go:117] "RemoveContainer" containerID="7b7d59e54b23ba650fc959a79a06d1f024ffdbd1ba2bc6ccbda40873310090ff" Oct 14 10:23:20 crc kubenswrapper[4870]: I1014 10:23:20.104945 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dgp7r" Oct 14 10:23:20 crc kubenswrapper[4870]: I1014 10:23:20.130862 4870 scope.go:117] "RemoveContainer" containerID="e977448a9c15e70548e23973e2c5ad111f0a9bc6aa61763a298974a049395e33" Oct 14 10:23:20 crc kubenswrapper[4870]: I1014 10:23:20.154360 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-dgp7r"] Oct 14 10:23:20 crc kubenswrapper[4870]: I1014 10:23:20.172938 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-dgp7r"] Oct 14 10:23:20 crc kubenswrapper[4870]: I1014 10:23:20.173629 4870 scope.go:117] "RemoveContainer" containerID="36a8021d709c8177703591c13607d42f823c796b41ea6ef076fc5bf4e61b372a" Oct 14 10:23:20 crc kubenswrapper[4870]: I1014 10:23:20.221887 4870 scope.go:117] "RemoveContainer" containerID="7b7d59e54b23ba650fc959a79a06d1f024ffdbd1ba2bc6ccbda40873310090ff" Oct 14 10:23:20 crc kubenswrapper[4870]: E1014 10:23:20.222381 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7b7d59e54b23ba650fc959a79a06d1f024ffdbd1ba2bc6ccbda40873310090ff\": container with ID starting with 7b7d59e54b23ba650fc959a79a06d1f024ffdbd1ba2bc6ccbda40873310090ff not found: ID does not exist" containerID="7b7d59e54b23ba650fc959a79a06d1f024ffdbd1ba2bc6ccbda40873310090ff" Oct 14 10:23:20 crc kubenswrapper[4870]: I1014 10:23:20.222454 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7b7d59e54b23ba650fc959a79a06d1f024ffdbd1ba2bc6ccbda40873310090ff"} err="failed to get container status \"7b7d59e54b23ba650fc959a79a06d1f024ffdbd1ba2bc6ccbda40873310090ff\": rpc error: code = NotFound desc = could not find container \"7b7d59e54b23ba650fc959a79a06d1f024ffdbd1ba2bc6ccbda40873310090ff\": container with ID starting with 7b7d59e54b23ba650fc959a79a06d1f024ffdbd1ba2bc6ccbda40873310090ff not found: ID does not exist" Oct 14 10:23:20 crc kubenswrapper[4870]: I1014 10:23:20.222491 4870 scope.go:117] "RemoveContainer" containerID="e977448a9c15e70548e23973e2c5ad111f0a9bc6aa61763a298974a049395e33" Oct 14 10:23:20 crc kubenswrapper[4870]: E1014 10:23:20.222928 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e977448a9c15e70548e23973e2c5ad111f0a9bc6aa61763a298974a049395e33\": container with ID starting with e977448a9c15e70548e23973e2c5ad111f0a9bc6aa61763a298974a049395e33 not found: ID does not exist" containerID="e977448a9c15e70548e23973e2c5ad111f0a9bc6aa61763a298974a049395e33" Oct 14 10:23:20 crc kubenswrapper[4870]: I1014 10:23:20.222995 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e977448a9c15e70548e23973e2c5ad111f0a9bc6aa61763a298974a049395e33"} err="failed to get container status \"e977448a9c15e70548e23973e2c5ad111f0a9bc6aa61763a298974a049395e33\": rpc error: code = NotFound desc = could not find container \"e977448a9c15e70548e23973e2c5ad111f0a9bc6aa61763a298974a049395e33\": container with ID starting with e977448a9c15e70548e23973e2c5ad111f0a9bc6aa61763a298974a049395e33 not found: ID does not exist" Oct 14 10:23:20 crc kubenswrapper[4870]: I1014 10:23:20.223052 4870 scope.go:117] "RemoveContainer" containerID="36a8021d709c8177703591c13607d42f823c796b41ea6ef076fc5bf4e61b372a" Oct 14 10:23:20 crc kubenswrapper[4870]: E1014 10:23:20.223432 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"36a8021d709c8177703591c13607d42f823c796b41ea6ef076fc5bf4e61b372a\": container with ID starting with 36a8021d709c8177703591c13607d42f823c796b41ea6ef076fc5bf4e61b372a not found: ID does not exist" containerID="36a8021d709c8177703591c13607d42f823c796b41ea6ef076fc5bf4e61b372a" Oct 14 10:23:20 crc kubenswrapper[4870]: I1014 10:23:20.223484 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"36a8021d709c8177703591c13607d42f823c796b41ea6ef076fc5bf4e61b372a"} err="failed to get container status \"36a8021d709c8177703591c13607d42f823c796b41ea6ef076fc5bf4e61b372a\": rpc error: code = NotFound desc = could not find container \"36a8021d709c8177703591c13607d42f823c796b41ea6ef076fc5bf4e61b372a\": container with ID starting with 36a8021d709c8177703591c13607d42f823c796b41ea6ef076fc5bf4e61b372a not found: ID does not exist" Oct 14 10:23:21 crc kubenswrapper[4870]: I1014 10:23:21.055644 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b66523d6-2bc0-430c-ad75-e67ee9eee39d" path="/var/lib/kubelet/pods/b66523d6-2bc0-430c-ad75-e67ee9eee39d/volumes" Oct 14 10:24:53 crc kubenswrapper[4870]: I1014 10:24:53.951345 4870 patch_prober.go:28] interesting pod/machine-config-daemon-7tvc7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 10:24:53 crc kubenswrapper[4870]: I1014 10:24:53.951854 4870 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 10:25:23 crc kubenswrapper[4870]: I1014 10:25:23.950920 4870 patch_prober.go:28] interesting pod/machine-config-daemon-7tvc7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 10:25:23 crc kubenswrapper[4870]: I1014 10:25:23.951450 4870 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 10:25:53 crc kubenswrapper[4870]: I1014 10:25:53.952542 4870 patch_prober.go:28] interesting pod/machine-config-daemon-7tvc7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 10:25:53 crc kubenswrapper[4870]: I1014 10:25:53.953618 4870 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 10:25:53 crc kubenswrapper[4870]: I1014 10:25:53.953711 4870 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" Oct 14 10:25:53 crc kubenswrapper[4870]: I1014 10:25:53.954996 4870 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"4e0c842e7d29f428c954b866c1f3929be48b4bb5118abceba1fbf69beed50855"} pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 14 10:25:53 crc kubenswrapper[4870]: I1014 10:25:53.955855 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerName="machine-config-daemon" containerID="cri-o://4e0c842e7d29f428c954b866c1f3929be48b4bb5118abceba1fbf69beed50855" gracePeriod=600 Oct 14 10:25:55 crc kubenswrapper[4870]: I1014 10:25:55.021501 4870 generic.go:334] "Generic (PLEG): container finished" podID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerID="4e0c842e7d29f428c954b866c1f3929be48b4bb5118abceba1fbf69beed50855" exitCode=0 Oct 14 10:25:55 crc kubenswrapper[4870]: I1014 10:25:55.021686 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" event={"ID":"874a5e45-dffd-4d17-b609-b7d3ed2eab07","Type":"ContainerDied","Data":"4e0c842e7d29f428c954b866c1f3929be48b4bb5118abceba1fbf69beed50855"} Oct 14 10:25:55 crc kubenswrapper[4870]: I1014 10:25:55.022168 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" event={"ID":"874a5e45-dffd-4d17-b609-b7d3ed2eab07","Type":"ContainerStarted","Data":"3a25412c639a58b6345a6cb926f93c78a7a62c68cd923f500f921497a90b6efb"} Oct 14 10:25:55 crc kubenswrapper[4870]: I1014 10:25:55.022206 4870 scope.go:117] "RemoveContainer" containerID="7468e2739d25e53085cbe91cf249f36e5760f983ae0240cf58bcc9bcb1b91355" Oct 14 10:26:25 crc kubenswrapper[4870]: I1014 10:26:25.575904 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-p74nh"] Oct 14 10:26:25 crc kubenswrapper[4870]: E1014 10:26:25.577422 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b66523d6-2bc0-430c-ad75-e67ee9eee39d" containerName="extract-utilities" Oct 14 10:26:25 crc kubenswrapper[4870]: I1014 10:26:25.577464 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="b66523d6-2bc0-430c-ad75-e67ee9eee39d" containerName="extract-utilities" Oct 14 10:26:25 crc kubenswrapper[4870]: E1014 10:26:25.577499 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b66523d6-2bc0-430c-ad75-e67ee9eee39d" containerName="registry-server" Oct 14 10:26:25 crc kubenswrapper[4870]: I1014 10:26:25.577512 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="b66523d6-2bc0-430c-ad75-e67ee9eee39d" containerName="registry-server" Oct 14 10:26:25 crc kubenswrapper[4870]: E1014 10:26:25.577563 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b66523d6-2bc0-430c-ad75-e67ee9eee39d" containerName="extract-content" Oct 14 10:26:25 crc kubenswrapper[4870]: I1014 10:26:25.577573 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="b66523d6-2bc0-430c-ad75-e67ee9eee39d" containerName="extract-content" Oct 14 10:26:25 crc kubenswrapper[4870]: I1014 10:26:25.577962 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="b66523d6-2bc0-430c-ad75-e67ee9eee39d" containerName="registry-server" Oct 14 10:26:25 crc kubenswrapper[4870]: I1014 10:26:25.580373 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-p74nh" Oct 14 10:26:25 crc kubenswrapper[4870]: I1014 10:26:25.590278 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-p74nh"] Oct 14 10:26:25 crc kubenswrapper[4870]: I1014 10:26:25.612886 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/46bd7024-9c7e-4906-a48c-6d6034c27bee-catalog-content\") pod \"community-operators-p74nh\" (UID: \"46bd7024-9c7e-4906-a48c-6d6034c27bee\") " pod="openshift-marketplace/community-operators-p74nh" Oct 14 10:26:25 crc kubenswrapper[4870]: I1014 10:26:25.613153 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/46bd7024-9c7e-4906-a48c-6d6034c27bee-utilities\") pod \"community-operators-p74nh\" (UID: \"46bd7024-9c7e-4906-a48c-6d6034c27bee\") " pod="openshift-marketplace/community-operators-p74nh" Oct 14 10:26:25 crc kubenswrapper[4870]: I1014 10:26:25.613193 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s9fnb\" (UniqueName: \"kubernetes.io/projected/46bd7024-9c7e-4906-a48c-6d6034c27bee-kube-api-access-s9fnb\") pod \"community-operators-p74nh\" (UID: \"46bd7024-9c7e-4906-a48c-6d6034c27bee\") " pod="openshift-marketplace/community-operators-p74nh" Oct 14 10:26:25 crc kubenswrapper[4870]: I1014 10:26:25.717831 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/46bd7024-9c7e-4906-a48c-6d6034c27bee-utilities\") pod \"community-operators-p74nh\" (UID: \"46bd7024-9c7e-4906-a48c-6d6034c27bee\") " pod="openshift-marketplace/community-operators-p74nh" Oct 14 10:26:25 crc kubenswrapper[4870]: I1014 10:26:25.717894 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s9fnb\" (UniqueName: \"kubernetes.io/projected/46bd7024-9c7e-4906-a48c-6d6034c27bee-kube-api-access-s9fnb\") pod \"community-operators-p74nh\" (UID: \"46bd7024-9c7e-4906-a48c-6d6034c27bee\") " pod="openshift-marketplace/community-operators-p74nh" Oct 14 10:26:25 crc kubenswrapper[4870]: I1014 10:26:25.718006 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/46bd7024-9c7e-4906-a48c-6d6034c27bee-catalog-content\") pod \"community-operators-p74nh\" (UID: \"46bd7024-9c7e-4906-a48c-6d6034c27bee\") " pod="openshift-marketplace/community-operators-p74nh" Oct 14 10:26:25 crc kubenswrapper[4870]: I1014 10:26:25.718881 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/46bd7024-9c7e-4906-a48c-6d6034c27bee-catalog-content\") pod \"community-operators-p74nh\" (UID: \"46bd7024-9c7e-4906-a48c-6d6034c27bee\") " pod="openshift-marketplace/community-operators-p74nh" Oct 14 10:26:25 crc kubenswrapper[4870]: I1014 10:26:25.719147 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/46bd7024-9c7e-4906-a48c-6d6034c27bee-utilities\") pod \"community-operators-p74nh\" (UID: \"46bd7024-9c7e-4906-a48c-6d6034c27bee\") " pod="openshift-marketplace/community-operators-p74nh" Oct 14 10:26:25 crc kubenswrapper[4870]: I1014 10:26:25.743280 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s9fnb\" (UniqueName: \"kubernetes.io/projected/46bd7024-9c7e-4906-a48c-6d6034c27bee-kube-api-access-s9fnb\") pod \"community-operators-p74nh\" (UID: \"46bd7024-9c7e-4906-a48c-6d6034c27bee\") " pod="openshift-marketplace/community-operators-p74nh" Oct 14 10:26:25 crc kubenswrapper[4870]: I1014 10:26:25.922553 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-p74nh" Oct 14 10:26:26 crc kubenswrapper[4870]: I1014 10:26:26.439496 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-p74nh"] Oct 14 10:26:26 crc kubenswrapper[4870]: I1014 10:26:26.465895 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p74nh" event={"ID":"46bd7024-9c7e-4906-a48c-6d6034c27bee","Type":"ContainerStarted","Data":"d38f95619fe47d3f5ec98d445804b964a480226a78e8feac1cceef089e738b26"} Oct 14 10:26:27 crc kubenswrapper[4870]: I1014 10:26:27.390908 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-x9w5p"] Oct 14 10:26:27 crc kubenswrapper[4870]: I1014 10:26:27.394530 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-x9w5p" Oct 14 10:26:27 crc kubenswrapper[4870]: I1014 10:26:27.408460 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-x9w5p"] Oct 14 10:26:27 crc kubenswrapper[4870]: I1014 10:26:27.476740 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/12a63494-1118-4826-a873-50fa5df427dd-catalog-content\") pod \"redhat-operators-x9w5p\" (UID: \"12a63494-1118-4826-a873-50fa5df427dd\") " pod="openshift-marketplace/redhat-operators-x9w5p" Oct 14 10:26:27 crc kubenswrapper[4870]: I1014 10:26:27.477202 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/12a63494-1118-4826-a873-50fa5df427dd-utilities\") pod \"redhat-operators-x9w5p\" (UID: \"12a63494-1118-4826-a873-50fa5df427dd\") " pod="openshift-marketplace/redhat-operators-x9w5p" Oct 14 10:26:27 crc kubenswrapper[4870]: I1014 10:26:27.477253 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t5r2x\" (UniqueName: \"kubernetes.io/projected/12a63494-1118-4826-a873-50fa5df427dd-kube-api-access-t5r2x\") pod \"redhat-operators-x9w5p\" (UID: \"12a63494-1118-4826-a873-50fa5df427dd\") " pod="openshift-marketplace/redhat-operators-x9w5p" Oct 14 10:26:27 crc kubenswrapper[4870]: I1014 10:26:27.481419 4870 generic.go:334] "Generic (PLEG): container finished" podID="46bd7024-9c7e-4906-a48c-6d6034c27bee" containerID="5872c468c4140a8c9bf9e3b2ba85afdb854e111177aa39951ae5930ec957fd3b" exitCode=0 Oct 14 10:26:27 crc kubenswrapper[4870]: I1014 10:26:27.481470 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p74nh" event={"ID":"46bd7024-9c7e-4906-a48c-6d6034c27bee","Type":"ContainerDied","Data":"5872c468c4140a8c9bf9e3b2ba85afdb854e111177aa39951ae5930ec957fd3b"} Oct 14 10:26:27 crc kubenswrapper[4870]: I1014 10:26:27.579689 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/12a63494-1118-4826-a873-50fa5df427dd-catalog-content\") pod \"redhat-operators-x9w5p\" (UID: \"12a63494-1118-4826-a873-50fa5df427dd\") " pod="openshift-marketplace/redhat-operators-x9w5p" Oct 14 10:26:27 crc kubenswrapper[4870]: I1014 10:26:27.579871 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/12a63494-1118-4826-a873-50fa5df427dd-utilities\") pod \"redhat-operators-x9w5p\" (UID: \"12a63494-1118-4826-a873-50fa5df427dd\") " pod="openshift-marketplace/redhat-operators-x9w5p" Oct 14 10:26:27 crc kubenswrapper[4870]: I1014 10:26:27.579900 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t5r2x\" (UniqueName: \"kubernetes.io/projected/12a63494-1118-4826-a873-50fa5df427dd-kube-api-access-t5r2x\") pod \"redhat-operators-x9w5p\" (UID: \"12a63494-1118-4826-a873-50fa5df427dd\") " pod="openshift-marketplace/redhat-operators-x9w5p" Oct 14 10:26:27 crc kubenswrapper[4870]: I1014 10:26:27.580474 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/12a63494-1118-4826-a873-50fa5df427dd-catalog-content\") pod \"redhat-operators-x9w5p\" (UID: \"12a63494-1118-4826-a873-50fa5df427dd\") " pod="openshift-marketplace/redhat-operators-x9w5p" Oct 14 10:26:27 crc kubenswrapper[4870]: I1014 10:26:27.580625 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/12a63494-1118-4826-a873-50fa5df427dd-utilities\") pod \"redhat-operators-x9w5p\" (UID: \"12a63494-1118-4826-a873-50fa5df427dd\") " pod="openshift-marketplace/redhat-operators-x9w5p" Oct 14 10:26:27 crc kubenswrapper[4870]: I1014 10:26:27.601560 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t5r2x\" (UniqueName: \"kubernetes.io/projected/12a63494-1118-4826-a873-50fa5df427dd-kube-api-access-t5r2x\") pod \"redhat-operators-x9w5p\" (UID: \"12a63494-1118-4826-a873-50fa5df427dd\") " pod="openshift-marketplace/redhat-operators-x9w5p" Oct 14 10:26:27 crc kubenswrapper[4870]: I1014 10:26:27.725994 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-x9w5p" Oct 14 10:26:28 crc kubenswrapper[4870]: I1014 10:26:28.201868 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-x9w5p"] Oct 14 10:26:28 crc kubenswrapper[4870]: W1014 10:26:28.217564 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod12a63494_1118_4826_a873_50fa5df427dd.slice/crio-ed6bc777a5659c1ff6d0dec04b3fe62e5618ef47630d190ba628a9c1bd9f9fcb WatchSource:0}: Error finding container ed6bc777a5659c1ff6d0dec04b3fe62e5618ef47630d190ba628a9c1bd9f9fcb: Status 404 returned error can't find the container with id ed6bc777a5659c1ff6d0dec04b3fe62e5618ef47630d190ba628a9c1bd9f9fcb Oct 14 10:26:28 crc kubenswrapper[4870]: I1014 10:26:28.494118 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x9w5p" event={"ID":"12a63494-1118-4826-a873-50fa5df427dd","Type":"ContainerStarted","Data":"9b0a94a44d27d209b611aefb7af8274da40abceb90270f6233764dfc8a0abe02"} Oct 14 10:26:28 crc kubenswrapper[4870]: I1014 10:26:28.494382 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x9w5p" event={"ID":"12a63494-1118-4826-a873-50fa5df427dd","Type":"ContainerStarted","Data":"ed6bc777a5659c1ff6d0dec04b3fe62e5618ef47630d190ba628a9c1bd9f9fcb"} Oct 14 10:26:28 crc kubenswrapper[4870]: I1014 10:26:28.499589 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p74nh" event={"ID":"46bd7024-9c7e-4906-a48c-6d6034c27bee","Type":"ContainerStarted","Data":"9ee7ad13e366e910afe556b0daa8475868a5d304ebd1c3df35079ce1e5d5a240"} Oct 14 10:26:28 crc kubenswrapper[4870]: I1014 10:26:28.773041 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-brr97"] Oct 14 10:26:28 crc kubenswrapper[4870]: I1014 10:26:28.776531 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-brr97" Oct 14 10:26:28 crc kubenswrapper[4870]: I1014 10:26:28.784361 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-brr97"] Oct 14 10:26:28 crc kubenswrapper[4870]: I1014 10:26:28.801558 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aabbe5b0-7808-44e8-a9e7-83eaaf919511-catalog-content\") pod \"redhat-marketplace-brr97\" (UID: \"aabbe5b0-7808-44e8-a9e7-83eaaf919511\") " pod="openshift-marketplace/redhat-marketplace-brr97" Oct 14 10:26:28 crc kubenswrapper[4870]: I1014 10:26:28.802040 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b8whs\" (UniqueName: \"kubernetes.io/projected/aabbe5b0-7808-44e8-a9e7-83eaaf919511-kube-api-access-b8whs\") pod \"redhat-marketplace-brr97\" (UID: \"aabbe5b0-7808-44e8-a9e7-83eaaf919511\") " pod="openshift-marketplace/redhat-marketplace-brr97" Oct 14 10:26:28 crc kubenswrapper[4870]: I1014 10:26:28.802337 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aabbe5b0-7808-44e8-a9e7-83eaaf919511-utilities\") pod \"redhat-marketplace-brr97\" (UID: \"aabbe5b0-7808-44e8-a9e7-83eaaf919511\") " pod="openshift-marketplace/redhat-marketplace-brr97" Oct 14 10:26:28 crc kubenswrapper[4870]: I1014 10:26:28.903361 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b8whs\" (UniqueName: \"kubernetes.io/projected/aabbe5b0-7808-44e8-a9e7-83eaaf919511-kube-api-access-b8whs\") pod \"redhat-marketplace-brr97\" (UID: \"aabbe5b0-7808-44e8-a9e7-83eaaf919511\") " pod="openshift-marketplace/redhat-marketplace-brr97" Oct 14 10:26:28 crc kubenswrapper[4870]: I1014 10:26:28.903527 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aabbe5b0-7808-44e8-a9e7-83eaaf919511-utilities\") pod \"redhat-marketplace-brr97\" (UID: \"aabbe5b0-7808-44e8-a9e7-83eaaf919511\") " pod="openshift-marketplace/redhat-marketplace-brr97" Oct 14 10:26:28 crc kubenswrapper[4870]: I1014 10:26:28.903599 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aabbe5b0-7808-44e8-a9e7-83eaaf919511-catalog-content\") pod \"redhat-marketplace-brr97\" (UID: \"aabbe5b0-7808-44e8-a9e7-83eaaf919511\") " pod="openshift-marketplace/redhat-marketplace-brr97" Oct 14 10:26:28 crc kubenswrapper[4870]: I1014 10:26:28.904312 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aabbe5b0-7808-44e8-a9e7-83eaaf919511-utilities\") pod \"redhat-marketplace-brr97\" (UID: \"aabbe5b0-7808-44e8-a9e7-83eaaf919511\") " pod="openshift-marketplace/redhat-marketplace-brr97" Oct 14 10:26:28 crc kubenswrapper[4870]: I1014 10:26:28.904350 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aabbe5b0-7808-44e8-a9e7-83eaaf919511-catalog-content\") pod \"redhat-marketplace-brr97\" (UID: \"aabbe5b0-7808-44e8-a9e7-83eaaf919511\") " pod="openshift-marketplace/redhat-marketplace-brr97" Oct 14 10:26:28 crc kubenswrapper[4870]: I1014 10:26:28.937856 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b8whs\" (UniqueName: \"kubernetes.io/projected/aabbe5b0-7808-44e8-a9e7-83eaaf919511-kube-api-access-b8whs\") pod \"redhat-marketplace-brr97\" (UID: \"aabbe5b0-7808-44e8-a9e7-83eaaf919511\") " pod="openshift-marketplace/redhat-marketplace-brr97" Oct 14 10:26:29 crc kubenswrapper[4870]: I1014 10:26:29.096546 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-brr97" Oct 14 10:26:29 crc kubenswrapper[4870]: I1014 10:26:29.509083 4870 generic.go:334] "Generic (PLEG): container finished" podID="12a63494-1118-4826-a873-50fa5df427dd" containerID="9b0a94a44d27d209b611aefb7af8274da40abceb90270f6233764dfc8a0abe02" exitCode=0 Oct 14 10:26:29 crc kubenswrapper[4870]: I1014 10:26:29.510402 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x9w5p" event={"ID":"12a63494-1118-4826-a873-50fa5df427dd","Type":"ContainerDied","Data":"9b0a94a44d27d209b611aefb7af8274da40abceb90270f6233764dfc8a0abe02"} Oct 14 10:26:29 crc kubenswrapper[4870]: I1014 10:26:29.627853 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-brr97"] Oct 14 10:26:30 crc kubenswrapper[4870]: I1014 10:26:30.528736 4870 generic.go:334] "Generic (PLEG): container finished" podID="aabbe5b0-7808-44e8-a9e7-83eaaf919511" containerID="6231f85ad4f94f3ab5320f32efc3bea0740e30cf00ee25c215164366a0429fbc" exitCode=0 Oct 14 10:26:30 crc kubenswrapper[4870]: I1014 10:26:30.529511 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-brr97" event={"ID":"aabbe5b0-7808-44e8-a9e7-83eaaf919511","Type":"ContainerDied","Data":"6231f85ad4f94f3ab5320f32efc3bea0740e30cf00ee25c215164366a0429fbc"} Oct 14 10:26:30 crc kubenswrapper[4870]: I1014 10:26:30.529578 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-brr97" event={"ID":"aabbe5b0-7808-44e8-a9e7-83eaaf919511","Type":"ContainerStarted","Data":"3ac7c8df61577ef0847f2c874da49680e1f6a52eeed05c14c4076ad6f4cca359"} Oct 14 10:26:30 crc kubenswrapper[4870]: I1014 10:26:30.533618 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x9w5p" event={"ID":"12a63494-1118-4826-a873-50fa5df427dd","Type":"ContainerStarted","Data":"45e04185ca3f57b7711f351786981d4eb326fb093a548cf0c974dfec3c17c9ad"} Oct 14 10:26:30 crc kubenswrapper[4870]: I1014 10:26:30.540191 4870 generic.go:334] "Generic (PLEG): container finished" podID="46bd7024-9c7e-4906-a48c-6d6034c27bee" containerID="9ee7ad13e366e910afe556b0daa8475868a5d304ebd1c3df35079ce1e5d5a240" exitCode=0 Oct 14 10:26:30 crc kubenswrapper[4870]: I1014 10:26:30.540245 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p74nh" event={"ID":"46bd7024-9c7e-4906-a48c-6d6034c27bee","Type":"ContainerDied","Data":"9ee7ad13e366e910afe556b0daa8475868a5d304ebd1c3df35079ce1e5d5a240"} Oct 14 10:26:31 crc kubenswrapper[4870]: I1014 10:26:31.553251 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-brr97" event={"ID":"aabbe5b0-7808-44e8-a9e7-83eaaf919511","Type":"ContainerStarted","Data":"f1640e8d1e0d18b376e1ca7bd82900817c5d88ddd06b34611c632e1e0fda3620"} Oct 14 10:26:31 crc kubenswrapper[4870]: I1014 10:26:31.556078 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p74nh" event={"ID":"46bd7024-9c7e-4906-a48c-6d6034c27bee","Type":"ContainerStarted","Data":"fb8dca5495a657298b5da9f1346d48e73b65ff8a64e1f923ac01376ded87ac4a"} Oct 14 10:26:31 crc kubenswrapper[4870]: I1014 10:26:31.617159 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-p74nh" podStartSLOduration=3.180375987 podStartE2EDuration="6.617135669s" podCreationTimestamp="2025-10-14 10:26:25 +0000 UTC" firstStartedPulling="2025-10-14 10:26:27.483554785 +0000 UTC m=+12323.180915156" lastFinishedPulling="2025-10-14 10:26:30.920314467 +0000 UTC m=+12326.617674838" observedRunningTime="2025-10-14 10:26:31.607668058 +0000 UTC m=+12327.305028429" watchObservedRunningTime="2025-10-14 10:26:31.617135669 +0000 UTC m=+12327.314496040" Oct 14 10:26:32 crc kubenswrapper[4870]: I1014 10:26:32.570013 4870 generic.go:334] "Generic (PLEG): container finished" podID="aabbe5b0-7808-44e8-a9e7-83eaaf919511" containerID="f1640e8d1e0d18b376e1ca7bd82900817c5d88ddd06b34611c632e1e0fda3620" exitCode=0 Oct 14 10:26:32 crc kubenswrapper[4870]: I1014 10:26:32.570142 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-brr97" event={"ID":"aabbe5b0-7808-44e8-a9e7-83eaaf919511","Type":"ContainerDied","Data":"f1640e8d1e0d18b376e1ca7bd82900817c5d88ddd06b34611c632e1e0fda3620"} Oct 14 10:26:33 crc kubenswrapper[4870]: I1014 10:26:33.582289 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-brr97" event={"ID":"aabbe5b0-7808-44e8-a9e7-83eaaf919511","Type":"ContainerStarted","Data":"ccd7153d21b88b740470968454416cdf98f5f1d66c975ae2ef094127fc9a3564"} Oct 14 10:26:33 crc kubenswrapper[4870]: I1014 10:26:33.687127 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-brr97" podStartSLOduration=2.990573882 podStartE2EDuration="5.68710307s" podCreationTimestamp="2025-10-14 10:26:28 +0000 UTC" firstStartedPulling="2025-10-14 10:26:30.533285103 +0000 UTC m=+12326.230645494" lastFinishedPulling="2025-10-14 10:26:33.229814301 +0000 UTC m=+12328.927174682" observedRunningTime="2025-10-14 10:26:33.602429624 +0000 UTC m=+12329.299790005" watchObservedRunningTime="2025-10-14 10:26:33.68710307 +0000 UTC m=+12329.384463431" Oct 14 10:26:34 crc kubenswrapper[4870]: I1014 10:26:34.599037 4870 generic.go:334] "Generic (PLEG): container finished" podID="12a63494-1118-4826-a873-50fa5df427dd" containerID="45e04185ca3f57b7711f351786981d4eb326fb093a548cf0c974dfec3c17c9ad" exitCode=0 Oct 14 10:26:34 crc kubenswrapper[4870]: I1014 10:26:34.599160 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x9w5p" event={"ID":"12a63494-1118-4826-a873-50fa5df427dd","Type":"ContainerDied","Data":"45e04185ca3f57b7711f351786981d4eb326fb093a548cf0c974dfec3c17c9ad"} Oct 14 10:26:35 crc kubenswrapper[4870]: I1014 10:26:35.617110 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x9w5p" event={"ID":"12a63494-1118-4826-a873-50fa5df427dd","Type":"ContainerStarted","Data":"f9dc8f5f83c29706a2525d37f680c3cfc5b2e36e404dcb9a0abe26ca2813634c"} Oct 14 10:26:35 crc kubenswrapper[4870]: I1014 10:26:35.638253 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-x9w5p" podStartSLOduration=3.017005204 podStartE2EDuration="8.638230839s" podCreationTimestamp="2025-10-14 10:26:27 +0000 UTC" firstStartedPulling="2025-10-14 10:26:29.511650283 +0000 UTC m=+12325.209010654" lastFinishedPulling="2025-10-14 10:26:35.132875888 +0000 UTC m=+12330.830236289" observedRunningTime="2025-10-14 10:26:35.634491578 +0000 UTC m=+12331.331851949" watchObservedRunningTime="2025-10-14 10:26:35.638230839 +0000 UTC m=+12331.335591230" Oct 14 10:26:35 crc kubenswrapper[4870]: I1014 10:26:35.924131 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-p74nh" Oct 14 10:26:35 crc kubenswrapper[4870]: I1014 10:26:35.924176 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-p74nh" Oct 14 10:26:35 crc kubenswrapper[4870]: I1014 10:26:35.996924 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-p74nh" Oct 14 10:26:36 crc kubenswrapper[4870]: I1014 10:26:36.715103 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-p74nh" Oct 14 10:26:37 crc kubenswrapper[4870]: I1014 10:26:37.726659 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-x9w5p" Oct 14 10:26:37 crc kubenswrapper[4870]: I1014 10:26:37.726849 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-x9w5p" Oct 14 10:26:38 crc kubenswrapper[4870]: I1014 10:26:38.784259 4870 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-x9w5p" podUID="12a63494-1118-4826-a873-50fa5df427dd" containerName="registry-server" probeResult="failure" output=< Oct 14 10:26:38 crc kubenswrapper[4870]: timeout: failed to connect service ":50051" within 1s Oct 14 10:26:38 crc kubenswrapper[4870]: > Oct 14 10:26:39 crc kubenswrapper[4870]: I1014 10:26:39.096812 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-brr97" Oct 14 10:26:39 crc kubenswrapper[4870]: I1014 10:26:39.097195 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-brr97" Oct 14 10:26:39 crc kubenswrapper[4870]: I1014 10:26:39.165122 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-p74nh"] Oct 14 10:26:39 crc kubenswrapper[4870]: I1014 10:26:39.165361 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-p74nh" podUID="46bd7024-9c7e-4906-a48c-6d6034c27bee" containerName="registry-server" containerID="cri-o://fb8dca5495a657298b5da9f1346d48e73b65ff8a64e1f923ac01376ded87ac4a" gracePeriod=2 Oct 14 10:26:39 crc kubenswrapper[4870]: I1014 10:26:39.184724 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-brr97" Oct 14 10:26:39 crc kubenswrapper[4870]: I1014 10:26:39.677804 4870 generic.go:334] "Generic (PLEG): container finished" podID="46bd7024-9c7e-4906-a48c-6d6034c27bee" containerID="fb8dca5495a657298b5da9f1346d48e73b65ff8a64e1f923ac01376ded87ac4a" exitCode=0 Oct 14 10:26:39 crc kubenswrapper[4870]: I1014 10:26:39.677892 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p74nh" event={"ID":"46bd7024-9c7e-4906-a48c-6d6034c27bee","Type":"ContainerDied","Data":"fb8dca5495a657298b5da9f1346d48e73b65ff8a64e1f923ac01376ded87ac4a"} Oct 14 10:26:39 crc kubenswrapper[4870]: I1014 10:26:39.678217 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p74nh" event={"ID":"46bd7024-9c7e-4906-a48c-6d6034c27bee","Type":"ContainerDied","Data":"d38f95619fe47d3f5ec98d445804b964a480226a78e8feac1cceef089e738b26"} Oct 14 10:26:39 crc kubenswrapper[4870]: I1014 10:26:39.678244 4870 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d38f95619fe47d3f5ec98d445804b964a480226a78e8feac1cceef089e738b26" Oct 14 10:26:39 crc kubenswrapper[4870]: I1014 10:26:39.737763 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-brr97" Oct 14 10:26:39 crc kubenswrapper[4870]: I1014 10:26:39.748149 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-p74nh" Oct 14 10:26:39 crc kubenswrapper[4870]: I1014 10:26:39.817005 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s9fnb\" (UniqueName: \"kubernetes.io/projected/46bd7024-9c7e-4906-a48c-6d6034c27bee-kube-api-access-s9fnb\") pod \"46bd7024-9c7e-4906-a48c-6d6034c27bee\" (UID: \"46bd7024-9c7e-4906-a48c-6d6034c27bee\") " Oct 14 10:26:39 crc kubenswrapper[4870]: I1014 10:26:39.817123 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/46bd7024-9c7e-4906-a48c-6d6034c27bee-catalog-content\") pod \"46bd7024-9c7e-4906-a48c-6d6034c27bee\" (UID: \"46bd7024-9c7e-4906-a48c-6d6034c27bee\") " Oct 14 10:26:39 crc kubenswrapper[4870]: I1014 10:26:39.817315 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/46bd7024-9c7e-4906-a48c-6d6034c27bee-utilities\") pod \"46bd7024-9c7e-4906-a48c-6d6034c27bee\" (UID: \"46bd7024-9c7e-4906-a48c-6d6034c27bee\") " Oct 14 10:26:39 crc kubenswrapper[4870]: I1014 10:26:39.818265 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/46bd7024-9c7e-4906-a48c-6d6034c27bee-utilities" (OuterVolumeSpecName: "utilities") pod "46bd7024-9c7e-4906-a48c-6d6034c27bee" (UID: "46bd7024-9c7e-4906-a48c-6d6034c27bee"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 10:26:39 crc kubenswrapper[4870]: I1014 10:26:39.825469 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/46bd7024-9c7e-4906-a48c-6d6034c27bee-kube-api-access-s9fnb" (OuterVolumeSpecName: "kube-api-access-s9fnb") pod "46bd7024-9c7e-4906-a48c-6d6034c27bee" (UID: "46bd7024-9c7e-4906-a48c-6d6034c27bee"). InnerVolumeSpecName "kube-api-access-s9fnb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 10:26:39 crc kubenswrapper[4870]: I1014 10:26:39.890827 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/46bd7024-9c7e-4906-a48c-6d6034c27bee-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "46bd7024-9c7e-4906-a48c-6d6034c27bee" (UID: "46bd7024-9c7e-4906-a48c-6d6034c27bee"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 10:26:39 crc kubenswrapper[4870]: I1014 10:26:39.919214 4870 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/46bd7024-9c7e-4906-a48c-6d6034c27bee-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 10:26:39 crc kubenswrapper[4870]: I1014 10:26:39.919254 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s9fnb\" (UniqueName: \"kubernetes.io/projected/46bd7024-9c7e-4906-a48c-6d6034c27bee-kube-api-access-s9fnb\") on node \"crc\" DevicePath \"\"" Oct 14 10:26:39 crc kubenswrapper[4870]: I1014 10:26:39.919267 4870 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/46bd7024-9c7e-4906-a48c-6d6034c27bee-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 10:26:40 crc kubenswrapper[4870]: I1014 10:26:40.689675 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-p74nh" Oct 14 10:26:40 crc kubenswrapper[4870]: I1014 10:26:40.750602 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-p74nh"] Oct 14 10:26:40 crc kubenswrapper[4870]: I1014 10:26:40.765134 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-p74nh"] Oct 14 10:26:41 crc kubenswrapper[4870]: I1014 10:26:41.053858 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="46bd7024-9c7e-4906-a48c-6d6034c27bee" path="/var/lib/kubelet/pods/46bd7024-9c7e-4906-a48c-6d6034c27bee/volumes" Oct 14 10:26:41 crc kubenswrapper[4870]: I1014 10:26:41.766424 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-brr97"] Oct 14 10:26:42 crc kubenswrapper[4870]: I1014 10:26:42.712314 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-brr97" podUID="aabbe5b0-7808-44e8-a9e7-83eaaf919511" containerName="registry-server" containerID="cri-o://ccd7153d21b88b740470968454416cdf98f5f1d66c975ae2ef094127fc9a3564" gracePeriod=2 Oct 14 10:26:43 crc kubenswrapper[4870]: I1014 10:26:43.288177 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-brr97" Oct 14 10:26:43 crc kubenswrapper[4870]: I1014 10:26:43.404322 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aabbe5b0-7808-44e8-a9e7-83eaaf919511-utilities\") pod \"aabbe5b0-7808-44e8-a9e7-83eaaf919511\" (UID: \"aabbe5b0-7808-44e8-a9e7-83eaaf919511\") " Oct 14 10:26:43 crc kubenswrapper[4870]: I1014 10:26:43.405332 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aabbe5b0-7808-44e8-a9e7-83eaaf919511-utilities" (OuterVolumeSpecName: "utilities") pod "aabbe5b0-7808-44e8-a9e7-83eaaf919511" (UID: "aabbe5b0-7808-44e8-a9e7-83eaaf919511"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 10:26:43 crc kubenswrapper[4870]: I1014 10:26:43.405530 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aabbe5b0-7808-44e8-a9e7-83eaaf919511-catalog-content\") pod \"aabbe5b0-7808-44e8-a9e7-83eaaf919511\" (UID: \"aabbe5b0-7808-44e8-a9e7-83eaaf919511\") " Oct 14 10:26:43 crc kubenswrapper[4870]: I1014 10:26:43.407991 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b8whs\" (UniqueName: \"kubernetes.io/projected/aabbe5b0-7808-44e8-a9e7-83eaaf919511-kube-api-access-b8whs\") pod \"aabbe5b0-7808-44e8-a9e7-83eaaf919511\" (UID: \"aabbe5b0-7808-44e8-a9e7-83eaaf919511\") " Oct 14 10:26:43 crc kubenswrapper[4870]: I1014 10:26:43.409123 4870 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aabbe5b0-7808-44e8-a9e7-83eaaf919511-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 10:26:43 crc kubenswrapper[4870]: I1014 10:26:43.413727 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aabbe5b0-7808-44e8-a9e7-83eaaf919511-kube-api-access-b8whs" (OuterVolumeSpecName: "kube-api-access-b8whs") pod "aabbe5b0-7808-44e8-a9e7-83eaaf919511" (UID: "aabbe5b0-7808-44e8-a9e7-83eaaf919511"). InnerVolumeSpecName "kube-api-access-b8whs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 10:26:43 crc kubenswrapper[4870]: I1014 10:26:43.424580 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aabbe5b0-7808-44e8-a9e7-83eaaf919511-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "aabbe5b0-7808-44e8-a9e7-83eaaf919511" (UID: "aabbe5b0-7808-44e8-a9e7-83eaaf919511"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 10:26:43 crc kubenswrapper[4870]: I1014 10:26:43.511282 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b8whs\" (UniqueName: \"kubernetes.io/projected/aabbe5b0-7808-44e8-a9e7-83eaaf919511-kube-api-access-b8whs\") on node \"crc\" DevicePath \"\"" Oct 14 10:26:43 crc kubenswrapper[4870]: I1014 10:26:43.511318 4870 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aabbe5b0-7808-44e8-a9e7-83eaaf919511-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 10:26:43 crc kubenswrapper[4870]: I1014 10:26:43.727516 4870 generic.go:334] "Generic (PLEG): container finished" podID="aabbe5b0-7808-44e8-a9e7-83eaaf919511" containerID="ccd7153d21b88b740470968454416cdf98f5f1d66c975ae2ef094127fc9a3564" exitCode=0 Oct 14 10:26:43 crc kubenswrapper[4870]: I1014 10:26:43.727562 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-brr97" event={"ID":"aabbe5b0-7808-44e8-a9e7-83eaaf919511","Type":"ContainerDied","Data":"ccd7153d21b88b740470968454416cdf98f5f1d66c975ae2ef094127fc9a3564"} Oct 14 10:26:43 crc kubenswrapper[4870]: I1014 10:26:43.727591 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-brr97" event={"ID":"aabbe5b0-7808-44e8-a9e7-83eaaf919511","Type":"ContainerDied","Data":"3ac7c8df61577ef0847f2c874da49680e1f6a52eeed05c14c4076ad6f4cca359"} Oct 14 10:26:43 crc kubenswrapper[4870]: I1014 10:26:43.727608 4870 scope.go:117] "RemoveContainer" containerID="ccd7153d21b88b740470968454416cdf98f5f1d66c975ae2ef094127fc9a3564" Oct 14 10:26:43 crc kubenswrapper[4870]: I1014 10:26:43.727632 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-brr97" Oct 14 10:26:43 crc kubenswrapper[4870]: I1014 10:26:43.794210 4870 scope.go:117] "RemoveContainer" containerID="f1640e8d1e0d18b376e1ca7bd82900817c5d88ddd06b34611c632e1e0fda3620" Oct 14 10:26:43 crc kubenswrapper[4870]: I1014 10:26:43.809354 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-brr97"] Oct 14 10:26:43 crc kubenswrapper[4870]: I1014 10:26:43.821332 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-brr97"] Oct 14 10:26:43 crc kubenswrapper[4870]: I1014 10:26:43.828564 4870 scope.go:117] "RemoveContainer" containerID="6231f85ad4f94f3ab5320f32efc3bea0740e30cf00ee25c215164366a0429fbc" Oct 14 10:26:43 crc kubenswrapper[4870]: I1014 10:26:43.890972 4870 scope.go:117] "RemoveContainer" containerID="ccd7153d21b88b740470968454416cdf98f5f1d66c975ae2ef094127fc9a3564" Oct 14 10:26:43 crc kubenswrapper[4870]: E1014 10:26:43.891465 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ccd7153d21b88b740470968454416cdf98f5f1d66c975ae2ef094127fc9a3564\": container with ID starting with ccd7153d21b88b740470968454416cdf98f5f1d66c975ae2ef094127fc9a3564 not found: ID does not exist" containerID="ccd7153d21b88b740470968454416cdf98f5f1d66c975ae2ef094127fc9a3564" Oct 14 10:26:43 crc kubenswrapper[4870]: I1014 10:26:43.891509 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ccd7153d21b88b740470968454416cdf98f5f1d66c975ae2ef094127fc9a3564"} err="failed to get container status \"ccd7153d21b88b740470968454416cdf98f5f1d66c975ae2ef094127fc9a3564\": rpc error: code = NotFound desc = could not find container \"ccd7153d21b88b740470968454416cdf98f5f1d66c975ae2ef094127fc9a3564\": container with ID starting with ccd7153d21b88b740470968454416cdf98f5f1d66c975ae2ef094127fc9a3564 not found: ID does not exist" Oct 14 10:26:43 crc kubenswrapper[4870]: I1014 10:26:43.891534 4870 scope.go:117] "RemoveContainer" containerID="f1640e8d1e0d18b376e1ca7bd82900817c5d88ddd06b34611c632e1e0fda3620" Oct 14 10:26:43 crc kubenswrapper[4870]: E1014 10:26:43.891826 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f1640e8d1e0d18b376e1ca7bd82900817c5d88ddd06b34611c632e1e0fda3620\": container with ID starting with f1640e8d1e0d18b376e1ca7bd82900817c5d88ddd06b34611c632e1e0fda3620 not found: ID does not exist" containerID="f1640e8d1e0d18b376e1ca7bd82900817c5d88ddd06b34611c632e1e0fda3620" Oct 14 10:26:43 crc kubenswrapper[4870]: I1014 10:26:43.891857 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f1640e8d1e0d18b376e1ca7bd82900817c5d88ddd06b34611c632e1e0fda3620"} err="failed to get container status \"f1640e8d1e0d18b376e1ca7bd82900817c5d88ddd06b34611c632e1e0fda3620\": rpc error: code = NotFound desc = could not find container \"f1640e8d1e0d18b376e1ca7bd82900817c5d88ddd06b34611c632e1e0fda3620\": container with ID starting with f1640e8d1e0d18b376e1ca7bd82900817c5d88ddd06b34611c632e1e0fda3620 not found: ID does not exist" Oct 14 10:26:43 crc kubenswrapper[4870]: I1014 10:26:43.891886 4870 scope.go:117] "RemoveContainer" containerID="6231f85ad4f94f3ab5320f32efc3bea0740e30cf00ee25c215164366a0429fbc" Oct 14 10:26:43 crc kubenswrapper[4870]: E1014 10:26:43.892175 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6231f85ad4f94f3ab5320f32efc3bea0740e30cf00ee25c215164366a0429fbc\": container with ID starting with 6231f85ad4f94f3ab5320f32efc3bea0740e30cf00ee25c215164366a0429fbc not found: ID does not exist" containerID="6231f85ad4f94f3ab5320f32efc3bea0740e30cf00ee25c215164366a0429fbc" Oct 14 10:26:43 crc kubenswrapper[4870]: I1014 10:26:43.892200 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6231f85ad4f94f3ab5320f32efc3bea0740e30cf00ee25c215164366a0429fbc"} err="failed to get container status \"6231f85ad4f94f3ab5320f32efc3bea0740e30cf00ee25c215164366a0429fbc\": rpc error: code = NotFound desc = could not find container \"6231f85ad4f94f3ab5320f32efc3bea0740e30cf00ee25c215164366a0429fbc\": container with ID starting with 6231f85ad4f94f3ab5320f32efc3bea0740e30cf00ee25c215164366a0429fbc not found: ID does not exist" Oct 14 10:26:45 crc kubenswrapper[4870]: I1014 10:26:45.052355 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aabbe5b0-7808-44e8-a9e7-83eaaf919511" path="/var/lib/kubelet/pods/aabbe5b0-7808-44e8-a9e7-83eaaf919511/volumes" Oct 14 10:26:47 crc kubenswrapper[4870]: I1014 10:26:47.814983 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-x9w5p" Oct 14 10:26:47 crc kubenswrapper[4870]: I1014 10:26:47.901222 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-x9w5p" Oct 14 10:26:48 crc kubenswrapper[4870]: I1014 10:26:48.060948 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-x9w5p"] Oct 14 10:26:49 crc kubenswrapper[4870]: I1014 10:26:49.829124 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-x9w5p" podUID="12a63494-1118-4826-a873-50fa5df427dd" containerName="registry-server" containerID="cri-o://f9dc8f5f83c29706a2525d37f680c3cfc5b2e36e404dcb9a0abe26ca2813634c" gracePeriod=2 Oct 14 10:26:50 crc kubenswrapper[4870]: I1014 10:26:50.483680 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-x9w5p" Oct 14 10:26:50 crc kubenswrapper[4870]: I1014 10:26:50.585567 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/12a63494-1118-4826-a873-50fa5df427dd-utilities\") pod \"12a63494-1118-4826-a873-50fa5df427dd\" (UID: \"12a63494-1118-4826-a873-50fa5df427dd\") " Oct 14 10:26:50 crc kubenswrapper[4870]: I1014 10:26:50.586299 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t5r2x\" (UniqueName: \"kubernetes.io/projected/12a63494-1118-4826-a873-50fa5df427dd-kube-api-access-t5r2x\") pod \"12a63494-1118-4826-a873-50fa5df427dd\" (UID: \"12a63494-1118-4826-a873-50fa5df427dd\") " Oct 14 10:26:50 crc kubenswrapper[4870]: I1014 10:26:50.586378 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/12a63494-1118-4826-a873-50fa5df427dd-catalog-content\") pod \"12a63494-1118-4826-a873-50fa5df427dd\" (UID: \"12a63494-1118-4826-a873-50fa5df427dd\") " Oct 14 10:26:50 crc kubenswrapper[4870]: I1014 10:26:50.587932 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/12a63494-1118-4826-a873-50fa5df427dd-utilities" (OuterVolumeSpecName: "utilities") pod "12a63494-1118-4826-a873-50fa5df427dd" (UID: "12a63494-1118-4826-a873-50fa5df427dd"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 10:26:50 crc kubenswrapper[4870]: I1014 10:26:50.593669 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/12a63494-1118-4826-a873-50fa5df427dd-kube-api-access-t5r2x" (OuterVolumeSpecName: "kube-api-access-t5r2x") pod "12a63494-1118-4826-a873-50fa5df427dd" (UID: "12a63494-1118-4826-a873-50fa5df427dd"). InnerVolumeSpecName "kube-api-access-t5r2x". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 10:26:50 crc kubenswrapper[4870]: I1014 10:26:50.664943 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/12a63494-1118-4826-a873-50fa5df427dd-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "12a63494-1118-4826-a873-50fa5df427dd" (UID: "12a63494-1118-4826-a873-50fa5df427dd"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 10:26:50 crc kubenswrapper[4870]: I1014 10:26:50.690023 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t5r2x\" (UniqueName: \"kubernetes.io/projected/12a63494-1118-4826-a873-50fa5df427dd-kube-api-access-t5r2x\") on node \"crc\" DevicePath \"\"" Oct 14 10:26:50 crc kubenswrapper[4870]: I1014 10:26:50.690066 4870 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/12a63494-1118-4826-a873-50fa5df427dd-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 10:26:50 crc kubenswrapper[4870]: I1014 10:26:50.690096 4870 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/12a63494-1118-4826-a873-50fa5df427dd-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 10:26:50 crc kubenswrapper[4870]: I1014 10:26:50.845481 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x9w5p" event={"ID":"12a63494-1118-4826-a873-50fa5df427dd","Type":"ContainerDied","Data":"f9dc8f5f83c29706a2525d37f680c3cfc5b2e36e404dcb9a0abe26ca2813634c"} Oct 14 10:26:50 crc kubenswrapper[4870]: I1014 10:26:50.845554 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-x9w5p" Oct 14 10:26:50 crc kubenswrapper[4870]: I1014 10:26:50.845574 4870 generic.go:334] "Generic (PLEG): container finished" podID="12a63494-1118-4826-a873-50fa5df427dd" containerID="f9dc8f5f83c29706a2525d37f680c3cfc5b2e36e404dcb9a0abe26ca2813634c" exitCode=0 Oct 14 10:26:50 crc kubenswrapper[4870]: I1014 10:26:50.845619 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x9w5p" event={"ID":"12a63494-1118-4826-a873-50fa5df427dd","Type":"ContainerDied","Data":"ed6bc777a5659c1ff6d0dec04b3fe62e5618ef47630d190ba628a9c1bd9f9fcb"} Oct 14 10:26:50 crc kubenswrapper[4870]: I1014 10:26:50.845560 4870 scope.go:117] "RemoveContainer" containerID="f9dc8f5f83c29706a2525d37f680c3cfc5b2e36e404dcb9a0abe26ca2813634c" Oct 14 10:26:50 crc kubenswrapper[4870]: I1014 10:26:50.899222 4870 scope.go:117] "RemoveContainer" containerID="45e04185ca3f57b7711f351786981d4eb326fb093a548cf0c974dfec3c17c9ad" Oct 14 10:26:50 crc kubenswrapper[4870]: I1014 10:26:50.920339 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-x9w5p"] Oct 14 10:26:50 crc kubenswrapper[4870]: I1014 10:26:50.938123 4870 scope.go:117] "RemoveContainer" containerID="9b0a94a44d27d209b611aefb7af8274da40abceb90270f6233764dfc8a0abe02" Oct 14 10:26:50 crc kubenswrapper[4870]: I1014 10:26:50.938180 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-x9w5p"] Oct 14 10:26:50 crc kubenswrapper[4870]: I1014 10:26:50.982983 4870 scope.go:117] "RemoveContainer" containerID="f9dc8f5f83c29706a2525d37f680c3cfc5b2e36e404dcb9a0abe26ca2813634c" Oct 14 10:26:50 crc kubenswrapper[4870]: E1014 10:26:50.983475 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f9dc8f5f83c29706a2525d37f680c3cfc5b2e36e404dcb9a0abe26ca2813634c\": container with ID starting with f9dc8f5f83c29706a2525d37f680c3cfc5b2e36e404dcb9a0abe26ca2813634c not found: ID does not exist" containerID="f9dc8f5f83c29706a2525d37f680c3cfc5b2e36e404dcb9a0abe26ca2813634c" Oct 14 10:26:50 crc kubenswrapper[4870]: I1014 10:26:50.983522 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f9dc8f5f83c29706a2525d37f680c3cfc5b2e36e404dcb9a0abe26ca2813634c"} err="failed to get container status \"f9dc8f5f83c29706a2525d37f680c3cfc5b2e36e404dcb9a0abe26ca2813634c\": rpc error: code = NotFound desc = could not find container \"f9dc8f5f83c29706a2525d37f680c3cfc5b2e36e404dcb9a0abe26ca2813634c\": container with ID starting with f9dc8f5f83c29706a2525d37f680c3cfc5b2e36e404dcb9a0abe26ca2813634c not found: ID does not exist" Oct 14 10:26:50 crc kubenswrapper[4870]: I1014 10:26:50.983549 4870 scope.go:117] "RemoveContainer" containerID="45e04185ca3f57b7711f351786981d4eb326fb093a548cf0c974dfec3c17c9ad" Oct 14 10:26:50 crc kubenswrapper[4870]: E1014 10:26:50.983813 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"45e04185ca3f57b7711f351786981d4eb326fb093a548cf0c974dfec3c17c9ad\": container with ID starting with 45e04185ca3f57b7711f351786981d4eb326fb093a548cf0c974dfec3c17c9ad not found: ID does not exist" containerID="45e04185ca3f57b7711f351786981d4eb326fb093a548cf0c974dfec3c17c9ad" Oct 14 10:26:50 crc kubenswrapper[4870]: I1014 10:26:50.983835 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"45e04185ca3f57b7711f351786981d4eb326fb093a548cf0c974dfec3c17c9ad"} err="failed to get container status \"45e04185ca3f57b7711f351786981d4eb326fb093a548cf0c974dfec3c17c9ad\": rpc error: code = NotFound desc = could not find container \"45e04185ca3f57b7711f351786981d4eb326fb093a548cf0c974dfec3c17c9ad\": container with ID starting with 45e04185ca3f57b7711f351786981d4eb326fb093a548cf0c974dfec3c17c9ad not found: ID does not exist" Oct 14 10:26:50 crc kubenswrapper[4870]: I1014 10:26:50.983849 4870 scope.go:117] "RemoveContainer" containerID="9b0a94a44d27d209b611aefb7af8274da40abceb90270f6233764dfc8a0abe02" Oct 14 10:26:50 crc kubenswrapper[4870]: E1014 10:26:50.984216 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9b0a94a44d27d209b611aefb7af8274da40abceb90270f6233764dfc8a0abe02\": container with ID starting with 9b0a94a44d27d209b611aefb7af8274da40abceb90270f6233764dfc8a0abe02 not found: ID does not exist" containerID="9b0a94a44d27d209b611aefb7af8274da40abceb90270f6233764dfc8a0abe02" Oct 14 10:26:50 crc kubenswrapper[4870]: I1014 10:26:50.984251 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9b0a94a44d27d209b611aefb7af8274da40abceb90270f6233764dfc8a0abe02"} err="failed to get container status \"9b0a94a44d27d209b611aefb7af8274da40abceb90270f6233764dfc8a0abe02\": rpc error: code = NotFound desc = could not find container \"9b0a94a44d27d209b611aefb7af8274da40abceb90270f6233764dfc8a0abe02\": container with ID starting with 9b0a94a44d27d209b611aefb7af8274da40abceb90270f6233764dfc8a0abe02 not found: ID does not exist" Oct 14 10:26:51 crc kubenswrapper[4870]: I1014 10:26:51.049636 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="12a63494-1118-4826-a873-50fa5df427dd" path="/var/lib/kubelet/pods/12a63494-1118-4826-a873-50fa5df427dd/volumes" Oct 14 10:27:38 crc kubenswrapper[4870]: I1014 10:27:38.494712 4870 generic.go:334] "Generic (PLEG): container finished" podID="2929a25a-a5e9-4a00-8b77-6939e3bf7d66" containerID="264b4c7a0a082d24909cc6d5594818be1076b75d69cff6b79e25a89988f46238" exitCode=1 Oct 14 10:27:38 crc kubenswrapper[4870]: I1014 10:27:38.494779 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"2929a25a-a5e9-4a00-8b77-6939e3bf7d66","Type":"ContainerDied","Data":"264b4c7a0a082d24909cc6d5594818be1076b75d69cff6b79e25a89988f46238"} Oct 14 10:27:40 crc kubenswrapper[4870]: I1014 10:27:40.052631 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Oct 14 10:27:40 crc kubenswrapper[4870]: I1014 10:27:40.202794 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"2929a25a-a5e9-4a00-8b77-6939e3bf7d66\" (UID: \"2929a25a-a5e9-4a00-8b77-6939e3bf7d66\") " Oct 14 10:27:40 crc kubenswrapper[4870]: I1014 10:27:40.203216 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vbxrd\" (UniqueName: \"kubernetes.io/projected/2929a25a-a5e9-4a00-8b77-6939e3bf7d66-kube-api-access-vbxrd\") pod \"2929a25a-a5e9-4a00-8b77-6939e3bf7d66\" (UID: \"2929a25a-a5e9-4a00-8b77-6939e3bf7d66\") " Oct 14 10:27:40 crc kubenswrapper[4870]: I1014 10:27:40.204022 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/2929a25a-a5e9-4a00-8b77-6939e3bf7d66-test-operator-ephemeral-workdir\") pod \"2929a25a-a5e9-4a00-8b77-6939e3bf7d66\" (UID: \"2929a25a-a5e9-4a00-8b77-6939e3bf7d66\") " Oct 14 10:27:40 crc kubenswrapper[4870]: I1014 10:27:40.204167 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2929a25a-a5e9-4a00-8b77-6939e3bf7d66-config-data\") pod \"2929a25a-a5e9-4a00-8b77-6939e3bf7d66\" (UID: \"2929a25a-a5e9-4a00-8b77-6939e3bf7d66\") " Oct 14 10:27:40 crc kubenswrapper[4870]: I1014 10:27:40.204286 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/2929a25a-a5e9-4a00-8b77-6939e3bf7d66-test-operator-ephemeral-temporary\") pod \"2929a25a-a5e9-4a00-8b77-6939e3bf7d66\" (UID: \"2929a25a-a5e9-4a00-8b77-6939e3bf7d66\") " Oct 14 10:27:40 crc kubenswrapper[4870]: I1014 10:27:40.204508 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/2929a25a-a5e9-4a00-8b77-6939e3bf7d66-openstack-config-secret\") pod \"2929a25a-a5e9-4a00-8b77-6939e3bf7d66\" (UID: \"2929a25a-a5e9-4a00-8b77-6939e3bf7d66\") " Oct 14 10:27:40 crc kubenswrapper[4870]: I1014 10:27:40.204637 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/2929a25a-a5e9-4a00-8b77-6939e3bf7d66-openstack-config\") pod \"2929a25a-a5e9-4a00-8b77-6939e3bf7d66\" (UID: \"2929a25a-a5e9-4a00-8b77-6939e3bf7d66\") " Oct 14 10:27:40 crc kubenswrapper[4870]: I1014 10:27:40.204751 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2929a25a-a5e9-4a00-8b77-6939e3bf7d66-ssh-key\") pod \"2929a25a-a5e9-4a00-8b77-6939e3bf7d66\" (UID: \"2929a25a-a5e9-4a00-8b77-6939e3bf7d66\") " Oct 14 10:27:40 crc kubenswrapper[4870]: I1014 10:27:40.204847 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/2929a25a-a5e9-4a00-8b77-6939e3bf7d66-ca-certs\") pod \"2929a25a-a5e9-4a00-8b77-6939e3bf7d66\" (UID: \"2929a25a-a5e9-4a00-8b77-6939e3bf7d66\") " Oct 14 10:27:40 crc kubenswrapper[4870]: I1014 10:27:40.207136 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2929a25a-a5e9-4a00-8b77-6939e3bf7d66-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "2929a25a-a5e9-4a00-8b77-6939e3bf7d66" (UID: "2929a25a-a5e9-4a00-8b77-6939e3bf7d66"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 10:27:40 crc kubenswrapper[4870]: I1014 10:27:40.208819 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2929a25a-a5e9-4a00-8b77-6939e3bf7d66-config-data" (OuterVolumeSpecName: "config-data") pod "2929a25a-a5e9-4a00-8b77-6939e3bf7d66" (UID: "2929a25a-a5e9-4a00-8b77-6939e3bf7d66"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 10:27:40 crc kubenswrapper[4870]: I1014 10:27:40.212243 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2929a25a-a5e9-4a00-8b77-6939e3bf7d66-kube-api-access-vbxrd" (OuterVolumeSpecName: "kube-api-access-vbxrd") pod "2929a25a-a5e9-4a00-8b77-6939e3bf7d66" (UID: "2929a25a-a5e9-4a00-8b77-6939e3bf7d66"). InnerVolumeSpecName "kube-api-access-vbxrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 10:27:40 crc kubenswrapper[4870]: I1014 10:27:40.213355 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage02-crc" (OuterVolumeSpecName: "test-operator-logs") pod "2929a25a-a5e9-4a00-8b77-6939e3bf7d66" (UID: "2929a25a-a5e9-4a00-8b77-6939e3bf7d66"). InnerVolumeSpecName "local-storage02-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 14 10:27:40 crc kubenswrapper[4870]: I1014 10:27:40.214251 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2929a25a-a5e9-4a00-8b77-6939e3bf7d66-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "2929a25a-a5e9-4a00-8b77-6939e3bf7d66" (UID: "2929a25a-a5e9-4a00-8b77-6939e3bf7d66"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 10:27:40 crc kubenswrapper[4870]: I1014 10:27:40.240611 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2929a25a-a5e9-4a00-8b77-6939e3bf7d66-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "2929a25a-a5e9-4a00-8b77-6939e3bf7d66" (UID: "2929a25a-a5e9-4a00-8b77-6939e3bf7d66"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 10:27:40 crc kubenswrapper[4870]: I1014 10:27:40.244815 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2929a25a-a5e9-4a00-8b77-6939e3bf7d66-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "2929a25a-a5e9-4a00-8b77-6939e3bf7d66" (UID: "2929a25a-a5e9-4a00-8b77-6939e3bf7d66"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 10:27:40 crc kubenswrapper[4870]: I1014 10:27:40.256678 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2929a25a-a5e9-4a00-8b77-6939e3bf7d66-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "2929a25a-a5e9-4a00-8b77-6939e3bf7d66" (UID: "2929a25a-a5e9-4a00-8b77-6939e3bf7d66"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 10:27:40 crc kubenswrapper[4870]: I1014 10:27:40.280245 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2929a25a-a5e9-4a00-8b77-6939e3bf7d66-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "2929a25a-a5e9-4a00-8b77-6939e3bf7d66" (UID: "2929a25a-a5e9-4a00-8b77-6939e3bf7d66"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 10:27:40 crc kubenswrapper[4870]: I1014 10:27:40.307125 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vbxrd\" (UniqueName: \"kubernetes.io/projected/2929a25a-a5e9-4a00-8b77-6939e3bf7d66-kube-api-access-vbxrd\") on node \"crc\" DevicePath \"\"" Oct 14 10:27:40 crc kubenswrapper[4870]: I1014 10:27:40.307157 4870 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/2929a25a-a5e9-4a00-8b77-6939e3bf7d66-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Oct 14 10:27:40 crc kubenswrapper[4870]: I1014 10:27:40.307168 4870 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2929a25a-a5e9-4a00-8b77-6939e3bf7d66-config-data\") on node \"crc\" DevicePath \"\"" Oct 14 10:27:40 crc kubenswrapper[4870]: I1014 10:27:40.307178 4870 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/2929a25a-a5e9-4a00-8b77-6939e3bf7d66-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Oct 14 10:27:40 crc kubenswrapper[4870]: I1014 10:27:40.307188 4870 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/2929a25a-a5e9-4a00-8b77-6939e3bf7d66-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Oct 14 10:27:40 crc kubenswrapper[4870]: I1014 10:27:40.307196 4870 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/2929a25a-a5e9-4a00-8b77-6939e3bf7d66-openstack-config\") on node \"crc\" DevicePath \"\"" Oct 14 10:27:40 crc kubenswrapper[4870]: I1014 10:27:40.307204 4870 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/2929a25a-a5e9-4a00-8b77-6939e3bf7d66-ca-certs\") on node \"crc\" DevicePath \"\"" Oct 14 10:27:40 crc kubenswrapper[4870]: I1014 10:27:40.307211 4870 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2929a25a-a5e9-4a00-8b77-6939e3bf7d66-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 14 10:27:40 crc kubenswrapper[4870]: I1014 10:27:40.307244 4870 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" " Oct 14 10:27:40 crc kubenswrapper[4870]: I1014 10:27:40.327623 4870 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage02-crc" (UniqueName: "kubernetes.io/local-volume/local-storage02-crc") on node "crc" Oct 14 10:27:40 crc kubenswrapper[4870]: I1014 10:27:40.409104 4870 reconciler_common.go:293] "Volume detached for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" DevicePath \"\"" Oct 14 10:27:40 crc kubenswrapper[4870]: I1014 10:27:40.527891 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"2929a25a-a5e9-4a00-8b77-6939e3bf7d66","Type":"ContainerDied","Data":"48b2bf1ed0fca4be2d9e44f216574a00cff664e4ae7a7e02de24a3197a1a84a0"} Oct 14 10:27:40 crc kubenswrapper[4870]: I1014 10:27:40.527955 4870 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="48b2bf1ed0fca4be2d9e44f216574a00cff664e4ae7a7e02de24a3197a1a84a0" Oct 14 10:27:40 crc kubenswrapper[4870]: I1014 10:27:40.528297 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Oct 14 10:27:44 crc kubenswrapper[4870]: I1014 10:27:44.390563 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Oct 14 10:27:44 crc kubenswrapper[4870]: E1014 10:27:44.391732 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46bd7024-9c7e-4906-a48c-6d6034c27bee" containerName="extract-utilities" Oct 14 10:27:44 crc kubenswrapper[4870]: I1014 10:27:44.391749 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="46bd7024-9c7e-4906-a48c-6d6034c27bee" containerName="extract-utilities" Oct 14 10:27:44 crc kubenswrapper[4870]: E1014 10:27:44.391768 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="12a63494-1118-4826-a873-50fa5df427dd" containerName="extract-content" Oct 14 10:27:44 crc kubenswrapper[4870]: I1014 10:27:44.391775 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="12a63494-1118-4826-a873-50fa5df427dd" containerName="extract-content" Oct 14 10:27:44 crc kubenswrapper[4870]: E1014 10:27:44.391794 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46bd7024-9c7e-4906-a48c-6d6034c27bee" containerName="registry-server" Oct 14 10:27:44 crc kubenswrapper[4870]: I1014 10:27:44.391802 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="46bd7024-9c7e-4906-a48c-6d6034c27bee" containerName="registry-server" Oct 14 10:27:44 crc kubenswrapper[4870]: E1014 10:27:44.391816 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aabbe5b0-7808-44e8-a9e7-83eaaf919511" containerName="extract-content" Oct 14 10:27:44 crc kubenswrapper[4870]: I1014 10:27:44.391826 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="aabbe5b0-7808-44e8-a9e7-83eaaf919511" containerName="extract-content" Oct 14 10:27:44 crc kubenswrapper[4870]: E1014 10:27:44.391842 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aabbe5b0-7808-44e8-a9e7-83eaaf919511" containerName="extract-utilities" Oct 14 10:27:44 crc kubenswrapper[4870]: I1014 10:27:44.391850 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="aabbe5b0-7808-44e8-a9e7-83eaaf919511" containerName="extract-utilities" Oct 14 10:27:44 crc kubenswrapper[4870]: E1014 10:27:44.391868 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aabbe5b0-7808-44e8-a9e7-83eaaf919511" containerName="registry-server" Oct 14 10:27:44 crc kubenswrapper[4870]: I1014 10:27:44.391881 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="aabbe5b0-7808-44e8-a9e7-83eaaf919511" containerName="registry-server" Oct 14 10:27:44 crc kubenswrapper[4870]: E1014 10:27:44.391905 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="12a63494-1118-4826-a873-50fa5df427dd" containerName="extract-utilities" Oct 14 10:27:44 crc kubenswrapper[4870]: I1014 10:27:44.391911 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="12a63494-1118-4826-a873-50fa5df427dd" containerName="extract-utilities" Oct 14 10:27:44 crc kubenswrapper[4870]: E1014 10:27:44.391925 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2929a25a-a5e9-4a00-8b77-6939e3bf7d66" containerName="tempest-tests-tempest-tests-runner" Oct 14 10:27:44 crc kubenswrapper[4870]: I1014 10:27:44.391931 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="2929a25a-a5e9-4a00-8b77-6939e3bf7d66" containerName="tempest-tests-tempest-tests-runner" Oct 14 10:27:44 crc kubenswrapper[4870]: E1014 10:27:44.391945 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="12a63494-1118-4826-a873-50fa5df427dd" containerName="registry-server" Oct 14 10:27:44 crc kubenswrapper[4870]: I1014 10:27:44.391951 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="12a63494-1118-4826-a873-50fa5df427dd" containerName="registry-server" Oct 14 10:27:44 crc kubenswrapper[4870]: E1014 10:27:44.391964 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46bd7024-9c7e-4906-a48c-6d6034c27bee" containerName="extract-content" Oct 14 10:27:44 crc kubenswrapper[4870]: I1014 10:27:44.391970 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="46bd7024-9c7e-4906-a48c-6d6034c27bee" containerName="extract-content" Oct 14 10:27:44 crc kubenswrapper[4870]: I1014 10:27:44.392215 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="2929a25a-a5e9-4a00-8b77-6939e3bf7d66" containerName="tempest-tests-tempest-tests-runner" Oct 14 10:27:44 crc kubenswrapper[4870]: I1014 10:27:44.392228 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="46bd7024-9c7e-4906-a48c-6d6034c27bee" containerName="registry-server" Oct 14 10:27:44 crc kubenswrapper[4870]: I1014 10:27:44.392241 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="12a63494-1118-4826-a873-50fa5df427dd" containerName="registry-server" Oct 14 10:27:44 crc kubenswrapper[4870]: I1014 10:27:44.392259 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="aabbe5b0-7808-44e8-a9e7-83eaaf919511" containerName="registry-server" Oct 14 10:27:44 crc kubenswrapper[4870]: I1014 10:27:44.393347 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 14 10:27:44 crc kubenswrapper[4870]: I1014 10:27:44.404119 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Oct 14 10:27:44 crc kubenswrapper[4870]: I1014 10:27:44.405055 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-vnjjw" Oct 14 10:27:44 crc kubenswrapper[4870]: I1014 10:27:44.518453 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"08ea5fd9-1afd-44de-ae26-61e52ca72e7a\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 14 10:27:44 crc kubenswrapper[4870]: I1014 10:27:44.518592 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fn2mk\" (UniqueName: \"kubernetes.io/projected/08ea5fd9-1afd-44de-ae26-61e52ca72e7a-kube-api-access-fn2mk\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"08ea5fd9-1afd-44de-ae26-61e52ca72e7a\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 14 10:27:44 crc kubenswrapper[4870]: I1014 10:27:44.621011 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fn2mk\" (UniqueName: \"kubernetes.io/projected/08ea5fd9-1afd-44de-ae26-61e52ca72e7a-kube-api-access-fn2mk\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"08ea5fd9-1afd-44de-ae26-61e52ca72e7a\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 14 10:27:44 crc kubenswrapper[4870]: I1014 10:27:44.621967 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"08ea5fd9-1afd-44de-ae26-61e52ca72e7a\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 14 10:27:44 crc kubenswrapper[4870]: I1014 10:27:44.622471 4870 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"08ea5fd9-1afd-44de-ae26-61e52ca72e7a\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 14 10:27:44 crc kubenswrapper[4870]: I1014 10:27:44.653219 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fn2mk\" (UniqueName: \"kubernetes.io/projected/08ea5fd9-1afd-44de-ae26-61e52ca72e7a-kube-api-access-fn2mk\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"08ea5fd9-1afd-44de-ae26-61e52ca72e7a\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 14 10:27:44 crc kubenswrapper[4870]: I1014 10:27:44.666409 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"08ea5fd9-1afd-44de-ae26-61e52ca72e7a\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 14 10:27:44 crc kubenswrapper[4870]: I1014 10:27:44.728978 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 14 10:27:45 crc kubenswrapper[4870]: I1014 10:27:45.267187 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Oct 14 10:27:45 crc kubenswrapper[4870]: I1014 10:27:45.595189 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"08ea5fd9-1afd-44de-ae26-61e52ca72e7a","Type":"ContainerStarted","Data":"dc097760972a94fba62b8b3f35d57ddc92808824a0cf487febf1760a75313356"} Oct 14 10:27:46 crc kubenswrapper[4870]: I1014 10:27:46.607756 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"08ea5fd9-1afd-44de-ae26-61e52ca72e7a","Type":"ContainerStarted","Data":"9b708bfe48f51016f6e65f3f101f88665ddf1561bab0cb5fc7212cd80699797e"} Oct 14 10:27:46 crc kubenswrapper[4870]: I1014 10:27:46.622910 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" podStartSLOduration=1.705443749 podStartE2EDuration="2.622884146s" podCreationTimestamp="2025-10-14 10:27:44 +0000 UTC" firstStartedPulling="2025-10-14 10:27:45.284040407 +0000 UTC m=+12400.981400818" lastFinishedPulling="2025-10-14 10:27:46.201480834 +0000 UTC m=+12401.898841215" observedRunningTime="2025-10-14 10:27:46.622106077 +0000 UTC m=+12402.319466458" watchObservedRunningTime="2025-10-14 10:27:46.622884146 +0000 UTC m=+12402.320244527" Oct 14 10:28:23 crc kubenswrapper[4870]: I1014 10:28:23.950986 4870 patch_prober.go:28] interesting pod/machine-config-daemon-7tvc7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 10:28:23 crc kubenswrapper[4870]: I1014 10:28:23.951613 4870 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 10:28:48 crc kubenswrapper[4870]: I1014 10:28:48.660784 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-zlss5/must-gather-p4ml9"] Oct 14 10:28:48 crc kubenswrapper[4870]: I1014 10:28:48.670341 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-zlss5/must-gather-p4ml9" Oct 14 10:28:48 crc kubenswrapper[4870]: I1014 10:28:48.672991 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-zlss5"/"default-dockercfg-hx82p" Oct 14 10:28:48 crc kubenswrapper[4870]: I1014 10:28:48.673217 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-zlss5"/"openshift-service-ca.crt" Oct 14 10:28:48 crc kubenswrapper[4870]: I1014 10:28:48.674411 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-zlss5/must-gather-p4ml9"] Oct 14 10:28:48 crc kubenswrapper[4870]: I1014 10:28:48.675388 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-zlss5"/"kube-root-ca.crt" Oct 14 10:28:48 crc kubenswrapper[4870]: I1014 10:28:48.749741 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/e78b6d95-1398-4f8a-9565-d64feb722dd0-must-gather-output\") pod \"must-gather-p4ml9\" (UID: \"e78b6d95-1398-4f8a-9565-d64feb722dd0\") " pod="openshift-must-gather-zlss5/must-gather-p4ml9" Oct 14 10:28:48 crc kubenswrapper[4870]: I1014 10:28:48.749935 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2tn5k\" (UniqueName: \"kubernetes.io/projected/e78b6d95-1398-4f8a-9565-d64feb722dd0-kube-api-access-2tn5k\") pod \"must-gather-p4ml9\" (UID: \"e78b6d95-1398-4f8a-9565-d64feb722dd0\") " pod="openshift-must-gather-zlss5/must-gather-p4ml9" Oct 14 10:28:48 crc kubenswrapper[4870]: I1014 10:28:48.851944 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/e78b6d95-1398-4f8a-9565-d64feb722dd0-must-gather-output\") pod \"must-gather-p4ml9\" (UID: \"e78b6d95-1398-4f8a-9565-d64feb722dd0\") " pod="openshift-must-gather-zlss5/must-gather-p4ml9" Oct 14 10:28:48 crc kubenswrapper[4870]: I1014 10:28:48.852069 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2tn5k\" (UniqueName: \"kubernetes.io/projected/e78b6d95-1398-4f8a-9565-d64feb722dd0-kube-api-access-2tn5k\") pod \"must-gather-p4ml9\" (UID: \"e78b6d95-1398-4f8a-9565-d64feb722dd0\") " pod="openshift-must-gather-zlss5/must-gather-p4ml9" Oct 14 10:28:48 crc kubenswrapper[4870]: I1014 10:28:48.852783 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/e78b6d95-1398-4f8a-9565-d64feb722dd0-must-gather-output\") pod \"must-gather-p4ml9\" (UID: \"e78b6d95-1398-4f8a-9565-d64feb722dd0\") " pod="openshift-must-gather-zlss5/must-gather-p4ml9" Oct 14 10:28:48 crc kubenswrapper[4870]: I1014 10:28:48.875981 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2tn5k\" (UniqueName: \"kubernetes.io/projected/e78b6d95-1398-4f8a-9565-d64feb722dd0-kube-api-access-2tn5k\") pod \"must-gather-p4ml9\" (UID: \"e78b6d95-1398-4f8a-9565-d64feb722dd0\") " pod="openshift-must-gather-zlss5/must-gather-p4ml9" Oct 14 10:28:48 crc kubenswrapper[4870]: I1014 10:28:48.990354 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-zlss5/must-gather-p4ml9" Oct 14 10:28:49 crc kubenswrapper[4870]: I1014 10:28:49.526038 4870 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 14 10:28:49 crc kubenswrapper[4870]: I1014 10:28:49.537772 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-zlss5/must-gather-p4ml9"] Oct 14 10:28:50 crc kubenswrapper[4870]: I1014 10:28:50.552533 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-zlss5/must-gather-p4ml9" event={"ID":"e78b6d95-1398-4f8a-9565-d64feb722dd0","Type":"ContainerStarted","Data":"9ab23d766378d391c83395cfa0663c079eb18d5f963b7fa6bebfa915693be161"} Oct 14 10:28:53 crc kubenswrapper[4870]: I1014 10:28:53.951613 4870 patch_prober.go:28] interesting pod/machine-config-daemon-7tvc7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 10:28:53 crc kubenswrapper[4870]: I1014 10:28:53.952043 4870 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 10:28:55 crc kubenswrapper[4870]: I1014 10:28:55.612936 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-zlss5/must-gather-p4ml9" event={"ID":"e78b6d95-1398-4f8a-9565-d64feb722dd0","Type":"ContainerStarted","Data":"2ac83eeacc2ecd24b9221a6f13d502525efc9fb0101a534cfb28db48dd794f3d"} Oct 14 10:28:55 crc kubenswrapper[4870]: I1014 10:28:55.613696 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-zlss5/must-gather-p4ml9" event={"ID":"e78b6d95-1398-4f8a-9565-d64feb722dd0","Type":"ContainerStarted","Data":"120f8b1146a59099dee946943f0c73cf49cdb3e20248bb24290f386d7aecad21"} Oct 14 10:28:55 crc kubenswrapper[4870]: I1014 10:28:55.629733 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-zlss5/must-gather-p4ml9" podStartSLOduration=2.657681347 podStartE2EDuration="7.629713791s" podCreationTimestamp="2025-10-14 10:28:48 +0000 UTC" firstStartedPulling="2025-10-14 10:28:49.525263434 +0000 UTC m=+12465.222623825" lastFinishedPulling="2025-10-14 10:28:54.497295868 +0000 UTC m=+12470.194656269" observedRunningTime="2025-10-14 10:28:55.626252186 +0000 UTC m=+12471.323612557" watchObservedRunningTime="2025-10-14 10:28:55.629713791 +0000 UTC m=+12471.327074162" Oct 14 10:29:00 crc kubenswrapper[4870]: I1014 10:29:00.394398 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-zlss5/crc-debug-vdmzq"] Oct 14 10:29:00 crc kubenswrapper[4870]: I1014 10:29:00.396450 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-zlss5/crc-debug-vdmzq" Oct 14 10:29:00 crc kubenswrapper[4870]: I1014 10:29:00.472997 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xlkbw\" (UniqueName: \"kubernetes.io/projected/eff1e73e-2f51-4204-bf14-31099277d4c0-kube-api-access-xlkbw\") pod \"crc-debug-vdmzq\" (UID: \"eff1e73e-2f51-4204-bf14-31099277d4c0\") " pod="openshift-must-gather-zlss5/crc-debug-vdmzq" Oct 14 10:29:00 crc kubenswrapper[4870]: I1014 10:29:00.473111 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/eff1e73e-2f51-4204-bf14-31099277d4c0-host\") pod \"crc-debug-vdmzq\" (UID: \"eff1e73e-2f51-4204-bf14-31099277d4c0\") " pod="openshift-must-gather-zlss5/crc-debug-vdmzq" Oct 14 10:29:00 crc kubenswrapper[4870]: I1014 10:29:00.574489 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xlkbw\" (UniqueName: \"kubernetes.io/projected/eff1e73e-2f51-4204-bf14-31099277d4c0-kube-api-access-xlkbw\") pod \"crc-debug-vdmzq\" (UID: \"eff1e73e-2f51-4204-bf14-31099277d4c0\") " pod="openshift-must-gather-zlss5/crc-debug-vdmzq" Oct 14 10:29:00 crc kubenswrapper[4870]: I1014 10:29:00.574613 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/eff1e73e-2f51-4204-bf14-31099277d4c0-host\") pod \"crc-debug-vdmzq\" (UID: \"eff1e73e-2f51-4204-bf14-31099277d4c0\") " pod="openshift-must-gather-zlss5/crc-debug-vdmzq" Oct 14 10:29:00 crc kubenswrapper[4870]: I1014 10:29:00.574761 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/eff1e73e-2f51-4204-bf14-31099277d4c0-host\") pod \"crc-debug-vdmzq\" (UID: \"eff1e73e-2f51-4204-bf14-31099277d4c0\") " pod="openshift-must-gather-zlss5/crc-debug-vdmzq" Oct 14 10:29:00 crc kubenswrapper[4870]: I1014 10:29:00.597624 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xlkbw\" (UniqueName: \"kubernetes.io/projected/eff1e73e-2f51-4204-bf14-31099277d4c0-kube-api-access-xlkbw\") pod \"crc-debug-vdmzq\" (UID: \"eff1e73e-2f51-4204-bf14-31099277d4c0\") " pod="openshift-must-gather-zlss5/crc-debug-vdmzq" Oct 14 10:29:00 crc kubenswrapper[4870]: I1014 10:29:00.714951 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-zlss5/crc-debug-vdmzq" Oct 14 10:29:00 crc kubenswrapper[4870]: W1014 10:29:00.772081 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podeff1e73e_2f51_4204_bf14_31099277d4c0.slice/crio-6f757a4e45e24edf6bcd7549006a4faf2e6abd811d8114acca5221e03fe2de77 WatchSource:0}: Error finding container 6f757a4e45e24edf6bcd7549006a4faf2e6abd811d8114acca5221e03fe2de77: Status 404 returned error can't find the container with id 6f757a4e45e24edf6bcd7549006a4faf2e6abd811d8114acca5221e03fe2de77 Oct 14 10:29:01 crc kubenswrapper[4870]: I1014 10:29:01.678798 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-zlss5/crc-debug-vdmzq" event={"ID":"eff1e73e-2f51-4204-bf14-31099277d4c0","Type":"ContainerStarted","Data":"6f757a4e45e24edf6bcd7549006a4faf2e6abd811d8114acca5221e03fe2de77"} Oct 14 10:29:09 crc kubenswrapper[4870]: I1014 10:29:09.819258 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-zlss5/crc-debug-vdmzq" event={"ID":"eff1e73e-2f51-4204-bf14-31099277d4c0","Type":"ContainerStarted","Data":"51d984c3b838f64ad984d3814421456cee1b210d1a66d555471d1facdedc3493"} Oct 14 10:29:09 crc kubenswrapper[4870]: I1014 10:29:09.842858 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-zlss5/crc-debug-vdmzq" podStartSLOduration=1.324965743 podStartE2EDuration="9.841976587s" podCreationTimestamp="2025-10-14 10:29:00 +0000 UTC" firstStartedPulling="2025-10-14 10:29:00.774747627 +0000 UTC m=+12476.472108018" lastFinishedPulling="2025-10-14 10:29:09.291758501 +0000 UTC m=+12484.989118862" observedRunningTime="2025-10-14 10:29:09.833309996 +0000 UTC m=+12485.530670367" watchObservedRunningTime="2025-10-14 10:29:09.841976587 +0000 UTC m=+12485.539336958" Oct 14 10:29:23 crc kubenswrapper[4870]: I1014 10:29:23.950702 4870 patch_prober.go:28] interesting pod/machine-config-daemon-7tvc7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 10:29:23 crc kubenswrapper[4870]: I1014 10:29:23.951327 4870 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 10:29:23 crc kubenswrapper[4870]: I1014 10:29:23.951387 4870 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" Oct 14 10:29:23 crc kubenswrapper[4870]: I1014 10:29:23.952371 4870 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"3a25412c639a58b6345a6cb926f93c78a7a62c68cd923f500f921497a90b6efb"} pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 14 10:29:23 crc kubenswrapper[4870]: I1014 10:29:23.952468 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerName="machine-config-daemon" containerID="cri-o://3a25412c639a58b6345a6cb926f93c78a7a62c68cd923f500f921497a90b6efb" gracePeriod=600 Oct 14 10:29:24 crc kubenswrapper[4870]: E1014 10:29:24.107164 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 10:29:25 crc kubenswrapper[4870]: I1014 10:29:25.018839 4870 generic.go:334] "Generic (PLEG): container finished" podID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerID="3a25412c639a58b6345a6cb926f93c78a7a62c68cd923f500f921497a90b6efb" exitCode=0 Oct 14 10:29:25 crc kubenswrapper[4870]: I1014 10:29:25.018944 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" event={"ID":"874a5e45-dffd-4d17-b609-b7d3ed2eab07","Type":"ContainerDied","Data":"3a25412c639a58b6345a6cb926f93c78a7a62c68cd923f500f921497a90b6efb"} Oct 14 10:29:25 crc kubenswrapper[4870]: I1014 10:29:25.019714 4870 scope.go:117] "RemoveContainer" containerID="4e0c842e7d29f428c954b866c1f3929be48b4bb5118abceba1fbf69beed50855" Oct 14 10:29:25 crc kubenswrapper[4870]: I1014 10:29:25.020425 4870 scope.go:117] "RemoveContainer" containerID="3a25412c639a58b6345a6cb926f93c78a7a62c68cd923f500f921497a90b6efb" Oct 14 10:29:25 crc kubenswrapper[4870]: E1014 10:29:25.020858 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 10:29:31 crc kubenswrapper[4870]: I1014 10:29:31.145692 4870 generic.go:334] "Generic (PLEG): container finished" podID="eff1e73e-2f51-4204-bf14-31099277d4c0" containerID="51d984c3b838f64ad984d3814421456cee1b210d1a66d555471d1facdedc3493" exitCode=0 Oct 14 10:29:31 crc kubenswrapper[4870]: I1014 10:29:31.145781 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-zlss5/crc-debug-vdmzq" event={"ID":"eff1e73e-2f51-4204-bf14-31099277d4c0","Type":"ContainerDied","Data":"51d984c3b838f64ad984d3814421456cee1b210d1a66d555471d1facdedc3493"} Oct 14 10:29:32 crc kubenswrapper[4870]: I1014 10:29:32.278179 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-zlss5/crc-debug-vdmzq" Oct 14 10:29:32 crc kubenswrapper[4870]: I1014 10:29:32.315016 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-zlss5/crc-debug-vdmzq"] Oct 14 10:29:32 crc kubenswrapper[4870]: I1014 10:29:32.325230 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-zlss5/crc-debug-vdmzq"] Oct 14 10:29:32 crc kubenswrapper[4870]: I1014 10:29:32.379181 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/eff1e73e-2f51-4204-bf14-31099277d4c0-host\") pod \"eff1e73e-2f51-4204-bf14-31099277d4c0\" (UID: \"eff1e73e-2f51-4204-bf14-31099277d4c0\") " Oct 14 10:29:32 crc kubenswrapper[4870]: I1014 10:29:32.379398 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xlkbw\" (UniqueName: \"kubernetes.io/projected/eff1e73e-2f51-4204-bf14-31099277d4c0-kube-api-access-xlkbw\") pod \"eff1e73e-2f51-4204-bf14-31099277d4c0\" (UID: \"eff1e73e-2f51-4204-bf14-31099277d4c0\") " Oct 14 10:29:32 crc kubenswrapper[4870]: I1014 10:29:32.380541 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/eff1e73e-2f51-4204-bf14-31099277d4c0-host" (OuterVolumeSpecName: "host") pod "eff1e73e-2f51-4204-bf14-31099277d4c0" (UID: "eff1e73e-2f51-4204-bf14-31099277d4c0"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 14 10:29:32 crc kubenswrapper[4870]: I1014 10:29:32.387695 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eff1e73e-2f51-4204-bf14-31099277d4c0-kube-api-access-xlkbw" (OuterVolumeSpecName: "kube-api-access-xlkbw") pod "eff1e73e-2f51-4204-bf14-31099277d4c0" (UID: "eff1e73e-2f51-4204-bf14-31099277d4c0"). InnerVolumeSpecName "kube-api-access-xlkbw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 10:29:32 crc kubenswrapper[4870]: I1014 10:29:32.482047 4870 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/eff1e73e-2f51-4204-bf14-31099277d4c0-host\") on node \"crc\" DevicePath \"\"" Oct 14 10:29:32 crc kubenswrapper[4870]: I1014 10:29:32.482370 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xlkbw\" (UniqueName: \"kubernetes.io/projected/eff1e73e-2f51-4204-bf14-31099277d4c0-kube-api-access-xlkbw\") on node \"crc\" DevicePath \"\"" Oct 14 10:29:33 crc kubenswrapper[4870]: I1014 10:29:33.057220 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eff1e73e-2f51-4204-bf14-31099277d4c0" path="/var/lib/kubelet/pods/eff1e73e-2f51-4204-bf14-31099277d4c0/volumes" Oct 14 10:29:33 crc kubenswrapper[4870]: I1014 10:29:33.169780 4870 scope.go:117] "RemoveContainer" containerID="51d984c3b838f64ad984d3814421456cee1b210d1a66d555471d1facdedc3493" Oct 14 10:29:33 crc kubenswrapper[4870]: I1014 10:29:33.169821 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-zlss5/crc-debug-vdmzq" Oct 14 10:29:33 crc kubenswrapper[4870]: I1014 10:29:33.582966 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-zlss5/crc-debug-t6vc4"] Oct 14 10:29:33 crc kubenswrapper[4870]: E1014 10:29:33.584066 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eff1e73e-2f51-4204-bf14-31099277d4c0" containerName="container-00" Oct 14 10:29:33 crc kubenswrapper[4870]: I1014 10:29:33.584085 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="eff1e73e-2f51-4204-bf14-31099277d4c0" containerName="container-00" Oct 14 10:29:33 crc kubenswrapper[4870]: I1014 10:29:33.584584 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="eff1e73e-2f51-4204-bf14-31099277d4c0" containerName="container-00" Oct 14 10:29:33 crc kubenswrapper[4870]: I1014 10:29:33.586037 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-zlss5/crc-debug-t6vc4" Oct 14 10:29:33 crc kubenswrapper[4870]: I1014 10:29:33.711132 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/22ac2f11-3574-4219-b555-f550ad652904-host\") pod \"crc-debug-t6vc4\" (UID: \"22ac2f11-3574-4219-b555-f550ad652904\") " pod="openshift-must-gather-zlss5/crc-debug-t6vc4" Oct 14 10:29:33 crc kubenswrapper[4870]: I1014 10:29:33.711178 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t8br8\" (UniqueName: \"kubernetes.io/projected/22ac2f11-3574-4219-b555-f550ad652904-kube-api-access-t8br8\") pod \"crc-debug-t6vc4\" (UID: \"22ac2f11-3574-4219-b555-f550ad652904\") " pod="openshift-must-gather-zlss5/crc-debug-t6vc4" Oct 14 10:29:33 crc kubenswrapper[4870]: I1014 10:29:33.812773 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/22ac2f11-3574-4219-b555-f550ad652904-host\") pod \"crc-debug-t6vc4\" (UID: \"22ac2f11-3574-4219-b555-f550ad652904\") " pod="openshift-must-gather-zlss5/crc-debug-t6vc4" Oct 14 10:29:33 crc kubenswrapper[4870]: I1014 10:29:33.812833 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t8br8\" (UniqueName: \"kubernetes.io/projected/22ac2f11-3574-4219-b555-f550ad652904-kube-api-access-t8br8\") pod \"crc-debug-t6vc4\" (UID: \"22ac2f11-3574-4219-b555-f550ad652904\") " pod="openshift-must-gather-zlss5/crc-debug-t6vc4" Oct 14 10:29:33 crc kubenswrapper[4870]: I1014 10:29:33.813029 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/22ac2f11-3574-4219-b555-f550ad652904-host\") pod \"crc-debug-t6vc4\" (UID: \"22ac2f11-3574-4219-b555-f550ad652904\") " pod="openshift-must-gather-zlss5/crc-debug-t6vc4" Oct 14 10:29:33 crc kubenswrapper[4870]: I1014 10:29:33.847225 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t8br8\" (UniqueName: \"kubernetes.io/projected/22ac2f11-3574-4219-b555-f550ad652904-kube-api-access-t8br8\") pod \"crc-debug-t6vc4\" (UID: \"22ac2f11-3574-4219-b555-f550ad652904\") " pod="openshift-must-gather-zlss5/crc-debug-t6vc4" Oct 14 10:29:33 crc kubenswrapper[4870]: I1014 10:29:33.902812 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-zlss5/crc-debug-t6vc4" Oct 14 10:29:34 crc kubenswrapper[4870]: I1014 10:29:34.190153 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-zlss5/crc-debug-t6vc4" event={"ID":"22ac2f11-3574-4219-b555-f550ad652904","Type":"ContainerStarted","Data":"8fadf44427ea2d2b7717c50d77a5f5407f5c05fea6433e75771276be8c659044"} Oct 14 10:29:35 crc kubenswrapper[4870]: I1014 10:29:35.208160 4870 generic.go:334] "Generic (PLEG): container finished" podID="22ac2f11-3574-4219-b555-f550ad652904" containerID="dbb97183cd4380a64764bf9a391948edfd49687016aa664d90f68d1d5e8c6884" exitCode=1 Oct 14 10:29:35 crc kubenswrapper[4870]: I1014 10:29:35.208222 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-zlss5/crc-debug-t6vc4" event={"ID":"22ac2f11-3574-4219-b555-f550ad652904","Type":"ContainerDied","Data":"dbb97183cd4380a64764bf9a391948edfd49687016aa664d90f68d1d5e8c6884"} Oct 14 10:29:35 crc kubenswrapper[4870]: I1014 10:29:35.267076 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-zlss5/crc-debug-t6vc4"] Oct 14 10:29:35 crc kubenswrapper[4870]: I1014 10:29:35.288814 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-zlss5/crc-debug-t6vc4"] Oct 14 10:29:36 crc kubenswrapper[4870]: I1014 10:29:36.344112 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-zlss5/crc-debug-t6vc4" Oct 14 10:29:36 crc kubenswrapper[4870]: I1014 10:29:36.379134 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/22ac2f11-3574-4219-b555-f550ad652904-host\") pod \"22ac2f11-3574-4219-b555-f550ad652904\" (UID: \"22ac2f11-3574-4219-b555-f550ad652904\") " Oct 14 10:29:36 crc kubenswrapper[4870]: I1014 10:29:36.379272 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/22ac2f11-3574-4219-b555-f550ad652904-host" (OuterVolumeSpecName: "host") pod "22ac2f11-3574-4219-b555-f550ad652904" (UID: "22ac2f11-3574-4219-b555-f550ad652904"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 14 10:29:36 crc kubenswrapper[4870]: I1014 10:29:36.379290 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t8br8\" (UniqueName: \"kubernetes.io/projected/22ac2f11-3574-4219-b555-f550ad652904-kube-api-access-t8br8\") pod \"22ac2f11-3574-4219-b555-f550ad652904\" (UID: \"22ac2f11-3574-4219-b555-f550ad652904\") " Oct 14 10:29:36 crc kubenswrapper[4870]: I1014 10:29:36.380406 4870 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/22ac2f11-3574-4219-b555-f550ad652904-host\") on node \"crc\" DevicePath \"\"" Oct 14 10:29:36 crc kubenswrapper[4870]: I1014 10:29:36.386590 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22ac2f11-3574-4219-b555-f550ad652904-kube-api-access-t8br8" (OuterVolumeSpecName: "kube-api-access-t8br8") pod "22ac2f11-3574-4219-b555-f550ad652904" (UID: "22ac2f11-3574-4219-b555-f550ad652904"). InnerVolumeSpecName "kube-api-access-t8br8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 10:29:36 crc kubenswrapper[4870]: I1014 10:29:36.481705 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t8br8\" (UniqueName: \"kubernetes.io/projected/22ac2f11-3574-4219-b555-f550ad652904-kube-api-access-t8br8\") on node \"crc\" DevicePath \"\"" Oct 14 10:29:37 crc kubenswrapper[4870]: I1014 10:29:37.056798 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22ac2f11-3574-4219-b555-f550ad652904" path="/var/lib/kubelet/pods/22ac2f11-3574-4219-b555-f550ad652904/volumes" Oct 14 10:29:37 crc kubenswrapper[4870]: I1014 10:29:37.242260 4870 scope.go:117] "RemoveContainer" containerID="dbb97183cd4380a64764bf9a391948edfd49687016aa664d90f68d1d5e8c6884" Oct 14 10:29:37 crc kubenswrapper[4870]: I1014 10:29:37.242405 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-zlss5/crc-debug-t6vc4" Oct 14 10:29:38 crc kubenswrapper[4870]: I1014 10:29:38.034243 4870 scope.go:117] "RemoveContainer" containerID="3a25412c639a58b6345a6cb926f93c78a7a62c68cd923f500f921497a90b6efb" Oct 14 10:29:38 crc kubenswrapper[4870]: E1014 10:29:38.034919 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 10:29:49 crc kubenswrapper[4870]: I1014 10:29:49.039276 4870 scope.go:117] "RemoveContainer" containerID="3a25412c639a58b6345a6cb926f93c78a7a62c68cd923f500f921497a90b6efb" Oct 14 10:29:49 crc kubenswrapper[4870]: E1014 10:29:49.045077 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 10:30:00 crc kubenswrapper[4870]: I1014 10:30:00.183827 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29340630-dw5fj"] Oct 14 10:30:00 crc kubenswrapper[4870]: E1014 10:30:00.184880 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="22ac2f11-3574-4219-b555-f550ad652904" containerName="container-00" Oct 14 10:30:00 crc kubenswrapper[4870]: I1014 10:30:00.184896 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="22ac2f11-3574-4219-b555-f550ad652904" containerName="container-00" Oct 14 10:30:00 crc kubenswrapper[4870]: I1014 10:30:00.185165 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="22ac2f11-3574-4219-b555-f550ad652904" containerName="container-00" Oct 14 10:30:00 crc kubenswrapper[4870]: I1014 10:30:00.186161 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29340630-dw5fj" Oct 14 10:30:00 crc kubenswrapper[4870]: I1014 10:30:00.188266 4870 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 14 10:30:00 crc kubenswrapper[4870]: I1014 10:30:00.188765 4870 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 14 10:30:00 crc kubenswrapper[4870]: I1014 10:30:00.212581 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29340630-dw5fj"] Oct 14 10:30:00 crc kubenswrapper[4870]: I1014 10:30:00.317952 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/31e995c2-04e9-49d3-a47d-a771beee2a10-secret-volume\") pod \"collect-profiles-29340630-dw5fj\" (UID: \"31e995c2-04e9-49d3-a47d-a771beee2a10\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340630-dw5fj" Oct 14 10:30:00 crc kubenswrapper[4870]: I1014 10:30:00.318192 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/31e995c2-04e9-49d3-a47d-a771beee2a10-config-volume\") pod \"collect-profiles-29340630-dw5fj\" (UID: \"31e995c2-04e9-49d3-a47d-a771beee2a10\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340630-dw5fj" Oct 14 10:30:00 crc kubenswrapper[4870]: I1014 10:30:00.318337 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gvh6f\" (UniqueName: \"kubernetes.io/projected/31e995c2-04e9-49d3-a47d-a771beee2a10-kube-api-access-gvh6f\") pod \"collect-profiles-29340630-dw5fj\" (UID: \"31e995c2-04e9-49d3-a47d-a771beee2a10\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340630-dw5fj" Oct 14 10:30:00 crc kubenswrapper[4870]: I1014 10:30:00.420573 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/31e995c2-04e9-49d3-a47d-a771beee2a10-config-volume\") pod \"collect-profiles-29340630-dw5fj\" (UID: \"31e995c2-04e9-49d3-a47d-a771beee2a10\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340630-dw5fj" Oct 14 10:30:00 crc kubenswrapper[4870]: I1014 10:30:00.420682 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gvh6f\" (UniqueName: \"kubernetes.io/projected/31e995c2-04e9-49d3-a47d-a771beee2a10-kube-api-access-gvh6f\") pod \"collect-profiles-29340630-dw5fj\" (UID: \"31e995c2-04e9-49d3-a47d-a771beee2a10\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340630-dw5fj" Oct 14 10:30:00 crc kubenswrapper[4870]: I1014 10:30:00.420825 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/31e995c2-04e9-49d3-a47d-a771beee2a10-secret-volume\") pod \"collect-profiles-29340630-dw5fj\" (UID: \"31e995c2-04e9-49d3-a47d-a771beee2a10\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340630-dw5fj" Oct 14 10:30:00 crc kubenswrapper[4870]: I1014 10:30:00.422400 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/31e995c2-04e9-49d3-a47d-a771beee2a10-config-volume\") pod \"collect-profiles-29340630-dw5fj\" (UID: \"31e995c2-04e9-49d3-a47d-a771beee2a10\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340630-dw5fj" Oct 14 10:30:00 crc kubenswrapper[4870]: I1014 10:30:00.428537 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/31e995c2-04e9-49d3-a47d-a771beee2a10-secret-volume\") pod \"collect-profiles-29340630-dw5fj\" (UID: \"31e995c2-04e9-49d3-a47d-a771beee2a10\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340630-dw5fj" Oct 14 10:30:00 crc kubenswrapper[4870]: I1014 10:30:00.442718 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gvh6f\" (UniqueName: \"kubernetes.io/projected/31e995c2-04e9-49d3-a47d-a771beee2a10-kube-api-access-gvh6f\") pod \"collect-profiles-29340630-dw5fj\" (UID: \"31e995c2-04e9-49d3-a47d-a771beee2a10\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29340630-dw5fj" Oct 14 10:30:00 crc kubenswrapper[4870]: I1014 10:30:00.505741 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29340630-dw5fj" Oct 14 10:30:00 crc kubenswrapper[4870]: I1014 10:30:00.949490 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29340630-dw5fj"] Oct 14 10:30:00 crc kubenswrapper[4870]: W1014 10:30:00.956585 4870 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod31e995c2_04e9_49d3_a47d_a771beee2a10.slice/crio-97d526a89a7d74ba6c90660fc5aef445431fcd2ff6479f9e01060bb40b611383 WatchSource:0}: Error finding container 97d526a89a7d74ba6c90660fc5aef445431fcd2ff6479f9e01060bb40b611383: Status 404 returned error can't find the container with id 97d526a89a7d74ba6c90660fc5aef445431fcd2ff6479f9e01060bb40b611383 Oct 14 10:30:01 crc kubenswrapper[4870]: I1014 10:30:01.544399 4870 generic.go:334] "Generic (PLEG): container finished" podID="31e995c2-04e9-49d3-a47d-a771beee2a10" containerID="aad2008191d2ed404d60762f3646d5f0c7532478dce28d0601b42a0b262c57ba" exitCode=0 Oct 14 10:30:01 crc kubenswrapper[4870]: I1014 10:30:01.544530 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29340630-dw5fj" event={"ID":"31e995c2-04e9-49d3-a47d-a771beee2a10","Type":"ContainerDied","Data":"aad2008191d2ed404d60762f3646d5f0c7532478dce28d0601b42a0b262c57ba"} Oct 14 10:30:01 crc kubenswrapper[4870]: I1014 10:30:01.544961 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29340630-dw5fj" event={"ID":"31e995c2-04e9-49d3-a47d-a771beee2a10","Type":"ContainerStarted","Data":"97d526a89a7d74ba6c90660fc5aef445431fcd2ff6479f9e01060bb40b611383"} Oct 14 10:30:02 crc kubenswrapper[4870]: I1014 10:30:02.883484 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29340630-dw5fj" Oct 14 10:30:03 crc kubenswrapper[4870]: I1014 10:30:03.034120 4870 scope.go:117] "RemoveContainer" containerID="3a25412c639a58b6345a6cb926f93c78a7a62c68cd923f500f921497a90b6efb" Oct 14 10:30:03 crc kubenswrapper[4870]: E1014 10:30:03.034347 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 10:30:03 crc kubenswrapper[4870]: I1014 10:30:03.079017 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gvh6f\" (UniqueName: \"kubernetes.io/projected/31e995c2-04e9-49d3-a47d-a771beee2a10-kube-api-access-gvh6f\") pod \"31e995c2-04e9-49d3-a47d-a771beee2a10\" (UID: \"31e995c2-04e9-49d3-a47d-a771beee2a10\") " Oct 14 10:30:03 crc kubenswrapper[4870]: I1014 10:30:03.079084 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/31e995c2-04e9-49d3-a47d-a771beee2a10-secret-volume\") pod \"31e995c2-04e9-49d3-a47d-a771beee2a10\" (UID: \"31e995c2-04e9-49d3-a47d-a771beee2a10\") " Oct 14 10:30:03 crc kubenswrapper[4870]: I1014 10:30:03.079142 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/31e995c2-04e9-49d3-a47d-a771beee2a10-config-volume\") pod \"31e995c2-04e9-49d3-a47d-a771beee2a10\" (UID: \"31e995c2-04e9-49d3-a47d-a771beee2a10\") " Oct 14 10:30:03 crc kubenswrapper[4870]: I1014 10:30:03.080148 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31e995c2-04e9-49d3-a47d-a771beee2a10-config-volume" (OuterVolumeSpecName: "config-volume") pod "31e995c2-04e9-49d3-a47d-a771beee2a10" (UID: "31e995c2-04e9-49d3-a47d-a771beee2a10"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 14 10:30:03 crc kubenswrapper[4870]: I1014 10:30:03.085055 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31e995c2-04e9-49d3-a47d-a771beee2a10-kube-api-access-gvh6f" (OuterVolumeSpecName: "kube-api-access-gvh6f") pod "31e995c2-04e9-49d3-a47d-a771beee2a10" (UID: "31e995c2-04e9-49d3-a47d-a771beee2a10"). InnerVolumeSpecName "kube-api-access-gvh6f". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 10:30:03 crc kubenswrapper[4870]: I1014 10:30:03.091823 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31e995c2-04e9-49d3-a47d-a771beee2a10-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "31e995c2-04e9-49d3-a47d-a771beee2a10" (UID: "31e995c2-04e9-49d3-a47d-a771beee2a10"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 14 10:30:03 crc kubenswrapper[4870]: I1014 10:30:03.181512 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gvh6f\" (UniqueName: \"kubernetes.io/projected/31e995c2-04e9-49d3-a47d-a771beee2a10-kube-api-access-gvh6f\") on node \"crc\" DevicePath \"\"" Oct 14 10:30:03 crc kubenswrapper[4870]: I1014 10:30:03.181540 4870 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/31e995c2-04e9-49d3-a47d-a771beee2a10-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 14 10:30:03 crc kubenswrapper[4870]: I1014 10:30:03.181553 4870 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/31e995c2-04e9-49d3-a47d-a771beee2a10-config-volume\") on node \"crc\" DevicePath \"\"" Oct 14 10:30:03 crc kubenswrapper[4870]: I1014 10:30:03.568078 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29340630-dw5fj" event={"ID":"31e995c2-04e9-49d3-a47d-a771beee2a10","Type":"ContainerDied","Data":"97d526a89a7d74ba6c90660fc5aef445431fcd2ff6479f9e01060bb40b611383"} Oct 14 10:30:03 crc kubenswrapper[4870]: I1014 10:30:03.568763 4870 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="97d526a89a7d74ba6c90660fc5aef445431fcd2ff6479f9e01060bb40b611383" Oct 14 10:30:03 crc kubenswrapper[4870]: I1014 10:30:03.568245 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29340630-dw5fj" Oct 14 10:30:03 crc kubenswrapper[4870]: I1014 10:30:03.982332 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29340585-q9shp"] Oct 14 10:30:03 crc kubenswrapper[4870]: I1014 10:30:03.993378 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29340585-q9shp"] Oct 14 10:30:05 crc kubenswrapper[4870]: I1014 10:30:05.059414 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0ffb284a-79dd-4b65-b824-c1c0211d4a0a" path="/var/lib/kubelet/pods/0ffb284a-79dd-4b65-b824-c1c0211d4a0a/volumes" Oct 14 10:30:17 crc kubenswrapper[4870]: I1014 10:30:17.037383 4870 scope.go:117] "RemoveContainer" containerID="3a25412c639a58b6345a6cb926f93c78a7a62c68cd923f500f921497a90b6efb" Oct 14 10:30:17 crc kubenswrapper[4870]: E1014 10:30:17.038993 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 10:30:31 crc kubenswrapper[4870]: I1014 10:30:31.039126 4870 scope.go:117] "RemoveContainer" containerID="3a25412c639a58b6345a6cb926f93c78a7a62c68cd923f500f921497a90b6efb" Oct 14 10:30:31 crc kubenswrapper[4870]: E1014 10:30:31.040323 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 10:30:36 crc kubenswrapper[4870]: I1014 10:30:36.782887 4870 scope.go:117] "RemoveContainer" containerID="d019c9eaf21cb0cbc817f0291648727457f79b6bd70d6e62f41211f6d66a537a" Oct 14 10:30:42 crc kubenswrapper[4870]: I1014 10:30:42.034434 4870 scope.go:117] "RemoveContainer" containerID="3a25412c639a58b6345a6cb926f93c78a7a62c68cd923f500f921497a90b6efb" Oct 14 10:30:42 crc kubenswrapper[4870]: E1014 10:30:42.035509 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 10:30:57 crc kubenswrapper[4870]: I1014 10:30:57.034729 4870 scope.go:117] "RemoveContainer" containerID="3a25412c639a58b6345a6cb926f93c78a7a62c68cd923f500f921497a90b6efb" Oct 14 10:30:57 crc kubenswrapper[4870]: E1014 10:30:57.035625 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 10:31:12 crc kubenswrapper[4870]: I1014 10:31:12.035239 4870 scope.go:117] "RemoveContainer" containerID="3a25412c639a58b6345a6cb926f93c78a7a62c68cd923f500f921497a90b6efb" Oct 14 10:31:12 crc kubenswrapper[4870]: E1014 10:31:12.037002 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 10:31:25 crc kubenswrapper[4870]: I1014 10:31:25.051785 4870 scope.go:117] "RemoveContainer" containerID="3a25412c639a58b6345a6cb926f93c78a7a62c68cd923f500f921497a90b6efb" Oct 14 10:31:25 crc kubenswrapper[4870]: E1014 10:31:25.052790 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 10:31:39 crc kubenswrapper[4870]: I1014 10:31:39.035660 4870 scope.go:117] "RemoveContainer" containerID="3a25412c639a58b6345a6cb926f93c78a7a62c68cd923f500f921497a90b6efb" Oct 14 10:31:39 crc kubenswrapper[4870]: E1014 10:31:39.036925 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 10:31:51 crc kubenswrapper[4870]: I1014 10:31:51.034172 4870 scope.go:117] "RemoveContainer" containerID="3a25412c639a58b6345a6cb926f93c78a7a62c68cd923f500f921497a90b6efb" Oct 14 10:31:51 crc kubenswrapper[4870]: E1014 10:31:51.036092 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 10:32:06 crc kubenswrapper[4870]: I1014 10:32:06.034001 4870 scope.go:117] "RemoveContainer" containerID="3a25412c639a58b6345a6cb926f93c78a7a62c68cd923f500f921497a90b6efb" Oct 14 10:32:06 crc kubenswrapper[4870]: E1014 10:32:06.034779 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 10:32:07 crc kubenswrapper[4870]: I1014 10:32:07.994001 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_ffbac7fa-fdbf-42d6-9941-472fa69c381b/init-config-reloader/0.log" Oct 14 10:32:08 crc kubenswrapper[4870]: I1014 10:32:08.212032 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_ffbac7fa-fdbf-42d6-9941-472fa69c381b/init-config-reloader/0.log" Oct 14 10:32:08 crc kubenswrapper[4870]: I1014 10:32:08.215535 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_ffbac7fa-fdbf-42d6-9941-472fa69c381b/alertmanager/0.log" Oct 14 10:32:08 crc kubenswrapper[4870]: I1014 10:32:08.243608 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_ffbac7fa-fdbf-42d6-9941-472fa69c381b/config-reloader/0.log" Oct 14 10:32:08 crc kubenswrapper[4870]: I1014 10:32:08.416295 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_81cb33bd-91aa-4f54-a805-596957cb2b26/aodh-api/0.log" Oct 14 10:32:08 crc kubenswrapper[4870]: I1014 10:32:08.466750 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_81cb33bd-91aa-4f54-a805-596957cb2b26/aodh-evaluator/0.log" Oct 14 10:32:08 crc kubenswrapper[4870]: I1014 10:32:08.573717 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_81cb33bd-91aa-4f54-a805-596957cb2b26/aodh-listener/0.log" Oct 14 10:32:08 crc kubenswrapper[4870]: I1014 10:32:08.599625 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_81cb33bd-91aa-4f54-a805-596957cb2b26/aodh-notifier/0.log" Oct 14 10:32:08 crc kubenswrapper[4870]: I1014 10:32:08.787288 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-d46667c8-kddgf_658a9360-4e8e-40ca-9575-5501fb52be26/barbican-api/0.log" Oct 14 10:32:08 crc kubenswrapper[4870]: I1014 10:32:08.820578 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-d46667c8-kddgf_658a9360-4e8e-40ca-9575-5501fb52be26/barbican-api-log/0.log" Oct 14 10:32:08 crc kubenswrapper[4870]: I1014 10:32:08.989598 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-68859d4c94-fx6jt_cc348a20-302d-407f-be5c-b690c39f12cf/barbican-keystone-listener/0.log" Oct 14 10:32:09 crc kubenswrapper[4870]: I1014 10:32:09.268713 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-7c9b74fd49-gbcpm_4446db55-22b0-4800-9ebc-f4b6b6578679/barbican-worker/0.log" Oct 14 10:32:09 crc kubenswrapper[4870]: I1014 10:32:09.498266 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-7c9b74fd49-gbcpm_4446db55-22b0-4800-9ebc-f4b6b6578679/barbican-worker-log/0.log" Oct 14 10:32:09 crc kubenswrapper[4870]: I1014 10:32:09.664079 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-68859d4c94-fx6jt_cc348a20-302d-407f-be5c-b690c39f12cf/barbican-keystone-listener-log/0.log" Oct 14 10:32:09 crc kubenswrapper[4870]: I1014 10:32:09.785959 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-openstack-openstack-cell1-flgnw_295dcd08-48f4-484b-a2e5-5c03db6afd73/bootstrap-openstack-openstack-cell1/0.log" Oct 14 10:32:09 crc kubenswrapper[4870]: I1014 10:32:09.901882 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-openstack-openstack-networker-772fx_139b40af-cdfc-417f-9807-a67634524bcb/bootstrap-openstack-openstack-networker/0.log" Oct 14 10:32:10 crc kubenswrapper[4870]: I1014 10:32:10.292428 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_a26548c9-166f-422a-ad41-24302cab4447/ceilometer-central-agent/0.log" Oct 14 10:32:10 crc kubenswrapper[4870]: I1014 10:32:10.346937 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_a26548c9-166f-422a-ad41-24302cab4447/ceilometer-notification-agent/0.log" Oct 14 10:32:10 crc kubenswrapper[4870]: I1014 10:32:10.489536 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_a26548c9-166f-422a-ad41-24302cab4447/sg-core/0.log" Oct 14 10:32:10 crc kubenswrapper[4870]: I1014 10:32:10.518842 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_a26548c9-166f-422a-ad41-24302cab4447/proxy-httpd/0.log" Oct 14 10:32:10 crc kubenswrapper[4870]: I1014 10:32:10.684116 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceph-client-openstack-openstack-cell1-f7d6q_aee008c6-821e-44b3-a557-11dda24d715b/ceph-client-openstack-openstack-cell1/0.log" Oct 14 10:32:11 crc kubenswrapper[4870]: I1014 10:32:11.087742 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_0a387a03-3520-4d50-b7c7-5e3abb17978d/cinder-api/0.log" Oct 14 10:32:11 crc kubenswrapper[4870]: I1014 10:32:11.302055 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_0a387a03-3520-4d50-b7c7-5e3abb17978d/cinder-api-log/0.log" Oct 14 10:32:11 crc kubenswrapper[4870]: I1014 10:32:11.329581 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-backup-0_0852377f-00a7-406c-91e0-83310c8c4abc/probe/0.log" Oct 14 10:32:11 crc kubenswrapper[4870]: I1014 10:32:11.586000 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_dad751a8-e411-4244-8bae-004bcad630a0/cinder-scheduler/0.log" Oct 14 10:32:11 crc kubenswrapper[4870]: I1014 10:32:11.626690 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_dad751a8-e411-4244-8bae-004bcad630a0/probe/0.log" Oct 14 10:32:11 crc kubenswrapper[4870]: I1014 10:32:11.980271 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-volume-volume1-0_599f0358-dd66-498a-94c2-b0cfe3053ac8/probe/0.log" Oct 14 10:32:12 crc kubenswrapper[4870]: I1014 10:32:12.156856 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-openstack-openstack-cell1-jglgw_2e2a40a0-7698-4294-a107-639c5de9e5f7/configure-network-openstack-openstack-cell1/0.log" Oct 14 10:32:12 crc kubenswrapper[4870]: I1014 10:32:12.160598 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-backup-0_0852377f-00a7-406c-91e0-83310c8c4abc/cinder-backup/0.log" Oct 14 10:32:12 crc kubenswrapper[4870]: I1014 10:32:12.434399 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-openstack-openstack-networker-8chgx_5a77a035-4684-43ed-b3a1-1e9dfa53dd63/configure-network-openstack-openstack-networker/0.log" Oct 14 10:32:12 crc kubenswrapper[4870]: I1014 10:32:12.522414 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-openstack-openstack-cell1-7928w_e3f48d1d-384a-4b16-964a-5834de86fdb3/configure-os-openstack-openstack-cell1/0.log" Oct 14 10:32:12 crc kubenswrapper[4870]: I1014 10:32:12.684350 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-openstack-openstack-cell1-p9242_6815b76f-1435-4980-9b2f-9c36e02d6f63/configure-os-openstack-openstack-cell1/0.log" Oct 14 10:32:12 crc kubenswrapper[4870]: I1014 10:32:12.935228 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-openstack-openstack-networker-gbdf9_0d62b70c-9f3a-4029-92ff-05143f46254e/configure-os-openstack-openstack-networker/0.log" Oct 14 10:32:13 crc kubenswrapper[4870]: I1014 10:32:13.037739 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-openstack-openstack-networker-nwsds_324ce306-ca93-447c-a19c-711764988cc9/configure-os-openstack-openstack-networker/0.log" Oct 14 10:32:13 crc kubenswrapper[4870]: I1014 10:32:13.176181 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-67dbcf57c9-fcxrn_e5c27b78-9614-44c4-b96e-db0380508659/init/0.log" Oct 14 10:32:13 crc kubenswrapper[4870]: I1014 10:32:13.461949 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-67dbcf57c9-fcxrn_e5c27b78-9614-44c4-b96e-db0380508659/init/0.log" Oct 14 10:32:13 crc kubenswrapper[4870]: I1014 10:32:13.566206 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-67dbcf57c9-fcxrn_e5c27b78-9614-44c4-b96e-db0380508659/dnsmasq-dns/0.log" Oct 14 10:32:13 crc kubenswrapper[4870]: I1014 10:32:13.677675 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-openstack-openstack-cell1-cfvhm_e4ecb749-bbab-40eb-82be-bd39305cb787/download-cache-openstack-openstack-cell1/0.log" Oct 14 10:32:13 crc kubenswrapper[4870]: I1014 10:32:13.840525 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-openstack-openstack-networker-pqffh_ea545890-c197-4826-8f15-380af2a44ad9/download-cache-openstack-openstack-networker/0.log" Oct 14 10:32:13 crc kubenswrapper[4870]: I1014 10:32:13.988365 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-volume-volume1-0_599f0358-dd66-498a-94c2-b0cfe3053ac8/cinder-volume/0.log" Oct 14 10:32:14 crc kubenswrapper[4870]: I1014 10:32:14.054529 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_95cfd5a9-9dfc-44a4-bb94-0d60963dffd7/glance-httpd/0.log" Oct 14 10:32:14 crc kubenswrapper[4870]: I1014 10:32:14.082904 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_95cfd5a9-9dfc-44a4-bb94-0d60963dffd7/glance-log/0.log" Oct 14 10:32:14 crc kubenswrapper[4870]: I1014 10:32:14.224657 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_bdd40f40-57d1-4dbf-8b03-f447844dbcd3/glance-httpd/0.log" Oct 14 10:32:14 crc kubenswrapper[4870]: I1014 10:32:14.269609 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_bdd40f40-57d1-4dbf-8b03-f447844dbcd3/glance-log/0.log" Oct 14 10:32:14 crc kubenswrapper[4870]: I1014 10:32:14.546040 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-api-f4c547d9b-gjq77_0e232c83-c81a-4c64-ad90-74b771732688/heat-api/0.log" Oct 14 10:32:14 crc kubenswrapper[4870]: I1014 10:32:14.618705 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-cfnapi-77b878966d-fm2hf_1aa3b494-7370-4251-afd5-01109e40cc85/heat-cfnapi/0.log" Oct 14 10:32:14 crc kubenswrapper[4870]: I1014 10:32:14.668599 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-engine-6bb7df6545-4x5wj_3ff05517-1df4-4a38-ad5c-c293546a56e3/heat-engine/0.log" Oct 14 10:32:15 crc kubenswrapper[4870]: I1014 10:32:15.028900 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-6fb9fc4bf-jmsxf_73bab307-e3b8-47df-bba3-fa624f4c4261/horizon/0.log" Oct 14 10:32:15 crc kubenswrapper[4870]: I1014 10:32:15.177763 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-6fb9fc4bf-jmsxf_73bab307-e3b8-47df-bba3-fa624f4c4261/horizon-log/0.log" Oct 14 10:32:15 crc kubenswrapper[4870]: I1014 10:32:15.234613 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-openstack-openstack-cell1-q4zrw_a90e5cfd-b6bb-45a3-8b31-cfa84c3ebf93/install-certs-openstack-openstack-cell1/0.log" Oct 14 10:32:15 crc kubenswrapper[4870]: I1014 10:32:15.361656 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-openstack-openstack-networker-lqg8j_5570af78-f5d2-414d-8437-1e9b8210f0de/install-certs-openstack-openstack-networker/0.log" Oct 14 10:32:15 crc kubenswrapper[4870]: I1014 10:32:15.438254 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-openstack-openstack-cell1-54frh_faed3d6f-4637-44f8-8925-0e82407fcd44/install-os-openstack-openstack-cell1/0.log" Oct 14 10:32:15 crc kubenswrapper[4870]: I1014 10:32:15.594595 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-openstack-openstack-networker-wdjn4_ed0d4c93-b579-4253-867f-d191d43c35c7/install-os-openstack-openstack-networker/0.log" Oct 14 10:32:15 crc kubenswrapper[4870]: I1014 10:32:15.825095 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29340541-wtcff_ec2b912e-ff75-44ed-b862-6bffd46f0ddf/keystone-cron/0.log" Oct 14 10:32:15 crc kubenswrapper[4870]: I1014 10:32:15.907571 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29340601-cnz9t_4ad4b83a-f2a8-439f-a02e-66424aa17de1/keystone-cron/0.log" Oct 14 10:32:16 crc kubenswrapper[4870]: I1014 10:32:16.138828 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_f1c15a86-a683-46ae-8ca4-431ebbf213c9/kube-state-metrics/0.log" Oct 14 10:32:16 crc kubenswrapper[4870]: I1014 10:32:16.396050 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-openstack-openstack-cell1-r6f6j_6795067a-45b7-42d4-abbb-44a3f094c02e/libvirt-openstack-openstack-cell1/0.log" Oct 14 10:32:16 crc kubenswrapper[4870]: I1014 10:32:16.789097 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-58b648bb8f-6s7rq_5df1fce8-53fe-42c0-a90c-683eb730b653/keystone-api/0.log" Oct 14 10:32:16 crc kubenswrapper[4870]: I1014 10:32:16.842249 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-api-0_0e863093-887f-414c-bfed-ffb1be0f0033/manila-api/0.log" Oct 14 10:32:16 crc kubenswrapper[4870]: I1014 10:32:16.984729 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-scheduler-0_f1580185-56b7-4ef3-b148-e3934a5df6eb/manila-scheduler/0.log" Oct 14 10:32:17 crc kubenswrapper[4870]: I1014 10:32:17.035271 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-scheduler-0_f1580185-56b7-4ef3-b148-e3934a5df6eb/probe/0.log" Oct 14 10:32:17 crc kubenswrapper[4870]: I1014 10:32:17.060701 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-api-0_0e863093-887f-414c-bfed-ffb1be0f0033/manila-api-log/0.log" Oct 14 10:32:17 crc kubenswrapper[4870]: I1014 10:32:17.271609 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-share-share1-0_dea5cc89-8770-43ab-a5f3-4986cfac76af/probe/0.log" Oct 14 10:32:17 crc kubenswrapper[4870]: I1014 10:32:17.274331 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-share-share1-0_dea5cc89-8770-43ab-a5f3-4986cfac76af/manila-share/0.log" Oct 14 10:32:17 crc kubenswrapper[4870]: I1014 10:32:17.845533 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-75695669c7-rj984_47c05f0e-09b2-453f-b777-72836ec666f3/neutron-httpd/0.log" Oct 14 10:32:18 crc kubenswrapper[4870]: I1014 10:32:18.052423 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-dhcp-openstack-openstack-cell1-tmmz7_0417d7ee-a386-402e-bfa8-72b8dc5cf38e/neutron-dhcp-openstack-openstack-cell1/0.log" Oct 14 10:32:18 crc kubenswrapper[4870]: I1014 10:32:18.231517 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-75695669c7-rj984_47c05f0e-09b2-453f-b777-72836ec666f3/neutron-api/0.log" Oct 14 10:32:18 crc kubenswrapper[4870]: I1014 10:32:18.304671 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-openstack-openstack-cell1-kh4dj_793a3762-db45-4a91-a194-7710517917e8/neutron-metadata-openstack-openstack-cell1/0.log" Oct 14 10:32:18 crc kubenswrapper[4870]: I1014 10:32:18.565377 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-openstack-openstack-networker-7fknh_6b3e50e6-d053-4f51-9a47-60dc660562a7/neutron-metadata-openstack-openstack-networker/0.log" Oct 14 10:32:18 crc kubenswrapper[4870]: I1014 10:32:18.661826 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-sriov-openstack-openstack-cell1-7zwk7_59c26d47-f3a1-4103-a00d-5ee15b387c4a/neutron-sriov-openstack-openstack-cell1/0.log" Oct 14 10:32:19 crc kubenswrapper[4870]: I1014 10:32:19.038304 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_fce04f20-89b3-463d-b0eb-f32adcf5c8de/nova-api-api/0.log" Oct 14 10:32:19 crc kubenswrapper[4870]: I1014 10:32:19.246015 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_fce04f20-89b3-463d-b0eb-f32adcf5c8de/nova-api-log/0.log" Oct 14 10:32:19 crc kubenswrapper[4870]: I1014 10:32:19.337847 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_52dbdd9e-e634-404b-8881-1a898ea568f0/nova-cell0-conductor-conductor/0.log" Oct 14 10:32:19 crc kubenswrapper[4870]: I1014 10:32:19.579589 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_60d06fa5-6fbe-4958-8a2e-a1781c615a46/nova-cell1-conductor-conductor/0.log" Oct 14 10:32:19 crc kubenswrapper[4870]: I1014 10:32:19.769682 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_becafd06-ed0f-4ecf-b9f1-2027b349da1b/nova-cell1-novncproxy-novncproxy/0.log" Oct 14 10:32:20 crc kubenswrapper[4870]: I1014 10:32:20.033713 4870 scope.go:117] "RemoveContainer" containerID="3a25412c639a58b6345a6cb926f93c78a7a62c68cd923f500f921497a90b6efb" Oct 14 10:32:20 crc kubenswrapper[4870]: E1014 10:32:20.034028 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 10:32:20 crc kubenswrapper[4870]: I1014 10:32:20.120849 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell7hdbp_2ff165ad-7fb3-4c5a-918b-94950ba97a0a/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell1/0.log" Oct 14 10:32:20 crc kubenswrapper[4870]: I1014 10:32:20.385109 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-openstack-openstack-cell1-z74pq_7f05f9ee-e89a-45a1-9f10-e68addbef873/nova-cell1-openstack-openstack-cell1/0.log" Oct 14 10:32:20 crc kubenswrapper[4870]: I1014 10:32:20.431625 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_9827ed43-9350-43db-8503-f4b9b0bd9c25/nova-metadata-log/0.log" Oct 14 10:32:20 crc kubenswrapper[4870]: I1014 10:32:20.656570 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_9827ed43-9350-43db-8503-f4b9b0bd9c25/nova-metadata-metadata/0.log" Oct 14 10:32:20 crc kubenswrapper[4870]: I1014 10:32:20.798746 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_dee94958-2c48-4715-964b-ebd6efe1f95e/nova-scheduler-scheduler/0.log" Oct 14 10:32:20 crc kubenswrapper[4870]: I1014 10:32:20.851023 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_8cd09b24-b663-4aea-9ef4-36cf69ea689c/mysql-bootstrap/0.log" Oct 14 10:32:21 crc kubenswrapper[4870]: I1014 10:32:21.102590 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_8cd09b24-b663-4aea-9ef4-36cf69ea689c/mysql-bootstrap/0.log" Oct 14 10:32:21 crc kubenswrapper[4870]: I1014 10:32:21.111117 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_8cd09b24-b663-4aea-9ef4-36cf69ea689c/galera/0.log" Oct 14 10:32:21 crc kubenswrapper[4870]: I1014 10:32:21.306923 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_05c1ec22-e0d6-45c4-85fc-d586c046b3b1/mysql-bootstrap/0.log" Oct 14 10:32:21 crc kubenswrapper[4870]: I1014 10:32:21.496361 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_05c1ec22-e0d6-45c4-85fc-d586c046b3b1/mysql-bootstrap/0.log" Oct 14 10:32:21 crc kubenswrapper[4870]: I1014 10:32:21.526875 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_05c1ec22-e0d6-45c4-85fc-d586c046b3b1/galera/0.log" Oct 14 10:32:21 crc kubenswrapper[4870]: I1014 10:32:21.721160 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_73cb6ac7-5103-401c-8d79-6d2d337b133c/openstackclient/0.log" Oct 14 10:32:21 crc kubenswrapper[4870]: I1014 10:32:21.793304 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_c8bd2488-7e3a-4dc2-87c9-dbaf25141f92/openstack-network-exporter/0.log" Oct 14 10:32:21 crc kubenswrapper[4870]: I1014 10:32:21.943883 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_c8bd2488-7e3a-4dc2-87c9-dbaf25141f92/ovn-northd/0.log" Oct 14 10:32:22 crc kubenswrapper[4870]: I1014 10:32:22.034852 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-openstack-openstack-cell1-6ggqp_6735cf95-3968-4805-b08c-e3d9231e0d6f/ovn-openstack-openstack-cell1/0.log" Oct 14 10:32:22 crc kubenswrapper[4870]: I1014 10:32:22.378904 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-openstack-openstack-networker-hw7gw_f27b6002-352b-4191-8b00-347f35a563bf/ovn-openstack-openstack-networker/0.log" Oct 14 10:32:22 crc kubenswrapper[4870]: I1014 10:32:22.438590 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_79b5761f-2312-4d15-9ded-d42c9f8f6009/openstack-network-exporter/0.log" Oct 14 10:32:22 crc kubenswrapper[4870]: I1014 10:32:22.526712 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_79b5761f-2312-4d15-9ded-d42c9f8f6009/ovsdbserver-nb/0.log" Oct 14 10:32:22 crc kubenswrapper[4870]: I1014 10:32:22.651264 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-1_2f70f991-789c-49bc-bd89-aeb6a69e6ac5/openstack-network-exporter/0.log" Oct 14 10:32:22 crc kubenswrapper[4870]: I1014 10:32:22.738799 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-1_2f70f991-789c-49bc-bd89-aeb6a69e6ac5/ovsdbserver-nb/0.log" Oct 14 10:32:22 crc kubenswrapper[4870]: I1014 10:32:22.852727 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-2_7d4f0ff6-a758-4177-a48e-679d3373589f/openstack-network-exporter/0.log" Oct 14 10:32:22 crc kubenswrapper[4870]: I1014 10:32:22.946274 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-2_7d4f0ff6-a758-4177-a48e-679d3373589f/ovsdbserver-nb/0.log" Oct 14 10:32:23 crc kubenswrapper[4870]: I1014 10:32:23.061379 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_0363cb3b-2556-47c2-a1d4-d1cec83c1354/openstack-network-exporter/0.log" Oct 14 10:32:23 crc kubenswrapper[4870]: I1014 10:32:23.202654 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_0363cb3b-2556-47c2-a1d4-d1cec83c1354/ovsdbserver-sb/0.log" Oct 14 10:32:23 crc kubenswrapper[4870]: I1014 10:32:23.295685 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-1_09981a06-37e3-4dbf-b991-5a5a74824641/openstack-network-exporter/0.log" Oct 14 10:32:23 crc kubenswrapper[4870]: I1014 10:32:23.390657 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-1_09981a06-37e3-4dbf-b991-5a5a74824641/ovsdbserver-sb/0.log" Oct 14 10:32:23 crc kubenswrapper[4870]: I1014 10:32:23.483493 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-2_cd6f2350-67d9-417b-90e9-599fd80326ab/openstack-network-exporter/0.log" Oct 14 10:32:23 crc kubenswrapper[4870]: I1014 10:32:23.594535 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-2_cd6f2350-67d9-417b-90e9-599fd80326ab/ovsdbserver-sb/0.log" Oct 14 10:32:23 crc kubenswrapper[4870]: I1014 10:32:23.911630 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-594cc89896-9lvzl_21457f84-984b-4745-823d-d4fb6bb2a6c0/placement-api/0.log" Oct 14 10:32:24 crc kubenswrapper[4870]: I1014 10:32:24.028326 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-594cc89896-9lvzl_21457f84-984b-4745-823d-d4fb6bb2a6c0/placement-log/0.log" Oct 14 10:32:24 crc kubenswrapper[4870]: I1014 10:32:24.139149 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_pre-adoption-validation-openstack-pre-adoption-openstack-c6xncd_2233d692-f4ea-4390-adde-275dab829e73/pre-adoption-validation-openstack-pre-adoption-openstack-cell1/0.log" Oct 14 10:32:24 crc kubenswrapper[4870]: I1014 10:32:24.286907 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_pre-adoption-validation-openstack-pre-adoption-openstack-nwfcbm_364d6666-a8e2-48f6-b42d-3d676de7f45a/pre-adoption-validation-openstack-pre-adoption-openstack-networ/0.log" Oct 14 10:32:24 crc kubenswrapper[4870]: I1014 10:32:24.459188 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_1920aea5-2efe-4faf-9a54-d5bb6f83cc08/init-config-reloader/0.log" Oct 14 10:32:24 crc kubenswrapper[4870]: I1014 10:32:24.599120 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_1920aea5-2efe-4faf-9a54-d5bb6f83cc08/config-reloader/0.log" Oct 14 10:32:24 crc kubenswrapper[4870]: I1014 10:32:24.632427 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_1920aea5-2efe-4faf-9a54-d5bb6f83cc08/init-config-reloader/0.log" Oct 14 10:32:24 crc kubenswrapper[4870]: I1014 10:32:24.690838 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_1920aea5-2efe-4faf-9a54-d5bb6f83cc08/prometheus/0.log" Oct 14 10:32:24 crc kubenswrapper[4870]: I1014 10:32:24.818316 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_1920aea5-2efe-4faf-9a54-d5bb6f83cc08/thanos-sidecar/0.log" Oct 14 10:32:24 crc kubenswrapper[4870]: I1014 10:32:24.859496 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_e07b718e-6d3b-491e-9229-d30e5228635c/setup-container/0.log" Oct 14 10:32:25 crc kubenswrapper[4870]: I1014 10:32:25.227552 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_e07b718e-6d3b-491e-9229-d30e5228635c/setup-container/0.log" Oct 14 10:32:25 crc kubenswrapper[4870]: I1014 10:32:25.250886 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_e07b718e-6d3b-491e-9229-d30e5228635c/rabbitmq/0.log" Oct 14 10:32:25 crc kubenswrapper[4870]: I1014 10:32:25.401057 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_a40d8a1a-605f-4db1-bf5f-4e4785efaa4e/setup-container/0.log" Oct 14 10:32:25 crc kubenswrapper[4870]: I1014 10:32:25.599919 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_a40d8a1a-605f-4db1-bf5f-4e4785efaa4e/rabbitmq/0.log" Oct 14 10:32:25 crc kubenswrapper[4870]: I1014 10:32:25.608128 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_a40d8a1a-605f-4db1-bf5f-4e4785efaa4e/setup-container/0.log" Oct 14 10:32:25 crc kubenswrapper[4870]: I1014 10:32:25.771137 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-openstack-openstack-cell1-cx8gp_f7861d1b-2199-407a-8abb-594c42ec2544/reboot-os-openstack-openstack-cell1/0.log" Oct 14 10:32:25 crc kubenswrapper[4870]: I1014 10:32:25.866020 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-openstack-openstack-networker-j54sj_2df5305c-8a09-4db0-9fe3-28ea067728a5/reboot-os-openstack-openstack-networker/0.log" Oct 14 10:32:26 crc kubenswrapper[4870]: I1014 10:32:26.077039 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-openstack-openstack-cell1-m4fnw_fe4c4a5a-274a-4adf-929a-46a1057c51ee/run-os-openstack-openstack-cell1/0.log" Oct 14 10:32:26 crc kubenswrapper[4870]: I1014 10:32:26.218352 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-openstack-openstack-networker-89k68_b3e9780b-d240-4d71-bb80-3e21346f7eb9/run-os-openstack-openstack-networker/0.log" Oct 14 10:32:26 crc kubenswrapper[4870]: I1014 10:32:26.345414 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-openstack-rmpkj_b48a6df4-d071-487d-b843-b60af6b5a141/ssh-known-hosts-openstack/0.log" Oct 14 10:32:26 crc kubenswrapper[4870]: I1014 10:32:26.600155 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-openstack-openstack-cell1-q2bvq_1358f0f9-0ff8-415f-9861-359ba368c8dd/telemetry-openstack-openstack-cell1/0.log" Oct 14 10:32:26 crc kubenswrapper[4870]: I1014 10:32:26.757693 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_3d2c8717-a111-4e18-99e4-46be8b4b5673/memcached/0.log" Oct 14 10:32:26 crc kubenswrapper[4870]: I1014 10:32:26.817350 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_08ea5fd9-1afd-44de-ae26-61e52ca72e7a/test-operator-logs-container/0.log" Oct 14 10:32:26 crc kubenswrapper[4870]: I1014 10:32:26.824837 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_2929a25a-a5e9-4a00-8b77-6939e3bf7d66/tempest-tests-tempest-tests-runner/0.log" Oct 14 10:32:27 crc kubenswrapper[4870]: I1014 10:32:27.013811 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tripleo-cleanup-tripleo-cleanup-openstack-cell1-tz9gr_ea75b5d6-ee4b-4a8d-a3af-4995af7bafb9/tripleo-cleanup-tripleo-cleanup-openstack-cell1/0.log" Oct 14 10:32:27 crc kubenswrapper[4870]: I1014 10:32:27.125852 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tripleo-cleanup-tripleo-cleanup-openstack-networker-ncc5q_b96078e0-b889-4b47-9bec-a1411c1fec36/tripleo-cleanup-tripleo-cleanup-openstack-networker/0.log" Oct 14 10:32:27 crc kubenswrapper[4870]: I1014 10:32:27.266129 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-openstack-openstack-cell1-lrc5r_34d9c179-4fe6-4431-b495-6803732cb3ee/validate-network-openstack-openstack-cell1/0.log" Oct 14 10:32:27 crc kubenswrapper[4870]: I1014 10:32:27.362696 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-openstack-openstack-networker-ps772_1cd91851-2684-4634-8d0e-78857205d00a/validate-network-openstack-openstack-networker/0.log" Oct 14 10:32:35 crc kubenswrapper[4870]: I1014 10:32:35.051102 4870 scope.go:117] "RemoveContainer" containerID="3a25412c639a58b6345a6cb926f93c78a7a62c68cd923f500f921497a90b6efb" Oct 14 10:32:35 crc kubenswrapper[4870]: E1014 10:32:35.052080 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 10:32:36 crc kubenswrapper[4870]: I1014 10:32:36.917911 4870 scope.go:117] "RemoveContainer" containerID="fb8dca5495a657298b5da9f1346d48e73b65ff8a64e1f923ac01376ded87ac4a" Oct 14 10:32:36 crc kubenswrapper[4870]: I1014 10:32:36.950212 4870 scope.go:117] "RemoveContainer" containerID="9ee7ad13e366e910afe556b0daa8475868a5d304ebd1c3df35079ce1e5d5a240" Oct 14 10:32:36 crc kubenswrapper[4870]: I1014 10:32:36.974322 4870 scope.go:117] "RemoveContainer" containerID="5872c468c4140a8c9bf9e3b2ba85afdb854e111177aa39951ae5930ec957fd3b" Oct 14 10:32:46 crc kubenswrapper[4870]: I1014 10:32:46.034894 4870 scope.go:117] "RemoveContainer" containerID="3a25412c639a58b6345a6cb926f93c78a7a62c68cd923f500f921497a90b6efb" Oct 14 10:32:46 crc kubenswrapper[4870]: E1014 10:32:46.035729 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 10:33:01 crc kubenswrapper[4870]: I1014 10:33:01.034983 4870 scope.go:117] "RemoveContainer" containerID="3a25412c639a58b6345a6cb926f93c78a7a62c68cd923f500f921497a90b6efb" Oct 14 10:33:01 crc kubenswrapper[4870]: E1014 10:33:01.036399 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 10:33:07 crc kubenswrapper[4870]: I1014 10:33:07.620721 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_32da80840a2017f27ed4ad61f02adc64a25aa18e8dad0409953372036al2fnh_36e53446-018c-48eb-8398-f28a109c9cb0/util/0.log" Oct 14 10:33:07 crc kubenswrapper[4870]: I1014 10:33:07.840556 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_32da80840a2017f27ed4ad61f02adc64a25aa18e8dad0409953372036al2fnh_36e53446-018c-48eb-8398-f28a109c9cb0/pull/0.log" Oct 14 10:33:07 crc kubenswrapper[4870]: I1014 10:33:07.844756 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_32da80840a2017f27ed4ad61f02adc64a25aa18e8dad0409953372036al2fnh_36e53446-018c-48eb-8398-f28a109c9cb0/util/0.log" Oct 14 10:33:07 crc kubenswrapper[4870]: I1014 10:33:07.854014 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_32da80840a2017f27ed4ad61f02adc64a25aa18e8dad0409953372036al2fnh_36e53446-018c-48eb-8398-f28a109c9cb0/pull/0.log" Oct 14 10:33:08 crc kubenswrapper[4870]: I1014 10:33:08.016110 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_32da80840a2017f27ed4ad61f02adc64a25aa18e8dad0409953372036al2fnh_36e53446-018c-48eb-8398-f28a109c9cb0/extract/0.log" Oct 14 10:33:08 crc kubenswrapper[4870]: I1014 10:33:08.019502 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_32da80840a2017f27ed4ad61f02adc64a25aa18e8dad0409953372036al2fnh_36e53446-018c-48eb-8398-f28a109c9cb0/util/0.log" Oct 14 10:33:08 crc kubenswrapper[4870]: I1014 10:33:08.041611 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_32da80840a2017f27ed4ad61f02adc64a25aa18e8dad0409953372036al2fnh_36e53446-018c-48eb-8398-f28a109c9cb0/pull/0.log" Oct 14 10:33:08 crc kubenswrapper[4870]: I1014 10:33:08.204070 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-658bdf4b74-76kw6_36e6bbc0-f57d-4486-9120-d2cd5a5ce638/kube-rbac-proxy/0.log" Oct 14 10:33:08 crc kubenswrapper[4870]: I1014 10:33:08.311850 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-7b7fb68549-xnk22_9f311a89-2483-42d6-9fe5-57b1900624d2/kube-rbac-proxy/0.log" Oct 14 10:33:08 crc kubenswrapper[4870]: I1014 10:33:08.346764 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-658bdf4b74-76kw6_36e6bbc0-f57d-4486-9120-d2cd5a5ce638/manager/0.log" Oct 14 10:33:08 crc kubenswrapper[4870]: I1014 10:33:08.509467 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-7b7fb68549-xnk22_9f311a89-2483-42d6-9fe5-57b1900624d2/manager/0.log" Oct 14 10:33:08 crc kubenswrapper[4870]: I1014 10:33:08.524625 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-85d5d9dd78-ld7xh_634060b4-d317-415d-bc9e-aa3ed6503b4f/kube-rbac-proxy/0.log" Oct 14 10:33:08 crc kubenswrapper[4870]: I1014 10:33:08.594887 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-85d5d9dd78-ld7xh_634060b4-d317-415d-bc9e-aa3ed6503b4f/manager/0.log" Oct 14 10:33:08 crc kubenswrapper[4870]: I1014 10:33:08.757423 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-84b9b84486-cl7ft_b4a4fd5f-360a-45ca-9684-74fe1c7631a6/kube-rbac-proxy/0.log" Oct 14 10:33:08 crc kubenswrapper[4870]: I1014 10:33:08.869242 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-84b9b84486-cl7ft_b4a4fd5f-360a-45ca-9684-74fe1c7631a6/manager/0.log" Oct 14 10:33:08 crc kubenswrapper[4870]: I1014 10:33:08.907483 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-858f76bbdd-9bdj8_519142a1-852b-4410-a511-13e2a36aed77/kube-rbac-proxy/0.log" Oct 14 10:33:09 crc kubenswrapper[4870]: I1014 10:33:09.027004 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-858f76bbdd-9bdj8_519142a1-852b-4410-a511-13e2a36aed77/manager/0.log" Oct 14 10:33:09 crc kubenswrapper[4870]: I1014 10:33:09.101995 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-7ffbcb7588-g2fdg_8ed9c494-ed5c-46d0-9434-1edbfcfba361/kube-rbac-proxy/0.log" Oct 14 10:33:09 crc kubenswrapper[4870]: I1014 10:33:09.128752 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-7ffbcb7588-g2fdg_8ed9c494-ed5c-46d0-9434-1edbfcfba361/manager/0.log" Oct 14 10:33:09 crc kubenswrapper[4870]: I1014 10:33:09.291119 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-656bcbd775-lp9f2_d4236126-b907-4f39-a491-5006f2a9c301/kube-rbac-proxy/0.log" Oct 14 10:33:09 crc kubenswrapper[4870]: I1014 10:33:09.420454 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-9c5c78d49-gjsj6_d27f8663-102d-4565-b7d5-00db25a8388a/kube-rbac-proxy/0.log" Oct 14 10:33:09 crc kubenswrapper[4870]: I1014 10:33:09.543860 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-656bcbd775-lp9f2_d4236126-b907-4f39-a491-5006f2a9c301/manager/0.log" Oct 14 10:33:09 crc kubenswrapper[4870]: I1014 10:33:09.548231 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-9c5c78d49-gjsj6_d27f8663-102d-4565-b7d5-00db25a8388a/manager/0.log" Oct 14 10:33:09 crc kubenswrapper[4870]: I1014 10:33:09.625847 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-55b6b7c7b8-jbt9q_5ad2a6ec-e2aa-45d9-b1a8-59ed908aaacd/kube-rbac-proxy/0.log" Oct 14 10:33:09 crc kubenswrapper[4870]: I1014 10:33:09.837626 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-5f67fbc655-szjkl_4bc1fceb-2d98-415e-b34e-82c2d6e1430f/kube-rbac-proxy/0.log" Oct 14 10:33:09 crc kubenswrapper[4870]: I1014 10:33:09.870287 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-55b6b7c7b8-jbt9q_5ad2a6ec-e2aa-45d9-b1a8-59ed908aaacd/manager/0.log" Oct 14 10:33:09 crc kubenswrapper[4870]: I1014 10:33:09.911204 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-5f67fbc655-szjkl_4bc1fceb-2d98-415e-b34e-82c2d6e1430f/manager/0.log" Oct 14 10:33:10 crc kubenswrapper[4870]: I1014 10:33:10.054956 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-f9fb45f8f-kvmkj_23eadd14-852e-4683-bae5-3720a36d9407/kube-rbac-proxy/0.log" Oct 14 10:33:10 crc kubenswrapper[4870]: I1014 10:33:10.124644 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-f9fb45f8f-kvmkj_23eadd14-852e-4683-bae5-3720a36d9407/manager/0.log" Oct 14 10:33:10 crc kubenswrapper[4870]: I1014 10:33:10.230015 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-79d585cb66-67fkk_586d846b-21c2-40f4-8da4-37392d812645/kube-rbac-proxy/0.log" Oct 14 10:33:10 crc kubenswrapper[4870]: I1014 10:33:10.314471 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-79d585cb66-67fkk_586d846b-21c2-40f4-8da4-37392d812645/manager/0.log" Oct 14 10:33:10 crc kubenswrapper[4870]: I1014 10:33:10.411772 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-5df598886f-zj4rk_34297fe0-c74f-4d69-8847-86346bfbea4f/kube-rbac-proxy/0.log" Oct 14 10:33:10 crc kubenswrapper[4870]: I1014 10:33:10.543119 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-69fdcfc5f5-2schz_34891227-4e7a-43d5-8dc0-3f56b54bb9fd/kube-rbac-proxy/0.log" Oct 14 10:33:10 crc kubenswrapper[4870]: I1014 10:33:10.714752 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-5df598886f-zj4rk_34297fe0-c74f-4d69-8847-86346bfbea4f/manager/0.log" Oct 14 10:33:10 crc kubenswrapper[4870]: I1014 10:33:10.718933 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-69fdcfc5f5-2schz_34891227-4e7a-43d5-8dc0-3f56b54bb9fd/manager/0.log" Oct 14 10:33:10 crc kubenswrapper[4870]: I1014 10:33:10.841044 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-55b7d4484882chz_3278d66b-0ec7-47b4-8499-1671fc13f4ee/kube-rbac-proxy/0.log" Oct 14 10:33:10 crc kubenswrapper[4870]: I1014 10:33:10.935853 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-55b7d4484882chz_3278d66b-0ec7-47b4-8499-1671fc13f4ee/manager/0.log" Oct 14 10:33:10 crc kubenswrapper[4870]: I1014 10:33:10.998947 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-7fb8c88b76-jzcb4_3645fdee-7690-4d37-a0f8-16ea440d9453/kube-rbac-proxy/0.log" Oct 14 10:33:11 crc kubenswrapper[4870]: I1014 10:33:11.267701 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-64895cd698-r8hz4_f7122ea8-dbd9-4f40-9eca-401b4a189140/kube-rbac-proxy/0.log" Oct 14 10:33:11 crc kubenswrapper[4870]: I1014 10:33:11.358022 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-64895cd698-r8hz4_f7122ea8-dbd9-4f40-9eca-401b4a189140/operator/0.log" Oct 14 10:33:11 crc kubenswrapper[4870]: I1014 10:33:11.591105 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-79df5fb58c-s54lf_9916d05d-8aa7-4e71-9ea6-07f2847fc4e6/kube-rbac-proxy/0.log" Oct 14 10:33:11 crc kubenswrapper[4870]: I1014 10:33:11.907197 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-79df5fb58c-s54lf_9916d05d-8aa7-4e71-9ea6-07f2847fc4e6/manager/0.log" Oct 14 10:33:11 crc kubenswrapper[4870]: I1014 10:33:11.908283 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-wfxzz_18d94d36-8f20-4399-b74e-29d831b2ffe8/registry-server/0.log" Oct 14 10:33:11 crc kubenswrapper[4870]: I1014 10:33:11.948763 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-68b6c87b68-dtn8j_96effd40-d810-4b88-9dec-a6e98993778f/kube-rbac-proxy/0.log" Oct 14 10:33:12 crc kubenswrapper[4870]: I1014 10:33:12.134239 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-68b6c87b68-dtn8j_96effd40-d810-4b88-9dec-a6e98993778f/manager/0.log" Oct 14 10:33:12 crc kubenswrapper[4870]: I1014 10:33:12.209287 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-5f97d8c699-lgqbd_de2a07da-ed81-4aac-bf11-da6b54ad095a/operator/0.log" Oct 14 10:33:12 crc kubenswrapper[4870]: I1014 10:33:12.383816 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-db6d7f97b-22qw9_c2b45ced-e9c2-4d6f-ac5e-1960e5a9296a/manager/0.log" Oct 14 10:33:12 crc kubenswrapper[4870]: I1014 10:33:12.398761 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-db6d7f97b-22qw9_c2b45ced-e9c2-4d6f-ac5e-1960e5a9296a/kube-rbac-proxy/0.log" Oct 14 10:33:12 crc kubenswrapper[4870]: I1014 10:33:12.481587 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-67cfc6749b-cblnh_01663eb4-cb0e-41fe-92c9-97e908f4ddff/kube-rbac-proxy/0.log" Oct 14 10:33:12 crc kubenswrapper[4870]: I1014 10:33:12.713797 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5458f77c4-n4csp_f47c02fb-6e94-4f5c-873f-ecaacd522121/manager/0.log" Oct 14 10:33:12 crc kubenswrapper[4870]: I1014 10:33:12.742110 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5458f77c4-n4csp_f47c02fb-6e94-4f5c-873f-ecaacd522121/kube-rbac-proxy/0.log" Oct 14 10:33:12 crc kubenswrapper[4870]: I1014 10:33:12.952817 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-7f554bff7b-68d4q_a1711137-ad44-4da0-ab90-5666fd111419/kube-rbac-proxy/0.log" Oct 14 10:33:13 crc kubenswrapper[4870]: I1014 10:33:13.004075 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-7f554bff7b-68d4q_a1711137-ad44-4da0-ab90-5666fd111419/manager/0.log" Oct 14 10:33:13 crc kubenswrapper[4870]: I1014 10:33:13.061500 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-67cfc6749b-cblnh_01663eb4-cb0e-41fe-92c9-97e908f4ddff/manager/0.log" Oct 14 10:33:14 crc kubenswrapper[4870]: I1014 10:33:14.042003 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-7fb8c88b76-jzcb4_3645fdee-7690-4d37-a0f8-16ea440d9453/manager/0.log" Oct 14 10:33:15 crc kubenswrapper[4870]: I1014 10:33:15.035511 4870 scope.go:117] "RemoveContainer" containerID="3a25412c639a58b6345a6cb926f93c78a7a62c68cd923f500f921497a90b6efb" Oct 14 10:33:15 crc kubenswrapper[4870]: E1014 10:33:15.035927 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 10:33:27 crc kubenswrapper[4870]: I1014 10:33:27.033954 4870 scope.go:117] "RemoveContainer" containerID="3a25412c639a58b6345a6cb926f93c78a7a62c68cd923f500f921497a90b6efb" Oct 14 10:33:27 crc kubenswrapper[4870]: E1014 10:33:27.034747 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 10:33:29 crc kubenswrapper[4870]: I1014 10:33:29.406349 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-6vdcc_8f610aec-a4a3-4d54-b805-4f01c0de5dce/control-plane-machine-set-operator/0.log" Oct 14 10:33:29 crc kubenswrapper[4870]: I1014 10:33:29.566317 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-nc79f_172ac2db-a387-4718-9874-fc770b806a17/kube-rbac-proxy/0.log" Oct 14 10:33:29 crc kubenswrapper[4870]: I1014 10:33:29.567961 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-nc79f_172ac2db-a387-4718-9874-fc770b806a17/machine-api-operator/0.log" Oct 14 10:33:40 crc kubenswrapper[4870]: I1014 10:33:40.034625 4870 scope.go:117] "RemoveContainer" containerID="3a25412c639a58b6345a6cb926f93c78a7a62c68cd923f500f921497a90b6efb" Oct 14 10:33:40 crc kubenswrapper[4870]: E1014 10:33:40.035556 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 10:33:42 crc kubenswrapper[4870]: I1014 10:33:42.759700 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-7d4cc89fcb-gvlll_d8377a33-a023-456d-811a-53be669f1178/cert-manager-controller/0.log" Oct 14 10:33:42 crc kubenswrapper[4870]: I1014 10:33:42.953386 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7d9f95dbf-ddfsr_06432b45-ce1d-4ace-b390-476fc76b06dc/cert-manager-cainjector/0.log" Oct 14 10:33:42 crc kubenswrapper[4870]: I1014 10:33:42.965389 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-d969966f-wfx5l_6ae15c75-9b7b-4583-8ac9-f60db231ae06/cert-manager-webhook/0.log" Oct 14 10:33:53 crc kubenswrapper[4870]: I1014 10:33:53.035341 4870 scope.go:117] "RemoveContainer" containerID="3a25412c639a58b6345a6cb926f93c78a7a62c68cd923f500f921497a90b6efb" Oct 14 10:33:53 crc kubenswrapper[4870]: E1014 10:33:53.036233 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 10:33:56 crc kubenswrapper[4870]: I1014 10:33:56.487456 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-6b874cbd85-w5zvx_9292dbae-ed2d-4a25-a889-b951b36614c6/nmstate-console-plugin/0.log" Oct 14 10:33:56 crc kubenswrapper[4870]: I1014 10:33:56.661184 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-fdff9cb8d-269h6_e4f05978-9de8-4cad-89d0-8620fbe004a3/kube-rbac-proxy/0.log" Oct 14 10:33:56 crc kubenswrapper[4870]: I1014 10:33:56.680026 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-ffdgs_65450853-71f4-44b7-bb21-1e5c650de334/nmstate-handler/0.log" Oct 14 10:33:56 crc kubenswrapper[4870]: I1014 10:33:56.732195 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-fdff9cb8d-269h6_e4f05978-9de8-4cad-89d0-8620fbe004a3/nmstate-metrics/0.log" Oct 14 10:33:56 crc kubenswrapper[4870]: I1014 10:33:56.870230 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-858ddd8f98-q77sx_98eac063-80b2-46f7-8096-d384fc10d632/nmstate-operator/0.log" Oct 14 10:33:56 crc kubenswrapper[4870]: I1014 10:33:56.955281 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-6cdbc54649-5gj9z_78703da9-8ebd-4610-9d48-61c17c1eddeb/nmstate-webhook/0.log" Oct 14 10:34:06 crc kubenswrapper[4870]: I1014 10:34:06.033503 4870 scope.go:117] "RemoveContainer" containerID="3a25412c639a58b6345a6cb926f93c78a7a62c68cd923f500f921497a90b6efb" Oct 14 10:34:06 crc kubenswrapper[4870]: E1014 10:34:06.034219 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 10:34:12 crc kubenswrapper[4870]: I1014 10:34:12.864245 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-68d546b9d8-bs82b_c7f0d440-b7cc-440a-9aac-e19c2bafab73/kube-rbac-proxy/0.log" Oct 14 10:34:13 crc kubenswrapper[4870]: I1014 10:34:13.130282 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-pbtkp_2f058085-57d8-48a8-9267-fde99e4c15fb/cp-frr-files/0.log" Oct 14 10:34:13 crc kubenswrapper[4870]: I1014 10:34:13.297723 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-pbtkp_2f058085-57d8-48a8-9267-fde99e4c15fb/cp-frr-files/0.log" Oct 14 10:34:13 crc kubenswrapper[4870]: I1014 10:34:13.391738 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-pbtkp_2f058085-57d8-48a8-9267-fde99e4c15fb/cp-reloader/0.log" Oct 14 10:34:13 crc kubenswrapper[4870]: I1014 10:34:13.406733 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-pbtkp_2f058085-57d8-48a8-9267-fde99e4c15fb/cp-metrics/0.log" Oct 14 10:34:13 crc kubenswrapper[4870]: I1014 10:34:13.457420 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-68d546b9d8-bs82b_c7f0d440-b7cc-440a-9aac-e19c2bafab73/controller/0.log" Oct 14 10:34:13 crc kubenswrapper[4870]: I1014 10:34:13.530635 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-pbtkp_2f058085-57d8-48a8-9267-fde99e4c15fb/cp-reloader/0.log" Oct 14 10:34:13 crc kubenswrapper[4870]: I1014 10:34:13.700170 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-pbtkp_2f058085-57d8-48a8-9267-fde99e4c15fb/cp-reloader/0.log" Oct 14 10:34:13 crc kubenswrapper[4870]: I1014 10:34:13.706063 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-pbtkp_2f058085-57d8-48a8-9267-fde99e4c15fb/cp-metrics/0.log" Oct 14 10:34:13 crc kubenswrapper[4870]: I1014 10:34:13.707135 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-pbtkp_2f058085-57d8-48a8-9267-fde99e4c15fb/cp-metrics/0.log" Oct 14 10:34:13 crc kubenswrapper[4870]: I1014 10:34:13.763410 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-pbtkp_2f058085-57d8-48a8-9267-fde99e4c15fb/cp-frr-files/0.log" Oct 14 10:34:13 crc kubenswrapper[4870]: I1014 10:34:13.895132 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-pbtkp_2f058085-57d8-48a8-9267-fde99e4c15fb/cp-frr-files/0.log" Oct 14 10:34:13 crc kubenswrapper[4870]: I1014 10:34:13.895186 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-pbtkp_2f058085-57d8-48a8-9267-fde99e4c15fb/cp-reloader/0.log" Oct 14 10:34:13 crc kubenswrapper[4870]: I1014 10:34:13.968083 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-pbtkp_2f058085-57d8-48a8-9267-fde99e4c15fb/cp-metrics/0.log" Oct 14 10:34:13 crc kubenswrapper[4870]: I1014 10:34:13.982025 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-pbtkp_2f058085-57d8-48a8-9267-fde99e4c15fb/controller/0.log" Oct 14 10:34:14 crc kubenswrapper[4870]: I1014 10:34:14.104235 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-pbtkp_2f058085-57d8-48a8-9267-fde99e4c15fb/frr-metrics/0.log" Oct 14 10:34:14 crc kubenswrapper[4870]: I1014 10:34:14.137032 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-pbtkp_2f058085-57d8-48a8-9267-fde99e4c15fb/kube-rbac-proxy/0.log" Oct 14 10:34:14 crc kubenswrapper[4870]: I1014 10:34:14.247072 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-pbtkp_2f058085-57d8-48a8-9267-fde99e4c15fb/kube-rbac-proxy-frr/0.log" Oct 14 10:34:14 crc kubenswrapper[4870]: I1014 10:34:14.296068 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-pbtkp_2f058085-57d8-48a8-9267-fde99e4c15fb/reloader/0.log" Oct 14 10:34:14 crc kubenswrapper[4870]: I1014 10:34:14.449218 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-64bf5d555-f8wq5_d0a20afe-74ac-4d3f-80ca-981fb5ce00a8/frr-k8s-webhook-server/0.log" Oct 14 10:34:14 crc kubenswrapper[4870]: I1014 10:34:14.623396 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-f459659f5-vgrhm_1150555d-bbee-49bc-b90c-69e30c971a19/manager/0.log" Oct 14 10:34:14 crc kubenswrapper[4870]: I1014 10:34:14.724860 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-7544c98cdc-r5nkv_c1cd26c0-7d04-45fb-b171-d3398fcf8d24/webhook-server/0.log" Oct 14 10:34:14 crc kubenswrapper[4870]: I1014 10:34:14.825511 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-mtvtd_010032d8-4503-4fb5-8158-066872d97d52/kube-rbac-proxy/0.log" Oct 14 10:34:15 crc kubenswrapper[4870]: I1014 10:34:15.800007 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-mtvtd_010032d8-4503-4fb5-8158-066872d97d52/speaker/0.log" Oct 14 10:34:17 crc kubenswrapper[4870]: I1014 10:34:17.463521 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-pbtkp_2f058085-57d8-48a8-9267-fde99e4c15fb/frr/0.log" Oct 14 10:34:18 crc kubenswrapper[4870]: I1014 10:34:18.154320 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-2lz4z"] Oct 14 10:34:18 crc kubenswrapper[4870]: E1014 10:34:18.154867 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="31e995c2-04e9-49d3-a47d-a771beee2a10" containerName="collect-profiles" Oct 14 10:34:18 crc kubenswrapper[4870]: I1014 10:34:18.154894 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="31e995c2-04e9-49d3-a47d-a771beee2a10" containerName="collect-profiles" Oct 14 10:34:18 crc kubenswrapper[4870]: I1014 10:34:18.155238 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="31e995c2-04e9-49d3-a47d-a771beee2a10" containerName="collect-profiles" Oct 14 10:34:18 crc kubenswrapper[4870]: I1014 10:34:18.157271 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2lz4z" Oct 14 10:34:18 crc kubenswrapper[4870]: I1014 10:34:18.167063 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-2lz4z"] Oct 14 10:34:18 crc kubenswrapper[4870]: I1014 10:34:18.259804 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/50031582-0241-47f8-a345-cd8757b6bee1-utilities\") pod \"certified-operators-2lz4z\" (UID: \"50031582-0241-47f8-a345-cd8757b6bee1\") " pod="openshift-marketplace/certified-operators-2lz4z" Oct 14 10:34:18 crc kubenswrapper[4870]: I1014 10:34:18.260207 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sl2jt\" (UniqueName: \"kubernetes.io/projected/50031582-0241-47f8-a345-cd8757b6bee1-kube-api-access-sl2jt\") pod \"certified-operators-2lz4z\" (UID: \"50031582-0241-47f8-a345-cd8757b6bee1\") " pod="openshift-marketplace/certified-operators-2lz4z" Oct 14 10:34:18 crc kubenswrapper[4870]: I1014 10:34:18.260266 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/50031582-0241-47f8-a345-cd8757b6bee1-catalog-content\") pod \"certified-operators-2lz4z\" (UID: \"50031582-0241-47f8-a345-cd8757b6bee1\") " pod="openshift-marketplace/certified-operators-2lz4z" Oct 14 10:34:18 crc kubenswrapper[4870]: I1014 10:34:18.361922 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/50031582-0241-47f8-a345-cd8757b6bee1-utilities\") pod \"certified-operators-2lz4z\" (UID: \"50031582-0241-47f8-a345-cd8757b6bee1\") " pod="openshift-marketplace/certified-operators-2lz4z" Oct 14 10:34:18 crc kubenswrapper[4870]: I1014 10:34:18.362025 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sl2jt\" (UniqueName: \"kubernetes.io/projected/50031582-0241-47f8-a345-cd8757b6bee1-kube-api-access-sl2jt\") pod \"certified-operators-2lz4z\" (UID: \"50031582-0241-47f8-a345-cd8757b6bee1\") " pod="openshift-marketplace/certified-operators-2lz4z" Oct 14 10:34:18 crc kubenswrapper[4870]: I1014 10:34:18.362078 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/50031582-0241-47f8-a345-cd8757b6bee1-catalog-content\") pod \"certified-operators-2lz4z\" (UID: \"50031582-0241-47f8-a345-cd8757b6bee1\") " pod="openshift-marketplace/certified-operators-2lz4z" Oct 14 10:34:18 crc kubenswrapper[4870]: I1014 10:34:18.362668 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/50031582-0241-47f8-a345-cd8757b6bee1-catalog-content\") pod \"certified-operators-2lz4z\" (UID: \"50031582-0241-47f8-a345-cd8757b6bee1\") " pod="openshift-marketplace/certified-operators-2lz4z" Oct 14 10:34:18 crc kubenswrapper[4870]: I1014 10:34:18.362938 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/50031582-0241-47f8-a345-cd8757b6bee1-utilities\") pod \"certified-operators-2lz4z\" (UID: \"50031582-0241-47f8-a345-cd8757b6bee1\") " pod="openshift-marketplace/certified-operators-2lz4z" Oct 14 10:34:18 crc kubenswrapper[4870]: I1014 10:34:18.395721 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sl2jt\" (UniqueName: \"kubernetes.io/projected/50031582-0241-47f8-a345-cd8757b6bee1-kube-api-access-sl2jt\") pod \"certified-operators-2lz4z\" (UID: \"50031582-0241-47f8-a345-cd8757b6bee1\") " pod="openshift-marketplace/certified-operators-2lz4z" Oct 14 10:34:18 crc kubenswrapper[4870]: I1014 10:34:18.491554 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2lz4z" Oct 14 10:34:19 crc kubenswrapper[4870]: I1014 10:34:19.034535 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-2lz4z"] Oct 14 10:34:19 crc kubenswrapper[4870]: I1014 10:34:19.534686 4870 generic.go:334] "Generic (PLEG): container finished" podID="50031582-0241-47f8-a345-cd8757b6bee1" containerID="a5394393128ebac03dcd8f58aa4626f4cf3366a2e5ae4f9d5f86550b3c74d5f7" exitCode=0 Oct 14 10:34:19 crc kubenswrapper[4870]: I1014 10:34:19.534782 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2lz4z" event={"ID":"50031582-0241-47f8-a345-cd8757b6bee1","Type":"ContainerDied","Data":"a5394393128ebac03dcd8f58aa4626f4cf3366a2e5ae4f9d5f86550b3c74d5f7"} Oct 14 10:34:19 crc kubenswrapper[4870]: I1014 10:34:19.535012 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2lz4z" event={"ID":"50031582-0241-47f8-a345-cd8757b6bee1","Type":"ContainerStarted","Data":"f7fa9540f7f200c9e17b3af66fa24831c6c1b06dc8b54e64d504a5ae72b148f3"} Oct 14 10:34:19 crc kubenswrapper[4870]: I1014 10:34:19.537227 4870 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 14 10:34:20 crc kubenswrapper[4870]: I1014 10:34:20.547048 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2lz4z" event={"ID":"50031582-0241-47f8-a345-cd8757b6bee1","Type":"ContainerStarted","Data":"8dbd38403269dbaed9fdd05bdc6e2d501a5583e2ae5f0f9a01b3ca924669f33f"} Oct 14 10:34:21 crc kubenswrapper[4870]: I1014 10:34:21.034720 4870 scope.go:117] "RemoveContainer" containerID="3a25412c639a58b6345a6cb926f93c78a7a62c68cd923f500f921497a90b6efb" Oct 14 10:34:21 crc kubenswrapper[4870]: E1014 10:34:21.035168 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 10:34:21 crc kubenswrapper[4870]: I1014 10:34:21.565915 4870 generic.go:334] "Generic (PLEG): container finished" podID="50031582-0241-47f8-a345-cd8757b6bee1" containerID="8dbd38403269dbaed9fdd05bdc6e2d501a5583e2ae5f0f9a01b3ca924669f33f" exitCode=0 Oct 14 10:34:21 crc kubenswrapper[4870]: I1014 10:34:21.565984 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2lz4z" event={"ID":"50031582-0241-47f8-a345-cd8757b6bee1","Type":"ContainerDied","Data":"8dbd38403269dbaed9fdd05bdc6e2d501a5583e2ae5f0f9a01b3ca924669f33f"} Oct 14 10:34:22 crc kubenswrapper[4870]: I1014 10:34:22.580151 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2lz4z" event={"ID":"50031582-0241-47f8-a345-cd8757b6bee1","Type":"ContainerStarted","Data":"0b7bd782baca100bd8eb9563e7a5a19a5aae8702a140f45869e0f93e573c1658"} Oct 14 10:34:22 crc kubenswrapper[4870]: I1014 10:34:22.611064 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-2lz4z" podStartSLOduration=2.161237185 podStartE2EDuration="4.611044143s" podCreationTimestamp="2025-10-14 10:34:18 +0000 UTC" firstStartedPulling="2025-10-14 10:34:19.536954031 +0000 UTC m=+12795.234314402" lastFinishedPulling="2025-10-14 10:34:21.986760989 +0000 UTC m=+12797.684121360" observedRunningTime="2025-10-14 10:34:22.600412953 +0000 UTC m=+12798.297773324" watchObservedRunningTime="2025-10-14 10:34:22.611044143 +0000 UTC m=+12798.308404514" Oct 14 10:34:28 crc kubenswrapper[4870]: I1014 10:34:28.491837 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-2lz4z" Oct 14 10:34:28 crc kubenswrapper[4870]: I1014 10:34:28.493725 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-2lz4z" Oct 14 10:34:28 crc kubenswrapper[4870]: I1014 10:34:28.543576 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-2lz4z" Oct 14 10:34:28 crc kubenswrapper[4870]: I1014 10:34:28.702879 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-2lz4z" Oct 14 10:34:28 crc kubenswrapper[4870]: I1014 10:34:28.786127 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-2lz4z"] Oct 14 10:34:29 crc kubenswrapper[4870]: I1014 10:34:29.641398 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69dkdfp_ebb7cfd0-0388-4aca-99f6-7669c6bc8d4e/util/0.log" Oct 14 10:34:29 crc kubenswrapper[4870]: I1014 10:34:29.790132 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69dkdfp_ebb7cfd0-0388-4aca-99f6-7669c6bc8d4e/util/0.log" Oct 14 10:34:29 crc kubenswrapper[4870]: I1014 10:34:29.840852 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69dkdfp_ebb7cfd0-0388-4aca-99f6-7669c6bc8d4e/pull/0.log" Oct 14 10:34:29 crc kubenswrapper[4870]: I1014 10:34:29.875657 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69dkdfp_ebb7cfd0-0388-4aca-99f6-7669c6bc8d4e/pull/0.log" Oct 14 10:34:30 crc kubenswrapper[4870]: I1014 10:34:30.081013 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69dkdfp_ebb7cfd0-0388-4aca-99f6-7669c6bc8d4e/pull/0.log" Oct 14 10:34:30 crc kubenswrapper[4870]: I1014 10:34:30.116591 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69dkdfp_ebb7cfd0-0388-4aca-99f6-7669c6bc8d4e/extract/0.log" Oct 14 10:34:30 crc kubenswrapper[4870]: I1014 10:34:30.124015 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69dkdfp_ebb7cfd0-0388-4aca-99f6-7669c6bc8d4e/util/0.log" Oct 14 10:34:30 crc kubenswrapper[4870]: I1014 10:34:30.272725 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2lsrmg_82fe0e08-918e-44f3-a5ac-940278d1e9c3/util/0.log" Oct 14 10:34:30 crc kubenswrapper[4870]: I1014 10:34:30.519204 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2lsrmg_82fe0e08-918e-44f3-a5ac-940278d1e9c3/util/0.log" Oct 14 10:34:30 crc kubenswrapper[4870]: I1014 10:34:30.527262 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2lsrmg_82fe0e08-918e-44f3-a5ac-940278d1e9c3/pull/0.log" Oct 14 10:34:30 crc kubenswrapper[4870]: I1014 10:34:30.536973 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2lsrmg_82fe0e08-918e-44f3-a5ac-940278d1e9c3/pull/0.log" Oct 14 10:34:30 crc kubenswrapper[4870]: I1014 10:34:30.657030 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-2lz4z" podUID="50031582-0241-47f8-a345-cd8757b6bee1" containerName="registry-server" containerID="cri-o://0b7bd782baca100bd8eb9563e7a5a19a5aae8702a140f45869e0f93e573c1658" gracePeriod=2 Oct 14 10:34:30 crc kubenswrapper[4870]: I1014 10:34:30.758493 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2lsrmg_82fe0e08-918e-44f3-a5ac-940278d1e9c3/extract/0.log" Oct 14 10:34:30 crc kubenswrapper[4870]: I1014 10:34:30.793221 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2lsrmg_82fe0e08-918e-44f3-a5ac-940278d1e9c3/pull/0.log" Oct 14 10:34:30 crc kubenswrapper[4870]: I1014 10:34:30.826232 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2lsrmg_82fe0e08-918e-44f3-a5ac-940278d1e9c3/util/0.log" Oct 14 10:34:31 crc kubenswrapper[4870]: I1014 10:34:31.080319 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dg96wq_b87acf8e-2f81-484d-99d4-eeb3e1278895/util/0.log" Oct 14 10:34:31 crc kubenswrapper[4870]: I1014 10:34:31.183876 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2lz4z" Oct 14 10:34:31 crc kubenswrapper[4870]: I1014 10:34:31.340254 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dg96wq_b87acf8e-2f81-484d-99d4-eeb3e1278895/pull/0.log" Oct 14 10:34:31 crc kubenswrapper[4870]: I1014 10:34:31.365240 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/50031582-0241-47f8-a345-cd8757b6bee1-utilities\") pod \"50031582-0241-47f8-a345-cd8757b6bee1\" (UID: \"50031582-0241-47f8-a345-cd8757b6bee1\") " Oct 14 10:34:31 crc kubenswrapper[4870]: I1014 10:34:31.365291 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/50031582-0241-47f8-a345-cd8757b6bee1-catalog-content\") pod \"50031582-0241-47f8-a345-cd8757b6bee1\" (UID: \"50031582-0241-47f8-a345-cd8757b6bee1\") " Oct 14 10:34:31 crc kubenswrapper[4870]: I1014 10:34:31.365388 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sl2jt\" (UniqueName: \"kubernetes.io/projected/50031582-0241-47f8-a345-cd8757b6bee1-kube-api-access-sl2jt\") pod \"50031582-0241-47f8-a345-cd8757b6bee1\" (UID: \"50031582-0241-47f8-a345-cd8757b6bee1\") " Oct 14 10:34:31 crc kubenswrapper[4870]: I1014 10:34:31.370025 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/50031582-0241-47f8-a345-cd8757b6bee1-utilities" (OuterVolumeSpecName: "utilities") pod "50031582-0241-47f8-a345-cd8757b6bee1" (UID: "50031582-0241-47f8-a345-cd8757b6bee1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 10:34:31 crc kubenswrapper[4870]: I1014 10:34:31.375611 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/50031582-0241-47f8-a345-cd8757b6bee1-kube-api-access-sl2jt" (OuterVolumeSpecName: "kube-api-access-sl2jt") pod "50031582-0241-47f8-a345-cd8757b6bee1" (UID: "50031582-0241-47f8-a345-cd8757b6bee1"). InnerVolumeSpecName "kube-api-access-sl2jt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 10:34:31 crc kubenswrapper[4870]: I1014 10:34:31.401899 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dg96wq_b87acf8e-2f81-484d-99d4-eeb3e1278895/pull/0.log" Oct 14 10:34:31 crc kubenswrapper[4870]: I1014 10:34:31.422756 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/50031582-0241-47f8-a345-cd8757b6bee1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "50031582-0241-47f8-a345-cd8757b6bee1" (UID: "50031582-0241-47f8-a345-cd8757b6bee1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 10:34:31 crc kubenswrapper[4870]: I1014 10:34:31.467835 4870 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/50031582-0241-47f8-a345-cd8757b6bee1-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 10:34:31 crc kubenswrapper[4870]: I1014 10:34:31.467876 4870 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/50031582-0241-47f8-a345-cd8757b6bee1-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 10:34:31 crc kubenswrapper[4870]: I1014 10:34:31.467890 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sl2jt\" (UniqueName: \"kubernetes.io/projected/50031582-0241-47f8-a345-cd8757b6bee1-kube-api-access-sl2jt\") on node \"crc\" DevicePath \"\"" Oct 14 10:34:31 crc kubenswrapper[4870]: I1014 10:34:31.477162 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dg96wq_b87acf8e-2f81-484d-99d4-eeb3e1278895/util/0.log" Oct 14 10:34:31 crc kubenswrapper[4870]: I1014 10:34:31.590864 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dg96wq_b87acf8e-2f81-484d-99d4-eeb3e1278895/util/0.log" Oct 14 10:34:31 crc kubenswrapper[4870]: I1014 10:34:31.614112 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dg96wq_b87acf8e-2f81-484d-99d4-eeb3e1278895/pull/0.log" Oct 14 10:34:31 crc kubenswrapper[4870]: I1014 10:34:31.675457 4870 generic.go:334] "Generic (PLEG): container finished" podID="50031582-0241-47f8-a345-cd8757b6bee1" containerID="0b7bd782baca100bd8eb9563e7a5a19a5aae8702a140f45869e0f93e573c1658" exitCode=0 Oct 14 10:34:31 crc kubenswrapper[4870]: I1014 10:34:31.675525 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2lz4z" event={"ID":"50031582-0241-47f8-a345-cd8757b6bee1","Type":"ContainerDied","Data":"0b7bd782baca100bd8eb9563e7a5a19a5aae8702a140f45869e0f93e573c1658"} Oct 14 10:34:31 crc kubenswrapper[4870]: I1014 10:34:31.675560 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2lz4z" event={"ID":"50031582-0241-47f8-a345-cd8757b6bee1","Type":"ContainerDied","Data":"f7fa9540f7f200c9e17b3af66fa24831c6c1b06dc8b54e64d504a5ae72b148f3"} Oct 14 10:34:31 crc kubenswrapper[4870]: I1014 10:34:31.675583 4870 scope.go:117] "RemoveContainer" containerID="0b7bd782baca100bd8eb9563e7a5a19a5aae8702a140f45869e0f93e573c1658" Oct 14 10:34:31 crc kubenswrapper[4870]: I1014 10:34:31.675761 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2lz4z" Oct 14 10:34:31 crc kubenswrapper[4870]: I1014 10:34:31.683226 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dg96wq_b87acf8e-2f81-484d-99d4-eeb3e1278895/extract/0.log" Oct 14 10:34:31 crc kubenswrapper[4870]: I1014 10:34:31.733953 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-2lz4z"] Oct 14 10:34:31 crc kubenswrapper[4870]: I1014 10:34:31.757097 4870 scope.go:117] "RemoveContainer" containerID="8dbd38403269dbaed9fdd05bdc6e2d501a5583e2ae5f0f9a01b3ca924669f33f" Oct 14 10:34:31 crc kubenswrapper[4870]: I1014 10:34:31.776764 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-2lz4z"] Oct 14 10:34:31 crc kubenswrapper[4870]: I1014 10:34:31.800667 4870 scope.go:117] "RemoveContainer" containerID="a5394393128ebac03dcd8f58aa4626f4cf3366a2e5ae4f9d5f86550b3c74d5f7" Oct 14 10:34:31 crc kubenswrapper[4870]: I1014 10:34:31.843278 4870 scope.go:117] "RemoveContainer" containerID="0b7bd782baca100bd8eb9563e7a5a19a5aae8702a140f45869e0f93e573c1658" Oct 14 10:34:31 crc kubenswrapper[4870]: E1014 10:34:31.844344 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0b7bd782baca100bd8eb9563e7a5a19a5aae8702a140f45869e0f93e573c1658\": container with ID starting with 0b7bd782baca100bd8eb9563e7a5a19a5aae8702a140f45869e0f93e573c1658 not found: ID does not exist" containerID="0b7bd782baca100bd8eb9563e7a5a19a5aae8702a140f45869e0f93e573c1658" Oct 14 10:34:31 crc kubenswrapper[4870]: I1014 10:34:31.844378 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0b7bd782baca100bd8eb9563e7a5a19a5aae8702a140f45869e0f93e573c1658"} err="failed to get container status \"0b7bd782baca100bd8eb9563e7a5a19a5aae8702a140f45869e0f93e573c1658\": rpc error: code = NotFound desc = could not find container \"0b7bd782baca100bd8eb9563e7a5a19a5aae8702a140f45869e0f93e573c1658\": container with ID starting with 0b7bd782baca100bd8eb9563e7a5a19a5aae8702a140f45869e0f93e573c1658 not found: ID does not exist" Oct 14 10:34:31 crc kubenswrapper[4870]: I1014 10:34:31.844420 4870 scope.go:117] "RemoveContainer" containerID="8dbd38403269dbaed9fdd05bdc6e2d501a5583e2ae5f0f9a01b3ca924669f33f" Oct 14 10:34:31 crc kubenswrapper[4870]: E1014 10:34:31.844771 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8dbd38403269dbaed9fdd05bdc6e2d501a5583e2ae5f0f9a01b3ca924669f33f\": container with ID starting with 8dbd38403269dbaed9fdd05bdc6e2d501a5583e2ae5f0f9a01b3ca924669f33f not found: ID does not exist" containerID="8dbd38403269dbaed9fdd05bdc6e2d501a5583e2ae5f0f9a01b3ca924669f33f" Oct 14 10:34:31 crc kubenswrapper[4870]: I1014 10:34:31.844798 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8dbd38403269dbaed9fdd05bdc6e2d501a5583e2ae5f0f9a01b3ca924669f33f"} err="failed to get container status \"8dbd38403269dbaed9fdd05bdc6e2d501a5583e2ae5f0f9a01b3ca924669f33f\": rpc error: code = NotFound desc = could not find container \"8dbd38403269dbaed9fdd05bdc6e2d501a5583e2ae5f0f9a01b3ca924669f33f\": container with ID starting with 8dbd38403269dbaed9fdd05bdc6e2d501a5583e2ae5f0f9a01b3ca924669f33f not found: ID does not exist" Oct 14 10:34:31 crc kubenswrapper[4870]: I1014 10:34:31.844813 4870 scope.go:117] "RemoveContainer" containerID="a5394393128ebac03dcd8f58aa4626f4cf3366a2e5ae4f9d5f86550b3c74d5f7" Oct 14 10:34:31 crc kubenswrapper[4870]: E1014 10:34:31.845148 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a5394393128ebac03dcd8f58aa4626f4cf3366a2e5ae4f9d5f86550b3c74d5f7\": container with ID starting with a5394393128ebac03dcd8f58aa4626f4cf3366a2e5ae4f9d5f86550b3c74d5f7 not found: ID does not exist" containerID="a5394393128ebac03dcd8f58aa4626f4cf3366a2e5ae4f9d5f86550b3c74d5f7" Oct 14 10:34:31 crc kubenswrapper[4870]: I1014 10:34:31.845173 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a5394393128ebac03dcd8f58aa4626f4cf3366a2e5ae4f9d5f86550b3c74d5f7"} err="failed to get container status \"a5394393128ebac03dcd8f58aa4626f4cf3366a2e5ae4f9d5f86550b3c74d5f7\": rpc error: code = NotFound desc = could not find container \"a5394393128ebac03dcd8f58aa4626f4cf3366a2e5ae4f9d5f86550b3c74d5f7\": container with ID starting with a5394393128ebac03dcd8f58aa4626f4cf3366a2e5ae4f9d5f86550b3c74d5f7 not found: ID does not exist" Oct 14 10:34:32 crc kubenswrapper[4870]: I1014 10:34:32.034035 4870 scope.go:117] "RemoveContainer" containerID="3a25412c639a58b6345a6cb926f93c78a7a62c68cd923f500f921497a90b6efb" Oct 14 10:34:32 crc kubenswrapper[4870]: I1014 10:34:32.250575 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-prqvc_e9f8d5e1-367a-4cb2-9119-b5e6ae554774/extract-utilities/0.log" Oct 14 10:34:32 crc kubenswrapper[4870]: I1014 10:34:32.530951 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-prqvc_e9f8d5e1-367a-4cb2-9119-b5e6ae554774/extract-content/0.log" Oct 14 10:34:32 crc kubenswrapper[4870]: I1014 10:34:32.563392 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-prqvc_e9f8d5e1-367a-4cb2-9119-b5e6ae554774/extract-utilities/0.log" Oct 14 10:34:32 crc kubenswrapper[4870]: I1014 10:34:32.580418 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-prqvc_e9f8d5e1-367a-4cb2-9119-b5e6ae554774/extract-content/0.log" Oct 14 10:34:32 crc kubenswrapper[4870]: I1014 10:34:32.687326 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" event={"ID":"874a5e45-dffd-4d17-b609-b7d3ed2eab07","Type":"ContainerStarted","Data":"3ecb7d8f7bc7b21ce0eb0f68bcda8198b66e9e2e6c18afa3bb02bf2b3ab41d20"} Oct 14 10:34:32 crc kubenswrapper[4870]: I1014 10:34:32.911798 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-prqvc_e9f8d5e1-367a-4cb2-9119-b5e6ae554774/extract-content/0.log" Oct 14 10:34:33 crc kubenswrapper[4870]: I1014 10:34:33.025398 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-prqvc_e9f8d5e1-367a-4cb2-9119-b5e6ae554774/extract-utilities/0.log" Oct 14 10:34:33 crc kubenswrapper[4870]: I1014 10:34:33.057755 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="50031582-0241-47f8-a345-cd8757b6bee1" path="/var/lib/kubelet/pods/50031582-0241-47f8-a345-cd8757b6bee1/volumes" Oct 14 10:34:33 crc kubenswrapper[4870]: I1014 10:34:33.236526 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-9qgf9_2baef9db-d79c-4f55-9c3f-cd4ac8488a8d/extract-utilities/0.log" Oct 14 10:34:33 crc kubenswrapper[4870]: I1014 10:34:33.727213 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-9qgf9_2baef9db-d79c-4f55-9c3f-cd4ac8488a8d/extract-content/0.log" Oct 14 10:34:33 crc kubenswrapper[4870]: I1014 10:34:33.770108 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-9qgf9_2baef9db-d79c-4f55-9c3f-cd4ac8488a8d/extract-content/0.log" Oct 14 10:34:33 crc kubenswrapper[4870]: I1014 10:34:33.783729 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-9qgf9_2baef9db-d79c-4f55-9c3f-cd4ac8488a8d/extract-utilities/0.log" Oct 14 10:34:34 crc kubenswrapper[4870]: I1014 10:34:34.123340 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-9qgf9_2baef9db-d79c-4f55-9c3f-cd4ac8488a8d/extract-content/0.log" Oct 14 10:34:34 crc kubenswrapper[4870]: I1014 10:34:34.125293 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-9qgf9_2baef9db-d79c-4f55-9c3f-cd4ac8488a8d/extract-utilities/0.log" Oct 14 10:34:34 crc kubenswrapper[4870]: I1014 10:34:34.370296 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cmmng9_54f5c716-17eb-4c40-88c3-95ec36d3f406/util/0.log" Oct 14 10:34:34 crc kubenswrapper[4870]: I1014 10:34:34.591626 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cmmng9_54f5c716-17eb-4c40-88c3-95ec36d3f406/util/0.log" Oct 14 10:34:34 crc kubenswrapper[4870]: I1014 10:34:34.642029 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cmmng9_54f5c716-17eb-4c40-88c3-95ec36d3f406/pull/0.log" Oct 14 10:34:34 crc kubenswrapper[4870]: I1014 10:34:34.704352 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cmmng9_54f5c716-17eb-4c40-88c3-95ec36d3f406/pull/0.log" Oct 14 10:34:34 crc kubenswrapper[4870]: I1014 10:34:34.714241 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-prqvc_e9f8d5e1-367a-4cb2-9119-b5e6ae554774/registry-server/0.log" Oct 14 10:34:34 crc kubenswrapper[4870]: I1014 10:34:34.864134 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cmmng9_54f5c716-17eb-4c40-88c3-95ec36d3f406/util/0.log" Oct 14 10:34:34 crc kubenswrapper[4870]: I1014 10:34:34.882202 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cmmng9_54f5c716-17eb-4c40-88c3-95ec36d3f406/pull/0.log" Oct 14 10:34:34 crc kubenswrapper[4870]: I1014 10:34:34.957232 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cmmng9_54f5c716-17eb-4c40-88c3-95ec36d3f406/extract/0.log" Oct 14 10:34:35 crc kubenswrapper[4870]: I1014 10:34:35.135375 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-7kgk8_1a1da733-c81b-4cdf-be2c-8f1997560247/marketplace-operator/0.log" Oct 14 10:34:35 crc kubenswrapper[4870]: I1014 10:34:35.195571 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-jqsnc_08bc1ff9-0341-4e01-93ec-25d41eebf646/extract-utilities/0.log" Oct 14 10:34:35 crc kubenswrapper[4870]: I1014 10:34:35.366186 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-jqsnc_08bc1ff9-0341-4e01-93ec-25d41eebf646/extract-content/0.log" Oct 14 10:34:35 crc kubenswrapper[4870]: I1014 10:34:35.416651 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-jqsnc_08bc1ff9-0341-4e01-93ec-25d41eebf646/extract-utilities/0.log" Oct 14 10:34:35 crc kubenswrapper[4870]: I1014 10:34:35.442578 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-jqsnc_08bc1ff9-0341-4e01-93ec-25d41eebf646/extract-content/0.log" Oct 14 10:34:35 crc kubenswrapper[4870]: I1014 10:34:35.679496 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-jqsnc_08bc1ff9-0341-4e01-93ec-25d41eebf646/extract-utilities/0.log" Oct 14 10:34:35 crc kubenswrapper[4870]: I1014 10:34:35.796941 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-jqsnc_08bc1ff9-0341-4e01-93ec-25d41eebf646/extract-content/0.log" Oct 14 10:34:35 crc kubenswrapper[4870]: I1014 10:34:35.884202 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-9qgf9_2baef9db-d79c-4f55-9c3f-cd4ac8488a8d/registry-server/0.log" Oct 14 10:34:35 crc kubenswrapper[4870]: I1014 10:34:35.997960 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-gdw8c_3f73653c-d771-4da5-a66d-aa5c974f2898/extract-utilities/0.log" Oct 14 10:34:36 crc kubenswrapper[4870]: I1014 10:34:36.200195 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-jqsnc_08bc1ff9-0341-4e01-93ec-25d41eebf646/registry-server/0.log" Oct 14 10:34:36 crc kubenswrapper[4870]: I1014 10:34:36.206819 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-gdw8c_3f73653c-d771-4da5-a66d-aa5c974f2898/extract-utilities/0.log" Oct 14 10:34:36 crc kubenswrapper[4870]: I1014 10:34:36.229944 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-gdw8c_3f73653c-d771-4da5-a66d-aa5c974f2898/extract-content/0.log" Oct 14 10:34:36 crc kubenswrapper[4870]: I1014 10:34:36.316982 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-gdw8c_3f73653c-d771-4da5-a66d-aa5c974f2898/extract-content/0.log" Oct 14 10:34:36 crc kubenswrapper[4870]: I1014 10:34:36.456058 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-gdw8c_3f73653c-d771-4da5-a66d-aa5c974f2898/extract-content/0.log" Oct 14 10:34:36 crc kubenswrapper[4870]: I1014 10:34:36.499503 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-gdw8c_3f73653c-d771-4da5-a66d-aa5c974f2898/extract-utilities/0.log" Oct 14 10:34:36 crc kubenswrapper[4870]: I1014 10:34:36.858217 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-gdw8c_3f73653c-d771-4da5-a66d-aa5c974f2898/registry-server/0.log" Oct 14 10:34:49 crc kubenswrapper[4870]: I1014 10:34:49.913907 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-7c8cf85677-fzq5k_ea8c24a3-71f9-43d5-8524-d3d95dd3b9f6/prometheus-operator/0.log" Oct 14 10:34:50 crc kubenswrapper[4870]: I1014 10:34:50.061210 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-769c7ff77b-ksw25_b0c9563b-efee-4741-88f3-b09a08320257/prometheus-operator-admission-webhook/0.log" Oct 14 10:34:50 crc kubenswrapper[4870]: I1014 10:34:50.145396 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-769c7ff77b-prngx_203800e4-7772-492d-942b-1175c02174e8/prometheus-operator-admission-webhook/0.log" Oct 14 10:34:50 crc kubenswrapper[4870]: I1014 10:34:50.285867 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_observability-operator-cc5f78dfc-xmgpw_8b4db008-9294-40b8-9b49-96273ac9627a/operator/0.log" Oct 14 10:34:50 crc kubenswrapper[4870]: I1014 10:34:50.345305 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_perses-operator-54bc95c9fb-ghvk7_c8bc8628-7639-4a34-b1f3-5de19a0614a8/perses-operator/0.log" Oct 14 10:36:53 crc kubenswrapper[4870]: I1014 10:36:53.951480 4870 patch_prober.go:28] interesting pod/machine-config-daemon-7tvc7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 10:36:53 crc kubenswrapper[4870]: I1014 10:36:53.954126 4870 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 10:37:23 crc kubenswrapper[4870]: I1014 10:37:23.951481 4870 patch_prober.go:28] interesting pod/machine-config-daemon-7tvc7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 10:37:23 crc kubenswrapper[4870]: I1014 10:37:23.952246 4870 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 10:37:25 crc kubenswrapper[4870]: I1014 10:37:25.307620 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-ntwkl"] Oct 14 10:37:25 crc kubenswrapper[4870]: E1014 10:37:25.308296 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="50031582-0241-47f8-a345-cd8757b6bee1" containerName="registry-server" Oct 14 10:37:25 crc kubenswrapper[4870]: I1014 10:37:25.308319 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="50031582-0241-47f8-a345-cd8757b6bee1" containerName="registry-server" Oct 14 10:37:25 crc kubenswrapper[4870]: E1014 10:37:25.308408 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="50031582-0241-47f8-a345-cd8757b6bee1" containerName="extract-utilities" Oct 14 10:37:25 crc kubenswrapper[4870]: I1014 10:37:25.308421 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="50031582-0241-47f8-a345-cd8757b6bee1" containerName="extract-utilities" Oct 14 10:37:25 crc kubenswrapper[4870]: E1014 10:37:25.308481 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="50031582-0241-47f8-a345-cd8757b6bee1" containerName="extract-content" Oct 14 10:37:25 crc kubenswrapper[4870]: I1014 10:37:25.308494 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="50031582-0241-47f8-a345-cd8757b6bee1" containerName="extract-content" Oct 14 10:37:25 crc kubenswrapper[4870]: I1014 10:37:25.308861 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="50031582-0241-47f8-a345-cd8757b6bee1" containerName="registry-server" Oct 14 10:37:25 crc kubenswrapper[4870]: I1014 10:37:25.311572 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ntwkl" Oct 14 10:37:25 crc kubenswrapper[4870]: I1014 10:37:25.323867 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-ntwkl"] Oct 14 10:37:25 crc kubenswrapper[4870]: I1014 10:37:25.383061 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c3c21d56-4c18-4f40-b061-37a18f4b3bd0-utilities\") pod \"community-operators-ntwkl\" (UID: \"c3c21d56-4c18-4f40-b061-37a18f4b3bd0\") " pod="openshift-marketplace/community-operators-ntwkl" Oct 14 10:37:25 crc kubenswrapper[4870]: I1014 10:37:25.383746 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c3c21d56-4c18-4f40-b061-37a18f4b3bd0-catalog-content\") pod \"community-operators-ntwkl\" (UID: \"c3c21d56-4c18-4f40-b061-37a18f4b3bd0\") " pod="openshift-marketplace/community-operators-ntwkl" Oct 14 10:37:25 crc kubenswrapper[4870]: I1014 10:37:25.383788 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fdb78\" (UniqueName: \"kubernetes.io/projected/c3c21d56-4c18-4f40-b061-37a18f4b3bd0-kube-api-access-fdb78\") pod \"community-operators-ntwkl\" (UID: \"c3c21d56-4c18-4f40-b061-37a18f4b3bd0\") " pod="openshift-marketplace/community-operators-ntwkl" Oct 14 10:37:25 crc kubenswrapper[4870]: I1014 10:37:25.485827 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c3c21d56-4c18-4f40-b061-37a18f4b3bd0-catalog-content\") pod \"community-operators-ntwkl\" (UID: \"c3c21d56-4c18-4f40-b061-37a18f4b3bd0\") " pod="openshift-marketplace/community-operators-ntwkl" Oct 14 10:37:25 crc kubenswrapper[4870]: I1014 10:37:25.485920 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fdb78\" (UniqueName: \"kubernetes.io/projected/c3c21d56-4c18-4f40-b061-37a18f4b3bd0-kube-api-access-fdb78\") pod \"community-operators-ntwkl\" (UID: \"c3c21d56-4c18-4f40-b061-37a18f4b3bd0\") " pod="openshift-marketplace/community-operators-ntwkl" Oct 14 10:37:25 crc kubenswrapper[4870]: I1014 10:37:25.485988 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c3c21d56-4c18-4f40-b061-37a18f4b3bd0-utilities\") pod \"community-operators-ntwkl\" (UID: \"c3c21d56-4c18-4f40-b061-37a18f4b3bd0\") " pod="openshift-marketplace/community-operators-ntwkl" Oct 14 10:37:25 crc kubenswrapper[4870]: I1014 10:37:25.486532 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c3c21d56-4c18-4f40-b061-37a18f4b3bd0-catalog-content\") pod \"community-operators-ntwkl\" (UID: \"c3c21d56-4c18-4f40-b061-37a18f4b3bd0\") " pod="openshift-marketplace/community-operators-ntwkl" Oct 14 10:37:25 crc kubenswrapper[4870]: I1014 10:37:25.486640 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c3c21d56-4c18-4f40-b061-37a18f4b3bd0-utilities\") pod \"community-operators-ntwkl\" (UID: \"c3c21d56-4c18-4f40-b061-37a18f4b3bd0\") " pod="openshift-marketplace/community-operators-ntwkl" Oct 14 10:37:25 crc kubenswrapper[4870]: I1014 10:37:25.514348 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fdb78\" (UniqueName: \"kubernetes.io/projected/c3c21d56-4c18-4f40-b061-37a18f4b3bd0-kube-api-access-fdb78\") pod \"community-operators-ntwkl\" (UID: \"c3c21d56-4c18-4f40-b061-37a18f4b3bd0\") " pod="openshift-marketplace/community-operators-ntwkl" Oct 14 10:37:25 crc kubenswrapper[4870]: I1014 10:37:25.651357 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ntwkl" Oct 14 10:37:26 crc kubenswrapper[4870]: I1014 10:37:26.297596 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-ntwkl"] Oct 14 10:37:26 crc kubenswrapper[4870]: I1014 10:37:26.919723 4870 generic.go:334] "Generic (PLEG): container finished" podID="c3c21d56-4c18-4f40-b061-37a18f4b3bd0" containerID="84e4e9a4ab77ba7d13dd9822e5a667b71e6c3707f291f5aa572aaf04e77701a6" exitCode=0 Oct 14 10:37:26 crc kubenswrapper[4870]: I1014 10:37:26.919786 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ntwkl" event={"ID":"c3c21d56-4c18-4f40-b061-37a18f4b3bd0","Type":"ContainerDied","Data":"84e4e9a4ab77ba7d13dd9822e5a667b71e6c3707f291f5aa572aaf04e77701a6"} Oct 14 10:37:26 crc kubenswrapper[4870]: I1014 10:37:26.919849 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ntwkl" event={"ID":"c3c21d56-4c18-4f40-b061-37a18f4b3bd0","Type":"ContainerStarted","Data":"2897a405d5e9d55f2ac5202fc7c113bf22c37465f85ff131c3e42e2dea63581e"} Oct 14 10:37:31 crc kubenswrapper[4870]: I1014 10:37:31.992470 4870 generic.go:334] "Generic (PLEG): container finished" podID="c3c21d56-4c18-4f40-b061-37a18f4b3bd0" containerID="dc09390b2a912d4329f7c4751bd8e00b87cc56d0e3b20fa6f2a089e522d342db" exitCode=0 Oct 14 10:37:31 crc kubenswrapper[4870]: I1014 10:37:31.992687 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ntwkl" event={"ID":"c3c21d56-4c18-4f40-b061-37a18f4b3bd0","Type":"ContainerDied","Data":"dc09390b2a912d4329f7c4751bd8e00b87cc56d0e3b20fa6f2a089e522d342db"} Oct 14 10:37:33 crc kubenswrapper[4870]: I1014 10:37:33.012954 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ntwkl" event={"ID":"c3c21d56-4c18-4f40-b061-37a18f4b3bd0","Type":"ContainerStarted","Data":"8b5ea406a227bf932447571602c7071a6d85e15286af8a8aaf2e02f284503d72"} Oct 14 10:37:33 crc kubenswrapper[4870]: I1014 10:37:33.046243 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-ntwkl" podStartSLOduration=2.198751991 podStartE2EDuration="8.046219027s" podCreationTimestamp="2025-10-14 10:37:25 +0000 UTC" firstStartedPulling="2025-10-14 10:37:26.922232763 +0000 UTC m=+12982.619593134" lastFinishedPulling="2025-10-14 10:37:32.769699759 +0000 UTC m=+12988.467060170" observedRunningTime="2025-10-14 10:37:33.034288156 +0000 UTC m=+12988.731648547" watchObservedRunningTime="2025-10-14 10:37:33.046219027 +0000 UTC m=+12988.743579408" Oct 14 10:37:35 crc kubenswrapper[4870]: I1014 10:37:35.652527 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-ntwkl" Oct 14 10:37:35 crc kubenswrapper[4870]: I1014 10:37:35.654714 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-ntwkl" Oct 14 10:37:35 crc kubenswrapper[4870]: I1014 10:37:35.704542 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-ntwkl" Oct 14 10:37:45 crc kubenswrapper[4870]: I1014 10:37:45.725641 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-ntwkl" Oct 14 10:37:45 crc kubenswrapper[4870]: I1014 10:37:45.807338 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-ntwkl"] Oct 14 10:37:45 crc kubenswrapper[4870]: I1014 10:37:45.866548 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-9qgf9"] Oct 14 10:37:45 crc kubenswrapper[4870]: I1014 10:37:45.866959 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-9qgf9" podUID="2baef9db-d79c-4f55-9c3f-cd4ac8488a8d" containerName="registry-server" containerID="cri-o://877dd9831079fffa4178ebbc2445a85a114bf775ab1effc8694fc1d707ff00a2" gracePeriod=2 Oct 14 10:37:46 crc kubenswrapper[4870]: E1014 10:37:46.081669 4870 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 877dd9831079fffa4178ebbc2445a85a114bf775ab1effc8694fc1d707ff00a2 is running failed: container process not found" containerID="877dd9831079fffa4178ebbc2445a85a114bf775ab1effc8694fc1d707ff00a2" cmd=["grpc_health_probe","-addr=:50051"] Oct 14 10:37:46 crc kubenswrapper[4870]: E1014 10:37:46.082186 4870 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 877dd9831079fffa4178ebbc2445a85a114bf775ab1effc8694fc1d707ff00a2 is running failed: container process not found" containerID="877dd9831079fffa4178ebbc2445a85a114bf775ab1effc8694fc1d707ff00a2" cmd=["grpc_health_probe","-addr=:50051"] Oct 14 10:37:46 crc kubenswrapper[4870]: E1014 10:37:46.082600 4870 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 877dd9831079fffa4178ebbc2445a85a114bf775ab1effc8694fc1d707ff00a2 is running failed: container process not found" containerID="877dd9831079fffa4178ebbc2445a85a114bf775ab1effc8694fc1d707ff00a2" cmd=["grpc_health_probe","-addr=:50051"] Oct 14 10:37:46 crc kubenswrapper[4870]: E1014 10:37:46.082632 4870 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 877dd9831079fffa4178ebbc2445a85a114bf775ab1effc8694fc1d707ff00a2 is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/community-operators-9qgf9" podUID="2baef9db-d79c-4f55-9c3f-cd4ac8488a8d" containerName="registry-server" Oct 14 10:37:46 crc kubenswrapper[4870]: I1014 10:37:46.256323 4870 generic.go:334] "Generic (PLEG): container finished" podID="2baef9db-d79c-4f55-9c3f-cd4ac8488a8d" containerID="877dd9831079fffa4178ebbc2445a85a114bf775ab1effc8694fc1d707ff00a2" exitCode=0 Oct 14 10:37:46 crc kubenswrapper[4870]: I1014 10:37:46.256375 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9qgf9" event={"ID":"2baef9db-d79c-4f55-9c3f-cd4ac8488a8d","Type":"ContainerDied","Data":"877dd9831079fffa4178ebbc2445a85a114bf775ab1effc8694fc1d707ff00a2"} Oct 14 10:37:46 crc kubenswrapper[4870]: I1014 10:37:46.447536 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9qgf9" Oct 14 10:37:46 crc kubenswrapper[4870]: I1014 10:37:46.599563 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2baef9db-d79c-4f55-9c3f-cd4ac8488a8d-utilities\") pod \"2baef9db-d79c-4f55-9c3f-cd4ac8488a8d\" (UID: \"2baef9db-d79c-4f55-9c3f-cd4ac8488a8d\") " Oct 14 10:37:46 crc kubenswrapper[4870]: I1014 10:37:46.599621 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dqzcn\" (UniqueName: \"kubernetes.io/projected/2baef9db-d79c-4f55-9c3f-cd4ac8488a8d-kube-api-access-dqzcn\") pod \"2baef9db-d79c-4f55-9c3f-cd4ac8488a8d\" (UID: \"2baef9db-d79c-4f55-9c3f-cd4ac8488a8d\") " Oct 14 10:37:46 crc kubenswrapper[4870]: I1014 10:37:46.599808 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2baef9db-d79c-4f55-9c3f-cd4ac8488a8d-catalog-content\") pod \"2baef9db-d79c-4f55-9c3f-cd4ac8488a8d\" (UID: \"2baef9db-d79c-4f55-9c3f-cd4ac8488a8d\") " Oct 14 10:37:46 crc kubenswrapper[4870]: I1014 10:37:46.602010 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2baef9db-d79c-4f55-9c3f-cd4ac8488a8d-utilities" (OuterVolumeSpecName: "utilities") pod "2baef9db-d79c-4f55-9c3f-cd4ac8488a8d" (UID: "2baef9db-d79c-4f55-9c3f-cd4ac8488a8d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 10:37:46 crc kubenswrapper[4870]: I1014 10:37:46.607246 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2baef9db-d79c-4f55-9c3f-cd4ac8488a8d-kube-api-access-dqzcn" (OuterVolumeSpecName: "kube-api-access-dqzcn") pod "2baef9db-d79c-4f55-9c3f-cd4ac8488a8d" (UID: "2baef9db-d79c-4f55-9c3f-cd4ac8488a8d"). InnerVolumeSpecName "kube-api-access-dqzcn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 10:37:46 crc kubenswrapper[4870]: I1014 10:37:46.676893 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2baef9db-d79c-4f55-9c3f-cd4ac8488a8d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2baef9db-d79c-4f55-9c3f-cd4ac8488a8d" (UID: "2baef9db-d79c-4f55-9c3f-cd4ac8488a8d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 10:37:46 crc kubenswrapper[4870]: I1014 10:37:46.702314 4870 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2baef9db-d79c-4f55-9c3f-cd4ac8488a8d-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 10:37:46 crc kubenswrapper[4870]: I1014 10:37:46.702355 4870 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2baef9db-d79c-4f55-9c3f-cd4ac8488a8d-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 10:37:46 crc kubenswrapper[4870]: I1014 10:37:46.702369 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dqzcn\" (UniqueName: \"kubernetes.io/projected/2baef9db-d79c-4f55-9c3f-cd4ac8488a8d-kube-api-access-dqzcn\") on node \"crc\" DevicePath \"\"" Oct 14 10:37:47 crc kubenswrapper[4870]: I1014 10:37:47.268757 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9qgf9" event={"ID":"2baef9db-d79c-4f55-9c3f-cd4ac8488a8d","Type":"ContainerDied","Data":"91caa1e999fcc488856c96556dc72c49dde7dde6a08713c0c4bbc551486f7575"} Oct 14 10:37:47 crc kubenswrapper[4870]: I1014 10:37:47.268808 4870 scope.go:117] "RemoveContainer" containerID="877dd9831079fffa4178ebbc2445a85a114bf775ab1effc8694fc1d707ff00a2" Oct 14 10:37:47 crc kubenswrapper[4870]: I1014 10:37:47.268947 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9qgf9" Oct 14 10:37:47 crc kubenswrapper[4870]: I1014 10:37:47.303210 4870 scope.go:117] "RemoveContainer" containerID="767a55328722b63ea6d75648d06c370895e1fe0a5429ed7a61f10f13379f3d6c" Oct 14 10:37:47 crc kubenswrapper[4870]: I1014 10:37:47.304757 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-9qgf9"] Oct 14 10:37:47 crc kubenswrapper[4870]: I1014 10:37:47.316221 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-9qgf9"] Oct 14 10:37:47 crc kubenswrapper[4870]: I1014 10:37:47.332054 4870 scope.go:117] "RemoveContainer" containerID="663cb46f4f51bf6fb5ceb2a4b0027b914a3dba9aa5486f8dadfab93dd3a1e061" Oct 14 10:37:48 crc kubenswrapper[4870]: I1014 10:37:48.796137 4870 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-hqgxs"] Oct 14 10:37:48 crc kubenswrapper[4870]: E1014 10:37:48.796993 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2baef9db-d79c-4f55-9c3f-cd4ac8488a8d" containerName="registry-server" Oct 14 10:37:48 crc kubenswrapper[4870]: I1014 10:37:48.797008 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="2baef9db-d79c-4f55-9c3f-cd4ac8488a8d" containerName="registry-server" Oct 14 10:37:48 crc kubenswrapper[4870]: E1014 10:37:48.797039 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2baef9db-d79c-4f55-9c3f-cd4ac8488a8d" containerName="extract-content" Oct 14 10:37:48 crc kubenswrapper[4870]: I1014 10:37:48.797045 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="2baef9db-d79c-4f55-9c3f-cd4ac8488a8d" containerName="extract-content" Oct 14 10:37:48 crc kubenswrapper[4870]: E1014 10:37:48.797063 4870 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2baef9db-d79c-4f55-9c3f-cd4ac8488a8d" containerName="extract-utilities" Oct 14 10:37:48 crc kubenswrapper[4870]: I1014 10:37:48.797070 4870 state_mem.go:107] "Deleted CPUSet assignment" podUID="2baef9db-d79c-4f55-9c3f-cd4ac8488a8d" containerName="extract-utilities" Oct 14 10:37:48 crc kubenswrapper[4870]: I1014 10:37:48.797313 4870 memory_manager.go:354] "RemoveStaleState removing state" podUID="2baef9db-d79c-4f55-9c3f-cd4ac8488a8d" containerName="registry-server" Oct 14 10:37:48 crc kubenswrapper[4870]: I1014 10:37:48.798916 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hqgxs" Oct 14 10:37:48 crc kubenswrapper[4870]: I1014 10:37:48.817017 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-hqgxs"] Oct 14 10:37:48 crc kubenswrapper[4870]: I1014 10:37:48.955530 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/71f9b1b2-2d5b-44b9-8af2-62645e9e6292-utilities\") pod \"redhat-operators-hqgxs\" (UID: \"71f9b1b2-2d5b-44b9-8af2-62645e9e6292\") " pod="openshift-marketplace/redhat-operators-hqgxs" Oct 14 10:37:48 crc kubenswrapper[4870]: I1014 10:37:48.955851 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/71f9b1b2-2d5b-44b9-8af2-62645e9e6292-catalog-content\") pod \"redhat-operators-hqgxs\" (UID: \"71f9b1b2-2d5b-44b9-8af2-62645e9e6292\") " pod="openshift-marketplace/redhat-operators-hqgxs" Oct 14 10:37:48 crc kubenswrapper[4870]: I1014 10:37:48.956090 4870 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fn95f\" (UniqueName: \"kubernetes.io/projected/71f9b1b2-2d5b-44b9-8af2-62645e9e6292-kube-api-access-fn95f\") pod \"redhat-operators-hqgxs\" (UID: \"71f9b1b2-2d5b-44b9-8af2-62645e9e6292\") " pod="openshift-marketplace/redhat-operators-hqgxs" Oct 14 10:37:49 crc kubenswrapper[4870]: I1014 10:37:49.045908 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2baef9db-d79c-4f55-9c3f-cd4ac8488a8d" path="/var/lib/kubelet/pods/2baef9db-d79c-4f55-9c3f-cd4ac8488a8d/volumes" Oct 14 10:37:49 crc kubenswrapper[4870]: I1014 10:37:49.059320 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/71f9b1b2-2d5b-44b9-8af2-62645e9e6292-utilities\") pod \"redhat-operators-hqgxs\" (UID: \"71f9b1b2-2d5b-44b9-8af2-62645e9e6292\") " pod="openshift-marketplace/redhat-operators-hqgxs" Oct 14 10:37:49 crc kubenswrapper[4870]: I1014 10:37:49.059525 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/71f9b1b2-2d5b-44b9-8af2-62645e9e6292-catalog-content\") pod \"redhat-operators-hqgxs\" (UID: \"71f9b1b2-2d5b-44b9-8af2-62645e9e6292\") " pod="openshift-marketplace/redhat-operators-hqgxs" Oct 14 10:37:49 crc kubenswrapper[4870]: I1014 10:37:49.059640 4870 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fn95f\" (UniqueName: \"kubernetes.io/projected/71f9b1b2-2d5b-44b9-8af2-62645e9e6292-kube-api-access-fn95f\") pod \"redhat-operators-hqgxs\" (UID: \"71f9b1b2-2d5b-44b9-8af2-62645e9e6292\") " pod="openshift-marketplace/redhat-operators-hqgxs" Oct 14 10:37:49 crc kubenswrapper[4870]: I1014 10:37:49.060469 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/71f9b1b2-2d5b-44b9-8af2-62645e9e6292-utilities\") pod \"redhat-operators-hqgxs\" (UID: \"71f9b1b2-2d5b-44b9-8af2-62645e9e6292\") " pod="openshift-marketplace/redhat-operators-hqgxs" Oct 14 10:37:49 crc kubenswrapper[4870]: I1014 10:37:49.060620 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/71f9b1b2-2d5b-44b9-8af2-62645e9e6292-catalog-content\") pod \"redhat-operators-hqgxs\" (UID: \"71f9b1b2-2d5b-44b9-8af2-62645e9e6292\") " pod="openshift-marketplace/redhat-operators-hqgxs" Oct 14 10:37:49 crc kubenswrapper[4870]: I1014 10:37:49.091137 4870 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fn95f\" (UniqueName: \"kubernetes.io/projected/71f9b1b2-2d5b-44b9-8af2-62645e9e6292-kube-api-access-fn95f\") pod \"redhat-operators-hqgxs\" (UID: \"71f9b1b2-2d5b-44b9-8af2-62645e9e6292\") " pod="openshift-marketplace/redhat-operators-hqgxs" Oct 14 10:37:49 crc kubenswrapper[4870]: I1014 10:37:49.132357 4870 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hqgxs" Oct 14 10:37:49 crc kubenswrapper[4870]: I1014 10:37:49.696415 4870 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-hqgxs"] Oct 14 10:37:50 crc kubenswrapper[4870]: I1014 10:37:50.317844 4870 generic.go:334] "Generic (PLEG): container finished" podID="71f9b1b2-2d5b-44b9-8af2-62645e9e6292" containerID="cdca1a50ead6c63601b6fb57dfe43600b465d5d03f21556b7dfa358f9d4e6aee" exitCode=0 Oct 14 10:37:50 crc kubenswrapper[4870]: I1014 10:37:50.317897 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hqgxs" event={"ID":"71f9b1b2-2d5b-44b9-8af2-62645e9e6292","Type":"ContainerDied","Data":"cdca1a50ead6c63601b6fb57dfe43600b465d5d03f21556b7dfa358f9d4e6aee"} Oct 14 10:37:50 crc kubenswrapper[4870]: I1014 10:37:50.318246 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hqgxs" event={"ID":"71f9b1b2-2d5b-44b9-8af2-62645e9e6292","Type":"ContainerStarted","Data":"b14aa776edddb8de03bf9c99358da219a7f089c38463487bba2a0d08b3523a15"} Oct 14 10:37:51 crc kubenswrapper[4870]: I1014 10:37:51.332660 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hqgxs" event={"ID":"71f9b1b2-2d5b-44b9-8af2-62645e9e6292","Type":"ContainerStarted","Data":"d1ae46eefa065572f74897dcf2bd77c3da8a89d33957ec360bf11413925503ed"} Oct 14 10:37:53 crc kubenswrapper[4870]: I1014 10:37:53.951199 4870 patch_prober.go:28] interesting pod/machine-config-daemon-7tvc7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 10:37:53 crc kubenswrapper[4870]: I1014 10:37:53.951842 4870 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 10:37:53 crc kubenswrapper[4870]: I1014 10:37:53.951909 4870 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" Oct 14 10:37:53 crc kubenswrapper[4870]: I1014 10:37:53.953094 4870 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"3ecb7d8f7bc7b21ce0eb0f68bcda8198b66e9e2e6c18afa3bb02bf2b3ab41d20"} pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 14 10:37:53 crc kubenswrapper[4870]: I1014 10:37:53.953194 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerName="machine-config-daemon" containerID="cri-o://3ecb7d8f7bc7b21ce0eb0f68bcda8198b66e9e2e6c18afa3bb02bf2b3ab41d20" gracePeriod=600 Oct 14 10:37:54 crc kubenswrapper[4870]: I1014 10:37:54.374626 4870 generic.go:334] "Generic (PLEG): container finished" podID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerID="3ecb7d8f7bc7b21ce0eb0f68bcda8198b66e9e2e6c18afa3bb02bf2b3ab41d20" exitCode=0 Oct 14 10:37:54 crc kubenswrapper[4870]: I1014 10:37:54.374707 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" event={"ID":"874a5e45-dffd-4d17-b609-b7d3ed2eab07","Type":"ContainerDied","Data":"3ecb7d8f7bc7b21ce0eb0f68bcda8198b66e9e2e6c18afa3bb02bf2b3ab41d20"} Oct 14 10:37:54 crc kubenswrapper[4870]: I1014 10:37:54.374965 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" event={"ID":"874a5e45-dffd-4d17-b609-b7d3ed2eab07","Type":"ContainerStarted","Data":"2f0b254d2a9a41bb03fd01ffa6dbabef116dab4b83d6c59f6fc2212d01cf0721"} Oct 14 10:37:54 crc kubenswrapper[4870]: I1014 10:37:54.374991 4870 scope.go:117] "RemoveContainer" containerID="3a25412c639a58b6345a6cb926f93c78a7a62c68cd923f500f921497a90b6efb" Oct 14 10:37:55 crc kubenswrapper[4870]: I1014 10:37:55.399995 4870 generic.go:334] "Generic (PLEG): container finished" podID="71f9b1b2-2d5b-44b9-8af2-62645e9e6292" containerID="d1ae46eefa065572f74897dcf2bd77c3da8a89d33957ec360bf11413925503ed" exitCode=0 Oct 14 10:37:55 crc kubenswrapper[4870]: I1014 10:37:55.400071 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hqgxs" event={"ID":"71f9b1b2-2d5b-44b9-8af2-62645e9e6292","Type":"ContainerDied","Data":"d1ae46eefa065572f74897dcf2bd77c3da8a89d33957ec360bf11413925503ed"} Oct 14 10:37:56 crc kubenswrapper[4870]: I1014 10:37:56.412662 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hqgxs" event={"ID":"71f9b1b2-2d5b-44b9-8af2-62645e9e6292","Type":"ContainerStarted","Data":"2bb08a56e60d9864ae0e7b87671a4445959b13516a733abadb77adf67c2349af"} Oct 14 10:37:56 crc kubenswrapper[4870]: I1014 10:37:56.434907 4870 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-hqgxs" podStartSLOduration=2.839477426 podStartE2EDuration="8.434886971s" podCreationTimestamp="2025-10-14 10:37:48 +0000 UTC" firstStartedPulling="2025-10-14 10:37:50.320634455 +0000 UTC m=+13006.017994846" lastFinishedPulling="2025-10-14 10:37:55.91604402 +0000 UTC m=+13011.613404391" observedRunningTime="2025-10-14 10:37:56.43322049 +0000 UTC m=+13012.130580861" watchObservedRunningTime="2025-10-14 10:37:56.434886971 +0000 UTC m=+13012.132247342" Oct 14 10:37:59 crc kubenswrapper[4870]: I1014 10:37:59.132845 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-hqgxs" Oct 14 10:37:59 crc kubenswrapper[4870]: I1014 10:37:59.133425 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-hqgxs" Oct 14 10:38:00 crc kubenswrapper[4870]: I1014 10:38:00.191212 4870 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-hqgxs" podUID="71f9b1b2-2d5b-44b9-8af2-62645e9e6292" containerName="registry-server" probeResult="failure" output=< Oct 14 10:38:00 crc kubenswrapper[4870]: timeout: failed to connect service ":50051" within 1s Oct 14 10:38:00 crc kubenswrapper[4870]: > Oct 14 10:38:09 crc kubenswrapper[4870]: I1014 10:38:09.214662 4870 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-hqgxs" Oct 14 10:38:09 crc kubenswrapper[4870]: I1014 10:38:09.275006 4870 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-hqgxs" Oct 14 10:38:09 crc kubenswrapper[4870]: I1014 10:38:09.469988 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-hqgxs"] Oct 14 10:38:10 crc kubenswrapper[4870]: I1014 10:38:10.592231 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-hqgxs" podUID="71f9b1b2-2d5b-44b9-8af2-62645e9e6292" containerName="registry-server" containerID="cri-o://2bb08a56e60d9864ae0e7b87671a4445959b13516a733abadb77adf67c2349af" gracePeriod=2 Oct 14 10:38:11 crc kubenswrapper[4870]: I1014 10:38:11.279869 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hqgxs" Oct 14 10:38:11 crc kubenswrapper[4870]: I1014 10:38:11.449479 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/71f9b1b2-2d5b-44b9-8af2-62645e9e6292-catalog-content\") pod \"71f9b1b2-2d5b-44b9-8af2-62645e9e6292\" (UID: \"71f9b1b2-2d5b-44b9-8af2-62645e9e6292\") " Oct 14 10:38:11 crc kubenswrapper[4870]: I1014 10:38:11.449522 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/71f9b1b2-2d5b-44b9-8af2-62645e9e6292-utilities\") pod \"71f9b1b2-2d5b-44b9-8af2-62645e9e6292\" (UID: \"71f9b1b2-2d5b-44b9-8af2-62645e9e6292\") " Oct 14 10:38:11 crc kubenswrapper[4870]: I1014 10:38:11.449606 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fn95f\" (UniqueName: \"kubernetes.io/projected/71f9b1b2-2d5b-44b9-8af2-62645e9e6292-kube-api-access-fn95f\") pod \"71f9b1b2-2d5b-44b9-8af2-62645e9e6292\" (UID: \"71f9b1b2-2d5b-44b9-8af2-62645e9e6292\") " Oct 14 10:38:11 crc kubenswrapper[4870]: I1014 10:38:11.450308 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/71f9b1b2-2d5b-44b9-8af2-62645e9e6292-utilities" (OuterVolumeSpecName: "utilities") pod "71f9b1b2-2d5b-44b9-8af2-62645e9e6292" (UID: "71f9b1b2-2d5b-44b9-8af2-62645e9e6292"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 10:38:11 crc kubenswrapper[4870]: I1014 10:38:11.476820 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/71f9b1b2-2d5b-44b9-8af2-62645e9e6292-kube-api-access-fn95f" (OuterVolumeSpecName: "kube-api-access-fn95f") pod "71f9b1b2-2d5b-44b9-8af2-62645e9e6292" (UID: "71f9b1b2-2d5b-44b9-8af2-62645e9e6292"). InnerVolumeSpecName "kube-api-access-fn95f". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 10:38:11 crc kubenswrapper[4870]: I1014 10:38:11.549983 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/71f9b1b2-2d5b-44b9-8af2-62645e9e6292-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "71f9b1b2-2d5b-44b9-8af2-62645e9e6292" (UID: "71f9b1b2-2d5b-44b9-8af2-62645e9e6292"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 10:38:11 crc kubenswrapper[4870]: I1014 10:38:11.550852 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/71f9b1b2-2d5b-44b9-8af2-62645e9e6292-catalog-content\") pod \"71f9b1b2-2d5b-44b9-8af2-62645e9e6292\" (UID: \"71f9b1b2-2d5b-44b9-8af2-62645e9e6292\") " Oct 14 10:38:11 crc kubenswrapper[4870]: I1014 10:38:11.551506 4870 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/71f9b1b2-2d5b-44b9-8af2-62645e9e6292-utilities\") on node \"crc\" DevicePath \"\"" Oct 14 10:38:11 crc kubenswrapper[4870]: I1014 10:38:11.551524 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fn95f\" (UniqueName: \"kubernetes.io/projected/71f9b1b2-2d5b-44b9-8af2-62645e9e6292-kube-api-access-fn95f\") on node \"crc\" DevicePath \"\"" Oct 14 10:38:11 crc kubenswrapper[4870]: W1014 10:38:11.554822 4870 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/71f9b1b2-2d5b-44b9-8af2-62645e9e6292/volumes/kubernetes.io~empty-dir/catalog-content Oct 14 10:38:11 crc kubenswrapper[4870]: I1014 10:38:11.554851 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/71f9b1b2-2d5b-44b9-8af2-62645e9e6292-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "71f9b1b2-2d5b-44b9-8af2-62645e9e6292" (UID: "71f9b1b2-2d5b-44b9-8af2-62645e9e6292"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 10:38:11 crc kubenswrapper[4870]: I1014 10:38:11.604387 4870 generic.go:334] "Generic (PLEG): container finished" podID="71f9b1b2-2d5b-44b9-8af2-62645e9e6292" containerID="2bb08a56e60d9864ae0e7b87671a4445959b13516a733abadb77adf67c2349af" exitCode=0 Oct 14 10:38:11 crc kubenswrapper[4870]: I1014 10:38:11.604472 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hqgxs" event={"ID":"71f9b1b2-2d5b-44b9-8af2-62645e9e6292","Type":"ContainerDied","Data":"2bb08a56e60d9864ae0e7b87671a4445959b13516a733abadb77adf67c2349af"} Oct 14 10:38:11 crc kubenswrapper[4870]: I1014 10:38:11.604490 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hqgxs" Oct 14 10:38:11 crc kubenswrapper[4870]: I1014 10:38:11.604524 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hqgxs" event={"ID":"71f9b1b2-2d5b-44b9-8af2-62645e9e6292","Type":"ContainerDied","Data":"b14aa776edddb8de03bf9c99358da219a7f089c38463487bba2a0d08b3523a15"} Oct 14 10:38:11 crc kubenswrapper[4870]: I1014 10:38:11.604558 4870 scope.go:117] "RemoveContainer" containerID="2bb08a56e60d9864ae0e7b87671a4445959b13516a733abadb77adf67c2349af" Oct 14 10:38:11 crc kubenswrapper[4870]: I1014 10:38:11.634684 4870 scope.go:117] "RemoveContainer" containerID="d1ae46eefa065572f74897dcf2bd77c3da8a89d33957ec360bf11413925503ed" Oct 14 10:38:11 crc kubenswrapper[4870]: I1014 10:38:11.654023 4870 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/71f9b1b2-2d5b-44b9-8af2-62645e9e6292-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 14 10:38:11 crc kubenswrapper[4870]: I1014 10:38:11.655044 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-hqgxs"] Oct 14 10:38:11 crc kubenswrapper[4870]: I1014 10:38:11.674215 4870 scope.go:117] "RemoveContainer" containerID="cdca1a50ead6c63601b6fb57dfe43600b465d5d03f21556b7dfa358f9d4e6aee" Oct 14 10:38:11 crc kubenswrapper[4870]: I1014 10:38:11.679005 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-hqgxs"] Oct 14 10:38:11 crc kubenswrapper[4870]: I1014 10:38:11.712029 4870 scope.go:117] "RemoveContainer" containerID="2bb08a56e60d9864ae0e7b87671a4445959b13516a733abadb77adf67c2349af" Oct 14 10:38:11 crc kubenswrapper[4870]: E1014 10:38:11.712460 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2bb08a56e60d9864ae0e7b87671a4445959b13516a733abadb77adf67c2349af\": container with ID starting with 2bb08a56e60d9864ae0e7b87671a4445959b13516a733abadb77adf67c2349af not found: ID does not exist" containerID="2bb08a56e60d9864ae0e7b87671a4445959b13516a733abadb77adf67c2349af" Oct 14 10:38:11 crc kubenswrapper[4870]: I1014 10:38:11.712500 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2bb08a56e60d9864ae0e7b87671a4445959b13516a733abadb77adf67c2349af"} err="failed to get container status \"2bb08a56e60d9864ae0e7b87671a4445959b13516a733abadb77adf67c2349af\": rpc error: code = NotFound desc = could not find container \"2bb08a56e60d9864ae0e7b87671a4445959b13516a733abadb77adf67c2349af\": container with ID starting with 2bb08a56e60d9864ae0e7b87671a4445959b13516a733abadb77adf67c2349af not found: ID does not exist" Oct 14 10:38:11 crc kubenswrapper[4870]: I1014 10:38:11.712532 4870 scope.go:117] "RemoveContainer" containerID="d1ae46eefa065572f74897dcf2bd77c3da8a89d33957ec360bf11413925503ed" Oct 14 10:38:11 crc kubenswrapper[4870]: E1014 10:38:11.712976 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d1ae46eefa065572f74897dcf2bd77c3da8a89d33957ec360bf11413925503ed\": container with ID starting with d1ae46eefa065572f74897dcf2bd77c3da8a89d33957ec360bf11413925503ed not found: ID does not exist" containerID="d1ae46eefa065572f74897dcf2bd77c3da8a89d33957ec360bf11413925503ed" Oct 14 10:38:11 crc kubenswrapper[4870]: I1014 10:38:11.713050 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d1ae46eefa065572f74897dcf2bd77c3da8a89d33957ec360bf11413925503ed"} err="failed to get container status \"d1ae46eefa065572f74897dcf2bd77c3da8a89d33957ec360bf11413925503ed\": rpc error: code = NotFound desc = could not find container \"d1ae46eefa065572f74897dcf2bd77c3da8a89d33957ec360bf11413925503ed\": container with ID starting with d1ae46eefa065572f74897dcf2bd77c3da8a89d33957ec360bf11413925503ed not found: ID does not exist" Oct 14 10:38:11 crc kubenswrapper[4870]: I1014 10:38:11.713090 4870 scope.go:117] "RemoveContainer" containerID="cdca1a50ead6c63601b6fb57dfe43600b465d5d03f21556b7dfa358f9d4e6aee" Oct 14 10:38:11 crc kubenswrapper[4870]: E1014 10:38:11.713404 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cdca1a50ead6c63601b6fb57dfe43600b465d5d03f21556b7dfa358f9d4e6aee\": container with ID starting with cdca1a50ead6c63601b6fb57dfe43600b465d5d03f21556b7dfa358f9d4e6aee not found: ID does not exist" containerID="cdca1a50ead6c63601b6fb57dfe43600b465d5d03f21556b7dfa358f9d4e6aee" Oct 14 10:38:11 crc kubenswrapper[4870]: I1014 10:38:11.713463 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cdca1a50ead6c63601b6fb57dfe43600b465d5d03f21556b7dfa358f9d4e6aee"} err="failed to get container status \"cdca1a50ead6c63601b6fb57dfe43600b465d5d03f21556b7dfa358f9d4e6aee\": rpc error: code = NotFound desc = could not find container \"cdca1a50ead6c63601b6fb57dfe43600b465d5d03f21556b7dfa358f9d4e6aee\": container with ID starting with cdca1a50ead6c63601b6fb57dfe43600b465d5d03f21556b7dfa358f9d4e6aee not found: ID does not exist" Oct 14 10:38:13 crc kubenswrapper[4870]: I1014 10:38:13.050766 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="71f9b1b2-2d5b-44b9-8af2-62645e9e6292" path="/var/lib/kubelet/pods/71f9b1b2-2d5b-44b9-8af2-62645e9e6292/volumes" Oct 14 10:38:40 crc kubenswrapper[4870]: I1014 10:38:40.961273 4870 generic.go:334] "Generic (PLEG): container finished" podID="e78b6d95-1398-4f8a-9565-d64feb722dd0" containerID="120f8b1146a59099dee946943f0c73cf49cdb3e20248bb24290f386d7aecad21" exitCode=0 Oct 14 10:38:40 crc kubenswrapper[4870]: I1014 10:38:40.962864 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-zlss5/must-gather-p4ml9" event={"ID":"e78b6d95-1398-4f8a-9565-d64feb722dd0","Type":"ContainerDied","Data":"120f8b1146a59099dee946943f0c73cf49cdb3e20248bb24290f386d7aecad21"} Oct 14 10:38:40 crc kubenswrapper[4870]: I1014 10:38:40.963695 4870 scope.go:117] "RemoveContainer" containerID="120f8b1146a59099dee946943f0c73cf49cdb3e20248bb24290f386d7aecad21" Oct 14 10:38:41 crc kubenswrapper[4870]: I1014 10:38:41.067772 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-zlss5_must-gather-p4ml9_e78b6d95-1398-4f8a-9565-d64feb722dd0/gather/0.log" Oct 14 10:38:53 crc kubenswrapper[4870]: I1014 10:38:53.466598 4870 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-zlss5/must-gather-p4ml9"] Oct 14 10:38:53 crc kubenswrapper[4870]: I1014 10:38:53.467931 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-zlss5/must-gather-p4ml9" podUID="e78b6d95-1398-4f8a-9565-d64feb722dd0" containerName="copy" containerID="cri-o://2ac83eeacc2ecd24b9221a6f13d502525efc9fb0101a534cfb28db48dd794f3d" gracePeriod=2 Oct 14 10:38:53 crc kubenswrapper[4870]: I1014 10:38:53.489260 4870 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-zlss5/must-gather-p4ml9"] Oct 14 10:38:54 crc kubenswrapper[4870]: I1014 10:38:54.005057 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-zlss5_must-gather-p4ml9_e78b6d95-1398-4f8a-9565-d64feb722dd0/copy/0.log" Oct 14 10:38:54 crc kubenswrapper[4870]: I1014 10:38:54.005876 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-zlss5/must-gather-p4ml9" Oct 14 10:38:54 crc kubenswrapper[4870]: I1014 10:38:54.126931 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2tn5k\" (UniqueName: \"kubernetes.io/projected/e78b6d95-1398-4f8a-9565-d64feb722dd0-kube-api-access-2tn5k\") pod \"e78b6d95-1398-4f8a-9565-d64feb722dd0\" (UID: \"e78b6d95-1398-4f8a-9565-d64feb722dd0\") " Oct 14 10:38:54 crc kubenswrapper[4870]: I1014 10:38:54.127002 4870 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/e78b6d95-1398-4f8a-9565-d64feb722dd0-must-gather-output\") pod \"e78b6d95-1398-4f8a-9565-d64feb722dd0\" (UID: \"e78b6d95-1398-4f8a-9565-d64feb722dd0\") " Oct 14 10:38:54 crc kubenswrapper[4870]: I1014 10:38:54.143966 4870 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-zlss5_must-gather-p4ml9_e78b6d95-1398-4f8a-9565-d64feb722dd0/copy/0.log" Oct 14 10:38:54 crc kubenswrapper[4870]: I1014 10:38:54.144859 4870 generic.go:334] "Generic (PLEG): container finished" podID="e78b6d95-1398-4f8a-9565-d64feb722dd0" containerID="2ac83eeacc2ecd24b9221a6f13d502525efc9fb0101a534cfb28db48dd794f3d" exitCode=143 Oct 14 10:38:54 crc kubenswrapper[4870]: I1014 10:38:54.144994 4870 scope.go:117] "RemoveContainer" containerID="2ac83eeacc2ecd24b9221a6f13d502525efc9fb0101a534cfb28db48dd794f3d" Oct 14 10:38:54 crc kubenswrapper[4870]: I1014 10:38:54.145231 4870 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-zlss5/must-gather-p4ml9" Oct 14 10:38:54 crc kubenswrapper[4870]: I1014 10:38:54.156712 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e78b6d95-1398-4f8a-9565-d64feb722dd0-kube-api-access-2tn5k" (OuterVolumeSpecName: "kube-api-access-2tn5k") pod "e78b6d95-1398-4f8a-9565-d64feb722dd0" (UID: "e78b6d95-1398-4f8a-9565-d64feb722dd0"). InnerVolumeSpecName "kube-api-access-2tn5k". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 14 10:38:54 crc kubenswrapper[4870]: I1014 10:38:54.209464 4870 scope.go:117] "RemoveContainer" containerID="120f8b1146a59099dee946943f0c73cf49cdb3e20248bb24290f386d7aecad21" Oct 14 10:38:54 crc kubenswrapper[4870]: I1014 10:38:54.234557 4870 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2tn5k\" (UniqueName: \"kubernetes.io/projected/e78b6d95-1398-4f8a-9565-d64feb722dd0-kube-api-access-2tn5k\") on node \"crc\" DevicePath \"\"" Oct 14 10:38:54 crc kubenswrapper[4870]: I1014 10:38:54.272918 4870 scope.go:117] "RemoveContainer" containerID="2ac83eeacc2ecd24b9221a6f13d502525efc9fb0101a534cfb28db48dd794f3d" Oct 14 10:38:54 crc kubenswrapper[4870]: E1014 10:38:54.274876 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2ac83eeacc2ecd24b9221a6f13d502525efc9fb0101a534cfb28db48dd794f3d\": container with ID starting with 2ac83eeacc2ecd24b9221a6f13d502525efc9fb0101a534cfb28db48dd794f3d not found: ID does not exist" containerID="2ac83eeacc2ecd24b9221a6f13d502525efc9fb0101a534cfb28db48dd794f3d" Oct 14 10:38:54 crc kubenswrapper[4870]: I1014 10:38:54.274931 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2ac83eeacc2ecd24b9221a6f13d502525efc9fb0101a534cfb28db48dd794f3d"} err="failed to get container status \"2ac83eeacc2ecd24b9221a6f13d502525efc9fb0101a534cfb28db48dd794f3d\": rpc error: code = NotFound desc = could not find container \"2ac83eeacc2ecd24b9221a6f13d502525efc9fb0101a534cfb28db48dd794f3d\": container with ID starting with 2ac83eeacc2ecd24b9221a6f13d502525efc9fb0101a534cfb28db48dd794f3d not found: ID does not exist" Oct 14 10:38:54 crc kubenswrapper[4870]: I1014 10:38:54.274965 4870 scope.go:117] "RemoveContainer" containerID="120f8b1146a59099dee946943f0c73cf49cdb3e20248bb24290f386d7aecad21" Oct 14 10:38:54 crc kubenswrapper[4870]: E1014 10:38:54.275401 4870 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"120f8b1146a59099dee946943f0c73cf49cdb3e20248bb24290f386d7aecad21\": container with ID starting with 120f8b1146a59099dee946943f0c73cf49cdb3e20248bb24290f386d7aecad21 not found: ID does not exist" containerID="120f8b1146a59099dee946943f0c73cf49cdb3e20248bb24290f386d7aecad21" Oct 14 10:38:54 crc kubenswrapper[4870]: I1014 10:38:54.275472 4870 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"120f8b1146a59099dee946943f0c73cf49cdb3e20248bb24290f386d7aecad21"} err="failed to get container status \"120f8b1146a59099dee946943f0c73cf49cdb3e20248bb24290f386d7aecad21\": rpc error: code = NotFound desc = could not find container \"120f8b1146a59099dee946943f0c73cf49cdb3e20248bb24290f386d7aecad21\": container with ID starting with 120f8b1146a59099dee946943f0c73cf49cdb3e20248bb24290f386d7aecad21 not found: ID does not exist" Oct 14 10:38:54 crc kubenswrapper[4870]: I1014 10:38:54.434826 4870 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e78b6d95-1398-4f8a-9565-d64feb722dd0-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "e78b6d95-1398-4f8a-9565-d64feb722dd0" (UID: "e78b6d95-1398-4f8a-9565-d64feb722dd0"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 14 10:38:54 crc kubenswrapper[4870]: I1014 10:38:54.438163 4870 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/e78b6d95-1398-4f8a-9565-d64feb722dd0-must-gather-output\") on node \"crc\" DevicePath \"\"" Oct 14 10:38:55 crc kubenswrapper[4870]: I1014 10:38:55.049715 4870 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e78b6d95-1398-4f8a-9565-d64feb722dd0" path="/var/lib/kubelet/pods/e78b6d95-1398-4f8a-9565-d64feb722dd0/volumes" Oct 14 10:40:23 crc kubenswrapper[4870]: I1014 10:40:23.950871 4870 patch_prober.go:28] interesting pod/machine-config-daemon-7tvc7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 10:40:23 crc kubenswrapper[4870]: I1014 10:40:23.951658 4870 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 10:40:53 crc kubenswrapper[4870]: I1014 10:40:53.951334 4870 patch_prober.go:28] interesting pod/machine-config-daemon-7tvc7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 10:40:53 crc kubenswrapper[4870]: I1014 10:40:53.952038 4870 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 10:41:23 crc kubenswrapper[4870]: I1014 10:41:23.951051 4870 patch_prober.go:28] interesting pod/machine-config-daemon-7tvc7 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 14 10:41:23 crc kubenswrapper[4870]: I1014 10:41:23.951868 4870 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 14 10:41:23 crc kubenswrapper[4870]: I1014 10:41:23.951970 4870 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" Oct 14 10:41:23 crc kubenswrapper[4870]: I1014 10:41:23.953512 4870 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"2f0b254d2a9a41bb03fd01ffa6dbabef116dab4b83d6c59f6fc2212d01cf0721"} pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 14 10:41:23 crc kubenswrapper[4870]: I1014 10:41:23.953622 4870 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerName="machine-config-daemon" containerID="cri-o://2f0b254d2a9a41bb03fd01ffa6dbabef116dab4b83d6c59f6fc2212d01cf0721" gracePeriod=600 Oct 14 10:41:24 crc kubenswrapper[4870]: E1014 10:41:24.090553 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 10:41:24 crc kubenswrapper[4870]: I1014 10:41:24.193309 4870 generic.go:334] "Generic (PLEG): container finished" podID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" containerID="2f0b254d2a9a41bb03fd01ffa6dbabef116dab4b83d6c59f6fc2212d01cf0721" exitCode=0 Oct 14 10:41:24 crc kubenswrapper[4870]: I1014 10:41:24.193366 4870 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" event={"ID":"874a5e45-dffd-4d17-b609-b7d3ed2eab07","Type":"ContainerDied","Data":"2f0b254d2a9a41bb03fd01ffa6dbabef116dab4b83d6c59f6fc2212d01cf0721"} Oct 14 10:41:24 crc kubenswrapper[4870]: I1014 10:41:24.193414 4870 scope.go:117] "RemoveContainer" containerID="3ecb7d8f7bc7b21ce0eb0f68bcda8198b66e9e2e6c18afa3bb02bf2b3ab41d20" Oct 14 10:41:24 crc kubenswrapper[4870]: I1014 10:41:24.194157 4870 scope.go:117] "RemoveContainer" containerID="2f0b254d2a9a41bb03fd01ffa6dbabef116dab4b83d6c59f6fc2212d01cf0721" Oct 14 10:41:24 crc kubenswrapper[4870]: E1014 10:41:24.194468 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" Oct 14 10:41:38 crc kubenswrapper[4870]: I1014 10:41:38.034633 4870 scope.go:117] "RemoveContainer" containerID="2f0b254d2a9a41bb03fd01ffa6dbabef116dab4b83d6c59f6fc2212d01cf0721" Oct 14 10:41:38 crc kubenswrapper[4870]: E1014 10:41:38.035614 4870 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7tvc7_openshift-machine-config-operator(874a5e45-dffd-4d17-b609-b7d3ed2eab07)\"" pod="openshift-machine-config-operator/machine-config-daemon-7tvc7" podUID="874a5e45-dffd-4d17-b609-b7d3ed2eab07" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515073424352024452 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015073424353017370 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015073371503016511 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015073371503015461 5ustar corecore